commit_message
stringlengths
3
2.32k
diff
stringlengths
186
49.5k
concern_count
int64
1
5
shas
stringlengths
44
220
types
stringlengths
6
45
exclude github.io from link checking to avoid rate limiting
["diff --git a/.github/workflows/ibis-docs-lint.yml b/.github/workflows/ibis-docs-lint.yml\nindex 90c5a27..db6457b 100644\n--- a/.github/workflows/ibis-docs-lint.yml\n+++ b/.github/workflows/ibis-docs-lint.yml\n@@ -101,6 +101,7 @@ jobs:\n --exclude-mail \\\n --exclude fonts.gstatic.com \\\n --exclude github.com \\\n+ --exclude github.io \\\n --no-progress \\\n --github-token ${{ steps.generate_token.outputs.token }}\n \n"]
1
["ce0539a32b927a3559feebf8f5307e3863e992a1"]
["ci"]
fix `get-deploy-tags.sh`
["diff --git a/.circleci/get-deploy-tags.sh b/.circleci/get-deploy-tags.sh\nindex f80c8cb..7ddfa62 100755\n--- a/.circleci/get-deploy-tags.sh\n+++ b/.circleci/get-deploy-tags.sh\n@@ -20,7 +20,7 @@\n set -euo pipefail\n \n DOCKER_IMAGE_TAG=${1}\n-DOCKER_IMAGE=\"quay.io/influxdb/fusion\"\n+DOCKER_IMAGE=\"quay.io/influxdb/iox\"\n APP_NAME=\"IOx\"\n \n DOCKER_IMAGE_DIGEST=\"$(docker image inspect \"${DOCKER_IMAGE}:${DOCKER_IMAGE_TAG}\" --format '{{ if eq (len .RepoDigests) 1 }}{{index .RepoDigests 0}}{{ end }}')\"\n"]
1
["6786fd5955b064021f5b6d6a630453351d683fae"]
["ci"]
fetch git history
["diff --git a/.github/workflows/ci.yml b/.github/workflows/ci.yml\nindex 6726e35..9114eeb 100644\n--- a/.github/workflows/ci.yml\n+++ b/.github/workflows/ci.yml\n@@ -172,6 +172,8 @@ jobs:\n \n steps:\n - uses: actions/checkout@v3\n+ with:\n+ fetch-depth: 0\n - run: corepack enable\n - uses: actions/setup-node@v3\n with:\ndiff --git a/scripts/bump-edge.ts b/scripts/bump-edge.ts\nindex aa33d76..1d1cbc9 100644\n--- a/scripts/bump-edge.ts\n+++ b/scripts/bump-edge.ts\n@@ -18,7 +18,7 @@ async function main () {\n \n const config = await loadChangelogConfig(process.cwd())\n \n- const latestTag = execaSync('git', ['describe', '--tags', '--abbrev=0', 'main']).stdout\n+ const latestTag = execaSync('git', ['describe', '--tags', '--abbrev=0']).stdout\n \n const commits = await getGitDiff(latestTag)\n const bumpType = determineSemverChange(parseCommits(commits, config), config)\n"]
1
["e11d55a4922978b89a2c50bf577124b09449e89c"]
["ci"]
fix unit tests
["diff --git a/src/components/__tests__/__snapshots__/BottomNavigation.test.js.snap b/src/components/__tests__/__snapshots__/BottomNavigation.test.js.snap\nindex 4d771d6..9f9683c 100644\n--- a/src/components/__tests__/__snapshots__/BottomNavigation.test.js.snap\n+++ b/src/components/__tests__/__snapshots__/BottomNavigation.test.js.snap\n@@ -9,9 +9,6 @@ exports[`renders custom icon and label in non-shifting bottom navigation 1`] = `\n Object {\n \"flex\": 1,\n },\n- Object {\n- \"backgroundColor\": \"#000000\",\n- },\n undefined,\n ]\n }\n@@ -132,6 +129,33 @@ exports[`renders custom icon and label in non-shifting bottom navigation 1`] = `\n ]\n }\n >\n+ <AnimatedComponent\n+ pointerEvents=\"none\"\n+ style={\n+ Array [\n+ Object {\n+ \"position\": \"absolute\",\n+ },\n+ Object {\n+ \"backgroundColor\": \"rgba(0, 0, 0, 0.12)\",\n+ \"borderRadius\": 48,\n+ \"height\": 96,\n+ \"left\": -54.666666666666664,\n+ \"opacity\": 0.002,\n+ \"top\": -20,\n+ \"transform\": Array [\n+ Object {\n+ \"translateX\": 0,\n+ },\n+ Object {\n+ \"scale\": 0.001,\n+ },\n+ ],\n+ \"width\": 96,\n+ },\n+ ]\n+ }\n+ />\n <TouchableWithoutFeedback\n key=\"key-0\"\n onPress={[Function]}\n@@ -599,9 +623,6 @@ exports[`renders custom icon and label in shifting bottom navigation 1`] = `\n Object {\n \"flex\": 1,\n },\n- Object {\n- \"backgroundColor\": \"#000000\",\n- },\n undefined,\n ]\n }\n@@ -783,12 +804,15 @@ exports[`renders custom icon and label in shifting bottom navigation 1`] = `\n \"backgroundColor\": undefined,\n \"borderRadius\": 0,\n \"height\": 0,\n- \"left\": 6,\n+ \"left\": -4,\n \"opacity\": 0,\n \"top\": 28,\n \"transform\": Array [\n Object {\n- \"scale\": 0.002,\n+ \"translateX\": 10,\n+ },\n+ Object {\n+ \"scale\": 0.008,\n },\n ],\n \"width\": 0,\n@@ -805,17 +829,20 @@ exports[`renders custom icon and label in shifting bottom navigation 1`] = `\n },\n Object {\n \"backgroundColor\": \"rgba(255, 255, 255, 0.12)\",\n- \"borderRadius\": 36,\n- \"height\": 72,\n- \"left\": -30,\n- \"opacity\": 0,\n- \"top\": -8,\n+ \"borderRadius\": 48,\n+ \"height\": 96,\n+ \"left\": -52,\n+ \"opacity\": 0.002,\n+ \"top\": -20,\n \"transform\": Array [\n Object {\n- \"scale\": 0.002,\n+ \"translateX\": 10,\n+ },\n+ Object {\n+ \"scale\": 0.001,\n },\n ],\n- \"width\": 72,\n+ \"width\": 96,\n },\n ]\n }\n@@ -1374,9 +1401,6 @@ exports[`renders non-shifting bottom navigation 1`] = `\n Object {\n \"flex\": 1,\n },\n- Object {\n- \"backgroundColor\": \"#000000\",\n- },\n undefined,\n ]\n }\n@@ -1497,6 +1521,33 @@ exports[`renders non-shifting bottom navigation 1`] = `\n ]\n }\n >\n+ <AnimatedComponent\n+ pointerEvents=\"none\"\n+ style={\n+ Array [\n+ Object {\n+ \"position\": \"absolute\",\n+ },\n+ Object {\n+ \"backgroundColor\": \"rgba(0, 0, 0, 0.12)\",\n+ \"borderRadius\": 48,\n+ \"height\": 96,\n+ \"left\": -54.666666666666664,\n+ \"opacity\": 0.002,\n+ \"top\": -20,\n+ \"transform\": Array [\n+ Object {\n+ \"translateX\": 0,\n+ },\n+ Object {\n+ \"scale\": 0.001,\n+ },\n+ ],\n+ \"width\": 96,\n+ },\n+ ]\n+ }\n+ />\n <TouchableWithoutFeedback\n key=\"key-0\"\n onPress={[Function]}\n@@ -2072,9 +2123,6 @@ exports[`renders shifting bottom navigation 1`] = `\n Object {\n \"flex\": 1,\n },\n- Object {\n- \"backgroundColor\": \"#000000\",\n- },\n undefined,\n ]\n }\n@@ -2256,12 +2304,15 @@ exports[`renders shifting bottom navigation 1`] = `\n \"backgroundColor\": undefined,\n \"borderRadius\": 0,\n \"height\": 0,\n- \"left\": 6,\n+ \"left\": -4,\n \"opacity\": 0,\n \"top\": 28,\n \"transform\": Array [\n Object {\n- \"scale\": 0.002,\n+ \"translateX\": 10,\n+ },\n+ Object {\n+ \"scale\": 0.008,\n },\n ],\n \"width\": 0,\n@@ -2278,17 +2329,20 @@ exports[`renders shifting bottom navigation 1`] = `\n },\n Object {\n \"backgroundColor\": \"rgba(255, 255, 255, 0.12)\",\n- \"borderRadius\": 36,\n- \"height\": 72,\n- \"left\": -30,\n- \"opacity\": 0,\n- \"top\": -8,\n+ \"borderRadius\": 48,\n+ \"height\": 96,\n+ \"left\": -52,\n+ \"opacity\": 0.002,\n+ \"top\": -20,\n \"transform\": Array [\n Object {\n- \"scale\": 0.002,\n+ \"translateX\": 10,\n+ },\n+ Object {\n+ \"scale\": 0.001,\n },\n ],\n- \"width\": 72,\n+ \"width\": 96,\n },\n ]\n }\n"]
1
["87427fe39d165bee2acedde8dbaa237cca3fb61e"]
["test"]
support multi deploy in ReplayStateTest In order to support cases for call-activity, the ReplayStateTest needs to be able to deploy multiple processes.
["diff --git a/engine/src/test/java/io/zeebe/engine/processing/streamprocessor/ReplayStateTest.java b/engine/src/test/java/io/zeebe/engine/processing/streamprocessor/ReplayStateTest.java\nindex 77e320f..0389291 100644\n--- a/engine/src/test/java/io/zeebe/engine/processing/streamprocessor/ReplayStateTest.java\n+++ b/engine/src/test/java/io/zeebe/engine/processing/streamprocessor/ReplayStateTest.java\n@@ -22,9 +22,9 @@ import io.zeebe.protocol.record.intent.ProcessInstanceIntent;\n import io.zeebe.protocol.record.value.BpmnElementType;\n import io.zeebe.test.util.record.RecordingExporter;\n import java.time.Duration;\n+import java.util.ArrayList;\n import java.util.Collection;\n import java.util.List;\n-import java.util.Optional;\n import java.util.function.Function;\n import org.assertj.core.api.SoftAssertions;\n import org.awaitility.Awaitility;\n@@ -170,7 +170,7 @@ public final class ReplayStateTest {\n @Test\n public void shouldRestoreState() {\n // given\n- testCase.process.ifPresent(process -> engine.deployment().withXmlResource(process).deploy());\n+ testCase.processes.forEach(process -> engine.deployment().withXmlResource(process).deploy());\n \n final Record<?> finalRecord = testCase.execution.apply(engine);\n \n@@ -227,7 +227,7 @@ public final class ReplayStateTest {\n \n private static final class TestCase {\n private final String description;\n- private Optional<BpmnModelInstance> process = Optional.empty();\n+ private final List<BpmnModelInstance> processes = new ArrayList<>();\n private Function<EngineRule, Record<?>> execution =\n engine -> RecordingExporter.records().getFirst();\n \n@@ -236,7 +236,7 @@ public final class ReplayStateTest {\n }\n \n private TestCase withProcess(final BpmnModelInstance process) {\n- this.process = Optional.of(process);\n+ processes.add(process);\n return this;\n }\n \n"]
1
["834a9e3cfd6439f295072e5808a02cf2a35ba083"]
["test"]
Use arm64v8 postfix for Cube Store :dev build
["diff --git a/.github/workflows/rust-cubestore-master.yml b/.github/workflows/rust-cubestore-master.yml\nindex 4a84984..bb07cd7 100644\n--- a/.github/workflows/rust-cubestore-master.yml\n+++ b/.github/workflows/rust-cubestore-master.yml\n@@ -115,9 +115,9 @@ jobs:\n if [[ $VERSION =~ ^v[0-9]{1,3}\\.[0-9]{1,3}\\.[0-9]{1,3}$ ]]; then\n MINOR=${VERSION%.*}\n MAJOR=${MINOR%.*}\n- TAGS=\"$TAGS,${DOCKER_IMAGE}:${MINOR},${DOCKER_IMAGE}:${MAJOR},${DOCKER_IMAGE}:latest\"\n+ TAGS=\"$TAGS,${DOCKER_IMAGE}:${MINOR},${DOCKER_IMAGE}:${MAJOR}\"\n elif [ \"${{ github.event_name }}\" = \"push\" ]; then\n- TAGS=\"$TAGS,${DOCKER_IMAGE}:build-1${GITHUB_RUN_NUMBER}\"\n+ TAGS=\"$TAGS,${DOCKER_IMAGE}:build-1${GITHUB_RUN_NUMBER}${{ matrix.postfix }}\"\n fi\n \n echo ::set-output name=version::${VERSION}\n"]
1
["10bdcb452ff9d2b884d45a9c43a4b8a20fc4a883"]
["ci"]
add method to extract snapshot name from filename also corrected pattern, where the period was meant to match a period, not any character. related to zeebe-io/zeebe#876
["diff --git a/logstreams/src/main/java/io/zeebe/logstreams/impl/snapshot/fs/FsSnapshotStorageConfiguration.java b/logstreams/src/main/java/io/zeebe/logstreams/impl/snapshot/fs/FsSnapshotStorageConfiguration.java\nindex d8f4d89..e54e85a 100644\n--- a/logstreams/src/main/java/io/zeebe/logstreams/impl/snapshot/fs/FsSnapshotStorageConfiguration.java\n+++ b/logstreams/src/main/java/io/zeebe/logstreams/impl/snapshot/fs/FsSnapshotStorageConfiguration.java\n@@ -23,8 +23,9 @@ public class FsSnapshotStorageConfiguration\n {\n protected static final String CHECKSUM_ALGORITHM = \"SHA1\";\n \n- protected static final String SNAPSHOT_FILE_NAME_TEMPLATE = \"%s\" + File.separatorChar + \"%s-%d.snapshot\";\n- protected static final String SNAPSHOT_FILE_NAME_PATTERN = \"%s-(\\\\d+).snapshot\";\n+ protected static final String SNAPSHOT_FILE_NAME_TEMPLATE = \"%s-%d.snapshot\";\n+ protected static final String SNAPSHOT_FILE_PATH_TEMPLATE = \"%s\" + File.separatorChar + SNAPSHOT_FILE_NAME_TEMPLATE;\n+ protected static final String SNAPSHOT_FILE_NAME_PATTERN = \"%s-(\\\\d+)\\\\.snapshot\";\n \n protected static final String CHECKSUM_FILE_NAME_TEMPLATE = \"%s\" + File.separatorChar + \"%s-%d.\" + CHECKSUM_ALGORITHM.toLowerCase();\n \n@@ -50,7 +51,7 @@ public class FsSnapshotStorageConfiguration\n \n public String snapshotFileName(String name, long logPosition)\n {\n- return String.format(SNAPSHOT_FILE_NAME_TEMPLATE, rootPath, name, logPosition);\n+ return String.format(SNAPSHOT_FILE_PATH_TEMPLATE, rootPath, name, logPosition);\n }\n \n public String checksumFileName(String name, long logPosition)\n@@ -86,7 +87,7 @@ public class FsSnapshotStorageConfiguration\n return String.format(CHECKSUM_CONTENT_TEMPLATE, checksum, dataFileName);\n }\n \n- public String extractDigetsFromChecksumContent(String content)\n+ public String extractDigestFromChecksumContent(String content)\n {\n final int indexOfSeparator = content.indexOf(CHECKSUM_CONTENT_SEPARATOR);\n if (indexOfSeparator < 0)\n@@ -108,9 +109,18 @@ public class FsSnapshotStorageConfiguration\n return content.substring(indexOfSeparator + CHECKSUM_CONTENT_SEPARATOR.length());\n }\n \n+ public String getSnapshotNameFromFileName(final String fileName)\n+ {\n+ final String suffixPattern = String.format(SNAPSHOT_FILE_NAME_PATTERN, \"\");\n+ final Pattern pattern = Pattern.compile(suffixPattern);\n+ final String[] parts = pattern.split(fileName);\n+\n+ return parts[0];\n+ }\n+\n public String getSnapshotFileNameTemplate()\n {\n- return SNAPSHOT_FILE_NAME_TEMPLATE;\n+ return SNAPSHOT_FILE_PATH_TEMPLATE;\n }\n \n public String getChecksumFileNameTemplate()\n"]
1
["7ab965c55d0e98fdb6179577d0db56599675e400"]
["feat"]
adds test for exec with http proxy Signed-off-by: Sam Alba <[email protected]>
["diff --git a/tests/tasks.bats b/tests/tasks.bats\nindex e3b95c4..4cfba15 100644\n--- a/tests/tasks.bats\n+++ b/tests/tasks.bats\n@@ -101,6 +101,14 @@ setup() {\n assert_line --partial --index 9 'actions.basicTest.stop'\n }\n \n+@test \"task: #Exec with HTTP proxy\" {\n+ cd ./tasks/exec\n+ export HTTPS_PROXY=\"https://localhost:4242/\"\n+ run \"$DAGGER\" \"do\" -p ./http_proxy.cue curlProxy\n+ assert_failure\n+ unset HTTP_PROXY\n+}\n+\n @test \"task: #Start #Stop params\" {\n cd ./tasks/exec\n \"$DAGGER\" \"do\" -p ./start_stop_exec.cue execParamsTest\n@@ -297,4 +305,4 @@ setup() {\n \n @test \"task: #Rm\" {\n \"$DAGGER\" \"do\" -p ./tasks/rm/rm.cue test\n-}\n\\ No newline at end of file\n+}\ndiff --git a/tests/tasks/exec/http_proxy.cue b/tests/tasks/exec/http_proxy.cue\nnew file mode 100644\nindex 0000000..05de4b9\n--- /dev/null\n+++ b/tests/tasks/exec/http_proxy.cue\n@@ -0,0 +1,25 @@\n+package main\n+\n+import (\n+\t\"dagger.io/dagger\"\n+\t\"dagger.io/dagger/core\"\n+)\n+\n+dagger.#Plan & {\n+\tactions: {\n+\t\timage: core.#Pull & {\n+\t\t\tsource: \"alpine:3.15.0@sha256:e7d88de73db3d3fd9b2d63aa7f447a10fd0220b7cbf39803c803f2af9ba256b3\"\n+\t\t}\n+\n+\t\tcurlProxy: core.#Exec & {\n+\t\t\tinput: image.output\n+\t\t\targs: [\n+\t\t\t\t\"sh\", \"-c\",\n+\t\t\t\t\"\"\"\n+\t\t\t\t\tapk add --no-cache curl\n+\t\t\t\t\tcurl -sfL -o /dev/null https://www.google.com/\n+\t\t\t\t\t\"\"\",\n+\t\t\t]\n+\t\t}\n+\t}\n+}\n"]
1
["6c7398993bc567ec84e4573b6ededbf50b1ef606"]
["test"]
extract lambdas
["diff --git a/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/AsyncSnapshotDirector.java b/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/AsyncSnapshotDirector.java\nindex 6ee5797..bcfcc72 100644\n--- a/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/AsyncSnapshotDirector.java\n+++ b/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/AsyncSnapshotDirector.java\n@@ -224,7 +224,6 @@ public final class AsyncSnapshotDirector extends Actor\n private void takeSnapshot() {\n final var transientSnapshotFuture =\n stateController.takeTransientSnapshot(lowerBoundSnapshotPosition);\n-\n transientSnapshotFuture.onComplete(\n (optionalTransientSnapshot, snapshotTakenError) -> {\n if (snapshotTakenError != null) {\n@@ -237,27 +236,31 @@ public final class AsyncSnapshotDirector extends Actor\n takingSnapshot = false;\n return;\n }\n- pendingSnapshot = optionalTransientSnapshot.get();\n- onRecovered();\n-\n- final ActorFuture<Long> lastWrittenPosition =\n- streamProcessor.getLastWrittenPositionAsync();\n- actor.runOnCompletion(\n- lastWrittenPosition,\n- (endPosition, error) -> {\n- if (error == null) {\n- LOG.info(LOG_MSG_WAIT_UNTIL_COMMITTED, endPosition, commitPosition);\n- lastWrittenEventPosition = endPosition;\n- persistingSnapshot = false;\n- persistSnapshotIfLastWrittenPositionCommitted();\n- } else {\n- resetStateOnFailure();\n- LOG.error(ERROR_MSG_ON_RESOLVE_WRITTEN_POS, error);\n- }\n- });\n+ onTransientSnapshotTaken(optionalTransientSnapshot.get());\n });\n }\n \n+ private void onTransientSnapshotTaken(final TransientSnapshot transientSnapshot) {\n+\n+ pendingSnapshot = transientSnapshot;\n+ onRecovered();\n+\n+ final ActorFuture<Long> lastWrittenPosition = streamProcessor.getLastWrittenPositionAsync();\n+ actor.runOnCompletion(lastWrittenPosition, this::onLastWrittenPositionReceived);\n+ }\n+\n+ private void onLastWrittenPositionReceived(final Long endPosition, final Throwable error) {\n+ if (error == null) {\n+ LOG.info(LOG_MSG_WAIT_UNTIL_COMMITTED, endPosition, commitPosition);\n+ lastWrittenEventPosition = endPosition;\n+ persistingSnapshot = false;\n+ persistSnapshotIfLastWrittenPositionCommitted();\n+ } else {\n+ resetStateOnFailure();\n+ LOG.error(ERROR_MSG_ON_RESOLVE_WRITTEN_POS, error);\n+ }\n+ }\n+\n private void onRecovered() {\n if (healthStatus != HealthStatus.HEALTHY) {\n healthStatus = HealthStatus.HEALTHY;\n"]
1
["14abf5c31523a551134aebe9e8f3505ef26ed421"]
["refactor"]
fix unstable MessageCorrelationTest
["diff --git a/broker-core/src/test/java/io/zeebe/broker/workflow/MessageCorrelationTest.java b/broker-core/src/test/java/io/zeebe/broker/workflow/MessageCorrelationTest.java\nindex 0f5fed9..796393c 100644\n--- a/broker-core/src/test/java/io/zeebe/broker/workflow/MessageCorrelationTest.java\n+++ b/broker-core/src/test/java/io/zeebe/broker/workflow/MessageCorrelationTest.java\n@@ -27,7 +27,6 @@ import static io.zeebe.test.util.MsgPackUtil.asMsgPack;\n import static org.assertj.core.api.Assertions.assertThat;\n import static org.assertj.core.api.Assertions.entry;\n \n-import io.zeebe.UnstableTest;\n import io.zeebe.broker.test.EmbeddedBrokerRule;\n import io.zeebe.model.bpmn.Bpmn;\n import io.zeebe.model.bpmn.BpmnModelInstance;\n@@ -50,7 +49,6 @@ import org.agrona.DirectBuffer;\n import org.junit.Before;\n import org.junit.Rule;\n import org.junit.Test;\n-import org.junit.experimental.categories.Category;\n import org.junit.rules.RuleChain;\n import org.junit.runner.RunWith;\n import org.junit.runners.Parameterized;\n@@ -165,7 +163,7 @@ public class MessageCorrelationTest {\n \"receive-message\", WorkflowInstanceIntent.ELEMENT_ACTIVATED);\n \n final SubscribedRecord messageSubscription =\n- findMessageSubscription(testClient, MessageSubscriptionIntent.OPENED);\n+ findMessageSubscription(MessageSubscriptionIntent.OPENED);\n assertThat(messageSubscription.valueType()).isEqualTo(ValueType.MESSAGE_SUBSCRIPTION);\n assertThat(messageSubscription.recordType()).isEqualTo(RecordType.EVENT);\n assertThat(messageSubscription.value())\n@@ -244,7 +242,7 @@ public class MessageCorrelationTest {\n final long workflowInstanceKey =\n testClient.createWorkflowInstance(\"wf\", asMsgPack(\"orderId\", \"order-123\"));\n \n- testClient.receiveFirstWorkflowInstanceEvent(WorkflowInstanceIntent.ELEMENT_ACTIVATED);\n+ findMessageSubscription(MessageSubscriptionIntent.OPENED);\n \n // when\n testClient.publishMessage(\"order canceled\", \"order-123\", asMsgPack(\"foo\", \"bar\"));\n@@ -308,13 +306,12 @@ public class MessageCorrelationTest {\n }\n \n @Test\n- @Category(UnstableTest.class) // => https://github.com/zeebe-io/zeebe/issues/1234\n public void shouldCorrelateMessageWithZeroTTL() throws Exception {\n // given\n final long workflowInstanceKey =\n testClient.createWorkflowInstance(\"wf\", asMsgPack(\"orderId\", \"order-123\"));\n \n- testClient.receiveElementInState(\"receive-message\", WorkflowInstanceIntent.ELEMENT_ACTIVATED);\n+ findMessageSubscription(MessageSubscriptionIntent.OPENED);\n \n // when\n testClient.publishMessage(\"order canceled\", \"order-123\", asMsgPack(\"foo\", \"bar\"), 0);\n@@ -499,10 +496,9 @@ public class MessageCorrelationTest {\n .containsEntry(\"activityInstanceKey\", catchEventEntered.key());\n }\n \n- private SubscribedRecord findMessageSubscription(\n- final TestPartitionClient client, final MessageSubscriptionIntent intent)\n+ private SubscribedRecord findMessageSubscription(final MessageSubscriptionIntent intent)\n throws AssertionError {\n- return client\n+ return testClient\n .receiveEvents()\n .filter(intent(intent))\n .findFirst()\n"]
1
["98bed2a8137930149559bc1cae9bd34a1a75e556"]
["test"]
use module path alias
["diff --git a/src/background/audio-manager.ts b/src/background/audio-manager.ts\nindex 54e8b24..11c5fba 100644\n--- a/src/background/audio-manager.ts\n+++ b/src/background/audio-manager.ts\n@@ -2,7 +2,7 @@\n * To make sure only one audio plays at a time\n */\n \n-import { timeout } from '../_helpers/promise-more'\n+import { timeout } from '@/_helpers/promise-more'\n \n declare global {\n interface Window {\ndiff --git a/src/background/context-menus.ts b/src/background/context-menus.ts\nindex 994b59e..7036362 100644\n--- a/src/background/context-menus.ts\n+++ b/src/background/context-menus.ts\n@@ -1,5 +1,5 @@\n-import { storage, openURL } from '../_helpers/browser-api'\n-import { AppConfig } from '../app-config'\n+import { storage, openURL } from '@/_helpers/browser-api'\n+import { AppConfig } from '@/app-config'\n \n import { Observable } from 'rxjs/Observable'\n import { fromPromise } from 'rxjs/observable/fromPromise'\ndiff --git a/src/background/initialization.ts b/src/background/initialization.ts\nindex 0e5b3ad..001ee73 100644\n--- a/src/background/initialization.ts\n+++ b/src/background/initialization.ts\n@@ -1,6 +1,6 @@\n-import { storage, openURL } from '../_helpers/browser-api'\n-import checkUpdate from '../_helpers/check-update'\n-import { AppConfig } from '../app-config'\n+import { storage, openURL } from '@/_helpers/browser-api'\n+import checkUpdate from '@/_helpers/check-update'\n+import { AppConfig } from '@/app-config'\n import { mergeConfig } from './merge-config'\n import { init as initMenus } from './context-menus'\n import { init as initPdf } from './pdf-sniffer'\ndiff --git a/src/background/merge-config.ts b/src/background/merge-config.ts\nindex afa1800..afdbd63 100644\n--- a/src/background/merge-config.ts\n+++ b/src/background/merge-config.ts\n@@ -1,4 +1,4 @@\n-import { appConfigFactory, AppConfig } from '../app-config'\n+import { appConfigFactory, AppConfig } from '@/app-config'\n import _ from 'lodash'\n \n /**\n@@ -24,7 +24,7 @@ function initConfig (): Promise<AppConfig> {\n const storageObj = { config: appConfigFactory() }\n \n Object.keys(storageObj.config.dicts.all).forEach(id => {\n- storageObj[id] = require('../components/dictionaries/' + id + '/config')\n+ storageObj[id] = require('@/components/dictionaries/' + id + '/config')\n })\n \n return browser.storage.sync.set(storageObj)\n@@ -70,7 +70,7 @@ function mergeHistorical (config): Promise<AppConfig> {\n \n const storageObj = { config: base }\n Object.keys(base.dicts.all).forEach(id => {\n- storageObj[id] = config.dicts.all[id] || require('../components/dictionaries/' + id + '/config')\n+ storageObj[id] = config.dicts.all[id] || require('@/components/dictionaries/' + id + '/config')\n })\n \n return browser.storage.sync.set(storageObj)\ndiff --git a/src/background/pdf-sniffer.ts b/src/background/pdf-sniffer.ts\nindex 6ba27cf..70aa38f 100644\n--- a/src/background/pdf-sniffer.ts\n+++ b/src/background/pdf-sniffer.ts\n@@ -2,8 +2,8 @@\n * Open pdf link directly\n */\n \n-import { storage } from '../_helpers/browser-api'\n-import { AppConfig } from '../app-config'\n+import { storage } from '@/_helpers/browser-api'\n+import { AppConfig } from '@/app-config'\n \n export function init (pdfSniff: boolean) {\n if (browser.webRequest.onBeforeRequest.hasListener(otherPdfListener)) {\ndiff --git a/src/background/server.ts b/src/background/server.ts\nindex 73b34b6..66ed5c0 100644\n--- a/src/background/server.ts\n+++ b/src/background/server.ts\n@@ -1,7 +1,7 @@\n-import { DictID } from '../app-config'\n-import { message, openURL } from '../_helpers/browser-api'\n+import { DictID } from '@/app-config'\n+import { message, openURL } from '@/_helpers/browser-api'\n import { play } from './audio-manager'\n-import { chsToChz } from '../_helpers/chs-to-chz'\n+import { chsToChz } from '@/_helpers/chs-to-chz'\n \n interface MessageOpenUrlWithEscape {\n type: 'OPEN_URL'\n@@ -63,7 +63,7 @@ function fetchDictResult (data: MessageFetchDictResult): Promise<void> {\n let search\n \n try {\n- search = require('../components/dictionaries/' + data.dict + '/engine.js')\n+ search = require('@/components/dictionaries/' + data.dict + '/engine.js')\n } catch (err) {\n return Promise.reject(err)\n }\ndiff --git a/test/unit/_helpers/browser-api.spec.ts b/test/unit/_helpers/browser-api.spec.ts\nindex 1f39145..e327169 100644\n--- a/test/unit/_helpers/browser-api.spec.ts\n+++ b/test/unit/_helpers/browser-api.spec.ts\n@@ -1,4 +1,4 @@\n-import { message, storage, openURL } from '../../../src/_helpers/browser-api'\n+import { message, storage, openURL } from '@/_helpers/browser-api'\n \n beforeEach(() => {\n browser.flush()\ndiff --git a/test/unit/_helpers/check-update.spec.ts b/test/unit/_helpers/check-update.spec.ts\nindex 2abfc57..fd0b678 100644\n--- a/test/unit/_helpers/check-update.spec.ts\n+++ b/test/unit/_helpers/check-update.spec.ts\n@@ -1,4 +1,4 @@\n-import checkUpdate from '../../../src/_helpers/check-update'\n+import checkUpdate from '@/_helpers/check-update'\n import fetchMock from 'jest-fetch-mock'\n \n describe('Check Update', () => {\ndiff --git a/test/unit/_helpers/chs-to-chz.spec.ts b/test/unit/_helpers/chs-to-chz.spec.ts\nindex 295c6ad..21d5229 100644\n--- a/test/unit/_helpers/chs-to-chz.spec.ts\n+++ b/test/unit/_helpers/chs-to-chz.spec.ts\n@@ -1,4 +1,4 @@\n-import chsToChz from '../../../src/_helpers/chs-to-chz'\n+import chsToChz from '@/_helpers/chs-to-chz'\n \n describe('Chs to Chz', () => {\n it('should convert chs to chz', () => {\ndiff --git a/test/unit/_helpers/fetch-dom.spec.ts b/test/unit/_helpers/fetch-dom.spec.ts\nindex a79dda0..bbfbf10 100644\n--- a/test/unit/_helpers/fetch-dom.spec.ts\n+++ b/test/unit/_helpers/fetch-dom.spec.ts\n@@ -1,4 +1,4 @@\n-import fetchDom from '../../../src/_helpers/fetch-dom'\n+import fetchDom from '@/_helpers/fetch-dom'\n \n class XMLHttpRequestMock {\n static queue: XMLHttpRequestMock[] = []\ndiff --git a/test/unit/_helpers/lang-check.spec.ts b/test/unit/_helpers/lang-check.spec.ts\nindex f3e668a..09f30bb 100644\n--- a/test/unit/_helpers/lang-check.spec.ts\n+++ b/test/unit/_helpers/lang-check.spec.ts\n@@ -1,4 +1,4 @@\n-import { isContainChinese, isContainEnglish } from '../../../src/_helpers/lang-check'\n+import { isContainChinese, isContainEnglish } from '@/_helpers/lang-check'\n \n describe('Language Check', () => {\n it('isContainChinese should return ture if text contains Chinese', () => {\ndiff --git a/test/unit/_helpers/promise-more.spec.ts b/test/unit/_helpers/promise-more.spec.ts\nindex 9601c7d..66dc8d9 100644\n--- a/test/unit/_helpers/promise-more.spec.ts\n+++ b/test/unit/_helpers/promise-more.spec.ts\n@@ -1,4 +1,4 @@\n-import * as pm from '../../../src/_helpers/promise-more'\n+import * as pm from '@/_helpers/promise-more'\n \n describe('Promise More', () => {\n beforeAll(() => {\ndiff --git a/test/unit/_helpers/selection.spec.ts b/test/unit/_helpers/selection.spec.ts\nindex 370239a..06812cf 100644\n--- a/test/unit/_helpers/selection.spec.ts\n+++ b/test/unit/_helpers/selection.spec.ts\n@@ -1,4 +1,4 @@\n-import selection from '../../../src/_helpers/selection'\n+import selection from '@/_helpers/selection'\n \n describe('Selection', () => {\n const bakSelection = window.getSelection\ndiff --git a/test/unit/_helpers/strip-script.spec.ts b/test/unit/_helpers/strip-script.spec.ts\nindex cce558f..355b382 100644\n--- a/test/unit/_helpers/strip-script.spec.ts\n+++ b/test/unit/_helpers/strip-script.spec.ts\n@@ -1,4 +1,4 @@\n-import stripScript from '../../../src/_helpers/strip-script'\n+import stripScript from '@/_helpers/strip-script'\n \n describe('Strip Script', () => {\n const expectedEl = document.createElement('div') as HTMLDivElement\ndiff --git a/test/unit/background/audio-manager.spec.ts b/test/unit/background/audio-manager.spec.ts\nindex b0096a6..b1266d7 100644\n--- a/test/unit/background/audio-manager.spec.ts\n+++ b/test/unit/background/audio-manager.spec.ts\n@@ -1,4 +1,4 @@\n-import audio from '../../../src/background/audio-manager'\n+import audio from '@/background/audio-manager'\n \n describe('Audio Manager', () => {\n const bakAudio = (window as any).Audio\ndiff --git a/test/unit/background/context-menus.spec.ts b/test/unit/background/context-menus.spec.ts\nindex 39e249c..d9049dc 100644\n--- a/test/unit/background/context-menus.spec.ts\n+++ b/test/unit/background/context-menus.spec.ts\n@@ -1,4 +1,4 @@\n-import { appConfigFactory, AppConfig } from '../../../src/app-config'\n+import { appConfigFactory, AppConfig } from '@/app-config'\n import sinon from 'sinon'\n \n function specialConfig () {\n@@ -11,7 +11,7 @@ describe('Context Menus', () => {\n beforeAll(() => {\n browser.flush()\n jest.resetModules()\n- require('../../../src/background/context-menus')\n+ require('@/background/context-menus')\n })\n afterAll(() => browser.flush())\n \n@@ -93,7 +93,7 @@ describe('Context Menus', () => {\n browser.contextMenus.create.callsFake((_, cb) => cb())\n config = specialConfig()\n jest.resetModules()\n- const { init } = require('../../../src/background/context-menus')\n+ const { init } = require('@/background/context-menus')\n init(config.contextMenus)\n })\n \n@@ -110,7 +110,7 @@ describe('Context Menus', () => {\n it('should not init setup when called multiple times', () => {\n expect(browser.contextMenus.removeAll.calledOnce).toBeTruthy()\n \n- const { init } = require('../../../src/background/context-menus')\n+ const { init } = require('@/background/context-menus')\n init(config.contextMenus)\n init(config.contextMenus)\n \ndiff --git a/test/unit/background/initialization.spec.ts b/test/unit/background/initialization.spec.ts\nindex 7bc0972..56a6389 100644\n--- a/test/unit/background/initialization.spec.ts\n+++ b/test/unit/background/initialization.spec.ts\n@@ -1,4 +1,4 @@\n-import { appConfigFactory, AppConfig } from '../../../src/app-config'\n+import { appConfigFactory, AppConfig } from '@/app-config'\n import fetchMock from 'jest-fetch-mock'\n import sinon from 'sinon'\n \n@@ -11,12 +11,12 @@ describe('Initialization', () => {\n const checkUpdate = jest.fn().mockReturnValue(Promise.resolve())\n \n beforeAll(() => {\n- const { message, storage } = require('../../../src/_helpers/browser-api')\n+ const { message, storage } = require('@/_helpers/browser-api')\n window.fetch = fetchMock\n \n browser.flush()\n jest.resetModules()\n- jest.doMock('../../../src/background/merge-config', () => {\n+ jest.doMock('@/background/merge-config', () => {\n return {\n mergeConfig (config) {\n mergeConfig(config)\n@@ -24,16 +24,16 @@ describe('Initialization', () => {\n }\n }\n })\n- jest.doMock('../../../src/background/context-menus', () => {\n+ jest.doMock('@/background/context-menus', () => {\n return { init: initMenus }\n })\n- jest.doMock('../../../src/background/pdf-sniffer', () => {\n+ jest.doMock('@/background/pdf-sniffer', () => {\n return { init: initPdf }\n })\n- jest.doMock('../../../src/_helpers/check-update', () => {\n+ jest.doMock('@/_helpers/check-update', () => {\n return checkUpdate\n })\n- jest.doMock('../../../src/_helpers/browser-api', () => {\n+ jest.doMock('@/_helpers/browser-api', () => {\n return {\n message,\n storage,\n@@ -41,13 +41,13 @@ describe('Initialization', () => {\n }\n })\n \n- require('../../../src/background/initialization')\n+ require('@/background/initialization')\n })\n afterAll(() => {\n browser.flush()\n- jest.dontMock('../../../src/background/merge-config')\n- jest.dontMock('../../../src/background/context-menus')\n- jest.dontMock('../../../src/_helpers/browser-api')\n+ jest.dontMock('@/background/merge-config')\n+ jest.dontMock('@/background/context-menus')\n+ jest.dontMock('@/_helpers/browser-api')\n window.fetch = bakFetch\n })\n \ndiff --git a/test/unit/background/merge-config.spec.ts b/test/unit/background/merge-config.spec.ts\nindex 73c047d..c0dce26 100644\n--- a/test/unit/background/merge-config.spec.ts\n+++ b/test/unit/background/merge-config.spec.ts\n@@ -1,5 +1,5 @@\n-import { appConfigFactory, AppConfig, AppConfigMutable } from '../../../src/app-config'\n-import mergeConfig from '../../../src/background/merge-config'\n+import { appConfigFactory, AppConfig, AppConfigMutable } from '@/app-config'\n+import mergeConfig from '@/background/merge-config'\n import sinon from 'sinon'\n \n describe('Merge Config', () => {\ndiff --git a/test/unit/background/pdf-sniffer.spec.ts b/test/unit/background/pdf-sniffer.spec.ts\nindex a0219d2..bb7726f 100644\n--- a/test/unit/background/pdf-sniffer.spec.ts\n+++ b/test/unit/background/pdf-sniffer.spec.ts\n@@ -1,5 +1,5 @@\n-import { appConfigFactory, AppConfig } from '../../../src/app-config'\n-import { init as initPdf } from '../../../src/background/pdf-sniffer'\n+import { appConfigFactory, AppConfig } from '@/app-config'\n+import { init as initPdf } from '@/background/pdf-sniffer'\n import sinon from 'sinon'\n \n function hasListenerPatch (fn) {\ndiff --git a/test/unit/background/server.spec.ts b/test/unit/background/server.spec.ts\nindex b8ef065..aa04525 100644\n--- a/test/unit/background/server.spec.ts\n+++ b/test/unit/background/server.spec.ts\n@@ -1,5 +1,5 @@\n-import { appConfigFactory, AppConfig } from '../../../src/app-config'\n-import * as browserWrap from '../../../src/_helpers/browser-api'\n+import { appConfigFactory, AppConfig } from '@/app-config'\n+import * as browserWrap from '@/_helpers/browser-api'\n import sinon from 'sinon'\n \n describe('Server', () => {\n@@ -13,26 +13,26 @@ describe('Server', () => {\n browserWrap.openURL = openURL\n \n beforeAll(() => {\n- jest.doMock('../../../src/_helpers/chs-to-chz', () => {\n+ jest.doMock('@/_helpers/chs-to-chz', () => {\n return { chsToChz }\n })\n- jest.doMock('../../../src/background/audio-manager', () => {\n+ jest.doMock('@/background/audio-manager', () => {\n return { play }\n })\n- jest.doMock('../../../src/_helpers/browser-api', () => {\n+ jest.doMock('@/_helpers/browser-api', () => {\n return browserWrap\n })\n- jest.doMock('../../../src/components/dictionaries/bing/engine.js', () => {\n+ jest.doMock('@/components/dictionaries/bing/engine.js', () => {\n return bingSearch\n })\n })\n \n afterAll(() => {\n browser.flush()\n- jest.dontMock('../../../src/_helpers/chs-to-chz')\n- jest.dontMock('../../../src/background/audio-manager')\n- jest.dontMock('../../../src/_helpers/browser-api')\n- jest.dontMock('../../../src/components/dictionaries/bing/engine.js')\n+ jest.dontMock('@/_helpers/chs-to-chz')\n+ jest.dontMock('@/background/audio-manager')\n+ jest.dontMock('@/_helpers/browser-api')\n+ jest.dontMock('@/components/dictionaries/bing/engine.js')\n })\n \n beforeEach(() => {\n@@ -46,7 +46,7 @@ describe('Server', () => {\n bingSearch.mockReset()\n bingSearch.mockImplementation(() => Promise.resolve())\n jest.resetModules()\n- require('../../../src/background/server')\n+ require('@/background/server')\n })\n \n it('should properly init', () => {\n"]
1
["8246d024f21d93cc092e19bede5f7b3a5325c8dc"]
["refactor"]
ensure "dist" dirs exist
["diff --git a/scripts/prepare.js b/scripts/prepare.js\nindex 9eb8cb8..f285825 100644\n--- a/scripts/prepare.js\n+++ b/scripts/prepare.js\n@@ -68,6 +68,9 @@ async function prepare() {\n names.push(json.name)\n }\n \n+ // Ensure all \"dist\" directories exist.\n+ dirs.forEach(dir => fs.ensureDirSync(join(dir, distId)))\n+\n log(``)\n for (let i = 0; i < names.length; i++) {\n const dir = dirs[i]\n"]
1
["ca060bf255a55b99000ddf0c67f7422f28b735a6"]
["build"]
run nix macos jobs on macos-13 to try and avoid SIP
["diff --git a/.github/actionlint.yaml b/.github/actionlint.yaml\nnew file mode 100644\nindex 0000000..5be7d17\n--- /dev/null\n+++ b/.github/actionlint.yaml\n@@ -0,0 +1,7 @@\n+self-hosted-runner:\n+ # Labels of self-hosted runner in array of strings.\n+ labels: [macos-13]\n+# Configuration variables in array of strings defined in your repository or\n+# organization. `null` means disabling configuration variables check.\n+# Empty array means no configuration variable is allowed.\n+config-variables: null\ndiff --git a/.github/workflows/nix.yml b/.github/workflows/nix.yml\nindex e37346c..dce77e1 100644\n--- a/.github/workflows/nix.yml\n+++ b/.github/workflows/nix.yml\n@@ -37,7 +37,7 @@ jobs:\n - \"3.10\"\n - \"3.11\"\n include:\n- - os: macos-latest\n+ - os: macos-13\n python-version: \"3.10\"\n steps:\n - name: checkout\ndiff --git a/.pre-commit-config.yaml b/.pre-commit-config.yaml\nindex 005a850..8db22e2 100644\n--- a/.pre-commit-config.yaml\n+++ b/.pre-commit-config.yaml\n@@ -3,7 +3,7 @@ ci:\n autofix_prs: false\n autoupdate_commit_msg: \"chore(deps): pre-commit.ci autoupdate\"\n skip:\n- - actionlint\n+ - actionlint-system\n - deadnix\n - just\n - nixpkgs-fmt\n@@ -17,9 +17,9 @@ default_stages:\n - commit\n repos:\n - repo: https://github.com/rhysd/actionlint\n- rev: v1.6.24\n+ rev: v1.6.25\n hooks:\n- - id: actionlint\n+ - id: actionlint-system\n - repo: https://github.com/psf/black\n rev: 23.3.0\n hooks:\n@@ -30,7 +30,7 @@ repos:\n - id: nbstripout\n exclude: .+/rendered/.+\n - repo: https://github.com/codespell-project/codespell\n- rev: v2.2.4\n+ rev: v2.2.5\n hooks:\n - id: codespell\n additional_dependencies:\n"]
1
["54cb6d4643b4a072ff997592a7fa14a69a6c068d"]
["ci"]
added vue3 readme
["diff --git a/core/main/README.md b/core/main/README.md\nindex e5e4c93..e9cfda9 100644\n--- a/core/main/README.md\n+++ b/core/main/README.md\n@@ -217,7 +217,7 @@ You can find the instructions [here](https://github.com/matteobruni/tsparticles/\n \n You can find the instructions [here](https://github.com/matteobruni/tsparticles/blob/master/components/svelte/README.md)\n \n-### VueJS\n+### VueJS 2.x\n \n #### `particles.vue`\n \n@@ -225,6 +225,14 @@ You can find the instructions [here](https://github.com/matteobruni/tsparticles/\n \n You can find the instructions [here](https://github.com/matteobruni/tsparticles/blob/master/components/vue/README.md)\n \n+### VueJS 3.x\n+\n+#### `particles.vue3`\n+\n+[![npm](https://img.shields.io/npm/v/particles.vue3)](https://www.npmjs.com/package/particles.vue3) [![npm](https://img.shields.io/npm/dm/particles.vue3)](https://www.npmjs.com/package/particles.vue3)\n+\n+You can find the instructions [here](https://github.com/matteobruni/tsparticles/blob/master/components/vue3/README.md)\n+\n ---\n \n ## **_Demo / Generator_**\ndiff --git a/core/main/tsconfig.json b/core/main/tsconfig.json\nindex 7916bc5..72399c0 100644\n--- a/core/main/tsconfig.json\n+++ b/core/main/tsconfig.json\n@@ -107,10 +107,14 @@\n \"source\": \"../../components/react/README.md\"\n },\n {\n- \"title\": \"Vue\",\n+ \"title\": \"Vue 2.x\",\n \"source\": \"../../components/vue/README.md\"\n },\n {\n+ \"title\": \"Vue 3.x\",\n+ \"source\": \"../../components/vue3/README.md\"\n+ },\n+ {\n \"title\": \"Svelte\",\n \"source\": \"../../components/svelte/README.md\"\n },\n"]
1
["e4c3e2cff769ce46d22d5c8f7dd527510443a8a7"]
["docs"]
bump version Signed-off-by: rjshrjndrn <[email protected]>
["diff --git a/scripts/helmcharts/init.sh b/scripts/helmcharts/init.sh\nindex 5a2b4b0..69a6944 100644\n--- a/scripts/helmcharts/init.sh\n+++ b/scripts/helmcharts/init.sh\n@@ -26,7 +26,7 @@ usr=$(whoami)\n \n # Installing k3s\n function install_k8s() {\n- curl -sL https://get.k3s.io | sudo K3S_KUBECONFIG_MODE=\"644\" INSTALL_K3S_VERSION='v1.22.8+k3s1' INSTALL_K3S_EXEC=\"--no-deploy=traefik\" sh -\n+ curl -sL https://get.k3s.io | sudo K3S_KUBECONFIG_MODE=\"644\" INSTALL_K3S_VERSION='v1.25.6+k3s1' INSTALL_K3S_EXEC=\"--disable=traefik\" sh -\n [[ -d ~/.kube ]] || mkdir ~/.kube\n sudo cp /etc/rancher/k3s/k3s.yaml ~/.kube/config\n sudo chmod 0644 ~/.kube/config\n"]
1
["9a25fe59dfb63d32505afcea3a164ff0b8ea4c71"]
["build"]
fix node test cases run into infinite loop
["diff --git a/packages/designer/tests/document/node/node.test.ts b/packages/designer/tests/document/node/node.test.ts\nindex dd20bd3..113360d 100644\n--- a/packages/designer/tests/document/node/node.test.ts\n+++ b/packages/designer/tests/document/node/node.test.ts\n@@ -26,7 +26,7 @@ import rootHeaderMetadata from '../../fixtures/component-metadata/root-header';\n import rootContentMetadata from '../../fixtures/component-metadata/root-content';\n import rootFooterMetadata from '../../fixtures/component-metadata/root-footer';\n \n-describe.skip('Node \u65b9\u6cd5\u6d4b\u8bd5', () => {\n+describe('Node \u65b9\u6cd5\u6d4b\u8bd5', () => {\n let editor: Editor;\n let designer: Designer;\n let project: Project;\n@@ -474,15 +474,16 @@ describe.skip('Node \u65b9\u6cd5\u6d4b\u8bd5', () => {\n it('didDropIn / didDropOut', () => {\n const form = doc.getNode('node_k1ow3cbo');\n designer.createComponentMeta(divMetadata);\n+ designer.createComponentMeta(formMetadata);\n const callbacks = form.componentMeta.getMetadata().configure.advanced?.callbacks;\n const fn1 = callbacks.onNodeAdd = jest.fn();\n const fn2 = callbacks.onNodeRemove = jest.fn();\n const textField = doc.getNode('node_k1ow3cc9');\n form.didDropIn(textField);\n- expect(fn1).toHaveBeenCalledWith(textField, form);\n+ expect(fn1).toHaveBeenCalledWith(textField.internalToShellNode(), form.internalToShellNode());\n \n form.didDropOut(textField);\n- expect(fn2).toHaveBeenCalledWith(textField, form);\n+ expect(fn2).toHaveBeenCalledWith(textField.internalToShellNode(), form.internalToShellNode());\n });\n \n it('hover', () => {\n"]
1
["d2c3f0ba6f85b659b76636a91ea9ab2b5a95a720"]
["test"]
getBorderSize() missing "width" The correct property name to use is "borderWidth", not just "border". "border" works in Chrome but was breaking in Firefox. Also had to change .ui-grid-header's box-sizing to content-box so IE11 would include the border in height calcs. AND finally IE11 was returning fractional heights so Grid parseInt()s the returned values.
["diff --git a/src/js/core/factories/Grid.js b/src/js/core/factories/Grid.js\nindex dcf10af..2be7842 100644\n--- a/src/js/core/factories/Grid.js\n+++ b/src/js/core/factories/Grid.js\n@@ -1525,7 +1525,7 @@ angular.module('ui.grid')\n var oldHeaderHeight = container.headerHeight;\n var headerHeight = gridUtil.outerElementHeight(container.header);\n \n- container.headerHeight = headerHeight;\n+ container.headerHeight = parseInt(headerHeight, 10);\n \n if (oldHeaderHeight !== headerHeight) {\n rebuildStyles = true;\n@@ -1534,7 +1534,9 @@ angular.module('ui.grid')\n // Get the \"inner\" header height, that is the height minus the top and bottom borders, if present. We'll use it to make sure all the headers have a consistent height\n var topBorder = gridUtil.getBorderSize(container.header, 'top');\n var bottomBorder = gridUtil.getBorderSize(container.header, 'bottom');\n- var innerHeaderHeight = headerHeight - topBorder - bottomBorder;\n+ var innerHeaderHeight = parseInt(headerHeight - topBorder - bottomBorder, 10);\n+\n+ innerHeaderHeight = innerHeaderHeight < 0 ? 0 : innerHeaderHeight;\n \n container.innerHeaderHeight = innerHeaderHeight;\n \ndiff --git a/src/js/core/services/ui-grid-util.js b/src/js/core/services/ui-grid-util.js\nindex 2c32cbe..cc7c36c 100644\n--- a/src/js/core/services/ui-grid-util.js\n+++ b/src/js/core/services/ui-grid-util.js\n@@ -757,6 +757,8 @@ module.service('gridUtil', ['$log', '$window', '$document', '$http', '$templateC\n borderType = 'border';\n }\n \n+ borderType += 'Width';\n+\n var val = parseInt(styles[borderType], 10);\n \n if (isNaN(val)) {\ndiff --git a/src/less/header.less b/src/less/header.less\nindex 5468a43..de8ff0b 100644\n--- a/src/less/header.less\n+++ b/src/less/header.less\n@@ -7,6 +7,7 @@\n \n .ui-grid-header {\n border-bottom: 1px solid @borderColor;\n+ box-sizing: content-box;;\n }\n \n .ui-grid-top-panel {\n"]
1
["174f25214caa10ec643db6c81aaa0f3511bf78f4"]
["fix"]
updated webpack in react
["diff --git a/components/react/package.json b/components/react/package.json\nindex bbeb9ee..43ddebc 100644\n--- a/components/react/package.json\n+++ b/components/react/package.json\n@@ -114,7 +114,7 @@\n \"ts-loader\": \"^9.2.9\",\n \"ts-node\": \"^10.7.0\",\n \"typescript\": \"^4.7.3\",\n- \"webpack\": \"^5.72.0\",\n+ \"webpack\": \"^5.73.0\",\n \"webpack-bundle-analyzer\": \"^4.5.0\",\n \"webpack-cli\": \"^4.9.2\",\n \"webpack-node-externals\": \"^3.0.0\"\ndiff --git a/yarn.lock b/yarn.lock\nindex a3fdb26..19a0716 100644\n--- a/yarn.lock\n+++ b/yarn.lock\n@@ -25212,7 +25212,7 @@ webpack@^4.38.0, webpack@^4.42.1:\n watchpack \"^1.7.4\"\n webpack-sources \"^1.4.1\"\n \n-webpack@^5.54.0, webpack@^5.71.0, webpack@^5.72.0:\n+webpack@^5.54.0, webpack@^5.71.0, webpack@^5.72.0, webpack@^5.73.0:\n version \"5.73.0\"\n resolved \"https://registry.yarnpkg.com/webpack/-/webpack-5.73.0.tgz#bbd17738f8a53ee5760ea2f59dce7f3431d35d38\"\n integrity sha512-svjudQRPPa0YiOYa2lM/Gacw0r6PvxptHj4FuEKQ2kX05ZLkjbVc5MnPs6its5j7IZljnIqSVo/OsY2X0IpHGA==\n"]
1
["78c446cbea61af2268b4c4da03a9ad4283f10049"]
["build"]
verify process can start at supported element types Verifies a PI can be started at specific element types. The test will deploy the process, start an instance at the desired start element and verify that it has been activated succesfully.
["diff --git a/engine/src/test/java/io/camunda/zeebe/engine/processing/processinstance/CreateProcessInstanceSupportedElementTest.java b/engine/src/test/java/io/camunda/zeebe/engine/processing/processinstance/CreateProcessInstanceSupportedElementTest.java\nnew file mode 100644\nindex 0000000..a505307\n--- /dev/null\n+++ b/engine/src/test/java/io/camunda/zeebe/engine/processing/processinstance/CreateProcessInstanceSupportedElementTest.java\n@@ -0,0 +1,233 @@\n+/*\n+ * Copyright Camunda Services GmbH and/or licensed to Camunda Services GmbH under\n+ * one or more contributor license agreements. See the NOTICE file distributed\n+ * with this work for additional information regarding copyright ownership.\n+ * Licensed under the Zeebe Community License 1.1. You may not use this file\n+ * except in compliance with the Zeebe Community License 1.1.\n+ */\n+package io.camunda.zeebe.engine.processing.processinstance;\n+\n+import static org.assertj.core.api.Assertions.assertThat;\n+import static org.assertj.core.groups.Tuple.tuple;\n+\n+import io.camunda.zeebe.engine.util.EngineRule;\n+import io.camunda.zeebe.model.bpmn.Bpmn;\n+import io.camunda.zeebe.model.bpmn.BpmnModelInstance;\n+import io.camunda.zeebe.protocol.record.Record;\n+import io.camunda.zeebe.protocol.record.intent.ProcessInstanceIntent;\n+import io.camunda.zeebe.protocol.record.value.BpmnElementType;\n+import io.camunda.zeebe.test.util.record.RecordingExporter;\n+import io.camunda.zeebe.test.util.record.RecordingExporterTestWatcher;\n+import java.util.Collection;\n+import java.util.Collections;\n+import java.util.List;\n+import java.util.Map;\n+import org.junit.ClassRule;\n+import org.junit.Rule;\n+import org.junit.Test;\n+import org.junit.runner.RunWith;\n+import org.junit.runners.Parameterized;\n+import org.junit.runners.Parameterized.Parameters;\n+\n+@RunWith(Parameterized.class)\n+public class CreateProcessInstanceSupportedElementTest {\n+\n+ @ClassRule public static final EngineRule ENGINE = EngineRule.singlePartition();\n+ private static final String PROCESS_ID = \"processId\";\n+ private static final String CHILD_PROCESS_ID = \"childProcessId\";\n+ private static final String START_ELEMENT_ID = \"startElement\";\n+ private static final String MESSAGE = \"message\";\n+ private static final String JOBTYPE = \"jobtype\";\n+\n+ @Rule\n+ public final RecordingExporterTestWatcher recordingExporterTestWatcher =\n+ new RecordingExporterTestWatcher();\n+\n+ private final Scenario scenario;\n+\n+ public CreateProcessInstanceSupportedElementTest(final Scenario scenario) {\n+ this.scenario = scenario;\n+ }\n+\n+ @Parameters(name = \"{0}\")\n+ public static Collection<Object> scenarios() {\n+ return List.of(\n+ new Scenario(\n+ BpmnElementType.SUB_PROCESS,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .subProcess(START_ELEMENT_ID)\n+ .embeddedSubProcess()\n+ .startEvent()\n+ .subProcessDone()\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.EVENT_SUB_PROCESS,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .eventSubProcess(\n+ START_ELEMENT_ID, e -> e.startEvent().timerWithDuration(\"PT1H\").endEvent())\n+ .startEvent()\n+ .endEvent()\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.INTERMEDIATE_CATCH_EVENT,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .intermediateCatchEvent(START_ELEMENT_ID)\n+ .message(b -> b.name(MESSAGE).zeebeCorrelationKeyExpression(\"correlationKey\"))\n+ .done(),\n+ Map.of(\"correlationKey\", \"value\")),\n+ new Scenario(\n+ BpmnElementType.INTERMEDIATE_THROW_EVENT,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .intermediateThrowEvent(START_ELEMENT_ID)\n+ .endEvent()\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.END_EVENT,\n+ Bpmn.createExecutableProcess(PROCESS_ID).startEvent().endEvent(START_ELEMENT_ID).done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.SERVICE_TASK,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .serviceTask(START_ELEMENT_ID, b -> b.zeebeJobType(JOBTYPE))\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.RECEIVE_TASK,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .receiveTask(START_ELEMENT_ID)\n+ .message(b -> b.name(MESSAGE).zeebeCorrelationKeyExpression(\"correlationKey\"))\n+ .done(),\n+ Map.of(\"correlationKey\", \"value\")),\n+ new Scenario(\n+ BpmnElementType.USER_TASK,\n+ Bpmn.createExecutableProcess(PROCESS_ID).startEvent().userTask(START_ELEMENT_ID).done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.MANUAL_TASK,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .manualTask(START_ELEMENT_ID)\n+ .endEvent()\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.EXCLUSIVE_GATEWAY,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .exclusiveGateway(START_ELEMENT_ID)\n+ .defaultFlow()\n+ .endEvent()\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.PARALLEL_GATEWAY,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .parallelGateway(START_ELEMENT_ID)\n+ .endEvent()\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.EVENT_BASED_GATEWAY,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .eventBasedGateway(START_ELEMENT_ID)\n+ .intermediateCatchEvent()\n+ .message(b -> b.name(MESSAGE).zeebeCorrelationKeyExpression(\"correlationKey\"))\n+ .moveToLastGateway()\n+ .intermediateCatchEvent()\n+ .timerWithDuration(\"PT1H\")\n+ .done(),\n+ Map.of(\"correlationKey\", \"value\")),\n+ new Scenario(\n+ BpmnElementType.MULTI_INSTANCE_BODY,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .serviceTask(\n+ START_ELEMENT_ID,\n+ t ->\n+ t.zeebeJobType(JOBTYPE)\n+ .multiInstance(m -> m.parallel().zeebeInputCollectionExpression(\"[1]\")))\n+ .endEvent()\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.CALL_ACTIVITY,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .callActivity(START_ELEMENT_ID, c -> c.zeebeProcessId(CHILD_PROCESS_ID))\n+ .endEvent()\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.BUSINESS_RULE_TASK,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .businessRuleTask(START_ELEMENT_ID, b -> b.zeebeJobType(JOBTYPE))\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.SCRIPT_TASK,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .scriptTask(START_ELEMENT_ID, b -> b.zeebeJobType(JOBTYPE))\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.SEND_TASK,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .sendTask(START_ELEMENT_ID, b -> b.zeebeJobType(JOBTYPE))\n+ .done(),\n+ Collections.emptyMap()));\n+ }\n+\n+ @Test\n+ public void testProcessInstanceCanStartAtElementType() {\n+ // given\n+ ENGINE.deployment().withXmlResource(scenario.modelInstance).deploy();\n+ if (scenario.type == BpmnElementType.CALL_ACTIVITY) {\n+ ENGINE.deployment().withXmlResource(getChildProcess()).deploy();\n+ }\n+\n+ // when\n+ final long instanceKey =\n+ ENGINE\n+ .processInstance()\n+ .ofBpmnProcessId(PROCESS_ID)\n+ .withStartInstruction(START_ELEMENT_ID)\n+ .withVariables(scenario.variables)\n+ .create();\n+\n+ // then\n+ assertThat(\n+ RecordingExporter.processInstanceRecords()\n+ .withProcessInstanceKey(instanceKey)\n+ .onlyEvents()\n+ .limit(\n+ r ->\n+ r.getValue().getBpmnElementType() == scenario.type\n+ && r.getIntent() == ProcessInstanceIntent.ELEMENT_ACTIVATED))\n+ .extracting(record -> record.getValue().getBpmnElementType(), Record::getIntent)\n+ .containsSequence(\n+ tuple(BpmnElementType.PROCESS, ProcessInstanceIntent.ELEMENT_ACTIVATING),\n+ tuple(BpmnElementType.PROCESS, ProcessInstanceIntent.ELEMENT_ACTIVATED),\n+ tuple(scenario.type, ProcessInstanceIntent.ELEMENT_ACTIVATING),\n+ tuple(scenario.type, ProcessInstanceIntent.ELEMENT_ACTIVATED));\n+ }\n+\n+ private BpmnModelInstance getChildProcess() {\n+ return Bpmn.createExecutableProcess(CHILD_PROCESS_ID).startEvent().endEvent().done();\n+ }\n+\n+ record Scenario(\n+ BpmnElementType type, BpmnModelInstance modelInstance, Map<String, Object> variables) {}\n+}\n"]
1
["a5ecfdf49b0d4c43fbbbf7947be7c0327ccb3415"]
["test"]
remove unused branches and ignore envrc file
["diff --git a/.github/workflows/ibis-backends-cloud.yml b/.github/workflows/ibis-backends-cloud.yml\nindex 2003e8e..7c7fd26 100644\n--- a/.github/workflows/ibis-backends-cloud.yml\n+++ b/.github/workflows/ibis-backends-cloud.yml\n@@ -5,9 +5,12 @@ on:\n # Skip the backend suite if all changes are in the docs directory\n paths-ignore:\n - \"docs/**\"\n+ - \"**/*.md\"\n+ - \"**/*.qmd\"\n+ - \"codecov.yml\"\n+ - \".envrc\"\n branches:\n - master\n- - quarto\n \n permissions:\n # this allows extractions/setup-just to list releases for `just` at a higher\ndiff --git a/.github/workflows/ibis-backends-skip-helper.yml b/.github/workflows/ibis-backends-skip-helper.yml\nindex 5d5f3f7..0471994 100644\n--- a/.github/workflows/ibis-backends-skip-helper.yml\n+++ b/.github/workflows/ibis-backends-skip-helper.yml\n@@ -9,20 +9,20 @@ on:\n - \"**/*.md\"\n - \"**/*.qmd\"\n - \"codecov.yml\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n pull_request:\n paths:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n - \"codecov.yml\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n merge_group:\n jobs:\n test_backends:\ndiff --git a/.github/workflows/ibis-backends.yml b/.github/workflows/ibis-backends.yml\nindex 4a1cae9..30e6c1a 100644\n--- a/.github/workflows/ibis-backends.yml\n+++ b/.github/workflows/ibis-backends.yml\n@@ -8,10 +8,10 @@ on:\n - \"**/*.md\"\n - \"**/*.qmd\"\n - \"codecov.yml\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n pull_request:\n # Skip the backend suite if all changes are docs\n paths-ignore:\n@@ -19,10 +19,10 @@ on:\n - \"**/*.md\"\n - \"**/*.qmd\"\n - \"codecov.yml\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n merge_group:\n \n permissions:\ndiff --git a/.github/workflows/ibis-docs-lint.yml b/.github/workflows/ibis-docs-lint.yml\nindex 1adda11..b528a30 100644\n--- a/.github/workflows/ibis-docs-lint.yml\n+++ b/.github/workflows/ibis-docs-lint.yml\n@@ -5,12 +5,10 @@ on:\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n pull_request:\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n merge_group:\n \n concurrency:\ndiff --git a/.github/workflows/ibis-main-skip-helper.yml b/.github/workflows/ibis-main-skip-helper.yml\nindex a5fdc6f..0fb5dea 100644\n--- a/.github/workflows/ibis-main-skip-helper.yml\n+++ b/.github/workflows/ibis-main-skip-helper.yml\n@@ -8,19 +8,19 @@ on:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n pull_request:\n paths:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n merge_group:\n jobs:\n test_core:\ndiff --git a/.github/workflows/ibis-main.yml b/.github/workflows/ibis-main.yml\nindex aa31436..0b1536a 100644\n--- a/.github/workflows/ibis-main.yml\n+++ b/.github/workflows/ibis-main.yml\n@@ -7,20 +7,20 @@ on:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n pull_request:\n # Skip the test suite if all changes are in the docs directory\n paths-ignore:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n merge_group:\n \n permissions:\ndiff --git a/.github/workflows/ibis-tpch-queries-skip-helper.yml b/.github/workflows/ibis-tpch-queries-skip-helper.yml\nindex 1f1c0bc..f10fb8d 100644\n--- a/.github/workflows/ibis-tpch-queries-skip-helper.yml\n+++ b/.github/workflows/ibis-tpch-queries-skip-helper.yml\n@@ -6,19 +6,19 @@ on:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n pull_request:\n paths:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n merge_group:\n \n concurrency:\ndiff --git a/.github/workflows/ibis-tpch-queries.yml b/.github/workflows/ibis-tpch-queries.yml\nindex b4f8a48..9e65a61 100644\n--- a/.github/workflows/ibis-tpch-queries.yml\n+++ b/.github/workflows/ibis-tpch-queries.yml\n@@ -6,19 +6,19 @@ on:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n pull_request:\n paths-ignore:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n merge_group:\n \n concurrency:\ndiff --git a/.github/workflows/nix-skip-helper.yml b/.github/workflows/nix-skip-helper.yml\nindex 677b4d7..e0ab8f7 100644\n--- a/.github/workflows/nix-skip-helper.yml\n+++ b/.github/workflows/nix-skip-helper.yml\n@@ -9,19 +9,19 @@ on:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n pull_request:\n paths:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n merge_group:\n \n jobs:\ndiff --git a/.github/workflows/nix.yml b/.github/workflows/nix.yml\nindex f2dd3f0..7ea9e26 100644\n--- a/.github/workflows/nix.yml\n+++ b/.github/workflows/nix.yml\n@@ -6,19 +6,19 @@ on:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n pull_request:\n paths-ignore:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n merge_group:\n \n concurrency:\n"]
1
["d0c6476df61b9c6ab07b87e1724ea7c5318595bb"]
["ci"]
fix error spacing
["diff --git a/cmd/infracost/main.go b/cmd/infracost/main.go\nindex 425aef1..fcc9eb5 100644\n--- a/cmd/infracost/main.go\n+++ b/cmd/infracost/main.go\n@@ -149,7 +149,7 @@ Example:\n \t\t\t}\n \n \t\t\tif appErr.Error() != \"\" {\n-\t\t\t\tfmt.Fprintf(os.Stderr, \"\\n%s\\n\", color.HiRedString(appErr.Error()))\n+\t\t\t\tfmt.Fprintf(os.Stderr, \"%s\\n\", color.HiRedString(appErr.Error()))\n \t\t\t}\n \t\t}\n \n"]
1
["c623b3622058b913290120b06ccdc779a4e4413d"]
["fix"]
post installers compatiblity with Windows #2520
["diff --git a/packages/cubejs-databricks-jdbc-driver/package.json b/packages/cubejs-databricks-jdbc-driver/package.json\nindex cc164f0..fd7ad45 100644\n--- a/packages/cubejs-databricks-jdbc-driver/package.json\n+++ b/packages/cubejs-databricks-jdbc-driver/package.json\n@@ -14,13 +14,16 @@\n },\n \"main\": \"dist/src/index.js\",\n \"typings\": \"dist/src/index.d.ts\",\n+ \"bin\": {\n+ \"databricks-jdbc-installer\": \"bin/post-install\"\n+ },\n \"scripts\": {\n \"build\": \"rm -rf dist && npm run tsc\",\n \"tsc\": \"tsc\",\n \"watch\": \"tsc -w\",\n \"lint\": \"eslint src/* --ext .ts\",\n \"lint:fix\": \"eslint --fix src/* --ext .ts\",\n- \"postinstall\": \"bin/post-install\"\n+ \"postinstall\": \"databricks-jdbc-installer\"\n },\n \"files\": [\n \"README.md\",\ndiff --git a/rust/package.json b/rust/package.json\nindex b139279..5bf6446 100644\n--- a/rust/package.json\n+++ b/rust/package.json\n@@ -8,7 +8,8 @@\n \"node\": \">=10.8.0\"\n },\n \"bin\": {\n- \"cubestore-dev\": \"bin/cubestore-dev\"\n+ \"cubestore-dev\": \"bin/cubestore-dev\",\n+ \"cubestore-installer\": \"bin/post-install\"\n },\n \"scripts\": {\n \"build\": \"rm -rf dist && npm run tsc\",\n@@ -18,7 +19,7 @@\n \"lint:fix\": \"eslint --fix js-wrapper/* --ext .ts,js\",\n \"unit\": \"jest\",\n \"unit:debug\": \"jest --runInBand\",\n- \"postinstall\": \"bin/post-install\"\n+ \"postinstall\": \"cubestore-installer\"\n },\n \"files\": [\n \"dist\",\ndiff --git a/yarn.lock b/yarn.lock\nindex d2a4038..b59bb77 100644\n--- a/yarn.lock\n+++ b/yarn.lock\n@@ -4036,9 +4036,9 @@\n integrity sha512-7btbphLrKvo5yl/5CC2OCxUSMx1wV1wvGT1qDXkSt7yi00/YW7E8k6qzXqJHsp+WU0eoG7r6MTQQXI9lIvd0qA==\n \n \"@types/fs-extra@^9.0.1\", \"@types/fs-extra@^9.0.2\", \"@types/fs-extra@^9.0.8\":\n- version \"9.0.10\"\n- resolved \"https://registry.yarnpkg.com/@types/fs-extra/-/fs-extra-9.0.10.tgz#8023a72e3d06cf54929ea47ec7634e47f33f4046\"\n- integrity sha512-O9T2LLkRDiTlalOBdjEkcnT0MRdT2+wglCl7pJUJ3mkWkR8hX4K+5bg2raQNJcLv4V8zGuTXe7Ud3wSqkTyuyQ==\n+ version \"9.0.11\"\n+ resolved \"https://registry.yarnpkg.com/@types/fs-extra/-/fs-extra-9.0.11.tgz#8cc99e103499eab9f347dbc6ca4e99fb8d2c2b87\"\n+ integrity sha512-mZsifGG4QeQ7hlkhO56u7zt/ycBgGxSVsFI/6lGTU34VtwkiqrrSDgw0+ygs8kFGWcXnFQWMrzF2h7TtDFNixA==\n dependencies:\n \"@types/node\" \"*\"\n \n@@ -5306,9 +5306,9 @@ acorn@^7.0.0, acorn@^7.1.0, acorn@^7.1.1, acorn@^7.4.0:\n integrity sha512-nQyp0o1/mNdbTO1PO6kHkwSrmgZ0MT/jCCpNiwbUjGoRN4dlBhqJtoQuCnEOKzgTVwg0ZWiCoQy6SxMebQVh8A==\n \n acorn@^8.1.0:\n- version \"8.1.0\"\n- resolved \"https://registry.yarnpkg.com/acorn/-/acorn-8.1.0.tgz#52311fd7037ae119cbb134309e901aa46295b3fe\"\n- integrity sha512-LWCF/Wn0nfHOmJ9rzQApGnxnvgfROzGilS8936rqN/lfcYkY9MYZzdMqN+2NJ4SlTc+m5HiSa+kNfDtI64dwUA==\n+ version \"8.1.1\"\n+ resolved \"https://registry.yarnpkg.com/acorn/-/acorn-8.1.1.tgz#fb0026885b9ac9f48bac1e185e4af472971149ff\"\n+ integrity sha512-xYiIVjNuqtKXMxlRMDc6mZUhXehod4a3gbZ1qRlM7icK4EbxUFNLhWoPblCvFtB2Y9CIqHP3CF/rdxLItaQv8g==\n \n adal-node@^0.1.28:\n version \"0.1.28\"\n@@ -5441,9 +5441,9 @@ ajv@^6.1.0, ajv@^6.10.0, ajv@^6.10.2, ajv@^6.12.2, ajv@^6.12.3, ajv@^6.12.4, ajv\n uri-js \"^4.2.2\"\n \n ajv@^8.0.1:\n- version \"8.0.5\"\n- resolved \"https://registry.yarnpkg.com/ajv/-/ajv-8.0.5.tgz#f07d6fdeffcdbb80485570ce3f1bc845fcc812b9\"\n- integrity sha512-RkiLa/AeJx7+9OvniQ/qeWu0w74A8DiPPBclQ6ji3ZQkv5KamO+QGpqmi7O4JIw3rHGUXZ6CoP9tsAkn3gyazg==\n+ version \"8.1.0\"\n+ resolved \"https://registry.yarnpkg.com/ajv/-/ajv-8.1.0.tgz#45d5d3d36c7cdd808930cc3e603cf6200dbeb736\"\n+ integrity sha512-B/Sk2Ix7A36fs/ZkuGLIR86EdjbgR6fsAcbx9lOP/QBSXujDNbVmIS/U4Itz5k8fPFDeVZl/zQ/gJW4Jrq6XjQ==\n dependencies:\n fast-deep-equal \"^3.1.1\"\n json-schema-traverse \"^1.0.0\"\n@@ -6828,15 +6828,15 @@ [email protected]:\n node-releases \"^1.1.61\"\n \n browserslist@^4.0.0, browserslist@^4.11.1, browserslist@^4.12.0, browserslist@^4.14.5, browserslist@^4.16.3, browserslist@^4.3.4, browserslist@^4.6.2, browserslist@^4.6.4, browserslist@^4.7.0, browserslist@^4.9.1:\n- version \"4.16.3\"\n- resolved \"https://registry.yarnpkg.com/browserslist/-/browserslist-4.16.3.tgz#340aa46940d7db878748567c5dea24a48ddf3717\"\n- integrity sha512-vIyhWmIkULaq04Gt93txdh+j02yX/JzlyhLYbV3YQCn/zvES3JnY7TifHHvvr1w5hTDluNKMkV05cs4vy8Q7sw==\n+ version \"4.16.4\"\n+ resolved \"https://registry.yarnpkg.com/browserslist/-/browserslist-4.16.4.tgz#7ebf913487f40caf4637b892b268069951c35d58\"\n+ integrity sha512-d7rCxYV8I9kj41RH8UKYnvDYCRENUlHRgyXy/Rhr/1BaeLGfiCptEdFE8MIrvGfWbBFNjVYx76SQWvNX1j+/cQ==\n dependencies:\n- caniuse-lite \"^1.0.30001181\"\n- colorette \"^1.2.1\"\n- electron-to-chromium \"^1.3.649\"\n+ caniuse-lite \"^1.0.30001208\"\n+ colorette \"^1.2.2\"\n+ electron-to-chromium \"^1.3.712\"\n escalade \"^3.1.1\"\n- node-releases \"^1.1.70\"\n+ node-releases \"^1.1.71\"\n \n [email protected]:\n version \"0.2.6\"\n@@ -7217,7 +7217,7 @@ caniuse-api@^3.0.0:\n lodash.memoize \"^4.1.2\"\n lodash.uniq \"^4.5.0\"\n \n-caniuse-lite@^1.0.0, caniuse-lite@^1.0.30000981, caniuse-lite@^1.0.30001032, caniuse-lite@^1.0.30001061, caniuse-lite@^1.0.30001109, caniuse-lite@^1.0.30001125, caniuse-lite@^1.0.30001181:\n+caniuse-lite@^1.0.0, caniuse-lite@^1.0.30000981, caniuse-lite@^1.0.30001032, caniuse-lite@^1.0.30001061, caniuse-lite@^1.0.30001109, caniuse-lite@^1.0.30001125, caniuse-lite@^1.0.30001208:\n version \"1.0.30001208\"\n resolved \"https://registry.yarnpkg.com/caniuse-lite/-/caniuse-lite-1.0.30001208.tgz#a999014a35cebd4f98c405930a057a0d75352eb9\"\n integrity sha512-OE5UE4+nBOro8Dyvv0lfx+SRtfVIOM9uhKqFmJeUbGriqhhStgp1A0OyBpgy3OUF8AhYCT+PVwPC1gMl2ZcQMA==\n@@ -9549,10 +9549,10 @@ ejs@^2.6.1:\n resolved \"https://registry.yarnpkg.com/ejs/-/ejs-2.7.4.tgz#48661287573dcc53e366c7a1ae52c3a120eec9ba\"\n integrity sha512-7vmuyh5+kuUyJKePhQfRQBhXV5Ce+RnaeeQArKu1EAMpL3WbgMt5WG6uQZpEVvYSSsxMXRKOewtDk9RaTKXRlA==\n \n-electron-to-chromium@^1.3.564, electron-to-chromium@^1.3.649:\n- version \"1.3.711\"\n- resolved \"https://registry.yarnpkg.com/electron-to-chromium/-/electron-to-chromium-1.3.711.tgz#92c3caf7ffed5e18bf63f66b4b57b4db2409c450\"\n- integrity sha512-XbklBVCDiUeho0PZQCjC25Ha6uBwqqJeyDhPLwLwfWRAo4x+FZFsmu1pPPkXT+B4MQMQoQULfyaMltDopfeiHQ==\n+electron-to-chromium@^1.3.564, electron-to-chromium@^1.3.712:\n+ version \"1.3.712\"\n+ resolved \"https://registry.yarnpkg.com/electron-to-chromium/-/electron-to-chromium-1.3.712.tgz#ae467ffe5f95961c6d41ceefe858fc36eb53b38f\"\n+ integrity sha512-3kRVibBeCM4vsgoHHGKHmPocLqtFAGTrebXxxtgKs87hNUzXrX2NuS3jnBys7IozCnw7viQlozxKkmty2KNfrw==\n \n elegant-spinner@^1.0.1:\n version \"1.0.1\"\n@@ -9945,9 +9945,9 @@ eslint-plugin-import@^2.16.0, eslint-plugin-import@^2.18.2, eslint-plugin-import\n tsconfig-paths \"^3.9.0\"\n \n eslint-plugin-jest@^24.1.0:\n- version \"24.3.4\"\n- resolved \"https://registry.yarnpkg.com/eslint-plugin-jest/-/eslint-plugin-jest-24.3.4.tgz#6d90c3554de0302e879603dd6405474c98849f19\"\n- integrity sha512-3n5oY1+fictanuFkTWPwSlehugBTAgwLnYLFsCllzE3Pl1BwywHl5fL0HFxmMjoQY8xhUDk8uAWc3S4JOHGh3A==\n+ version \"24.3.5\"\n+ resolved \"https://registry.yarnpkg.com/eslint-plugin-jest/-/eslint-plugin-jest-24.3.5.tgz#71f0b580f87915695c286c3f0eb88cf23664d044\"\n+ integrity sha512-XG4rtxYDuJykuqhsOqokYIR84/C8pRihRtEpVskYLbIIKGwPNW2ySxdctuVzETZE+MbF/e7wmsnbNVpzM0rDug==\n dependencies:\n \"@typescript-eslint/experimental-utils\" \"^4.0.1\"\n \n@@ -12140,12 +12140,11 @@ [email protected]:\n micromatch \"^3.1.10\"\n \n http-proxy-middleware@^1.0.0:\n- version \"1.1.0\"\n- resolved \"https://registry.yarnpkg.com/http-proxy-middleware/-/http-proxy-middleware-1.1.0.tgz#b896b2cc6836019af4a4f2d5f7b21b99c77ea13f\"\n- integrity sha512-OnjU5vyVgcZVe2AjLJyMrk8YLNOC2lspCHirB5ldM+B/dwEfZ5bgVTrFyzE9R7xRWAP/i/FXtvIqKjTNEZBhBg==\n+ version \"1.1.1\"\n+ resolved \"https://registry.yarnpkg.com/http-proxy-middleware/-/http-proxy-middleware-1.1.1.tgz#48900a68cd9d388c735d1dd97302c919b7e94a13\"\n+ integrity sha512-FIDg9zPvOwMhQ3XKB2+vdxK6WWbVAH7s5QpqQCif7a1TNL76GNAATWA1sy6q2gSfss8UJ/Nwza3N6QnFkKclpA==\n dependencies:\n \"@types/http-proxy\" \"^1.17.5\"\n- camelcase \"^6.2.0\"\n http-proxy \"^1.18.1\"\n is-glob \"^4.0.1\"\n is-plain-obj \"^3.0.0\"\n@@ -14341,9 +14340,9 @@ jsdom@^15.2.1:\n xml-name-validator \"^3.0.0\"\n \n jsdom@^16.4.0:\n- version \"16.5.2\"\n- resolved \"https://registry.yarnpkg.com/jsdom/-/jsdom-16.5.2.tgz#583fac89a0aea31dbf6237e7e4bedccd9beab472\"\n- integrity sha512-JxNtPt9C1ut85boCbJmffaQ06NBnzkQY/MWO3YxPW8IWS38A26z+B1oBvA9LwKrytewdfymnhi4UNH3/RAgZrg==\n+ version \"16.5.3\"\n+ resolved \"https://registry.yarnpkg.com/jsdom/-/jsdom-16.5.3.tgz#13a755b3950eb938b4482c407238ddf16f0d2136\"\n+ integrity sha512-Qj1H+PEvUsOtdPJ056ewXM4UJPCi4hhLA8wpiz9F2YvsRBhuFsXxtrIFAgGBDynQA9isAMGE91PfUYbdMPXuTA==\n dependencies:\n abab \"^2.0.5\"\n acorn \"^8.1.0\"\n@@ -15590,12 +15589,12 @@ micromatch@^3.1.10, micromatch@^3.1.4:\n to-regex \"^3.0.2\"\n \n micromatch@^4.0.2:\n- version \"4.0.3\"\n- resolved \"https://registry.yarnpkg.com/micromatch/-/micromatch-4.0.3.tgz#fdad8352bf0cbeb89b391b5d244bc22ff3dd4ec8\"\n- integrity sha512-ueuSaP4i67F/FAUac9zzZ0Dz/5KeKDkITYIS/k4fps+9qeh1SkeH6gbljcqz97mNBOsaWZ+iv2UobMKK/yD+aw==\n+ version \"4.0.4\"\n+ resolved \"https://registry.yarnpkg.com/micromatch/-/micromatch-4.0.4.tgz#896d519dfe9db25fce94ceb7a500919bf881ebf9\"\n+ integrity sha512-pRmzw/XUcwXGpD9aI9q/0XOwLNygjETJ8y0ao0wdqprrzDa4YnxLcz7fQRZr8voh8V10kGhABbNcHVk5wHgWwg==\n dependencies:\n braces \"^3.0.1\"\n- picomatch \"^2.2.1\"\n+ picomatch \"^2.2.3\"\n \n miller-rabin@^4.0.0:\n version \"4.0.1\"\n@@ -16356,7 +16355,7 @@ node-pre-gyp@^0.11.0:\n semver \"^5.3.0\"\n tar \"^4\"\n \n-node-releases@^1.1.61, node-releases@^1.1.70:\n+node-releases@^1.1.61, node-releases@^1.1.71:\n version \"1.1.71\"\n resolved \"https://registry.yarnpkg.com/node-releases/-/node-releases-1.1.71.tgz#cb1334b179896b1c89ecfdd4b725fb7bbdfc7dbb\"\n integrity sha512-zR6HoT6LrLCRBwukmrVbHv0EpEQjksO6GmFcZQQuCAy139BEsoVKPYnf3jongYW83fAa1torLGYwxxky/p28sg==\n@@ -17571,10 +17570,10 @@ [email protected]:\n dependencies:\n split2 \"^3.1.1\"\n \n-picomatch@^2.0.4, picomatch@^2.2.1, picomatch@^2.2.2:\n- version \"2.2.2\"\n- resolved \"https://registry.yarnpkg.com/picomatch/-/picomatch-2.2.2.tgz#21f333e9b6b8eaff02468f5146ea406d345f4dad\"\n- integrity sha512-q0M/9eZHzmr0AulXyPwNfZjtwZ/RBZlbN3K3CErVrk50T2ASYI7Bye0EvekFY3IP1Nt2DHu0re+V2ZHIpMkuWg==\n+picomatch@^2.0.4, picomatch@^2.2.1, picomatch@^2.2.2, picomatch@^2.2.3:\n+ version \"2.2.3\"\n+ resolved \"https://registry.yarnpkg.com/picomatch/-/picomatch-2.2.3.tgz#465547f359ccc206d3c48e46a1bcb89bf7ee619d\"\n+ integrity sha512-KpELjfwcCDUb9PeigTs2mBJzXUPzAuP2oPcA989He8Rte0+YUAjw1JVedDhuTKPkHjSYzMN3npC9luThGYEKdg==\n \n pify@^2.0.0, pify@^2.2.0, pify@^2.3.0:\n version \"2.3.0\"\n@@ -18446,9 +18445,9 @@ postcss@^7, postcss@^7.0.0, postcss@^7.0.1, postcss@^7.0.14, postcss@^7.0.17, po\n supports-color \"^6.1.0\"\n \n postcss@^8.1.0, postcss@^8.2.8:\n- version \"8.2.9\"\n- resolved \"https://registry.yarnpkg.com/postcss/-/postcss-8.2.9.tgz#fd95ff37b5cee55c409b3fdd237296ab4096fba3\"\n- integrity sha512-b+TmuIL4jGtCHtoLi+G/PisuIl9avxs8IZMSmlABRwNz5RLUUACrC+ws81dcomz1nRezm5YPdXiMEzBEKgYn+Q==\n+ version \"8.2.10\"\n+ resolved \"https://registry.yarnpkg.com/postcss/-/postcss-8.2.10.tgz#ca7a042aa8aff494b334d0ff3e9e77079f6f702b\"\n+ integrity sha512-b/h7CPV7QEdrqIxtAf2j31U5ef05uBDuvoXv6L51Q4rcS1jdlXAVKJv+atCFdUXYl9dyTHGyoMzIepwowRJjFw==\n dependencies:\n colorette \"^1.2.2\"\n nanoid \"^3.1.22\"\n@@ -19318,9 +19317,9 @@ rc-tree@^4.0.0, rc-tree@~4.1.0:\n rc-virtual-list \"^3.0.1\"\n \n rc-trigger@^5.0.0, rc-trigger@^5.0.4, rc-trigger@^5.1.2, rc-trigger@^5.2.1:\n- version \"5.2.3\"\n- resolved \"https://registry.yarnpkg.com/rc-trigger/-/rc-trigger-5.2.3.tgz#8c55046ab432d7b52d51c69afb57ebb5bbe37e17\"\n- integrity sha512-6Fokao07HUbqKIDkDRFEM0AGZvsvK0Fbp8A/KFgl1ngaqfO1nY037cISCG1Jm5fxImVsXp9awdkP7Vu5cxjjog==\n+ version \"5.2.4\"\n+ resolved \"https://registry.yarnpkg.com/rc-trigger/-/rc-trigger-5.2.4.tgz#f1cca4a6c1f378a5d6fadec010292250772069d3\"\n+ integrity sha512-nLZa4XYo3hOAVauQr7HsGrBtE8/pyoIWhHZnpr7x/H/dd6pPeRzH0//+1TzaBAXylbFgsY6hogKAMeJwaKeDFw==\n dependencies:\n \"@babel/runtime\" \"^7.11.2\"\n classnames \"^2.2.6\"\n@@ -20516,9 +20515,9 @@ rollup@^1.31.1:\n acorn \"^7.1.0\"\n \n rollup@^2.40.0, rollup@^2.8.0:\n- version \"2.45.0\"\n- resolved \"https://registry.yarnpkg.com/rollup/-/rollup-2.45.0.tgz#bfcce2347c96f15f5c78ac860bc38e3349ba27c9\"\n- integrity sha512-JJznbtGIsHZfKH0Sa9RpCAy5JarH8SWvBzRAGuRkgzAafb8e8D7VSMJ0O1Bsix1nn91koN/Ecvl2+ZWhljcuTw==\n+ version \"2.45.1\"\n+ resolved \"https://registry.yarnpkg.com/rollup/-/rollup-2.45.1.tgz#eae2b94dc2088b4e0a3b7197a5a1ee0bdd589d5c\"\n+ integrity sha512-vPD+JoDj3CY8k6m1bLcAFttXMe78P4CMxoau0iLVS60+S9kLsv2379xaGy4NgYWu+h2WTlucpoLPAoUoixFBag==\n optionalDependencies:\n fsevents \"~2.3.1\"\n \n@@ -22971,9 +22970,9 @@ typescript@~4.1.5:\n integrity sha512-6OSu9PTIzmn9TCDiovULTnET6BgXtDYL4Gg4szY+cGsc3JP1dQL8qvE8kShTRx1NIw4Q9IBHlwODjkjWEtMUyA==\n \n ua-parser-js@^0.7.18:\n- version \"0.7.27\"\n- resolved \"https://registry.yarnpkg.com/ua-parser-js/-/ua-parser-js-0.7.27.tgz#b54f8ce9eb6c7abf3584edeaf9a3d8b3bd92edba\"\n- integrity sha512-eXMaRYK2skomGocoX0x9sBXzx5A1ZVQgXfrW4mTc8dT0zS7olEcyfudAzRC5tIIRgLxQ69B6jut3DI+n5hslPA==\n+ version \"0.7.28\"\n+ resolved \"https://registry.yarnpkg.com/ua-parser-js/-/ua-parser-js-0.7.28.tgz#8ba04e653f35ce210239c64661685bf9121dec31\"\n+ integrity sha512-6Gurc1n//gjp9eQNXjD9O3M/sMwVtN5S8Lv9bvOYBfKfDNiIIhqiyi01vMBO45u4zkDE420w/e0se7Vs+sIg+g==\n \n [email protected]:\n version \"3.4.10\"\n@@ -22984,9 +22983,9 @@ [email protected]:\n source-map \"~0.6.1\"\n \n uglify-js@^3.1.4, uglify-js@^3.4.9:\n- version \"3.13.3\"\n- resolved \"https://registry.yarnpkg.com/uglify-js/-/uglify-js-3.13.3.tgz#ce72a1ad154348ea2af61f50933c76cc8802276e\"\n- integrity sha512-otIc7O9LyxpUcQoXzj2hL4LPWKklO6LJWoJUzNa8A17Xgi4fOeDC8FBDOLHnC/Slo1CQgsZMcM6as0M76BZaig==\n+ version \"3.13.4\"\n+ resolved \"https://registry.yarnpkg.com/uglify-js/-/uglify-js-3.13.4.tgz#592588bb9f47ae03b24916e2471218d914955574\"\n+ integrity sha512-kv7fCkIXyQIilD5/yQy8O+uagsYIOt5cZvs890W40/e/rvjMSzJw81o9Bg0tkURxzZBROtDQhW2LFjOGoK3RZw==\n \n [email protected]:\n version \"0.0.6\"\n"]
1
["7e9bd7c86df1032d53e752654fe4a446951480bb"]
["build"]
Deploy utilities from correct folder Signed-off-by: rjshrjndrn <[email protected]>
["diff --git a/.github/workflows/utilities.yaml b/.github/workflows/utilities.yaml\nindex 92e130c..afbc850 100644\n--- a/.github/workflows/utilities.yaml\n+++ b/.github/workflows/utilities.yaml\n@@ -43,7 +43,7 @@ jobs:\n PUSH_IMAGE=1 bash build.sh\n - name: Deploy to kubernetes\n run: |\n- cd scripts/helm/\n+ cd scripts/helmcharts/\n sed -i \"s#openReplayContainerRegistry.*#openReplayContainerRegistry: \\\"${{ secrets.OSS_REGISTRY_URL }}\\\"#g\" vars.yaml\n sed -i \"s#minio_access_key.*#minio_access_key: \\\"${{ secrets.OSS_MINIO_ACCESS_KEY }}\\\" #g\" vars.yaml\n sed -i \"s#minio_secret_key.*#minio_secret_key: \\\"${{ secrets.OSS_MINIO_SECRET_KEY }}\\\" #g\" vars.yaml\n"]
1
["2ebf04099353ef70395b8c8f5e130f70e1ed0814"]
["ci"]
updated to lerna v6
["diff --git a/lerna.json b/lerna.json\nindex 29cb6ed..4488fb6 100644\n--- a/lerna.json\n+++ b/lerna.json\n@@ -25,7 +25,6 @@\n \"npmClient\": \"pnpm\",\n \"useWorkspaces\": true,\n \"conventionalCommits\": true,\n- \"useNx\": true,\n \"command\": {\n \"version\": {\n \"message\": \"chore(release): published new version\"\ndiff --git a/nx.json b/nx.json\nindex 00997ba..b3f2627 100644\n--- a/nx.json\n+++ b/nx.json\n@@ -33,32 +33,6 @@\n }\n }\n },\n- \"targetDependencies\": {\n- \"build\": [\n- {\n- \"target\": \"build\",\n- \"projects\": \"dependencies\"\n- }\n- ],\n- \"build:ci\": [\n- {\n- \"target\": \"build:ci\",\n- \"projects\": \"dependencies\"\n- }\n- ],\n- \"prepare\": [\n- {\n- \"target\": \"prepare\",\n- \"projects\": \"dependencies\"\n- }\n- ],\n- \"package\": [\n- {\n- \"target\": \"package\",\n- \"projects\": \"dependencies\"\n- }\n- ]\n- },\n \"affected\": {\n \"defaultBase\": \"main\"\n },\n@@ -66,5 +40,28 @@\n \"@nrwl/js\": {\n \"analyzeSourceFiles\": false\n }\n+ },\n+ \"$schema\": \"./node_modules/nx/schemas/nx-schema.json\",\n+ \"targetDefaults\": {\n+ \"build\": {\n+ \"dependsOn\": [\n+ \"^build\"\n+ ]\n+ },\n+ \"build:ci\": {\n+ \"dependsOn\": [\n+ \"^build:ci\"\n+ ]\n+ },\n+ \"prepare\": {\n+ \"dependsOn\": [\n+ \"^prepare\"\n+ ]\n+ },\n+ \"package\": {\n+ \"dependsOn\": [\n+ \"^package\"\n+ ]\n+ }\n }\n }\n"]
1
["aa30370b796c1f46cc60aa56b2918c3d27c0cf88"]
["build"]
add --ignore-existing to all npx commands
["diff --git a/docs/getting-started/getting-started.md b/docs/getting-started/getting-started.md\nindex dc6db37..3ef9d0a 100644\n--- a/docs/getting-started/getting-started.md\n+++ b/docs/getting-started/getting-started.md\n@@ -13,7 +13,7 @@ npm install -g @angular/cli\n **Using `npx`**\n \n ```bash\n-npx create-nx-workspace myworkspace\n+npx --ignore-existing create-nx-workspace myworkspace\n ```\n \n **Using `npm init`**\ndiff --git a/docs/guides/react-and-angular.md b/docs/guides/react-and-angular.md\nindex c1929a2..a5651ff 100644\n--- a/docs/guides/react-and-angular.md\n+++ b/docs/guides/react-and-angular.md\n@@ -11,7 +11,7 @@ To show how Nx does it, let's build two applications (one in Angular, and one in\n Let's start by creating a new Nx workspace. The easiest way to do this is to use npx.\n \n ```bash\n-npx create-nx-workspace happynrwl --preset=empty\n+npx --ignore-existing create-nx-workspace happynrwl --preset=empty\n ```\n \n ## Creating an Angular Application\ndiff --git a/docs/guides/react.md b/docs/guides/react.md\nindex e1647fd..eac848e 100644\n--- a/docs/guides/react.md\n+++ b/docs/guides/react.md\n@@ -16,13 +16,13 @@ Nx has first class support for React: you can create React applications and libr\n Create a new Nx workspace. The easiest way to do it is to use npx.\n \n ```bash\n-npx create-nx-workspace happynrwl --preset=empty\n+npx --ignore-existing create-nx-workspace happynrwl --preset=empty\n ```\n \n You can also create a workspace with a React application in place by running:\n \n ```bash\n-npx create-nx-workspace happynrwl --preset=react\n+npx --ignore-existing create-nx-workspace happynrwl --preset=react\n ```\n \n ## Generating a React Application\ndiff --git a/docs/tutorial/01-create-application.md b/docs/tutorial/01-create-application.md\nindex ea87ecf..967a56e 100644\n--- a/docs/tutorial/01-create-application.md\n+++ b/docs/tutorial/01-create-application.md\n@@ -7,7 +7,7 @@ In this tutorial you will use Nx to build a full-stack application out of common\n **Start by creating a new workspace.**\n \n ```bash\n-npx create-nx-workspace myorg\n+npx --ignore-existing create-nx-workspace myorg\n ```\n \n When asked about 'preset', select `empty`.\n"]
1
["fc9af4d0b93d69be4e201ffb18da04324e8a4a87"]
["docs"]
remove unnecessary import
["diff --git a/transport/src/main/java/io/camunda/zeebe/transport/stream/impl/LogicalId.java b/transport/src/main/java/io/camunda/zeebe/transport/stream/impl/LogicalId.java\nindex 14c6f30..ebaef60 100644\n--- a/transport/src/main/java/io/camunda/zeebe/transport/stream/impl/LogicalId.java\n+++ b/transport/src/main/java/io/camunda/zeebe/transport/stream/impl/LogicalId.java\n@@ -8,7 +8,6 @@\n package io.camunda.zeebe.transport.stream.impl;\n \n import io.camunda.zeebe.util.buffer.BufferUtil;\n-import org.agrona.BitUtil;\n import org.agrona.concurrent.UnsafeBuffer;\n \n /**\n"]
1
["84529bcb10c6fe02e2c0079d069ab6c6ac7683d6"]
["refactor"]
expose the means by which we process each reward cycle's affirmation maps at reward cycle boundaries
["diff --git a/src/burnchains/burnchain.rs b/src/burnchains/burnchain.rs\nindex 92105d6..60c608a 100644\n--- a/src/burnchains/burnchain.rs\n+++ b/src/burnchains/burnchain.rs\n@@ -851,8 +851,26 @@ impl Burnchain {\n );\n \n burnchain_db.store_new_burnchain_block(burnchain, indexer, &block)?;\n- let block_height = block.block_height();\n+ Burnchain::process_affirmation_maps(\n+ burnchain,\n+ burnchain_db,\n+ indexer,\n+ block.block_height(),\n+ )?;\n+\n+ let header = block.header();\n+ Ok(header)\n+ }\n \n+ /// Update the affirmation maps for the previous reward cycle's commits.\n+ /// This is a no-op unless the given burnchain block height falls on a reward cycle boundary. In that\n+ /// case, the previous reward cycle's block commits' affirmation maps are all re-calculated.\n+ pub fn process_affirmation_maps<B: BurnchainHeaderReader>(\n+ burnchain: &Burnchain,\n+ burnchain_db: &mut BurnchainDB,\n+ indexer: &B,\n+ block_height: u64,\n+ ) -> Result<(), burnchain_error> {\n let this_reward_cycle = burnchain\n .block_height_to_reward_cycle(block_height)\n .unwrap_or(0);\n@@ -872,10 +890,7 @@ impl Burnchain {\n );\n update_pox_affirmation_maps(burnchain_db, indexer, prev_reward_cycle, burnchain)?;\n }\n-\n- let header = block.header();\n-\n- Ok(header)\n+ Ok(())\n }\n \n /// Hand off the block to the ChainsCoordinator _and_ process the sortition\n"]
1
["d7972da833257c073403dec3c2ac3a7f297e328a"]
["refactor"]
add page balckwhitelist and pdf
["diff --git a/src/_locales/common/messages.json b/src/_locales/common/messages.json\nindex e8524ac..3a596d6 100644\n--- a/src/_locales/common/messages.json\n+++ b/src/_locales/common/messages.json\n@@ -4,11 +4,21 @@\n \"zh_CN\": \"\u6dfb\u52a0\",\n \"zh_TW\": \"\u65b0\u589e\"\n },\n+ \"blacklist\": {\n+ \"en\": \"Blacklist\",\n+ \"zh_CN\": \"\u9ed1\u540d\u5355\",\n+ \"zh_TW\": \"\u9ed1\u540d\u55ae\"\n+ },\n \"cancel\": {\n \"en\": \"Cancel\",\n \"zh_CN\": \"\u53d6\u6d88\",\n \"zh_TW\": \"\u53d6\u6d88\"\n },\n+ \"changes_confirm\": {\n+ \"en\": \"Changes not saved. Close anyway?\",\n+ \"zh_CN\": \"\u4fee\u6539\u672a\u4fdd\u5b58\u3002\u786e\u8ba4\u5173\u95ed\uff1f\",\n+ \"zh_TW\": \"\u4fee\u6539\u672a\u4fdd\u5b58\u3002\u78ba\u5b9a\u95dc\u9589\uff1f\"\n+ },\n \"confirm\": {\n \"en\": \"Confirm\",\n \"zh_CN\": \"\u786e\u8ba4\",\n@@ -93,5 +103,10 @@\n \"en\": \"words\",\n \"zh_CN\": \"\u4e2a\",\n \"zh_TW\": \"\u4e2a\"\n+ },\n+ \"whitelist\": {\n+ \"en\": \"Whitelist\",\n+ \"zh_CN\": \"\u767d\u540d\u5355\",\n+ \"zh_TW\": \"\u767d\u540d\u55ae\"\n }\n }\ndiff --git a/src/_locales/options/messages.json b/src/_locales/options/messages.json\nindex ada2488..e7d699a 100644\n--- a/src/_locales/options/messages.json\n+++ b/src/_locales/options/messages.json\n@@ -119,6 +119,11 @@\n \"zh_CN\": \"\u53cd\u9988\u95ee\u9898\",\n \"zh_TW\": \"\u8edf\u9ad4\u4f7f\u7528\u7591\u554f\u548c\u5efa\u8a00\"\n },\n+ \"match_pattern_description\": {\n+ \"en\": \"Specify URLs as match patterns. <a href=\\\"https://developer.mozilla.org/en-US/Add-ons/WebExtensions/Match_patterns#Examples\\\" target=\\\"_blank\\\">Examples</a>. Empty fields will be removed.\",\n+ \"zh_CN\": \"\u7f51\u5740\u652f\u6301\u5339\u914d\u6a21\u5f0f\uff08<a href=\\\"https://developer.mozilla.org/zh-CN/Add-ons/WebExtensions/Match_patterns#\u8303\u4f8b\\\" target=\\\"_blank\\\">\u4f8b\u5b50</a>\uff09\u3002\u7559\u7a7a\u4fdd\u5b58\u5373\u53ef\u6e05\u9664\u3002\",\n+ \"zh_TW\": \"\u7db2\u5740\u652f\u63f4\u5339\u914d\u6a21\u5f0f\uff08<a href=\\\"https://developer.mozilla.org/zh-CN/Add-ons/WebExtensions/Match_patterns#\u8303\u4f8b\\\" target=\\\"_blank\\\">\u4f8b\u5b50</a>\uff09\u3002\u7559\u7a7a\u5132\u5b58\u5373\u53ef\u6e05\u9664\u3002\"\n+ },\n \"msg_updated\": {\n \"en\": \"Successfully updated\",\n \"zh_CN\": \"\u8bbe\u7f6e\u5df2\u66f4\u65b0\",\n@@ -319,6 +324,21 @@\n \"zh_CN\": \"\u5f00\u542f\u540e\uff0c\u672c\u6269\u5c55\u4f1a\u81ea\u52a8\u8bc6\u522b\u8f93\u5165\u6846\u4ee5\u53ca\u5e38\u89c1\u7f16\u8f91\u5668\uff0c\u5982 CodeMirror\u3001ACE \u548c Monaco\u3002\",\n \"zh_TW\": \"\u958b\u555f\u540e\uff0c\u672c\u7a0b\u5f0f\u6703\u81ea\u52d5\u8b58\u5225\u8f38\u5165\u6846\u4ee5\u53ca\u5e38\u898b\u7de8\u8f2f\u5668\uff0c\u5982 CodeMirror\u3001ACE \u548c Monaco\u3002\"\n },\n+ \"opt_pdf_blackwhitelist_help\": {\n+ \"en\": \"Blacklisted PDF links will not jump to Saladict PDF Viewer.\",\n+ \"zh_CN\": \"\u9ed1\u540d\u5355\u5339\u914d\u7684 PDF \u94fe\u63a5\u5c06\u4e0d\u4f1a\u8df3\u8f6c\u5230 Saladict \u6253\u5f00\u3002\",\n+ \"zh_TW\": \"\u9ed1\u540d\u55ae\u5339\u914d\u7684 PDF \u9023\u7d50\u5c07\u4e0d\u6703\u8df3\u8f49\u5230 Saladict \u958b\u555f\u3002\"\n+ },\n+ \"opt_pdf_sniff\": {\n+ \"en\": \"Enable PDF Sniffer\",\n+ \"zh_CN\": \"\u9ed8\u8ba4\u7528\u672c\u6269\u5c55\u6d4f\u89c8 PDF\",\n+ \"zh_TW\": \"\u4f7f\u7528\u672c\u61c9\u7528\u7a0b\u5f0f\u700f\u89bd PDF\"\n+ },\n+ \"opt_pdf_sniff_help\": {\n+ \"en\": \"If turned on\uff0c PDF links will be automatically captured.\",\n+ \"zh_CN\": \"\u5f00\u542f\u540e\u6240\u6709 PDF \u94fe\u63a5\u5c06\u81ea\u52a8\u8df3\u8f6c\u5230\u672c\u6269\u5c55\u6253\u5f00\uff08\u5305\u62ec\u672c\u5730\uff0c\u5982\u679c\u5728\u6269\u5c55\u7ba1\u7406\u9875\u9762\u52fe\u9009\u4e86\u5141\u8bb8\uff09\u3002\",\n+ \"zh_TW\": \"\u958b\u555f\u5f8c\u6240\u6709 PDF \u9023\u7d50\u5c07\u81ea\u52d5\u8df3\u8f49\u5230\u672c\u64f4\u5145\u5957\u4ef6\u958b\u555f\uff08\u5305\u62ec\u672c\u5730\uff0c\u5982\u679c\u5728\u64f4\u5145\u5957\u4ef6\u7ba1\u7406\u9801\u9762\u52fe\u9078\u4e86\u5141\u8a31\uff09\u3002\"\n+ },\n \"opt_profile_change\": {\n \"en\": \"This option may change base on \\\"Profile\\\".\",\n \"zh_CN\": \"\u6b64\u9009\u9879\u4f1a\u56e0\u300c\u60c5\u666f\u6a21\u5f0f\u300d\u800c\u6539\u53d8\u3002\",\n@@ -329,6 +349,16 @@\n \"zh_CN\": \"\u8f93\u5165\u65f6\u663e\u793a\u5019\u9009\",\n \"zh_TW\": \"\u8f38\u5165\u6642\u986f\u793a\u5019\u9078\"\n },\n+ \"opt_sel_blackwhitelist\": {\n+ \"en\": \"Selection Black/White List\",\n+ \"zh_CN\": \"\u5212\u8bcd\u9ed1\u767d\u540d\u5355\",\n+ \"zh_TW\": \"\u9078\u8a5e\u9ed1\u767d\u540d\u55ae\"\n+ },\n+ \"opt_sel_blackwhitelist_help\": {\n+ \"en\": \"Saladict will not react to selection in blacklisted pages.\",\n+ \"zh_CN\": \"\u9ed1\u540d\u5355\u5339\u914d\u7684\u9875\u9762 Saladict \u5c06\u4e0d\u4f1a\u54cd\u5e94\u9f20\u6807\u5212\u8bcd\u3002\",\n+ \"zh_TW\": \"\u9ed1\u540d\u55ae\u5339\u914d\u7684\u9801\u9762 Saladict \u5c07\u4e0d\u6703\u97ff\u61c9\u6ed1\u9f20\u5283\u8a5e\u3002\"\n+ },\n \"opt_sel_lang\": {\n \"en\": \"Selection Languages\",\n \"zh_CN\": \"\u5212\u8bcd\u8bed\u8a00\",\ndiff --git a/src/options/components/options/BlackWhiteList/index.tsx b/src/options/components/options/BlackWhiteList/index.tsx\nnew file mode 100644\nindex 0000000..52708dd\n--- /dev/null\n+++ b/src/options/components/options/BlackWhiteList/index.tsx\n@@ -0,0 +1,69 @@\n+import React from 'react'\n+import { Props } from '../typings'\n+import { formItemLayout } from '../helpers'\n+import MatchPatternModal from '../../MatchPatternModal'\n+\n+import { FormComponentProps } from 'antd/lib/form'\n+import { Form, Button } from 'antd'\n+\n+export type BlackWhiteListProps = Props & FormComponentProps\n+\n+interface BlackWhiteListState {\n+ editingArea: '' | 'pdfWhitelist' | 'pdfBlacklist' | 'whitelist' | 'blacklist'\n+}\n+\n+export class BlackWhiteList extends React.Component<BlackWhiteListProps, BlackWhiteListState> {\n+ constructor (props: BlackWhiteListProps) {\n+ super(props)\n+ this.state = {\n+ editingArea: ''\n+ }\n+ }\n+\n+ closeModal = () => {\n+ this.setState({ editingArea: '' })\n+ }\n+\n+ render () {\n+ const { t, config } = this.props\n+\n+ return (\n+ <Form>\n+ <Form.Item\n+ {...formItemLayout}\n+ label={t('opt_sel_blackwhitelist')}\n+ help={t('opt_sel_blackwhitelist_help')}\n+ >\n+ <Button\n+ style={{ marginRight: 10 }}\n+ onClick={() => this.setState({ editingArea: 'blacklist' })}\n+ >{t('common:blacklist')}</Button>\n+ <Button\n+ onClick={() => this.setState({ editingArea: 'whitelist' })}\n+ >{t('common:whitelist')}</Button>\n+ </Form.Item>\n+ <Form.Item\n+ {...formItemLayout}\n+ label={`PDF ${t('nav_BlackWhiteList')}`}\n+ help={t('opt_pdf_blackwhitelist_help')}\n+ >\n+ <Button\n+ style={{ marginRight: 10 }}\n+ onClick={() => this.setState({ editingArea: 'pdfBlacklist' })}\n+ >PDF {t('common:blacklist')}</Button>\n+ <Button\n+ onClick={() => this.setState({ editingArea: 'pdfWhitelist' })}\n+ >PDF {t('common:whitelist')}</Button>\n+ </Form.Item>\n+ <MatchPatternModal\n+ t={t}\n+ config={config}\n+ area={this.state.editingArea}\n+ onClose={this.closeModal}\n+ />\n+ </Form>\n+ )\n+ }\n+}\n+\n+export default BlackWhiteList\ndiff --git a/src/options/components/options/PDF/index.tsx b/src/options/components/options/PDF/index.tsx\nnew file mode 100644\nindex 0000000..3e7772d\n--- /dev/null\n+++ b/src/options/components/options/PDF/index.tsx\n@@ -0,0 +1,72 @@\n+import React from 'react'\n+import { Props } from '../typings'\n+import { updateConfigOrProfile, formItemLayout } from '../helpers'\n+import MatchPatternModal from '../../MatchPatternModal'\n+\n+import { FormComponentProps } from 'antd/lib/form'\n+import { Form, Switch, Button } from 'antd'\n+\n+export type PDFProps = Props & FormComponentProps\n+\n+interface PDFState {\n+ editingArea: '' | 'pdfWhitelist' | 'pdfBlacklist'\n+}\n+\n+export class PDF extends React.Component<PDFProps, PDFState> {\n+ constructor (props: PDFProps) {\n+ super(props)\n+\n+ this.state = {\n+ editingArea: ''\n+ }\n+ }\n+\n+ closeModal = () => {\n+ this.setState({ editingArea: '' })\n+ }\n+\n+ render () {\n+ const { t, config } = this.props\n+ const { getFieldDecorator } = this.props.form\n+\n+ return (\n+ <Form>\n+ <Form.Item\n+ {...formItemLayout}\n+ label={t('opt_pdf_sniff')}\n+ help={t('opt_pdf_sniff_help')}\n+ >{\n+ getFieldDecorator('config#pdfSniff', {\n+ initialValue: config.pdfSniff,\n+ valuePropName: 'checked',\n+ })(\n+ <Switch />\n+ )\n+ }</Form.Item>\n+ <Form.Item\n+ {...formItemLayout}\n+ label={t('nav_BlackWhiteList')}\n+ help={t('opt_pdf_blackwhitelist_help')}\n+ >\n+ <Button\n+ style={{ marginRight: 10 }}\n+ onClick={() => this.setState({ editingArea: 'pdfBlacklist' })}\n+ >PDF {t('common:blacklist')}</Button>\n+ <Button\n+ onClick={() => this.setState({ editingArea: 'pdfWhitelist' })}\n+ >PDF {t('common:whitelist')}</Button>\n+ </Form.Item>\n+ <MatchPatternModal\n+ t={t}\n+ config={config}\n+ area={this.state.editingArea}\n+ onClose={this.closeModal}\n+ />\n+ </Form>\n+ )\n+ }\n+}\n+\n+export default Form.create<PDFProps>({\n+ onValuesChange: updateConfigOrProfile\n+})(PDF)\n"]
1
["9b1c0fc20b614513384a1e562317dbf076eb8ef0"]
["feat"]
change notice from 'danger' > 'info' Signed-off-by: Raju Udava <[email protected]>
["diff --git a/packages/noco-docs/docs/030.workspaces/040.actions-on-workspace.md b/packages/noco-docs/docs/030.workspaces/040.actions-on-workspace.md\nindex 17a1d85..b8c3f52 100644\n--- a/packages/noco-docs/docs/030.workspaces/040.actions-on-workspace.md\n+++ b/packages/noco-docs/docs/030.workspaces/040.actions-on-workspace.md\n@@ -20,7 +20,7 @@ To update the workspace name:\n ## Delete workspace\n If you determine that a workspace is no longer necessary, you have the option to permanently remove it from your settings. Deleting a workspace will delete all the bases and data associated with it.\n \n-:::danger\n+:::info\n **This action cannot be undone.**\n :::\n \ndiff --git a/packages/noco-docs/docs/040.bases/070.actions-on-base.md b/packages/noco-docs/docs/040.bases/070.actions-on-base.md\nindex b8e5723..7207971 100644\n--- a/packages/noco-docs/docs/040.bases/070.actions-on-base.md\n+++ b/packages/noco-docs/docs/040.bases/070.actions-on-base.md\n@@ -69,7 +69,7 @@ To duplicate a base, you can follow these straightforward steps:\n \n If you determine that a base is no longer necessary, you have the option to permanently remove it from your workspace. Deleting a base will delete all the tables and data associated with it.\n \n-:::danger\n+:::info\n **This action cannot be undone.**\n :::\n \ndiff --git a/packages/noco-docs/docs/050.tables/060.actions-on-table.md b/packages/noco-docs/docs/050.tables/060.actions-on-table.md\nindex 3cf03d3..8ae9ade 100644\n--- a/packages/noco-docs/docs/050.tables/060.actions-on-table.md\n+++ b/packages/noco-docs/docs/050.tables/060.actions-on-table.md\n@@ -46,7 +46,7 @@ A new table will be generated, mirroring the original table's schema and content\n \n ## Delete table\n \n-:::danger\n+:::info\n **This action cannot be undone.**\n :::\n \ndiff --git a/packages/noco-docs/docs/070.fields/060.actions-on-field.md b/packages/noco-docs/docs/070.fields/060.actions-on-field.md\nindex 600c6fd..fe2cfa8 100644\n--- a/packages/noco-docs/docs/070.fields/060.actions-on-field.md\n+++ b/packages/noco-docs/docs/070.fields/060.actions-on-field.md\n@@ -83,7 +83,7 @@ New field will be created to the right of the original field.\n New field will be created to the left of the original field.\n \n ### Delete field\n-:::danger\n+:::info\n **This action cannot be undone.**\n :::\n \ndiff --git a/packages/noco-docs/docs/080.records/070.actions-on-record.md b/packages/noco-docs/docs/080.records/070.actions-on-record.md\nindex a9245ff..6d4774a 100644\n--- a/packages/noco-docs/docs/080.records/070.actions-on-record.md\n+++ b/packages/noco-docs/docs/080.records/070.actions-on-record.md\n@@ -54,8 +54,8 @@ On the bulk update modal,\n 5. Click on the `Bulk Update all` button\n 6. A confirmation dialog will be displayed. Click on `Confirm` to update the records.\n \n-:::danger\n-This operation cannot be undone.\n+:::info\n+**This action cannot be undone.**\n :::\n \n ![Bulk Update](/img/v2/records/bulk-update-1.png)\ndiff --git a/packages/noco-docs/docs/090.views/090.actions-on-view.md b/packages/noco-docs/docs/090.views/090.actions-on-view.md\nindex c6c6ab2..7d23959 100644\n--- a/packages/noco-docs/docs/090.views/090.actions-on-view.md\n+++ b/packages/noco-docs/docs/090.views/090.actions-on-view.md\n@@ -41,7 +41,7 @@ The view context menu provides a set of tools to interact with the view. The vie\n \n ## Delete view\n \n-:::danger\n+:::info\n **This action cannot be undone.**\n :::\n \n"]
1
["2ba752d45350a676babe553dd68f019af81b512b"]
["docs"]
entries updates
["diff --git a/packages/docz-core/src/DataServer.ts b/packages/docz-core/src/DataServer.ts\nindex 0dad341..d1d95fb 100644\n--- a/packages/docz-core/src/DataServer.ts\n+++ b/packages/docz-core/src/DataServer.ts\n@@ -34,13 +34,13 @@ export class DataServer {\n public async processEntries(): Promise<void> {\n const config = this.config\n const entries = new Entries(config)\n- const map = await entries.getMap()\n const watcher = chokidar.watch(this.config.files, {\n ignored: /(^|[\\/\\\\])\\../,\n })\n \n- const handleConnection = (socket: WS) => {\n- const update = this.updateEntries(socket)\n+ const handleConnection = async (socket: WS) => {\n+ const update = this.updateEntries(entries, socket)\n+ const map = await entries.getMap()\n \n watcher.on('change', async () => update(this.config))\n watcher.on('unlink', async () => update(this.config))\n@@ -51,12 +51,14 @@ export class DataServer {\n })\n \n socket.send(this.entriesData(map))\n+ await Entries.writeImports(map)\n }\n \n this.server.on('connection', handleConnection)\n this.server.on('close', () => watcher.close())\n \n- await Entries.write(config, map)\n+ await Entries.writeGenerated(config)\n+ await Entries.writeImports(await entries.getMap())\n }\n \n public async processThemeConfig(): Promise<void> {\n@@ -88,14 +90,16 @@ export class DataServer {\n return this.dataObj('docz.config', config.themeConfig)\n }\n \n- private updateEntries(socket: WS): (config: Config) => Promise<void> {\n+ private updateEntries(\n+ entries: Entries,\n+ socket: WS\n+ ): (config: Config) => Promise<void> {\n return async config => {\n if (isSocketOpened(socket)) {\n- const newEntries = new Entries(config)\n- const newMap = await newEntries.getMap()\n+ const map = await entries.getMap()\n \n- await Entries.rewrite(newMap)\n- socket.send(this.entriesData(newMap))\n+ await Entries.writeImports(map)\n+ socket.send(this.entriesData(map))\n }\n }\n }\ndiff --git a/packages/docz-core/src/Entries.ts b/packages/docz-core/src/Entries.ts\nindex 76178eb..6e1a370 100644\n--- a/packages/docz-core/src/Entries.ts\n+++ b/packages/docz-core/src/Entries.ts\n@@ -77,14 +77,13 @@ const writeImports = async (entries: EntryMap): Promise<void> => {\n export type EntryMap = Record<string, Entry>\n \n export class Entries {\n- public static async write(config: Config, entries: EntryMap): Promise<void> {\n+ public static async writeGenerated(config: Config): Promise<void> {\n mkd(paths.docz)\n await writeGeneratedFiles(config)\n- await writeImports(entries)\n }\n \n- public static async rewrite(map: EntryMap): Promise<void> {\n- await writeImports(map)\n+ public static async writeImports(entries: EntryMap): Promise<void> {\n+ await writeImports(entries)\n }\n \n public all: EntryMap\ndiff --git a/packages/docz-core/templates/app.tpl.js b/packages/docz-core/templates/app.tpl.js\nindex 22ad59b..bbb9081 100644\n--- a/packages/docz-core/templates/app.tpl.js\n+++ b/packages/docz-core/templates/app.tpl.js\n@@ -18,15 +18,6 @@ class App extends React.Component {\n state = {\n config: {},\n entries: {},\n- imports: {},\n- }\n-\n- static getDerivedStateFromProps(nextProps, prevState) {\n- return {\n- config: prevState.config,\n- entries: prevState.entries,\n- imports: nextProps.imports\n- }\n }\n \n async componentDidMount() {\n@@ -44,7 +35,8 @@ class App extends React.Component {\n }\n \n render() {\n- return <Theme {...this.state} wrapper={Wrapper} />\n+ const { imports } = this.props\n+ return <Theme {...this.state} imports={imports} wrapper={Wrapper} />\n }\n }\n \n"]
1
["7147ac1f43a3ca454c79a6709dda2c35162ec88c"]
["fix"]
increase timeout of multiregion failover test Due to the nature of the test, restarts and failovers can take long. If the recovery takes longer than 15m, then the test will fail unnecessarily. Since we are not really testing for how was it can recover, it is ok to increase the maxInstanceDuration.
["diff --git a/.github/workflows/e2e-testbench.yaml b/.github/workflows/e2e-testbench.yaml\nindex 708f97f..fd0b918 100644\n--- a/.github/workflows/e2e-testbench.yaml\n+++ b/.github/workflows/e2e-testbench.yaml\n@@ -31,6 +31,11 @@ on:\n default: null\n required: false\n type: string\n+ maxInstanceDuration:\n+ description: 'If an instance takes longer than the given duration to complete, test will fail.'\n+ default: '15m'\n+ required: false\n+ type: string\n \n workflow_call:\n inputs:\n@@ -59,6 +64,11 @@ on:\n default: null\n required: false\n type: string\n+ maxInstanceDuration:\n+ description: 'If an instance takes longer than the given duration to complete, test will fail.'\n+ default: '15m'\n+ required: false\n+ type: string\n \n jobs:\n e2e:\n@@ -81,7 +91,7 @@ jobs:\n {\n \\\"maxTestDuration\\\": \\\"${{ inputs.maxTestDuration || 'P5D' }}\\\",\n \\\"starter\\\": [ {\\\"rate\\\": 50, \\\"processId\\\": \\\"one-task-one-timer\\\" } ],\n- \\\"verifier\\\" : { \\\"maxInstanceDuration\\\" : \\\"15m\\\" },\n+ \\\"verifier\\\" : { \\\"maxInstanceDuration\\\" : \\\"${{ inputs.maxInstanceDuration }}\\\" },\n \\\"fault\\\": ${{ inputs.fault || 'null' }}\n }\n }\ndiff --git a/.github/workflows/weekly-e2e.yml b/.github/workflows/weekly-e2e.yml\nindex 93aaeb5..4bd0afd 100644\n--- a/.github/workflows/weekly-e2e.yml\n+++ b/.github/workflows/weekly-e2e.yml\n@@ -31,4 +31,5 @@ jobs:\n maxTestDuration: P1D\n clusterPlan: Multiregion test simulation\n fault: \\\"2-region-dataloss-failover\\\"\n+ maxInstanceDuration: 40m\n secrets: inherit\n"]
1
["ee824ddd71cbc4ccc26f7c6876d379c4927b79e6"]
["ci"]
add `to_sql` Co-authored-by: Gil Forsyth <[email protected]>
["diff --git a/docs/api/expressions/top_level.md b/docs/api/expressions/top_level.md\nindex efaffbd..34b529e 100644\n--- a/docs/api/expressions/top_level.md\n+++ b/docs/api/expressions/top_level.md\n@@ -28,7 +28,7 @@ These methods and objects are available directly in the `ibis` module.\n ::: ibis.or_\n ::: ibis.param\n ::: ibis.show_sql\n-::: ibis.sql\n+::: ibis.to_sql\n ::: ibis.random\n ::: ibis.range_window\n ::: ibis.row_number\n"]
1
["e2821a56c7d867b8b591f1777019843a2ffca797"]
["docs"]
build updates
["diff --git a/demo/vanilla_new/css/404.min.css b/demo/vanilla_new/css/404.min.css\nindex a3485b4..e69de29 100644\n--- a/demo/vanilla_new/css/404.min.css\n+++ b/demo/vanilla_new/css/404.min.css\n@@ -1 +0,0 @@\n-@import url(https://fonts.googleapis.com/css?family=Share+Tech+Mono%7CSpace+Mono);a,abbr,acronym,address,applet,article,aside,audio,b,big,blockquote,body,canvas,caption,center,cite,code,dd,del,details,dfn,div,dl,dt,em,embed,fieldset,figcaption,figure,footer,form,h1,h2,h3,h4,h5,h6,header,hgroup,html,i,iframe,img,ins,kbd,label,legend,li,mark,menu,nav,object,ol,output,p,pre,q,ruby,s,samp,section,small,span,strike,strong,sub,summary,sup,table,tbody,td,tfoot,th,thead,time,tr,tt,u,ul,var,video{margin:0;padding:0;border:0;font-size:100%;font:inherit;vertical-align:baseline}article,aside,details,figcaption,figure,footer,header,hgroup,main,menu,nav,section{display:block}body{line-height:1}ol,ul{list-style:none}blockquote,q{quotes:none}blockquote:after,blockquote:before,q:after,q:before{content:\"\";content:none}table{border-collapse:collapse;border-spacing:0}body{padding:0;margin:0;font-size:18px}.container{min-height:100vh;position:relative;padding:240px 0;box-sizing:border-box}.overlay{position:absolute;top:0;left:0;width:100%;height:100%;overflow:hidden}.content{position:absolute;top:50%;left:50%;width:100%;transform:translate(-50%,-50%)}.message{text-align:center;color:#000}.message-heading{font-family:\"Share Tech Mono\";font-weight:900;text-transform:uppercase;letter-spacing:.7em;font-size:2rem;padding:0 0 0 1.4em}.message-description{font-family:\"Space Mono\";line-height:42px;font-size:15px;letter-spacing:.15rem;padding:0 20px;max-width:600px;margin:auto}.links{max-width:600px;margin:40px auto 0;text-align:center}.links a{width:170px;display:inline-block;padding:15px 0;margin:0 15px;border:1px solid #000;color:#000;text-decoration:none;font-family:\"Space Mono\";text-transform:uppercase;font-size:11px;letter-spacing:.1rem;position:relative}.links a:before{content:\"\";height:42px;background:#000;position:absolute;top:0;right:0;width:0;transition:all .3s}.links a:after{transition:all .3s;z-index:999;position:relative;content:\"back to hompage\"}.links a:hover:before{width:170px}.links a:hover:after{color:#fff}.links a:nth-child(2){background:#fff;color:#000}.links a:nth-child(2):before{background:#212121;left:0}.links a:nth-child(2):after{content:\"report error\"}.links a:nth-child(2):hover:after{color:#fff}.social{position:absolute;bottom:15px;left:15px}.social-list{margin:0;padding:0;list-style-type:none}.social-list li{display:inline-block;margin:5px 10px}.social-list li a{color:#000}@media (max-width:480px){.message-heading{font-size:1rem;margin-bottom:30px}.message-description{font-size:.7rem;line-height:2rem}.links a{margin:10px;width:280px}.social{left:50%;margin-left:-55px}}\ndiff --git a/demo/vanilla_new/css/main.min.css b/demo/vanilla_new/css/main.min.css\nindex 043eb4f..e69de29 100644\n--- a/demo/vanilla_new/css/main.min.css\n+++ b/demo/vanilla_new/css/main.min.css\n@@ -1 +0,0 @@\n-html{height:100%;overflow:hidden}body{line-height:1;height:100%;overflow:hidden;background:#000}#floating-container{right:20px;top:20px;position:fixed;z-index:4000}\ndiff --git a/demo/vanilla_new/js/404.min.js b/demo/vanilla_new/js/404.min.js\nindex 3642106..e69de29 100644\n--- a/demo/vanilla_new/js/404.min.js\n+++ b/demo/vanilla_new/js/404.min.js\n@@ -1 +0,0 @@\n-tsParticles.loadJSON(\"tsparticles\",\"/configs/404.json\");\ndiff --git a/website/css/404.min.css b/website/css/404.min.css\nindex a3485b4..e69de29 100644\n--- a/website/css/404.min.css\n+++ b/website/css/404.min.css\n@@ -1 +0,0 @@\n-@import url(https://fonts.googleapis.com/css?family=Share+Tech+Mono%7CSpace+Mono);a,abbr,acronym,address,applet,article,aside,audio,b,big,blockquote,body,canvas,caption,center,cite,code,dd,del,details,dfn,div,dl,dt,em,embed,fieldset,figcaption,figure,footer,form,h1,h2,h3,h4,h5,h6,header,hgroup,html,i,iframe,img,ins,kbd,label,legend,li,mark,menu,nav,object,ol,output,p,pre,q,ruby,s,samp,section,small,span,strike,strong,sub,summary,sup,table,tbody,td,tfoot,th,thead,time,tr,tt,u,ul,var,video{margin:0;padding:0;border:0;font-size:100%;font:inherit;vertical-align:baseline}article,aside,details,figcaption,figure,footer,header,hgroup,main,menu,nav,section{display:block}body{line-height:1}ol,ul{list-style:none}blockquote,q{quotes:none}blockquote:after,blockquote:before,q:after,q:before{content:\"\";content:none}table{border-collapse:collapse;border-spacing:0}body{padding:0;margin:0;font-size:18px}.container{min-height:100vh;position:relative;padding:240px 0;box-sizing:border-box}.overlay{position:absolute;top:0;left:0;width:100%;height:100%;overflow:hidden}.content{position:absolute;top:50%;left:50%;width:100%;transform:translate(-50%,-50%)}.message{text-align:center;color:#000}.message-heading{font-family:\"Share Tech Mono\";font-weight:900;text-transform:uppercase;letter-spacing:.7em;font-size:2rem;padding:0 0 0 1.4em}.message-description{font-family:\"Space Mono\";line-height:42px;font-size:15px;letter-spacing:.15rem;padding:0 20px;max-width:600px;margin:auto}.links{max-width:600px;margin:40px auto 0;text-align:center}.links a{width:170px;display:inline-block;padding:15px 0;margin:0 15px;border:1px solid #000;color:#000;text-decoration:none;font-family:\"Space Mono\";text-transform:uppercase;font-size:11px;letter-spacing:.1rem;position:relative}.links a:before{content:\"\";height:42px;background:#000;position:absolute;top:0;right:0;width:0;transition:all .3s}.links a:after{transition:all .3s;z-index:999;position:relative;content:\"back to hompage\"}.links a:hover:before{width:170px}.links a:hover:after{color:#fff}.links a:nth-child(2){background:#fff;color:#000}.links a:nth-child(2):before{background:#212121;left:0}.links a:nth-child(2):after{content:\"report error\"}.links a:nth-child(2):hover:after{color:#fff}.social{position:absolute;bottom:15px;left:15px}.social-list{margin:0;padding:0;list-style-type:none}.social-list li{display:inline-block;margin:5px 10px}.social-list li a{color:#000}@media (max-width:480px){.message-heading{font-size:1rem;margin-bottom:30px}.message-description{font-size:.7rem;line-height:2rem}.links a{margin:10px;width:280px}.social{left:50%;margin-left:-55px}}\ndiff --git a/website/css/main.min.css b/website/css/main.min.css\nindex 818002f..e69de29 100644\n--- a/website/css/main.min.css\n+++ b/website/css/main.min.css\n@@ -1 +0,0 @@\n-@font-face{font-family:Polya;src:url(https://raw.githubusercontent.com/matteobruni/tsparticles/gh-pages/fonts/Polya.otf)}html{height:100%;overflow:hidden}body{line-height:1;height:100%;overflow:hidden;background:#000}.github{bottom:10px;right:10px;padding:0 12px 6px;position:fixed;border-radius:10px;background:#fff;border:1px solid #000}.github a{color:#000}.github a:active,.github a:hover,.github a:link,.github a:visited{color:#000;text-decoration:none}.github img{height:30px}.github #gh-project{font-size:20px;padding-left:5px;font-weight:700;vertical-align:bottom}.toggle-sidebar{top:50%;left:0;font-size:20px;color:#000;position:absolute;padding:3px;border-top-right-radius:5px;border-bottom-right-radius:5px;background:#e7e7e7;border:1px solid #000;border-left:none}#editor{background:#fff}[hidden]{display:none}#repulse-div{width:200px;height:200px;background-color:rgba(255,255,255,.5);border-radius:100px;position:absolute;top:50%;left:50%;margin-left:-100px;margin-top:-100px;z-index:200}@media (min-width:1600px) and (-webkit-device-pixel-ratio:1){.col-xxl-3{-ms-flex:0 0 25%;flex:0 0 25%;max-width:25%}}.btn-react{color:#fff;background-color:#61dafb;border-color:#fff}.btn-react:hover{color:#fff;background-color:#5aa3c4;border-color:#ccc}.btn-react.focus,.btn-react:focus{color:#fff;background-color:#5aa3c4;border-color:#ccc;box-shadow:0 0 0 .2rem rgba(90,163,196,.5)}.btn-react.disabled,.btn-react:disabled{color:#fff;background-color:#61dafb;border-color:#ccc}.btn-react:not(:disabled):not(.disabled).active,.btn-react:not(:disabled):not(.disabled):active,.show>.btn-react.dropdown-toggle{color:#fff;background-color:#5aa3c4;border-color:#ccc}.btn-react:not(:disabled):not(.disabled).active:focus,.btn-react:not(:disabled):not(.disabled):active:focus,.show>.btn-react.dropdown-toggle:focus{box-shadow:0 0 0 .2rem rgba(90,163,196,.5)}#stats,.count-particles{-webkit-user-select:none}#stats{overflow:hidden}#stats-graph canvas{border-radius:3px 3px 0 0}.count-particles{border-radius:0 0 3px 3px}\ndiff --git a/website/css/presets.min.css b/website/css/presets.min.css\nindex 6c2ae2c..e69de29 100644\n--- a/website/css/presets.min.css\n+++ b/website/css/presets.min.css\n@@ -1 +0,0 @@\n-html{height:100%;overflow:hidden}body{line-height:1;height:100%;overflow:hidden}\n"]
1
["9acf7a062ee9c0538c2cd4661c1f5da61ab06316"]
["build"]
make sure root is being watched and setRoot called when it changes
["diff --git a/packages/core/src/components/nav/nav.tsx b/packages/core/src/components/nav/nav.tsx\nindex 5aaacb6..27241ee 100644\n--- a/packages/core/src/components/nav/nav.tsx\n+++ b/packages/core/src/components/nav/nav.tsx\n@@ -1,4 +1,4 @@\n-import { Component, Element, Event, EventEmitter, Listen, Method, Prop } from '@stencil/core';\n+import { Component, Element, Event, EventEmitter, Listen, Method, Prop, Watch } from '@stencil/core';\n import {\n Animation,\n AnimationController,\n@@ -103,10 +103,19 @@ export class Nav implements PublicNav, NavContainer {\n }\n this.init = true;\n if (!this.useRouter) {\n+ console.log('componentDidLoadImpl: ', this.root);\n componentDidLoadImpl(this);\n }\n }\n \n+ @Watch('root')\n+ updateRootComponent(): any {\n+ console.log('updateRootComponent: ', this.root);\n+ if (this.init) {\n+ return this.setRoot(this.root);\n+ }\n+ }\n+\n getViews(): PublicViewController[] {\n return getViews(this);\n }\ndiff --git a/packages/core/src/components/nav/test/set-root/index.html b/packages/core/src/components/nav/test/set-root/index.html\nnew file mode 100644\nindex 0000000..823c9ed\n--- /dev/null\n+++ b/packages/core/src/components/nav/test/set-root/index.html\n@@ -0,0 +1,110 @@\n+<!DOCTYPE html>\n+<html dir=\"ltr\">\n+<head>\n+ <meta charset=\"UTF-8\">\n+ <title>Nav</title>\n+ <meta name=\"viewport\" content=\"width=device-width, initial-scale=1.0, minimum-scale=1.0, maximum-scale=1.0, user-scalable=no\">\n+ <script src=\"/dist/ionic.js\"></script>\n+</head>\n+<body onload=\"initiaize()\">\n+ <ion-app>\n+ <ion-nav root=\"page-one\"></ion-nav>\n+ </ion-app>\n+</body>\n+\n+<script>\n+\n+ class PageOne extends HTMLElement {\n+ async connectedCallback() {\n+ this.innerHTML = `\n+ <ion-page>\n+ <ion-header>\n+ <ion-toolbar>\n+ <ion-title>Page One</ion-title>\n+ </ion-toolbar>\n+ </ion-header>\n+ <ion-content padding>\n+ <h1>Page One</h1>\n+ <ion-button class=\"next\">Go to Page Two</ion-button>\n+ </ion-content>\n+ </ion-page>`;\n+\n+ const button = this.querySelector('ion-button');\n+ button.addEventListener('click', async () => {\n+ this.closest('ion-nav').push('page-two');\n+ });\n+ }\n+ }\n+\n+ class PageTwo extends HTMLElement {\n+ async connectedCallback() {\n+ this.innerHTML = `\n+ <ion-page>\n+ <ion-header>\n+ <ion-toolbar>\n+ <ion-title>Page Two</ion-title>\n+ </ion-toolbar>\n+ </ion-header>\n+ <ion-content padding>\n+ <h1>Page Two</h1>\n+ <ion-button class=\"next\">Go to Page Three</ion-button>\n+ <ion-button class=\"previous\">Go Back</ion-button>\n+ </ion-content>\n+ </ion-page>`;\n+\n+ const previousButton = this.querySelector('ion-button.previous');\n+ previousButton.addEventListener('click', async () => {\n+ await this.closest('ion-nav').pop();\n+ });\n+\n+ const nextButton = this.querySelector('ion-button.next');\n+ nextButton.addEventListener('click', async () => {\n+ await this.closest('ion-nav').push('page-three');\n+ });\n+ }\n+ }\n+\n+ class PageThree extends HTMLElement {\n+ async connectedCallback() {\n+ this.innerHTML = `\n+ <ion-page>\n+ <ion-header>\n+ <ion-toolbar>\n+ <ion-title>Page Three</ion-title>\n+ </ion-toolbar>\n+ </ion-header>\n+ <ion-content padding>\n+ <h1>Page Three</h1>\n+ <ion-button class=\"previous\">Go Back</ion-button>\n+ </ion-content>\n+ </ion-page>`;\n+\n+ const previousButton = this.querySelector('ion-button.previous');\n+ previousButton.addEventListener('click', async () => {\n+ await this.closest('ion-nav').pop();\n+ });\n+ }\n+ }\n+\n+ customElements.define('page-one', PageOne);\n+ customElements.define('page-two', PageTwo);\n+ customElements.define('page-three', PageThree);\n+\n+ async function initiaize() {\n+ const nav = document.querySelector('ion-nav');\n+ await nav.componentOnReady();\n+ nav.root = 'page-one';\n+\n+ setInterval(() => {\n+ if (nav.root === 'page-one') {\n+ nav.root = 'page-two';\n+ } else if ( nav.root === 'page-two') {\n+ nav.root = 'page-three';\n+ } else {\n+ nav.root = 'page-one';\n+ }\n+ }, 1000);\n+ }\n+\n+</script>\n+</html>\n"]
1
["4be836f5655fb5356fde5ddd7437125f8574705d"]
["refactor"]
create dashboards from imported templates
["diff --git a/CHANGELOG.md b/CHANGELOG.md\nindex 590f5ea..bd74e95 100644\n--- a/CHANGELOG.md\n+++ b/CHANGELOG.md\n@@ -1,6 +1,7 @@\n ## v2.0.0-alpha.6 [unreleased]\n \n ### Features\n+1. [12496](https://github.com/influxdata/influxdb/pull/12496): Add ability to import a dashboard\n \n ### Bug Fixes\n \ndiff --git a/ui/src/dashboards/actions/v2/index.ts b/ui/src/dashboards/actions/v2/index.ts\nindex ad0d13d..35babc7 100644\n--- a/ui/src/dashboards/actions/v2/index.ts\n+++ b/ui/src/dashboards/actions/v2/index.ts\n@@ -16,6 +16,7 @@ import {\n removeDashboardLabels as removeDashboardLabelsAJAX,\n updateView as updateViewAJAX,\n } from 'src/dashboards/apis/v2'\n+import {client} from 'src/utils/api'\n \n // Actions\n import {notify} from 'src/shared/actions/notifications'\n@@ -25,6 +26,10 @@ import {\n DeleteTimeRangeAction,\n } from 'src/dashboards/actions/v2/ranges'\n import {setView, SetViewAction} from 'src/dashboards/actions/v2/views'\n+import {\n+ importDashboardSucceeded,\n+ importDashboardFailed,\n+} from 'src/shared/copy/notifications'\n \n // Utils\n import {\n@@ -38,7 +43,7 @@ import * as copy from 'src/shared/copy/notifications'\n // Types\n import {RemoteDataState} from 'src/types'\n import {PublishNotificationAction} from 'src/types/actions/notifications'\n-import {CreateCell} from '@influxdata/influx'\n+import {CreateCell, IDashboardTemplate} from '@influxdata/influx'\n import {Dashboard, NewView, Cell} from 'src/types/v2'\n import {ILabel} from '@influxdata/influx'\n \n@@ -201,6 +206,19 @@ export const getDashboardsAsync = () => async (\n }\n }\n \n+export const createDashboardFromTemplate = (\n+ template: IDashboardTemplate,\n+ orgID: string\n+) => async dispatch => {\n+ try {\n+ await client.dashboards.createFromTemplate(template, orgID)\n+\n+ dispatch(notify(importDashboardSucceeded()))\n+ } catch (error) {\n+ dispatch(notify(importDashboardFailed(error)))\n+ }\n+}\n+\n export const importDashboardAsync = (dashboard: Dashboard) => async (\n dispatch: Dispatch<Action>\n ): Promise<void> => {\ndiff --git a/ui/src/dashboards/components/ImportDashboardOverlay.tsx b/ui/src/dashboards/components/ImportDashboardOverlay.tsx\nindex 37ef80b..84216c3 100644\n--- a/ui/src/dashboards/components/ImportDashboardOverlay.tsx\n+++ b/ui/src/dashboards/components/ImportDashboardOverlay.tsx\n@@ -1,70 +1,80 @@\n+// Libraries\n import React, {PureComponent} from 'react'\n import _ from 'lodash'\n+import {connect} from 'react-redux'\n \n-import Container from 'src/clockface/components/overlays/OverlayContainer'\n-import Heading from 'src/clockface/components/overlays/OverlayHeading'\n-import Body from 'src/clockface/components/overlays/OverlayBody'\n-import DragAndDrop from 'src/shared/components/DragAndDrop'\n+// Constants\n import {dashboardImportFailed} from 'src/shared/copy/notifications'\n \n-import {Dashboard} from 'src/types/v2'\n-import {Notification} from 'src/types/notifications'\n+// Actions\n+import {notify as notifyAction} from 'src/shared/actions/notifications'\n+import {getDashboardsAsync} from 'src/dashboards/actions/v2'\n \n-interface Props {\n+// Types\n+import ImportOverlay from 'src/shared/components/ImportOverlay'\n+import {createDashboardFromTemplate as createDashboardFromTemplateAction} from 'src/dashboards/actions/v2'\n+\n+interface OwnProps {\n onDismissOverlay: () => void\n- onImportDashboard: (dashboard: Dashboard) => void\n- notify: (message: Notification) => void\n+ orgID: string\n+ isVisible: boolean\n+}\n+interface DispatchProps {\n+ notify: typeof notifyAction\n+ createDashboardFromTemplate: typeof createDashboardFromTemplateAction\n+ populateDashboards: typeof getDashboardsAsync\n }\n+\n+type Props = OwnProps & DispatchProps\n+\n class ImportDashboardOverlay extends PureComponent<Props> {\n constructor(props: Props) {\n super(props)\n }\n \n public render() {\n- const {onDismissOverlay} = this.props\n+ const {isVisible, onDismissOverlay} = this.props\n \n return (\n- <Container maxWidth={800}>\n- <Heading title=\"Import Dashboard\" onDismiss={onDismissOverlay} />\n- <Body>\n- <DragAndDrop\n- submitText=\"Upload Dashboard\"\n- fileTypesToAccept={this.validFileExtension}\n- handleSubmit={this.handleUploadDashboard}\n- />\n- </Body>\n- </Container>\n+ <ImportOverlay\n+ isVisible={isVisible}\n+ onDismissOverlay={onDismissOverlay}\n+ resourceName=\"Dashboard\"\n+ onSubmit={this.handleUploadDashboard}\n+ />\n )\n }\n \n- private get validFileExtension(): string {\n- return '.json'\n- }\n-\n- private handleUploadDashboard = (\n- uploadContent: string,\n- fileName: string\n- ): void => {\n- const {notify, onImportDashboard, onDismissOverlay} = this.props\n- const fileExtensionRegex = new RegExp(`${this.validFileExtension}$`)\n- if (!fileName.match(fileExtensionRegex)) {\n- notify(dashboardImportFailed('Please import a JSON file'))\n- return\n- }\n+ private handleUploadDashboard = async (\n+ uploadContent: string\n+ ): Promise<void> => {\n+ const {\n+ notify,\n+ createDashboardFromTemplate,\n+ onDismissOverlay,\n+ populateDashboards,\n+ orgID,\n+ } = this.props\n \n try {\n- const {dashboard} = JSON.parse(uploadContent)\n+ const template = JSON.parse(uploadContent)\n \n- if (!_.isEmpty(dashboard)) {\n- onImportDashboard(dashboard)\n- onDismissOverlay()\n- } else {\n- notify(dashboardImportFailed('No dashboard found in file'))\n- }\n+ await createDashboardFromTemplate(template, orgID)\n+ await populateDashboards()\n+\n+ onDismissOverlay()\n } catch (error) {\n notify(dashboardImportFailed(error))\n }\n }\n }\n+const mdtp: DispatchProps = {\n+ notify: notifyAction,\n+ createDashboardFromTemplate: createDashboardFromTemplateAction,\n+ populateDashboards: getDashboardsAsync,\n+}\n \n-export default ImportDashboardOverlay\n+export default connect<{}, DispatchProps, OwnProps>(\n+ null,\n+ mdtp\n+)(ImportDashboardOverlay)\ndiff --git a/ui/src/dashboards/components/dashboard_index/DashboardsIndex.tsx b/ui/src/dashboards/components/dashboard_index/DashboardsIndex.tsx\nindex d6b299f..1ff7f47 100644\n--- a/ui/src/dashboards/components/dashboard_index/DashboardsIndex.tsx\n+++ b/ui/src/dashboards/components/dashboard_index/DashboardsIndex.tsx\n@@ -2,15 +2,15 @@\n import React, {PureComponent} from 'react'\n import {InjectedRouter} from 'react-router'\n import {connect} from 'react-redux'\n-import {isEmpty} from 'lodash'\n+import {get} from 'lodash'\n \n // Components\n import DashboardsIndexContents from 'src/dashboards/components/dashboard_index/DashboardsIndexContents'\n import {Page} from 'src/pageLayout'\n import SearchWidget from 'src/shared/components/search_widget/SearchWidget'\n import AddResourceDropdown from 'src/shared/components/AddResourceDropdown'\n-import ImportOverlay from 'src/shared/components/ImportOverlay'\n import ExportOverlay from 'src/shared/components/ExportOverlay'\n+import ImportDashboardOverlay from 'src/dashboards/components/ImportDashboardOverlay'\n \n // APIs\n import {createDashboard, cloneDashboard} from 'src/dashboards/apis/v2/'\n@@ -32,10 +32,7 @@ import {DEFAULT_DASHBOARD_NAME} from 'src/dashboards/constants/index'\n import {\n dashboardSetDefaultFailed,\n dashboardCreateFailed,\n- dashboardImported,\n- dashboardImportFailed,\n } from 'src/shared/copy/notifications'\n-import {cantImportInvalidResource} from 'src/shared/copy/v2/notifications'\n \n // Types\n import {Notification} from 'src/types/notifications'\n@@ -197,24 +194,6 @@ class DashboardIndex extends PureComponent<Props, State> {\n this.props.handleDeleteDashboard(dashboard)\n }\n \n- private handleImportDashboard = async (\n- importString: string\n- ): Promise<void> => {\n- const {notify} = this.props\n- try {\n- const resource = JSON.parse(importString)\n-\n- if (isEmpty(resource)) {\n- notify(cantImportInvalidResource('Dashboard'))\n- return\n- }\n- this.handleToggleImportOverlay()\n- notify(dashboardImported())\n- } catch (error) {\n- notify(dashboardImportFailed(error))\n- }\n- }\n-\n private handleFilterDashboards = (searchTerm: string): void => {\n this.setState({searchTerm})\n }\n@@ -229,13 +208,13 @@ class DashboardIndex extends PureComponent<Props, State> {\n \n private get importOverlay(): JSX.Element {\n const {isImportingDashboard} = this.state\n+ const {orgs} = this.props\n \n return (\n- <ImportOverlay\n- isVisible={isImportingDashboard}\n- resourceName=\"Dashboard\"\n+ <ImportDashboardOverlay\n onDismissOverlay={this.handleToggleImportOverlay}\n- onSubmit={this.handleImportDashboard}\n+ orgID={get(orgs, '0.id', '')}\n+ isVisible={isImportingDashboard}\n />\n )\n }\ndiff --git a/ui/src/organizations/components/Dashboards.tsx b/ui/src/organizations/components/Dashboards.tsx\nindex 08c8402..90c2514 100644\n--- a/ui/src/organizations/components/Dashboards.tsx\n+++ b/ui/src/organizations/components/Dashboards.tsx\n@@ -6,13 +6,10 @@ import _ from 'lodash'\n \n // Components\n import DashboardsIndexContents from 'src/dashboards/components/dashboard_index/DashboardsIndexContents'\n-import {OverlayTechnology, Input, Tabs} from 'src/clockface'\n+import {Input, Tabs} from 'src/clockface'\n import {Button, ComponentColor, IconFont} from '@influxdata/clockface'\n import ImportDashboardOverlay from 'src/dashboards/components/ImportDashboardOverlay'\n \n-// Utils\n-import {getDeep} from 'src/utils/wrappers'\n-\n // APIs\n import {createDashboard, cloneDashboard} from 'src/dashboards/apis/v2/'\n \n@@ -39,7 +36,7 @@ import {DEFAULT_DASHBOARD_NAME} from 'src/dashboards/constants/index'\n \n // Types\n import {Notification} from 'src/types/notifications'\n-import {Links, Cell, Dashboard, AppState, Organization} from 'src/types/v2'\n+import {Links, Dashboard, AppState, Organization} from 'src/types/v2'\n \n // Decorators\n import {ErrorHandling} from 'src/shared/decorators/errors'\n@@ -205,46 +202,20 @@ class Dashboards extends PureComponent<Props, State> {\n this.props.handleDeleteDashboard(dashboard)\n }\n \n- private handleImportDashboard = async (\n- dashboard: Dashboard\n- ): Promise<void> => {\n- const defaultCell = {\n- x: 0,\n- y: 0,\n- w: 4,\n- h: 4,\n- }\n-\n- const name = _.get(dashboard, 'name', DEFAULT_DASHBOARD_NAME)\n- const cellsWithDefaultsApplied = getDeep<Cell[]>(\n- dashboard,\n- 'cells',\n- []\n- ).map(c => ({...defaultCell, ...c}))\n-\n- await this.props.handleImportDashboard({\n- ...dashboard,\n- name,\n- cells: cellsWithDefaultsApplied,\n- })\n- }\n-\n private handleToggleOverlay = (): void => {\n this.setState({isImportingDashboard: !this.state.isImportingDashboard})\n }\n \n private get renderImportOverlay(): JSX.Element {\n- const {notify} = this.props\n const {isImportingDashboard} = this.state\n+ const {orgs} = this.props\n \n return (\n- <OverlayTechnology visible={isImportingDashboard}>\n- <ImportDashboardOverlay\n- onDismissOverlay={this.handleToggleOverlay}\n- onImportDashboard={this.handleImportDashboard}\n- notify={notify}\n- />\n- </OverlayTechnology>\n+ <ImportDashboardOverlay\n+ onDismissOverlay={this.handleToggleOverlay}\n+ orgID={_.get(orgs, '0.id', '')}\n+ isVisible={isImportingDashboard}\n+ />\n )\n }\n }\ndiff --git a/ui/src/shared/components/ImportOverlay.tsx b/ui/src/shared/components/ImportOverlay.tsx\nindex 10f1d50..476fa70 100644\n--- a/ui/src/shared/components/ImportOverlay.tsx\n+++ b/ui/src/shared/components/ImportOverlay.tsx\n@@ -10,6 +10,7 @@ import {\n OverlayHeading,\n OverlayFooter,\n Radio,\n+ ComponentStatus,\n } from 'src/clockface'\n import {Button, ComponentColor} from '@influxdata/clockface'\n \n@@ -93,6 +94,7 @@ export default class ImportOverlay extends PureComponent<Props, State> {\n submitText=\"Upload\"\n handleSubmit={this.handleSetImportContent}\n submitOnDrop={true}\n+ submitOnUpload={true}\n onCancel={this.clearImportContent}\n />\n )\n@@ -110,18 +112,21 @@ export default class ImportOverlay extends PureComponent<Props, State> {\n private get submitButton(): JSX.Element {\n const {resourceName} = this.props\n const {selectedImportOption, importContent} = this.state\n- if (\n+ const isEnabled =\n selectedImportOption === ImportOption.Paste ||\n (selectedImportOption === ImportOption.Upload && importContent)\n- ) {\n- return (\n- <Button\n- text={`Import JSON as ${resourceName}`}\n- onClick={this.submit}\n- color={ComponentColor.Primary}\n- />\n- )\n- }\n+ const status = isEnabled\n+ ? ComponentStatus.Default\n+ : ComponentStatus.Disabled\n+\n+ return (\n+ <Button\n+ text={`Import JSON as ${resourceName}`}\n+ onClick={this.submit}\n+ color={ComponentColor.Primary}\n+ status={status}\n+ />\n+ )\n }\n \n private submit = () => {\ndiff --git a/ui/src/shared/copy/notifications.ts b/ui/src/shared/copy/notifications.ts\nindex 4492078..86700e2 100644\n--- a/ui/src/shared/copy/notifications.ts\n+++ b/ui/src/shared/copy/notifications.ts\n@@ -799,9 +799,18 @@ export const importTaskSucceeded = (): Notification => ({\n })\n \n export const importTaskFailed = (error: string): Notification => ({\n- ...defaultSuccessNotification,\n+ ...defaultErrorNotification,\n message: `Failed to import task: ${error}`,\n })\n+export const importDashboardSucceeded = (): Notification => ({\n+ ...defaultSuccessNotification,\n+ message: `Successfully imported dashboard.`,\n+})\n+\n+export const importDashboardFailed = (error: string): Notification => ({\n+ ...defaultErrorNotification,\n+ message: `Failed to import dashboard: ${error}`,\n+})\n \n // Labels\n export const getLabelsFailed = (): Notification => ({\n"]
1
["9114362b39f5194209cd0b330af7076333f3db77"]
["feat"]
lint README
["diff --git a/README.md b/README.md\nindex a163c83..9cd12bc 100644\n--- a/README.md\n+++ b/README.md\n@@ -26,7 +26,7 @@ Ibis has three primary components:\n Ibis aims to be a future-proof solution to interacting with data using Python and can accomplish this goal through its main features:\n \n - **Familiar API**: Ibis\u2019s API design borrows from popular APIs like pandas and dplyr that most users already know and like to use.\n-- **Consistent syntax**: Ibis aims to be universal Python API for tabular data, big or small.\n+- **Consistent syntax**: Ibis aims to be universal Python API for tabular data, big or small.\n - **Deferred execution**: Ibis pushes code execution to the query engine and only moves required data into memory when it has to.\n This leads to more faster, more efficient analytics workflows\n - **Interactive mode**: Ibis also provides an interactive mode, in which users can quickly diagnose problems, do exploratory data analysis, and mock up workflows locally.\n"]
1
["cab2cf4d3fffb0ec2b56d455e67ac5fa992b4104"]
["docs"]
convert to record
["diff --git a/broker/src/test/java/io/camunda/zeebe/broker/exporter/stream/ExporterDirectorDistributionTest.java b/broker/src/test/java/io/camunda/zeebe/broker/exporter/stream/ExporterDirectorDistributionTest.java\nindex cc998c6..65c8550 100755\n--- a/broker/src/test/java/io/camunda/zeebe/broker/exporter/stream/ExporterDirectorDistributionTest.java\n+++ b/broker/src/test/java/io/camunda/zeebe/broker/exporter/stream/ExporterDirectorDistributionTest.java\n@@ -167,13 +167,8 @@ public final class ExporterDirectorDistributionTest {\n * <p>This makes sure that even if we miss one export position event, we distribute the event\n * later again, which makes tests less flaky.\n */\n- private static final class ClockShifter implements ConditionEvaluationListener<Void> {\n-\n- private final ControlledActorClock clock;\n-\n- public ClockShifter(final ControlledActorClock clock) {\n- this.clock = clock;\n- }\n+ private record ClockShifter(ControlledActorClock clock)\n+ implements ConditionEvaluationListener<Void> {\n \n @Override\n public void conditionEvaluated(final EvaluatedCondition<Void> condition) {\n"]
1
["3346331a963766c8193170fb130adad2e658ada2"]
["refactor"]
don't delay rendering if initialLayout is not specified
["diff --git a/packages/react-native-tab-view/example/src/BottomBarIconTextExample.js b/packages/react-native-tab-view/example/src/BottomBarIconTextExample.js\nindex fcc4708..397e6e6 100644\n--- a/packages/react-native-tab-view/example/src/BottomBarIconTextExample.js\n+++ b/packages/react-native-tab-view/example/src/BottomBarIconTextExample.js\n@@ -1,7 +1,7 @@\n /* @flow */\n \n import React, { Component } from 'react';\n-import { Animated, View, Text, Dimensions, StyleSheet } from 'react-native';\n+import { Animated, View, Text, StyleSheet } from 'react-native';\n import { TabViewAnimated, TabBar } from 'react-native-tab-view';\n import { Ionicons } from '@exponent/vector-icons';\n \n@@ -13,7 +13,6 @@ const styles = StyleSheet.create({\n backgroundColor: '#222',\n },\n tab: {\n- opacity: 1,\n padding: 0,\n },\n icon: {\n@@ -50,11 +49,6 @@ const styles = StyleSheet.create({\n },\n });\n \n-const initialLayout = {\n- height: 0,\n- width: Dimensions.get('window').width,\n-};\n-\n export default class TopBarIconExample extends Component {\n \n static title = 'Bottom bar with indicator';\n@@ -80,14 +74,16 @@ export default class TopBarIconExample extends Component {\n };\n \n _renderIndicator = (props) => {\n- const { width, position } = props;\n+ const { width, opacity, position } = props;\n \n- const translateX = Animated.multiply(position, new Animated.Value(width));\n+ const translateX = Animated.multiply(position, width);\n \n return (\n <Animated.View\n- style={[ styles.indicator, { width: width - 8, transform: [ { translateX } ] } ]}\n- />\n+ style={[ styles.container, { width, opacity, transform: [ { translateX } ] } ]}\n+ >\n+ <View style={styles.indicator} />\n+ </Animated.View>\n );\n };\n \n@@ -146,7 +142,6 @@ export default class TopBarIconExample extends Component {\n renderScene={this._renderScene}\n renderFooter={this._renderFooter}\n onRequestChangeTab={this._handleChangeTab}\n- initialLayout={initialLayout}\n />\n );\n }\ndiff --git a/packages/react-native-tab-view/example/src/CoverflowExample.js b/packages/react-native-tab-view/example/src/CoverflowExample.js\nindex 8950c0e..2336591 100644\n--- a/packages/react-native-tab-view/example/src/CoverflowExample.js\n+++ b/packages/react-native-tab-view/example/src/CoverflowExample.js\n@@ -2,7 +2,7 @@\n /* eslint-disable import/no-commonjs */\n \n import React, { Component } from 'react';\n-import { Animated, View, Image, Text, Dimensions, StyleSheet } from 'react-native';\n+import { Animated, View, Image, Text, StyleSheet } from 'react-native';\n import { TabViewAnimated, TabViewPagerPan } from 'react-native-tab-view';\n \n const styles = StyleSheet.create({\n@@ -48,11 +48,6 @@ const ALBUMS = {\n 'Lost Horizons': require('../assets/album-art-8.jpg'),\n };\n \n-const initialLayout = {\n- height: 0,\n- width: Dimensions.get('window').width,\n-};\n-\n export default class CoverflowExample extends Component {\n \n static title = 'Coverflow';\n@@ -142,7 +137,6 @@ export default class CoverflowExample extends Component {\n renderPager={this._renderPager}\n renderScene={this._renderScene}\n onRequestChangeTab={this._handleChangeTab}\n- initialLayout={initialLayout}\n />\n );\n }\ndiff --git a/packages/react-native-tab-view/example/src/ScrollViewsExample.js b/packages/react-native-tab-view/example/src/ScrollViewsExample.js\nindex 94fefbb..5be3b69 100644\n--- a/packages/react-native-tab-view/example/src/ScrollViewsExample.js\n+++ b/packages/react-native-tab-view/example/src/ScrollViewsExample.js\n@@ -28,11 +28,6 @@ const styles = StyleSheet.create({\n },\n });\n \n-const initialLayout = {\n- height: 0,\n- width: Dimensions.get('window').width,\n-};\n-\n export default class TopBarTextExample extends Component {\n \n static title = 'Scroll views';\n@@ -104,6 +99,7 @@ export default class TopBarTextExample extends Component {\n renderLabel={this._renderLabel(props)}\n indicatorStyle={styles.indicator}\n tabStyle={styles.tab}\n+ tabWidth={80}\n style={styles.tabbar}\n />\n );\n@@ -130,7 +126,6 @@ export default class TopBarTextExample extends Component {\n renderScene={this._renderScene}\n renderHeader={this._renderHeader}\n onRequestChangeTab={this._handleChangeTab}\n- initialLayout={initialLayout}\n />\n );\n }\ndiff --git a/packages/react-native-tab-view/example/src/TopBarIconExample.js b/packages/react-native-tab-view/example/src/TopBarIconExample.js\nindex d13755f..5464981 100644\n--- a/packages/react-native-tab-view/example/src/TopBarIconExample.js\n+++ b/packages/react-native-tab-view/example/src/TopBarIconExample.js\n@@ -1,7 +1,7 @@\n /* @flow */\n \n import React, { Component } from 'react';\n-import { View, Dimensions, StyleSheet } from 'react-native';\n+import { View, StyleSheet } from 'react-native';\n import { TabViewAnimated, TabBarTop } from 'react-native-tab-view';\n import { Ionicons } from '@exponent/vector-icons';\n \n@@ -22,11 +22,6 @@ const styles = StyleSheet.create({\n },\n });\n \n-const initialLayout = {\n- height: 0,\n- width: Dimensions.get('window').width,\n-};\n-\n export default class TopBarIconExample extends Component {\n \n static title = 'Icon only top bar';\n@@ -93,7 +88,6 @@ export default class TopBarIconExample extends Component {\n renderScene={this._renderScene}\n renderHeader={this._renderHeader}\n onRequestChangeTab={this._handleChangeTab}\n- initialLayout={initialLayout}\n />\n );\n }\ndiff --git a/packages/react-native-tab-view/example/src/TopBarTextExample.js b/packages/react-native-tab-view/example/src/TopBarTextExample.js\nindex 30307ad..454533d 100644\n--- a/packages/react-native-tab-view/example/src/TopBarTextExample.js\n+++ b/packages/react-native-tab-view/example/src/TopBarTextExample.js\n@@ -1,7 +1,7 @@\n /* @flow */\n \n import React, { Component } from 'react';\n-import { View, Dimensions, StyleSheet } from 'react-native';\n+import { View, StyleSheet } from 'react-native';\n import { TabViewAnimated, TabBarTop } from 'react-native-tab-view';\n \n const styles = StyleSheet.create({\n@@ -25,11 +25,6 @@ const styles = StyleSheet.create({\n },\n });\n \n-const initialLayout = {\n- height: 0,\n- width: Dimensions.get('window').width,\n-};\n-\n export default class TopBarTextExample extends Component {\n \n static title = 'Scrollable top bar';\n@@ -90,7 +85,6 @@ export default class TopBarTextExample extends Component {\n renderScene={this._renderScene}\n renderHeader={this._renderHeader}\n onRequestChangeTab={this._handleChangeTab}\n- initialLayout={initialLayout}\n />\n );\n }\ndiff --git a/packages/react-native-tab-view/src/TabBar.js b/packages/react-native-tab-view/src/TabBar.js\nindex 615e85a..a03d8e5 100644\n--- a/packages/react-native-tab-view/src/TabBar.js\n+++ b/packages/react-native-tab-view/src/TabBar.js\n@@ -92,6 +92,7 @@ type Props = SceneRendererProps & {\n \n type State = {\n offset: Animated.Value;\n+ visibility: Animated.Value;\n }\n \n export default class TabBar extends Component<DefaultProps, Props, State> {\n@@ -115,8 +116,15 @@ export default class TabBar extends Component<DefaultProps, Props, State> {\n \n state: State = {\n offset: new Animated.Value(0),\n+ visibility: new Animated.Value(0),\n };\n \n+ componentWillMount() {\n+ if (this.props.layout.width || this.props.tabWidth) {\n+ this.state.visibility.setValue(1);\n+ }\n+ }\n+\n componentDidMount() {\n this._adjustScroll(this.props.navigationState.index);\n this._positionListener = this.props.subscribe('position', this._adjustScroll);\n@@ -126,6 +134,16 @@ export default class TabBar extends Component<DefaultProps, Props, State> {\n if (this.props.navigationState !== nextProps.navigationState) {\n this._resetScrollOffset(nextProps);\n }\n+\n+ if (\n+ (this.props.tabWidth !== nextProps.tabWidth && nextProps.tabWidth) ||\n+ (this.props.layout.width !== nextProps.layout.width && nextProps.layout.width)\n+ ) {\n+ Animated.timing(this.state.visibility, {\n+ toValue: 1,\n+ duration: 150,\n+ }).start();\n+ }\n }\n \n componentWillUnmount() {\n@@ -282,7 +300,8 @@ export default class TabBar extends Component<DefaultProps, Props, State> {\n {this.props.renderIndicator ?\n this.props.renderIndicator({\n ...this.props,\n- width: tabWidth,\n+ width: new Animated.Value(tabWidth),\n+ opacity: this.state.visibility,\n }) :\n null\n }\n@@ -307,10 +326,10 @@ export default class TabBar extends Component<DefaultProps, Props, State> {\n {routes.map((route, i) => {\n const focused = index === i;\n const outputRange = inputRange.map(inputIndex => inputIndex === i ? 1 : 0.7);\n- const opacity = position.interpolate({\n+ const opacity = Animated.multiply(this.state.visibility, position.interpolate({\n inputRange,\n outputRange,\n- });\n+ }));\n const scene = {\n route,\n focused,\n@@ -348,14 +367,14 @@ export default class TabBar extends Component<DefaultProps, Props, State> {\n }}\n >\n <View style={styles.container}>\n- <Animated.View style={[ styles.tabitem, { opacity, width: tabWidth }, tabStyle, this.props.tabStyle ]}>\n+ <Animated.View style={[ styles.tabitem, { opacity }, tabWidth ? { width: tabWidth } : null, tabStyle, this.props.tabStyle ]}>\n {icon}\n {label}\n </Animated.View>\n {badge ?\n- <View style={styles.badge}>\n+ <Animated.View style={[ styles.badge, { opacity: this.state.visibility } ]}>\n {badge}\n- </View> : null\n+ </Animated.View> : null\n }\n </View>\n </TouchableItem>\ndiff --git a/packages/react-native-tab-view/src/TabBarTop.js b/packages/react-native-tab-view/src/TabBarTop.js\nindex 0960d4e..84dd6e2 100644\n--- a/packages/react-native-tab-view/src/TabBarTop.js\n+++ b/packages/react-native-tab-view/src/TabBarTop.js\n@@ -28,7 +28,8 @@ const styles = StyleSheet.create({\n });\n \n type IndicatorProps = SceneRendererProps & {\n- width: number;\n+ width: Animated.Valye;\n+ opacity: Animated.Value;\n }\n \n type Props = SceneRendererProps & {\n@@ -50,13 +51,13 @@ export default class TabBarTop extends Component<void, Props, void> {\n );\n \n _renderIndicator = (props: IndicatorProps) => {\n- const { width, position } = props;\n+ const { width, opacity, position } = props;\n \n- const translateX = Animated.multiply(position, new Animated.Value(width));\n+ const translateX = Animated.multiply(position, width);\n \n return (\n <Animated.View\n- style={[ styles.indicator, { width, transform: [ { translateX } ] }, this.props.indicatorStyle ]}\n+ style={[ styles.indicator, { width, opacity, transform: [ { translateX } ] }, this.props.indicatorStyle ]}\n />\n );\n };\ndiff --git a/packages/react-native-tab-view/src/TabViewAnimated.js b/packages/react-native-tab-view/src/TabViewAnimated.js\nindex d484816..4499748 100644\n--- a/packages/react-native-tab-view/src/TabViewAnimated.js\n+++ b/packages/react-native-tab-view/src/TabViewAnimated.js\n@@ -94,19 +94,17 @@ export default class TabViewAnimated extends Component<DefaultProps, Props, Stat\n };\n \n _renderItems = (props: SceneRendererProps) => {\n- if (props.layout.width === 0) {\n- return null;\n- }\n-\n const { renderPager, renderHeader, renderFooter } = this.props;\n+ const { navigationState, layout } = props;\n+ const currentRoute = navigationState.routes[navigationState.index];\n \n return (\n <View style={styles.container}>\n {renderHeader && renderHeader(props)}\n {renderPager({\n ...props,\n- children: props.navigationState.routes.map((route, index) => (\n- <View key={route.key} style={{ width: props.layout.width }}>\n+ children: layout.width ? navigationState.routes.map((route, index) => (\n+ <View key={route.key} style={{ width: layout.width }}>\n {this._renderScene({\n ...props,\n route,\n@@ -114,7 +112,16 @@ export default class TabViewAnimated extends Component<DefaultProps, Props, Stat\n focused: index === props.navigationState.index,\n })}\n </View>\n- )),\n+ )) : (\n+ <View key={currentRoute.key} style={styles.container}>\n+ {this._renderScene({\n+ ...props,\n+ route: currentRoute,\n+ index: navigationState.index,\n+ focused: true,\n+ })}\n+ </View>\n+ ),\n })}\n {renderFooter && renderFooter(props)}\n </View>\n"]
1
["e9233ae3f7811707945fc2de60971595d83c578d"]
["fix"]
switch to callback ref
["diff --git a/src/notebook/components/transforms/html.js b/src/notebook/components/transforms/html.js\nindex 83fc1fb..021cc65 100644\n--- a/src/notebook/components/transforms/html.js\n+++ b/src/notebook/components/transforms/html.js\n@@ -8,16 +8,16 @@ type Props = {\n \n export default class HTMLDisplay extends React.Component {\n props: Props;\n+ el: HTMLElement;\n \n componentDidMount(): void {\n- if (this.refs.here) {\n- if (document.createRange && Range && Range.prototype.createContextualFragment) {\n- const range = document.createRange();\n- const fragment = range.createContextualFragment(this.props.data);\n- ReactDOM.findDOMNode(this.refs.here).appendChild(fragment);\n- } else {\n- ReactDOM.findDOMNode(this.refs.here).innerHTML = this.props.data;\n- }\n+ // Create a range to ensure that scripts are invoked from within the HTML\n+ if (document.createRange && Range && Range.prototype.createContextualFragment) {\n+ const range = document.createRange();\n+ const fragment = range.createContextualFragment(this.props.data);\n+ this.el.appendChild(fragment);\n+ } else {\n+ this.el.innerHTML = this.props.data;\n }\n }\n \n@@ -27,7 +27,7 @@ export default class HTMLDisplay extends React.Component {\n \n render(): ?React.Element<any> {\n return (\n- <div ref=\"here\" />\n+ <div ref={(el) => { this.el = el; }} />\n );\n }\n }\n"]
1
["ee4bf61fb8836e249fb4ef3507dc938e70696b3f"]
["refactor"]
wire up fixed null encoding
["diff --git a/read_buffer/src/row_group.rs b/read_buffer/src/row_group.rs\nindex 91c9fb5..ca77f3c 100644\n--- a/read_buffer/src/row_group.rs\n+++ b/read_buffer/src/row_group.rs\n@@ -958,24 +958,15 @@ impl From<RecordBatch> for RowGroup {\n }\n Some(InfluxColumnType::Field(_)) => {\n let column_data = match arrow_column.data_type() {\n- arrow::datatypes::DataType::Int64 => Column::from(\n- arrow_column\n- .as_any()\n- .downcast_ref::<arrow::array::Int64Array>()\n- .unwrap(),\n- ),\n- arrow::datatypes::DataType::Float64 => Column::from(\n- arrow_column\n- .as_any()\n- .downcast_ref::<arrow::array::Float64Array>()\n- .unwrap(),\n- ),\n- arrow::datatypes::DataType::UInt64 => Column::from(\n- arrow_column\n- .as_any()\n- .downcast_ref::<arrow::array::UInt64Array>()\n- .unwrap(),\n- ),\n+ arrow::datatypes::DataType::Int64 => {\n+ Column::from(arrow::array::Int64Array::from(arrow_column.data()))\n+ }\n+ arrow::datatypes::DataType::Float64 => {\n+ Column::from(arrow::array::Float64Array::from(arrow_column.data()))\n+ }\n+ arrow::datatypes::DataType::UInt64 => {\n+ Column::from(arrow::array::UInt64Array::from(arrow_column.data()))\n+ }\n dt => unimplemented!(\n \"data type {:?} currently not supported for field columns\",\n dt\n"]
1
["28b596b8834d1b51be3ac6a2ac30df28f37702d8"]
["refactor"]
tests should pass now Make the code more safe
["diff --git a/goreleaserlib/goreleaser.go b/goreleaserlib/goreleaser.go\nindex 28ba3f4..67ba95d 100644\n--- a/goreleaserlib/goreleaser.go\n+++ b/goreleaserlib/goreleaser.go\n@@ -27,6 +27,15 @@ import (\n \tyaml \"gopkg.in/yaml.v2\"\n )\n \n+var (\n+\tnormalPadding = cli.Default.Padding\n+\tincreasedPadding = normalPadding * 2\n+)\n+\n+func init() {\n+\tlog.SetHandler(cli.Default)\n+}\n+\n var pipes = []pipeline.Piper{\n \tdefaults.Pipe{}, // load default configs\n \tgit.Pipe{}, // get and validate git repo state\n@@ -89,17 +98,15 @@ func Release(flags Flags) error {\n \t\tctx.Publish = false\n \t}\n \tctx.RmDist = flags.Bool(\"rm-dist\")\n-\tlogger, _ := log.Log.(*log.Logger)\n-\thandler, _ := logger.Handler.(*cli.Handler)\n \tfor _, pipe := range pipes {\n-\t\thandler.Padding = 3\n+\t\tcli.Default.Padding = normalPadding\n \t\tlog.Infof(\"\\033[1m%s\\033[0m\", strings.ToUpper(pipe.String()))\n-\t\thandler.Padding = 6\n+\t\tcli.Default.Padding = increasedPadding\n \t\tif err := handle(pipe.Run(ctx)); err != nil {\n \t\t\treturn err\n \t\t}\n \t}\n-\thandler.Padding = 3\n+\tcli.Default.Padding = normalPadding\n \treturn nil\n }\n \ndiff --git a/main.go b/main.go\nindex b9b961d..7ced1dd 100644\n--- a/main.go\n+++ b/main.go\n@@ -18,7 +18,7 @@ var (\n )\n \n func init() {\n-\tlog.SetHandler(lcli.New(os.Stdout))\n+\tlog.SetHandler(lcli.Default)\n }\n \n func main() {\n"]
1
["5636313d7c9cfbd9f48578fd104771d65eae9720"]
["fix"]
disable edit/delete if primary key missing Signed-off-by: Pranav C <[email protected]>
["diff --git a/packages/nc-gui/components/project/spreadsheet/components/expandedForm.vue b/packages/nc-gui/components/project/spreadsheet/components/expandedForm.vue\nindex 5f9841f..c414c8c 100644\n--- a/packages/nc-gui/components/project/spreadsheet/components/expandedForm.vue\n+++ b/packages/nc-gui/components/project/spreadsheet/components/expandedForm.vue\n@@ -413,6 +413,9 @@ export default {\n \n await this.reload()\n } else if (Object.keys(updatedObj).length) {\n+ if (!id) {\n+ return this.$toast.info('Update not allowed for table which doesn\\'t have primary Key').goAway(3000)\n+ }\n await this.api.update(id, updatedObj, this.oldRow)\n } else {\n return this.$toast.info('No columns to update').goAway(3000)\ndiff --git a/packages/nc-gui/components/project/spreadsheet/rowsXcDataTable.vue b/packages/nc-gui/components/project/spreadsheet/rowsXcDataTable.vue\nindex c2b4b81..1b9d6a0 100644\n--- a/packages/nc-gui/components/project/spreadsheet/rowsXcDataTable.vue\n+++ b/packages/nc-gui/components/project/spreadsheet/rowsXcDataTable.vue\n@@ -62,7 +62,15 @@\n <v-spacer class=\"h-100\" @dblclick=\"debug=true\" />\n \n <debug-metas v-if=\"debug\" class=\"mr-3\" />\n-\n+ <v-tooltip bottom>\n+ <template #activator=\"{on}\">\n+ <v-icon v-if=\"!isPkAvail\" color=\"warning\" small class=\"mr-3\" v-on=\"on\">\n+ mdi-information-outline\n+ </v-icon>\n+ </template>\n+ <span class=\"caption\"> Update & Delete not allowed since the table doesn't have any primary key\n+ </span>\n+ </v-tooltip>\n <lock-menu v-if=\"_isUIAllowed('view-type')\" v-model=\"viewStatus.type\" />\n <x-btn tooltip=\"Reload view data\" outlined small text @click=\"reload\">\n <v-icon small class=\"mr-1\" color=\"grey darken-3\">\n@@ -208,6 +216,7 @@\n :meta=\"meta\"\n :is-virtual=\"selectedView.type === 'vtable'\"\n :api=\"api\"\n+ :is-pk-avail=\"isPkAvail\"\n @onNewColCreation=\"onNewColCreation\"\n @onCellValueChange=\"onCellValueChange\"\n @insertNewRow=\"insertNewRow\"\n@@ -631,8 +640,8 @@ export default {\n if (\n !this.meta || (\n (this.meta.hasMany && this.meta.hasMany.length) ||\n- (this.meta.manyToMany && this.meta.manyToMany.length) ||\n- (this.meta.belongsTo && this.meta.belongsTo.length))\n+ (this.meta.manyToMany && this.meta.manyToMany.length) ||\n+ (this.meta.belongsTo && this.meta.belongsTo.length))\n ) {\n return this.$toast.info('Please delete relations before deleting table.').goAway(3000)\n }\n@@ -817,6 +826,10 @@ export default {\n \n const id = this.meta.columns.filter(c => c.pk).map(c => rowObj[c._cn]).join('___')\n \n+ if (!id) {\n+ return this.$toast.info('Update not allowed for table which doesn\\'t have primary Key').goAway(3000)\n+ }\n+\n const newData = await this.api.update(id, {\n [column._cn]: rowObj[column._cn]\n }, { [column._cn]: oldRow[column._cn] })\n@@ -841,6 +854,11 @@ export default {\n const rowObj = this.rowContextMenu.row\n if (!this.rowContextMenu.rowMeta.new) {\n const id = this.meta && this.meta.columns && this.meta.columns.filter(c => c.pk).map(c => rowObj[c._cn]).join('___')\n+\n+ if (!id) {\n+ return this.$toast.info('Delete not allowed for table which doesn\\'t have primary Key').goAway(3000)\n+ }\n+\n await this.api.delete(id)\n }\n this.data.splice(this.rowContextMenu.index, 1)\n@@ -859,6 +877,11 @@ export default {\n }\n if (!rowMeta.new) {\n const id = this.meta.columns.filter(c => c.pk).map(c => rowObj[c._cn]).join('___')\n+\n+ if (!id) {\n+ return this.$toast.info('Delete not allowed for table which doesn\\'t have primary Key').goAway(3000)\n+ }\n+\n await this.api.delete(id)\n }\n this.data.splice(row, 1)\n@@ -991,6 +1014,9 @@ export default {\n }\n },\n computed: {\n+ isPkAvail() {\n+ return this.meta && this.meta.columns.some(c => c.pk)\n+ },\n isGallery() {\n return this.selectedView && this.selectedView.show_as === 'gallery'\n },\ndiff --git a/packages/nc-gui/components/project/spreadsheet/views/xcGridView.vue b/packages/nc-gui/components/project/spreadsheet/views/xcGridView.vue\nindex 5497d05..c198784 100644\n--- a/packages/nc-gui/components/project/spreadsheet/views/xcGridView.vue\n+++ b/packages/nc-gui/components/project/spreadsheet/views/xcGridView.vue\n@@ -27,7 +27,7 @@\n @xcresized=\"resizingCol = null\"\n >\n <!-- :style=\"columnsWidth[col._cn] ? `min-width:${columnsWidth[col._cn]}; max-width:${columnsWidth[col._cn]}` : ''\"\n- -->\n+-->\n \n <virtual-header-cell\n v-if=\"col.virtual\"\n@@ -136,13 +136,13 @@\n :key=\"row + columnObj.alias\"\n class=\"cell pointer\"\n :class=\"{\n- 'active' : !isPublicView && selected.col === col && selected.row === row && isEditable ,\n+ 'active' :!isPublicView && selected.col === col && selected.row === row && isEditable ,\n 'primary-column' : primaryValueColumn === columnObj._cn,\n 'text-center': isCentrallyAligned(columnObj),\n 'required': isRequired(columnObj,rowObj)\n }\"\n :data-col=\"columnObj.alias\"\n- @dblclick=\"makeEditable(col,row,columnObj.ai)\"\n+ @dblclick=\"makeEditable(col,row,columnObj.ai,rowMeta)\"\n @click=\"makeSelected(col,row);\"\n @contextmenu=\"showRowContextMenu($event,rowObj,rowMeta,row,col, columnObj)\"\n >\n@@ -162,7 +162,8 @@\n \n <editable-cell\n v-else-if=\"\n- !isLocked\n+ (isPkAvail ||rowMeta.new) &&\n+ !isLocked\n && !isPublicView\n && (editEnabled.col === col && editEnabled.row === row)\n || enableEditable(columnObj)\n@@ -190,11 +191,11 @@\n :db-alias=\"nodes.dbAlias\"\n :value=\"rowObj[columnObj._cn]\"\n :sql-ui=\"sqlUi\"\n- @enableedit=\"makeSelected(col,row);makeEditable(col,row,columnObj.ai)\"\n+ @enableedit=\"makeSelected(col,row);makeEditable(col,row,columnObj.ai, rowMeta)\"\n />\n </td>\n </tr>\n- <tr v-if=\"!isLocked && !isPublicView && isEditable && relationType !== 'bt'\">\n+ <tr v-if=\"isPkAvail && !isLocked && !isPublicView && isEditable && relationType !== 'bt'\">\n <td :colspan=\"visibleColLength + 1\" class=\"text-left pointer\" @click=\"insertNewRow(true)\">\n <v-tooltip top>\n <template #activator=\"{on}\">\n@@ -214,7 +215,9 @@\n <!-- <div is=\"style\" v-html=\"resizeColStyle\" />-->\n <dynamic-style>\n <template v-if=\"resizingCol\">\n- [data-col=\"{{ resizingCol }}\"]{min-width:{{ resizingColWidth }};max-width:{{ resizingColWidth }};width:{{ resizingColWidth }};}\n+ [data-col=\"{{ resizingCol }}\"]{min-width:{{ resizingColWidth }};max-width:{{\n+ resizingColWidth\n+ }};width:{{ resizingColWidth }};}\n </template>\n </dynamic-style>\n </div>\n@@ -261,7 +264,8 @@ export default {\n table: String,\n isVirtual: Boolean,\n isLocked: Boolean,\n- columnsWidth: { type: Object }\n+ columnsWidth: { type: Object },\n+ isPkAvail: Boolean\n },\n data: () => ({\n resizingCol: null,\n@@ -426,6 +430,10 @@ export default {\n return\n }\n if (e.key && e.key.length === 1) {\n+ if (!this.isPkAvail && !this.data[this.selected.row].rowMeta.new) {\n+ return this.$toast.info('Update not allowed for table which doesn\\'t have primary Key').goAway(3000)\n+ }\n+\n this.$set(this.data[this.selected.row].row, this.availableColumns[this.selected.col]._cn, '')\n this.editEnabled = { ...this.selected }\n }\n@@ -466,10 +474,14 @@ export default {\n this.editEnabled = {}\n }\n },\n- makeEditable(col, row) {\n+ makeEditable(col, row, _, rowMeta) {\n if (this.isPublicView || !this.isEditable) {\n return\n }\n+\n+ if (!this.isPkAvail && !rowMeta.new) {\n+ return this.$toast.info('Update not allowed for table which doesn\\'t have primary Key').goAway(3000)\n+ }\n if (this.availableColumns[col].ai) {\n return this.$toast.info('Auto Increment field is not editable').goAway(3000)\n }\n"]
1
["4d92f352741b04c8709319dfe5c8419654f3682c"]
["feat"]
update the formatting for python integration example
["diff --git a/website/docs/integration/python.md b/website/docs/integration/python.md\nindex 064cae3..b6b720d 100644\n--- a/website/docs/integration/python.md\n+++ b/website/docs/integration/python.md\n@@ -13,6 +13,7 @@ header = \"All notable changes to this project will be documented in this file.\"\n body = \"...\"\n footer = \"<!-- generated by git-cliff -->\"\n # see [changelog] section for more keys\n+\n [tool.git-cliff.git]\n conventional_commits = true\n commit_parsers = []\n"]
1
["3ee672483790ec71c700907a6e93af4698492026"]
["docs"]
add travis file
["diff --git a/.travis.yml b/.travis.yml\nnew file mode 100644\nindex 0000000..c08cc34\n--- /dev/null\n+++ b/.travis.yml\n@@ -0,0 +1,11 @@\n+sudo: false\n+\n+language: java\n+jdk: oraclejdk8\n+\n+branches:\n+ only:\n+ - master\n+\n+notifications:\n+ email: false\n"]
1
["d0814a928601706635287fd3d9d3451d156b821a"]
["ci"]
add documentation to use react-native-paper with CRA (#874)
["diff --git a/docs/pages/4.react-native-web.md b/docs/pages/4.react-native-web.md\nindex 69e4e52..8d6ae2a 100644\n--- a/docs/pages/4.react-native-web.md\n+++ b/docs/pages/4.react-native-web.md\n@@ -16,6 +16,63 @@ To install `react-native-web`, run:\n yarn add react-native-web react-dom react-art\n ```\n \n+### Using CRA ([Create React App](https://github.com/facebook/create-react-app))\n+\n+Install [`react-app-rewired`](https://github.com/timarney/react-app-rewired) to override `webpack` configuration:\n+\n+```sh\n+yarn add --dev react-app-rewired\n+```\n+\n+[Configure `babel-loader`](#2-configure-babel-loader) using a new file called `config-overrides.js`:\n+\n+```js\n+module.exports = function override(config, env) {\n+ config.module.rules.push({\n+ test: /\\.js$/,\n+ exclude: /node_modules[/\\\\](?!react-native-paper|react-native-vector-icons|react-native-safe-area-view)/,\n+ use: {\n+ loader: \"babel-loader\",\n+ options: {\n+ // Disable reading babel configuration\n+ babelrc: false,\n+ configFile: false,\n+\n+ // The configration for compilation\n+ presets: [\n+ [\"@babel/preset-env\", { useBuiltIns: \"usage\" }],\n+ \"@babel/preset-react\",\n+ \"@babel/preset-flow\"\n+ ],\n+ plugins: [\n+ \"@babel/plugin-proposal-class-properties\",\n+ \"@babel/plugin-proposal-object-rest-spread\"\n+ ]\n+ }\n+ }\n+ });\n+\n+ return config;\n+};\n+```\n+\n+Change your script in `package.json`:\n+\n+```diff\n+/* package.json */\n+\n+ \"scripts\": {\n+- \"start\": \"react-scripts start\",\n++ \"start\": \"react-app-rewired start\",\n+- \"build\": \"react-scripts build\",\n++ \"build\": \"react-app-rewired build\",\n+- \"test\": \"react-scripts test --env=jsdom\",\n++ \"test\": \"react-app-rewired test --env=jsdom\"\n+}\n+```\n+\n+### Custom webpack setup\n+\n To install `webpack`, run:\n \n ```sh\n"]
1
["ee7cc5d5a940fba774e715b1f029c6361110b108"]
["docs"]
add prewatch script to core
["diff --git a/.gitignore b/.gitignore\nindex 3445558..72257cb 100644\n--- a/.gitignore\n+++ b/.gitignore\n@@ -1,4 +1,5 @@\n *.o\n+.env\n settheory\n constraint\n Main\ndiff --git a/packages/core/package.json b/packages/core/package.json\nindex 9ba8b93..79bd532 100644\n--- a/packages/core/package.json\n+++ b/packages/core/package.json\n@@ -14,6 +14,7 @@\n \"build:parsers\": \"nearleyc src/parser/Domain.ne > src/parser/DomainParser.ts && nearleyc src/parser/Substance.ne > src/parser/SubstanceParser.ts && nearleyc src/parser/Style.ne > src/parser/StyleParser.ts\",\n \"prebuild\": \"yarn build:parsers\",\n \"prestart\": \"yarn build:parsers\",\n+ \"prewatch\": \"yarn build:parsers\",\n \"test\": \"jest --watchAll=false\",\n \"test:watch\": \"jest --watchAll\",\n \"build\": \"rollup -c\",\n"]
1
["aa0152baa4376b1087c86499a7c289b668d5ad55"]
["build"]
rename step
["diff --git a/.github/workflows/ci.yml b/.github/workflows/ci.yml\nindex e81d897..5c3ee6b 100644\n--- a/.github/workflows/ci.yml\n+++ b/.github/workflows/ci.yml\n@@ -45,7 +45,7 @@ jobs:\n - name: Install dependencies\n run: pnpm install\n \n- - name: Typecheck\n+ - name: Build (stub)\n run: pnpm build:stub\n \n - name: Typecheck\n"]
1
["34875bc0e59b43d9041903101c823d25ec194a21"]
["ci"]
tests
["diff --git a/client/src/components/Profile/__test__/EducationCard.test.tsx b/client/src/components/Profile/__test__/EducationCard.test.tsx\nindex 44b6e00..14539dd 100644\n--- a/client/src/components/Profile/__test__/EducationCard.test.tsx\n+++ b/client/src/components/Profile/__test__/EducationCard.test.tsx\n@@ -53,7 +53,7 @@ describe('EducationCard', () => {\n });\n \n describe('filterPermissions', () => {\n- it('should left only contacts in \"permissionsSettings\" object', () => {\n+ it('should left only \"isEducationVisible\" in \"permissionsSettings\" object', () => {\n const permissionsSettings = {\n isProfileVisible: { all: true },\n isAboutVisible: { all: true, mentor: true, student: true },\ndiff --git a/client/src/components/Profile/__test__/MainCard.test.tsx b/client/src/components/Profile/__test__/MainCard.test.tsx\nindex 8fb2840..552804b 100644\n--- a/client/src/components/Profile/__test__/MainCard.test.tsx\n+++ b/client/src/components/Profile/__test__/MainCard.test.tsx\n@@ -3,6 +3,8 @@ import { shallow } from 'enzyme';\n import { shallowToJson } from 'enzyme-to-json';\n import MainCard from '../MainCard';\n \n+// TODO: Known Issue: https://stackoverflow.com/questions/59942808/how-can-i-use-jest-coverage-in-next-js-styled-jsx\n+\n describe('MainCard', () => {\n describe('Should render correctly', () => {\n it('if is editing mode disabled', () => {\n@@ -21,49 +23,89 @@ describe('MainCard', () => {\n );\n expect(shallowToJson(output)).toMatchSnapshot();\n });\n+ it('if is editing mode enabled', () => {\n+ const output = shallow(\n+ <MainCard\n+ data={{\n+ name: 'Petr Pervyi',\n+ githubId: 'piter',\n+ locationName: 'SPB',\n+ locationId: '1',\n+ }}\n+ isEditingModeEnabled={true}\n+ onPermissionsSettingsChange={() => {}}\n+ onProfileSettingsChange={() => {}}\n+ />,\n+ );\n+ expect(shallowToJson(output)).toMatchSnapshot();\n+ });\n });\n \n- // const wrapper = shallow(\n- // <MainCard\n- // data={{\n- // name: 'Petr Pervyi',\n- // githubId: 'piter',\n- // locationName: 'SPB',\n- // locationId: '1',\n- // }}\n- // isEditingModeEnabled={false}\n- // onPermissionsSettingsChange={() => {}}\n- // onProfileSettingsChange={() => {}}\n- // />);\n- // const instance = wrapper.instance();\n- // describe('showVisibilitySettings', () => {\n- // it('should set \"state.isVisibilitySettingsVisible\" as \"true\"', () => {\n- // expect(instance.state.isVisibilitySettingsVisible).toBe(false);\n- // instance.showVisibilitySettings();\n- // expect(instance.state.isVisibilitySettingsVisible).toBe(true);\n- // });\n- // });\n- // describe('hideVisibilitySettings', () => {\n- // it('should set \"state.isVisibilitySettingsVisible\" as \"false\"', () => {\n- // instance.state.isVisibilitySettingsVisible = true;\n- // expect(instance.state.isVisibilitySettingsVisible).toBe(true);\n- // instance.hideVisibilitySettings();\n- // expect(instance.state.isVisibilitySettingsVisible).toBe(false);\n- // });\n- // });\n- // describe('showProfileSettings', () => {\n- // it('should set \"state.isProfileSettingsVisible\" as \"true\"', () => {\n- // expect(instance.state.isProfileSettingsVisible).toBe(false);\n- // instance.showProfileSettings();\n- // expect(instance.state.isProfileSettingsVisible).toBe(true);\n- // });\n- // });\n- // describe('hideProfileSettings', () => {\n- // it('should set \"state.isProfileSettingsVisible\" as \"false\"', () => {\n- // instance.state.isProfileSettingsVisible = true;\n- // expect(instance.state.isProfileSettingsVisible).toBe(true);\n- // instance.hideProfileSettings();\n- // expect(instance.state.isProfileSettingsVisible).toBe(false);\n- // });\n- // });\n+ const wrapper = shallow(\n+ <MainCard\n+ data={{\n+ name: 'Petr Pervyi',\n+ githubId: 'piter',\n+ locationName: 'SPB',\n+ locationId: '1',\n+ }}\n+ isEditingModeEnabled={false}\n+ onPermissionsSettingsChange={() => {}}\n+ onProfileSettingsChange={() => {}}\n+ />);\n+ const instance = wrapper.instance();\n+ describe('showVisibilitySettings', () => {\n+ it('should set \"state.isVisibilitySettingsVisible\" as \"true\"', () => {\n+ expect(instance.state.isVisibilitySettingsVisible).toBe(false);\n+ instance.showVisibilitySettings();\n+ expect(instance.state.isVisibilitySettingsVisible).toBe(true);\n+ });\n+ });\n+ describe('hideVisibilitySettings', () => {\n+ it('should set \"state.isVisibilitySettingsVisible\" as \"false\"', () => {\n+ instance.state.isVisibilitySettingsVisible = true;\n+ expect(instance.state.isVisibilitySettingsVisible).toBe(true);\n+ instance.hideVisibilitySettings();\n+ expect(instance.state.isVisibilitySettingsVisible).toBe(false);\n+ });\n+ });\n+ describe('showProfileSettings', () => {\n+ it('should set \"state.isProfileSettingsVisible\" as \"true\"', () => {\n+ expect(instance.state.isProfileSettingsVisible).toBe(false);\n+ instance.showProfileSettings();\n+ expect(instance.state.isProfileSettingsVisible).toBe(true);\n+ });\n+ });\n+ describe('hideProfileSettings', () => {\n+ it('should set \"state.isProfileSettingsVisible\" as \"false\"', () => {\n+ instance.state.isProfileSettingsVisible = true;\n+ expect(instance.state.isProfileSettingsVisible).toBe(true);\n+ instance.hideProfileSettings();\n+ expect(instance.state.isProfileSettingsVisible).toBe(false);\n+ });\n+ });\n+ describe('filterPermissions', () => {\n+ it('should left only \"isProfileVisible\" in \"permissionsSettings\" object', () => {\n+ const permissionsSettings = {\n+ isProfileVisible: { all: true },\n+ isAboutVisible: { all: true, mentor: true, student: true },\n+ isEducationVisible: { all: true, mentor: true, student: true },\n+ isEnglishVisible: { all: false, student: false },\n+ isEmailVisible: { all: true, student: true },\n+ isTelegramVisible: { all: false, student: false },\n+ isSkypeVisible: { all: true, student: true },\n+ isPhoneVisible: { all: false, student: false },\n+ isContactsNotesVisible: { all: true, student: true },\n+ isLinkedInVisible: { all: false, mentor: false, student: false },\n+ isPublicFeedbackVisible: { all: true, mentor: true, student: true },\n+ isMentorStatsVisible: { all: true, mentor: true, student: true },\n+ isStudentStatsVisible: { all: true, student: true },\n+ };\n+ const instance = wrapper.instance();\n+ const result = instance.filterPermissions(permissionsSettings);\n+ expect(result).toEqual({\n+ isProfileVisible: { all: true },\n+ });\n+ });\n+ });\n });\ndiff --git a/client/src/components/Profile/__test__/__snapshots__/MainCard.test.tsx.snap b/client/src/components/Profile/__test__/__snapshots__/MainCard.test.tsx.snap\nindex 40331eb..fef20dd 100644\n--- a/client/src/components/Profile/__test__/__snapshots__/MainCard.test.tsx.snap\n+++ b/client/src/components/Profile/__test__/__snapshots__/MainCard.test.tsx.snap\n@@ -71,3 +71,158 @@ exports[`MainCard Should render correctly if is editing mode disabled 1`] = `\n </Card>\n </Fragment>\n `;\n+\n+exports[`MainCard Should render correctly if is editing mode enabled 1`] = `\n+<Fragment>\n+ <Card\n+ actions={\n+ Array [\n+ <ForwardRef(EditOutlined)\n+ onClick={[Function]}\n+ />,\n+ <ForwardRef(SettingOutlined)\n+ onClick={[Function]}\n+ />,\n+ ]\n+ }\n+ >\n+ <GithubAvatar\n+ githubId=\"piter\"\n+ size={96}\n+ style={\n+ Object {\n+ \"display\": \"block\",\n+ \"margin\": \"0 auto 10px\",\n+ }\n+ }\n+ />\n+ <Title\n+ level={1}\n+ style={\n+ Object {\n+ \"fontSize\": 24,\n+ \"margin\": 0,\n+ \"textAlign\": \"center\",\n+ }\n+ }\n+ >\n+ Petr Pervyi\n+ </Title>\n+ <Paragraph\n+ style={\n+ Object {\n+ \"marginBottom\": 20,\n+ \"textAlign\": \"center\",\n+ }\n+ }\n+ >\n+ <a\n+ href=\"https://github.com/piter\"\n+ style={\n+ Object {\n+ \"fontSize\": 16,\n+ \"marginLeft\": \"-14px\",\n+ }\n+ }\n+ target=\"_blank\"\n+ >\n+ <ForwardRef(GithubFilled) />\n+ \n+ piter\n+ </a>\n+ </Paragraph>\n+ <Paragraph\n+ style={\n+ Object {\n+ \"margin\": 0,\n+ \"textAlign\": \"center\",\n+ }\n+ }\n+ >\n+ <span\n+ style={\n+ Object {\n+ \"marginLeft\": \"-14px\",\n+ }\n+ }\n+ >\n+ <ForwardRef(EnvironmentFilled) />\n+ \n+ SPB\n+ </span>\n+ </Paragraph>\n+ <PermissionsSettingsDrawer\n+ hideSettings={[Function]}\n+ isSettingsVisible={false}\n+ onPermissionsSettingsChange={[Function]}\n+ />\n+ <ProfileSettingsDrawer\n+ content={\n+ <div>\n+ <p\n+ style={\n+ Object {\n+ \"fontSize\": 18,\n+ \"marginBottom\": 5,\n+ }\n+ }\n+ >\n+ <Text\n+ strong={true}\n+ >\n+ Name:\n+ </Text>\n+ </p>\n+ <p\n+ style={\n+ Object {\n+ \"marginBottom\": 20,\n+ }\n+ }\n+ >\n+ <Input\n+ onChange={[Function]}\n+ placeholder=\"Firstname Lastname\"\n+ type=\"text\"\n+ value=\"Petr Pervyi\"\n+ />\n+ </p>\n+ <p\n+ style={\n+ Object {\n+ \"fontSize\": 18,\n+ \"marginBottom\": 5,\n+ }\n+ }\n+ >\n+ <Text\n+ strong={true}\n+ >\n+ Location:\n+ </Text>\n+ </p>\n+ <div\n+ style={\n+ Object {\n+ \"marginBottom\": 5,\n+ }\n+ }\n+ >\n+ <LocationSelect\n+ defaultValue=\"1\"\n+ onChange={[Function]}\n+ style={\n+ Object {\n+ \"width\": \"100%\",\n+ }\n+ }\n+ />\n+ </div>\n+ </div>\n+ }\n+ hideSettings={[Function]}\n+ isSettingsVisible={false}\n+ />\n+ </Card>\n+</Fragment>\n+`;\ndiff --git a/client/src/jest.config.js b/client/src/jest.config.js\nindex df39788..654f9f3 100644\n--- a/client/src/jest.config.js\n+++ b/client/src/jest.config.js\n@@ -7,4 +7,5 @@ module.exports = {\n '^services(.*)$': '<rootDir>/services/$1',\n '^utils(.*)$': '<rootDir>/utils/$1',\n },\n+ verbose: true,\n };\n"]
1
["f87659953e9af59bc7cb314a22dd076d988ef607"]
["test"]
create mock img server
["diff --git a/scripts/gulp/tasks/test.ts b/scripts/gulp/tasks/test.ts\nindex 8014b12..d10c1aa 100644\n--- a/scripts/gulp/tasks/test.ts\n+++ b/scripts/gulp/tasks/test.ts\n@@ -26,12 +26,18 @@ task('test.imageserver', () => {\n function handleRequest(req, res) {\n const urlParse = url.parse(req.url, true);\n \n+ res.setHeader('Access-Control-Allow-Origin', '*');\n+ res.setHeader('Access-Control-Allow-Methods', 'GET');\n+ res.setHeader('Connection', 'keep-alive');\n+ res.setHeader('Age', '0');\n+ res.setHeader('cache-control', 'no-store');\n+\n if (urlParse.pathname === '/reset') {\n console.log('Image Server Reset');\n console.log('---------------------------');\n requestedUrls.length = 0;\n start = Date.now();\n- res.setHeader('Access-Control-Allow-Origin', '*');\n+ res.setHeader('Content-Type', 'text/plain');\n res.end('reset');\n return;\n }\n@@ -48,9 +54,8 @@ task('test.imageserver', () => {\n \n setTimeout(() => {\n res.setHeader('Content-Type', 'image/svg+xml');\n- res.setHeader('Access-Control-Allow-Origin', '*');\n res.end(`<svg xmlns=\"http://www.w3.org/2000/svg\" xmlns:xlink=\"http://www.w3.org/1999/xlink\"\n- style=\"background-color: ${color}; width: ${width}px; height: ${height}px;\">\n+ viewBox=\"0 0 ${width} ${height}\" style=\"background-color: ${color};\">\n <text x=\"5\" y=\"22\" style=\"font-family: Courier; font-size: 24px\">${id}</text>\n </svg>`);\n }, delay);\n"]
1
["32b76173a259ea1993298289b436cf10c1e800bf"]
["test"]
only restart if pages directory itself is changed resolves #429
["diff --git a/packages/cli/src/commands/dev.ts b/packages/cli/src/commands/dev.ts\nindex 35d859e..d6d91ed 100644\n--- a/packages/cli/src/commands/dev.ts\n+++ b/packages/cli/src/commands/dev.ts\n@@ -1,4 +1,4 @@\n-import { resolve } from 'upath'\n+import { resolve, relative } from 'upath'\n import chokidar from 'chokidar'\n import debounce from 'debounce-promise'\n import type { Nuxt } from '@nuxt/kit'\n@@ -27,9 +27,9 @@ export default defineNuxtCommand({\n const { loadNuxt, buildNuxt } = requireModule('@nuxt/kit', rootDir) as typeof import('@nuxt/kit')\n \n let currentNuxt: Nuxt\n- const load = async (isRestart: boolean) => {\n+ const load = async (isRestart: boolean, reason?: string) => {\n try {\n- const message = `${isRestart ? 'Restarting' : 'Starting'} nuxt...`\n+ const message = `${reason ? reason + '. ' : ''}${isRestart ? 'Restarting' : 'Starting'} nuxt...`\n server.setApp(createLoadingHandler(message))\n if (isRestart) {\n console.log(message)\n@@ -59,12 +59,8 @@ export default defineNuxtCommand({\n const dLoad = debounce(load, 250)\n const watcher = chokidar.watch([rootDir], { ignoreInitial: true, depth: 1 })\n watcher.on('all', (_event, file) => {\n- // Ignore any changes to files within the Nuxt build directory\n- if (file.includes(currentNuxt.options.buildDir)) {\n- return\n- }\n- if (file.includes('nuxt.config') || file.includes('modules') || file.includes('pages')) {\n- dLoad(true)\n+ if (file.match(/nuxt\\.config\\.(js|ts|mjs|cjs)$|pages$/)) {\n+ dLoad(true, `${relative(rootDir, file)} updated`)\n }\n })\n \n"]
1
["cbce777addb3dd118232a9f28db9d425d4c937b2"]
["fix"]
added components pages to typedoc output
["diff --git a/core/main/tsconfig.json b/core/main/tsconfig.json\nindex c4474a7..7916bc5 100644\n--- a/core/main/tsconfig.json\n+++ b/core/main/tsconfig.json\n@@ -96,11 +96,35 @@\n \"particles\": {\n \"groups\": [\n {\n- \"title\": \"Documentation\",\n+ \"title\": \"Components\",\n \"pages\": [\n {\n- \"title\": \"My Page\",\n- \"source\": \"./markdown/pages/index.md\"\n+ \"title\": \"Angular\",\n+ \"source\": \"../../components/angular/README.md\"\n+ },\n+ {\n+ \"title\": \"React\",\n+ \"source\": \"../../components/react/README.md\"\n+ },\n+ {\n+ \"title\": \"Vue\",\n+ \"source\": \"../../components/vue/README.md\"\n+ },\n+ {\n+ \"title\": \"Svelte\",\n+ \"source\": \"../../components/svelte/README.md\"\n+ },\n+ {\n+ \"title\": \"jQuery\",\n+ \"source\": \"../../components/jquery/README.md\"\n+ },\n+ {\n+ \"title\": \"Preact\",\n+ \"source\": \"../../components/preact/README.md\"\n+ },\n+ {\n+ \"title\": \"Inferno\",\n+ \"source\": \"../../components/inferno/README.md\"\n }\n ]\n }\n"]
1
["fca2c198c6486c4d586b1af1832be46f19667235"]
["docs"]
set name for topology module
["diff --git a/topology/pom.xml b/topology/pom.xml\nindex 389508e..ee6239a 100644\n--- a/topology/pom.xml\n+++ b/topology/pom.xml\n@@ -16,6 +16,7 @@\n </parent>\n \n <artifactId>zeebe-cluster-topology</artifactId>\n+ <name>Zeebe Cluster Topology</name>\n \n <properties>\n <proto.dir>${maven.multiModuleProjectDirectory}/topology/src/main/resources/proto</proto.dir>\n"]
1
["8911a972222dc80a242f3f1d9b3596321b3fdeaa"]
["build"]
coordinator accepts a request transformer instead of a list of operations The request transformer can generate the operations from the current topology. This helps to - ensure that the operations are generated based on the latest topology. When concurrent changes happens, coordinator can detect it. Previously it was unclear because by the time handle apply operations, the cluster topology might have changed. - return the simulated final topology as part of the result
["diff --git a/topology/src/main/java/io/camunda/zeebe/topology/changes/TopologyChangeCoordinator.java b/topology/src/main/java/io/camunda/zeebe/topology/changes/TopologyChangeCoordinator.java\nindex 8bb5c3d..f8f5e24 100644\n--- a/topology/src/main/java/io/camunda/zeebe/topology/changes/TopologyChangeCoordinator.java\n+++ b/topology/src/main/java/io/camunda/zeebe/topology/changes/TopologyChangeCoordinator.java\n@@ -10,6 +10,7 @@ package io.camunda.zeebe.topology.changes;\n import io.camunda.zeebe.scheduler.future.ActorFuture;\n import io.camunda.zeebe.topology.state.ClusterTopology;\n import io.camunda.zeebe.topology.state.TopologyChangeOperation;\n+import io.camunda.zeebe.util.Either;\n import java.util.List;\n \n public interface TopologyChangeCoordinator {\n@@ -39,4 +40,16 @@ public interface TopologyChangeCoordinator {\n ActorFuture<Boolean> hasCompletedChanges(final long version);\n \n ActorFuture<ClusterTopology> getCurrentTopology();\n+\n+ ActorFuture<TopologyChangeResult> applyOperations(TopologyChangeRequest request);\n+\n+ record TopologyChangeResult(\n+ ClusterTopology currentTopology,\n+ ClusterTopology finalTopology,\n+ List<TopologyChangeOperation> operations) {}\n+\n+ interface TopologyChangeRequest {\n+ Either<Exception, List<TopologyChangeOperation>> operations(\n+ final ClusterTopology currentTopology);\n+ }\n }\ndiff --git a/topology/src/main/java/io/camunda/zeebe/topology/changes/TopologyChangeCoordinatorImpl.java b/topology/src/main/java/io/camunda/zeebe/topology/changes/TopologyChangeCoordinatorImpl.java\nindex 13ec754..877fc3c 100644\n--- a/topology/src/main/java/io/camunda/zeebe/topology/changes/TopologyChangeCoordinatorImpl.java\n+++ b/topology/src/main/java/io/camunda/zeebe/topology/changes/TopologyChangeCoordinatorImpl.java\n@@ -103,6 +103,62 @@ public class TopologyChangeCoordinatorImpl implements TopologyChangeCoordinator \n return clusterTopologyManager.getClusterTopology();\n }\n \n+ @Override\n+ public ActorFuture<TopologyChangeResult> applyOperations(final TopologyChangeRequest request) {\n+ final ActorFuture<TopologyChangeResult> future = executor.createFuture();\n+ clusterTopologyManager\n+ .getClusterTopology()\n+ .onComplete(\n+ (currentClusterTopology, errorOnGettingTopology) -> {\n+ if (errorOnGettingTopology != null) {\n+ future.completeExceptionally(errorOnGettingTopology);\n+ return;\n+ }\n+\n+ final var operationsEither = request.operations(currentClusterTopology);\n+ if (operationsEither.isLeft()) {\n+ future.completeExceptionally(operationsEither.getLeft());\n+ return;\n+ }\n+ final var operations = operationsEither.get();\n+ if (operations.isEmpty()) {\n+ // No operations to apply\n+ future.complete(\n+ new TopologyChangeResult(\n+ currentClusterTopology, currentClusterTopology, operations));\n+ return;\n+ }\n+\n+ final ActorFuture<ClusterTopology> validation =\n+ validateTopologyChangeRequest(currentClusterTopology, operations);\n+\n+ validation.onComplete(\n+ (simulatedFinalTopology, validationError) -> {\n+ if (validationError != null) {\n+ future.completeExceptionally(validationError);\n+ return;\n+ }\n+\n+ // if the validation was successful, apply the changes\n+ final ActorFuture<ClusterTopology> applyFuture = executor.createFuture();\n+ applyTopologyChange(\n+ operations, currentClusterTopology, simulatedFinalTopology, applyFuture);\n+\n+ applyFuture.onComplete(\n+ (ignore, error) -> {\n+ if (error == null) {\n+ future.complete(\n+ new TopologyChangeResult(\n+ currentClusterTopology, simulatedFinalTopology, operations));\n+ } else {\n+ future.completeExceptionally(error);\n+ }\n+ });\n+ });\n+ });\n+ return future;\n+ }\n+\n private ActorFuture<ClusterTopology> validateTopologyChangeRequest(\n final ClusterTopology currentClusterTopology,\n final List<TopologyChangeOperation> operations) {\n"]
1
["dec860436916ef216998f80f8b2f9c39d00c064d"]
["feat"]
e2e
["diff --git a/.github/workflows/kibbeh-e2e_tests.yaml b/.github/workflows/kibbeh-e2e_tests.yaml\nindex 52bf3ed..74fe785 100644\n--- a/.github/workflows/kibbeh-e2e_tests.yaml\n+++ b/.github/workflows/kibbeh-e2e_tests.yaml\n@@ -4,7 +4,7 @@ on:\n branches:\n - staging\n paths:\n- - '.github/workflows/end-to-end-tests.yaml'\n+ - '.github/workflows/kibbeh-e2e_tests.yaml'\n - 'kousa/lib/**'\n - 'kibbeh/src/**'\n pull_request:\n"]
1
["02f9e79a755a1dd95e661c674d6966837f378442"]
["ci"]
update drone
["diff --git a/.drone.yml b/.drone.yml\nindex 53e3329..306516b 100644\n--- a/.drone.yml\n+++ b/.drone.yml\n@@ -21,10 +21,10 @@ steps:\n from_secret: docker_username\n password:\n from_secret: docker_password\n- tags: pg\n+ tags: latest\n when: \n branch:\n- - pg\n+ - master\n event: \n - push\n - \n@@ -40,14 +40,14 @@ steps:\n port: 22\n script: \n - \"sleep 10\"\n- - \"docker pull rsschool/api:pg\"\n+ - \"docker pull rsschool/api:latest\"\n - \"docker-compose stop api\"\n - \"docker-compose rm -f api\"\n - \"docker-compose up -d api\"\n - \"docker system prune -f\"\n when: \n branch: \n- - pg\n+ - master\n event: \n - push\n volumes:\n"]
1
["88129ee45b1d49bc4ff887f3b488464cc7097e29"]
["build"]
remove writers from interface
["diff --git a/engine/src/main/java/io/camunda/zeebe/engine/Engine.java b/engine/src/main/java/io/camunda/zeebe/engine/Engine.java\nindex 91f1b41..eb4b9a8 100644\n--- a/engine/src/main/java/io/camunda/zeebe/engine/Engine.java\n+++ b/engine/src/main/java/io/camunda/zeebe/engine/Engine.java\n@@ -81,8 +81,6 @@ public class Engine implements RecordProcessor<EngineContext> {\n \n engineContext.setLifecycleListeners(typedRecordProcessors.getLifecycleListeners());\n recordProcessorMap = typedRecordProcessors.getRecordProcessorMap();\n-\n- engineContext.setWriters(writers);\n }\n \n @Override\ndiff --git a/engine/src/main/java/io/camunda/zeebe/engine/EngineContext.java b/engine/src/main/java/io/camunda/zeebe/engine/EngineContext.java\nindex a8e5538..a27b6e6 100644\n--- a/engine/src/main/java/io/camunda/zeebe/engine/EngineContext.java\n+++ b/engine/src/main/java/io/camunda/zeebe/engine/EngineContext.java\n@@ -15,7 +15,6 @@ import io.camunda.zeebe.engine.processing.streamprocessor.StreamProcessorListene\n import io.camunda.zeebe.engine.processing.streamprocessor.TypedRecordProcessorFactory;\n import io.camunda.zeebe.engine.processing.streamprocessor.writers.LegacyTypedResponseWriter;\n import io.camunda.zeebe.engine.processing.streamprocessor.writers.LegacyTypedStreamWriter;\n-import io.camunda.zeebe.engine.processing.streamprocessor.writers.Writers;\n import io.camunda.zeebe.engine.state.EventApplier;\n import io.camunda.zeebe.engine.state.mutable.MutableZeebeState;\n import java.util.Collections;\n@@ -34,7 +33,6 @@ public final class EngineContext {\n private final TypedRecordProcessorFactory typedRecordProcessorFactory;\n private List<StreamProcessorLifecycleAware> lifecycleListeners = Collections.EMPTY_LIST;\n private StreamProcessorListener streamProcessorListener;\n- private Writers writers;\n \n public EngineContext(\n final int partitionId,\n@@ -102,12 +100,4 @@ public final class EngineContext {\n public void setStreamProcessorListener(final StreamProcessorListener streamProcessorListener) {\n this.streamProcessorListener = streamProcessorListener;\n }\n-\n- public Writers getWriters() {\n- return writers;\n- }\n-\n- public void setWriters(final Writers writers) {\n- this.writers = writers;\n- }\n }\ndiff --git a/engine/src/main/java/io/camunda/zeebe/engine/api/ReadonlyStreamProcessorContext.java b/engine/src/main/java/io/camunda/zeebe/engine/api/ReadonlyStreamProcessorContext.java\nindex f30c7cc..834b421 100644\n--- a/engine/src/main/java/io/camunda/zeebe/engine/api/ReadonlyStreamProcessorContext.java\n+++ b/engine/src/main/java/io/camunda/zeebe/engine/api/ReadonlyStreamProcessorContext.java\n@@ -8,7 +8,6 @@\n package io.camunda.zeebe.engine.api;\n \n import io.camunda.zeebe.engine.processing.streamprocessor.writers.LegacyTypedStreamWriter;\n-import io.camunda.zeebe.engine.processing.streamprocessor.writers.Writers;\n import io.camunda.zeebe.engine.state.mutable.MutableZeebeState;\n import io.camunda.zeebe.logstreams.log.LogStream;\n \n@@ -27,11 +26,6 @@ public interface ReadonlyStreamProcessorContext {\n LegacyTypedStreamWriter getLogStreamWriter();\n \n /**\n- * @return the specific writers, like command, response, etc\n- */\n- Writers getWriters();\n-\n- /**\n * @return the state, where the data is stored during processing\n */\n MutableZeebeState getZeebeState();\ndiff --git a/engine/src/main/java/io/camunda/zeebe/streamprocessor/StreamProcessor.java b/engine/src/main/java/io/camunda/zeebe/streamprocessor/StreamProcessor.java\nindex 844e487..49fd8e2 100755\n--- a/engine/src/main/java/io/camunda/zeebe/streamprocessor/StreamProcessor.java\n+++ b/engine/src/main/java/io/camunda/zeebe/streamprocessor/StreamProcessor.java\n@@ -346,7 +346,6 @@ public class StreamProcessor extends Actor implements HealthMonitorable, LogReco\n if (listener != null) {\n streamProcessorContext.listener(engineContext.getStreamProcessorListener());\n }\n- streamProcessorContext.writers(engineContext.getWriters());\n }\n \n private long recoverFromSnapshot() {\ndiff --git a/engine/src/main/java/io/camunda/zeebe/streamprocessor/StreamProcessorContext.java b/engine/src/main/java/io/camunda/zeebe/streamprocessor/StreamProcessorContext.java\nindex d02b273..b527d3c 100644\n--- a/engine/src/main/java/io/camunda/zeebe/streamprocessor/StreamProcessorContext.java\n+++ b/engine/src/main/java/io/camunda/zeebe/streamprocessor/StreamProcessorContext.java\n@@ -17,7 +17,6 @@ import io.camunda.zeebe.engine.processing.streamprocessor.writers.CommandRespons\n import io.camunda.zeebe.engine.processing.streamprocessor.writers.LegacyTypedResponseWriterImpl;\n import io.camunda.zeebe.engine.processing.streamprocessor.writers.LegacyTypedStreamWriter;\n import io.camunda.zeebe.engine.processing.streamprocessor.writers.NoopLegacyTypedStreamWriter;\n-import io.camunda.zeebe.engine.processing.streamprocessor.writers.Writers;\n import io.camunda.zeebe.engine.state.EventApplier;\n import io.camunda.zeebe.engine.state.KeyGeneratorControls;\n import io.camunda.zeebe.engine.state.ZeebeDbState;\n@@ -55,7 +54,7 @@ public final class StreamProcessorContext implements ReadonlyStreamProcessorCont\n private StreamProcessorMode streamProcessorMode = StreamProcessorMode.PROCESSING;\n private ProcessingScheduleService processingScheduleService;\n private MutableLastProcessedPositionState lastProcessedPositionState;\n- private Writers writers;\n+\n private LogStreamBatchWriter logStreamBatchWriter;\n private CommandResponseWriter commandResponseWriter;\n \n@@ -85,11 +84,6 @@ public final class StreamProcessorContext implements ReadonlyStreamProcessorCont\n }\n \n @Override\n- public Writers getWriters() {\n- return writers;\n- }\n-\n- @Override\n public MutableZeebeState getZeebeState() {\n return zeebeState;\n }\n@@ -216,10 +210,6 @@ public final class StreamProcessorContext implements ReadonlyStreamProcessorCont\n return streamProcessorMode;\n }\n \n- public void writers(final Writers writers) {\n- this.writers = writers;\n- }\n-\n public void logStreamBatchWriter(final LogStreamBatchWriter batchWriter) {\n logStreamBatchWriter = batchWriter;\n }\n"]
1
["d2709cab63295109dcd1a49f57da9418110e9044"]
["refactor"]
fix build
["diff --git a/server/Dockerfile b/server/Dockerfile\nindex 2f203bb..a84c31e 100755\n--- a/server/Dockerfile\n+++ b/server/Dockerfile\n@@ -9,9 +9,11 @@ ENV TZ utc\n WORKDIR /src\n \n COPY package.json /src\n+COPY package-lock.json /src\n+COPY tsconfig.json /src\n RUN npm install --production --no-optional\n \n COPY public /src/public\n COPY dist /src\n \n-CMD [ \"node\", \"./server/index.js\" ]\n+CMD [ \"node\", \"-r\", \"tsconfig-paths/register\", \"./server/index.js\" ]\ndiff --git a/server/package-lock.json b/server/package-lock.json\nindex 6cacfa2..236f1bb 100644\n--- a/server/package-lock.json\n+++ b/server/package-lock.json\n@@ -2164,8 +2164,7 @@\n \"@types/json5\": {\n \"version\": \"0.0.29\",\n \"resolved\": \"https://registry.npmjs.org/@types/json5/-/json5-0.0.29.tgz\",\n- \"integrity\": \"sha1-7ihweulOEdK4J7y+UnC86n8+ce4=\",\n- \"dev\": true\n+ \"integrity\": \"sha1-7ihweulOEdK4J7y+UnC86n8+ce4=\"\n },\n \"@types/jsonwebtoken\": {\n \"version\": \"8.3.5\",\n@@ -9246,8 +9245,7 @@\n \"strip-bom\": {\n \"version\": \"3.0.0\",\n \"resolved\": \"https://registry.npmjs.org/strip-bom/-/strip-bom-3.0.0.tgz\",\n- \"integrity\": \"sha1-IzTBjpx1n3vdVv3vfprj1YjmjtM=\",\n- \"dev\": true\n+ \"integrity\": \"sha1-IzTBjpx1n3vdVv3vfprj1YjmjtM=\"\n },\n \"strip-final-newline\": {\n \"version\": \"2.0.0\",\n@@ -9524,7 +9522,6 @@\n \"version\": \"3.9.0\",\n \"resolved\": \"https://registry.npmjs.org/tsconfig-paths/-/tsconfig-paths-3.9.0.tgz\",\n \"integrity\": \"sha512-dRcuzokWhajtZWkQsDVKbWyY+jgcLC5sqJhg2PSgf4ZkH2aHPvaOY8YWGhmjb68b5qqTfasSsDO9k7RUiEmZAw==\",\n- \"dev\": true,\n \"requires\": {\n \"@types/json5\": \"^0.0.29\",\n \"json5\": \"^1.0.1\",\n@@ -9536,7 +9533,6 @@\n \"version\": \"1.0.1\",\n \"resolved\": \"https://registry.npmjs.org/json5/-/json5-1.0.1.tgz\",\n \"integrity\": \"sha512-aKS4WQjPenRxiQsC93MNfjx+nbF4PAdYzmd/1JIj8HYzqfbu86beTuNgXDzPknWk0n0uARlyewZo4s++ES36Ow==\",\n- \"dev\": true,\n \"requires\": {\n \"minimist\": \"^1.2.0\"\n }\n@@ -9544,8 +9540,7 @@\n \"minimist\": {\n \"version\": \"1.2.5\",\n \"resolved\": \"https://registry.npmjs.org/minimist/-/minimist-1.2.5.tgz\",\n- \"integrity\": \"sha512-FM9nNUYrRBAELZQT3xeZQ7fmMOBg6nWNmJKTcgsJeaLstP/UODVpGsr5OhXhhXg6f+qtJ8uiZ+PUxkDWcgIXLw==\",\n- \"dev\": true\n+ \"integrity\": \"sha512-FM9nNUYrRBAELZQT3xeZQ7fmMOBg6nWNmJKTcgsJeaLstP/UODVpGsr5OhXhhXg6f+qtJ8uiZ+PUxkDWcgIXLw==\"\n }\n }\n },\ndiff --git a/server/package.json b/server/package.json\nindex 35426e9..896e9b3 100644\n--- a/server/package.json\n+++ b/server/package.json\n@@ -41,6 +41,7 @@\n \"pino-cloudwatch\": \"0.7.0\",\n \"pino-multi-stream\": \"4.2.0\",\n \"reflect-metadata\": \"0.1.13\",\n+ \"tsconfig-paths\": \"3.9.0\",\n \"typeorm\": \"0.2.37\"\n },\n \"devDependencies\": {\n@@ -69,7 +70,6 @@\n \"pino-pretty\": \"3.6.1\",\n \"ts-jest\": \"27.0.7\",\n \"ts-node-dev\": \"1.1.8\",\n- \"tsconfig-paths\": \"3.9.0\",\n \"typescript\": \"4.3.5\"\n },\n \"jest-junit\": {\n"]
1
["a827777f41e90b6332c191d05bae8db525de6f38"]
["build"]
update basic test with colors
["diff --git a/core/src/components/label/test/basic/index.html b/core/src/components/label/test/basic/index.html\nindex d0b566c..377e58c 100644\n--- a/core/src/components/label/test/basic/index.html\n+++ b/core/src/components/label/test/basic/index.html\n@@ -19,12 +19,32 @@\n </ion-header>\n \n <ion-content>\n+ <div padding>\n+ <ion-label>Default</ion-label>\n+\n+ <ion-label color=\"secondary\">Secondary</ion-label>\n+\n+ <ion-label color=\"tertiary\">Tertiary</ion-label>\n+\n+ <ion-label color=\"danger\">Danger</ion-label>\n+\n+ <ion-label class=\"custom\">Custom</ion-label>\n+ </div>\n+\n <ion-list>\n <ion-item>\n <ion-label>Default</ion-label>\n <ion-input></ion-input>\n </ion-item>\n <ion-item>\n+ <ion-label color=\"tertiary\">Tertiary</ion-label>\n+ <ion-input></ion-input>\n+ </ion-item>\n+ <ion-item>\n+ <ion-label class=\"custom\">Custom</ion-label>\n+ <ion-input></ion-input>\n+ </ion-item>\n+ <ion-item>\n <ion-label text-wrap>Wrap label this label just goes on and on and on</ion-label>\n <ion-input></ion-input>\n </ion-item>\n@@ -42,6 +62,12 @@\n </ion-item>\n </ion-list>\n </ion-content>\n+\n+ <style>\n+ .custom {\n+ color: lightblue;\n+ }\n+ </style>\n </ion-app>\n </body>\n \n"]
1
["c3b5dc77ff3d89d389f6f3a868b17d0a8ca63074"]
["test"]
update version (nightly.0)
["diff --git a/Cargo.lock b/Cargo.lock\nindex e6f659c..cf93556 100644\n--- a/Cargo.lock\n+++ b/Cargo.lock\n@@ -94,7 +94,7 @@ dependencies = [\n \n [[package]]\n name = \"els\"\n-version = \"0.1.23\"\n+version = \"0.1.24-nightly.0\"\n dependencies = [\n \"erg_common\",\n \"erg_compiler\",\n@@ -105,7 +105,7 @@ dependencies = [\n \n [[package]]\n name = \"erg\"\n-version = \"0.6.11\"\n+version = \"0.6.12-nightly.0\"\n dependencies = [\n \"els\",\n \"erg_common\",\n@@ -115,7 +115,7 @@ dependencies = [\n \n [[package]]\n name = \"erg_common\"\n-version = \"0.6.11\"\n+version = \"0.6.12-nightly.0\"\n dependencies = [\n \"backtrace-on-stack-overflow\",\n \"crossterm\",\n@@ -126,7 +126,7 @@ dependencies = [\n \n [[package]]\n name = \"erg_compiler\"\n-version = \"0.6.11\"\n+version = \"0.6.12-nightly.0\"\n dependencies = [\n \"erg_common\",\n \"erg_parser\",\n@@ -134,7 +134,7 @@ dependencies = [\n \n [[package]]\n name = \"erg_parser\"\n-version = \"0.6.11\"\n+version = \"0.6.12-nightly.0\"\n dependencies = [\n \"erg_common\",\n \"unicode-xid\",\ndiff --git a/Cargo.toml b/Cargo.toml\nindex c58299b..6e51b6e 100644\n--- a/Cargo.toml\n+++ b/Cargo.toml\n@@ -20,7 +20,7 @@ members = [\n ]\n \n [workspace.package]\n-version = \"0.6.11\"\n+version = \"0.6.12-nightly.0\"\n authors = [\"erg-lang team <[email protected]>\"]\n license = \"MIT OR Apache-2.0\"\n edition = \"2021\"\n@@ -64,10 +64,10 @@ full-repl = [\"erg_common/full-repl\"]\n full = [\"els\", \"full-repl\", \"unicode\", \"pretty\"]\n \n [workspace.dependencies]\n-erg_common = { version = \"0.6.11\", path = \"./crates/erg_common\" }\n-erg_parser = { version = \"0.6.11\", path = \"./crates/erg_parser\" }\n-erg_compiler = { version = \"0.6.11\", path = \"./crates/erg_compiler\" }\n-els = { version = \"0.1.23\", path = \"./crates/els\" }\n+erg_common = { version = \"0.6.12-nightly.0\", path = \"./crates/erg_common\" }\n+erg_parser = { version = \"0.6.12-nightly.0\", path = \"./crates/erg_parser\" }\n+erg_compiler = { version = \"0.6.12-nightly.0\", path = \"./crates/erg_compiler\" }\n+els = { version = \"0.1.24-nightly.0\", path = \"./crates/els\" }\n \n [dependencies]\n erg_common = { workspace = true }\ndiff --git a/crates/els/Cargo.toml b/crates/els/Cargo.toml\nindex 5f005a1..e1a9964 100644\n--- a/crates/els/Cargo.toml\n+++ b/crates/els/Cargo.toml\n@@ -2,7 +2,7 @@\n name = \"els\"\n description = \"An Erg compiler frontend for IDEs, implements LSP.\"\n documentation = \"http://docs.rs/els\"\n-version = \"0.1.23\"\n+version = \"0.1.24-nightly.0\"\n authors.workspace = true\n license.workspace = true\n edition.workspace = true\n"]
1
["92e940efeee199b1e0bbbc3c9eea7f3dc8221619"]
["build"]
add activatedElementInstanceKeys to modification record
["diff --git a/protocol-impl/src/test/java/io/camunda/zeebe/protocol/impl/JsonSerializableToJsonTest.java b/protocol-impl/src/test/java/io/camunda/zeebe/protocol/impl/JsonSerializableToJsonTest.java\nindex 33410da..edd0588 100644\n--- a/protocol-impl/src/test/java/io/camunda/zeebe/protocol/impl/JsonSerializableToJsonTest.java\n+++ b/protocol-impl/src/test/java/io/camunda/zeebe/protocol/impl/JsonSerializableToJsonTest.java\n@@ -787,7 +787,8 @@ final class JsonSerializableToJsonTest {\n }\n }],\n \"elementId\": \"activity\"\n- }]\n+ }],\n+ \"activatedElementInstanceKeys\": []\n }\n \"\"\"\n },\n@@ -803,7 +804,8 @@ final class JsonSerializableToJsonTest {\n {\n \"processInstanceKey\": 1,\n \"terminateInstructions\": [],\n- \"activateInstructions\": []\n+ \"activateInstructions\": [],\n+ \"activatedElementInstanceKeys\": []\n }\n \"\"\"\n },\n"]
1
["f7cc7b263afeb27eef393b7497db8dad8ebb0518"]
["test"]
fix cypress on windows
["diff --git a/packages/cypress/src/builders/cypress/cypress.impl.spec.ts b/packages/cypress/src/builders/cypress/cypress.impl.spec.ts\nindex 22851fb..c9296fe 100644\n--- a/packages/cypress/src/builders/cypress/cypress.impl.spec.ts\n+++ b/packages/cypress/src/builders/cypress/cypress.impl.spec.ts\n@@ -77,7 +77,7 @@ describe('Cypress builder', () => {\n await run.result;\n await run.stop();\n expect(fork).toHaveBeenCalledWith(\n- '/root/node_modules/.bin/tsc',\n+ '/root/node_modules/typescript/bin/tsc',\n ['-p', '/root/apps/my-app-e2e/tsconfig.json'],\n { stdio: [0, 1, 2, 'ipc'] }\n );\ndiff --git a/packages/cypress/src/builders/cypress/cypress.impl.ts b/packages/cypress/src/builders/cypress/cypress.impl.ts\nindex 9d9ded3..d3917c2 100644\n--- a/packages/cypress/src/builders/cypress/cypress.impl.ts\n+++ b/packages/cypress/src/builders/cypress/cypress.impl.ts\n@@ -115,7 +115,7 @@ function compileTypescriptFiles(\n let args = ['-p', path.join(context.workspaceRoot, tsConfigPath)];\n const tscPath = path.join(\n context.workspaceRoot,\n- '/node_modules/.bin/tsc'\n+ '/node_modules/typescript/bin/tsc'\n );\n if (isWatching) {\n args.push('--watch');\n"]
1
["eebee9ab0bb6d4255ad0402d8422364e96bfef61"]
["fix"]
render-svg
["diff --git a/package.json b/package.json\nindex 3f8e5fa..cc4e398 100644\n--- a/package.json\n+++ b/package.json\n@@ -42,6 +42,7 @@\n \"rollup\": \"^2.34.2\",\n \"rollup-plugin-copy\": \"^3.3.0\",\n \"rollup-plugin-dts\": \"^2.0.0\",\n+ \"rollup-plugin-terser\": \"^7.0.2\",\n \"rollup-plugin-typescript2\": \"^0.29.0\",\n \"ts-jest\": \"^26.4.4\",\n \"tsup\": \"^3.10.1\",\ndiff --git a/packages/renderer-svg/package.json b/packages/renderer-svg/package.json\nindex fa9c049..6a0654c 100644\n--- a/packages/renderer-svg/package.json\n+++ b/packages/renderer-svg/package.json\n@@ -1,16 +1,27 @@\n {\n- \"name\": \"shiki-renderer-svg\",\n+ \"name\": \"@antfu/shiki-renderer-svg\",\n \"version\": \"0.2.0\",\n \"description\": \"SVG renderer for shiki\",\n \"author\": \"Pine Wu <[email protected]>\",\n \"homepage\": \"https://github.com/octref/shiki/tree/master/packages/renderer-svg\",\n \"license\": \"MIT\",\n- \"main\": \"dist/index.js\",\n- \"types\": \"dist/index.d.ts\",\n \"repository\": {\n \"type\": \"git\",\n \"url\": \"git+https://github.com/shikijs/shiki.git\"\n },\n+ \"main\": \"dist/index.js\",\n+ \"module\": \"dist/index.mjs\",\n+ \"types\": \"dist/index.d.ts\",\n+ \"unpkg\": \"dist/index.iife.min.js\",\n+ \"jsdelivr\": \"dist/index.iife.min.js\",\n+ \"files\": [\n+ \"dist\"\n+ ],\n+ \"scripts\": {\n+ \"prepublishOnly\": \"npm run build\",\n+ \"build\": \"rollup -c\",\n+ \"watch\": \"rollup -c -w\"\n+ },\n \"dependencies\": {\n \"puppeteer\": \"^5.2.1\"\n },\ndiff --git a/packages/renderer-svg/rollup.config.js b/packages/renderer-svg/rollup.config.js\nnew file mode 100644\nindex 0000000..d4e45ce\n--- /dev/null\n+++ b/packages/renderer-svg/rollup.config.js\n@@ -0,0 +1,67 @@\n+import { nodeResolve } from '@rollup/plugin-node-resolve'\n+import commonjs from '@rollup/plugin-commonjs'\n+import dts from 'rollup-plugin-dts'\n+import typescript from 'rollup-plugin-typescript2'\n+import replace from '@rollup/plugin-replace'\n+import { terser } from 'rollup-plugin-terser'\n+\n+const external = ['shiki', 'puppeteer']\n+\n+export default [\n+ {\n+ input: 'src/index.ts',\n+ external,\n+ output: [\n+ {\n+ file: 'dist/index.js',\n+ format: 'cjs'\n+ },\n+ {\n+ file: 'dist/index.mjs',\n+ format: 'esm'\n+ }\n+ ],\n+ plugins: [\n+ replace({\n+ __BROWSER__: JSON.stringify(false)\n+ }),\n+ typescript(),\n+ nodeResolve(),\n+ commonjs()\n+ ]\n+ },\n+ {\n+ input: 'src/index.ts',\n+ output: [\n+ {\n+ file: 'dist/index.iife.js',\n+ format: 'iife',\n+ name: 'ShikiRenderSVG'\n+ },\n+ {\n+ file: 'dist/index.iife.min.js',\n+ format: 'iife',\n+ name: 'ShikiRenderSVG',\n+ plugins: [terser()]\n+ }\n+ ],\n+ plugins: [\n+ replace({\n+ __BROWSER__: JSON.stringify(true)\n+ }),\n+ typescript(),\n+ nodeResolve(),\n+ commonjs()\n+ ]\n+ },\n+ {\n+ input: 'src/index.ts',\n+ output: [\n+ {\n+ file: 'dist/index.d.ts',\n+ format: 'es'\n+ }\n+ ],\n+ plugins: [dts()]\n+ }\n+]\ndiff --git a/packages/renderer-svg/src/global.d.ts b/packages/renderer-svg/src/global.d.ts\nnew file mode 100644\nindex 0000000..08c128f\n--- /dev/null\n+++ b/packages/renderer-svg/src/global.d.ts\n@@ -0,0 +1 @@\n+declare var __BROWSER__: boolean\ndiff --git a/packages/renderer-svg/src/index.ts b/packages/renderer-svg/src/index.ts\nindex ae77136..8f92312 100644\n--- a/packages/renderer-svg/src/index.ts\n+++ b/packages/renderer-svg/src/index.ts\n@@ -1,4 +1,4 @@\n-import { IThemedToken } from 'shiki'\n+import type { IThemedToken } from 'shiki'\n import { measureMonospaceTypeface } from './measureMonospaceTypeface'\n \n interface SVGRendererOptions {\ndiff --git a/packages/renderer-svg/src/measureMonospaceTypeface.ts b/packages/renderer-svg/src/measureMonospaceTypeface.ts\nindex e28a1ff..6ab834d 100644\n--- a/packages/renderer-svg/src/measureMonospaceTypeface.ts\n+++ b/packages/renderer-svg/src/measureMonospaceTypeface.ts\n@@ -1,58 +1,61 @@\n-import puppeteer from 'puppeteer'\n+function measureFont(fontName: string, fontSize: number) {\n+ /**\n+ * Measure `M` for width\n+ */\n+ var c = document.createElement('canvas')\n+ var ctx = c.getContext('2d')!\n+ ctx.font = `${fontSize}px \"${fontName}\"`\n \n-export async function measureMonospaceTypeface(\n- fontName: string,\n- fontSize: number\n-): Promise<{ width: number; height: number }> {\n- const browser = await puppeteer.launch({ headless: true })\n- const page = await browser.newPage()\n- const measurement = await page.evaluate(measureFont, fontName, fontSize)\n- await browser.close()\n- return measurement\n+ const capMMeasurement = ctx.measureText('M')\n \n- function measureFont(fontName: string, fontSize: number) {\n- /**\n- * Measure `M` for width\n- */\n- var c = document.createElement('canvas')\n- var ctx = c.getContext('2d')!\n- ctx.font = `${fontSize}px \"${fontName}\"`\n-\n- const capMMeasurement = ctx.measureText('M')\n+ /**\n+ * Measure A-Z, a-z for height\n+ * A - 65\n+ * Z - 90\n+ * a - 97\n+ * z - 122\n+ */\n+ const characters = []\n+ for (let i = 65; i <= 90; i++) {\n+ characters.push(String.fromCharCode(i))\n+ }\n+ for (let i = 97; i <= 122; i++) {\n+ characters.push(String.fromCharCode(i))\n+ }\n \n- /**\n- * Measure A-Z, a-z for height\n- * A - 65\n- * Z - 90\n- * a - 97\n- * z - 122\n- */\n- const characters = []\n- for (let i = 65; i <= 90; i++) {\n- characters.push(String.fromCharCode(i))\n+ let highC, lowC\n+ let highestAscent = 0\n+ let lowestDescent = 0\n+ characters.forEach(c => {\n+ const m = ctx.measureText(c)\n+ if (m.actualBoundingBoxAscent > highestAscent) {\n+ highestAscent = m.actualBoundingBoxAscent\n+ highC = c\n }\n- for (let i = 97; i <= 122; i++) {\n- characters.push(String.fromCharCode(i))\n+ if (m.actualBoundingBoxDescent > lowestDescent) {\n+ lowestDescent = m.actualBoundingBoxDescent\n+ lowC = c\n }\n+ })\n \n- let highC, lowC\n- let highestAscent = 0\n- let lowestDescent = 0\n- characters.forEach(c => {\n- const m = ctx.measureText(c)\n- if (m.actualBoundingBoxAscent > highestAscent) {\n- highestAscent = m.actualBoundingBoxAscent\n- highC = c\n- }\n- if (m.actualBoundingBoxDescent > lowestDescent) {\n- lowestDescent = m.actualBoundingBoxDescent\n- lowC = c\n- }\n- })\n+ return {\n+ width: capMMeasurement.width,\n+ height: highestAscent + lowestDescent\n+ }\n+}\n \n- return {\n- width: capMMeasurement.width,\n- height: highestAscent + lowestDescent\n- }\n+export async function measureMonospaceTypeface(\n+ fontName: string,\n+ fontSize: number\n+): Promise<{ width: number; height: number }> {\n+ if (__BROWSER__) {\n+ return measureFont(fontName, fontSize)\n+ } else {\n+ const puppeteer = await import('puppeteer')\n+ const browser = await puppeteer.launch({ headless: true })\n+ const page = await browser.newPage()\n+ const measurement = await page.evaluate(measureFont, fontName, fontSize)\n+ await browser.close()\n+ return measurement\n }\n }\ndiff --git a/packages/renderer-svg/tsconfig.json b/packages/renderer-svg/tsconfig.json\nindex 3613212..bc50ce3 100644\n--- a/packages/renderer-svg/tsconfig.json\n+++ b/packages/renderer-svg/tsconfig.json\n@@ -1,9 +1,10 @@\n {\n- \"extends\": \"../../tsconfig.json\",\n \"compilerOptions\": {\n- \"composite\": true,\n- \"rootDir\": \"src\",\n- \"outDir\": \"dist\",\n- \"lib\": [\"dom\"]\n+ \"module\": \"esnext\",\n+ \"target\": \"es2017\",\n+ \"esModuleInterop\": true,\n+ \"moduleResolution\": \"node\",\n+ \"lib\": [\"esnext\", \"DOM\"],\n+ \"sourceMap\": true\n }\n }\ndiff --git a/packages/shiki/rollup.config.js b/packages/shiki/rollup.config.js\nindex b8ba9e3..9078ea2 100644\n--- a/packages/shiki/rollup.config.js\n+++ b/packages/shiki/rollup.config.js\n@@ -4,6 +4,7 @@ import dts from 'rollup-plugin-dts'\n import typescript from 'rollup-plugin-typescript2'\n import copy from 'rollup-plugin-copy'\n import replace from '@rollup/plugin-replace'\n+import { terser } from 'rollup-plugin-terser'\n import { version } from './package.json'\n \n const external = ['onigasm', 'vscode-textmate']\n@@ -22,7 +23,14 @@ export default [\n format: 'esm'\n }\n ],\n- plugins: [typescript(), nodeResolve(), commonjs()]\n+ plugins: [\n+ replace({\n+ __BROWSER__: JSON.stringify(false)\n+ }),\n+ typescript(),\n+ nodeResolve(),\n+ commonjs()\n+ ]\n },\n {\n input: 'src/index.ts',\n@@ -58,7 +66,15 @@ export default [\n ]\n }\n ],\n- plugins: [typescript(), nodeResolve(), commonjs()]\n+ plugins: [\n+ replace({\n+ __BROWSER__: JSON.stringify(true)\n+ }),\n+ typescript(),\n+ nodeResolve(),\n+ commonjs(),\n+ terser()\n+ ]\n },\n {\n input: 'src/index.ts',\ndiff --git a/packages/shiki/src/global.d.ts b/packages/shiki/src/global.d.ts\nnew file mode 100644\nindex 0000000..08c128f\n--- /dev/null\n+++ b/packages/shiki/src/global.d.ts\n@@ -0,0 +1 @@\n+declare var __BROWSER__: boolean\ndiff --git a/packages/shiki/src/loader.ts b/packages/shiki/src/loader.ts\nindex 934cfbd..d9c3128 100644\n--- a/packages/shiki/src/loader.ts\n+++ b/packages/shiki/src/loader.ts\n@@ -5,11 +5,16 @@ import type { ILanguageRegistration, IShikiTheme } from './types'\n export const isBrowser = typeof window !== 'undefined' && typeof window.document !== 'undefined'\n \n let CDN_ROOT = '__CDN_ROOT__'\n+let ONIGASM_WASM = ''\n \n export function setCDN(root: string) {\n CDN_ROOT = root\n }\n \n+export function setOnigasmWASM(path: string) {\n+ ONIGASM_WASM = path\n+}\n+\n let _onigasmPromise: Promise<IOnigLib> = null\n \n export async function getOnigasm(): Promise<IOnigLib> {\n@@ -17,7 +22,7 @@ export async function getOnigasm(): Promise<IOnigLib> {\n let loader: Promise<any>\n \n if (isBrowser) {\n- loader = Onigasm.loadWASM(_resolvePath('onigasm.wasm', 'dist/'))\n+ loader = Onigasm.loadWASM(ONIGASM_WASM || _resolvePath('onigasm.wasm', 'dist/'))\n } else {\n const path = require('path')\n const onigasmPath = path.join(require.resolve('onigasm'), '../onigasm.wasm')\ndiff --git a/yarn.lock b/yarn.lock\nindex c143969..dfd7540 100644\n--- a/yarn.lock\n+++ b/yarn.lock\n@@ -2487,6 +2487,11 @@ combined-stream@^1.0.6, combined-stream@~1.0.6:\n dependencies:\n delayed-stream \"~1.0.0\"\n \n+commander@^2.20.0, commander@~2.20.3:\n+ version \"2.20.3\"\n+ resolved \"https://registry.yarnpkg.com/commander/-/commander-2.20.3.tgz#fd485e84c03eb4881c20722ba48035e8531aeb33\"\n+ integrity sha512-GpVkmM8vF2vQUkj2LvZmD35JxeJOLCwJ9cUkugyk2nuhbv3+mJvpLYYt+0+USMxE+oj+ey/lJEnhZw75x/OMcQ==\n+\n commander@^4.0.0:\n version \"4.1.1\"\n resolved \"https://registry.yarnpkg.com/commander/-/commander-4.1.1.tgz#9fd602bd936294e9e9ef46a3f4d6964044b18068\"\n@@ -2497,11 +2502,6 @@ commander@^6.2.0:\n resolved \"https://registry.yarnpkg.com/commander/-/commander-6.2.0.tgz#b990bfb8ac030aedc6d11bc04d1488ffef56db75\"\n integrity sha512-zP4jEKbe8SHzKJYQmq8Y9gYjtO/POJLgIdKgV7B9qNmABVFVc+ctqSX6iXh4mCpJfRBOabiZ2YKPg8ciDw6C+Q==\n \n-commander@~2.20.3:\n- version \"2.20.3\"\n- resolved \"https://registry.yarnpkg.com/commander/-/commander-2.20.3.tgz#fd485e84c03eb4881c20722ba48035e8531aeb33\"\n- integrity sha512-GpVkmM8vF2vQUkj2LvZmD35JxeJOLCwJ9cUkugyk2nuhbv3+mJvpLYYt+0+USMxE+oj+ey/lJEnhZw75x/OMcQ==\n-\n commondir@^1.0.1:\n version \"1.0.1\"\n resolved \"https://registry.yarnpkg.com/commondir/-/commondir-1.0.1.tgz#ddd800da0c66127393cca5950ea968a3aaf1253b\"\n@@ -4799,7 +4799,7 @@ jest-watcher@^26.6.2:\n jest-util \"^26.6.2\"\n string-length \"^4.0.1\"\n \n-jest-worker@^26.6.2:\n+jest-worker@^26.2.1, jest-worker@^26.6.2:\n version \"26.6.2\"\n resolved \"https://registry.yarnpkg.com/jest-worker/-/jest-worker-26.6.2.tgz#7f72cbc4d643c365e27b9fd775f9d0eaa9c7a8ed\"\n integrity sha512-KWYVV1c4i+jbMpaBC+U++4Va0cp8OisU185o73T1vo99hqi7w8tSJfUXYswwqqrjzwxa6KpRK54WhPvwf5w6PQ==\n@@ -6444,6 +6444,13 @@ quick-lru@^4.0.1:\n resolved \"https://registry.yarnpkg.com/quick-lru/-/quick-lru-4.0.1.tgz#5b8878f113a58217848c6482026c73e1ba57727f\"\n integrity sha512-ARhCpm70fzdcvNQfPoy49IaanKkTlRWF2JMzqhcJbhSFRZv7nPTvZJdcY7301IPmvW+/p0RgIWnQDLJxifsQ7g==\n \n+randombytes@^2.1.0:\n+ version \"2.1.0\"\n+ resolved \"https://registry.yarnpkg.com/randombytes/-/randombytes-2.1.0.tgz#df6f84372f0270dc65cdf6291349ab7a473d4f2a\"\n+ integrity sha512-vYl3iOX+4CKUWuxGi9Ukhie6fsqXqS9FE2Zaic4tNFD2N2QQaXOMFbuKK4QmDHC0JO6B1Zp41J0LpT0oR68amQ==\n+ dependencies:\n+ safe-buffer \"^5.1.0\"\n+\n react-is@^17.0.1:\n version \"17.0.1\"\n resolved \"https://registry.yarnpkg.com/react-is/-/react-is-17.0.1.tgz#5b3531bd76a645a4c9fb6e693ed36419e3301339\"\n@@ -6812,6 +6819,16 @@ rollup-plugin-dts@^2.0.0:\n optionalDependencies:\n \"@babel/code-frame\" \"^7.10.4\"\n \n+rollup-plugin-terser@^7.0.2:\n+ version \"7.0.2\"\n+ resolved \"https://registry.yarnpkg.com/rollup-plugin-terser/-/rollup-plugin-terser-7.0.2.tgz#e8fbba4869981b2dc35ae7e8a502d5c6c04d324d\"\n+ integrity sha512-w3iIaU4OxcF52UUXiZNsNeuXIMDvFrr+ZXK6bFZ0Q60qyVfq4uLptoS4bbq3paG3x216eQllFZX7zt6TIImguQ==\n+ dependencies:\n+ \"@babel/code-frame\" \"^7.10.4\"\n+ jest-worker \"^26.2.1\"\n+ serialize-javascript \"^4.0.0\"\n+ terser \"^5.0.0\"\n+\n rollup-plugin-typescript2@^0.29.0:\n version \"0.29.0\"\n resolved \"https://registry.yarnpkg.com/rollup-plugin-typescript2/-/rollup-plugin-typescript2-0.29.0.tgz#b7ad83f5241dbc5bdf1e98d9c3fca005ffe39e1a\"\n@@ -6873,7 +6890,7 @@ safe-buffer@^5.0.1, safe-buffer@^5.1.1, safe-buffer@^5.1.2, safe-buffer@~5.1.0, \n resolved \"https://registry.yarnpkg.com/safe-buffer/-/safe-buffer-5.1.2.tgz#991ec69d296e0313747d59bdfd2b745c35f8828d\"\n integrity sha512-Gd2UZBJDkXlY7GbJxfsE8/nvKkUEU1G38c1siN6QP6a9PT9MmHB8GnpscSmMJSoF8LOIrt8ud/wPtojys4G6+g==\n \n-safe-buffer@^5.2.0, safe-buffer@~5.2.0:\n+safe-buffer@^5.1.0, safe-buffer@^5.2.0, safe-buffer@~5.2.0:\n version \"5.2.1\"\n resolved \"https://registry.yarnpkg.com/safe-buffer/-/safe-buffer-5.2.1.tgz#1eaf9fa9bdb1fdd4ec75f58f9cdb4e6b7827eec6\"\n integrity sha512-rp3So07KcdmmKbGvgaNxQSJr7bGVSVk5S9Eq1F+ppbRo70+YeaDxkw5Dd8NPN+GD6bjnYm2VuPuCXmpuYvmCXQ==\n@@ -6937,6 +6954,13 @@ semver@^6.0.0, semver@^6.2.0, semver@^6.3.0:\n resolved \"https://registry.yarnpkg.com/semver/-/semver-6.3.0.tgz#ee0a64c8af5e8ceea67687b133761e1becbd1d3d\"\n integrity sha512-b39TBaTSfV6yBrapU89p5fKekE2m/NwnDocOVruQFS1/veMgdzuPcnOM34M6CwxW8jH/lxEa5rBoDeUwu5HHTw==\n \n+serialize-javascript@^4.0.0:\n+ version \"4.0.0\"\n+ resolved \"https://registry.yarnpkg.com/serialize-javascript/-/serialize-javascript-4.0.0.tgz#b525e1238489a5ecfc42afacc3fe99e666f4b1aa\"\n+ integrity sha512-GaNA54380uFefWghODBWEGisLZFj00nS5ACs6yHa9nLqlLpVLO8ChDGeKRjZnV4Nh4n0Qi7nhYZD/9fCPzEqkw==\n+ dependencies:\n+ randombytes \"^2.1.0\"\n+\n set-blocking@^2.0.0, set-blocking@~2.0.0:\n version \"2.0.0\"\n resolved \"https://registry.yarnpkg.com/set-blocking/-/set-blocking-2.0.0.tgz#045f9782d011ae9a6803ddd382b24392b3d890f7\"\n@@ -7140,7 +7164,7 @@ source-map-resolve@^0.5.0:\n source-map-url \"^0.4.0\"\n urix \"^0.1.0\"\n \n-source-map-support@^0.5.6:\n+source-map-support@^0.5.6, source-map-support@~0.5.19:\n version \"0.5.19\"\n resolved \"https://registry.yarnpkg.com/source-map-support/-/source-map-support-0.5.19.tgz#a98b62f86dcaf4f67399648c085291ab9e8fed61\"\n integrity sha512-Wonm7zOCIJzBGQdB+thsPar0kYuCIzYvxZwlBa87yi/Mdjv7Tip2cyVbLj5o0cFPN4EVkuTwb3GDDyUx2DGnGw==\n@@ -7163,7 +7187,7 @@ source-map@^0.6.0, source-map@^0.6.1, source-map@~0.6.1:\n resolved \"https://registry.yarnpkg.com/source-map/-/source-map-0.6.1.tgz#74722af32e9614e9c287a8d0bbde48b5e2f1a263\"\n integrity sha512-UjgapumWlbMhkBgzT7Ykc5YXUT46F0iKu8SGXq0bcwP5dz/h0Plj6enJqjz1Zbq2l5WaqYnrVbwWOWMyF3F47g==\n \n-source-map@^0.7.3:\n+source-map@^0.7.3, source-map@~0.7.2:\n version \"0.7.3\"\n resolved \"https://registry.yarnpkg.com/source-map/-/source-map-0.7.3.tgz#5302f8169031735226544092e64981f751750383\"\n integrity sha512-CkCj6giN3S+n9qrYiBTX5gystlENnRW5jZeNLHpe6aue+SrHcG5VYwujhW9s4dY31mEGsxBDrHR6oI69fTXsaQ==\n@@ -7552,6 +7576,15 @@ terminal-link@^2.0.0:\n ansi-escapes \"^4.2.1\"\n supports-hyperlinks \"^2.0.0\"\n \n+terser@^5.0.0:\n+ version \"5.5.1\"\n+ resolved \"https://registry.yarnpkg.com/terser/-/terser-5.5.1.tgz#540caa25139d6f496fdea056e414284886fb2289\"\n+ integrity sha512-6VGWZNVP2KTUcltUQJ25TtNjx/XgdDsBDKGt8nN0MpydU36LmbPPcMBd2kmtZNNGVVDLg44k7GKeHHj+4zPIBQ==\n+ dependencies:\n+ commander \"^2.20.0\"\n+ source-map \"~0.7.2\"\n+ source-map-support \"~0.5.19\"\n+\n test-exclude@^6.0.0:\n version \"6.0.0\"\n resolved \"https://registry.yarnpkg.com/test-exclude/-/test-exclude-6.0.0.tgz#04a8698661d805ea6fa293b6cb9e63ac044ef15e\"\n"]
1
["ace6b981c8098a68092d4a10e75daae7b8bfee9b"]
["feat"]
offset tests for min and max read cursors
["diff --git a/storage/reads/array_cursor_test.go b/storage/reads/array_cursor_test.go\nindex 7c7ad0c..c1e6ff9 100644\n--- a/storage/reads/array_cursor_test.go\n+++ b/storage/reads/array_cursor_test.go\n@@ -1541,6 +1541,34 @@ func TestWindowMinArrayCursor(t *testing.T) {\n \t\t\t},\n \t\t},\n \t\t{\n+\t\t\tname: \"window offset\",\n+\t\t\tevery: time.Hour,\n+\t\t\toffset: 30 * time.Minute,\n+\t\t\tinputArrays: []*cursors.IntegerArray{\n+\t\t\t\tmakeIntegerArray(\n+\t\t\t\t\t16,\n+\t\t\t\t\tmustParseTime(\"2010-01-01T00:00:00Z\"), 15*time.Minute,\n+\t\t\t\t\tfunc(i int64) int64 {\n+\t\t\t\t\t\tbase := (i / 4) * 100\n+\t\t\t\t\t\tm := (i % 4) * 15\n+\t\t\t\t\t\treturn base + m\n+\t\t\t\t\t},\n+\t\t\t\t),\n+\t\t\t},\n+\t\t\twant: []*cursors.IntegerArray{\n+\t\t\t\t{\n+\t\t\t\t\tTimestamps: []int64{\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T00:00:00Z\").UnixNano(),\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T00:30:00Z\").UnixNano(),\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T01:30:00Z\").UnixNano(),\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T02:30:00Z\").UnixNano(),\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T03:30:00Z\").UnixNano(),\n+\t\t\t\t\t},\n+\t\t\t\t\tValues: []int64{0, 30, 130, 230, 330},\n+\t\t\t\t},\n+\t\t\t},\n+\t\t},\n+\t\t{\n \t\t\tname: \"window desc values\",\n \t\t\tevery: time.Hour,\n \t\t\tinputArrays: []*cursors.IntegerArray{\n@@ -1560,6 +1588,34 @@ func TestWindowMinArrayCursor(t *testing.T) {\n \t\t\t},\n \t\t},\n \t\t{\n+\t\t\tname: \"window offset desc values\",\n+\t\t\tevery: time.Hour,\n+\t\t\toffset: 30 * time.Minute,\n+\t\t\tinputArrays: []*cursors.IntegerArray{\n+\t\t\t\tmakeIntegerArray(\n+\t\t\t\t\t16,\n+\t\t\t\t\tmustParseTime(\"2010-01-01T00:00:00Z\"), 15*time.Minute,\n+\t\t\t\t\tfunc(i int64) int64 {\n+\t\t\t\t\t\tbase := (i / 4) * 100\n+\t\t\t\t\t\tm := 60 - (i%4)*15\n+\t\t\t\t\t\treturn base + m\n+\t\t\t\t\t},\n+\t\t\t\t),\n+\t\t\t},\n+\t\t\twant: []*cursors.IntegerArray{\n+\t\t\t\t{\n+\t\t\t\t\tTimestamps: []int64{\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T00:15:00Z\").UnixNano(),\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T00:45:00Z\").UnixNano(),\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T01:45:00Z\").UnixNano(),\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T02:45:00Z\").UnixNano(),\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T03:45:00Z\").UnixNano(),\n+\t\t\t\t\t},\n+\t\t\t\t\tValues: []int64{45, 15, 115, 215, 315},\n+\t\t\t\t},\n+\t\t\t},\n+\t\t},\n+\t\t{\n \t\t\tname: \"window min int\",\n \t\t\tevery: time.Hour,\n \t\t\tinputArrays: []*cursors.IntegerArray{\n@@ -1693,6 +1749,34 @@ func TestWindowMaxArrayCursor(t *testing.T) {\n \t\t\t},\n \t\t},\n \t\t{\n+\t\t\tname: \"window offset\",\n+\t\t\tevery: time.Hour,\n+\t\t\toffset: 30 * time.Minute,\n+\t\t\tinputArrays: []*cursors.IntegerArray{\n+\t\t\t\tmakeIntegerArray(\n+\t\t\t\t\t16,\n+\t\t\t\t\tmustParseTime(\"2010-01-01T00:00:00Z\"), 15*time.Minute,\n+\t\t\t\t\tfunc(i int64) int64 {\n+\t\t\t\t\t\tbase := (i / 4) * 100\n+\t\t\t\t\t\tm := (i % 4) * 15\n+\t\t\t\t\t\treturn base + m\n+\t\t\t\t\t},\n+\t\t\t\t),\n+\t\t\t},\n+\t\t\twant: []*cursors.IntegerArray{\n+\t\t\t\t{\n+\t\t\t\t\tTimestamps: []int64{\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T00:15:00Z\").UnixNano(),\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T01:15:00Z\").UnixNano(),\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T02:15:00Z\").UnixNano(),\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T03:15:00Z\").UnixNano(),\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T03:45:00Z\").UnixNano(),\n+\t\t\t\t\t},\n+\t\t\t\t\tValues: []int64{15, 115, 215, 315, 345},\n+\t\t\t\t},\n+\t\t\t},\n+\t\t},\n+\t\t{\n \t\t\tname: \"window desc values\",\n \t\t\tevery: time.Hour,\n \t\t\tinputArrays: []*cursors.IntegerArray{\n@@ -1712,6 +1796,34 @@ func TestWindowMaxArrayCursor(t *testing.T) {\n \t\t\t},\n \t\t},\n \t\t{\n+\t\t\tname: \"window offset desc values\",\n+\t\t\tevery: time.Hour,\n+\t\t\toffset: 30 * time.Minute,\n+\t\t\tinputArrays: []*cursors.IntegerArray{\n+\t\t\t\tmakeIntegerArray(\n+\t\t\t\t\t16,\n+\t\t\t\t\tmustParseTime(\"2010-01-01T00:00:00Z\"), 15*time.Minute,\n+\t\t\t\t\tfunc(i int64) int64 {\n+\t\t\t\t\t\tbase := (i / 4) * 100\n+\t\t\t\t\t\tm := 60 - (i%4)*15\n+\t\t\t\t\t\treturn base + m\n+\t\t\t\t\t},\n+\t\t\t\t),\n+\t\t\t},\n+\t\t\twant: []*cursors.IntegerArray{\n+\t\t\t\t{\n+\t\t\t\t\tTimestamps: []int64{\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T00:00:00Z\").UnixNano(),\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T01:00:00Z\").UnixNano(),\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T02:00:00Z\").UnixNano(),\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T03:00:00Z\").UnixNano(),\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T03:30:00Z\").UnixNano(),\n+\t\t\t\t\t},\n+\t\t\t\t\tValues: []int64{60, 160, 260, 360, 330},\n+\t\t\t\t},\n+\t\t\t},\n+\t\t},\n+\t\t{\n \t\t\tname: \"window min int\",\n \t\t\tevery: time.Hour,\n \t\t\tinputArrays: []*cursors.IntegerArray{\n"]
1
["b7e2330fa3a8d7b8a9bff01b707c44e64b845c7b"]
["test"]
right side menus
["diff --git a/ionic/components/menu/menu-types.scss b/ionic/components/menu/menu-types.scss\nindex dbbfdda..5e4f990 100644\n--- a/ionic/components/menu/menu-types.scss\n+++ b/ionic/components/menu/menu-types.scss\n@@ -35,3 +35,7 @@ ion-menu[type=overlay] {\n }\n }\n }\n+\n+ion-menu[type=overlay][side=right] {\n+ left: 8px;\n+}\ndiff --git a/ionic/components/menu/menu-types.ts b/ionic/components/menu/menu-types.ts\nindex 360aeb4..0666a38 100644\n--- a/ionic/components/menu/menu-types.ts\n+++ b/ionic/components/menu/menu-types.ts\n@@ -10,7 +10,7 @@ import {Animation} from 'ionic/animations/animation';\n */\n export class MenuType {\n \n- constructor(menu: Menu) {\n+ constructor() {\n this.open = new Animation();\n this.close = new Animation();\n }\n@@ -88,16 +88,17 @@ class MenuRevealType extends MenuType {\n let duration = 250;\n \n let openedX = (menu.width() * (menu.side == 'right' ? -1 : 1)) + 'px';\n+ let closedX = '0px'\n \n this.open.easing(easing).duration(duration);\n this.close.easing(easing).duration(duration);\n \n let contentOpen = new Animation(menu.getContentElement());\n- contentOpen.fromTo(TRANSLATE_X, CENTER, openedX);\n+ contentOpen.fromTo(TRANSLATE_X, closedX, openedX);\n this.open.add(contentOpen);\n \n let contentClose = new Animation(menu.getContentElement());\n- contentClose.fromTo(TRANSLATE_X, openedX, CENTER);\n+ contentClose.fromTo(TRANSLATE_X, openedX, closedX);\n this.close.add(contentClose);\n }\n }\n@@ -117,13 +118,23 @@ class MenuOverlayType extends MenuType {\n let duration = 250;\n let backdropOpacity = 0.5;\n \n- let closedX = (menu.width() * (menu.side == 'right' ? 1 : -1)) + 'px';\n+ let closedX, openedX;\n+ if (menu.side == 'right') {\n+ // right side\n+ closedX = menu.platform.width() + 'px';\n+ openedX = (menu.platform.width() - menu.width() - 8) + 'px';\n+\n+ } else {\n+ // left side\n+ closedX = -menu.width() + 'px';\n+ openedX = '8px';\n+ }\n \n this.open.easing(easing).duration(duration);\n this.close.easing(easing).duration(duration);\n \n let menuOpen = new Animation(menu.getMenuElement());\n- menuOpen.fromTo(TRANSLATE_X, closedX, '8px');\n+ menuOpen.fromTo(TRANSLATE_X, closedX, openedX);\n this.open.add(menuOpen);\n \n let backdropOpen = new Animation(menu.getBackdropElement());\n@@ -131,7 +142,7 @@ class MenuOverlayType extends MenuType {\n this.open.add(backdropOpen);\n \n let menuClose = new Animation(menu.getMenuElement());\n- menuClose.fromTo(TRANSLATE_X, '8px', closedX);\n+ menuClose.fromTo(TRANSLATE_X, openedX, closedX);\n this.close.add(menuClose);\n \n let backdropClose = new Animation(menu.getBackdropElement());\n@@ -144,4 +155,3 @@ Menu.register('overlay', MenuOverlayType);\n \n const OPACITY = 'opacity';\n const TRANSLATE_X = 'translateX';\n-const CENTER = '0px';\ndiff --git a/ionic/components/menu/menu.ts b/ionic/components/menu/menu.ts\nindex d911b2e..e980069 100644\n--- a/ionic/components/menu/menu.ts\n+++ b/ionic/components/menu/menu.ts\n@@ -4,6 +4,7 @@ import {Ion} from '../ion';\n import {IonicApp} from '../app/app';\n import {IonicConfig} from '../../config/config';\n import {IonicComponent} from '../../config/annotations';\n+import {IonicPlatform} from '../../platform/platform';\n import * as gestures from './menu-gestures';\n \n \n@@ -35,10 +36,16 @@ import * as gestures from './menu-gestures';\n })\n export class Menu extends Ion {\n \n- constructor(app: IonicApp, elementRef: ElementRef, config: IonicConfig) {\n+ constructor(\n+ app: IonicApp,\n+ elementRef: ElementRef,\n+ config: IonicConfig,\n+ platform: IonicPlatform\n+ ) {\n super(elementRef, config);\n-\n this.app = app;\n+ this.platform = platform;\n+\n this.opening = new EventEmitter('opening');\n this.isOpen = false;\n this._disableTime = 0;\n@@ -46,9 +53,9 @@ export class Menu extends Ion {\n \n onInit() {\n super.onInit();\n- this.contentElement = (this.content instanceof Node) ? this.content : this.content.getNativeElement();\n+ this._cntEle = (this.content instanceof Node) ? this.content : this.content.getNativeElement();\n \n- if (!this.contentElement) {\n+ if (!this._cntEle) {\n return console.error('Menu: must have a [content] element to listen for drag events on. Example:\\n\\n<ion-menu [content]=\"content\"></ion-menu>\\n\\n<ion-content #content></ion-content>');\n }\n \n@@ -61,8 +68,8 @@ export class Menu extends Ion {\n this._initGesture();\n this._initType(this.type);\n \n- this.contentElement.classList.add('menu-content');\n- this.contentElement.classList.add('menu-content-' + this.type);\n+ this._cntEle.classList.add('menu-content');\n+ this._cntEle.classList.add('menu-content-' + this.type);\n \n let self = this;\n this.onContentClick = function(ev) {\n@@ -161,11 +168,11 @@ export class Menu extends Ion {\n \n this.isOpen = isOpen;\n \n- this.contentElement.classList[isOpen ? 'add' : 'remove']('menu-content-open');\n+ this._cntEle.classList[isOpen ? 'add' : 'remove']('menu-content-open');\n \n- this.contentElement.removeEventListener('click', this.onContentClick);\n+ this._cntEle.removeEventListener('click', this.onContentClick);\n if (isOpen) {\n- this.contentElement.addEventListener('click', this.onContentClick);\n+ this._cntEle.addEventListener('click', this.onContentClick);\n \n } else {\n this.getNativeElement().classList.remove('show-menu');\n@@ -220,7 +227,7 @@ export class Menu extends Ion {\n * @return {Element} The Menu's associated content element.\n */\n getContentElement() {\n- return this.contentElement;\n+ return this._cntEle;\n }\n \n /**\n@@ -239,7 +246,7 @@ export class Menu extends Ion {\n this.app.unregister(this.id);\n this._gesture && this._gesture.destroy();\n this._type && this._type.onDestroy();\n- this.contentElement = null;\n+ this._cntEle = null;\n }\n \n }\ndiff --git a/ionic/components/menu/test/basic/index.ts b/ionic/components/menu/test/basic/index.ts\nindex 698cec4..65952ff 100644\n--- a/ionic/components/menu/test/basic/index.ts\n+++ b/ionic/components/menu/test/basic/index.ts\n@@ -36,9 +36,9 @@ class E2EApp {\n ];\n }\n \n- openPage(menu, page) {\n+ openPage(page) {\n // close the menu when clicking a link from the menu\n- menu.close();\n+ this.app.getComponent('leftMenu').close();\n \n // Reset the content nav to have just this page\n // we wouldn't want the back button to show in this scenario\ndiff --git a/ionic/components/menu/test/basic/main.html b/ionic/components/menu/test/basic/main.html\nindex 9bdeb5c..4905ae6 100644\n--- a/ionic/components/menu/test/basic/main.html\n+++ b/ionic/components/menu/test/basic/main.html\n@@ -1,4 +1,4 @@\n-<ion-menu #menu [content]=\"content\">\n+<ion-menu [content]=\"content\" id=\"leftMenu\">\n \n <ion-toolbar secondary>\n <ion-title>Left Menu</ion-title>\n@@ -8,11 +8,35 @@\n \n <ion-list>\n \n- <button ion-item *ng-for=\"#p of pages\" (click)=\"openPage(menu, p)\">\n+ <button ion-item *ng-for=\"#p of pages\" (click)=\"openPage(p)\">\n {{p.title}}\n </button>\n \n- <button ion-item menu-toggle no-forward-icon class=\"e2eCloseMenu\">\n+ <button ion-item menu-toggle=\"leftMenu\" no-forward-icon class=\"e2eCloseMenu\">\n+ Close Menu\n+ </button>\n+\n+ </ion-list>\n+ </ion-content>\n+\n+</ion-menu>\n+\n+\n+<ion-menu side=\"right\" [content]=\"content\" id=\"rightMenu\">\n+\n+ <ion-toolbar secondary>\n+ <ion-title>Right Menu</ion-title>\n+ </ion-toolbar>\n+\n+ <ion-content>\n+\n+ <ion-list>\n+\n+ <button ion-item *ng-for=\"#p of pages\" (click)=\"openPage(p)\">\n+ {{p.title}}\n+ </button>\n+\n+ <button ion-item menu-toggle=\"rightMenu\" no-forward-icon class=\"e2eCloseMenu\">\n Close Menu\n </button>\n \ndiff --git a/ionic/components/menu/test/basic/page1.html b/ionic/components/menu/test/basic/page1.html\nindex 1881d9e..2bc5c79 100644\n--- a/ionic/components/menu/test/basic/page1.html\n+++ b/ionic/components/menu/test/basic/page1.html\n@@ -1,7 +1,7 @@\n \n <ion-navbar *navbar>\n \n- <a menu-toggle>\n+ <a menu-toggle=\"leftMenu\">\n <icon menu></icon>\n </a>\n \n@@ -21,19 +21,23 @@\n </button>\n </ion-nav-items>\n \n- <a menu-toggle secondary>\n+ <a menu-toggle=\"rightMenu\" secondary>\n <icon menu></icon>\n </a>\n \n </ion-navbar>\n \n \n-<ion-content #content padding>\n+<ion-content padding>\n \n <h3>Page 1</h3>\n \n <p>\n- <button class=\"e2eContentToggleMenu\" menu-toggle>Toggle Menu</button>\n+ <button class=\"e2eContentToggleMenu\" menu-toggle=\"leftMenu\">Toggle Left Menu</button>\n+ </p>\n+\n+ <p>\n+ <button class=\"e2eContentToggleMenu\" menu-toggle=\"rightMenu\">Toggle Right Menu</button>\n </p>\n \n <f></f><f></f><f></f><f></f><f></f><f></f><f></f><f></f>\ndiff --git a/ionic/components/menu/test/basic/page2.html b/ionic/components/menu/test/basic/page2.html\nindex 9801c4f..098f3e1 100644\n--- a/ionic/components/menu/test/basic/page2.html\n+++ b/ionic/components/menu/test/basic/page2.html\n@@ -1,7 +1,7 @@\n \n <ion-navbar *navbar>\n \n- <a menu-toggle>\n+ <a menu-toggle=\"leftMenu\">\n <icon menu></icon>\n </a>\n \n@@ -11,12 +11,12 @@\n \n </ion-navbar>\n \n-<ion-content #content padding>\n+<ion-content padding>\n \n <h3>Page 2</h3>\n \n <p>\n- <button menu-toggle class=\"e2eContentToggleMenu\">Toggle Menu</button>\n+ <button menu-toggle=\"leftMenu\" class=\"e2eContentToggleMenu\">Toggle Left Menu</button>\n </p>\n \n <p>\ndiff --git a/ionic/components/menu/test/basic/page3.html b/ionic/components/menu/test/basic/page3.html\nindex a2d65e2..079a3e9 100644\n--- a/ionic/components/menu/test/basic/page3.html\n+++ b/ionic/components/menu/test/basic/page3.html\n@@ -1,7 +1,7 @@\n \n <ion-navbar *navbar>\n \n- <a menu-toggle>\n+ <a menu-toggle=\"leftMenu\">\n <icon menu></icon>\n </a>\n \n@@ -12,12 +12,12 @@\n </ion-navbar>\n \n \n-<ion-content #content padding>\n+<ion-content padding>\n \n <h3>Page 3</h3>\n \n <p>\n- <button menu-toggle>Toggle Menu</button>\n+ <button menu-toggle=\"leftMenu\">Toggle Left Menu</button>\n </p>\n \n <f></f><f></f><f></f><f></f><f></f><f></f><f></f><f></f>\ndiff --git a/ionic/components/toolbar/modes/md.scss b/ionic/components/toolbar/modes/md.scss\nindex 984e758..339169a 100644\n--- a/ionic/components/toolbar/modes/md.scss\n+++ b/ionic/components/toolbar/modes/md.scss\n@@ -43,6 +43,12 @@ $toolbar-md-button-font-size: 1.4rem !default;\n }\n }\n \n+ [menu-toggle][secondary],\n+ [menu-toggle][secondary].activated {\n+ margin: 0 2px;\n+ min-width: 28px;\n+ }\n+\n }\n \n ion-title {\n"]
1
["1a60540f2bcda48d33f015e31f3728ac2c59a159"]
["feat"]
fix typos (#90)
["diff --git a/README.md b/README.md\nindex de15ac5..5ad8b47 100755\n--- a/README.md\n+++ b/README.md\n@@ -16,13 +16,13 @@ content that will be loaded, similar to Facebook cards loaders.\n \n ## Features\n \n-* :gear: **Complety customizable:** you can change the colors, speed and sizes;\n+* :gear: **Completely customizable:** you can change the colors, speed and sizes;\n * :pencil2: **Create your own loading:** use the\n [create-react-content-loader](https://danilowoz.github.io/create-react-content-loader/) to create\n- your customs loadings easily;\n+ your custom loadings easily;\n * :ok_hand: **You can use right now:** there are a lot of presets to use the loader, see the\n [options](#options);\n-* :rocket: **Perfomance:** react-content-loader uses pure SVG to work, so it's works without any extra scritpt,\n+* :rocket: **Performance:** react-content-loader uses pure SVG to work, so it works without any extra scripts,\n canvas, etc;\n \n ## Usage\n"]
1
["88257ee720ed8ba136d49087c0d31373e8397dd5"]
["docs"]
update CI images from docker buster to bullseye This will break `perf_image` until the new CI image is built due to the newly required `--all-tags` parameter to `docker push` that isn't available for the docker version we run on buster.
["diff --git a/.circleci/config.yml b/.circleci/config.yml\nindex f8a53ba..c378c7e 100644\n--- a/.circleci/config.yml\n+++ b/.circleci/config.yml\n@@ -336,7 +336,7 @@ jobs:\n # Disabling for now, and tracked further investigations\n # in https://github.com/influxdata/k8s-idpe/issues/3038\n docker_layer_caching: false\n- version: 19.03.14\n+ version: 20.10.7\n - run: |\n sudo apt-get update\n sudo apt-get install -y docker.io\n@@ -355,7 +355,7 @@ jobs:\n BRANCH=$(git rev-parse --abbrev-ref HEAD | tr '/' '.')\n COMMIT_SHA=$(git rev-parse --short HEAD)\n docker build -t quay.io/influxdb/iox:$COMMIT_SHA -t quay.io/influxdb/iox:main -f docker/Dockerfile.iox .\n- docker push quay.io/influxdb/iox\n+ docker push --all-tags quay.io/influxdb/iox\n echo \"export COMMIT_SHA=${COMMIT_SHA}\" >> $BASH_ENV\n - run:\n name: Deploy tags\ndiff --git a/Dockerfile b/Dockerfile\nindex 8c23ea2..1df1fd2 100644\n--- a/Dockerfile\n+++ b/Dockerfile\n@@ -17,7 +17,7 @@ RUN \\\n cp /influxdb_iox/target/release/influxdb_iox /root/influxdb_iox && \\\n du -cshx /usr/local/cargo/registry /usr/local/cargo/git /influxdb_iox/target\n \n-FROM debian:buster-slim\n+FROM debian:bullseye-slim\n \n RUN apt-get update \\\n && apt-get install -y libssl1.1 libgcc1 libc6 ca-certificates --no-install-recommends \\\ndiff --git a/docker/Dockerfile.ci b/docker/Dockerfile.ci\nindex db0a8ca..cf9cd15 100644\n--- a/docker/Dockerfile.ci\n+++ b/docker/Dockerfile.ci\n@@ -12,7 +12,7 @@\n \n ARG RUST_VERSION\n # Build actual image used for CI pipeline\n-FROM rust:${RUST_VERSION}-slim-buster\n+FROM rust:${RUST_VERSION}-slim-bullseye\n \n # When https://github.com/rust-lang/rustup/issues/2686 is fixed, run the command added that\n # will install everything in rust-toolchain.toml here so that components are in the container\n@@ -42,7 +42,7 @@ COPY docker/redpanda.gpg /tmp/redpanda.gpg\n # Generated from https://packages.vectorized.io/nzc4ZYQK3WRGd9sy/redpanda/cfg/setup/bash.deb.sh\n RUN apt-key add /tmp/redpanda.gpg \\\n && rm /tmp/redpanda.gpg \\\n- && curl ${CURL_FLAGS} \"https://packages.vectorized.io/nzc4ZYQK3WRGd9sy/redpanda/config.deb.txt?distro=debian&codename=buster&version=10&arch=x86_64\" \\\n+ && curl ${CURL_FLAGS} \"https://packages.vectorized.io/nzc4ZYQK3WRGd9sy/redpanda/config.deb.txt?distro=debian&codename=bullseye&version=10&arch=x86_64\" \\\n > /etc/apt/sources.list.d/vectorized-redpanda.list \\\n && apt-get update \\\n && apt-get install -y redpanda \\\ndiff --git a/docker/Dockerfile.iox b/docker/Dockerfile.iox\nindex 42414db..ae1f38e 100644\n--- a/docker/Dockerfile.iox\n+++ b/docker/Dockerfile.iox\n@@ -1,7 +1,7 @@\n ###\n # Dockerfile used for deploying IOx\n ##\n-FROM debian:buster-slim\n+FROM debian:bullseye-slim\n \n RUN apt-get update \\\n && apt-get install -y libssl1.1 libgcc1 libc6 ca-certificates gettext-base --no-install-recommends \\\n"]
1
["640cd88df3069a97d8244398414338dd317c5470"]
["ci"]
treeview width fix Signed-off-by: Raju Udava <[email protected]>
["diff --git a/tests/playwright/pages/Dashboard/TreeView.ts b/tests/playwright/pages/Dashboard/TreeView.ts\nindex 9cc622b..75c02c0 100644\n--- a/tests/playwright/pages/Dashboard/TreeView.ts\n+++ b/tests/playwright/pages/Dashboard/TreeView.ts\n@@ -23,10 +23,24 @@ export class TreeViewPage extends BasePage {\n }\n \n async verifyVisibility({ isVisible }: { isVisible: boolean }) {\n- if (isVisible) {\n- await expect(this.get()).toBeVisible();\n+ await this.rootPage.waitForTimeout(1000);\n+\n+ const domElement = await this.get();\n+ // get width of treeview dom element\n+ const width = (await domElement.boundingBox()).width;\n+\n+ // if (isVisible) {\n+ // await expect(this.get()).toBeVisible();\n+ // } else {\n+ // await expect(this.get()).not.toBeVisible();\n+ // }\n+\n+ // border for treeview is 1px\n+ // if not-visible, width should be < 5;\n+ if (!isVisible) {\n+ expect(width).toBeLessThan(5);\n } else {\n- await expect(this.get()).not.toBeVisible();\n+ expect(width).toBeGreaterThan(5);\n }\n }\n \n"]
1
["efeb30f26252ef4791ef2a02d83827b7f0c45462"]
["test"]
use new freespace config for disk space recory test
["diff --git a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/DiskSpaceRecoveryIT.java b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/DiskSpaceRecoveryIT.java\nindex 0854323..bfc7b7e 100644\n--- a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/DiskSpaceRecoveryIT.java\n+++ b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/DiskSpaceRecoveryIT.java\n@@ -47,7 +47,8 @@ final class DiskSpaceRecoveryIT {\n .withZeebeData(volume)\n .withEnv(\"ZEEBE_BROKER_DATA_LOGSEGMENTSIZE\", \"1MB\")\n .withEnv(\"ZEEBE_BROKER_NETWORK_MAXMESSAGESIZE\", \"1MB\")\n- .withEnv(\"ZEEBE_BROKER_DATA_DISKUSAGECOMMANDWATERMARK\", \"0.5\");\n+ .withEnv(\"ZEEBE_BROKER_DATA_DISK_FREESPACE_PROCESSING\", \"10MB\")\n+ .withEnv(\"ZEEBE_BROKER_DATA_DISK_FREESPACE_REPLICATION\", \"1MB\");\n \n private ZeebeClient client;\n \n@@ -127,7 +128,9 @@ final class DiskSpaceRecoveryIT {\n ContainerEngine.builder()\n .withDebugReceiverPort(SocketUtil.getNextAddress().getPort())\n .withContainer(\n- container.withEnv(\"ZEEBE_BROKER_DATA_DISKUSAGECOMMANDWATERMARK\", \"0.0001\"))\n+ container\n+ .withEnv(\"ZEEBE_BROKER_DATA_DISK_FREESPACE_PROCESSING\", \"16MB\")\n+ .withEnv(\"ZEEBE_BROKER_DATA_DISK_FREESPACE_REPLICATION\", \"10MB\"))\n .build();\n \n @BeforeEach\n"]
1
["672cd2b9775fb6dac2d522cb3f4469db47c0556b"]
["test"]
fix golden tests for aws_vpn_connection
["diff --git a/internal/providers/terraform/aws/testdata/vpn_connection_test/vpn_connection_test.tf b/internal/providers/terraform/aws/testdata/vpn_connection_test/vpn_connection_test.tf\nindex d895677..cf10e3f 100644\n--- a/internal/providers/terraform/aws/testdata/vpn_connection_test/vpn_connection_test.tf\n+++ b/internal/providers/terraform/aws/testdata/vpn_connection_test/vpn_connection_test.tf\n@@ -12,6 +12,7 @@ provider \"aws\" {\n resource \"aws_vpn_connection\" \"vpn_connection\" {\n customer_gateway_id = \"dummy-customer-gateway-id\"\n type = \"ipsec.1\"\n+ vpn_gateway_id = \"vpn-gateway-id\"\n }\n \n resource \"aws_vpn_connection\" \"transit\" {\n@@ -23,10 +24,11 @@ resource \"aws_vpn_connection\" \"transit\" {\n resource \"aws_vpn_connection\" \"vpn_connection_withUsage\" {\n customer_gateway_id = \"dummy-customer-gateway-id2\"\n type = \"ipsec.1\"\n+ vpn_gateway_id = \"vpn-gateway-id\"\n }\n \n resource \"aws_vpn_connection\" \"transit_withUsage\" {\n customer_gateway_id = \"dummy-customer-gateway-id2\"\n type = \"ipsec.1\"\n transit_gateway_id = \"dummy-transit-gateway-id2\"\n-}\n\\ No newline at end of file\n+}\n"]
1
["9b059dd8245e72f0bf8c40fc633f9ef6fccae405"]
["test"]
verify process responses for deploy command Tests should generally only fail for 1 reason, but the first test case (`shouldDeployResourceFromFile`) verifies multiple unrelated things. To align with the other test cases in this class, it makes sense that this test case only verifies that the gateway service was called with a specific request. We can extract the verification of the response into a separate test. This can also be applied to the shouldDeployMultipleResources test case.
["diff --git a/clients/java/src/test/java/io/camunda/zeebe/client/process/DeployResourceTest.java b/clients/java/src/test/java/io/camunda/zeebe/client/process/DeployResourceTest.java\nindex 1d96c74..b65d9f3 100644\n--- a/clients/java/src/test/java/io/camunda/zeebe/client/process/DeployResourceTest.java\n+++ b/clients/java/src/test/java/io/camunda/zeebe/client/process/DeployResourceTest.java\n@@ -22,7 +22,6 @@ import static org.assertj.core.api.Assertions.assertThatThrownBy;\n \n import io.camunda.zeebe.client.api.command.ClientException;\n import io.camunda.zeebe.client.api.response.DeploymentEvent;\n-import io.camunda.zeebe.client.api.response.Process;\n import io.camunda.zeebe.client.impl.command.StreamUtil;\n import io.camunda.zeebe.client.impl.response.ProcessImpl;\n import io.camunda.zeebe.client.util.ClientTest;\n@@ -35,7 +34,6 @@ import java.io.IOException;\n import java.io.InputStream;\n import java.nio.charset.StandardCharsets;\n import java.time.Duration;\n-import java.util.List;\n import org.junit.Test;\n \n public final class DeployResourceTest extends ClientTest {\n@@ -49,25 +47,15 @@ public final class DeployResourceTest extends ClientTest {\n @Test\n public void shouldDeployResourceFromFile() {\n // given\n- final long key = 123L;\n- final String filename = DeployResourceTest.class.getResource(BPMN_1_FILENAME).getPath();\n- gatewayService.onDeployResourceRequest(\n- key, deployedResource(deployedProcess(BPMN_1_PROCESS_ID, 12, 423, filename)));\n- final Process expected = new ProcessImpl(423, BPMN_1_PROCESS_ID, 12, filename);\n+ final String path = DeployResourceTest.class.getResource(BPMN_1_FILENAME).getPath();\n \n // when\n- final DeploymentEvent response =\n- client.newDeployCommand().addResourceFile(filename).send().join();\n+ client.newDeployCommand().addResourceFile(path).send().join();\n \n // then\n- assertThat(response.getKey()).isEqualTo(key);\n-\n- final List<Process> processes = response.getProcesses();\n- assertThat(processes).containsOnly(expected);\n-\n final DeployResourceRequest request = gatewayService.getLastRequest();\n final Resource resource = request.getResources(0);\n- assertThat(resource.getName()).isEqualTo(filename);\n+ assertThat(resource.getName()).isEqualTo(path);\n assertThat(resource.getContent().toByteArray()).isEqualTo(getBytes(BPMN_1_FILENAME));\n }\n \n@@ -114,7 +102,6 @@ public final class DeployResourceTest extends ClientTest {\n // then\n final DeployResourceRequest request = gatewayService.getLastRequest();\n final Resource resource = request.getResources(0);\n-\n assertThat(resource.getName()).isEqualTo(filename);\n assertThat(resource.getContent().toByteArray()).isEqualTo(getBytes(BPMN_1_FILENAME));\n }\n@@ -135,7 +122,6 @@ public final class DeployResourceTest extends ClientTest {\n // then\n final DeployResourceRequest request = gatewayService.getLastRequest();\n final Resource resource = request.getResources(0);\n-\n assertThat(resource.getName()).isEqualTo(filename);\n assertThat(resource.getContent().toByteArray()).isEqualTo(getBytes(BPMN_1_FILENAME));\n }\n@@ -152,7 +138,6 @@ public final class DeployResourceTest extends ClientTest {\n // then\n final DeployResourceRequest request = gatewayService.getLastRequest();\n final Resource resource = request.getResources(0);\n-\n assertThat(resource.getName()).isEqualTo(filename);\n assertThat(resource.getContent().toByteArray()).isEqualTo(getBytes(BPMN_1_FILENAME));\n }\n@@ -174,7 +159,6 @@ public final class DeployResourceTest extends ClientTest {\n // then\n final DeployResourceRequest request = gatewayService.getLastRequest();\n final Resource resource = request.getResources(0);\n-\n assertThat(resource.getName()).isEqualTo(filename);\n assertThat(resource.getContent().toByteArray()).isEqualTo(expectedBytes);\n }\n@@ -183,13 +167,58 @@ public final class DeployResourceTest extends ClientTest {\n public void shouldDeployMultipleResources() {\n // given\n final long key = 345L;\n-\n final String filename1 = BPMN_1_FILENAME.substring(1);\n final String filename2 = BPMN_2_FILENAME.substring(1);\n+ gatewayService.onDeployResourceRequest(\n+ key,\n+ deployedResource(deployedProcess(BPMN_1_PROCESS_ID, 1, 1, filename1)),\n+ deployedResource(deployedProcess(BPMN_2_PROCESS_ID, 1, 2, filename2)));\n \n- final Process expected1 = new ProcessImpl(1, BPMN_1_PROCESS_ID, 1, filename1);\n- final Process expected2 = new ProcessImpl(2, BPMN_2_PROCESS_ID, 1, filename2);\n+ // when\n+ client\n+ .newDeployCommand()\n+ .addResourceFromClasspath(filename1)\n+ .addResourceFromClasspath(filename2)\n+ .send()\n+ .join();\n \n+ // then\n+ final DeployResourceRequest request = gatewayService.getLastRequest();\n+ assertThat(request.getResourcesList()).hasSize(2);\n+\n+ final Resource resource1 = request.getResources(0);\n+ assertThat(resource1.getName()).isEqualTo(filename1);\n+ assertThat(resource1.getContent().toByteArray()).isEqualTo(getBytes(BPMN_1_FILENAME));\n+\n+ final Resource resource2 = request.getResources(1);\n+ assertThat(resource2.getName()).isEqualTo(filename2);\n+ assertThat(resource2.getContent().toByteArray()).isEqualTo(getBytes(BPMN_2_FILENAME));\n+ }\n+\n+ @Test\n+ public void shouldDeployProcessAsResource() {\n+ // given\n+ final long key = 123L;\n+ final String filename = DeployResourceTest.class.getResource(BPMN_1_FILENAME).getPath();\n+ gatewayService.onDeployResourceRequest(\n+ key, deployedResource(deployedProcess(BPMN_1_PROCESS_ID, 12, 423, filename)));\n+\n+ // when\n+ final DeploymentEvent response =\n+ client.newDeployCommand().addResourceFile(filename).send().join();\n+\n+ // then\n+ assertThat(response.getKey()).isEqualTo(key);\n+ assertThat(response.getProcesses())\n+ .containsExactly(new ProcessImpl(423, BPMN_1_PROCESS_ID, 12, filename));\n+ }\n+\n+ @Test\n+ public void shouldDeployMultipleProcessesAsResources() {\n+ // given\n+ final long key = 345L;\n+ final String filename1 = BPMN_1_FILENAME.substring(1);\n+ final String filename2 = BPMN_2_FILENAME.substring(1);\n gatewayService.onDeployResourceRequest(\n key,\n deployedResource(deployedProcess(BPMN_1_PROCESS_ID, 1, 1, filename1)),\n@@ -206,21 +235,10 @@ public final class DeployResourceTest extends ClientTest {\n \n // then\n assertThat(response.getKey()).isEqualTo(key);\n-\n- final List<Process> processes = response.getProcesses();\n- assertThat(processes).containsOnly(expected1, expected2);\n-\n- final DeployResourceRequest request = gatewayService.getLastRequest();\n- assertThat(request.getResourcesList()).hasSize(2);\n-\n- Resource resource = request.getResources(0);\n-\n- assertThat(resource.getName()).isEqualTo(filename1);\n- assertThat(resource.getContent().toByteArray()).isEqualTo(getBytes(BPMN_1_FILENAME));\n-\n- resource = request.getResources(1);\n- assertThat(resource.getName()).isEqualTo(filename2);\n- assertThat(resource.getContent().toByteArray()).isEqualTo(getBytes(BPMN_2_FILENAME));\n+ assertThat(response.getProcesses())\n+ .containsExactly(\n+ new ProcessImpl(1, BPMN_1_PROCESS_ID, 1, filename1),\n+ new ProcessImpl(2, BPMN_2_PROCESS_ID, 1, filename2));\n }\n \n @Test\n"]
1
["390eadc270d027493722cdbe9c8f4140d027e473"]
["test"]
fix height calc
["diff --git a/src/content/redux/modules/widget.ts b/src/content/redux/modules/widget.ts\nindex 96b54f1..bcfe3bb 100644\n--- a/src/content/redux/modules/widget.ts\n+++ b/src/content/redux/modules/widget.ts\n@@ -93,7 +93,7 @@ export const initState: WidgetState = {\n : _initConfig.panelWidth,\n height: isSaladictPopupPage\n ? 400\n- : 30 + _initConfig.dicts.selected.length * 30,\n+ : 30 + 30, // menubar + 1 dict hegiht\n },\n bowlRect: {\n x: 0,\n@@ -565,7 +565,7 @@ function listenNewSelection (\n mouseX,\n mouseY,\n lastPanelRect.width,\n- 30 + state.config.dicts.selected.length * 30,\n+ 30 + state.dictionaries.active.length * 30,\n )\n }\n }\n"]
1
["148cd56d096ba972e9706653c47052a07d5f9d08"]
["fix"]
temporary do no run "verify-ffmpeg.py" on Mac CI (#14986)
["diff --git a/vsts.yml b/vsts.yml\nindex 6cb0eb3..a058238 100644\n--- a/vsts.yml\n+++ b/vsts.yml\n@@ -86,13 +86,13 @@ jobs:\n killall Electron\n fi\n displayName: Make sure Electron isn't running from previous tests\n-\n- - bash: |\n- cd src\n- python electron/script/verify-ffmpeg.py --source-root \"$PWD\" --build-dir out/Default --ffmpeg-path out/ffmpeg\n- displayName: Verify non proprietary ffmpeg\n- condition: and(succeeded(), eq(variables['RUN_TESTS'], '1'))\n- timeoutInMinutes: 5\n+# FIXME(alexeykuzmin)\n+# - bash: |\n+# cd src\n+# python electron/script/verify-ffmpeg.py --source-root \"$PWD\" --build-dir out/Default --ffmpeg-path out/ffmpeg\n+# displayName: Verify non proprietary ffmpeg\n+# condition: and(succeeded(), eq(variables['RUN_TESTS'], '1'))\n+# timeoutInMinutes: 5\n \n - bash: |\n cd src\n"]
1
["9187415f5ee35d2e88dd834e413fc16bf19c5db1"]
["ci"]
101: fix import key cmd Signed-off-by: Sam Alba <[email protected]>
["diff --git a/docs/learn/101-use.md b/docs/learn/101-use.md\nindex 283c1c1..2ec10f9 100644\n--- a/docs/learn/101-use.md\n+++ b/docs/learn/101-use.md\n@@ -41,8 +41,7 @@ cd ./examples/todoapp\n The example app contains encrypted secrets and other pre-configured inputs, here is how to decrypt them:\n \n ```sh\n-curl -sfL https://releases.dagger.io/examples/key.txt >> ~/.config/dagger/keys.txt\n-dagger input list\n+dagger input list || curl -sfL https://releases.dagger.io/examples/key.txt >> ~/.config/dagger/keys.txt\n ```\n \n **Step 4**: Deploy!\n"]
1
["2b01808ec86fe9d8b4a93141a1b7f95e11fd6010"]
["docs"]
import flux-lsp v0.5.21
["diff --git a/ui/package.json b/ui/package.json\nindex 7a44aad..a36fc3d 100644\n--- a/ui/package.json\n+++ b/ui/package.json\n@@ -134,7 +134,7 @@\n \"dependencies\": {\n \"@influxdata/clockface\": \"2.3.4\",\n \"@influxdata/flux\": \"^0.5.1\",\n- \"@influxdata/flux-lsp-browser\": \"0.5.20\",\n+ \"@influxdata/flux-lsp-browser\": \"0.5.21\",\n \"@influxdata/giraffe\": \"0.29.0\",\n \"@influxdata/influx\": \"0.5.5\",\n \"@influxdata/influxdb-templates\": \"0.9.0\",\ndiff --git a/ui/yarn.lock b/ui/yarn.lock\nindex 99ae766..e6e2a47 100644\n--- a/ui/yarn.lock\n+++ b/ui/yarn.lock\n@@ -752,10 +752,10 @@\n resolved \"https://registry.yarnpkg.com/@influxdata/clockface/-/clockface-2.3.4.tgz#9c496601253e1d49cbeae29a7b9cfb54862785f6\"\n integrity sha512-mmz3YElK8Ho+1onEafuas6sVhIT638JA4NbDTO3bVJgK1TG7AnU4rQP+c6fj7vZSfvrIwtOwGaMONJTaww5o6w==\n \n-\"@influxdata/[email protected]\":\n- version \"0.5.20\"\n- resolved \"https://registry.yarnpkg.com/@influxdata/flux-lsp-browser/-/flux-lsp-browser-0.5.20.tgz#150d261bab869e130f6d00ee73ea4e859e8969e4\"\n- integrity sha512-gUy19t/QndkJPmyv7Lb56zXxaW5v7R9TslTHt0hB0GJjo7lmYkRfkD7DELdFHrD2e/CLtcNQBnczIMIGkII8Bw==\n+\"@influxdata/[email protected]\":\n+ version \"0.5.21\"\n+ resolved \"https://registry.yarnpkg.com/@influxdata/flux-lsp-browser/-/flux-lsp-browser-0.5.21.tgz#d5632f45e925c09bae9501a00fbef2ed55567f9e\"\n+ integrity sha512-lcUwKX1yj0QqGiusQFOVi7UPsvp6+qNX7Cwf9qqS5/dRwoh7c++nFVRdGNrSWlsbyRrPaAWBoZWEnghSnIf6DQ==\n \n \"@influxdata/flux@^0.5.1\":\n version \"0.5.1\"\n"]
1
["bfe32bf10e9b6d699f694fbd095af0b3f2e6275f"]
["build"]
add flag to wait for workflow instance result - with the flag withResult the create instance command will wait for the workflow to complete - optional a list of variable names can be specified to limit the fetched variables
["diff --git a/clients/zbctl/cmd/createInstance.go b/clients/zbctl/cmd/createInstance.go\nindex 016f115..85ac0be 100644\n--- a/clients/zbctl/cmd/createInstance.go\n+++ b/clients/zbctl/cmd/createInstance.go\n@@ -15,13 +15,15 @@ package cmd\n \n import (\n \t\"github.com/zeebe-io/zeebe/clients/go/commands\"\n+\t\"strings\"\n \n \t\"github.com/spf13/cobra\"\n )\n \n var (\n-\tcreateInstanceVersionFlag int32\n-\tcreateInstanceVariablesFlag string\n+\tcreateInstanceVersionFlag int32\n+\tcreateInstanceVariablesFlag string\n+\tcreateInstanceWithResultFlag []string\n )\n \n var createInstanceCmd = &cobra.Command{\n@@ -39,12 +41,29 @@ var createInstanceCmd = &cobra.Command{\n \t\t\treturn err\n \t\t}\n \n-\t\tresponse, err := zbCmd.Send()\n-\t\tif err != nil {\n-\t\t\treturn err\n-\t\t}\n+\t\tif createInstanceWithResultFlag == nil {\n+\t\t\tresponse, err := zbCmd.Send()\n+\t\t\tif err != nil {\n+\t\t\t\treturn err\n+\t\t\t}\n+\n+\t\t\treturn printJson(response)\n+\t\t} else {\n+\t\t\tvariableNames := []string{}\n+\t\t\tfor _, variableName := range createInstanceWithResultFlag {\n+\t\t\t\ttrimedVariableName := strings.TrimSpace(variableName)\n+\t\t\t\tif trimedVariableName != \"\" {\n+\t\t\t\t\tvariableNames = append(variableNames, trimedVariableName)\n+\t\t\t\t}\n+\t\t\t}\n+\t\t\tresponse, err := zbCmd.WithResult().FetchVariables(variableNames...).Send()\n+\t\t\tif err != nil {\n+\t\t\t\treturn err\n+\t\t\t}\n+\n+\t\t\treturn printJson(response)\n \n-\t\treturn printJson(response)\n+\t\t}\n \t},\n }\n \n@@ -58,4 +77,11 @@ func init() {\n \tcreateInstanceCmd.\n \t\tFlags().\n \t\tInt32Var(&createInstanceVersionFlag, \"version\", commands.LatestVersion, \"Specify version of workflow which should be executed.\")\n+\n+\tcreateInstanceCmd.\n+\t\tFlags().\n+\t\tStringSliceVar(&createInstanceWithResultFlag, \"withResult\", nil, \"Specify to await result of workflow, optional a list of variable names can be provided to limit the returned variables\")\n+\n+\t// hack to use --withResult without values\n+\tcreateInstanceCmd.Flag(\"withResult\").NoOptDefVal = \" \"\n }\n"]
1
["f3107f1a8eb124b55e775d23416540f49204a19e"]
["feat"]
licensing
["diff --git a/atomix/cluster/src/test/java/io/atomix/cluster/messaging/impl/NettyMessagingServiceTlsTest.java b/atomix/cluster/src/test/java/io/atomix/cluster/messaging/impl/NettyMessagingServiceTlsTest.java\nindex a4aee6b..bb523fa 100644\n--- a/atomix/cluster/src/test/java/io/atomix/cluster/messaging/impl/NettyMessagingServiceTlsTest.java\n+++ b/atomix/cluster/src/test/java/io/atomix/cluster/messaging/impl/NettyMessagingServiceTlsTest.java\n@@ -1,3 +1,18 @@\n+/*\n+ * Copyright \u00a9 2020 camunda services GmbH ([email protected])\n+ *\n+ * Licensed under the Apache License, Version 2.0 (the \"License\");\n+ * you may not use this file except in compliance with the License.\n+ * You may obtain a copy of the License at\n+ *\n+ * http://www.apache.org/licenses/LICENSE-2.0\n+ *\n+ * Unless required by applicable law or agreed to in writing, software\n+ * distributed under the License is distributed on an \"AS IS\" BASIS,\n+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n+ * See the License for the specific language governing permissions and\n+ * limitations under the License.\n+ */\n package io.atomix.cluster.messaging.impl;\n \n import static org.assertj.core.api.Assertions.assertThat;\n"]
1
["cbe62140ce219da84772e21e7cfb4b5c2a25c1b8"]
["docs"]
updates the readme to improve the readability and contributing sections
["diff --git a/.github/CONTRIBUTING.md b/.github/CONTRIBUTING.md\nindex 3c4dd8d..f8b8514 100644\n--- a/.github/CONTRIBUTING.md\n+++ b/.github/CONTRIBUTING.md\n@@ -21,7 +21,8 @@ Contributions are always welcome! Please use the following guidelines when contr\n - `chore` - Catch all or things that have to do with the build system, etc\n - `examples` - Changes to existing example, or a new example\n * The `COMPONENT` is optional, and may be a single file, directory, or logical component. Can be omitted if commit applies globally\n-5. Run the tests (`cargo test --no-std-features && cargo test --features yaml`)\n+5. Run the tests (`cargo test --features \"yaml unstable\"`)\n+5. Run the lints (`cargo build --features lints`) (requires a nightly compiler)\n 6. `git rebase` into concise commits and remove `--fixup`s (`git rebase -i HEAD~NUM` where `NUM` is number of commits back)\n 7. Push your changes back to your fork (`git push origin $your-branch`)\n 8. Create a pull request! (You can also create the pull request first, and we'll merge when ready. This a good way to discuss proposed changes.)\ndiff --git a/README.md b/README.md\nindex 9e6efce..b74405d 100644\n--- a/README.md\n+++ b/README.md\n@@ -31,7 +31,9 @@ Table of Contents\n * [More Information](#more-information)\n * [Video Tutorials](#video-tutorials)\n * [How to Contribute](#how-to-contribute)\n- * [Running the tests](#running-the-tests)\n+ * [Testing Code](#testing-code)\n+ * [Linting Code](#linting-code)\n+ * [Debugging Code](#debugging-code)\n * [Goals](#goals)\n * [Compatibility Policy](#compatibility-policy)\n * [Minimum Version of Rust](#minimum-version-of-rust)\n@@ -43,288 +45,83 @@ Created by [gh-md-toc](https://github.com/ekalinin/github-markdown-toc)\n \n ## What's New\n \n-Here's what's new in v2.18.0\n+Here's the highlights from v2.0.0 to v2.18.0\n \n * **Completions:** Adds completion support for Microsoft PowerShell! (Thanks to @Arnavion)\n-\n-Here's what's new in v2.17.1\n-\n-* Fixes a bug where using low index multiples was propagated to subcommands\n-\n-Here's what's new in v2.17.0\n-\n * Allows specifying the second to last positional argument as `multiple(true)` (i.e. things such as `mv <files>... <target>`)\n * Adds an `App::get_name` and `App::get_bin_name`\n-\n-Here's what's new in v2.16.4\n-\n-* Fixes bug that caused panic on subcommands with aliases\n * Conflicting argument errors are now symetrical, meaning more consistent and better usage suggestions\n-* Fixes typo in example `13a_enum_values_automatic`\n-* Fixes failing yaml example (#715)\n-* Fixes the `debug` feature (#716)\n-\n-Here's the highlights for v2.16.3\n-\n-* Fixes a bug where the derived display order isn't propagated\n-* **yaml-example:** fixes some inconsistent args in the example\n-\n-Here's the highlights for v2.16.2\n-\n-* Fixes a bug where single quotes are not escaped\n-\n-Here's the highlights for v2.16.1\n-\n-* **Help Message:** fixes a regression bug where args with multiple(true) threw off alignment\n-\n-Here's the highlights for v2.16.0\n-\n * **Completions:** adds automatic ZSH completion script generation support! :tada: :tada:\n-\n-Here's a gif of them in action!\n-\n-![zsh-comppletions](http://i.imgur.com/rwlMbAv.gif)\n-\n-Here's the highlights for v2.15.0\n-\n * **AppSettings:** adds new setting `AppSettings::AllowNegativeNumbers` which functions like `AllowLeadingHyphen` except only allows undefined negative numbers to pass parsing.\n-* Improves some of the documentation of `AppSettings` by moving variants into roughly alphabetical order\n-\n-Here's the highlights for v2.14.1 (Huge thanks to all the contributors who put in a lot of work this cycle! Especially @tormol @nabijaczleweli and @wdv4758h)\n-\n * Stabilize `clap_app!` macro (i.e. no longer need to use `unstable` feature)\n-* Fixes a bug that made determining when to auto-wrap long help messages inconsistent\n-* Fixes fish completions for nested subcommands\n-* Improve documentation around features\n-* Reword docs for `ErrorKind` and `App::settings`\n-* Fix tests that fail when the `suggestions` feature is disabled\n-* Fix the `OsString`-using doc-tests\n-* Tag non-rust code blocks as such instead of ignoring them\n-* Improve some errors about subcommands\n-* Makes sure the doc-tests don't fail before \"missing file\" in YAML tests\n * Deprecate `App::with_defaults`\n-* Make lints not enable other nightly-requiring features\n-\n-Here's the highlights for v2.14.0\n-\n-* One can now alias arguments either visibly (whichc appears in the help text) or invisibly just like subcommands!\n+* One can now alias arguments either visibly (which appears in the help text) or invisibly just like subcommands!\n * The `from_usage` parser now correctly handles non-ascii names / options and help!\n-* Fixes a bug in the `require_delimiter` code which caused some incorrect parses\n-* Fixes various typos in the docs\n-* Various other small performance improvements and enhancements\n-\n-Here's the highlights for v2.13.0\n-\n * **Value Delimiters:** fixes the confusion around implicitly setting value delimiters. (The default is to *not* use a delimiter unless explicitly set)\n-* **Docs:** Updates README.md with new website information and updated video tutorials info\n-* **Docs:** Updates the docs about removing implicit `value_delimiter(true)`\n-* **Docs:** Adds better examples on using default values\n-\n-\n-Here's the highlights for v2.12.1\n-\n-* Fixes a regression-bug where the old `{n}` newline char stopped being replaced a properly re-aligned newline\n-\n-Here's the highlights for v2.12.0\n-\n * Changes the default value delimiter rules (i.e. the default is `use_delimiter(false)` *unless* a setting/method that implies multiple values was used) **[Bugfix that *may* \"break\" code]**\n * If code breaks, simply add `Arg::use_delimiter(true)` to the affected args\n-* Updates the docs for the `Arg::multiple` method WRT value delimiters and default settings\n * Adds ability to hide the possible values from the help text on a per argument basis, instead of command wide\n * Allows for limiting detected terminal width (i.e. wrap at `x` length, unless the terminal width is *smaller*)\n-* Removes some redundant `contains()` checks for minor performance improvements\n-* Fixes a bug where valid args aren't recognized with the `AppSettings::AllowLeadingHyphen` setting\n * `clap` now ignores hard newlines in help messages and properly re-aligns text, but still wraps if the term width is too small\n-* Makes some minor changes to when next line help is automatically used\n * Adds support for the setting `Arg::require_delimiter` from YAML\n-* Removes the verbage about using `'{n}'` to insert newlines in help text from the docs (the normal `\\n` can now be used)\n-* Documents `AppSetting::DisableVersion`\n-\n-Here's the highlights for v2.11.3\n-\n * `clap` no longer requires one to use `{n}` inside help text to insert a newline that is properly aligned. One can now use the normal `\\n`.\n * `clap` now ignores hard newlines in help messages and properly re-aligns text, but still wraps if the term width is too small\n-* Supports setting `Arg::require_delimiter` from YAML\n-\n-Here's the highlights for v2.11.2\n-\n-* Makes some minor changes to when next line help is automatically used for improved wrapping\n-\n-Here's the highlights for v2.11.1\n-\n-* Fixes an issue where settings weren't propogated down through grand-child subcommands\n * Errors can now have custom description\n * Uses `term_size` instead of home-grown solution on Windows\n-* Updates deps with some minor bug fixes\n-\n-\n-Here's the highlights for v2.11.0\n-\n * Adds the ability to wrap help text intelligently on Windows!\n-* Moves docs to [docs.rs!](https://docs.rs/clap/)\n-* Fixes some usage strings that contain both args in groups and ones that conflict with each other\n-* Uses standard conventions for bash completion files, namely `{bin}.bash-completion`\n+* Moves docs to [docs.rs!](https://docs.rs/clap/)!\n * Automatically moves help text to the next line and wraps when term width is determined to be too small, or help text is too long\n * Vastly improves *development* error messages when using YAML\n-* Adds `App::with_defaults` to automatically use `crate_authors!` and `crate_version!` macros\n-* Other minor improvements and bug fixes\n-\n-Here's the highlights for v2.10.4\n-\n-* Fixes a bug where help is wrapped incorrectly and causing a panic with some non-English characters\n-\n-Here's the highlights for v2.10.3\n-\n-* Fixes a bug with non-English characters in help text wrapping, where the character is stripped or causes a panic\n-* Fixes an issue with `strsim` which caused a panic in some scenarios\n * Adds a shorthand way to ignore help text wrapping and use source formatting (i.e. `App::set_term_width(0)`)\n-\n-Here's the highlights for v2.10.2\n-\n-* Fixes a critical bug where the help message is printed twice\n-\n-Here's the highlights for v2.10.1\n-\n * **Help Subcommand:** fixes misleading usage string when using multi-level subcommmands such as `myprog help subcmd1 subcmd2`\n * **YAML:** allows using lists or single values with certain arg declarations for increased ergonomics\n-\n-\n-Here's the highlights for v2.10.0\n-\n-\n * **Fish Shell Completions:** one can generate a basic fish completions script at compile time!\n-* **External SubCommands:** fixes a bug which now correctly preserves external subcommand name along with args to said command (Minor breaking change that breaks no known real world code)\n-* **YAML Documentation:** fixes example 17's incorrect reference to arg_groups instead of groups\n-\n-\n-Here's the highlights for v2.9.3\n-\n * Adds the ability to generate completions to an `io::Write` object\n * Adds an `App::unset_setting` and `App::unset_settings`\n-* Fixes bug where only first arg in list of `required_unless_one` is recognized\n-* Fixes a typo bug `SubcommandsRequired`->`SubcommandRequired`\n-\n-\n-Here's the highlights for v2.9.2\n-\n-\n-* fixes bug where --help and --version short weren't added to the completion list\n-* improves completions allowing multiple bins to have seperate completion files\n-\n-Here's the highlights for v2.9.0\n-\n * **Completions:** one can now [generate a bash completions](https://docs.rs/clap/2.9.0/clap/struct.App.html#method.gen_completions) script at compile time! These completions work with options using [possible values](https://docs.rs/clap/2.9.0/clap/struct.Arg.html#method.possible_values), [subcommand aliases](https://docs.rs/clap/2.9.0/clap/struct.App.html#method.aliases), and even multiple levels of subcommands\n-* Minor bug fixes when using `AppSettings::TrailingVarArg` and `AppSettings::AllowLeadingHyphen`\n-\n-Here's the highlights for v2.8.0\n-\n * **Arg:** adds new optional setting [`Arg::require_delimiter`](https://docs.rs/clap/2.8.0/clap/struct.Arg.html#method.require_delimiter) which requires val delimiter to parse multiple values\n * The terminal sizing portion has been factored out into a separate crate, [term_size](https://crates.io/crates/term_size)\n-* Minor bug fixes\n-\n-\n-Here's the highlights for v2.7.1\n-\n-* **Options:**\n- * options using multiple values and delimiters no longer parse additional values after a trailing space (i.e. `prog -o 1,2 file.txt` parses as `1,2` for `-o` and `file.txt` for a positional arg)\n- * using options using multiple values and with an `=` no longer parse args after the trailing space as values (i.e. `prog -o=1 file.txt` parses as `1` for `-o` and `file.txt` for a positional arg)\n-\n-Here's the highlights for v2.7.0\n-\n+* Options using multiple values and delimiters no longer parse additional values after a trailing space (i.e. `prog -o 1,2 file.txt` parses as `1,2` for `-o` and `file.txt` for a positional arg)\n+* Using options using multiple values and with an `=` no longer parse args after the trailing space as values (i.e. `prog -o=1 file.txt` parses as `1` for `-o` and `file.txt` for a positional arg)\n * **Usage Strings:** `[FLAGS]` and `[ARGS]` are no longer blindly added to usage strings, instead only when applicable\n * `arg_enum!`: allows using more than one meta item, or things like `#[repr(C)]` with `arg_enum!`s\n * `App::print_help`: now prints the same as would have been printed by `--help` or the like\n-* **Help Messages:**\n- * prevents invoking `<cmd> help help` and displaying incorrect help message\n- * subcommand help messages requested via `<cmd> help <sub>` now correctly match `<cmd> <sub> --help`\n-* **`ArgGroup`s:**\n- * one can now specify groups which require AT LEAST one of the args\n- * allows adding multiple ArgGroups per Arg\n- * **Documentation:** vastly improves `ArgGroup` docs by adding better examples\n-* **Documentation:** fixes a bunch of typos in the documentation\n-\n-Here's the highlights for v2.6.0\n-\n+* Prevents invoking `<cmd> help help` and displaying incorrect help message\n+* Subcommand help messages requested via `<cmd> help <sub>` now correctly match `<cmd> <sub> --help`\n+* One can now specify groups which require AT LEAST one of the args\n+* Allows adding multiple ArgGroups per Arg\n * **Global Settings:** One can now set an `AppSetting` which is propogated down through child subcommands\n * **Terminal Wrapping:** Allows wrapping at specified term width (Even on Windows!) (can now set an absolute width to \"smart\" wrap at)\n * **SubCommands/Aliases:** adds support for visible aliases for subcommands (i.e. aliases that are dipslayed in the help message)\n * **Subcommands/Aliases:** when viewing the help of an alias, it now display help of the aliased subcommand\n-* Improves the default usage string when only a single positional arg is present\n * Adds new setting to stop delimiting values with `--` or `AppSettings::TrailingVarArg`\n-* `App::before_help` and `App::after_help` now correctly wrap\n-* Fixes bug where positional args are printed out of order when using templates\n-* Fixes bug where one can't override the auto-generated version or help flags\n-* Fixes issue where `App::before_help` wasn't printed\n-* Fixes a failing windows build\n-* Fixes bug where new color settings couldn't be converted from strings\n-* Adds missing YAML methods for App and Arg\n-* Allows printing version to any io::Write object\n-* Removes extra newline from help and version output\n-\n-Here's what's new in v.2.5.2\n-\n-* Removes trailing newlines from help and version output\n-* Allows printing version to any io::Write object\n-* Inter-links all types and pages\n-* Makes all publicly available types viewable in docs\n-* Fixes bug where one can't override version or help flags\n-* Fixes bug where args are printed out of order when using templates\n-* Fixes issue where `App::before_help` wasn't printed properly\n-\n-Here's what's new in v.2.5.0\n-\n * Subcommands now support aliases - think of them as hidden subcommands that dispatch to said subcommand automatically\n-\n-Here's what's new in v2.4.3\n-\n-* Bug Fixes\n- * Usage strings get de-deuplicated when there are args which are also part ``ArgGroup`s`\n- * Fixed times when `ArgGroup`s are duplicated in usage strings\n-* Improvements\n- * Positional arguments which are part of a group are now formatted in a more readable way (fewer brackets)\n- * Positional arguments use the standard `<>` brackets to reduce confusion\n- * The default help string for the `help` subcommand has been shortened to fit in 80 columns\n-\n-Here's the highlights from v2.4.0\n-\n+* Fixed times when `ArgGroup`s are duplicated in usage strings\n * **Before Help:** adds support for displaying info before help message\n * **Required Unless:** adds support for allowing args that are required unless certain other args are present\n-* Bug fixes\n-\n-Here's the highlights from v2.3.0\n-\n * **New Help Template Engine!**: Now you have full control over the layout of your help message. Major thanks to @hgrecco\n * **Pull crate Authors from Cargo.toml**: One can now use the `crate_authors!` macro to automatically pull the crate authors from their Cargo.toml file\n * **Colored Help Messages**: Help messages can now be optionally colored (See the `AppSettings::ColoredHelp` setting). Screenshot below.\n-* A bunch of bug fixes\n-\n-Here's the highlights from v2.2.1\n-\n * **Help text auto wraps and aligns at for subcommands too!** - Long help strings of subcommands will now properly wrap and align to term width on Linux and OS X. This can be turned off as well.\n-* Bug fixes\n-\n-An example of the optional colored help:\n-\n-![screenshot](http://i.imgur.com/7fs2h5j.png)\n-\n-Here's the highlights from v2.2.0\n-\n * **Help text auto wraps and aligns at term width!** - Long help strings will now properly wrap and align to term width on Linux and OS X (and presumably Unix too). This can be turned off as well.\n * **Can customize the order of opts, flags, and subcommands in help messages** - Instead of using the default alphabetical order, you can now re-arrange the order of your args and subcommands in help message. This helps to emphasize more popular or important options.\n- * **Can auto-derive the order from declaration order** - Have a bunch of args or subcommmands to re-order? You can now just derive the order from the declaration order!\n+* **Can auto-derive the order from declaration order** - Have a bunch of args or subcommmands to re-order? You can now just derive the order from the declaration order!\n * **Help subcommand now accepts other subcommands as arguments!** - Similar to other CLI precedents, the `help` subcommand can now accept other subcommands as arguments to display their help message. i.e. `$ myprog help mysubcmd` (*Note* these can even be nested heavily such as `$ myprog help subcmd1 subcmd2 subcmd3` etc.)\n+* **Default Values**: Args can now specify default values\n+* **Next Line Help**: Args can have help strings on the line following the argument (useful for long arguments, or those with many values). This can be set command-wide or for individual args\n \n-* Other minor bug fixes\n+Here's a gif of them in action!\n+\n+![zsh-comppletions](http://i.imgur.com/rwlMbAv.gif)\n \n An example of the help text wrapping at term width:\n \n ![screenshot](http://i.imgur.com/PAJzJJG.png)\n \n-In v2.1.2\n+An example of the optional colored help:\n+\n+![screenshot](http://i.imgur.com/7fs2h5j.png)\n \n- * **Default Values**: Args can now specify default values\n- * **Next Line Help**: Args can have help strings on the line following the argument (useful for long arguments, or those with many values). This can be set command-wide or for individual args\n- * **Documentation Examples**: The examples in the documentation have been vastly improved\n \n For full details, see [CHANGELOG.md](https://github.com/kbknapp/clap-rs/blob/master/CHANGELOG.md)\n \n@@ -697,6 +494,7 @@ features = [ \"suggestions\", \"color\" ]\n #### Opt-in features\n \n * **\"yaml\"**: Enables building CLIs from YAML documents. (builds dependency `yaml-rust`)\n+* **\"unstable\"**: Enables unstable `clap` features that may change from release to release\n \n ### Dependencies Tree\n \n@@ -707,6 +505,7 @@ The following graphic depicts `clap`s dependency graph (generated using [cargo-g\n * **Blue** Color: Dev dependency, only used while developing.\n \n ![clap dependencies](clap_dep_graph.png)\n+\n ### More Information\n \n You can find complete documentation on the [docs.rs](https://docs.rs/clap/) for this project.\n@@ -727,20 +526,65 @@ Another really great way to help is if you find an interesting, or helpful way i\n \n Please read [CONTRIBUTING.md](.github/CONTRIBUTING.md) before you start contributing.\n \n+\n+### Testing Code\n+\n To test with all features both enabled and disabled, you can run theese commands:\n \n ```sh\n $ cargo test --no-default-features\n-$ cargo test --features yaml\n+$ cargo test --features \"yaml unstable\"\n ```\n \n-If you have a nightly compiler you can append `--features lints` to both commands\n-to get style warnings and code smells; If you get one from code you think is fine,\n-you can ignore it by prepending `#[cfg_attr(feature=\"lints\", allow(lint_name))]`\n-to the function or impl block.\n+Alternatively, if you have [`just`](https://github.com/casey/just) installed you can run the prebuilt recipies. *Not* using `just` is prfeclty fine as well, it simply bundles commands automatically.\n+\n+For example, to test the code, as above simply run:\n+\n+```sh\n+$ just run-tests`\n+```\n+\n+From here on, I will lis the appropriate `cargo` command as well as the `just` command.\n+\n+Sometimes it's helpful to only run a subset of the tests, which can be done via:\n+\n+```sh\n+$ cargo test --test <test_name>\n+\n+# Or\n+\n+$ just run-test <test_name>\n+```\n \n-If you are debugging (or just trying to understand the code) you can enable the\n-\"debug\" feature which will trace function calls and brances in some parts of the code.\n+### Linting Code\n+\n+During the CI process `clap` runs against many different lints using [`clippy`](https://github.com/Manishearth/rust-clippy). In order to check if these lints pass on your own computer prior to submitting a PR you'll need a nightly compiler.\n+\n+In order to check the code for lints run either:\n+\n+```sh\n+$ rustup override add nightly\n+$ cargo build --features lints\n+$ rustup override remove\n+\n+# Or\n+\n+$ just lint\n+```\n+\n+### Debugging Code\n+\n+Another helpful technique is to see the `clap` debug output while developing features. In order to see the debug output while running the full test suite or individual tests, run:\n+\n+```sh\n+$ cargo test --features debug\n+\n+# Or for individual tests\n+$ cargo test --test <test_name> --features debug\n+\n+# The corresponding just command for individual debugging tests is:\n+$ just debug <test_name>\n+```\n \n ### Goals\n \n"]
1
["eb51316cdfdc7258d287ba13b67ef2f42bd2b8f6"]
["docs"]
enable performance test trigger This reverts commit 146c7b58154a5b3de957f87e3b193447e0576547.
["diff --git a/Jenkinsfile b/Jenkinsfile\nindex 399f8b8..c3f8fde 100644\n--- a/Jenkinsfile\n+++ b/Jenkinsfile\n@@ -120,6 +120,12 @@ pipeline {\n }\n }\n \n+ stage('Trigger Performance Tests') {\n+ when { branch 'develop' }\n+ steps {\n+ build job: 'zeebe-cluster-performance-tests', wait: false\n+ }\n+ }\n }\n \n post {\n"]
1
["80944b7a513b442afcb2d0d6c7d71c0d79365dba"]
["ci"]
better pin mode view
["diff --git a/src/content/redux/modules/widget.ts b/src/content/redux/modules/widget.ts\nindex 5df38c2..b8a1080 100644\n--- a/src/content/redux/modules/widget.ts\n+++ b/src/content/redux/modules/widget.ts\n@@ -627,7 +627,9 @@ function listenNewSelection (\n isSaladictOptionsPage\n ) {\n dispatch(searchText({ info: selectionInfo }))\n- } else {\n+ } else if (!shouldPanelShow) {\n+ // Everything stays the same if the panel is still visible (e.g. pin mode)\n+ // Otherwise clean up all dicts\n dispatch(restoreDicts())\n }\n })\n"]
1
["9c0aae64774a6fd864622474cb645371fee114b5"]
["refactor"]
add donation section to footer
["diff --git a/client/src/components/Feedback.tsx b/client/src/components/Feedback.tsx\nindex 117b21d..0d7e7a9 100644\n--- a/client/src/components/Feedback.tsx\n+++ b/client/src/components/Feedback.tsx\n@@ -16,12 +16,7 @@ const publicRoutes = [\n name: `\ud83d\udcdd Feedback on RS School`,\n link: `https://docs.google.com/forms/d/1F4NeS0oBq-CY805aqiPVp6CIrl4_nIYJ7Z_vUcMOFrQ/viewform`,\n newTab: true,\n- },\n- {\n- name: `\ud83d\udcb0 Make a donation`,\n- link: `https://www.patreon.com/therollingscopes`,\n- newTab: true,\n- },\n+ }\n ];\n \n type LinkInfo = { name: string; link: string; newTab: boolean };\ndiff --git a/client/src/components/FooterLayout.tsx b/client/src/components/FooterLayout.tsx\nindex 79c0f39..56661b4 100644\n--- a/client/src/components/FooterLayout.tsx\n+++ b/client/src/components/FooterLayout.tsx\n@@ -1,5 +1,5 @@\n import * as React from 'react';\n-import { Col, Layout, Row, Divider } from 'antd';\n+import { Col, Layout, Row, Divider, Button } from 'antd';\n import { Feedback } from './Feedback';\n import { Help } from './Help';\n import { SocialNetworks } from './SocialNetworks';\n@@ -23,9 +23,17 @@ class FooterLayout extends React.Component<any, any> {\n </Col>\n </Row>\n <Divider />\n- <div className=\"text-center\">\n- <small>&copy; The Rolling Scopes 2019</small>\n- </div>\n+ <h5> Thank you for your support! \ud83c\udf89</h5>\n+ <p>\n+ <object type=\"image/svg+xml\"\n+ data=\"https://opencollective.com/rsschool/backers.svg?avatarHeight=28&button=false\"></object>\n+ </p>\n+ <p>\n+ <Button size=\"small\" href=\"https://opencollective.com/rsschool#section-contribute\" target=\"_blank\" ghost>\n+ \u2764\ufe0f Make a donation\n+ </Button>\n+ </p>\n+ <p className=\"text-center\"><small>&copy; The Rolling Scopes 2019</small></p>\n </Footer>\n </div>\n );\ndiff --git a/client/src/styles/main.scss b/client/src/styles/main.scss\nindex cd61fcd..6e37ea6 100644\n--- a/client/src/styles/main.scss\n+++ b/client/src/styles/main.scss\n@@ -46,4 +46,7 @@ body,\n padding-right: 0;\n font-size: .7rem;\n }\n+ .ant-btn {\n+ font-size: .7rem;\n+ }\n }\n"]
1
["7704121d0c0bfce49f01c2b41cbc64a642cbb399"]
["feat"]
create mock img server,ensure "dist" dirs exist
["diff --git a/scripts/gulp/tasks/test.ts b/scripts/gulp/tasks/test.ts\nindex 8014b12..d10c1aa 100644\n--- a/scripts/gulp/tasks/test.ts\n+++ b/scripts/gulp/tasks/test.ts\n@@ -26,12 +26,18 @@ task('test.imageserver', () => {\n function handleRequest(req, res) {\n const urlParse = url.parse(req.url, true);\n \n+ res.setHeader('Access-Control-Allow-Origin', '*');\n+ res.setHeader('Access-Control-Allow-Methods', 'GET');\n+ res.setHeader('Connection', 'keep-alive');\n+ res.setHeader('Age', '0');\n+ res.setHeader('cache-control', 'no-store');\n+\n if (urlParse.pathname === '/reset') {\n console.log('Image Server Reset');\n console.log('---------------------------');\n requestedUrls.length = 0;\n start = Date.now();\n- res.setHeader('Access-Control-Allow-Origin', '*');\n+ res.setHeader('Content-Type', 'text/plain');\n res.end('reset');\n return;\n }\n@@ -48,9 +54,8 @@ task('test.imageserver', () => {\n \n setTimeout(() => {\n res.setHeader('Content-Type', 'image/svg+xml');\n- res.setHeader('Access-Control-Allow-Origin', '*');\n res.end(`<svg xmlns=\"http://www.w3.org/2000/svg\" xmlns:xlink=\"http://www.w3.org/1999/xlink\"\n- style=\"background-color: ${color}; width: ${width}px; height: ${height}px;\">\n+ viewBox=\"0 0 ${width} ${height}\" style=\"background-color: ${color};\">\n <text x=\"5\" y=\"22\" style=\"font-family: Courier; font-size: 24px\">${id}</text>\n </svg>`);\n }, delay);\n", "diff --git a/scripts/prepare.js b/scripts/prepare.js\nindex 9eb8cb8..f285825 100644\n--- a/scripts/prepare.js\n+++ b/scripts/prepare.js\n@@ -68,6 +68,9 @@ async function prepare() {\n names.push(json.name)\n }\n \n+ // Ensure all \"dist\" directories exist.\n+ dirs.forEach(dir => fs.ensureDirSync(join(dir, distId)))\n+\n log(``)\n for (let i = 0; i < names.length; i++) {\n const dir = dirs[i]\n"]
2
["32b76173a259ea1993298289b436cf10c1e800bf", "ca060bf255a55b99000ddf0c67f7422f28b735a6"]
["test", "build"]
use ng2 loadNextToLocation,fix build
["diff --git a/ionic/components/nav/nav-controller.ts b/ionic/components/nav/nav-controller.ts\nindex 8e23c4c..37ac0f4 100644\n--- a/ionic/components/nav/nav-controller.ts\n+++ b/ionic/components/nav/nav-controller.ts\n@@ -527,41 +527,13 @@ export class NavController extends Ion {\n * @private\n * TODO\n */\n- createViewComponentRef(type, hostProtoViewRef, viewContainer, viewCtrlBindings) {\n- let bindings = this.bindings.concat(viewCtrlBindings);\n-\n- // the same guts as DynamicComponentLoader.loadNextToLocation\n- var hostViewRef =\n- viewContainer.createHostView(hostProtoViewRef, viewContainer.length, bindings);\n- var newLocation = this._viewManager.getHostElement(hostViewRef);\n- var component = this._viewManager.getComponent(newLocation);\n-\n- var dispose = () => {\n- var index = viewContainer.indexOf(hostViewRef);\n- if (index !== -1) {\n- viewContainer.remove(index);\n- }\n- };\n-\n- // TODO: make-shift ComponentRef_, this is pretty much going to\n- // break in future versions of ng2, keep an eye on it\n- return {\n- location: newLocation,\n- instance: component,\n- dispose: dispose\n- };\n- }\n-\n- /**\n- * @private\n- * TODO\n- */\n- getBindings(viewCtrl) {\n- // create bindings to this ViewController and its NavParams\n- return this.bindings.concat(Injector.resolve([\n+ loadNextToAnchor(type, location, viewCtrl) {\n+ let bindings = this.bindings.concat(Injector.resolve([\n bind(ViewController).toValue(viewCtrl),\n bind(NavParams).toValue(viewCtrl.params),\n ]));\n+\n+ return this._loader.loadNextToLocation(type, location, bindings);\n }\n \n /**\ndiff --git a/ionic/components/nav/nav.ts b/ionic/components/nav/nav.ts\nindex a98a4ef..063eeb9 100644\n--- a/ionic/components/nav/nav.ts\n+++ b/ionic/components/nav/nav.ts\n@@ -192,65 +192,70 @@ export class Nav extends NavController {\n if (structure.tabs) {\n // the component being loaded is an <ion-tabs>\n // Tabs is essentially a pane, cuz it has its own navbar and content containers\n- let contentContainerRef = this._viewManager.getViewContainer(this.anchorElementRef());\n- let viewComponentRef = this.createViewComponentRef(componentType, hostProtoViewRef, contentContainerRef, this.getBindings(viewCtrl));\n- viewComponentRef.instance._paneView = true;\n+ this.loadNextToAnchor(componentType, this.anchorElementRef(), viewCtrl).then(componentRef => {\n \n- viewCtrl.disposals.push(() => {\n- viewComponentRef.dispose();\n- });\n+ componentRef.instance._paneView = true;\n+\n+ viewCtrl.disposals.push(() => {\n+ componentRef.dispose();\n+ });\n+\n+ viewCtrl.onReady().then(() => {\n+ done();\n+ });\n \n- viewCtrl.onReady().then(() => {\n- done();\n });\n \n } else {\n // normal ion-view going into pane\n this.getPane(structure, viewCtrl, (pane) => {\n // add the content of the view into the pane's content area\n- let viewComponentRef = this.createViewComponentRef(componentType, hostProtoViewRef, pane.contentContainerRef, this.getBindings(viewCtrl));\n- viewCtrl.disposals.push(() => {\n- viewComponentRef.dispose();\n+ this.loadNextToAnchor(componentType, pane.contentAnchorRef, viewCtrl).then(componentRef => {\n \n- // remove the pane if there are no view items left\n- pane.totalViews--;\n- if (pane.totalViews === 0) {\n- pane.dispose && pane.dispose();\n- }\n- });\n+ viewCtrl.disposals.push(() => {\n+ componentRef.dispose();\n \n- // count how many ViewControllers are in this pane\n- pane.totalViews++;\n+ // remove the pane if there are no view items left\n+ pane.totalViews--;\n+ if (pane.totalViews === 0) {\n+ pane.dispose && pane.dispose();\n+ }\n+ });\n \n- // a new ComponentRef has been created\n- // set the ComponentRef's instance to this ViewController\n- viewCtrl.setInstance(viewComponentRef.instance);\n+ // count how many ViewControllers are in this pane\n+ pane.totalViews++;\n \n- // remember the ElementRef to the content that was just created\n- viewCtrl.viewElementRef(viewComponentRef.location);\n+ // a new ComponentRef has been created\n+ // set the ComponentRef's instance to this ViewController\n+ viewCtrl.setInstance(componentRef.instance);\n \n- // get the NavController's container for navbars, which is\n- // the place this NavController will add each ViewController's navbar\n- let navbarContainerRef = pane.navbarContainerRef;\n+ // remember the ElementRef to the content that was just created\n+ viewCtrl.viewElementRef(componentRef.location);\n \n- // get this ViewController's navbar TemplateRef, which may not\n- // exist if the ViewController's template didn't have an <ion-navbar *navbar>\n- let navbarTemplateRef = viewCtrl.getNavbarTemplateRef();\n+ // get the NavController's container for navbars, which is\n+ // the place this NavController will add each ViewController's navbar\n+ let navbarContainerRef = pane.navbarContainerRef;\n \n- // create the navbar view if the pane has a navbar container, and the\n- // ViewController's instance has a navbar TemplateRef to go to inside of it\n- if (navbarContainerRef && navbarTemplateRef) {\n- let navbarView = navbarContainerRef.createEmbeddedView(navbarTemplateRef, -1);\n+ // get this ViewController's navbar TemplateRef, which may not\n+ // exist if the ViewController's template didn't have an <ion-navbar *navbar>\n+ let navbarTemplateRef = viewCtrl.getNavbarTemplateRef();\n \n- viewCtrl.disposals.push(() => {\n- let index = navbarContainerRef.indexOf(navbarView);\n- if (index > -1) {\n- navbarContainerRef.remove(index);\n- }\n- });\n- }\n+ // create the navbar view if the pane has a navbar container, and the\n+ // ViewController's instance has a navbar TemplateRef to go to inside of it\n+ if (navbarContainerRef && navbarTemplateRef) {\n+ let navbarView = navbarContainerRef.createEmbeddedView(navbarTemplateRef, -1);\n+\n+ viewCtrl.disposals.push(() => {\n+ let index = navbarContainerRef.indexOf(navbarView);\n+ if (index > -1) {\n+ navbarContainerRef.remove(index);\n+ }\n+ });\n+ }\n+\n+ done();\n+ });\n \n- done();\n });\n }\n }\n@@ -273,7 +278,7 @@ export class Nav extends NavController {\n \n } else {\n // create a new nav pane\n- this._loader.loadNextToLocation(Pane, this.anchorElementRef(), this.getBindings(viewCtrl)).then(componentRef => {\n+ this._loader.loadNextToLocation(Pane, this.anchorElementRef(), this.bindings).then(componentRef => {\n \n // get the pane reference\n pane = this.newPane;\n@@ -354,17 +359,6 @@ export class Nav extends NavController {\n \n /**\n * @private\n- * TODO\n- * @param {TODO} elementBinder TODO\n- * @param {TODO} id TODO\n- * @return {TODO} TODO\n- */\n-function isComponent(elementBinder, id) {\n- return (elementBinder && elementBinder.componentDirective && elementBinder.componentDirective.metadata.id == id);\n-}\n-\n-/**\n- * @private\n */\n @Directive({selector: 'template[pane-anchor]'})\n class NavPaneAnchor {\n@@ -393,9 +387,9 @@ class NavBarAnchor {\n class ContentAnchor {\n constructor(\n @Host() @Inject(forwardRef(() => Pane)) pane: Pane,\n- viewContainerRef: ViewContainerRef\n+ elementRef: ElementRef\n ) {\n- pane.contentContainerRef = viewContainerRef;\n+ pane.contentAnchorRef = elementRef;\n }\n }\n \ndiff --git a/ionic/components/tabs/tab.ts b/ionic/components/tabs/tab.ts\nindex aa21cad..af5d190 100644\n--- a/ionic/components/tabs/tab.ts\n+++ b/ionic/components/tabs/tab.ts\n@@ -153,40 +153,44 @@ export class Tab extends NavController {\n \n loadContainer(componentType, hostProtoViewRef, viewCtrl, done) {\n \n- let viewComponentRef = this.createViewComponentRef(componentType, hostProtoViewRef, this.contentContainerRef, this.getBindings(viewCtrl));\n- viewCtrl.disposals.push(() => {\n- viewComponentRef.dispose();\n- });\n+ this.loadNextToAnchor(componentType, this.contentAnchorRef, viewCtrl).then(componentRef => {\n \n- // a new ComponentRef has been created\n- // set the ComponentRef's instance to this ViewController\n- viewCtrl.setInstance(viewComponentRef.instance);\n+ viewCtrl.disposals.push(() => {\n+ componentRef.dispose();\n+ });\n \n- // remember the ElementRef to the content that was just created\n- viewCtrl.viewElementRef(viewComponentRef.location);\n+ // a new ComponentRef has been created\n+ // set the ComponentRef's instance to this ViewController\n+ viewCtrl.setInstance(componentRef.instance);\n \n- // get the NavController's container for navbars, which is\n- // the place this NavController will add each ViewController's navbar\n- let navbarContainerRef = this.tabs.navbarContainerRef;\n+ // remember the ElementRef to the content that was just created\n+ viewCtrl.viewElementRef(componentRef.location);\n \n- // get this ViewController's navbar TemplateRef, which may not\n- // exist if the ViewController's template didn't have an <ion-navbar *navbar>\n- let navbarTemplateRef = viewCtrl.getNavbarTemplateRef();\n+ // get the NavController's container for navbars, which is\n+ // the place this NavController will add each ViewController's navbar\n+ let navbarContainerRef = this.tabs.navbarContainerRef;\n \n- // create the navbar view if the pane has a navbar container, and the\n- // ViewController's instance has a navbar TemplateRef to go to inside of it\n- if (navbarContainerRef && navbarTemplateRef) {\n- let navbarView = navbarContainerRef.createEmbeddedView(navbarTemplateRef, -1);\n+ // get this ViewController's navbar TemplateRef, which may not\n+ // exist if the ViewController's template didn't have an <ion-navbar *navbar>\n+ let navbarTemplateRef = viewCtrl.getNavbarTemplateRef();\n \n- viewCtrl.disposals.push(() => {\n- let index = navbarContainerRef.indexOf(navbarView);\n- if (index > -1) {\n- navbarContainerRef.remove(index);\n- }\n- });\n- }\n+ // create the navbar view if the pane has a navbar container, and the\n+ // ViewController's instance has a navbar TemplateRef to go to inside of it\n+ if (navbarContainerRef && navbarTemplateRef) {\n+ let navbarView = navbarContainerRef.createEmbeddedView(navbarTemplateRef, -1);\n+\n+ viewCtrl.disposals.push(() => {\n+ let index = navbarContainerRef.indexOf(navbarView);\n+ if (index > -1) {\n+ navbarContainerRef.remove(index);\n+ }\n+ });\n+ }\n+\n+ done();\n+\n+ });\n \n- done();\n }\n \n }\n@@ -194,10 +198,7 @@ export class Tab extends NavController {\n \n @Directive({selector: 'template[content-anchor]'})\n class TabContentAnchor {\n- constructor(\n- @Host() tab: Tab,\n- viewContainerRef: ViewContainerRef\n- ) {\n- tab.contentContainerRef = viewContainerRef;\n+ constructor(@Host() tab: Tab, elementRef: ElementRef) {\n+ tab.contentAnchorRef = elementRef;\n }\n }\n", "diff --git a/server/Dockerfile b/server/Dockerfile\nindex 2f203bb..a84c31e 100755\n--- a/server/Dockerfile\n+++ b/server/Dockerfile\n@@ -9,9 +9,11 @@ ENV TZ utc\n WORKDIR /src\n \n COPY package.json /src\n+COPY package-lock.json /src\n+COPY tsconfig.json /src\n RUN npm install --production --no-optional\n \n COPY public /src/public\n COPY dist /src\n \n-CMD [ \"node\", \"./server/index.js\" ]\n+CMD [ \"node\", \"-r\", \"tsconfig-paths/register\", \"./server/index.js\" ]\ndiff --git a/server/package-lock.json b/server/package-lock.json\nindex 6cacfa2..236f1bb 100644\n--- a/server/package-lock.json\n+++ b/server/package-lock.json\n@@ -2164,8 +2164,7 @@\n \"@types/json5\": {\n \"version\": \"0.0.29\",\n \"resolved\": \"https://registry.npmjs.org/@types/json5/-/json5-0.0.29.tgz\",\n- \"integrity\": \"sha1-7ihweulOEdK4J7y+UnC86n8+ce4=\",\n- \"dev\": true\n+ \"integrity\": \"sha1-7ihweulOEdK4J7y+UnC86n8+ce4=\"\n },\n \"@types/jsonwebtoken\": {\n \"version\": \"8.3.5\",\n@@ -9246,8 +9245,7 @@\n \"strip-bom\": {\n \"version\": \"3.0.0\",\n \"resolved\": \"https://registry.npmjs.org/strip-bom/-/strip-bom-3.0.0.tgz\",\n- \"integrity\": \"sha1-IzTBjpx1n3vdVv3vfprj1YjmjtM=\",\n- \"dev\": true\n+ \"integrity\": \"sha1-IzTBjpx1n3vdVv3vfprj1YjmjtM=\"\n },\n \"strip-final-newline\": {\n \"version\": \"2.0.0\",\n@@ -9524,7 +9522,6 @@\n \"version\": \"3.9.0\",\n \"resolved\": \"https://registry.npmjs.org/tsconfig-paths/-/tsconfig-paths-3.9.0.tgz\",\n \"integrity\": \"sha512-dRcuzokWhajtZWkQsDVKbWyY+jgcLC5sqJhg2PSgf4ZkH2aHPvaOY8YWGhmjb68b5qqTfasSsDO9k7RUiEmZAw==\",\n- \"dev\": true,\n \"requires\": {\n \"@types/json5\": \"^0.0.29\",\n \"json5\": \"^1.0.1\",\n@@ -9536,7 +9533,6 @@\n \"version\": \"1.0.1\",\n \"resolved\": \"https://registry.npmjs.org/json5/-/json5-1.0.1.tgz\",\n \"integrity\": \"sha512-aKS4WQjPenRxiQsC93MNfjx+nbF4PAdYzmd/1JIj8HYzqfbu86beTuNgXDzPknWk0n0uARlyewZo4s++ES36Ow==\",\n- \"dev\": true,\n \"requires\": {\n \"minimist\": \"^1.2.0\"\n }\n@@ -9544,8 +9540,7 @@\n \"minimist\": {\n \"version\": \"1.2.5\",\n \"resolved\": \"https://registry.npmjs.org/minimist/-/minimist-1.2.5.tgz\",\n- \"integrity\": \"sha512-FM9nNUYrRBAELZQT3xeZQ7fmMOBg6nWNmJKTcgsJeaLstP/UODVpGsr5OhXhhXg6f+qtJ8uiZ+PUxkDWcgIXLw==\",\n- \"dev\": true\n+ \"integrity\": \"sha512-FM9nNUYrRBAELZQT3xeZQ7fmMOBg6nWNmJKTcgsJeaLstP/UODVpGsr5OhXhhXg6f+qtJ8uiZ+PUxkDWcgIXLw==\"\n }\n }\n },\ndiff --git a/server/package.json b/server/package.json\nindex 35426e9..896e9b3 100644\n--- a/server/package.json\n+++ b/server/package.json\n@@ -41,6 +41,7 @@\n \"pino-cloudwatch\": \"0.7.0\",\n \"pino-multi-stream\": \"4.2.0\",\n \"reflect-metadata\": \"0.1.13\",\n+ \"tsconfig-paths\": \"3.9.0\",\n \"typeorm\": \"0.2.37\"\n },\n \"devDependencies\": {\n@@ -69,7 +70,6 @@\n \"pino-pretty\": \"3.6.1\",\n \"ts-jest\": \"27.0.7\",\n \"ts-node-dev\": \"1.1.8\",\n- \"tsconfig-paths\": \"3.9.0\",\n \"typescript\": \"4.3.5\"\n },\n \"jest-junit\": {\n"]
2
["085ee958c48d695ba50822d8767d615fd9e887fa", "a827777f41e90b6332c191d05bae8db525de6f38"]
["refactor", "build"]
verify process can start at supported element types Verifies a PI can be started at specific element types. The test will deploy the process, start an instance at the desired start element and verify that it has been activated succesfully.,refactor to get ride of cloneDeep
["diff --git a/engine/src/test/java/io/camunda/zeebe/engine/processing/processinstance/CreateProcessInstanceSupportedElementTest.java b/engine/src/test/java/io/camunda/zeebe/engine/processing/processinstance/CreateProcessInstanceSupportedElementTest.java\nnew file mode 100644\nindex 0000000..a505307\n--- /dev/null\n+++ b/engine/src/test/java/io/camunda/zeebe/engine/processing/processinstance/CreateProcessInstanceSupportedElementTest.java\n@@ -0,0 +1,233 @@\n+/*\n+ * Copyright Camunda Services GmbH and/or licensed to Camunda Services GmbH under\n+ * one or more contributor license agreements. See the NOTICE file distributed\n+ * with this work for additional information regarding copyright ownership.\n+ * Licensed under the Zeebe Community License 1.1. You may not use this file\n+ * except in compliance with the Zeebe Community License 1.1.\n+ */\n+package io.camunda.zeebe.engine.processing.processinstance;\n+\n+import static org.assertj.core.api.Assertions.assertThat;\n+import static org.assertj.core.groups.Tuple.tuple;\n+\n+import io.camunda.zeebe.engine.util.EngineRule;\n+import io.camunda.zeebe.model.bpmn.Bpmn;\n+import io.camunda.zeebe.model.bpmn.BpmnModelInstance;\n+import io.camunda.zeebe.protocol.record.Record;\n+import io.camunda.zeebe.protocol.record.intent.ProcessInstanceIntent;\n+import io.camunda.zeebe.protocol.record.value.BpmnElementType;\n+import io.camunda.zeebe.test.util.record.RecordingExporter;\n+import io.camunda.zeebe.test.util.record.RecordingExporterTestWatcher;\n+import java.util.Collection;\n+import java.util.Collections;\n+import java.util.List;\n+import java.util.Map;\n+import org.junit.ClassRule;\n+import org.junit.Rule;\n+import org.junit.Test;\n+import org.junit.runner.RunWith;\n+import org.junit.runners.Parameterized;\n+import org.junit.runners.Parameterized.Parameters;\n+\n+@RunWith(Parameterized.class)\n+public class CreateProcessInstanceSupportedElementTest {\n+\n+ @ClassRule public static final EngineRule ENGINE = EngineRule.singlePartition();\n+ private static final String PROCESS_ID = \"processId\";\n+ private static final String CHILD_PROCESS_ID = \"childProcessId\";\n+ private static final String START_ELEMENT_ID = \"startElement\";\n+ private static final String MESSAGE = \"message\";\n+ private static final String JOBTYPE = \"jobtype\";\n+\n+ @Rule\n+ public final RecordingExporterTestWatcher recordingExporterTestWatcher =\n+ new RecordingExporterTestWatcher();\n+\n+ private final Scenario scenario;\n+\n+ public CreateProcessInstanceSupportedElementTest(final Scenario scenario) {\n+ this.scenario = scenario;\n+ }\n+\n+ @Parameters(name = \"{0}\")\n+ public static Collection<Object> scenarios() {\n+ return List.of(\n+ new Scenario(\n+ BpmnElementType.SUB_PROCESS,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .subProcess(START_ELEMENT_ID)\n+ .embeddedSubProcess()\n+ .startEvent()\n+ .subProcessDone()\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.EVENT_SUB_PROCESS,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .eventSubProcess(\n+ START_ELEMENT_ID, e -> e.startEvent().timerWithDuration(\"PT1H\").endEvent())\n+ .startEvent()\n+ .endEvent()\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.INTERMEDIATE_CATCH_EVENT,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .intermediateCatchEvent(START_ELEMENT_ID)\n+ .message(b -> b.name(MESSAGE).zeebeCorrelationKeyExpression(\"correlationKey\"))\n+ .done(),\n+ Map.of(\"correlationKey\", \"value\")),\n+ new Scenario(\n+ BpmnElementType.INTERMEDIATE_THROW_EVENT,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .intermediateThrowEvent(START_ELEMENT_ID)\n+ .endEvent()\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.END_EVENT,\n+ Bpmn.createExecutableProcess(PROCESS_ID).startEvent().endEvent(START_ELEMENT_ID).done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.SERVICE_TASK,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .serviceTask(START_ELEMENT_ID, b -> b.zeebeJobType(JOBTYPE))\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.RECEIVE_TASK,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .receiveTask(START_ELEMENT_ID)\n+ .message(b -> b.name(MESSAGE).zeebeCorrelationKeyExpression(\"correlationKey\"))\n+ .done(),\n+ Map.of(\"correlationKey\", \"value\")),\n+ new Scenario(\n+ BpmnElementType.USER_TASK,\n+ Bpmn.createExecutableProcess(PROCESS_ID).startEvent().userTask(START_ELEMENT_ID).done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.MANUAL_TASK,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .manualTask(START_ELEMENT_ID)\n+ .endEvent()\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.EXCLUSIVE_GATEWAY,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .exclusiveGateway(START_ELEMENT_ID)\n+ .defaultFlow()\n+ .endEvent()\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.PARALLEL_GATEWAY,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .parallelGateway(START_ELEMENT_ID)\n+ .endEvent()\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.EVENT_BASED_GATEWAY,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .eventBasedGateway(START_ELEMENT_ID)\n+ .intermediateCatchEvent()\n+ .message(b -> b.name(MESSAGE).zeebeCorrelationKeyExpression(\"correlationKey\"))\n+ .moveToLastGateway()\n+ .intermediateCatchEvent()\n+ .timerWithDuration(\"PT1H\")\n+ .done(),\n+ Map.of(\"correlationKey\", \"value\")),\n+ new Scenario(\n+ BpmnElementType.MULTI_INSTANCE_BODY,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .serviceTask(\n+ START_ELEMENT_ID,\n+ t ->\n+ t.zeebeJobType(JOBTYPE)\n+ .multiInstance(m -> m.parallel().zeebeInputCollectionExpression(\"[1]\")))\n+ .endEvent()\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.CALL_ACTIVITY,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .callActivity(START_ELEMENT_ID, c -> c.zeebeProcessId(CHILD_PROCESS_ID))\n+ .endEvent()\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.BUSINESS_RULE_TASK,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .businessRuleTask(START_ELEMENT_ID, b -> b.zeebeJobType(JOBTYPE))\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.SCRIPT_TASK,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .scriptTask(START_ELEMENT_ID, b -> b.zeebeJobType(JOBTYPE))\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.SEND_TASK,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .sendTask(START_ELEMENT_ID, b -> b.zeebeJobType(JOBTYPE))\n+ .done(),\n+ Collections.emptyMap()));\n+ }\n+\n+ @Test\n+ public void testProcessInstanceCanStartAtElementType() {\n+ // given\n+ ENGINE.deployment().withXmlResource(scenario.modelInstance).deploy();\n+ if (scenario.type == BpmnElementType.CALL_ACTIVITY) {\n+ ENGINE.deployment().withXmlResource(getChildProcess()).deploy();\n+ }\n+\n+ // when\n+ final long instanceKey =\n+ ENGINE\n+ .processInstance()\n+ .ofBpmnProcessId(PROCESS_ID)\n+ .withStartInstruction(START_ELEMENT_ID)\n+ .withVariables(scenario.variables)\n+ .create();\n+\n+ // then\n+ assertThat(\n+ RecordingExporter.processInstanceRecords()\n+ .withProcessInstanceKey(instanceKey)\n+ .onlyEvents()\n+ .limit(\n+ r ->\n+ r.getValue().getBpmnElementType() == scenario.type\n+ && r.getIntent() == ProcessInstanceIntent.ELEMENT_ACTIVATED))\n+ .extracting(record -> record.getValue().getBpmnElementType(), Record::getIntent)\n+ .containsSequence(\n+ tuple(BpmnElementType.PROCESS, ProcessInstanceIntent.ELEMENT_ACTIVATING),\n+ tuple(BpmnElementType.PROCESS, ProcessInstanceIntent.ELEMENT_ACTIVATED),\n+ tuple(scenario.type, ProcessInstanceIntent.ELEMENT_ACTIVATING),\n+ tuple(scenario.type, ProcessInstanceIntent.ELEMENT_ACTIVATED));\n+ }\n+\n+ private BpmnModelInstance getChildProcess() {\n+ return Bpmn.createExecutableProcess(CHILD_PROCESS_ID).startEvent().endEvent().done();\n+ }\n+\n+ record Scenario(\n+ BpmnElementType type, BpmnModelInstance modelInstance, Map<String, Object> variables) {}\n+}\n", "diff --git a/config/webpack.config.prod.js b/config/webpack.config.prod.js\nindex 3d2e5a6..e5219bd 100644\n--- a/config/webpack.config.prod.js\n+++ b/config/webpack.config.prod.js\n@@ -56,7 +56,7 @@ const extractTextPluginOptions = shouldUseRelativeAssetPaths\n const entries = fs.readdirSync(paths.appSrc)\n .filter(name => !name.startsWith('_'))\n .map(name => ({name, dirPath: path.join(paths.appSrc, name)}))\n- .filter(({name, dirPath}) => !/^assets|components|manifest|typings$/.test(name) && fs.lstatSync(dirPath).isDirectory())\n+ .filter(({name, dirPath}) => !/^assets|components|manifest|typings|app-config$/.test(name) && fs.lstatSync(dirPath).isDirectory())\n \n // This is the production configuration.\n // It compiles slowly and is focused on producing a fast and minimal bundle.\ndiff --git a/src/app-config/context-menus.ts b/src/app-config/context-menus.ts\nnew file mode 100644\nindex 0000000..a733b01\n--- /dev/null\n+++ b/src/app-config/context-menus.ts\n@@ -0,0 +1,27 @@\n+export function getAllContextMenus () {\n+ const allContextMenus = {\n+ google_page_translate: 'x',\n+ youdao_page_translate: 'x',\n+ google_search: 'https://www.google.com/#newwindow=1&q=%s',\n+ baidu_search: 'https://www.baidu.com/s?ie=utf-8&wd=%s',\n+ bing_search: 'https://www.bing.com/search?q=%s',\n+ google_translate: 'https://translate.google.cn/#auto/zh-CN/%s',\n+ etymonline: 'http://www.etymonline.com/index.php?search=%s',\n+ merriam_webster: 'http://www.merriam-webster.com/dictionary/%s',\n+ oxford: 'http://www.oxforddictionaries.com/us/definition/english/%s',\n+ cambridge: 'http://dictionary.cambridge.org/spellcheck/english-chinese-simplified/?q=%s',\n+ youdao: 'http://dict.youdao.com/w/%s',\n+ dictcn: 'https://dict.eudic.net/dicts/en/%s',\n+ iciba: 'http://www.iciba.com/%s',\n+ liangan: 'https://www.moedict.tw/~%s',\n+ guoyu: 'https://www.moedict.tw/%s',\n+ longman_business: 'http://www.ldoceonline.com/search/?q=%s',\n+ bing_dict: 'https://cn.bing.com/dict/?q=%s'\n+ }\n+\n+ // Just for type check. Keys in allContextMenus are useful so no actual assertion\n+ // tslint:disable-next-line:no-unused-expression\n+ allContextMenus as { [id: string]: string }\n+\n+ return allContextMenus\n+}\ndiff --git a/src/app-config/dicts.ts b/src/app-config/dicts.ts\nnew file mode 100644\nindex 0000000..905d2de\n--- /dev/null\n+++ b/src/app-config/dicts.ts\n@@ -0,0 +1,398 @@\n+import { DeepReadonly } from '@/typings/helpers'\n+\n+export function getALlDicts () {\n+ const allDicts = {\n+ bing: {\n+ /**\n+ * Full content page to jump to when user clicks the title.\n+ * %s will be replaced with the current word.\n+ * %z will be replaced with the traditional Chinese version of the current word.\n+ */\n+ page: 'https://cn.bing.com/dict/search?q=%s',\n+ /**\n+ * If set to true, the dict start searching automatically.\n+ * Otherwise it'll only start seaching when user clicks the unfold button.\n+ * Default MUST be true and let user decide.\n+ */\n+ defaultUnfold: true,\n+ /**\n+ * This is the default height when the dict first renders the result.\n+ * If the content height is greater than the preferred height,\n+ * the preferred height is used and a mask with a view-more button is shown.\n+ * Otherwise the content height is used.\n+ */\n+ preferredHeight: 240,\n+ /**\n+ * Only start searching if the selection contains the language.\n+ * Better set default to true and let user decide.\n+ */\n+ selectionLang: {\n+ eng: true,\n+ chs: true\n+ },\n+ /** Optional dict custom options. Can only be boolean or number. */\n+ options: {\n+ tense: true,\n+ phsym: true,\n+ cdef: true,\n+ related: true,\n+ sentence: 4\n+ }\n+ },\n+ business: {\n+ /**\n+ * Full content page to jump to when user clicks the title.\n+ * %s will be replaced with the current word.\n+ * %z will be replaced with the traditional Chinese version of the current word\n+ */\n+ page: 'http://www.ldoceonline.com/search/?q=%s',\n+ /**\n+ * If set to true, the dict start searching automatically.\n+ * Otherwise it'll only start seaching when user clicks the unfold button.\n+ * Default MUST be true and let user decide.\n+ */\n+ defaultUnfold: true,\n+ /**\n+ * This is the default height when the dict first renders the result.\n+ * If the content height is greater than the preferred height,\n+ * the preferred height is used and a mask with a view-more button is shown.\n+ * Otherwise the content height is used.\n+ */\n+ preferredHeight: 265,\n+ /**\n+ * Only start searching if the selection contains the language.\n+ * Better set default to true and let user decide.\n+ */\n+ selectionLang: {\n+ eng: true,\n+ chs: true\n+ }\n+ },\n+ cobuild: {\n+ /**\n+ * Full content page to jump to when user clicks the title.\n+ * %s will be replaced with the current word.\n+ * %z will be replaced with the traditional Chinese version of the current word\n+ */\n+ page: 'https://www.collinsdictionary.com/dictionary/%s',\n+ /**\n+ * If set to true, the dict start searching automatically.\n+ * Otherwise it'll only start seaching when user clicks the unfold button.\n+ * Default MUST be true and let user decide.\n+ */\n+ defaultUnfold: true,\n+ /**\n+ * This is the default height when the dict first renders the result.\n+ * If the content height is greater than the preferred height,\n+ * the preferred height is used and a mask with a view-more button is shown.\n+ * Otherwise the content height is used.\n+ */\n+ preferredHeight: 300,\n+ /**\n+ * Only start searching if the selection contains the language.\n+ * Better set default to true and let user decide.\n+ */\n+ selectionLang: {\n+ eng: true,\n+ chs: true\n+ },\n+ /** Optional dict custom options. Can only be boolean or number. */\n+ options: {\n+ sentence: 4\n+ }\n+ },\n+ dictcn: {\n+ /**\n+ * Full content page to jump to when user clicks the title.\n+ * %s will be replaced with the current word.\n+ * %z will be replaced with the traditional Chinese version of the current word\n+ */\n+ page: 'http://dict.cn/%s',\n+ /**\n+ * If set to true, the dict start searching automatically.\n+ * Otherwise it'll only start seaching when user clicks the unfold button.\n+ * Default MUST be true and let user decide.\n+ */\n+ defaultUnfold: true,\n+ /**\n+ * This is the default height when the dict first renders the result.\n+ * If the content height is greater than the preferred height,\n+ * the preferred height is used and a mask with a view-more button is shown.\n+ * Otherwise the content height is used.\n+ */\n+ preferredHeight: 300,\n+ /**\n+ * Only start searching if the selection contains the language.\n+ * Better set default to true and let user decide.\n+ */\n+ selectionLang: {\n+ eng: true,\n+ chs: true\n+ },\n+ /** Optional dict custom options. Can only be boolean or number. */\n+ options: {\n+ chart: true,\n+ etym: true\n+ }\n+ },\n+ etymonline: {\n+ /**\n+ * Full content page to jump to when user clicks the title.\n+ * %s will be replaced with the current word.\n+ * %z will be replaced with the traditional Chinese version of the current word\n+ */\n+ page: 'http://www.etymonline.com/search?q=%s',\n+ /**\n+ * If set to true, the dict start searching automatically.\n+ * Otherwise it'll only start seaching when user clicks the unfold button.\n+ * Default MUST be true and let user decide.\n+ */\n+ defaultUnfold: true,\n+ /**\n+ * This is the default height when the dict first renders the result.\n+ * If the content height is greater than the preferred height,\n+ * the preferred height is used and a mask with a view-more button is shown.\n+ * Otherwise the content height is used.\n+ */\n+ preferredHeight: 265,\n+ /**\n+ * Only start searching if the selection contains the language.\n+ * Better set default to true and let user decide.\n+ */\n+ selectionLang: {\n+ eng: true,\n+ chs: true\n+ },\n+ /** Optional dict custom options. Can only be boolean or number. */\n+ options: {\n+ resultnum: 2\n+ }\n+ },\n+ google: {\n+ /**\n+ * Full content page to jump to when user clicks the title.\n+ * %s will be replaced with the current word.\n+ * %z will be replaced with the traditional Chinese version of the current word\n+ */\n+ page: 'https://translate.google.com/#auto/zh-CN/%s',\n+ /**\n+ * If set to true, the dict start searching automatically.\n+ * Otherwise it'll only start seaching when user clicks the unfold button.\n+ * Default MUST be true and let user decide.\n+ */\n+ defaultUnfold: true,\n+ /**\n+ * This is the default height when the dict first renders the result.\n+ * If the content height is greater than the preferred height,\n+ * the preferred height is used and a mask with a view-more button is shown.\n+ * Otherwise the content height is used.\n+ */\n+ preferredHeight: 110,\n+ /**\n+ * Only start searching if the selection contains the language.\n+ * Better set default to true and let user decide.\n+ */\n+ selectionLang: {\n+ eng: true,\n+ chs: true\n+ }\n+ },\n+ guoyu: {\n+ /**\n+ * Full content page to jump to when user clicks the title.\n+ * %s will be replaced with the current word.\n+ * %z will be replaced with the traditional Chinese version of the current word\n+ */\n+ page: 'https://www.moedict.tw/%z',\n+ /**\n+ * If set to true, the dict start searching automatically.\n+ * Otherwise it'll only start seaching when user clicks the unfold button.\n+ * Default MUST be true and let user decide.\n+ */\n+ defaultUnfold: true,\n+ /**\n+ * This is the default height when the dict first renders the result.\n+ * If the content height is greater than the preferred height,\n+ * the preferred height is used and a mask with a view-more button is shown.\n+ * Otherwise the content height is used.\n+ */\n+ preferredHeight: 265,\n+ /**\n+ * Only start searching if the selection contains the language.\n+ * Better set default to true and let user decide.\n+ */\n+ selectionLang: {\n+ eng: true,\n+ chs: true\n+ }\n+ },\n+ liangan: {\n+ /**\n+ * Full content page to jump to when user clicks the title.\n+ * %s will be replaced with the current word.\n+ * %z will be replaced with the traditional Chinese version of the current word\n+ */\n+ page: 'https://www.moedict.tw/~%z',\n+ /**\n+ * If set to true, the dict start searching automatically.\n+ * Otherwise it'll only start seaching when user clicks the unfold button.\n+ * Default MUST be true and let user decide.\n+ */\n+ defaultUnfold: true,\n+ /**\n+ * This is the default height when the dict first renders the result.\n+ * If the content height is greater than the preferred height,\n+ * the preferred height is used and a mask with a view-more button is shown.\n+ * Otherwise the content height is used.\n+ */\n+ preferredHeight: 265,\n+ /**\n+ * Only start searching if the selection contains the language.\n+ * Better set default to true and let user decide.\n+ */\n+ selectionLang: {\n+ eng: true,\n+ chs: true\n+ }\n+ },\n+ macmillan: {\n+ /**\n+ * Full content page to jump to when user clicks the title.\n+ * %s will be replaced with the current word.\n+ * %z will be replaced with the traditional Chinese version of the current word\n+ */\n+ page: 'http://www.macmillandictionary.com/dictionary/british/%s',\n+ /**\n+ * If set to true, the dict start searching automatically.\n+ * Otherwise it'll only start seaching when user clicks the unfold button.\n+ * Default MUST be true and let user decide.\n+ */\n+ defaultUnfold: true,\n+ /**\n+ * This is the default height when the dict first renders the result.\n+ * If the content height is greater than the preferred height,\n+ * the preferred height is used and a mask with a view-more button is shown.\n+ * Otherwise the content height is used.\n+ */\n+ preferredHeight: 265,\n+ /**\n+ * Only start searching if the selection contains the language.\n+ * Better set default to true and let user decide.\n+ */\n+ selectionLang: {\n+ eng: true,\n+ chs: true\n+ }\n+ },\n+ urban: {\n+ /**\n+ * Full content page to jump to when user clicks the title.\n+ * %s will be replaced with the current word.\n+ * %z will be replaced with the traditional Chinese version of the current word\n+ */\n+ page: 'http://www.urbandictionary.com/define.php?term=%s',\n+ /**\n+ * If set to true, the dict start searching automatically.\n+ * Otherwise it'll only start seaching when user clicks the unfold button.\n+ * Default MUST be true and let user decide.\n+ */\n+ defaultUnfold: true,\n+ /**\n+ * This is the default height when the dict first renders the result.\n+ * If the content height is greater than the preferred height,\n+ * the preferred height is used and a mask with a view-more button is shown.\n+ * Otherwise the content height is used.\n+ */\n+ preferredHeight: 180,\n+ /**\n+ * Only start searching if the selection contains the language.\n+ * Better set default to true and let user decide.\n+ */\n+ selectionLang: {\n+ eng: true,\n+ chs: true\n+ },\n+ /** Optional dict custom options. Can only be boolean or number. */\n+ options: {\n+ resultnum: 4\n+ }\n+ },\n+ vocabulary: {\n+ /**\n+ * Full content page to jump to when user clicks the title.\n+ * %s will be replaced with the current word.\n+ * %z will be replaced with the traditional Chinese version of the current word\n+ */\n+ page: 'https://www.vocabulary.com/dictionary/%s',\n+ /**\n+ * If set to true, the dict start searching automatically.\n+ * Otherwise it'll only start seaching when user clicks the unfold button.\n+ * Default MUST be true and let user decide.\n+ */\n+ defaultUnfold: true,\n+ /**\n+ * This is the default height when the dict first renders the result.\n+ * If the content height is greater than the preferred height,\n+ * the preferred height is used and a mask with a view-more button is shown.\n+ * Otherwise the content height is used.\n+ */\n+ preferredHeight: 180,\n+ /**\n+ * Only start searching if the selection contains the language.\n+ * Better set default to true and let user decide.\n+ */\n+ selectionLang: {\n+ eng: true,\n+ chs: true\n+ }\n+ },\n+ zdic: {\n+ /**\n+ * Full content page to jump to when user clicks the title.\n+ * %s will be replaced with the current word.\n+ * %z will be replaced with the traditional Chinese version of the current word\n+ */\n+ page: 'http://www.zdic.net/search/?c=1&q=%s',\n+ /**\n+ * If set to true, the dict start searching automatically.\n+ * Otherwise it'll only start seaching when user clicks the unfold button.\n+ * Default MUST be true and let user decide.\n+ */\n+ defaultUnfold: true,\n+ /**\n+ * This is the default height when the dict first renders the result.\n+ * If the content height is greater than the preferred height,\n+ * the preferred height is used and a mask with a view-more button is shown.\n+ * Otherwise the content height is used.\n+ */\n+ preferredHeight: 400,\n+ /**\n+ * Only start searching if the selection contains the language.\n+ * Better set default to true and let user decide.\n+ */\n+ selectionLang: {\n+ eng: true,\n+ chs: true\n+ }\n+ },\n+ }\n+\n+ // Just for type check. Keys in allDicts are useful so no actual assertion\n+ // tslint:disable-next-line:no-unused-expression\n+ allDicts as {\n+ [id: string]: {\n+ page: string\n+ defaultUnfold: boolean\n+ preferredHeight: number\n+ selectionLang: {\n+ eng: boolean\n+ chs: boolean\n+ }\n+ options?: {\n+ [option: string]: number | boolean\n+ }\n+ }\n+ }\n+\n+ return allDicts\n+}\ndiff --git a/src/app-config/index.ts b/src/app-config/index.ts\nindex 350cd8f..879a312 100644\n--- a/src/app-config/index.ts\n+++ b/src/app-config/index.ts\n@@ -1,5 +1,6 @@\n-import cloneDeep from 'lodash/cloneDeep'\n-import { DeepReadonly } from './typings/helpers'\n+import { DeepReadonly } from '@/typings/helpers'\n+import { getALlDicts } from './dicts'\n+import { getAllContextMenus } from './context-menus'\n \n const langUI = (browser.i18n.getUILanguage() || 'en').replace('-', '_')\n const langCode = /^zh_CN|zh_TW|en$/.test(langUI)\n@@ -8,220 +9,11 @@ const langCode = /^zh_CN|zh_TW|en$/.test(langUI)\n : langUI\n : 'en'\n \n-const allDicts = {\n- bing: {\n- page: 'https://cn.bing.com/dict/search?q=%s',\n- defaultUnfold: true,\n- preferredHeight: 240,\n- selectionLang: {\n- eng: true,\n- chs: true\n- },\n- options: {\n- tense: true,\n- phsym: true,\n- cdef: true,\n- related: true,\n- sentence: 4\n- }\n- },\n- business: {\n- page: 'http://www.ldoceonline.com/search/?q=%s',\n- defaultUnfold: true,\n- preferredHeight: 265,\n- selectionLang: {\n- eng: true,\n- chs: true\n- }\n- },\n- cobuild: {\n- page: 'https://www.collinsdictionary.com/dictionary/%s',\n- defaultUnfold: true,\n- preferredHeight: 300,\n- selectionLang: {\n- eng: true,\n- chs: true\n- },\n- options: {\n- sentence: 4\n- }\n- },\n- dictcn: {\n- page: 'http://dict.cn/%s',\n- defaultUnfold: true,\n- preferredHeight: 300,\n- selectionLang: {\n- eng: true,\n- chs: true\n- },\n- options: {\n- chart: true,\n- etym: true\n- }\n- },\n- etymonline: {\n- page: 'http://www.etymonline.com/search?q=%s',\n- defaultUnfold: true,\n- preferredHeight: 265,\n- selectionLang: {\n- eng: true,\n- chs: true\n- },\n- options: {\n- resultnum: 2\n- }\n- },\n- eudic: {\n- page: 'https://dict.eudic.net/dicts/en/%s',\n- defaultUnfold: true,\n- preferredHeight: 265,\n- selectionLang: {\n- eng: true,\n- chs: true\n- }\n- },\n- google: {\n- page: 'https://translate.google.com/#auto/zh-CN/%s',\n- defaultUnfold: true,\n- preferredHeight: 110,\n- selectionLang: {\n- eng: true,\n- chs: true\n- }\n- },\n- guoyu: {\n- page: 'https://www.moedict.tw/%z',\n- defaultUnfold: true,\n- preferredHeight: 265,\n- selectionLang: {\n- eng: true,\n- chs: true\n- }\n- },\n- howjsay: {\n- page: 'http://www.howjsay.com/index.php?word=%s',\n- defaultUnfold: true,\n- preferredHeight: 265,\n- selectionLang: {\n- eng: true,\n- chs: true\n- },\n- options: {\n- related: true\n- }\n- },\n- liangan: {\n- page: 'https://www.moedict.tw/~%z',\n- defaultUnfold: true,\n- preferredHeight: 265,\n- selectionLang: {\n- eng: true,\n- chs: true\n- }\n- },\n- macmillan: {\n- page: 'http://www.macmillandictionary.com/dictionary/british/%s',\n- defaultUnfold: true,\n- preferredHeight: 265,\n- selectionLang: {\n- eng: true,\n- chs: true\n- }\n- },\n- urban: {\n- page: 'http://www.urbandictionary.com/define.php?term=%s',\n- defaultUnfold: true,\n- preferredHeight: 180,\n- selectionLang: {\n- eng: true,\n- chs: true\n- },\n- options: {\n- resultnum: 4\n- }\n- },\n- vocabulary: {\n- page: 'https://www.vocabulary.com/dictionary/%s',\n- defaultUnfold: true,\n- preferredHeight: 180,\n- selectionLang: {\n- eng: true,\n- chs: true\n- }\n- },\n- wordreference: {\n- page: 'http://www.wordreference.com/definition/%s',\n- defaultUnfold: true,\n- preferredHeight: 180,\n- selectionLang: {\n- eng: true,\n- chs: true\n- },\n- options: {\n- etym: true,\n- idiom: true\n- }\n- },\n- zdic: {\n- page: 'http://www.zdic.net/search/?c=1&q=%s',\n- defaultUnfold: true,\n- preferredHeight: 400,\n- selectionLang: {\n- eng: true,\n- chs: true\n- }\n- },\n-}\n-\n-// Just for type check. Keys in allDicts are useful so no actual assertion\n-// tslint:disable-next-line:no-unused-expression\n-allDicts as {\n- [id: string]: {\n- /** url for the complete result */\n- page: string\n- /** lazy load */\n- defaultUnfold: boolean\n- /** content below the preferrred height will be hidden by default */\n- preferredHeight: number\n- /** only search when the selection contains the language */\n- selectionLang: {\n- eng: boolean\n- chs: boolean\n- }\n- /** other options */\n- options?: {\n- [option: string]: number | boolean\n- }\n- }\n-}\n-\n-export type DictID = keyof typeof allDicts\n-\n-const allContextMenus = {\n- google_page_translate: 'x',\n- youdao_page_translate: 'x',\n- google_search: 'https://www.google.com/#newwindow=1&q=%s',\n- baidu_search: 'https://www.baidu.com/s?ie=utf-8&wd=%s',\n- bing_search: 'https://www.bing.com/search?q=%s',\n- google_translate: 'https://translate.google.cn/#auto/zh-CN/%s',\n- etymonline: 'http://www.etymonline.com/index.php?search=%s',\n- merriam_webster: 'http://www.merriam-webster.com/dictionary/%s',\n- oxford: 'http://www.oxforddictionaries.com/us/definition/english/%s',\n- cambridge: 'http://dictionary.cambridge.org/spellcheck/english-chinese-simplified/?q=%s',\n- youdao: 'http://dict.youdao.com/w/%s',\n- dictcn: 'https://dict.eudic.net/dicts/en/%s',\n- iciba: 'http://www.iciba.com/%s',\n- liangan: 'https://www.moedict.tw/~%s',\n- guoyu: 'https://www.moedict.tw/%s',\n- longman_business: 'http://www.ldoceonline.com/search/?q=%s',\n- bing_dict: 'https://cn.bing.com/dict/?q=%s'\n-}\n-\n-// Just for type check. Keys in allContextMenus are useful so no actual assertion\n-// tslint:disable-next-line:no-unused-expression\n-allContextMenus as { [id: string]: string }\n+export type DictConfigsMutable = ReturnType<typeof getALlDicts>\n+export type DictConfigs = DeepReadonly<DictConfigsMutable>\n+export type DictID = keyof DictConfigsMutable\n \n-export type ContextMenuDictID = keyof typeof allContextMenus\n+export type ContextMenuDictID = keyof ReturnType<typeof getAllContextMenus>\n \n export const enum TCDirection {\n center,\n@@ -238,10 +30,6 @@ export const enum TCDirection {\n /** '' means no preload */\n export type PreloadSource = '' | 'clipboard' | 'selection'\n \n-export type DictConfigs = DeepReadonly<DictConfigsMutable>\n-\n-export type DictConfigsMutable = typeof allDicts\n-\n export type AppConfig = DeepReadonly<AppConfigMutable>\n \n export interface AppConfigMutable {\n@@ -418,7 +206,7 @@ export function appConfigFactory (): AppConfig {\n },\n en: {\n dict: '',\n- list: ['bing', 'dictcn', 'howjsay', 'macmillan', 'eudic', 'urban'],\n+ list: ['bing', 'dictcn', 'macmillan', 'urban'],\n accent: 'uk' as ('us' | 'uk')\n }\n },\n@@ -426,11 +214,11 @@ export function appConfigFactory (): AppConfig {\n dicts: {\n selected: ['bing', 'urban', 'vocabulary', 'dictcn'],\n // settings of each dict will be auto-generated\n- all: cloneDeep(allDicts)\n+ all: getALlDicts()\n },\n contextMenus: {\n selected: ['oxford', 'google_translate', 'merriam_webster', 'cambridge', 'google_search', 'google_page_translate', 'youdao_page_translate'],\n- all: cloneDeep(allContextMenus)\n+ all: getAllContextMenus()\n }\n }\n }\n"]
2
["a5ecfdf49b0d4c43fbbbf7947be7c0327ccb3415", "d986b530775edd8ef1f4e445a5d4b0016f409722"]
["test", "refactor"]
Fix readme Signed-off-by: Ben Johnson <[email protected]>,fix `get-deploy-tags.sh`
["diff --git a/README.md b/README.md\nindex 587d655..da746bb 100644\n--- a/README.md\n+++ b/README.md\n@@ -38,15 +38,20 @@ simple and unified.\n * [**Installation**][docs.installation] - [containers][docs.containers], [operating systems][docs.operating_systems], [package managers][docs.package_managers], [from archives][docs.from-archives], [from source][docs.from-source]\n * [**Configuration**][docs.configuration]\n * [**Deployment**][docs.deployment] - [topologies][docs.topologies], [roles][docs.roles]\n+* [**Guides**][docs.guides] - [getting started][docs.guides.getting_started]\n \n-#### [Components](https://vector.dev/components)\n+#### Reference\n \n-* [**Sources**][docs.sources] - \n-* [**Transforms**][docs.transforms]\n-* [**Sinks**][docs.sinks]\n+* [**Sources**][docs.sources] - [docker][docs.sources.docker], [file][docs.sources.file], [journald][docs.sources.journald], [kafka][docs.sources.kafka]\n+* [**Transforms**][docs.transforms] - [json_parser][docs.transforms.json_parser], [log_to_metric][docs.transforms.log_to_metric], [lua][docs.transforms.lua], [regex_parser][docs.transforms.regex_parser]\n+* [**Sinks**][docs.sinks] - [aws_cloudwatch_logs][docs.sinks.aws_cloudwatch_logs], [aws_cloudwatch_metrics][docs.sinks.aws_cloudwatch_metrics], [aws_kinesis_streams][docs.sinks.aws_kinesis_streams], [aws_s3][docs.sinks.aws_s3], [clickhouse][docs.sinks.clickhouse], [elasticsearch][docs.sinks.elasticsearch], and [15 more][docs.sinks]\n \n-* [**Administration**][docs.administration] - [process management][docs.process-management], [monitoring][docs.monitoring], [updating][docs.updating], [validating][docs.validating]\n-* [**Guides**][docs.guides]\n+#### Administration\n+\n+* [**Process management**][docs.process-management]\n+* [**Monitoring**][docs.monitoring]\n+* [**Updating**][docs.updating]\n+* [**Validating**][docs.validating]\n \n #### Resources\n \n@@ -105,88 +110,6 @@ curl --proto '=https' --tlsv1.2 -sSf https://sh.vector.dev | sh\n \n Or view [platform specific installation instructions][docs.installation].\n \n-\n-## Sources\n-\n-| Name | Description |\n-|:------|:------------|\n-| [**`docker`**][docs.sources.docker] | Ingests data through the docker engine daemon and outputs [`log`][docs.data-model#log] events. |\n-| [**`file`**][docs.sources.file] | Ingests data through one or more local files and outputs [`log`][docs.data-model#log] events. |\n-| [**`journald`**][docs.sources.journald] | Ingests data through log records from journald and outputs [`log`][docs.data-model#log] events. |\n-| [**`kafka`**][docs.sources.kafka] | Ingests data through Kafka 0.9 or later and outputs [`log`][docs.data-model#log] events. |\n-| [**`statsd`**][docs.sources.statsd] | Ingests data through the StatsD UDP protocol and outputs [`metric`][docs.data-model#metric] events. |\n-| [**`stdin`**][docs.sources.stdin] | Ingests data through standard input (STDIN) and outputs [`log`][docs.data-model#log] events. |\n-| [**`syslog`**][docs.sources.syslog] | Ingests data through the Syslog 5424 protocol and outputs [`log`][docs.data-model#log] events. |\n-| [**`tcp`**][docs.sources.tcp] | Ingests data through the TCP protocol and outputs [`log`][docs.data-model#log] events. |\n-| [**`udp`**][docs.sources.udp] | Ingests data through the UDP protocol and outputs [`log`][docs.data-model#log] events. |\n-| [**`vector`**][docs.sources.vector] | Ingests data through another upstream [`vector` sink][docs.sinks.vector] and outputs [`log`][docs.data-model#log] and [`metric`][docs.data-model#metric] events. |\n-\n-[+ request a new source][urls.new_source]\n-\n-\n-## Transforms\n-\n-| Name | Description |\n-|:------|:------------|\n-| [**`add_fields`**][docs.transforms.add_fields] | Accepts [`log`][docs.data-model#log] events and allows you to add one or more log fields. |\n-| [**`add_tags`**][docs.transforms.add_tags] | Accepts [`metric`][docs.data-model#metric] events and allows you to add one or more metric tags. |\n-| [**`coercer`**][docs.transforms.coercer] | Accepts [`log`][docs.data-model#log] events and allows you to coerce log fields into fixed types. |\n-| [**`field_filter`**][docs.transforms.field_filter] | Accepts [`log`][docs.data-model#log] and [`metric`][docs.data-model#metric] events and allows you to filter events by a log field's value. |\n-| [**`grok_parser`**][docs.transforms.grok_parser] | Accepts [`log`][docs.data-model#log] events and allows you to parse a log field value with [Grok][urls.grok]. |\n-| [**`json_parser`**][docs.transforms.json_parser] | Accepts [`log`][docs.data-model#log] events and allows you to parse a log field value as JSON. |\n-| [**`log_to_metric`**][docs.transforms.log_to_metric] | Accepts [`log`][docs.data-model#log] events and allows you to convert logs into one or more metrics. |\n-| [**`lua`**][docs.transforms.lua] | Accepts [`log`][docs.data-model#log] events and allows you to transform events with a full embedded [Lua][urls.lua] engine. |\n-| [**`regex_parser`**][docs.transforms.regex_parser] | Accepts [`log`][docs.data-model#log] events and allows you to parse a log field's value with a [Regular Expression][urls.regex]. |\n-| [**`remove_fields`**][docs.transforms.remove_fields] | Accepts [`log`][docs.data-model#log] events and allows you to remove one or more log fields. |\n-| [**`remove_tags`**][docs.transforms.remove_tags] | Accepts [`metric`][docs.data-model#metric] events and allows you to remove one or more metric tags. |\n-| [**`sampler`**][docs.transforms.sampler] | Accepts [`log`][docs.data-model#log] events and allows you to sample events with a configurable rate. |\n-| [**`split`**][docs.transforms.split] | Accepts [`log`][docs.data-model#log] events and allows you to split a field's value on a given separator and zip the tokens into ordered field names. |\n-| [**`tokenizer`**][docs.transforms.tokenizer] | Accepts [`log`][docs.data-model#log] events and allows you to tokenize a field's value by splitting on white space, ignoring special wrapping characters, and zip the tokens into ordered field names. |\n-\n-[+ request a new transform][urls.new_transform]\n-\n-\n-## Sinks\n-\n-| Name | Description |\n-|:------|:------------|\n-| [**`aws_cloudwatch_logs`**][docs.sinks.aws_cloudwatch_logs] | [Batches](#buffers-and-batches) [`log`][docs.data-model#log] events to [AWS CloudWatch Logs][urls.aws_cw_logs] via the [`PutLogEvents` API endpoint](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutLogEvents.html). |\n-| [**`aws_cloudwatch_metrics`**][docs.sinks.aws_cloudwatch_metrics] | [Streams](#streaming) [`metric`][docs.data-model#metric] events to [AWS CloudWatch Metrics][urls.aws_cw_metrics] via the [`PutMetricData` API endpoint](https://docs.aws.amazon.com/AmazonCloudWatch/latest/APIReference/API_PutMetricData.html). |\n-| [**`aws_kinesis_streams`**][docs.sinks.aws_kinesis_streams] | [Batches](#buffers-and-batches) [`log`][docs.data-model#log] events to [AWS Kinesis Data Stream][urls.aws_kinesis_data_streams] via the [`PutRecords` API endpoint](https://docs.aws.amazon.com/kinesis/latest/APIReference/API_PutRecords.html). |\n-| [**`aws_s3`**][docs.sinks.aws_s3] | [Batches](#buffers-and-batches) [`log`][docs.data-model#log] events to [AWS S3][urls.aws_s3] via the [`PutObject` API endpoint](https://docs.aws.amazon.com/AmazonS3/latest/API/RESTObjectPUT.html). |\n-| [**`blackhole`**][docs.sinks.blackhole] | [Streams](#streaming) [`log`][docs.data-model#log] and [`metric`][docs.data-model#metric] events to a blackhole that simply discards data, designed for testing and benchmarking purposes. |\n-| [**`clickhouse`**][docs.sinks.clickhouse] | [Batches](#buffers-and-batches) [`log`][docs.data-model#log] events to [Clickhouse][urls.clickhouse] via the [`HTTP` Interface][urls.clickhouse_http]. |\n-| [**`console`**][docs.sinks.console] | [Streams](#streaming) [`log`][docs.data-model#log] and [`metric`][docs.data-model#metric] events to [standard output streams][urls.standard_streams], such as `STDOUT` and `STDERR`. |\n-| [**`datadog_metrics`**][docs.sinks.datadog_metrics] | [Batches](#buffers-and-batches) [`metric`][docs.data-model#metric] events to [Datadog][urls.datadog] metrics service using [HTTP API](https://docs.datadoghq.com/api/?lang=bash#metrics). |\n-| [**`elasticsearch`**][docs.sinks.elasticsearch] | [Batches](#buffers-and-batches) [`log`][docs.data-model#log] events to [Elasticsearch][urls.elasticsearch] via the [`_bulk` API endpoint](https://www.elastic.co/guide/en/elasticsearch/reference/current/docs-bulk.html). |\n-| [**`file`**][docs.sinks.file] | [Streams](#streaming) [`log`][docs.data-model#log] events to a file. |\n-| [**`http`**][docs.sinks.http] | [Batches](#buffers-and-batches) [`log`][docs.data-model#log] events to a generic HTTP endpoint. |\n-| [**`kafka`**][docs.sinks.kafka] | [Streams](#streaming) [`log`][docs.data-model#log] events to [Apache Kafka][urls.kafka] via the [Kafka protocol][urls.kafka_protocol]. |\n-| [**`prometheus`**][docs.sinks.prometheus] | [Exposes](#exposing-and-scraping) [`metric`][docs.data-model#metric] events to [Prometheus][urls.prometheus] metrics service. |\n-| [**`splunk_hec`**][docs.sinks.splunk_hec] | [Batches](#buffers-and-batches) [`log`][docs.data-model#log] events to a [Splunk HTTP Event Collector][urls.splunk_hec]. |\n-| [**`statsd`**][docs.sinks.statsd] | [Streams](#streaming) [`metric`][docs.data-model#metric] events to [StatsD][urls.statsd] metrics service. |\n-| [**`tcp`**][docs.sinks.tcp] | [Streams](#streaming) [`log`][docs.data-model#log] events to a TCP connection. |\n-| [**`vector`**][docs.sinks.vector] | [Streams](#streaming) [`log`][docs.data-model#log] events to another downstream [`vector` source][docs.sources.vector]. |\n-\n-[+ request a new sink][urls.new_sink]\n-\n-\n-## License\n-\n-Copyright 2019, Vector Authors. All rights reserved.\n-\n-Licensed under the Apache License, Version 2.0 (the \"License\"); you may not\n-use these files except in compliance with the License. You may obtain a copy\n-of the License at\n-\n-http://www.apache.org/licenses/LICENSE-2.0\n-\n-Unless required by applicable law or agreed to in writing, software\n-distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n-WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n-License for the specific language governing permissions and limitations under\n-the License.\n-\n ---\n \n <p align=\"center\">\n@@ -200,8 +123,6 @@ the License.\n [docs.configuration]: https://vector.dev/docs/setup/configuration\n [docs.containers]: https://vector.dev/docs/setup/installation/containers\n [docs.correctness]: https://vector.dev/docs/about/correctness\n-[docs.data-model#log]: https://vector.dev/docs/about/data-model#log\n-[docs.data-model#metric]: https://vector.dev/docs/about/data-model#metric\n [docs.data-model.log]: https://vector.dev/docs/about/data-model/log\n [docs.data-model.metric]: https://vector.dev/docs/about/data-model/metric\n [docs.data_model]: https://vector.dev/docs/about/data-model\n@@ -209,6 +130,7 @@ the License.\n [docs.from-archives]: https://vector.dev/docs/setup/installation/manual/from-archives\n [docs.from-source]: https://vector.dev/docs/setup/installation/manual/from-source\n [docs.guarantees]: https://vector.dev/docs/about/guarantees\n+[docs.guides.getting_started]: https://vector.dev/docs/setup/guides/getting-started\n [docs.guides]: https://vector.dev/docs/setup/guides\n [docs.installation]: https://vector.dev/docs/setup/installation\n [docs.monitoring]: https://vector.dev/docs/administration/monitoring\n@@ -224,72 +146,25 @@ the License.\n [docs.sinks.aws_cloudwatch_metrics]: https://vector.dev/docs/reference/sinks/aws_cloudwatch_metrics\n [docs.sinks.aws_kinesis_streams]: https://vector.dev/docs/reference/sinks/aws_kinesis_streams\n [docs.sinks.aws_s3]: https://vector.dev/docs/reference/sinks/aws_s3\n-[docs.sinks.blackhole]: https://vector.dev/docs/reference/sinks/blackhole\n [docs.sinks.clickhouse]: https://vector.dev/docs/reference/sinks/clickhouse\n-[docs.sinks.console]: https://vector.dev/docs/reference/sinks/console\n-[docs.sinks.datadog_metrics]: https://vector.dev/docs/reference/sinks/datadog_metrics\n [docs.sinks.elasticsearch]: https://vector.dev/docs/reference/sinks/elasticsearch\n-[docs.sinks.file]: https://vector.dev/docs/reference/sinks/file\n-[docs.sinks.http]: https://vector.dev/docs/reference/sinks/http\n-[docs.sinks.kafka]: https://vector.dev/docs/reference/sinks/kafka\n-[docs.sinks.prometheus]: https://vector.dev/docs/reference/sinks/prometheus\n-[docs.sinks.splunk_hec]: https://vector.dev/docs/reference/sinks/splunk_hec\n-[docs.sinks.statsd]: https://vector.dev/docs/reference/sinks/statsd\n-[docs.sinks.tcp]: https://vector.dev/docs/reference/sinks/tcp\n-[docs.sinks.vector]: https://vector.dev/docs/reference/sinks/vector\n [docs.sinks]: https://vector.dev/docs/reference/sinks\n [docs.sources.docker]: https://vector.dev/docs/reference/sources/docker\n [docs.sources.file]: https://vector.dev/docs/reference/sources/file\n [docs.sources.journald]: https://vector.dev/docs/reference/sources/journald\n [docs.sources.kafka]: https://vector.dev/docs/reference/sources/kafka\n-[docs.sources.statsd]: https://vector.dev/docs/reference/sources/statsd\n-[docs.sources.stdin]: https://vector.dev/docs/reference/sources/stdin\n-[docs.sources.syslog]: https://vector.dev/docs/reference/sources/syslog\n-[docs.sources.tcp]: https://vector.dev/docs/reference/sources/tcp\n-[docs.sources.udp]: https://vector.dev/docs/reference/sources/udp\n-[docs.sources.vector]: https://vector.dev/docs/reference/sources/vector\n [docs.sources]: https://vector.dev/docs/reference/sources\n [docs.topologies]: https://vector.dev/docs/setup/deployment/topologies\n-[docs.transforms.add_fields]: https://vector.dev/docs/reference/transforms/add_fields\n-[docs.transforms.add_tags]: https://vector.dev/docs/reference/transforms/add_tags\n-[docs.transforms.coercer]: https://vector.dev/docs/reference/transforms/coercer\n-[docs.transforms.field_filter]: https://vector.dev/docs/reference/transforms/field_filter\n-[docs.transforms.grok_parser]: https://vector.dev/docs/reference/transforms/grok_parser\n [docs.transforms.json_parser]: https://vector.dev/docs/reference/transforms/json_parser\n [docs.transforms.log_to_metric]: https://vector.dev/docs/reference/transforms/log_to_metric\n [docs.transforms.lua]: https://vector.dev/docs/reference/transforms/lua\n [docs.transforms.regex_parser]: https://vector.dev/docs/reference/transforms/regex_parser\n-[docs.transforms.remove_fields]: https://vector.dev/docs/reference/transforms/remove_fields\n-[docs.transforms.remove_tags]: https://vector.dev/docs/reference/transforms/remove_tags\n-[docs.transforms.sampler]: https://vector.dev/docs/reference/transforms/sampler\n-[docs.transforms.split]: https://vector.dev/docs/reference/transforms/split\n-[docs.transforms.tokenizer]: https://vector.dev/docs/reference/transforms/tokenizer\n [docs.transforms]: https://vector.dev/docs/reference/transforms\n [docs.updating]: https://vector.dev/docs/administration/updating\n [docs.use_cases]: https://vector.dev/docs/use_cases\n [docs.validating]: https://vector.dev/docs/administration/validating\n-[urls.aws_cw_logs]: https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/WhatIsCloudWatchLogs.html\n-[urls.aws_cw_metrics]: https://docs.aws.amazon.com/AmazonCloudWatch/latest/monitoring/working_with_metrics.html\n-[urls.aws_kinesis_data_streams]: https://aws.amazon.com/kinesis/data-streams/\n-[urls.aws_s3]: https://aws.amazon.com/s3/\n-[urls.clickhouse]: https://clickhouse.yandex/\n-[urls.clickhouse_http]: https://clickhouse.yandex/docs/en/interfaces/http/\n-[urls.datadog]: https://www.datadoghq.com\n-[urls.elasticsearch]: https://www.elastic.co/products/elasticsearch\n-[urls.grok]: http://grokdebug.herokuapp.com/\n-[urls.kafka]: https://kafka.apache.org/\n-[urls.kafka_protocol]: https://kafka.apache.org/protocol\n-[urls.lua]: https://www.lua.org/\n [urls.mailing_list]: https://vector.dev/mailing_list/\n-[urls.new_sink]: https://github.com/timberio/vector/issues/new?labels=Type%3A+New+Feature\n-[urls.new_source]: https://github.com/timberio/vector/issues/new?labels=Type%3A+New+Feature\n-[urls.new_transform]: https://github.com/timberio/vector/issues/new?labels=Type%3A+New+Feature\n-[urls.prometheus]: https://prometheus.io/\n-[urls.regex]: https://en.wikipedia.org/wiki/Regular_expression\n [urls.rust]: https://www.rust-lang.org/\n-[urls.splunk_hec]: http://dev.splunk.com/view/event-collector/SP-CAAAE6M\n-[urls.standard_streams]: https://en.wikipedia.org/wiki/Standard_streams\n-[urls.statsd]: https://github.com/statsd/statsd\n [urls.test_harness]: https://github.com/timberio/vector-test-harness/\n [urls.v0.5.0]: https://github.com/timberio/vector/releases/tag/v0.5.0\n [urls.vector_changelog]: https://github.com/timberio/vector/blob/master/CHANGELOG.md\ndiff --git a/README.md.erb b/README.md.erb\nindex 3b14aa0..cc241eb 100644\n--- a/README.md.erb\n+++ b/README.md.erb\n@@ -38,15 +38,20 @@ simple and unified.\n * [**Installation**][docs.installation] - [containers][docs.containers], [operating systems][docs.operating_systems], [package managers][docs.package_managers], [from archives][docs.from-archives], [from source][docs.from-source]\n * [**Configuration**][docs.configuration]\n * [**Deployment**][docs.deployment] - [topologies][docs.topologies], [roles][docs.roles]\n+* [**Guides**][docs.guides] - [getting started][docs.guides.getting_started]\n \n-#### [Components](https://vector.dev/components)\n+#### Reference\n \n-* [**Sources**][docs.sources] - \n-* [**Transforms**][docs.transforms]\n-* [**Sinks**][docs.sinks]\n+* [**Sources**][docs.sources] - <%= common_component_links(:source) %>\n+* [**Transforms**][docs.transforms] - <%= common_component_links(:transform) %>\n+* [**Sinks**][docs.sinks] - <%= common_component_links(:sink) %>\n \n-* [**Administration**][docs.administration] - [process management][docs.process-management], [monitoring][docs.monitoring], [updating][docs.updating], [validating][docs.validating]\n-* [**Guides**][docs.guides]\n+#### Administration\n+\n+* [**Process management**][docs.process-management]\n+* [**Monitoring**][docs.monitoring]\n+* [**Updating**][docs.updating]\n+* [**Validating**][docs.validating]\n \n #### Resources\n \n@@ -105,44 +110,6 @@ Run the following in your terminal, then follow the on-screen instructions.\n \n Or view [platform specific installation instructions][docs.installation].\n \n-\n-## Sources\n-\n-<%= components_table(metadata.sources.to_h.values.sort) %>\n-\n-[+ request a new source][urls.new_source]\n-\n-\n-## Transforms\n-\n-<%= components_table(metadata.transforms.to_h.values.sort) %>\n-\n-[+ request a new transform][urls.new_transform]\n-\n-\n-## Sinks\n-\n-<%= components_table(metadata.sinks.to_h.values.sort) %>\n-\n-[+ request a new sink][urls.new_sink]\n-\n-\n-## License\n-\n-Copyright <%= Time.now.year %>, Vector Authors. All rights reserved.\n-\n-Licensed under the Apache License, Version 2.0 (the \"License\"); you may not\n-use these files except in compliance with the License. You may obtain a copy\n-of the License at\n-\n-http://www.apache.org/licenses/LICENSE-2.0\n-\n-Unless required by applicable law or agreed to in writing, software\n-distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n-WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n-License for the specific language governing permissions and limitations under\n-the License.\n-\n ---\n \n <p align=\"center\">\ndiff --git a/scripts/generate/templates.rb b/scripts/generate/templates.rb\nindex e5e7ce7..c793ae0 100644\n--- a/scripts/generate/templates.rb\n+++ b/scripts/generate/templates.rb\n@@ -89,6 +89,23 @@ class Templates\n render(\"#{partials_path}/_commit_type_toc_item.md\", binding).gsub(/,$/, \"\")\n end\n \n+ def common_component_links(type, limit = 5)\n+ common = metadata.send(\"#{type.to_s.pluralize}_list\").select(&:common?)\n+\n+ links =\n+ common[0..limit].collect do |component|\n+ \"[#{component.name}][docs.#{type.to_s.pluralize}.#{component.name}]\"\n+ end\n+\n+ num_leftover = common.size - links.size\n+\n+ if num_leftover > 0\n+ links << \"and [15 more][docs.#{type.to_s.pluralize}]\"\n+ end\n+\n+ links.join(\", \")\n+ end\n+\n def component_config_example(component)\n render(\"#{partials_path}/_component_config_example.md\", binding).strip\n end\ndiff --git a/scripts/util/metadata/component.rb b/scripts/util/metadata/component.rb\nindex 0873b2e..4dc5650 100644\n--- a/scripts/util/metadata/component.rb\n+++ b/scripts/util/metadata/component.rb\n@@ -9,6 +9,7 @@ class Component\n include Comparable\n \n attr_reader :beta,\n+ :common,\n :function_category,\n :id,\n :name,\n@@ -18,6 +19,7 @@ class Component\n \n def initialize(hash)\n @beta = hash[\"beta\"] == true\n+ @common = hash[\"common\"] == true\n @function_category = hash.fetch(\"function_category\")\n @name = hash.fetch(\"name\")\n @type ||= self.class.name.downcase\n@@ -71,6 +73,10 @@ class Component\n beta == true\n end\n \n+ def common?\n+ common == true\n+ end\n+\n def context_options\n options_list.select(&:context?)\n end\ndiff --git a/website/src/components/VectorComponents/index.js b/website/src/components/VectorComponents/index.js\nindex b6c5c13..d3c9adf 100644\n--- a/website/src/components/VectorComponents/index.js\n+++ b/website/src/components/VectorComponents/index.js\n@@ -154,7 +154,7 @@ function VectorComponents(props) {\n //\n \n const [onlyAtLeastOnce, setOnlyAtLeastOnce] = useState(queryObj['at-least-once'] == 'true');\n- const [onlyFunctions, setOnlyFunctions] = useState(new Set(queryObj['providers']));\n+ const [onlyFunctions, setOnlyFunctions] = useState(new Set(queryObj['functions']));\n const [onlyLog, setOnlyLog] = useState(queryObj['log'] == 'true');\n const [onlyMetric, setOnlyMetric] = useState(queryObj['metric'] == 'true');\n const [onlyProductionReady, setOnlyProductionReady] = useState(queryObj['prod-ready'] == 'true');\n", "diff --git a/.circleci/get-deploy-tags.sh b/.circleci/get-deploy-tags.sh\nindex f80c8cb..7ddfa62 100755\n--- a/.circleci/get-deploy-tags.sh\n+++ b/.circleci/get-deploy-tags.sh\n@@ -20,7 +20,7 @@\n set -euo pipefail\n \n DOCKER_IMAGE_TAG=${1}\n-DOCKER_IMAGE=\"quay.io/influxdb/fusion\"\n+DOCKER_IMAGE=\"quay.io/influxdb/iox\"\n APP_NAME=\"IOx\"\n \n DOCKER_IMAGE_DIGEST=\"$(docker image inspect \"${DOCKER_IMAGE}:${DOCKER_IMAGE_TAG}\" --format '{{ if eq (len .RepoDigests) 1 }}{{index .RepoDigests 0}}{{ end }}')\"\n"]
2
["662c5d1346ea2b01c0bc3c11c648cbdf92035fe2", "6786fd5955b064021f5b6d6a630453351d683fae"]
["docs", "ci"]
add important to override paragraphs in items,pin version of actionlint used
["diff --git a/packages/core/src/components/text/text.ios.scss b/packages/core/src/components/text/text.ios.scss\nindex a3c58e2..2a020ab 100644\n--- a/packages/core/src/components/text/text.ios.scss\n+++ b/packages/core/src/components/text/text.ios.scss\n@@ -9,8 +9,9 @@\n @each $color-name, $color-base, $color-contrast in get-colors($colors-ios) {\n \n .text-ios-#{$color-name},\n- .text-ios-#{$color-name} a {\n- color: $color-base;\n+ .text-ios-#{$color-name} a,\n+ .text-ios-#{$color-name} p {\n+ color: $color-base !important\n }\n \n }\ndiff --git a/packages/core/src/components/text/text.md.scss b/packages/core/src/components/text/text.md.scss\nindex b397acb..050af1a 100644\n--- a/packages/core/src/components/text/text.md.scss\n+++ b/packages/core/src/components/text/text.md.scss\n@@ -9,8 +9,9 @@\n @each $color-name, $color-base, $color-contrast in get-colors($colors-md) {\n \n .text-md-#{$color-name},\n- .text-md-#{$color-name} a {\n- color: $color-base;\n+ .text-md-#{$color-name} a,\n+ .text-md-#{$color-name} p {\n+ color: $color-base !important;\n }\n \n }\n", "diff --git a/.github/workflows/introspect.yml b/.github/workflows/introspect.yml\nindex b6d9125..82d22a5 100644\n--- a/.github/workflows/introspect.yml\n+++ b/.github/workflows/introspect.yml\n@@ -25,5 +25,5 @@ jobs:\n # From https://github.com/rhysd/actionlint/blob/main/docs/usage.md#use-actionlint-on-github-actions\n - name: Check workflow files\n run: |\n- bash <(curl https://raw.githubusercontent.com/rhysd/actionlint/590d3bd9dde0c91f7a66071d40eb84716526e5a6/scripts/download-actionlint.bash)\n+ bash <(curl https://raw.githubusercontent.com/rhysd/actionlint/590d3bd9dde0c91f7a66071d40eb84716526e5a6/scripts/download-actionlint.bash) 1.6.25\n ./actionlint -color -shellcheck=\"\"\n"]
2
["7ab363f7ba2807b3eb9895e47f4fcd058f43ae5e", "b702adc245f679ae20d84de39f0d63b14aabed5d"]
["test", "ci"]
add tenant property to ProcessMessageSubscriptionRecord Adds the tenant id property + getter/setter to the record.,document the use of export buckets for large pre-aggregations Co-authored-by: Ray Paik <[email protected]> Co-authored-by: Artyom Keydunov <[email protected]> Co-authored-by: Dmitry Patsura <[email protected]>
["diff --git a/protocol-impl/src/main/java/io/camunda/zeebe/protocol/impl/record/value/message/ProcessMessageSubscriptionRecord.java b/protocol-impl/src/main/java/io/camunda/zeebe/protocol/impl/record/value/message/ProcessMessageSubscriptionRecord.java\nindex 0b07e49..790a974 100644\n--- a/protocol-impl/src/main/java/io/camunda/zeebe/protocol/impl/record/value/message/ProcessMessageSubscriptionRecord.java\n+++ b/protocol-impl/src/main/java/io/camunda/zeebe/protocol/impl/record/value/message/ProcessMessageSubscriptionRecord.java\n@@ -37,6 +37,8 @@ public final class ProcessMessageSubscriptionRecord extends UnifiedRecordValue\n private final BooleanProperty interruptingProp = new BooleanProperty(\"interrupting\", true);\n private final StringProperty correlationKeyProp = new StringProperty(\"correlationKey\", \"\");\n private final StringProperty elementIdProp = new StringProperty(\"elementId\", \"\");\n+ private final StringProperty tenantIdProp =\n+ new StringProperty(\"tenantId\", TenantOwned.DEFAULT_TENANT_IDENTIFIER);\n \n public ProcessMessageSubscriptionRecord() {\n declareProperty(subscriptionPartitionIdProp)\n@@ -48,7 +50,8 @@ public final class ProcessMessageSubscriptionRecord extends UnifiedRecordValue\n .declareProperty(interruptingProp)\n .declareProperty(bpmnProcessIdProp)\n .declareProperty(correlationKeyProp)\n- .declareProperty(elementIdProp);\n+ .declareProperty(elementIdProp)\n+ .declareProperty(tenantIdProp);\n }\n \n public void wrap(final ProcessMessageSubscriptionRecord record) {\n@@ -62,6 +65,7 @@ public final class ProcessMessageSubscriptionRecord extends UnifiedRecordValue\n setBpmnProcessId(record.getBpmnProcessIdBuffer());\n setCorrelationKey(record.getCorrelationKeyBuffer());\n setElementId(record.getElementIdBuffer());\n+ setTenantId(record.getTenantId());\n }\n \n @JsonIgnore\n@@ -191,7 +195,11 @@ public final class ProcessMessageSubscriptionRecord extends UnifiedRecordValue\n \n @Override\n public String getTenantId() {\n- // todo(#13289): replace dummy implementation\n- return TenantOwned.DEFAULT_TENANT_IDENTIFIER;\n+ return bufferAsString(tenantIdProp.getValue());\n+ }\n+\n+ public ProcessMessageSubscriptionRecord setTenantId(final String tenantId) {\n+ tenantIdProp.setValue(tenantId);\n+ return this;\n }\n }\n", "diff --git a/docs/content/Caching/Using-Pre-Aggregations.md b/docs/content/Caching/Using-Pre-Aggregations.md\nindex 7882a25..a927241 100644\n--- a/docs/content/Caching/Using-Pre-Aggregations.md\n+++ b/docs/content/Caching/Using-Pre-Aggregations.md\n@@ -65,8 +65,8 @@ In development mode, Cube.js enables background refresh by default and will\n refresh all pre-aggregations marked with the\n [`scheduledRefresh`](/pre-aggregations#scheduled-refresh) parameter.\n \n-Please consult the [Production Checklist][ref-production-checklist-refresh] for\n-best practices on running background refresh in production environments.\n+Please consult the [Production Checklist][ref-prod-list-refresh] for best\n+practices on running background refresh in production environments.\n \n ```js\n cube(`Orders`, {\n@@ -193,10 +193,20 @@ CUBEJS_EXT_DB_TYPE=<SUPPORTED_DB_TYPE_HERE>\n \n <!-- prettier-ignore-start -->\n [[warning |]]\n-| Please be aware of the limitations when using internal and external (outside of Cube Store) pre-aggregations.\n+| Please be aware of the limitations when using internal and external (outside\n+| of Cube Store) pre-aggregations.\n <!-- prettier-ignore-end -->\n \n-![](https://raw.githubusercontent.com/cube-js/cube.js/master/docs/content/Caching/pre-aggregations.png)\n+<div\n+ style=\"text-align: center\"\n+>\n+ <img\n+ alt=\"Internal vs External vs External with Cube Store diagram\"\n+ src=\"https://raw.githubusercontent.com/cube-js/cube.js/master/docs/content/Caching/pre-aggregations.png\"\n+ style=\"border: none\"\n+ width=\"100%\"\n+ />\n+</div>\n \n #### Some known limitations when using Postgres/MySQL as a storage layer listed below.\n \n@@ -245,15 +255,75 @@ slow to return results.\n (such as AWS Athena and BigQuery). Repeatedly querying for this data can easily\n rack up costs.\n \n+## Optimizing Pre-Aggregation Build Times\n+\n+<!-- prettier-ignore-start -->\n+[[info | ]]\n+| For ideal performance, pre-aggregations should be built using a dedicated\n+| Refresh Worker. [See here for more details][ref-prod-list-refresh].\n+<!-- prettier-ignore-end -->\n+\n+By default, Cube.js will use the source database as a temporary staging area for\n+writing pre-aggregations to determine column types. The data is loaded back into\n+memory before writing them to Cube Store (or an external database).\n+\n+![](build-regular.png)\n+\n+If the dataset is large (more than 100k rows), then Cube.js can face issues when\n+the Node runtime runs out of memory.\n+\n+### Batching\n+\n+Batching is a more performant strategy where Cube.js sends compressed CSVs for\n+Cube Store to ingest.\n+\n+![](build-batching.png)\n+\n+The performance scales to the amount of memory available on the Cube.js\n+instance. Support is currently available for:\n+\n+- [AWS Athena][ref-connect-db-athena] (coming soon)\n+- [AWS Redshift][ref-connect-db-redshift]\n+- [BigQuery][ref-connect-db-bigquery]\n+- [MySQL][ref-connect-db-mysql]\n+- [Postgres][ref-connect-db-postgres]\n+\n+### Export bucket\n+\n+When dealing with larger pre-aggregations (more than 100k rows), performance can\n+be significantly improved by using an export bucket. This allows the source\n+database to persist data directly into cloud storage, which is then loaded into\n+Cube Store in parallel:\n+\n+![](build-export-bucket.png)\n+\n+Export buckets are currently supported for the following databases:\n+\n+- [AWS Athena][ref-connect-db-athena] (coming soon)\n+- [AWS Redshift][ref-connect-db-redshift]\n+- [BigQuery][ref-connect-db-bigquery]\n+- [Snowflake][ref-connect-db-snowflake]\n+\n+When using cloud storage, it is important to correctly configure any data\n+retention policies to clean up the data in the export bucket as Cube.js does not\n+currently manage this. For most use-cases, 1 day is sufficient.\n+\n [wiki-partitioning]: https://en.wikipedia.org/wiki/Partition_(database)\n+[ref-config-connect-db]: /connecting-to-the-database\n+[ref-config-env]: /reference/environment-variables#cube-store\n+[ref-connect-db-athena]: /connecting-to-the-database#notes-aws-athena\n+[ref-connect-db-redshift]: /connecting-to-the-database#notes-aws-redshift\n+[ref-connect-db-bigquery]: /connecting-to-the-database#notes-google-big-query\n+[ref-connect-db-mysql]: /connecting-to-the-database#notes-my-sql\n+[ref-connect-db-postgres]: /connecting-to-the-database#notes-aws-rds-postgres\n+[ref-connect-db-snowflake]: /connecting-to-the-database#notes-snowflake\n [ref-schema-timedimension]: /types-and-formats#dimensions-types-time\n [ref-preaggs]: /pre-aggregations\n [ref-preagg-sched-refresh]: /pre-aggregations#scheduled-refresh\n [ref-preagg-time-part]: /pre-aggregations#rollup-time-partitioning\n [ref-preagg-segment-part]: /pre-aggregations#rollup-segment-partitioning\n [ref-preaggs-refresh-key]: /pre-aggregations#refresh-key\n+[ref-prod-list-refresh]: /deployment/production-checklist#set-up-refresh-worker\n [ref-config-extdbtype]: /config#options-reference-external-db-type\n [ref-config-driverfactory]: /config#options-reference-driver-factory\n [ref-config-extdriverfactory]: /config#options-reference-external-driver-factory\n-[ref-production-checklist-refresh]:\n- /deployment/production-checklist#set-up-refresh-worker\ndiff --git a/docs/content/Caching/build-batching.png b/docs/content/Caching/build-batching.png\nnew file mode 100755\nindex 0000000..d1e28b3\nBinary files /dev/null and b/docs/content/Caching/build-batching.png differ\ndiff --git a/docs/content/Caching/build-export-bucket.png b/docs/content/Caching/build-export-bucket.png\nnew file mode 100755\nindex 0000000..7da2425\nBinary files /dev/null and b/docs/content/Caching/build-export-bucket.png differ\ndiff --git a/docs/content/Caching/build-regular.png b/docs/content/Caching/build-regular.png\nnew file mode 100644\nindex 0000000..af4c3a2\nBinary files /dev/null and b/docs/content/Caching/build-regular.png differ\ndiff --git a/docs/content/Configuration/Connecting-to-the-Database.md b/docs/content/Configuration/Connecting-to-the-Database.md\nindex 321518f..a16ccc4 100644\n--- a/docs/content/Configuration/Connecting-to-the-Database.md\n+++ b/docs/content/Configuration/Connecting-to-the-Database.md\n@@ -49,20 +49,21 @@ CUBEJS_API_SECRET=secret\n The table below shows which environment variables are used for different\n databases:\n \n-| Database | Credentials |\n-| ------------------------------------------------------ | ---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- |\n-| PostgreSQL, MySQL, AWS Redshift, Hive/SparkSQL, Oracle | `CUBEJS_DB_HOST`, `CUBEJS_DB_PORT`, `CUBEJS_DB_NAME`, `CUBEJS_DB_USER`, `CUBEJS_DB_PASS` |\n-| MS SQL | `CUBEJS_DB_HOST`, `CUBEJS_DB_PORT`, `CUBEJS_DB_NAME`, `CUBEJS_DB_USER`, `CUBEJS_DB_PASS`, `CUBEJS_DB_DOMAIN` |\n-| ClickHouse | `CUBEJS_DB_HOST`, `CUBEJS_DB_PORT`, `CUBEJS_DB_NAME`, `CUBEJS_DB_USER`, `CUBEJS_DB_PASS`, `CUBEJS_DB_SSL`, `CUBEJS_DB_CLICKHOUSE_READONLY` |\n-| AWS Athena | `CUBEJS_AWS_KEY`, `CUBEJS_AWS_SECRET`, `CUBEJS_AWS_REGION`, `CUBEJS_AWS_S3_OUTPUT_LOCATION` |\n-| Google BigQuery | `CUBEJS_DB_BQ_PROJECT_ID`, `CUBEJS_DB_BQ_KEY_FILE or CUBEJS_DB_BQ_CREDENTIALS`, `CUBEJS_DB_BQ_LOCATION`, `CUBEJS_DB_BQ_EXPORT_BUCKET` |\n-| MongoDB | `CUBEJS_DB_HOST`, `CUBEJS_DB_NAME`, `CUBEJS_DB_PORT`, `CUBEJS_DB_USER`, `CUBEJS_DB_PASS`, `CUBEJS_DB_SSL`, `CUBEJS_DB_SSL_CA`, `CUBEJS_DB_SSL_CERT`, `CUBEJS_DB_SSL_CIPHERS`, `CUBEJS_DB_SSL_PASSPHRASE` |\n-| Snowflake | `CUBEJS_DB_SNOWFLAKE_ACCOUNT`, `CUBEJS_DB_SNOWFLAKE_REGION`, `CUBEJS_DB_SNOWFLAKE_WAREHOUSE`, `CUBEJS_DB_SNOWFLAKE_ROLE`, `CUBEJS_DB_SNOWFLAKE_CLIENT_SESSION_KEEP_ALIVE`, `CUBEJS_DB_NAME`, `CUBEJS_DB_USER`, `CUBEJS_DB_PASS`, `CUBEJS_DB_SNOWFLAKE_AUTHENTICATOR`, `CUBEJS_DB_SNOWFLAKE_PRIVATE_KEY_PATH`, `CUBEJS_DB_SNOWFLAKE_PRIVATE_KEY_PASS` |\n-| Presto | `CUBEJS_DB_HOST`, `CUBEJS_DB_PORT`, `CUBEJS_DB_CATALOG`, `CUBEJS_DB_SCHEMA`, `CUBEJS_DB_USER`, `CUBEJS_DB_PASS` |\n-| Druid | `CUBEJS_DB_URL`, `CUBEJS_DB_USER`, `CUBEJS_DB_PASS`, `CUBEJS_DB_SSL` |\n-| SQLite | `CUBEJS_DB_NAME` |\n-| Databricks | `CUBEJS_DB_NAME`, `CUBEJS_DB_DATABRICKS_URL` |\n-| Elasticsearch | `CUBEJS_DB_URL`, `CUBEJS_DB_ELASTIC_QUERY_FORMAT`,`CUBEJS_DB_ELASTIC_OPENDISTRO` ,`CUBEJS_DB_ELASTIC_APIKEY_ID`,`CUBEJS_DB_ELASTIC_APIKEY_KEY` |\n+| Database | Credentials |\n+| ---------------------------------------- | ---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- |\n+| PostgreSQL, MySQL, Hive/SparkSQL, Oracle | `CUBEJS_DB_HOST`, `CUBEJS_DB_PORT`, `CUBEJS_DB_NAME`, `CUBEJS_DB_USER`, `CUBEJS_DB_PASS` |\n+| AWS Redshift | `CUBEJS_DB_HOST`, `CUBEJS_DB_PORT`, `CUBEJS_DB_NAME`, `CUBEJS_DB_USER`, `CUBEJS_DB_PASS`, |\n+| MS SQL | `CUBEJS_DB_HOST`, `CUBEJS_DB_PORT`, `CUBEJS_DB_NAME`, `CUBEJS_DB_USER`, `CUBEJS_DB_PASS`, `CUBEJS_DB_DOMAIN` |\n+| ClickHouse | `CUBEJS_DB_HOST`, `CUBEJS_DB_PORT`, `CUBEJS_DB_NAME`, `CUBEJS_DB_USER`, `CUBEJS_DB_PASS`, `CUBEJS_DB_SSL`, `CUBEJS_DB_CLICKHOUSE_READONLY` |\n+| AWS Athena | `CUBEJS_AWS_KEY`, `CUBEJS_AWS_SECRET`, `CUBEJS_AWS_REGION`, `CUBEJS_AWS_S3_OUTPUT_LOCATION` |\n+| Google BigQuery | `CUBEJS_DB_BQ_PROJECT_ID`, `CUBEJS_DB_BQ_KEY_FILE or CUBEJS_DB_BQ_CREDENTIALS`, `CUBEJS_DB_BQ_LOCATION`, |\n+| MongoDB | `CUBEJS_DB_HOST`, `CUBEJS_DB_NAME`, `CUBEJS_DB_PORT`, `CUBEJS_DB_USER`, `CUBEJS_DB_PASS`, `CUBEJS_DB_SSL`, `CUBEJS_DB_SSL_CA`, `CUBEJS_DB_SSL_CERT`, `CUBEJS_DB_SSL_CIPHERS`, `CUBEJS_DB_SSL_PASSPHRASE` |\n+| Snowflake | `CUBEJS_DB_SNOWFLAKE_ACCOUNT`, `CUBEJS_DB_SNOWFLAKE_REGION`, `CUBEJS_DB_SNOWFLAKE_WAREHOUSE`, `CUBEJS_DB_SNOWFLAKE_ROLE`, `CUBEJS_DB_SNOWFLAKE_CLIENT_SESSION_KEEP_ALIVE`, `CUBEJS_DB_NAME`, `CUBEJS_DB_USER`, `CUBEJS_DB_PASS`, `CUBEJS_DB_SNOWFLAKE_AUTHENTICATOR`, `CUBEJS_DB_SNOWFLAKE_PRIVATE_KEY_PATH`, `CUBEJS_DB_SNOWFLAKE_PRIVATE_KEY_PASS` |\n+| Presto | `CUBEJS_DB_HOST`, `CUBEJS_DB_PORT`, `CUBEJS_DB_CATALOG`, `CUBEJS_DB_SCHEMA`, `CUBEJS_DB_USER`, `CUBEJS_DB_PASS` |\n+| Druid | `CUBEJS_DB_URL`, `CUBEJS_DB_USER`, `CUBEJS_DB_PASS`, `CUBEJS_DB_SSL` |\n+| SQLite | `CUBEJS_DB_NAME` |\n+| Databricks | `CUBEJS_DB_NAME`, `CUBEJS_DB_DATABRICKS_URL` |\n+| Elasticsearch | `CUBEJS_DB_URL`, `CUBEJS_DB_ELASTIC_QUERY_FORMAT`,`CUBEJS_DB_ELASTIC_OPENDISTRO` ,`CUBEJS_DB_ELASTIC_APIKEY_ID`,`CUBEJS_DB_ELASTIC_APIKEY_KEY` |\n \n ## Multiple Databases\n \n@@ -195,18 +196,25 @@ You can learn more about acquiring Google BigQuery credentials\n [here][link-bigquery-getting-started] and [here][link-bigquery-credentials].\n \n You can set the dataset location using the `CUBEJS_DB_BQ_LOCATION` environment\n-variable.\n+variable. All supported regions [can be found\n+here][link-bigquery-regional-locations].\n \n ```dotenv\n CUBEJS_DB_BQ_LOCATION=us-central1\n ```\n \n-You can find more supported regions [here][link-bigquery-regional-locations].\n+#### Configuring an export bucket\n \n-If your pre-aggregations dataset is too big to fit in memory, we **strongly**\n-recommend configuring `CUBEJS_DB_BQ_EXPORT_BUCKET`. This will allow Cube.js to\n-materialize results on an \"export\" bucket which are then loaded into BigQuery,\n-providing better performance.\n+<!-- prettier-ignore-start -->\n+[[warning |]]\n+| BigQuery only supports using Google Cloud Storage for export buckets.\n+<!-- prettier-ignore-end -->\n+\n+##### Google Cloud Storage\n+\n+For [improved pre-aggregation performance with large\n+datasets][ref-caching-large-preaggs], enable the export bucket functionality by\n+configuring Cube.js with the following environment variables:\n \n <!-- prettier-ignore-start -->\n [[info |]]\n@@ -216,7 +224,8 @@ providing better performance.\n <!-- prettier-ignore-end -->\n \n ```dotenv\n-CUBEJS_DB_BQ_EXPORT_BUCKET=export_data_58148478376\n+CUBEJS_DB_EXPORT_BUCKET=export_data_58148478376\n+CUBEJS_DB_EXPORT_BUCKET_TYPE=gcp\n ```\n \n ### MSSQL\n@@ -279,6 +288,73 @@ To connect to a Elasticsearch database, use `CUBEJS_DB_URL` with the username\n and password embedded in the URL, if required. If you're not using Elastic\n Cloud, you **must** specify `CUBEJS_DB_ELASTIC_QUERY_FORMAT`.\n \n+### AWS Redshift\n+\n+#### Configuring an export bucket\n+\n+<!-- prettier-ignore-start -->\n+[[warning |]]\n+| AWS Redshift only supports using AWS S3 for export buckets.\n+<!-- prettier-ignore-end -->\n+\n+##### AWS S3\n+\n+For [improved pre-aggregation performance with large\n+datasets][ref-caching-large-preaggs], enable the export bucket functionality by\n+configuring Cube.js with the following environment variables:\n+\n+<!-- prettier-ignore-start -->\n+[[info |]]\n+| Ensure the AWS credentials are correctly configured in IAM to allow reads and\n+| writes to the export bucket.\n+<!-- prettier-ignore-end -->\n+\n+```dotenv\n+CUBEJS_DB_EXPORT_BUCKET_TYPE=s3\n+CUBEJS_DB_EXPORT_BUCKET=my.bucket.on.s3\n+CUBEJS_DB_EXPORT_BUCKET_AWS_KEY=<AWS_KEY>\n+CUBEJS_DB_EXPORT_BUCKET_AWS_SECRET=<AWS_SECRET>\n+CUBEJS_DB_EXPORT_BUCKET_AWS_REGION=<AWS_REGION>\n+```\n+\n+### Snowflake\n+\n+#### Configuring an export bucket\n+\n+Snowflake supports using both AWS S3 and Google Cloud Storage for export bucket\n+functionality.\n+\n+##### AWS S3\n+\n+<!-- prettier-ignore-start -->\n+[[info |]]\n+| Ensure the AWS credentials are correctly configured in IAM to allow reads and\n+| writes to the export bucket.\n+<!-- prettier-ignore-end -->\n+\n+```dotenv\n+CUBEJS_DB_EXPORT_BUCKET_TYPE=s3\n+CUBEJS_DB_EXPORT_BUCKET=my.bucket.on.s3\n+CUBEJS_DB_EXPORT_BUCKET_AWS_KEY=<AWS_KEY>\n+CUBEJS_DB_EXPORT_BUCKET_AWS_SECRET=<AWS_SECRET>\n+CUBEJS_DB_EXPORT_BUCKET_AWS_REGION=<AWS_REGION>\n+```\n+\n+##### Google Cloud Storage\n+\n+Before configuring Cube.js, an [integration must be created and configured in\n+Snowflake][link-snowflake-gcs-integration]. Take note of the integration name\n+(`gcs_int` from the example link) as you'll need it to configure Cube.js.\n+\n+Once the Snowflake integration is set up, configure Cube.js using the following:\n+\n+```dotenv\n+CUBEJS_DB_EXPORT_BUCKET=snowflake-export-bucket\n+CUBEJS_DB_EXPORT_BUCKET_TYPE=gcp\n+CUBEJS_DB_EXPORT_GCS_CREDENTIALS=<BASE64_ENCODED_SERVICE_CREDENTIALS_JSON\n+CUBEJS_DB_EXPORT_INTEGRATION=gcs_int\n+```\n+\n [link-java-guide]:\n https://github.com/cube-js/cube.js/blob/master/packages/cubejs-jdbc-driver/README.md#java-installation\n [link-cubejs-driver-guide]:\n@@ -300,8 +376,11 @@ Cloud, you **must** specify `CUBEJS_DB_ELASTIC_QUERY_FORMAT`.\n https://console.cloud.google.com/apis/credentials/serviceaccountkey\n [link-heroku-postgres-issue]:\n https://help.heroku.com/3DELT3RK/why-can-t-my-third-party-utility-connect-to-heroku-postgres-with-ssl\n+[link-snowflake-gcs-integration]:\n+ https://docs.snowflake.com/en/user-guide/data-load-gcs-config.html\n+[link-bigquery-regional-locations]:\n+ https://cloud.google.com/bigquery/docs/locations#regional-locations\n [ref-cubejs-cli]: /using-the-cubejs-cli\n [ref-enabling-ssl]: #enabling-ssl\n [ref-env-var]: /reference/environment-variables#database-connection\n-[link-bigquery-regional-locations]:\n- https://cloud.google.com/bigquery/docs/locations#regional-locations\n+[ref-caching-large-preaggs]: /using-pre-aggregations#large-pre-aggregations\ndiff --git a/docs/content/Configuration/Environment-Variables-Reference.md b/docs/content/Configuration/Environment-Variables-Reference.md\nindex 692d2c7..6888697 100644\n--- a/docs/content/Configuration/Environment-Variables-Reference.md\n+++ b/docs/content/Configuration/Environment-Variables-Reference.md\n@@ -124,6 +124,18 @@ databases [in this guide][link-connecting-to-db].\n | `CUBEJS_DB_SNOWFLAKE_PRIVATE_KEY_PASS` | Snowflake | The password for the private RSA key. Only required for encrypted keys | A valid password for the encrypted private RSA key |\n | `CUBEJS_DB_DATABRICKS_URL` | Databricks | The URL for a JDBC connection | A valid JDBC URL |\n \n+## Export Bucket\n+\n+| Environment variable | Description | Possible Values |\n+| ------------------------------------ | -------------------------------------------------------------------------------------------- | ---------------------------------------------------------------- |\n+| `CUBEJS_DB_EXPORT_BUCKET` | The name of a bucket in cloud storage | `exports-20210505` |\n+| `CUBEJS_DB_EXPORT_BUCKET_TYPE` | The cloud provider where the bucket is hosted | `gcs`, `s3` |\n+| `CUBEJS_DB_EXPORT_BUCKET_AWS_KEY` | The AWS Access Key ID to use for the export bucket | A valid AWS Access Key ID |\n+| `CUBEJS_DB_EXPORT_BUCKET_AWS_SECRET` | The AWS Secret Access Key to use for the export bucket | A valid AWS Secret Access Key |\n+| `CUBEJS_DB_EXPORT_BUCKET_AWS_REGION` | The AWS region of the export bucket | [A valid AWS region][link-aws-regions] |\n+| `CUBEJS_DB_EXPORT_GCS_CREDENTIALS` | A Base64 encoded JSON key file for connecting to Google Cloud | A valid Google Cloud JSON key file encoded as a Base64 string |\n+| `CUBEJS_DB_EXPORT_INTEGRATION` | The name of the integration used in the database. Only required when using Snowflake and GCS | A valid string matching the name of the integration in Snowflake |\n+\n ## Cube Store\n \n | Environment variable | Description | Possible Values |\n"]
2
["497a7d0676564774c18435e082e7de9132a251c7", "81f37be838d5e3af738908b1bcbf59fea2b45989"]
["feat", "docs"]
added resize observer, this will replace window.resize if available,create mock img server
["diff --git a/engine/src/Utils/EventListeners.ts b/engine/src/Utils/EventListeners.ts\nindex 9e7b189..a29cab4 100644\n--- a/engine/src/Utils/EventListeners.ts\n+++ b/engine/src/Utils/EventListeners.ts\n@@ -47,6 +47,7 @@ export class EventListeners {\n \n private canPush: boolean;\n private resizeTimeout?: NodeJS.Timeout;\n+ private resizeObserver?: ResizeObserver;\n \n /**\n * Events listener constructor\n@@ -144,7 +145,31 @@ export class EventListeners {\n }\n \n if (options.interactivity.events.resize) {\n- manageListener(window, Constants.resizeEvent, this.resizeHandler, add);\n+ if (typeof ResizeObserver !== \"undefined\") {\n+ if (this.resizeObserver && !add) {\n+ if (container.canvas.element) {\n+ this.resizeObserver.unobserve(container.canvas.element);\n+ }\n+\n+ this.resizeObserver.disconnect();\n+\n+ delete this.resizeObserver;\n+ } else if (!this.resizeObserver && add && container.canvas.element) {\n+ this.resizeObserver = new ResizeObserver((entries) => {\n+ const entry = entries.find((e) => e.target === container.canvas.element);\n+\n+ if (!entry) {\n+ return;\n+ }\n+\n+ this.handleWindowResize();\n+ });\n+\n+ this.resizeObserver.observe(container.canvas.element);\n+ }\n+ } else {\n+ manageListener(window, Constants.resizeEvent, this.resizeHandler, add);\n+ }\n }\n \n if (document) {\n", "diff --git a/scripts/gulp/tasks/test.ts b/scripts/gulp/tasks/test.ts\nindex 8014b12..d10c1aa 100644\n--- a/scripts/gulp/tasks/test.ts\n+++ b/scripts/gulp/tasks/test.ts\n@@ -26,12 +26,18 @@ task('test.imageserver', () => {\n function handleRequest(req, res) {\n const urlParse = url.parse(req.url, true);\n \n+ res.setHeader('Access-Control-Allow-Origin', '*');\n+ res.setHeader('Access-Control-Allow-Methods', 'GET');\n+ res.setHeader('Connection', 'keep-alive');\n+ res.setHeader('Age', '0');\n+ res.setHeader('cache-control', 'no-store');\n+\n if (urlParse.pathname === '/reset') {\n console.log('Image Server Reset');\n console.log('---------------------------');\n requestedUrls.length = 0;\n start = Date.now();\n- res.setHeader('Access-Control-Allow-Origin', '*');\n+ res.setHeader('Content-Type', 'text/plain');\n res.end('reset');\n return;\n }\n@@ -48,9 +54,8 @@ task('test.imageserver', () => {\n \n setTimeout(() => {\n res.setHeader('Content-Type', 'image/svg+xml');\n- res.setHeader('Access-Control-Allow-Origin', '*');\n res.end(`<svg xmlns=\"http://www.w3.org/2000/svg\" xmlns:xlink=\"http://www.w3.org/1999/xlink\"\n- style=\"background-color: ${color}; width: ${width}px; height: ${height}px;\">\n+ viewBox=\"0 0 ${width} ${height}\" style=\"background-color: ${color};\">\n <text x=\"5\" y=\"22\" style=\"font-family: Courier; font-size: 24px\">${id}</text>\n </svg>`);\n }, delay);\n"]
2
["4197f2654e8767039dbfd66eca34f261ee3d88c8", "32b76173a259ea1993298289b436cf10c1e800bf"]
["feat", "test"]
update deps,remove unused
["diff --git a/package.json b/package.json\nindex 3696132..ba002d2 100755\n--- a/package.json\n+++ b/package.json\n@@ -34,7 +34,6 @@\n \"koa-router\": \"~7.4.0\",\n \"koa-session\": \"~5.10.1\",\n \"koa-static\": \"~5.0.0\",\n- \"koa2-ratelimit\": \"~0.8.0\",\n \"koa2-swagger-ui\": \"~2.13.2\",\n \"node-fetch\": \"^2.5.0\",\n \"passport-github\": \"~1.1.0\",\ndiff --git a/src/index.ts b/src/index.ts\nindex 847e8aa..8c9baff 100755\n--- a/src/index.ts\n+++ b/src/index.ts\n@@ -4,7 +4,6 @@ import * as Koa from 'koa';\n import * as bodyParser from 'koa-bodyparser';\n import * as session from 'koa-session';\n import * as serve from 'koa-static';\n-// import { RateLimit } from 'koa2-ratelimit';\n import { Server } from 'net';\n \n import { setupPassport } from './auth';\ndiff --git a/src/typings/koa-router.d.ts b/src/typings/koa-router.d.ts\ndeleted file mode 100644\nindex f891ed8..0000000\n--- a/src/typings/koa-router.d.ts\n+++ /dev/null\n@@ -1,16 +0,0 @@\n-import * as koa from 'koa-router';\n-import { IUserSession } from '../models';\n-import { ILogger } from '../logger';\n-\n-declare module 'koa-router' {\n- export interface IRouterContext {\n- state: { user: IUserSession | undefined };\n- logger: ILogger;\n- }\n-}\n-\n-declare module 'koa' {\n- export interface Context {\n- logger: ILogger;\n- }\n-}\ndiff --git a/src/typings/koa2-ratelimit/index.d.ts b/src/typings/koa2-ratelimit/index.d.ts\ndeleted file mode 100644\nindex cc73228..0000000\n--- a/src/typings/koa2-ratelimit/index.d.ts\n+++ /dev/null\n@@ -1,13 +0,0 @@\n-declare module 'koa2-ratelimit' {\n- import * as Koa from 'koa';\n- interface Props {\n- interval: { min?: number; hour?: number }; // 15 minutes = 15*60*1000\n- max: number;\n- message?: string;\n- prefixKey?: string;\n- getUserId?: (ctx: Koa.Context) => Promise<string>;\n- }\n- export namespace RateLimit {\n- function middleware(options: Props): Koa.Middleware;\n- }\n-}\ndiff --git a/src/typings/mockingoose/index.d.ts b/src/typings/mockingoose/index.d.ts\ndeleted file mode 100644\nindex 757c4e4..0000000\n--- a/src/typings/mockingoose/index.d.ts\n+++ /dev/null\n@@ -1,5 +0,0 @@\n-declare module 'mockingoose' {\n- const mockingoose: any;\n-\n- export default mockingoose;\n-}\n", "diff --git a/src/content/redux/modules/dictionaries.ts b/src/content/redux/modules/dictionaries.ts\nindex 88f7215..570d397 100644\n--- a/src/content/redux/modules/dictionaries.ts\n+++ b/src/content/redux/modules/dictionaries.ts\n@@ -3,7 +3,6 @@ import { DictID, appConfigFactory, AppConfig } from '@/app-config'\n import isEqual from 'lodash/isEqual'\n import { saveWord } from '@/_helpers/record-manager'\n import { getDefaultSelectionInfo, SelectionInfo, isSameSelection } from '@/_helpers/selection'\n-import { createActiveConfigStream } from '@/_helpers/config-manager'\n import { isContainChinese, isContainEnglish, testerPunct, isContainMinor, testerChinese, testJapanese, testKorean } from '@/_helpers/lang-check'\n import { MsgType, MsgFetchDictResult } from '@/typings/message'\n import { StoreState, DispatcherThunk, Dispatcher } from './index'\ndiff --git a/src/content/redux/modules/widget.ts b/src/content/redux/modules/widget.ts\nindex 53ad550..68e0a3d 100644\n--- a/src/content/redux/modules/widget.ts\n+++ b/src/content/redux/modules/widget.ts\n@@ -1,9 +1,9 @@\n import * as recordManager from '@/_helpers/record-manager'\n import { StoreState, DispatcherThunk, Dispatcher } from './index'\n-import appConfigFactory, { TCDirection, AppConfig, DictID } from '@/app-config'\n+import appConfigFactory, { TCDirection, DictID } from '@/app-config'\n import { message, storage } from '@/_helpers/browser-api'\n-import { createActiveConfigStream, createConfigIDListStream } from '@/_helpers/config-manager'\n-import { MsgSelection, MsgType, MsgTempDisabledState, MsgEditWord, MsgOpenUrl, MsgFetchDictResult } from '@/typings/message'\n+import { createConfigIDListStream } from '@/_helpers/config-manager'\n+import { MsgType, MsgTempDisabledState, MsgEditWord, MsgOpenUrl, MsgFetchDictResult } from '@/typings/message'\n import { searchText, restoreDicts } from '@/content/redux/modules/dictionaries'\n import { SelectionInfo, getDefaultSelectionInfo } from '@/_helpers/selection'\n import { Mutable } from '@/typings/helpers'\n"]
2
["f46c6c6c26a14312aa05a77ff2a14aebd74e13ac", "a50b51999015e210918d9c8e95fd4cac347353be"]
["build", "refactor"]
Fix readme Signed-off-by: Ben Johnson <[email protected]>,convert to record
["diff --git a/README.md b/README.md\nindex 587d655..da746bb 100644\n--- a/README.md\n+++ b/README.md\n@@ -38,15 +38,20 @@ simple and unified.\n * [**Installation**][docs.installation] - [containers][docs.containers], [operating systems][docs.operating_systems], [package managers][docs.package_managers], [from archives][docs.from-archives], [from source][docs.from-source]\n * [**Configuration**][docs.configuration]\n * [**Deployment**][docs.deployment] - [topologies][docs.topologies], [roles][docs.roles]\n+* [**Guides**][docs.guides] - [getting started][docs.guides.getting_started]\n \n-#### [Components](https://vector.dev/components)\n+#### Reference\n \n-* [**Sources**][docs.sources] - \n-* [**Transforms**][docs.transforms]\n-* [**Sinks**][docs.sinks]\n+* [**Sources**][docs.sources] - [docker][docs.sources.docker], [file][docs.sources.file], [journald][docs.sources.journald], [kafka][docs.sources.kafka]\n+* [**Transforms**][docs.transforms] - [json_parser][docs.transforms.json_parser], [log_to_metric][docs.transforms.log_to_metric], [lua][docs.transforms.lua], [regex_parser][docs.transforms.regex_parser]\n+* [**Sinks**][docs.sinks] - [aws_cloudwatch_logs][docs.sinks.aws_cloudwatch_logs], [aws_cloudwatch_metrics][docs.sinks.aws_cloudwatch_metrics], [aws_kinesis_streams][docs.sinks.aws_kinesis_streams], [aws_s3][docs.sinks.aws_s3], [clickhouse][docs.sinks.clickhouse], [elasticsearch][docs.sinks.elasticsearch], and [15 more][docs.sinks]\n \n-* [**Administration**][docs.administration] - [process management][docs.process-management], [monitoring][docs.monitoring], [updating][docs.updating], [validating][docs.validating]\n-* [**Guides**][docs.guides]\n+#### Administration\n+\n+* [**Process management**][docs.process-management]\n+* [**Monitoring**][docs.monitoring]\n+* [**Updating**][docs.updating]\n+* [**Validating**][docs.validating]\n \n #### Resources\n \n@@ -105,88 +110,6 @@ curl --proto '=https' --tlsv1.2 -sSf https://sh.vector.dev | sh\n \n Or view [platform specific installation instructions][docs.installation].\n \n-\n-## Sources\n-\n-| Name | Description |\n-|:------|:------------|\n-| [**`docker`**][docs.sources.docker] | Ingests data through the docker engine daemon and outputs [`log`][docs.data-model#log] events. |\n-| [**`file`**][docs.sources.file] | Ingests data through one or more local files and outputs [`log`][docs.data-model#log] events. |\n-| [**`journald`**][docs.sources.journald] | Ingests data through log records from journald and outputs [`log`][docs.data-model#log] events. |\n-| [**`kafka`**][docs.sources.kafka] | Ingests data through Kafka 0.9 or later and outputs [`log`][docs.data-model#log] events. |\n-| [**`statsd`**][docs.sources.statsd] | Ingests data through the StatsD UDP protocol and outputs [`metric`][docs.data-model#metric] events. |\n-| [**`stdin`**][docs.sources.stdin] | Ingests data through standard input (STDIN) and outputs [`log`][docs.data-model#log] events. |\n-| [**`syslog`**][docs.sources.syslog] | Ingests data through the Syslog 5424 protocol and outputs [`log`][docs.data-model#log] events. |\n-| [**`tcp`**][docs.sources.tcp] | Ingests data through the TCP protocol and outputs [`log`][docs.data-model#log] events. |\n-| [**`udp`**][docs.sources.udp] | Ingests data through the UDP protocol and outputs [`log`][docs.data-model#log] events. |\n-| [**`vector`**][docs.sources.vector] | Ingests data through another upstream [`vector` sink][docs.sinks.vector] and outputs [`log`][docs.data-model#log] and [`metric`][docs.data-model#metric] events. |\n-\n-[+ request a new source][urls.new_source]\n-\n-\n-## Transforms\n-\n-| Name | Description |\n-|:------|:------------|\n-| [**`add_fields`**][docs.transforms.add_fields] | Accepts [`log`][docs.data-model#log] events and allows you to add one or more log fields. |\n-| [**`add_tags`**][docs.transforms.add_tags] | Accepts [`metric`][docs.data-model#metric] events and allows you to add one or more metric tags. |\n-| [**`coercer`**][docs.transforms.coercer] | Accepts [`log`][docs.data-model#log] events and allows you to coerce log fields into fixed types. |\n-| [**`field_filter`**][docs.transforms.field_filter] | Accepts [`log`][docs.data-model#log] and [`metric`][docs.data-model#metric] events and allows you to filter events by a log field's value. |\n-| [**`grok_parser`**][docs.transforms.grok_parser] | Accepts [`log`][docs.data-model#log] events and allows you to parse a log field value with [Grok][urls.grok]. |\n-| [**`json_parser`**][docs.transforms.json_parser] | Accepts [`log`][docs.data-model#log] events and allows you to parse a log field value as JSON. |\n-| [**`log_to_metric`**][docs.transforms.log_to_metric] | Accepts [`log`][docs.data-model#log] events and allows you to convert logs into one or more metrics. |\n-| [**`lua`**][docs.transforms.lua] | Accepts [`log`][docs.data-model#log] events and allows you to transform events with a full embedded [Lua][urls.lua] engine. |\n-| [**`regex_parser`**][docs.transforms.regex_parser] | Accepts [`log`][docs.data-model#log] events and allows you to parse a log field's value with a [Regular Expression][urls.regex]. |\n-| [**`remove_fields`**][docs.transforms.remove_fields] | Accepts [`log`][docs.data-model#log] events and allows you to remove one or more log fields. |\n-| [**`remove_tags`**][docs.transforms.remove_tags] | Accepts [`metric`][docs.data-model#metric] events and allows you to remove one or more metric tags. |\n-| [**`sampler`**][docs.transforms.sampler] | Accepts [`log`][docs.data-model#log] events and allows you to sample events with a configurable rate. |\n-| [**`split`**][docs.transforms.split] | Accepts [`log`][docs.data-model#log] events and allows you to split a field's value on a given separator and zip the tokens into ordered field names. |\n-| [**`tokenizer`**][docs.transforms.tokenizer] | Accepts [`log`][docs.data-model#log] events and allows you to tokenize a field's value by splitting on white space, ignoring special wrapping characters, and zip the tokens into ordered field names. |\n-\n-[+ request a new transform][urls.new_transform]\n-\n-\n-## Sinks\n-\n-| Name | Description |\n-|:------|:------------|\n-| [**`aws_cloudwatch_logs`**][docs.sinks.aws_cloudwatch_logs] | [Batches](#buffers-and-batches) [`log`][docs.data-model#log] events to [AWS CloudWatch Logs][urls.aws_cw_logs] via the [`PutLogEvents` API endpoint](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutLogEvents.html). |\n-| [**`aws_cloudwatch_metrics`**][docs.sinks.aws_cloudwatch_metrics] | [Streams](#streaming) [`metric`][docs.data-model#metric] events to [AWS CloudWatch Metrics][urls.aws_cw_metrics] via the [`PutMetricData` API endpoint](https://docs.aws.amazon.com/AmazonCloudWatch/latest/APIReference/API_PutMetricData.html). |\n-| [**`aws_kinesis_streams`**][docs.sinks.aws_kinesis_streams] | [Batches](#buffers-and-batches) [`log`][docs.data-model#log] events to [AWS Kinesis Data Stream][urls.aws_kinesis_data_streams] via the [`PutRecords` API endpoint](https://docs.aws.amazon.com/kinesis/latest/APIReference/API_PutRecords.html). |\n-| [**`aws_s3`**][docs.sinks.aws_s3] | [Batches](#buffers-and-batches) [`log`][docs.data-model#log] events to [AWS S3][urls.aws_s3] via the [`PutObject` API endpoint](https://docs.aws.amazon.com/AmazonS3/latest/API/RESTObjectPUT.html). |\n-| [**`blackhole`**][docs.sinks.blackhole] | [Streams](#streaming) [`log`][docs.data-model#log] and [`metric`][docs.data-model#metric] events to a blackhole that simply discards data, designed for testing and benchmarking purposes. |\n-| [**`clickhouse`**][docs.sinks.clickhouse] | [Batches](#buffers-and-batches) [`log`][docs.data-model#log] events to [Clickhouse][urls.clickhouse] via the [`HTTP` Interface][urls.clickhouse_http]. |\n-| [**`console`**][docs.sinks.console] | [Streams](#streaming) [`log`][docs.data-model#log] and [`metric`][docs.data-model#metric] events to [standard output streams][urls.standard_streams], such as `STDOUT` and `STDERR`. |\n-| [**`datadog_metrics`**][docs.sinks.datadog_metrics] | [Batches](#buffers-and-batches) [`metric`][docs.data-model#metric] events to [Datadog][urls.datadog] metrics service using [HTTP API](https://docs.datadoghq.com/api/?lang=bash#metrics). |\n-| [**`elasticsearch`**][docs.sinks.elasticsearch] | [Batches](#buffers-and-batches) [`log`][docs.data-model#log] events to [Elasticsearch][urls.elasticsearch] via the [`_bulk` API endpoint](https://www.elastic.co/guide/en/elasticsearch/reference/current/docs-bulk.html). |\n-| [**`file`**][docs.sinks.file] | [Streams](#streaming) [`log`][docs.data-model#log] events to a file. |\n-| [**`http`**][docs.sinks.http] | [Batches](#buffers-and-batches) [`log`][docs.data-model#log] events to a generic HTTP endpoint. |\n-| [**`kafka`**][docs.sinks.kafka] | [Streams](#streaming) [`log`][docs.data-model#log] events to [Apache Kafka][urls.kafka] via the [Kafka protocol][urls.kafka_protocol]. |\n-| [**`prometheus`**][docs.sinks.prometheus] | [Exposes](#exposing-and-scraping) [`metric`][docs.data-model#metric] events to [Prometheus][urls.prometheus] metrics service. |\n-| [**`splunk_hec`**][docs.sinks.splunk_hec] | [Batches](#buffers-and-batches) [`log`][docs.data-model#log] events to a [Splunk HTTP Event Collector][urls.splunk_hec]. |\n-| [**`statsd`**][docs.sinks.statsd] | [Streams](#streaming) [`metric`][docs.data-model#metric] events to [StatsD][urls.statsd] metrics service. |\n-| [**`tcp`**][docs.sinks.tcp] | [Streams](#streaming) [`log`][docs.data-model#log] events to a TCP connection. |\n-| [**`vector`**][docs.sinks.vector] | [Streams](#streaming) [`log`][docs.data-model#log] events to another downstream [`vector` source][docs.sources.vector]. |\n-\n-[+ request a new sink][urls.new_sink]\n-\n-\n-## License\n-\n-Copyright 2019, Vector Authors. All rights reserved.\n-\n-Licensed under the Apache License, Version 2.0 (the \"License\"); you may not\n-use these files except in compliance with the License. You may obtain a copy\n-of the License at\n-\n-http://www.apache.org/licenses/LICENSE-2.0\n-\n-Unless required by applicable law or agreed to in writing, software\n-distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n-WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n-License for the specific language governing permissions and limitations under\n-the License.\n-\n ---\n \n <p align=\"center\">\n@@ -200,8 +123,6 @@ the License.\n [docs.configuration]: https://vector.dev/docs/setup/configuration\n [docs.containers]: https://vector.dev/docs/setup/installation/containers\n [docs.correctness]: https://vector.dev/docs/about/correctness\n-[docs.data-model#log]: https://vector.dev/docs/about/data-model#log\n-[docs.data-model#metric]: https://vector.dev/docs/about/data-model#metric\n [docs.data-model.log]: https://vector.dev/docs/about/data-model/log\n [docs.data-model.metric]: https://vector.dev/docs/about/data-model/metric\n [docs.data_model]: https://vector.dev/docs/about/data-model\n@@ -209,6 +130,7 @@ the License.\n [docs.from-archives]: https://vector.dev/docs/setup/installation/manual/from-archives\n [docs.from-source]: https://vector.dev/docs/setup/installation/manual/from-source\n [docs.guarantees]: https://vector.dev/docs/about/guarantees\n+[docs.guides.getting_started]: https://vector.dev/docs/setup/guides/getting-started\n [docs.guides]: https://vector.dev/docs/setup/guides\n [docs.installation]: https://vector.dev/docs/setup/installation\n [docs.monitoring]: https://vector.dev/docs/administration/monitoring\n@@ -224,72 +146,25 @@ the License.\n [docs.sinks.aws_cloudwatch_metrics]: https://vector.dev/docs/reference/sinks/aws_cloudwatch_metrics\n [docs.sinks.aws_kinesis_streams]: https://vector.dev/docs/reference/sinks/aws_kinesis_streams\n [docs.sinks.aws_s3]: https://vector.dev/docs/reference/sinks/aws_s3\n-[docs.sinks.blackhole]: https://vector.dev/docs/reference/sinks/blackhole\n [docs.sinks.clickhouse]: https://vector.dev/docs/reference/sinks/clickhouse\n-[docs.sinks.console]: https://vector.dev/docs/reference/sinks/console\n-[docs.sinks.datadog_metrics]: https://vector.dev/docs/reference/sinks/datadog_metrics\n [docs.sinks.elasticsearch]: https://vector.dev/docs/reference/sinks/elasticsearch\n-[docs.sinks.file]: https://vector.dev/docs/reference/sinks/file\n-[docs.sinks.http]: https://vector.dev/docs/reference/sinks/http\n-[docs.sinks.kafka]: https://vector.dev/docs/reference/sinks/kafka\n-[docs.sinks.prometheus]: https://vector.dev/docs/reference/sinks/prometheus\n-[docs.sinks.splunk_hec]: https://vector.dev/docs/reference/sinks/splunk_hec\n-[docs.sinks.statsd]: https://vector.dev/docs/reference/sinks/statsd\n-[docs.sinks.tcp]: https://vector.dev/docs/reference/sinks/tcp\n-[docs.sinks.vector]: https://vector.dev/docs/reference/sinks/vector\n [docs.sinks]: https://vector.dev/docs/reference/sinks\n [docs.sources.docker]: https://vector.dev/docs/reference/sources/docker\n [docs.sources.file]: https://vector.dev/docs/reference/sources/file\n [docs.sources.journald]: https://vector.dev/docs/reference/sources/journald\n [docs.sources.kafka]: https://vector.dev/docs/reference/sources/kafka\n-[docs.sources.statsd]: https://vector.dev/docs/reference/sources/statsd\n-[docs.sources.stdin]: https://vector.dev/docs/reference/sources/stdin\n-[docs.sources.syslog]: https://vector.dev/docs/reference/sources/syslog\n-[docs.sources.tcp]: https://vector.dev/docs/reference/sources/tcp\n-[docs.sources.udp]: https://vector.dev/docs/reference/sources/udp\n-[docs.sources.vector]: https://vector.dev/docs/reference/sources/vector\n [docs.sources]: https://vector.dev/docs/reference/sources\n [docs.topologies]: https://vector.dev/docs/setup/deployment/topologies\n-[docs.transforms.add_fields]: https://vector.dev/docs/reference/transforms/add_fields\n-[docs.transforms.add_tags]: https://vector.dev/docs/reference/transforms/add_tags\n-[docs.transforms.coercer]: https://vector.dev/docs/reference/transforms/coercer\n-[docs.transforms.field_filter]: https://vector.dev/docs/reference/transforms/field_filter\n-[docs.transforms.grok_parser]: https://vector.dev/docs/reference/transforms/grok_parser\n [docs.transforms.json_parser]: https://vector.dev/docs/reference/transforms/json_parser\n [docs.transforms.log_to_metric]: https://vector.dev/docs/reference/transforms/log_to_metric\n [docs.transforms.lua]: https://vector.dev/docs/reference/transforms/lua\n [docs.transforms.regex_parser]: https://vector.dev/docs/reference/transforms/regex_parser\n-[docs.transforms.remove_fields]: https://vector.dev/docs/reference/transforms/remove_fields\n-[docs.transforms.remove_tags]: https://vector.dev/docs/reference/transforms/remove_tags\n-[docs.transforms.sampler]: https://vector.dev/docs/reference/transforms/sampler\n-[docs.transforms.split]: https://vector.dev/docs/reference/transforms/split\n-[docs.transforms.tokenizer]: https://vector.dev/docs/reference/transforms/tokenizer\n [docs.transforms]: https://vector.dev/docs/reference/transforms\n [docs.updating]: https://vector.dev/docs/administration/updating\n [docs.use_cases]: https://vector.dev/docs/use_cases\n [docs.validating]: https://vector.dev/docs/administration/validating\n-[urls.aws_cw_logs]: https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/WhatIsCloudWatchLogs.html\n-[urls.aws_cw_metrics]: https://docs.aws.amazon.com/AmazonCloudWatch/latest/monitoring/working_with_metrics.html\n-[urls.aws_kinesis_data_streams]: https://aws.amazon.com/kinesis/data-streams/\n-[urls.aws_s3]: https://aws.amazon.com/s3/\n-[urls.clickhouse]: https://clickhouse.yandex/\n-[urls.clickhouse_http]: https://clickhouse.yandex/docs/en/interfaces/http/\n-[urls.datadog]: https://www.datadoghq.com\n-[urls.elasticsearch]: https://www.elastic.co/products/elasticsearch\n-[urls.grok]: http://grokdebug.herokuapp.com/\n-[urls.kafka]: https://kafka.apache.org/\n-[urls.kafka_protocol]: https://kafka.apache.org/protocol\n-[urls.lua]: https://www.lua.org/\n [urls.mailing_list]: https://vector.dev/mailing_list/\n-[urls.new_sink]: https://github.com/timberio/vector/issues/new?labels=Type%3A+New+Feature\n-[urls.new_source]: https://github.com/timberio/vector/issues/new?labels=Type%3A+New+Feature\n-[urls.new_transform]: https://github.com/timberio/vector/issues/new?labels=Type%3A+New+Feature\n-[urls.prometheus]: https://prometheus.io/\n-[urls.regex]: https://en.wikipedia.org/wiki/Regular_expression\n [urls.rust]: https://www.rust-lang.org/\n-[urls.splunk_hec]: http://dev.splunk.com/view/event-collector/SP-CAAAE6M\n-[urls.standard_streams]: https://en.wikipedia.org/wiki/Standard_streams\n-[urls.statsd]: https://github.com/statsd/statsd\n [urls.test_harness]: https://github.com/timberio/vector-test-harness/\n [urls.v0.5.0]: https://github.com/timberio/vector/releases/tag/v0.5.0\n [urls.vector_changelog]: https://github.com/timberio/vector/blob/master/CHANGELOG.md\ndiff --git a/README.md.erb b/README.md.erb\nindex 3b14aa0..cc241eb 100644\n--- a/README.md.erb\n+++ b/README.md.erb\n@@ -38,15 +38,20 @@ simple and unified.\n * [**Installation**][docs.installation] - [containers][docs.containers], [operating systems][docs.operating_systems], [package managers][docs.package_managers], [from archives][docs.from-archives], [from source][docs.from-source]\n * [**Configuration**][docs.configuration]\n * [**Deployment**][docs.deployment] - [topologies][docs.topologies], [roles][docs.roles]\n+* [**Guides**][docs.guides] - [getting started][docs.guides.getting_started]\n \n-#### [Components](https://vector.dev/components)\n+#### Reference\n \n-* [**Sources**][docs.sources] - \n-* [**Transforms**][docs.transforms]\n-* [**Sinks**][docs.sinks]\n+* [**Sources**][docs.sources] - <%= common_component_links(:source) %>\n+* [**Transforms**][docs.transforms] - <%= common_component_links(:transform) %>\n+* [**Sinks**][docs.sinks] - <%= common_component_links(:sink) %>\n \n-* [**Administration**][docs.administration] - [process management][docs.process-management], [monitoring][docs.monitoring], [updating][docs.updating], [validating][docs.validating]\n-* [**Guides**][docs.guides]\n+#### Administration\n+\n+* [**Process management**][docs.process-management]\n+* [**Monitoring**][docs.monitoring]\n+* [**Updating**][docs.updating]\n+* [**Validating**][docs.validating]\n \n #### Resources\n \n@@ -105,44 +110,6 @@ Run the following in your terminal, then follow the on-screen instructions.\n \n Or view [platform specific installation instructions][docs.installation].\n \n-\n-## Sources\n-\n-<%= components_table(metadata.sources.to_h.values.sort) %>\n-\n-[+ request a new source][urls.new_source]\n-\n-\n-## Transforms\n-\n-<%= components_table(metadata.transforms.to_h.values.sort) %>\n-\n-[+ request a new transform][urls.new_transform]\n-\n-\n-## Sinks\n-\n-<%= components_table(metadata.sinks.to_h.values.sort) %>\n-\n-[+ request a new sink][urls.new_sink]\n-\n-\n-## License\n-\n-Copyright <%= Time.now.year %>, Vector Authors. All rights reserved.\n-\n-Licensed under the Apache License, Version 2.0 (the \"License\"); you may not\n-use these files except in compliance with the License. You may obtain a copy\n-of the License at\n-\n-http://www.apache.org/licenses/LICENSE-2.0\n-\n-Unless required by applicable law or agreed to in writing, software\n-distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n-WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n-License for the specific language governing permissions and limitations under\n-the License.\n-\n ---\n \n <p align=\"center\">\ndiff --git a/scripts/generate/templates.rb b/scripts/generate/templates.rb\nindex e5e7ce7..c793ae0 100644\n--- a/scripts/generate/templates.rb\n+++ b/scripts/generate/templates.rb\n@@ -89,6 +89,23 @@ class Templates\n render(\"#{partials_path}/_commit_type_toc_item.md\", binding).gsub(/,$/, \"\")\n end\n \n+ def common_component_links(type, limit = 5)\n+ common = metadata.send(\"#{type.to_s.pluralize}_list\").select(&:common?)\n+\n+ links =\n+ common[0..limit].collect do |component|\n+ \"[#{component.name}][docs.#{type.to_s.pluralize}.#{component.name}]\"\n+ end\n+\n+ num_leftover = common.size - links.size\n+\n+ if num_leftover > 0\n+ links << \"and [15 more][docs.#{type.to_s.pluralize}]\"\n+ end\n+\n+ links.join(\", \")\n+ end\n+\n def component_config_example(component)\n render(\"#{partials_path}/_component_config_example.md\", binding).strip\n end\ndiff --git a/scripts/util/metadata/component.rb b/scripts/util/metadata/component.rb\nindex 0873b2e..4dc5650 100644\n--- a/scripts/util/metadata/component.rb\n+++ b/scripts/util/metadata/component.rb\n@@ -9,6 +9,7 @@ class Component\n include Comparable\n \n attr_reader :beta,\n+ :common,\n :function_category,\n :id,\n :name,\n@@ -18,6 +19,7 @@ class Component\n \n def initialize(hash)\n @beta = hash[\"beta\"] == true\n+ @common = hash[\"common\"] == true\n @function_category = hash.fetch(\"function_category\")\n @name = hash.fetch(\"name\")\n @type ||= self.class.name.downcase\n@@ -71,6 +73,10 @@ class Component\n beta == true\n end\n \n+ def common?\n+ common == true\n+ end\n+\n def context_options\n options_list.select(&:context?)\n end\ndiff --git a/website/src/components/VectorComponents/index.js b/website/src/components/VectorComponents/index.js\nindex b6c5c13..d3c9adf 100644\n--- a/website/src/components/VectorComponents/index.js\n+++ b/website/src/components/VectorComponents/index.js\n@@ -154,7 +154,7 @@ function VectorComponents(props) {\n //\n \n const [onlyAtLeastOnce, setOnlyAtLeastOnce] = useState(queryObj['at-least-once'] == 'true');\n- const [onlyFunctions, setOnlyFunctions] = useState(new Set(queryObj['providers']));\n+ const [onlyFunctions, setOnlyFunctions] = useState(new Set(queryObj['functions']));\n const [onlyLog, setOnlyLog] = useState(queryObj['log'] == 'true');\n const [onlyMetric, setOnlyMetric] = useState(queryObj['metric'] == 'true');\n const [onlyProductionReady, setOnlyProductionReady] = useState(queryObj['prod-ready'] == 'true');\n", "diff --git a/broker/src/test/java/io/camunda/zeebe/broker/exporter/stream/ExporterDirectorDistributionTest.java b/broker/src/test/java/io/camunda/zeebe/broker/exporter/stream/ExporterDirectorDistributionTest.java\nindex cc998c6..65c8550 100755\n--- a/broker/src/test/java/io/camunda/zeebe/broker/exporter/stream/ExporterDirectorDistributionTest.java\n+++ b/broker/src/test/java/io/camunda/zeebe/broker/exporter/stream/ExporterDirectorDistributionTest.java\n@@ -167,13 +167,8 @@ public final class ExporterDirectorDistributionTest {\n * <p>This makes sure that even if we miss one export position event, we distribute the event\n * later again, which makes tests less flaky.\n */\n- private static final class ClockShifter implements ConditionEvaluationListener<Void> {\n-\n- private final ControlledActorClock clock;\n-\n- public ClockShifter(final ControlledActorClock clock) {\n- this.clock = clock;\n- }\n+ private record ClockShifter(ControlledActorClock clock)\n+ implements ConditionEvaluationListener<Void> {\n \n @Override\n public void conditionEvaluated(final EvaluatedCondition<Void> condition) {\n"]
2
["662c5d1346ea2b01c0bc3c11c648cbdf92035fe2", "3346331a963766c8193170fb130adad2e658ada2"]
["docs", "refactor"]
add workingDirectory option to shell.openExternal() (#15065) Allows passing `workingDirectory` to the underlying `ShellExecuteW` API on Windows._x000D_ _x000D_ the motivation is that by default `ShellExecute` would use the current working directory, which would get locked on Windows and can prevent autoUpdater from working correctly. We need to be able specify a different `workingDirectory` to prevent this situation.,lint source on ci
["diff --git a/atom/browser/atom_browser_client.cc b/atom/browser/atom_browser_client.cc\nindex 97e5f26..df0774b 100644\n--- a/atom/browser/atom_browser_client.cc\n+++ b/atom/browser/atom_browser_client.cc\n@@ -611,7 +611,7 @@ void OnOpenExternal(const GURL& escaped_url, bool allowed) {\n #else\n escaped_url,\n #endif\n- true);\n+ platform_util::OpenExternalOptions());\n }\n \n void HandleExternalProtocolInUI(\ndiff --git a/atom/common/api/atom_api_shell.cc b/atom/common/api/atom_api_shell.cc\nindex 1323cd6..7c67c7a 100644\n--- a/atom/common/api/atom_api_shell.cc\n+++ b/atom/common/api/atom_api_shell.cc\n@@ -60,11 +60,12 @@ bool OpenExternal(\n const GURL& url,\n #endif\n mate::Arguments* args) {\n- bool activate = true;\n+ platform_util::OpenExternalOptions options;\n if (args->Length() >= 2) {\n- mate::Dictionary options;\n- if (args->GetNext(&options)) {\n- options.Get(\"activate\", &activate);\n+ mate::Dictionary obj;\n+ if (args->GetNext(&obj)) {\n+ obj.Get(\"activate\", &options.activate);\n+ obj.Get(\"workingDirectory\", &options.working_dir);\n }\n }\n \n@@ -72,13 +73,13 @@ bool OpenExternal(\n base::Callback<void(v8::Local<v8::Value>)> callback;\n if (args->GetNext(&callback)) {\n platform_util::OpenExternal(\n- url, activate,\n+ url, options,\n base::Bind(&OnOpenExternalFinished, args->isolate(), callback));\n return true;\n }\n }\n \n- return platform_util::OpenExternal(url, activate);\n+ return platform_util::OpenExternal(url, options);\n }\n \n #if defined(OS_WIN)\ndiff --git a/atom/common/platform_util.h b/atom/common/platform_util.h\nindex 6fd8405..6686a4f 100644\n--- a/atom/common/platform_util.h\n+++ b/atom/common/platform_util.h\n@@ -8,6 +8,7 @@\n #include <string>\n \n #include \"base/callback_forward.h\"\n+#include \"base/files/file_path.h\"\n #include \"build/build_config.h\"\n \n #if defined(OS_WIN)\n@@ -16,10 +17,6 @@\n \n class GURL;\n \n-namespace base {\n-class FilePath;\n-}\n-\n namespace platform_util {\n \n typedef base::Callback<void(const std::string&)> OpenExternalCallback;\n@@ -32,6 +29,11 @@ bool ShowItemInFolder(const base::FilePath& full_path);\n // Must be called from the UI thread.\n bool OpenItem(const base::FilePath& full_path);\n \n+struct OpenExternalOptions {\n+ bool activate = true;\n+ base::FilePath working_dir;\n+};\n+\n // Open the given external protocol URL in the desktop's default manner.\n // (For example, mailto: URLs in the default mail user agent.)\n bool OpenExternal(\n@@ -40,7 +42,7 @@ bool OpenExternal(\n #else\n const GURL& url,\n #endif\n- bool activate);\n+ const OpenExternalOptions& options);\n \n // The asynchronous version of OpenExternal.\n void OpenExternal(\n@@ -49,7 +51,7 @@ void OpenExternal(\n #else\n const GURL& url,\n #endif\n- bool activate,\n+ const OpenExternalOptions& options,\n const OpenExternalCallback& callback);\n \n // Move a file to trash.\ndiff --git a/atom/common/platform_util_linux.cc b/atom/common/platform_util_linux.cc\nindex 63ee0bd..f17cbda 100644\n--- a/atom/common/platform_util_linux.cc\n+++ b/atom/common/platform_util_linux.cc\n@@ -80,7 +80,7 @@ bool OpenItem(const base::FilePath& full_path) {\n return XDGOpen(full_path.value(), false);\n }\n \n-bool OpenExternal(const GURL& url, bool activate) {\n+bool OpenExternal(const GURL& url, const OpenExternalOptions& options) {\n // Don't wait for exit, since we don't want to wait for the browser/email\n // client window to close before returning\n if (url.SchemeIs(\"mailto\"))\n@@ -90,10 +90,10 @@ bool OpenExternal(const GURL& url, bool activate) {\n }\n \n void OpenExternal(const GURL& url,\n- bool activate,\n+ const OpenExternalOptions& options,\n const OpenExternalCallback& callback) {\n // TODO(gabriel): Implement async open if callback is specified\n- callback.Run(OpenExternal(url, activate) ? \"\" : \"Failed to open\");\n+ callback.Run(OpenExternal(url, options) ? \"\" : \"Failed to open\");\n }\n \n bool MoveItemToTrash(const base::FilePath& full_path) {\ndiff --git a/atom/common/platform_util_mac.mm b/atom/common/platform_util_mac.mm\nindex b83b1e1..4cda8bf 100644\n--- a/atom/common/platform_util_mac.mm\n+++ b/atom/common/platform_util_mac.mm\n@@ -139,16 +139,16 @@ bool OpenItem(const base::FilePath& full_path) {\n launchIdentifiers:NULL];\n }\n \n-bool OpenExternal(const GURL& url, bool activate) {\n+bool OpenExternal(const GURL& url, const OpenExternalOptions& options) {\n DCHECK([NSThread isMainThread]);\n NSURL* ns_url = net::NSURLWithGURL(url);\n if (ns_url)\n- return OpenURL(ns_url, activate).empty();\n+ return OpenURL(ns_url, options.activate).empty();\n return false;\n }\n \n void OpenExternal(const GURL& url,\n- bool activate,\n+ const OpenExternalOptions& options,\n const OpenExternalCallback& callback) {\n NSURL* ns_url = net::NSURLWithGURL(url);\n if (!ns_url) {\n@@ -157,13 +157,13 @@ void OpenExternal(const GURL& url,\n }\n \n __block OpenExternalCallback c = callback;\n- dispatch_async(dispatch_get_global_queue(DISPATCH_QUEUE_PRIORITY_DEFAULT, 0),\n- ^{\n- __block std::string error = OpenURL(ns_url, activate);\n- dispatch_async(dispatch_get_main_queue(), ^{\n- c.Run(error);\n- });\n- });\n+ dispatch_async(\n+ dispatch_get_global_queue(DISPATCH_QUEUE_PRIORITY_DEFAULT, 0), ^{\n+ __block std::string error = OpenURL(ns_url, options.activate);\n+ dispatch_async(dispatch_get_main_queue(), ^{\n+ c.Run(error);\n+ });\n+ });\n }\n \n bool MoveItemToTrash(const base::FilePath& full_path) {\ndiff --git a/atom/common/platform_util_win.cc b/atom/common/platform_util_win.cc\nindex 34576be..5712200 100644\n--- a/atom/common/platform_util_win.cc\n+++ b/atom/common/platform_util_win.cc\n@@ -294,15 +294,18 @@ bool OpenItem(const base::FilePath& full_path) {\n return ui::win::OpenFileViaShell(full_path);\n }\n \n-bool OpenExternal(const base::string16& url, bool activate) {\n+bool OpenExternal(const base::string16& url,\n+ const OpenExternalOptions& options) {\n // Quote the input scheme to be sure that the command does not have\n // parameters unexpected by the external program. This url should already\n // have been escaped.\n base::string16 escaped_url = L\"\\\"\" + url + L\"\\\"\";\n+ auto working_dir = options.working_dir.value();\n \n- if (reinterpret_cast<ULONG_PTR>(ShellExecuteW(\n- NULL, L\"open\", escaped_url.c_str(), NULL, NULL, SW_SHOWNORMAL)) <=\n- 32) {\n+ if (reinterpret_cast<ULONG_PTR>(\n+ ShellExecuteW(nullptr, L\"open\", escaped_url.c_str(), nullptr,\n+ working_dir.empty() ? nullptr : working_dir.c_str(),\n+ SW_SHOWNORMAL)) <= 32) {\n // We fail to execute the call. We could display a message to the user.\n // TODO(nsylvain): we should also add a dialog to warn on errors. See\n // bug 1136923.\n@@ -312,10 +315,10 @@ bool OpenExternal(const base::string16& url, bool activate) {\n }\n \n void OpenExternal(const base::string16& url,\n- bool activate,\n+ const OpenExternalOptions& options,\n const OpenExternalCallback& callback) {\n // TODO(gabriel): Implement async open if callback is specified\n- callback.Run(OpenExternal(url, activate) ? \"\" : \"Failed to open\");\n+ callback.Run(OpenExternal(url, options) ? \"\" : \"Failed to open\");\n }\n \n bool MoveItemToTrash(const base::FilePath& path) {\ndiff --git a/docs/api/shell.md b/docs/api/shell.md\nindex a469f94..b38348a 100644\n--- a/docs/api/shell.md\n+++ b/docs/api/shell.md\n@@ -37,9 +37,10 @@ Open the given file in the desktop's default manner.\n ### `shell.openExternal(url[, options, callback])`\n \n * `url` String - Max 2081 characters on windows, or the function returns false.\n-* `options` Object (optional) _macOS_\n- * `activate` Boolean - `true` to bring the opened application to the\n- foreground. The default is `true`.\n+* `options` Object (optional)\n+ * `activate` Boolean (optional) - `true` to bring the opened application to the\n+ foreground. The default is `true`. _macOS_\n+ * `workingDirectory` String (optional) - The working directory. _Windows_\n * `callback` Function (optional) _macOS_ - If specified will perform the open asynchronously.\n * `error` Error\n \n", "diff --git a/.travis.yml b/.travis.yml\nindex d56185e..96510cb 100644\n--- a/.travis.yml\n+++ b/.travis.yml\n@@ -2,5 +2,6 @@ language: node_js\n node_js:\n - 'stable'\n script:\n+ - yarn lint\n - yarn build\n - yarn test\n"]
2
["a9475f359061fcd6cd53557599fedf0df5e9ee00", "2ac99c0a66a1adc18ee4ef660608f814823dd198"]
["feat", "ci"]
fixed tick interval,add prewatch script to core
["diff --git a/backend/services/integrations/main.go b/backend/services/integrations/main.go\nindex 4a5e764..35c3ff2 100644\n--- a/backend/services/integrations/main.go\n+++ b/backend/services/integrations/main.go\n@@ -54,7 +54,7 @@ func main() {\n \tsigchan := make(chan os.Signal, 1)\n \tsignal.Notify(sigchan, syscall.SIGINT, syscall.SIGTERM)\n \n-\ttick := time.Tick(intervals.INTEGRATIONS_REQUEST_INTERVAL)\n+\ttick := time.Tick(intervals.INTEGRATIONS_REQUEST_INTERVAL * time.Millisecond)\n \n \tlog.Printf(\"Integration service started\\n\")\n \tmanager.RequestAll()\n@@ -66,7 +66,7 @@ func main() {\n \t\t\tpg.Close()\n \t\t\tos.Exit(0)\n \t\tcase <-tick:\n-\t\t\t// log.Printf(\"Requesting all...\\n\")\n+\t\t\tlog.Printf(\"Requesting all...\\n\")\n \t\t\tmanager.RequestAll()\n \t\tcase event := <-manager.Events:\n \t\t\t// log.Printf(\"New integration event: %v\\n\", *event.RawErrorEvent)\n", "diff --git a/.gitignore b/.gitignore\nindex 3445558..72257cb 100644\n--- a/.gitignore\n+++ b/.gitignore\n@@ -1,4 +1,5 @@\n *.o\n+.env\n settheory\n constraint\n Main\ndiff --git a/packages/core/package.json b/packages/core/package.json\nindex 9ba8b93..79bd532 100644\n--- a/packages/core/package.json\n+++ b/packages/core/package.json\n@@ -14,6 +14,7 @@\n \"build:parsers\": \"nearleyc src/parser/Domain.ne > src/parser/DomainParser.ts && nearleyc src/parser/Substance.ne > src/parser/SubstanceParser.ts && nearleyc src/parser/Style.ne > src/parser/StyleParser.ts\",\n \"prebuild\": \"yarn build:parsers\",\n \"prestart\": \"yarn build:parsers\",\n+ \"prewatch\": \"yarn build:parsers\",\n \"test\": \"jest --watchAll=false\",\n \"test:watch\": \"jest --watchAll\",\n \"build\": \"rollup -c\",\n"]
2
["7dc3b70fe40fc7de255a28bb3098bcb8c0d35365", "aa0152baa4376b1087c86499a7c289b668d5ad55"]
["fix", "build"]
retry uploading pdb files on appveyor (#21561),fixa few issues
["diff --git a/appveyor.yml b/appveyor.yml\nindex 9aca21e..8b54543 100644\n--- a/appveyor.yml\n+++ b/appveyor.yml\n@@ -146,12 +146,12 @@ build_script:\n - ps: >-\n if ($env:GN_CONFIG -eq 'release') {\n python electron\\script\\zip-symbols.py\n- appveyor PushArtifact out/Default/symbols.zip\n+ appveyor-retry appveyor PushArtifact out/Default/symbols.zip\n } else {\n # It's useful to have pdb files when debugging testing builds that are\n # built on CI.\n 7z a pdb.zip out\\Default\\*.pdb\n- appveyor PushArtifact pdb.zip\n+ appveyor-retry appveyor PushArtifact pdb.zip\n }\n - python electron/script/zip_manifests/check-zip-manifest.py out/Default/dist.zip electron/script/zip_manifests/dist_zip.win.%TARGET_ARCH%.manifest\n test_script:\n", "diff --git a/README.md b/README.md\nindex d944d22..5099f03 100644\n--- a/README.md\n+++ b/README.md\n@@ -10,9 +10,8 @@ React state management with a minimal API. Made with :heart: and ES6 Proxies.\n \n <!-- toc -->\n \n-* [Motivation](#motivation)\n+* [Introduction](#introduction)\n * [Installation](#installation)\n- + [Setting up a quick project](#setting-up-a-quick-project)\n * [Usage](#usage)\n + [Creating stores](#creating-stores)\n + [Creating reactive views](#creating-reactive-views)\n@@ -35,12 +34,14 @@ React state management with a minimal API. Made with :heart: and ES6 Proxies.\n Easy State consists of two wrapper functions only. `store` creates state stores and `view` creates reactive components, which re-render whenever state stores are mutated. The rest is just plain JavaScript.\n \n ```js\n-import React, from 'react'\n+import React from 'react'\n import { store, view } from 'react-easy-state'\n \n+// stores are normal objects\n const clock = store({ time: new Date() })\n setInterval(() => clock.time = new Date(), 1000)\n \n+// reactive components re-render on store mutations\n function ClockComp () {\n return <div>{clock.time}</div>\n }\n"]
2
["7152173d26293f4638920b17ce2dfa8ae995193b", "b8a664c1b10f4e30a3e221a14211a3cdaf90b7f4"]
["ci", "docs"]
note about archive without license and readme closes #602,add donation section to footer
["diff --git a/docs/060-archive.md b/docs/060-archive.md\nindex e5d7cfb..bf0d569 100644\n--- a/docs/060-archive.md\n+++ b/docs/060-archive.md\n@@ -85,3 +85,24 @@ Then you can run:\n ```console\n GOVERSION_NR=$(go version | awk '{print $3;}') goreleaser\n ```\n+\n+## Packaging only the binaries\n+\n+Since GoReleaser will always add the `README` and `LICENSE` files to the\n+archive if the file list is empty, you'll need to provide a filled `files`\n+on the archive section.\n+\n+A working hack is to use something like this:\n+\n+```yaml\n+# goreleaser.yml\n+archive:\n+ files:\n+ - none*\n+```\n+\n+This would add all files matching the glob `none*`, provide that you don't\n+have any files matching that glob, only the binary will be added to the\n+archive.\n+\n+For more information, check [#602](https://github.com/goreleaser/goreleaser/issues/602)\n", "diff --git a/client/src/components/Feedback.tsx b/client/src/components/Feedback.tsx\nindex 117b21d..0d7e7a9 100644\n--- a/client/src/components/Feedback.tsx\n+++ b/client/src/components/Feedback.tsx\n@@ -16,12 +16,7 @@ const publicRoutes = [\n name: `\ud83d\udcdd Feedback on RS School`,\n link: `https://docs.google.com/forms/d/1F4NeS0oBq-CY805aqiPVp6CIrl4_nIYJ7Z_vUcMOFrQ/viewform`,\n newTab: true,\n- },\n- {\n- name: `\ud83d\udcb0 Make a donation`,\n- link: `https://www.patreon.com/therollingscopes`,\n- newTab: true,\n- },\n+ }\n ];\n \n type LinkInfo = { name: string; link: string; newTab: boolean };\ndiff --git a/client/src/components/FooterLayout.tsx b/client/src/components/FooterLayout.tsx\nindex 79c0f39..56661b4 100644\n--- a/client/src/components/FooterLayout.tsx\n+++ b/client/src/components/FooterLayout.tsx\n@@ -1,5 +1,5 @@\n import * as React from 'react';\n-import { Col, Layout, Row, Divider } from 'antd';\n+import { Col, Layout, Row, Divider, Button } from 'antd';\n import { Feedback } from './Feedback';\n import { Help } from './Help';\n import { SocialNetworks } from './SocialNetworks';\n@@ -23,9 +23,17 @@ class FooterLayout extends React.Component<any, any> {\n </Col>\n </Row>\n <Divider />\n- <div className=\"text-center\">\n- <small>&copy; The Rolling Scopes 2019</small>\n- </div>\n+ <h5> Thank you for your support! \ud83c\udf89</h5>\n+ <p>\n+ <object type=\"image/svg+xml\"\n+ data=\"https://opencollective.com/rsschool/backers.svg?avatarHeight=28&button=false\"></object>\n+ </p>\n+ <p>\n+ <Button size=\"small\" href=\"https://opencollective.com/rsschool#section-contribute\" target=\"_blank\" ghost>\n+ \u2764\ufe0f Make a donation\n+ </Button>\n+ </p>\n+ <p className=\"text-center\"><small>&copy; The Rolling Scopes 2019</small></p>\n </Footer>\n </div>\n );\ndiff --git a/client/src/styles/main.scss b/client/src/styles/main.scss\nindex cd61fcd..6e37ea6 100644\n--- a/client/src/styles/main.scss\n+++ b/client/src/styles/main.scss\n@@ -46,4 +46,7 @@ body,\n padding-right: 0;\n font-size: .7rem;\n }\n+ .ant-btn {\n+ font-size: .7rem;\n+ }\n }\n"]
2
["7648760108613f771a5e6e40bb87a8f1fcee21ad", "7704121d0c0bfce49f01c2b41cbc64a642cbb399"]
["docs", "feat"]
Template using kube api version Signed-off-by: rjshrjndrn <[email protected]>,build improvements
["diff --git a/.github/workflows/api-ee.yaml b/.github/workflows/api-ee.yaml\nindex c014f34..2a12e0d 100644\n--- a/.github/workflows/api-ee.yaml\n+++ b/.github/workflows/api-ee.yaml\n@@ -8,7 +8,7 @@ on:\n default: 'false'\n push:\n branches:\n- - dev\n+ - test_ci\n paths:\n - ee/api/**\n - api/**\n@@ -112,7 +112,8 @@ jobs:\n # Deploy command\n kubectl config set-context --namespace=app --current\n kubectl config get-contexts\n- helm template openreplay -n app openreplay -f vars.yaml -f /tmp/image_override.yaml --set ingress-nginx.enabled=false --set skipMigration=true --no-hooks | kubectl apply -f -\n+ k_version=$(kubectl version --short 2>/dev/null | awk '/Server/{print $NF}')\n+ helm template openreplay -n app openreplay -f vars.yaml -f /tmp/image_override.yaml --set ingress-nginx.enabled=false --set skipMigration=true --no-hooks --kube-version=$k_version | kubectl apply -f -\n env:\n DOCKER_REPO: ${{ secrets.EE_REGISTRY_URL }}\n # We're not passing -ee flag, because helm will add that.\n", "diff --git a/.travis.yml b/.travis.yml\nindex 9e1b926..3144244 100644\n--- a/.travis.yml\n+++ b/.travis.yml\n@@ -1,5 +1,6 @@\n language: node_js\n dist: trusty\n+sudo: required\n node_js:\n - '6.9.5'\n before_install:\ndiff --git a/e2e/schematics/command-line.test.ts b/e2e/schematics/command-line.test.ts\nindex 16d8b34..ea91494 100644\n--- a/e2e/schematics/command-line.test.ts\n+++ b/e2e/schematics/command-line.test.ts\n@@ -68,8 +68,6 @@ describe('Command line', () => {\n \n updateFile('apps/myapp/src/app/app.component.spec.ts', `import '@nrwl/mylib';`);\n \n- updateRunAffectedToWorkInE2ESetup();\n-\n const affectedApps = runCommand('npm run affected:apps -- --files=\"libs/mylib/index.ts\"');\n expect(affectedApps).toContain('myapp');\n expect(affectedApps).not.toContain('myapp2');\n@@ -147,11 +145,3 @@ describe('Command line', () => {\n 1000000\n );\n });\n-\n-function updateRunAffectedToWorkInE2ESetup() {\n- const runAffected = readFile('node_modules/@nrwl/schematics/src/command-line/affected.js');\n- const newRunAffected = runAffected\n- .replace('ng build', '../../node_modules/.bin/ng build')\n- .replace('ng e2e', '../../node_modules/.bin/ng e2e');\n- updateFile('node_modules/@nrwl/schematics/src/command-line/affected.js', newRunAffected);\n-}\ndiff --git a/e2e/schematics/workspace.test.ts b/e2e/schematics/workspace.test.ts\nindex 8a41070..8749926 100644\n--- a/e2e/schematics/workspace.test.ts\n+++ b/e2e/schematics/workspace.test.ts\n@@ -82,7 +82,7 @@ describe('Nrwl Convert to Nx Workspace', () => {\n \n it('should generate a workspace and not change dependencies or devDependencies if they already exist', () => {\n // create a new AngularCLI app\n- runNgNew('--skip-install');\n+ runNgNew();\n const nxVersion = '0.0.0';\n const schematicsVersion = '0.0.0';\n const ngrxVersion = '0.0.0';\ndiff --git a/e2e/utils.ts b/e2e/utils.ts\nindex 422d866..a03104f 100644\n--- a/e2e/utils.ts\n+++ b/e2e/utils.ts\n@@ -17,8 +17,7 @@ export function newProject(): void {\n copyMissingPackages();\n execSync('mv ./tmp/proj ./tmp/proj_backup');\n }\n- execSync('cp -r ./tmp/proj_backup ./tmp/proj');\n- setUpSynLink();\n+ execSync('cp -a ./tmp/proj_backup ./tmp/proj');\n }\n \n export function copyMissingPackages(): void {\n@@ -26,14 +25,9 @@ export function copyMissingPackages(): void {\n modulesToCopy.forEach(m => copyNodeModule(projectName, m));\n }\n \n-export function setUpSynLink(): void {\n- execSync(`ln -s ../@nrwl/schematics/src/command-line/nx.js tmp/${projectName}/node_modules/.bin/nx`);\n- execSync(`chmod +x tmp/${projectName}/node_modules/.bin/nx`);\n-}\n-\n function copyNodeModule(path: string, name: string) {\n execSync(`rm -rf tmp/${path}/node_modules/${name}`);\n- execSync(`cp -r node_modules/${name} tmp/${path}/node_modules/${name}`);\n+ execSync(`cp -a node_modules/${name} tmp/${path}/node_modules/${name}`);\n }\n \n export function runCLI(\n@@ -43,7 +37,7 @@ export function runCLI(\n }\n ): string {\n try {\n- return execSync(`../../node_modules/.bin/ng ${command}`, {\n+ return execSync(`./node_modules/.bin/ng ${command}`, {\n cwd: `./tmp/${projectName}`\n })\n .toString()\n@@ -67,7 +61,7 @@ export function newLib(name: string): string {\n }\n \n export function runSchematic(command: string): string {\n- return execSync(`../../node_modules/.bin/schematics ${command}`, {\n+ return execSync(`./node_modules/.bin/schematics ${command}`, {\n cwd: `./tmp/${projectName}`\n }).toString();\n }\ndiff --git a/package.json b/package.json\nindex bef54f8..9186a58 100644\n--- a/package.json\n+++ b/package.json\n@@ -6,7 +6,7 @@\n \"private\": true,\n \"scripts\": {\n \"build\": \"./scripts/build.sh\",\n- \"e2e\": \"yarn build && ./scripts/e2e.sh\",\n+ \"e2e\": \"./scripts/e2e.sh\",\n \"format\": \"./scripts/format.sh\",\n \"linknpm\": \"./scripts/link.sh\",\n \"package\": \"./scripts/package.sh\",\n@@ -14,7 +14,7 @@\n \"copy\": \"./scripts/copy.sh\",\n \"test:schematics\": \"yarn build && ./scripts/test_schematics.sh\",\n \"test:nx\": \"yarn build && ./scripts/test_nx.sh\",\n- \"test\": \"yarn build && ./scripts/test_nx.sh && ./scripts/test_schematics.sh\",\n+ \"test\": \"yarn linknpm && ./scripts/test_nx.sh && ./scripts/test_schematics.sh\",\n \"checkformat\": \"./scripts/check-format.sh\",\n \"publish_npm\": \"./scripts/publish.sh\"\n },\ndiff --git a/packages/schematics/src/collection/workspace/index.ts b/packages/schematics/src/collection/workspace/index.ts\nindex 8f8897f..c70d161 100644\n--- a/packages/schematics/src/collection/workspace/index.ts\n+++ b/packages/schematics/src/collection/workspace/index.ts\n@@ -254,20 +254,7 @@ function moveFiles(options: Schema) {\n \n function copyAngularCliTgz() {\n return (host: Tree) => {\n- copyFile(\n- path.join(\n- 'node_modules',\n- '@nrwl',\n- 'schematics',\n- 'src',\n- 'collection',\n- 'application',\n- 'files',\n- '__directory__',\n- '.angular_cli.tgz'\n- ),\n- '.'\n- );\n+ copyFile(path.join(__dirname, '..', 'application', 'files', '__directory__', '.angular_cli.tgz'), '.');\n return host;\n };\n }\ndiff --git a/packages/schematics/src/command-line/affected.ts b/packages/schematics/src/command-line/affected.ts\nindex b7f9173..89a4f72 100644\n--- a/packages/schematics/src/command-line/affected.ts\n+++ b/packages/schematics/src/command-line/affected.ts\n@@ -1,5 +1,7 @@\n import { execSync } from 'child_process';\n import { getAffectedApps, parseFiles } from './shared';\n+import * as path from 'path';\n+import * as resolve from 'resolve';\n \n export function affected(args: string[]): void {\n const command = args[0];\n@@ -39,7 +41,7 @@ function build(apps: string[], rest: string[]) {\n if (apps.length > 0) {\n console.log(`Building ${apps.join(', ')}`);\n apps.forEach(app => {\n- execSync(`ng build ${rest.join(' ')} -a=${app}`, { stdio: [0, 1, 2] });\n+ execSync(`${ngPath()} build ${rest.join(' ')} -a=${app}`, { stdio: [0, 1, 2] });\n });\n } else {\n console.log('No apps to build');\n@@ -50,9 +52,13 @@ function e2e(apps: string[], rest: string[]) {\n if (apps.length > 0) {\n console.log(`Testing ${apps.join(', ')}`);\n apps.forEach(app => {\n- execSync(`ng e2e ${rest.join(' ')} -a=${app}`, { stdio: [0, 1, 2] });\n+ execSync(`${ngPath()} e2e ${rest.join(' ')} -a=${app}`, { stdio: [0, 1, 2] });\n });\n } else {\n- console.log('No apps to tst');\n+ console.log('No apps to test');\n }\n }\n+\n+function ngPath() {\n+ return `${path.dirname(path.dirname(path.dirname(resolve.sync('@angular/cli', { basedir: __dirname }))))}/bin/ng`;\n+}\ndiff --git a/scripts/build.sh b/scripts/build.sh\nindex ac533b5..9b8891b 100755\n--- a/scripts/build.sh\n+++ b/scripts/build.sh\n@@ -3,6 +3,8 @@\n rm -rf build\n ngc\n rsync -a --exclude=*.ts packages/ build/packages\n+chmod +x build/packages/schematics/bin/create-nx-workspace.js\n+chmod +x build/packages/schematics/src/command-line/nx.js\n rm -rf build/packages/install\n cp README.md build/packages/schematics\n cp README.md build/packages/nx\n\\ No newline at end of file\n"]
2
["c3531347fe5a4cc82d426db195026a5bdad15e7a", "e0a977b2d316e7612b5d72cb02cd7d78e75dbc55"]
["ci", "build"]
Fix Cube Store build on Windows,export order
["diff --git a/.github/workflows/publish.yml b/.github/workflows/publish.yml\nindex a986501..db33097 100644\n--- a/.github/workflows/publish.yml\n+++ b/.github/workflows/publish.yml\n@@ -409,7 +409,7 @@ jobs:\n run: vcpkg integrate install; vcpkg install openssl:x64-windows\n - name: Instal LLVM for Windows\n if: ${{ startsWith(matrix.os, 'windows') }}\n- run: choco install -y llvm --version 9.0.1\n+ run: choco install -y --force llvm --version 9.0.1\n - name: Set Env Variables for Windows\n uses: allenevans/[email protected]\n if: ${{ startsWith(matrix.os, 'windows') }}\ndiff --git a/.github/workflows/rust.yml b/.github/workflows/rust.yml\nindex d45cdf9..8d87ed6 100644\n--- a/.github/workflows/rust.yml\n+++ b/.github/workflows/rust.yml\n@@ -158,7 +158,7 @@ jobs:\n run: vcpkg integrate install; vcpkg install openssl:x64-windows\n - name: Instal LLVM for Windows\n if: ${{ startsWith(matrix.os, 'windows') }}\n- run: choco install -y llvm --version 9.0.1\n+ run: choco install -y --force llvm --version 9.0.1\n - name: Set Env Variables for Windows\n uses: allenevans/[email protected]\n if: ${{ startsWith(matrix.os, 'windows') }}\n", "diff --git a/src/Object/_api.ts b/src/Object/_api.ts\nindex a1e5c8c..70e9fea 100644\n--- a/src/Object/_api.ts\n+++ b/src/Object/_api.ts\n@@ -38,9 +38,9 @@ export {Optional} from './Optional'\n export {OptionalKeys} from './OptionalKeys'\n export {Overwrite} from './Overwrite'\n export {Partial} from './Partial'\n+export {Path} from './Path'\n export {Paths} from './Paths'\n export {PathValid} from './PathValid'\n-export {Path} from './Path'\n export {Pick} from './Pick'\n export {Readonly} from './Readonly'\n export {ReadonlyKeys} from './ReadonlyKeys'\n"]
2
["e34bb04baac7574e67bc566d13dea72092e0cfa3", "879edb6ed90f88b9ae6a3c2e8878ae1be48e0c88"]
["ci", "refactor"]
101: fix import key cmd Signed-off-by: Sam Alba <[email protected]>,brew tests/multiple darwin builds/gh enterprise
["diff --git a/docs/learn/101-use.md b/docs/learn/101-use.md\nindex 283c1c1..2ec10f9 100644\n--- a/docs/learn/101-use.md\n+++ b/docs/learn/101-use.md\n@@ -41,8 +41,7 @@ cd ./examples/todoapp\n The example app contains encrypted secrets and other pre-configured inputs, here is how to decrypt them:\n \n ```sh\n-curl -sfL https://releases.dagger.io/examples/key.txt >> ~/.config/dagger/keys.txt\n-dagger input list\n+dagger input list || curl -sfL https://releases.dagger.io/examples/key.txt >> ~/.config/dagger/keys.txt\n ```\n \n **Step 4**: Deploy!\n", "diff --git a/pipeline/brew/brew.go b/pipeline/brew/brew.go\nindex ec27182..15ed189 100644\n--- a/pipeline/brew/brew.go\n+++ b/pipeline/brew/brew.go\n@@ -1,5 +1,3 @@\n-// Package brew implements the Pipe, providing formula generation and\n-// uploading it to a configured repo.\n package brew\n \n import (\n@@ -10,13 +8,12 @@ import (\n \t\"strings\"\n \t\"text/template\"\n \n-\t\"github.com/goreleaser/goreleaser/internal/artifact\"\n-\n \t\"github.com/apex/log\"\n \n \t\"github.com/goreleaser/goreleaser/checksum\"\n \t\"github.com/goreleaser/goreleaser/config\"\n \t\"github.com/goreleaser/goreleaser/context\"\n+\t\"github.com/goreleaser/goreleaser/internal/artifact\"\n \t\"github.com/goreleaser/goreleaser/internal/client\"\n \t\"github.com/goreleaser/goreleaser/pipeline\"\n )\n@@ -106,14 +103,14 @@ func doRun(ctx *context.Context, client client.Client) error {\n \t\tartifact.And(\n \t\t\tartifact.ByGoos(\"darwin\"),\n \t\t\tartifact.ByGoarch(\"amd64\"),\n-\t\t\tartifact.ByGoarch(\"\"),\n+\t\t\tartifact.ByGoarm(\"\"),\n \t\t\tartifact.ByType(artifact.UploadableArchive),\n \t\t),\n \t).List()\n \tif len(archives) == 0 {\n \t\treturn ErrNoDarwin64Build\n \t}\n-\tif len(archives) > 0 {\n+\tif len(archives) > 1 {\n \t\treturn ErrTooManyDarwin64Builds\n \t}\n \tvar path = filepath.Join(ctx.Config.Brew.Folder, ctx.Config.ProjectName+\".rb\")\n@@ -145,8 +142,7 @@ func doBuildFormula(data templateData) (out bytes.Buffer, err error) {\n }\n \n func dataFor(ctx *context.Context, client client.Client, artifact artifact.Artifact) (result templateData, err error) {\n-\tvar file = artifact.Path\n-\tsum, err := checksum.SHA256(file)\n+\tsum, err := checksum.SHA256(artifact.Path)\n \tif err != nil {\n \t\treturn\n \t}\n@@ -163,7 +159,7 @@ func dataFor(ctx *context.Context, client client.Client, artifact artifact.Artif\n \t\tTag: ctx.Git.CurrentTag,\n \t\tVersion: ctx.Version,\n \t\tCaveats: ctx.Config.Brew.Caveats,\n-\t\tFile: file,\n+\t\tFile: artifact.Name,\n \t\tSHA256: sum,\n \t\tDependencies: ctx.Config.Brew.Dependencies,\n \t\tConflicts: ctx.Config.Brew.Conflicts,\ndiff --git a/pipeline/brew/brew_test.go b/pipeline/brew/brew_test.go\nindex 7e513bf..9066935 100644\n--- a/pipeline/brew/brew_test.go\n+++ b/pipeline/brew/brew_test.go\n@@ -9,6 +9,7 @@ import (\n \n \t\"github.com/goreleaser/goreleaser/config\"\n \t\"github.com/goreleaser/goreleaser/context\"\n+\t\"github.com/goreleaser/goreleaser/internal/artifact\"\n \t\"github.com/goreleaser/goreleaser/internal/testlib\"\n \t\"github.com/stretchr/testify/assert\"\n )\n@@ -93,7 +94,8 @@ func TestRunPipe(t *testing.T) {\n \t\tGit: context.GitInfo{\n \t\t\tCurrentTag: \"v1.0.1\",\n \t\t},\n-\t\tVersion: \"1.0.1\",\n+\t\tVersion: \"1.0.1\",\n+\t\tArtifacts: artifact.New(),\n \t\tConfig: config.Project{\n \t\t\tDist: folder,\n \t\t\tProjectName: \"run-pipe\",\n@@ -124,31 +126,53 @@ func TestRunPipe(t *testing.T) {\n \t\tPublish: true,\n \t}\n \tvar path = filepath.Join(folder, \"bin.tar.gz\")\n-\tctx.AddBinary(\"darwinamd64\", \"bin\", \"bin\", path)\n+\tctx.Artifacts.Add(artifact.Artifact{\n+\t\tName: \"bin.tar.gz\",\n+\t\tPath: path,\n+\t\tGoos: \"darwin\",\n+\t\tGoarch: \"amd64\",\n+\t\tType: artifact.UploadableArchive,\n+\t})\n \tclient := &DummyClient{}\n \tassert.Error(t, doRun(ctx, client))\n \tassert.False(t, client.CreatedFile)\n \n \t_, err = os.Create(path)\n \tassert.NoError(t, err)\n-\tassert.NoError(t, doRun(ctx, client))\n-\tassert.True(t, client.CreatedFile)\n \n-\tbts, err := ioutil.ReadFile(\"testdata/run_pipe.rb\")\n-\tassert.NoError(t, err)\n-\t// ioutil.WriteFile(\"testdata/run_pipe.rb\", []byte(client.Content), 0644)\n+\tt.Run(\"default git url\", func(tt *testing.T) {\n+\t\tassert.NoError(tt, doRun(ctx, client))\n+\t\tassert.True(tt, client.CreatedFile)\n+\n+\t\tbts, err := ioutil.ReadFile(\"testdata/run_pipe.rb\")\n+\t\tassert.NoError(tt, err)\n+\t\t// TODO: make writing this file toggleable somehow?\n+\t\t// ioutil.WriteFile(\"testdata/run_pipe.rb\", []byte(client.Content), 0644)\n+\t\tassert.Equal(tt, string(bts), client.Content)\n+\t})\n \n-\tassert.Equal(t, string(bts), client.Content)\n+\tt.Run(\"github enterprise url\", func(tt *testing.T) {\n+\t\tctx.Config.GitHubURLs.Download = \"http://github.example.org\"\n+\t\tassert.NoError(tt, doRun(ctx, client))\n+\t\tassert.True(tt, client.CreatedFile)\n+\n+\t\tbts, err := ioutil.ReadFile(\"testdata/run_pipe_enterprise.rb\")\n+\t\tassert.NoError(tt, err)\n+\t\t// TODO: make writing this file toggleable somehow?\n+\t\t// ioutil.WriteFile(\"testdata/run_pipe_enterprise.rb\", []byte(client.Content), 0644)\n+\t\tassert.Equal(tt, string(bts), client.Content)\n+\t})\n }\n \n+// TODO: this test is irrelevant and can probavly be removed\n func TestRunPipeFormatOverride(t *testing.T) {\n \tfolder, err := ioutil.TempDir(\"\", \"goreleasertest\")\n \tassert.NoError(t, err)\n \tvar path = filepath.Join(folder, \"bin.zip\")\n \t_, err = os.Create(path)\n \tassert.NoError(t, err)\n-\tvar ctx = &context.Context{\n-\t\tConfig: config.Project{\n+\tvar ctx = context.New(\n+\t\tconfig.Project{\n \t\t\tDist: folder,\n \t\t\tArchive: config.Archive{\n \t\t\t\tFormat: \"tar.gz\",\n@@ -166,9 +190,15 @@ func TestRunPipeFormatOverride(t *testing.T) {\n \t\t\t\t},\n \t\t\t},\n \t\t},\n-\t\tPublish: true,\n-\t}\n-\tctx.AddBinary(\"darwinamd64\", \"bin\", \"bin\", path)\n+\t)\n+\tctx.Publish = true\n+\tctx.Artifacts.Add(artifact.Artifact{\n+\t\tName: \"bin.zip\",\n+\t\tPath: path,\n+\t\tGoos: \"darwin\",\n+\t\tGoarch: \"amd64\",\n+\t\tType: artifact.UploadableArchive,\n+\t})\n \tclient := &DummyClient{}\n \tassert.NoError(t, doRun(ctx, client))\n \tassert.True(t, client.CreatedFile)\n@@ -195,6 +225,40 @@ func TestRunPipeNoDarwin64Build(t *testing.T) {\n \tassert.False(t, client.CreatedFile)\n }\n \n+func TestRunPipeMultipleDarwin64Build(t *testing.T) {\n+\tvar ctx = context.New(\n+\t\tconfig.Project{\n+\t\t\tArchive: config.Archive{\n+\t\t\t\tFormat: \"tar.gz\",\n+\t\t\t},\n+\t\t\tBrew: config.Homebrew{\n+\t\t\t\tGitHub: config.Repo{\n+\t\t\t\t\tOwner: \"test\",\n+\t\t\t\t\tName: \"test\",\n+\t\t\t\t},\n+\t\t\t},\n+\t\t},\n+\t)\n+\tctx.Publish = true\n+\tctx.Artifacts.Add(artifact.Artifact{\n+\t\tName: \"bin1\",\n+\t\tPath: \"doesnt mather\",\n+\t\tGoos: \"darwin\",\n+\t\tGoarch: \"amd64\",\n+\t\tType: artifact.UploadableArchive,\n+\t})\n+\tctx.Artifacts.Add(artifact.Artifact{\n+\t\tName: \"bin2\",\n+\t\tPath: \"doesnt mather\",\n+\t\tGoos: \"darwin\",\n+\t\tGoarch: \"amd64\",\n+\t\tType: artifact.UploadableArchive,\n+\t})\n+\tclient := &DummyClient{}\n+\tassert.Equal(t, ErrTooManyDarwin64Builds, doRun(ctx, client))\n+\tassert.False(t, client.CreatedFile)\n+}\n+\n func TestRunPipeBrewNotSetup(t *testing.T) {\n \tvar ctx = &context.Context{\n \t\tConfig: config.Project{},\n@@ -206,9 +270,8 @@ func TestRunPipeBrewNotSetup(t *testing.T) {\n }\n \n func TestRunPipeBinaryRelease(t *testing.T) {\n-\tvar ctx = &context.Context{\n-\t\tPublish: true,\n-\t\tConfig: config.Project{\n+\tvar ctx = context.New(\n+\t\tconfig.Project{\n \t\t\tArchive: config.Archive{\n \t\t\t\tFormat: \"binary\",\n \t\t\t},\n@@ -219,8 +282,15 @@ func TestRunPipeBinaryRelease(t *testing.T) {\n \t\t\t\t},\n \t\t\t},\n \t\t},\n-\t}\n-\tctx.AddBinary(\"darwinamd64\", \"foo\", \"bar\", \"baz\")\n+\t)\n+\tctx.Publish = true\n+\tctx.Artifacts.Add(artifact.Artifact{\n+\t\tName: \"bin\",\n+\t\tPath: \"doesnt mather\",\n+\t\tGoos: \"darwin\",\n+\t\tGoarch: \"amd64\",\n+\t\tType: artifact.Binary,\n+\t})\n \tclient := &DummyClient{}\n \ttestlib.AssertSkipped(t, doRun(ctx, client))\n \tassert.False(t, client.CreatedFile)\ndiff --git a/pipeline/brew/doc.go b/pipeline/brew/doc.go\nnew file mode 100644\nindex 0000000..2cddc12\n--- /dev/null\n+++ b/pipeline/brew/doc.go\n@@ -0,0 +1,3 @@\n+// Package brew implements the Pipe, providing formula generation and\n+// uploading it to a configured repo.\n+package brew\ndiff --git a/pipeline/brew/testdata/run_pipe_enterprise.rb b/pipeline/brew/testdata/run_pipe_enterprise.rb\nnew file mode 100644\nindex 0000000..4b24ce0\n--- /dev/null\n+++ b/pipeline/brew/testdata/run_pipe_enterprise.rb\n@@ -0,0 +1,33 @@\n+class RunPipe < Formula\n+ desc \"A run pipe test formula\"\n+ homepage \"https://github.com/goreleaser\"\n+ url \"http://github.example.org/test/test/releases/download/v1.0.1/bin.tar.gz\"\n+ version \"1.0.1\"\n+ sha256 \"e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855\"\n+ \n+ depends_on \"zsh\"\n+ depends_on \"bash\"\n+ \n+ conflicts_with \"gtk+\"\n+ conflicts_with \"qt\"\n+\n+ def install\n+ bin.install \"foo\"\n+ end\n+\n+ def caveats\n+ \"don't do this\"\n+ end\n+\n+ plist_options :startup => false\n+\n+ def plist; <<-EOS.undent\n+ <xml>whatever</xml>\n+ EOS\n+ end\n+\n+ test do\n+ system \"true\"\n+ system \"#{bin}/foo -h\"\n+ end\n+end\n"]
2
["2b01808ec86fe9d8b4a93141a1b7f95e11fd6010", "f433bcb59c36571e22d4e86c612e0a6a52f73c09"]
["docs", "feat"]
rename step,serialize access to StreamObserver
["diff --git a/.github/workflows/ci.yml b/.github/workflows/ci.yml\nindex e81d897..5c3ee6b 100644\n--- a/.github/workflows/ci.yml\n+++ b/.github/workflows/ci.yml\n@@ -45,7 +45,7 @@ jobs:\n - name: Install dependencies\n run: pnpm install\n \n- - name: Typecheck\n+ - name: Build (stub)\n run: pnpm build:stub\n \n - name: Typecheck\n", "diff --git a/gateway/src/main/java/io/camunda/zeebe/gateway/impl/stream/ClientStreamAdapter.java b/gateway/src/main/java/io/camunda/zeebe/gateway/impl/stream/ClientStreamAdapter.java\nindex ae2b1c0..8ed64e5 100644\n--- a/gateway/src/main/java/io/camunda/zeebe/gateway/impl/stream/ClientStreamAdapter.java\n+++ b/gateway/src/main/java/io/camunda/zeebe/gateway/impl/stream/ClientStreamAdapter.java\n@@ -22,6 +22,7 @@ import io.camunda.zeebe.transport.stream.api.ClientStreamer;\n import io.camunda.zeebe.util.VisibleForTesting;\n import io.grpc.Status;\n import io.grpc.StatusRuntimeException;\n+import io.grpc.internal.SerializingExecutor;\n import io.grpc.stub.ServerCallStreamObserver;\n import io.grpc.stub.StreamObserver;\n import java.util.concurrent.CompletableFuture;\n@@ -83,12 +84,12 @@ public class ClientStreamAdapter {\n @VisibleForTesting(\"Allow unit testing behavior job handling behavior\")\n static final class ClientStreamConsumerImpl implements ClientStreamConsumer {\n private final StreamObserver<ActivatedJob> responseObserver;\n- private final Executor executor;\n+ private final SerializingExecutor executor;\n \n public ClientStreamConsumerImpl(\n final StreamObserver<ActivatedJob> responseObserver, final Executor executor) {\n this.responseObserver = responseObserver;\n- this.executor = executor;\n+ this.executor = new SerializingExecutor(executor);\n }\n \n @Override\n"]
2
["34875bc0e59b43d9041903101c823d25ec194a21", "22044d58302513f5cf22b06151c4a367bbb88f6e"]
["ci", "fix"]
fix test Write another record so the commit position is updated and we can take a snapshot,explain `ChunkOrder` query test scenario
["diff --git a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/clustering/ReaderCloseTest.java b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/clustering/ReaderCloseTest.java\nindex 24f1316..881c727 100644\n--- a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/clustering/ReaderCloseTest.java\n+++ b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/clustering/ReaderCloseTest.java\n@@ -70,6 +70,14 @@ public class ReaderCloseTest {\n .getCluster()\n .getNodeId();\n clusteringRule.forceClusterToHaveNewLeader(followerId);\n+ // because of https://github.com/camunda-cloud/zeebe/issues/8329\n+ // we need to add another record so we can do a snapshot\n+ clientRule\n+ .getClient()\n+ .newPublishMessageCommand()\n+ .messageName(\"test\")\n+ .correlationKey(\"test\")\n+ .send();\n \n // when\n clusteringRule.triggerAndWaitForSnapshots();\n@@ -78,6 +86,7 @@ public class ReaderCloseTest {\n for (final Broker broker : clusteringRule.getBrokers()) {\n assertThatFilesOfDeletedSegmentsDoesNotExist(broker);\n }\n+ assertThat(leaderId).isNotEqualTo(clusteringRule.getLeaderForPartition(1).getNodeId());\n }\n \n private void assertThatFilesOfDeletedSegmentsDoesNotExist(final Broker leader)\n", "diff --git a/query_tests/src/scenarios.rs b/query_tests/src/scenarios.rs\nindex f0e352b..86df0e9 100644\n--- a/query_tests/src/scenarios.rs\n+++ b/query_tests/src/scenarios.rs\n@@ -1170,6 +1170,21 @@ impl DbSetup for ChunkOrder {\n .clear_lifecycle_action()\n .unwrap();\n \n+ // Now we have the the following chunks (same partition and table):\n+ //\n+ // | ID | order | tag: region | field: user | time |\n+ // | -- | ----- | ----------- | ----------- | ---- |\n+ // | 1 | 1 | \"west\" | 2 | 100 |\n+ // | 2 | 0 | \"west\" | 1 | 100 |\n+ //\n+ // The result after deduplication should be:\n+ //\n+ // | tag: region | field: user | time |\n+ // | ----------- | ----------- | ---- |\n+ // | \"west\" | 2 | 100 |\n+ //\n+ // So the query engine must use `order` as a primary key to sort chunks, NOT `id`.\n+\n let scenario = DbScenario {\n scenario_name: \"chunks where chunk ID alone cannot be used for ordering\".into(),\n db,\n"]
2
["47df74d40becf915a9d89cdb887abd259b77def0", "9a60af7fa3b480e2e04bacd646112cad9aaab6d7"]
["test", "docs"]
Fix windows build,fix golden tests for aws_vpn_connection
["diff --git a/src/fs/mounts/mod.rs b/src/fs/mounts/mod.rs\nindex a7f8188..662e2f5 100644\n--- a/src/fs/mounts/mod.rs\n+++ b/src/fs/mounts/mod.rs\n@@ -29,11 +29,14 @@ impl std::error::Error for Error {}\n \n impl std::fmt::Display for Error {\n fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {\n+ // Allow unreachable_patterns for windows build\n+ #[allow(unreachable_patterns)]\n match self {\n #[cfg(target_os = \"macos\")]\n Error::GetFSStatError(err) => write!(f, \"getfsstat failed: {err}\"),\n #[cfg(target_os = \"linux\")]\n- Error::IOError(err) => write!(f, \"failed to read /proc/mounts: {err}\")\n+ Error::IOError(err) => write!(f, \"failed to read /proc/mounts: {err}\"),\n+ _ => write!(f, \"Unknown error\"),\n }\n }\n }\n\\ No newline at end of file\ndiff --git a/src/main.rs b/src/main.rs\nindex 483e14d..ca28081 100644\n--- a/src/main.rs\n+++ b/src/main.rs\n@@ -62,6 +62,8 @@ mod theme;\n // to `clap` is complete.\n lazy_static! {\n static ref ALL_MOUNTS: HashMap<PathBuf, mounts::MountedFs> = {\n+ // Allow unused_mut for windows\n+ #[allow(unused_mut)]\n let mut mount_map: HashMap<PathBuf, mounts::MountedFs> = HashMap::new();\n \n #[cfg(any(target_os = \"linux\", target_os = \"macos\"))]\n", "diff --git a/internal/providers/terraform/aws/testdata/vpn_connection_test/vpn_connection_test.tf b/internal/providers/terraform/aws/testdata/vpn_connection_test/vpn_connection_test.tf\nindex d895677..cf10e3f 100644\n--- a/internal/providers/terraform/aws/testdata/vpn_connection_test/vpn_connection_test.tf\n+++ b/internal/providers/terraform/aws/testdata/vpn_connection_test/vpn_connection_test.tf\n@@ -12,6 +12,7 @@ provider \"aws\" {\n resource \"aws_vpn_connection\" \"vpn_connection\" {\n customer_gateway_id = \"dummy-customer-gateway-id\"\n type = \"ipsec.1\"\n+ vpn_gateway_id = \"vpn-gateway-id\"\n }\n \n resource \"aws_vpn_connection\" \"transit\" {\n@@ -23,10 +24,11 @@ resource \"aws_vpn_connection\" \"transit\" {\n resource \"aws_vpn_connection\" \"vpn_connection_withUsage\" {\n customer_gateway_id = \"dummy-customer-gateway-id2\"\n type = \"ipsec.1\"\n+ vpn_gateway_id = \"vpn-gateway-id\"\n }\n \n resource \"aws_vpn_connection\" \"transit_withUsage\" {\n customer_gateway_id = \"dummy-customer-gateway-id2\"\n type = \"ipsec.1\"\n transit_gateway_id = \"dummy-transit-gateway-id2\"\n-}\n\\ No newline at end of file\n+}\n"]
2
["81ca000c6a7e7435809081c60be37dda23458ec8", "9b059dd8245e72f0bf8c40fc633f9ef6fccae405"]
["build", "test"]
replace tuple with record,correctly read new last flushed index
["diff --git a/engine/src/main/java/io/camunda/zeebe/engine/processing/processinstance/CreateProcessInstanceProcessor.java b/engine/src/main/java/io/camunda/zeebe/engine/processing/processinstance/CreateProcessInstanceProcessor.java\nindex fa6f8d4..2185b1e 100644\n--- a/engine/src/main/java/io/camunda/zeebe/engine/processing/processinstance/CreateProcessInstanceProcessor.java\n+++ b/engine/src/main/java/io/camunda/zeebe/engine/processing/processinstance/CreateProcessInstanceProcessor.java\n@@ -37,7 +37,6 @@ import io.camunda.zeebe.protocol.record.intent.ProcessInstanceCreationIntent;\n import io.camunda.zeebe.protocol.record.intent.ProcessInstanceIntent;\n import io.camunda.zeebe.protocol.record.value.BpmnElementType;\n import io.camunda.zeebe.util.Either;\n-import io.camunda.zeebe.util.collection.Tuple;\n import java.util.Arrays;\n import java.util.HashMap;\n import java.util.Map;\n@@ -236,21 +235,22 @@ public final class CreateProcessInstanceProcessor\n return startInstructions.stream()\n .map(\n instruction ->\n- Tuple.of(\n+ new ElementIdAndType(\n instruction.getElementId(),\n process.getElementById(instruction.getElementIdBuffer()).getElementType()))\n- .filter(elementTuple -> UNSUPPORTED_ELEMENT_TYPES.contains(elementTuple.getRight()))\n+ .filter(\n+ elementIdAndType -> UNSUPPORTED_ELEMENT_TYPES.contains(elementIdAndType.elementType))\n .findAny()\n .map(\n- elementTypeTuple ->\n+ elementIdAndType ->\n Either.left(\n new Rejection(\n RejectionType.INVALID_ARGUMENT,\n (\"Expected to create instance of process with start instructions but the element with id '%s' targets unsupported element type '%s'. \"\n + \"Supported element types are: %s\")\n .formatted(\n- elementTypeTuple.getLeft(),\n- elementTypeTuple.getRight(),\n+ elementIdAndType.elementId,\n+ elementIdAndType.elementType,\n Arrays.stream(BpmnElementType.values())\n .filter(\n elementType ->\n@@ -493,4 +493,6 @@ public final class CreateProcessInstanceProcessor\n }\n \n record Rejection(RejectionType type, String reason) {}\n+\n+ record ElementIdAndType(String elementId, BpmnElementType elementType) {}\n }\n", "diff --git a/restore/src/main/java/io/camunda/zeebe/restore/PartitionRestoreService.java b/restore/src/main/java/io/camunda/zeebe/restore/PartitionRestoreService.java\nindex 69b06b6..a4fcb77 100644\n--- a/restore/src/main/java/io/camunda/zeebe/restore/PartitionRestoreService.java\n+++ b/restore/src/main/java/io/camunda/zeebe/restore/PartitionRestoreService.java\n@@ -112,7 +112,7 @@ public class PartitionRestoreService {\n SegmentedJournal.builder()\n .withDirectory(dataDirectory.toFile())\n .withName(partition.name())\n- .withLastWrittenIndex(-1)\n+ .withLastFlushedIndex(-1)\n .build()) {\n \n resetJournal(checkpointPosition, journal);\n"]
2
["bb2ccc1a778452aebf233cf78b20f1f4bab4354b", "5ffc5794808647de14f945141692be26ad143006"]
["refactor", "fix"]