message
stringlengths
6
474
diff
stringlengths
8
5.22k
avois warning on library build
@@ -44,7 +44,7 @@ DEP_LIB := $(OBJ_LIB:.o=.d) LST_LIB := $(SRC_LIB_C:.c=.lst) INCS_LIB := -I$(INCLUDE_LIB) -I$(SRC_LIB) -I$(RES_LIB) -DEFAULT_FLAGS_LIB := -m68000 -Wall -Wextra -Wno-shift-negative-value -fno-builtin $(INCS_LIB) -B$(BIN) +DEFAULT_FLAGS_LIB := -m68000 -Wall -Wextra -Wno-shift-negative-value -fno-builtin -Wno-unused-parameter $(INCS_LIB) -B$(BIN) FLAGSZ80_LIB := -i$(SRC_LIB) -i$(INCLUDE_LIB)
add timer to metric types
@@ -46,6 +46,7 @@ func (mt MetricType) String() string { var metricTypesToCodes map[string]MetricType = map[string]MetricType{ "counter": Count, "gauge": Gauge, + "timer": Timer, } // Reader reads dogstatsd metrics from a file
driver/usb_mux/pi3usb3x532.h: Format with clang-format BRANCH=none TEST=none
* dp0-1 : rx1, tx1 * hpd+/-: rfu2, rfu1 */ -#define PI3USB3X532_MODE_DP_USB_SWAP (PI3USB3X532_MODE_DP_USB | \ - PI3USB3X532_BIT_SWAP) +#define PI3USB3X532_MODE_DP_USB_SWAP \ + (PI3USB3X532_MODE_DP_USB | PI3USB3X532_BIT_SWAP) /* Get Vendor ID */ int pi3usb3x532_check_vendor(const struct usb_mux *me, int *val);
Add module feature in KConfig.
@@ -118,14 +118,11 @@ menu "Memory Management" if RT_USING_HEAP - choice USING_HEAP - config RT_USING_SMALL_MEM bool "The memory management for small memory" config RT_USING_SLAB bool "Using SLAB memory management for large memory" - endchoice endif @@ -153,4 +150,8 @@ menu "Kernel Device Object" endmenu +config RT_USING_MODULE + bool "The dynamic module feature" + default n + endmenu
btc: fix issue with channel-js part 2
@@ -3,7 +3,7 @@ import { store } from './store'; export class Subscription { start() { - if (api.authTokens) { + if (api.ship) { this.initializeBtcWallet(); this.initializeCurrencyPoll(); } else { @@ -12,7 +12,8 @@ export class Subscription { } initializeBtcWallet() { - api.bind('/all', 'PUT', api.authTokens.ship, 'btc-wallet', + console.log('initialize'); + api.bind('/all', 'PUT', api.ship, 'btc-wallet', this.handleEvent.bind(this), this.handleError.bind(this)); }
update requirements for using appscope
@@ -38,19 +38,13 @@ AppScope 1.2, Cribl Stream 4.0, Cribl Edge 4.0, and Cribl Search 1.0 are mutuall ### Known Limitations -**Only** these runtimes are supported: - -- Open JVM 7 and later, Oracle JVM 7 and later, go1.9 through go1.19. - AppScope cannot: - Unload the libscope library, once loaded. - Instrument static executables that are not written in Go. - Instrument Go executables on ARM. -- Attach to any static application. +- Instrument Go executables built with go 1.8 and earlier. +- Instrument static stripped Go executables built with go1.12 and earlier. +- Instrument Java executables that use Open JVM 6 and earlier, Oracle JVM 6 and earlier. When an executable that's being scoped has been [stripped](https://en.wikipedia.org/wiki/Strip_(Unix)), it is not possible for `libscope.so` to obtain a file descriptor for an SSL session, and in turn, AppScope cannot include IP and port number fields in HTTP events. - -Static executables can be scoped only if they are written in Go. - -Static stripped Go executables can be scoped only when built with go1.13 or newer.
Use Y_PYTHON as executable name/prefix when sys.executable is empty. Otherwise the importer accepts importing anything.
@@ -11,7 +11,7 @@ import __res as __resource env_entry_point = 'Y_PYTHON_ENTRY_POINT' env_source_root = 'Y_PYTHON_SOURCE_ROOT' -executable = sys.executable +executable = sys.executable or 'Y_PYTHON' sys.modules['run_import_hook'] = __resource find_py_module = lambda mod: __resource.find('/py_modules/' + mod)
Get rid of unsigned integer overflow in column calculation While unsigned integer overflow is well-defined, Android sanitizers treat it as an error. We also have some in the SipHash implementation but those won't be easy to get rid of.
@@ -1768,13 +1768,14 @@ PREFIX(updatePosition)(const ENCODING *enc, const char *ptr, const char *end, # define LEAD_CASE(n) \ case BT_LEAD##n: \ ptr += n; \ + pos->columnNumber++; \ break; LEAD_CASE(2) LEAD_CASE(3) LEAD_CASE(4) # undef LEAD_CASE case BT_LF: - pos->columnNumber = (XML_Size)-1; + pos->columnNumber = 0; pos->lineNumber++; ptr += MINBPC(enc); break; @@ -1783,13 +1784,13 @@ PREFIX(updatePosition)(const ENCODING *enc, const char *ptr, const char *end, ptr += MINBPC(enc); if (HAS_CHAR(enc, ptr, end) && BYTE_TYPE(enc, ptr) == BT_LF) ptr += MINBPC(enc); - pos->columnNumber = (XML_Size)-1; + pos->columnNumber = 0; break; default: ptr += MINBPC(enc); + pos->columnNumber++; break; } - pos->columnNumber++; } }
Sort tests when generating cases
@@ -82,7 +82,7 @@ class BaseTarget: @classmethod def generate_tests(cls): """Generate test cases for the target subclasses.""" - for subclass in cls.__subclasses__(): + for subclass in sorted(cls.__subclasses__(), key=lambda c: c.__name__): yield from subclass.generate_tests()
session BUGFIX avoid unsafe variable reading Fixes
@@ -283,11 +283,6 @@ nc_session_rpc_lock(struct nc_session *session, int timeout, const char *func) } } } else if (!timeout) { - if (*session->opts.server.rpc_inuse) { - /* immediate timeout */ - return 0; - } - /* LOCK */ ret = pthread_mutex_trylock(session->opts.server.rpc_lock); if (!ret) {
readme: Fix links to gadgets documentation. Fixes: ("docs: Rename guides to gadgets")
@@ -28,35 +28,35 @@ Inspektor Gadget tools are known as gadgets. You can deploy one, two or many gad Explore the following documentation to find out which tools can help you in your investigations. - `advise`: - - [`network-policy`](docs/guides/advise/network-policy.md) - - [`seccomp-profile`](docs/guides/advise/seccomp-profile.md) + - [`network-policy`](docs/gadgets/advise/network-policy.md) + - [`seccomp-profile`](docs/gadgets/advise/seccomp-profile.md) - `audit`: - - [`seccomp`](docs/guides/audit/seccomp.md) + - [`seccomp`](docs/gadgets/audit/seccomp.md) - `profile`: - - [`block-io`](docs/guides/profile/block-io.md) - - [`cpu`](docs/guides/profile/cpu.md) + - [`block-io`](docs/gadgets/profile/block-io.md) + - [`cpu`](docs/gadgets/profile/cpu.md) - `snapshot`: - - [`process`](docs/guides/snapshot/process.md) - - [`socket`](docs/guides/snapshot/socket.md) + - [`process`](docs/gadgets/snapshot/process.md) + - [`socket`](docs/gadgets/snapshot/socket.md) - `top`: - - [`block-io`](docs/guides/top/block-io.md) - - [`ebpf`](docs/guides/top/ebpf.md) - - [`file`](docs/guides/top/file.md) - - [`tcp`](docs/guides/top/tcp.md) + - [`block-io`](docs/gadgets/top/block-io.md) + - [`ebpf`](docs/gadgets/top/ebpf.md) + - [`file`](docs/gadgets/top/file.md) + - [`tcp`](docs/gadgets/top/tcp.md) - `trace`: - - [`bind`](docs/guides/trace/bind.md) - - [`capabilities`](docs/guides/trace/capabilities.md) - - [`dns`](docs/guides/trace/dns.md) - - [`exec`](docs/guides/trace/exec.md) - - [`fsslower`](docs/guides/trace/fsslower.md) - - [`mount`](docs/guides/trace/mount.md) - - [`oomkill`](docs/guides/trace/oomkill.md) - - [`open`](docs/guides/trace/open.md) - - [`signal`](docs/guides/trace/signal.md) - - [`sni`](docs/guides/trace/sni.md) - - [`tcp`](docs/guides/trace/tcp.md) - - [`tcpconnect`](docs/guides/trace/tcpconnect.md) -- [`traceloop`](docs/guides/traceloop.md) + - [`bind`](docs/gadgets/trace/bind.md) + - [`capabilities`](docs/gadgets/trace/capabilities.md) + - [`dns`](docs/gadgets/trace/dns.md) + - [`exec`](docs/gadgets/trace/exec.md) + - [`fsslower`](docs/gadgets/trace/fsslower.md) + - [`mount`](docs/gadgets/trace/mount.md) + - [`oomkill`](docs/gadgets/trace/oomkill.md) + - [`open`](docs/gadgets/trace/open.md) + - [`signal`](docs/gadgets/trace/signal.md) + - [`sni`](docs/gadgets/trace/sni.md) + - [`tcp`](docs/gadgets/trace/tcp.md) + - [`tcpconnect`](docs/gadgets/trace/tcpconnect.md) +- [`traceloop`](docs/gadgets/traceloop.md) ## Installation
testcase/filesystem : Test more negative cases of setvbuf() 1. Fail to file open 2. Invalid mode check 3. If a buffer pointer is provided, then it must have a non-zero size 4. Reuse buffer 5. EBUSY test IMXRT1050-EVKB log : [tc_libc_stdio_setvbuf] PASS
@@ -2445,12 +2445,44 @@ static void tc_libc_stdio_setvbuf(void) char buffer[64]; char *filename = VFS_FILE_PATH; int ret; + int fp_fd; /* setvbuf_test: DEFAULT buffering */ fp = fopen(filename, "w"); TC_ASSERT_NEQ("fopen", fp, NULL); + /* setvbuf_test: fail to file open */ + + fp_fd = fp->fs_fd; + fp->fs_fd = ERROR; + ret = setvbuf(fp, buffer, _IOFBF, 64); + fp->fs_fd = fp_fd; + TC_ASSERT_LEQ_CLEANUP("setvbuf", ret, 0, fclose(fp)); + TC_ASSERT_EQ_CLEANUP("setvbuf", errno, EBADF, fclose(fp)); + + /* setvbuf_test: Invalid mode check */ + + ret = setvbuf(fp, NULL, ERROR, 0); + TC_ASSERT_LEQ_CLEANUP("setvbuf", ret, 0, fclose(fp)); + TC_ASSERT_EQ_CLEANUP("setvbuf", errno, EINVAL, fclose(fp)); + + /* setvbuf_test: EINVAL error check + * In case that if a buffer pointer is provided and the buffer size is zero + */ + + ret = setvbuf(fp, buffer, _IOFBF, 0); + TC_ASSERT_LEQ_CLEANUP("setvbuf", ret, 0, fclose(fp)); + TC_ASSERT_EQ_CLEANUP("setvbuf", errno, EINVAL, fclose(fp)); + + /* setvbuf_test: EBUSY test */ + + fp->fs_bufpos++; + ret = setvbuf(fp, NULL, _IOFBF, 64); + fp->fs_bufpos = fp->fs_bufstart; + TC_ASSERT_LEQ_CLEANUP("setvbuf", ret, 0, fclose(fp)); + TC_ASSERT_EQ_CLEANUP("setvbuf", errno, EBUSY, fclose(fp)); + /* setvbuf_test: NO buffering */ ret = setvbuf(fp, NULL, _IONBF, 0); @@ -2458,6 +2490,11 @@ static void tc_libc_stdio_setvbuf(void) /* setvbuf_test: FULL buffering */ + ret = setvbuf(fp, NULL, _IOFBF, 0); + TC_ASSERT_LEQ_CLEANUP("setvbuf", ret, 0, fclose(fp)); + + /* setvbuf_test: Reuse full buffering */ + ret = setvbuf(fp, NULL, _IOFBF, 0); TC_ASSERT_LEQ_CLEANUP("setvbuf", ret, 0, fclose(fp)); ret = setvbuf(fp, NULL, _IONBF, 0); @@ -2475,8 +2512,8 @@ static void tc_libc_stdio_setvbuf(void) ret = setvbuf(fp, buffer, _IOFBF, 64); TC_ASSERT_LEQ_CLEANUP("setvbuf", ret, 0, fclose(fp)); ret = setvbuf(fp, NULL, _IONBF, 0); + TC_ASSERT_LEQ_CLEANUP("setvbuf", ret, 0, fclose(fp)); fclose(fp); - TC_ASSERT_LEQ("setvbuf", ret, 0); TC_SUCCESS_RESULT(); }
tests: Use 0.5 msec flush interval instead of 1 sec This reflects the feeback in the pull request. Modify to use a sub- second flush interval and revise the test schedule accordingly.
@@ -95,41 +95,35 @@ void do_test(char *system, ...) TEST_CHECK(out_ffd >= 0); TEST_CHECK(flb_output_set(ctx, out_ffd, "match", "test", NULL) == 0); - TEST_CHECK(flb_service_set(ctx, "Flush", "3", + TEST_CHECK(flb_service_set(ctx, "Flush", "0.5", "Grace", "1", NULL) == 0); /* The following test tries to check if an input plugin generates * data in a timely manner. * - * 0 1 2 3 4 5 6 7 8 (sec) - * |--C--*--F--C--*--F--C--| + * 0 1 2 3 4 (sec) + * |--F--F--F--C--F--F--F--C--| * - * F ... Flush (3 sec interval) + * F ... Flush (0.5 sec interval) * C ... Condition checks * * Since CI servers can be sometimes very slow, we wait slightly a - * little more (1 sec) before checking the condition. + * little more before checking the condition. */ /* Start test */ TEST_CHECK(flb_start(ctx) == 0); - /* 1 sec passed. No data should be flushed */ - sleep(1); + /* 2 sec passed. It must have flushed */ + sleep(2); flb_info("[test] check status 1"); ret = get_result(); - TEST_CHECK(ret == 0); - - /* 4 sec passed. It must have flushed once */ - sleep(3); - flb_info("[test] check status 2"); - ret = get_result(); TEST_CHECK(ret > 0); - /* 7 sec passed. It must have flushed again */ - sleep(3); - flb_info("[test] check status 3"); + /* 4 sec passed. It must have flushed */ + sleep(2); + flb_info("[test] check status 2"); ret = get_result(); TEST_CHECK(ret > 0); @@ -140,13 +134,15 @@ void do_test(char *system, ...) void flb_test_in_disk_flush() { do_test("disk", - "interval_sec", "1", + "interval_sec", "0", + "interval_nsec", "500000000", NULL); } void flb_test_in_proc_flush() { do_test("proc", - "interval_sec", "1", + "interval_sec", "0", + "interval_nsec", "500000000", "proc_name", "flb_test_in_proc", "alert", "true", "mem", "on", @@ -156,7 +152,8 @@ void flb_test_in_proc_flush() void flb_test_in_head_flush() { do_test("head", - "Interval_Sec", "1", + "interval_sec", "0", + "interval_nsec", "500000000", "File", "/dev/urandom", NULL); }
Run All: Allow script to access terminal directly Before this change the output of `ninja run_all` was buffered. Consequently we could not see which sub-command caused problems if we terminated the script before it finished. This behaviour was especially problematic, if Travis killed the script cause it ran for too long.
@@ -35,10 +35,15 @@ macro (do_test source) ) endmacro (do_test) +if (NOT CMAKE_VERSION VERSION_LESS 3.2) + set (USES_TERMINAL USES_TERMINAL) +endif (NOT CMAKE_VERSION VERSION_LESS 3.2) + add_custom_target(run_all COMMAND "${CMAKE_SOURCE_DIR}/scripts/run_all" "$<CONFIGURATION>" - WORKING_DIRECTORY "${CMAKE_BINARY_DIR}") + WORKING_DIRECTORY "${CMAKE_BINARY_DIR}" + ${USES_TERMINAL}) add_custom_target(run_nokdbtests COMMAND "${CMAKE_SOURCE_DIR}/scripts/run_nokdbtests"
Always use TLSv1.0 for record layer version in TLSv1.3 TLSv1.3 freezes the record layer version and ensures that it is always set to TLSv1.0. Some implementations check this.
@@ -784,7 +784,7 @@ int do_ssl3_write(SSL *s, int type, const unsigned char *buf, /* Clear our SSL3_RECORD structures */ memset(wr, 0, sizeof wr); for (j = 0; j < numpipes; j++) { - unsigned int version = s->version; + unsigned int version = SSL_IS_TLS13(s) ? TLS1_VERSION : s->version; unsigned char *compressdata = NULL; size_t maxcomplen; unsigned int rectype;
Add some string/check-set tests.
(assert (deep= (parser/status p) (parser/status p2)) "parser 2") (assert (deep= (parser/state p) (parser/state p2)) "parser 3") +# String check-set +(assert (string/check-set "abc" "a") "string/check-set 1") +(assert (not (string/check-set "abc" "z")) "string/check-set 2") +(assert (string/check-set "abc" "abc") "string/check-set 3") +(assert (not (string/check-set "abc" "")) "string/check-set 4") +(assert (not (string/check-set "" "aabc")) "string/check-set 5") + (end-suite)
updating nagios core to v4.3.1
%global _hardened_build 1 Name: %{pname}%{PROJ_DELIM} -Version: 4.1.1 +Version: 4.3.1 Release: 1%{?dist} Summary: Host/service/network monitoring program Group: %{PROJ_NAME}/admin
Rename conn_send_client_initial as conn_write_client_initial
@@ -1047,8 +1047,8 @@ static ssize_t conn_write_handshake_ack_pkt(ngtcp2_conn *conn, uint8_t *dest, } /* - * conn_send_client_initial writes Client Initial packet in the buffer - * pointed by |dest| whose length is |destlen|. + * conn_write_client_initial writes Client Initial packet in the + * buffer pointed by |dest| whose length is |destlen|. * * This function returns the number of bytes written in |dest| if it * succeeds, or one of the following negative error codes: @@ -1060,7 +1060,7 @@ static ssize_t conn_write_handshake_ack_pkt(ngtcp2_conn *conn, uint8_t *dest, * NGTCP2_ERR_NOBUF * Buffer is too small. */ -static ssize_t conn_send_client_initial(ngtcp2_conn *conn, uint8_t *dest, +static ssize_t conn_write_client_initial(ngtcp2_conn *conn, uint8_t *dest, size_t destlen, ngtcp2_tstamp ts) { uint64_t pkt_num = 0; const uint8_t *payload; @@ -1452,7 +1452,7 @@ ssize_t ngtcp2_conn_send(ngtcp2_conn *conn, uint8_t *dest, size_t destlen, switch (conn->state) { case NGTCP2_CS_CLIENT_INITIAL: - nwrite = conn_send_client_initial(conn, dest, destlen, ts); + nwrite = conn_write_client_initial(conn, dest, destlen, ts); if (nwrite < 0) { break; }
gall: review cleanups
?: ?=([* %pass * %g %deal * * %leave *] move) =/ =wire p.move.move ?> ?=([%use @ @ %out @ @ *] wire) - =/ short-wire t.t.t.t.t.t.wire + =/ sub-wire t.t.t.t.t.t.wire =/ =dock [q.p q]:q.move.move =. outbound.watches.yoke - (~(del by outbound.watches.yoke) [short-wire dock]) + (~(del by outbound.watches.yoke) [sub-wire dock]) $(moves t.moves, new-moves [move new-moves]) ?. ?=([* %pass * %g %deal * * ?(%watch %watch-as) *] move) $(moves t.moves, new-moves [move new-moves]) :: =. p.move.move (weld sys-wire [(scot %ud sub-nonce.yoke) sub-wire]) - %= $ + %_ $ moves t.moves new-moves [move new-moves] sub-nonce.yoke +(sub-nonce.yoke)
brya: Remove duplicate keyboard support defines This removes some duplicate keyboard defines that crept in from splitting larger patches and merging. sorry. BRANCH=none TEST=buildall passes
#define CONFIG_EXTPOWER_GPIO -/* Common Keyboard Defines */ -#define CONFIG_CMD_KEYBOARD -#define CONFIG_KEYBOARD_BOARD_CONFIG -#define CONFIG_KEYBOARD_COL2_INVERTED -#define CONFIG_KEYBOARD_KEYPAD -#define CONFIG_KEYBOARD_PROTOCOL_8042 -#ifdef CONFIG_KEYBOARD_VIVALDI -#define CONFIG_KEYBOARD_PWRBTN_ASSERTS_KSI2 -#else -#define CONFIG_KEYBOARD_PWRBTN_ASSERTS_KSI3 -#endif - /* Host communication */ #define CONFIG_HOSTCMD_ESPI #define CONFIG_HOSTCMD_ESPI_VW_SLP_S4
Conan build time optimization. Build release version to avoid building everything.
@@ -21,7 +21,7 @@ jobs: - name: Setup Conan Profile run: | conan profile new default --detect - conan profile update settings.build_type=RelWithDebInfo default + conan profile update settings.build_type=Release default - name: Install Dependencies env: CONAN_BUILD_OPTIONS: | @@ -41,6 +41,9 @@ jobs: source activate_run.sh ctest --verbose source deactivate_run.sh + - name: Test Installed Celix + run: | + conan create -pr:b default -pr:h default -tf examples/conan_test_package -tbf test-build -o celix:celix_cxx17=True -o celix:celix_install_deprecated_api=True --require-override=libcurl/7.64.1 --require-override=openssl/1.1.1s . build-brew: runs-on: macOS-latest
multiline: always validate stream_id with lru_parser
@@ -524,7 +524,6 @@ static int ml_append_try_parser(struct flb_ml_parser_ins *parser, /* Parse incoming content */ ret = flb_parser_do(parser->ml_parser->parser, (char *) buf, size, &out_buf, &out_size, &out_time); - if (flb_time_to_double(&out_time) == 0.0) { flb_time_copy(&out_time, tm); } @@ -623,7 +622,8 @@ int flb_ml_append(struct flb_ml *ml, uint64_t stream_id, mk_list_foreach(head_group, &group->parsers) { parser_i = mk_list_entry(head_group, struct flb_ml_parser_ins, _head); - if (lru_parser && parser_i == lru_parser) { + if (lru_parser && lru_parser == parser_i && + lru_parser->last_stream_id == stream_id) { continue; } @@ -639,7 +639,6 @@ int flb_ml_append(struct flb_ml *ml, uint64_t stream_id, else { parser_i = NULL; } - } if (!processed) {
c2d2: increase the usart1_to_usb size Set the usart1_to_usb size to 1024 to match the other consoles. BRANCH=servo TEST=none
@@ -220,7 +220,7 @@ static struct usart_config const usart1; struct usb_stream_config const usart1_usb; static struct queue const usart1_to_usb = - QUEUE_DIRECT(128, uint8_t, usart1.producer, usart1_usb.consumer); + QUEUE_DIRECT(1024, uint8_t, usart1.producer, usart1_usb.consumer); static struct queue const usb_to_usart1 = QUEUE_DIRECT(64, uint8_t, usart1_usb.producer, usart1.consumer);
Fix use-after-free in deallocate_vmap
@@ -368,8 +368,9 @@ vmap allocate_vmap(rangemap rm, range q, struct vmap k) void deallocate_vmap(rangemap rm, vmap vm) { if (vm->fsf) { + filesystem fs = fsfile_get_fs(vm->fsf); fsfile_release(vm->fsf); - filesystem_release(fsfile_get_fs(vm->fsf)); + filesystem_release(fs); } deallocate(rm->h, vm, sizeof(struct vmap)); }
test: fix the e2e test to use the configured manifest
@@ -20,10 +20,10 @@ func (e *e2e) testCaseReDeployOperator([]string) { e.seccompOnlyTestCase() // Clean up the operator - e.cleanupOperator(manifest) + e.cleanupOperator(e.operatorManifest) // Deploy the operator again - e.deployOperator(manifest) + e.deployOperator(e.operatorManifest) } func (e *e2e) testCaseReDeployNamespaceOperator([]string) {
denied to use toredis in new projects
@@ -160,3 +160,8 @@ ALLOW sandbox/projects/mail/Load/ShootingComparison -> contrib/python/jsonlib ALLOW sandbox/projects/mail/Load/WebApiAmmoGen -> contrib/python/jsonlib ALLOW sandbox/projects/mail/Load/WebApiShooting -> contrib/python/jsonlib DENY .* -> contrib/python/jsonlib + +ALLOW antiadblock/libs/tornado_redis -> contrib/python/toredis +ALLOW yabs/amazon/cache_proxy -> contrib/python/toredis +ALLOW yql/udfs/common/python/python_arc -> contrib/python/toredis +DENY .* -> contrib/python/toredis
docs: explains when one should use "easy commit"
@@ -39,6 +39,9 @@ References: # Easy Commit +If you have problems to submit clean pull requests without introducing merge commits, +you can use this method, otherwise please use your git commands. + This section explains how to submit changes and pull requests the easy and cleanest way without causing any merge conflicts.
Reword documentation of CMAC operations Change the wording of the documentation for some CMAC functions, as the existing wording, while technically correct, can be easy to misunderstand. The reworded docs explain the flow of a CMAC computation a little more fully.
@@ -67,9 +67,17 @@ struct mbedtls_cmac_context_t #endif /* !MBEDTLS_CMAC_ALT */ /** - * \brief This function sets the CMAC key, and prepares to authenticate + * \brief This function starts a new CMAC computation + * by setting the CMAC key, and preparing to authenticate * the input data. - * Must be called with an initialized cipher context. + * It must be called with an initialized cipher context. + * + * Once this function has completed, data can be supplied + * to the CMAC computation by calling + * mbedtls_cipher_cmac_update(). + * + * To start a CMAC computation using the same key as a previous + * CMAC computation, use mbedtls_cipher_cmac_finish(). * * \note When the CMAC implementation is supplied by an alternate * implementation (through #MBEDTLS_CMAC_ALT), some ciphers @@ -95,9 +103,15 @@ int mbedtls_cipher_cmac_starts( mbedtls_cipher_context_t *ctx, * \brief This function feeds an input buffer into an ongoing CMAC * computation. * - * It is called between mbedtls_cipher_cmac_starts() or - * mbedtls_cipher_cmac_reset(), and mbedtls_cipher_cmac_finish(). - * Can be called repeatedly. + * The CMAC computation must have previously been started + * by calling mbedtls_cipher_cmac_starts() or + * mbedtls_cipher_cmac_reset(). + * + * Call this function as many times as needed to input the + * data to be authenticated. + * Once all of the required data has been input, + * call mbedtls_cipher_cmac_finish() to obtain the result + * of the CMAC operation. * * \param ctx The cipher context used for the CMAC operation. * \param input The buffer holding the input data. @@ -111,12 +125,13 @@ int mbedtls_cipher_cmac_update( mbedtls_cipher_context_t *ctx, const unsigned char *input, size_t ilen ); /** - * \brief This function finishes the CMAC operation, and writes - * the result to the output buffer. + * \brief This function finishes an ongoing CMAC operation, and + * writes the result to the output buffer. * - * It is called after mbedtls_cipher_cmac_update(). - * It can be followed by mbedtls_cipher_cmac_reset() and - * mbedtls_cipher_cmac_update(), or mbedtls_cipher_free(). + * It should be followed either by + * mbedtls_cipher_cmac_reset(), which starts another CMAC + * operation with the same key, or mbedtls_cipher_free(), + * which clears the cipher context. * * \param ctx The cipher context used for the CMAC operation. * \param output The output buffer for the CMAC checksum result. @@ -129,12 +144,14 @@ int mbedtls_cipher_cmac_finish( mbedtls_cipher_context_t *ctx, unsigned char *output ); /** - * \brief This function prepares the authentication of another - * message with the same key as the previous CMAC - * operation. - * - * It is called after mbedtls_cipher_cmac_finish() - * and before mbedtls_cipher_cmac_update(). + * \brief This function starts a new CMAC operation with the same + * key as the previous one. + * + * It should be called after finishing the previous CMAC + * operation with mbedtls_cipher_cmac_finish(). + * After calling this function, + * call mbedtls_cipher_cmac_update() to supply the new + * CMAC operation with data. * * \param ctx The cipher context used for the CMAC operation. *
tests/float/cmath_fun.py: Fix truncation of small real part of complex.
@@ -50,6 +50,6 @@ for f_name, f, test_vals in functions: else: # some test (eg cmath.sqrt(-0.5)) disagree with CPython with tiny real part real = ret.real - if abs(real) < 1e15: + if abs(real) < 1e-6: real = 0. print("complex(%.5g, %.5g)" % (real, ret.imag))
Enable PBR for Release builds
@@ -52,7 +52,6 @@ else() set(CMAKE_BUILD_TYPE Debug) add_definitions( -DDEBUG_BUILD - -DPLAYBACK_RECORD_SUPPORTED ) endif() @@ -114,6 +113,7 @@ add_definitions( -DOS_BUILD -DPCD_CACHE_ENABLED -D__VERSION_NUMBER__=${IPMCTL_VERSION_STRING} + -DPLAYBACK_RECORD_SUPPORTED ) if(MSVC)
libc: Partially support ADDFB2 ioctl
@@ -796,6 +796,7 @@ int socketpair(int domain, int type, int proto, int *fds) { } #include <libdrm/drm.h> +#include <libdrm/drm_fourcc.h> int ioctl(int fd, unsigned long request, void *arg) { // frigg::infoLogger() << "mlibc: ioctl with" @@ -1183,6 +1184,57 @@ int ioctl(int fd, unsigned long request, void *arg) { return resp.result(); } + case DRM_IOCTL_MODE_ADDFB2: { + auto param = reinterpret_cast<drm_mode_fb_cmd2 *>(arg); + HelAction actions[3]; + globalQueue.trim(); + + __ensure(param->pixel_format == DRM_FORMAT_XRGB8888); + __ensure(!param->flags); + __ensure(!param->modifier[0]); + __ensure(!param->offsets[0]); + + managarm::fs::CntRequest<MemoryAllocator> req(getAllocator()); + req.set_req_type(managarm::fs::CntReqType::PT_IOCTL); + req.set_command(DRM_IOCTL_MODE_ADDFB); + + req.set_drm_width(param->width); + req.set_drm_height(param->height); + req.set_drm_pitch(param->pitches[0]); + req.set_drm_bpp(32); + req.set_drm_depth(24); + req.set_drm_handle(param->handles[0]); + + frigg::String<MemoryAllocator> ser(getAllocator()); + req.SerializeToString(&ser); + actions[0].type = kHelActionOffer; + actions[0].flags = kHelItemAncillary; + actions[1].type = kHelActionSendFromBuffer; + actions[1].flags = kHelItemChain; + actions[1].buffer = ser.data(); + actions[1].length = ser.size(); + actions[2].type = kHelActionRecvInline; + actions[2].flags = 0; + HEL_CHECK(helSubmitAsync(handle, actions, 3, + globalQueue.getQueue(), 0, 0)); + + auto element = globalQueue.dequeueSingle(); + auto offer = parseSimple(element); + auto send_req = parseSimple(element); + auto recv_resp = parseInline(element); + + HEL_CHECK(offer->error); + HEL_CHECK(send_req->error); + HEL_CHECK(recv_resp->error); + + managarm::fs::SvrResponse<MemoryAllocator> resp(getAllocator()); + resp.ParseFromArray(recv_resp->data, recv_resp->length); + __ensure(resp.error() == managarm::fs::Errors::SUCCESS); + + param->fb_id = resp.drm_fb_id(); + + return resp.result(); + } case DRM_IOCTL_MODE_RMFB: { auto param = reinterpret_cast<int *>(arg); HelAction actions[3];
Support for ecbuild v3.4. Changes to ecbuild_add_option and REQUIRED_PACKAGES
@@ -99,10 +99,11 @@ ecbuild_add_option( FEATURE JPG_LIBOPENJPEG CONDITION ENABLE_JPG DEFAULT ON ) +find_package( PNG ) ecbuild_add_option( FEATURE PNG DESCRIPTION "Support for PNG decoding/encoding" DEFAULT OFF - REQUIRED_PACKAGES PNG ) + CONDITION PNG_FOUND ) if( HAVE_PNG ) set( HAVE_LIBPNG 1 ) # compatibility with autotools @@ -111,21 +112,29 @@ else() set( HAVE_LIBPNG 0 ) endif() +find_package( NetCDF ) ecbuild_add_option( FEATURE NETCDF DESCRIPTION "Support for GRIB to NetCDF conversion" DEFAULT ON - REQUIRED_PACKAGES NetCDF + CONDITION NetCDF_FOUND NO_TPL ) +find_package( AEC QUIET ) ecbuild_add_option( FEATURE AEC DESCRIPTION "Support for Adaptive Entropy Coding" DEFAULT OFF - REQUIRED_PACKAGES AEC ) + CONDITION AEC_FOUND ) +#find_package( Python 2.6 NO_LIBS ) +ecbuild_find_python( VERSION 2.6 NO_LIBS ) +find_package( NumPy ) +message(STATUS "..............Python_FOUND=${PYTHON_FOUND} and NumPy_FOUND=${NUMPY_FOUND}") ecbuild_add_option( FEATURE PYTHON DESCRIPTION "Build the ecCodes Python2 interface (deprecated)" DEFAULT OFF - REQUIRED_PACKAGES "Python VERSION 2.6 NO_LIBS" NumPy ) + #REQUIRED_PACKAGES "Python VERSION 2.6 NO_LIBS" NumPy + CONDITION Python_FOUND AND NumPy_FOUND + ) # For Python2 we build our own bindings (using SWIG) in the build directory # but for Python3 one has to add the eccodes from pip3 AFTER the install if( PYTHON_VERSION_MAJOR EQUAL 3 )
khan: handle multiple %fards in one event
$>(?(%arow %avow) gift) :: thread result == == :: +$ khan-state :: - [%0 hey=duct] :: current unix duct + $: %0 :: state v0 + hey=duct :: unix duct + tic=@ud :: tid counter + == :: -- :: => |% =* state - |= [now=@da eny=@uvJ rof=roof] =* khan-gate . +=. tic 0 ^? |% :: +call: handle a +task request :: %fard =/ tid=@ta - (cat 3 'khan-fyrd--' (scot %uv (sham eny))) + %^ cat 3 + 'khan-fyrd--' + (scot %uv (sham (mix tic eny))) + =. tic +(tic) =* fyd p.task =/ =beak (get-beak bear.fyd now) =/ args [~ `tid beak name.fyd q.args.fyd]
while we still have the in-process crash, default plugin session to auto-start named pipe. For existing prefs, warn if it's set to in-process and do auto-start pipe instead anyway, but don't modify the exisitng prefs.
@@ -168,10 +168,7 @@ namespace { OptionVars() : asyncMode("AsynchronousMode", 1) - , sessionType( - "SessionType", - static_cast<int>(HAPI_SESSION_INPROCESS) - ) + , sessionType( "SessionType", 2) // named pipe , thriftServer("ThriftServer", "localhost") , thriftPort("ThriftPort", 9090) , sessionPipeCustom("SessionPipeCustom", 0) @@ -214,10 +211,22 @@ initializeSession(const OptionVars& optionVars) const SessionType::Enum sessionType = static_cast<SessionType::Enum>( optionVars.sessionType.get() ); + // In Process is currently crashing due to library conflicts + // switch to named pipe with autostart, but leave existing prefs alone + // revert this if we can resolve the library conflicts + SessionType::Enum actualSessionType = sessionType; + bool overrideInProcess = false; + if( sessionType == SessionType::ST_INPROCESS) { + MGlobal::displayInfo( + "Houdini Engine In Process session not currently supported, switching to auto-start named pipe session"); + actualSessionType = SessionType::ST_THRIFT_PIPE; + overrideInProcess = true; + } + Util::theHAPISession.reset( new Util::HAPISession ); HAPI_Result sessionResult = HAPI_RESULT_FAILURE; - switch (sessionType) + switch (actualSessionType) { case SessionType::ST_INPROCESS: MGlobal::displayInfo( @@ -270,7 +279,7 @@ initializeSession(const OptionVars& optionVars) MString msgPipe; - if( !optionVars.sessionPipeCustom.get() ) + if( !optionVars.sessionPipeCustom.get() || overrideInProcess) { HAPI_ThriftServerOptions serverOptions; serverOptions.autoClose = true;
authors: add copy-on-write
@@ -161,7 +161,7 @@ notification system, I/O bindings ## Maximilian Irlinger -documentation, Python bindings for elektraMerge +documentation, Python bindings for elektraMerge, copy-on-write functionality of Key and KeySet - email: [email protected] - GitHub user: [atmaxinger](https://github.com/atmaxinger)
ENHANCE: skip build_complete_strings() when nsegtok is 0.
@@ -579,9 +579,11 @@ int tokenize_sblocks(mblck_list_t *blist, int length, char delimiter, int keycnt if (finish_flag == false) { return -1; /* some errors */ } + if (nsegtok > 0) { if (build_complete_strings(blist, segtoks, nsegtok, tokens, ntokens) != 0) { return -2; /* out of memory */ } + } return 0; /* OK */ } #endif
Add link to LLVM bug
@@ -473,6 +473,7 @@ ParallelRegion::Verify() * is safe. Given an instruction reading from memory, * IsLoadUnconditionallySafe should return whether it is safe under * (unconditional, unpredicated) speculative execution. + * See https://bugs.llvm.org/show_bug.cgi?id=46666 */ void ParallelRegion::AddParallelLoopMetadata(
scripts: Ensure non-zero exit code on error
@@ -82,12 +82,12 @@ if [ "$EC2FASTINSTALL" = "true" ]; then FASTINSTALL=true echo "Using fast pre-compiled install for riscv-tools" else - echo "Error: hash of precompiled toolchain doesn't match the riscv-tools submodule hash." - exit + error 'error: hash of precompiled toolchain does not match the riscv-tools submodule hash' + exit -1 fi else - echo "Error: No precompiled toolchain for esp-tools or other non-native riscv-tools." - exit + error "error: unsupported precompiled toolchain: ${TOOLCHAIN}" + exit -1 fi fi
added units for SeasonalClimate Files
@@ -305,10 +305,24 @@ def ProcessClimatefile(climatefile, data, body, GridOutputOrder): def ProcessSeasonalClimatefile(prefix, data, body, name): file = list(csv.reader(open('SeasonalClimateFiles/' + prefix + '.' + name + '.0'))) key_name = body + '_' + name - if key_name in data: - data[key_name].append(file) + units = '' + if (name == 'DailyInsol' or name == 'SeasonalFIn' or + name == 'SeasonalFOut'or name == 'SeasonalDivF'): + units = 'W/m^2' + if name == 'PlanckB': + units = 'W/m^2/K' + if name == 'SeasonalIceBalance': + units = 'kg/m^2/s' + if name == 'SeasonalTemp': + units = 'deg C' + if name == 'SeasonalFMerid': + units = 'W' + + if key_name not in data: + data[key_name]= [units, file] else: - data[key_name] = [file] + data[key_name].append(file) + return data def CreateHDF5(data, system_name, infiles, logfile, quiet, h5filename): @@ -346,8 +360,8 @@ def CreateHDF5(data, system_name, infiles, logfile, quiet, h5filename): with h5py.File(h5filename, 'w') as h: for k, v in data.items(): - #print("Key:",k) - #print("Length of Value:",len(v)) + print("Key:",k) + print("Length of Value:",len(v)) if len(v) == 2: v_attr = v[0] v_value = [v[1]] @@ -355,8 +369,8 @@ def CreateHDF5(data, system_name, infiles, logfile, quiet, h5filename): else: v_value = v[0] v_attr = '' - #print("Units:",v_attr) - #print() + print("Units:",v_attr) + print() h.create_dataset(k, data=np.array(v_value,dtype='S'),compression = 'gzip') h[k].attrs['Units'] = v_attr
update ya tool crypta version with task
}, "crypta": { "formula": { - "sandbox_id": 334422043, + "sandbox_id": 338265308, "match": "crypta" }, "executable": {
decisions: update man pages
.\" generated with Ronn-NG/v0.10.1 .\" http://github.com/apjanke/ronn-ng/tree/0.10.1.pre1 -.TH "ELEKTRA\-LIBS" "7" "January 2021" "" +.TH "ELEKTRA\-LIBS" "7" "May 2022" "" .SH "NAME" \fBelektra\-libs\fR \- libs overview .SH "Highlevel APIs" @@ -24,7 +24,9 @@ Using asynchronous I/O bindings \fIhttps://www\.libelektra\.org/examples/notific Reload KDB when Elektra's configuration has changed \fIhttps://www\.libelektra\.org/examples/notificationreload\fR .IP "" 0 .SH "Base Elektra Libraries" -Since version \fB0\.8\.15 \fI/doc/decisions/library_split\.md\fR\fR \fBlibelektra \fIelektra/\fR\fR is split into following libraries: +TODO: \fIOutdated\fR +.P +Since version \fB0\.8\.15\fR \fBlibelektra \fIelektra/\fR\fR is split into following libraries: .P .SS "Libkdb" .nf
Restructure docs for the CC2538 CPU
* OF THE POSSIBILITY OF SUCH DAMAGE. */ /** - * \addtogroup platform - * @{ - * - * \defgroup cc2538-platforms TI cc2538-powered platforms - * - * Documentation for all platforms powered by the TI cc2538 System-on-Chip - * @{ - * - * \defgroup cc2538 The TI cc2538 System-on-Chip - * CPU-Specific functionality - available to all cc2538-based platforms + * \addtogroup cc2538 * @{ * * \defgroup cc2538-cpu cc2538 CPU #endif /* CPU_H_ */ /** - * @} - * @} * @} * @} */
out_es: fix fractional second formatting
@@ -237,7 +237,7 @@ static char *elasticsearch_format(void *data, size_t bytes, s = strftime(time_formatted, sizeof(time_formatted) - 1, ctx->time_key_format, &tm); len = snprintf(time_formatted + s, sizeof(time_formatted) - 1 - s, - ".%" PRIu64 "Z", (uint64_t) tms.tm.tv_nsec); + ".%03" PRIu64 "Z", (uint64_t) tms.tm.tv_nsec); s += len; msgpack_pack_str(&tmp_pck, s);
typo in the workflow
@@ -39,7 +39,7 @@ jobs: - name: Get Tag id: tag run: | - if [ -z "${GITHUB_REF%%ref/tags/v*}" -a "true" = "${{ steps.version.outputs.is-semver }}" ]; then + if [ -z "${GITHUB_REF%%refs/tags/v*}" -a "true" = "${{ steps.version.outputs.is-semver }}" ]; then echo "::set-output name=tag::${{ steps.version.outputs.version-without-v }}" echo "::set-output name=push::true" if [ -z "${{ steps.version.outputs.prerelease }}" ]; then
use smaller fonts for subsubsection to improve readability
@@ -28,7 +28,7 @@ sudo apt-get install -y libcurl4-openssl-dev libssl-dev \* If you do not have Ubuntu or Debian but have Windows 10, you can install WSL2 and get either Ubuntu or Debian [here](https://docs.microsoft.com/en-us/windows/wsl/install-win10). -### Compile +#### Compile ``` make bot-echo ``` @@ -44,10 +44,10 @@ make bot-echo 3. Run `./bot-echo.exe` in the same folder of `bot.config` -### Test bot-echo +#### Test bot-echo Type any message in any public channel of the server that the bot is invited. -### Terminate bot-echo +#### Terminate bot-echo Close the Terminal that bot-echo is running or type "Ctrl-C" to kill it.
sh2 update fixed wrong offset
@@ -685,9 +685,9 @@ void Init() injector::WriteMemory(pattern.count(3).get(0).get<void>(2), &fneg190, true); injector::WriteMemory(pattern.count(3).get(2).get<void>(2), &fneg190, true); pattern = hook::pattern("C7 05 ? ? ? ? ? ? ? ? E9 ? ? ? ? 8B 15"); //004A314F - injector::WriteMemory<float>(pattern.get_first(2), fneg190, true); + injector::WriteMemory<float>(pattern.get_first(6), fneg190, true); pattern = hook::pattern("C7 05 ? ? ? ? ? ? ? ? E9 ? ? ? ? 6A 00 68 00 00 04 00 6A 00"); //004A2DD2 - injector::WriteMemory<float>(pattern.get_first(2), fneg190, true); + injector::WriteMemory<float>(pattern.get_first(6), fneg190, true); } // Fixes lying figure cutscene bug; original value 00000005; issue #349
fix RGB brightness missing
@@ -135,7 +135,8 @@ Blockly.Arduino.display_rgb_init=function(){ var Brightness = Blockly.Arduino.valueToCode(this, 'Brightness',Blockly.Arduino.ORDER_ATOMIC); Blockly.Arduino.definitions_['include_Adafruit_NeoPixel'] = '#include <Adafruit_NeoPixel.h>'; Blockly.Arduino.definitions_['var_declare_rgb_display' + dropdown_rgbpin] = 'Adafruit_NeoPixel rgb_display_' + dropdown_rgbpin + '= Adafruit_NeoPixel(' + value_ledcount + ','+dropdown_rgbpin+',NEO_GRB + NEO_KHZ800);'; - Blockly.Arduino.setups_['setup_rgb_display_begin_' + dropdown_rgbpin] = 'rgb_display_' + dropdown_rgbpin + '.begin();\nrgb_display_' + dropdown_rgbpin + '.setBrightness('+Brightness+');'; + Blockly.Arduino.setups_['setup_rgb_display_begin_' + dropdown_rgbpin] = 'rgb_display_' + dropdown_rgbpin + '.begin();'; + Blockly.Arduino.setups_['setup_rgb_display_setBrightness' + dropdown_rgbpin] = 'rgb_display_' + dropdown_rgbpin + '.setBrightness('+Brightness+');'; return ''; }; Blockly.Arduino.display_rgb=function(){
Add ksceKernelSearchModuleByName
@@ -102,6 +102,8 @@ int ksceKernelStopUnloadModuleForPid(SceUID pid, SceUID modid, SceSize args, voi int ksceKernelMountBootfs(const char *bootImagePath); int ksceKernelUmountBootfs(void); +int ksceKernelSearchModuleByName(const char* module_name, const char* path, int pid); + #ifdef __cplusplus } #endif
Update Autotools instructions. Not fully tested, but these updated instructions should cause less issues than the previous one.
Building Libtcod using Autotools ================================ -The following instructions have been tested on 32 and 64-bit versions of -Ubuntu 14.04 and Fedora 22. - Dependencies ------------ -For Ubuntu 14.04, install these dependencies: +For Ubuntu 16.04, install these dependencies: - $ sudo apt-get install curl build-essential make cmake autoconf automake libtool mercurial libasound2-dev libpulse-dev libaudio-dev libx11-dev libxext-dev libxrandr-dev libxcursor-dev libxi-dev libxinerama-dev libxxf86vm-dev libxss-dev libgl1-mesa-dev libesd0-dev libdbus-1-dev libudev-dev libgles1-mesa-dev libgles2-mesa-dev libegl1-mesa-dev + $ sudo apt install build-essential autoconf automake libtool git libsdl2-dev For Fedora 22: @@ -26,32 +23,21 @@ It is recommended strongly that you install SDL2 using your package manager. However, if you are unable to work out the package name, then you can take the harder route and build it yourself. -Download the supported SDL2 revision, build and install it if you must: - - $ curl -o sdl.tar.gz http://hg.libsdl.org/SDL/archive/007dfe83abf8.tar.gz - $ tar -xf sdl.tar.gz - $ cd SDL-007dfe83abf8/ - $ mkdir -p build - $ cd build - $ ../configure - $ make - $ sudo make install - -This will place the libraries at `/usr/local/lib/` and the development headers -at `/usr/local/include/SDL2/`. +https://wiki.libsdl.org/Installation Building Libtcod ---------------- -Download the latest libtcod version, build it and install it: +Download the latest libtcod version, build it, and install it: - $ hg clone https://bitbucket.org/libtcod/libtcod + $ git clone https://github.com/libtcod/libtcod.git $ cd libtcod/build/autotools $ autoreconf -i - $ ./configure CFLAGS='-O2' + $ ./configure $ make $ sudo make install This will place libtcod static and shared libraries in the `/usr/local/lib` directory, and header files in the `/usr/local/include/libtcod` directory. -Note that the same makefile is used for 32 and 64 bit distributions. +Once installed you can get the necessary compile flags for a libtcod project +with the `pkg-config libtcod --cflags --libs` command.
Docs - update PostGIS version compatibility
<row class="- topic/row "> <entry colname="col1" class="- topic/entry ">MADlib machine learning<sup>1</sup></entry> - <entry colname="col3" class="- topic/entry ">MADlib 1.16</entry> + <entry colname="col3" class="- topic/entry ">MADlib 1.17, 1.16</entry> </row> <row> <entry>PL/Java</entry> <row class="- topic/row "> <entry colname="col1" class="- topic/entry ">PostGIS Spatial and Geographic Objects for Greenplum Database 6.0.x</entry> - <entry colname="col3" class="- topic/entry ">2.1.5+pivotal.2-2</entry> + <entry colname="col3" class="- topic/entry ">2.5.4+pivotal.1, + 2.1.5+pivotal.2-2</entry> </row> </tbody> </tgroup>
symtrack: fixing typo in comment with parameter description
@@ -6702,7 +6702,7 @@ int SYMTRACK(_set_bandwidth)(SYMTRACK() _q, \ \ /* Adjust internal NCO by requested frequency */ \ /* _q : symtrack object */ \ -/* _dphi : NCO phase adjustment [radians] */ \ +/* _dphi : NCO frequency adjustment [radians/sample] */ \ int SYMTRACK(_adjust_frequency)(SYMTRACK() _q, \ T _dphi); \ \
Added parsing support for Ascii password files so the dimm lockstate can be changed
@@ -1683,6 +1683,19 @@ ParseSourcePassFile( goto Finish; } + // Verify if it is Unicode file: + //If it is not a Unicode File Convert the File String + if (*((CHAR16 *)pFileBuffer) != UTF_16_BOM) { + pFileString = AllocateZeroPool((FileBufferSize * sizeof(CHAR16)) + sizeof(L'\0')); + if (pFileString == NULL) { + Print(FORMAT_STR_NL, CLI_ERR_OUT_OF_MEMORY); + ReturnCode = EFI_OUT_OF_RESOURCES; + goto Finish; + } + ReturnCode = SafeAsciiStrToUnicodeStr((const CHAR8 *)pFileBuffer, (UINT32)FileBufferSize, pFileString); + Index = 0; + } + else { // Add size of L'\0' (UTF16) char pFileString = ReallocatePool(FileBufferSize, FileBufferSize + sizeof(L'\0'), pFileBuffer); if (pFileString == NULL) { @@ -1690,19 +1703,13 @@ ParseSourcePassFile( ReturnCode = EFI_OUT_OF_RESOURCES; goto Finish; } + Index = 1; NumberOfChars = (UINT32)(FileBufferSize / sizeof(CHAR16)); pFileString[NumberOfChars] = L'\0'; - NumberOfChars++; - - // Verify if it is Unicode file: - if (pFileString[0] != UTF_16_BOM) { - Print(L"Error: The file is not in UTF16 format, BOM header missing.\n"); - ReturnCode = EFI_INVALID_PARAMETER; - goto Finish; } - // Split input file to lines (but ignore byte order mark) - ppLinesBuffer = StrSplit(&pFileString[1], L'\n', &NumberOfLines); + // Split input file to lines + ppLinesBuffer = StrSplit(&pFileString[Index], L'\n', &NumberOfLines); if (ppLinesBuffer == NULL || NumberOfLines == 0) { Print(L"Error: The file is empty.\n"); ReturnCode = EFI_INVALID_PARAMETER; @@ -1763,8 +1770,9 @@ Finish: for (Index = 0; ppLinesBuffer != NULL && Index < NumberOfLines; ++Index) { FREE_POOL_SAFE(ppLinesBuffer[Index]); } - FREE_POOL_SAFE(ppLinesBuffer); + FREE_POOL_SAFE(pFileBuffer); FREE_POOL_SAFE(pFileString); + FREE_POOL_SAFE(ppLinesBuffer); FREE_POOL_SAFE(pReadBuffer); NVDIMM_EXIT_I64(ReturnCode); return ReturnCode;
Fix compile-time warn in pg_basebackup code. Change sprintf function to snprintf, to check resulting buffer size overflow.
@@ -1986,7 +1986,7 @@ BaseBackup(void) char *path = unconstify(char *, get_tablespace_mapping(PQgetvalue(res, i, 1))); char path_with_subdir[MAXPGPATH]; - sprintf(path_with_subdir, "%s/%d/%s", path, target_gp_dbid, GP_TABLESPACE_VERSION_DIRECTORY); + snprintf(path_with_subdir, MAXPGPATH, "%s/%d/%s", path, target_gp_dbid, GP_TABLESPACE_VERSION_DIRECTORY); verify_dir_is_empty_or_create(path_with_subdir, &made_tablespace_dirs, &found_tablespace_dirs); }
luci-app-serverchan: bump to 1.89
include $(TOPDIR)/rules.mk PKG_NAME:=luci-app-serverchan -PKG_VERSION:=1.87 +PKG_VERSION:=1.89 PKG_RELEASE:=9 include $(INCLUDE_DIR)/package.mk
OcAppleKernelLib: Include non-local VTable entries when counting.
@@ -507,8 +507,9 @@ InternalInitializeVtablePatchData ( IN OUT OC_MACHO_CONTEXT *MachoContext, IN CONST MACH_NLIST_64 *VtableSymbol, IN OUT UINT32 *MaxSize, - OUT UINT32 *NumEntries, OUT UINT64 **VtableDataPtr, + OUT UINT32 *NumEntries, + OUT UINT32 *NumSymbols, OUT MACH_NLIST_64 **SolveSymbols ) { @@ -517,7 +518,7 @@ InternalInitializeVtablePatchData ( UINT32 VtableMaxSize; CONST MACH_HEADER_64 *MachHeader; UINT64 *VtableData; - UINT32 Index; + UINT32 SymIndex; UINT32 EntryOffset; UINT32 MaxSymbols; MACH_NLIST_64 *Symbol; @@ -546,7 +547,7 @@ InternalInitializeVtablePatchData ( MaxSymbols = (*MaxSize / sizeof (*SolveSymbols)); VtableMaxSize /= VTABLE_ENTRY_SIZE_64; - Index = 0; + SymIndex = 0; for ( EntryOffset = VTABLE_HEADER_LEN_64; @@ -564,18 +565,19 @@ InternalInitializeVtablePatchData ( // If the VTable entry is 0 and it is not referenced by a Relocation, // it is the end of the table. // - *NumEntries = Index; - *MaxSize -= (Index * sizeof (*SolveSymbols)); + *MaxSize -= (SymIndex * sizeof (*SolveSymbols)); *VtableDataPtr = VtableData; + *NumEntries = (EntryOffset - VTABLE_HEADER_LEN_64); + *NumSymbols = SymIndex; return TRUE; } - if (Index >= MaxSymbols) { + if (SymIndex >= MaxSymbols) { return FALSE; } - SolveSymbols[Index] = Symbol; - ++Index; + SolveSymbols[SymIndex] = Symbol; + ++SymIndex; } } @@ -597,6 +599,7 @@ InternalPatchByVtables64 ( UINT32 Index; UINT32 NumTables; UINT32 NumEntries; + UINT32 NumEntriesTemp; UINT32 NumPatched; BOOLEAN Result; CONST MACH_NLIST_64 *Smcp; @@ -662,29 +665,33 @@ InternalPatchByVtables64 ( MachoContext, EntryWalker->Vtable, &MaxSize, - &EntryWalker->MetaSymsIndex, &EntryWalker->VtableData, + &NumEntriesTemp, + &EntryWalker->MetaSymsIndex, EntryWalker->SolveSymbols ); if (!Result) { return FALSE; } + NumEntries += NumEntriesTemp; + Result = InternalInitializeVtablePatchData ( MachoContext, EntryWalker->MetaVtable, &MaxSize, - &EntryWalker->NumSolveSymbols, &EntryWalker->MetaVtableData, + &NumEntriesTemp, + &EntryWalker->NumSolveSymbols, &EntryWalker->SolveSymbols[EntryWalker->MetaSymsIndex] ); if (!Result) { return FALSE; } - EntryWalker->NumSolveSymbols += EntryWalker->MetaSymsIndex; + NumEntries += NumEntriesTemp; - NumEntries += EntryWalker->NumSolveSymbols; + EntryWalker->NumSolveSymbols += EntryWalker->MetaSymsIndex; ++NumTables; EntryWalker = GET_NEXT_OC_VTABLE_PATCH_ENTRY (EntryWalker);
Fix test/recipes/01-test_symbol_presence.t to allow for stripped libraries It's a small change to the 'nm' call, to have it look at dynamic symbols rather than the normal ones. Fixes
@@ -48,12 +48,12 @@ foreach my $libname (@libnames) { *OSTDOUT = *STDOUT; open STDERR, ">", devnull(); open STDOUT, ">", devnull(); - my @nm_lines = map { s|\R$||; $_ } `nm -Pg $shlibpath 2> /dev/null`; + my @nm_lines = map { s|\R$||; $_ } `nm -DPg $shlibpath 2> /dev/null`; close STDERR; close STDOUT; *STDERR = *OSTDERR; *STDOUT = *OSTDOUT; - skip "Can't run 'nm -Pg $shlibpath' => $?... ignoring", 2 + skip "Can't run 'nm -DPg $shlibpath' => $?... ignoring", 2 unless $? == 0; my $bldtop = bldtop_dir();
grid: remove native webkit appearance on SystemMenu
@@ -57,7 +57,7 @@ export const SystemMenu = ({ className, menu, open, navOpen }: SystemMenuProps) as={Link} to="/system-menu" className={classNames( - 'circle-button default-ring', + 'appearance-none circle-button default-ring', open && 'text-gray-300', navOpen && menu !== 'system-preferences' && @@ -66,7 +66,7 @@ export const SystemMenu = ({ className, menu, open, navOpen }: SystemMenuProps) className )} > - <Adjust className="w-6 h-6 fill-current" /> + <Adjust className="w-6 h-6 fill-current text-gray" /> <span className="sr-only">System Menu</span> </DropdownMenu.Trigger> <Route path="/system-menu">
[mod_dirlisting] hide unused variable on MacOS (quiet compiler warning)
@@ -851,7 +851,9 @@ static int http_list_directory(request_st * const r, plugin_data * const p, buff char * const path = malloc(dlen + name_max + 1); force_assert(NULL != path); memcpy(path, dir->ptr, dlen+1); + #if defined(HAVE_XATTR) || defined(HAVE_EXTATTR) || !defined(_ATFILE_SOURCE) char *path_file = path + dlen; + #endif log_error_st * const errh = r->conf.errh; struct dirent *dent;
hfuzz_cc: logging
@@ -183,11 +183,11 @@ static int ldMode(int argc, char **argv) int main(int argc, char **argv) { if (argc <= 1) { - LOG_I("No arguments provided"); + LOG_I("'%s': No arguments provided", argv[0]); return execCC(argc, argv); } if (argc > (ARGS_MAX - 4)) { - LOG_F("Too many positional arguments: %d", argc); + LOG_F("'%s': Too many positional arguments: %d", argv[0], argc); return EXIT_FAILURE; }
server session FEATURE handle expired accounts Fixes
@@ -700,6 +700,9 @@ auth_password_get_pwd_hash(const char *username) if (!spwd) { VRB("Failed to retrieve the shadow entry for \"%s\".", username); return NULL; + } else if ((spwd->sp_expire > -1) && (spwd->sp_expire <= (time(NULL) / (60 * 60 * 24)))) { + WRN("User \"%s\" account has expired.", username); + return NULL; } pass_hash = spwd->sp_pwdp;
controller: fix typo in get_lport_type_str()
@@ -881,7 +881,7 @@ get_lport_type_str(enum en_lport_type lport_type) case LP_CHASSISREDIRECT: return "CHASSISREDIRECT"; case LP_L3GATEWAY: - return "L3GATEWAT"; + return "L3GATEWAY"; case LP_LOCALNET: return "PATCH"; case LP_LOCALPORT:
Coder spec: move function argument error check One less pallene file to compile. Now the same functions are used to test for argument errors and working function calls.
@@ -60,69 +60,6 @@ describe("Pallene coder /", function() end) end) - describe("Function arguments /", function() - setup(compile([[ - function id_int(x: integer): integer - return x - end - - function id_float(x: float): float - return x - end - ]])) - - it("missing arguments", function() - run_test([[ - local ok, err = pcall(test.id_int) - assert(string.find(err, - "wrong number of arguments to function 'id_int', " .. - "expected 1 but received 0", - nil, true)) - ]]) - end) - - it("too many arguments", function() - run_test([[ - local ok, err = pcall(test.id_int, 10, 20) - assert(string.find(err, - "wrong number of arguments to function 'id_int', " .. - "expected 1 but received 2", - nil, true)) - ]]) - end) - - it("type of argument", function() - run_test([[ - local ok, err = pcall(test.id_float, "abc") - assert(string.find(err, - "wrong type for argument x, " .. - "expected float but found string", - nil, true)) - ]]) - end) - - -- See if error messages show float/integer instead of "number": - it("expected float but found integer", function() - run_test([[ - local ok, err = pcall(test.id_float, 10) - assert(string.find(err, - "wrong type for argument x, " .. - "expected float but found integer", - nil, true)) - ]]) - end) - - it("expected float but found integer", function() - run_test([[ - local ok, err = pcall(test.id_int, 3.14) - assert(string.find(err, - "wrong type for argument x, " .. - "expected integer but found float", - nil, true)) - ]]) - end) - end) - describe("Literals /", function() setup(compile([[ function f_nil(): nil return nil end @@ -243,8 +180,41 @@ describe("Pallene coder /", function() it("void functions", function() run_test([[ assert(11 == test.next_x()) ]]) end) + + -- Errors + + it("missing arguments", function() + run_test([[ + local ok, err = pcall(test.g1) + assert(string.find(err, + "wrong number of arguments to function 'g1', " .. + "expected 1 but received 0", + nil, true)) + ]]) + end) + + it("too many arguments", function() + run_test([[ + local ok, err = pcall(test.g1, 10, 20) + assert(string.find(err, + "wrong number of arguments to function 'g1', " .. + "expected 1 but received 2", + nil, true)) + ]]) end) + it("type of argument", function() + -- Also sees if error messages say "float" and "integer" + -- instead of "number" + run_test([[ + local ok, err = pcall(test.g1, 3.14) + assert(string.find(err, + "wrong type for argument x, " .. + "expected integer but found float", + nil, true)) + ]]) + end) + end) describe("Variables /", function() setup(compile([[
Changed an OSPRay exception from ospray::Exception to EXCEPTION1
@@ -339,6 +339,10 @@ avtOSPRaySamplePointExtractor::DoSampling(vtkDataSet *ds, int idx) // Kevin Griffin, Fri Apr 22 16:31:57 PDT 2016 // Added support for polygons. // +// Eric Brugger, Thu Apr 18 14:15:53 PDT 2019 +// Converted the generation of the exception from ospray::Exception +// to EXCEPTION1. +// // **************************************************************************** void @@ -420,13 +424,9 @@ avtOSPRaySamplePointExtractor::RasterBasedSample(vtkDataSet *ds, int num) //--------------------------------------------------------- if (num == 0) { const std::string msg = - "Dataset type " + std::to_string((int)(ds->GetDataObjectType())) + " " - "is not a VTK_RECTILINEAR_GRID. " - "Currently the RayCasting:OSPRay renderer " - "only supports rectilinear grid, " - "thus the volume cannot be rendered\n"; - //ospray::Warning(msg); - ospray::Exception(msg); + "The data wasn't rendered because OSPRay currently only " + "supports volume rendering rectilinear grids."; + EXCEPTION1(ImproperUseException, msg); } } }
VirtIOFS: RHBZ#2024518: fix FUSE_READ input buffer size FUSE_READ input buffer contains only header.
@@ -1588,7 +1588,7 @@ static NTSTATUS Read(FSP_FILE_SYSTEM *FileSystem, PVOID FileContext0, read_in.read.lock_owner = 0; read_in.read.flags = 0; - FUSE_HEADER_INIT(&read_in.hdr, FUSE_READ, FileContext->NodeId, Length); + FUSE_HEADER_INIT(&read_in.hdr, FUSE_READ, FileContext->NodeId, sizeof(read_in.read)); Status = VirtFsFuseRequest(VirtFs->Device, &read_in, sizeof(read_in), read_out, sizeof(*read_out) + Length);
connection_ssl_new log errors to stderr added error logs for evhtp_connection_ssl_new when any context allocation errors occur. (M<3D)
@@ -5318,7 +5318,7 @@ evhtp_connection_ssl_new(struct event_base * evbase, { evhtp_connection_t * conn; struct sockaddr_in sin; - int8_t err; + const char * errstr; if (evbase == NULL) { @@ -5330,20 +5330,24 @@ evhtp_connection_ssl_new(struct event_base * evbase, } conn->evbase = evbase; - err = -1; + errstr = NULL; do { if ((conn->ssl = SSL_new(ctx)) == NULL) { + errstr = "unable to allocate SSL context"; + break; } if ((conn->bev = ssl_sk_new_(evbase, -1, conn->ssl, BUFFEREVENT_SSL_CONNECTING, BEV_OPT_CLOSE_ON_FREE)) == NULL) { + errstr = "unable to allocate bev context"; break; } if (bufferevent_enable(conn->bev, EV_READ) == -1) { + errstr = "unable to enable reading"; break; } @@ -5359,14 +5363,15 @@ evhtp_connection_ssl_new(struct event_base * evbase, if (ssl_sk_connect_(conn->bev, (struct sockaddr *)&sin, sizeof(sin)) == -1) { + errstr = "sk_connect_ failure"; break; } - - err = 0; } while (0); - if (err == -1) { + if (errstr != NULL) { + log_error("%s", errstr); + evhtp_safe_free(conn, evhtp_connection_free); return NULL;
ci: add single parent commit case on check-signed-off-by script
# See the License for the specific language governing permissions and # limitations under the License. -# this retrieves the merge commit created by GH parents=(`git log -n 1 --format=%p HEAD`) -if [[ "${#parents[@]}" -ne 2 ]]; then - echo "This PR's merge commit is missing a parent!" - exit 1 -fi - +if [[ "${#parents[@]}" -eq 1 ]]; then + # CI doesn't use a merge commit + commits=$(git show -s --format=%h ${parents}) +elif [[ "${#parents[@]}" -eq 2 ]]; then + # CI uses a merge commit, eg GH / Travis from="${parents[0]}" into="${parents[1]}" commits=$(git show -s --format=%h ${from}..${into}) +else + echo "Unexpected behavior, cannot verify more than 2 parent commits!" + exit 1 +fi has_commits=false for sha in $commits; do
Rust: Remove error/key TODO
@@ -162,8 +162,6 @@ impl<'a> KDBErrorWrapper<'a> { KDBErrorWrapper { error_key } } - // TODO: For which of these error/* can we be sure that they exist? - /// Returns the error number. pub fn number(&self) -> String { self.error_key
dbuild: Trim white-space from hostname A username of github is used for hostname. Trim white-space from hostname for a case with spaces in the username.
@@ -413,7 +413,7 @@ function BUILD() DOCKER_OPT="-i" fi - HOSTNAME="-h=`git config user.name`" # set github username instead of hostname, "-h=`hostname`" + HOSTNAME="-h=`git config user.name | tr -d ' '`" # set github username instead of hostname, "-h=`hostname`" LOCALTIME="-v /etc/localtime:/etc/localtime:ro" DOCKER_IMAGES=`docker images | grep tizenrt/tizenrt | awk '{print $2}'`
Fix CID 168603
@@ -1315,13 +1315,13 @@ sctp_shutdown(struct socket *so) if (!((inp->sctp_flags & SCTP_PCB_FLAGS_TCPTYPE) || (inp->sctp_flags & SCTP_PCB_FLAGS_IN_TCPPOOL))) { /* Restore the flags that the soshutdown took away. */ -#if (defined(__FreeBSD__) && __FreeBSD_version >= 502115) || defined(__Windows__) SOCKBUF_LOCK(&so->so_rcv); +#if (defined(__FreeBSD__) && __FreeBSD_version >= 502115) || defined(__Windows__) so->so_rcv.sb_state &= ~SBS_CANTRCVMORE; - SOCKBUF_UNLOCK(&so->so_rcv); #else so->so_state &= ~SS_CANTRCVMORE; #endif + SOCKBUF_UNLOCK(&so->so_rcv); /* This proc will wakeup for read and do nothing (I hope) */ SCTP_INP_RUNLOCK(inp); SCTP_LTRACE_ERR_RET(inp, NULL, NULL, SCTP_FROM_SCTP_USRREQ, EOPNOTSUPP);
Fixed issue where the incorrect bone was used for sole effects.
|- ^- (pair (list move) (list sole-effect)) ?~ moves [~ ~] =+ mor=$(moves t.moves) - ?: ?& =(ost.hid p.i.moves) + ?: ?& =(id.cli p.i.moves) ?=({$diff $sole-effect *} q.i.moves) == [p.mor [+>.q.i.moves q.mor]] ?~ q.yop ~ ?~(t.q.yop `i.q.yop `[%mor (flop `(list sole-effect)`q.yop)]) == + ?~ foc + moz + ?~ id.cli + ~& %no-sole + moz ::x produce moves or sole-effects and moves. - ?~(foc moz [[ost.hid %diff %sole-effect u.foc] moz]) + [[id.cli %diff %sole-effect u.foc] moz] :: ++ ra-abet :: complete core ::x applies talk reports, then produces moves and updated state. :: |= fec/sole-effect ^+ +> - +>(moves :_(moves [ost.hid %diff %sole-effect fec])) + +>(moves :_(moves [id.she %diff %sole-effect fec])) :: ++ sh-update ::x adds a talk-update to ++ra's moves
docs(codes): Windows Support: Language/International Inputs
@@ -3383,7 +3383,7 @@ export default [ ], documentation: "https://usb.org/sites/default/files/hut1_2.pdf#page=86", os: { - windows: null, + windows: false, linux: true, android: false, macos: null, @@ -3404,7 +3404,7 @@ export default [ ], documentation: "https://usb.org/sites/default/files/hut1_2.pdf#page=86", os: { - windows: null, + windows: false, linux: true, android: false, macos: null, @@ -3425,7 +3425,7 @@ export default [ ], documentation: "https://usb.org/sites/default/files/hut1_2.pdf#page=86", os: { - windows: null, + windows: false, linux: true, android: false, macos: null, @@ -3446,7 +3446,7 @@ export default [ ], documentation: "https://usb.org/sites/default/files/hut1_2.pdf#page=86", os: { - windows: null, + windows: false, linux: true, android: false, macos: null, @@ -3467,7 +3467,7 @@ export default [ ], documentation: "https://usb.org/sites/default/files/hut1_2.pdf#page=87", os: { - windows: null, + windows: false, linux: true, android: false, macos: null, @@ -3488,7 +3488,7 @@ export default [ ], documentation: "https://usb.org/sites/default/files/hut1_2.pdf#page=87", os: { - windows: null, + windows: false, linux: true, android: false, macos: null, @@ -3572,7 +3572,7 @@ export default [ ], documentation: "https://usb.org/sites/default/files/hut1_2.pdf#page=87", os: { - windows: null, + windows: true, linux: true, android: false, macos: null, @@ -3593,7 +3593,7 @@ export default [ ], documentation: "https://usb.org/sites/default/files/hut1_2.pdf#page=87", os: { - windows: null, + windows: true, linux: true, android: false, macos: null, @@ -3614,7 +3614,7 @@ export default [ ], documentation: "https://usb.org/sites/default/files/hut1_2.pdf#page=87", os: { - windows: null, + windows: false, linux: true, android: false, macos: null, @@ -3635,7 +3635,7 @@ export default [ ], documentation: "https://usb.org/sites/default/files/hut1_2.pdf#page=87", os: { - windows: null, + windows: false, linux: true, android: false, macos: null, @@ -3656,7 +3656,7 @@ export default [ ], documentation: "https://usb.org/sites/default/files/hut1_2.pdf#page=87", os: { - windows: null, + windows: false, linux: true, android: false, macos: null,
Add ifdef guard around water audio
@@ -502,7 +502,9 @@ void update(uint32_t time_ms) { if (water_dist < 0) { water_dist = 0; } +#ifdef __AUDIO__ audio::channels[0].volume = 4000 + (sin(float(time_ms) / 1000.0f) * 3000); +#endif if (game_state == enum_state::menu) { if(pressed & button::B) {
Add WorldId offset to AgentLobby.cs
@@ -13,6 +13,7 @@ namespace FFXIVClientStructs.FFXIV.Client.UI.Agent { [FieldOffset(0x0)] public AgentInterface AgentInterface; [FieldOffset(0x818)] public ulong SelectedCharacterId; [FieldOffset(0x820)] public byte DataCenter; + [FieldOffset(0x824)] public ushort WorldId; [FieldOffset(0x840)] public uint IdleTime; } }
chip/stm32/usart-stm32f4.c: Format with clang-format BRANCH=none TEST=none
@@ -31,7 +31,6 @@ static void usart_variant_enable(struct usart_config const *config) { configs[config->hw->index] = config; - /* Use single-bit sampling */ STM32_USART_CR3(config->hw->base) |= STM32_USART_CR3_ONEBIT;
minor change to java project
<?xml version="1.0" encoding="UTF-8"?> <classpath> <classpathentry kind="src" path="src"/> - <classpathentry kind="con" path="org.eclipse.jdt.launching.JRE_CONTAINER/org.eclipse.jdt.internal.debug.ui.launcher.StandardVMType/jdk7"/> + <classpathentry kind="con" path="org.eclipse.jdt.launching.JRE_CONTAINER"/> <classpathentry kind="output" path="bin"/> </classpath>
Dockerfile: minor fix
@@ -98,8 +98,8 @@ RUN ant -q -f ${CONTIKI_NG}/tools/cooja/build.xml jar # Working directory WORKDIR ${CONTIKI_NG} -# Enable IPv6 -- must be done at runtime, not in Dockerfile -RUN echo "sudo sysctl -w net.ipv6.conf.all.disable_ipv6=0 > /dev/null" >> /home/user/.profile +# Enable IPv6 -- must be done at runtime, hence added to .profile +RUN echo "sudo sysctl -w net.ipv6.conf.all.disable_ipv6=0 > /dev/null" >> ${HOME}/.profile # Start a bash CMD bash --login
ixfr-out, fix to not alloc skipped compression pointer insertions.
@@ -141,14 +141,14 @@ static uint16_t pktcompression_find(struct pktcompression* pcomp, static void pktcompression_insert(struct pktcompression* pcomp, uint8_t* dname, size_t len, uint16_t offset) { - struct rrcompress_entry* entry = pktcompression_alloc(pcomp, - sizeof(*entry)); - if(!entry) - return; + struct rrcompress_entry* entry; if(len > 65535) return; if(offset > 16384) return; /* too far for a compression pointer */ + entry = pktcompression_alloc(pcomp, sizeof(*entry)); + if(!entry) + return; memset(&entry->node, 0, sizeof(entry->node)); entry->node.key = entry; entry->dname = dname;
README.md: Use version 2.05.30
@@ -34,11 +34,11 @@ https://github.com/dresden-elektronik/deconz-rest-plugin/wiki/Supported-Devices ### Install deCONZ 1. Download deCONZ package - wget http://www.dresden-elektronik.de/rpi/deconz/beta/deconz-2.05.29-qt5.deb + wget http://www.dresden-elektronik.de/rpi/deconz/beta/deconz-2.05.30-qt5.deb 2. Install deCONZ package - sudo dpkg -i deconz-2.05.29-qt5.deb + sudo dpkg -i deconz-2.05.30-qt5.deb **Important** this step might print some errors *that's ok* and will be fixed in the next step. @@ -53,11 +53,11 @@ The deCONZ package already contains the REST API plugin, the development package 1. Download deCONZ development package - wget http://www.dresden-elektronik.de/rpi/deconz-dev/deconz-dev-2.05.29.deb + wget http://www.dresden-elektronik.de/rpi/deconz-dev/deconz-dev-2.05.30.deb 2. Install deCONZ development package - sudo dpkg -i deconz-dev-2.05.29.deb + sudo dpkg -i deconz-dev-2.05.30.deb 3. Install missing dependencies @@ -72,7 +72,7 @@ The deCONZ package already contains the REST API plugin, the development package 2. Checkout related version tag cd deconz-rest-plugin - git checkout -b mybranch V2_05_29 + git checkout -b mybranch V2_05_30 3. Compile the plugin
vere: checks on boot that we have write permissions for $pier
@@ -249,6 +249,10 @@ _main_getopt(c3_i argc, c3_c** argv) fprintf(stderr, "normal usage: %s %s\n", argv[0], u3_Host.dir_c); exit(1); } + else if ( 0 != access(u3_Host.dir_c, W_OK) ) { + fprintf(stderr, "urbit: write permissions are required for %s\n", u3_Host.dir_c); + exit(1); + } } c3_t imp_t = ((0 != u3_Host.ops_u.who_c) &&
mINI: Use macros to describe return values
#include <kdbhelper.h> +/* -- Macros ---------------------------------------------------------------------------------------------------------------------------- */ + +#define KEYSET_UNCHANGED 0 +#define KEYSET_MODIFIED 1 + /* -- Functions ------------------------------------------------------------------------------------------------------------------------- */ // =========== @@ -41,19 +46,15 @@ int elektraMiniGet (Plugin * handle ELEKTRA_UNUSED, KeySet * returned, Key * par ksAppend (returned, contract); ksDel (contract); - return 1; // success + return KEYSET_MODIFIED; } - // get all keys - return 1; // success + return KEYSET_UNCHANGED; } int elektraMiniSet (Plugin * handle ELEKTRA_UNUSED, KeySet * returned ELEKTRA_UNUSED, Key * parentKey ELEKTRA_UNUSED) { - // set all keys - // this function is optional - - return 1; // success + return KEYSET_UNCHANGED; }
core: fixed format bug by casting to void*
@@ -1663,7 +1663,7 @@ static void elektraSetCommit (Split * split, Key * parentKey) } keySetName (parentKey, keyName (split->parents[i])); - ELEKTRA_LOG_DEBUG ("elektraSetCommit: %p # %zu with %s - %s\n", backend, p, keyName (parentKey), + ELEKTRA_LOG_DEBUG ("elektraSetCommit: %p # %zu with %s - %s\n", (void *) backend, p, keyName (parentKey), keyString (parentKey)); /* TODO: Remove use of deprecated internal iterator! */
* added SPR_defragVRAM() function
@@ -172,7 +172,7 @@ typedef struct u16 numSprite; VDPSpriteInf **vdpSpritesInf; Collision *collision; - TileSet *tileset; + TileSet *tileset; // TODO: have a tileset per VDP sprite (when rescomp will be optimized for better LZ4W compression) s16 w; s16 h; u16 timer; @@ -420,9 +420,14 @@ Sprite* SPR_addSprite(const SpriteDefinition *spriteDef, s16 x, s16 y, u16 attri void SPR_releaseSprite(Sprite* sprite); /** * \brief - * Returns the number of active sprite (number of sprite added with SPR_addSprite(..) method). + * Returns the number of active sprite (number of sprite added with SPR_addSprite(..) or SPR_addSpriteEx(..) methods). */ u16 SPR_getNumActiveSprite(); +/** + * \brief + * Defragment allocated VRAM for sprites, that can help when sprite allocation fail (SPR_addSprite(..) or SPR_addSpriteEx(..) return <i>NULL</i>). + */ +void SPR_defragVRAM(); /** * \brief
fixed mv(move) command parameter for travis.
@@ -53,7 +53,7 @@ if [ $? -ne 0 ]; then echo -e "\xe2\x9c\x96"; else echo -e "\xe2\x9c\x93"; fi # install 3rd party libraries git clone https://github.com/JonHub/Filters.git echo -n "INSTALL Filters Library: " -DEPENDENCY_OUTPUT=$(mv $HOME/Arduino/libraries/Filters 2>&1) +DEPENDENCY_OUTPUT=$(mv Filters $HOME/Arduino/libraries/Filters 2>&1) if [ $? -ne 0 ]; then echo -e "\xe2\x9c\x96"; else echo -e "\xe2\x9c\x93"; fi # install random lib so the arduino IDE grabs a new library index
Initialize vxlan-gpe bypass mode
@@ -1200,6 +1200,22 @@ VLIB_CLI_COMMAND (set_interface_ip6_vxlan_gpe_bypass_command, static) = { }; /* *INDENT-ON* */ +/* *INDENT-OFF* */ +VNET_FEATURE_INIT (ip4_vxlan_gpe_bypass, static) = +{ + .arc_name = "ip4-unicast", + .node_name = "ip4-vxlan-gpe-bypass", + .runs_before = VNET_FEATURES ("ip4-lookup"), +}; + +VNET_FEATURE_INIT (ip6_vxlan_gpe_bypass, static) = +{ + .arc_name = "ip6-unicast", + .node_name = "ip6-vxlan-gpe-bypass", + .runs_before = VNET_FEATURES ("ip6-lookup"), +}; +/* *INDENT-ON* */ + /** * @brief Feature init function for VXLAN GPE *
Fix initialization of FDW batching in ExecInitModifyTable ExecInitModifyTable has to initialize batching for all result relations, not just the first one. Furthermore, when junk filters were necessary, the pointer pointed past the mtstate->resultRelInfo array. Per reports from multiple non-x86 animals (florican, locust, ...). Discussion:
@@ -2797,9 +2797,16 @@ ExecInitModifyTable(ModifyTable *node, EState *estate, int eflags) * Determine if the FDW supports batch insert and determine the batch * size (a FDW may support batching, but it may be disabled for the * server/table). + * + * We only do this for INSERT, so that for UPDATE/DELETE the batch + * size remains set to 0. */ + if (operation == CMD_INSERT) + { + resultRelInfo = mtstate->resultRelInfo; + for (i = 0; i < nplans; i++) + { if (!resultRelInfo->ri_usesFdwDirectModify && - operation == CMD_INSERT && resultRelInfo->ri_FdwRoutine != NULL && resultRelInfo->ri_FdwRoutine->GetForeignModifyBatchSize && resultRelInfo->ri_FdwRoutine->ExecForeignBatchInsert) @@ -2810,6 +2817,10 @@ ExecInitModifyTable(ModifyTable *node, EState *estate, int eflags) Assert(resultRelInfo->ri_BatchSize >= 1); + resultRelInfo++; + } + } + /* * Lastly, if this is not the primary (canSetTag) ModifyTable node, add it * to estate->es_auxmodifytables so that it will be run to completion by
VERSION bump to version 2.1.62
@@ -64,7 +64,7 @@ endif() # micro version is changed with a set of small changes or bugfixes anywhere in the project. set(SYSREPO_MAJOR_VERSION 2) set(SYSREPO_MINOR_VERSION 1) -set(SYSREPO_MICRO_VERSION 61) +set(SYSREPO_MICRO_VERSION 62) set(SYSREPO_VERSION ${SYSREPO_MAJOR_VERSION}.${SYSREPO_MINOR_VERSION}.${SYSREPO_MICRO_VERSION}) # Version of the library
BugID:25799401: fix issue of adding app Config.in twice
@@ -101,13 +101,14 @@ def write_depends_config(config_file, board, app=None): if comp_info: mandatory_deps = get_comp_mandatory_depends(comp_info, comps) mandatory_deps.append(board) - if app: - mandatory_deps.append(app) for comp in mandatory_deps: if comp in comp_info: config = comp_info[comp]["config_file"] if config: mandatory_configs.append(config) + # do not add app config_file to mandatory_configs + if app: + mandatory_deps.append(app) optional_deps = get_comp_optional_depends(comp_info, mandatory_deps) for comp in optional_deps: if comp["comp_name"] in comp_info:
Feat:fix the problem"undefined reference to `fibocom_task_entry'"
@@ -6,10 +6,8 @@ DESCRIPTION ===========================================================================*/ /* MA510 Includes ---------------------------------------------------------*/ -//#include "test_app_module.h" #include "test_utils.h" #include "qflog_utils.h" -#include "odm_ght_log.h" #include "qapi_uart.h" #include "qapi_timer.h"
oc_csr: return CSR in PEM format instead of DER
@@ -30,38 +30,22 @@ get_csr(oc_request_t *request, oc_interface_mask_t iface_mask, void *data) size_t device = request->resource->device; -#ifdef OC_DYNAMIC_ALLOCATION - unsigned char *csr = - (unsigned char *)calloc(OC_PDU_SIZE, sizeof(unsigned char)); - if (!csr) { - oc_send_response(request, OC_STATUS_INTERNAL_SERVER_ERROR); - return; - } -#else /* OC_DYNAMIC_ALLOCATION */ - unsigned char csr[OC_PDU_SIZE]; -#endif /* !OC_DYNAMIC_ALLOCATION */ + unsigned char csr[4096]; - int csr_len = oc_certs_generate_csr(device, csr, OC_PDU_SIZE); - if (csr_len < 0) { + int ret = oc_certs_generate_csr(device, csr, OC_PDU_SIZE); + if (ret != 0) { oc_send_response(request, OC_STATUS_INTERNAL_SERVER_ERROR); -#ifdef OC_DYNAMIC_ALLOCATION - free(csr); -#endif /* OC_DYNAMIC_ALLOCATION */ return; } oc_rep_start_root_object(); oc_process_baseline_interface( oc_core_get_resource_by_index(OCF_SEC_CSR, device)); - oc_rep_set_byte_string(root, csr, csr, csr_len); - oc_rep_set_text_string(root, encoding, "oic.sec.encoding.der"); + oc_rep_set_text_string(root, csr, (const char *)csr); + oc_rep_set_text_string(root, encoding, "oic.sec.encoding.pem"); oc_rep_end_root_object(); oc_send_response(request, OC_STATUS_OK); - -#ifdef OC_DYNAMIC_ALLOCATION - free(csr); -#endif /* OC_DYNAMIC_ALLOCATION */ } #else /* OC_PKI */
session: fix app attach on error Type: fix
@@ -661,13 +661,12 @@ vl_api_app_attach_t_handler (vl_api_app_attach_t * mp) } done: - - ctrl_thread = vlib_num_workers ()? 1 : 0; - ctrl_mq = session_main_get_vpp_event_queue (ctrl_thread); /* *INDENT-OFF* */ REPLY_MACRO2 (VL_API_APP_ATTACH_REPLY, ({ if (!rv) { + ctrl_thread = vlib_num_workers ()? 1 : 0; + ctrl_mq = session_main_get_vpp_event_queue (ctrl_thread); segp = a->segment; rmp->app_index = clib_host_to_net_u32 (a->app_index); rmp->app_mq = pointer_to_uword (a->app_evt_q);
ci: Run the cancel workflow on generic workers and update action No need to bottleneck on the self hosted worker(s) when this is a light job that needs to run as soon as possible so that it can cancel early.
@@ -15,9 +15,9 @@ on: jobs: cancel-redundant-workflows: - runs-on: [self-hosted, Linux] + runs-on: ubuntu-latest steps: - - uses: styfle/[email protected] + - uses: styfle/[email protected] with: all_but_latest: true workflow_id: ${{ github.event.workflow.id }}
closing procedure improvements
@@ -1611,10 +1611,7 @@ io_readable(neat_ctx *ctx, neat_flow *flow, retval = recv(flow->socket->fd, buffer, 1, MSG_PEEK); if (retval <= 0) { neat_log(ctx, NEAT_LOG_INFO, "%s - TCP connection peek: %d - connection closed", __func__, retval); - if (flow->operations->on_close) { - READYCALLBACKSTRUCT; - flow->operations->on_close(flow->operations); - } + neat_notify_close(flow); return READ_WITH_ZERO; } } @@ -6414,15 +6411,23 @@ neat_notify_close(neat_flow *flow) neat_error_code code = NEAT_ERROR_OK; neat_ctx *ctx = flow->ctx; - neat_log(ctx, NEAT_LOG_DEBUG, "%s", __func__); - if (!flow->operations || !flow->operations->on_close) { + if (flow->state == NEAT_FLOW_CLOSED) { + neat_log(ctx, NEAT_LOG_WARNING, "%s - flow already closed - skipping", __func__); return; } + flow->state = NEAT_FLOW_CLOSED; + + neat_log(ctx, NEAT_LOG_DEBUG, "%s", __func__); + if (flow->operations && flow->operations->on_close) { READYCALLBACKSTRUCT; flow->operations->on_close(flow->operations); } + // this was the last callback - free all ressources + neat_free_flow(flow); +} + // Notify application about network changes. // Code should identify what happened. void @@ -6465,8 +6470,6 @@ neat_close(struct neat_ctx *ctx, struct neat_flow *flow) } #endif - neat_free_flow(flow); - return NEAT_OK; }
vioserial: Remove all uses of bEnableInterruptSuppression The field has no effect and will be removed.
@@ -374,7 +374,6 @@ VIOSerialGetQueueParamCallback( { PPORTS_DEVICE pContext = CONTAINING_RECORD(pVDevice, PORTS_DEVICE, VDevice); - pQueueParam->bEnableInterruptSuppression = false; if (uQueueIndex == 2 || uQueueIndex == 3) { // control queues pQueueParam->Interrupt = pContext->WdfInterrupt;
Fixes a forgotten (gcc only) example function
@@ -75,11 +75,10 @@ static void gccExample(activator_data_t *data) { result = calc->calc(calc->handle, 1); } - celix_service_use_options_t opts; - memset(&opts, 0, sizeof(opts)); + celix_service_use_options_t opts = CELIX_EMPTY_SERVICE_USE_OPTIONS; - opts.serviceName = EXAMPLE_CALC_NAME; - opts.callbackHandle = NULL; //can be null for trampolines + opts.filter.serviceName = EXAMPLE_CALC_NAME; + opts.callbackHandle = NULL; //can be null opts.useWithProperties = use; bool called = celix_bundleContext_useServiceWithOptions(data->ctx, &opts);
Fix SYNOPSIS for ASN1_ENUMERATED_get_int64 and ASN1_ENUMERATED_set_int64
@@ -22,10 +22,10 @@ ASN1_INTEGER_get_int64, ASN1_INTEGER_get, ASN1_INTEGER_set_int64, ASN1_INTEGER_s ASN1_INTEGER *BN_to_ASN1_INTEGER(const BIGNUM *bn, ASN1_INTEGER *ai); BIGNUM *ASN1_INTEGER_to_BN(const ASN1_INTEGER *ai, BIGNUM *bn); - int ASN1_ENUMERATED_get_int64(int64_t *pr, const ASN1_INTEGER *a); + int ASN1_ENUMERATED_get_int64(int64_t *pr, const ASN1_ENUMERATED *a); long ASN1_ENUMERATED_get(const ASN1_ENUMERATED *a); - int ASN1_ENUMERATED_set_int64(ASN1_INTEGER *a, int64_t r); + int ASN1_ENUMERATED_set_int64(ASN1_ENUMERATED *a, int64_t r); int ASN1_ENUMERATED_set(ASN1_ENUMERATED *a, long v); ASN1_ENUMERATED *BN_to_ASN1_ENUMERATED(BIGNUM *bn, ASN1_ENUMERATED *ai);
Fixed issue where messages got stored in the wrong order.
?~ old :: new message %_ +>.$ - grams [gam grams] + grams (welp grams [gam ~]) count +(count) known (~(put by known) uid.tot.gam count) ==
ruby: suppress format overflow warning in generated SWIG code
@@ -56,6 +56,7 @@ if (DEPENDENCY_PHASE) set (SWIG_COMPILE_FLAGS "${SWIG_COMPILE_FLAGS} -Wno-literal-suffix") set (SWIG_COMPILE_FLAGS "${SWIG_COMPILE_FLAGS} -Wno-attributes") set (SWIG_COMPILE_FLAGS "${SWIG_COMPILE_FLAGS} -Wno-pedantic") + set (SWIG_COMPILE_FLAGS "${SWIG_COMPILE_FLAGS} -Wno-format-overflow") endif () set_source_files_properties ("ruby.cpp" PROPERTIES COMPILE_FLAGS "${SWIG_COMPILE_FLAGS}")
Register core/file abstract type
@@ -416,4 +416,7 @@ void janet_lib_io(JanetTable *env) { janet_core_def(env, "stdin", makef(stdin, IO_READ | IO_NOT_CLOSEABLE | IO_SERIALIZABLE), JDOC("The standard input file.")); + + janet_register_abstract_type(&cfun_io_filetype); + }
drv_time: fix debug timer on f0
@@ -14,6 +14,7 @@ volatile unsigned long systickcount = 0; #warning SYS_CLOCK_FREQ_HZ not present #endif +#ifdef F405 void debug_timer_init() { DWT->CTRL |= DWT_CTRL_CYCCNTENA_Msk; @@ -38,8 +39,6 @@ void debug_timer_delay_us(uint32_t us) { } } -#ifdef F405 - // divider by 8 is enabled in this systick config static __INLINE uint32_t SysTick_Config2(uint32_t ticks) { if (ticks > SysTick_LOAD_RELOAD_Msk) @@ -147,8 +146,6 @@ void time_init() { while (1) ; } - - debug_timer_init(); } // called at least once per 16ms or time will overflow @@ -208,4 +205,19 @@ void delay(uint32_t data) { void SysTick_Handler(void) { } + +void debug_timer_init() { +} + +uint32_t debug_timer_micros() { + return gettime(); +} + +uint32_t debug_timer_millis() { + return (debug_timer_micros()) / 1000; +} + +void debug_timer_delay_us(uint32_t us) { + delay(us); +} #endif
vppapigen: supports backwards compatible marking of enums enum bar_enum { BAR1 = 0, BAR2, BAR3 [backwards_compatible], BAR4 = 9 [backwards_compatible], }; This allows adding backwards compatible (as guaranteed by the developer) enums. The enums marked backwards compatible are not considered in the CRC calculation. Type: improvement
@@ -142,11 +142,6 @@ class VPPAPILexer(object): t_ignore = ' \t' -def crc_block_combine(block, crc): - s = str(block).encode() - return binascii.crc32(s, crc) & 0xffffffff - - def vla_is_last_check(name, block): vla = False for i, b in enumerate(block): @@ -227,7 +222,12 @@ class Using(): else: a = {'type': alias.fieldtype} self.alias = a - self.crc = str(alias).encode() + # + # Should have been: + # self.crc = str(alias).encode() + # but to be backwards compatible use the block ([]) + # + self.crc = str(self.block).encode() global_type_add(name, self) def __repr__(self): @@ -303,15 +303,27 @@ class Enum(): self.vla = False count = 0 - for i, b in enumerate(block): - if type(b) is list: - count = b[1] + block2 = [] + block3 = [] + bc_set = False + + for b in block: + if 'value' in b: + count = b['value'] else: count += 1 - block[i] = [b, count] - - self.block = block - self.crc = str(block).encode() + block2.append([b['id'], count]) + try: + if b['option']['backwards_compatible']: + pass + bc_set = True + except KeyError: + block3.append([b['id'], count]) + if bc_set: + raise ValueError("Backward compatible enum must be last {!r} {!r}" + .format(name, b['id'])) + self.block = block2 + self.crc = str(block3).encode() global_type_add(name, self) def __repr__(self): @@ -621,11 +633,19 @@ class VPPAPIParser(object): def p_enum_statement(self, p): '''enum_statement : ID '=' NUM ',' - | ID ',' ''' - if len(p) == 5: - p[0] = [p[1], p[3]] + | ID ',' + | ID '[' field_options ']' ',' + | ID '=' NUM '[' field_options ']' ',' ''' + if len(p) == 3: + p[0] = {'id': p[1]} + elif len(p) == 5: + p[0] = {'id': p[1], 'value': p[3]} + elif len(p) == 6: + p[0] = {'id': p[1], 'option': p[3]} + elif len(p) == 8: + p[0] = {'id': p[1], 'value': p[3], 'option': p[5]} else: - p[0] = p[1] + self._parse_error('ERROR', self._token_coord(p, 1)) def p_field_options(self, p): '''field_options : field_option @@ -934,7 +954,7 @@ def foldup_blocks(block, crc): # Recursively t = global_types[b.fieldtype] try: - crc = crc_block_combine(t.block, crc) + crc = binascii.crc32(t.crc, crc) & 0xffffffff crc = foldup_blocks(t.block, crc) except AttributeError: pass