message
stringlengths 6
474
| diff
stringlengths 8
5.22k
|
---|---|
Try Francois travis fix | @@ -18,9 +18,11 @@ matrix:
python: 3.6
env: TOXENV=py36
- os: osx
+ osx_image: xcode8.3
language: generic
env: TOXENV=py27
- os: osx
+ osx_image: xcode8.3
language: generic
env: TOXENV=py36
install:
|
Modify flash type for sssrw mount to smartfs for artik055s
Set sssrw mount devname to proper point
Mdofiy partition name from none to smartfs properly | @@ -236,7 +236,7 @@ CONFIG_ARTIK05X_AUTOMOUNT_USERFS=y
CONFIG_ARTIK05X_AUTOMOUNT_USERFS_DEVNAME="/dev/smart0p8"
CONFIG_ARTIK05X_AUTOMOUNT_USERFS_MOUNTPOINT="/mnt"
CONFIG_ARTIK05X_AUTOMOUNT_SSSRW=y
-CONFIG_ARTIK05X_AUTOMOUNT_SSSRW_DEVNAME="/dev/smart0p10"
+CONFIG_ARTIK05X_AUTOMOUNT_SSSRW_DEVNAME="/dev/smart0p11"
CONFIG_ARTIK05X_AUTOMOUNT_SSSRW_MOUNTPOINT="/sss"
CONFIG_ARTIK05X_AUTOMOUNT_ROMFS=y
CONFIG_ARTIK05X_AUTOMOUNT_ROMFS_DEVNAME="/dev/mtdblock9"
@@ -249,7 +249,7 @@ CONFIG_ARCH_USE_FLASH=y
CONFIG_FLASH_PARTITION=y
CONFIG_FLASH_MINOR=0
CONFIG_FLASH_PART_LIST="16,48,192,32,512,2400,1536,1536,1000,400,8,512,"
-CONFIG_FLASH_PART_TYPE="none,ftl,none,none,none,none,none,ftl,smartfs,romfs,config,none,"
+CONFIG_FLASH_PART_TYPE="none,ftl,none,none,none,none,none,ftl,smartfs,romfs,config,smartfs,"
CONFIG_FLASH_PART_NAME="bl1,sssro,bl2,sssfw,wlanfw,os,factory,ota,user,rom,nvram,sssrw,"
#
|
fix mman.h ref | @@ -23,7 +23,12 @@ terms of the MIT license. A copy of the license can be found in the file
#include <sys/mman.h> // mmap
#include <unistd.h> // sysconf
#if defined(__linux__)
+#include <features.h>
+#if defined(__GLIBC__)
#include <linux/mman.h> // linux mmap flags
+#else
+#include <sys/mman.h>
+#endif
#endif
#if defined(__APPLE__)
#include <mach/vm_statistics.h>
|
use a static commit hash instead of `master` for *.d files | @@ -5,6 +5,8 @@ SET(CMAKE_CXX_STANDARD 11)
SET(CMAKE_CXX_STANDARD_REQUIRED ON)
SET(CMAKE_CXX_EXTENSIONS OFF)
+SET(H2O_COMMIT "fb0d55a")
+
PROJECT(h2olog)
IF(NOT CMAKE_BUILD_TYPE)
@@ -38,13 +40,13 @@ SET(SOURCE_FILES
ADD_CUSTOM_COMMAND(
OUTPUT "${CMAKE_SOURCE_DIR}/quicly-probes.d"
- COMMAND curl -sLf 'https://raw.githubusercontent.com/h2o/h2o/master/deps/quicly/quicly-probes.d' >${CMAKE_SOURCE_DIR}/quicly-probes.d
+ COMMAND curl -sLf 'https://raw.githubusercontent.com/h2o/h2o/${H2O_COMMIT}/deps/quicly/quicly-probes.d' >${CMAKE_SOURCE_DIR}/quicly-probes.d
)
LIST(APPEND DTRACE_FILES "${CMAKE_SOURCE_DIR}/quicly-probes.d")
ADD_CUSTOM_COMMAND(
OUTPUT "${CMAKE_SOURCE_DIR}/h2o-probes.d"
- COMMAND curl -sLf 'https://raw.githubusercontent.com/h2o/h2o/master/h2o-probes.d' >${CMAKE_SOURCE_DIR}/h2o-probes.d
+ COMMAND curl -sLf 'https://raw.githubusercontent.com/h2o/h2o/${H2O_COMMIT}/h2o-probes.d' >${CMAKE_SOURCE_DIR}/h2o-probes.d
)
LIST(APPEND DTRACE_FILES "${CMAKE_SOURCE_DIR}/h2o-probes.d")
|
dm: Remove hardcode vm_config size from vm_get_config()
vm_config size can be calced by platform_info.sw.max_vms *
platform_info.sw.vm_config_size.
Change vm_get_config() to call IOCTL ACRN_IOCTL_GET_PLATFORM_INFO
twice, first to get platform_info, second to get the vm_configs
content. | @@ -673,29 +673,36 @@ vm_irqfd(struct vmctx *ctx, struct acrn_irqfd *args)
int
vm_get_config(struct vmctx *ctx, struct acrn_vm_config_header *vm_cfg, struct acrn_platform_info *plat_info)
{
-#define VM_CFG_BUFF_SIZE 0x8000
int i, err = 0;
- uint8_t *configs_buff;
+ int configs_size;
+ uint8_t *configs_buff = NULL;
struct acrn_vm_config_header *pcfg;
struct acrn_platform_info platform_info;
if ((ctx == NULL) || (vm_cfg == NULL))
return -1;
- configs_buff = calloc(1, VM_CFG_BUFF_SIZE);
+ /* The first IOCTL to get max_vm and vm_config size of a VM */
+ bzero(&platform_info, sizeof(platform_info));
+ err = ioctl(ctx->fd, ACRN_IOCTL_GET_PLATFORM_INFO, &platform_info);
+ if (err) {
+ pr_err("%s: IOCTL first time failed!\n", __func__);
+ goto exit;
+ }
+
+ configs_size = platform_info.sw.max_vms * platform_info.sw.vm_config_size;
+ configs_buff = calloc(1, configs_size);
if (configs_buff == NULL) {
pr_err("%s, Allocate memory fail.\n", __func__);
return -1;
}
- bzero(&platform_info, sizeof(platform_info));
platform_info.sw.vm_configs_addr = configs_buff;
err = ioctl(ctx->fd, ACRN_IOCTL_GET_PLATFORM_INFO, &platform_info);
if (err) {
- pr_err("%s, failed: get platform info.\n", __func__);
+ pr_err("%s: IOCTL second time failed!\n", __func__);
goto exit;
}
- assert(VM_CFG_BUFF_SIZE > (platform_info.sw.max_vms * platform_info.sw.vm_config_size));
for (i = 0; i < platform_info.sw.max_vms; i++) {
pcfg = (struct acrn_vm_config_header *)(configs_buff + (i * platform_info.sw.vm_config_size));
|
Fix widget bugs | @@ -199,10 +199,12 @@ struct Provider: IntentTimelineProvider {
executingScripts.remove(at: i)
}
ProcessInfo.processInfo.performExpiringActivity(withReason: "Because I want to") { (expired) in
- if widgetsQueue <= 0 {
+
if !expired {
Thread.sleep(forTimeInterval: 3)
}
+
+ if widgetsQueue <= 0 {
exit(0)
}
}
@@ -240,7 +242,7 @@ struct Provider: IntentTimelineProvider {
if (Python.shared.isSetup || RuntimeCommunicator.shared.isContainerAppRunning) && !runningScript {
run()
} else {
- _ = Timer.scheduledTimer(withTimeInterval: 0.2, repeats: true, block: { (timer) in
+ _ = Timer.scheduledTimer(withTimeInterval: 0.01, repeats: true, block: { (timer) in
if (Python.shared.isSetup || RuntimeCommunicator.shared.isContainerAppRunning) && !runningScript {
run()
timer.invalidate()
@@ -277,8 +279,6 @@ struct InAppProvider: IntentTimelineProvider {
func getTimeline(for configuration: SetContentInAppIntent, in context: Context, completion: @escaping (Timeline<ScriptEntry>) -> Void) {
- widgetsQueue += 1
-
if let category = configuration.category {
let saved = UserDefaults.shared?.value(forKey: "savedWidgets") as? [String:Data]
@@ -296,8 +296,6 @@ struct InAppProvider: IntentTimelineProvider {
} else {
completion(Timeline(entries: [ScriptEntry(date: Date(), output: NSLocalizedString("noSelectedScript", comment: ""))], policy: .never))
}
-
- widgetsQueue -= 1
}
}
#endif
|
Run CI scripts without AVM_DISABLE_FP=on with OTP-22+ | @@ -93,19 +93,15 @@ jobs:
- otp: "21"
elixir_version: "1.7"
- cmake_opts: ""
- otp: "22"
elixir_version: "1.8"
- cmake_opts: "-DAVM_DISABLE_FP=on"
- otp: "23"
- elixir_version: "1.10.4"
- cmake_opts: "-DAVM_DISABLE_FP=on"
+ elixir_version: "1.11"
- otp: "24"
- elixir_version: "1.11"
- cmake_opts: "-DAVM_DISABLE_FP=on"
+ elixir_version: "1.14"
- os: "ubuntu-18.04"
cc: "gcc-4.8"
@@ -123,13 +119,22 @@ jobs:
cxx: "g++-10"
cflags: "-m32 -O3"
otp: "23"
- elixir_version: "1.10.4"
- cmake_opts: "-DOPENSSL_CRYPTO_LIBRARY=/usr/lib/i386-linux-gnu/libcrypto.so
- -DAVM_DISABLE_FP=on"
+ elixir_version: "1.11"
+ cmake_opts: "-DOPENSSL_CRYPTO_LIBRARY=/usr/lib/i386-linux-gnu/libcrypto.so"
arch: "i386"
compiler_pkgs: "gcc-10 g++-10 gcc-10-multilib g++-10-multilib libc6-dev-i386
libc6-dbg:i386 zlib1g-dev:i386 libssl-dev:i386"
+ # keep one FP disabled configuration
+ - os: "ubuntu-20.04"
+ cc: "gcc-10"
+ cxx: "g++-10"
+ cflags: "-O3"
+ otp: "24"
+ elixir_version: "1.14"
+ cmake_opts: "-DAVM_DISABLE_FP=on"
+ compiler_pkgs: "gcc-10 g++-10"
+
env:
CC: ${{ matrix.cc }}
CXX: ${{ matrix.cxx }}
|
Core: allow NULL argument to stored_namemap_free(). | @@ -72,10 +72,12 @@ static void stored_namemap_free(void *vnamemap)
{
OSSL_NAMEMAP *namemap = vnamemap;
+ if (namemap != NULL) {
/* Pretend it isn't stored, or ossl_namemap_free() will do nothing */
namemap->stored = 0;
ossl_namemap_free(namemap);
}
+}
static const OPENSSL_CTX_METHOD stored_namemap_method = {
stored_namemap_new,
|
Fix symbol shadow | @@ -117,16 +117,16 @@ static void parseit(void)
static int shouldfail(void)
{
int roll = (int)(random() % 100);
- int shouldfail = roll < md_fail_percent;
+ int shoulditfail = roll < md_fail_percent;
char buff[80];
if (md_tracefd > 0) {
BIO_snprintf(buff, sizeof(buff),
"%c C%ld %%%d R%d\n",
- shouldfail ? '-' : '+', md_count, md_fail_percent, roll);
+ shoulditfail ? '-' : '+', md_count, md_fail_percent, roll);
write(md_tracefd, buff, strlen(buff));
#ifndef OPENSSL_NO_CRYPTO_MDEBUG_BACKTRACE
- if (shouldfail) {
+ if (shoulditfail) {
void *addrs[30];
int num = backtrace(addrs, OSSL_NELEM(addrs));
@@ -141,7 +141,7 @@ static int shouldfail(void)
parseit();
}
- return shouldfail;
+ return shoulditfail;
}
void ossl_malloc_setup_failures(void)
|
[mod_auth] http_auth_const_time_memeq improvement
employ volatile, which might matter with some compilers (or might not)
explicitly check that string lengths match
(or else might match string where last char of short string matches
repeated chars in longer string) | @@ -56,11 +56,22 @@ int http_auth_const_time_memeq (const char *a, const size_t alen, const char *b,
/* constant time memory compare, unless compiler figures it out
* (similar to mod_secdownload.c:const_time_memeq()) */
/* round to next multiple of 64 to avoid potentially leaking exact
- * password length when subject to high precision timing attacks) */
+ * password length when subject to high precision timing attacks)
+ * (not necessary when comparing digests, which have defined lengths)
+ */
+ /* Note: some libs provide similar funcs but might not obscure length, e.g.
+ * OpenSSL:
+ * int CRYPTO_memcmp(const void * in_a, const void * in_b, size_t len)
+ * Note: some OS provide similar funcs but might not obscure length, e.g.
+ * OpenBSD: int timingsafe_bcmp(const void *b1, const void *b2, size_t len)
+ * NetBSD: int consttime_memequal(void *b1, void *b2, size_t len)
+ */
+ const volatile unsigned char * const av = (const unsigned char *)a;
+ const volatile unsigned char * const bv = (const unsigned char *)b;
size_t lim = ((alen >= blen ? alen : blen) + 0x3F) & ~0x3F;
- int diff = 0;
+ int diff = (alen != blen); /*(never match if string length mismatch)*/
for (size_t i = 0, j = 0; lim; --lim) {
- diff |= (a[i] ^ b[j]);
+ diff |= (av[i] ^ bv[j]);
i += (i < alen);
j += (j < blen);
}
|
Clarify homebrew installation under OSX
homebrew/science needs to be tapped first. | @@ -70,12 +70,12 @@ Charles for doing this.
apt-get install bedtools
-
**Homebrew**. Carlos Borroto has made BEDTools available on the bedtools
package manager for OSX.
.. code-block:: bash
+ brew tap homebrew/science
brew install bedtools
**MacPorts**. Alternatively, the MacPorts ports system can be used to install BEDTools on OSX.
|
Rollback d3d12 custom hooking :( | @@ -164,8 +164,7 @@ void D3D12::Hook()
}
else
{
- m_realPresentD3D12 = (decltype(m_realPresentD3D12))ApplyHook(kiero::getSwapChainVtable(), 8, &Present);
- if (m_realPresentD3D12 == nullptr)
+ if (kiero::bind(140, reinterpret_cast<void**>(&m_realPresentD3D12), &Present) != kiero::Status::Success)
{
spdlog::error("{0} Present hook failed!", d3d12type);
++d3d12FailedHooksCount;
@@ -176,8 +175,8 @@ void D3D12::Hook()
++d3d12CompleteHooksCount;
}
- m_realResizeBuffersD3D12 = (decltype(m_realResizeBuffersD3D12))ApplyHook(kiero::getSwapChainVtable(), 13, &ResizeBuffers);
- if (m_realResizeBuffersD3D12 == nullptr) // 13
+ if (kiero::bind(145, reinterpret_cast<void**>(&m_realResizeBuffersD3D12), &ResizeBuffers) !=
+ kiero::Status::Success)
{
spdlog::error("{0} ResizeBuffers hook failed!", d3d12type);
++d3d12FailedHooksCount;
@@ -189,8 +188,8 @@ void D3D12::Hook()
}
}
- m_realExecuteCommandLists = (decltype(m_realExecuteCommandLists))ApplyHook(kiero::getCommandQueueVtable(), 10, &ExecuteCommandLists);
- if (m_realExecuteCommandLists == nullptr) // 10
+ if (kiero::bind(54, reinterpret_cast<void**>(&m_realExecuteCommandLists), &ExecuteCommandLists) !=
+ kiero::Status::Success)
{
spdlog::error("{0} ExecuteCommandLists hook failed!", d3d12type);
++d3d12FailedHooksCount;
|
nrf/gccollect: Use the SP register instead of MSP.
Using the current stack pointer directly saves 8 bytes of code.
We need the *current* register anyway for GC (which is always MSP). | #include "py/gc.h"
#include "gccollect.h"
-static inline uint32_t get_msp(void)
-{
- register uint32_t result;
- __asm volatile ("MRS %0, msp\n" : "=r" (result) );
- return(result);
+static inline uintptr_t get_sp(void) {
+ uintptr_t result;
+ __asm__ ("mov %0, sp\n" : "=r" (result) );
+ return result;
}
void gc_collect(void) {
// start the GC
gc_collect_start();
- mp_uint_t sp = get_msp(); // Get stack pointer
+ // Get stack pointer
+ uintptr_t sp = get_sp();
// trace the stack, including the registers (since they live on the stack in this function)
gc_collect_root((void**)sp, ((uint32_t)&_ram_end - sp) / sizeof(uint32_t));
|
When initializing all the CAN busses, make sure the are also cleared
Thanks to 4vanetten | @@ -165,6 +165,7 @@ bool can_set_speed(uint8_t can_number) {
void can_init_all(void) {
bool ret = true;
for (uint8_t i=0U; i < CAN_MAX; i++) {
+ can_clear(can_queues[i]);
ret &= can_init(i);
}
UNUSED(ret);
|
test-suite: update numpy test for multiple module files | @@ -7,13 +7,21 @@ if [ -s ./common/TEST_ENV ];then
source ./common/TEST_ENV
fi
+declare -A python_module_prefixes=( ["python2"]="py2" ["python3"]="py3")
+if [ "x$DISTRO_FAMILY" == "xCentOS" -o "x$DISTRO_FAMILY" == "xRHEL" ];then
+ declare -A python_package_prefixes=( ["python2"]="python" ["python3"]="python34")
+else
+ declare -A python_package_prefixes=( ["python2"]="python" ["python3"]="python3")
+fi
+
PKG=NUMPY
-module=$python_module_prefix-numpy
-testname=$python-Numpy
-rpm=$python_package_prefix-numpy-$LMOD_FAMILY_COMPILER${DELIM}
+testname=Numpy
-@test "[$testname] Verify $PKG module is loaded and matches rpm version ($LMOD_FAMILY_COMPILER)" {
+@test "[$testname] Verify $PKG modules can be loaded and match rpm version ($LMOD_FAMILY_COMPILER)" {
+ for python in "${!python_module_prefixes[@]}"; do
+ module=${python_module_prefixes[$python]}-numpy
+ rpm=${python_package_prefixes[$python]}-numpy-$LMOD_FAMILY_COMPILER${DELIM}
module list $module | grep "1) $module" >& .cmd_output || exit 1
run grep $module .cmd_output
assert_success
@@ -24,6 +32,7 @@ rpm=$python_package_prefix-numpy-$LMOD_FAMILY_COMPILER${DELIM}
assert_output " 1) $module/$version"
rm -f .cmd_output
+ done
}
@test "[$testname] Verify module ${PKG}_DIR is defined and exists ($LMOD_FAMILY_COMPILER)" {
|
vere: scry out next behn timer for backstop
Instead of potentially waiting ten minutes in the problematic case, we
scry out the next timer from behn and set to that (if we haven't set
a new timer while we were waiting for the scry). | c3_o alm; // alarm
} u3_behn;
+static void _behn_scry_cb(void* vod_p, u3_noun nun);
+
/* _behn_time_cb(): timer callback.
*/
static void
@@ -27,7 +29,7 @@ _behn_time_cb(uv_timer_t* tim_u)
u3_behn* teh_u = tim_u->data;
teh_u->alm = c3n;
- // start another timer for 10 minutes
+ // take initiative to start the next timer, just in case
//
// This is a backstop to deal with the case where a %doze is not
// properly sent, for example after a crash. If the timer continues
@@ -35,9 +37,10 @@ _behn_time_cb(uv_timer_t* tim_u)
// transient error, this will get us past it.
//
{
- c3_d gap_d = 10 * 60 * 1000;
- teh_u->alm = c3y;
- uv_timer_start(&teh_u->tim_u, _behn_time_cb, gap_d, 0);
+ u3_noun pax = u3i_trel(u3i_string("timers"), u3i_string("next"), u3_nul);
+ u3_lord_peek_last(teh_u->car_u.pir_u->god_u, u3_nul,
+ c3_s2('b', 'x'), u3_nul, pax,
+ teh_u, _behn_scry_cb);
}
// send timer event
@@ -82,6 +85,29 @@ _behn_ef_doze(u3_behn* teh_u, u3_noun wen)
u3z(wen);
}
+/* _behn_scry_cb(): next timer scry result callback.
+*/
+static void
+_behn_scry_cb(void* vod_p, u3_noun nun)
+{
+ u3_behn* teh_u = vod_p;
+ u3_weak tim = u3r_at(7, nun);
+
+ if (c3y == teh_u->alm) {
+ // timer already set while we were scrying, no-op
+ }
+ else if (u3_none == tim) {
+ // fall back to a timer for 10 minutes
+ //
+ c3_d gap_d = 10 * 60 * 1000;
+ teh_u->alm = c3y;
+ uv_timer_start(&teh_u->tim_u, _behn_time_cb, gap_d, 0);
+ } else {
+ _behn_ef_doze(teh_u, u3k(tim));
+ }
+ u3z(nun);
+}
+
/* _behn_io_talk(): notify %behn that we're live
*/
static void
|
ipmi-watchdog: Simplify our completion function
This makes no functional changes, just refactors the completion function
to be used for reset only, since it does nothing but free the message
for set calls. This will be useful for future changes to reduce nesting
depth. | @@ -48,27 +48,18 @@ static struct timer wdt_timer;
static bool wdt_stopped;
static bool wdt_ticking;
-static void ipmi_wdt_complete(struct ipmi_msg *msg)
-{
- if (msg->cmd == IPMI_CMD(IPMI_RESET_WDT) && wdt_ticking)
- schedule_timer(&wdt_timer, msecs_to_tb(
- (WDT_TIMEOUT - WDT_MARGIN)*100));
-
- ipmi_free_msg(msg);
-}
-
static void set_wdt(uint8_t action, uint16_t count, uint8_t pretimeout,
bool dont_stop)
{
struct ipmi_msg *ipmi_msg;
ipmi_msg = ipmi_mkmsg(IPMI_DEFAULT_INTERFACE, IPMI_SET_WDT,
- ipmi_wdt_complete, NULL, NULL, 6, 0);
+ ipmi_free_msg, NULL, NULL, 6, 0);
if (!ipmi_msg) {
prerror("Unable to allocate set wdt message\n");
return;
}
- ipmi_msg->error = ipmi_wdt_complete;
+ ipmi_msg->error = ipmi_free_msg;
ipmi_msg->data[0] = TIMER_USE_POST |
TIMER_USE_DONT_LOG |
(dont_stop ? TIMER_USE_DONT_STOP : 0);
@@ -80,12 +71,24 @@ static void set_wdt(uint8_t action, uint16_t count, uint8_t pretimeout,
ipmi_queue_msg(ipmi_msg);
}
+static void reset_wdt_complete(struct ipmi_msg *msg)
+{
+ const uint64_t reset_delay_ms = (WDT_TIMEOUT - WDT_MARGIN) * 100;
+
+ /* If we are inside of skiboot we need to periodically restart the
+ * timer. Reschedule a reset so it happens before the timeout. */
+ if (wdt_ticking)
+ schedule_timer(&wdt_timer, msecs_to_tb(reset_delay_ms));
+
+ ipmi_free_msg(msg);
+}
+
static struct ipmi_msg *wdt_reset_mkmsg(void)
{
struct ipmi_msg *ipmi_msg;
ipmi_msg = ipmi_mkmsg(IPMI_DEFAULT_INTERFACE, IPMI_RESET_WDT,
- ipmi_wdt_complete, NULL, NULL, 0, 0);
+ reset_wdt_complete, NULL, NULL, 0, 0);
if (!ipmi_msg) {
prerror("Unable to allocate reset wdt message\n");
return NULL;
|
handle behn error in acme | :: +wake: timer wakeup event
::
++ wake
- |= [wir=wire ~]
+ |= [wir=wire error=(unit tang)]
^- (quip move _this)
+ ?^ error ~| %acme-timer-failed !!
?> ?=([%acme *] wir)
abet:(retry:event t.wir)
:: +poke-acme-order: create new order for a set of domains
|
Fix bug with calc importance of empty nodes | @@ -185,8 +185,8 @@ TVector<double> CalcEffect(
TVector<double> parentAvrg;
for(int dimension = 0; dimension < approxDimension; dimension++) {
- const double val1 = leftInfo.Values[dimension];
- const double val2 = rightInfo.Values[dimension];
+ const double val1 = count1 ? leftInfo.Values[dimension] : 0;
+ const double val2 = count2 ? rightInfo.Values[dimension] : 0;
const double avrg = (val1 * count1 + val2 * count2) / (sumCount ? sumCount : 1);
const double dif = Sqr(val1 - avrg) * count1 + Sqr(val2 - avrg) * count2;
|
doc: fix kdb validate shell recorder test | @@ -308,43 +308,25 @@ To check if an existing set of keys can be read and written with the current val
```sh
# mount test config file and set a value
-sudo kdb mount test.ini user:/tests/validate range dump
+sudo kdb mount testvalidate.ini /tests/validate range dump
kdb set user:/tests/validate/x 10
# add range check to value
-kdb meta-set user:/tests/validate/x check/range "1-10"
+kdb meta-set spec:/tests/validate/x check/range "1-10"
# check if validate passes
-kdb validate user:/tests/validate
+kdb validate /tests/validate
-# get actual path to mounted file
-CONFIG_FILE=$(kdb file user:/tests/validate)
-echo $CONFIG_FILE
-
-# unmount to change to invalid configuration
-sudo kdb umount user:/tests/validate
-
-# overwrite config with invalid config
-sudo truncate -s 0 $CONFIG_FILE
-# force add range check that that makes value invalid
-sudo echo 'kdbOpen 2' >> $CONFIG_FILE
-sudo echo '$key string 1 2' >> $CONFIG_FILE
-sudo echo 'x' >> $CONFIG_FILE
-sudo echo '51' >> $CONFIG_FILE
-sudo echo '$meta 11 4' >> $CONFIG_FILE
-sudo echo 'check/range' >> $CONFIG_FILE
-sudo echo '1-50' >> $CONFIG_FILE
-
-# mount again
-sudo kdb mount test.ini user:/tests/validate range dump
+# change allowed range
+kdb meta-set -f spec:/tests/validate/x check/range "1-5"
# validation fails now
-kdb validate user:/tests/validate
+kdb validate /tests/validate
# RET:1
# clean up
-kdb rm -r user:/tests/validate
-sudo kdb umount user:/tests/validate
+kdb rm -r /tests/validate
+sudo kdb umount /tests/validate
$end
|
mangle: remove resize as resizing is done by inflate/deflate | @@ -828,11 +828,6 @@ static void mangle_CloneByte(run_t* run) {
run->dynamicFile[off2] = tmp;
}
-static void mangle_Resize(run_t* run) {
- size_t sz = util_rndGet(1, run->global->mutate.maxFileSz);
- input_setSize(run, sz);
-}
-
static void mangle_Expand(run_t* run) {
size_t off = util_rndGet(0, run->dynamicFileSz - 1);
size_t len = util_rndGet(1, run->dynamicFileSz - off);
@@ -881,7 +876,6 @@ static void mangle_ASCIIVal(run_t* run) {
void mangle_mangleContent(run_t* run) {
static void (*const mangleFuncs[])(run_t * run) = {
- mangle_Resize,
mangle_Byte,
mangle_Bit,
mangle_Bytes,
@@ -903,7 +897,6 @@ void mangle_mangleContent(run_t* run) {
};
static void (*const manglePrintableFuncs[])(run_t * run) = {
- mangle_Resize,
mangle_PrintableByte,
mangle_BitPrintable,
mangle_PrintableBytes,
|
fix:fix ethereum test case(test_002InitWallet_0005SetNodeUrlSuccess) | @@ -668,7 +668,7 @@ START_TEST(test_002InitWallet_0005SetNodeUrlSuccess)
/* 2-2. verify the global variables that be affected */
ck_assert_str_eq(wallet_ptr->network_info.node_url_ptr, wallet.node_url_str);
- BoatIotSdkDeInit();
+ BoatEthWalletDeInit(wallet_ptr);
}
END_TEST
|
lyd mods BUGFIX use correct context with errors | @@ -224,7 +224,6 @@ sr_lydmods_add_op_deps(struct lyd_node *sr_mod, const struct lysc_node *op_root)
struct lyd_node *sr_op_deps, *ly_cur_deps;
struct ly_set *set = NULL;
char *data_path = NULL, *xpath = NULL;
- const struct ly_ctx *ly_ctx = op_root->module->ctx;
struct sr_lydmods_deps_dfs_arg dfs_arg;
int is_rpc;
@@ -252,7 +251,7 @@ sr_lydmods_add_op_deps(struct lyd_node *sr_mod, const struct lysc_node *op_root)
/* RPC/notification with path */
if (lyd_new_list(sr_mod, NULL, is_rpc ? "rpc" : "notification", 0, &sr_op_deps, data_path)) {
- sr_errinfo_new_ly(&err_info, ly_ctx);
+ sr_errinfo_new_ly(&err_info, LYD_CTX(sr_mod));
goto cleanup;
}
@@ -260,7 +259,7 @@ sr_lydmods_add_op_deps(struct lyd_node *sr_mod, const struct lysc_node *op_root)
switch (op_root->nodetype) {
case LYS_NOTIF:
if (lyd_new_inner(sr_op_deps, NULL, "deps", 0, &ly_cur_deps)) {
- sr_errinfo_new_ly(&err_info, ly_ctx);
+ sr_errinfo_new_ly(&err_info, LYD_CTX(sr_op_deps));
goto cleanup;
}
@@ -278,7 +277,7 @@ sr_lydmods_add_op_deps(struct lyd_node *sr_mod, const struct lysc_node *op_root)
case LYS_ACTION:
/* input */
if (lyd_new_inner(sr_op_deps, NULL, "in", 0, &ly_cur_deps)) {
- sr_errinfo_new_ly(&err_info, ly_ctx);
+ sr_errinfo_new_ly(&err_info, LYD_CTX(sr_op_deps));
goto cleanup;
}
op_root = lysc_node_child(op_root);
@@ -294,7 +293,7 @@ sr_lydmods_add_op_deps(struct lyd_node *sr_mod, const struct lysc_node *op_root)
/* output */
if (lyd_new_inner(sr_op_deps, NULL, "out", 0, &ly_cur_deps)) {
- sr_errinfo_new_ly(&err_info, ly_ctx);
+ sr_errinfo_new_ly(&err_info, LYD_CTX(sr_op_deps));
goto cleanup;
}
op_root = op_root->next;
|
CI: Fix cache key on Linux lane | @@ -49,9 +49,9 @@ jobs:
uses: actions/cache@v1
with:
path: ./build-ImageMagick
- key: v1-imagemagick-${{ matrix.imagemagick-version.full }}
+ key: v1-linux-imagemagick-${{ matrix.imagemagick-version.full }}
restore-keys: |
- v1-imagemagick-${{ matrix.imagemagick-version.full }}
+ v1-linux-imagemagick-${{ matrix.imagemagick-version.full }}
- name: Set up Ruby ${{ matrix.ruby-version }}
uses: ruby/setup-ruby@master
with:
|
kas: remove 'image-mklibs' from USER_CLASSES list
mklibs is no longer supported, see [1]
[1] | @@ -45,7 +45,7 @@ local_conf_header:
CONF_VERSION = "1"
PACKAGE_CLASSES = "package_rpm"
SDKMACHINE = "x86_64"
- USER_CLASSES = "buildstats image-mklibs image-prelink"
+ USER_CLASSES = "buildstats image-prelink"
PATCHRESOLVE = "noop"
debug-tweaks: |
EXTRA_IMAGE_FEATURES = "debug-tweaks"
|
Fix library version only being set on library unload | @@ -51,6 +51,10 @@ MsQuicLibraryLoad(
CxPlatListInitializeHead(&MsQuicLib.Bindings);
QuicTraceRundownCallback = QuicTraceRundown;
MsQuicLib.Loaded = TRUE;
+ MsQuicLib.Version[0] = VER_MAJOR;
+ MsQuicLib.Version[1] = VER_MINOR;
+ MsQuicLib.Version[2] = VER_PATCH;
+ MsQuicLib.Version[3] = VER_BUILD_ID;
}
}
@@ -68,10 +72,6 @@ MsQuicLibraryUnload(
QUIC_LIB_VERIFY(MsQuicLib.OpenRefCount == 0);
QUIC_LIB_VERIFY(!MsQuicLib.InUse);
MsQuicLib.Loaded = FALSE;
- MsQuicLib.Version[0] = VER_MAJOR;
- MsQuicLib.Version[1] = VER_MINOR;
- MsQuicLib.Version[2] = VER_PATCH;
- MsQuicLib.Version[3] = VER_BUILD_ID;
CxPlatDispatchLockUninitialize(&MsQuicLib.StatelessRetryKeysLock);
CxPlatDispatchLockUninitialize(&MsQuicLib.DatapathLock);
CxPlatLockUninitialize(&MsQuicLib.Lock);
|
fix one more compiler warning | #include <stdlib.h>
#include <stdint.h>
#include <stdio.h>
+
#include <limits.h>
+#include <unistd.h>
#include <sys/types.h>
#include <sys/socket.h>
#include <sys/un.h>
+
#include <netdb.h>
#include <errno.h>
|
tcpm: it83xx: Add support for TYPEC_CC_OPEN
BRANCH=None
TEST=Verify compilation on it83xx w/ IT83XX_PD_EVB = 1.
Commit-Ready: Shawn N
Tested-by: Shawn N | @@ -234,6 +234,14 @@ static void it83xx_enable_vconn(enum usbpd_port port, int enabled)
}
}
+static void it83xx_enable_cc(enum usbpd_port port, int enable)
+{
+ if (enable)
+ CLEAR_MASK(IT83XX_USBPD_CCGCR(port), (1 << 4));
+ else
+ SET_MASK(IT83XX_USBPD_CCGCR(port), (1 << 4));
+}
+
static void it83xx_set_power_role(enum usbpd_port port, int power_role)
{
/* PD_ROLE_SINK 0, PD_ROLE_SOURCE 1 */
@@ -310,12 +318,27 @@ static void it83xx_select_polarity(enum usbpd_port port,
CLEAR_MASK(IT83XX_USBPD_CCGCR(port), (1 << 0));
}
-static void it83xx_set_cc(enum usbpd_port port, int pull)
+static int it83xx_set_cc(enum usbpd_port port, int pull)
{
- if (pull == TYPEC_CC_RD)
+ int enable_cc = 1;
+
+ switch (pull) {
+ case TYPEC_CC_RD:
it83xx_set_power_role(port, PD_ROLE_SINK);
- else if (pull == TYPEC_CC_RP)
+ break;
+ case TYPEC_CC_RP:
it83xx_set_power_role(port, PD_ROLE_SOURCE);
+ break;
+ case TYPEC_CC_OPEN:
+ /* Power-down CC1 & CC2 to remove Rp/Rd */
+ enable_cc = 0;
+ break;
+ default:
+ return EC_ERROR_UNIMPLEMENTED;
+ }
+
+ it83xx_enable_cc(port, enable_cc);
+ return EC_SUCCESS;
}
static int it83xx_tcpm_init(int port)
@@ -363,9 +386,7 @@ static int it83xx_tcpm_select_rp_value(int port, int rp_sel)
static int it83xx_tcpm_set_cc(int port, int pull)
{
- it83xx_set_cc(port, pull);
-
- return EC_SUCCESS;
+ return it83xx_set_cc(port, pull);
}
static int it83xx_tcpm_set_polarity(int port, int polarity)
|
Skyrim: Remove unconnected pin
This pin is NC on more recent board versions.
BRANCH=None
TEST=zmake build skyrim | gpios = <&gpioa 5 GPIO_INPUT>;
};
/* TODO: Add interrupt handler */
- sc_0_int_l {
- gpios = <&gpio6 3 GPIO_INPUT_PULL_UP>;
- };
- /* TODO: Add interrupt handler */
usb_hub_fault_q_odl {
gpios = <&gpioe 5 GPIO_INPUT_PULL_UP>;
};
|
parallel-libs/superlu_dist: add optional MKL support | @@ -78,6 +78,9 @@ module load metis ptscotch
%if "%{compiler_family}" != "intel"
module load openblas
+%define blas_lib -lopenblas
+%else
+%define blas_lib -L$MKL_LIB_PATH -lmkl_intel_ilp64 -lmkl_sequential -lmkl_core -lpthread -lm -ldl
%endif
make SuperLUroot=$(pwd)
@@ -87,7 +90,7 @@ mkdir tmp
mpif90 -z muldefs -shared -Wl,-soname=%{libname}.so.%{major} \
-o ./%{libname}.so.%{version} tmp/*.o -fopenmp -L$METIS_LIB \
-L$OPENBLAS_LIB -L$PTSCOTCH_LIB -lptscotchparmetis \
- -lptscotch -lptscotcherr -lscotch -lmetis -lopenblas \
+ -lptscotch -lptscotcherr -lscotch -lmetis %{blas_lib} \
-lbz2 %{?__global_ldflags}
|
YAML CPP: Stop using namespace `ckdb` for all code | #include <kdbplugin.h>
-using namespace ckdb;
extern "C" {
-int elektraYamlcppGet (Plugin * handle, KeySet * ks, Key * parentKey);
-int elektraYamlcppSet (Plugin * handle, KeySet * ks, Key * parentKey);
+int elektraYamlcppGet (ckdb::Plugin * handle, ckdb::KeySet * ks, ckdb::Key * parentKey);
+int elektraYamlcppSet (ckdb::Plugin * handle, ckdb::KeySet * ks, ckdb::Key * parentKey);
-Plugin * ELEKTRA_PLUGIN_EXPORT (yamlcpp);
+ckdb::Plugin * ELEKTRA_PLUGIN_EXPORT (yamlcpp);
} // end extern "C"
|
readme: add python3 and cmake to requirements | @@ -19,6 +19,8 @@ For convenience, you can alternatively run the `make.sh` script.
### For building h2olog
- LLVM and clang (>= 3.7.1)
+- CMake for generating the build files
+- Python 3 for the [code generator](https://github.com/toru/h2olog/blob/v2/misc/gen-bpf.py)
- [BCC](https://iovisor.github.io/bcc/) (>= 0.11.0) [installed](https://github.com/iovisor/bcc/blob/master/INSTALL.md) on your system
### For running h2olog
|
Add documentation mentioning init-software | @@ -13,6 +13,15 @@ official RISC-V ISA reference implementation. Qemu is a high-performance
functional simulator that can run nearly as fast as native code, but can be
challenging to modify.
+To initialize additional software repositories, such as wrappers for Coremark,
+SPEC2017, and workloads for the NVDLA, run the following script. The
+submodules are located in the ``software`` directory.
+
+.. code-block:: shell
+
+ ./scripts/init-software.sh
+
+
.. toctree::
:maxdepth: 2
:caption: Contents:
|
Fix policy for jaeger client.
Note: mandatory check (NEED_CHECK) was skipped | @@ -279,6 +279,7 @@ ALLOW .* -> vendor/github.com/go-cmd/cmd
ALLOW .* -> vendor/github.com/NVIDIA/gpu-monitoring-tools/bindings/go
# Opentracing interface and implementation.
+ALLOW .* -> vendor/github.com/uber/jaeger-client-go
ALLOW .* -> vendor/github.com/jaegertracing/jaeger-client-go
ALLOW .* -> vendor/github.com/opentracing/opentracing-go
|
Add references to sway-output(5) in sway(5)
update ref in the swaybg_command description
add ref to sway-output(5) in See Also
add an `output` command description | @@ -69,7 +69,7 @@ The following commands may only be used in the configuration file.
*swaybg\_command* <command>
Executes custom background _command_. Default is _swaybg_. Refer to
- *output* below for more information.
+ *sway-output*(5) for more information.
It can be disabled by setting the command to a single dash:
_swaybg\_command -_
@@ -495,6 +495,12 @@ The default colors are:
Prevents windows matching <criteria> from being focused automatically when
they're created. This has no effect on the first window in a workspace.
+*output* <output\_name> <output-subcommands...>
+ For details on output subcommands, see *sway-output*(5).
+
+ \* may be used in lieu of a specific output name to configure all outputs.
+ A list of output names may be obtained via *swaymsg -t get\_outputs*.
+
*popup\_during\_fullscreen* smart|ignore|leave\_fullscreen
Determines what to do when a fullscreen view opens a dialog.
If _smart_ (the default), the dialog will be displayed. If _ignore_, the
@@ -669,4 +675,4 @@ The following attributes may be matched with:
# SEE ALSO
-*sway*(1) *sway-input*(5) *sway-bar*(5)
+*sway*(1) *sway-input*(5) *sway-output*(5) *sway-bar*(5)
|
Correct bootloader | @@ -154,7 +154,7 @@ SUPPORTED_CONFIGURATIONS = [
( 0x8080, 'lpc11u35_ff1705_l151_if', None, None ), # TODO - set target to 'L-TEK-FF1705' when mbed-os supports this
( 0x8081, 'lpc11u35_ff_lpc546xx_if', None, None ), # TODO - set target to 'L-TEK-FF-LPC546XX' when mbed-os supports this
( 0x700, 'stm32f103xb_stm32f103rb_if', 'stm32f103xb_bl', 'ST-Nucleo-F103RB' ),
- ( 0x720, 'stm32f103xb_stm32f401re_if', 'stm32f401re_bl', 'ST-Nucleo-F401RE' ),
+ ( 0x720, 'stm32f103xb_stm32f401re_if', 'stm32f103xb_bl', 'ST-Nucleo-F401RE' ),
( 0x720, 'stm32f103xb_stm32f072rb_if', 'stm32f103xb_bl', 'ST-Nucleo-F072RB' ),
( 0x735, 'stm32f103xb_stm32f334r8_if', 'stm32f103xb_bl', 'ST-Nucleo-F334R8' ),
( 0x740, 'stm32f103xb_stm32f411re_if', 'stm32f103xb_bl', 'ST-Nucleo-F411RE' ),
|
Update token capabilities page with capability guid name | @@ -2017,6 +2017,7 @@ INT_PTR CALLBACK PhpTokenCapabilitiesPageProc(
}
else if (subAuthoritiesCount == SECURITY_CAPABILITY_RID_COUNT)
{
+ PPH_STRING capabilityName;
GUID capabilityGuid;
ULONG firstPart;
ULONG secondPart;
@@ -2034,7 +2035,13 @@ INT_PTR CALLBACK PhpTokenCapabilitiesPageProc(
*((PULONG)&capabilityGuid.Data4[0]) = thirdPart;
*((PULONG)&capabilityGuid.Data4[4]) = lastPart;
- name = PhFormatGuid(&capabilityGuid);
+ if (name = PhFormatGuid(&capabilityGuid))
+ {
+ if (capabilityName = PhGetCapabilityGuidName(name))
+ {
+ PhMoveReference(&name, capabilityName);
+ }
+ }
}
}
}
|
zephyr: Fix 'edefines' typo
Fix a typo that crept into the Kconfig.
BRANCH=none
TEST=build for volteer | @@ -32,7 +32,7 @@ config ZEPHYR
config.h not being available in some headers.
# Define PLATFORM_EC_... options to enable EC features. Each Kconfig should be
-# matched by a line in zephyr/shim/include/config_chip.h which #edefines the
+# matched by a line in zephyr/shim/include/config_chip.h which #defines the
# corresponding EC CONFIG if this Kconfig is enabled.
#
# Please keep these in alphabetical order
|
Display version info
Show version info from metadata, instead of number of blocks. | @@ -393,12 +393,7 @@ void render(uint32_t time) {
screen.text(wrapped_desc, minimal_font, desc_rect);
screen.text(selected_game_metadata.author, minimal_font, Point(172, 208));
-
- int num_blocks = calc_num_blocks(game_list[persist.selected_menu_item].size);
- char buf[20];
- snprintf(buf, 20, "%i block%s", num_blocks, num_blocks == 1 ? "" : "s");
- screen.text(buf, minimal_font, Point(172, 216));
-
+ screen.text(selected_game_metadata.version, minimal_font, Point(172, 224));
}
else {
screen.pen = Pen(235, 245, 255);
|
Docs: Cleaning of fields Generic for SampleCustom.plist because aren't used | <key>AdviseWindows</key>
<false/>
<key>MLB</key>
- <string>M0000000000000001</string>
+ <string></string>
<key>ProcessorType</key>
<integer>0</integer>
<key>ROM</key>
- <data>ESIzRFVm</data>
+ <data></data>
<key>SpoofVendor</key>
- <true/>
+ <false/>
<key>SystemProductName</key>
- <string>iMac19,1</string>
+ <string></string>
<key>SystemSerialNumber</key>
- <string>W00000000001</string>
+ <string></string>
<key>SystemUUID</key>
- <string>00000000-0000-0000-0000-000000000000</string>
+ <string></string>
</dict>
<key>DataHub</key>
<dict>
|
Remove changelog description of UPX to remove from 0.7.4. | @@ -12,9 +12,6 @@ See the AppScope repo to view [all issues](https://github.com/criblio/appscope/i
- **Fix**: [#264](https://github.com/criblio/appscope/issues/264),[#525](https://github.com/criblio/appscope/issues/525) `scope k8s` now (1) supports specifying the namespace to install into, and (2) when the [ConfigMap](https://kubernetes.io/docs/concepts/configuration/configmap/) in that namespace is edited, automatically propagates the changes to all namespaces instrumented with the label `scope=enabled`.
-- **Fix**: [#531](https://github.com/criblio/appscope/issues/531) The AppScope build system now uses UPX for compression, reducing the binary size to about 20 MB.
-
-
## AppScope 0.7.3
2021-09-01 - Maintenance/Hotfix Pre-Release
|
Enable TMPFS and NSH_QUOTE to get CONFIG_NSH_CMDPARMS working | @@ -125,8 +125,9 @@ config NSH_QUOTE
config NSH_CMDPARMS
bool "Enable commands as parameters"
- default !DEFAULT_SMALL
+ default !DEFAULT_SMALL && NSH_QUOTE
depends on !DISABLE_MOUNTPOINT
+ depends on FS_TMPFS
---help---
If selected, then the output from commands, from file applications, and
from NSH built-in commands can be used as arguments to other
@@ -142,6 +143,8 @@ config NSH_CMDPARMS
Because this feature commits significant resources, it is disabled by
default.
+ This features requires TMPFS mounted at /tmp and NSH_QUOTE enabled.
+
config NSH_MAXARGUMENTS
int "Maximum number of command arguments"
default 7
|
[catboost/py] add epsilon dataset to storage;
[catboost/py] add epsilon dataset to storage;
([arc::pullid] b1d19b2b-1f1a64bf-6c97a784-30946227) | @@ -73,7 +73,7 @@ def _get_cache_path():
return os.path.join(os.getcwd(), 'catboost_cached_datasets')
-def _cached_dataset_load(url, md5, dataset_name, train_file, test_file, header='infer'):
+def _cached_dataset_load(url, md5, dataset_name, train_file, test_file, sep=',', header='infer'):
dir_path = os.path.join(_get_cache_path(), dataset_name)
train_path = os.path.join(dir_path, train_file)
test_path = os.path.join(dir_path, test_file)
@@ -86,7 +86,7 @@ def _cached_dataset_load(url, md5, dataset_name, train_file, test_file, header='
_extract(file_path, dir_path)
finally:
os.remove(file_path)
- return pd.read_csv(train_path, header=header), pd.read_csv(test_path, header=header)
+ return pd.read_csv(train_path, header=header, sep=sep), pd.read_csv(test_path, header=header, sep=sep)
def titanic():
@@ -110,6 +110,27 @@ def msrank():
return _cached_dataset_load(url, md5, dataset_name, train_file, test_file, header=None)
+def epsilon():
+ """
+ Download "epsilon" [1] data set.
+
+ Will return two pandas.DataFrame-s, first with train part (epsilon_normalized) and second with
+ test part (epsilon_normalized.t) of the dataset. Object class will be located in the first
+ column of dataset.
+
+ NOTE: This is a preprocessed version of the dataset. It was converted from libsvm format into
+ tsv (CatBoost doesn't support libsvm format out of the box).
+
+ [1]: https://www.csie.ntu.edu.tw/~cjlin/libsvmtools/datasets/binary.html#epsilon
+ """
+ urls = (
+ 'https://proxy.sandbox.yandex-team.ru/785711439',
+ 'https://storage.mds.yandex.net/get-devtools-opensource/250854/epsilon.tar.gz', )
+ md5 = '5bbfac403ac673da7d7ee84bd532e973'
+ dataset_name, train_file, test_file = 'epsilon', 'train.tsv', 'test.tsv'
+ return _cached_dataset_load(urls, md5, dataset_name, train_file, test_file, sep='\t', header=None)
+
+
def adult():
"""
Download "Adult Data Set" [1] from UCI Machine Learning Repository.
|
Fix typo in comment of syncrep.c
Author: Kyotaro Horiguchi
Discussion: | @@ -1065,8 +1065,8 @@ SyncRepUpdateSyncStandbysDefined(void)
/*
* If synchronous_standby_names has been reset to empty, it's futile
- * for backends to continue to waiting. Since the user no longer
- * wants synchronous replication, we'd better wake them up.
+ * for backends to continue waiting. Since the user no longer wants
+ * synchronous replication, we'd better wake them up.
*/
if (!sync_standbys_defined)
{
|
OpenCore: Fix kernel version comparison | @@ -24,8 +24,6 @@ WITHOUT WARRANTIES OR REPRESENTATIONS OF ANY KIND, EITHER EXPRESS OR IMPLIED.
#include <Library/PrintLib.h>
#include <Library/UefiBootServicesTableLib.h>
-#include <Protocol/DevicePath.h>
-
STATIC OC_STORAGE_CONTEXT *mOcStorage;
STATIC OC_GLOBAL_CONFIG *mOcConfiguration;
@@ -58,6 +56,8 @@ OcKernelReadDarwinVersion (
return;
}
+ Offset += L_STR_LEN ("Darwin Kernel Version ");
+
for (Index = 0; Index < DarwinVersionSize - 1; ++Index, ++Offset) {
if (Offset >= KernelSize || Kernel[Offset] == ':') {
break;
@@ -209,7 +209,7 @@ OcKernelApplyPatches (
MatchKernel = OC_BLOB_GET (&UserPatch->MatchKernel);
- if (AsciiStrnCmp (DarwinVersion, MatchKernel, UserPatch->MatchKernel.Size) != 0) {
+ if (AsciiStrnCmp (DarwinVersion, MatchKernel, AsciiStrLen (MatchKernel)) != 0) {
DEBUG ((
DEBUG_INFO,
"OC: Kernel patcher skips %a patch at %u due to version %a vs %a",
@@ -324,7 +324,7 @@ OcKernelBlockKexts (
MatchKernel = OC_BLOB_GET (&Kext->MatchKernel);
- if (AsciiStrnCmp (DarwinVersion, MatchKernel, Kext->MatchKernel.Size) != 0) {
+ if (AsciiStrnCmp (DarwinVersion, MatchKernel, AsciiStrLen (MatchKernel)) != 0) {
DEBUG ((
DEBUG_INFO,
"OC: Kernel blocker skips %a block at %u due to version %a vs %a",
@@ -393,7 +393,7 @@ OcKernelProcessPrelinked (
BundleName = OC_BLOB_GET (&Kext->BundleName);
MatchKernel = OC_BLOB_GET (&Kext->MatchKernel);
- if (AsciiStrnCmp (DarwinVersion, MatchKernel, Kext->MatchKernel.Size) != 0) {
+ if (AsciiStrnCmp (DarwinVersion, MatchKernel, AsciiStrLen (MatchKernel)) != 0) {
DEBUG ((
DEBUG_INFO,
"OC: Prelink injection skips %a kext at %u due to version %a vs %a",
|
VFS: fix possible memory leak in UART | @@ -89,6 +89,8 @@ static esp_line_endings_t s_rx_mode =
ESP_LINE_ENDINGS_LF;
#endif
+static void uart_end_select();
+
// Functions used to write bytes to UART. Default to "basic" functions.
static tx_func_t s_uart_tx_func[UART_NUM] = {
&uart_tx_char, &uart_tx_char, &uart_tx_char
@@ -337,25 +339,25 @@ static esp_err_t uart_start_select(int nfds, fd_set *readfds, fd_set *writefds,
if (_readfds || _writefds || _errorfds || _readfds_orig || _writefds_orig || _errorfds_orig || _signal_sem) {
taskEXIT_CRITICAL(uart_get_selectlock());
- _lock_release(&s_one_select_lock);
+ uart_end_select();
return ESP_ERR_INVALID_STATE;
}
if ((_readfds_orig = malloc(sizeof(fd_set))) == NULL) {
taskEXIT_CRITICAL(uart_get_selectlock());
- _lock_release(&s_one_select_lock);
+ uart_end_select();
return ESP_ERR_NO_MEM;
}
if ((_writefds_orig = malloc(sizeof(fd_set))) == NULL) {
taskEXIT_CRITICAL(uart_get_selectlock());
- _lock_release(&s_one_select_lock);
+ uart_end_select();
return ESP_ERR_NO_MEM;
}
if ((_errorfds_orig = malloc(sizeof(fd_set))) == NULL) {
taskEXIT_CRITICAL(uart_get_selectlock());
- _lock_release(&s_one_select_lock);
+ uart_end_select();
return ESP_ERR_NO_MEM;
}
|
riscv: use far jump to init/fini | .hidden __mlibc_do_dtors
.section .init
- j __mlibc_do_ctors
+ tail __mlibc_do_ctors
.section .fini
- j __mlibc_do_dtors
+ tail __mlibc_do_dtors
.section .ctors
.hidden __CTOR_END__
|
Jenkins: Add `-Werror` to `debian-stable-minimal` | @@ -563,7 +563,8 @@ def generateFullBuildStages() {
tasks << buildAndTest(
"debian-stable-minimal",
DOCKER_IMAGES.stretch_minimal,
- ['PLUGINS': 'NODEP'],
+ ['PLUGINS': 'NODEP'] +
+ CMAKE_FLAGS_WERROR,
[TEST.ALL]
)
|
rewrote the script section | @@ -36,7 +36,11 @@ git:
lfs_skip_smudge: true
script:
- - if [ "$TRAVIS_BRANCH" = "master" ]; then make coverage; else; pytest -v -s tests; fi
+ - if [ "$TRAVIS_BRANCH" = "master" ]; then
+ make coverage;
+ else;
+ pytest -v -s tests;
+ fi
after_success:
|
Undefine ASMNAME/NAME/CNAME before defining them
to avoid redefinition warning when environment variables like CFLAGS are being used (fixes | @@ -1154,6 +1154,7 @@ KERNELDIR = $(TOPDIR)/kernel/$(ARCH)
include $(TOPDIR)/Makefile.$(ARCH)
+CCOMMON_OPT += -UASMNAME -UASMFNAME -UNAME -UCNAME -UCHAR_NAME -UCHAR_CNAME
CCOMMON_OPT += -DASMNAME=$(FU)$(*F) -DASMFNAME=$(FU)$(*F)$(BU) -DNAME=$(*F)$(BU) -DCNAME=$(*F) -DCHAR_NAME=\"$(*F)$(BU)\" -DCHAR_CNAME=\"$(*F)\"
ifeq ($(CORE), PPC440)
|
forward strdup, strndup, and realpath too | @@ -50,10 +50,10 @@ terms of the MIT license. A copy of the license can be found in the file
MI_INTERPOSE_MI(malloc),
MI_INTERPOSE_MI(calloc),
MI_INTERPOSE_MI(realloc),
+ MI_INTERPOSE_MI(free),
MI_INTERPOSE_MI(strdup),
MI_INTERPOSE_MI(strndup),
- MI_INTERPOSE_MI(realpath),
- MI_INTERPOSE_MI(free)
+ MI_INTERPOSE_MI(realpath)
};
#else
// On all other systems forward to our API
@@ -61,6 +61,9 @@ terms of the MIT license. A copy of the license can be found in the file
void* calloc(size_t size, size_t n) mi_attr_noexcept MI_FORWARD2(mi_calloc, size, n);
void* realloc(void* p, size_t newsize) mi_attr_noexcept MI_FORWARD2(mi_realloc, p, newsize);
void free(void* p) mi_attr_noexcept MI_FORWARD0(mi_free, p);
+ char* strdup(const char* s) MI_FORWARD1(mi_strdup, s);
+ char* strndup(const char* s, size_t n) MI_FORWARD2(mi_strndup, s, n);
+ char* realpath(const char* fname, char* resolved_name) MI_FORWARD2(mi_realpath, fname, resolved_name);
#endif
#if (defined(__GNUC__) || defined(__clang__)) && !defined(__MACH__)
@@ -134,11 +137,6 @@ size_t malloc_size(void* p) MI_FORWARD1(mi_usable_size,p);
size_t malloc_usable_size(void *p) MI_FORWARD1(mi_usable_size,p);
void cfree(void* p) MI_FORWARD0(mi_free, p);
-#ifdef __APPLE__
-char* strdup(const char* s) MI_FORWARD1(mi_strdup,s)
-char* strndup(const char* s, size_t n) MI_FORWARD2(mi_strndup,s,n)
-char* realpath(const char* fname, char* resolved_name) MI_FORWARD2(mi_realpath,fname,resolved_name)
-#endif
int posix_memalign(void** p, size_t alignment, size_t size) {
// TODO: the spec says we should return EINVAL also if alignment is not a power of 2.
|
Fix Font atlas expansion; | @@ -1776,22 +1776,22 @@ Font* lovrFontCreate(FontInfo* info) {
map_init(&font->glyphLookup, 36);
map_init(&font->kerning, 36);
+ font->pixelDensity = lovrRasterizerGetHeight(info->rasterizer);
+ font->lineSpacing = 1.f;
+ font->padding = (uint32_t) ceil(info->spread / 2.);
+
// Initial atlas size must be big enough to hold any of the glyphs
float box[4];
font->atlasWidth = 1;
font->atlasHeight = 1;
lovrRasterizerGetBoundingBox(info->rasterizer, box);
- uint32_t width = (uint32_t) ceilf(box[2] - box[0]);
- uint32_t height = (uint32_t) ceilf(box[3] - box[1]);
- while (font->atlasWidth < 2 * width || font->atlasHeight < 2 * height) {
+ uint32_t maxWidth = (uint32_t) ceilf(box[2] - box[0]) + 2 * font->padding;
+ uint32_t maxHeight = (uint32_t) ceilf(box[3] - box[1]) + 2 * font->padding;
+ while (font->atlasWidth < 2 * maxWidth || font->atlasHeight < 2 * maxHeight) {
font->atlasWidth <<= 1;
font->atlasHeight <<= 1;
}
- font->pixelDensity = lovrRasterizerGetHeight(info->rasterizer);
- font->lineSpacing = 1.f;
- font->padding = (uint32_t) ceil(info->spread / 2.);
-
return font;
}
@@ -1853,9 +1853,13 @@ static Glyph* lovrFontGetGlyph(Font* font, uint32_t codepoint) {
uint32_t pixelWidth = 2 * font->padding + (uint32_t) ceilf(width);
uint32_t pixelHeight = 2 * font->padding + (uint32_t) ceilf(height);
+ // If the glyph exceeds the width, start a new row
if (font->atlasX + pixelWidth > font->atlasWidth) {
font->atlasX = font->atlasWidth == font->atlasHeight ? 0 : font->atlasWidth >> 1;
font->atlasY += font->rowHeight;
+ }
+
+ // If the glyph exceeds the height, expand the atlas
if (font->atlasY + pixelHeight > font->atlasHeight) {
if (font->atlasWidth == font->atlasHeight) {
font->atlasX = font->atlasWidth;
@@ -1869,7 +1873,6 @@ static Glyph* lovrFontGetGlyph(Font* font, uint32_t codepoint) {
font->rowHeight = 0;
}
}
- }
glyph->x = font->atlasX + font->padding;
glyph->y = font->atlasY + font->padding;
@@ -1879,7 +1882,7 @@ static Glyph* lovrFontGetGlyph(Font* font, uint32_t codepoint) {
glyph->uv[3] = (uint16_t) ((float) (glyph->y + height) / font->atlasHeight * 65535.f + .5f);
font->atlasX += pixelWidth;
- font->rowHeight = MAX(font->rowHeight, font->atlasY + pixelHeight);
+ font->rowHeight = MAX(font->rowHeight, pixelHeight);
return glyph;
}
|
Sort points vertically in ticLine | @@ -967,8 +967,15 @@ void tic_api_circb(tic_mem* memory, s32 xm, s32 ym, s32 radius, u8 color)
typedef void(*linePixelFunc)(tic_mem* memory, s32 x, s32 y, u8 color);
static void ticLine(tic_mem* memory, s32 x0, s32 y0, s32 x1, s32 y1, u8 color, linePixelFunc func)
{
+ if (y1 < y0)
+ {
+ s32 swap;
+ swap=x0; x0=x1; x1=swap;
+ swap=y0; y0=y1; y1=swap;
+ }
+
s32 dx = abs(x1 - x0), sx = x0 < x1 ? 1 : -1;
- s32 dy = abs(y1 - y0), sy = y0 < y1 ? 1 : -1;
+ s32 dy = y1 - y0;
s32 err = (dx > dy ? dx : -dy) / 2, e2;
for(;;)
@@ -977,7 +984,7 @@ static void ticLine(tic_mem* memory, s32 x0, s32 y0, s32 x1, s32 y1, u8 color, l
if (x0 == x1 && y0 == y1) break;
e2 = err;
if (e2 >-dx) { err -= dy; x0 += sx; }
- if (e2 < dy) { err += dx; y0 += sy; }
+ if (e2 < dy) { err += dx; y0 += 1; }
}
}
|
Ignore result from AT+CIPSERVERMAXCONN as not all AT versions support this yet. | @@ -625,7 +625,7 @@ espi_parse_received(esp_recv_t* rcv) {
*/
if (IS_CURR_CMD(ESP_CMD_TCPIP_CIPSEND)) {
if (esp.msg->msg.conn_send.conn == conn) {
- /** \todo: Find better idea */
+ /** \todo: Find better idea to handle what to do in this case */
//is_error = 1; /* Set as error to stop processing or waiting for connection */
}
}
@@ -1040,12 +1040,11 @@ espi_process_sub_cmd(esp_msg_t* msg, uint8_t is_ok, uint8_t is_error, uint8_t is
*/
if (msg->cmd_def == ESP_CMD_TCPIP_CIPSERVER && msg->msg.tcpip_server.port > 0) {
if (msg->cmd == ESP_CMD_TCPIP_CIPSERVERMAXCONN) {
- if (is_ok) {
+ /* Since not all AT versions support CIPSERVERMAXCONN command, ignore result */
msg->cmd = ESP_CMD_TCPIP_CIPSERVER;
if (espi_initiate_cmd(msg) == espOK) { /* Try to start with new connection */
return espCONT;
}
- }
} else if (msg->cmd == ESP_CMD_TCPIP_CIPSERVER) {
if (is_ok) {
esp.cb_server = msg->msg.tcpip_server.cb; /* Set server callback function */
|
modem/autotest: adding negative test cases to types | @@ -149,5 +149,12 @@ void autotest_modemcf_types()
CONTEND_EQUALITY(liquid_modem_is_specific(LIQUID_MODEM_ARB256OPT), 1);
CONTEND_EQUALITY(liquid_modem_is_specific(LIQUID_MODEM_ARB64VT), 1);
#endif
+
+ // test some negative cases
+ CONTEND_EQUALITY(liquid_modem_is_psk (LIQUID_MODEM_QPSK), 0);
+ CONTEND_EQUALITY(liquid_modem_is_dpsk(LIQUID_MODEM_QPSK), 0);
+ CONTEND_EQUALITY(liquid_modem_is_ask (LIQUID_MODEM_QPSK), 0);
+ CONTEND_EQUALITY(liquid_modem_is_qam (LIQUID_MODEM_QPSK), 0);
+ CONTEND_EQUALITY(liquid_modem_is_apsk(LIQUID_MODEM_QPSK), 0);
}
|
Increase "adb forward" connection attempts
5 seconds might not be sufficient:
<https://github.com/Genymobile/scrcpy/issues/213>
Increase to 10 seconds (it does not harm). | @@ -207,7 +207,7 @@ socket_t server_connect_to(struct server *server) {
if (!server->tunnel_forward) {
server->device_socket = net_accept(server->server_socket);
} else {
- Uint32 attempts = 50;
+ Uint32 attempts = 100;
Uint32 delay = 100; // ms
server->device_socket = connect_to_server(server->local_port, attempts, delay);
}
|
nimble/ll: Fix race in scan stop
We should better disable PHY before cleaning up scansm. This way we
avoid e.g. scenario when extended scanner was active, but PDU was
received after ext_scanning flag was cleared so such PDU would be
marked incorrectly. | @@ -1125,6 +1125,22 @@ ble_ll_scan_sm_stop(int chk_disable)
scansm = &g_ble_ll_scan_sm;
os_cputime_timer_stop(&scansm->scan_timer);
+ /* Only set state if we are currently in a scan window */
+ if (chk_disable) {
+ OS_ENTER_CRITICAL(sr);
+ lls = ble_ll_state_get();
+
+ if ((lls == BLE_LL_STATE_SCANNING) ||
+ (lls == BLE_LL_STATE_INITIATING && chk_disable == 1)) {
+ /* Disable phy */
+ ble_phy_disable();
+
+ /* Set LL state to standby */
+ ble_ll_state_set(BLE_LL_STATE_STANDBY);
+ }
+ OS_EXIT_CRITICAL(sr);
+ }
+
OS_ENTER_CRITICAL(sr);
/* Disable scanning state machine */
@@ -1149,22 +1165,6 @@ ble_ll_scan_sm_stop(int chk_disable)
/* Count # of times stopped */
STATS_INC(ble_ll_stats, scan_stops);
- /* Only set state if we are currently in a scan window */
- if (chk_disable) {
- OS_ENTER_CRITICAL(sr);
- lls = ble_ll_state_get();
-
- if ((lls == BLE_LL_STATE_SCANNING) ||
- (lls == BLE_LL_STATE_INITIATING && chk_disable == 1)) {
- /* Disable phy */
- ble_phy_disable();
-
- /* Set LL state to standby */
- ble_ll_state_set(BLE_LL_STATE_STANDBY);
- }
- OS_EXIT_CRITICAL(sr);
- }
-
/* No need for RF anymore */
OS_ENTER_CRITICAL(sr);
ble_ll_rfmgmt_scan_changed(false, 0);
|
Fix add NULL check for `group->obj_focus` | @@ -218,7 +218,7 @@ void lv_group_focus_obj(lv_obj_t * obj)
if(g->frozen != 0) return;
- if(obj == *g->obj_focus) return;
+ if(g->obj_focus != NULL && obj == *g->obj_focus) return;
/*On defocus edit mode must be leaved*/
lv_group_set_editing(g, false);
|
FURYF4OSD: add blackbox defintion | #define MAX7456_SPI_PORT SPI_PORT2
#define MAX7456_NSS PIN_B12
+#define USE_M25P16
+#define M25P16_SPI_PORT SPI_PORT3
+#define M25P16_NSS_PIN PIN_B3
+
// VOLTAGE DIVIDER
#define VBAT_PIN PIN_C1
#define VBAT_DIVIDER_R1 10000
|
Use wildcards in nuspec for platform-independent packaging. | </metadata>
<files>
<file src="lib/TinySpline.dll" target="lib/net45" />
- <file src="lib/libtinysplinecsharp.so" target="runtimes/@TINYSPLINE_PLATFORM@/lib/native" />
+ <file src="lib/*tinysplinecsharp.*" target="runtimes/@TINYSPLINE_PLATFORM@/lib/native" />
</files>
</package>
|
Work CD-CI
Improve output of GCC toolchain being downloaded. | @@ -7,8 +7,6 @@ steps:
targetType: 'inline'
script: |
- Write-Host "Downloading ARM GNU GCC toolchain..."
-
if("$(GccArm_Version)" -ne "")
{
if("$(GccArm_Version)" -eq "5-2016-q3-update")
@@ -29,6 +27,8 @@ steps:
$url = "https://bintray.com/nfbot/internal-build-tools/download_file?file_path=gcc-arm-none-eabi-7-2018-q2-update-win32.7z"
}
+ Write-Host "Downloading ARM GNU GCC toolchain $(GccArm_Version)..."
+
$output = "$PSScriptRoot\gcc-arm.7z"
(New-Object Net.WebClient).DownloadFile($url, $output)
errorActionPreference: 'stop'
|
apps/system/utils : Modify heapinfo explanation in README.md
Heapinfo was expanded to show kernel, user and specific binary heap information.
Add explanation about target for heapinfo. | @@ -325,11 +325,18 @@ This command shows heap memory usages per thread. This has arguments which confi
```bash
TASH>>heapinfo --help
-Usage: heapinfo [OPTIONS]
+Usage: heapinfo [-TARGET] [-OPTION]
Display information of heap memory
Options:
- -i Initialize the heapinfo
+[-TARGET]
+ -k Kernel Heap
+ (-k target is available when CONFIG_BUILD_PROTECTED is enabled)
+ -u User Heap
+ (-u target is available when CONFIG_BUILD_PROTECTED is enabled)
+ -b BIN_NAME Heap for BIN_NAME binary
+ (-b target is available when CONFIG_APP_BINARY_SEPARATION is enabled)
+[-OPTION]
-a Show the all allocation details
-p PID Show the specific PID allocation details
-f Show the free list
@@ -339,7 +346,6 @@ Options:
(-e option is available when CONFIG_MM_NHEAPS is greater than 1)
-r Show the all region information
(-r option is available when CONFIG_MM_REGIONS is greater than 1)
- -k OPTION Show the kernel heap memory allocation details based on above options
TASH>>heapinfo
|
Fix gall scry
cf. fixes | ==
::
++ scry
- |= {fur/(unit (set monk)) ren/@tas who/ship syd/desk lot/coin tyl/path}
+ |= {fur/(unit (set monk)) ren/@tas why/shop syd/desk lot/coin tyl/path}
^- (unit (unit cage))
+ ?. ?=($& -.why) ~
+ =* his p.why
?: ?& =(%u ren)
=(~ tyl)
=([%$ %da now] lot)
- (~(has by pol.all) who)
- (~(has by bum:(~(got by pol.all) who)) syd)
+ (~(has by pol.all) his)
+ (~(has by bum:(~(got by pol.all) his)) syd)
==
``[%null !>(~)]
- ?. (~(has by pol.all) who)
+ ?. (~(has by pol.all) his)
~
?. =([%$ %da now] lot)
~
- ?. (~(has by bum:(~(got by pol.all) who)) syd)
+ ?. (~(has by bum:(~(got by pol.all) his)) syd)
[~ ~]
?. ?=(^ tyl)
~
- (mo-peek:(mo-abed:mo who *duct) syd high+`who ren tyl)
+ (mo-peek:(mo-abed:mo his *duct) syd high+`his ren tyl)
::
++ stay :: save w+o cache
`axle`all
|
[iovisor/bcc] trace: Incorrect symbol offsets when using build_id
Fix - bcc_bsymcache API returned absolute address instead of
offset from start of the symbol | @@ -424,7 +424,7 @@ bool BuildSyms::Module::resolve_addr(uint64_t offset, struct bcc_symbol* sym,
sym->name = (*it).name->c_str();
if (demangle)
sym->demangle_name = sym->name;
- sym->offset = (*it).start;
+ sym->offset = offset - (*it).start;
sym->module = module_name_.c_str();
return true;
}
|
Fix memory leak on lookup failure | @@ -129,6 +129,9 @@ static EVP_PKEY_CTX *int_ctx_new(EVP_PKEY *pkey, ENGINE *e, int id)
pmeth = EVP_PKEY_meth_find(id);
if (pmeth == NULL) {
+#ifndef OPENSSL_NO_ENGINE
+ ENGINE_finish(e);
+#endif
EVPerr(EVP_F_INT_CTX_NEW, EVP_R_UNSUPPORTED_ALGORITHM);
return NULL;
}
|
chore(docs) add "v" prefix | {% block footer %}
+<style>
+ .wy-side-nav-search > div[role="search"] {
+ color: black;
+ }
+</style>
<script type="text/javascript">
$(document).ready(function() {
$(".toggle > *").hide();
@@ -22,7 +27,7 @@ function add_version_selector()
const versions = text.split("\n").filter(version => version.trim().length > 0);
let p = document.getElementById("rtd-search-form").parentElement;
p.innerHTML = `
- <select name="versions" id="versions" onchange="ver_sel()" style="border-radius:5px; margin-bottom:15px">
+ v: <select name="versions" id="versions" onchange="ver_sel()" style="border-radius:5px; margin-bottom:15px">
${versions.map(version => {
const versionName = (version.indexOf(".") != -1) ? version : (version + " (latest minor)");
return `<option value="${version}">${versionName}</option>`;
|
Set "Maintenance Req'd" bit in GDL90 heartbeat when there exists a system error. | @@ -574,6 +574,11 @@ func makeHeartbeat() []byte {
}
msg[1] = msg[1] | 0x10 //FIXME: Addr talkback.
+ // "Maintenance Req'd". Add flag if there are any current critical system errors.
+ if len(globalStatus.Errors) > 0 {
+ msg[1] = msg[1] | 0x40
+ }
+
nowUTC := time.Now().UTC()
// Seconds since 0000Z.
midnightUTC := time.Date(nowUTC.Year(), nowUTC.Month(), nowUTC.Day(), 0, 0, 0, 0, time.UTC)
|
arch: fix typo in MSVC x86_64 detection
Fixes | /* AMD64 / x86_64
<https://en.wikipedia.org/wiki/X86-64> */
-#if defined(__amd64__) || defined(__amd64) || defined(__x86_64__) || defined(__x86_64) || defined(_M_X66) || defined(_M_AMD64)
+#if defined(__amd64__) || defined(__amd64) || defined(__x86_64__) || defined(__x86_64) || defined(_M_X64) || defined(_M_AMD64)
# define SIMDE_ARCH_AMD64 1000
#endif
|
[core] reduce trace on Upgrade backend connection
reduce trace on Upgrade'd backend connection when ECONNRESET received,
which, for example, apparently might occur if a backend calls close()
on socket without first calling shutdown(fd, SHUT_WR) -- seen on Linux
kernel 5.16.15 where lighttpd received ECONNRESET when trying to read()
(instead of receiving EOF). | @@ -2412,7 +2412,8 @@ static handler_t gw_recv_response_error(gw_handler_ctx * const hctx, request_st
"socket: %s for %s?%.*s, closing connection",
(long long)hctx->wb.bytes_out, proc->connection_name->ptr,
r->uri.path.ptr, BUFFER_INTLEN_PTR(&r->uri.query));
- } else if (!light_btst(r->resp_htags, HTTP_HEADER_UPGRADE)) {
+ } else if (!light_btst(r->resp_htags, HTTP_HEADER_UPGRADE)
+ && !r->h2_connect_ext) {
log_error(r->conf.errh, __FILE__, __LINE__,
"response already sent out, but backend returned error on "
"socket: %s for %s?%.*s, terminating connection",
|
update ci to exit at the end instead of partway through | fix flattened toolchain repos | @@ -21,29 +21,66 @@ git config submodule.vlsi/hammer-synopsys-plugins.update none
git submodule update --init
status=$(git submodule status)
+all_names=()
+
search () {
for submodule in "${submodules[@]}"
do
echo "Running check on submodule $submodule in $dir"
- hash=$(echo "$status" | grep $submodule | awk '{print$1}' | grep -o "[[:alnum:]]*")
- echo "Searching for $hash in origin/master of $submodule"
- git -C $dir/$submodule branch -r --contains "$hash" | grep "origin/master" # needs init'ed submodules
+ hash=$(echo "$status" | grep "$dir.*$submodule " | awk '{print$1}' | grep -o "[[:alnum:]]*")
+ echo "Searching for $hash in origin/$branch of $submodule"
+ (git -C $dir/$submodule branch -r --contains "$hash" | grep "origin/$branch") && true # needs init'ed submodules
+ all_names+=("$dir/$submodule $hash $?")
done
}
submodules=("boom" "hwacha" "icenet" "rocket-chip" "sifive-blocks" "sifive-cache" "testchipip")
dir="generators"
+branch="master"
+
+search
+
+submodules=("riscv-gnu-toolchain" "riscv-isa-sim" "riscv-pk" "riscv-tests")
+dir="toolchains/esp-tools"
+branch="master"
+
+search
+
+
+submodules=("riscv-gnu-toolchain" "riscv-isa-sim" "riscv-pk" "riscv-tests" "riscv-gnu-toolchain-prebuilt")
+dir="toolchains/riscv-tools"
+branch="master"
search
-submodules=("esp-tools" "riscv-tools")
-dir="toolchains"
+# riscv-openocd doesn't use its master branch
+submodules=("riscv-openocd")
+dir="toolchains/riscv-tools"
+branch="riscv"
search
submodules=("barstools" "chisel3" "firrtl" "torture")
dir="tools"
+branch="master"
search
+# turn off verbose printing to make this easier to read
+set +x
+
+# print all result strings
+for str in "${all_names[@]}";
+do
+ echo "$str"
+done
+
+# check if there was a non-zero return code
+for str in "${all_names[@]}";
+do
+ if [ ! 0 = $(echo "$str" | awk '{print$3}') ]; then
+ exit 1
+ fi
+done
+
echo "Done checking all submodules"
|
cmd/kubectl: Add possible sort values for filetop CLI. | @@ -101,7 +101,7 @@ var filetopCmd = &cobra.Command{
func init() {
filetopCmd.Flags().IntVarP(&maxRows, "maxrows", "r", 20, "Maximum rows to print")
- filetopCmd.Flags().StringVarP(&sortByStr, "sort", "", "rbytes", "Sort column")
+ filetopCmd.Flags().StringVarP(&sortByStr, "sort", "", "rbytes", fmt.Sprintf("Sort column (%s)", strings.Join(types.SortBySlice, ", ")))
filetopCmd.Flags().BoolVarP(&allFiles, "all-files", "a", false, "Include non-regular file types (sockets, FIFOs, etc)")
rootCmd.AddCommand(filetopCmd)
|
contact-store: enforce timestamp ordering on %initial | |= [rolo=rolodex:store is-public=?]
^- (quip card _state)
=/ our-contact (~(got by rolodex) our.bowl)
- =. rolodex (~(uni by rolodex) rolo)
+ ::
+ =/ diff-rolo=rolodex:store
+ %- ~(gas by *rolodex:store)
+ %+ skim ~(tap in rolo)
+ |= [=ship =contact:store]
+ ?~ local-con=(~(get by rolodex) ship) %.y
+ (gth last-updated.contact last-updated.u.local-con)
+ =. rolodex (~(uni by rolodex) diff-rolo)
=. rolodex (~(put by rolodex) our.bowl our-contact)
:_ state(rolodex rolodex)
(send-diff [%initial rolodex is-public] %.n)
|
lib: msgpack-c: do not install static library | @@ -230,7 +230,7 @@ ENDIF ()
IF (MSGPACK_ENABLE_SHARED)
SET (MSGPACK_INSTALLTARGETS msgpackc msgpackc-static)
ELSE()
- SET (MSGPACK_INSTALLTARGETS msgpackc-static)
+ # SET (MSGPACK_INSTALLTARGETS msgpackc-static)
ENDIF ()
INSTALL (TARGETS ${MSGPACK_INSTALLTARGETS} RUNTIME DESTINATION ${CMAKE_INSTALL_BINDIR}
|
Blacklist support
Initial blacklist: Battle.net.exe, EpicGamesLauncher.exe, IGOProxy.exe, IGOProxy64.exe, Origin.exe, Steam.exe | @@ -2457,8 +2457,45 @@ static const struct {
#undef ADD_HOOK
};
+static bool checkBlacklisted()
+{
+ std::vector<std::string> blacklist {
+ "Battle.net.exe",
+ "EpicGamesLauncher.exe",
+ "IGOProxy.exe",
+ "IGOProxy64.exe",
+ "Origin.exe",
+ "OriginThinSetupInternal.exe",
+ "Steam.exe",
+ };
+
+#ifdef _GNU_SOURCE_OFF
+ std::string p(program_invocation_name);
+ std::string procName = p.substr(p.find_last_of("/\\") + 1);
+#else
+ std::string p = get_exe_path();
+ std::string procName;
+ if (ends_with(p, "wine-preloader") || ends_with(p, "wine64-preloader")) {
+ get_wine_exe_name(procName, true);
+ } else {
+ procName = p.substr(p.find_last_of("/\\") + 1);
+ }
+#endif
+
+ return std::find(blacklist.begin(), blacklist.end(), procName) != blacklist.end();
+}
+
+static bool isBlacklisted = checkBlacklisted();
+
static void *find_ptr(const char *name)
{
+ std::string f(name);
+
+ if (isBlacklisted && (f != "vkCreateInstance" && f != "vkDestroyInstance" && f != "vkCreateDevice" && f != "vkDestroyDevice"))
+ {
+ return NULL;
+ }
+
for (uint32_t i = 0; i < ARRAY_SIZE(name_to_funcptr_map); i++) {
if (strcmp(name, name_to_funcptr_map[i].name) == 0)
return name_to_funcptr_map[i].ptr;
|
Update USAGE-GUIDE.md for multicert | @@ -559,6 +559,8 @@ int s2n_config_add_cert_chain_and_key_to_store(struct s2n_config *config,
**s2n_config_add_cert_chain_and_key_to_store** is the preferred method of associating a certificate chain and private key pair with an **s2n_config** object. At present, this may only be called once for each config object. It is not recommended to free or modify the **cert_key_pair** as any subsequent changes will be reflected in the config.
+**s2n_config_add_cert_chain_and_key_to_store** may be called multiple times to support multiple key types(RSA, ECDSA) and multiple domains. On the server side, the certificate selected will be based on the incoming SNI value and the client's capabilities(supported ciphers). In the case of no certificate matching the client's SNI extension or if no SNI extension was sent by the client, the certificate from the **first** call to **s2n_config_add_cert_chain_and_key_to_store** will be selected.
+
### s2n\_config\_add\_dhparams
```c
|
Add ommitted assignment for fd in dup2 tc
Although negative tc is failed below, fd1 can be positive value because of tc success before
So fd should be assigned as negative value | @@ -425,24 +425,25 @@ static void fs_vfs_dup2_tc(void)
/* now fd1 points fd2 */
ret = dup2(fd2, fd1);
- TC_ASSERT_GEQ_CLEANUP("dup2", ret, 0, close(fd2));
+ close(fd2);
+ TC_ASSERT_NEQ("dup2", ret, ERROR);
+ TC_ASSERT_GEQ("dup2", fd1, 0);
len = strlen(VFS_TEST_CONTENTS_3);
ret = write(fd1, str, len);
close(fd1);
- TC_ASSERT_EQ_CLEANUP("write", ret, len, close(fd2));
- close(fd2);
+ TC_ASSERT_EQ("write", ret, len);
fd2 = open(filename2, O_RDONLY);
TC_ASSERT_GEQ("open", fd2, 0);
ret = read(fd2, buf, len);
close(fd2);
-
TC_ASSERT_GT("read", ret, 0);
TC_ASSERT_EQ("read", strcmp(buf, VFS_TEST_CONTENTS_3), 0);
/* Nagative case with invalid argument, invalid fd. It will return ERROR */
+ fd1 = -1;
ret = dup2(CONFIG_NFILE_DESCRIPTORS + CONFIG_NSOCKET_DESCRIPTORS, fd1);
TC_ASSERT_LT_CLEANUP("dup2", fd1, 0, close(fd1));
TC_ASSERT_EQ("dup2", ret, ERROR);
|
Expost static and dynamic Swift packages | @@ -14,6 +14,16 @@ let package = Package(
name: "TrustKit",
targets: ["TrustKit"]
),
+ .library(
+ name: "TrustKitDynamic",
+ type: .dynamic,
+ targets: ["TrustKit"]
+ ),
+ .library(
+ name: "TrustKitStatic",
+ type: .static,
+ targets: ["TrustKit"]
+ ),
],
dependencies: [
],
|
Use MEM_CONTEXT_BEGIN() block instead of memContextSwitch().
This is the standard pattern but was missed here. | @@ -827,12 +827,12 @@ storageS3PathRemoveCallback(StorageS3 *this, void *callbackData, const String *n
// If there is something to delete then create the request
if (data->xml == NULL)
{
- MemContext *memContextOld = memContextSwitch(data->memContext);
-
+ MEM_CONTEXT_BEGIN(data->memContext)
+ {
data->xml = xmlDocumentNew(S3_XML_TAG_DELETE_STR);
xmlNodeContentSet(xmlNodeAdd(xmlDocumentRoot(data->xml), S3_XML_TAG_QUIET_STR), TRUE_STR);
-
- memContextSwitch(memContextOld);
+ }
+ MEM_CONTEXT_END();
}
// Add to delete list
|
Fix state restoration | @@ -102,15 +102,16 @@ extension SceneDelegate {
_ = Timer.scheduledTimer(withTimeInterval: 0.1, repeats: true, block: { (timer) in
if let doc = self.documentBrowserViewController, doc.view.window != nil {
- let sidebar = doc.makeSidebarNavigation(url: url, run: false, isShortcut: false)
- let vc = UIHostingController(rootView: sidebar)
+ let sidebar = doc.makeSidebarNavigation(url: url, run: false, isShortcut: false, restoreSelection: true)
+ let vc = UIHostingController(rootView: AnyView(sidebar))
vc.modalPresentationStyle = .fullScreen
vc.modalTransitionStyle = .crossDissolve
sidebar.viewControllerStore.vc = vc
doc.present(vc, animated: true, completion: nil)
- timer.invalidate()
}
+
+ timer.invalidate()
})
} catch {
print(error.localizedDescription)
|
all: sets scrollbar colors
fixes urbit/landscape#87 | @@ -39,6 +39,24 @@ const Root = styled.div`
}
display: flex;
flex-flow: column nowrap;
+
+ * {
+ scrollbar-width: thin;
+ scrollbar-color: ${ p => p.theme.colors.gray } ${ p => p.theme.colors.white };
+ }
+
+ /* Works on Chrome/Edge/Safari */
+ *::-webkit-scrollbar {
+ width: 12px;
+ }
+ *::-webkit-scrollbar-track {
+ background: ${ p => p.theme.colors.white };
+ }
+ *::-webkit-scrollbar-thumb {
+ background-color: ${ p => p.theme.colors.gray };
+ border-radius: 1rem;
+ border: 3px solid ${ p => p.theme.colors.white };
+ }
`;
const StatusBarWithRouter = withRouter(StatusBar);
|
timer_reset: do not do anything if the timer has not expired | @@ -74,9 +74,8 @@ timer_set(struct timer *t, clock_time_t interval)
* given to the timer_set() function. The start point of the interval
* is the exact time that the timer last expired. Therefore, this
* function will cause the timer to be stable over time, unlike the
- * timer_restart() function.
- *
- * \note Must not be executed before timer expired
+ * timer_restart() function. If this is executed before the
+ * timer expired, this function has no effect.
*
* \param t A pointer to the timer.
* \sa timer_restart()
@@ -84,8 +83,10 @@ timer_set(struct timer *t, clock_time_t interval)
void
timer_reset(struct timer *t)
{
+ if(timer_expired(t)) {
t->start += t->interval;
}
+}
/*---------------------------------------------------------------------------*/
/**
* Restart the timer from the current point in time
|
extstore: make ext_recache_rate=0 work
would divide by zero before :) now short circuits whole function. | @@ -642,8 +642,10 @@ conn *conn_new(const int sfd, enum conn_states init_state,
static void recache_or_free(conn *c, io_wrap *wrap) {
item *it;
it = (item *)wrap->io.buf;
+ bool do_free = true;
// If request was ultimately a miss, unlink the header.
if (wrap->miss) {
+ do_free = false;
size_t ntotal = ITEM_ntotal(wrap->hdr_it);
item_unlink(wrap->hdr_it);
slabs_free(it, ntotal, slabs_clsid(ntotal));
@@ -652,10 +654,9 @@ static void recache_or_free(conn *c, io_wrap *wrap) {
if (wrap->badcrc)
c->thread->stats.badcrc_from_extstore++;
pthread_mutex_unlock(&c->thread->stats.mutex);
- } else {
+ } else if (settings.ext_recache_rate) {
// hashvalue is cuddled during store
uint32_t hv = (uint32_t)it->time;
- bool do_free = true;
// opt to throw away rather than wait on a lock.
void *hold_lock = item_trylock(hv);
if (hold_lock != NULL) {
@@ -678,12 +679,12 @@ static void recache_or_free(conn *c, io_wrap *wrap) {
pthread_mutex_unlock(&c->thread->stats.mutex);
}
}
-
- if (do_free)
- slabs_free(it, ITEM_ntotal(it), ITEM_clsid(it));
if (hold_lock)
item_trylock_unlock(hold_lock);
}
+ if (do_free)
+ slabs_free(it, ITEM_ntotal(it), ITEM_clsid(it));
+
wrap->io.buf = NULL; // sanity.
wrap->io.next = NULL;
wrap->next = NULL;
|
Install prerequisites | @@ -14,6 +14,8 @@ jobs:
steps:
- uses: actions/checkout@v1
+ - name: Install prerequisites
+ run: sudo apt-get install gcc-multilib
- name: Run CMake
env:
CC: ${{ matrix.compiler }}
|
Enable node modules in bootstrap again, Guix will try to copy them manually from cherow. | @@ -59,12 +59,8 @@ install(FILES
COMPONENT runtime
)
-if(NOT OPTION_BUILD_GUIX)
- # In Guix Bootstrap dependencies (aka Cherow
- # at the moment of writting) must be installed manually
install(DIRECTORY
${LOADER_LIBRARY_PATH}/node_modules
DESTINATION ${INSTALL_LIB}
COMPONENT runtime
)
-endif()
|
Have raczlib check the dictionary TTag | @@ -316,7 +316,7 @@ func (r *Reader) nextChunk() error {
dict := []byte(nil)
if !chunk.CSecondary.Empty() {
- if dict, err = r.loadDictionary(chunk.CSecondary); err != nil {
+ if dict, err = r.loadDictionary(chunk.CSecondary, chunk.TTag); err != nil {
return err
}
}
@@ -358,15 +358,15 @@ func (r *Reader) nextChunk() error {
//
// For a description of the RAC+Zlib secondary-data format, see
// https://github.com/google/wuffs/blob/master/doc/spec/rac-spec.md#rac--zlib
-func (r *Reader) loadDictionary(cRange rac.Range) ([]byte, error) {
+func (r *Reader) loadDictionary(cRange rac.Range, tTag uint8) ([]byte, error) {
// Load from the MRU cache, if it was loaded from the same cRange.
if (cRange == r.cachedDictionaryCRange) && !cRange.Empty() {
return r.cachedDictionary, nil
}
r.cachedDictionaryCRange = rac.Range{}
- // Check the cRange size.
- if (cRange.Size() < 6) || (cRange[1] > r.CompressedSize) {
+ // Check the cRange size and the tTag.
+ if (cRange.Size() < 6) || (cRange[1] > r.CompressedSize) || (tTag != 0xFF) {
r.err = errInvalidDictionary
return nil, r.err
}
|
Turn fsync off for behave terraform tests
This should speed up the tests a bit, since they are currently spending a significant amount of time in createdb statements. | @@ -7,5 +7,7 @@ export PGPORT=5432
export MASTER_DATA_DIRECTORY=/data/gpdata/master/gpseg-1
export PGDATABASE=gptest
createdb gptest
+gpconfig --skipvalidation -c fsync -v off
+gpstop -u
cd /home/gpadmin/gpdb_src/gpMgmt
make -f Makefile.behave behave flags="$BEHAVE_FLAGS"
|
Fix pattern set/check rollover | @@ -58,15 +58,15 @@ u32 pat_redzone = 0xcafefade;
static void set_pattern(void *v, bytes sz, void *p, bytes psz)
{
u8 *bp = v;
- for (; sz > 0; sz -= psz, bp += psz)
- runtime_memcpy(bp, p, MIN(psz, sz));
+ for (s64 ssz = sz; ssz > 0; ssz -= psz, bp += psz)
+ runtime_memcpy(bp, p, MIN(psz, ssz));
}
static boolean check_pattern(void *v, bytes sz, void *p, bytes psz)
{
u8 *bp = v;
- for (; sz > 0; sz -= psz, bp += psz)
- if (runtime_memcmp(bp, p, MIN(psz, sz)) != 0)
+ for (s64 ssz = sz; ssz > 0; ssz -= psz, bp += psz)
+ if (runtime_memcmp(bp, p, MIN(psz, ssz)) != 0)
return false;
return true;
}
|
merged changes and updated news line | @@ -280,6 +280,8 @@ you up to date with the multi-language support provided by Elektra.
- `kdb get -v` now displays if the resulting value is a default-value defined by the metadata of the key. _(Thomas Bretterbauer)_
- `kdb cp` now succeeds if target-keys already have the same values as the source-keys. _(Thomas Bretterbauer)_
+- `web-ui` issue #2441 is fixed in form of a workaround _(Josef Wechselauer)_
+- <<TODO>>
- <<TODO>>
- <<TODO>>
|
fix (encrypt-async): remove crypto node from test-async-encrypt example | @@ -167,7 +167,7 @@ test-encrypt@test @nonic ctr=off x_crypto
test-gen-encrypt@test @nonic ctr=off x_crypto
test-ipsec@test @nonic ctr=off x_crypto async=pd crypto_burst_size=1
-test-async-encrypt@test @nonic ctr=off x_crypto async=pd crypto_burst_size=1 cores=1 crypto_node=openssl
+test-async-encrypt@test @nonic ctr=off x_crypto async=pd crypto_burst_size=1 cores=1
test-async-encrypt-decrypt@test @nonic ctr=off x_crypto async=pd crypto_burst_size=1 cores=1
portfwd-fix-async-encrypt@test @nonic ctr=off x_crypto async=pd crypto_burst_size=1 cores=1
encrypt-async@test @nonic ctr=off x_crypto async=pd crypto_burst_size=1 cores=1
|
Configure: recognize div3w modules and add -DBN_DIV3W. | @@ -1354,6 +1354,7 @@ unless ($disabled{asm}) {
push @{$config{lib_defines}}, "OPENSSL_BN_ASM_MONT" if ($target{bn_asm_src} =~ /-mont/);
push @{$config{lib_defines}}, "OPENSSL_BN_ASM_MONT5" if ($target{bn_asm_src} =~ /-mont5/);
push @{$config{lib_defines}}, "OPENSSL_BN_ASM_GF2m" if ($target{bn_asm_src} =~ /-gf2m/);
+ push @{$config{lib_defines}}, "BN_DIV3W" if ($target{bn_asm_src} =~ /-div3w/);
if ($target{sha1_asm_src}) {
push @{$config{lib_defines}}, "SHA1_ASM" if ($target{sha1_asm_src} =~ /sx86/ || $target{sha1_asm_src} =~ /sha1/);
|
compiler-families/intel-compilers-devel: include mkl-devel, needed for certain package builds | @@ -35,7 +35,7 @@ Source2: oneAPI.repo
Requires: gcc libstdc++-devel
Requires(pre): intel-oneapi-compiler-dpcpp-cpp-and-cpp-classic >= %{min_intel_ver}
Requires: intel-oneapi-compiler-dpcpp-cpp-and-cpp-classic >= %{min_intel_ver}
-Requires: intel-oneapi-mkl
+Requires: intel-oneapi-mkl intel-oneapi-mkl-devel
Requires: intel-oneapi-compiler-fortran
Recommends: intel-hpckit >= %{min_intel_vers}
|
Provide lastupdated attribute in event for sensor state updates | @@ -1050,6 +1050,13 @@ void DeRestPluginPrivate::handleSensorEvent(const Event &e)
map["id"] = e.id();
QVariantMap state;
state[e.what() + 6] = item->toVariant();
+
+ item = sensor->item(RStateLastUpdated);
+ if (item)
+ {
+ state["lastupdated"] = item->toVariant();
+ }
+
map["state"] = state;
webSocketServer->broadcastTextMessage(Json::serialize(map));
|
Update file modified and file size process' tree columns to native paths | @@ -1244,8 +1244,8 @@ static VOID PhpUpdateProcessNodeFileAttributes(
{
FILE_NETWORK_OPEN_INFORMATION networkOpenInfo;
- if (ProcessNode->ProcessItem->FileName && NT_SUCCESS(PhQueryFullAttributesFileWin32(
- ProcessNode->ProcessItem->FileName->Buffer,
+ if (ProcessNode->ProcessItem->FileName && NT_SUCCESS(PhQueryFullAttributesFile(
+ ProcessNode->ProcessItem->FileName,
&networkOpenInfo
)))
{
|
Using $(CURDIR) to fix the absolute path issue | @@ -44,7 +44,7 @@ RPM:
mkdir -p RPMS SOURCES
cp $(TARBALL) SOURCES/vpp-$(VERSION)-$(RELEASE).tar.xz
rpmbuild -bb \
- --define "_topdir $(PWD)" \
+ --define "_topdir $(CURDIR)" \
--define "_version $(VERSION)" \
--define "_release $(RELEASE)" \
vpp-suse.spec
|
vlib: chdir to runtime_dir
Type: improvement | @@ -513,6 +513,9 @@ unix_config (vlib_main_t * vm, unformat_input_t * input)
if (error)
return error;
+ if (chdir ((char *) um->runtime_dir) < 0)
+ return clib_error_return_unix (0, "chdir('%s')", um->runtime_dir);
+
error = setup_signal_handlers (um);
if (error)
return error;
|
Fix build error on Linux/Windows. | @@ -3260,23 +3260,23 @@ test_network_get_cb(
test_networks[i].up = true;
test_networks[i].config4 = PAPPL_NETCONF_DHCP;
- test_networks[i].addr4.sin_len = sizeof(struct sockaddr_in);
+// test_networks[i].addr4.sin_len = sizeof(struct sockaddr_in);
test_networks[i].addr4.sin_family = AF_INET;
test_networks[i].addr4.sin_addr.s_addr = htonl(0x0a000102 + i);
- test_networks[i].mask4.sin_len = sizeof(struct sockaddr_in);
+// test_networks[i].mask4.sin_len = sizeof(struct sockaddr_in);
test_networks[i].mask4.sin_family = AF_INET;
test_networks[i].mask4.sin_addr.s_addr = htonl(0xffffff00);
- test_networks[i].router4.sin_len = sizeof(struct sockaddr_in);
+// test_networks[i].router4.sin_len = sizeof(struct sockaddr_in);
test_networks[i].router4.sin_family = AF_INET;
test_networks[i].router4.sin_addr.s_addr = htonl(0x0a000101);
- test_networks[i].dns4[0].sin_len = sizeof(struct sockaddr_in);
+// test_networks[i].dns4[0].sin_len = sizeof(struct sockaddr_in);
test_networks[i].dns4[0].sin_family = AF_INET;
test_networks[i].dns4[0].sin_addr.s_addr = htonl(0x0a000101);
- test_networks[i].linkaddr6.sin6_len = sizeof(struct sockaddr_in6);
+// test_networks[i].linkaddr6.sin6_len = sizeof(struct sockaddr_in6);
test_networks[i].linkaddr6.sin6_family = AF_INET6;
test_networks[i].linkaddr6.sin6_addr.s6_addr[0] = 0xfe;
test_networks[i].linkaddr6.sin6_addr.s6_addr[1] = 0x80;
|
config_tools: remove ivshmem region id from ui
remove ivshmem region id from ui. | <div class="IVSH_REGIONS" v-if="defaultVal && defaultVal.length>0">
<div class="IVSH_REGION" v-for="(IVSHMEM_VMO, index) in defaultVal">
<div class="IVSH_REGION_CONTENT">
- <b style="margin-bottom: 2rem">InterVM shared memory region {{ index + 1 }}</b>
+ <b style="margin-bottom: 2rem">InterVM shared memory region</b>
<b-row class="align-items-center my-2 mt-4">
<b-col md="2">
@@ -197,7 +197,7 @@ export default {
return (value != null) && regexp.test(value);
},
validation(value) {
- return (value != null) && (value.length != 0);
+ return (value != null) && (value.length !== 0);
},
addSharedVM(vms, index) {
// add new item after current item
@@ -241,6 +241,7 @@ export default {
};
</script>
+<!--suppress CssUnusedSymbol -->
<style scoped>
label:before {
content: '*';
|
Add size check for PhGetMappedImageDataEntry | @@ -522,7 +522,7 @@ NTSTATUS PhGetMappedImageDataEntry(
dataDirectory = &optionalHeader->DataDirectory[Index];
- if (dataDirectory->VirtualAddress)
+ if (dataDirectory->VirtualAddress && dataDirectory->Size)
{
*Entry = dataDirectory;
return STATUS_SUCCESS;
@@ -540,14 +540,14 @@ NTSTATUS PhGetMappedImageDataEntry(
dataDirectory = &optionalHeader->DataDirectory[Index];
- if (dataDirectory->VirtualAddress)
+ if (dataDirectory->VirtualAddress && dataDirectory->Size)
{
*Entry = dataDirectory;
return STATUS_SUCCESS;
}
}
- return STATUS_INVALID_PARAMETER;
+ return STATUS_NOT_FOUND;
}
PVOID PhGetMappedImageDirectoryEntry(
|
Fixed Copy/Paste errors of "Snap redux" (PR
Some variable declarations got lost by the changes which made the code
not building. | @@ -1960,6 +1960,8 @@ check_admin_access(cupsd_client_t *con) // I - Client connection
SnapdClient *client = NULL; // Data structure of snapd access
GError *error = NULL; // Glib error
int ret = 1; // Return value
+ SnapdSnap *snap = NULL; // Data structure of client Snap
+ SnapdClient *snapd = NULL; // Data structure of snapd access
# ifdef AF_LOCAL
|
optimized_nop: make clang happy with D=0 case | @@ -529,6 +529,8 @@ static void nary_opt(void* _data, void* ptr[])
}
+
+
/**
* Optimized n-op.
*
@@ -544,6 +546,25 @@ static void nary_opt(void* _data, void* ptr[])
*/
void optimized_nop(unsigned int N, unsigned int io, unsigned int D, const long dim[D], const long (*nstr[N])[D], void* const nptr[N], size_t sizes[N], md_nary_opt_fun_t too, void* data_ptr)
{
+ assert(N > 0);
+
+ if (0 == D) {
+
+ long dim1[1] = { 1 };
+ long tstrs[N][1];
+ long (*nstr1[N])[1];
+
+ for (unsigned int i = 0; i < N; i++) {
+
+ tstrs[i][0] = 0;
+ nstr1[i] = &tstrs[i];
+ }
+
+ optimized_nop(N, io, 1, dim1, (void*)nstr1, nptr, sizes, too, data_ptr);
+
+ return;
+ }
+
long tdims[D];
md_copy_dims(D, tdims, dim);
@@ -554,6 +575,7 @@ void optimized_nop(unsigned int N, unsigned int io, unsigned int D, const long d
for (unsigned int i = 0; i < N; i++) {
md_copy_strides(D, tstrs[i], *nstr[i]);
+
nstr1[i] = &tstrs[i];
nptr1[i] = nptr[i];
}
|
Subsets and Splits
No saved queries yet
Save your SQL queries to embed, download, and access them later. Queries will appear here once saved.