message
stringlengths 6
474
| diff
stringlengths 8
5.22k
|
---|---|
parser: on exception release time offset (CID 300959) | @@ -582,6 +582,9 @@ int flb_parser_conf_file(const char *file, struct flb_config *config)
if (time_key) {
flb_sds_destroy(time_key);
}
+ if (time_offset) {
+ flb_sds_destroy(time_offset);
+ }
if (types_str) {
flb_sds_destroy(types_str);
}
|
ci: refactor pipeline to reduce copy-paste | -_pipeline: &pipeline
- info_script: |
- meson --version
- ninja --version
+.compile: &compile
submodule-update_script: git submodule update --init --recursive
configure_script: meson -Db_lundef=false build
compile_script: ninja -C build
+.test: &test
+ test_script: ninja -C build test
+
+.pipeline: &pipeline
+ info_script: |
+ meson --version
+ ninja --version
+ <<: *compile
+ <<: *test
+
Debian (gcc)_task:
container:
image: snaipe/ci-meson:debian-10
@@ -14,7 +21,6 @@ Debian (gcc)_task:
apt-get install -y pkg-config cmake libgit2-dev libffi-dev libnanomsg-dev
pip3 install cram==0.7
<<: *pipeline
- test_script: ninja -C build test
Alpine (gcc,x86_64)_task:
container:
@@ -23,14 +29,12 @@ Alpine (gcc,x86_64)_task:
apk add --no-cache cmake libgit2-dev libffi-dev
pip3 install cram==0.7
<<: *pipeline
- test_script: ninja -C build test
Alpine (gcc,i386)_task:
container:
image: snaipe/ci-meson:alpine-x86
setup_script: *alpine-deps
<<: *pipeline
- test_script: ninja -C build test
Alpine (gcc,arm)_task:
container:
@@ -39,7 +43,6 @@ Alpine (gcc,arm)_task:
CIRRUS_SHELL: /sbin/qemu-sh
setup_script: *alpine-deps
<<: *pipeline
- test_script: ninja -C build test
Alpine (gcc,aarch64)_task:
container:
@@ -48,7 +51,6 @@ Alpine (gcc,aarch64)_task:
CIRRUS_SHELL: /sbin/qemu-sh
setup_script: *alpine-deps
<<: *pipeline
- test_script: ninja -C build test
MacOS_task:
macos_instance:
@@ -59,7 +61,6 @@ MacOS_task:
brew install meson cmake pkg-config libffi libgit2 nanomsg
sudo pip3 install cram==0.7
<<: *pipeline
- test_script: ninja -C build test
FreeBSD_task:
freebsd_instance:
@@ -70,7 +71,6 @@ FreeBSD_task:
python3 -m pip install --upgrade pip
python3 -m pip install cram==0.7
<<: *pipeline
- test_script: ninja -C build test
Windows (mingw-gcc)_task:
windows_container:
@@ -85,5 +85,5 @@ Windows (mingw-gcc)_task:
git config --global core.autocrlf false
patch --forward --directory="C:\\Python\\Lib\\site-packages\\cram" < ci/cram-fix-winenv.patch
patch --forward --directory="C:\\Python\\Lib\\site-packages\\cram" < ci/cram-ignore-cr.patch
- <<: *pipeline
+ <<: *compile
test_script: meson devenv -C build ninja test
|
imports VERSION from version.py into setup.py | @@ -2,9 +2,24 @@ from setuptools import setup, find_packages
PACKAGE = 'catboost'
+def execfile(filepath, globals=None, locals=None):
+ if globals is None:
+ globals = {}
+ globals.update({
+ "__file__": filepath,
+ "__name__": "__main__",
+ })
+ with open(filepath, 'rb') as file:
+ exec(compile(file.read(), filepath, 'exec'), globals, locals)
+
+def version():
+ globals={}
+ execfile('./catboost/version.py', globals)
+ return globals['VERSION']
+
setup(
name=PACKAGE,
- version="0.13",
+ version=version(),
author="CatBoost Developers",
description="Python package for catboost",
|
Report HealthState = Non-Functional if DDRT is not trained
On the ipmctl show -dimm display, we need to sometimes
override what the HealthState value is (from the FIS SMART and
Health Info command) due to other circumstances. This change adds
the non-functional property to these overriding values. | @@ -726,12 +726,15 @@ GetDimmInfo (
pDimmInfo->ErrorMask |= DIMM_INFO_ERROR_UID;
}
+ // Set defaults first
+ pDimmInfo->ManageabilityState = MANAGEMENT_VALID_CONFIG;
+ pDimmInfo->HealthState = HEALTH_UNKNOWN;
+ // Then change as needed.
if (!IsDimmManageable(pDimm)) {
pDimmInfo->ManageabilityState = MANAGEMENT_INVALID_CONFIG;
pDimmInfo->HealthState = HEALTH_UNMANAGEABLE;
- } else {
- pDimmInfo->ManageabilityState = MANAGEMENT_VALID_CONFIG;
- pDimmInfo->HealthState = HEALTH_UNKNOWN;
+ } else if (TRUE == pDimm->NonFunctional) {
+ pDimmInfo->HealthState = HEALTH_NON_FUNCTIONAL;
}
pDimmInfo->IsNew = pDimm->IsNew;
@@ -871,7 +874,9 @@ GetDimmInfo (
if (EFI_ERROR(ReturnCode)) {
pDimmInfo->ErrorMask |= DIMM_INFO_ERROR_SMART_AND_HEALTH;
}
- if (HEALTH_UNMANAGEABLE != pDimmInfo->HealthState) {
+ // Fill in the DCPMM's understanding of its own HealthState if we didn't have any
+ // other opinions earlier
+ if (HEALTH_UNKNOWN == pDimmInfo->HealthState) {
ConvertHealthBitmask(HealthInfo.HealthStatus, &pDimmInfo->HealthState);
}
pDimmInfo->HealthStatusReason = HealthInfo.HealthStatusReason;
|
Floats working fro value mapping | @@ -241,8 +241,8 @@ int64_t _lv_pow(int64_t base, int8_t exp)
*/
LV_ATTRIBUTE_FAST_MEM int16_t _lv_map(int16_t x, float min_in, float max_in, float min, float max)
{
- int16_t slope = (max - min) / (max_in - min_in);
- int16_t bias = min - slope * min_in;
+ float slope = (max - min) / (max_in - min_in);
+ float bias = min - slope * min_in;
return bias + slope * x;
}
|
remove DTRACE_TESTS=1 from `asan` target | @@ -44,7 +44,6 @@ ossl3.0:
asan:
docker run $(DOCKER_RUN_OPTS) h2oserver/h2o-ci:ubuntu2004 \
- env DTRACE_TESTS=1 \
make -f $(SRC_DIR).ro/misc/docker-ci/check.mk _check \
CMAKE_ARGS='-DCMAKE_C_COMPILER=clang -DCMAKE_CXX_COMPILER=clang++ -DCMAKE_C_FLAGS=-fsanitize=address -DCMAKE_CXX_FLAGS=-fsanitize=address' \
BUILD_ARGS='$(BUILD_ARGS)' \
|
Change Tika server start command to include classpath (required to
declare and load additionnal packages like NER) | @@ -16,7 +16,7 @@ cmd_start() {
exit 1
fi
echo "Starting Tika Server"
- nohup "$JAVA_HOME/bin/java" -Dlog4j2.configurationFile=file:$TIKA_SERVER_HOME/conf/log4j2.properties.xml -Duser.timezone=UTC $TIKA_MEM -jar $TIKA_SERVER_HOME/bin/tika-server.jar -c $TIKA_SERVER_HOME/conf/tika-config.xml >/dev/null 2>&1 &
+ nohup "$JAVA_HOME/bin/java" -Dlog4j2.configurationFile=file:$TIKA_SERVER_HOME/conf/log4j2.properties.xml -Duser.timezone=UTC $TIKA_MEM -cp $TIKA_SERVER_HOME/bin/tika-server.jar org.apache.tika.server.core.TikaServerCli -c $TIKA_SERVER_HOME/conf/tika-config.xml >/dev/null 2>&1 &
echo $! > $TIKA_SERVER_PID_FILE
echo "Tika Server started with PID $(cat $TIKA_SERVER_PID_FILE)"
return 0
|
reduce page retire words to 32 | @@ -396,7 +396,7 @@ void _mi_page_retire(mi_page_t* page) {
// is the only page left with free blocks. It is not clear
// how to check this efficiently though... for now we just check
// if its neighbours are almost fully used.
- if (mi_likely(page->block_size <= MI_SMALL_SIZE_MAX)) {
+ if (mi_likely(page->block_size <= 32*MI_INTPTR_SIZE)) {
if (mi_page_mostly_used(page->prev) && mi_page_mostly_used(page->next)) {
_mi_stat_counter_increase(&_mi_stats_main.page_no_retire,1);
return; // dont't retire after all
|
hv: fix potential NULL pointer reference in hc_assgin_ptdev
this patch validates input 'vdev->pdev' before
reference to avoid potenial hypervisor crash.
[v2] update:
Combine condition check for 'vdev' and 'vdev->pdev' | @@ -836,7 +836,7 @@ int32_t hcall_assign_ptdev(struct acrn_vm *vm, uint16_t vmid, uint64_t param)
spinlock_obtain(&vm->vpci.lock);
vdev = pci_find_vdev(&vm->vpci, bdf);
- if (vdev == NULL) {
+ if ((vdev == NULL) || (vdev->pdev == NULL)) {
pr_fatal("%s %x:%x.%x not found\n", __func__, bdf.bits.b, bdf.bits.d, bdf.bits.f);
ret = -EPERM;
} else {
|
Modify is_tls13_capable() to take account of the servername cb
A servername cb may change the available certificates, so if we have one
set then we cannot rely on the configured certificates to determine if we
are capable of negotiating TLSv1.3 or not.
Fixes | @@ -1515,8 +1515,8 @@ static int ssl_method_error(const SSL *s, const SSL_METHOD *method)
/*
* Only called by servers. Returns 1 if the server has a TLSv1.3 capable
- * certificate type, or has PSK or a certificate callback configured. Otherwise
- * returns 0.
+ * certificate type, or has PSK or a certificate callback configured, or has
+ * a servername callback configure. Otherwise returns 0.
*/
static int is_tls13_capable(const SSL *s)
{
@@ -1525,6 +1525,17 @@ static int is_tls13_capable(const SSL *s)
int curve;
#endif
+ if (!ossl_assert(s->ctx != NULL) || !ossl_assert(s->session_ctx != NULL))
+ return 0;
+
+ /*
+ * A servername callback can change the available certs, so if a servername
+ * cb is set then we just assume TLSv1.3 will be ok
+ */
+ if (s->ctx->ext.servername_cb != NULL
+ || s->session_ctx->ext.servername_cb != NULL)
+ return 1;
+
#ifndef OPENSSL_NO_PSK
if (s->psk_server_callback != NULL)
return 1;
|
Remove second cropping of multiple choice text | @@ -136,8 +136,8 @@ export const actorFramesPerDir = (spriteType, numFrames) => {
};
export const combineMultipleChoiceText = (args) => {
- const trueText = args.trueText.slice(0, 17) || "Choice A";
- const falseText = args.falseText.slice(0, 17) || "Choice B";
+ const trueText = args.trueText || "Choice A";
+ const falseText = args.falseText || "Choice B";
return `${trueText}\n${falseText}`;
};
|
admin/lmod: clean up .spec to not use undefined %{luaver} | @@ -24,12 +24,14 @@ Url: https://github.com/TACC/Lmod
Source0: https://github.com/TACC/Lmod/archive/%{version}.tar.gz#$/%{pname}-%{version}.tar.gz
# Known dependencies
-Requires: lua >= %{luaver}
+Requires: lua
Requires: tcl
-BuildRequires: lua >= %{luaver}
-BuildRequires: lua-devel >= %{luaver}
+BuildRequires: lua
+BuildRequires: lua-devel
BuildRequires: lua-libs
+BuildRequires: rsync
+BuildRequires: tcl tcl-devel
%if 0%{?rhel} > 7
BuildRequires: lua-libs
@@ -42,12 +44,6 @@ Requires: lua-posix
# SUSE Leap
BuildRequires: lua53-luafilesystem
BuildRequires: lua53-luaposix
-%endif
-
-BuildRequires: rsync
-BuildRequires: tcl tcl-devel
-
-%if 0%{?sle_version} || 0%{?suse_version}
BuildRequires: procps
%endif
@@ -55,7 +51,7 @@ BuildRequires: procps
Conflicts: Modules
%else
%if 0%{?rhel} > 7
-# Starting with RHEL8 packages in RHEL8 depending on
+# Starting with RHEL8, packages in RHEL8 depending on
# environment modules no longer depend on the package
# but on the virtual provide 'environment(modules)'.
# By extending the MODULEPATH of this lmod we can easily
|
"one statistics flush" throws assert when one counter not added | @@ -1289,6 +1289,9 @@ vnet_lisp_flush_stats (void)
vlib_combined_counter_main_t *cm = &lgm->counters;
u32 i;
+ if (cm->counters == NULL)
+ return 0;
+
for (i = 0; i < vlib_combined_counter_n_counters (cm); i++)
vlib_zero_combined_counter (cm, i);
|
VERSION bump to version 2.0.266 | @@ -61,7 +61,7 @@ set (CMAKE_RUNTIME_OUTPUT_DIRECTORY ${CMAKE_BINARY_DIR})
# set version of the project
set(LIBYANG_MAJOR_VERSION 2)
set(LIBYANG_MINOR_VERSION 0)
-set(LIBYANG_MICRO_VERSION 265)
+set(LIBYANG_MICRO_VERSION 266)
set(LIBYANG_VERSION ${LIBYANG_MAJOR_VERSION}.${LIBYANG_MINOR_VERSION}.${LIBYANG_MICRO_VERSION})
# set version of the library
set(LIBYANG_MAJOR_SOVERSION 2)
|
Fix find handles regex type filter | @@ -768,17 +768,12 @@ static BOOLEAN MatchTypeString(
_In_ PPH_HANDLE_SEARCH_CONTEXT Context,
_In_ PPH_STRINGREF Input
)
-{
- if (Context->SearchRegexCompiledExpression && Context->SearchRegexMatchData)
- return TRUE;
- else
{
if (PhEqualString2(Context->SearchTypeString, L"Everything", FALSE))
return TRUE;
return PhFindStringInStringRef(Input, &Context->SearchTypeString->sr, TRUE) != -1;
}
-}
typedef struct _SEARCH_HANDLE_CONTEXT
{
|
CI: windows fixes again | @@ -48,7 +48,7 @@ jobs:
uses: microsoft/[email protected]
- name: build
working-directory: ./projects/vs2019/
- run: msbuild ./wren_cli.vcxproj /property:Configuration=Release /property:Platform=64bit
+ run: msbuild ./wren_cli.sln /property:Configuration=Release /property:Platform=64bit
- uses: actions/upload-artifact@v2
with:
name: windows-bin
|
Fix build warning regarding define syntax in debug.h | @@ -137,7 +137,8 @@ Once LOGM is approved, each module should have its own index
#define LOGM_PRI (1)
#define LOGM_EN (2)
-#if defined(CONFIG_DEBUG_ERROR) && defined(CONFIG_LOGM)
+#ifdef CONFIG_DEBUG_ERROR
+#ifdef CONFIG_LOGM
#define dbg(format, ...) \
logm(LOGM_EN, LOGM_IDX, LOGM_PRI, EXTRA_FMT format EXTRA_ARG, ##__VA_ARGS__)
@@ -147,7 +148,7 @@ Once LOGM is approved, each module should have its own index
#define lldbg(format, ...) \
logm(LOGM_EN, LOGM_IDX, LOGM_PRI, EXTRA_FMT format EXTRA_ARG, ##__VA_ARGS__)
-#elif CONFIG_DEBUG_ERROR
+#else
/**
* @brief Error debug
* @since Tizen RT v1.0
@@ -167,20 +168,22 @@ Once LOGM is approved, each module should have its own index
#else
#define lldbg(x...)
#endif
+#endif
#else
#define dbg(x...)
#define lldbg(x...)
#endif
-#if defined(CONFIG_DEBUG_WARN) && defined(CONFIG_LOGM)
+#ifdef CONFIG_DEBUG_WARN
+#ifdef CONFIG_LOGM
#define wdbg(format, ...) \
logm(LOGM_EN, LOGM_IDX, LOGM_PRI, EXTRA_FMT format EXTRA_ARG, ##__VA_ARGS__)
#define llwdbg(format, ...) \
logm(LOGM_EN, LOGM_IDX, LOGM_PRI, EXTRA_FMT format EXTRA_ARG, ##__VA_ARGS__)
-#elif CONFIG_DEBUG_WARN
+#else
/**
* @brief Warning debug
* @since Tizen RT v1.0
@@ -198,20 +201,22 @@ Once LOGM is approved, each module should have its own index
#else
#define llwdbg(x...)
#endif
+#endif
#else
#define wdbg(x...)
#define llwdbg(x...)
#endif
-#if defined(CONFIG_DEBUG_VERBOSE) && defined(CONFIG_LOGM)
+#ifdef CONFIG_DEBUG_VERBOSE
+#ifdef CONFIG_LOGM
#define vdbg(format, ...) \
logm(LOGM_EN, LOGM_IDX, LOGM_PRI, EXTRA_FMT format EXTRA_ARG, ##__VA_ARGS__)
#define llvdbg(format, ...) \
logm(LOGM_EN, LOGM_IDX, LOGM_PRI, EXTRA_FMT format EXTRA_ARG, ##__VA_ARGS__)
-#elif CONFIG_DEBUG_VERBOSE
+#else
/**
* @brief Informational(Verbose) debug
* @since Tizen RT v1.0
@@ -230,6 +235,7 @@ Once LOGM is approved, each module should have its own index
#else
#define llvdbg(x...)
#endif
+#endif
#else
#define vdbg(x...)
|
BugID:16179547: modified the uData queue stack size | #define UDATA_QUEUE_MAXSLOTS 8
+#ifndef UDATA_TASK_STACK_SIZE
#define UDATA_TASK_STACK_SIZE 4096 // 4kByte
+#endif
#define UDATA_TASK_PRIO \
(AOS_DEFAULT_APP_PRI - 2) // higher prio than normal app's
#define UDATA_QUEUE_MAX_MSG_SIZE (sizeof(sensor_msg_pkg_t))
|
Add tests for the effect of --extra-version-suffix
should append to both psql 'select version()' and also to 'gpssh --version' | @@ -9,3 +9,6 @@ ${CURRENT_DIR}/gpdb_src/concourse/scripts/setup_gpadmin_user.bash
su - gpadmin ${CURRENT_DIR}/gpdb_src/concourse/scripts/deb_init_cluster.bash
su - gpadmin ${CURRENT_DIR}/gpdb_src/concourse/scripts/deb_test_cluster.bash
+su - gpadmin -c "source /opt/gpdb/greenplum_path.sh && psql -t -U gpadmin template1 -c 'select version()' | grep '\-oss'"
+# gpssh replaces "-" with " " in version string
+su - gpadmin -c "source /opt/gpdb/greenplum_path.sh && gpssh --version | grep ' oss'"
|
error messages should go to stderr | @@ -69,7 +69,7 @@ int grib_tool_init(grib_runtime_options* options)
options->dump_mode = "default";
if (opt > 1) {
- printf("%s: simultaneous j/O/D options not allowed\n", tool_name);
+ fprintf(stderr, "%s: simultaneous j/O/D options not allowed\n", tool_name);
exit(1);
}
|
appveyor.yml: Let 'nmake' do builds in parallel on all CPU cores | @@ -62,6 +62,7 @@ build_script:
- ps: >-
If ($env:Configuration -Match "shared" -or $env:EXTENDED_TESTS) {
cmd /c "%NMAKE% build_all_generated 2>&1"
+ # Unfortunately, CL=/MP would not have parallelizing effect
cmd /c "%NMAKE% PERL=no-perl 2>&1"
}
- cd ..
@@ -70,6 +71,7 @@ test_script:
- cd _build
- ps: >-
If ($env:Configuration -Match "shared" -or $env:EXTENDED_TESTS) {
+ # Unfortunately, HARNESS_JOBS=4 would not have parallelizing effect
if ($env:EXTENDED_TESTS) {
cmd /c "%NMAKE% test HARNESS_VERBOSE_FAILURE=yes 2>&1"
} Else {
|
include/throttle_ap.h: Format with clang-format
BRANCH=none
TEST=none
Tricium: disable | @@ -62,8 +62,7 @@ struct prochot_cfg {
defined(CONFIG_THROTTLE_AP_ON_BAT_DISCHG_CURRENT) || \
defined(CONFIG_THROTTLE_AP_ON_BAT_VOLTAGE)
-void throttle_ap(enum throttle_level level,
- enum throttle_type type,
+void throttle_ap(enum throttle_level level, enum throttle_type type,
enum throttle_sources source);
/**
@@ -103,11 +102,11 @@ void throttle_ap_c10_input_interrupt(enum gpio_signal signal);
static inline void throttle_ap(enum throttle_level level,
enum throttle_type type,
enum throttle_sources source)
-{}
+{
+}
#endif
-void throttle_gpu(enum throttle_level level,
- enum throttle_type type,
+void throttle_gpu(enum throttle_level level, enum throttle_type type,
enum throttle_sources source);
#endif /* __CROS_EC_THROTTLE_AP_H */
|
New ws test | @@ -41,6 +41,22 @@ coroutine void nohttp_client(int s) {
errno_assert(rc == 0);
}
+coroutine void http_client(int s) {
+ s = ws_attach_client(s, "/", "www.example.org", WS_BINARY, -1);
+ errno_assert(s >= 0);
+ int rc = msend(s, "ABC", 3, -1);
+ errno_assert(rc == 0);
+ char buf[3];
+ ssize_t sz = mrecv(s, buf, sizeof(buf), -1);
+ errno_assert(sz >= 0);
+ assert(sz == 3);
+ assert(buf[0] == 'D' && buf[1] == 'E' && buf[2] == 'F');
+ s = ws_detach(s, -1);
+ errno_assert(s >= 0);
+ rc = hclose(s);
+ errno_assert(rc == 0);
+}
+
int main(void) {
int p[2];
int rc = ipc_pair(p);
@@ -63,6 +79,28 @@ int main(void) {
rc = hclose(cr);
errno_assert(rc == 0);
+ rc = ipc_pair(p);
+ errno_assert(rc == 0);
+ cr = go(http_client(p[1]));
+ errno_assert(cr >= 0);
+ char resource[256];
+ char host[256];
+ s = ws_attach_server(p[0], WS_BINARY, resource, sizeof(resource),
+ host, sizeof(host), -1);
+ errno_assert(s >= 0);
+ sz = mrecv(s, buf, sizeof(buf), -1);
+ errno_assert(sz >= 0);
+ assert(sz == 3);
+ assert(buf[0] == 'A' && buf[1] == 'B' && buf[2] == 'C');
+ rc = msend(s, "DEF", 3, -1);
+ errno_assert(rc == 0);
+ s = ws_detach(s, -1);
+ errno_assert(s >= 0);
+ rc = hclose(s);
+ errno_assert(rc == 0);
+ rc = hclose(cr);
+ errno_assert(rc == 0);
+
return 0;
}
|
Fix test title in storage/remote test. | @@ -183,7 +183,7 @@ testRun(void)
TEST_RESULT_STR_Z(info.group, testGroup(), " check group");
// -------------------------------------------------------------------------------------------------------------------------
- TEST_TITLE("protocol output that is not tested elsewhere (basic)");
+ TEST_TITLE("protocol output that is not tested elsewhere (detail)");
info = (StorageInfo){.level = storageInfoLevelDetail, .type = storageTypeLink, .linkDestination = STRDEF("../")};
TEST_RESULT_VOID(storageRemoteInfoWrite(server, &info), "write link info");
|
HV: remove one lock for ctx->flags operation.
for ctx->flags is protected by scheduler lock, so not need
to set lock again.
Acked-by: Eddie Dong | @@ -112,7 +112,7 @@ void make_reschedule_request(uint16_t pcpu_id, uint16_t delmode)
{
struct sched_context *ctx = &per_cpu(sched_ctx, pcpu_id);
- bitmap_set_lock(NEED_RESCHEDULE, &ctx->flags);
+ bitmap_set_nolock(NEED_RESCHEDULE, &ctx->flags);
if (get_pcpu_id() != pcpu_id) {
switch (delmode) {
case DEL_MODE_IPI:
@@ -192,7 +192,7 @@ void schedule(void)
get_schedule_lock(pcpu_id);
next = get_next_sched_obj(ctx);
- bitmap_clear_lock(NEED_RESCHEDULE, &ctx->flags);
+ bitmap_clear_nolock(NEED_RESCHEDULE, &ctx->flags);
if (prev == next) {
release_schedule_lock(pcpu_id);
|
ci: increase parallel UT count for few jobs | @@ -478,7 +478,7 @@ UT_001:
UT_002:
extends: .unit_test_esp32_template
- parallel: 14
+ parallel: 15
tags:
- ESP32_IDF
- UT_T1_1
@@ -683,7 +683,7 @@ UT_S2_SDSPI:
UT_C3:
extends: .unit_test_esp32c3_template
- parallel: 32
+ parallel: 33
tags:
- ESP32C3_IDF
- UT_T1_1
|
remove info about .. spelling:: from last section | @@ -369,8 +369,6 @@ Things To Consider Going Forward
*zone*/*cell*/*element* and *node*/*point*/*vertex*
* Additional features of Sphinx to consider adopting...
- * Using ``.. spelling::`` directives to limit scope of special case words to
- the ``.rst`` files in which they occur.
* ``:guilable:`` role for referring to GUI widgets
* ``:command:`` role for OS level commands
* ``:file:`` role for referring to file names
|
Update Date string every second | @@ -45,12 +45,13 @@ static inline void add_date(http_s *r) {
if (!mod_hash)
mod_hash = fiobj_sym_hash("last-modified", 13);
- if (facil_last_tick().tv_sec >= last_date_added + 60) {
- FIOBJ tmp = fiobj_str_buf(32);
- fiobj_str_resize(
- tmp, http_time2str(fiobj_obj2cstr(tmp).data, facil_last_tick().tv_sec));
+ if (facil_last_tick().tv_sec > last_date_added) {
+ FIOBJ tmp = NULL;
spn_lock(&date_lock);
- if (facil_last_tick().tv_sec >= last_date_added + 60) {
+ if (facil_last_tick().tv_sec > last_date_added) { /* retest inside lock */
+ tmp = fiobj_str_buf(32);
+ fiobj_str_resize(tmp, http_time2str(fiobj_obj2cstr(tmp).data,
+ facil_last_tick().tv_sec));
last_date_added = facil_last_tick().tv_sec;
FIOBJ other = current_date;
current_date = tmp;
@@ -1102,7 +1103,7 @@ size_t http_time2str(char *target, const time_t t) {
}
if (last_tick > cached_tick) {
struct tm tm;
- cached_tick = last_tick | 1;
+ cached_tick = last_tick; /* refresh every second */
http_gmtime(last_tick, &tm);
chached_len = http_date2str(cached_httpdate, &tm);
}
|
Add unit test for ippDeleteValues. | @@ -18,6 +18,7 @@ Changes in CUPS v2.4.3 (TBA)
- Fixed invalid memory access during generating IPP Everywhere queue (Issue #466)
- Fixed memory leaks in `create_local_bg_thread()` (Issue #466)
- Fixed TLS certificate generation bugs.
+- `ippDeleteValues` would not delete the last value (Issue #556)
- Ignore some of IPP defaults if the application sends its PPD alternative (Issue #484)
- Now look for default printer on network if needed (Issue #452)
- Now report fax attributes and values as needed (Issue #459)
|
Add actorGetPosition to codegen | @@ -1030,6 +1030,28 @@ class ScriptBuilder {
this._addNL();
};
+ actorGetPosition = (variableX: string, variableY: string) => {
+ const variableXAlias = this.getVariableAlias(variableX);
+ const variableYAlias = this.getVariableAlias(variableY);
+
+ this._addComment(`Store Position In Variables`);
+ this._actorGetPosition("ACTOR");
+
+ this._rpn() //
+ .ref("^/(ACTOR + 1)/")
+ .int16(8 * 16)
+ .operator(".DIV")
+ .ref("^/(ACTOR + 2)/")
+ .int16(8 * 16)
+ .operator(".DIV")
+ .stop();
+
+ this._set(variableXAlias, ".ARG1");
+ this._set(variableYAlias, ".ARG0");
+ this._stackPop(2);
+ this._addNL();
+ };
+
actorPush = (continueUntilCollision = false) => {
const stackPtr = this.stackPtr;
const upLabel = this.getNextLabel();
@@ -2329,12 +2351,6 @@ class ScriptBuilder {
output.push(cmd(ACTOR_SET_POSITION_TO_VALUE));
};
- actorGetPosition = (variableX, variableY) => {
- const output = this.output;
- this.vectorsLoad(variableX, variableY);
- output.push(cmd(ACTOR_GET_POSITION));
- };
-
actorSetDirection = (direction = "down") => {
const output = this.output;
output.push(cmd(ACTOR_SET_DIRECTION));
|
CMSIS-DSP: Bug correction in SDF.
Problems with slidding buffer. | @@ -221,6 +221,7 @@ class SlidingBuffer: public GenericNode<IN,windowSize-overlap,IN,windowSize>
public:
SlidingBuffer(FIFOBase<IN> &src,FIFOBase<IN> &dst):GenericNode<IN,windowSize-overlap,IN,windowSize>(src,dst)
{
+ static_assert((windowSize-overlap)>0, "Overlap is too big");
memory.resize(overlap);
};
@@ -229,7 +230,7 @@ public:
IN *b=this->getWriteBuffer();
memcpy((void*)b,(void*)memory.data(),overlap*sizeof(IN));
memcpy((void*)(b+overlap),(void*)a,(windowSize-overlap)*sizeof(IN));
- memcpy((void*)memory.data(),(void*)(a+windowSize-(overlap<<1)),overlap*sizeof(IN)) ;
+ memcpy((void*)memory.data(),(void*)(b+windowSize-overlap),overlap*sizeof(IN)) ;
return(0);
};
protected:
@@ -238,12 +239,13 @@ protected:
};
template<typename IN,int windowSize, int overlap>
-class OverlapAdd: public GenericNode<IN,windowSize,IN,overlap>
+class OverlapAdd: public GenericNode<IN,windowSize,IN,windowSize-overlap>
{
public:
OverlapAdd(FIFOBase<IN> &src,FIFOBase<IN> &dst):GenericNode<IN,windowSize,IN,overlap>(src,dst)
{
- memory.resize(windowSize);
+ static_assert((windowSize-overlap)>0, "Overlap is too big");
+ memory.resize(overlap);
};
int run(){
@@ -251,18 +253,33 @@ public:
IN *a=this->getReadBuffer();
IN *b=this->getWriteBuffer();
- memmove((void*)memory.data(),(void*)(memory.data()+overlap),(windowSize-overlap)*sizeof(IN));
- for(i=0;i<windowSize-overlap;i++)
+ for(i=0;i<overlap;i++)
{
memory[i] = a[i] + memory[i];
}
- for(;i<windowSize;i++)
+
+ if (2*overlap - windowSize > 0)
{
- memory[i] = a[i];
+
+ memcpy((void*)b,(void*)memory.data(),(windowSize-overlap)*sizeof(IN));
+
+ memmove(memory.data(),memory.data()+windowSize-overlap,(2*overlap - windowSize)*sizeof(IN));
+ memcpy(memory.data()+2*overlap - windowSize,a+overlap,(windowSize-overlap)*sizeof(IN));
}
+ else if (2*overlap - windowSize < 0)
+ {
+ memcpy((void*)b,(void*)memory.data(),overlap*sizeof(IN));
+ memcpy((void*)(b+overlap),(void*)(a+overlap),(windowSize - 2*overlap)*sizeof(IN));
+ memcpy((void*)memory.data(),(void*)(a+windowSize-overlap),overlap*sizeof(IN));
+ }
+ else
+ {
memcpy((void*)b,(void*)memory.data(),overlap*sizeof(IN));
+ memcpy((void*)memory.data(),(void*)(a+overlap),overlap*sizeof(IN));
+ }
+
return(0);
};
protected:
|
[catboost/R] Fix S3 generic/method consistency warning
Fix this WARNING issued by R check:
* checking S3 generic/method consistency ... WARNING
print:
function(x, ...)
print.catboost.Pool:
function(pool, ...) | @@ -367,13 +367,13 @@ tail.catboost.Pool <- function(x, n = 10, ...) {
#'
#' Print dimensions of catboost.Pool.
#'
-#' @param pool a catboost.Pool object
+#' @param x a catboost.Pool object
#'
#' Default value: Required argument
#' @param ... not currently used
#' @export
-print.catboost.Pool <- function(pool, ...) {
- cat("catboost.Pool\n", nrow(pool), " rows, ", ncol(pool), " columns", sep = "")
+print.catboost.Pool <- function(x, ...) {
+ cat("catboost.Pool\n", nrow(x), " rows, ", ncol(x), " columns", sep = "")
}
|
make test: Stop downloading pip.
Use the installed version of pip. If a newer version of pip is needed,
it can be specified in requirements.txt. This is to improve idempotence
by providing some control over upstream changes. | @@ -79,7 +79,6 @@ PYTHON_DEPENDS=$(PYTHON_EXTRA_DEPENDS) -r requirements.txt
SCAPY_SOURCE=$(shell find $(VENV_PATH)/lib/$(PYTHON) -name site-packages)
BUILD_COV_DIR=$(TEST_DIR)/coverage
-GET_PIP_SCRIPT=$(TEST_RUN_DIR)/get-pip.py
PIP_INSTALL_DONE=$(TEST_RUN_DIR)/pip-install.done
PIP_PATCH_DONE=$(TEST_RUN_DIR)/pip-patch.done
PAPI_INSTALL_DONE=$(TEST_RUN_DIR)/papi-install.done
@@ -92,11 +91,8 @@ else
PYTHON_INTERP=$(PYTHON)
endif
-$(GET_PIP_SCRIPT):
+$(PIP_INSTALL_DONE):
@mkdir -p $(TEST_RUN_DIR)
- @bash -c "cd $(TEST_RUN_DIR) && curl -O https://bootstrap.pypa.io/get-pip.py"
-
-$(PIP_INSTALL_DONE): $(GET_PIP_SCRIPT)
@virtualenv $(VENV_PATH) -p $(PYTHON_INTERP)
@bash -c "source $(VENV_PATH)/bin/activate && $(PYTHON_INTERP) -m pip install $(PYTHON_DEPENDS)"
@touch $@
@@ -111,7 +107,7 @@ $(PIP_PATCH_DONE): $(PIP_INSTALL_DONE)
@touch $@
$(PAPI_INSTALL_DONE): $(PIP_PATCH_DONE)
- @bash -c "source $(VENV_PATH)/bin/activate && cd $(WS_ROOT)/src/vpp-api/python && $(PYTHON_INTERP) setup.py install"
+ @bash -c "source $(VENV_PATH)/bin/activate && $(PYTHON_INTERP) -m pip install -e $(WS_ROOT)/src/vpp-api/python"
@touch $@
define retest-func
|
Align the "; Next union/fragment" comments with their symbols | @@ -469,9 +469,11 @@ static uint16_t writeMapBank(struct SortedSections const *sectList,
if (sect->nextu) {
// Announce the following "piece"
if (sect->nextu->modifier == SECTION_UNION)
- fprintf(mapFile, "\t\t; Next union\n");
+ fprintf(mapFile,
+ "\t ; Next union\n");
else if (sect->nextu->modifier == SECTION_FRAGMENT)
- fprintf(mapFile, "\t\t; Next fragment\n");
+ fprintf(mapFile,
+ "\t ; Next fragment\n");
}
sect = sect->nextu; // Also print symbols in the following "pieces"
|
pick up atomic bit clear fix | @@ -345,12 +345,12 @@ void kernel_unlock();
// wakeup
static inline void atomic_set_bit(u64 *target, u64 bit)
{
- __asm__("lock btsq %1, %0": "+m"(*target): "r"(bit));
+ asm volatile("lock btsq %1, %0": "+m"(*target): "r"(bit) : "memory");
}
static inline void atomic_clear_bit(u64 *target, u64 bit)
{
- __asm__("lock btcq %1, %0": "+m"(*target):"r"(bit));
+ asm volatile("lock btrq %1, %0": "+m"(*target):"r"(bit) : "memory");
}
extern u64 idle_cpu_mask;
|
replace fd with mmapfd in adc-test-server.c | @@ -42,8 +42,8 @@ int main ()
return 1;
}
- slcr = mmap(NULL, sysconf(_SC_PAGESIZE), PROT_READ|PROT_WRITE, MAP_SHARED, fd, 0xF8000000);
- axi_hp0 = mmap(NULL, sysconf(_SC_PAGESIZE), PROT_READ|PROT_WRITE, MAP_SHARED, fd, 0xF8008000);
+ slcr = mmap(NULL, sysconf(_SC_PAGESIZE), PROT_READ|PROT_WRITE, MAP_SHARED, mmapfd, 0xF8000000);
+ axi_hp0 = mmap(NULL, sysconf(_SC_PAGESIZE), PROT_READ|PROT_WRITE, MAP_SHARED, mmapfd, 0xF8008000);
cfg = mmap(NULL, sysconf(_SC_PAGESIZE), PROT_READ|PROT_WRITE, MAP_SHARED, mmapfd, 0x40000000);
sts = mmap(NULL, sysconf(_SC_PAGESIZE), PROT_READ|PROT_WRITE, MAP_SHARED, mmapfd, 0x40001000);
ram = mmap(NULL, 2048*sysconf(_SC_PAGESIZE), PROT_READ|PROT_WRITE, MAP_SHARED, mmapfd, 0x1E000000);
|
netutils/netlib/netlib_getarptab.c: Do not initialize response buffer, it is write-only. Also add a check to assue that the sequence number in the response is the same as the sequence number in the request. | @@ -168,19 +168,7 @@ ssize_t netlib_get_arptable(FAR struct arp_entry_s *arptab, unsigned int nentrie
goto errout_with_socket;
}
- /* Initialize the response buffer.
- * REVISIT: Linux examples that I have seen just pass a raw buffer. I am
- * not sure how they associate the requested data to the recv() without a
- * sequence number.
- */
-
- memset(&resp->hdr, 0, sizeof(resp->hdr));
- resp->hdr.nlmsg_len = NLMSG_LENGTH(sizeof(struct ndmsg));
- resp->hdr.nlmsg_seq = thiseq;
- resp->hdr.nlmsg_type = RTM_GETNEIGH;
-
- memset(&resp->msg, 0, sizeof(resp->msg));
- resp->msg.ndm_family = AF_INET;
+ /* Read the response */
nrecvd = recv(fd, resp, allocsize, 0);
if (nrecvd < 0)
@@ -201,6 +189,18 @@ ssize_t netlib_get_arptable(FAR struct arp_entry_s *arptab, unsigned int nentrie
goto errout_with_socket;
}
+ /* The sequence number in the response should match the sequence
+ * number in the request (since we created the socket, this should
+ * always be tree).
+ */
+
+ if (resp->hdr.nlmsg_seq != thiseq)
+ {
+ fprintf(stderr, "ERROR: Bad sequence number in response\n");
+ ret = -EIO;
+ goto errout_with_socket;
+ }
+
/* Copy the ARP table data to the caller's buffer */
paysize = resp->attr.rta_len;
|
hv:enable the O2 option for HV
commit removed O2 option improperly,
this patch re-enable it. | @@ -74,6 +74,7 @@ CFLAGS += -m64 -mno-mmx -mno-sse -mno-sse2 -mno-80387 -mno-fp-ret-in-387
CFLAGS += -mno-red-zone
CFLAGS += -nostdinc -nostdlib -fno-common
CFLAGS += -Werror
+CFLAGS += -O2
ifeq (y, $(CONFIG_RELOC))
CFLAGS += -fpie
else
|
Refactor Certificate | @@ -508,7 +508,7 @@ int mbedtls_ssl_tls13_parse_certificate( mbedtls_ssl_context *ssl,
MBEDTLS_SSL_CHK_BUF_READ_PTR( p, certificate_list_end, extensions_len );
extensions_end = p + extensions_len;
- handshake->received_extensions = MBEDTLS_SSL_EXT_NONE;
+ handshake->received_extensions = MBEDTLS_SSL_EXT_MASK_NONE;
while( p < extensions_end )
{
@@ -537,18 +537,16 @@ int mbedtls_ssl_tls13_parse_certificate( mbedtls_ssl_context *ssl,
switch( extension_type )
{
default:
- MBEDTLS_SSL_DEBUG_MSG( 3,
- ( "Certificate: received %s(%u) extension ( ignored )",
- mbedtls_tls13_get_extension_name( extension_type ),
- extension_type ) );
+ MBEDTLS_SSL_PRINT_EXT_TYPE(
+ 3, MBEDTLS_SSL_HS_CERTIFICATE,
+ extension_type, "( ignored )" );
break;
}
p += extension_data_len;
}
- MBEDTLS_SSL_TLS1_3_PRINT_EXTS(
- 3, MBEDTLS_SSL_HS_CERTIFICATE, handshake->received_extensions );
+ MBEDTLS_SSL_PRINT_RECEIVED_EXTS( 3, MBEDTLS_SSL_HS_CERTIFICATE );
}
exit:
|
add Fedora building instructions | @@ -105,6 +105,18 @@ git clone --recursive https://github.com/nesbox/TIC-80 && cd TIC-80/build
cmake ..
make -j4
```
+
+### Fedora
+
+run the following commands in the Terminal
+```
+sudo dnf -y groupinstall "Development Tools"
+sudo dnf -y install git cmake libglvnd-devel libglvnd-gles freeglut-devel
+git clone --recursive https://github.com/nesbox/TIC-80 && cd TIC-80/build
+cmake ..
+make -j4
+```
+
Install with [Install instructions](#install-instructions)
### Raspberry Pi (Retropie)
|
dnsmasq: add support for ttl cache and filtering ipv6 record | @@ -184,6 +184,11 @@ return L.view.extend({
s.taboption('files', form.DynamicList, 'addnhosts',
_('Additional Hosts files')).optional = true;
+ o = s.taboption('advanced', form.Flag, 'filter_aaaa',
+ _('Disable IPv6 DNS forwards'),
+ _('Filter IPv6(AAAA) DNS Query Name Resolve'));
+ o.optional = true;
+
o = s.taboption('advanced', form.Flag, 'quietdhcp',
_('Suppress logging'),
_('Suppress logging of the routine operation of these protocols'));
@@ -342,6 +347,13 @@ return L.view.extend({
o.datatype = 'range(0,10000)';
o.placeholder = 150;
+ o = s.taboption('advanced', form.Value, 'mini_ttl',
+ _('Minimum TTL to send to clients'),
+ _('Modify DNS entries minimum TTL (max is 86400, 0 is no modify)'));
+ o.optional = true;
+ o.datatype = 'range(0,86400)';
+ o.placeholder = 0;
+
s.taboption('tftp', form.Flag, 'enable_tftp',
_('Enable TFTP server')).optional = true;
|
smaller window image | [](LICENSE.md)
[](https://bestpractices.coreinfrastructure.org/projects/2799)
-
+
**OpenEXR** is a high dynamic-range (HDR) image file format originally
developed by Industrial Light & Magic (ILM) in 2003 for use in
@@ -56,7 +56,7 @@ The OpenEXR distribution consists of the following sub-modules:
* **Contrib** - Various plugins and utilities, contributed by the community.
A collection of OpenEXR images is available from the adjacent repository
-[openexr-images](https://github.com/openexr/openexr-images).
+https://github.com/openexr/openexr-images.
## Supported Platforms
|
More safec constraint avoidance | @@ -1040,7 +1040,7 @@ static ACVP_RESULT match_oes_page(ACVP_CTX *ctx,
for (k = 0; k < oe->dependencies.count; k++) {
int diff = 0;
const char *parsed_url = json_array_get_string(dependency_urls, k);
-
+ if (parsed_url) {
strcmp_s(oe->dependencies.deps[k]->url, ACVP_ATTR_URL_MAX,
parsed_url, &diff);
if (diff != 0) {
@@ -1048,6 +1048,7 @@ static ACVP_RESULT match_oes_page(ACVP_CTX *ctx,
break;
}
}
+ }
if (equal) {
/*
@@ -1280,11 +1281,13 @@ static int compare_emails(ACVP_STRING_LIST *email_list,
const char *tmp_email = NULL;
tmp_email = json_array_get_string(candidate_emails, i);
+ if (tmp_email) {
strcmp_s(email, ACVP_OE_STR_MAX, tmp_email, &diff);
if (diff != 0) {
*match = 0;
return ACVP_SUCCESS;
}
+ }
email_list = email_list->next;
i++;
@@ -1351,18 +1354,22 @@ static ACVP_RESULT compare_phone_numbers(ACVP_OE_PHONE_LIST *phone_list,
}
tmp_number = json_object_get_string(obj, "number");
+ if (tmp_number) {
strcmp_s(phone_list->number, ACVP_OE_STR_MAX, tmp_number, &diff);
if (diff != 0) {
*match = 0;
return ACVP_SUCCESS;
}
+ }
tmp_type = json_object_get_string(obj, "type");
+ if (tmp_type) {
strcmp_s(phone_list->type, ACVP_OE_STR_MAX, tmp_type, &diff);
if (diff != 0) {
*match = 0;
return ACVP_SUCCESS;
}
+ }
phone_list = phone_list->next;
i++;
|
Delete recently documented functions from missingcrypto | @@ -768,7 +768,6 @@ PKCS12_SAFEBAG_create_cert
PKCS12_SAFEBAG_create_crl
PKCS12_SAFEBAG_create_pkcs8_encrypt
PKCS12_SAFEBAG_get0_attr
-PKCS12_SAFEBAG_get0_attrs
PKCS12_SAFEBAG_get0_p8inf
PKCS12_SAFEBAG_get0_pkcs8
PKCS12_SAFEBAG_get0_safes
@@ -778,21 +777,14 @@ PKCS12_SAFEBAG_get1_crl
PKCS12_SAFEBAG_get_bag_nid
PKCS12_SAFEBAG_get_nid
PKCS12_SAFEBAG_it
-PKCS12_add_CSPName_asc
PKCS12_add_cert
-PKCS12_add_friendlyname_asc
-PKCS12_add_friendlyname_uni
-PKCS12_add_friendlyname_utf8
PKCS12_add_key
-PKCS12_add_localkeyid
PKCS12_add_safe
PKCS12_add_safes
PKCS12_decrypt_skey
PKCS12_gen_mac
PKCS12_get0_mac
PKCS12_get_attr
-PKCS12_get_attr_gen
-PKCS12_get_friendlyname
PKCS12_init
PKCS12_it
PKCS12_item_decrypt_d2i
|
Ruby libtools: install kdbtools + fix build process | @@ -57,8 +57,29 @@ else()
set (CMAKE_LIBRARY_OUTPUT_DIRECTORY "${CMAKE_SWIG_OUTDIR}")
set (CMAKE_SWIG_FLAGS "-O;-autorename;-DSWIG_NO_EXPORT_ITERATOR_METHODS;-DSWIG_WITHOUT_OVERRIDE")
+ # specify the SWIG_TYPE_TABLE to use (has to be in sync with the ruby plugin)
+ set (SWIG_COMPILE_FLAGS "${SWIG_COMPILE_FLAGS} -DSWIG_TYPE_TABLE=kdb")
+
+ # disable certain compiler warnings for SWIG generated files
+ set (SWIG_COMPILE_FLAGS "${SWIG_COMPILE_FLAGS} -Wno-unused-parameter -Wno-unused-but-set-variable")
+ set (SWIG_COMPILE_FLAGS "${SWIG_COMPILE_FLAGS} -Wno-sign-compare")
+
+ # add 'kdb' module
swig_add_module (swig-ruby ruby kdb.i)
+ # set the compiler settings for the generated file
+ # (has to be done for each module separately)
+ set_source_files_properties (
+ ${swig_generated_sources} PROPERTIES
+ COMPILE_FLAGS "${SWIG_COMPILE_FLAGS}"
+ )
+
+ # add the 'kdbtools' module
swig_add_module (swig-ruby-tools ruby kdbtools.i)
+ set_source_files_properties (
+ ${swig_generated_sources} PROPERTIES
+ COMPILE_FLAGS "${SWIG_COMPILE_FLAGS}"
+ )
+
swig_link_libraries (swig-ruby ${RUBY_LIBRARY} elektra-core elektra-kdb)
set_target_properties (swig-ruby PROPERTIES
OUTPUT_NAME _kdb
@@ -71,11 +92,6 @@ else()
PREFIX ""
)
- set_source_files_properties (
- ${swig_generated_file_fullname} PROPERTIES
- COMPILE_FLAGS "${SWIG_COMPILE_FLAGS} -Wno-unused-parameter -Wno-unused-but-set-variable -DSWIG_TYPE_TABLE=kdb -Wno-sign-compare -Wno-old-style-cast"
- )
-
# CMAKE_INSTALL_PREFIX dependent install location
# if we hit one of the usual cases (/usr/local or /usr) install the lib
@@ -100,11 +116,11 @@ else()
endif()
install (
- TARGETS swig-ruby
+ TARGETS swig-ruby swig-ruby-tools
LIBRARY DESTINATION ${RUBY_LIB_INSTALL_DIR}
)
install (
- FILES ${CMAKE_CURRENT_SOURCE_DIR}/kdb.rb
+ FILES ${CMAKE_CURRENT_SOURCE_DIR}/kdb.rb ${CMAKE_CURRENT_SOURCE_DIR}/kdbtools.rb
DESTINATION ${RUBY_MODULE_INSTALL_DIR}
)
|
Reset iVerbose in examples/AtmEscKepler-36 to 0. | @@ -36,7 +36,7 @@ saOutputOrder Time -EnvelopeMass Radius
system = """#
sSystemName kepler36
-iVerbose 5
+iVerbose 0
bOverwrite 1
saBodyFiles star.in %s
sUnitMass solar
|
GCC 11 build fix
During Adafruit Bootloader compilation, I spotted bellow error which do not allow me build project.
``` c
inlined from 'hfclk_running' at lib/tinyusb/src/portable/nordic/nrf5x/dcd_nrf5x.c:785:13:
lib/tinyusb/src/portable/nordic/nrf5x/dcd_nrf5x.c:792:31: error: 'is_running' may be used uninitialized [-Werror=maybe-uninitialized]
792 | return (is_running ? true : false);
| ~~~~~~~~~~~~~~~~~~~^~~~~~~~
``` | @@ -787,7 +787,7 @@ static bool hfclk_running(void)
#ifdef SOFTDEVICE_PRESENT
if ( is_sd_enabled() )
{
- uint32_t is_running;
+ uint32_t is_running = 0;
(void) sd_clock_hfclk_is_running(&is_running);
return (is_running ? true : false);
}
|
BUFR encoding memory leaks (part 5) | @@ -2944,10 +2944,6 @@ static int process_elements(grib_accessor* a, int flag, long onlySubset, long st
grib_vsarray_delete(c, self->stringValues);
self->stringValues = NULL;
}
- // if (do_clean == 1 && self->tempDoubleValues) {
- // grib_vdarray_delete_content(c, self->tempDoubleValues);
- // grib_vdarray_delete(c, self->tempDoubleValues);
- // }
if (flag != PROCESS_ENCODE) {
self->numericValues = grib_vdarray_new(c, 1000, 1000);
@@ -3003,7 +2999,6 @@ static int process_elements(grib_accessor* a, int flag, long onlySubset, long st
elementsDescriptorsIndex = grib_iarray_new(c, DYN_ARRAY_SIZE_INIT, DYN_ARRAY_SIZE_INCR);
if (!self->compressedData) {
dval = grib_darray_new(c, DYN_ARRAY_SIZE_INIT, DYN_ARRAY_SIZE_INCR);
- //printf("DBG:: NEWED dval=%p\n", (void*)dval);
}
}
else {
|
Parameter: Disable mpfit-velocity | @@ -92,7 +92,7 @@ typedef struct MPFITData {
int record_reprojection_error;
FLT current_bias, obj_up_variance, lh_up_variance, stationary_obj_up_variance;
- int model_velocity;
+ bool model_velocity;
bool globalDataAvailable;
struct survive_async_optimizer *async_optimizer;
@@ -100,7 +100,7 @@ typedef struct MPFITData {
} MPFITData;
STRUCT_CONFIG_SECTION(MPFITData)
-STRUCT_CONFIG_ITEM("mpfit-model-velocity", "Model velocity in non mpfit process", 1, t->model_velocity)
+STRUCT_CONFIG_ITEM("mpfit-model-velocity", "Model velocity in non mpfit process", false, t->model_velocity)
STRUCT_CONFIG_ITEM("mpfit-current-bias", "", 0, t->current_bias)
STRUCT_CONFIG_ITEM("mpfit-record-reprojection-error", "", 0, t->record_reprojection_error)
STRUCT_CONFIG_ITEM("mpfit-object-up-variance",
|
config_tools: hide syntactic error message from configurator ui
hide syntactic error message from configurator ui | @@ -381,7 +381,7 @@ export default {
translateErrors(errors, scenarioData) {
let formErrors = {}
- let translate = error => {
+ let translate = errorType => error => {
error.paths.forEach(path => {
let formPath = path.split('/')[2];
// translate form path to scenario vmid
@@ -398,15 +398,17 @@ export default {
if (!formErrors.hasOwnProperty(vmid)) {
formErrors[vmid] = []
}
+ if (errorType === 'semantic') {
formErrors[vmid].push(error)
+ }
})
}
if (errors.syntactic_errors.length > 0) {
- errors.syntactic_errors.forEach(translate)
+ errors.syntactic_errors.forEach(translate('syntactic'))
}
- if (errors.semantic_errors.length !== 0) {
- errors.semantic_errors.forEach(translate)
+ if (errors.semantic_errors.length > 0) {
+ errors.semantic_errors.forEach(translate('semantic'))
}
return formErrors
|
globally disable tcp proxy in +http-config | ++ host (each turf @if) :: http host
++ hoke %+ each {$localhost ~} :: local host
?($.0.0.0.0 $.127.0.0.1) ::
- :: +http-config: full http-server configuration
- ::
- += http-config
- $: :: secure: PEM-encoded RSA private key and cert or cert chain
- ::
- secure=(unit [key=wain cert=wain])
- :: proxy: reverse TCP proxy HTTP(s)
- ::
- proxy=?
- :: log: keep HTTP(s) access logs
- ::
- log=?
- :: redirect: send 301 redirects to upgrade HTTP to HTTPS
- ::
- :: Note: requires certificate.
- ::
- redirect=?
- ==
- :: +http-rule: update configuration
- ::
- += http-rule
- $% :: %cert: set or clear certificate and keypair
- ::
- [%cert cert=(unit [key=wain cert=wain])]
- :: %turf: add or remove established dns binding
- ::
- [%turf action=?(%put %del) =turf]
- ==
++ httq :: raw http request
$: p/meth :: method
q/@t :: unparsed url
[%live insecure=@ud secure=(unit @ud)]
:: update http configuration
::
- [%rule =http-rule:eyre]
+ [%rule =http-rule]
:: starts handling an inbound http request
::
[%request secure=? =address =request:http]
secure=(unit [key=wain cert=wain])
:: proxy: reverse TCP proxy HTTP(s)
::
- proxy=?
+ proxy=_|
:: log: keep HTTP(s) access logs
::
log=?
::
redirect=?
==
+ :: +http-rule: update configuration
+ ::
+ +$ http-rule
+ $% :: %cert: set or clear certificate and keypair
+ ::
+ [%cert cert=(unit [key=wain cert=wain])]
+ :: %turf: add or remove established dns binding
+ ::
+ [%turf action=?(%put %del) =turf]
+ ==
:: +address: client IP address
::
+$ address
|
cc3200/mods: Include stream.h to get definition of mp_stream_p_t. | #ifndef MICROPY_INCLUDED_CC3200_MODS_MODUSOCKET_H
#define MICROPY_INCLUDED_CC3200_MODS_MODUSOCKET_H
+#include "py/stream.h"
+
extern const mp_obj_dict_t socket_locals_dict;
extern const mp_stream_p_t socket_stream_p;
|
Removing tabs and fixing doxygen formatting | @@ -421,7 +421,7 @@ int mbedtls_ecdsa_sign( mbedtls_ecp_group *grp, mbedtls_mpi *r, mbedtls_mpi *s,
/*
* Deterministic signature wrapper
*
- * \note The f_rng_blind parameter must not be \c NULL.
+ * note: The f_rng_blind parameter must not be NULL.
*
*/
static int ecdsa_sign_det_restartable( mbedtls_ecp_group *grp,
|
highlevel: fix memory leak | @@ -88,6 +88,7 @@ Elektra * elektraOpen (const char * application, KeySet * defaults, KeySet * con
{
keyDel (contractCut);
ksDel (highlevelContract);
+ ksDel (contract); // consume contract, like kdbEnsure would
return NULL;
}
}
|
IsFlat: inline when possible | @@ -977,7 +977,8 @@ static void SwapOut(VP8EncIterator* const it) {
SwapPtr(&it->yuv_out_, &it->yuv_out2_);
}
-static int IsFlat(const int16_t* levels, int num_blocks, int thresh) {
+static WEBP_INLINE int IsFlat(const int16_t* levels, int num_blocks,
+ int thresh) {
int score = 0;
while (num_blocks-- > 0) { // TODO(skal): refine positional scoring?
int i;
|
fix llvm compilation error
The following changes from llvm 10
changed MCDisassembler::getInstruction() signature.
Do the corresponding change in bcc_debug.cc to fix the issue. | @@ -201,8 +201,13 @@ void SourceDebugger::dump() {
string src_dbg_str;
llvm::raw_string_ostream os(src_dbg_str);
for (uint64_t Index = 0; Index < FuncSize; Index += Size) {
+#if LLVM_MAJOR_VERSION >= 10
+ S = DisAsm->getInstruction(Inst, Size, Data.slice(Index), Index,
+ nulls());
+#else
S = DisAsm->getInstruction(Inst, Size, Data.slice(Index), Index,
nulls(), nulls());
+#endif
if (S != MCDisassembler::Success) {
os << "Debug Error: disassembler failed: " << std::to_string(S)
<< '\n';
|
bugfix for op_jmpuw. | @@ -923,15 +923,22 @@ static inline int op_jmpuw( mrbc_vm *vm, mrbc_value *regs )
// Check ensure
const mrbc_irep_catch_handler *handler = catch_handler_find(vm, 0, MRBC_CATCH_FILTER_ENSURE);
- if( handler ){
- vm->inst = vm->cur_irep->inst + bin_to_uint32(handler->target);
- vm->exc.tt = MRBC_TT_BREAK;
- vm->exc.jmpuw = vm->inst + (int16_t)a;
- } else {
+ if( !handler ) {
vm->inst += (int16_t)a;
- vm->exc = mrbc_nil_value();
+ return 0;
}
+ // found the ensure.
+ uint32_t jump_point = vm->inst - vm->cur_irep->inst + (int16_t)a;
+ if( (bin_to_uint32(handler->begin) < jump_point) &&
+ (jump_point <= bin_to_uint32(handler->end)) ) {
+ vm->inst += (int16_t)a;
+ return 0;
+ }
+
+ vm->exc.tt = MRBC_TT_BREAK;
+ vm->exc.jmpuw = vm->inst + (int16_t)a;
+ vm->inst = vm->cur_irep->inst + bin_to_uint32(handler->target);
return 0;
}
@@ -951,12 +958,8 @@ static inline int op_except( mrbc_vm *vm, mrbc_value *regs )
FETCH_B();
mrbc_decref( ®s[a] );
- if( mrbc_israised(vm) ){
regs[a] = vm->exc;
vm->exc = mrbc_nil_value();
- } else {
- regs[a] = mrbc_nil_value();
- }
return 0;
}
@@ -1007,14 +1010,13 @@ static inline int op_raiseif( mrbc_vm *vm, mrbc_value *regs )
{
FETCH_B();
- // mrbc_printf("OP_RAISEIF tt:%d\n", regs[0].tt);
-
mrb_value *exc = ®s[a];
if( exc->tt == MRBC_TT_BREAK ){
vm->inst = exc->jmpuw;
} else {
- mrbc_incref( ®s[a] );
+ assert( mrbc_type(regs[a]) == MRBC_TT_EXCEPTION ||
+ mrbc_type(regs[a]) == MRBC_TT_NIL );
vm->exc = regs[a];
}
@@ -2700,6 +2702,7 @@ void mrbc_vm_begin( struct VM *vm )
vm->target_class = mrbc_class_object;
vm->exc = mrbc_nil_value();
+ vm->exc_message = mrbc_nil_value();
vm->error_code = 0;
vm->flag_preemption = 0;
|
OcAppleKernelLib: Consider positive replace count as successful and log replace mismatch | @@ -160,8 +160,16 @@ PatcherApplyGenericPatch (
Patch->Skip
);
- if ((ReplaceCount > 0 && Patch->Count == 0)
- || (ReplaceCount == Patch->Count && Patch->Count > 0)) {
+ if (ReplaceCount > 0 && Patch->Count > 0 && ReplaceCount != Patch->Count) {
+ DEBUG ((
+ DEBUG_INFO,
+ "Performed only %u replacements out of %u\n",
+ ReplaceCount,
+ Patch->Count
+ ));
+ }
+
+ if (ReplaceCount > 0) {
return EFI_SUCCESS;
}
|
For integration tests, restore download from cdn.
This effectively undoes pull request | @@ -55,9 +55,9 @@ stages:
targetFolder: ${{ format('dl/scope/branch/{0}/', variables['Build.SourceBranchName']) }}
filesAcl: 'public-read'
- stage: integration_tests
- displayName: Run integraion tests
+ displayName: Run integration tests
dependsOn: build
- #condition: and(succeeded(), eq(variables['Build.SourceBranch'], 'refs/heads/master'))
+ condition: and(succeeded(), eq(variables['Build.SourceBranch'], 'refs/heads/master'))
jobs:
- template: test/testContainers/azure/it-test-job-template.yml
parameters:
|
BugID:17338742:delete old benchmarks and perf | @@ -9,9 +9,4 @@ GLOBAL_DEFINES += AOS_NO_WIFI
$(NAME)_COMPONENTS := yloop cli
-ifeq ($(BENCHMARKS),1)
-$(NAME)_COMPONENTS += benchmarks
-GLOBAL_DEFINES += CONFIG_CMD_BENCHMARKS
-endif
-
GLOBAL_INCLUDES += ./
|
FIXed KeyboardDisplayRequiresUserAction on CWebViewImpl | @@ -203,6 +203,9 @@ public:
//Android only
virtual void save( const rho::String& format, const rho::String& path, int tabIndex, rho::apiGenerator::CMethodResult& oResult){}
//
+
+ void getKeyboardDisplayRequiresUserAction(rho::apiGenerator::CMethodResult& oResult){}
+ void setKeyboardDisplayRequiresUserAction( bool keyboardDisplayRequiresUserAction, rho::apiGenerator::CMethodResult& oResult){}
};
////////////////////////////////////////////////////////////////////////
|
docs; fix blog template missing links and wrong paths | <nav class="small">
<table>
<tr>
- <td><a href="getting-started.html">Getting Started</a></td>
- <td><a href="contributing.html">Contributing</a></td>
+ <td>
+ <ul>
+ <li><a href="../getting-started.html">Getting Started</a></li>
+ <li><a href="../contributing.html">Contributing</a></li>
+ <li><a href="../blog">Blog</a></li>
+ <li><a href="../try">Try it!</a></li>
+ </ul>
+ </td>
</tr>
</table>
</nav>
|
Use SSL by default | @@ -54,6 +54,8 @@ DYLIB_MAKE_CMD=$(CC) -shared -Wl,-soname,$(DYLIB_MINOR_NAME) -o $(DYLIBNAME) $(L
STLIBNAME=$(LIBNAME).$(STLIBSUFFIX)
STLIB_MAKE_CMD=$(AR) rcs $(STLIBNAME)
+USE_SSL:=1
+
ifdef USE_SSL
# This is the prefix of openssl on my system. This should be the sane default
# based on the platform
|
Encrypt Coverity Scan variables for use with NLnet Labs repository | @@ -97,8 +97,8 @@ submit_to_coverity_scan: &submit_to_coverity_scan
"https://scan.coverity.com/builds"
env:
- COVERITY_EMAIL: ENCRYPTED[23892f3093e665828e3d67570ed20ee54c4740eb44b26770eb2e43b44427d3c2cf9452d829d6fc226553b0cf6f2a5192]
- COVERITY_SCAN_TOKEN: ENCRYPTED[58f70d80a6be95eb448c57f958c7d5f4b4d93a3c5bc6c5deff246bc7c5e8d692906de8568d27475718f4ab2d2e000541]
+ COVERITY_EMAIL: ENCRYPTED[effa3340c97e8cf92c0dbb564187d35b6829580cc2577b176d6c6fc9b775745f7130c56f5bd9ab2472f4ae818b6f3791]
+ COVERITY_SCAN_TOKEN: ENCRYPTED[8f67f850ca3d464ea87fa8dee17bbb0cfb2a991b6f401fd593fe0744eece838e325af438d62ee2d46c4e18a2bd5c873f]
task:
only_if: $CIRRUS_CRON != ''
|
error: Removed strerror from method which does not set errno | @@ -75,8 +75,8 @@ int elektraPortInfo (Key * toCheck, Key * parentKey)
service = getservbyname (keyString (toCheck), NULL); // NULL means we accept both tcp and udp
if (service == NULL)
{
- ELEKTRA_SET_VALIDATION_SEMANTIC_ERRORF (parentKey, "Could not find service with name %s on key %s. Reason: %s",
- keyString (toCheck), keyName (toCheck), strerror (errno));
+ ELEKTRA_SET_VALIDATION_SEMANTIC_ERRORF (parentKey, "Could not find service with name %s on key %s",
+ keyString (toCheck), keyName (toCheck));
return -1;
}
portNumberNetworkByteOrder = service->s_port;
|
test for HTTPS added | @@ -21,6 +21,7 @@ tests_general.append([1, 0, workdir + 'client_http_get -u /cgi-bin/he -v 2 not.r
tests_general.append([1, 0, workdir + 'client_http_get -u /cgi-bin/he -v 2 buildbot.nplab.de'])
tests_general.append([0, 0, workdir + 'client_http_get -n 2 -u /files/4M bsd10.nplab.de'])
if (platform.system() == "FreeBSD") or (platform.system() == "Linux"):
+ tests_general.append([1, 0, workdir + 'client_http_get -P ' + workdir = 'prop_tcp_security.json -p 443 -v 2 www.neat-project.org'])
tests_general.append([0, 0, workdir + 'tneat -v 2 -P ' + workdir + 'prop_sctp_dtls.json interop.fh-muenster.de'])
#tests_general.append([0, 0, 'python3.5 ../../policy/pmtests.py'])
|
Wii port now correctly exit to homebrew channel | @@ -41,7 +41,7 @@ char rootDir[MAX_FILENAME_LEN]; // note: this one ends with a slash
*/
char* getFullPath(char *relPath)
{
- static char filename[256];
+ static char filename[MAX_FILENAME_LEN];
strcpy(filename, rootDir);
strcat(filename, relPath);
return filename;
@@ -54,7 +54,7 @@ void borExit(int reset)
else if(reset == WII_RESET) SYS_ResetSystem(SYS_HOTRESET, 0, 0);
else exit(reset);
#else
- SYS_ResetSystem(SYS_RETURNTOMENU, 0, 0); //exit(reset);
+ exit(reset); //SYS_ResetSystem(SYS_RETURNTOMENU, 0, 0);
#endif
}
|
Update duckduckgo format | @@ -179,7 +179,7 @@ VOID PhAddDefaultSettings(
PhpAddIntegerSetting(L"SampleCount", L"200"); // 512
PhpAddIntegerSetting(L"SampleCountAutomatic", L"1");
PhpAddIntegerSetting(L"ScrollToNewProcesses", L"0");
- PhpAddStringSetting(L"SearchEngine", L"https://www.duckduckgo.com/?q=\"%s\"");
+ PhpAddStringSetting(L"SearchEngine", L"https://duckduckgo.com/?q=\"%s\"");
PhpAddStringSetting(L"SegmentHeapListViewColumns", L"");
PhpAddStringSetting(L"SegmentHeapListViewSort", L"0,1");
PhpAddIntegerPairSetting(L"SegmentHeapWindowPosition", L"0,0");
|
docs: fix minor typo in COPY command. | @@ -400,13 +400,13 @@ COPY {table [(<varname>column</varname> [, ...])] | (<varname>query</varname>)}
with the <codeph>COPY TO</codeph> command. The <codeph>COPY</codeph> command might return
table distribution policy errors, if you attempt to restore table data and the table
distribution policy was changed after the <codeph>COPY FROM...ON SEGMENT</codeph> was
- run.</p><note>If you run <codeph>COPY FROM...ON SEGMENT</codeph>and the server configuration
- parameter <codeph>gp_enable_segment_copy_checking</codeph> is <codeph>false</codeph>, manual
- redistribution of table data might be required. See the <codeph>ALTER TABLE</codeph> clause
- <codeph>WITH REORGANIZE</codeph>.</note><p>When you specify the <codeph>LOG
- ERRORS</codeph> clause, Greenplum Database captures errors that occur while reading the
- external table data. You can view and manage the captured error log data. </p><ul
- id="ul_wk3_jdj_bp">
+ run.</p><note>If you run <codeph>COPY FROM...ON SEGMENT</codeph> and the server
+ configuration parameter <codeph>gp_enable_segment_copy_checking</codeph> is
+ <codeph>false</codeph>, manual redistribution of table data might be required. See the
+ <codeph>ALTER TABLE</codeph> clause <codeph>WITH REORGANIZE</codeph>.</note><p>When you
+ specify the <codeph>LOG ERRORS</codeph> clause, Greenplum Database captures errors that
+ occur while reading the external table data. You can view and manage the captured error log
+ data. </p><ul id="ul_wk3_jdj_bp">
<li>Use the built-in SQL function
<codeph>gp_read_error_log('<varname>table_name</varname>')</codeph>. It requires
<codeph>SELECT</codeph> privilege on <varname>table_name</varname>. This example
|
Improve Luos engine compilation script prints | @@ -12,6 +12,8 @@ visited_key = "__LUOS_CORE_SCRIPT_CALLED"
global_env = DefaultEnvironment()
if not visited_key in global_env:
+ click.secho("")
+ click.secho("Luos engine build configuration:", underline=True)
# install pyluos
try:
import pyluos
@@ -19,7 +21,6 @@ if not visited_key in global_env:
except ImportError: # module doesn't exist, deal with it.
env.Execute("$PYTHONEXE -m pip install pyluos")
pass
- click.secho("Luos engine build configuration:\n", underline=True)
try:
from pyluos import version
click.secho("\t* Pyluos revision " +
@@ -44,10 +45,10 @@ for item in env.get("CPPDEFINES", []):
if (path.exists("network/robus/HAL/" + item[1]) and path.exists("engine/HAL/" + item[1])):
if not visited_key in global_env:
click.secho(
- "\t* %s HAL selected for Luos and Robus.\n" % item[1], fg="green")
+ "\t* %s HAL selected for Luos and Robus." % item[1], fg="green")
else:
if not visited_key in global_env:
- click.secho("\t* %s HAL not found\n" % item[1], fg="red")
+ click.secho("\t* %s HAL not found" % item[1], fg="red")
env.Append(CPPPATH=[realpath("network/robus/HAL/" + item[1])])
env.Append(CPPPATH=[realpath("engine/HAL/" + item[1])])
env.Replace(SRC_FILTER=sources)
@@ -66,7 +67,7 @@ for item in env.get("CPPDEFINES", []):
current_os = pf.system()
if find_MOCK_HAL == False:
click.secho(
- "\t* Mock HAL for %s is selected for Luos and Robus.\n" % current_os, fg="green")
+ "\t* Mock HAL for %s is selected for Luos and Robus." % current_os, fg="green")
find_MOCK_HAL = True
find_HAL = True
env.Replace(SRC_FILTER=sources)
@@ -103,6 +104,6 @@ for item in env.get("CPPDEFINES", []):
if not visited_key in global_env:
if (find_HAL == False):
click.echo(click.style("\t* No HAL selected. Please add a ", fg="red") + click.style(
- "-DLUOSHAL", fg="red", bold=True) + click.style(" compilation flag\n", fg="red"))
+ "-DLUOSHAL", fg="red", bold=True) + click.style(" compilation flag", fg="red"))
global_env[visited_key] = True
|
out BUGFIX trim print and opque nodes
Fixes cesnet/netopeer2#952 | @@ -43,7 +43,7 @@ ly_should_print(const struct lyd_node *node, uint32_t options)
if (node->flags & LYD_DEFAULT) {
/* implicit default node/NP container with only default nodes */
return 0;
- } else if (node->schema->nodetype & LYD_NODE_TERM) {
+ } else if (node->schema && (node->schema->nodetype & LYD_NODE_TERM)) {
if (lyd_is_default(node)) {
/* explicit default node */
return 0;
|
README.md: add missing pointer to tools/mountsnoop.py
add missing pointer to tools/mountsnoop.py | @@ -118,8 +118,9 @@ pair of .c and .py files, and some are directories of files.
- tools/[killsnoop](tools/killsnoop.py): Trace signals issued by the kill() syscall. [Examples](tools/killsnoop_example.txt).
- tools/[llcstat](tools/llcstat.py): Summarize CPU cache references and misses by process. [Examples](tools/llcstat_example.txt).
- tools/[mdflush](tools/mdflush.py): Trace md flush events. [Examples](tools/mdflush_example.txt).
-- tools/[mysqld_qslower](tools/mysqld_qslower.py): Trace MySQL server queries slower than a threshold. [Examples](tools/mysqld_qslower_example.txt).
- tools/[memleak](tools/memleak.py): Display outstanding memory allocations to find memory leaks. [Examples](tools/memleak_example.txt).
+- tools/[mountsnoop](tools/mountsnoop.py): Trace mount and umount syscalls system-wide. [Examples](tools/mountsnoop_example.txt).
+- tools/[mysqld_qslower](tools/mysqld_qslower.py): Trace MySQL server queries slower than a threshold. [Examples](tools/mysqld_qslower_example.txt).
- tools/[nfsslower](tools/nfsslower.py): Trace slow NFS operations. [Examples](tools/nfsslower_example.txt).
- tools/[nfsdist](tools/nfsdist.py): Summarize NFS operation latency distribution as a histogram. [Examples](tools/nfsdist_example.txt).
- tools/[offcputime](tools/offcputime.py): Summarize off-CPU time by kernel stack trace. [Examples](tools/offcputime_example.txt).
|
fixed touch for player 1 on android (multiple buttons bug) | @@ -589,6 +589,7 @@ void control_update(s_playercontrols ** playercontrols, int numplayers)
}
#ifdef ANDROID
+ if (player <= 0) {
for(i=0;i<JOY_MAX_INPUTS;i++)
{
t = touch_control.settings[i];
@@ -596,6 +597,7 @@ void control_update(s_playercontrols ** playercontrols, int numplayers)
if(keystate_def[t]) k |= (1<<i);
}
}
+ }
#endif
if(usejoy)
|
BugId:19028262: Fix Config.in for mcu esp32 | @@ -2,12 +2,12 @@ config AOS_MCU_ESP32
bool
select AOS_COMP_LWIP
select AOS_COMP_ALICRYPTO
- select AOS_COMP_BT_HOST if BLE_CONFIG = 0
+ select AOS_COMP_BT_HOST if BLE_CONFIG != 0
select AOS_COMP_NETMGR
select AOS_COMP_IMBEDTLS
select AOS_ARCH_XTENSA_LX6 if OSAL_CONFIG = rhino
select AOS_COMP_RHINO if OSAL_CONFIG = rhino
- select AOS_COMP_UMESH if MESH_CONFIG = 0
+ select AOS_COMP_UMESH if MESH_CONFIG != 0
help
driver & sdk for platform/mcu esp32
|
doc: fix STDOUT for storage plugin tutorial and cleanup | @@ -230,9 +230,11 @@ kdb meta-ls /tests/hosts/ipv4/localhost
kdb meta-get /tests/hosts/ipv4/localhost 'comment/#0'
#> test comment
-```
-As you can see the Key that corresponds to the respective line in the hosts file has additional meta-information. This way comments in the configuration file can easily be imported into the KDB. Be aware that the comment also contains the trailing space preceding the text in the comment, which might be confusing. You can also opt to strip preceding and trailing whitespaces entirely.
+# Undo modifications to the key database
+kdb rm -r /tests/hosts
+sudo kdb umount /tests/hosts
+```
## Ordering of Elements
@@ -255,6 +257,7 @@ kdb ls /tests/hosts/ipv4
# Checking the created Meta KeySet
kdb meta-ls /tests/hosts/ipv4/localhost.1
+#> comment/#0
#> order
# Getting the content of the order
@@ -285,6 +288,10 @@ cat `kdb file /tests/hosts`
#> 127.0.0.1 localhost.2
#> 127.0.0.1 localhost.3
#> 127.0.0.1 localhost.4
+
+# Undo modifications to the key database
+kdb rm -r /tests/hosts
+sudo kdb umount /tests/hosts
```
As you can see by setting the order meta Key in the respective KDB entries, we can manipulate the order in which entries get written to the hosts file. Also when importing from the initial hosts file, the plugin stores the correct order in the meta KeySet.
|
Fix input color for ReactSelect in dark theme | @@ -242,6 +242,11 @@ textarea {
font-size: 11px;
}
+.FormField .ReactSelect__input {
+ color: var(--input-text-color);
+ font-size: 11px;
+}
+
.FormField .ReactSelect__control--is-focused {
border: 1px solid transparent !important;
box-shadow: 0 0 0px 2px var(--highlight-color);
|
[chainmaker][#436]add macro define | @@ -25,9 +25,11 @@ extern Suite *make_wallet_suite(void);
extern Suite *make_parameters_suite(void);
extern Suite *make_contract_suite(void);
-char chainmaker_sign_key_buf[1024];
-char chainmaker_sign_cert_buf[1024];
-char chainmaker_ca_cert_buf[1024];
+#define CERT_PRIKEY_LEN 1024
+
+char chainmaker_sign_key_buf[CERT_PRIKEY_LEN];
+char chainmaker_sign_cert_buf[CERT_PRIKEY_LEN];
+char chainmaker_ca_cert_buf[CERT_PRIKEY_LEN];
int read_key_cert_content(char* key_ptr, char* cert_ptr, char* ca_ptr)
{
|
Persistence configuration is optional if payment-processing is disabled and the pool is in relay-mode | @@ -502,10 +502,12 @@ namespace MiningCore
private static void ConfigurePersistence(ContainerBuilder builder)
{
- if (clusterConfig.Persistence == null)
+ if (clusterConfig.Persistence == null &&
+ clusterConfig.PaymentProcessing?.Enabled == true &&
+ string.IsNullOrEmpty(clusterConfig.ShareRelayPublisherUrl))
logger.ThrowLogPoolStartupException("Persistence is not configured!");
- if (clusterConfig.Persistence.Postgres != null)
+ if (clusterConfig.Persistence?.Postgres != null)
ConfigurePostgres(clusterConfig.Persistence.Postgres, builder);
}
|
fix: fix errors of 'LDS' and 'LES'. | @@ -903,16 +903,16 @@ _trace:
_0xC4: /* LES Gd,Ed */
nextop = F8;
GET_ED;
- emu->segs[_ES] = ED->word[0];
+ emu->segs[_ES] = *(__uint16_t*)(((char*)ED)+4);
emu->segs_serial[_ES] = 0;
- GD.dword[0] = *(uint32_t*)(((void*)ED)+2);
+ GD.dword[0] = *(uint32_t*)ED;
NEXT;
_0xC5: /* LDS Gd,Ed */
nextop = F8;
GET_ED;
- emu->segs[_DS] = ED->word[0];
+ emu->segs[_DS] = *(__uint16_t*)(((char*)ED)+4);
emu->segs_serial[_DS] = 0;
- GD.dword[0] = *(uint32_t*)(((void*)ED)+2);
+ GD.dword[0] = *(uint32_t*)ED;
NEXT;
_0xC6: /* MOV Eb,Ib */
nextop = F8;
|
removed option -m (no longer needed) | @@ -438,7 +438,7 @@ eigenpfadname = strdupa(argv[0]);
eigenname = basename(eigenpfadname);
setbuf(stdout, NULL);
-while ((auswahl = getopt(argc, argv, "i:O:o:m:hv")) != -1)
+while ((auswahl = getopt(argc, argv, "i:O:o:hv")) != -1)
{
switch (auswahl)
{
|
BugID:18072832:starterkit debug feature porting | @@ -155,6 +155,7 @@ void SysTick_Handler(void)
krhino_intrpt_exit();
}
+#if (DEBUG_CONFIG_PANIC != 1)
void HardFault_Handler(void)
{
while (1)
@@ -164,6 +165,7 @@ void HardFault_Handler(void)
// #endif
}
}
+#endif
/**
* @brief Retargets the C library printf function to the USART.
|
make: enable parallel compilation | @@ -18,7 +18,7 @@ TARGET = riscv64
VERSION = 2.80
CORE_VERSION = 1.0
SRCDIR := $(CURDIR)
-MAKEFLAGS += --no-print-directory
+MAKEFLAGS += --no-print-directory --output-sync
SUBSYSTEMS = lib vm proc test
@@ -66,18 +66,13 @@ all: subsystems main.o syscalls.o $(EMBED) $(BIN)
printf " data=%-5d text=%-5d rodata=%-5d\n" $$datasz $$textsz $$rodatasz)
-subsystems:
- @for i in $(SUBDIRS); do\
- d=`pwd`;\
- echo "\033[1;32mCOMPILE $$i\033[0m";\
- if ! cd $$i; then\
- exit 1;\
- fi;\
- if ! make; then\
+subsystems: $(ARCHS)
+
+%/$(ARCH):
+ @+echo "\033[1;32mCOMPILE $(@D)\033[0m";\
+ if ! make -C "$(@D)"; then\
exit 1;\
fi;\
- cd $$d;\
- done;
$(BIN): $(ARCHS) $(EMBED)
@@ -109,7 +104,7 @@ depend:
if ! cd $$i; then\
exit 1;\
fi;\
- if ! make -s depend; then\
+ if ! $(MAKE) -s depend; then\
exit 1;\
fi;\
cd $$d;\
@@ -125,7 +120,7 @@ clean:
if ! cd $$i; then\
exit 1;\
fi;\
- if ! make clean; then\
+ if ! $(MAKE) clean; then\
exit 1;\
fi;\
cd $$d;\
|
[chainmaker] add wallet config struct | @@ -28,7 +28,8 @@ api_ethereum.h is header file for BoAT IoT SDK ethereum's interface.
/*! @defgroup eth-api boat ethereum-API
* @{
*/
-
+#define BOAT_CHAINMAKER_CERT_MAX_LEN 1024
+#define BOAT_CHAINMAKER_ROOTCA_MAX_NUM 3
typedef struct TBoatChainmakerSender {
// organization identifier of the member
char* orgId;
@@ -79,7 +80,35 @@ typedef struct TBoatChainmkaerTxRequest {
char* signature;
} BoatChainmkaerTxRequest;
-
+//! chainmaker certificate information config structure
+typedef struct TBoatChainmakerCertInfoCfg {
+ BUINT32 length; //!< length of certificate content, this length contains the terminator '\0'.
+ BCHAR content[BOAT_CHAINMAKER_CERT_MAX_LEN]; //!< content of certificate.
+} BoatChainmakerCertInfoCfg;
+
+typedef struct TBoatChainmakerNode{
+ BCHAR* addr;
+ BUINT32 connCnt;
+ bool useTLS;
+ BCHAR* tlsHostName;
+
+} BoatChainmakerNode;
+
+// chainmaker wallet config structure
+typedef struct TBoatChainmakerWalletConfig
+{
+ BoatWalletPriKeyCtx_config user_pri_key_config;
+ // certificate content of account
+ BoatChainmakerCertInfoCfg user_cert_content;
+ BoatWalletPriKeyCtx_config user_sign_pri_key_config;
+ //certificate content of TLS
+ BoatChainmakerCertInfoCfg user_sign_cert_content;
+
+ BUINT32 rootCaNumber; //!< The number of rootCA file to be set
+ BoatChainmakerCertInfoCfg roo_ca_cnotent[BOAT_CHAINMAKER_ROOTCA_MAX_NUM];//!< certificate content of rootCA
+
+ BoatChainmakerNode node_info;
+}BoatChainmakerWalletConfig;
|
phb4: Split clearing REGB into new function
So others can call it. | @@ -1690,6 +1690,17 @@ static void phb4_rc_err_clear(struct phb4 *p)
0xffffffff);
}
+static void phb4_err_clear_regb(struct phb4 *p)
+{
+ uint64_t val64;
+
+ val64 = phb4_read_reg(p, PHB_REGB_ERR_STATUS);
+ phb4_write_reg(p, PHB_REGB_ERR_STATUS, val64);
+ phb4_write_reg(p, PHB_REGB_ERR1_STATUS, 0x0ul);
+ phb4_write_reg(p, PHB_REGB_ERR_LOG_0, 0x0ul);
+ phb4_write_reg(p, PHB_REGB_ERR_LOG_1, 0x0ul);
+}
+
/*
* The function can be called during error recovery for all classes of
* errors. This is new to PHB4; previous revisions had separate
@@ -1716,11 +1727,7 @@ static void phb4_err_clear(struct phb4 *p)
phb4_write_reg(p, PHB_PBL_ERR_LOG_1, 0x0ul);
/* Rec 24...31: Clear REGB errors */
- val64 = phb4_read_reg(p, PHB_REGB_ERR_STATUS);
- phb4_write_reg(p, PHB_REGB_ERR_STATUS, val64);
- phb4_write_reg(p, PHB_REGB_ERR1_STATUS, 0x0ul);
- phb4_write_reg(p, PHB_REGB_ERR_LOG_0, 0x0ul);
- phb4_write_reg(p, PHB_REGB_ERR_LOG_1, 0x0ul);
+ phb4_err_clear_regb(p);
/* Rec 32...59: Clear PHB error trap */
val64 = phb4_read_reg(p, PHB_TXE_ERR_STATUS);
|
Try fixing issue with Travis-CI by removing postgresql-9.3 package. | @@ -20,6 +20,7 @@ matrix:
install: |
sudo dpkg --add-architecture i386 && sudo rm -rf /etc/apt/sources.list.d/
sudo apt-get update
+ sudo apt-get remove postgresql-9.3
sudo apt-get upgrade -y gcc
sudo apt-get install -y gcc-multilib autoconf automake libtool libjansson-dev:i386 libmagic-dev:i386 libssl-dev:i386
- os: linux
|
net/lwip: fix exception for prefix lifetime
The prefix will be immediately deleted when
it receives RA message with prefix lifetime 0 | @@ -2388,7 +2388,9 @@ static u8_t nd6_prefix_lifetime_isvalid(const struct nd6_prefix_list_entry *pref
if (prefix_entry == NULL)
return 0;
- if ((lifetime_ms > prefix_entry->invalidation_timer) || (lifetime_ms > TIME_HOUR_TO_MS(2))) {
+ if ((lifetime_ms > prefix_entry->invalidation_timer) ||
+ (lifetime_ms > TIME_HOUR_TO_MS(2)) ||
+ (lifetime_ms == 0)) {
return 1;
}
|
TEST: added bop mget test case having key list with wrong delimiter. | #!/usr/bin/perl
use strict;
-use Test::More tests => 69;
+use Test::More tests => 70;
use FindBin qw($Bin);
use lib "$Bin/lib";
use MemcachedTest;
@@ -525,6 +525,13 @@ for ($kidx = 1; $kidx < $kcnt; $kidx += 1) {
$klen = length($val);
$cmd = "bop mget $klen $kcnt 0..1000 10";
mem_cmd_is($sock, $cmd, $val, $rst);
+$val = "$kstr" . "0";
+for ($kidx = 1; $kidx < $kcnt; $kidx += 1) {
+ $val .= "?$kstr$kidx";
+}
+$klen = length($val);
+$cmd = "bop mget $klen $kcnt 0..1000 10";
+mem_cmd_is($sock, $cmd, $val, "CLIENT_ERROR bad data chunk");
# after test
release_memcached($engine, $server);
|
out_file: use different API to do msgpack to json conversion | @@ -209,22 +209,23 @@ static void cb_file_flush(void *data, size_t bytes,
while (msgpack_unpack_next(&result, data, bytes, &off)) {
alloc_size = (off - last_off) + 128; /* JSON is larger than msgpack */
last_off = off;
- buf = (char *)flb_calloc(1, alloc_size);
- if (buf == NULL) {
- flb_errno();
- msgpack_unpacked_destroy(&result);
- fclose(fp);
- FLB_OUTPUT_RETURN(FLB_RETRY);
- }
flb_time_pop_from_msgpack(&tm, &result, &obj);
+
switch (ctx->format){
case FLB_OUT_FILE_FMT_JSON:
- if (flb_msgpack_obj_to_json(buf, alloc_size, obj) >= 0) {
+ buf = flb_msgpack_to_json_str(alloc_size, &result);
+ if (buf) {
fprintf(fp, "%s: [%f, %s]\n",
tag,
flb_time_to_double(&tm),
buf);
+ flb_free(buf);
+ }
+ else {
+ msgpack_unpacked_destroy(&result);
+ fclose(fp);
+ FLB_OUTPUT_RETURN(FLB_RETRY);
}
break;
case FLB_OUT_FILE_FMT_CSV:
@@ -234,10 +235,8 @@ static void cb_file_flush(void *data, size_t bytes,
ltsv_output(fp, &tm, obj, ctx);
break;
}
- flb_free(buf);
}
msgpack_unpacked_destroy(&result);
-
fclose(fp);
FLB_OUTPUT_RETURN(FLB_OK);
|
input: disable the skipFactor computation. It's uncertain how much it helps, so do proper testing first before re-enabling it | @@ -461,8 +461,16 @@ static inline int input_speedFactor(run_t* run, dynfile_t* dynfile) {
}
static inline int input_skipFactor(run_t* run, dynfile_t* dynfile, int* speed_factor) {
+ /*
+ * TODO: measure impact of the skipFactor on the speed of fuzzing.
+ * It's currently unsure how much it helps, so disable it for now,
+ * and re-enable once proper test has been conducted
+ */
int penalty = 0;
+ *speed_factor = HF_CAP(input_speedFactor(run, dynfile) / 2, -15, 0);
+ return penalty;
+#if 0
{
*speed_factor = HF_CAP(input_speedFactor(run, dynfile) / 2, -15, 0);
penalty += *speed_factor;
@@ -511,6 +519,7 @@ static inline int input_skipFactor(run_t* run, dynfile_t* dynfile, int* speed_fa
}
return penalty;
+#endif
}
bool input_prepareDynamicInput(run_t* run, bool needs_mangle) {
|
put changes in the correct file | },
"commanddir": {
"title": "commanddir",
- "description": "The directory in which the command that launched the scoped app was run.",
+ "description": "The directory AppScope polls for a `scope.<pid>` file, as explained in the Dynamic Configuration section of the Using the CLI page of the AppScope docs.",
"type": "string"
},
"data": {
|
Changelog entry for:
Merge PR from noloader: Make ICANN Update CA and DS Trust Anchor
static data. | - Merge PR #408 from fobser: Prevent a few more yacc clashes.
- Merge PR #275 from Roland van Rijswijk-Deij: Add feature to return the
original instead of a decrementing TTL ('serve-original-ttl')
+ - Merge PR #355 from noloader: Make ICANN Update CA and DS Trust Anchor
+ static data.
22 January 2022: Willem
- Padding of queries and responses with DNS over TLS as specified in
|
fix compile test error | package chain
import (
+ "github.com/aergoio/aergo/contract"
"io/ioutil"
"math"
"os"
@@ -57,31 +58,31 @@ func TestErrorInExecuteTx(t *testing.T) {
tx := &types.Tx{}
- err := executeTx(bs, tx, 0, 0)
+ err := executeTx(bs, tx, 0, 0, contract.ChainService)
assert.EqualError(t, err, types.ErrTxFormatInvalid.Error(), "execute empty tx")
tx.Body = &types.TxBody{}
- err = executeTx(bs, tx, 0, 0)
+ err = executeTx(bs, tx, 0, 0, contract.ChainService)
assert.EqualError(t, err, types.ErrTxFormatInvalid.Error(), "execute empty tx body")
tx.Body.Account = makeTestAddress(t)
tx.Body.Recipient = makeTestAddress(t)
- err = executeTx(bs, tx, 0, 0)
+ err = executeTx(bs, tx, 0, 0, contract.ChainService)
assert.EqualError(t, err, types.ErrTxHasInvalidHash.Error(), "execute tx body with account")
signTestAddress(t, tx)
- err = executeTx(bs, tx, 0, 0)
+ err = executeTx(bs, tx, 0, 0, contract.ChainService)
assert.EqualError(t, err, types.ErrTxNonceTooLow.Error(), "execute tx body with account")
tx.Body.Nonce = 1
tx.Body.Amount = math.MaxUint64
signTestAddress(t, tx)
- err = executeTx(bs, tx, 0, 0)
+ err = executeTx(bs, tx, 0, 0, contract.ChainService)
assert.EqualError(t, err, types.ErrInsufficientBalance.Error(), "execute tx body with nonce")
tx.Body.Amount = types.MaxAER
signTestAddress(t, tx)
- err = executeTx(bs, tx, 0, 0)
+ err = executeTx(bs, tx, 0, 0, contract.ChainService)
assert.EqualError(t, err, types.ErrInsufficientBalance.Error(), "execute tx body with nonce")
}
@@ -96,13 +97,13 @@ func TestBasicExecuteTx(t *testing.T) {
tx.Body.Recipient = makeTestAddress(t)
tx.Body.Nonce = 1
signTestAddress(t, tx)
- err := executeTx(bs, tx, 0, 0)
+ err := executeTx(bs, tx, 0, 0, contract.ChainService)
assert.NoError(t, err, "execute amount 0")
tx.Body.Nonce = 2
tx.Body.Amount = 1000
signTestAddress(t, tx)
- err = executeTx(bs, tx, 0, 0)
+ err = executeTx(bs, tx, 0, 0, contract.ChainService)
assert.NoError(t, err, "execute amount 1000")
tx.Body.Nonce = 3
@@ -111,7 +112,7 @@ func TestBasicExecuteTx(t *testing.T) {
tx.Body.Type = types.TxType_GOVERNANCE
tx.Body.Payload = []byte{'s'}
signTestAddress(t, tx)
- err = executeTx(bs, tx, 0, 0)
+ err = executeTx(bs, tx, 0, 0, contract.ChainService)
assert.NoError(t, err, "execute governance type")
}
|
Update unk_C0 member comment | @@ -43,7 +43,7 @@ typedef struct SceKblParam { // size is 0x100 or 0x200. must 0x200 on 3.60.
* 0x00000001:bsod reboot(or other serious factors)
* 0x00000010:bsod poweroff
* 0x00000400:always?
- * 0x00020000:unknown
+ * 0x00020000:maybe low battery
*/
uint32_t unk_C0;
uint32_t wakeup_factor;
|
[hardware] Change bitwise to logic operators | @@ -96,7 +96,7 @@ module tcdm_adapter
) i_metadata_register (
.clk_i (clk_i ),
.rst_ni (rst_ni ),
- .valid_i(in_valid_i & in_ready_o & !in_write_i),
+ .valid_i(in_valid_i && in_ready_o && !in_write_i),
.ready_o(meta_ready ),
.data_i (in_meta_i ),
.valid_o(meta_valid ),
@@ -131,12 +131,12 @@ module tcdm_adapter
// Ready to output data if both meta and read data
// are available (the read data will always be last)
- assign in_valid_o = meta_valid & rdata_valid;
+ assign in_valid_o = meta_valid && rdata_valid;
// Only pop the data from the registers once both registers are ready
- assign pop_resp = in_ready_i & in_valid_o;
+ assign pop_resp = in_ready_i && in_valid_o;
// Generate out_gnt one cycle after sending read request to the bank
- `FFARN(out_gnt, (out_req_o && !out_write_o) | sc_successful, 1'b0, clk_i, rst_ni);
+ `FFARN(out_gnt, (out_req_o && !out_write_o) || sc_successful, 1'b0, clk_i, rst_ni);
// ----------------
// LR/SC
@@ -195,7 +195,7 @@ module tcdm_adapter
// it makes a new reservation in program order or issues any SC.
if (amo_op_t'(in_amo_i) == AMOLR &&
(!reservation_q.valid || reservation_q.core == unique_core_id)) begin
- reservation_d.valid = 1'b1 & LrScEnable;
+ reservation_d.valid = 1'b1 && LrScEnable;
reservation_d.addr = in_address_i;
reservation_d.core = unique_core_id;
end
@@ -216,7 +216,7 @@ module tcdm_adapter
if (reservation_q.valid && amo_op_t'(in_amo_i) == AMOSC
&& reservation_q.core == unique_core_id) begin
reservation_d.valid = 1'b0;
- sc_successful = (reservation_q.addr == in_address_i) & LrScEnable;
+ sc_successful = (reservation_q.addr == in_address_i) && LrScEnable;
end
end
end // always_comb
@@ -232,10 +232,10 @@ module tcdm_adapter
always_comb begin
// feed-through
- in_ready_o = in_valid_o & ~in_ready_i ? 1'b0 : 1'b1;
- out_req_o = in_valid_i & in_ready_o;
+ in_ready_o = in_valid_o && !in_ready_i ? 1'b0 : 1'b1;
+ out_req_o = in_valid_i && in_ready_o;
out_add_o = in_address_i;
- out_write_o = in_write_i | (sc_successful & (amo_op_t'(in_amo_i) == AMOSC));
+ out_write_o = in_write_i || (sc_successful && (amo_op_t'(in_amo_i) == AMOSC));
out_wdata_o = in_wdata_i;
out_be_o = in_be_i;
|
Express linop_reshape by operator_reshape | @@ -364,55 +364,18 @@ extern struct linop_s* linop_extract_create(unsigned int N, const long pos[N], c
return linop_create(N, out_dims, N, in_dims, CAST_UP(PTR_PASS(data)), extract_forward, extract_adjoint, NULL, NULL, extract_free);
}
-
-
-
-
-struct reshape_op_s {
-
- INTERFACE(linop_data_t);
-
- unsigned int N;
- const long* dims;
-};
-
-static DEF_TYPEID(reshape_op_s);
-
-static void reshape_forward(const linop_data_t* _data, complex float* dst, const complex float* src)
-{
- const struct reshape_op_s* data = CAST_DOWN(reshape_op_s, _data);
-
- md_copy(data->N, data->dims, dst, src, CFL_SIZE);
-}
-
-static void reshape_free(const linop_data_t* _data)
-{
- const struct reshape_op_s* data = CAST_DOWN(reshape_op_s, _data);
-
- xfree(data->dims);
-
- xfree(data);
-}
-
-
struct linop_s* linop_reshape_create(unsigned int A, const long out_dims[A], int B, const long in_dims[B])
{
- PTR_ALLOC(struct reshape_op_s, data);
- SET_TYPEID(reshape_op_s, data);
-
- assert(md_calc_size(A, out_dims) == md_calc_size(B, in_dims));
+ PTR_ALLOC(struct linop_s, c);
- unsigned int N = A;
- data->N = N;
- long* dims = *TYPE_ALLOC(long[N]);
- md_copy_dims(N, dims, out_dims);
- data->dims = dims;
+ c->forward = operator_reshape_create(A, out_dims, B, in_dims);
+ c->adjoint = operator_reshape_create(B, in_dims, A, out_dims);
+ c->normal = operator_reshape_create(B, in_dims, B, in_dims);
+ c->norm_inv = NULL;
- return linop_create(A, out_dims, B, in_dims, CAST_UP(PTR_PASS(data)), reshape_forward, reshape_forward, reshape_forward, NULL, reshape_free);
+ return PTR_PASS(c);
}
-
-
struct transpose_op_s {
INTERFACE(linop_data_t);
@@ -1246,5 +1209,3 @@ struct linop_s* linop_conv_create(int N, unsigned int flags, enum conv_type ctyp
return linop_create(N, odims, N, idims, CAST_UP(PTR_PASS(data)), linop_conv_forward, linop_conv_adjoint, NULL, NULL, linop_conv_free);
}
-
-
|
Fix potential multi-threading consistency issues | @@ -187,7 +187,6 @@ namespace carto {
return _kineticEventHandler;
}
- // TODO: add explicit 'ViewState' argument to all calculateCameraEvent methods
void MapRenderer::calculateCameraEvent(CameraPanEvent& cameraEvent, float durationSeconds, bool updateKinetic) {
if (durationSeconds > 0) {
if (cameraEvent.isUseDelta()) {
@@ -207,13 +206,15 @@ namespace carto {
{
std::lock_guard<std::recursive_mutex> lock(_mutex);
- oldFocusPos = _options->getProjectionSurface()->calculateMapPos(_viewState.getFocusPos());
+ std::shared_ptr<ProjectionSurface> projectionSurface = getProjectionSurface();
+
+ oldFocusPos = projectionSurface->calculateMapPos(_viewState.getFocusPos());
// Calculate new focusPos, cameraPos and upVec
cameraEvent.calculate(*_options, _viewState);
// Calculate parameters for kinetic events
- newFocusPos = _options->getProjectionSurface()->calculateMapPos(_viewState.getFocusPos());
+ newFocusPos = projectionSurface->calculateMapPos(_viewState.getFocusPos());
zoom = _viewState.getZoom();
// In case of seamless panning horizontal teleport, offset the delta focus pos
@@ -742,7 +743,7 @@ namespace carto {
}
cglib::vec3<double> origin = viewState.getCameraPos();
- cglib::vec3<double> target = _options->getProjectionSurface()->calculatePosition(targetPos);
+ cglib::vec3<double> target = viewState.getProjectionSurface()->calculatePosition(targetPos);
cglib::ray3<double> ray(origin, target - origin);
// Normal layer click detection is done in the layer order
|
Universal meson build. | project('urbit', 'c', meson_version: '>=0.29.0')
+legacy_meson = false
+
+if(meson.version() == '0.29.0')
+ legacy_meson = true
+elif(meson.version() != '0.45.0')
+ error('Meson 0.29.0 is last legacy version supported. Otherwise please upgrade to 0.45.0.')
+endif
+
jets_a_src = [
'jets/a/add.c',
'jets/a/dec.c',
@@ -221,7 +229,12 @@ os_link_flags = []
if osdet == 'linux'
conf_data.set('U3_OS_linux', true)
+ if(legacy_meson)
pthread_dep = find_library('pthread')
+ else
+ pthread_dep = meson.get_compiler('c').find_library('pthread')
+ endif
+
ncurses_dep = dependency('ncurses')
os_deps = os_deps + [pthread_dep, ncurses_dep]
@@ -230,7 +243,12 @@ elif osdet == 'darwin'
os_c_flags = os_c_flags + ['-bind_at_load']
# os_link_flags = ['-framework CoreServices', '-framework CoreFoundation']
+ if(legacy_meson)
ncurses_dep = find_library('ncurses')
+ else
+ ncurses_dep = meson.get_compiler('c').find_library('ncurses')
+ endif
+
os_deps = os_deps + [ncurses_dep]
elif osdet == 'bsd'
@@ -261,9 +279,13 @@ openssl_dep = dependency('openssl', version: '>=1.0.0')
curl_dep = dependency('libcurl', version: '>=7.35.0')
libuv_dep = dependency('libuv', version: '>=1.8.0')
+if(legacy_meson)
gmp_dep = find_library('gmp')
sigsegv_dep = find_library('sigsegv')
-
+else
+ gmp_dep = meson.get_compiler('c').find_library('gmp')
+ sigsegv_dep = meson.get_compiler('c').find_library('sigsegv')
+endif
# For these libs we provide fallback bundle
cmark_dep = dependency('libcmark', version: '0.12.0', fallback: ['commonmark-legacy', 'cmark_dep'])
|
gpio: Bugfix - Move esp_intr_free() out of the critical section in gpio_uninstall_isr_service()
Closes
Fix the bug that if the API was called from one core to free the interrupt source on the other core, it would trigger interrupt watchdog. | @@ -491,15 +491,21 @@ esp_err_t gpio_isr_handler_remove(gpio_num_t gpio_num)
void gpio_uninstall_isr_service(void)
{
+ gpio_isr_func_t *gpio_isr_func_free = NULL;
+ gpio_isr_handle_t gpio_isr_handle_free = NULL;
+ portENTER_CRITICAL(&gpio_context.gpio_spinlock);
if (gpio_context.gpio_isr_func == NULL) {
+ portEXIT_CRITICAL(&gpio_context.gpio_spinlock);
return;
}
- portENTER_CRITICAL(&gpio_context.gpio_spinlock);
- esp_intr_free(gpio_context.gpio_isr_handle);
- free(gpio_context.gpio_isr_func);
+ gpio_isr_func_free = gpio_context.gpio_isr_func;
gpio_context.gpio_isr_func = NULL;
+ gpio_isr_handle_free = gpio_context.gpio_isr_handle;
+ gpio_context.gpio_isr_handle = NULL;
gpio_context.isr_core_id = GPIO_ISR_CORE_ID_UNINIT;
portEXIT_CRITICAL(&gpio_context.gpio_spinlock);
+ esp_intr_free(gpio_isr_handle_free);
+ free(gpio_isr_func_free);
return;
}
@@ -532,7 +538,12 @@ esp_err_t gpio_isr_register(void (*fn)(void *), void *arg, int intr_alloc_flags,
#else /* CONFIG_FREERTOS_UNICORE */
ret = esp_ipc_call_blocking(gpio_context.isr_core_id, gpio_isr_register_on_core_static, (void *)&p);
#endif /* !CONFIG_FREERTOS_UNICORE */
- if(ret != ESP_OK || p.ret != ESP_OK) {
+ if (ret != ESP_OK) {
+ ESP_LOGE(GPIO_TAG, "esp_ipc_call_blocking failed (0x%x)", ret);
+ return ESP_ERR_NOT_FOUND;
+ }
+ if (p.ret != ESP_OK) {
+ ESP_LOGE(GPIO_TAG, "esp_intr_alloc failed (0x%x)", p.ret);
return ESP_ERR_NOT_FOUND;
}
return ESP_OK;
|
change to use list_from_json | @@ -11,7 +11,7 @@ int main (int argc, char ** argv) {
char * json = orka_load_whole_file(argv[1], &len);
discord::guild::member::dati ** p;
- discord::guild::member::from_json_list(json, len, &p);
+ discord::guild::member::list_from_json(json, len, &p);
return 0;
}
|
Explicitely set the locale for time displayed in reports ; fixes | {
// Convert the date to a string
NSDateFormatter *dateFormatter = [[NSDateFormatter alloc] init];
+
+ // Explicitely set the locale to avoid an iOS 8 bug
+ // http://stackoverflow.com/questions/29374181/nsdateformatter-hh-returning-am-pm-on-ios-8-device
+ [dateFormatter setLocale:[[NSLocale alloc] initWithLocaleIdentifier:@"en_US_POSIX"]];
+
[dateFormatter setDateFormat:@"yyyy-MM-dd'T'HH:mm:ss'Z'"];
[dateFormatter setTimeZone:[NSTimeZone timeZoneWithAbbreviation:@"UTC"]];
NSString *currentTimeStr = [dateFormatter stringFromDate: self.dateTime];
|
Subsets and Splits
No saved queries yet
Save your SQL queries to embed, download, and access them later. Queries will appear here once saved.