message
stringlengths
6
474
diff
stringlengths
8
5.22k
Fix Makefile to correctly enable running uao with storage schedule.
@@ -89,7 +89,7 @@ storage_queryfinish_and_transactionmanagement: $(TESTER) $(DISCOVER) -t tincrepo/mpp/gpdb/tests \ -s dispatch/queryfinish \ -s storage/transaction_management \ - -s storage/basic + -s storage/basic \ -s storage/uao storage_persistent_filerep_accessmethods_and_vacuum:
Line: Adapt Markdown Shell Recorder test We do not assume that a non-root user executes the test any more. This update fixes part of issue
@@ -76,14 +76,10 @@ cat `kdb file /tests/line` #> here kdb ls /tests/line -#> user/tests/line -#> user/tests/line/#0 -#> user/tests/line/#1 -#> user/tests/line/#2 +# STDOUT-REGEX: line.+line/#0.+line/#1.+line/#2 kdb set /tests/line/#1 huhu -#> Using name user/tests/line/#1 -#> Set string to "huhu" +# STDOUT-REGEX: .+Set string to "huhu" kdb export /tests/line line #> something
pagecache_write_sg: apply error to merge in case of alloc failure; completion added to page in READING state should be from runqueue, not bhqueue
@@ -643,7 +643,9 @@ closure_function(1, 3, void, pagecache_write_sg, pp = allocate_page_nodelocked(pn, pi); if (pp == INVALID_ADDRESS) { pagecache_unlock_node(pn); - apply(completion, timm("result", "failed to allocate pagecache_page")); + const char *err = "failed to allocate pagecache_page"; + apply(sh, timm("result", err)); /* close out merge, record write error */ + apply(completion, timm("result", err)); return; } @@ -666,7 +668,7 @@ closure_function(1, 3, void, pagecache_write_sg, realloc_pagelocked(pc, pp); refcount_reserve(&pp->refcount); if (page_state(pp) == PAGECACHE_PAGESTATE_READING) - enqueue_page_completion_statelocked(pc, pp, apply_merge(m), true /* complete on bhqueue */); + enqueue_page_completion_statelocked(pc, pp, apply_merge(m), false); pagecache_unlock_state(pc); } pagecache_unlock_node(pn);
Expose mach_absolute_time and mach_timebase_info. Fixes trollius on Darwin. Note: mandatory check (NEED_CHECK) was skipped
@@ -93,29 +93,13 @@ namespace { } BEGIN_SYMS("c") -#if defined(_linux_) -SYM(prctl) -SYM(sched_getaffinity) -SYM(sched_setaffinity) -SYM(inotify_init) -SYM(inotify_add_watch) -SYM(inotify_rm_watch) -#endif -#if defined(_darwin_) -SYM(sysctlbyname) -#endif -SYM(ftok) -#if __has_include(<aio.h>) -SYM(aio_error) -SYM(aio_read) -SYM(aio_return) -SYM(aio_suspend) -#endif + SYM(calloc) SYM(clock_gettime) SYM_2("clock_getres", ClockGetres) SYM(closedir) SYM(freeifaddrs) +SYM(ftok) SYM(getifaddrs) SYM(getnameinfo) SYM(getpwnam) @@ -129,5 +113,28 @@ SYM(strdup) SYM(sendfile) SYM(strtod) SYM_2("__errno_location", ErrnoLocation) + +#if defined(_linux_) +SYM(prctl) +SYM(sched_getaffinity) +SYM(sched_setaffinity) +SYM(inotify_init) +SYM(inotify_add_watch) +SYM(inotify_rm_watch) +#endif + +#if defined(_darwin_) +SYM(mach_absolute_time) +SYM(mach_timebase_info) +SYM(sysctlbyname) +#endif + +#if __has_include(<aio.h>) +SYM(aio_error) +SYM(aio_read) +SYM(aio_return) +SYM(aio_suspend) +#endif + END_SYMS() #endif
Docker tag/push fix.
@@ -38,5 +38,5 @@ build:dockerimage: - docker pull $CI_BUILD_IMAGE:latest || true - docker build --cache-from $CI_BUILD_IMAGE:latest -t $CI_BUILD_IMAGE:$CI_COMMIT_SHORT_SHA . - docker push $CI_BUILD_IMAGE:$CI_COMMIT_SHORT_SHA - - docker tag $CI_BUILD_IMAGE:$CI_COMMIT_SHORT_SHA $CI_BUILD_IMAGE:latest && git push $CI_BUILD_IMAGE:latest + - docker tag $CI_BUILD_IMAGE:$CI_COMMIT_SHORT_SHA $CI_BUILD_IMAGE:latest && docker push $CI_BUILD_IMAGE:latest - if [ -n "$CI_COMMIT_TAG" ]; then docker tag $CI_BUILD_IMAGE:$CI_COMMIT_SHORT_SHA $CI_BUILD_IMAGE:$CI_COMMIT_TAG && docker push $CI_BUILD_IMAGE:$CI_COMMIT_TAG; fi
fskframesync: adding noise back into example
@@ -47,8 +47,8 @@ int main(int argc, char*argv[]) srand( time(NULL) ); // options - float SNRdB = 40.0f; // signal-to-noise ratio - float noise_floor = -30.0f; // noise floor + float SNRdB = 20.0f; // signal-to-noise ratio + float noise_floor = -20.0f; // noise floor float dphi = 0.01f; // carrier frequency offset float theta = 0.0f; // carrier phase offset float dt = -0.2f; // fractional sample timing offset @@ -117,7 +117,7 @@ int main(int argc, char*argv[]) // add noise, channel gain for (i=0; i<buf_len; i++) - buf_rx[i] = buf_tx[i]*gamma + 0.0f*nstd*(randnf() + randnf()*_Complex_I)*M_SQRT1_2; + buf_rx[i] = buf_tx[i]*gamma + nstd*(randnf() + randnf()*_Complex_I)*M_SQRT1_2; // synchronize/receive the frame fskframesync_execute_block(fs, buf_rx, buf_len);
Documentation: Use American English spelling
@@ -170,7 +170,7 @@ verified or added to build Elektra correctly: added to exclude the `debian` branch from being build. The reason for this is that the `debian` branch is not providing a Jenkinsfile. -* `Advanced clone behaviours` should be added and the path to the git mirror +* `Advanced clone behaviors` should be added and the path to the git mirror needs to be specified: `/home/jenkins/git_mirrors/libelektra`. This reference repository is created and maintained by our [daily buildjob](https://build.libelektra.org/jenkins/job/libelektra-daily/).
add development guides
@@ -153,3 +153,52 @@ authenticate users, e.g. by [username/password auth](https://www.digitalocean.co - `src/containers/` - contains components that are connected to Redux - `src/css/` - contains CSS styles - `src/reducers/` - contains Redux reducers (used to process actions) + + +## Development Guides + +### Updating dependencies + +Lockfiles (`package-lock.json`) can be updated by simply deleting the current +lock file and running `npm install`, which creates a new lock file. + +Check for outdated dependencies via `npm outdated`. Dependencies can then be +updated by running `npm update`. + +### Adding support for new metadata + +- Create a new sub dialog by, for example, copying the `NumberSubDialog.jsx` + file (or similar) to a new file in the + `client/src/components/TreeItem/dialogs` folder. + +- Include the sub dialog by adding it to the `SettingsDialog.jsx` file in the + same folder. For example, it could be added before the + AdditionalMetakeysSubDialog at the end of the file: + +```diff ++ <NewSubDialog ++ onChange={this.handleEdit('check/something')} ++ value={this.getMeta('check/something', '')} ++ saved={this.getSaved('check/something')} ++ /> + <AdditionalMetakeysSubDialog + handleEdit={this.handleEdit.bind(this)} + getMeta={this.getMeta.bind(this)} + getSaved={this.getSaved.bind(this)} + meta={this.props.meta} + deleteMeta={this.props.deleteMeta} + /> + </FocusTrapDialog> +``` + +- Mark the meta keys as handled by adding them to the `HANDLED_METADATA` array + in `client/src/components/TreeItem/dialogs/utils.js`: + +```diff +export const HANDLED_METADATA = [ + ..., + 'visibility', + 'binary', ++ 'check/something', +] +```
Benchmark: Use warmup phase for YAML benchmark
@@ -26,7 +26,7 @@ for PLUGIN in "${PLUGINS[@]}"; do cp "$DATA_DIRECTORY/test.yaml" "$DATA_DIRECTORY/test.$PLUGIN.in" done -hyperfine \ +hyperfine --warmup 3 \ "\"$BENCHMARK_TOOL\" \"$DATA_DIRECTORY\" user ${PLUGINS[0]} get" \ "\"$BENCHMARK_TOOL\" \"$DATA_DIRECTORY\" user ${PLUGINS[1]} get" \ "\"$BENCHMARK_TOOL\" \"$DATA_DIRECTORY\" user ${PLUGINS[2]} get" \
[mod_magnet] set script env in func first upvalue set script env in first upvalue (_ENV upvalue) for func once, upon script init, since script-env table is preserved and reused (and script-env is cleared between script invocations)
@@ -2510,6 +2510,9 @@ static handler_t magnet_attract(request_st * const r, plugin_data * const p, scr lua_createtable(L, 0, 1); /* (sp += 1) */ magnet_mainenv_metatable(L); /* (sp += 1) */ lua_setmetatable(L, -2); /* (sp -= 1) */ + /* set script env in first upvalue (_ENV upvalue) for func */ + lua_pushvalue(L, env_ndx); /* (sp += 1) */ + magnet_setfenv_mainfn(L, func_ndx); /* (sp -= 1) */ /* insert lighty table at index 4 (lighty_table_ndx = 4) */ magnet_init_lighty_table(L); /* lighty.* (sp += 1) */ } @@ -2544,17 +2547,11 @@ static handler_t magnet_attract(request_st * const r, plugin_data * const p, scr */ lua_pushvalue(L, lighty_table_ndx); /* (sp += 1) */ lua_setfield(L, env_ndx, "lighty"); /* lighty.* (sp -= 1) */ - lua_pushvalue(L, env_ndx); /* (sp += 1) */ - magnet_setfenv_mainfn(L, 1); /* (sp -= 1) */ /* pcall will destroy the func value, duplicate it */ lua_pushvalue(L, func_ndx); /* (sp += 1) */ int ret = lua_pcall(L, 0, 1, errfunc_ndx); /* (sp -= 1; sp += 1) */ - /* reset environment */ - lua_pushglobaltable(L); /* (sp += 1) */ - magnet_setfenv_mainfn(L, 1); /* (sp -= 1) */ - handler_t result = HANDLER_GO_ON; if (0 != ret) { log_error(r->conf.errh, __FILE__, __LINE__,
[net][sal] Delete the judgment of the default network interface device when socket is created
@@ -387,14 +387,7 @@ static int socket_init(int family, int type, int protocol, struct sal_socket **r sock->type = type; sock->protocol = protocol; - /* get socket operations from network interface device */ - if (netdv_def == RT_NULL) - { - LOG_E("not find default network interface device for socket create."); - return -3; - } - - if (netdev_is_up(netdv_def) && netdev_is_link_up(netdv_def)) + if (netdv_def && netdev_is_up(netdv_def) && netdev_is_link_up(netdv_def)) { /* check default network interface device protocol family */ pf = (struct sal_proto_family *) netdv_def->sal_user_data;
Fix few ereport calls to properly set errcode in url_curl.c
@@ -1214,7 +1214,7 @@ url_curl_fopen(char *url, bool forwrite, extvar_t *ev, CopyState pstate) if (!is_file_exists(extssl_cer_full)) ereport(ERROR, - (errcode(errcode_for_file_access()), + (errcode_for_file_access(), errmsg("could not open certificate file \"%s\": %m", extssl_cer_full))); @@ -1235,7 +1235,7 @@ url_curl_fopen(char *url, bool forwrite, extvar_t *ev, CopyState pstate) if (!is_file_exists(extssl_key_full)) ereport(ERROR, - (errcode(errcode_for_file_access()), + (errcode_for_file_access(), errmsg("could not open private key file \"%s\": %m", extssl_key_full))); @@ -1250,7 +1250,7 @@ url_curl_fopen(char *url, bool forwrite, extvar_t *ev, CopyState pstate) if (!is_file_exists(extssl_cas_full)) ereport(ERROR, - (errcode(errcode_for_file_access()), + (errcode_for_file_access(), errmsg("could not open private key file \"%s\": %m", extssl_cas_full)));
just removed SystemTap from build requirements
@@ -24,7 +24,6 @@ For convenience, you can alternatively run the `make.sh` script. - CMake for generating the build files - Python 3 for the [code generator](https://github.com/toru/h2olog/blob/v2/misc/gen-quic-bpf.py) - [BCC](https://iovisor.github.io/bcc/) (>= 0.11.0) [installed](https://github.com/iovisor/bcc/blob/master/INSTALL.md) on your system -- [SystemTap](https://sourceware.org/systemtap/) for DTrace shim to build H2O with DTrace support For Ubuntu 20.04 or later, you can install dependencies with:
Changed functions to pre/post actions when forking.
@@ -4345,8 +4345,13 @@ static void wsgi_python_child_init(apr_pool_t *p) * do it if Python was initialised in parent process. */ - if (wsgi_python_initialized && !wsgi_python_after_fork) + if (wsgi_python_initialized && !wsgi_python_after_fork) { +#if PY_MAJOR_VERSION > 3 || (PY_MAJOR_VERSION == 3 && PY_MINOR_VERSION >= 7) + PyOS_AfterFork_Child(); +#else PyOS_AfterFork(); +#endif + } /* Finalise any Python objects required by child process. */ @@ -10422,6 +10427,12 @@ static int wsgi_start_process(apr_pool_t *p, WSGIDaemonProcess *daemon) wsgi_exit_daemon_process(0); } + if (wsgi_python_initialized) { +#if PY_MAJOR_VERSION > 3 || (PY_MAJOR_VERSION == 3 && PY_MINOR_VERSION >= 7) + PyOS_AfterFork_Parent(); +#endif + } + apr_pool_note_subprocess(p, &daemon->process, APR_KILL_AFTER_TIMEOUT); apr_proc_other_child_register(&daemon->process, wsgi_manage_process, daemon, NULL, p);
update sniffer project.
@@ -162,7 +162,7 @@ void cb_timer(opentimers_id_t id) { cb_timer // callback ); app_vars.outputOrInput = (app_vars.outputOrInput+1)%2; - openserial_triggerDebugprint(); // FIXME: replace by task + scheduler_push_task(task_openserial_debugPrint,TASKPRIO_OPENSERIAL); } // ================================ task =======================================
build: add missing dnf-plugins-core package on centos-8 Type: fix
@@ -285,7 +285,7 @@ ifeq ($(OS_ID),rhel) @sudo -E yum install $(CONFIRM) $(RPM_DEPENDS) @sudo -E debuginfo-install $(CONFIRM) glibc openssl-libs mbedtls-devel zlib else ifeq ($(OS_ID)-$(OS_VERSION_ID),centos-8) - @sudo -E dnf install $(CONFIRM) epel-release + @sudo -E dnf install $(CONFIRM) dnf-plugins-core epel-release @sudo -E dnf config-manager --set-enabled PowerTools @sudo -E dnf groupinstall $(CONFIRM) $(RPM_DEPENDS_GROUPS) @sudo -E dnf install $(CONFIRM) $(RPM_DEPENDS)
[BSP] correct stm32l412-st-nucleo capacity info
/* Program Entry, set to mark it as "used" and avoid gc */ MEMORY { - ROM (rx) : ORIGIN = 0x08000000, LENGTH = 256k /* 256KB flash */ - RAM1 (rw) : ORIGIN = 0x20000000, LENGTH = 48k /* 48K sram */ - RAM2 (rw) : ORIGIN = 0x10000000, LENGTH = 16k /* 16K sram */ + ROM (rx) : ORIGIN = 0x08000000, LENGTH = 128k /* 128KB flash */ + RAM1 (rw) : ORIGIN = 0x20000000, LENGTH = 32k /* 32K sram */ + RAM2 (rw) : ORIGIN = 0x10000000, LENGTH = 8k /* 8K sram */ } ENTRY(Reset_Handler) _system_stack_size = 0x200; @@ -90,7 +90,7 @@ SECTIONS . = ALIGN(4); /* This is used by the startup in order to initialize the .data secion */ _edata = . ; - } >RAM2 + } >RAM1 .stack : { @@ -99,7 +99,7 @@ SECTIONS . = . + _system_stack_size; . = ALIGN(4); _estack = .; - } >RAM2 + } >RAM1 __bss_start = .; .bss : @@ -117,7 +117,7 @@ SECTIONS _ebss = . ; *(.bss.init) - } > RAM2 + } > RAM1 __bss_end = .; _end = .;
py/objexcept: With PRINT_EXC gone, refactor/optimize code.
@@ -103,18 +103,10 @@ mp_obj_t mp_alloc_emergency_exception_buf(mp_obj_t size_in) { void mp_obj_exception_print(const mp_print_t *print, mp_obj_t o_in, mp_print_kind_t kind) { mp_obj_exception_t *o = MP_OBJ_TO_PTR(o_in); mp_print_kind_t k = kind & ~PRINT_EXC_SUBCLASS; - bool is_subclass = kind & PRINT_EXC_SUBCLASS; - if (!is_subclass && (k == PRINT_REPR || k == PRINT_EXC)) { - mp_print_str(print, qstr_str(o->base.type->name)); - } - - if (k == PRINT_EXC) { - mp_print_str(print, ": "); - } - if (k == PRINT_STR || k == PRINT_EXC) { + if (k == PRINT_STR) { if (o->args == NULL || o->args->len == 0) { - mp_print_str(print, ""); + // In this case, str() is empty string, nothing to print. return; } else if (o->args->len == 1) { #if MICROPY_PY_UERRNO @@ -130,6 +122,16 @@ void mp_obj_exception_print(const mp_print_t *print, mp_obj_t o_in, mp_print_kin mp_obj_print_helper(print, o->args->items[0], PRINT_STR); return; } + // Fallthru + } + + // PRINT_STR with multiple args, PRINT_REPR, etc. + + if (kind == (PRINT_REPR & ~PRINT_EXC_SUBCLASS)) { + // Print type name only for pure PRINT_REPR, not + // PRINT_REPR | PRINT_EXC_SUBCLASS. + // In the latter case, type name is printed in objtype:instance_print(). + mp_print_str(print, qstr_str(o->base.type->name)); } mp_obj_tuple_print(print, MP_OBJ_FROM_PTR(o->args), kind); }
Fix coverity unchecked return value Just extending the comment so coveriety knows as well.
@@ -252,6 +252,7 @@ static int ssl_set_cert(CERT *c, X509 *x) /* * The return code from EVP_PKEY_copy_parameters is deliberately * ignored. Some EVP_PKEY types cannot do this. + * coverity[check_return] */ EVP_PKEY_copy_parameters(pkey, c->pkeys[i].privatekey); ERR_clear_error();
Tests: temporary dir removed after tests execution.
@@ -671,4 +671,10 @@ def pytest_sessionfinish(session): option.restart = True unit_stop() + + public_dir(option.cache_dir) shutil.rmtree(option.cache_dir) + + if not option.save_log: + public_dir(option.temp_dir) + shutil.rmtree(option.temp_dir)
When freezing an asset, if a mesh output is connected to auxiliary nodes or downstream history rather than a mesh shape, insert an upstream mesh so that there is a node to freeze the mesh data on.
@@ -22,9 +22,25 @@ houdiniEngine_freezeAsset(string $assetNode) int $attrIndex = 0; for($con in $outputObjCon) { string $src = `connectionInfo -sfd $con`; + string $discon = $con; + + // if a mesh output has auxiliary nodes or downstream history + // need a temporary mesh shape to backstop the history + if(endsWith($src, "outputPartMeshData")) { + if(nodeType(plugNode($con)) != "mesh") { + string $proxyMesh = `createNode mesh -n "houdiniAssetProxyMesh#" -parent $assetNode`; + connectAttr $src ($proxyMesh + ".inMesh"); + + // force a compute of the mesh before disconnecting so that it has data + polyEvaluate $proxyMesh; + $discon = $proxyMesh + ".inMesh"; + connectAttr -f ($proxyMesh + ".outMesh") $con; + setAttr ($proxyMesh + ".intermediateObject") 1; + } + } setAttr ( $assetNode + ".cachedSrc[" + $attrIndex +"]") -type "string" $src; setAttr ( $assetNode + ".cachedDst[" + $attrIndex +"]") -type "string" $con; - disconnectAttr $src $con; + disconnectAttr $src $discon; $attrIndex++; } @@ -98,7 +114,19 @@ houdiniEngine_unfreezeAsset(string $assetNode) for($ai = 0; $ai < $numCon; $ai++) { string $src = `getAttr ( $assetNode + ".cachedSrc[" + $ai +"]")`; string $dst = `getAttr ( $assetNode + ".cachedDst[" + $ai +"]")`; - connectAttr $src $dst; + + // if an output that needs reconnecting has an input mesh, it's either + // the mesh we added to backstop downstream history when we did the freeze + string $dstNode = plugNode($dst); + string $inputProxies[] = `listConnections -type "mesh" -d off -shapes on $dstNode`; + + connectAttr -f $src $dst; + + // delete them only after they're been forced to disconnect, otherwise + // maya's delete might take some of the downstream history with it + for($proxy in $inputProxies) { + delete $proxy; + } } setAttr ( $assetNode + ".nodeState") 0;
webp-container-spec.txt,cosmetics: normalize formatting prefer Exif to EXIF when not referring to the chunk normalize use of code format for canvas / frame width & height
@@ -36,7 +36,7 @@ for: * **Lossless compression.** An image can be losslessly compressed, using the WebP Lossless Format. - * **Metadata.** An image may have metadata stored in EXIF or XMP formats. + * **Metadata.** An image may have metadata stored in Exif or XMP formats. * **Transparency.** An image may have transparency, i.e., an alpha channel. @@ -94,7 +94,7 @@ _1-based_ RIFF File Format ---------------- -The WebP file format is based on the RIFF (resource interchange file format) +The WebP file format is based on the RIFF (Resource Interchange File Format) document format. The basic element of a RIFF file is a _chunk_. It consists of: @@ -261,7 +261,7 @@ An extended format file consists of: * Image data. - * An optional 'EXIF' chunk with EXIF metadata. + * An optional 'EXIF' chunk with Exif metadata. * An optional 'XMP ' chunk with XMP metadata. @@ -317,9 +317,9 @@ Alpha (L): 1 bit : Set if any of the frames of the image contain transparency information ("alpha"). -EXIF metadata (E): 1 bit +Exif metadata (E): 1 bit -: Set if the file contains EXIF metadata. +: Set if the file contains Exif metadata. XMP metadata (X): 1 bit @@ -341,12 +341,12 @@ Reserved: 24 bits Canvas Width Minus One: 24 bits : _1-based_ width of the canvas in pixels. - The actual canvas width is '1 + Canvas Width Minus One' + The actual canvas width is `1 + Canvas Width Minus One`. Canvas Height Minus One: 24 bits : _1-based_ height of the canvas in pixels. - The actual canvas height is '1 + Canvas Height Minus One' + The actual canvas height is `1 + Canvas Height Minus One`. The product of _Canvas Width_ and _Canvas Height_ MUST be at most `2^32 - 1`. @@ -423,21 +423,21 @@ If the _Animation flag_ is not set, then this chunk SHOULD NOT be present. Frame X: 24 bits (_uint24_) -: The X coordinate of the upper left corner of the frame is `Frame X * 2` +: The X coordinate of the upper left corner of the frame is `Frame X * 2`. Frame Y: 24 bits (_uint24_) -: The Y coordinate of the upper left corner of the frame is `Frame Y * 2` +: The Y coordinate of the upper left corner of the frame is `Frame Y * 2`. Frame Width Minus One: 24 bits (_uint24_) : The _1-based_ width of the frame. - The frame width is `1 + Frame Width Minus One` + The frame width is `1 + Frame Width Minus One`. Frame Height Minus One: 24 bits (_uint24_) : The _1-based_ height of the frame. - The frame height is `1 + Frame Height Minus One` + The frame height is `1 + Frame Height Minus One`. Frame Duration: 24 bits (_uint24_) @@ -677,12 +677,12 @@ EXIF chunk: +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+ | ChunkHeader('EXIF') | +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+ - | EXIF Metadata | + | Exif Metadata | +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+ -EXIF Metadata: _Chunk Size_ bytes +Exif Metadata: _Chunk Size_ bytes -: image metadata in EXIF format. +: image metadata in Exif format. XMP chunk: @@ -798,7 +798,7 @@ RIFF/WEBP +- XMP (metadata) ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ -An animated image with EXIF metadata may look as follows: +An animated image with Exif metadata may look as follows: ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ RIFF/WEBP
Add some links to README
@@ -26,7 +26,6 @@ data. Its main features and design principles are: Jansson is licensed under the `MIT license`_; see LICENSE in the source distribution for details. - Compilation and Installation ---------------------------- @@ -74,6 +73,14 @@ Then, point your browser to ``doc/_build/html/index.html``. Sphinx_ 1.0 or newer is required to generate the documentation. +Community +--------- + +* `Documentation <http://jansson.readthedocs.io/en/latest/>`_ +* `Issue tracker <https://github.com/akheron/jansson/issues>`_ +* `Mailing list <http://groups.google.com/group/jansson-users>`_ +* `Wiki <https://github.com/akheron/jansson/wiki>`_ contains some development documentation + .. _Jansson: http://www.digip.org/jansson/ .. _`Comprehensive documentation`: http://jansson.readthedocs.io/en/latest/ .. _`MIT license`: http://www.opensource.org/licenses/mit-license.php
correct the sac and dac value.
@@ -116,8 +116,6 @@ owerror_t forwarding_send(OpenQueueEntry_t* msg) { dac = IPHC_DAC_STATELESS; } else { myprefix = idmanager_getMyID(ADDR_PREFIX); - sac = IPHC_SAC_STATEFUL; - dac = IPHC_DAC_STATEFUL; } memcpy(&(msg->l3_sourceAdd.addr_128b[0]),myprefix->prefix,8); memcpy(&(msg->l3_sourceAdd.addr_128b[8]),myadd64->addr_64b,8); @@ -151,6 +149,8 @@ owerror_t forwarding_send(OpenQueueEntry_t* msg) { //XV -poipoi we want to check if the source address prefix is the same as destination prefix if (packetfunctions_sameAddress(&temp_dest_prefix,&temp_src_prefix)) { // same prefix use 64B address + sac = IPHC_SAC_STATEFUL; + dac = IPHC_DAC_STATEFUL; sam = IPHC_SAM_64B; dam = IPHC_DAM_64B; p_dest = &temp_dest_mac64b; @@ -159,6 +159,8 @@ owerror_t forwarding_send(OpenQueueEntry_t* msg) { //not the same prefix. so the packet travels to another network //check if this is a source routing pkt. in case it is then the DAM is elided as it is in the SrcRouting header. if (packetfunctions_isBroadcastMulticast(&(msg->l3_destinationAdd))==FALSE){ + sac = IPHC_SAC_STATELESS; + dac = IPHC_DAC_STATELESS; sam = IPHC_SAM_128B; dam = IPHC_DAM_128B; p_dest = &(msg->l3_destinationAdd);
Fixed the return code of RAND_query_egd_bytes when connect fails.
@@ -133,6 +133,7 @@ int RAND_query_egd_bytes(const char *path, unsigned char *buf, int bytes) break; # endif default: + ret = -1; goto err; /* failure */ } } @@ -230,9 +231,7 @@ int RAND_egd_bytes(const char *path, int bytes) int num, ret = 0; num = RAND_query_egd_bytes(path, NULL, bytes); - if (num < 1) - goto err; - if (RAND_status() == 1) + if (num < 1 || RAND_status() == 1) ret = num; err: return (ret);
pipeline: extract pre-run test setup to it's own separate task. Ideally, pretest setup should be explicit for their own jobs, so we extracted out the setup to be more modularized. Any tests should be able to use this pattern instead of having to add it in run_tinc.yml or run_behave.yml
@@ -2600,6 +2600,25 @@ jobs: <<: *ccp_gen_cluster_default_params on_failure: <<: *ccp_destroy + - task: pre_run_test_setup + tags: ["ccp"] + image: centos-gpdb-dev-6 + config: + platform: linux + inputs: + - name: ccp_src + - name: cluster_env_files + run: + path: bash + args: + - -c + - | + set -ex + ccp_src/aws/setup_ssh_to_cluster.sh + ssh -t mdw "HOST1=mdw HOST2=sdw1 HOST3=sdw2 HOST4=sdw3 HOST5=sdw4 \ + bash -c 'ssh-keygen -f ~/.ssh/id_rsa -y > ~/.ssh/id_rsa.pub; for i in mdw sdw{1..4}; do ssh centos@\$i \"sudo chmod 777 /usr/local\"; done'" + on_failure: + <<: *debug_sleep - task: run_tests tags: ["ccp"] file: gpdb_src/concourse/tasks/run_tinc.yml @@ -2607,7 +2626,6 @@ jobs: params: TINC_TARGET: gpexpand_1 CUSTOM_ENV: HOST1=mdw HOST2=sdw1 HOST3=sdw2 HOST4=sdw3 HOST5=sdw4 - PRE_TEST_SCRIPT: bash -c 'ssh-keygen -f ~/.ssh/id_rsa -y > ~/.ssh/id_rsa.pub; for i in mdw sdw{1..4}; do ssh centos@$i "sudo chmod 777 /usr/local"; done' on_failure: <<: *debug_sleep - *ccp_destroy @@ -2643,14 +2661,32 @@ jobs: <<: *ccp_gen_cluster_default_params on_failure: <<: *ccp_destroy + - task: pre_run_test_setup + tags: ["ccp"] + image: centos-gpdb-dev-6 + config: + platform: linux + inputs: + - name: ccp_src + - name: cluster_env_files + run: + path: bash + args: + - -c + - | + set -ex + ccp_src/aws/setup_ssh_to_cluster.sh + ssh -t mdw "HOST1=mdw HOST2=sdw1 HOST3=sdw2 HOST4=sdw3 HOST5=sdw4 \ + bash -c 'ssh-keygen -f ~/.ssh/id_rsa -y > ~/.ssh/id_rsa.pub; for i in mdw sdw{1..4}; do ssh centos@\$i \"sudo chmod 777 /usr/local\"; done'" + on_failure: + <<: *debug_sleep - task: run_tests tags: ["ccp"] file: gpdb_src/concourse/tasks/run_tinc.yml image: centos-gpdb-dev-6 params: - TINC_TARGET: gpexpand_2 CUSTOM_ENV: HOST1=mdw HOST2=sdw1 HOST3=sdw2 HOST4=sdw3 HOST5=sdw4 - PRE_TEST_SCRIPT: bash -c 'ssh-keygen -f ~/.ssh/id_rsa -y > ~/.ssh/id_rsa.pub; for i in mdw sdw{1..4}; do ssh centos@$i "sudo chmod 777 /usr/local"; done' + TINC_TARGET: gpexpand_2 on_failure: <<: *debug_sleep - *ccp_destroy
Makefile.include: add rules to generate preprocessed versions of C files, or assembly code, for any file
@@ -287,12 +287,35 @@ $(OBJECTDIR)/%.o: %.S | $(OBJECTDIR) $(Q)$(AS) $(ASFLAGS) -o $@ $< endif +ifndef CUSTOM_RULE_C_TO_OBJECTDIR_S +$(OBJECTDIR)/%.s: %.c | $(OBJECTDIR) + $(TRACE_CC) + $(Q)$(CC) $(CFLAGS) -S $< -o $@ +endif + +ifndef CUSTOM_RULE_C_TO_OBJECTDIR_E +$(OBJECTDIR)/%.e: %.c | $(OBJECTDIR) + $(TRACE_CC) + $(Q)$(CC) $(CFLAGS) -E $< -o $@ +endif + ifndef CUSTOM_RULE_C_TO_O %.o: %.c $(TRACE_CC) $(Q)$(CC) $(CFLAGS) -c $< -o $@ endif +ifndef CUSTOM_RULE_C_TO_S +%.s: %.c + $(TRACE_CC) + $(Q)$(CC) $(CFLAGS) -S $< -o $@ +endif + +ifndef CUSTOM_RULE_C_TO_E +%.e: %.c + $(TRACE_CC) + $(Q)$(CC) $(CFLAGS) -E $< -o $@ +endif ifndef CUSTOM_RULE_C_TO_CO %.co: %.c
Test some basic lua functions
@@ -34,7 +34,7 @@ main = defaultMain $ testGroup "hslua" tests tests :: [TestTree] tests = [ testGroup "Base tests" HsLuaSpec.tests - , testGroup "Monadic functions" [Foreign.Lua.FunctionsTest.tests] + , Foreign.Lua.FunctionsTest.tests , testGroup "Sendings and receiving values from the stack" [ Foreign.Lua.TypesTest.tests , Foreign.Lua.Types.FromLuaStackTest.tests
Update IoT Esp8266 sample readme with latest versions verified
@@ -32,7 +32,7 @@ This is a "to-the-point" guide outlining how to run an Azure SDK for Embedded C - Configuration instructions for the Arduino IDE to compile a sample using the Azure SDK for Embedded C. - Configuration, build, and run instructions for the IoT Hub telemetry sample. -_The following was run on Windows 10 and Ubuntu Desktop 20.04 environments, with Arduino IDE 1.8.12 and Esp8266 module 2.7.4._ +_The following was run on Windows 10 and Ubuntu Desktop 20.04 environments, with Arduino IDE 1.8.15 and Esp8266 module 3.0.1._ ## Prerequisites
fix yndexing for cpp source files compiled with extra flags $YNDEXER_OUTPUT must be the last in the list of arguments for C/C++ command line since yndexer.py expects the last argument defines yndexer output file
@@ -1230,8 +1230,8 @@ class GnuCompiler(Compiler): append('EXTRA_OUTPUT') style = ['${hide;kv:"p CC"} ${hide;kv:"pc green"}'] - cxx_args = ['$GCCFILTER', '$YNDEXER_ARGS', '$CXX_COMPILER', '$C_FLAGS_PLATFORM', '$GCC_COMPILE_FLAGS', '$CXXFLAGS', '$EXTRA_OUTPUT', '$TOOLCHAIN_ENV', '$YNDEXER_OUTPUT', '$SRCFLAGS'] + style - c_args = ['$GCCFILTER', '$YNDEXER_ARGS', '$C_COMPILER', '$C_FLAGS_PLATFORM', '$GCC_COMPILE_FLAGS', '$CFLAGS', '$CONLYFLAGS', '$EXTRA_OUTPUT', '$TOOLCHAIN_ENV', '$YNDEXER_OUTPUT', '$SRCFLAGS'] + style + cxx_args = ['$GCCFILTER', '$YNDEXER_ARGS', '$CXX_COMPILER', '$C_FLAGS_PLATFORM', '$GCC_COMPILE_FLAGS', '$CXXFLAGS', '$EXTRA_OUTPUT', '$SRCFLAGS', '$TOOLCHAIN_ENV', '$YNDEXER_OUTPUT'] + style + c_args = ['$GCCFILTER', '$YNDEXER_ARGS', '$C_COMPILER', '$C_FLAGS_PLATFORM', '$GCC_COMPILE_FLAGS', '$CFLAGS', '$CONLYFLAGS', '$EXTRA_OUTPUT', '$SRCFLAGS', '$TOOLCHAIN_ENV', '$YNDEXER_OUTPUT'] + style print 'macro _SRC_cpp(SRC, SRCFLAGS...) {\n .CMD=%s\n}' % ' '.join(cxx_args) print 'macro _SRC_c(SRC, SRCFLAGS...) {\n .CMD=%s\n}' % ' '.join(c_args)
BugID:17865306: Remove warnings about MBINS
@@ -19,9 +19,9 @@ config AOS_BUILD_APP help This option holds the name of the application that is going to build. -config MBINS - string - option env="MBINS" +#config MBINS +# string +# option env="MBINS" ## --- Generated Automatically --- source "board/Config.in"
mINI: Remove `preview` and `unfinished` tags
- infos/provides = storage/ini - infos/recommends = - infos/placements = getstorage setstorage -- infos/status = experimental limited maintained nodep preview unfinished +- infos/status = experimental limited maintained nodep - infos/metadata = - infos/description = A minimal plugin for simple INI files
fix kappa metric
@@ -29,12 +29,12 @@ static TVector<TVector<int>> GetWeights(EKappaMetricType type, int classCount) { } } -static TVector<TVector<double>> GetExpectedMatrix(TConstArrayRef<TVector<int>> matrix, int classCount) { +static TVector<TVector<double>> GetExpectedMatrix(TConstArrayRef<TVector<double>> matrix, int classCount) { TVector<TVector<double>> expected(classCount, TVector<double>(classCount)); - TVector<int> rows(classCount, 0); - TVector<int> columns(classCount, 0); - int all = 0; + TVector<double> rows(classCount, 0); + TVector<double> columns(classCount, 0); + double all = 0; for (int i = 0; i < classCount; ++i) { for (int j = 0; j < classCount; ++j) { @@ -46,17 +46,17 @@ static TVector<TVector<double>> GetExpectedMatrix(TConstArrayRef<TVector<int>> m for (int i = 0; i < classCount; ++i) { for (int j = 0; j < classCount; ++j) { - expected[i][j] = rows[i] * columns[j] / static_cast<double>(all); + expected[i][j] = rows[i] * columns[j] / all; } } return expected; } -static TVector<TVector<int>> UnzipConfusionMatrix(TMetricHolder metric, int classCount) { - TVector<TVector<int>> matrix(classCount, TVector<int>(classCount)); +static TVector<TVector<double>> UnzipConfusionMatrix(TMetricHolder metric, int classCount) { + TVector<TVector<double>> matrix(classCount, TVector<double>(classCount)); for (int i = 0; i < classCount; ++i) { for (int j = 0; j < classCount; ++j) { - matrix[i][j] = static_cast<int>(metric.Stats[i * classCount + j]); + matrix[i][j] = metric.Stats[i * classCount + j]; } } return matrix; @@ -64,7 +64,7 @@ static TVector<TVector<int>> UnzipConfusionMatrix(TMetricHolder metric, int clas double CalcKappa(TMetricHolder confusionMatrix, int classCount, EKappaMetricType type) { - TVector<TVector<int>> matrix = UnzipConfusionMatrix(confusionMatrix, classCount); + TVector<TVector<double>> matrix = UnzipConfusionMatrix(confusionMatrix, classCount); TVector<TVector<int>> weights = GetWeights(type, classCount); TVector<TVector<double>> expected = GetExpectedMatrix(matrix, classCount);
RHEL8: README.Linux.md update on master Updating RHEL8 installation steps on README.Linux.md, as few additional RHEL8 specific steps needs to be updated.
echo 'source scl_source enable devtoolset-7' >> ~/.bashrc ``` -## For RHEL +## For RHEL: -Use dependency script for CentOS. +- Install Development Tools. + - For RHEL 8: Install `Development Tools`: -- If you want to install `devtoolset-7`: + ```bash + sudo yum group install -y "Development Tools" + ``` + + - For RHEL versions (< 8.0): Install `devtoolset-7`: ```bash sudo yum-config-manager --enable rhui-REGION-rhel-server-rhscl sudo yum install -y devtoolset-7-toolchain ``` +- Install dependencies using README.CentOS.bash script. + - For RHEL 8: Execute additional step before running README.CentOS.bash script. + + Note: Make sure installation of `Development Tools` includes `git` and `make` else install these tools manually. + + ```bash + sudo yum install -y https://dl.fedoraproject.org/pub/epel/epel-release-latest-8.noarch.rpm + ``` + + - Install dependencies using README.CentOS.bash script. + + ```bash + ./README.CentOS.bash + ``` + ## For Ubuntu: - Install Dependencies
fix various code format issues
@@ -395,17 +395,12 @@ static int ssl_tls13_write_selected_identity_ext( mbedtls_ssl_context *ssl, MBEDTLS_SSL_DEBUG_MSG( 3, ( "server hello, adding pre_shared_key extension" ) ); MBEDTLS_SSL_CHK_BUF_PTR( p, end, 6 ); - /* Extension Type */ MBEDTLS_PUT_UINT16_BE( MBEDTLS_TLS_EXT_PRE_SHARED_KEY, p, 0 ); - - /* Extension Length */ MBEDTLS_PUT_UINT16_BE( 2, p, 2 ); /* NOTE: This will need to be adjusted once we support multiple PSKs * being offered by the client. */ selected_identity = 0; - - /* Write selected_identity */ MBEDTLS_PUT_UINT16_BE( selected_identity, p, 4 ); *olen = 6; @@ -1702,7 +1697,8 @@ static int ssl_tls13_write_server_hello_body( mbedtls_ssl_context *ssl, } #if defined(MBEDTLS_KEY_EXCHANGE_SOME_PSK_ENABLED) - MBEDTLS_SSL_DEBUG_MSG( 2,( " mbedtls_ssl_tls13_some_psk_enabled %d", mbedtls_ssl_tls13_some_psk_enabled( ssl ) ) ); + MBEDTLS_SSL_DEBUG_MSG( 2, ( " mbedtls_ssl_tls13_some_psk_enabled %d", + mbedtls_ssl_tls13_some_psk_enabled( ssl ) ) ); if( mbedtls_ssl_tls13_some_psk_enabled( ssl ) ) { ret = ssl_tls13_write_selected_identity_ext( ssl, p, end, &output_len );
metadata-pull-hook: disabling remove on nack
++ on-pull-nack |= [=resource =tang] ^- (quip card _this) - =/ =associations:metadata - (metadata-for-group:met resource) - :_ this - %+ turn ~(tap by associations) - |= [=md-resource:metadata =association:metadata] - %+ poke-our:pass:io:hc %metadata-store - :- %metadata-update-2 - !> ^- update:metadata - [%remove resource md-resource] + `this :: ++ on-pull-kick |= =resource
include/debug.h: Modify debug defines of task manager Add missing task manager debug defines at debug.h and modify task manager debug defines.
@@ -733,20 +733,19 @@ Once LOGM is approved, each module should have its own index #define medllvdbg(...) #endif -#ifdef CONFIG_DEBUG_TASK_MANAGER #ifdef CONFIG_DEBUG_TASK_MANAGER_ERROR #define tmdbg(format, ...) dbg(format, ##__VA_ARGS__) +#define tmlldbg(format, ...) lldbg(format, ##__VA_ARGS__) #else #define tmdbg(...) +#define tmlldbg(...) #endif #ifdef CONFIG_DEBUG_TASK_MANAGER_INFO #define tmvdbg(format, ...) vdbg(format, ##__VA_ARGS__) +#define tmllvdbg(format, ...) llvdbg(format, ##__VA_ARGS__) #else #define tmvdbg(...) -#endif -#else -#define tmdbg(...) -#define tmvdbg(...) +#define tmllvdbg(...) #endif #else /* CONFIG_CPP_HAVE_VARARGS */ @@ -1150,6 +1149,22 @@ Once LOGM is approved, each module should have its own index #define medllvdbg (...) #endif +#ifdef CONFIG_DEBUG_TASK_MANAGER_ERROR +#define tmdbg dbg +#define tmlldbg lldbg +#else +#define tmdbg (void) +#define tmlldbg (void) +#endif + +#ifdef CONFIG_DEBUG_TASK_MANAGER_INFO +#define tmvdbg vdbg +#define tmllvdbg llvdbg +#else +#define tmvdbg (void) +#define tmllvdbg (void) +#endif + #endif /* CONFIG_CPP_HAVE_VARARGS */ /* Buffer dumping macros do not depend on varargs */
fix error in Display readme
@@ -163,7 +163,7 @@ picodisplay.rectangle(x, y, w, h) Draws a circle filled with the current pen colour to the buffer. The `x` and `y` parameters specify the centre of the circle, `r` specifies the radius in pixels. ```python -picodisplay.rectangle(x, y, w, h) +picodisplay.circle(x, y, r) ``` ![Circle function explanation image](/micropython/modules/pico_display/images/circle.png)
Remove deadlock that was caused by calling pthread_rwlock_wrlock() on same thread twice. This can be reproduce only with musl. CLA: trivial
@@ -108,7 +108,7 @@ static int dummy_dup(CRYPTO_EX_DATA *to, const CRYPTO_EX_DATA *from, int crypto_free_ex_index_ex(OPENSSL_CTX *ctx, int class_index, int idx) { - EX_CALLBACKS *ip = get_and_lock(ctx, class_index); + EX_CALLBACKS *ip; EX_CALLBACK *a; int toret = 0; OSSL_EX_DATA_GLOBAL *global = openssl_ctx_get_ex_data_global(ctx);
Create-operator timings to wshfl
@@ -719,14 +719,40 @@ int main_wshfl(int argc, char* argv[]) coeff_dims[4] = md; coeff_dims[6] = tk; - debug_printf(DP_INFO, "Creating linear operators... "); + debug_printf(DP_INFO, "Creating linear operators: \n"); + + double t1; + double t2; + + t1 = timestamp(); const struct linop_s* E = linop_espirit_create(sx, sy, sz, nc, md, tk, maps); + t2 = timestamp(); + debug_printf(DP_INFO, "\tE: %f seconds.\n", t2 - t1); + + t1 = timestamp(); const struct linop_s* R = linop_reshape_create(wx, sx, sy, sz, nc, tk); + t2 = timestamp(); + debug_printf(DP_INFO, "\tR: %f seconds.\n", t2 - t1); + + t1 = timestamp(); const struct linop_s* Fx = linop_fx_create(wx, sy, sz, nc, tk, false); + t2 = timestamp(); + debug_printf(DP_INFO, "\tFx: %f seconds.\n", t2 - t1); + + t1 = timestamp(); const struct linop_s* W = linop_wave_create(wx, sy, sz, nc, tk, wave); + t2 = timestamp(); + debug_printf(DP_INFO, "\tW: %f seconds.\n", t2 - t1); + + t1 = timestamp(); const struct linop_s* Fyz = linop_fyz_create(wx, sy, sz, nc, tk, false); + t2 = timestamp(); + debug_printf(DP_INFO, "\tFyz: %f seconds.\n", t2 - t1); + + t1 = timestamp(); const struct linop_s* K = linop_kern_create(gpun >= 0, reorder_dims, reorder, phi_dims, phi, kernel_dims, kernel, table_dims); - debug_printf(DP_INFO, "Done.\n"); + t2 = timestamp(); + debug_printf(DP_INFO, "\tK: %f seconds.\n", t2 - t1); if (fwd != NULL) {
Configure: fixed posix_spawn() detection with glic 2.30. In particular, it was previously broken on Ubuntu 19.10 and Fedora 31. See for details: https://sourceware.org/git/?p=glibc.git;a=commitdiff;h=2ab5741
@@ -186,8 +186,8 @@ nxt_feature_libs= nxt_feature_test="#include <spawn.h> #include <unistd.h> - int main() { - (void) posix_spawn(NULL, NULL, NULL, NULL, NULL, NULL); + int main(int argc, char *argv[]) { + (void) posix_spawn(NULL, \"\", NULL, NULL, argv, NULL); return 0; }" . auto/feature
vere: bypass terminal logging on lmdb worker thread
@@ -96,7 +96,7 @@ c3_o _perform_put_on_database_raw(MDB_txn* transaction_u, c3_w ret_w = mdb_put(transaction_u, database_u, &key_val, &value_val, flags); if (ret_w != 0) { - u3l_log("lmdb: write failed: %s\n", mdb_strerror(ret_w)); + fprintf(stderr, "lmdb: write failed: %s\n", mdb_strerror(ret_w)); return c3n; } @@ -118,7 +118,7 @@ c3_o _perform_get_on_database_raw(MDB_txn* transaction_u, c3_w ret_w = mdb_get(transaction_u, database_u, &key_val, value); if (ret_w != 0) { - u3l_log("lmdb: read failed: %s\n", mdb_strerror(ret_w)); + fprintf(stderr, "lmdb: read failed: %s\n", mdb_strerror(ret_w)); return c3n; } @@ -277,7 +277,7 @@ static void _u3_lmdb_write_event_cb(uv_work_t* req) { 0, /* flags */ &transaction_u); if (0 != ret_w) { - u3l_log("lmdb: txn_begin fail: %s\n", mdb_strerror(ret_w)); + fprintf(stderr, "lmdb: txn_begin fail: %s\n", mdb_strerror(ret_w)); return; } @@ -289,7 +289,7 @@ static void _u3_lmdb_write_event_cb(uv_work_t* req) { flags_w, &database_u); if (0 != ret_w) { - u3l_log("lmdb: dbi_open fail: %s\n", mdb_strerror(ret_w)); + fprintf(stderr, "lmdb: dbi_open fail: %s\n", mdb_strerror(ret_w)); return; } @@ -307,7 +307,7 @@ static void _u3_lmdb_write_event_cb(uv_work_t* req) { request->malloced_event_data_size[i]); if (success == c3n) { - u3l_log("lmdb: failed to write event %" PRIu64 "\n", event_number); + fprintf(stderr, "lmdb: failed to write event %" PRIu64 "\n", event_number); mdb_txn_abort(transaction_u); data->success = c3n; return; @@ -317,12 +317,12 @@ static void _u3_lmdb_write_event_cb(uv_work_t* req) { ret_w = mdb_txn_commit(transaction_u); if (0 != ret_w) { if ( request->event_count == 1 ) { - u3l_log("lmdb: failed to commit event %" PRIu64 ": %s\n", + fprintf(stderr, "lmdb: failed to commit event %" PRIu64 ": %s\n", request->first_event, mdb_strerror(ret_w)); } else { c3_d through = request->first_event + request->event_count - 1ULL; - u3l_log("lmdb: failed to commit events %" PRIu64 " through %" PRIu64 + fprintf(stderr, "lmdb: failed to commit events %" PRIu64 " through %" PRIu64 ": %s\n", request->first_event, through,
added a check for valid time step to assure the file is and H5part file
@@ -145,6 +145,17 @@ avtH5PartFileFormat::avtH5PartFileFormat(const char *filename, EXCEPTION1(InvalidFilesException, filename); } + // Check for the first time step - if it does not exit bailout as + // the file is not a H5Part file. + activeTimeStep = 0; + if (!H5PartHasStep(file, activeTimeStep)) + { + H5PartCloseFile(file); + EXCEPTION1(InvalidFilesException, "Cannot find first time step 0."); + } + + // Assume a valid H5Part file from here out. + // Get number of time steps in file numTimestepsInFile = H5PartGetNumSteps(file);
doc: add info about GSOC 2018
This page is the first point for anyone interested in contributing to Elektra! +## GSoC 2018 + +This page is also the central point for students interested in contributing to Elektra during +[Google Summer of Code 2018](https://summerofcode.withgoogle.com). + +Google Summer of Code is a opportunity for you to get paid while working on free software +initiatives during summer: + +- [Timeline is available here.](https://developers.google.com/open-source/gsoc/timeline) +- [Flyers can be found here.](https://developers.google.com/open-source/gsoc/resources/media) + +> Disclaimer: We do not yet know if we will participate! + +### Which GSoC projects are available? + +We think that GSoC is a perfect opportunity to bridge Elektra and other free +software projects. So the ideal GSoC project is to elektrify some free software. + +Elektra may be used for applications (frameworks) like KDE, GNOME, +and LXCE or for any applications that need to read configuration files, +access environment variables, or command-line parameters. + +Some possible tasks are: + +- Integrate applications (or frameworks) into Elektra. +- Arbitrary nesting of plugins. +- Expore further ways of how configuration can be validated. +- The Qt-GUI could be improved by integrating validation. + +## What are the requirements to participate? + +- At least one PR submitted to Elektra (see below for details). +- Fluent in English or German +- Good writing skills in English +- Profound programming skills with focus on good code quality, documentation and unit tests. +- Ideally experience with an application (framework) or even author of a FLOSS application + + + + + ## How can I get started? We prepared
Update the link for the nightly test status. Update the link for the nightly test status to the web site on github.io.
Source code repository for the VisIt Scientific Visualization and Data Analysis Application -[Project Website](https://visit.llnl.gov) | [Nightly Test Status](http://portal.nersc.gov/project/visit/) +[Project Website](https://visit.llnl.gov) | [Nightly Test Status](https://visit-dav.github.io/dashboard/) ## Documentation
Add `---` <hr> parsing
++ flow marl:dynamic :: node or generator ++ mite :: context $? $down :: outer embed + $rule :: horizontal ruler $list :: unordered list $lime :: list item $lord :: ordered list ++ trig-style :: $? $fini :: terminator $done :: end of input + $rule :: --- horizontal ruler $lint :: + line item $lite :: - line item $head :: # heading =/ nex/@ud ?- p.cur $down 2 + $rule 0 $head 0 $expr 2 $list 0 (flop q.cur) =- [[- ~] (flop q.cur)]~ ?- p.cur + $rule %hr $list %ul $lord %ol $lime %li :: either a one-line header or a paragraph %. [p.u.lub yex] ?- p.cur + $rule (full hrul):parse $expr expr:parse $head head:parse @ para:parse :: if error, propagate correctly ?~ q.vex ~&(%err-prop ..$(err `p.vex)) :: - :: finish tag if it's a header - =< ?:(=(%head p.cur) fold ..$) + :: finish tag if it's a header or rule + =< ?:(?=(?($head $rule) p.cur) fold ..$) :: :: save good result, clear buffer ..$(lub ~, q.cur (weld p.u.q.vex q.cur)) :: detect bad block structure ?. ?- p.cur :: - :: only one line in a heading - $head | + :: only one line in a heading/ruler + ?($rule $head) | :: :: literals need to end with a blank line ?($code $poem $expr) ++ apse ^+ . :: by prefix style ?- sty.pic $fini !! :: terminator + $rule (push %rule) :: horizontal ruler $head (push %head) :: heading $bloc (entr %bloc) :: blockquote line $expr (entr %expr) :: hoon expression (full (easy %done)) :: end of input (cold ~ (just `@`10)) :: blank line (cold %fini bas) :: terminator + (cold %rule ;~(plug hep hep hep)) :: --- horizontal ruler (cold %head ;~(plug (star hax) ace)) :: # heading (cold %lite ;~(plug hep ace)) :: - line item (cold %lint ;~(plug lus ace)) :: + line item == -- :: + ++ hrul :: empty besides fence + (cold ~ ;~(plug gay hep hep hep (star hep))) + :: ++ para :: paragraph %+ cook |=(flow [[%p ~] +<]~)
Update CHANGELOG.md with the logger function setup api name change
@@ -30,6 +30,7 @@ and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0 `oeapkman --optee exec gcc -c file.c` cross-compile `file.c` to target OP-TEE. - See [samples/apkman](samples/apkman) for a complete example demonstrating use of the `sqlite` database library within enclaves. - Support for `compiler-rt`. `oelibc` includes LLVM's `compiler-rt-10.0.1`. +- Update logging function setup API name for SGX Quote Provider plugin to `sgx_ql_set_logging_callback` and mark API name `sgx_ql_set_logging_function` as deprecated. ## Changed - Updated libcxx to version 10.0.1
Fix logging with module loading support
@@ -797,10 +797,12 @@ static oe_result_t _link_elf_image( else { if ((symbol->st_info >> 4) != STB_WEAK) - OE_RAISE( + OE_RAISE_MSG( OE_UNSUPPORTED_ENCLAVE_IMAGE, "symbol %s not found\n", name); + else + OE_TRACE_WARNING("Weak symbol %s is not resolved\n"); } } /* Patch non-symbolic relocation records */
Fix i2c return value when it fails to read Assign I/O error when i2c fails to read bytes, and jump to fail laber with -EIO return value.
@@ -765,7 +765,7 @@ int s5j_i2c_transfer(struct i2c_dev_s *dev, struct i2c_msg_s *msgv, int msgc) ret = readbytes(priv, pmsg); if (ret < pmsg->length) { if (ret >= 0) { - return -EIO; + ret = -EIO; } goto fail; }
Tighten up SSL_get1_supported_ciphers() docs This function is really emulating what would happen in client mode, and does not necessarily reflect what is usable for a server SSL. Make this a bit more explicit, and do some wordsmithing while here.
@@ -29,16 +29,16 @@ is returned. SSL_CTX_get_ciphers() returns the stack of available SSL_CIPHERs for B<ctx>. SSL_get1_supported_ciphers() returns the stack of enabled SSL_CIPHERs for -B<ssl>, sorted by preference. +B<ssl> as would be sent in a ClientHello (that is, sorted by preference). The list depends on settings like the cipher list, the supported protocol versions, the security level, and the enabled signature algorithms. SRP and PSK ciphers are only enabled if the appropriate callbacks or settings have been applied. -This is the list that will be sent by the client to the server. -The list supported by the server might include more ciphers in case there is a -hole in the list of supported protocols. -The server will also not use ciphers from this list depending on the -configured certificates and DH parameters. +The list of ciphers that would be sent in a ClientHello can differ from +the list of ciphers that would be acceptable when acting as a server. +For example, additional ciphers may be usable by a server if there is +a gap in the list of supported protocols, and some ciphers may not be +usable by a server if there is not a suitable certificate configured. If B<ssl> is NULL or no ciphers are available, NULL is returned. SSL_get_client_ciphers() returns the stack of available SSL_CIPHERs matching the
sim: Remove complexity from an if Clippy suggests that having a closure in the condition of an if can be confusing in regards to code formatting. Move the conditional outside of the if into a temp variable.
@@ -1564,11 +1564,14 @@ fn install_ptable(flash: &mut SimMultiFlash, areadesc: &AreaDesc) { // aren't marked as the BootLoader partition, avoid adding the // partition table. This makes it harder to view the image, but // avoids messing up images already written. - if areadesc.iter_areas().any(|area| { + let skip_ptable = areadesc + .iter_areas() + .any(|area| { area.device_id == id && area.off == 0 && area.flash_id != FlashId::BootLoader - }) { + }); + if skip_ptable { if log_enabled!(Info) { let special: Vec<FlashId> = areadesc.iter_areas() .filter(|area| area.device_id == id && area.off == 0)
core/test/run-buddy: add one more test to get to 100% lines covered We're now only missing some assert() branches in buddy testing.
-/* Copyright 2016 IBM Corp. +/* Copyright 2016-2017 IBM Corp. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. @@ -38,6 +38,7 @@ int main(void) buddy_reserve(b, 127, 0); buddy_reserve(b, 0, 4); + assert(buddy_reserve(b, 0, 4) == false); a[0] = buddy_alloc(b, 0); assert(a[0] >= 0);
deep sleep: fix sleep gpio wakeup fail issue
@@ -585,6 +585,9 @@ esp_err_t gpio_wakeup_disable(gpio_num_t gpio_num) #endif portENTER_CRITICAL(&gpio_context.gpio_spinlock); gpio_hal_wakeup_disable(gpio_context.gpio_hal, gpio_num); +#if SOC_GPIO_SUPPORT_SLP_SWITCH && CONFIG_ESP32C3_LIGHTSLEEP_GPIO_RESET_WORKAROUND + gpio_hal_sleep_sel_en(gpio_context.gpio_hal, gpio_num); +#endif portEXIT_CRITICAL(&gpio_context.gpio_spinlock); return ret; } @@ -907,6 +910,9 @@ esp_err_t gpio_deep_sleep_wakeup_enable(gpio_num_t gpio_num, gpio_int_type_t int } portENTER_CRITICAL(&gpio_context.gpio_spinlock); gpio_hal_deepsleep_wakeup_enable(gpio_context.gpio_hal, gpio_num, intr_type); +#if SOC_GPIO_SUPPORT_SLP_SWITCH && CONFIG_ESP32C3_LIGHTSLEEP_GPIO_RESET_WORKAROUND + gpio_hal_sleep_sel_dis(gpio_context.gpio_hal, gpio_num); +#endif portEXIT_CRITICAL(&gpio_context.gpio_spinlock); return ESP_OK; } @@ -919,6 +925,9 @@ esp_err_t gpio_deep_sleep_wakeup_disable(gpio_num_t gpio_num) } portENTER_CRITICAL(&gpio_context.gpio_spinlock); gpio_hal_deepsleep_wakeup_disable(gpio_context.gpio_hal, gpio_num); +#if SOC_GPIO_SUPPORT_SLP_SWITCH && CONFIG_ESP32C3_LIGHTSLEEP_GPIO_RESET_WORKAROUND + gpio_hal_sleep_sel_en(gpio_context.gpio_hal, gpio_num); +#endif portEXIT_CRITICAL(&gpio_context.gpio_spinlock); return ESP_OK; }
examples/png: msan poison for input data
@@ -55,6 +55,12 @@ void png_user_free(png_structp png_ptr, png_voidp ptr) void png_user_read_data(png_structp png_ptr, png_bytep data, png_size_t length) { +#if defined(__clang__) +#if __has_feature(memory_sanitizer) + __msan_poison(data, length); +#endif /* __has_feature(memory_sanitizer) */ +#endif /* defined(__clang__) */ + user_file_t* f = (user_file_t*)png_ptr->io_ptr; if (length > f->len) {
add test timeout with ctx lock
******************************************************************************/ #include "selftest.h" +#include "context.h" +#include "stdbool.h" #include "luos_hal.h" /******************************************************************************* @@ -78,11 +80,18 @@ result_t selftest_com(void) uint32_t tickstart = LuosHAL_GetSystick(); while (!rx_flag) { - if ((LuosHAL_GetSystick() - tickstart) > 3000) + if ((LuosHAL_GetSystick() - tickstart) > 2000) { return KO; } } + while ((LuosHAL_GetSystick() - tickstart) < 2000) + ; + + if (ctx.tx.lock == true) + { + return KO; + } return OK; }
switch entries in announcements list entries
@@ -56,8 +56,8 @@ void announces_debug(FILE *fp) { fprintf(fp, "Announcements:\n" ); while (value) { - fprintf(fp, " id: %s\n", str_id( value->id)); fprintf(fp, " query: %s\n", value->query); + fprintf(fp, " id: %s\n", str_id( value->id)); fprintf(fp, " port: %d\n", value->port); if (value->refresh < now) { fprintf(fp, " refresh: now\n");
fix error 'undefined reference to entry' after menuconfig
@@ -3,6 +3,8 @@ menu "Hardware Drivers Config" config SOC_STM32L4R5ZI bool select SOC_SERIES_STM32L4 + select RT_USING_COMPONENTS_INIT + select RT_USING_USER_MAIN default y menu "Onboard Peripheral Drivers"
OcMachoLib: Revert incorrect changes from last commit
@@ -288,13 +288,8 @@ MachoGetClassNameFromMetaClassPointer ( PrefixSize = L_STR_LEN (OSOBJ_PREFIX); SuffixSize = L_STR_LEN (METACLASS_TOKEN); - BodySize = AsciiStrLen (MetaClassName); - if (PrefixSize + SuffixSize > BodySize) { - return FALSE; - } - - ClassNameLength = (BodySize - PrefixSize - SuffixSize); + ClassNameLength = (AsciiStrLen (MetaClassName) - PrefixSize - SuffixSize); if ((ClassNameLength + 1) > ClassNameSize) { return FALSE; }
Add support for ROCm AOMP compilation with additional paths to resolve hip libs.
@@ -7,8 +7,8 @@ ifeq ($(AOMP),) endif CFLAGS = -c -std=c++11 -D__HIP_PLATFORM_HCC__ -I$(AOMP)/../include -I$(AOMP)/include -# Set the rpath to pick up AOMP libs in case rocm was installed. -LFLAGS = -L$(AOMP)/lib -lhip_hcc -Wl,-rpath,$(AOMP)/lib +# Set the rpath to pick up AOMP libs in case rocm was installed. Also support looking in /opt/rocm/lib as AOMP in ROCm no longer builds hip. +LFLAGS = -L$(AOMP)/lib -L$(AOMP)/../lib -lhip_hcc -Wl,-rpath,$(AOMP)/lib -Wl,-rpath,$(AOMP)/../lib # find the current gpu, default to vega10 if unknown MYGPU = $(shell $(AOMP)/bin/mygpu -d gfx900)
board/plankton/usb_pd_pdo.c: Format with clang-format BRANCH=none TEST=none
#include "usb_pd.h" #include "usb_pd_pdo.h" -#define PDO_FIXED_FLAGS (PDO_FIXED_DATA_SWAP | PDO_FIXED_UNCONSTRAINED |\ - PDO_FIXED_COMM_CAP) +#define PDO_FIXED_FLAGS \ + (PDO_FIXED_DATA_SWAP | PDO_FIXED_UNCONSTRAINED | PDO_FIXED_COMM_CAP) /* Source PDOs */ const uint32_t pd_src_pdo[] = {
Cost function for Valhalla wheelchair routing
#include <valhalla/proto/trippath.pb.h> #include <valhalla/proto/tripdirections.pb.h> +namespace valhalla { namespace sif { + + cost_ptr_t CreateWheelchairCost(const boost::property_tree::ptree& config) { + boost::property_tree::ptree wheelchairConfig(config); + wheelchairConfig.put("type", "wheelchair"); + return CreatePedestrianCost(wheelchairConfig); + } + +} } + namespace valhalla { namespace meili { inline float local_tile_size(const valhalla::baldr::GraphReader& graphreader) { @@ -137,7 +147,7 @@ namespace valhalla { namespace meili { cost_factory_.Register("auto", sif::CreateAutoCost); cost_factory_.Register("bicycle", sif::CreateBicycleCost); cost_factory_.Register("pedestrian", sif::CreatePedestrianCost); - cost_factory_.Register("wheelchair", sif::CreatePedestrianCost); + cost_factory_.Register("wheelchair", sif::CreateWheelchairCost); cost_factory_.Register("multimodal", CreateUniversalCost); } @@ -313,7 +323,7 @@ namespace valhalla { namespace thor { factory.Register("bus", sif::CreateBusCost); factory.Register("bicycle", sif::CreateBicycleCost); factory.Register("pedestrian", sif::CreatePedestrianCost); - factory.Register("wheelchair", sif::CreatePedestrianCost); + factory.Register("wheelchair", sif::CreateWheelchairCost); factory.Register("truck", sif::CreateTruckCost); factory.Register("transit", sif::CreateTransitCost);
filter_modify: condition key_value_matches regex bug In configuring Condition 'Key_Value_Matches' KEY REGEX Accidentally KEY is consired regex not REGEX parameter. This can crash program.
@@ -169,7 +169,7 @@ static int setup(struct filter_modify_ctx *ctx, } else if (strcasecmp(sentry->value, "key_value_matches") == 0) { condition->conditiontype = KEY_VALUE_MATCHES; - condition->a_is_regex = true; + condition->b_is_regex = true; } else if (strcasecmp(sentry->value, "key_value_does_not_match") == 0) {
ExtendedTools: Remove duplicate PhGetTaskbarDpi for tray icons
@@ -291,12 +291,6 @@ VOID EtpGpuIconUpdateCallback( HANDLE maxGpuProcessId; PPH_PROCESS_ITEM maxGpuProcessItem; PH_FORMAT format[8]; - LONG dpiValue; - - dpiValue = PhGetTaskbarDpi(); - - drawInfo.Width = PhGetSystemMetrics(SM_CXSMICON, dpiValue); - drawInfo.Height = PhGetSystemMetrics(SM_CYSMICON, dpiValue); // Icon @@ -334,7 +328,7 @@ VOID EtpGpuIconUpdateCallback( else maxGpuProcessItem = NULL; - PhInitFormatS(&format[0], L"GPU Usage: "); + PhInitFormatS(&format[0], L"GPU usage: "); PhInitFormatF(&format[1], (DOUBLE)EtGpuNodeUsage * 100, 2); PhInitFormatC(&format[2], '%'); @@ -411,12 +405,6 @@ VOID EtpDiskIconUpdateCallback( HANDLE maxDiskProcessId; PPH_PROCESS_ITEM maxDiskProcessItem; PH_FORMAT format[6]; - LONG dpiValue; - - dpiValue = PhGetTaskbarDpi(); - - drawInfo.Width = PhGetSystemMetrics(SM_CXSMICON, dpiValue); - drawInfo.Height = PhGetSystemMetrics(SM_CYSMICON, dpiValue); // Icon @@ -543,12 +531,6 @@ VOID EtpNetworkIconUpdateCallback( HANDLE maxNetworkProcessId; PPH_PROCESS_ITEM maxNetworkProcessItem; PH_FORMAT format[6]; - LONG dpiValue; - - dpiValue = PhGetTaskbarDpi(); - - drawInfo.Width = PhGetSystemMetrics(SM_CXSMICON, dpiValue); - drawInfo.Height = PhGetSystemMetrics(SM_CYSMICON, dpiValue); // Icon @@ -699,7 +681,7 @@ VOID EtpGpuTextIconUpdateCallback( else maxGpuProcessItem = NULL; - PhInitFormatS(&format[0], L"GPU Usage: "); + PhInitFormatS(&format[0], L"GPU usage: "); PhInitFormatF(&format[1], (DOUBLE)EtGpuNodeUsage * 100, 2); PhInitFormatC(&format[2], '%'); @@ -747,14 +729,8 @@ VOID EtpDiskTextIconUpdateCallback( PPH_PROCESS_ITEM maxDiskProcessItem; PH_FORMAT format[6]; PPH_STRING text; - LONG dpiValue; static ULONG64 maxValue = 100000 * 1024; // minimum scaling of 100 MB. - dpiValue = PhGetTaskbarDpi(); - - drawInfo.Width = PhGetSystemMetrics(SM_CXSMICON, dpiValue); - drawInfo.Height = PhGetSystemMetrics(SM_CYSMICON, dpiValue); - // Icon Icon->Pointers->BeginBitmap(&drawInfo.Width, &drawInfo.Height, &bitmap, &bits, &hdc, &oldBitmap); @@ -831,14 +807,8 @@ VOID EtpNetworkTextIconUpdateCallback( PPH_PROCESS_ITEM maxNetworkProcessItem; PH_FORMAT format[6]; PPH_STRING text; - LONG dpiValue; static ULONG64 maxValue = 1024 * 1024; // minimum scaling of 1 MB. - dpiValue = PhGetTaskbarDpi(); - - drawInfo.Width = PhGetSystemMetrics(SM_CXSMICON, dpiValue); - drawInfo.Height = PhGetSystemMetrics(SM_CYSMICON, dpiValue); - // Icon Icon->Pointers->BeginBitmap(&drawInfo.Width, &drawInfo.Height, &bitmap, &bits, &hdc, &oldBitmap);
reference: fix test_baseName
@@ -184,14 +184,14 @@ static void test_getArrayNext (void) ksDel (array); } -void test_baseName () +static void test_baseName (void) { printf ("Test validate base name"); succeed_if (elektraArrayValidateBaseNameString ("#") == 0, "Start not detected correctly"); succeed_if (elektraArrayValidateBaseNameString ("#0") == 1, "#0 should be valid"); - succeed_if (elektraArrayValidateBaseNameString ("#_10") == 0, "#_10 should be valid"); - succeed_if (elektraArrayValidateBaseNameString ("#_________1234567890") == 0, "#_________1234567890 should be valid"); + succeed_if (elektraArrayValidateBaseNameString ("#_10") == 1, "#_10 should be valid"); + succeed_if (elektraArrayValidateBaseNameString ("#_________1234567890") == 1, "#_________1234567890 should be valid"); succeed_if (elektraArrayValidateBaseNameString ("#__________12345678901") == -1, "#__________12345678901 should not be valid"); succeed_if (elektraArrayValidateBaseNameString ("monkey") == -1, "monkey should not be valid"); } @@ -209,6 +209,7 @@ int main (int argc, char ** argv) test_startArray (); test_getArray (); test_getArrayNext (); + test_baseName (); printf ("\ntest_array RESULTS: %d test(s) done. %d error(s).\n", nbTest, nbError);
Fix spelling and formatting consistency
This guide details the steps required to migrate from Mbed TLS version 2.x to Mbed TLS version 3.0 or greater. Unlike normal releases, Mbed TLS 3.0 breaks -compatibility with previous versions, so users (and alt implementors) might +compatibility with previous versions, so users (and alt implementers) might need to change their own code in order to make it work with Mbed TLS 3.0. Here's the list of breaking changes; each entry should help you answer these @@ -580,13 +580,13 @@ extension if it contains any unsupported certificate policies. ### Remove `MBEDTLS_X509_CHECK_*_KEY_USAGE` options from `mbedtls_config.h` This change affects users who have chosen the configuration options to disable the -library's verification of the `keyUsage` and `extendedKeyUsage` fields of x509 +library's verification of the `keyUsage` and `extendedKeyUsage` fields of X.509 certificates. The `MBEDTLS_X509_CHECK_KEY_USAGE` and `MBEDTLS_X509_CHECK_EXTENDED_KEY_USAGE` -configuration options are removed and the X509 code now behaves as if they were +configuration options are removed and the X.509 code now behaves as if they were always enabled. It is consequently not possible anymore to disable at compile -time the verification of the `keyUsage` and `extendedKeyUsage` fields of X509 +time the verification of the `keyUsage` and `extendedKeyUsage` fields of X.509 certificates. The verification of the `keyUsage` and `extendedKeyUsage` fields is important,
Fix effort calculation
@@ -175,11 +175,7 @@ namespace MiningCore.Blockchain.Bitcoin public Task CalculateBlockEffortAsync(Block block, ulong accumulatedBlockShareDiff) { - // adjust diff - var coinProps = BitcoinProperties.GetCoinProperties(poolConfig.Coin.Type, poolConfig.Coin.Algorithm); - var adjusted = accumulatedBlockShareDiff / coinProps.ShareMultiplier; - - block.Effort = adjusted / block.NetworkDifficulty; + block.Effort = (double)accumulatedBlockShareDiff / block.NetworkDifficulty; return Task.FromResult(true); }
[threads smoke test]-Added clang version logic. Clang 9 now considers this to be SPMD instead of Generic (Clang 8)
#include <stdio.h> #include <omp.h> #include <string.h> +#include <stdlib.h> +/*Starting with clang version 9 this test is considered to be SPMD mode. Previously, clang 8 considered this to be generic, +so now this test determines the clang version to verify the results.*/ int main() { + FILE * file; + char ln[256]; + + //Default Clang version set to 8 + int clangVersion = 8; + + //Determine which Clang version + system("$AOMP/bin/clang --version > clang.txt"); + file = fopen("clang.txt", "r"); + if (!file) + printf("Clang version file not found!\n"); + else{ + char *found = NULL; + while((fgets(ln, 256, file) != NULL)){ + found = strstr(ln, "9."); + if(found){ + clangVersion = 9; + printf("Clang version: %s\n", found); + break; + } + } + } + fclose(file); + system("rm clang.txt"); //Determine which GPU type (NVIDIA or AMD) char* nvidia= "sm"; char* aomp_gpu= getenv("AOMP_GPU"); @@ -13,9 +40,10 @@ int main() if(aomp_gpu && strstr(aomp_gpu, nvidia) != NULL) isAMDGPU = 0; - if(isAMDGPU) + if(isAMDGPU && clangVersion < 9) masterWarpThread = 959; - else + + if(!isAMDGPU) masterWarpThread = 991; int thread_id[1024] ; @@ -34,11 +62,15 @@ int main() maxThrd = thread_id[i]; } printf("Max thread id %d\n", maxThrd); - if (maxThrd == masterWarpThread) { + if (maxThrd == masterWarpThread && clangVersion < 9) { + printf("Passed\n"); + return 0; + } else if (maxThrd == 1023 && clangVersion > 8){ printf("Passed\n"); return 0; } else{ printf("Failed\n"); return 1; + } }
[core] do not remove pid-file in test mode do not remove pid-file in test mode (whether test passes or fails) (thx m4t)
#endif static int oneshot_fd = 0; -static volatile int pid_fd = -1; +static volatile int pid_fd = -2; static server_socket_array graceful_sockets; static volatile sig_atomic_t graceful_restart = 0; static volatile sig_atomic_t graceful_shutdown = 0; @@ -417,7 +417,7 @@ static void server_free(server *srv) { } static void remove_pid_file(server *srv) { - if (pid_fd < -2) return; + if (pid_fd <= -2) return; if (!buffer_string_is_empty(srv->srvconf.pid_file) && 0 <= pid_fd) { if (0 != ftruncate(pid_fd, 0)) { log_error_write(srv, __FILE__, __LINE__, "sbds", @@ -1153,13 +1153,13 @@ static int server_main (server * const srv, int argc, char **argv) { } if (test_config) { + buffer_reset(srv->srvconf.pid_file); if (1 == test_config) { printf("Syntax OK\n"); } else { /*(test_config > 1)*/ test_config = 0; srv->srvconf.preflight_check = 1; srv->srvconf.dont_daemonize = 1; - buffer_reset(srv->srvconf.pid_file); } } @@ -1238,6 +1238,7 @@ static int server_main (server * const srv, int argc, char **argv) { } /* open pid file BEFORE chroot */ + if (-2 == pid_fd) pid_fd = -1; /*(initial startup state)*/ if (-1 == pid_fd && !buffer_string_is_empty(srv->srvconf.pid_file)) { if (-1 == (pid_fd = fdevent_open_cloexec(srv->srvconf.pid_file->ptr, O_WRONLY | O_CREAT | O_EXCL | O_TRUNC, S_IRUSR | S_IWUSR | S_IRGRP | S_IROTH))) { struct stat st;
Fix check of cpuid_asm_src config attribute The contents of that attribute is C file names, not object file names. This went undetected because quite a lot of systems have assembler implementations anyway, so setting OPENSSL_CPUID_OBJ was correct for them.
@@ -1290,7 +1290,7 @@ if ($target{sys_id} ne "") unless ($disabled{asm}) { $target{cpuid_asm_src}=$table{DEFAULTS}->{cpuid_asm_src} if ($config{processor} eq "386"); - push @{$config{defines}}, "OPENSSL_CPUID_OBJ" if ($target{cpuid_asm_src} ne "mem_clr.o"); + push @{$config{defines}}, "OPENSSL_CPUID_OBJ" if ($target{cpuid_asm_src} ne "mem_clr.c"); $target{bn_asm_src} =~ s/\w+-gf2m.c// if (defined($disabled{ec2m}));
drivers: Fix compiler errors riscv64-unknown-elf-ld: staging/libdrivers.a(rwbuffer.o): in function `rwb_flush': rwbuffer.c:(.text.rwb_flush+0x12): undefined reference to `rwb_forcetake'
@@ -71,18 +71,20 @@ static ssize_t rwb_read_(FAR struct rwbuffer_s *rwb, off_t startblock, * Name: rwb_semtake ****************************************************************************/ -#if defined(CONFIG_DRVR_WRITEBUFFER) && CONFIG_DRVR_WRDELAY != 0 +#if defined(CONFIG_DRVR_WRITEBUFFER) static int rwb_semtake(FAR sem_t *sem) { return nxsem_wait_uninterruptible(sem); } +#else +# define rwb_semtake(s) OK #endif /**************************************************************************** * Name: rwb_forcetake ****************************************************************************/ -#if defined(CONFIG_DRVR_WRITEBUFFER) && CONFIG_DRVR_WRDELAY != 0 +#if defined(CONFIG_DRVR_WRITEBUFFER) static int rwb_forcetake(FAR sem_t *sem) { int result; @@ -106,13 +108,19 @@ static int rwb_forcetake(FAR sem_t *sem) return ret; } +#else +# define rwb_forcetake(s) OK #endif /**************************************************************************** * Name: rwb_semgive ****************************************************************************/ +#if defined(CONFIG_DRVR_WRITEBUFFER) # define rwb_semgive(s) nxsem_post(s) +#else +# define rwb_semgive(s) +#endif /**************************************************************************** * Name: rwb_overlap
Minor fix to wifi documentation Another fix
@@ -604,7 +604,7 @@ wifi.sta.config(station_cfg) - [`wifi.sta.clearconfig()`](#wifistaclearconfig) - [`wifi.sta.connect()`](#wifistaconnect) - [`wifi.sta.disconnect()`](#wifistadisconnect) -- [`wifi.sta.apinfo()`](#wifistaapinfo) +- [`wifi.sta.getapinfo()`](#wifistagetapinfo) ## wifi.sta.connect() @@ -768,8 +768,8 @@ print("the index of the currently selected AP is: "..wifi.sta.getapindex()) #### See also - [`wifi.sta.getapindex()`](#wifistagetapindex) -- [`wifi.sta.apinfo()`](#wifistaapinfo) -- [`wifi.sta.apchange()`](#wifistaapchange) +- [`wifi.sta.getapinfo()`](#wifistagetapinfo) +- [`wifi.sta.changeap()`](#wifistachangeap) ## wifi.sta.getapinfo()
Fix final warnigns on OSX builds. The OSX deployment target should also be set for building rt/_myrrt.o
@@ -55,6 +55,7 @@ case $OS in sysinit='setenv("MACOSX_DEPLOYMENT_TARGET", "10.6", 0)' echo 'export SYS=osx' >> config.mk echo export INST_MAN=$prefix/share/man/man >> config.mk + echo export MACOSX_DEPLOYMENT_TARGET=10.6 >> config.mk echo 'const Sys = "OSX"' >> mbld/config.myr echo 'const Linkcmd = ["ld", ' \ '"-pagezero_size", "0x100000000",' \
Bug fix for 2Comp and LIFO tank mixing models (issue
@@ -32,6 +32,7 @@ double mixtank(Project *, int, double, double ,double); // Imported functions extern void addseg(Project *, int, double, double); +extern void reversesegs(Project *, int); // Local functions static double piperate(Project *, int); @@ -539,7 +540,7 @@ void tankmix2(Project *pr, int i, double vin, double win, double vnet) vt = MAX(0.0, (mixzone->v + vnet - vmz)); if (vin > 0.0) { - mixzone->c = ((stagzone->c) * (stagzone->v) + win) / + mixzone->c = ((mixzone->c) * (mixzone->v) + win) / (mixzone->v + vin); } if (vt > 0.0) @@ -683,18 +684,14 @@ void tankmix4(Project *pr, int i, double vin, double win, double vnet) seg = qual->LastSeg[k]; if (vnet > 0.0) { - // ... quality is the same, so just add flow volume to last seg + // ... inflow quality is same as last segment's quality, + // so just add inflow volume to last segment if (fabs(seg->c - cin) < qual->Ctol) seg->v += vnet; - // ... otherwise add a new last seg to tank which points to old last seg - else - { - qual->LastSeg[k] = NULL; - addseg(pr, k, vnet, cin); - qual->LastSeg[k]->prev = seg; - } + // ... otherwise add a new last segment with inflow quality + else addseg(pr, k, vnet, cin); - // ... update reported tank quality + // Update reported tank quality tank->C = qual->LastSeg[k]->c; } @@ -704,28 +701,48 @@ void tankmix4(Project *pr, int i, double vin, double win, double vnet) vsum = 0.0; wsum = 0.0; vnet = -vnet; + + // Reverse segment chain so segments are processed from last to first + reversesegs(pr, k); + + // While there is still volume to remove while (vnet > 0.0) { - seg = qual->LastSeg[k]; + // ... start with reversed first segment + seg = qual->FirstSeg[k]; if (seg == NULL) break; + + // ... find volume to remove from it vseg = seg->v; vseg = MIN(vseg, vnet); - if (seg == qual->FirstSeg[k]) vseg = vnet; + if (seg == qual->LastSeg[k]) vseg = vnet; + + // ... update total volume & mass removed vsum += vseg; wsum += (seg->c) * vseg; + + // ... update remiaing volume to remove vnet -= vseg; - if (vnet >= 0.0 && vseg >= seg->v) // Seg used up + + // ... if no more volume left in current segment + if (vnet >= 0.0 && vseg >= seg->v) { + // ... replace current segment with previous one if (seg->prev) { - qual->LastSeg[k] = seg->prev; + qual->FirstSeg[k] = seg->prev; seg->prev = qual->FreeSeg; qual->FreeSeg = seg; } } - else seg->v -= vseg; // Remaining volume in segment + + // ... otherwise reduce volume of current segment + else seg->v -= vseg; } + // Restore original orientation of segment chain + reversesegs(pr, k); + // Reported tank quality is mixture of flow released and any inflow tank->C = (wsum + win) / (vsum + vin); }
Adjust leap mobile height
@@ -216,7 +216,7 @@ export class Omnibox extends Component { renderResults() { const { props, state } = this; return <Box - maxHeight="400px" + maxHeight={['200px', "400px"]} overflowY="auto" overflowX="hidden" borderBottomLeftRadius='2'
Debug: Mention early environment setup
@@ -16,6 +16,8 @@ The general approach is as follows: #### Preparing Source Code +To get started, the compilation environment should be set up. Such commands can be found in `Docs/Configuratio.pdf`, section `3.3 Contribution`. + By default EDK II optimises produced binaries, so to build a "real" debug binary one should target `NOOPT`. Do be aware that it strongly affects resulting binary size:
posix: Add cfgetispeed()/cfgetospeed()
#include <bits/ensure.h> #include <mlibc/sysdeps.hpp> -speed_t cfgetispeed(const struct termios *) { - __ensure(!"Not implemented"); - __builtin_unreachable(); +speed_t cfgetispeed(const struct termios *tios) { + return tios->ibaud; } -speed_t cfgetospeed(const struct termios *) { - __ensure(!"Not implemented"); - __builtin_unreachable(); +speed_t cfgetospeed(const struct termios *tios) { + return tios->obaud; } int cfsetispeed(struct termios *, speed_t) { __ensure(!"Not implemented");
Compiler friendlier
### Ver. 0.4.4 (next) +**Fix**: (`facil`) Fixed compiling error on older `gcc` v.4.8.4, discovered on Ubuntu trusty/64. + **Fix**: Fix enhanced CPU cycles introduced in the v.0.4.3 update. Now CPU cycles are lower and thread throttling handles empty queues more effectively. ### Ver. 0.4.3
make test: remove excessive prints
@@ -78,8 +78,6 @@ class TestL2Flood(VppTestCase): for i in self.pg_interfaces[1:12]: rx0 = i.get_capture(65, timeout=1) - self.logger.error(self.vapi.cli("sh trace")) - # # input on pg4 (SHG=1) expect copies on pg0->3 (SHG=0) # and pg8->11 (SHG=2)
core/minute-ia/interrupts.c: Format with clang-format BRANCH=none TEST=none
@@ -228,15 +228,10 @@ DEFINE_EXN_HANDLER(20); * watchdog timer expiration. However, this time, hardware does not * push errorcode, and we must account for that by pushing zero. */ -noreturn __keep -void exception_panic_wdt(uint32_t cs) +noreturn __keep void exception_panic_wdt(uint32_t cs) { - exception_panic( - CONFIG_MIA_WDT_VEC, - 0, - (uint32_t)__builtin_return_address(0), - cs, - 0); + exception_panic(CONFIG_MIA_WDT_VEC, 0, + (uint32_t)__builtin_return_address(0), cs, 0); } void set_interrupt_gate(uint8_t num, isr_handler_t func, uint8_t flags) @@ -384,11 +379,9 @@ void handle_lapic_lvt_error(void) /* LAPIC LVT error is not an IRQ and can not use DECLARE_IRQ() to call. */ void _lapic_error_handler(void); -__asm__ ( - ".section .text._lapic_error_handler\n" +__asm__(".section .text._lapic_error_handler\n" "_lapic_error_handler:\n" - "pusha\n" - ASM_LOCK_PREFIX "addl $1, __in_isr\n" + "pusha\n" ASM_LOCK_PREFIX "addl $1, __in_isr\n" "movl %esp, %eax\n" "movl $stack_end, %esp\n" "push %eax\n" @@ -399,11 +392,14 @@ __asm__ ( #endif "call handle_lapic_lvt_error\n" "pop %esp\n" - "movl $0x00, (0xFEE000B0)\n" /* Set EOI for LAPIC */ + "movl $0x00, (0xFEE000B0)\n" /* Set + EOI + for + LAPIC + */ ASM_LOCK_PREFIX "subl $1, __in_isr\n" "popa\n" - "iret\n" - ); + "iret\n"); /* Should only be called in interrupt context */ void unhandled_vector(void) @@ -454,8 +450,7 @@ void init_interrupts(void) /* Setup gates for IRQs declared by drivers using DECLARE_IRQ */ for (p = __irq_data; p < __irq_data_end; p++) - set_interrupt_gate(IRQ_TO_VEC(p->irq), - p->handler, + set_interrupt_gate(IRQ_TO_VEC(p->irq), p->handler, IDT_DESC_FLAGS); /* Software generated IRQ */
ssse3: add NEON implementation of simde_mm_mulhrs_pi16
@@ -808,10 +808,25 @@ simde_mm_mulhrs_pi16 (simde__m64 a, simde__m64 b) { a_ = simde__m64_to_private(a), b_ = simde__m64_to_private(b); + #if defined(SIMDE_ARM_NEON_A32V7_NATIVE) + // Has issues due to saturation + // r_.neon_i16 = vqrdmulh_s16(a, b); + + // Multiply + int32x4_t mul = vmull_s16(a_.neon_i16, b_.neon_i16); + + // Rounding narrowing shift right + // narrow = (int16_t)((mul + 16384) >> 15); + int16x4_t narrow = vrshrn_n_s32(mul, 15); + + // Join together + r_.neon_i16 = narrow; + #else SIMDE_VECTORIZE for (size_t i = 0 ; i < (sizeof(r_.i16) / sizeof(r_.i16[0])) ; i++) { r_.i16[i] = HEDLEY_STATIC_CAST(int16_t, (((HEDLEY_STATIC_CAST(int32_t, a_.i16[i]) * HEDLEY_STATIC_CAST(int32_t, b_.i16[i])) + 0x4000) >> 15)); } + #endif return simde__m64_from_private(r_); #endif
test NEW deviation causing submodule prefix to be used Refs
@@ -3102,6 +3102,24 @@ test_deviation(void **state) assert_int_equal(LY_TYPE_IDENT, leaf->type->basetype); assert_string_equal("ident", ((struct lysc_type_identityref *)leaf->type)->bases[0]->name); + assert_int_equal(LY_SUCCESS, lys_parse_mem(ctx, "module v {namespace urn:v;prefix v;" + "identity ident; identity ident2 { base ident; }" + "}", LYS_IN_YANG, NULL)); + ly_ctx_set_module_imp_clb(ctx, test_imp_clb, "submodule w-sub { belongs-to w { prefix w; }" + "import v { prefix v_pref; }" + "leaf l { type string; default \"v_pref:ident2\"; }" + "}"); + assert_int_equal(LY_SUCCESS, lys_parse_mem(ctx, "module w {namespace urn:w;prefix w;" + "include w-sub;" + "}", LYS_IN_YANG, &mod)); + assert_int_equal(LY_SUCCESS, lys_parse_mem(ctx, "module x {namespace urn:x;prefix x;" + "import w { prefix w_pref; } import v { prefix v_pref; }" + "deviation /w_pref:l { deviate replace { type identityref { base v_pref:ident; } } }" + "}", LYS_IN_YANG, NULL)); + assert_non_null(leaf = (struct lysc_node_leaf*)mod->compiled->data); + assert_string_equal("l", leaf->name); + assert_int_equal(LY_TYPE_IDENT, leaf->type->basetype); + assert_int_equal(LY_ENOTFOUND, lys_parse_mem(ctx, "module aa1 {namespace urn:aa1;prefix aa1;import a {prefix a;}" "deviation /a:top/a:z {deviate not-supported;}}", LYS_IN_YANG, &mod)); logbuf_assert("Deviation(s) target node \"/a:top/a:z\" from module \"aa1\" was not found.");
test(ethereum): Add test_001CreateWallet_0012CreateOnetimeWalletWithLoadExistedWallet
@@ -347,6 +347,24 @@ START_TEST(test_001CreateWallet_0011CreateSevenWallet) } END_TEST +START_TEST(test_001CreateWallet_0012CreateOnetimeWalletWithLoadExistedWallet) +{ + BSINT32 rtnVal; + BoatEthWalletConfig wallet = get_ethereum_wallet_settings(); + extern BoatIotSdkContext g_boat_iot_sdk_context; + wallet.prikeyCtx_config.load_existed_wallet = BOAT_TRUE; + /* 1. execute unit test */ + rtnVal = BoatWalletCreate(BOAT_PROTOCOL_ETHEREUM, NULL, &wallet, sizeof(BoatEthWalletConfig)); + /* 2. verify test result */ + /* 2-1. verify the return value */ + ck_assert_int_eq(rtnVal, 0); + + /* 2-2. verify the global variables that be affected */ + ck_assert(g_boat_iot_sdk_context.wallet_list[0].is_used == true); + BoatIotSdkDeInit(); +} +END_TEST + START_TEST(test_002InitWallet_0001SetEIP155CompSuccess) { BSINT32 rtnVal; @@ -697,6 +715,7 @@ Suite *make_wallet_suite(void) tcase_add_test(tc_wallet_api, test_001CreateWallet_0009CreateWalletWithInternalGeneration); tcase_add_test(tc_wallet_api, test_001CreateWallet_0010CreateSixWallet); tcase_add_test(tc_wallet_api, test_001CreateWallet_0011CreateSevenWallet); + tcase_add_test(tc_wallet_api, test_001CreateWallet_0012CreateOnetimeWalletWithLoadExistedWallet); tcase_add_test(tc_wallet_api, test_002InitWallet_0001SetEIP155CompSuccess); tcase_add_test(tc_wallet_api, test_002InitWallet_0002SetEIP155CompFailureNullParam);
esp32/modsocket: Make socket.accept return a tuple: (client, addr).
#include "py/runtime.h" #include "py/mphal.h" #include "py/stream.h" +#include "lib/netutils/netutils.h" #include "lwip/sockets.h" #include "lwip/netdb.h" @@ -117,14 +118,30 @@ STATIC MP_DEFINE_CONST_FUN_OBJ_2(socket_listen_obj, socket_listen); STATIC mp_obj_t socket_accept(const mp_obj_t arg0) { socket_obj_t *self = MP_OBJ_TO_PTR(arg0); - int x = lwip_accept(self->fd, NULL, NULL); - if (x >= 0) { + + struct sockaddr addr; + socklen_t addr_len = sizeof(addr); + int x = lwip_accept(self->fd, &addr, &addr_len); + if (x < 0) { + exception_from_errno(errno); + } + + // create new socket object socket_obj_t *sock = m_new_obj_with_finaliser(socket_obj_t); sock->base.type = self->base.type; sock->fd = x; - return MP_OBJ_FROM_PTR(sock); - } - return mp_const_none; + sock->domain = self->domain; + sock->type = self->type; + sock->proto = self->proto; + + // make the return value + uint8_t *ip = (uint8_t*)&((struct sockaddr_in*)&addr)->sin_addr; + mp_uint_t port = ((struct sockaddr_in*)&addr)->sin_port; + mp_obj_tuple_t *client = mp_obj_new_tuple(2, NULL); + client->items[0] = sock; + client->items[1] = netutils_format_inet_addr(ip, port, NETUTILS_BIG); + + return client; } STATIC MP_DEFINE_CONST_FUN_OBJ_1(socket_accept_obj, socket_accept);
Added dummy support for ARB_fog_linear
@@ -3588,6 +3588,8 @@ void parseToken(sCurStatus* curStatusPtr, int vertex, char **error_msg) { case TOK_END_OF_INST: if (!strcmp(curStatusPtr->curValue.newOpt.optName, "ARB_precision_hint_fastest")) { // Nothing to do + } else if (!strcmp(curStatusPtr->curValue.newOpt.optName, "ARB_fog_linear")) { + // To do? } else { FAIL("Unknown option"); }
Renamed slightly misleading parameter names
@@ -73,9 +73,9 @@ typedef enum { S2N_SERVER, S2N_CLIENT } s2n_mode; extern struct s2n_connection *s2n_connection_new(s2n_mode mode); extern int s2n_connection_set_config(struct s2n_connection *conn, struct s2n_config *config); -extern int s2n_connection_set_fd(struct s2n_connection *conn, int readfd); +extern int s2n_connection_set_fd(struct s2n_connection *conn, int fd); extern int s2n_connection_set_read_fd(struct s2n_connection *conn, int readfd); -extern int s2n_connection_set_write_fd(struct s2n_connection *conn, int readfd); +extern int s2n_connection_set_write_fd(struct s2n_connection *conn, int writefd); typedef int s2n_recv_fn(void *io_context, uint8_t *buf, uint32_t len); typedef int s2n_send_fn(void *io_context, const uint8_t *buf, uint32_t len);
py_tf: fix headers when module is disabled.
#include "py/objlist.h" #include "py/objtuple.h" +#include "imlib_config.h" #include "py_helper.h" +#ifdef IMLIB_ENABLE_TF #include "py_assert.h" #include "py_image.h" #include "ff_wrapper.h" #include "libtf_person_detect_model_data.h" #include "py_tf.h" -#ifdef IMLIB_ENABLE_TF - #define PY_TF_PUTCHAR_BUFFER_LEN 1023 void py_tf_alloc_putchar_buffer()
u3: print errors while deleting snapshot patch
@@ -433,10 +433,16 @@ _ce_patch_delete(void) c3_c ful_c[8193]; snprintf(ful_c, 8192, "%s/.urb/chk/control.bin", u3P.dir_c); - c3_unlink(ful_c); + if ( unlink(ful_c) ) { + fprintf(stderr, "loom: failed to delete control.bin: %s\r\n", + strerror(errno)); + } snprintf(ful_c, 8192, "%s/.urb/chk/memory.bin", u3P.dir_c); - c3_unlink(ful_c); + if ( unlink(ful_c) ) { + fprintf(stderr, "loom: failed to remove memory.bin: %s\r\n", + strerror(errno)); + } } /* _ce_patch_verify(): check patch data mug.
Bump actions checkout version
@@ -15,7 +15,7 @@ jobs: os: [ubuntu-latest, ubuntu-16.04] steps: - - uses: actions/checkout@v1 + - uses: actions/checkout@v2 - name: Make dictu and run tests (No HTTP) run: | make debug DISABLE_HTTP=1 @@ -36,7 +36,7 @@ jobs: os: [macOS-latest] steps: - - uses: actions/checkout@v1 + - uses: actions/checkout@v2 - name: Make dictu and run tests (No HTTP) run: | make debug DISABLE_HTTP=1
vulkan: Remove unused VkPhysicalDeviceFeatures
@@ -1749,7 +1749,6 @@ static VkResult overlay_CreateDevice( // Advance the link info for the next element on the chain chain_info->u.pLayerInfo = chain_info->u.pLayerInfo->pNext; - VkPhysicalDeviceFeatures device_features = {}; std::vector<const char*> enabled_extensions(pCreateInfo->ppEnabledExtensionNames, pCreateInfo->ppEnabledExtensionNames +
UWP: uninstalling application before installing
@@ -651,14 +651,37 @@ namespace "run" do end namespace "uwp" do + def uninstallApplication(deviceGuid, targetAppxFN) + args = [] + args << "uninstall" + args << "-file" + args << "\""+targetAppxFN+"\"" + args << "-g" + args << "\"" + deviceGuid + "\"" + puts Jake.run($winAppDeploy, args) + end - def deployApplication(deviceGuid, targetAppxFN) + def updateApplication(deviceGuid, targetAppxFN) + args = [] + args << "update" + args << "-file" + args << "\""+targetAppxFN+"\"" + args << "-g" + args << "\"" + deviceGuid + "\"" + puts Jake.run($winAppDeploy, args) + end + + def deployApplication(deviceGuid, targetAppxFN, dependencyDir) args = [] args << "install" args << "-file" args << "\""+targetAppxFN+"\"" args << "-g" args << "\"" + deviceGuid + "\"" + if (dependencyDir.nil?) + args << "-dependency" + args << dependencyDir + end puts Jake.run($winAppDeploy, args) end @@ -702,8 +725,12 @@ namespace "run" do deviceGuid = devicesResult[/\w+-\w+-\w+-\w+-\w+/] if ((deviceGuid != nil) && (deviceGuid != "")) - Dir[$rhodes_bin_dir + '/AppxPackageDir/*/**/Dependencies/ARM/*.appx'].each { |f| deployApplication(deviceGuid, f) if File.file?(f) } - deployApplication(deviceGuid, $targetAppxFileName) + dependencyDir = nil + Dir[$rhodes_bin_dir + '/AppxPackageDir/*/**/Dependencies/ARM/'].each { |f| dependencyDir = f if Dir.exist?(f) } + + uninstallApplication(deviceGuid, $targetAppxFileName) + deployApplication(deviceGuid, $targetAppxFileName, dependencyDir) + launchApplication(deviceGuid, $targetAppxFileName) else puts "Error: no available devices connected"
Fix print-color-mode-default support so that it can be configured and defaults to the printer's value (Issue
@@ -4,6 +4,8 @@ CHANGES - OpenPrinting CUPS 2.4.0 - 2021-11-29 Changes in CUPS v2.4.1 (TBA) ---------------------------- +- The default color mode now is now configurable and defaults to the printer's + reported default mode (Issue #277) - Configuration script now checks linking for -Wl,-pie flags (Issue #303) - Fixed memory leaks - in testi18n (Issue #313), in `cups_enum_dests()` (Issue #317), in `_cupsEncodeOption()` and `http_tls_upgrade()` (Issue #322)
Added more SceLsdb NIDs
@@ -6001,13 +6001,18 @@ modules: kernel: false nid: 0x6BC25E17 functions: + sceLsdbGetAttribute: 0xD494B2C3 sceLsdbGetAppVer: 0x63AB6A8F sceLsdbGetBootable: 0x46BF2766 sceLsdbGetCategory: 0x6469796F sceLsdbGetDiscId: 0x1F569CFC sceLsdbGetGameDataId: 0x2FFE0E3F + sceLsdbGetMetaContentsPath: 0x9117289F sceLsdbGetName: 0xD02A8B85 - sceLsdbGetOriginalPath: 0x92D14842 sceLsdbGetParentalLevel: 0x226B12F7 sceLsdbGetSelfPath: 0xD6B57313 + sceLsdbGetSystemVersion: 0x9AE94B9F + sceLsdbGetStitle: 0x3B064DF5 + sceLsdbGetTitle: 0x4141EBCD + sceLsdbGetOriginalPath: 0x92D14842 sceLsdbGetType: 0xDEC358E4
Add a new job to run icw with gpos using gpdb memory allocation functions
@@ -10,6 +10,7 @@ groups: - compile_gpdb_aix7_remote - icw_planner_centos6 - icw_gporca_centos6 + - icw_gporca_centos6_gpos_memory - icw_gporca_centos7 - icw_gporca_sles11 - icw_planner_ictcp_centos6 @@ -783,6 +784,25 @@ jobs: TEST_OS: centos CONFIGURE_FLAGS: {{configure_flags}} +- name: icw_gporca_centos6_gpos_memory + plan: + - aggregate: + - get: gpdb_src + passed: [compile_gpdb_centos6] + - get: bin_gpdb + resource: bin_gpdb_centos6 + passed: [compile_gpdb_centos6] + trigger: true + - get: centos-gpdb-dev-6 + - task: ic_gpdb + file: gpdb_src/concourse/tasks/ic_gpdb.yml + image: centos-gpdb-dev-6 + params: + MAKE_TEST_COMMAND: PGOPTIONS='-c optimizer=on' installcheck-world + BLDWRAP_POSTGRES_CONF_ADDONS: "fsync=off | optimizer_use_gpdb_allocators=on" + TEST_OS: centos + CONFIGURE_FLAGS: {{configure_flags}} + - name: icw_gporca_centos7 serial: true plan:
complete AUTHORS
@@ -55,6 +55,5 @@ reporting bugs, providing fixes, suggesting useful features or other: Ricardo Martin Camarero <https://github.com/rickyepoderi> Filip Vujicic <https://github.com/FilipVujicic> Janusz Ulanowski <https://github.com/janul> - AIMOTO Norihito + Aimoto Norihito <https://github.com/oss-aimoto> Andy Lindeman <https://github.com/alindeman> -
Don't activate paths more than once when writting MPCIDs
@@ -77,9 +77,11 @@ protoop_arg_t write_mp_new_connection_id_frame(picoquic_cnx_t* cnx) consumed = byte_index; + if (p->state < path_active) { /* Now that we sent the MP NEW CONNECTION ID frame, we should be active to receive packets */ mp_receive_path_active(cnx, p, picoquic_current_time()); } + } /* Do not freem mncic yet, do it in notify! */ /* my_free(cnx, mncic); */
parsers: conf: remove typo of Time_Format in syslog-rfc3164 parser config
Regex /^\<(?<pri>[0-9]+)\>(?<time>[^ ]* {1,2}[^ ]* [^ ]*) (?<host>[^ ]*) (?<ident>[a-zA-Z0-9_\/\.\-]*)(?:\[(?<pid>[0-9]+)\])?(?:[^\:]*\:)? *(?<message>.*)$/ Time_Key time Time_Format %b %d %H:%M:%S - Time_Format %Y-%m-%dT%H:%M:%S.%L Time_Keep On [PARSER]
contributing / interface: 'confirm' -> 'apply'
@@ -23,7 +23,7 @@ this folder, [pkg/interface][interface]. You can find an `urbitrc-sample` here for reference. Then `cd` into the application's folder and `npm install` the dependencies, then `gulp watch` to watch for changes. -On your development ship, ensure you `|commit %home` to confirm your changes. +On your development ship, ensure you `|commit %home` to apply your changes. Once you're done and ready to make a pull request, running `gulp bundle-prod` will make the production files and deposit them in [pkg/arvo][arvo]. Create a pull request with both the production files, and the source code you were
groups-js: seperate user's ship from rest of contacts
@@ -20,8 +20,11 @@ export class ContactSidebar extends Component { let responsiveClasses = props.activeDrawer === "contacts" ? "db" : "dn db-ns"; - let me = (window.ship in props.defaultContacts) ? - props.defaultContacts[window.ship] : { color: '0x0', nickname: null}; + let me = (window.ship in props.contacts) + ? props.contacts[window.ship] + : (window.ship in props.defaultContacts) + ? props.defaultContacts[window.ship] + : { color: '0x0', nickname: null }; let shareSheet = !(window.ship in props.contacts) ? @@ -32,11 +35,23 @@ export class ContactSidebar extends Component { path={props.path} selected={props.path + "/" + window.ship === props.selectedContact} /> - ) : (<div></div>); + ) : ( + <> + <h2 className="f9 pt4 pr4 pb2 pl4 gray2 c-default">You</h2> + <ContactItem + ship={window.ship} + nickname={me.nickname} + color={me.color} + path={props.path} + selected={props.path + "/" + window.ship === props.selectedContact} + /> + </> + ); group.delete(window.ship); let contactItems = Object.keys(props.contacts) + .filter(c => c !== window.ship) .map((contact) => { group.delete(contact); let path = props.path + "/" + contact;
Work CI-CD Fix condition for WIN builds so they publish to nuget only on builds from main. ***NO_CI***
@@ -855,12 +855,16 @@ jobs: ArtifactName: deployables ArtifactType: Container - # push NuGet packages to Azure Artifacts feed (always happens except on PR builds) + # push NuGet packages to Azure Artifacts feed (always happens when building from main, except on PR builds) - task: NuGetCommand@2 condition: >- and( succeeded(), - eq(variables['System.PullRequest.PullRequestId'], '') + eq(variables['System.PullRequest.PullRequestId'], ''), + or( + eq(variables['ForceUpload'], true), + eq(variables['Build.SourceBranchName'], 'main') + ) ) continueOnError: true displayName: Push NuGet packages to Azure Artifacts @@ -871,12 +875,16 @@ jobs: publishFeedCredentials: 'AzureArtifacts-nf-interpreter' allowPackageConflicts: true - # push NuGet packages to NuGet (always happens except on PR builds) + # push NuGet packages to NuGet (always happens when building from main, except on PR builds) - task: NuGetCommand@2 condition: >- and( succeeded(), - eq(variables['System.PullRequest.PullRequestId'], '') + eq(variables['System.PullRequest.PullRequestId'], ''), + or( + eq(variables['ForceUpload'], true), + eq(variables['Build.SourceBranchName'], 'main') + ) ) continueOnError: true displayName: Push NuGet packages to NuGet @@ -1013,12 +1021,16 @@ jobs: ArtifactName: deployables ArtifactType: Container - # push NuGet packages to Azure Artifacts feed (always happens except on PR builds) + # push NuGet packages to Azure Artifacts feed (always happens when building from main, except on PR builds) - task: NuGetCommand@2 condition: >- and( succeeded(), - eq(variables['System.PullRequest.PullRequestId'], '') + eq(variables['System.PullRequest.PullRequestId'], ''), + or( + eq(variables['ForceUpload'], true), + eq(variables['Build.SourceBranchName'], 'main') + ) ) continueOnError: true displayName: Push NuGet packages to Azure Artifacts @@ -1029,12 +1041,16 @@ jobs: publishFeedCredentials: 'AzureArtifacts-nf-interpreter' allowPackageConflicts: true - # push NuGet packages to NuGet (always happens except on PR builds) + # push NuGet packages to NuGet (always happens when building from main, except on PR builds) - task: NuGetCommand@2 condition: >- and( succeeded(), - eq(variables['System.PullRequest.PullRequestId'], '') + eq(variables['System.PullRequest.PullRequestId'], ''), + or( + eq(variables['ForceUpload'], true), + eq(variables['Build.SourceBranchName'], 'main') + ) ) continueOnError: true displayName: Push NuGet packages to NuGet
Tests: more timeout tests.
@@ -25,6 +25,31 @@ Connection: close self.assertEqual(resp['status'], 408, 'status header read timeout') + @unittest.expectedFailure + def test_settings_header_read_timeout_update(self): + self.load('empty') + + self.conf({'http': { 'header_read_timeout': 4 }}, '/settings') + + (resp, sock) = self.http(b"""GET / HTTP/1.1 +""", start=True, raw=True) + + time.sleep(2) + + (resp, sock) = self.http(b"""Host: localhost +""", start=True, sock=sock, raw=True) + + time.sleep(2) + + resp = self.http(b"""Connection: close + +""", sock=sock, raw=True) + + time.sleep(2) + + self.assertEqual(resp['status'], 408, + 'status header read timeout update') + def test_settings_body_read_timeout(self): self.load('empty') @@ -43,6 +68,32 @@ Connection: close self.assertEqual(resp['status'], 408, 'status body read timeout') + def test_settings_body_read_timeout_update(self): + self.load('empty') + + self.conf({'http': { 'body_read_timeout': 4 }}, '/settings') + + (resp, sock) = self.http(b"""POST / HTTP/1.1 +Host: localhost +Content-Length: 10 +Connection: close + +""", start=True, raw=True) + + time.sleep(2) + + (resp, sock) = self.http(b"""012""", start=True, sock=sock, raw=True) + + time.sleep(2) + + (resp, sock) = self.http(b"""345""", start=True, sock=sock, raw=True) + + time.sleep(2) + + resp = self.http(b"""6789""", sock=sock, raw=True) + + self.assertEqual(resp['status'], 200, 'status body read timeout update') + def test_settings_send_timeout(self): self.load('mirror')