message
stringlengths 6
474
| diff
stringlengths 8
5.22k
|
---|---|
PACSign: add SDM hash to payload for content types
In addition to PR, add the SDM hash to the RKH payload for content
types SR, SR_TEST and PR_TEST. | @@ -912,7 +912,10 @@ class RHP_reader(_READER_BASE):
while payload.size() < 48:
payload.append_byte(0)
- if self.bitstream_type == database.CONTENT_PR:
+ needs_hash = [database.CONTENT_SR, database.CONTENT_PR,
+ database.CONTENT_SR_TEST, database.CONTENT_PR_TEST]
+
+ if self.bitstream_type in needs_hash:
# Add hash to payload and pad
payload.append_data(self.s10_root_hash.data)
|
Minimize pre-allocation for outgoing buffer
Current implementation would pre-initialize a whole memory page for the
response buffer, which is wasteful for short AJAX requests. | @@ -99,7 +99,7 @@ static int write_header(FIOBJ o, void *w_) {
return 0;
}
-static FIOBJ headers2str(http_s *h) {
+static FIOBJ headers2str(http_s *h, uintptr_t padding) {
if (!h->method && !!h->status_str)
return FIOBJ_INVALID;
@@ -108,7 +108,11 @@ static FIOBJ headers2str(http_s *h) {
connection_hash = fio_siphash("connection", 10);
struct header_writer_s w;
- w.dest = fiobj_str_buf(0);
+ {
+ const uintptr_t header_length_guess =
+ fiobj_hash_count(h->private_data.out_headers) * 96;
+ w.dest = fiobj_str_buf(header_length_guess + padding);
+ }
http1pr_s *p = handle2pr(h);
if (p->is_client == 0) {
@@ -176,7 +180,7 @@ static FIOBJ headers2str(http_s *h) {
/** Should send existing headers and data */
static int http1_send_body(http_s *h, void *data, uintptr_t length) {
- FIOBJ packet = headers2str(h);
+ FIOBJ packet = headers2str(h, length);
if (!packet) {
http1_after_finish(h);
return -1;
@@ -189,7 +193,7 @@ static int http1_send_body(http_s *h, void *data, uintptr_t length) {
/** Should send existing headers and file */
static int http1_sendfile(http_s *h, int fd, uintptr_t length,
uintptr_t offset) {
- FIOBJ packet = headers2str(h);
+ FIOBJ packet = headers2str(h, 0);
if (!packet) {
close(fd);
http1_after_finish(h);
@@ -221,7 +225,7 @@ static int http1_sendfile(http_s *h, int fd, uintptr_t length,
/** Should send existing headers or complete streaming */
static void htt1p_finish(http_s *h) {
- FIOBJ packet = headers2str(h);
+ FIOBJ packet = headers2str(h, 0);
if (packet)
fiobj_send_free((handle2pr(h)->p.uuid), packet);
else {
|
zephyr: drivers: displight: Use 'period' from PWM spec inplace of 'frequency'
Update driver to use 'period' from PWM spec inplace of 'frequency'
BRANCH=none
TEST=zmake testall | @@ -21,7 +21,7 @@ BUILD_ASSERT(DT_NUM_INST_STATUS_OKAY(DT_DRV_COMPAT) == 1,
#define DISPLIGHT_PWM_NODE DT_INST_PWMS_CTLR(0)
#define DISPLIGHT_PWM_CHANNEL DT_INST_PWMS_CHANNEL(0)
#define DISPLIGHT_PWM_FLAGS DT_INST_PWMS_FLAGS(0)
-#define DISPLIGHT_PWM_PERIOD_NS (NSEC_PER_SEC / DT_INST_PROP(0, frequency))
+#define DISPLIGHT_PWM_PERIOD_NS DT_INST_PWMS_PERIOD(0)
static int displight_percent;
|
Specification: Fix minor spelling mistakes | @@ -120,7 +120,7 @@ severity:warning
ingroup:kdb
number:19
-description:names of Plugins must start have the position number as second char
+description:names of Plugins must start with the position number as second char
severity:warning
ingroup:kdb
|
Fix coverity CID - Wrong sizeof() arg in rsa_freectx() | @@ -839,7 +839,7 @@ static void rsa_freectx(void *vprsactx)
OPENSSL_free(prsactx->propq);
free_tbuf(prsactx);
- OPENSSL_clear_free(prsactx, sizeof(prsactx));
+ OPENSSL_clear_free(prsactx, sizeof(*prsactx));
}
static void *rsa_dupctx(void *vprsactx)
|
Utilities: Added -x M335 to gen_pack to supress false positives on CMSIS-Drivers. | @@ -189,7 +189,7 @@ POPD
:: Checking
-Win32\PackChk.exe %RELEASE_PATH%\ARM.CMSIS.pdsc -n %RELEASE_PATH%\PackName.txt -x M353 -x M364
+Win32\PackChk.exe %RELEASE_PATH%\ARM.CMSIS.pdsc -n %RELEASE_PATH%\PackName.txt -x M353 -x M364 -x M335
:: --Check if PackChk.exe has completed successfully
IF %errorlevel% neq 0 GOTO ErrPackChk
|
TinyEXR: Fix leak in ParseEXRVersionFromFile() | @@ -13289,6 +13289,7 @@ int ParseEXRVersionFromFile(EXRVersion *version, const char *filename) {
fseek(fp, 0, SEEK_SET);
if (file_size < tinyexr::kEXRVersionSize) {
+ fclose(fp);
return TINYEXR_ERROR_INVALID_FILE;
}
|
Fixed renaming issue with apache-mynewt-core. | @@ -558,7 +558,7 @@ log_fcb_walk_sector(struct log *log, log_walk_func_t walk_func,
* last entry), add a bookmark pointing to this walk's start location.
*/
if (log_offset->lo_ts >= 0) {
- fcb_log_add_bmark(fcb_log, &loc, log_offset->lo_index);
+ log_fcb_add_bmark(fcb_log, &loc, log_offset->lo_index);
}
#endif
do {
|
Tweak attribute list to improve test coverage | @@ -5082,7 +5082,10 @@ START_TEST(test_alloc_realloc_many_attributes)
" l='12'"
" m='13'"
" n='14'"
- " p='15'>"
+ " p='15'"
+ " q='16'"
+ " r='17'"
+ " s='18'>"
"</doc>";
int i;
#define MAX_REALLOC_COUNT 10
|
Lost fixes to nodemcu-partition.py | @@ -126,7 +126,7 @@ def load_PT(data, args):
"""
PTrec,recs = unpack_RCR(data)
- flash_size = fs.args if args.fs is not None else DEFAULT_FLASH_SIZE
+ flash_size = args.fs if args.fs is not None else DEFAULT_FLASH_SIZE
# The partition table format is a set of 3*uint32 fields (type, addr, size),
# with the optional last slot being an end marker (0,size,0) where size is
@@ -367,7 +367,7 @@ def main():
# ---------- Write to a temp file and use esptool to write it to flash ---------- #
spiffs_file = arg.sf
- espargs = base + ['', str(sa), spiffs_file]
+ espargs = base + ['write_flash', str(sa), spiffs_file]
esptool.main(espargs)
# ---------- Clean up temp directory ---------- #
|
Enable fp-contract=fast for all NO_INVARIANCE builds | @@ -167,6 +167,10 @@ macro(astcenc_set_properties NAME)
target_compile_definitions(${NAME}
PRIVATE
ASTCENC_NO_INVARIANCE=1)
+
+ target_compile_options(${NAME}
+ PRIVATE
+ $<$<NOT:$<CXX_COMPILER_ID:MSVC>>:-ffp-contract=fast>)
endif()
if(${CLI})
@@ -273,8 +277,7 @@ macro(astcenc_set_properties NAME)
if(${NO_INVARIANCE})
target_compile_options(${NAME}
PRIVATE
- $<$<NOT:$<CXX_COMPILER_ID:MSVC>>:-mfma>
- $<$<NOT:$<CXX_COMPILER_ID:MSVC>>:-ffp-contract=fast>)
+ $<$<NOT:$<CXX_COMPILER_ID:MSVC>>:-mfma>)
endif()
endif()
|
config_format: add new optional head for linked list | @@ -102,6 +102,10 @@ struct flb_cf {
/* set the last error found */
char *error_str;
+
+
+ /* a list head entry in case the caller want's to link contexts */
+ struct mk_list _head;
};
|
Fix for when building against LibreSSL | @@ -638,7 +638,7 @@ static EVP_CIPHER * aes_256_ctr_cipher = NULL;
void _libssh2_openssl_crypto_init(void)
{
-#if OPENSSL_VERSION_NUMBER >= 0x10100000L
+#if (OPENSSL_VERSION_NUMBER >= 0x10100000L && !defined(LIBSSH2_LIBRESSL))
#ifndef OPENSSL_NO_ENGINE
ENGINE_load_builtin_engines();
ENGINE_register_all_complete();
|
Reduce namespace length further | @@ -203,7 +203,7 @@ extern "C" {
#define MAX_NUM_HASH_ARRAY 2000
#define GRIB_NAMESPACE 10
-#define MAX_NAMESPACE_LEN 128
+#define MAX_NAMESPACE_LEN 64
#define GRIB_MY_BUFFER 0
#define GRIB_USER_BUFFER 1
|
sds: add new flb_sds_casecmp() api | @@ -87,6 +87,15 @@ static inline int flb_sds_cmp(flb_sds_t s, const char *str, int len)
return strncmp(s, str, len);
}
+static inline int flb_sds_casecmp(flb_sds_t s, const char *str, int len)
+{
+ if (flb_sds_len(s) != len) {
+ return -1;
+ }
+
+ return strncasecmp(s, str, len);
+}
+
flb_sds_t flb_sds_create(const char *str);
flb_sds_t flb_sds_create_len(const char *str, int len);
flb_sds_t flb_sds_create_size(size_t size);
|
Fix type of img file in guide | The following procedure describes how to register mobile apps under your account settings. While the number of mobile apps is unlimited, application quota consumption rules are enforced based on your account plan. [Contact us](mailto:[email protected]) if you have questions about your mobile account options.
1. Access your [Account settings](https://carto.com/docs/carto-editor/your-account/#how-to-access-your-account-options) from the CARTO Dashboard. Your profile information appears.
- <span class="wrap-border"><img src="../../img/avatar.jpg" alt="Access mobile apps from API keys" /></span>
+ <span class="wrap-border"><img src="../../img/avatar.gif" alt="Access mobile apps from API keys" /></span>
2. Click _API keys_. The API key page opens, displaying options for _CARTO_ or _Mobile Apps_.
3. Click _Mobile apps_ to add mobile applications to your account
|
adding records syntax to EBNF | @@ -140,15 +140,21 @@ have the same name. The body is a sequence of statements.
Here is the complete syntax of Titan in extended BNF. As usual in extended BNF, {A} means 0 or more As, and \[A\] means an optional A.
- program ::= {tlfunc | tlvar}
+ program ::= {tlfunc | tlvar | tlrecord}
tlfunc ::= [local] function Name '(' [parlist] ')' ':' type block end
tlvar ::= [local] Name [':' type] '=' Numeral
+ tlrecord ::= record Name recordfields end
+
parlist ::= Name ':' type {',' Name ':' type}
- type ::= integer | float | boolean | string | '{' type '}' |
+ type ::= integer | float | boolean | string | '{' type '}'
+
+ recordfields ::= recordfield {recordfield}
+
+ recordfield ::= Name ':' type
block ::= {stat} [retstat]
@@ -164,7 +170,7 @@ Here is the complete syntax of Titan in extended BNF. As usual in extended BNF,
retstat ::= return exp [';']
- var ::= Name | prefixexp '[' exp ']'
+ var ::= Name | prefixexp '[' exp ']' | prefixexp '.' Name
explist ::= exp {',' exp}
|
gtkui protocol combo (change): set default_protocol
menu Remote -> Open Location does not attempt a connection
using the selected protocol.. if you select another protocol
in the toolbar protocol combo
this commit fixes the issue | @@ -40,7 +40,8 @@ pthread_mutex_t log_mutex = PTHREAD_MUTEX_INITIALIZER;
pthread_t main_thread_id;
GList * viewedit_processes = NULL;
-gboolean on_key_press_combo_toolbar(GtkWidget *widget, GdkEventKey *event, gpointer data);
+static gboolean on_key_press_combo_toolbar(GtkWidget *widget, GdkEventKey *event, gpointer data);
+static void on_combo_protocol_change_cb (GtkComboBox *cb, gpointer data);
static int combo_key_pressed = 0;
static int
@@ -855,7 +856,10 @@ CreateConnectToolbar (GtkWidget * parent)
toolbar_combo_protocol = gtk_combo_box_text_new ();
gtk_box_pack_start (GTK_BOX (tempwid), toolbar_combo_protocol, TRUE, FALSE, 0);
-
+ g_signal_connect (G_OBJECT (toolbar_combo_protocol),
+ "changed",
+ G_CALLBACK (on_combo_protocol_change_cb),
+ NULL);
num = 0;
j = 0;
gftp_lookup_global_option ("default_protocol", &default_protocol);
@@ -1352,7 +1356,15 @@ _get_selected_protocol ()
return i;
}
-gboolean
+static void
+on_combo_protocol_change_cb (GtkComboBox *cb, gpointer data)
+{
+ GtkComboBoxText *combo = GTK_COMBO_BOX_TEXT(cb);
+ char *txt = gtk_combo_box_text_get_active_text (combo);
+ gftp_set_global_option ("default_protocol", txt);
+}
+
+static gboolean
on_key_press_combo_toolbar(GtkWidget *widget, GdkEventKey *event, gpointer data)
{
if (event->type == GDK_KEY_PRESS) {
|
Include complex rather than complex.h in C++ contexts
to avoid name clashes e.g. with boost headers that use I as a generic placeholder.
Fixes as suggested by aprokop in that issue ticket. | @@ -86,7 +86,11 @@ lapack_complex_float lapack_make_complex_float( float re, float im );
/* Complex type (double precision) */
#ifndef lapack_complex_double
+#ifndef __cplusplus
#include <complex.h>
+#else
+#include <complex>
+#endif
#define lapack_complex_double double _Complex
#endif
|
Conan build-time optimization. | @@ -27,6 +27,13 @@ jobs:
CC: ${{ matrix.compiler[0] }}
CXX: ${{ matrix.compiler[1] }}
run: |
+ # build profile
+ conan profile new release --detect
+ conan profile update settings.build_type=Release release
+ #Note no backwards compatiblity for gcc5 needed, setting libcxx to c++11.
+ conan profile update settings.compiler.libcxx=libstdc++11 release
+ conan profile show release
+ # host profile
conan profile new default --detect
conan profile update settings.build_type=${{ matrix.type }} default
#Note no backwards compatiblity for gcc5 needed, setting libcxx to c++11.
@@ -42,7 +49,7 @@ jobs:
-o celix:build_all=True
run: |
#force require libcurl 7.64.1, due to a sha256 verify issue in libcurl/7.87.0
- conan install . celix/ci -pr:b default -pr:h default -if build ${CONAN_BUILD_OPTIONS} -b missing --require-override=libcurl/7.64.1 --require-override=openssl/1.1.1s
+ conan install . celix/ci -pr:b release -pr:h default -if build ${CONAN_BUILD_OPTIONS} -b missing --require-override=libcurl/7.64.1 --require-override=openssl/1.1.1s
- name: Build
env:
CC: ${{ matrix.compiler[0] }}
@@ -61,7 +68,7 @@ jobs:
CC: ${{ matrix.compiler[0] }}
CXX: ${{ matrix.compiler[1] }}
run: |
- conan create -pr:b default -pr:h default -tf examples/conan_test_package -tbf test-build -o celix:celix_cxx17=True -o celix:celix_install_deprecated_api=True --require-override=libcurl/7.64.1 --require-override=openssl/1.1.1s .
+ conan create -pr:b release -pr:h default -tf examples/conan_test_package -tbf test-build -o celix:celix_cxx17=True -o celix:celix_install_deprecated_api=True --require-override=libcurl/7.64.1 --require-override=openssl/1.1.1s .
build-apt:
runs-on: ubuntu-20.04
|
stm32l4/timer: Improved getting timer value
Previous method always did two addtional reads on timer fetch fail
JIRA: | @@ -60,13 +60,16 @@ static u32 timer_getCnt(void)
{
u32 cnt[2];
- do {
/* From documentation: "It should be noted that for a reliable LPTIM_CNT
* register read access, two consecutive read accesses must be performed and compared.
* A read access can be considered reliable when the
* values of the two consecutive read accesses are equal." */
+
+ cnt[0] = *(timer_common.lptim + lptim_cnt);
+
+ do {
+ cnt[1] = cnt[0];
cnt[0] = *(timer_common.lptim + lptim_cnt);
- cnt[1] = *(timer_common.lptim + lptim_cnt);
} while (cnt[0] != cnt[1]);
return cnt[0] & 0xffffu;
@@ -109,6 +112,8 @@ time_t hal_getTimer(void)
lower = timer_getCnt();
if (*(timer_common.lptim + lptim_isr) & (1 << 1)) {
+ /* Check if we have unhandled overflow event.
+ * If so, upper is one less than it should be */
if (timer_getCnt() >= lower)
++upper;
}
|
hosts: fixes in README.md | -- infos = Information about HOSTS plugin is in keys below
+- infos = Information about hosts plugin is in keys below
- infos/author = Markus Raab <[email protected]>
- infos/licence = BSD
- infos/provides = storage/hosts
@@ -18,34 +18,30 @@ with hostnames, one line per IP address. The format is described in `hosts(5)`.
### Hostnames
-Canonical hostnames are stored as key names with the IP address as key
-value.
+Canonical hostnames are stored as key names with their IP address
+as value.
### Aliases
Aliases are stored as sub keys with a read only duplicate of the
-associated ip address as value.
+associated IP address as value.
### Comments
-Comments are stored according to the comment metadata specification (see [/doc/METADATA.ini](/doc/METADATA.ini) for more information)
+Comments are stored according to the comment metadata specification
+(see [/doc/METADATA.ini](/doc/METADATA.ini) for more information).
-### Multi-Line Comments
+### Ordering
-Since line breaks are preserved, you can identify multi line comments
-by their trailing line break.
-
-
-## Restrictions
-
-The ordering of the hosts is stored in metakeys of type "order".
-The value is an ascending number. Ordering of aliases is NOT preserved.
+The ordering of the hosts is stored in metakeys of type `order`.
+The value is an ascending number. Ordering of aliases is
+*not* preserved.
## Examples
Mount the plugin:
- $ kdb mount --with-recommends /etc/hosts system/hosts hosts
+ $ sudo kdb mount --with-recommends /etc/hosts system/hosts hosts
Print out all known hosts and their aliases:
@@ -59,9 +55,11 @@ Check if a comment is belonging to host "localhost":
$ kdb lsmeta system/hosts/ipv4/localhost
-Try to change the host "localhost", should fail because it is not an ipv4 address:
+Try to change the host "localhost", should fail because it is not an
+IPv4 address:
+
+ $ sudo kdb set system/hosts/ipv4/localhost ::1
- $ kdb set system/hosts/ipv4/localhost ::1
```sh
# Backup-and-Restore:/examples/hosts
|
Add comment to HTML test file. | <!DOCTYPE html>
<html lang="en">
<head>
+ <!-- This file is used for prototyping and testing CSS and HTML changes -->
<title>Test Page for PAPPL Stylesheet</title>
<link rel="shortcut icon" href="favicon.png" type="image/png">
<link rel="stylesheet" href="style.css">
|
run_tests.pl: Improve diagnostics on the use of HARNESS_JOBS | @@ -47,6 +47,7 @@ my %tapargs =
);
$tapargs{jobs} = $jobs if $jobs > 1;
+print "Using HARNESS_JOBS=$jobs\n" if $jobs > 1;
# Additional OpenSSL special TAP arguments. Because we can't pass them via
# TAP::Harness->new(), they will be accessed directly, see the
@@ -57,7 +58,7 @@ $openssl_args{'failure_verbosity'} = $ENV{HARNESS_VERBOSE} ? 0 :
$ENV{HARNESS_VERBOSE_FAILURE_PROGRESS} ? 2 :
1; # $ENV{HARNESS_VERBOSE_FAILURE}
print "Warning: HARNESS_JOBS > 1 overrides HARNESS_VERBOSE\n"
- if $jobs > 1;
+ if $jobs > 1 && $ENV{HARNESS_VERBOSE};
print "Warning: HARNESS_VERBOSE overrides HARNESS_VERBOSE_FAILURE*\n"
if ($ENV{HARNESS_VERBOSE} && ($ENV{HARNESS_VERBOSE_FAILURE}
|| $ENV{HARNESS_VERBOSE_FAILURE_PROGRESS}));
@@ -76,7 +77,7 @@ sub reorder {
my $key = pop;
# for parallel test runs, do slow tests first
- if (defined $jobs && $jobs > 1 && $key =~ m/test_ssl_new|test_fuzz/) {
+ if ($jobs > 1 && $key =~ m/test_ssl_new|test_fuzz/) {
$key =~ s/(\d+)-/00-/;
}
return $key;
|
Fix batching edge case; | @@ -524,6 +524,13 @@ void lovrGraphicsBatch(BatchRequest* req) {
*(req->indices) = lovrGraphicsMapBuffer(STREAM_INDEX, req->indexCount);
*(req->baseVertex) = state.cursors[STREAM_VERTEX];
}
+
+ // The buffer mapping here could have triggered a flush, so if we were hoping to batch with
+ // something but the batch count is zero now, we just start a new batch. Maybe there's a better
+ // way to detect this.
+ if (batch && state.batchCount == 0) {
+ batch = NULL;
+ }
}
// Start a new batch
|
matekf411rx: readd brushed. | "defines": {
"BRUSHLESS_TARGET": ""
}
+ },
+ {
+ "name": "brushed",
+ "defines": {
+ "BRUSHED_TARGET": ""
+ }
}
]
},
|
Added all dipdap boards in BOARD_ID_SUPPORTING_PAGE_ERASE for tests | @@ -233,6 +233,11 @@ BOARD_ID_SUPPORTING_PAGE_ERASE = set([
0x0311, # K66F
0x1101, # Nordic-nRF52-DK
0x1102, # Nordic-nRF52840-DK
+ 0x3104, # dipdap_sdt52832b
+ 0x3108, # dipdap_sdt32429b
+ 0x3110, # dipdap_sdt32439b
+ 0x3105, # dipdap_sdt64b
+ 0x3103, # dipdap_sdt51822b
0x5500, # GR-PEACH
0x5501, # GR-LYCHEE
])
|
mavlink_opticalflow: add green flashing LED | @@ -17,6 +17,19 @@ MAV_OPTICAL_FLOW_confidence_threshold = 0.1 # Below 0.1 or so (YMMV) and the re
##############################################################################
+# LED control
+led = pyb.LED(2) # Red LED = 1, Green LED = 2, Blue LED = 3, IR LEDs = 4.
+led_state = 0
+
+def update_led():
+ global led_state
+ led_state = led_state + 1
+ if led_state == 10:
+ led.on()
+ elif led_state >= 20:
+ led.off()
+ led_state = 0
+
# Link Setup
uart = pyb.UART(3, uart_baudrate, timeout_char = 1000)
@@ -66,6 +79,7 @@ def send_optical_flow_packet(x, y, c):
checksum(temp, MAV_OPTICAL_FLOW_extra_crc))
packet_sequence += 1
uart.write(temp)
+ update_led()
sensor.reset() # Reset and initialize the sensor.
sensor.set_pixformat(sensor.RGB565) # Set pixel format to RGB565 (or GRAYSCALE)
|
Always add DYLD_LIBRARY_PATH and LD_LIBRARY_PATH into env. | @@ -225,11 +225,8 @@ class Platform(object):
return vs
@property
- def library_path_variable(self):
- if self.is_linux:
- return 'LD_LIBRARY_PATH'
- if self.is_macos:
- return 'DYLD_LIBRARY_PATH'
+ def library_path_variables(self):
+ return ['LD_LIBRARY_PATH', 'DYLD_LIBRARY_PATH']
def find_in_dict(self, dict_, default=None):
if dict_ is None:
@@ -1170,7 +1167,8 @@ class GnuToolchain(Toolchain):
if self.tc.is_clang:
if self.tc.is_from_arcadia:
- self.env.setdefault(build.host.library_path_variable, []).append('{}/lib'.format(self.tc.name_marker))
+ for lib_path in build.host.library_path_variables:
+ self.env.setdefault(lib_path, []).append('{}/lib'.format(self.tc.name_marker))
target_triple = select(default=None, selectors=[
(target.is_linux and target.is_x86_64, 'x86_64-linux-gnu'),
@@ -1235,7 +1233,8 @@ class GnuToolchain(Toolchain):
self.platform_projects.append(project)
self.c_flags_platform.append('-B{}/{}'.format(var, bin))
if ldlibs:
- self.env.setdefault(self.build.host.library_path_variable, []).append(ldlibs)
+ for lib_path in self.build.host.library_path_variables:
+ self.env.setdefault(lib_path, []).append(ldlibs)
def print_toolchain(self):
super(GnuToolchain, self).print_toolchain()
|
groups: sidebar, additional check for safety
When accepting an invite, the sidebar would crash the application
because we were accessing 'metadata' of undefined -- accessing
too many nested property levels without enough safety.
This commit just adds a check for the intermediate property. | @@ -62,6 +62,7 @@ export class GroupSidebar extends Component {
let groupChannel = `${path}/contacts${path}`
if (
props.associations[path] &&
+ props.associations[path][groupChannel] &&
props.associations[path][groupChannel].metadata
) {
name =
|
keyset: undo changes | @@ -2554,26 +2554,16 @@ int ksClose (KeySet * ks)
Key * k;
ksRewind (ks);
- int keysNotFreed = 0;
while ((k = ksNext (ks)) != 0)
{
- int refs = keyDecRef (k);
-
- if (refs > 0)
- {
- keysNotFreed++;
- }
-
+ keyDecRef (k);
keyDel (k);
}
- printf ("ksClose did NOT free %d keys\n", keysNotFreed);
-
if (ks->array && !test_bit (ks->flags, KS_FLAG_MMAP_ARRAY))
{
elektraFree (ks->array);
}
-
clear_bit (ks->flags, (keyflag_t) KS_FLAG_MMAP_ARRAY);
ks->array = 0;
|
Allow any subpackage of plgo project to use contribs | ALLOW market/sre/tools/config-primer/src/internal/blogic -> contrib/go/patched/hugo
# CONTRIB-1581. responsible: grihabor@, g:strm-admin
-ALLOW strm/plgo/pkg/cdn -> contrib/go/patched/hashring
+ALLOW strm/plgo -> contrib/go/patched/hashring
# STRM-1124. responsible: grihabor@, g:strm-admin
-ALLOW strm/plgo/pkg/playlist -> contrib/go/patched/m3u8
+ALLOW strm/plgo -> contrib/go/patched/m3u8
ALLOW contrib/go/patched/m3u8/example -> contrib/go/patched/m3u8
DENY .* -> contrib/go/patched/
|
Check Uncrustify returncode in code_style.py | @@ -106,8 +106,12 @@ def check_style_is_correct(src_file_list: List[str]) -> bool:
style_correct = True
for src_file in src_file_list:
uncrustify_cmd = [UNCRUSTIFY_EXE] + UNCRUSTIFY_ARGS + [src_file]
- subprocess.run(uncrustify_cmd, stdout=subprocess.PIPE, \
+ result = subprocess.run(uncrustify_cmd, stdout=subprocess.PIPE, \
stderr=subprocess.PIPE, check=False)
+ if result.returncode != 0:
+ print_err("Uncrustify returned " + str(result.returncode) + \
+ " correcting file " + src_file)
+ return False
# Uncrustify makes changes to the code and places the result in a new
# file with the extension ".uncrustify". To get the changes (if any)
@@ -135,15 +139,22 @@ def fix_style_single_pass(src_file_list: List[str]) -> None:
code_change_args = UNCRUSTIFY_ARGS + ["--no-backup"]
for src_file in src_file_list:
uncrustify_cmd = [UNCRUSTIFY_EXE] + code_change_args + [src_file]
- subprocess.run(uncrustify_cmd, check=False, stdout=STDOUT_UTF8, \
- stderr=STDERR_UTF8)
+ result = subprocess.run(uncrustify_cmd, check=False, \
+ stdout=STDOUT_UTF8, stderr=STDERR_UTF8)
+ if result.returncode != 0:
+ print_err("Uncrustify with file returned: " + \
+ str(result.returncode) + " correcting file " + \
+ src_file)
+ return False
def fix_style(src_file_list: List[str]) -> int:
"""
Fix the code style. This takes 2 passes of Uncrustify.
"""
- fix_style_single_pass(src_file_list)
- fix_style_single_pass(src_file_list)
+ if fix_style_single_pass(src_file_list) != True:
+ return 1
+ if fix_style_single_pass(src_file_list) != True:
+ return 1
# Guard against future changes that cause the codebase to require
# more passes.
|
Add st official checksum for binary file | @@ -1380,6 +1380,16 @@ static void md5_calculate(mapped_file_t *mf, const char *path) {
printf("\n");
}
+static void stlink_checksum(mapped_file_t *mp, const char* path) {
+ /* checksum that backward compatible with official ST tools */
+ uint32_t sum = 0;
+ uint8_t *mp_byte = (uint8_t *)mp->base;
+ for (size_t i = 0; i < mp->len; ++i) {
+ sum += mp_byte[i];
+ }
+ printf("file %s stlink checksum: 0x%08x\n", path, sum);
+}
+
static void stlink_fwrite_finalize(stlink_t *sl, stm32_addr_t addr) {
unsigned int val;
/* set stack*/
@@ -1462,6 +1472,7 @@ int stlink_fwrite_sram(stlink_t * sl, const char* path, stm32_addr_t addr) {
}
md5_calculate(&mf, path);
+ stlink_checksum(&mf, path);
/* check addr range is inside the sram */
if (addr < sl->sram_base) {
@@ -2688,6 +2699,7 @@ int stlink_fwrite_flash(stlink_t *sl, const char* path, stm32_addr_t addr) {
}
md5_calculate(&mf, path);
+ stlink_checksum(&mf, path);
if (sl->opt) {
idx = (unsigned int) mf.len;
@@ -3254,6 +3266,7 @@ int stlink_fwrite_option_bytes(stlink_t *sl, const char* path, stm32_addr_t addr
}
md5_calculate(&mf, path);
+ stlink_checksum(&mf, path);
err = stlink_write_option_bytes(sl, addr, mf.base, (uint32_t) mf.len);
stlink_fwrite_finalize(sl, addr);
|
xfconf: Add file workaround | #include <kdbhelper.h>
#include <kdblogger.h>
+#include <libgen.h>
#include <xfconf/xfconf.h>
int elektraXfconfOpen (Plugin * handle ELEKTRA_UNUSED, Key * errorKey ELEKTRA_UNUSED)
{
+ ELEKTRA_LOG ("try to initialize xfconf\n");
GError * err = NULL;
if (xfconf_init (&err))
{
+ ELEKTRA_LOG_DEBUG ("succeed initielize xfconf\n");
return ELEKTRA_PLUGIN_STATUS_SUCCESS;
}
else
{
- ELEKTRA_LOG_NOTICE ("unable to initialize xfconf(%d): %s\n", err->code, err->message);
+ ELEKTRA_LOG ("unable to initialize xfconf(%d): %s\n", err->code, err->message);
g_error_free (err);
return ELEKTRA_PLUGIN_STATUS_ERROR;
}
@@ -36,8 +39,10 @@ int elektraXfconfClose (Plugin * handle ELEKTRA_UNUSED, Key * errorKey ELEKTRA_U
int elektraXfconfGet (Plugin * handle ELEKTRA_UNUSED, KeySet * returned, Key * parentKey)
{
+ ELEKTRA_LOG ("issued get\n");
if (!elektraStrCmp (keyName (parentKey), "system:/elektra/modules/xfconf"))
{
+ ELEKTRA_LOG_DEBUG ("getting system modules values\n");
KeySet * contract =
ksNew (30, keyNew ("system:/elektra/modules/xfconf", KEY_VALUE, "xfconf plugin waits for your orders", KEY_END),
keyNew ("system:/elektra/modules/xfconf/exports", KEY_END),
@@ -57,6 +62,33 @@ int elektraXfconfGet (Plugin * handle ELEKTRA_UNUSED, KeySet * returned, Key * p
}
// get all keys
+ // KeySet * config = elektraPluginGetConfig (handle);
+ // todo: remove workaround which requires a channel to exist as a file
+ char * absolutePath = elektraStrDup (keyString (parentKey));
+ const char * channelName = basename (absolutePath);
+ ELEKTRA_LOG_DEBUG ("fetch keys from channel: %s\n", channelName);
+ XfconfChannel * channel = xfconf_channel_get (channelName);
+ if (channel == NULL)
+ {
+ ELEKTRA_LOG_DEBUG ("retrieved NULL attempting getting channel: %s\n", channelName);
+ }
+ GHashTable * properties = xfconf_channel_get_properties (channel, NULL);
+ if (properties == NULL)
+ {
+ ELEKTRA_LOG_DEBUG ("retrieved NULL attempting getting properties\n");
+ }
+ GList * channelKeys = g_hash_table_get_keys (properties);
+ while (channelKeys != NULL)
+ {
+ char * keyName = elektraStrDup (channelKeys->data);
+ Key * key = keyNew (keyName, KEY_END);
+ char * keyValue = g_hash_table_lookup (properties, channelKeys->data);
+ ELEKTRA_LOG_DEBUG ("found %s -> %s\n", keyName, keyValue);
+ keySetString (key, keyValue);
+ ksAppendKey (returned, key);
+ channelKeys = channelKeys->next;
+ }
+
return ELEKTRA_PLUGIN_STATUS_NO_UPDATE;
}
|
Change to use better error message for >1 file arg | @@ -29,6 +29,9 @@ int main(int argc, char** argv) {
if (optind == argc - 1) {
input_file = argv[optind];
+ } else if (optind < argc - 1) {
+ fprintf(stderr, "Expected only 1 file argument");
+ return 1;
}
if (input_file == NULL) {
|
Fix Pass:translate; | @@ -135,8 +135,8 @@ static int l_lovrPassOrigin(lua_State* L) {
static int l_lovrPassTranslate(lua_State* L) {
float translation[4];
- Pass* pass = luax_checktype(L, 2, Pass);
- luax_readvec3(L, 1, translation, NULL);
+ Pass* pass = luax_checktype(L, 1, Pass);
+ luax_readvec3(L, 2, translation, NULL);
lovrPassTranslate(pass, translation);
return 0;
}
|
Improve USB device open log
For consistency with "List USB devices", log "Open USB device". | @@ -195,7 +195,7 @@ sc_usb_connect(struct sc_usb *usb, libusb_device *device,
const struct sc_usb_callbacks *cbs, void *cbs_userdata) {
int result = libusb_open(device, &usb->handle);
if (result < 0) {
- LOGE("Open device: libusb error: %s", libusb_strerror(result));
+ LOGE("Open USB device: libusb error: %s", libusb_strerror(result));
return false;
}
|
Remove inject error from UEFI spec Clarify FW Update will continue on error | @@ -113,19 +113,19 @@ Load FW on DIMM (DimmID): (Valid|Invalid|Downgrade) [(with
confirmation or the force option)]
If the firmware is being downgraded and the force option is not provided, the user will
-be prompted to confirm the downgrade for each DCPMM.
-
+be prompted to confirm the downgrade for each DCPMM. Otherwise, for each DCPMM,
+the CLI will indicate the status of the operation.
[verse]
Downgrade firmware on DIMM (DimmID)? (y or [n]) Downgrade firmware
on DIMM (DimmID)? (y or [n])
...
-Otherwise, for each DCPMM, the CLI will indicate the status of the operation. If a failure
-occurs when updating multiple DCPMMs, the process will exit and not continue updating
-the remaining DCPMMs. The firmware will not become active until the next power cycle.
-Use the command Section 2.5.3, "Show Device Firmware" to view more detailed
-information about the active and staged firmware.
+If a failure occurs when updating multiple DCPMMs, the process will continue
+attempting to update the remaining DCPMMs requested. The firmware will not
+become active until the next power cycle. Use the command Section 2.5.3,
+"Show Device Firmware" to view more detailed information about the active
+and staged firmware.
[verse]
Load FW on DIMM (DimmID): Success, a power cycle is required to
|
reclaim memory on restart | @@ -1235,6 +1235,12 @@ u3_sist_boot(void)
}
if ( c3n == u3_Host.ops_u.nuu ) {
+ // reclaim memory from persistent caches
+ //
+ u3m_reclaim();
+
+ // restore from event log, replaying if necessary
+ //
_sist_rest();
if ( c3y == u3A->fak ) {
|
fix module objectfile handling | @@ -79,7 +79,7 @@ function _add_objectfile_to_link_arguments(target, objectfile)
if table.contains(cache, objectfile) then
return
end
- table.insert(cache, path.translate(objectfile))
+ table.insert(cache, objectfile)
common.localcache():set(cachekey, cache)
common.localcache():save(cachekey)
end
@@ -519,7 +519,7 @@ function build_modules_for_batchjobs(target, batchjobs, objectfiles, modules, op
name = name or module.cppfile,
flags = _flags,
progress = (index * 100) / total})
- _add_objectfile_to_link_arguments(target, objectfile)
+ _add_objectfile_to_link_arguments(target, path(objectfile))
elseif requiresflags then
requiresflags = get_requiresflags(target, module.requires)
target:fileconfig_add(cppfile, {force = {cxxflags = table.join(flags, requiresflags)}})
@@ -584,7 +584,7 @@ function build_modules_for_batchcmds(target, batchcmds, objectfiles, modules, op
batchcmds:mkdir(path.directory(objectfile))
batchcmds:vrunv(compinst:program(), table.join(compinst:compflags({target = target}), table.join(flags, requiresflags or {})), {envs = msvc:runenvs()})
batchcmds:add_depfiles(cppfile)
- _add_objectfile_to_link_arguments(target, objectfile)
+ _add_objectfile_to_link_arguments(target, path(objectfile))
if provide then
_add_module_to_mapper(target, referenceflag, name, name, objectfile, provide.bmi, requiresflags)
end
|
Add gcc10/arm64 DYNAMIC_ARCH build | @@ -190,3 +190,25 @@ steps:
- make -C ctest $COMMON_FLAGS
- make -C utest $COMMON_FLAGS
- make -C cpp_thread_test dgemm_tester
+---
+kind: pipeline
+name: arm64_gcc10
+
+platform:
+ os: linux
+ arch: arm64
+
+steps:
+- name: Build and Test
+ image: ubuntu:20.04
+ environment:
+ CC: gcc-10
+ COMMON_FLAGS: 'TARGET=ARMV8 DYNAMIC_ARCH=1'
+ commands:
+ - echo "MAKE_FLAGS:= $COMMON_FLAGS"
+ - apt-get update -y
+ - apt-get install -y make $CC gfortran-10 perl python g++
+ - $CC --version
+ - make QUIET_MAKE=1 $COMMON_FLAGS
+ - make -C test $COMMON_FLAGS
+
|
Avoid invalid read of linked list when closing a client inside a foreach. | @@ -2509,20 +2509,17 @@ clear_fifo_packet (WSPipeIn * pipein) {
/* Broadcast to all connected clients the given message. */
static int
-ws_broadcast_fifo (void *value, void *user_data) {
- WSClient *client = value;
- WSServer *server = user_data;
+ws_broadcast_fifo (WSClient * client, WSServer * server) {
WSPacket *packet = server->pipein->packet;
LOG (("Broadcasting to %d [%s] ", client->listener, client->remote_ip));
- if (client == NULL || user_data == NULL)
+ if (client == NULL)
return 1;
- /* no handshake for this client */
+
if (client->headers == NULL || client->headers->ws_accept == NULL) {
+ /* no handshake for this client */
LOG (("No headers. Closing %d [%s]\n", client->listener, client->remote_ip));
-
- handle_tcp_close (client->listener, client, server);
- return 0;
+ return -1;
}
LOG ((" - Sending...\n"));
@@ -2531,6 +2528,35 @@ ws_broadcast_fifo (void *value, void *user_data) {
return 0;
}
+static void
+ws_broadcast_fifo_to_clients (WSServer * server) {
+ WSClient *client = NULL;
+ void *data = NULL;
+ uint32_t *close_list = NULL;
+ int n = 0, idx = 0, i = 0, listener = 0;
+
+ if ((n = list_count (server->colist)) == 0)
+ return;
+
+ close_list = xcalloc (n, sizeof (uint32_t));
+ /* *INDENT-OFF* */
+ GSLIST_FOREACH (server->colist, data, {
+ client = data;
+ if (ws_broadcast_fifo(client, server) == -1)
+ close_list[idx++] = client->listener;
+ });
+ /* *INDENT-ON* */
+
+ client = NULL;
+ for (i = 0; i < idx; ++i) {
+ listener = close_list[i];
+ if ((client = ws_get_client_from_list (listener, &server->colist)))
+ handle_tcp_close (listener, client, server);
+ }
+
+ free (close_list);
+}
+
/* Send a message from the incoming named pipe to specific client
* given the socket id. */
static void
@@ -2669,7 +2695,7 @@ handle_strict_fifo (WSServer * server) {
if (listener != 0)
ws_send_strict_fifo_to_client (server, listener, *pa);
else
- list_foreach (server->colist, ws_broadcast_fifo, server);
+ ws_broadcast_fifo_to_clients (server);
clear_fifo_packet (pi);
}
@@ -2702,7 +2728,7 @@ handle_fixed_fifo (WSServer * server) {
}
/* broadcast message to all clients */
- list_foreach (server->colist, ws_broadcast_fifo, server);
+ ws_broadcast_fifo_to_clients (server);
clear_fifo_packet (pi);
}
|
Move from jcenter() to mavenCentral()
Refs <https://developer.android.com/studio/build/jcenter-migration>
Refs <https://jfrog.com/blog/into-the-sunset-bintray-jcenter-gocenter-and-chartcenter/> | @@ -4,7 +4,7 @@ buildscript {
repositories {
google()
- jcenter()
+ mavenCentral()
}
dependencies {
classpath 'com.android.tools.build:gradle:7.0.3'
@@ -17,7 +17,7 @@ buildscript {
allprojects {
repositories {
google()
- jcenter()
+ mavenCentral()
}
tasks.withType(JavaCompile) {
options.compilerArgs << "-Xlint:deprecation"
|
iter/thresh.h: change bool to _Bool | struct operator_p_s;
extern const struct operator_p_s* prox_thresh_create(unsigned int D, const long dim[__VLA(D)], const float lambda, const unsigned long flags, _Bool gpu);
-extern const struct operator_p_s* prox_niht_thresh_create(unsigned int D, const long dim[D], const unsigned int k, const unsigned long flags, bool gpu);
+extern const struct operator_p_s* prox_niht_thresh_create(unsigned int D, const long dim[D], const unsigned int k, const unsigned long flags, _Bool gpu);
extern void thresh_free(const struct operator_p_s* data);
|
Undo build.yaml changles | @@ -16,6 +16,10 @@ jobs:
- kyria_right
- lily58_left
- lily58_right
+<<<<<<< HEAD
+
+=======
+>>>>>>> parent of c3164f4... Updated build.yaml
include:
- board: proton_c
shield: clueboard_california
|
Point to correct path of riscv-tools directory | @@ -26,7 +26,7 @@ the following steps are necessary.
export RISCV=/path/to/install/dir
export PATH=$RISCV/bin:$PATH
- cd riscv-tools
+ cd rocket-chip/riscv-tools
./build.sh
### Compiling and running the Verilator simulation
|
ikev2: use remote proposals when installing tunnel
Type: fix | @@ -1723,7 +1723,7 @@ ikev2_create_tunnel_interface (vlib_main_t * vm,
{
ip46_address_set_ip4 (&a.local_ip, &sa->iaddr);
ip46_address_set_ip4 (&a.remote_ip, &sa->raddr);
- proposals = child->i_proposals;
+ proposals = child->r_proposals;
a.local_spi = child->r_proposals[0].spi;
a.remote_spi = child->i_proposals[0].spi;
}
@@ -1731,7 +1731,7 @@ ikev2_create_tunnel_interface (vlib_main_t * vm,
{
ip46_address_set_ip4 (&a.local_ip, &sa->raddr);
ip46_address_set_ip4 (&a.remote_ip, &sa->iaddr);
- proposals = child->r_proposals;
+ proposals = child->i_proposals;
a.local_spi = child->i_proposals[0].spi;
a.remote_spi = child->r_proposals[0].spi;
}
|
test: add platone test case test_001CreateWallet_0006CreateOneTimeWalletFailureShortSize
fix the issue aitos-io#1176
teambition task id: | @@ -199,6 +199,24 @@ START_TEST(test_001CreateWallet_0005CreateLoadWalletFailureNoExist)
}
END_TEST
+START_TEST(test_001CreateWallet_0006CreateOneTimeWalletFailureShortSize)
+{
+ BSINT32 rtnVal;
+ BoatPlatoneWalletConfig wallet = get_platone_wallet_settings();
+ extern BoatIotSdkContext g_boat_iot_sdk_context;
+
+ /* 1. execute unit test */
+ rtnVal = BoatWalletCreate(BOAT_PROTOCOL_PLATONE, NULL, &wallet, sizeof(BoatPlatoneWalletConfig) - 10);
+
+ /* 2. verify test result */
+ /* 2-1. verify the return value */
+ ck_assert_int_eq(rtnVal, BOAT_ERROR);
+
+ /* 2-2. verify the global variables that be affected */
+ ck_assert(g_boat_iot_sdk_context.wallet_list[0].is_used == false);
+}
+END_TEST
+
Suite *make_wallet_suite(void)
{
/* Create Suite */
@@ -215,6 +233,7 @@ Suite *make_wallet_suite(void)
tcase_add_test(tc_wallet_api, test_001CreateWallet_0003CreatePersistWalletSuccess);
tcase_add_test(tc_wallet_api, test_001CreateWallet_0004CreateLoadWalletSuccess);
tcase_add_test(tc_wallet_api, test_001CreateWallet_0005CreateLoadWalletFailureNoExist);
+ tcase_add_test(tc_wallet_api, test_001CreateWallet_0006CreateOneTimeWalletFailureShortSize);
return s_wallet;
}
|
linux-raspberrypi: Replace /lib with ${nonarch_base_libdir}
Use standard /lib variable name and avoid
QA errors when usermerge DISTRO_FEATURE is enabled. | @@ -162,7 +162,7 @@ do_compile_append_raspberrypi3-64() {
}
do_install_prepend() {
- install -d ${D}/lib/firmware
+ install -d ${D}${nonarch_base_libdir}/firmware
}
do_deploy_append() {
|
BUMP pymongocrypt 1.1.1 | # See the License for the specific language governing permissions and
# limitations under the License.
-__version__ = '1.1.1.dev0'
+__version__ = '1.1.1'
_MIN_LIBMONGOCRYPT_VERSION = '1.2.0'
|
u3: optimizes +stir jet, reducing incremental allocations in fold | u3_noun tub)
{
// lef: list of successful [fel] parse results
- // wag: initial accumulator
+ // wag: initial accumulator (deconstructed)
//
- u3_noun wag, lef = u3_nul;
+ u3_noun lef = u3_nul;
+ u3_noun p_wag, puq_wag, quq_wag;
{
u3_noun vex, p_vex, q_vex;
u3x_cell(vex, &p_vex, &q_vex);
}
- wag = u3nq(u3k(p_vex), u3_nul, u3k(rud), u3k(tub));
+ p_wag = u3k(p_vex);
+ puq_wag = u3k(rud);
+ quq_wag = u3k(tub);
u3z(vex);
u3j_gate_lose(&fel_u);
}
// fold [lef] into [wag] by way of [raq]
//
if ( u3_nul != lef ) {
- u3_noun p_vex, puq_vex, gaw, p_wag, puq_wag, quq_wag;
+ u3_noun p_vex, puq_vex;
u3_noun i, t = lef;
u3j_site raq_u;
u3j_gate_prep(&raq_u, u3k(raq));
do {
u3x_cell(t, &i, &t);
u3x_qual(i, &p_vex, 0, &puq_vex, 0);
- u3x_qual(wag, &p_wag, 0, &puq_wag, &quq_wag);
- gaw = u3nq(_last(p_vex, p_wag),
- u3_nul,
- u3j_gate_slam(&raq_u,
- u3nc(u3k(puq_vex), u3k(puq_wag))),
- u3k(quq_wag));
- u3z(wag);
- wag = gaw;
+ {
+ u3_noun p_gaw = _last(p_vex, p_wag);
+ u3z(p_wag);
+ p_wag = p_gaw;
+ }
+
+ puq_wag = u3j_gate_slam(&raq_u, u3nc(u3k(puq_vex), puq_wag));
} while ( u3_nul != t );
u3z(lef);
u3j_gate_lose(&raq_u);
}
- return wag;
+ return u3nq(p_wag, u3_nul, puq_wag, quq_wag);
}
u3_noun
|
Badge bugfix | @@ -88,6 +88,7 @@ def draw_badge():
# Replace with drawing an image
display.pen(15)
+ display.thickness(1)
display.rectangle(WIDTH - IMAGE_WIDTH, 0, IMAGE_WIDTH, HEIGHT)
# Draw a border around the image
|
Update usage docs for client authentication | @@ -314,7 +314,7 @@ supported status request type is OCSP, **S2N_STATUS_REQUEST_OCSP**.
```c
-typedef enum { S2N_CERT_AUTH_NONE, S2N_CERT_AUTH_REQUIRED } s2n_cert_auth_type;
+typedef enum { S2N_CERT_AUTH_NONE, S2N_CERT_AUTH_REQUIRED, S2N_CERT_AUTH_OPTIONAL } s2n_cert_auth_type;
```
**s2n_cert_auth_type** is used to declare what type of client certificiate authentication to use.
Currently the default for s2n is for neither the server side or the client side to use Client (aka Mutual) authentication.
@@ -649,9 +649,11 @@ Client Auth Related API's are not recommended for normal users. Use of these API
int s2n_config_set_client_auth_type(struct s2n_config *config, s2n_cert_auth_type cert_auth_type);
int s2n_connection_set_client_auth_type(struct s2n_connection *conn, s2n_cert_auth_type cert_auth_type);
```
-Sets whether or not a Client Certificate should be required to complete the TLS Connection.
-If this is set to **S2N_CERT_AUTH_REQUIRED** then a **verify_cert_trust_chain_fn** callback should be provided as well since the current
-default is for s2n to accept all RSA Certs on the client side, and deny all certs on the server side.
+Sets whether or not a Client Certificate should be required to complete the TLS Connection. If this is set to
+**S2N_CERT_AUTH_OPTIONAL** the server will request a client certificate but allow the client to not provide one.
+If this is set to **S2N_CERT_AUTH_REQUIRED** or **S2N_CERT_AUTH_OPTIONAL** then a **verify_cert_trust_chain_fn** callback should be provided as well since the current
+default is for s2n to accept all RSA Certs on the client side, and deny all certs on the server side. Rejecting a
+client certificate when using **S2N_CERT_AUTH_OPTIONAL** will terminate the handshake.
### verify_cert_trust_chain_fn
|
Remove some logically dead code
Found by coverity. This is an artifact left over from the original
decaf import which generated the source code for different curves. For
curve 448 this is dead. | @@ -253,7 +253,6 @@ c448_error_t c448_ed448_verify(
curve448_point_decode_like_eddsa_and_mul_by_ratio(pk_point, pubkey);
curve448_scalar_t challenge_scalar;
curve448_scalar_t response_scalar;
- unsigned int c;
if (C448_SUCCESS != error)
return error;
@@ -291,9 +290,6 @@ c448_error_t c448_ed448_verify(
&signature[EDDSA_448_PUBLIC_BYTES],
EDDSA_448_PRIVATE_BYTES);
- for (c = 1; c < C448_EDDSA_DECODE_RATIO; c <<= 1)
- curve448_scalar_add(response_scalar, response_scalar, response_scalar);
-
/* pk_point = -c(x(P)) + (cx + k)G = kG */
curve448_base_double_scalarmul_non_secret(pk_point,
response_scalar,
|
debian changelog for v0.10.0 tag | +bcc (0.10.0-1) unstable; urgency=low
+
+ * Support for kernel up to 5.1
+ * corresponding libbpf submodule release is v0.0.3
+ * support for reading kernel headers from /proc
+ * libbpf.{a,so} renamed to libcc_bpf.{a,so}
+ * new common options for some tools
+ * new tool: drsnoop
+ * s390 USDT support
+
+ -- Brenden Blanco <[email protected]> Tue, 28 May 2019 17:00:00 +0000
+
bcc (0.9.0-1) unstable; urgency=low
* Adds support for BTF
|
Rename SceJsonError to SceJsonErrorCode
Necessary change to conform to the rest of the SDK. | /**
* Enumerator of different errors in this library.
*/
-typedef enum SceJsonError
+typedef enum SceJsonErrorCode
{
/**
* The module has not been initialised.
@@ -42,7 +42,7 @@ typedef enum SceJsonError
* Invalid character in the parsed data.
*/
SCE_JSON_PARSER_ERROR_INVALID_TOKEN = 0x80920101,
-} SceJsonError;
+} SceJsonErrorCode;
namespace sce
{
|
Added a number of explicit casts to avoid compiler warnings | @@ -467,9 +467,9 @@ int http_SendMessage(SOCKINFO *info, int *TimeOut, const char *fmt, ...)
if (Instr->ReadSendSize >= 0)
amount_to_be_read = Instr->ReadSendSize;
else
- amount_to_be_read = Data_Buf_Size;
- if (amount_to_be_read < WEB_SERVER_BUF_SIZE)
- Data_Buf_Size = amount_to_be_read;
+ amount_to_be_read = (off_t)Data_Buf_Size;
+ if (amount_to_be_read < (off_t)WEB_SERVER_BUF_SIZE)
+ Data_Buf_Size = (size_t)amount_to_be_read;
ChunkBuf = malloc((size_t)
(Data_Buf_Size + CHUNK_HEADER_SIZE +
CHUNK_TAIL_SIZE));
@@ -504,7 +504,7 @@ int http_SendMessage(SOCKINFO *info, int *TimeOut, const char *fmt, ...)
while (amount_to_be_read) {
if (Instr) {
int nr;
- size_t n = amount_to_be_read >= Data_Buf_Size ?
+ size_t n = amount_to_be_read >= (off_t)Data_Buf_Size ?
Data_Buf_Size : (size_t)amount_to_be_read;
if (Instr->IsVirtualFile) {
nr = virtualDirCallback.read(Fp, file_buf, n, Instr->Cookie);
@@ -512,10 +512,10 @@ int http_SendMessage(SOCKINFO *info, int *TimeOut, const char *fmt, ...)
} else {
num_read = fread(file_buf, (size_t)1, n, Fp);
}
- amount_to_be_read -= num_read;
+ amount_to_be_read -= (off_t)num_read;
if (Instr->ReadSendSize < 0) {
/* read until close */
- amount_to_be_read = Data_Buf_Size;
+ amount_to_be_read = (off_t)Data_Buf_Size;
}
} else {
num_read = fread(file_buf, (size_t)1, Data_Buf_Size, Fp);
|
netutils/cjson: fix unpackage error when local gerrit does not exist | @@ -55,6 +55,7 @@ $(CJSON_TARBALL):
$(CJSON_UNPACKNAME): $(CJSON_TARBALL)
@echo "Unpacking: $(CJSON_TARBALL) -> $(CJSON_UNPACKNAME)"
$(Q) $(UNPACK) $(CJSON_TARBALL)
+ $(Q) mv cJSON-$(CJSON_VERSION) $(CJSON_UNPACKNAME)
$(Q) touch $(CJSON_UNPACKNAME)
endif
|
feat(venachain):just for local debug | #CC := $(CURDIR)/../../../build/usr/bin/arm-linux-gcc
#AR := $(CURDIR)/../../../build/usr/bin/arm-linux-ar
+CC := gcc
+AR := ar
+
# Commands
BOAT_RM := rm -rf
BOAT_MKDIR := mkdir
|
chat: removed console.log | @@ -188,7 +188,6 @@ export class ChatInput extends Component {
}
render() {
- console.log('hi');
const { props, state } = this;
const color = props.ownerContact
|
Fix os.outputof for lua 5.3.4 | local pipe = io.popen(cmd .. " 2>&1")
local result = pipe:read('*a')
- local b, exitcode = pipe:close()
- if not b then
- exitcode = -1
- end
-
+ local success, what, code = pipe:close()
+ if success then
-- chomp trailing newlines
if result then
result = string.gsub(result, "[\r\n]+$", "")
end
- return result, exitcode
+ return result, code, what
+ else
+ return nil, code, what
+ end
end
|
Only output metric watch types if enabled. | @@ -2279,6 +2279,7 @@ createMetricWatchArrayJson(config_t* cfg)
metric_watch_t category;
for(category = CFG_MTC_FS; category <= CFG_MTC_STATSD; ++category) {
cJSON* item;
+ if (!cfgMtcWatchEnable(cfg, category)) continue;
if (!(item = createMetricWatchObjectJson(cfg, mtcWatchTypeMap[category].str))) goto err;
cJSON_AddItemToArray(root, item);
}
|
Removed unused LOG_DEBUG from parser.c. | @@ -2489,8 +2489,6 @@ process_log (GLogItem * logitem) {
if (logitem->ignorelevel != IGNORE_LEVEL_REQ) {
count_valid (logitem->numdate);
}
- LOG_DEBUG (("\n\n"));
-
}
/* Process a line from the log and store it accordingly taking into
|
Updated Readme Appveyor Badge [default dev branch] | @@ -2,7 +2,7 @@ EPANET {#epanet-readme}
======
## Build Status
-[](https://ci.appveyor.com/project/OpenWaterAnalytics/epanet/branch/dev)
+[](https://ci.appveyor.com/project/OpenWaterAnalytics/epanet)
[](https://travis-ci.org/OpenWaterAnalytics/EPANET)
The EPANET Library is a pressurized pipe network hydraulic and water quality analysis toolkit written in C.
@@ -16,6 +16,3 @@ A step-by-step tutorial on how to contribute to EPANET using GitHub is also [ava
__Note:__ This repository is not affiliated with, or endorsed by, the USEPA. For the last "official" release of EPANET (2.00.12 UI and Toolkit) please go to the [EPA's GitHub repo](https://github.com/USEPA/Water-Distribution-Network-Model) or [the USEPA website](http://www2.epa.gov/water-research/epanet). It is also not the graphical user interface version. This is the hydraulic and water quality solver engine.
However, if you are interested in extending EPANET for academic, personal, or commercial use, then you've come to the right place. For community discussion, FAQ, and roadmapping of the project, go to the [Community Forum](http://community.wateranalytics.org/category/epanet).
\ No newline at end of file
-
-testchange
-2
\ No newline at end of file
|
now we use WPA-PMKID-PBKDF2 and WPA-PMKID-PMK | @@ -14,7 +14,7 @@ Support for John the Ripper hash-modes: WPAPSK-PMK, PBKDF2-HMAC-SHA1, chap, netn
After capturing, upload the "uncleaned" cap here
(https://wpa-sec.stanev.org/?submit) to see if your ap or the client is vulnerable
-by using common wordlists. Convert the cap to hccapx and/or to WPA*-PMKID-PBKDF2 hashline (16800)
+by using common wordlists. Convert the cap to hccapx and/or to WPA-PMKID-PBKDF2 hashline (16800)
and check if wlan-key or plainmasterkey was transmitted unencrypted.
|
DOCS: gpcrondump creates table list when table and schema filters are specified. From 4.3.x | status file for SQL execution errors and displays a warning if an error is found. The
default location of the backup status files are in the
<codeph>db_dumps/<varname>date</varname>/</codeph> directory.</p>
+ <p>If you specify an option that includes or excludes tables or schemas, such as
+ <codeph>-t</codeph>, <codeph>-T</codeph>, <codeph>-s</codeph>, or <codeph>-S</codeph>, the
+ schema qualified names of the tables that are backed up are listed in the file
+ <codeph>gp_dump_<varname>timestamp</varname>_table</codeph>. The file is stored in the
+ backup directory of the master segment. </p>
<p><codeph>gpcrondump</codeph> allows you to schedule routine backups of a Greenplum database
using <codeph>cron</codeph> (a scheduling utility for UNIX operating systems). Cron jobs
that call <codeph>gpcrondump</codeph> should be scheduled on the master host.</p>
|
interface: fix show_or_clear_hw_interfaces
Type: fix
Fixes: | @@ -77,7 +77,7 @@ show_or_clear_hw_interfaces (vlib_main_t * vm,
int i, verbose = -1, show_bond = 0;
if (!unformat_user (input, unformat_line_input, line_input))
- return 0;
+ goto skip_unformat;
while (unformat_check_input (line_input) != UNFORMAT_END_OF_INPUT)
{
@@ -117,6 +117,7 @@ show_or_clear_hw_interfaces (vlib_main_t * vm,
unformat_free (line_input);
+skip_unformat:
/* Gather interfaces. */
if (vec_len (hw_if_indices) == 0)
pool_foreach (hi, im->hw_interfaces)
|
Change to debug log for http disconnect in OTA | @@ -1201,7 +1201,7 @@ OTA_Err_t _AwsIotOTA_CleanupData_HTTP( OTA_AgentContext_t * pAgentCtx )
if( IOT_HTTPS_OK != httpsStatus )
{
- IotLogError( "Failed to disconnect from S3 server. Error code: %d.", httpsStatus );
+ IotLogDebug( "Failed to disconnect from S3 server. Error code: %d.", httpsStatus );
}
_httpDownloader.httpConnection.connectionHandle = NULL;
|
hv: bug fix on synchronizing with APs
Using eax will truncate the high 32bit part of 64bit virtual address.
And the type of sync is unsigned long, so using rbx instead of ebx. | @@ -732,11 +732,11 @@ static void print_hv_banner(void)
static void pcpu_sync_sleep(unsigned long *sync, int mask_bit)
{
- int wake_sync = (1 << mask_bit);
+ uint64_t wake_sync = (1UL << mask_bit);
if (get_monitor_cap()) {
/* Wait for the event to be set using monitor/mwait */
- asm volatile ("1: cmpl %%ebx,(%%eax)\n"
+ asm volatile ("1: cmpq %%rbx,(%%rax)\n"
" je 2f\n"
" monitor\n"
" mwait\n"
@@ -748,7 +748,7 @@ static void pcpu_sync_sleep(unsigned long *sync, int mask_bit)
: "cc");
} else {
/* Wait for the event to be set using pause */
- asm volatile ("1: cmpl %%ebx,(%%eax)\n"
+ asm volatile ("1: cmpq %%rbx,(%%rax)\n"
" je 2f\n"
" pause\n"
" jmp 1b\n"
|
BugID:23881073: sync httpdns bug fix from genie | @@ -59,6 +59,10 @@ static inline void httpdns_resolv_unlock(void)
static inline void httpnds_resolv_lock_init(void)
{
+ if (resolv_lock.initted == PTHREAD_INITIALIZED_OBJ) {
+ return;
+ }
+
pthread_mutex_init(&resolv_lock, NULL);
}
@@ -433,7 +437,7 @@ struct addrinfo* httpdns_build_addrinfo(dns_cache_t *cache, int port)
int httpdns_prefetch(char * host_name, dns_cache_t ** cache)
{
- if((NULL == cache) || (NULL == host_name)) {
+ if((NULL == cache) || (NULL == host_name) || (NULL == resolv)) {
return 0;
}
httpdns_resolv_lock();
@@ -693,7 +697,6 @@ static int localdns_parse(char *host_name)
{
int ret = -1;
struct addrinfo hints;
- char portstr[16];
struct addrinfo * result = NULL;
dns_cache_t * cache = NULL;
@@ -708,6 +711,9 @@ static int localdns_parse(char *host_name)
}
cache = httpdns_build_cache(result, host_name);
+
+ freeaddrinfo(result);
+
if (!cache) {
HTTPDNS_ERR("%s httpdns_build_cache failed", __func__);
return -1;
|
Update default WAL segment size comment. | @@ -22,7 +22,8 @@ Page size can only be changed at compile time and is not known to be well-tested
/***********************************************************************************************************************************
Define default wal segment size
-Page size can only be changed at compile time and and is not known to be well-tested, so only the default page size is supported.
+Before PostgreSQL 11 WAL segment size could only be changed at compile time and is not known to be well-tested, so only the default
+WAL segment size is supported for versions below 11.
***********************************************************************************************************************************/
#define PG_WAL_SEGMENT_SIZE_DEFAULT ((unsigned int)(16 * 1024 * 1024))
|
DEV: Refactor APS indication | @@ -974,8 +974,11 @@ void DeRestPluginPrivate::apsdeDataIndicationDevice(const deCONZ::ApsDataIndicat
{
ResourceItem *item = r->itemForIndex(i);
DBG_Assert(item);
- if (item && item->ddfItemHandle() != DeviceDescription::Item::InvalidItemHandle)
+ if (!item)
{
+ continue;
+ }
+
ParseFunction_t parseFunction = item->parseFunction();
const auto &ddfItem = DDF_GetItem(item);
@@ -986,7 +989,16 @@ void DeRestPluginPrivate::apsdeDataIndicationDevice(const deCONZ::ApsDataIndicat
parseFunction = DA_GetParseFunction(ddfItem.parseParameters);
}
- if (parseFunction && parseFunction(r, item, ind, zclFrame, ddfItem.parseParameters))
+ if (!parseFunction)
+ {
+ if (!ddfItem.parseParameters.isNull())
+ {
+ DBG_Printf(DBG_INFO, "parse function for %s not found: %s\n", item->descriptor().suffix, qPrintable(ddfItem.parseParameters.toString()));
+ }
+ continue;
+ }
+
+ if (parseFunction(r, item, ind, zclFrame, ddfItem.parseParameters))
{
if (item->awake())
{
@@ -1007,11 +1019,6 @@ void DeRestPluginPrivate::apsdeDataIndicationDevice(const deCONZ::ApsDataIndicat
DB_StoreSubDeviceItem(r, item);
}
}
- else if (!parseFunction)
- {
- DBG_Printf(DBG_INFO, "parse function not found: %s\n", qPrintable(ddfItem.parseParameters.toString()));
- }
- }
}
}
|
Fix LEDs for configured modifier states | @@ -401,6 +401,14 @@ void sway_keyboard_configure(struct sway_keyboard *keyboard) {
}
if (locked_mods) {
wlr_keyboard_notify_modifiers(wlr_device->keyboard, 0, 0, locked_mods, 0);
+ uint32_t leds = 0;
+ for (uint32_t i = 0; i < WLR_LED_COUNT; ++i) {
+ if (xkb_state_led_index_is_active(wlr_device->keyboard->xkb_state,
+ wlr_device->keyboard->led_indexes[i])) {
+ leds |= (1 << i);
+ }
+ }
+ wlr_keyboard_led_update(wlr_device->keyboard, leds);
}
if (input_config && input_config->repeat_delay != INT_MIN
|
Add a doxygen ci check | @@ -48,6 +48,11 @@ jobs:
echo "Excerpt of the diff follows"
git diff | head -n 20
fi
+ - run: >
+ if doxygen Doxyfile | grep 'warning: '; then
+ echo "Doxygen warning (see above) -- misformatted docs?"
+ exit 1
+ fi
- run: >
cd doc &&
pip3 install -r source/requirements.txt &&
|
rpz-triggers, spelling fix in comment. | @@ -1954,7 +1954,7 @@ rpz_apply_nsip_trigger(struct module_qstate* ms, struct rpz* r,
break;
case RPZ_TCP_ONLY_ACTION:
/* basically a passthru here but the tcp-only will be
- * honored before the query gets send */
+ * honored before the query gets sent. */
ms->respip_action_info->action = respip_truncate;
ret = NULL;
break;
@@ -2003,7 +2003,7 @@ rpz_apply_nsdname_trigger(struct module_qstate* ms, struct rpz* r,
break;
case RPZ_TCP_ONLY_ACTION:
/* basically a passthru here but the tcp-only will be
- * honored before the query gets send */
+ * honored before the query gets sent. */
ms->respip_action_info->action = respip_truncate;
ret = NULL;
break;
|
build: exclude target dependencies | @@ -181,7 +181,7 @@ option(MSGPACK_ENABLE_CXX OFF)
option(MSGPACK_ENABLE_SHARED OFF)
option(MSGPACK_BUILD_TESTS OFF)
option(MSGPACK_BUILD_EXAMPLES OFF)
-add_subdirectory(lib/msgpack-c-0b7cabd)
+add_subdirectory(lib/msgpack-c-0b7cabd EXCLUDE_FROM_ALL)
# Lib: build the core libraries used by Fluent-Bit
FLB_DEFINITION(JSMN_PARENT_LINKS)
@@ -198,7 +198,7 @@ macro(MK_SET_OPTION option value)
endmacro()
MK_SET_OPTION(MK_SYSTEM_MALLOC ON)
-add_subdirectory(lib/monkey/mk_core)
+add_subdirectory(lib/monkey/mk_core EXCLUDE_FROM_ALL)
# SSL/TLS: add encryption support
if(FLB_OUT_TD)
@@ -210,7 +210,7 @@ if(FLB_TLS)
option(ENABLE_TESTING OFF)
option(ENABLE_PROGRAMS OFF)
option(INSTALL_MBEDTLS_HEADERS OFF)
- add_subdirectory(lib/mbedtls-2.4.2)
+ add_subdirectory(lib/mbedtls-2.4.2 EXCLUDE_FROM_ALL)
include_directories(lib/mbedtls-2.4.2/include)
endif()
|
docs: adjust the list | ### Blockchain Official Websites
+ [Ethereum](https://ethereum.org/)
+ [Ganache: an Ethereum Simulator](https://www.trufflesuite.com/truffle/)
++ [PlatON](https://www.platon.network/)
+ [PlatON Enterprise](https://github.com/PlatONEnterprise/)
-+ [Hyperledger Fabric](https://www.hyperledger.org/use/fabric)
-+ [Hyperledger Fabric Docs](https://hyperledger-fabric.readthedocs.io/)
+ [FISCO BCOS](http://fisco-bcos.org/)
+ [FISCO BCOS Github](https://github.com/FISCO-BCOS)
-+ [PlatON](https://www.platon.network/)
++ [Hyperledger Fabric](https://www.hyperledger.org/use/fabric)
++ [Hyperledger Fabric Docs](https://hyperledger-fabric.readthedocs.io/)
# Supported Module List
|
Bump CMake project version to 3.1.0 | @@ -24,7 +24,7 @@ if(MSVC)
add_compile_options("/wd4324") # Disable structure was padded due to alignment specifier
endif()
-project(astcencoder VERSION 3.0.0)
+project(astcencoder VERSION 3.1.0)
set(CMAKE_CXX_STANDARD 14)
set(CMAKE_CXX_STANDARD_REQUIRED ON)
|
upadted cmake | @@ -7,7 +7,7 @@ python:
addons:
apt:
sources:
- - george-edison55-precise-backports
+ - george-edison55/cmake-3.x
packages:
- fftw3
- fftw3-dev
@@ -16,12 +16,12 @@ addons:
- swig
- python-numpy
- python3-numpy
- - cmake-data
- cmake
install: true
script:
+ - cmake --version
- python setup.py build
- python setup.py test
- sudo make -Cbuild install
|
changes: fix woring that mentions SHA* one shot functions are deprecated | @@ -690,8 +690,8 @@ breaking changes, and mappings for the large list of deprecated functions.
*Paul Dale*
- * The low-level MD2, MD4, MD5, MDC2, RIPEMD160, SHA1, SHA224, SHA256,
- SHA384, SHA512 and Whirlpool digest functions have been deprecated.
+ * The low-level MD2, MD4, MD5, MDC2, RIPEMD160 and Whirlpool digest
+ functions have been deprecated.
*Paul Dale and David von Oheimb*
|
Add PhGetThreadPriorityBoost | @@ -806,6 +806,51 @@ PhSetThreadPagePriority(
);
}
+FORCEINLINE
+NTSTATUS
+PhGetThreadPriorityBoost(
+ _In_ HANDLE ThreadHandle,
+ _Out_ PBOOLEAN PriorityBoost
+ )
+{
+ NTSTATUS status;
+ ULONG priorityBoost;
+
+ status = NtQueryInformationThread(
+ ThreadHandle,
+ ThreadPriorityBoost,
+ &priorityBoost,
+ sizeof(ULONG),
+ NULL
+ );
+
+ if (NT_SUCCESS(status))
+ {
+ *PriorityBoost = !!priorityBoost;
+ }
+
+ return status;
+}
+
+FORCEINLINE
+NTSTATUS
+PhSetThreadPriorityBoost(
+ _In_ HANDLE ThreadHandle,
+ _In_ BOOLEAN PriorityBoost
+ )
+{
+ ULONG priorityBoost;
+
+ priorityBoost = PriorityBoost ? 1 : 0;
+
+ return NtSetInformationThread(
+ ThreadHandle,
+ ThreadPriorityBoost,
+ &priorityBoost,
+ sizeof(ULONG)
+ );
+}
+
/**
* Gets a thread's cycle count.
*
|
add function to remap kernel image to another address | @@ -212,6 +212,22 @@ static void map_heap_pages(page_directory_t* dir) {
}
}
+static void vmm_remap_kernel(page_directory_t* dir, uint32_t dest_virt_addr) {
+ boot_info_t* info = boot_info_get();
+ uint32_t kernel_base = info->kernel_image_start;
+ uint32_t kernel_size = info->kernel_image_size;
+
+ for (int i = 0; i < kernel_size; i += PAGING_PAGE_SIZE) {
+ //figure out the address of the physical frame
+ uint32_t frame_addr = i + kernel_base;
+
+ //map this kernel frame to a virtual page
+ uint32_t dest_page_addr = dest_virt_addr + i;
+ page_t* dest_page = vmm_get_page_for_virtual_address(dir, dest_page_addr);
+ vmm_map_page_to_frame(dest_page, frame_addr);
+ }
+}
+
void paging_install() {
printf_info("Initializing paging...");
|
Update snapcraft builds. | name: ipp
-base: core18
+base: core20
version: 1.0b1
summary: IPP Sample Implementation
description: |
@@ -11,42 +11,42 @@ grade: stable
confinement: strict
icon: server/printer.png
+architectures:
+ - build-on: [ arm64, armhf, amd64 ]
+ run-on: [ arm64, armhf, amd64 ]
+
apps:
- server:
- command: sbin/server
+ ipp3dprinter:
+ command: bin/ipp3dprinter
plugs: [avahi-observe, home, network, network-bind]
- eveprinter:
- command: bin/eveprinter
+ ippdoclint:
+ command: bin/ippdoclint
+ plugs: [home]
+ ippeveprinter:
+ command: bin/ippeveprinter
plugs: [avahi-observe, home, network, network-bind]
- find:
- command: bin/find
+ ippfind:
+ command: bin/ippfind
plugs: [avahi-observe, home, network]
- proxy:
- command: sbin/proxy
+ ippproxy:
+ command: sbin/ippproxy
plugs: [avahi-observe, home, network]
- tool:
- command: bin/tool
+ ippserver:
+ command: sbin/ippserver
+ plugs: [avahi-observe, home, network, network-bind]
+ ipptool:
+ command: bin/ipptool
+ plugs: [home, network]
+ ipptransform:
+ command: bin/ipptransform
+ plugs: [home, network]
+ ipptransform3d:
+ command: bin/ipptransform3d
plugs: [home, network]
parts:
- mupdf:
- plugin: make
- make-install-var: prefix
- make-parameters: [HAVE_X11=no, HAVE_GLFW=no, HAVE_GLUT=no, USE_SYSTEM_LIBS=yes]
- source: https://mupdf.com/downloads/archive/mupdf-1.12.0-source.tar.gz
- prime:
- - -bin/mu*
- - -include/mu*
- - -lib/libmu*
- - -share/doc/mupdf
- - -share/man/man1/mu*
- build-packages: [libfreetype6-dev, libharfbuzz-dev, libjbig2dec0-dev, libjpeg-dev, liblcms2-dev, libopenjp2-7-dev, libpng-dev, zlib1g-dev]
-
main:
- after: [mupdf]
plugin: autotools
- configflags: [--with-name-prefix=]
source: .
- build-packages: [cura-engine, libavahi-client-dev, libfreetype6-dev, libgnutls28-dev, libharfbuzz-dev, libjbig2dec0-dev, libjpeg-dev, liblcms2-dev, libopenjp2-7-dev, libpng-dev, zlib1g-dev]
- stage-packages: [libavahi-client3, libjpeg-turbo8, libopenjp2-7]
-
+ build-packages: [cura-engine, libavahi-client-dev, libjpeg-dev, libpam-dev, libpng-dev, libssl-dev, libusb-1.0-0-dev, zlib1g-dev]
+ stage-packages: [libavahi-client3, libjpeg-turbo8]
|
fix: fix test_002InitWallet_0003SetChainIdSuccess
issue | @@ -628,6 +628,7 @@ START_TEST(test_002InitWallet_0003SetChainIdSuccess)
BoatEthWallet *wallet_ptr = BoatMalloc(sizeof(BoatEthWallet));
BoatEthWalletConfig wallet = get_ethereum_wallet_settings();
+ ck_assert_ptr_ne(wallet_ptr, NULL);
/* 1. execute unit test */
rtnVal = BoatEthWalletSetChainId(wallet_ptr, wallet.chain_id);
|
Detect `main.py` as entrypoint for modules | @@ -31,6 +31,8 @@ def main():
return
module_path = spec.origin
+ __main__path = os.path.dirname(module_path)+"/main.py"
+ if not os.path.isfile(__main__path):
__main__path = os.path.dirname(module_path)+"/__main__.py"
if os.path.isfile(__main__path) and os.path.basename(module_path) == "__init__.py":
module_path = __main__path
|
ports/stm32: Fix HAL MSP deinit functions. | @@ -462,6 +462,16 @@ void HAL_DCMI_MspDeInit(DCMI_HandleTypeDef* hdcmi)
for (int i=0; i<NUM_DCMI_PINS; i++) {
HAL_GPIO_DeInit(dcmi_pins[i].port, dcmi_pins[i].pin);
}
+
+ #if defined(DCMI_RESET_PIN)
+ HAL_GPIO_DeInit(DCMI_RESET_PORT, DCMI_RESET_PIN);
+ #endif
+ #if defined(DCMI_FSYNC_PIN)
+ HAL_GPIO_DeInit(DCMI_FSYNC_PORT, DCMI_FSYNC_PIN);
+ #endif
+ #if defined(DCMI_PWDN_PIN)
+ HAL_GPIO_DeInit(DCMI_PWDN_PORT, DCMI_PWDN_PIN);
+ #endif
}
@@ -521,7 +531,7 @@ void HAL_SPI_MspInit(SPI_HandleTypeDef *hspi)
#endif
}
-void HAL_SPI_MspDeinit(SPI_HandleTypeDef *hspi)
+void HAL_SPI_MspDeInit(SPI_HandleTypeDef *hspi)
{
}
@@ -744,7 +754,7 @@ void HAL_DAC_MspInit(DAC_HandleTypeDef *hdac)
#endif
}
-void HAL_DAC_MspDeinit(DAC_HandleTypeDef *hdac)
+void HAL_DAC_MspDeInit(DAC_HandleTypeDef *hdac)
{
#if defined(OMV_SPI_LCD_BL_DAC)
if (hdac->Instance == OMV_SPI_LCD_BL_DAC) {
|
Update Visual Studio project file | <GenerateDebugInformation>true</GenerateDebugInformation>
<AdditionalDependencies>libxml2.lib;libpng16.lib;jpeg.lib;freetype271.lib;zlib.lib;kernel32.lib;user32.lib;gdi32.lib;</AdditionalDependencies>
<AdditionalLibraryDirectories>$(SolutionDir)..\..\vendor\lib</AdditionalLibraryDirectories>
- <IgnoreSpecificDefaultLibraries>LIBCMT;LIBCMTD;MSVCRT;</IgnoreSpecificDefaultLibraries>
+ <IgnoreSpecificDefaultLibraries>LIBCMT</IgnoreSpecificDefaultLibraries>
<LinkStatus>false</LinkStatus>
+ <ShowProgress>NotSet</ShowProgress>
</Link>
<ResourceCompile>
<Culture>0x0804</Culture>
<AdditionalDependencies>libxml2.lib;libpng16.lib;jpeg.lib;freetype271.lib;zlib.lib;kernel32.lib;user32.lib;gdi32.lib;</AdditionalDependencies>
<AdditionalLibraryDirectories>$(SolutionDir)..\..\vendor\lib</AdditionalLibraryDirectories>
<IgnoreSpecificDefaultLibraries>LIBCMT</IgnoreSpecificDefaultLibraries>
+ <ShowProgress>NotSet</ShowProgress>
</Link>
<Lib>
<AdditionalDependencies>libpng16.lib;libxml2.lib;jpeg.lib;freetype271.lib;zlib.lib;</AdditionalDependencies>
|
Avoid integer overflow when starting HB timer | @@ -2383,14 +2383,19 @@ sctp_timer_start(int t_type, struct sctp_inpcb *inp, struct sctp_tcb *stcb,
}
rndval = sctp_select_initial_TSN(&inp->sctp_ep);
jitter = rndval % to_ticks;
- if (jitter >= (to_ticks >> 1)) {
- to_ticks = to_ticks + (jitter - (to_ticks >> 1));
+ to_ticks >>= 1;
+ if (jitter < (UINT32_MAX - to_ticks)) {
+ to_ticks += jitter;
} else {
- to_ticks = to_ticks - jitter;
+ to_ticks = UINT32_MAX;
}
if (!(net->dest_state & SCTP_ADDR_UNCONFIRMED) &&
!(net->dest_state & SCTP_ADDR_PF)) {
+ if (net->heart_beat_delay < (UINT32_MAX - to_ticks)) {
to_ticks += net->heart_beat_delay;
+ } else {
+ to_ticks = UINT32_MAX;
+ }
}
/*
* Now we must convert the to_ticks that are now in
|
adc: fix common calls | @@ -47,10 +47,12 @@ void adc_init(void) {
LL_ADC_CommonInitTypeDef adc_common_init;
adc_common_init.CommonClock = LL_ADC_CLOCK_SYNC_PCLK_DIV2;
- LL_ADC_CommonInit(ADC1, &adc_common_init);
+ LL_ADC_CommonInit(ADC1_COMMON, &adc_common_init);
LL_ADC_REG_InitTypeDef adc_reg_init;
adc_reg_init.TriggerSource = LL_ADC_REG_TRIG_SOFTWARE;
+ adc_reg_init.SequencerLength = LL_ADC_REG_SEQ_SCAN_DISABLE;
+ adc_reg_init.SequencerDiscont = LL_ADC_REG_SEQ_DISCONT_DISABLE;
adc_reg_init.ContinuousMode = LL_ADC_REG_CONV_SINGLE;
adc_reg_init.DMATransfer = LL_ADC_REG_DMA_TRANSFER_NONE;
LL_ADC_REG_Init(ADC1, &adc_reg_init);
@@ -61,7 +63,7 @@ void adc_init(void) {
adc_init.SequencersScanMode = LL_ADC_SEQ_SCAN_DISABLE;
LL_ADC_Init(ADC1, &adc_init);
- LL_ADC_SetCommonPathInternalCh(ADC1, LL_ADC_PATH_INTERNAL_VREFINT);
+ LL_ADC_SetCommonPathInternalCh(ADC1_COMMON, LL_ADC_PATH_INTERNAL_VREFINT);
LL_ADC_Enable(ADC1);
LL_ADC_REG_SetSequencerRanks(ADC1, LL_ADC_REG_RANK_1, LL_ADC_CHANNEL_VREFINT);
|
chat-fe: correct light mode autocomplete styling | @@ -10,8 +10,12 @@ import { uuid, uxToHex } from '/lib/util';
const DEFAULT_INPUT_HEIGHT = 28;
+
function getAdvance(a, b) {
let res = '';
+ if(!a) {
+ return b;
+ }
for (let i = 0; i < Math.min(a.length, b.length); i++) {
if (a[i] !== b[i]) {
return res;
@@ -29,26 +33,23 @@ function ChatInputSuggestions({ suggestions, onSelect, selected }) {
left: '48px'
}}
className={
- 'absolute bg-white bg-gray0-d ' +
- 'w7 pv3 z-1 mt1 ba b--white-d overflow-y-scroll'
+ 'absolute black white-d bg-white bg-gray0-d ' +
+ 'w7 pv1 z-1 mt1 ba b--gray1-d b--gray4'
}
>
{suggestions.map(ship => (
<div
className={cn(
- 'list mono f8 pv1 ph3 pointer' + ' hover-bg-gray4 relative',
+ 'list mono f8 pv1 ph3 pointer' + ' hover-bg-gray4 relative bb b--gray1-d b--gray4',
{
'white-d': ship !== selected,
'black-d': ship === selected,
'bg-gray0-d': ship !== selected,
- 'bg-gray1-d': ship === selected
+ 'bg-white': ship !== selected,
+ 'bg-gray1-d': ship === selected,
+ 'bg-gray5': ship === selected
}
)}
- style={{
- borderBottom: '1px solid #4d4d4d',
- zIndex: '99',
- fontFamily: 'monospace'
- }}
key={ship}
>
<Sigil
@@ -57,7 +58,7 @@ function ChatInputSuggestions({ suggestions, onSelect, selected }) {
color="#000000"
classes="mix-blend-diff v-mid"
/>
- <span className="v-mid ml2 mw5 truncate dib mix-blend-diff">
+ <span className="v-mid ml2 mw5 truncate dib">
{'~' + ship}
</span>
</div>
@@ -139,21 +140,20 @@ export class ChatInput extends Component {
const match = /~([a-z\-]*)$/.exec(message);
if (!match) {
- return null;
+ this.setState({ patpSuggestions: [] })
}
- const envelopes = ['hastuc', 'hastuc-dibtux', 'hasfun'];
- // const suggestions = _.chain(props.envelopes)
- // .map("author")
- // .uniq()
- const suggestions = _.chain(envelopes)
+ const suggestions = _.chain(this.props.envelopes)
+ .map("author")
+ .uniq()
.filter(s => s.startsWith(match[1]))
+ .take(3)
.value();
const advance = _.chain(suggestions)
.map(s => s.replace(match[0], ''))
- .reduce(getAdvance)
+ .reduce(getAdvance, )
.value();
let newState = {
|
Rename args --> opts
To avoid confusion with the similarly-named "arg" | @@ -19,7 +19,7 @@ local pallenec = {}
local compiler_name = arg[0]
-- Command-line options
-local args
+local opts
do
local p = argparse("pallenec", "Pallene compiler")
p:argument("source_file", "File to compile")
@@ -39,20 +39,20 @@ do
p:option("-o --output", "Output file path")
- args = p:parse()
+ opts = p:parse()
end
local function compile(in_ext, out_ext)
- local ok, errs = driver.compile(compiler_name, args.O, in_ext, out_ext, args.source_file,
- args.output)
+ local ok, errs = driver.compile(compiler_name, opts.O, in_ext, out_ext, opts.source_file,
+ opts.output)
if not ok then util.abort(table.concat(errs, "\n")) end
end
local function compile_up_to(stop_after)
- local input, err = driver.load_input(args.source_file)
+ local input, err = driver.load_input(opts.source_file)
if err then util.abort(err) end
- local out, errs = driver.compile_internal(args.source_file, input, stop_after, args.O)
+ local out, errs = driver.compile_internal(opts.source_file, input, stop_after, opts.O)
if not out then util.abort(table.concat(errs, "\n")) end
return out
@@ -64,10 +64,10 @@ local function do_print_ir()
end
function pallenec.main()
- if args.emit_c then compile("pln", "c")
- elseif args.emit_lua then compile("pln", "lua")
- elseif args.compile_c then compile("c" , "so")
- elseif args.print_ir then do_print_ir()
+ if opts.emit_c then compile("pln", "c")
+ elseif opts.emit_lua then compile("pln", "lua")
+ elseif opts.compile_c then compile("c" , "so")
+ elseif opts.print_ir then do_print_ir()
else --[[default]] compile("pln", "so")
end
end
|
Fix pyyaml windows build
Local fork at gpMgmt/bin/ext/yaml was removed by Unpack
it from gpMgmt/bin/pythonSrc/ext just like pygresql. | @@ -11,7 +11,8 @@ type nul > %GPDB_INSTALL_PATH%\bin\gppylib\__init__.py
copy ..\..\..\..\..\gpMgmt\bin\gppylib\gpversion.py %GPDB_INSTALL_PATH%\bin\gppylib\
perl -pi.bak -e "s,\$Revision\$,%VERSION%," %GPDB_INSTALL_PATH%\bin\gpload.py
copy ..\..\..\..\..\gpMgmt\bin\gpload.bat %GPDB_INSTALL_PATH%\bin
-copy ..\..\..\..\..\gpMgmt\bin\ext\yaml\* %GPDB_INSTALL_PATH%\lib\python\yaml
+for %%f in (..\..\..\..\..\gpMgmt\bin\pythonSrc\ext\PyYAML-*.tar.gz) do tar -xf %%f
+for /D %%d in (PyYAML-*) do copy %%d\lib\yaml\* %GPDB_INSTALL_PATH%\lib\python\yaml
perl -p -e "s,__VERSION_PLACEHOLDER__,%VERSION%," greenplum-clients.wxs > greenplum-clients-%VERSION%.wxs
candle.exe -nologo greenplum-clients-%VERSION%.wxs -out greenplum-clients-%VERSION%.wixobj -dSRCDIR=%GPDB_INSTALL_PATH% -dVERSION=%VERSION%
light.exe -nologo -sval greenplum-clients-%VERSION%.wixobj -out greenplum-clients-x86_64.msi
\ No newline at end of file
|
fix(draw): fix set_px_cb memory write overflow crash. | @@ -420,8 +420,6 @@ static void map_set_px(lv_color_t * dest_buf, lv_coord_t dest_stride, const lv_a
int32_t src_stride = lv_area_get_width(src_area);
- dest_buf += dest_stride * clip_area->y1 + clip_area->x1;
-
src_buf += src_stride * (clip_area->y1 - src_area->y1);
src_buf += (clip_area->x1 - src_area->x1);
@@ -430,7 +428,7 @@ static void map_set_px(lv_color_t * dest_buf, lv_coord_t dest_stride, const lv_a
if(mask == NULL) {
for(y = 0; y < clip_h; y++) {
- for(x = 0; x <= clip_w; x++) {
+ for(x = 0; x < clip_w; x++) {
disp->driver->set_px_cb(disp->driver, (void *)dest_buf, dest_stride, clip_area->x1 + x, clip_area->y1 + y, src_buf[x],
opa);
}
@@ -439,7 +437,7 @@ static void map_set_px(lv_color_t * dest_buf, lv_coord_t dest_stride, const lv_a
}
else {
for(y = 0; y < clip_h; y++) {
- for(x = 0; x <= clip_w; x++) {
+ for(x = 0; x < clip_w; x++) {
if(mask[x]) {
disp->driver->set_px_cb(disp->driver, (void *)dest_buf, dest_stride, clip_area->x1 + x, clip_area->y1 + y, src_buf[x],
(uint32_t)((uint32_t)opa * mask[x]) >> 8);
|
fix(demo): fix Wformat warning | @@ -902,7 +902,7 @@ void lv_demo_benchmark_run_scene(int_fast16_t scene_no)
scene_act = scene_no >> 1;
if(scenes[scene_act].create_cb) {
- lv_label_set_text_fmt(title, "%"LV_PRId32"/%d: %s%s", scene_act * 2 + (opa_mode ? 1 : 0), (dimof(scenes) * 2) - 2,
+ lv_label_set_text_fmt(title, "%"LV_PRId32"/%d: %s%s", scene_act * 2 + (opa_mode ? 1 : 0), (int)(dimof(scenes) * 2) - 2,
scenes[scene_act].name, opa_mode ? " + opa" : "");
if(opa_mode) {
lv_label_set_text_fmt(subtitle, "Result of \"%s\": %"LV_PRId32" FPS", scenes[scene_act].name,
@@ -950,7 +950,7 @@ static void scene_next_task_cb(lv_timer_t * timer)
if(scenes[scene_act].create_cb) {
lv_label_set_text_fmt(title, "%"LV_PRId32"/%d: %s%s", scene_act * 2 + (opa_mode ? 1 : 0),
- (dimof(scenes) * 2) - 2, scenes[scene_act].name, opa_mode ? " + opa" : "");
+ (int)(dimof(scenes) * 2) - 2, scenes[scene_act].name, opa_mode ? " + opa" : "");
if(opa_mode) {
lv_label_set_text_fmt(subtitle, "Result of \"%s\": %"LV_PRId32" FPS", scenes[scene_act].name,
scenes[scene_act].fps_normal);
|
CMSIS-DSP: Corrected wrong table in fft.cmake | @@ -224,7 +224,7 @@ if (CONFIGTABLE AND RFFT_F32_32)
endif()
if (CONFIGTABLE AND RFFT_F32_64)
- target_compile_definitions(${PROJECT} PUBLIC ARM_TABLE_REALCOEF_Q31)
+ target_compile_definitions(${PROJECT} PUBLIC ARM_TABLE_REALCOEF_F32)
# For cfft_radix4_init
target_compile_definitions(${PROJECT} PUBLIC ARM_TABLE_BITREV_1024)
target_compile_definitions(${PROJECT} PUBLIC ARM_TABLE_TWIDDLECOEF_F32_4096)
|
external/dhcpd_lwip: Fix pthread_create return check in dhcps | @@ -343,7 +343,7 @@ int dhcp_server_start(char *intf, dhcp_sta_joined_cb dhcp_join_cb)
return ERROR;
}
ret = pthread_create(&g_dhcpd_tid, &attr, _dhcpd_join_handler, (void *)data);
- if (ret < 0) {
+ if (ret != OK) {
free(data->intf);
free(data);
ndbg("dhcps create dhcp handler fail(%d) errno %d\n", ret, errno);
|
external/grpc: fix grpc thread name
fix grpc thread naming issue | @@ -117,7 +117,7 @@ int gpr_thd_new(gpr_thd_id* t, const char* thd_name,
free(a);
dec_thd_count();
} else {
- pthread_setname_np(thread_started, a->name);
+ pthread_setname_np(p, a->name);
}
*t = (gpr_thd_id)p;
return thread_started;
|
Fix TLSv1.3 alert handling
In TLSv1.3 we should ignore the severity level of an alert according to
the spec. | @@ -1497,6 +1497,7 @@ int ssl3_read_bytes(SSL *s, int type, int *recvd_type, unsigned char *buf,
if (SSL3_RECORD_get_type(rr) == SSL3_RT_ALERT) {
unsigned int alert_level, alert_descr;
+ int tls13_user_cancelled;
unsigned char *alert_bytes = SSL3_RECORD_get_data(rr)
+ SSL3_RECORD_get_off(rr);
PACKET alert;
@@ -1524,7 +1525,9 @@ int ssl3_read_bytes(SSL *s, int type, int *recvd_type, unsigned char *buf,
cb(s, SSL_CB_READ_ALERT, j);
}
- if (alert_level == SSL3_AL_WARNING) {
+ tls13_user_cancelled = SSL_IS_TLS13(s)
+ && alert_descr == SSL_AD_USER_CANCELLED;
+ if (alert_level == SSL3_AL_WARNING || tls13_user_cancelled) {
s->s3->warn_alert = alert_descr;
SSL3_RECORD_set_read(rr);
@@ -1535,19 +1538,20 @@ int ssl3_read_bytes(SSL *s, int type, int *recvd_type, unsigned char *buf,
return -1;
}
- if (alert_descr == SSL_AD_CLOSE_NOTIFY) {
- s->shutdown |= SSL_RECEIVED_SHUTDOWN;
- return 0;
- }
/*
* Apart from close_notify the only other warning alert in TLSv1.3
* is user_cancelled - which we just ignore.
*/
- if (SSL_IS_TLS13(s) && alert_descr != SSL_AD_USER_CANCELLED) {
- SSLfatal(s, SSL_AD_ILLEGAL_PARAMETER, SSL_F_SSL3_READ_BYTES,
- SSL_R_UNKNOWN_ALERT_TYPE);
- return -1;
+ if (tls13_user_cancelled)
+ goto start;
}
+
+ if (alert_descr == SSL_AD_CLOSE_NOTIFY
+ && (SSL_IS_TLS13(s) || alert_level == SSL3_AL_WARNING)) {
+ s->shutdown |= SSL_RECEIVED_SHUTDOWN;
+ return 0;
+ }
+
/*
* This is a warning but we receive it if we requested
* renegotiation and the peer denied it. Terminate with a fatal
@@ -1556,12 +1560,15 @@ int ssl3_read_bytes(SSL *s, int type, int *recvd_type, unsigned char *buf,
* future we might have a renegotiation where we don't care if
* the peer refused it where we carry on.
*/
- if (alert_descr == SSL_AD_NO_RENEGOTIATION) {
+ if (alert_descr == SSL_AD_NO_RENEGOTIATION
+ && !SSL_IS_TLS13(s)
+ && alert_level == SSL3_AL_WARNING) {
SSLfatal(s, SSL_AD_HANDSHAKE_FAILURE, SSL_F_SSL3_READ_BYTES,
SSL_R_NO_RENEGOTIATION);
return -1;
}
- } else if (alert_level == SSL3_AL_FATAL) {
+
+ if (alert_level == SSL3_AL_FATAL || SSL_IS_TLS13(s)) {
char tmp[16];
s->rwstate = SSL_NOTHING;
@@ -1579,8 +1586,6 @@ int ssl3_read_bytes(SSL *s, int type, int *recvd_type, unsigned char *buf,
SSL_R_UNKNOWN_ALERT_TYPE);
return -1;
}
-
- goto start;
}
if (s->shutdown & SSL_SENT_SHUTDOWN) { /* but we have not received a
|
grib_index_release/grib_index_delete does not close the file | @@ -280,7 +280,7 @@ const char* grib_get_package_name()
return "ecCodes";
}
-#define DEFAULT_FILE_POOL_MAX_OPENED_FILES 200
+#define DEFAULT_FILE_POOL_MAX_OPENED_FILES 0
static grib_context default_grib_context = {
0, /* inited */
@@ -538,13 +538,10 @@ grib_context* grib_context_get_default()
default_grib_context.grib_samples_path);
default_grib_context.keys_count = 0;
- default_grib_context.keys = grib_hash_keys_new(&(default_grib_context),
- &(default_grib_context.keys_count));
+ default_grib_context.keys = grib_hash_keys_new(&(default_grib_context), &(default_grib_context.keys_count));
- default_grib_context.concepts_index = grib_itrie_new(&(default_grib_context),
- &(default_grib_context.concepts_count));
- default_grib_context.hash_array_index = grib_itrie_new(&(default_grib_context),
- &(default_grib_context.hash_array_count));
+ default_grib_context.concepts_index = grib_itrie_new(&(default_grib_context), &(default_grib_context.concepts_count));
+ default_grib_context.hash_array_index = grib_itrie_new(&(default_grib_context), &(default_grib_context.hash_array_count));
default_grib_context.def_files = grib_trie_new(&(default_grib_context));
default_grib_context.lists = grib_trie_new(&(default_grib_context));
default_grib_context.classes = grib_trie_new(&(default_grib_context));
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.