message
stringlengths
6
474
diff
stringlengths
8
5.22k
Fix bug with Image -> Texture copies;
@@ -943,7 +943,7 @@ static int l_lovrPassCopy(lua_State* L) { srcOffset[0] = luax_optu32(L, 4, 0); srcOffset[1] = luax_optu32(L, 5, 0); dstOffset[0] = luax_optu32(L, 6, 0); - srcOffset[1] = luax_optu32(L, 7, 0); + dstOffset[1] = luax_optu32(L, 7, 0); extent[0] = luax_optu32(L, 8, ~0u); extent[1] = luax_optu32(L, 9, ~0u); srcOffset[2] = luax_optu32(L, 10, 1) - 1;
CCode: Update variable names
@@ -284,10 +284,10 @@ int elektraCcodeGet (Plugin * handle, KeySet * returned, Key * parentKey) ksRewind (returned); while ((key = ksNext (returned)) != 0) { - size_t const valsize = keyGetValueSize (key); - if (valsize > mapping->bufferSize) + size_t const size = keyGetValueSize (key); + if (size > mapping->bufferSize) { - mapping->bufferSize = valsize; + mapping->bufferSize = size; mapping->buffer = new unsigned char[mapping->bufferSize]; } @@ -306,10 +306,10 @@ int elektraCcodeSet (Plugin * handle, KeySet * returned, Key * parentKey ELEKTRA ksRewind (returned); while ((key = ksNext (returned)) != 0) { - size_t const valsize = keyGetValueSize (key); - if (valsize * 2 > mapping->bufferSize) + size_t const size = keyGetValueSize (key) * 2; + if (size > mapping->bufferSize) { - mapping->bufferSize = valsize * 2; + mapping->bufferSize = size; mapping->buffer = new unsigned char[mapping->bufferSize]; }
pluginprocess: fix typo in comment
@@ -212,8 +212,7 @@ public: { if (this->gptr () == this->egptr ()) { - // read from the pipe directly, using an ifstream on - // /dev/fd/<fd> fails on macOS + // read from the pipe directly into the buffer ssize_t r = read (fd_, buffer_, 4096); this->setg (this->buffer_, this->buffer_, this->buffer_ + r); } @@ -248,7 +247,7 @@ int elektraDumpGet (ckdb::Plugin *, ckdb::KeySet * returned, ckdb::Key * parentK keyDel (root); int errnosave = errno; - // We use dump for the processplugin library. Unfortunately on macOS reading from /dev/fd/<fd> via + // We use dump for the pluginprocess library. Unfortunately on macOS reading from /dev/fd/<fd> via // ifstream fails, thus we read directly from unnamed pipes using a custom buffer and read const char pipe[] = "/dev/fd/"; if (!strncmp (keyString (parentKey), pipe, strlen (pipe)))
webxr: also support velocity of head device;
@@ -273,9 +273,17 @@ var webxr = { }, webxr_getVelocity: function(device, velocity, angularVelocity) { - if (state.hands[device]) { + var pose; + + if (device === 0 /* DEVICE_HEAD */) { + pose = state.viewer; + } else if (state.hands[device]) { var space = state.hands[device].gripSpace || state.hands[device].targetRaySpace; - var pose = state.frame.getPose(space, state.space); + pose = state.frame.getPose(space, state.space); + } else { + return false; + } + if (pose && pose.linearVelocity && pose.angularVelocity) { HEAPF32[(velocity >> 2) + 0] = pose.linearVelocity.x; HEAPF32[(velocity >> 2) + 1] = pose.linearVelocity.y; @@ -285,7 +293,6 @@ var webxr = { HEAPF32[(angularVelocity >> 2) + 2] = pose.angularVelocity.z; return true; } - } return false; },
Bluedroid: Fix BLE provisioning failure with bluedroid stack
@@ -1242,6 +1242,8 @@ void GATT_Deregister (tGATT_IF gatt_if) tGATT_REG *p_reg = gatt_get_regcb(gatt_if); tGATT_TCB *p_tcb; tGATT_CLCB *p_clcb; + list_node_t *p_node = NULL; + list_node_t *p_next = NULL; #if (GATTS_INCLUDED == TRUE) UINT8 ii; tGATT_SR_REG *p_sreg; @@ -1268,9 +1270,9 @@ void GATT_Deregister (tGATT_IF gatt_if) #endif ///GATTS_INCLUDED == TRUE /* When an application deregisters, check remove the link associated with the app */ - list_node_t *p_node = NULL; - for(p_node = list_begin(gatt_cb.p_tcb_list); p_node; p_node = list_next(p_node)) { + for(p_node = list_begin(gatt_cb.p_tcb_list); p_node; p_node = p_next) { p_tcb = list_node(p_node); + p_next = list_next(p_node); if (p_tcb->in_use) { if (gatt_get_ch_state(p_tcb) != GATT_CH_CLOSE) { gatt_update_app_use_link_flag(gatt_if, p_tcb, FALSE, FALSE); @@ -1280,9 +1282,11 @@ void GATT_Deregister (tGATT_IF gatt_if) } } - list_node_t *p_node = NULL; - for(p_node = list_begin(gatt_cb.p_clcb_list); p_node; p_node = list_next(p_node)) { - p_clcb = list_node(p_node); + list_node_t *p_node_clcb = NULL; + list_node_t *p_node_next = NULL; + for(p_node_clcb = list_begin(gatt_cb.p_clcb_list); p_node_clcb; p_node_clcb = p_node_next) { + p_clcb = list_node(p_node_clcb); + p_node_next = list_next(p_node_clcb); if (p_clcb->in_use && (p_clcb->p_reg->gatt_if == gatt_if) && (p_clcb->p_tcb->tcb_idx == p_tcb->tcb_idx)) {
Solve more bugs from Python async handler.
@@ -1396,35 +1396,37 @@ PyObject *py_task_callback_handler_impl_unsafe(PyGILState_STATE gstate, PyObject Py_DECREF(result); + value ret = NULL; + if (PyErr_Occurred() != NULL) { loader_impl_py py_impl = loader_impl_get(callback_state->impl); py_loader_impl_error_print(py_impl); - /* TODO: Handle CancelledError or any exceptions raised by the user or propagate them up?? */ - PyGILState_Release(gstate); - callback_state->reject_callback(v, callback_state->context); + ret = callback_state->reject_callback(v, callback_state->context); gstate = PyGILState_Ensure(); - - Py_DECREF(callback_state->coroutine); - free(callback_state); - - Py_RETURN_NONE; } else { PyGILState_Release(gstate); - value ret = callback_state->resolve_callback(v, callback_state->context); + ret = callback_state->resolve_callback(v, callback_state->context); gstate = PyGILState_Ensure(); + } - PyObject *ret_py = py_loader_impl_value_to_capi(callback_state->impl, value_type_id(ret), ret); + loader_impl impl = callback_state->impl; Py_DECREF(callback_state->coroutine); free(callback_state); - return ret_py; + if (ret == NULL) + { + Py_RETURN_NONE; + } + else + { + return py_loader_impl_value_to_capi(impl, value_type_id(ret), ret); } }
common/keyboard_8042: Track aux chan enable and aux irq enable Start tracking these so we can correctly filter mouse messages. BRANCH=none TEST=Verified 8042 command prints current state
@@ -95,9 +95,11 @@ struct host_byte { static struct queue const from_host = QUEUE_NULL(8, struct host_byte); static int i8042_keyboard_irq_enabled; +static int i8042_aux_irq_enabled; /* i8042 global settings */ static int keyboard_enabled; /* default the keyboard is disabled. */ +static int aux_chan_enabled; /* default the mouse is disabled. */ static int keystroke_enabled; /* output keystrokes */ static uint8_t resend_command[MAX_SCAN_CODE_LEN]; static uint8_t resend_command_len; @@ -212,6 +214,18 @@ static void keyboard_enable_irq(int enable) lpc_keyboard_resume_irq(); } +/** + * Enable mouse IRQ generation. + * + * @param enable Enable (!=0) or disable (0) IRQ generation. + */ +static void aux_enable_irq(int enable) +{ + CPRINTS("AUX IRQ %s", enable ? "enable" : "disable"); + + i8042_aux_irq_enabled = enable; +} + /** * Send a scan code to the host. * @@ -431,6 +445,16 @@ static void keyboard_enable(int enable) keyboard_enabled = enable; } +static void aux_enable(int enable) +{ + if (!aux_chan_enabled && enable) + CPRINTS("AUX enabled"); + else if (aux_chan_enabled && !enable) + CPRINTS("AUX disabled"); + + aux_chan_enabled = enable; +} + static uint8_t read_ctl_ram(uint8_t addr) { if (addr < ARRAY_SIZE(controller_ram)) @@ -462,16 +486,23 @@ static void update_ctl_ram(uint8_t addr, uint8_t data) /* Enable IRQ before enable keyboard (queue chars to host) */ if (!(orig & I8042_ENIRQ1) && (data & I8042_ENIRQ1)) keyboard_enable_irq(1); + if (!(orig & I8042_ENIRQ12) && (data & I8042_ENIRQ12)) + aux_enable_irq(1); /* Handle the I8042_KBD_DIS bit */ keyboard_enable(!(data & I8042_KBD_DIS)); + /* Handle the I8042_AUX_DIS bit */ + aux_enable(!(data & I8042_AUX_DIS)); + /* * Disable IRQ after disable keyboard so that every char must * have informed the host. */ if ((orig & I8042_ENIRQ1) && !(data & I8042_ENIRQ1)) keyboard_enable_irq(0); + if ((orig & I8042_ENIRQ12) && !(data & I8042_ENIRQ12)) + aux_enable_irq(0); } } @@ -1046,8 +1077,10 @@ static int command_8042_internal(int argc, char **argv) ccprintf("data_port_state=%d\n", data_port_state); ccprintf("i8042_keyboard_irq_enabled=%d\n", i8042_keyboard_irq_enabled); + ccprintf("i8042_aux_irq_enabled=%d\n", i8042_aux_irq_enabled); ccprintf("keyboard_enabled=%d\n", keyboard_enabled); ccprintf("keystroke_enabled=%d\n", keystroke_enabled); + ccprintf("aux_chan_enabled=%d\n", aux_chan_enabled); ccprintf("resend_command[]={"); for (i = 0; i < resend_command_len; i++)
Lambert azimuthal
@@ -145,7 +145,18 @@ static int get_native_type(grib_accessor* a) { return GRIB_TYPE_STRING; } - +#if 0 +static int proj_mercator(grib_handle* h, char* result) +{ + int err = 0; + double LaDInDegrees = 0; + if ((err = grib_get_double_internal(h, "LaDInDegrees", &LaDInDegrees)) != GRIB_SUCCESS) + return err; + sprintf(result, "+proj=merc +lat_ts=%lf +lat_0=0 +lon_0=0 +x_0=0 +y_0=0 +a=%lf +b=%lf", + LaDInDegrees, earthMajorAxisInMetres, earthMinorAxisInMetres); + return GRIB_SUCCESS; +} +#endif static int unpack_string(grib_accessor* a, char* v, size_t* len) { grib_accessor_proj_string* self = (grib_accessor_proj_string*)a; @@ -167,6 +178,9 @@ static int unpack_string(grib_accessor* a, char* v, size_t* len) earthMinorAxisInMetres = earthMajorAxisInMetres = radius; } + /* Default: lat/lon grid */ + sprintf(v,"+proj=latlong +a=%lf +b=%lf", earthMajorAxisInMetres, earthMinorAxisInMetres); + if (strcmp(grid_type, "mercator") == 0) { double LaDInDegrees = 0; if ((err = grib_get_double_internal(h, "LaDInDegrees", &LaDInDegrees)) != GRIB_SUCCESS) @@ -201,12 +215,23 @@ static int unpack_string(grib_accessor* a, char* v, size_t* len) sprintf(v,"+proj=lcc +lon_0=%lf +lat_0=%lf +lat_1=%lf +lat_2=%lf +a=%lf +b=%lf",LoVInDegrees, LaDInDegrees, Latin1InDegrees,Latin2InDegrees, earthMajorAxisInMetres, earthMinorAxisInMetres); } + else if (strcmp(grid_type, "lambert_azimuthal_equal_area") == 0) { + double standardParallel, centralLongitude; + if ((err = grib_get_double_internal(h, "standardParallel", &standardParallel)) != GRIB_SUCCESS) + return err; + if ((err = grib_get_double_internal(h, "centralLongitude", &centralLongitude)) != GRIB_SUCCESS) + return err; + sprintf(v,"+proj=laea +lon_0=%lf +lat_0=%lf +a=%lf +b=%lf", + centralLongitude, standardParallel, earthMajorAxisInMetres, earthMinorAxisInMetres); + } else { grib_context_log(a->context, GRIB_LOG_ERROR, "proj string for grid '%s' not implemented", grid_type); *len = 0; return GRIB_NOT_IMPLEMENTED; } - *len = strlen(v) + 1; + size = strlen(v); + Assert(size > 0); + *len = size + 1; return err; }
fix for mpi
@@ -310,6 +310,7 @@ function build_uintah info "../src/configure CXX=\"$PAR_COMPILER_CXX\" CC=\"$PAR_COMPILER\" \ CFLAGS=\"$CFLAGS $C_OPT_FLAGS\" CXXFLAGS=\"$CXXFLAGS $CXX_OPT_FLAGS\" \ MPI_EXTRA_LIB_FLAG=\"$PAR_LIBRARY_NAMES\" \ + --with-mpi=built-in \ $FORTRANARGS \ $ZLIB_ARGS \ --prefix=\"$VISITDIR/uintah/$UINTAH_VERSION/$VISITARCH\" \ @@ -320,6 +321,7 @@ function build_uintah sh -c "../src/configure CXX=\"$PAR_COMPILER_CXX\" CC=\"$PAR_COMPILER\" \ CFLAGS=\"$CFLAGS $C_OPT_FLAGS\" CXXFLAGS=\"$CXXFLAGS $CXX_OPT_FLAGS\" \ MPI_EXTRA_LIB_FLAG=\"$PAR_LIBRARY_NAMES\" \ + --with-mpi=built-in \ $FORTRANARGS \ $ZLIB_ARGS \ --prefix=\"$VISITDIR/uintah/$UINTAH_VERSION/$VISITARCH\" \
docs/fingerprint: Update audit report link BRANCH=none TEST=view in gitiles
@@ -745,7 +745,7 @@ The log file is `/var/log/cros_fp.log`. [RM0433]: https://www.st.com/content/ccc/resource/technical/document/reference_manual/group0/c9/a3/76/fa/55/46/45/fa/DM00314099/files/DM00314099.pdf/jcr:content/translations/en.DM00314099.pdf [sandboxing]: https://chromium.googlesource.com/chromiumos/docs/+/HEAD/sandboxing.md [seccomp filter]: https://chromium.googlesource.com/chromiumos/platform2/+/HEAD/biod/init/seccomp/biod-seccomp-amd64.policy -[Security Audit Report]: https://drive.google.com/a/google.com/file/d/0B1HHKpeDpzYnMDdocGxwWUhpckpWM0hMU0tPa2ZjdEFnLU53/ +[Security Audit Report]: https://drive.google.com/file/d/0B1HHKpeDpzYnMDdocGxwWUhpckpWM0hMU0tPa2ZjdEFnLU53/view?usp=sharing&resourcekey=0-utAJosm8Lwvx9TOz7F4i7w [state machine]: https://chromium.googlesource.com/chromiumos/platform/ec/+/90d177e3f0ae729bea7e24934a3c6ef9f2520d45/common/fpsensor.c#252 [statically allocated]: https://chromium.googlesource.com/chromiumos/platform/ec/+/90d177e3f0ae729bea7e24934a3c6ef9f2520d45/common/fpsensor.c#57 [system key]: https://chromium.googlesource.com/chromiumos/platform2/+/23b79133514ac2cd986bce21c398fb6658bda248/cryptohome/mount_encrypted/encryption_key.h#125
removing undocumented and incompatible tempo message from [mtr]
@@ -698,13 +698,6 @@ static void mtr_write(t_mtr *x, t_symbol *s) hammerpanel_save(x->x_filehandle, canvas_getdir(x->x_glist), 0); } -static void mtr_tempo(t_mtr *x, t_floatarg f) -{ - int ntracks = x->x_ntracks; - t_mtrack **tpp = x->x_tracks; - while (ntracks--) mtrack_tempo(*tpp++, f); -} - static void mtr_free(t_mtr *x) { if (x->x_tracks) @@ -852,7 +845,5 @@ void mtr_setup(void) gensym("read"), A_DEFSYM, 0); class_addmethod(mtr_class, (t_method)mtr_write, gensym("write"), A_DEFSYM, 0); - class_addmethod(mtr_class, (t_method)mtr_tempo, - gensym("tempo"), A_FLOAT, 0); hammerfile_setup(mtr_class, 0); }
Fixing PSA return status
@@ -922,6 +922,7 @@ static int ssl_tls13_write_certificate_verify_body( mbedtls_ssl_context *ssl, size_t signature_len = 0; unsigned char verify_hash[ MBEDTLS_MD_MAX_SIZE ]; size_t verify_hash_len; + psa_status_t status = PSA_ERROR_CORRUPTION_DETECTED; *out_len = 0; @@ -984,11 +985,12 @@ static int ssl_tls13_write_certificate_verify_body( mbedtls_ssl_context *ssl, /* Hash verify buffer with indicated hash function */ psa_algorithm = mbedtls_psa_translate_md( md_alg ); - if( psa_hash_compute( psa_algorithm, + status = psa_hash_compute( psa_algorithm, verify_buffer, verify_buffer_len, verify_hash,sizeof( verify_hash ), - &verify_hash_len ) != PSA_SUCCESS ) + &verify_hash_len ); + if( status != PSA_SUCCESS ) return( psa_ssl_status_to_mbedtls( status ) ); MBEDTLS_SSL_DEBUG_BUF( 3, "verify hash", verify_hash, verify_hash_len );
Another increase of char array size
@@ -119,7 +119,7 @@ static void testAddressesSingleHyphen (void) static void testAddressesNumber (void) { - char intChar[15]; + char intChar[16]; convertLong (intChar, 0); testAddressSet (intChar, 1);
Values including `max` shall be allowed
@@ -4360,7 +4360,7 @@ static int do_send(quicly_conn_t *conn, quicly_send_context_t *s) conn->super.stats.num_handshake_timeouts++; goto CloseNow; } - if (conn->super.stats.num_packets.initial_handshake_sent >= conn->super.ctx->max_initial_handshake_packets) { + if (conn->super.stats.num_packets.initial_handshake_sent > conn->super.ctx->max_initial_handshake_packets) { QUICLY_PROBE(INITIAL_HANDSHAKE_PACKET_EXCEED, conn, conn->stash.now, conn->super.stats.num_packets.initial_handshake_sent); conn->super.stats.num_initial_handshake_exceeded++; goto CloseNow;
fixed typo in action_wrapper.vhd_source
@@ -350,7 +350,7 @@ ARCHITECTURE STRUCTURE OF action_wrapper IS axi_nvme_buser : IN std_logic_vector(0 DOWNTO 0); -- only for NVME_USED=TRUE axi_nvme_rid : IN std_logic_vector(0 DOWNTO 0); -- only for NVME_USED=TRUE axi_nvme_ruser : IN std_logic_vector(0 DOWNTO 0); -- only for NVME_USED=TRUE - axi_nvme_wuser : OUT std_logic_vector(0 DOWNTO 0) -- only for NVME_USED=TRUE + axi_nvme_wuser : OUT std_logic_vector(0 DOWNTO 0); -- only for NVME_USED=TRUE -- Ports of Axi Slave Bus Interface AXI_CTRL_REG axi_ctrl_reg_awaddr : IN std_logic_vector(C_S_AXI_CTRL_REG_ADDR_WIDTH-1 DOWNTO 0);
fix(user-agent): error description typo
@@ -141,7 +141,7 @@ http_reason_print(int httpcode) case HTTP_METHOD_NOT_ALLOWED: return "The HTTP method used is not valid for the location specified."; case HTTP_TOO_MANY_REQUESTS: - return "You got synced."; + return "You got ratelimited."; case HTTP_GATEWAY_UNAVAILABLE: return "There was not a gateway available to process your request. Wait a bit and retry."; default:
remove shadow warning when building in static mode
@@ -246,10 +246,10 @@ static void mi_segment_os_free(mi_segment_t* segment, mi_segments_tld_t* tld) { if (MI_SECURE>0) { // _mi_os_unprotect(segment, mi_segment_size(segment)); // ensure no more guard pages are set // unprotect the guard pages; we cannot just unprotect the whole segment size as part may be decommitted - size_t os_page_size = _mi_os_page_size(); - _mi_os_unprotect((uint8_t*)segment + mi_segment_info_size(segment) - os_page_size, os_page_size); - uint8_t* end = (uint8_t*)segment + mi_segment_size(segment) - os_page_size; - _mi_os_unprotect(end, os_page_size); + size_t os_pagesize = _mi_os_page_size(); + _mi_os_unprotect((uint8_t*)segment + mi_segment_info_size(segment) - os_pagesize, os_pagesize); + uint8_t* end = (uint8_t*)segment + mi_segment_size(segment) - os_pagesize; + _mi_os_unprotect(end, os_pagesize); } // purge delayed decommits now? (no, leave it to the cache) @@ -712,12 +712,12 @@ static mi_segment_t* mi_segment_init(mi_segment_t* segment, size_t required, mi_ if (MI_SECURE>0) { // in secure mode, we set up a protected page in between the segment info // and the page data - size_t os_page_size = _mi_os_page_size(); - mi_assert_internal(mi_segment_info_size(segment) - os_page_size >= pre_size); - _mi_os_protect((uint8_t*)segment + mi_segment_info_size(segment) - os_page_size, os_page_size); - uint8_t* end = (uint8_t*)segment + mi_segment_size(segment) - os_page_size; - mi_segment_ensure_committed(segment, end, os_page_size, tld->stats); - _mi_os_protect(end, os_page_size); + size_t os_pagesize = _mi_os_page_size(); + mi_assert_internal(mi_segment_info_size(segment) - os_pagesize >= pre_size); + _mi_os_protect((uint8_t*)segment + mi_segment_info_size(segment) - os_pagesize, os_pagesize); + uint8_t* end = (uint8_t*)segment + mi_segment_size(segment) - os_pagesize; + mi_segment_ensure_committed(segment, end, os_pagesize, tld->stats); + _mi_os_protect(end, os_pagesize); if (slice_entries == segment_slices) segment->slice_entries--; // don't use the last slice :-( guard_slices = 1; }
BugID:26073282: fix gbk path issue in create component script.
@@ -13,6 +13,12 @@ if sys.version_info[0] < 3: pass scriptdir = os.path.dirname(os.path.abspath(__file__)) +try: + from aos.util import locale_to_unicode + scriptdir = locale_to_unicode(scriptdir) +except: + pass + TEMPLATE_DIR = "templates/new_component_template" COMPONENTS_FOLDER = "components" AOS_MAKEFILE = "aos.mk"
Simply remove that line to avoid confusion
@@ -100,8 +100,6 @@ if (${CLAP_BUILD_TESTS}) MACOSX_BUNDLE_INFO_PLIST ${CMAKE_CURRENT_SOURCE_DIR}/src/plugins.plist.in ) elseif(CMAKE_SYSTEM_NAME STREQUAL "Windows") - # this is already exported by using CLAP_EXPORT - # target_link_options(clap-plugin-template PRIVATE /EXPORT:clap_entry) set_target_properties(clap-plugin-template PROPERTIES SUFFIX ".clap" PREFIX "") endif() endif()
docs: create table- fixed default storage options information * docs: create table- fixed default storage options information default storage options can be set for system, database, user. * Small edit to keep it in line with the 4.3.x commit.
<li> <cmdname>ORIENTATION</cmdname> </li> - </ul><p>The defaults can be set for a database, schema, and user. For information about - setting storage options, see the server configuration parameter - <codeph>gp_default_storage_options</codeph>.</p></li> + </ul><p>The defaults can be set for the system, a database, or a user. For information about + setting storage options, see the server configuration parameter <codeph><xref + href="../../ref_guide/config_params/guc-list.xml#gp_default_storage_options" + >gp_default_storage_options</xref></codeph>.</p></li> </ul> <note type="important">The current Greenplum Database legacy optimizer allows list partitions with multi-column (composite) partition keys. GPORCA does not support composite keys, so
Update: fixes and additions
@@ -184,8 +184,8 @@ def process(line): (obj, x, y, w, h, c) = detection x_real = x+w/2 y_real = y+h #down side of bb - if y_real > y_real_temp and (pickobj == None or pickobj == obj): - (obj_temp, x_real_temp, y_real_temp, w_temp, h_temp, c_temp) = (obj, x_real_temp, y_real_temp, w, h, c) + if y_real > y_real_temp: + (obj_temp, x_real_temp, y_real_temp, w_temp, h_temp, c_temp) = (obj, x_real, y_real, w, h, c) if y_real_temp != -1: TransbotPerceiveVisual(obj, x_real_temp, y_real_temp, trans, rot) print("//seen: ", obj, x_real, y_real) @@ -198,7 +198,7 @@ def process(line): if HadExternalAction: break #external action triggered, done executions = [] - if line.endswith(".") or line.endswith(". :|:"): + if line.endswith(".") or line.endswith(". :|:") or line.endswith("?") or line.endswith("? :|:"): NAR.AddInput(line) elif line == "*pick_with_feedback": pick_with_feedback() @@ -229,6 +229,8 @@ def process(line): NAR.AddInput("a. :|:") for i in range(50): TransbotExecute(NAR.AddInput("G! :|:")["executions"]) + elif line.isdigit() or line.startswith("*") or line.endswith("}"): + NAR.AddInput(line) def shell_step(lastLine = ""): #Get input line and forward potential command
rpz triggers, spelling fix.
@@ -77,7 +77,7 @@ rpz_action_to_string(enum rpz_action a) case RPZ_DISABLED_ACTION: return "rpz-disabled"; case RPZ_CNAME_OVERRIDE_ACTION: return "rpz-cname-override"; case RPZ_NO_OVERRIDE_ACTION: return "rpz-no-override"; - default: return "rpz-unkown-action"; + default: return "rpz-unknown-action"; } }
Ignore patches applied to 3rdparty/mbedtls
path = 3rdparty/mbedtls/mbedtls url = https://github.com/openenclave/openenclave-mbedtls.git branch = openenclave-mbedtls-2.28 + ignore = all [submodule "3rdparty/openssl/intel-sgx-ssl"] path = 3rdparty/openssl/intel-sgx-ssl url = https://github.com/intel/intel-sgx-ssl
rt1739: add es2 workaround Add a workaround to fix the incorrect CCD_MODE_ODL behavior on RT1739 ES2. TEST=manually BRANCH=none Tested-by: Ting Shen
@@ -183,6 +183,14 @@ static int rt1739_workaround(int port) case RT1739_DEVICE_ID_ES2: CPRINTS("RT1739 ES2"); + /* enter hidden mode */ + RETURN_ERROR(write_reg(port, 0xF1, 0x62)); + RETURN_ERROR(write_reg(port, 0xF0, 0x86)); + /* turn off SWENB output */ + RETURN_ERROR(write_reg(port, 0xE0, 0x07)); + /* leave hidden mode */ + RETURN_ERROR(write_reg(port, 0xF1, 0)); + RETURN_ERROR(write_reg(port, 0xF0, 0)); break; default:
OpenCoreMisc: Sync with OcSupportPkg changes
@@ -76,10 +76,12 @@ OcToolLoadEntry ( IN VOID *Context, IN OC_BOOT_ENTRY *ChosenEntry, OUT VOID **Data, - OUT UINT32 *DataSize + OUT UINT32 *DataSize, + OUT EFI_DEVICE_PATH_PROTOCOL **DevicePath OPTIONAL ) { CHAR16 ToolPath[64]; + OC_STORAGE_CONTEXT *Storage; UnicodeSPrint ( ToolPath, @@ -88,8 +90,10 @@ OcToolLoadEntry ( ChosenEntry->PathName ); + Storage = (OC_STORAGE_CONTEXT *) Context; + *Data = OcStorageReadFileUnicode ( - (OC_STORAGE_CONTEXT *) Context, + Storage, ToolPath, DataSize ); @@ -102,6 +106,10 @@ OcToolLoadEntry ( return EFI_NOT_FOUND; } + if (DevicePath != NULL) { + *DevicePath = Storage->DummyDevicePath; + } + return EFI_SUCCESS; }
Update vlsi/Makefile [skip ci]
@@ -70,7 +70,7 @@ SRAM_CONF=$(build_dir)/sram_generator-output.json srams: sram_generator sram_generator: $(SRAM_CONF) -# This should be built alongside $(SMEMS_FILE) +# This should be built alongside $(TOP_SMEMS_FILE) $(SMEMS_HAMMER): $(TOP_SMEMS_FILE) $(SRAM_GENERATOR_CONF): $(SMEMS_HAMMER)
OcMachoLib: Fix address range check.
@@ -469,7 +469,7 @@ MachoGetSectionByAddress64 ( Segment = MachoGetNextSegment64 (Context, Segment) ) { if ((Address >= Segment->VirtualAddress) - && (Address < (Segment->VirtualAddress + Address >= Segment->FileSize))) { + && (Address < (Segment->VirtualAddress + Segment->Size))) { Section = &Segment->Sections[0]; for (Index = 0; Index < Segment->NumberOfSections; ++Index) {
test_jvpp: improve error message when JVpp JARS are missing The java command fails with missing class error, when some of the JARs given by -cp are missing, which may be missleading. This patch fixes that by adding os.path.isfile check to test_jvpp.py.
@@ -106,9 +106,16 @@ class TestJVpp(VppTestCase): REGISTRY_JAR_PREFIX, version) self.logger.info("JVpp Registry jar path : {0}" .format(registry_jar_path)) + if (not os.path.isfile(registry_jar_path)): + raise Exception( + "JVpp Registry jar has not been found: {0}" + .format(registry_jar_path)) api_jar_path = self.full_jar_name(install_dir, api_jar_name, version) self.logger.info("Api jar path : {0}".format(api_jar_path)) + if (not os.path.isfile(api_jar_path)): + raise Exception( + "Api jar has not been found: {0}".format(api_jar_path)) # passes shm prefix as parameter to create connection with same value command = ["java", "-cp",
rms/pmix: latest also requiring libxml2
@@ -24,6 +24,7 @@ BuildRequires: libevent-devel BuildRequires: gcc-c++ BuildRequires: python3 BuildRequires: hwloc%{PROJ_DELIM} +BuildRequires: libxml2-devel #!BuildIgnore: post-build-checks %global install_path %{OHPC_ADMIN}/%{pname}
Fix fps not showing when using full config
@@ -520,6 +520,7 @@ parse_overlay_env(struct overlay_params *params, params->enabled[OVERLAY_PARAM_ENABLED_histogram] = 0; params->enabled[OVERLAY_PARAM_ENABLED_gpu_load_change] = 0; params->enabled[OVERLAY_PARAM_ENABLED_cpu_load_change] = 0; + params->enabled[OVERLAY_PARAM_ENABLED_fps_only] = 0; params->enabled[OVERLAY_PARAM_ENABLED_read_cfg] = read_cfg; } #define OVERLAY_PARAM_BOOL(name) \
test/usb_tcpmv2_td_pd_ll_e3.c: Format with clang-format BRANCH=none TEST=none
@@ -29,8 +29,8 @@ static int td_pd_ll_e3(enum pd_data_role data_role) /* * a) Run PROC.PD.E1 Bring-up according to the UUT role. */ - TEST_EQ(proc_pd_e1(data_role, INITIAL_AND_ALREADY_ATTACHED), - EC_SUCCESS, "%d"); + TEST_EQ(proc_pd_e1(data_role, INITIAL_AND_ALREADY_ATTACHED), EC_SUCCESS, + "%d"); /* * Make sure we are idle. Reject everything that is pending @@ -42,9 +42,7 @@ static int td_pd_ll_e3(enum pd_data_role data_role) * and do not send GoodCrc for nRetryCount + 1 times * (nRetryCount equals 3 since PD 2.1). */ - partner_send_msg(TCPCI_MSG_SOP, - PD_CTRL_GET_SINK_CAP, - 0, 0, NULL); + partner_send_msg(TCPCI_MSG_SOP, PD_CTRL_GET_SINK_CAP, 0, 0, NULL); retries = (partner_get_pd_rev() == PD_REV30) ? 2 : 3; TEST_EQ(verify_tcpci_tx_retry_count(TCPCI_MSG_SOP, 0, PD_DATA_SINK_CAP,
formatting changes/PR feedback no logic changes
@@ -54,7 +54,8 @@ try{ "1804 SGX1FLC Package Release" : { LinuxPackaging('1804', 'Release') }, "1804 SGX1FLC Package Release LVI" : { LinuxPackaging('1804', 'Release', 'ControlFlow') }, "1804 SGX1FLC Package RelWithDebInfo" : { LinuxPackaging('1804', 'RelWithDebInfo') }, - "1804 SGX1FLC Package RelWithDebInfo LVI" : { LinuxPackaging('1804', 'RelWithDebInfo', 'ControlFlow') },"Windows 2019 Debug" : { WindowsPackaging('2019','Debug') }, + "1804 SGX1FLC Package RelWithDebInfo LVI" : { LinuxPackaging('1804', 'RelWithDebInfo', 'ControlFlow') }, + "Windows 2019 Debug" : { WindowsPackaging('2019','Debug') }, "Windows 2019 Debug LVI" : { WindowsPackaging('2019','Debug', 'ControlFlow') }, "Windows 2019 Release" : { WindowsPackaging('2019','Release') }, "Windows 2019 Release LVI" : { WindowsPackaging('2019','Release', 'ControlFlow') }
Layout is fixed
@@ -1487,9 +1487,9 @@ void xdag_print_block_list(struct block_internal **block_list, int count, int pr char time_buf[64]; if(!print_only_addresses) { - fprintf(out, "--------------------------------------------------------------------------------------\n"); + fprintf(out, "-----------------------------------------------------------------------\n"); fprintf(out, "address time state\n"); - fprintf(out, "--------------------------------------------------------------------------------------\n"); + fprintf(out, "-----------------------------------------------------------------------\n"); } for(int i = 0; i < count; ++i) { @@ -1500,7 +1500,7 @@ void xdag_print_block_list(struct block_internal **block_list, int count, int pr fprintf(out, "%s\n", address); } else { block_time_to_string(block, time_buf); - fprintf(out, "%s\t%s\t\t%s\n", address, time_buf, get_block_state_info(block)); + fprintf(out, "%s %s %s\n", address, time_buf, get_block_state_info(block)); } } }
grib_dump error on GRIB edition 1 with raw packing
SUPER = grib_accessor_class_gen IMPLEMENTS = init IMPLEMENTS = unpack_long + IMPLEMENTS = get_native_type MEMBERS=const char* values MEMBERS=const char* precision END_CLASS_DEF @@ -45,6 +46,7 @@ or edit "accessor.class" and rerun ./make_class.pl */ +static int get_native_type(grib_accessor*); static int unpack_long(grib_accessor*, long* val, size_t* len); static void init(grib_accessor*, const long, grib_arguments*); static void init_class(grib_accessor_class*); @@ -75,7 +77,7 @@ static grib_accessor_class _grib_accessor_class_number_of_values_data_raw_packin 0, /* get number of values */ 0, /* get number of bytes */ 0, /* get offset to bytes */ - 0, /* get native type */ + &get_native_type, /* get native type */ 0, /* get sub_section */ 0, /* grib_pack procedures long */ 0, /* grib_pack procedures long */ @@ -115,7 +117,6 @@ static void init_class(grib_accessor_class* c) c->value_count = (*(c->super))->value_count; c->byte_count = (*(c->super))->byte_count; c->byte_offset = (*(c->super))->byte_offset; - c->get_native_type = (*(c->super))->get_native_type; c->sub_section = (*(c->super))->sub_section; c->pack_missing = (*(c->super))->pack_missing; c->is_missing = (*(c->super))->is_missing; @@ -189,3 +190,8 @@ static int unpack_long(grib_accessor* a, long* val, size_t* len) return err; } + +static int get_native_type(grib_accessor* a) +{ + return GRIB_TYPE_LONG; +}
remove unused function decl
@@ -1805,8 +1805,6 @@ int h2o_socket_ebpf_init_key(h2o_ebpf_map_key_t *key, void *_sock) return h2o_socket_ebpf_init_key_raw(key, sock_type, (void *)&local, (void *)&remote); } -static void on_track_ebpf_lookup_timer(h2o_timer_t *timeout); - static void report_ebpf_lookup_errors(h2o_error_reporter_t *reporter, uint64_t total_successes, uint64_t cur_successes) { fprintf(stderr,
BugID:24381709:Fixed the bug that uart demo could not send and receive data
@@ -24,7 +24,7 @@ static void task_recvdata_entry(void *arg) { int i; int ret; - char rev_buf[10]; + char rev_buf[1]; int rev_length; while (1) { @@ -34,7 +34,7 @@ static void task_recvdata_entry(void *arg) break; } - printf("recv data for uart:\r\n"); + printf("recv data from uart:\r\n"); for(i = 0; i < rev_length; i++) { printf("%c ", rev_buf[i]); } @@ -64,6 +64,8 @@ static int uart_app_init() printf("init uart error\r\n"); return -1; } + + return ret; } void hal_uart_app_run(void)
avf: fix simultaneous txq wrap and tx retry Type: fix Fixes: Ticket:
@@ -376,7 +376,7 @@ VNET_DEVICE_CLASS_TX_FN (avf_device_class) (vlib_main_t * vm, u32 thread_index = vm->thread_index; u8 qid = thread_index; avf_txq_t *txq = vec_elt_at_index (ad->txqs, qid % ad->num_queue_pairs); - u16 next = txq->next; + u16 next; u16 mask = txq->size - 1; u32 *buffers = vlib_frame_vector_args (frame); u16 n_enq, n_left, n_desc, *slot; @@ -387,6 +387,7 @@ VNET_DEVICE_CLASS_TX_FN (avf_device_class) (vlib_main_t * vm, n_left = frame->n_vectors; retry: + next = txq->next; /* release consumed bufs */ if (txq->n_enqueued) {
doc: prettify opal IMC counters calls
.. _opal-imc-counters: +.. _OPAL_IMC_COUNTERS_INIT: + OPAL_IMC_COUNTERS_INIT -============================== +====================== OPAL call interface to initialize In-memory collection infrastructure. Call does multiple scom writes on each invocation for Core/Trace IMC initialization. And for the @@ -37,18 +39,17 @@ Parameters Returns ------- -OPAL_PARAMETER +:ref:`OPAL_PARAMETER` In case of unsupported ``type`` - -OPAL_HARDWARE +:ref:`OPAL_HARDWARE` If any error in setting up the hardware. - -OPAL_SUCCESS +:ref:`OPAL_SUCCESS` On succesfully initialized or even if init operation is a no-op. +.. _OPAL_IMC_COUNTERS_START: OPAL_IMC_COUNTERS_START -============================ +======================= OPAL call interface for starting the In-Memory Collection counters for a specified domain (NEST/CORE/TRACE). @@ -65,15 +66,14 @@ Parameters Returns ------- -OPAL_PARAMETER +:ref:`OPAL_PARAMETER` In case of Unsupported ``type`` - -OPAL_HARDWARE +:ref:`OPAL_HARDWARE` If any error in setting up the hardware. - -OPAL_SUCCESS +:ref:`OPAL_SUCCESS` On successful execution of the operation for the given ``type``. +.. _OPAL_IMC_COUNTERS_STOP: OPAL_IMC_COUNTERS_STOP ====================== @@ -96,11 +96,9 @@ Parameters Returns ------- -OPAL_PARAMETER +:ref:`OPAL_PARAMETER` In case of Unsupported ``type`` - -OPAL_HARDWARE +:ref:`OPAL_HARDWARE` If any error in setting up the hardware. - -OPAL_SUCCESS +:ref:`OPAL_SUCCESS` On successful execution of the operation for the given ``type``.
PEERDIR 4 test_tool
@@ -435,6 +435,7 @@ when ($NO_GPL == "yes") { NO_GPL_FLAG=--no-gpl } NEED_PLATFORM_PEERDIRS=yes +PEERDIR_TEST_TOOL=yes PYTHON2=no PYTHON3=no @@ -612,7 +613,7 @@ module BASE_UNIT { CFLAGS+=-DCYTHON_TRACE=1 -DCYTHON_TRACE_NOGIL=1 } - when ($TESTS_REQUESTED && $NEED_PLATFORM_PEERDIRS == "yes") { + when ($PEERDIR_TEST_TOOL == "yes") { PEERDIR+=build/platform/test_tool } @@ -1179,6 +1180,7 @@ module RESOURCES_LIBRARY: _LIBRARY { .RESTRICTED=ALLOCATOR SIZE TAG DATA TEST_DATA DEPENDS FORK_TESTS FORK_SUBTESTS SPLIT_FACTOR TEST_CWD RUN TIMEOUT SRCS PEERDIR SET(NEED_PLATFORM_PEERDIRS no) + SET(PEERDIR_TEST_TOOL no) DISABLE(WITH_VALGRIND) NO_CODENAVIGATION() NO_PLATFORM() @@ -3570,6 +3572,9 @@ module GO_PROGRAM: GO_BASE_UNIT { } PEERDIR(${GOSTD}/runtime) + when ($PEERDIR_TEST_TOOL == "yes") { + PEERDIR+=build/platform/test_tool + } } module GO_TEST: GO_PROGRAM { @@ -3581,10 +3586,6 @@ module GO_TEST: GO_PROGRAM { PEERDIR(${GOSTD}/testing) ADD_YTEST($REALPRJNAME go.test) - - when ($TESTS_REQUESTED && $NEED_PLATFORM_PEERDIRS == "yes") { - PEERDIR+=build/platform/test_tool - } } JAVA_IGNORE_CLASSPATH_CLASH_VALUE=
Auto generate sprite name if file moved as long as current name matches previous filename
@@ -180,12 +180,16 @@ const loadSprite = createAsyncThunk<{ data: SpriteSheet }, string>( if (existingAsset) { delete inodeToRecentSpriteSheet[data.inode]; + const oldAutoName = existingAsset.filename.replace(/.png/i, ""); return { data: { ...existingAsset, ...data, id: existingAsset.id, - name: existingAsset.name || data.name, + name: + existingAsset.name !== oldAutoName + ? existingAsset.name || data.name + : data.name, states: existingAsset.states, }, };
[build] fix meson.build when building all TLS mods x-ref: "[lighttpd] -mod-openssl fails" "[lighttpd] -mod-wolfssl fails"
@@ -432,15 +432,20 @@ if get_option('with_mysql') endif libssl = [] -libx509 = [] libcrypto = [] +libsslcrypto = [] libgnutls = [] +libmbedtls = [] +libmbedcrypto = [] +libmbedx509 = [] +libwolfssl = [] if get_option('with_openssl') # manual search: # header: openssl/ssl.h # function: SSL_new (-lssl) # function: BIO_f_base64 (-lcrypto) libssl = [ dependency('libssl') ] + libsslcrypto = [ dependency('libcrypto') ] libcrypto = [ dependency('libcrypto') ] conf_data.set('HAVE_OPENSSL_SSL_H', true) conf_data.set('HAVE_LIBSSL', true) @@ -449,7 +454,7 @@ if get_option('with_wolfssl') # manual search: # header: wolfssl/ssl.h # function: wolfSSL_Init (-lwolfssl) - libssl = [ dependency('wolfssl') ] + libwolfssl = [ dependency('wolfssl') ] libcrypto = [ dependency('wolfssl') ] conf_data.set('HAVE_WOLFSSL_SSL_H', true) endif @@ -459,8 +464,9 @@ if get_option('with_mbedtls') # function: mbedtls_cipher_info_from_type (-lmbedtls) # function: mbedtls_x509_get_name (-lmbedx509) # function: mbedtls_base64_encode (-lmbedcrypto) - libssl = [ compiler.find_library('mbedtls') ] - libx509 = [ compiler.find_library('mbedx509') ] + libmbedtls = [ compiler.find_library('mbedtls') ] + libmbedx509 = [ compiler.find_library('mbedx509') ] + libmbedcrypto = [ compiler.find_library('mbedcrypto') ] libcrypto = [ compiler.find_library('mbedcrypto') ] conf_data.set('HAVE_LIBMBEDCRYPTO', true) endif @@ -1070,19 +1076,19 @@ endif if get_option('with_openssl') modules += [ - [ 'mod_openssl', [ 'mod_openssl.c' ], libssl + libcrypto ], + [ 'mod_openssl', [ 'mod_openssl.c' ], libssl + libsslcrypto ], ] endif if get_option('with_wolfssl') modules += [ - [ 'mod_wolfssl', [ 'mod_wolfssl.c' ], libcrypto ], + [ 'mod_wolfssl', [ 'mod_wolfssl.c' ], libwolfssl ], ] endif if get_option('with_mbedtls') modules += [ - [ 'mod_mbedtls', [ 'mod_mbedtls.c' ], libssl + libx509 + libcrypto ], + [ 'mod_mbedtls', [ 'mod_mbedtls.c' ], libmbedtls + libmbedx509 + libmbedcrypto ], ] endif
Fix integer overflow in _sdsMakeRoomFor (CVE-2021-41099) The existing overflow checks handled the greedy growing, but didn't handle a case where the addition of the header size is what causes the overflow.
@@ -239,7 +239,7 @@ void sdsclear(sds s) { sds _sdsMakeRoomFor(sds s, size_t addlen, int greedy) { void *sh, *newsh; size_t avail = sdsavail(s); - size_t len, newlen; + size_t len, newlen, reqlen; char type, oldtype = s[-1] & SDS_TYPE_MASK; int hdrlen; size_t usable; @@ -249,7 +249,7 @@ sds _sdsMakeRoomFor(sds s, size_t addlen, int greedy) { len = sdslen(s); sh = (char*)s-sdsHdrSize(oldtype); - newlen = (len+addlen); + reqlen = newlen = (len+addlen); assert(newlen > len); /* Catch size_t overflow */ if (greedy == 1) { if (newlen < SDS_MAX_PREALLOC) @@ -266,7 +266,7 @@ sds _sdsMakeRoomFor(sds s, size_t addlen, int greedy) { if (type == SDS_TYPE_5) type = SDS_TYPE_8; hdrlen = sdsHdrSize(type); - assert(hdrlen + newlen + 1 > len); /* Catch size_t overflow */ + assert(hdrlen + newlen + 1 > reqlen); /* Catch size_t overflow */ if (oldtype==type) { newsh = s_realloc_usable(sh, hdrlen+newlen+1, &usable); if (newsh == NULL) return NULL;
Get rid of internal_recurse.
@@ -35,13 +35,12 @@ def onrun_java_program(unit, *args): Custom code generation @link: https://wiki.yandex-team.ru/yatool/java/#kodogeneracijarunjavaprogram """ - flat, kv = common.sort_by_keywords( - {'IN': -1, 'IN_DIR': -1, 'OUT': -1, 'OUT_DIR': -1, 'CWD': 1, 'CLASSPATH': -1, 'ADD_SRCS_TO_CLASSPATH': 0}, - args - ) - for cp in kv.get('CLASSPATH', []): - unit.oninternal_recurse(cp) + flat, kv = common.sort_by_keywords({'IN': -1, 'IN_DIR': -1, 'OUT': -1, 'OUT_DIR': -1, 'CWD': 1, 'CLASSPATH': -1, 'ADD_SRCS_TO_CLASSPATH': 0}, args) + depends = kv.get('CLASSPATH', []) + kv.get('JAR', []) + if depends: + # XXX: hack to force ymake to build dependencies + unit.on_run_java(['TOOL'] + depends + ["OUT", "fake.out.{}".format(hash(tuple(depends)))]) prev = unit.get(['RUN_JAVA_PROGRAM_VALUE']) or '' new_val = (prev + ' ' + base64.b64encode(json.dumps(list(args), encoding='utf-8'))).strip()
libhfuzz/memorycmp: cast characters to int before passing them to the cmp_func()
@@ -48,7 +48,7 @@ static inline int HF_strcmp(const char* s1, const char* s2, uintptr_t addr) { static inline int HF_strcasecmp( const char* s1, const char* s2, int (*cmp_func)(int), uintptr_t addr) { size_t i; - for (i = 0; cmp_func((unsigned char)s1[i]) == cmp_func((unsigned char)s2[i]); i++) { + for (i = 0; cmp_func((int)(unsigned char)s1[i]) == cmp_func((int)(unsigned char)s2[i]); i++) { if (s1[i] == '\0' || s2[i] == '\0') { break; } @@ -57,7 +57,7 @@ static inline int HF_strcasecmp( instrumentUpdateCmpMap(HF_cmphash(addr, s1, s2), i); instrumentAddConstStr(s1); instrumentAddConstStr(s2); - return ((int)cmp_func((unsigned char)s1[i]) - (int)cmp_func((unsigned char)s2[i])); + return (cmp_func((int)(unsigned char)s1[i]) - cmp_func((int)(unsigned char)s2[i])); } static inline int HF_strncmp( @@ -86,8 +86,8 @@ static inline int HF_strncasecmp( const char* s1, const char* s2, size_t n, int (*cmp_func)(int), bool constfb, uintptr_t addr) { size_t i; for (i = 0; i < n; i++) { - if ((cmp_func((unsigned char)s1[i]) != cmp_func((unsigned char)s2[i])) || s1[i] == '\0' || - s2[i] == '\0') { + if ((cmp_func((int)(unsigned char)s1[i]) != cmp_func((int)(unsigned char)s2[i])) || + s1[i] == '\0' || s2[i] == '\0') { break; } } @@ -102,7 +102,7 @@ static inline int HF_strncasecmp( return 0; } - return cmp_func((unsigned char)s1[i]) - cmp_func((unsigned char)s2[i]); + return cmp_func((int)(unsigned char)s1[i]) - cmp_func((int)(unsigned char)s2[i]); } static inline char* HF_strstr(const char* haystack, const char* needle, uintptr_t addr) {
BIO_lookup_ex: Retry with AI_ADDRCONFIG cleared if getaddrinfo fails The lookup for ::1 with getaddrinfo() might return error even if the ::1 would work if AI_ADDRCONFIG flag is used. Fixes:
@@ -696,6 +696,7 @@ int BIO_lookup_ex(const char *host, const char *service, int lookup_type, /* Note that |res| SHOULD be a 'struct addrinfo **' thanks to * macro magic in bio_lcl.h */ + retry: switch ((gai_ret = getaddrinfo(host, service, &hints, res))) { # ifdef EAI_SYSTEM case EAI_SYSTEM: @@ -706,6 +707,19 @@ int BIO_lookup_ex(const char *host, const char *service, int lookup_type, case 0: ret = 1; /* Success */ break; +# if (defined(EAI_FAMILY) || defined(EAI_ADDRFAMILY)) && defined(AI_ADDRCONFIG) +# ifdef EAI_FAMILY + case EAI_FAMILY: +# endif +# ifdef EAI_ADDRFAMILY + case EAI_ADDRFAMILY: +# endif + if (hints.ai_flags & AI_ADDRCONFIG) { + hints.ai_flags &= ~AI_ADDRCONFIG; + goto retry; + } + /* fall through */ +# endif default: BIOerr(BIO_F_BIO_LOOKUP_EX, ERR_R_SYS_LIB); ERR_add_error_data(1, gai_strerror(gai_ret));
Docs: Changing plr example as it was incorrect
@@ -257,8 +257,7 @@ LANGUAGE 'plr'; --table DROP TABLE IF EXISTS wj_model_results_roi; CREATE TABLE wj_model_results_roi AS SELECT * - FROM wj_plr_RE((SELECT wj_droi2_array), - (SELECT cs FROM wj_droi2_array));</codeblock> + FROM wj_plr_RE('{1,1,1}', '{"a", "b", "c"}');</codeblock> </body> </topic> </topic>
Fix conf parsing overflow After adding a new option to the 'struct conf_pool' list, a previously undiscovered bug came to light where we were casting data types of 'short' to 'int' and 'bool' to 'int' which would give us the wrong result. This patch fixes these bugs.
@@ -974,13 +974,38 @@ static char *conf_set_num(struct conf *cf, struct command *cmd, void *conf) { return CONF_OK; } +static char *conf_set_short(struct conf *cf, struct command *cmd, void *conf) { + uint8_t *p; + int num; + uint8_t *np; + struct string *value; + + p = conf; + np = (uint8_t *)(p + cmd->offset); + + if (*np != CONF_UNSET_NUM) { + return "is a duplicate"; + } + + value = array_top(&cf->arg); + + num = dn_atoi(value->data, value->len); + if (num < 0) { + return "is not a number"; + } + + *np = num; + + return CONF_OK; +} + static char *conf_set_bool(struct conf *cf, struct command *cmd, void *conf) { uint8_t *p; - int *bp; + bool *bp; struct string *value, true_str, false_str; p = conf; - bp = (int *)(p + cmd->offset); + bp = (bool *)(p + cmd->offset); if (*bp != CONF_UNSET_NUM) { return "is a duplicate"; @@ -1158,13 +1183,13 @@ static struct command conf_commands[] = { {string("max_msgs"), conf_set_num, offsetof(struct conf_pool, alloc_msgs_max)}, - {string("datastore_connections"), conf_set_num, + {string("datastore_connections"), conf_set_short, offsetof(struct conf_pool, datastore_connections)}, - {string("local_peer_connections"), conf_set_num, + {string("local_peer_connections"), conf_set_short, offsetof(struct conf_pool, local_peer_connections)}, - {string("remote_peer_connections"), conf_set_num, + {string("remote_peer_connections"), conf_set_short, offsetof(struct conf_pool, remote_peer_connections)}, null_command};
Temporarily add debug output for install-gsl
@@ -8,5 +8,7 @@ else tar -xzf gsl-2.4.tar.gz cd gsl-2.4 && ./configure --prefix=/usr && make && sudo make install fi +pwd +ls export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:/usr/lib export LD_RUN_PATH=$LD_RUN_PATH:/usr/lib
removed undefined calls
@@ -1082,12 +1082,12 @@ QvisSimulationWindow::UpdateWindow(bool doAll) else if(uiValues->GetName() == "STRIP_CHART_CLEAR_MENU") { - stripChartMgr->clearMenu(); + // stripChartMgr->clearMenu(); } else if(uiValues->GetName() == "STRIP_CHART_ADD_MENU_ITEM") { - stripChartMgr->addMenuItem( uiValues->GetSvalue().c_str() ); + // stripChartMgr->addMenuItem( uiValues->GetSvalue().c_str() ); } else if(uiValues->GetName() == "STRIP_CHART_CLEAR")
Fix Misra 13.4 on safety code
@@ -166,7 +166,8 @@ void update_sample(struct sample_t *sample, int sample_new) { sample->values[0] = sample_new; // get the minimum and maximum measured samples - sample->min = sample->max = sample->values[0]; + sample->min = sample->values[0]; + sample->max = sample->values[0]; for (int i = 1; i < sizeof(sample->values)/sizeof(sample->values[0]); i++) { if (sample->values[i] < sample->min) { sample->min = sample->values[i];
set TOP property for system_wrapper
@@ -81,6 +81,8 @@ make_wrapper -files [get_files $bd_path/system.bd] -top add_files -norecurse $bd_path/hdl/system_wrapper.v +set_property TOP system_wrapper [current_fileset] + set files [glob -nocomplain cores/common_modules/*.v] if {[llength $files] > 0} { add_files -norecurse $files
pbio/dcmotor: port is not part of setup Make this pattern consistent with servo. Port is only needed to get object, not setup. (and even this may ultimately go away.)
static pbio_dcmotor_t dcmotors[PBDRV_CONFIG_NUM_MOTOR_CONTROLLER]; -static pbio_error_t pbio_dcmotor_setup(pbio_dcmotor_t *dcmotor, pbio_port_t port, pbio_direction_t direction) { +static pbio_error_t pbio_dcmotor_setup(pbio_dcmotor_t *dcmotor, pbio_direction_t direction) { pbio_error_t err; @@ -27,14 +27,13 @@ static pbio_error_t pbio_dcmotor_setup(pbio_dcmotor_t *dcmotor, pbio_port_t port } // Get device ID to ensure we are dealing with a supported device - err = pbdrv_motor_get_id(port, &dcmotor->id); + err = pbdrv_motor_get_id(dcmotor->port, &dcmotor->id); if (err != PBIO_SUCCESS) { return err; } // Set direction and state dcmotor->direction = direction; - dcmotor->port = port; dcmotor->state = PBIO_DCMOTOR_COAST; // Set duty scaling and offsets @@ -49,9 +48,10 @@ pbio_error_t pbio_dcmotor_get(pbio_port_t port, pbio_dcmotor_t **dcmotor, pbio_d // Get pointer to dcmotor *dcmotor = &dcmotors[port - PBDRV_CONFIG_FIRST_MOTOR_PORT]; + (*dcmotor)->port = port; // Initialize and set up pwm properties - return pbio_dcmotor_setup(*dcmotor, port, direction); + return pbio_dcmotor_setup(*dcmotor, direction); } pbio_error_t pbio_dcmotor_set_settings(pbio_dcmotor_t *dcmotor, int32_t stall_torque_limit_pct, int32_t duty_offset_pct) {
[bsp][stm32] fix drv_pwm.c device register usefault
@@ -542,7 +542,7 @@ static int stm32_pwm_init(void) LOG_D("%s init success", stm32_pwm_obj[i].name); /* register pwm device */ - if (rt_device_pwm_register(rt_calloc(1, sizeof(struct rt_device_pwm)), stm32_pwm_obj[i].name, &drv_ops, &stm32_pwm_obj[i].tim_handle) == RT_EOK) + if (rt_device_pwm_register(&stm32_pwm_obj[i].pwm_device, stm32_pwm_obj[i].name, &drv_ops, &stm32_pwm_obj[i].tim_handle) == RT_EOK) { LOG_D("%s register success", stm32_pwm_obj[i].name);
build: bump minimum Linux kernel headers version to 5.4 opae-sdk requires [1] IOVA range capability support [2] added in Linux 5.4. [1] [2]
@@ -208,7 +208,7 @@ if (OPAE_BUILD_LIBOPAEVFIO) set(PLATFORM_SUPPORTS_VFIO FALSE CACHE BOOL "Platform supports vfio driver" FORCE) message(WARNING "Could not compile VFIO. - This most likely means that kernel( version >= 4.6) headers aren't installed. + This most likely means that kernel( version >= 5.4) headers aren't installed. See errors in platform_vfio_errors.txt") file(WRITE ${CMAKE_BINARY_DIR}/platform_vfio_errors.txt ${TRY_COMPILE_OUTPUT}) endif(VFIO_CHECK)
framework/media : fix a logical error at MediaWorker::startWorker() Even though a worker thread creation fails, the flag of 'mIsRunning' is set to true. Fix this logical error by updating 'mIsRunning' to false in case that the thread creation fails.
@@ -45,6 +45,7 @@ void MediaWorker::startWorker() if (ret != OK) { medvdbg("Fail to create worker thread, return value : %d\n", ret); --mRefCnt; + mIsRunning = false; return; } pthread_setname_np(mWorkerThread, mThreadName);
Let's Amazon Echo control On/Off devices like Ubisys S1/S2
@@ -281,6 +281,8 @@ bool DeRestPluginPrivate::lightToMap(const ApiRequest &req, const LightNode *lig map["uniqueid"] = lightNode->uniqueId(); map["name"] = lightNode->name(); map["modelid"] = lightNode->modelId(); // real model id + map["manufacturername"] = lightNode->manufacturer(); + if (!auth.strict) { map["hascolor"] = lightNode->hasColor(); @@ -291,16 +293,17 @@ bool DeRestPluginPrivate::lightToMap(const ApiRequest &req, const LightNode *lig // Amazon Echo quirks mode if (auth.strict && auth.devicetype.startsWith(QLatin1String("Echo"))) { - // OSRAM plug - if (lightNode->type() == QLatin1String("On/Off plug-in unit")) + // OSRAM plug + Ubisys S1/S2 + if (lightNode->type().startsWith(QLatin1String("On/Off"))) { + map["modelid"] = QLatin1String("LWB010"); + map["manufacturername"] = QLatin1String("Philips"); map["type"] = QLatin1String("Dimmable light"); state["bri"] = (double)254; } } map["swversion"] = lightNode->swBuildId(); - map["manufacturername"] = lightNode->manufacturer(); QString etag = lightNode->etag; etag.remove('"'); // no quotes allowed in string
build: fix spelling
@@ -529,7 +529,7 @@ def withDockerEnv(stage_name, image, cl) { def cpu_count = cpuCount() withEnv(["MAKEFLAGS='-j${cpu_count+2} -l${cpu_count*2}'", "CTEST_PARALLEL_LEVEL='${cpu_count+2}'"]) { - echo "Staring ${env.STAGE_NAME} on ${env.NODE_NAME}" + echo "Starting ${env.STAGE_NAME} on ${env.NODE_NAME}" checkout scm docker.image(imageFullName(image)).inside { cl() } }
removed hardcoding of /opt/rocm. The EXTRA_LDFLAGS now use -L/lib and AOMPROCM is set to /tmp/grodgers/rocm/aomp if not already set.
@@ -16,8 +16,9 @@ endif CLANG = clang++ -lamdhip64 $(RPTH) $(USE_MALLOC) $(LARGE) $(ITERS) OMP_BIN = $(AOMP)/bin/$(CLANG) CC = $(OMP_BIN) $(VERBOSE) +AOMPROCM ?= $(AOMP) ifeq ($(EPSDB),1) -EXTRA_LDFLAGS = -L/opt/rocm/lib +EXTRA_LDFLAGS = -L$(AOMPROCM)/lib endif #-ccc-print-phases #"-\#\#\#"
Change alua transition in process message to a debug The message "Lock acquisition operation is already in process" should be a dbg message.
@@ -559,7 +559,7 @@ int alua_implicit_transition(struct tcmu_device *dev, struct tcmulib_cmd *cmd) if (rdev->lock_state == TCMUR_DEV_LOCK_LOCKED) { goto done; } else if (rdev->lock_state == TCMUR_DEV_LOCK_LOCKING) { - tcmu_dev_info(dev, "Lock acquisition operation is already in process.\n"); + tcmu_dev_dbg(dev, "Lock acquisition operation is already in process.\n"); ret = TCMU_STS_BUSY; goto done; }
Adjust the stack pointer before calling the C code to account for the current stack frame.
// Call the 'C' handler with the current stack pointer // as argument, e.g.: // go_write(stackptr) - mov %rsp, 0x08(%rsp) + mov %rsp, %r11 + add $0x20, %r11 + mov %r11, 0x08(%rsp) lea \c_func@GOTPCREL(%rip), %r11 mov (%r11), %r11 mov %r11, (%rsp)
Sync improvements from downstream
@@ -1599,7 +1599,7 @@ static CFStringRef copy_printer_interface_deviceid(printer_interface_t printer, /* This request takes the 0 based configuration index. IOKit returns a 1 based configuration index */ configurationIndex -= 1; - bzero(&request, sizeof(request)); + memset(&request, 0, sizeof(request)); request.bmRequestType = USBmakebmRequestType(kUSBIn, kUSBClass, kUSBInterface); request.bRequest = kUSBPrintClassGetDeviceID; @@ -1641,7 +1641,7 @@ static CFStringRef copy_printer_interface_deviceid(printer_interface_t printer, IOUSBDevRequestTO request; IOUSBDeviceDescriptor desc; - bzero(&request, sizeof(request)); + memset(&request, 0, sizeof(request)); request.bmRequestType = USBmakebmRequestType( kUSBIn, kUSBStandard, kUSBDevice ); request.bRequest = kUSBRqGetDescriptor; @@ -1704,15 +1704,10 @@ static CFStringRef copy_printer_interface_deviceid(printer_interface_t printer, { CFStringAppend(extras, ret); CFRelease(ret); - - ret = extras; } - else - { ret = extras; } } - } if (ret != NULL) { @@ -1754,7 +1749,7 @@ static CFStringRef copy_printer_interface_indexed_description(printer_interface_ UInt8 description[256]; // Max possible descriptor length IOUSBDevRequestTO request; - bzero(description, 2); + memset(description, 0, 2); request.bmRequestType = USBmakebmRequestType(kUSBIn, kUSBStandard, kUSBDevice); request.bRequest = kUSBRqGetDescriptor; @@ -1768,7 +1763,7 @@ static CFStringRef copy_printer_interface_indexed_description(printer_interface_ err = (*printer)->ControlRequestTO(printer, 0, &request); if (err != kIOReturnSuccess && err != kIOReturnOverrun) { - bzero(description, request.wLength); + memset(description, 0, request.wLength); // Let's try again full length. Here's why: // On USB 2.0 controllers, we will not get an overrun error. We just get a "babble" error @@ -1801,7 +1796,7 @@ static CFStringRef copy_printer_interface_indexed_description(printer_interface_ request.wValue = (kUSBStringDesc << 8) | index; request.wIndex = language; - bzero(description, length); + memset(description, 0, length); request.wLength = (UInt16)length; request.pData = &description; request.completionTimeout = 0;
Add links to wikipedia artciles
@@ -55,14 +55,14 @@ For an example of a project using SIMDe, see There are currently complete implementations of the following instruction sets: - * MMX - * SSE - * SSE2 - * SSE3 - * SSSE3 - * SSE4.1 - * AVX - * FMA + * [MMX](https://en.wikipedia.org/wiki/MMX_(instruction_set)) + * [SSE](https://en.wikipedia.org/wiki/Streaming_SIMD_Extensions) + * [SSE2](https://en.wikipedia.org/wiki/SSE2) + * [SSE3](https://en.wikipedia.org/wiki/SSE3) + * [SSSE3](https://en.wikipedia.org/wiki/SSSE3) + * [SSE4.1](https://en.wikipedia.org/wiki/SSE4#SSE4.1) + * [AVX](https://en.wikipedia.org/wiki/Advanced_Vector_Extensions) + * [FMA](https://en.wikipedia.org/wiki/FMA_instruction_set) As well as partial support for many others; see the [instruction-set-support](https://github.com/nemequ/simde/issues?q=is%3Aissue+is%3Aopen+label%3Ainstruction-set-support+sort%3Aupdated-desc)
Trying table
@@ -40,33 +40,21 @@ build_libm.sh - Built the libm DBCL (Device BC Library) These scripts install into $HOME/rocm/aomp (or $AOMP if set). The repositories needed by AOMP are: - Component DIRECTORY NAME * AOMP REPOSITORY ** - --------- ---------------------------- --------------------------- - roct $HOME/git/aomp/roct-thunk-interfaces [roct-thunk-interfaces] (https://github.com/radeonopencompute/roct-thunk-interface) - - rocr $HOME/git/aomp/rocr-runtime %roc/rocr-runtime - - llvm $HOME/git/aomp/clang %rocdev/clang - - llvm $HOME/git/aomp/llvm %rocdev/llvm - - llvm $HOME/git/aomp/lld %rocdev/lld - - utils $HOME/git/aomp/aomp %rocdev/aomp - - hcc $HOME/git/aomp/hcc %rocdev/hcc - - hip $HOME/git/aomp/hip %rocdev/hip - - atmi $HOME/git/aomp/atmi %roc/atmi - - openmp $HOME/git/aomp/openmp %rocdev/openmp - - libdevice $HOME/git/aomp/rocm-device-libs %roc/rocm-device-libs - - libm $HOME/git/aomp/aomp %rocdev/aomp - - $HOME/git/aomp/openmpapps %roclib/openmpapps +| Component | DIRECTORY NAME * | AOMP REPOSITORY ** +| --------- | ---------------- | ------------------ +| roct | $HOME/git/aomp/roct-thunk-interfaces | [roct-thunk-interfaces] (https://github.com/radeonopencompute/roct-thunk-interface) +| rocr | $HOME/git/aomp/rocr-runtime | %roc/rocr-runtime +| llvm | $HOME/git/aomp/clang | %rocdev/clang +| llvm | $HOME/git/aomp/llvm | %rocdev/llvm +| llvm | $HOME/git/aomp/lld | %rocdev/lld +| utils | $HOME/git/aomp/aomp | %rocdev/aomp +| hcc | $HOME/git/aomp/hcc | %rocdev/hcc +| hip | $HOME/git/aomp/hip | %rocdev/hip +| atmi | $HOME/git/aomp/atmi | %roc/atmi +| openmp | $HOME/git/aomp/openmp | %rocdev/openmp +| libdevice | $HOME/git/aomp/rocm-device-libs | %roc/rocm-device-libs +| libm | $HOME/git/aomp/aomp | %rocdev/aomp +| | $HOME/git/aomp/openmpapps | %roclib/openmpapps * By default, repositories clone here ** Replace %roc with "https://github.com/RadeonOpenCompute"
Set flag to proper value Set flag to proper value. Was previously 0xFFFFFFF and has been corrected to 0xFFFFFFFF.
@@ -35,8 +35,8 @@ const mbedtls_x509_crt_profile compat_profile = MBEDTLS_X509_ID_FLAG( MBEDTLS_MD_SHA256 ) | MBEDTLS_X509_ID_FLAG( MBEDTLS_MD_SHA384 ) | MBEDTLS_X509_ID_FLAG( MBEDTLS_MD_SHA512 ), - 0xFFFFFFF, /* Any PK alg */ - 0xFFFFFFF, /* Any curve */ + 0xFFFFFFFF, /* Any PK alg */ + 0xFFFFFFFF, /* Any curve */ 1024, }; @@ -53,8 +53,8 @@ const mbedtls_x509_crt_profile profile_rsa3072 = const mbedtls_x509_crt_profile profile_sha512 = { MBEDTLS_X509_ID_FLAG( MBEDTLS_MD_SHA512 ), - 0xFFFFFFF, /* Any PK alg */ - 0xFFFFFFF, /* Any curve */ + 0xFFFFFFFF, /* Any PK alg */ + 0xFFFFFFFF, /* Any curve */ 1024, };
I updated one of the onion peel operator tests. The test suite should now pass again.
@@ -109,9 +109,6 @@ def TestUCD(): ResetView() SetViewExtentsType("actual") AddOperator("OnionPeel") - op = OnionPeelAttributes() - op.index = 1 - SetOperatorOptions(op) DrawPlots() Test("ops_onionpeel_04")
[io] enable to modify an existing nslaw in the scene to change the parameters of the nslaw can be useful for multiple simulation sequence
@@ -1164,7 +1164,7 @@ class MechanicsHdf5(object): collision_group2=0): """ Add a nonsmooth law for contact between 2 groups. - Only NewtonImpactFrictionNSL are supported. + Only NewtonImpactRollingFrictionNSL are supported. name is an user identifiant and must be unique, mu is the coefficient of friction, e is the coefficient of restitution on the contact normal, @@ -1174,6 +1174,11 @@ class MechanicsHdf5(object): if name not in self._nslaws_data: nslaw = self._nslaws_data.create_dataset(name, (0,)) nslaw.attrs['type'] = 'NewtonImpactRollingFrictionNSL' + else: + nslaw=self._nslaws_data[name] + if nslaw.attrs['type'] != 'NewtonImpactRollingFrictionNSL': + self.print_verbose('[warning] a nslaw is already existing with the same name ', name ,' but not the same type') + nslaw.attrs['mu'] = mu nslaw.attrs['mu_r'] = mu_r nslaw.attrs['e'] = e @@ -1194,6 +1199,11 @@ class MechanicsHdf5(object): if name not in self._nslaws_data: nslaw = self._nslaws_data.create_dataset(name, (0,)) nslaw.attrs['type'] = 'NewtonImpactFrictionNSL' + else: + nslaw=self._nslaws_data[name] + if nslaw.attrs['type'] != 'NewtonImpactFrictionNSL': + self.print_verbose('[warning] a nslaw is already existing with the same name ', name ,' but not the same type') + nslaw.attrs['mu'] = mu nslaw.attrs['e'] = e nslaw.attrs['gid1'] = collision_group1 @@ -1218,6 +1228,11 @@ class MechanicsHdf5(object): if name not in self._nslaws_data: nslaw = self._nslaws_data.create_dataset(name, (0,)) nslaw.attrs['type'] = 'NewtonImpactNSL' + else: + nslaw=self._nslaws_data[name] + if nslaw.attrs['type'] != 'NewtonImpactNSL': + self.print_verbose('[warning] a nslaw is already existing with the same name ', name ,' but not the same type') + nslaw.attrs['e'] = e nslaw.attrs['gid1'] = collision_group1 nslaw.attrs['gid2'] = collision_group2 @@ -1241,6 +1256,11 @@ class MechanicsHdf5(object): if name not in self._nslaws_data: nslaw = self._nslaws_data.create_dataset(name, (0,)) nslaw.attrs['type'] = 'RelayNSL' + else: + nslaw=self._nslaws_data[name] + if nslaw.attrs['type'] != 'RelayNSL': + self.print_verbose('[warning] a nslaw is already existing with the same name', name ,' but not the same type') + nslaw.attrs['size'] = size nslaw.attrs['lb'] = lb nslaw.attrs['ub'] = ub
admin/docs: bump version for 2.6.0
%include %{_sourcedir}/OHPC_macros Name: docs%{PROJ_DELIM} -Version: 2.5.0 +Version: 2.6.0 Release: 1 Summary: OpenHPC documentation License: BSD-3-Clause
OpenCoreNvram: Fix NVRAM GUID reading
@@ -87,7 +87,7 @@ OcLoadNvramSupport ( // // FIXME: Checking string length manually is due to inadequate assertions. // - AsciiVariableGuid = OC_BLOB_GET (VariableMap->Keys[GuidIndex]); + AsciiVariableGuid = OC_BLOB_GET (Config->Nvram.Add.Keys[GuidIndex]); if (AsciiStrLen (AsciiVariableGuid) == GUID_STRING_LENGTH) { Status = AsciiStrToGuid (AsciiVariableGuid, &VariableGuid); } else {
ci: sort build configuration output
@@ -100,7 +100,7 @@ jobs: west build -s zmk/app -b ${{ matrix.board }} -- -DZMK_CONFIG=${GITHUB_WORKSPACE}/${{ inputs.config_path }} ${{ steps.variables.outputs.extra-cmake-args }} ${{ matrix.cmake-args }} - name: ${{ steps.variables.outputs.display-name }} Kconfig file - run: cat build/zephyr/.config | grep -v "^#" | grep -v "^$" + run: cat build/zephyr/.config | grep -v "^#" | grep -v "^$" | sort - name: Rename artifacts run: |
doc: define hierarchy and order comparision for use cases
@@ -8,9 +8,16 @@ This folder contains the use cases for `libelektra-core`. 2. associates each key with 1. values: an arbitrary byte sequence 2. metadata: another ordered, hierarchical associative array, which only associates keys with values -3. orders keys first by namespace then lexicographically -4. supports the operations: insert, lookup, prefix lookup, access by index and delete +3. orders keys first by namespace then lexicographically with respect to hierarchy +4. supports the operations: insert, remove, lookup, prefix/hierarchy lookup and access by index (which enables iteration) Additionally, `libelektra-core` provides a data structure, called `Key`, that represents a single key-value pair in the associative array, but can also be used standalone. +To support the hierarchical and ordered nature of a `KeySet` there are two fundamental comparison operations that can be performed on two `Key`s: + +1. Order Comparison: + Establishes the linear order of `Key`s and thereby defines the iteration order of a `KeySet`. +2. Hierarchy Comparison: + Establishes the hierarchy of `Key`s, by defining based on their names whether one `Key` is a descendant of another. + The individual use cases provide details on these data structures and the operations it supports.
decisions: clarify when not needed
@@ -46,6 +46,9 @@ Substantial decisions, however, must be made in a transparent and participative Rather it is important that decisions: - contain everything relevant, and - what is written is technically correct. +- Usually no decisions are needed for libraries, plugins, tools or bindings if: + - they are similar to already existing modules (e.g. yet another checker plugin) + - if they reimplement some existing module (e.g. reimplementation in other programming languages) - Reviewers are only required to read the files in the PR. They may ignore the discussions surrounding the decision. Reviews should focus on the "Problem" section first, and only when that is agreed upon focus in the other parts. @@ -67,8 +70,6 @@ Substantial decisions, however, must be made in a transparent and participative - People focus on getting the best solutions and not to wish for the impossible. - People creating decision PRs have strong motives to also finish it. They take extra effort on them to be clear about the problem and find the best solution. -- After 1.0, due to stability guarantees, not so many decisions are required anymore. - People are free to reimplement the libraries or plugins of Elektra if they disagree with Elektra's [quality goals](/doc/GOALS.md) and still help with our vision. ## Considered Alternatives
common: fix SIMDE_FLOAT64_C macro when SIMDE_FLOAT64_TYPE is defined
@@ -496,7 +496,7 @@ typedef SIMDE_FLOAT32_TYPE simde_float32; # define SIMDE_FLOAT64_TYPE double # define SIMDE_FLOAT64_C(value) value #else -# define SIMDE_FLOAT32_C(value) ((SIMDE_FLOAT64_TYPE) value) +# define SIMDE_FLOAT64_C(value) ((SIMDE_FLOAT64_TYPE) value) #endif typedef SIMDE_FLOAT64_TYPE simde_float64;
Add fallback when rk unaligned with padlock
@@ -83,6 +83,10 @@ int mbedtls_padlock_xcryptecb( mbedtls_aes_context *ctx, unsigned char buf[256]; rk = ctx->buf + ctx->rk_offset; + + if( ( (long) rk & 15 ) != 0 ) + return( MBEDTLS_ERR_PADLOCK_DATA_MISALIGNED ); + blk = MBEDTLS_PADLOCK_ALIGN16( buf ); memcpy( blk, input, 16 ); @@ -125,11 +129,13 @@ int mbedtls_padlock_xcryptcbc( mbedtls_aes_context *ctx, uint32_t *ctrl; unsigned char buf[256]; + rk = ctx->buf + ctx->rk_offset; + if( ( (long) input & 15 ) != 0 || - ( (long) output & 15 ) != 0 ) + ( (long) output & 15 ) != 0 || + ( (long) rk & 15 ) != 0 ) return( MBEDTLS_ERR_PADLOCK_DATA_MISALIGNED ); - rk = ctx->buf + ctx->rk_offset; iw = MBEDTLS_PADLOCK_ALIGN16( buf ); memcpy( iw, iv, 16 );
Update EVP_KDF-X942-ASN1.pod Replaced OSSL_KDF_PARAM_KEY with OSSL_KDF_PARAM_SECRET as that seems to be the intended value from the code (OSSL_KDF_PARAM_KEY is also supported but looks like a fallback). Fixed name for OSSL_KDF_PARAM_X942_USE_KEYBITS. CLA: trivial
@@ -30,7 +30,7 @@ The supported parameters are: These parameters work as described in L<EVP_KDF(3)/PARAMETERS>. -=item "key" (B<OSSL_KDF_PARAM_KEY>) <octet string> +=item "secret" (B<OSSL_KDF_PARAM_SECRET>) <octet string> The shared secret used for key derivation. This parameter sets the secret. @@ -60,7 +60,7 @@ An optional octet string containing public info contributed by the responder. An optional octet string containing some additional, mutually-known public information. Setting this value also sets "use-keybits" to 0. -=item "use-keybits" (B<OSSL_KDF_PARAM_X942_SUPP_PRIVINFO>) <integer> +=item "use-keybits" (B<OSSL_KDF_PARAM_X942_USE_KEYBITS>) <integer> The default value of 1 will use the KEK key length (in bits) as the "supp-pubinfo". A value of 0 disables setting the "supp-pubinfo".
Sync Juntalis' findings
@@ -180,6 +180,32 @@ public unsafe partial struct MJIManager { /// <inheritdoc cref="Facility1"/> [FieldOffset(0x1EC)] public MJIBuildingPlacement Cabin; + /// <summary> + /// The current day in the Craftworks cycle, from 0 to 6. + /// </summary> + /// <remarks> + /// 0 represents reset day (Tuesday). + /// </remarks> + [FieldOffset(0x27A)] public byte CurrentCycleDay; + + /// <summary> + /// An array containing the currently-configured rest days for the Isleworks. Contains values 0 - 13, and is + /// always in order. + /// </summary> + /// <remarks> + /// Like CurrentCycleDay, 0 represents Reset Day. 7, likewise, represents the next reset. This field may not be + /// populated until the Craftworks have been opened at least once. + /// </remarks> + [FieldOffset(0x27A)] public fixed byte CraftworksRestDays[4]; + + /// <summary> + /// The current groove level of the Isleworks. + /// </summary> + /// <remarks> + /// May not be present until the Isleworks is loaded at least once by the player. + /// </remarks> + [FieldOffset(0x2C8)] public byte CurrentGroove; + /// <summary> /// A reference to the current set of popularity scores given to craftworks on the player's island. The actual /// popularity scores can be pulled from the MJICraftworksPopularity sheet using this value as a Row ID. @@ -262,7 +288,8 @@ public unsafe partial struct MJIManager { [StructLayout(LayoutKind.Explicit, Size = 0x10)] public struct MJIBuildingPlacement { /// <summary> - /// Should line up with the row ids of the MJIBuildingPlace sheet. + /// At load, the location of this specific building. Will update if a building is changed, but the exact + /// mechanism of the update (and why it does such) is not currently known. /// </summary> [FieldOffset(0x4)] public uint PlaceId; @@ -386,7 +413,7 @@ public struct MJILandmarkPlacement { /// <summary> /// The RowID of the landmark currently present at the specified location. /// </summary> - [FieldOffset(0x5)] public byte LandmarkId; + [FieldOffset(0x9)] public byte LandmarkId; } public enum CraftworkSupply {
ARMv8: adding comments on the needed steps to initialize the core
@@ -93,16 +93,57 @@ static void print_string(char *string) string++; } } +#include <barrelfish_kpi/arm_core_data.h> - +/** + * @brief initializes an application core + * + * @param state pointer to the armv8_core_data structure + * + * This function is intended to bring the core to the same state as if it + * has been booted by the UEFI boot loader. + */ void boot_app_init(lpaddr_t state) { pl011_uart_initialize(&uart, (mackerel_addr_t)0x87E024000000UL); + struct armv8_core_data *cd = (struct armv8_core_data *)state; + + if (cd->boot_magic == ARMV8_BOOTMAGIC_PSCI) { + print_string("######### Hello world: ARMV8_BOOTMAGIC_PSCI\n"); + } else if (cd->boot_magic == ARMV8_BOOTMAGIC_PARKING) { + print_string("######### Hello world: ARMV8_BOOTMAGIC_PSCI\n"); + } else { + print_string("######### Hello world: UNKNOWN PROTOCOL\n"); + } + + uint8_t current_el = get_current_el(); + if (current_el == 3) { + print_string("EL=3\n"); + } else if (current_el == 2) { + print_string("EL=2\n"); + } else if (current_el == 1) { + print_string("EL=1\n"); + } + + /* disable interrupts */ + + /* set the TCR */ + + /* set the ttbr0/1 */ + + /* flush caches */ + + /* enable MMU */ + + /* invalidate TLB */ + + /* enable interrupts */ - print_string("######### Hello world\n"); while(1) ; + + boot_bsp_init(cd->boot_magic, state, cd->kernel_stack); }
chat-cli: Add debug poke for connecting to store
~& %chat-cli-prep ?^ old [~ this(+<+ u.old)] - :- [ost.bowl %peer /chat-store [our-self %chat-store] /all]~ + :- [connect ~] %_ this audience [[our-self /] ~ ~] settings (sy %showtime %notify ~) width 80 == +:: +connect: connect to the chat-store +:: +++ connect + ^- move + [ost.bowl %peer /chat-store [our-self %chat-store] /all] :: +true-self: moons to planets :: ++ true-self =+ who=(slaw %p i.path) ?~ who [our-self path] [u.who path] +:: +poke-noun: debug helpers +:: +++ poke-noun + |= a=* + ^- (quip move _this) + ?: ?=(%connect a) + [[connect ~] this] + [~ this] :: +poke-sole-action: handle cli input :: ++ poke-sole-action
improve qt finding on windows
@@ -81,8 +81,10 @@ function _find_sdkdir(sdkdir, sdkver) { "HKEY_CLASSES_ROOT\\Applications\\QtProject.QtCreator.c\\shell\\Open\\Command", "HKEY_CLASSES_ROOT\\Applications\\QtProject.QtCreator.cpp\\shell\\Open\\Command", + "HKEY_CLASSES_ROOT\\Applications\\QtProject.QtCreator.pro\\shell\\Open\\Command", "HKEY_CURRENT_USER\\SOFTWARE\\Classes\\Applications\\QtProject.QtCreator.c\\shell\\Open\\Command", - "HKEY_CURRENT_USER\\SOFTWARE\\Classes\\Applications\\QtProject.QtCreator.cpp\\shell\\Open\\Command" + "HKEY_CURRENT_USER\\SOFTWARE\\Classes\\Applications\\QtProject.QtCreator.cpp\\shell\\Open\\Command", + "HKEY_CURRENT_USER\\SOFTWARE\\Classes\\Applications\\QtProject.QtCreator.pro\\shell\\Open\\Command" } for _, reg in ipairs(regs) do table.insert(paths, function ()
docs: seccomp: Update instruction to get seccomp operator relase 0.4.0. This release packages seccompprofile in v1beta1.
@@ -262,13 +262,13 @@ Operator](https://github.com/kubernetes-sigs/security-profiles-operator). To install the operator, use the following commands: <!-- -In our code, we include seccompprofile/v1alpha1, thus we apply from v0.3.0 to +In our code, we include seccompprofile/v1beta1, thus we apply from v0.4.0 to ensure operator.yaml applies on our code. --> ```bash $ kubectl apply -f https://github.com/jetstack/cert-manager/releases/download/v1.6.0/cert-manager.yaml $ kubectl --namespace cert-manager wait --for condition=ready pod -l app.kubernetes.io/instance=cert-manager -$ kubectl apply -f https://raw.githubusercontent.com/kubernetes-sigs/security-profiles-operator/v0.3.0/deploy/operator.yaml +$ kubectl apply -f https://raw.githubusercontent.com/kubernetes-sigs/security-profiles-operator/v0.4.0/deploy/operator.yaml ``` Once installed, the seccomp gadget can generate `seccompprofile` resources
doc: adding more content to
\title{Skate in Barrelfish} \author{Barrelfish project} % \date{\today} % Uncomment (if needed) - date is automatic -\tnnumber{20} +\tnnumber{020} \tnkey{Skate} + \lstdefinelanguage{skate}{ morekeywords={schema,typedef,fact,enum}, sensitive=true, @@ -123,7 +124,7 @@ The Skate compiler is written in Haskell using the Parsec parsing library. It generates C header files from the Skate files. In addition it supports the generation of Schema documentation. -The source code for Skate can be found in \texttt{$SOURCE/tools/skate}. +The source code for Skate can be found in \texttt{SOURCE/tools/skate}. \section{Command line options} @@ -133,6 +134,7 @@ The source code for Skate can be found in \texttt{$SOURCE/tools/skate}. $ skate <options> INFILE.skt \end{verbatim} + Where options is one of \begin{description} \item[-o] \textit{filename} The output file name @@ -214,6 +216,78 @@ fact \> query \> flags \> constants \> \> \\ \label{chap:declaration} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% +In this chapter we define the layout of a skate file, which declarations it +must contain and what other declarations it can have. + +\section{The Skate File} +A Skate file must consist of zero or more \emph{import} declarations (see +~\ref{sec:decl:schema}) followed by a single \emph{schema} declaration +(see~\ref{sec:decl:schema}) which contains the actual fact definitions, etc. + +\begin{syntax} +(Import)* +Schema +\end{syntax} + +%\begin{align*} +% skatefile & \rightarrow ( Import )^{\textrm{*}} (Schema) +%\end{align*} + +\section{Imports}\label{sec:decl:import} +An import declaration makes the definitions in a different device file +available in the current device definition, as described below. The +syntax of an import declaration is as follows: + +\begin{syntax} +import \synit{schema}; +\end{syntax} + +\begin{description} +\item[schema] is the name of the schema to import definitions from. +\end{description} + +The Skate compiler will search for a file with the appropriate name and +parse this at the same time as the main file, along with this file's +imports, and so on. Cyclic dependencies between device files will not +cause errors, but at present are unlikely to result in C header files +which will successfully compile. + +\section{Schemas}\label{sec:decl:schema} + +\begin{syntax} +schema \synit{name} "\synit{description}" +\verb+{+ + \synit{declaration}; + \ldots +\verb+}+; +\end{syntax} + +\begin{description} +\item[name] is an identifier for the Schema type, and will be used to + generate identifiers in the target language (typically C). + The name of the name of the schema \emph{must} correspond to the + filename of the file, including case sensitivity: for example, + the file \texttt{cpuid.schema} will define a schema type + of name \texttt{cpuid}. + +\item [description] is a string literal in double quotes, which + describes the schema type being specified, for example \texttt{"CPUID + Information Schema"}. + +\item [declaration] One of the declarations described in~\ref{sec:decl:decls}. + +\end{description} + +\section{Declarations}\label{sec:decl:decls} + +\subsection{Facts} + +\subsection{Flags} + +\subsection{Constants} + +\subsection{Queries} +
Update TERMS.md Change format.
@@ -14,45 +14,65 @@ Scope CCL has a commitment to support key DESC analyses. In this regard, CCL aims to -support cosmological models that are priorities for WGs (e.g., as defined by the "Beyond wCDM" taskforce); + -ensure sufficient flexibility such that modifications can be implemented at a later stage as needed (e.g., models for observational systematics or changes to how the non-linear power spectrum is computed for various parts of the analysis); + -support core predictions (e.g., shear correlation functions), but not necessarily every statistic that could be computed from those core predictions (e.g., COSEBIs). CCL interface support is as follows. -Public APIs are maintained in Python only (i.e., version bumping is not necessary following changes in the C interface). + -The C implementation can be used for speed as needed. + -Documentation should be maintained in the main README, readthedocs and the benchmarks folder. + -Examples will be maintained in a separate repository. Boundaries of CCL (with respect to various models and systematics) -Not all models need to have equal support. CCL aims to support the DESC analyses, not to support all possible statistics in all models. + -CCL should provide a generic framework to enable modeling of systematics effects. However, overly specific models for systematics are best located in other repositories (e.g., firecrown). Boundaries of CCL (with respect to other TJP software) -TJP will necessarily develop other software tools to support DESC cosmological analyses. CCL development should ensure internal consistency between the various tools by defining a common parameterization of theoretical quantities (e.g., kernels for Limber integration, the metric potentials, etc.). + -CCL development should also proceed in a manner to avoid significant duplication of effort within the collaboration. Standards to establish the quality of code that can be part of CCL --------------------------------------- -Contributions to CCL should be well-documented in python (only doc-string level docs are required, readthedocs does the rest). C code is expected to be commented, though not documented. + -Features added to CCL should pass required tests. + 1.one (publicly available) benchmark with an independent prediction should be provided to test the specific feature that has been added. The benchmark code can use CCL input as long as that is not affecting the validation procedure. + 2.Tests should run in continuous integration scheme used for CCL. + 3.Unit tests should be written as identified in code review. + -All PRs are code reviewed by at least one person (two are suggested for PRs that bring in new features or propose significant changes to the CCL API). -CCL uses semantic versioning. + +-CCL uses semantic versioning. + -We should aim to tag new versions with bug fixes on a monthly basis with the expectation that no release is made if no new bugs have been fixed. + -Critical bug fixes should be released almost immediately via bump of the micro version and a new release. Guidelines towards contribution --------------------------------------- -Contributions should be made via PRs, and such PRs should be reviewed as described above. + -Response to code review requests should be prompt and happen within a few days of being requested on github by the author of the PR. Code review by members of the CCL development team should happen within a week of the request being accepted. + -CONTRIBUTING file should be consulted by those making PRs. + -Developers should use standard DESC channels to interact with the team (e.g., Slack, telecons, etc.). + -Developers should seek consensus with the team as needed about new features being incorporated (especially for API changes or additions). This process should happen via the CCL telecons, slack channel, or github comments generally before a new feature is merged. More minor changes, like bug fixes or refactoring, do not necessarily need this process. + -External contributions. They come in different forms: bug fixes or new features. They should proceed in full awareness of the LSST DESC Publication Policy. In particular, the use of new features added to the library, but not officially released, may require more formal steps to be taken within the DESC. External contributors and DESC members wishing to use CCL for non-DESC projects should consult with the TJP working group conveners, ideally before the work has started, but definitely before any publication or posting of the work to the arXiv.
BUMP pymongocrypt 1.3.2.dev0
# See the License for the specific language governing permissions and # limitations under the License. -__version__ = '1.3.1' +__version__ = '1.3.2.dev0' _MIN_LIBMONGOCRYPT_VERSION = '1.5.2'
fixing problems with vtune for mpi engine
@@ -1350,14 +1350,18 @@ class VTuneProfiler(Debugger): vtargs = self.DebuggerArguments() if len(vtargs) == 0: - vtargs = ["-collect", "hotspots", - "-result-dir", "./vtune_%s_hotspots.dir/"%(visitargs[0].split('/')[-1]), - "-trace-mpi", "-quiet"] + try: + os.mkdir("./vtune_%s"%self.launcher.generalArgs.exe_name) + except OSError: + pass + vtargs = ["-collect", "hotspots", "-quiet", "-user-data-dir", + "./vtune_%s"%self.launcher.generalArgs.exe_name] + if self.launcher.generalArgs.exe_name == "engine_par": + vtargs += ["-trace-mpi", "-r", "%d"%os.getpid()] if len(launcherargs) > 0: # parallel launch - args = launcherargs + xterm_args \ - + self.Executable() + vtargs + visitargs + args = launcherargs + self.Executable() + vtargs + visitargs else: # serial launch args = self.Executable() + vtargs + visitargs
output: added handling of case where the coro wasn't completely initialized
@@ -84,7 +84,10 @@ static FLB_INLINE void flb_coro_destroy(struct flb_coro *coro) VALGRIND_STACK_DEREGISTER(coro->valgrind_stack_id); #endif + if (coro->callee != NULL) { co_delete(coro->callee); + } + flb_free(coro); }
Add console clear method.
@@ -105,7 +105,7 @@ typedef struct TCOD_ConsoleTile { @code{.cpp} tcod::ConsolePtr console = tcod::new_console({{80, 50}}); console->at({1, 1}).ch = '@'; // Bounds-checked references to a tile. - (*console)[{1, 1}].bg = {0, 0, 255, 255}; // Access a tile without bounds checking. + (*console)[{1, 1}].bg = {0, 0, 255, 255}; // Access a tile without bounds checking, colors are RGBA. if (console->in_bounds({100, 100})) {} // Test if an index is in bounds. for (auto& tile : *console) tile.fg = {255, 255, 0, 255}; // Iterate over all tiles on a console. for (auto& tile : *console) tile = {0x20, {255, 255, 255, 255}, {0, 0, 0, 255}}; // Clear all tiles. @@ -133,6 +133,35 @@ struct TCOD_Console { @brief Return a const pointer to the end of this consoles tile data. */ [[nodiscard]] auto end() const noexcept -> const TCOD_ConsoleTile* { return tiles + elements; } + /*************************************************************************** + @brief Clear a console by setting all tiles to the provided TCOD_ConsoleTile object. + + @param tile A TCOD_ConsoleTile refernce which will be used to clear the console. + + @code{.cpp} + // New consoles start already cleared with the space character, a white foreground, and a black background. + auto console = tcod::new_console(80, 50) + console->clear() // Clear with the above mentioned defaults. + console->clear({0x20, {255, 255, 255, 255}, {0, 0, 0, 255}}); // Same as the above. + console->clear(0x20, {255, 255, 255}, {0, 0, 0}) // Also same as the above. + @endcode + */ + void clear(const TCOD_ConsoleTile& tile = {0x20, {255, 255, 255, 255}, {0, 0, 0, 255}}) noexcept { + for (auto& it : *this) it = tile; + } + /*************************************************************************** + @brief Clear the console with the given character and RGB colors. + + @param ch A Unicode codepoint, if unsure then use the space codepoint ``int{0x20}`` + @param fg The foreground color to clear with. + While it doesn't seem useful at first to set the foreground color at the same time as clearing with space, + this will still determine what the default color will be if a character is set without also overwriting the + colors for that tile. + @param bg The background color to clear with. + */ + void clear(int ch, const TCOD_ColorRGB& fg, const TCOD_ColorRGB& bg) noexcept { + clear({ch, TCOD_ColorRGBA{fg.r, fg.g, fg.b, 255}, TCOD_ColorRGBA{bg.r, bg.g, bg.b, 255}}); + } /*************************************************************************** @brief Return a reference to the tile at `xy`. */
add "SUPER" parameter.
@@ -63,6 +63,7 @@ end # class: "Range", # file: "c_range_method_table.h", # func: "mrbc_init_class_range", +# super: "mrbc_class_object", # methods: [ # { name: "first", func: "c_range_first", if_exp: "" } # ... @@ -84,13 +85,13 @@ def parse_source_string( src ) next end - # e.g. CLASS("xxx") + # e.g. CLASS, METHOD and etc. if /^([A-Z]+)\s*\((.*)\)/ =~ txt key = $1 args = $2.split(",").map {|s| s.strip } flag_arg_ok = true case key - when "CLASS", "FILE", "FUNC" + when "CLASS", "FILE", "FUNC", "SUPER" if args.size == 1 ret[key.downcase.to_sym] = args[0] else @@ -99,7 +100,7 @@ def parse_source_string( src ) when "METHOD" if args.size == 2 - ret[:methods] << { name: args[0], func: args[1] } + ret[:methods] << { name: strip_double_quot(args[0]), func: args[1] } ret[:methods].last[:if_exp] = if_exp.dup if !if_exp.empty? else flag_arg_ok = false @@ -119,6 +120,8 @@ def parse_source_string( src ) flag_error = true } + ret[:super] ||= "mrbc_class_object" + return flag_error ? nil : ret end @@ -168,7 +171,7 @@ def output_header_file( param ) param[:methods].each {|m| file.puts m[:if_exp].join if m[:if_exp] - file.puts " MRBC_SYMID_#{rename_for_symbol( strip_double_quot(m[:name]))}," + file.puts " MRBC_SYMID_#{rename_for_symbol(m[:name])}," m[:if_exp].size.times { file.puts "#endif" } if m[:if_exp] } file.puts " };" @@ -182,16 +185,13 @@ def output_header_file( param ) file.puts " };" file.puts - file.puts " return mrbc_define_builtin_class(#{param[:class]}, 0, method_symbols, method_functions, sizeof(method_symbols)/sizeof(mrbc_sym) );" + file.puts " return mrbc_define_builtin_class(#{param[:class]}, #{param[:super]}, method_symbols, method_functions, sizeof(method_symbols)/sizeof(mrbc_sym) );" file.puts "}" file.close end - - - ## # main # @@ -217,7 +217,11 @@ if !src exit 1 end +while src param = parse_source_string( src ) exit 1 if !param exit 1 if !check_error( param ) output_header_file( param ) + + src = get_method_table_source( file ) +end
BIO_do_accept: correct error return value `BIO_do_accept` was returning incorrect values when unable to bind to a port. Fixes CLA: trivial
@@ -222,19 +222,20 @@ static int acpt_state(BIO *b, BIO_ACCEPT *c) break; case ACPT_S_CREATE_SOCKET: - ret = BIO_socket(BIO_ADDRINFO_family(c->addr_iter), + s = BIO_socket(BIO_ADDRINFO_family(c->addr_iter), BIO_ADDRINFO_socktype(c->addr_iter), BIO_ADDRINFO_protocol(c->addr_iter), 0); - if (ret == (int)INVALID_SOCKET) { + if (s == (int)INVALID_SOCKET) { ERR_raise_data(ERR_LIB_SYS, get_last_socket_error(), "calling socket(%s, %s)", c->param_addr, c->param_serv); BIOerr(BIO_F_ACPT_STATE, BIO_R_UNABLE_TO_CREATE_SOCKET); goto exit_loop; } - c->accept_sock = ret; - b->num = ret; + c->accept_sock = s; + b->num = s; c->state = ACPT_S_LISTEN; + s = -1; break; case ACPT_S_LISTEN:
Add missing documentation of OSSL_FUNC_store_export_object()
@@ -105,6 +105,15 @@ further loading. OSSL_FUNC_store_close() frees the provider side context I<ctx>. +When a provider-native object is created by a store manager it would be unsuitable +for direct use with a foreign provider. The export function allows for +exporting the object to that foreign provider if the foreign provider +supports the type of the object and provides an import function. + +OSSL_FUNC_store_export_object() should export the object of size I<objref_sz> +referenced by I<objref> as an B<OSSL_PARAM> array and pass that to the +I<export_cb> as well as the given I<export_cbarg>. + =head2 Load Parameters =over 4
Just fsync those affected files during pg_rewind. In gpdb we've ensured that the database is in sync state before rewinding, so we could just fsync those affected files after rewinding.
@@ -812,6 +812,12 @@ digestControlFile(ControlFileData *ControlFile, char *src, size_t size) * most dirty buffers to disk. Additionally fsync_pgdata uses a two-pass * approach (only initiating writeback in the first pass), which often reduces * the overall amount of IO noticeably. + * + * gpdb: We assume that all files are synchronized before rewinding and thus we + * just need to synchronize those affected files. This is a resonable + * assumption for gpdb since we've ensured that the db state is clean shutdown + * in pg_rewind by running single mode postgres if needed and also we do not + * copy an unsynchronized dababase without sync as the target base. */ static void syncTargetDirectory(void) @@ -819,7 +825,59 @@ syncTargetDirectory(void) if (!do_sync || dry_run) return; - fsync_pgdata(datadir_target, PG_VERSION_NUM); + file_entry_t *entry; + int i; + + if (chdir(datadir_target) < 0) + { + pg_log_error("could not change directory to \"%s\": %m", datadir_target); + exit(1); + } + + for (i = 0; i < filemap->narray; i++) + { + entry = filemap->array[i]; + + if (entry->target_pages_to_overwrite.bitmapsize > 0) + fsync_fname(entry->path, false); + else + { + switch (entry->action) + { + case FILE_ACTION_COPY: + case FILE_ACTION_TRUNCATE: + case FILE_ACTION_COPY_TAIL: + fsync_fname(entry->path, false); + break; + + case FILE_ACTION_CREATE: + fsync_fname(entry->path, + entry->source_type == FILE_TYPE_DIRECTORY); + /* FALLTHROUGH */ + case FILE_ACTION_REMOVE: + /* + * Fsync the parent directory if we either create or delete + * files/directories in the parent directory. The parent + * directory might be missing as expected, so fsync it could + * fail but we ignore that error. + */ + fsync_parent_path(entry->path); + break; + + case FILE_ACTION_NONE: + break; + + default: + pg_fatal("no action decided for \"%s\"", entry->path); + break; + } + } + } + + /* fsync some files that are (possibly) written by pg_rewind. */ + fsync_fname("global/pg_control", false); + fsync_fname("backup_label", false); + fsync_fname("postgresql.auto.conf", false); } /*
Sockeye: Complete new AST for Frontend
@@ -17,58 +17,77 @@ module SockeyeASTFrontend where data SockeyeSpec = SockeyeSpec { modules :: [Module] - , nodeDecls :: [NodeDecl] + , net :: [NetSpec] } -data ParamType = Index | Address +data ParamType = IndexParam | AddressParam data ModuleParam = ModuleParam { paramName :: !String - , paramType :: !Maybe ParamType + , paramType :: Maybe ParamType } data Module = Module - { inputPorts :: [NodeId] - , outputPorts :: [NodeId] + { inputPorts :: [Identifier] + , outputPorts :: [Identifier] , parameters :: [ModuleParam] - , body :: [NodeDecl] + , body :: [NetSpec] } +data ModuleParamMap + = ModuleParamMap + { port :: Identifier + , nodeId :: Identifier + } + +data ModuleInstantiation + = ModuleInstantiation + { nameSpace :: Identifier + , arguments :: [Word] + , inputMappings :: [ModuleParamMap] + , outputMappings :: [ModuleParamMap] + } + +data NetSpec + = NodeDeclSpec NodeDecl + | ModuleInstSpec ModuleInstantiation + + + data NodeDecl = NodeDecl - { nodeId :: NodeId + { nodeIds :: [Identifier] , nodeSpec :: NodeSpec } -data NodeIdIndex = Index - { index :: !Word } - | Param - { name :: !String } +data IdentifierIndex + = NumberIndex !Word + | ParamIndex !String -data NodeId = Single - { prefix :: String } +data Identifier + = Single + { prefix :: !String } | Indexed { prefix :: !String } | Multi { prefix :: !String - , start :: NodeIdIndex - , end :: NodeIdIndex + , start :: IdentifierIndex + , end :: IdentifierIndex } data NodeType = Memory | Device data NodeSpec = NodeSpec - { nodeType :: !Maybe NodeType + { nodeType :: Maybe NodeType , accept :: [BlockSpec] , translate :: [MapSpec] - , overlay :: !String + , overlay :: Identifier } -data Address = Address - { address :: Word } - | Param - { name :: String } +data Address = Address !Word + | Param !String -data BlockSpec = Singleton +data BlockSpec + = Singleton { address :: !Address } | Range { base :: !Address @@ -79,11 +98,12 @@ data BlockSpec = Singleton , bits :: !Word } -data MapDest = Direct - { destNode :: !NodeId } +data MapDest + = Direct + { destNode :: Identifier } | BaseAddress - { destNode :: !NodeId - , destBase :: !Address + { destNode :: Identifier + , destBase :: Address } data MapSpec = MapSpec
sched/cpuload: Define variable for time constant as static variable Define g_cpuload_timeconstant as 'static' variable because this variable is used in sched_cpuload.c only
*/ volatile uint32_t g_cpuload_total[SCHED_NCPULOAD]; -volatile uint32_t g_cpuload_timeconstant[SCHED_NCPULOAD] = { +static volatile uint32_t g_cpuload_timeconstant[SCHED_NCPULOAD] = { #ifdef CONFIG_SCHED_MULTI_CPULOAD CONFIG_SCHED_CPULOAD_TIMECONSTANT_SHORT, CONFIG_SCHED_CPULOAD_TIMECONSTANT_MID,
fix (test-encrypt): remove async mode from non-async test in examples.cfg
@@ -81,8 +81,8 @@ test-ipsec@veth @veth ctr=off x_crypto async=off cryp ; -------------------------------------------------- ; Veth PCAP (input goes from pcap file) variant of the examples -encrypt-async@vethpcap @vethpcap ctr=off x_crypto async=pd crypto_burst_size=64 -test-ipsec@vethpcap @vethpcap ctr=off x_crypto async=pd crypto_burst_size=64 +encrypt-async@vethpcap @vethpcap ctr=off x_crypto async=pd crypto_burst_size=1 +test-ipsec@vethpcap @vethpcap ctr=off x_crypto async=pd crypto_burst_size=1 portfwd-fix@vethpcap @vethpcap ; -------------------------------------------------- @@ -163,7 +163,7 @@ test-trivial@test @nonic ctr=off test-varbit-2@test @nonic ctr=off test-hmac@test @nonic ctr=off x_crypto -test-encrypt@test @nonic ctr=off x_crypto async=pd crypto_burst_size=64 +test-encrypt@test @nonic ctr=off x_crypto test-gen-encrypt@test @nonic ctr=off x_crypto test-ipsec@test @nonic ctr=off x_crypto async=pd crypto_burst_size=1
Squashed 'opae-libs/' changes from c5f4c659..d1bc4cc9 opae-libs: change version to 2.0.10 git-subtree-dir: opae-libs git-subtree-split:
@@ -30,7 +30,7 @@ project(opae-libs) set(OPAE_VERSION_LOCAL "" CACHE STRING "OPAE local version") set(OPAE_VERSION_MAJOR 2 CACHE STRING "OPAE major version" FORCE) set(OPAE_VERSION_MINOR 0 CACHE STRING "OPAE minor version" FORCE) -set(OPAE_VERSION_REVISION 9${OPAE_VERSION_LOCAL} CACHE STRING "OPAE revision version" FORCE) +set(OPAE_VERSION_REVISION 10${OPAE_VERSION_LOCAL} CACHE STRING "OPAE revision version" FORCE) set(OPAE_VERSION ${OPAE_VERSION_MAJOR}.${OPAE_VERSION_MINOR}.${OPAE_VERSION_REVISION} CACHE STRING "OPAE version" FORCE)
increase Replaces version number
@@ -18,7 +18,7 @@ Build-Depends: debhelper-compat (= 13), check (>= 0.10.0), pkg-config (>= 0.29), libsecret-1-dev (>= 0.18.4), - libcjson-dev (>= 1.7.14), + libcjson-dev (>= 1.7.10-1.1), Package: oidc-agent-cli Architecture: any @@ -26,7 +26,7 @@ Depends: ${shlibs:Depends}, ${misc:Depends}, liboidc-agent4 (= ${binary:Version}) Suggests: qrencode, oidc-agent-desktop -Replaces: oidc-agent (<< 4.0.2-1) +Replaces: oidc-agent (<< 4.1.0-1) Description: Commandline tool for obtaining OpenID Connect Access tokens on the commandline This tool consists of five programs: - oidc-agent that handles communication with the OIDC provider @@ -65,7 +65,7 @@ Depends: ${misc:Depends}, oidc-agent-cli, xterm | x-terminal-emulator, yad -Replaces: oidc-agent (<< 4.0.2-1), oidc-agent-prompt (<< 4.0.2-1) +Replaces: oidc-agent (<< 4.1.0-1), oidc-agent-prompt (<< 4.1.0-1) Description: oidc-agent desktop integration Desktop integration files for oidc-gen and oidc-agent and for creating the user dialog.
BugID:17132041:fix bug Negative array index read
@@ -159,17 +159,17 @@ int uData_new_servicetask(const char *name, void (*fn)(void *),void *arg, int uData_observe_servicetask_tag(int taskid,sensor_tag_e tag) { - int actual_num = g_uData_own_task_tag[taskid].actual_num; int i = 0; - LOG("uData_set_servicetask_tag taskid=%d,actual_num=%d,tag=%d,g_uData_own_task_cnt=%d\n",taskid,actual_num,tag,g_uData_own_task_cnt); if(taskid >= g_uData_own_task_cnt || taskid < 0) { return -1; } + int actual_num = g_uData_own_task_tag[taskid].actual_num; if( actual_num >= UDATA_SERVICE_TAG_NUM) { return -1; } + LOG("uData_set_servicetask_tag taskid=%d,actual_num=%d,tag=%d,g_uData_own_task_cnt=%d\n",taskid,actual_num,tag,g_uData_own_task_cnt); for(i=0; i < actual_num; i++) { if(g_uData_own_task_tag[taskid].related_tag[i] == tag)
Disable Distribution Stage for Feature Branches
@@ -423,6 +423,6 @@ stages: dependsOn: - build_windows - build_linux - condition: and(in(variables['Build.Reason'], 'IndividualCI', 'BatchedCI'), succeeded()) + condition: and(in(variables['Build.Reason'], 'IndividualCI', 'BatchedCI'), not(startsWith(variables['Build.SourceBranch'], 'refs/heads/feature/')), succeeded()) jobs: - template: ./templates/build-distribution.yml
merge -s should only affect how things are merged, not reported
@@ -84,12 +84,6 @@ bool RecordOutputMgr::printKeyAndTerminate(RecordKeyVector &keyList) { //bed3 format, which is surprisingly difficult to do. Had to use the following: const Bed3Interval *bed3 = static_cast<const Bed3Interval *>(keyList.getKey()); bed3->Bed3Interval::print(_outBuf); - - //in addition, if we're doing stranded merges, we need to print the strand sign. - if (_context->getDesiredStrand() != FileRecordMergeMgr::ANY_STRAND) { - _outBuf.append("\t"); - _outBuf.append(keyList.getKey()->getStrand()); - } return false; } printBamType bamCode = printBamRecord(keyList);
scheduler: on timer callback, do not consume byte in OSX
@@ -44,7 +44,7 @@ static inline int consume_byte(flb_pipefd_t fd) /* We need to consume the byte */ ret = flb_pipe_r(fd, &val, sizeof(val)); if (ret <= 0) { - perror("read"); + flb_errno(); return -1; } @@ -318,7 +318,9 @@ int flb_sched_event_handler(struct flb_config *config, struct mk_event *event) } else if (timer->type == FLB_SCHED_TIMER_FRAME) { sched = timer->data; +#ifndef __APPLE__ consume_byte(sched->frame_fd); +#endif schedule_request_promote(sched); } else if (timer->type == FLB_SCHED_TIMER_CUSTOM) {
ci: travis: switch to bionic
-dist: xenial +dist: bionic language: c # The leak sanitizer uses ptrace, which doesn't work under the # travis container infrastructure, so we enable 'sudo' @@ -48,7 +48,7 @@ matrix: ci/do-ut || true - os: linux env: FLB_OPT="-DFLB_COVERAGE=On" - dist: xenial + dist: bionic sudo: true language: c compiler: gcc @@ -60,7 +60,7 @@ matrix: - os: linux services: - docker - dist: xenial + dist: bionic sudo: true language: c compiler: gcc
docs: Update skeleton README.md to support different PAL versions Use liberpal-skeleton-v${SKELETON_PAL_VERSION}.so to replace liberpal-skeleton.so in README.
@@ -14,7 +14,7 @@ Note that this step is only required when using SGX out-of-tree driver. ```shell cd "${path_to_inclavare_containers}/rune/libenclave/internal/runtime/pal/skeleton" make -cp liberpal-skeleton.so /usr/lib +cp liberpal-skeleton-v*.so /usr/lib ``` ## Build skeleton container image @@ -68,7 +68,7 @@ Runtimes: rune runc ```shell docker run -it --rm --runtime=rune \ -e ENCLAVE_TYPE=intelSgx \ - -e ENCLAVE_RUNTIME_PATH=/usr/lib/liberpal-skeleton.so \ + -e ENCLAVE_RUNTIME_PATH=/usr/lib/liberpal-skeleton-v${SKELETON_PAL_VERSION}.so \ -e ENCLAVE_RUNTIME_ARGS="debug" \ skeleton-enclave ``` @@ -111,7 +111,7 @@ In order to run the skeleton bundle with `rune`, you need to configure enclave r ```json "annotations": { "enclave.type": "intelSgx", - "enclave.runtime.path": "/usr/lib/liberpal-skeleton.so", + "enclave.runtime.path": "/usr/lib/liberpal-skeleton-v${SKELETON_PAL_VERSION}.so", "enclave.runtime.args": "debug" } ```