message
stringlengths 6
474
| diff
stringlengths 8
5.22k
|
---|---|
Fix compilation; | @@ -78,7 +78,7 @@ add_executable(lovr ${LOVR_SRC} ${LOVR_HEADSET})
if(EMSCRIPTEN)
set(CMAKE_EXECUTABLE_SUFFIX ".html")
- add_definitions(-s USE_ZLIB=1 -s ASSERTIONS=2 -s FULL_ES2=1 -O3 -s AGGRESSIVE_VARIABLE_ELIMINATION=1 -s OUTLINING_LIMIT=15000)
+ add_definitions(-s USE_ZLIB=1 -s ASSERTIONS=2 -s FULL_ES2=1 -O3 -s AGGRESSIVE_VARIABLE_ELIMINATION=1 -s OUTLINING_LIMIT=15000 -s NO_EXIT_RUNTIME=1)
set(ZLIB_FOUND 1)
set(ZLIB_LIBRARIES "-s USE_ZLIB=1")
set(ZLIB_INCLUDE_DIR .)
|
Add LDFLAGS to Lflags in generated janet.pc | @@ -285,7 +285,7 @@ $(PKG_CONFIG_PATH)/janet.pc: $(JANET_TARGET)
echo "Description: Library for the Janet programming language." >> $@
$(JANET_TARGET) -e '(print "Version: " janet/version)' >> $@
echo 'Cflags: -I$${includedir}' >> $@
- echo 'Libs: -L$${libdir} -ljanet' >> $@
+ echo 'Libs: -L$${libdir} -ljanet $(LDFLAGS)' >> $@
echo 'Libs.private: $(CLIBS)' >> $@
install: $(JANET_TARGET) $(PKG_CONFIG_PATH)/janet.pc
|
Add e_invincible_state enum. | @@ -137,6 +137,17 @@ typedef enum
AUTOKILL_ATTACK_HIT = (1 << 1)
} e_autokill_state;
+// Caskey, Damon V.
+// 2019-02-05
+typedef enum
+{
+ INVINCIBLE_NONE = 0,
+ INVINCIBLE_INTANGIBLE = (1 << 0),
+ INVINCIBLE_HP_MINIMUM = (1 << 1),
+ INVINCIBLE_HP_NULLIFY = (1 << 2),
+ INVINCIBLE_HP_RESET = (1 << 3)
+} e_invincible_state;
+
// Caskey, Damon V.
// 2019-01-25
//
@@ -1523,7 +1534,7 @@ if(n<1) n = 1;
(other->animation->vulnerable[other->animpos] && \
(!self->animation->move || self->animation->move[self->animpos]->axis.x == 0) && \
(!self->animation->move || self->animation->move[self->animpos]->axis.z == 0 ) && \
- !(other->nograb || other->invincible || other->link || \
+ !(other->nograb || other->invincible & INVINCIBLE_INTANGIBLE || other->link || \
other->model->animal || inair(other) || \
(self->modeldata.type == TYPE_PLAYER && other->modeldata.type == TYPE_PLAYER && savedata.mode)))
@@ -2504,7 +2515,6 @@ typedef struct entity
bool idling; // ~~
bool inbackpain; // Playing back pain/fall/rise/riseattack/die animation. ~~
bool inpain; // Hit and block stun. ~~
- bool invincible; // Can't be hit with attacks. ~~
bool jumping; // ~~
bool running; // ~~
bool turning; // Turning around. ~~
@@ -2515,6 +2525,7 @@ typedef struct entity
e_autokill_state autokill; // Kill entity on condition. ~~
e_duck_state ducking; // In or transitioning to/from duck. ~~
e_edge_state edge; // At an edge (unbalanced).
+ e_invincible_state invincible; // Attack invulnerability. ~~
e_direction normaldamageflipdir; // Used to reset backpain direction. ~~
e_rising_state rising; // Rise/Rise attacking. ~~
int seal; // If 0+, entity can't perform special with >= energy cost. ~~
|
VERSION bump to version 1.2.2 | @@ -27,7 +27,7 @@ endif()
# micro version is changed with a set of small changes or bugfixes anywhere in the project.
set(SYSREPO_MAJOR_VERSION 1)
set(SYSREPO_MINOR_VERSION 2)
-set(SYSREPO_MICRO_VERSION 1)
+set(SYSREPO_MICRO_VERSION 2)
set(SYSREPO_VERSION ${SYSREPO_MAJOR_VERSION}.${SYSREPO_MINOR_VERSION}.${SYSREPO_MICRO_VERSION})
# Version of the library
|
globally disable hashboard (pending actual hashes) | @@ -80,7 +80,10 @@ _main_getopt(c3_i argc, c3_c** argv)
u3_Host.ops_u.etn = c3n;
u3_Host.ops_u.gab = c3n;
u3_Host.ops_u.git = c3n;
- u3_Host.ops_u.has = c3n;
+ // always disable hashboard
+ // XX temporary, remove once hashes are added
+ //
+ u3_Host.ops_u.has = c3y;
u3_Host.ops_u.net = c3y;
u3_Host.ops_u.nuu = c3n;
u3_Host.ops_u.pro = c3n;
|
[P2P] Fix rarely occurred test failure
There was a timing bug in unit test. | @@ -295,9 +295,9 @@ func TestMultiLoads(t *testing.T) {
go consumeForLongterm(c, tt.testTime+time.Minute, doneC, finish)
wg := sync.WaitGroup{}
expire := time.Now().Add(tt.testTime)
+ wg.Add(threadsize)
for j := 0; j < threadsize; j++ {
go func(tid int) {
- wg.Add(1)
i := 0
for time.Now().Before(expire) {
c.In() <- mos[i%arrSize]
@@ -316,8 +316,8 @@ func TestMultiLoads(t *testing.T) {
lock.Unlock()
fmt.Printf("In %d , out %d , drop %d, consecutive drop %d\n", actStat.incnt, actStat.outcnt, actStat.dropcnt, actStat.consecdrop)
- // last one is in channel and not consumed
- assert.Equal(t, uint64(consumeCount+1), actStat.outcnt)
+ // There are two cases, one is last one is in channel and not consumed, and another is consumed all items.
+ assert.True(t, actStat.outcnt-uint64(consumeCount) <= 1 )
// in should equal to sum of out, drop, and remained in queue
assert.Equal(t, actStat.incnt, actStat.outcnt+actStat.dropcnt+uint64(rqueue.Size()))
|
hack/use-sbx-platform: add the steps for ubuntu | @@ -37,8 +37,12 @@ cd ../../QuoteVerification
make clean
USE_SBX_PLATFORM=1 make
# replace with modified QVL/QVE
+[ -d /usr/lib/x86_64-linux-gnu ] &&
+ sudo cp -f dcap_quoteverify/linux/libsgx_dcap_quoteverify.so.1 /usr/lib/x86_64-linux-gnu/libsgx_dcap_quoteverify.so.1.* ||
sudo cp -f dcap_quoteverify/linux/libsgx_dcap_quoteverify.so.1 /usr/lib64/libsgx_dcap_quoteverify.so.1.*
sudo ldconfig
+[ -d /usr/lib/x86_64-linux-gnu ] &&
+ sudo cp -f QvE/libsgx_qve.signed.so /usr/lib/x86_64-linux-gnu ||
sudo cp -f QvE/libsgx_qve.signed.so /usr/lib64
# replace the original
make clean -C dcap_tvl
|
Disable RHEL package builds since upstream is broken. | @@ -15,10 +15,10 @@ matrix:
include:
- env: PGB_CI="test --vm=u12"
- env: PGB_CI="test --vm=f30 --param=no-package --param=c-only"
- - env: PGB_CI="test --vm=co6 --param=module=mock --param=module=real"
+ - env: PGB_CI="test --vm=co6 --param=no-package --param=module=mock --param=module=real"
- env: PGB_CI="test --vm=u18 --param=container-only"
- env: PGB_CI=" doc --vm=u18"
- - env: PGB_CI="test --vm=co7 --param=module=mock --param=module=real"
+ - env: PGB_CI="test --vm=co7 --param=no-package --param=module=mock --param=module=real"
- dist: bionic
env:
- PGB_CI="test --vm=none --param=tz=America/New_York"
|
ssl_msg.c: fix parm in call to mbedtls_ssl_decrypt_buf() | @@ -1332,7 +1332,7 @@ int mbedtls_ssl_decrypt_buf( mbedtls_ssl_context const *ssl,
add_data, add_data_len,
data, rec->data_len + transform->taglen,
data, rec->buf_len - (data - rec->buf),
- &rec->data_len );
+ &olen );
if( status != PSA_SUCCESS )
return( psa_status_to_mbedtls( status ) );
|
increase MAC range from +-0x10 to +-0x20 | @@ -2164,8 +2164,8 @@ static char pskstring[PSKSTRING_LEN_MAX] = {};
me = macaddr &0xffffff;
fprintf(fhout, "05%6d\n", me);
oui = macaddr &0xffffff000000L;
-nic = (macaddr -0x10) &0xffffffL;
-for(c = 0; c <= 0x10; c++) writebssid(fhout, oui +((nic +c) &0xffffffL));
+nic = (macaddr -0x20) &0xffffffL;
+for(c = 0; c <= 0x20; c++) writebssid(fhout, oui +((nic +c) &0xffffffL));
swap = (nic >> 8) & 0xffff;
{
swap = (swap & 0xf000) >> 12 | (swap & 0x0f00) >> 4 | (swap & 0x00f0) << 4 | (swap & 0x000f) << 12;
|
Removed unnecessary operations in ++so-resub. | (so-delta-our %follow | [src ~ ~])
=- (so-delta-our %follow & [[cir.src -] ~ ~])
^- range
- ?. (~(has by sequence) cir.src) ran.src
- =- `[[%ud (~(got by sequence) cir.src)] -]
+ ?~ seq ran.src
+ =- `[[%ud u.seq] -]
?~ ran.src ~
tal.u.ran.src
::
|
Remove the #ifdef ANDROID blocks in LogManagerProvider.hpp, as even on android they're not used. There's a better design for this. | @@ -40,11 +40,6 @@ namespace ARIASDK_NS_BEGIN
ILogConfiguration& cfg,
status_t& status,
IHttpClient* httpClient,
-#ifdef ANDROID
- JNIEnv *env,
- jclass contextClass,
- jobject contextObject,
-#endif
uint64_t targetVersion = MAT::Version)
{
cfg["name"] = id;
@@ -76,11 +71,6 @@ namespace ARIASDK_NS_BEGIN
char const* id,
bool wantController,
status_t& status,
-#ifdef ANDROID
- JNIEnv *env,
- jclass contextClass,
- jobject contextObject,
-#endif
uint64_t targetVersion = MAT::Version)
{
UNREFERENCED_PARAMETER(targetVersion);
@@ -89,12 +79,6 @@ namespace ARIASDK_NS_BEGIN
return Get(
id,
status
-#ifdef ANDROID
- ,
- env,
- contextClass,
- contextObject
-#endif
);
};
#endif
@@ -107,11 +91,6 @@ namespace ARIASDK_NS_BEGIN
/// </summary>
static ILogManager* MATSDK_SPEC CreateLogManager(char const* id,
status_t& status,
-#ifdef ANDROID
- JNIEnv *env,
- jclass contextClass,
- jobject contextObject,
-#endif
uint64_t targetVersion = MAT::Version)
{
UNREFERENCED_PARAMETER(targetVersion);
@@ -119,12 +98,6 @@ namespace ARIASDK_NS_BEGIN
return Get(
id,
status
-#ifdef ANDROID
- ,
- env,
- contextClass,
- contextObject
-#endif
);
}
@@ -174,21 +147,11 @@ namespace ARIASDK_NS_BEGIN
ILogConfiguration & cfg,
status_t &status,
IHttpClient* httpClient
-#ifdef ANDROID
- , JNIEnv * env
- , jclass contextClass
- , jobject contextObject
-#endif
);
static ILogManager* MATSDK_SPEC Get(
const char * id,
status_t& status
-#ifdef ANDROID
- JNIEnv *env,
- jclass contextClass,
- jobject contextObject,
-#endif
);
};
|
Add feature support for AES for PSA crypto config
Basic support for enabling AES through PSA_WANT_KEY_TYPE_AES. | @@ -220,6 +220,12 @@ extern "C" {
#endif /* !MBEDTLS_PSA_ACCEL_KEY_TYPE_RSA_PUBLIC_KEY */
#endif /* PSA_WANT_KEY_TYPE_RSA_PUBLIC_KEY */
+#if defined(PSA_WANT_KEY_TYPE_AES)
+#if !defined(MBEDTLS_PSA_ACCEL_KEY_TYPE_AES)
+#define MBEDTLS_AES_C
+#endif /* !MBEDTLS_PSA_ACCEL_KEY_TYPE_AES */
+#endif /* PSA_WANT_KEY_TYPE_AES */
+
#else /* MBEDTLS_PSA_CRYPTO_CONFIG */
/*
@@ -328,6 +334,10 @@ extern "C" {
#define PSA_WANT_ALG_SHA_512 1
#endif
+#if defined(MBEDTLS_AES_C)
+#define PSA_WANT_KEY_TYPE_AES 1
+#endif
+
#endif /* MBEDTLS_PSA_CRYPTO_CONFIG */
#ifdef __cplusplus
|
Validate parser parameter to XML_SetStartDoctypeDeclHandler | @@ -1485,6 +1485,7 @@ XML_SetDoctypeDeclHandler(XML_Parser parser,
void XMLCALL
XML_SetStartDoctypeDeclHandler(XML_Parser parser,
XML_StartDoctypeDeclHandler start) {
+ if (parser != NULL)
startDoctypeDeclHandler = start;
}
|
OpenPartitionDxe: Fix calculation | @@ -588,7 +588,7 @@ PartitionCheckGptEntryArrayCRC (
return FALSE;
}
- Size = PartHeader->NumberOfPartitionEntries * PartHeader->SizeOfPartitionEntry;
+ Size = ((UINTN) PartHeader->NumberOfPartitionEntries) * PartHeader->SizeOfPartitionEntry;
Status = gBS->CalculateCrc32 (Ptr, Size, &Crc);
if (EFI_ERROR (Status)) {
|
EFAS: rename of key | @@ -20,13 +20,13 @@ unsigned[1] monthOfForecast = month : dump;
unsigned[1] dayOfForecast = day : dump;
unsigned[1] hourOfForecast = hour : dump;
unsigned[1] minuteOfForecast = minute : dump;
-constant secondsOfForecast = 0;
+constant secondOfForecast = 0;
meta dateOfForecast g2date(yearOfForecast,monthOfForecast, dayOfForecast) : dump;
-meta timeOfForecast time (hourOfForecast,minuteOfForecast,secondsOfForecast) : dump;
+meta timeOfForecast time (hourOfForecast,minuteOfForecast,secondOfForecast) : dump;
# Calculate the Julian number for the forecast date and time.
# This will be a floating point number with units of 'day'
-meta julianForecastDay julian_day(dateOfForecast,hourOfForecast,minuteOfForecast,secondsOfForecast): hidden;
+meta julianForecastDay julian_day(dateOfForecast,hourOfForecast,minuteOfForecast,secondOfForecast): hidden;
# Calculate the difference between the forecast date and reference date
transient diffInDays = (julianForecastDay - julianDay) : hidden; # float
@@ -62,16 +62,16 @@ codetable[1] efasDomain "grib2/tables/local/ecmf/efas_domain.table" = 255 : dump
# "global" = { efas_domain = 1; }
#} : hidden;
-# Climatology Date and Time
-# The climatology valid date is the date of the new official implementation of the EFAS cycle.
-unsigned[2] yearOfClimatology = year : dump;
-unsigned[1] monthOfClimatology = month : dump;
-unsigned[1] dayOfClimatology = day : dump;
-unsigned[1] hourOfClimatology = hour : dump;
-unsigned[1] minuteOfClimatology = minute : dump;
-constant secondsOfClimatology = 0;
-meta dateOfClimatology g2date(yearOfClimatology,monthOfClimatology, dayOfClimatology) : dump;
-meta timeOfClimatology time (hourOfClimatology,minuteOfClimatology,secondsOfClimatology) : dump;
+# Reanalysis Date and Time
+# This date is the date of the new official implementation of the EFAS cycle.
+unsigned[2] yearOfReanalysis = year : dump;
+unsigned[1] monthOfReanalysis = month : dump;
+unsigned[1] dayOfReanalysis = day : dump;
+unsigned[1] hourOfReanalysis = hour : dump;
+unsigned[1] minuteOfReanalysis = minute : dump;
+constant secondOfReanalysis = 0;
+meta dateOfReanalysis g2date(yearOfReanalysis,monthOfReanalysis, dayOfReanalysis) : dump;
+meta timeOfReanalysis time (hourOfReanalysis,minuteOfReanalysis,secondOfReanalysis) : dump;
# Note: the key inputOriginatingCentre is in the PDTNs 70, 71, 72 and 73
#concept efas_forecast {
|
mdns-example: fail gracefully if mdns response not received within timeout
If mdns answer hasn't been received within timeout, Value error would be raised, but the mdns-server-thread would still run, blocking CI jobs. Fixed by moving the raise statement within try-finally block | @@ -117,11 +117,11 @@ def test_examples_protocol_mdns(env, extra_data):
stop_mdns_server.set()
thread1.join()
raise ValueError('ENV_TEST_FAILURE: Cannot connect to AP')
+ try:
# 3. check the mdns name is accessible
if not esp_answered.wait(timeout=30):
raise ValueError('Test has failed: did not receive mdns answer within timeout')
# 4. check DUT output if mdns advertized host is resolved
- try:
dut1.expect(re.compile(r"mdns-test: Query A: tinytester.local resolved to: 127.0.0.1"), timeout=30)
dut1.expect(re.compile(r"mdns-test: gethostbyname: tinytester-lwip.local resolved to: 127.0.0.1"), timeout=30)
dut1.expect(re.compile(r"mdns-test: getaddrinfo: tinytester-lwip.local resolved to: 127.0.0.1"), timeout=30)
|
nimble/l2cap: Fix return error code
os_mbuf_foo() returns OS_ERR codes and we want to return BLE_HS_ERR
codes.
This patch fixes it | @@ -440,6 +440,7 @@ ble_l2cap_coc_continue_tx(struct ble_l2cap_chan *chan)
BLE_HS_LOG(DEBUG, "Sending SDU len=%d\n", OS_MBUF_PKTLEN(tx->sdu));
rc = os_mbuf_append(txom, &l, sizeof(uint16_t));
if (rc) {
+ rc = BLE_HS_ENOMEM;
BLE_HS_LOG(DEBUG, "Could not append data rc=%d", rc);
goto failed;
}
@@ -452,6 +453,7 @@ ble_l2cap_coc_continue_tx(struct ble_l2cap_chan *chan)
rc = os_mbuf_appendfrom(txom, tx->sdu, tx->data_offset,
len - sdu_size_offset);
if (rc) {
+ rc = BLE_HS_ENOMEM;
BLE_HS_LOG(DEBUG, "Could not append data rc=%d", rc);
goto failed;
}
|
Misra 15_7: fix what seems to be a false positive of cppcheck | @@ -320,6 +320,8 @@ int usb_cb_control_msg(USB_Setup_TypeDef *setup, uint8_t *resp, bool hardwired)
// if the ignition hook depends on something other than the started GPIO
// we have to disable power savings (fix for GM and Tesla)
set_power_save_state(POWER_SAVE_STATUS_DISABLED);
+ } else {
+ // power mode is already POWER_SAVE_STATUS_DISABLED and CAN TXs are active
}
can_init_all();
}
|
docs: update wifi country code docs. | @@ -1365,13 +1365,7 @@ The following table depicts which country info is used in different Wi-Fi modes
For scan:
- - If schan+nchan-1 >11 :
-
- Use active scan from schan to 11 and use passive scan from 12 to 14.
-
- - If schan+nchan-1 <= 11 :
-
- Use active scan from schan to schan+nchan-1 and use passive scan from schan+nchan to 14.
+ Use active scan from 1 to 11 and use passive scan from 12 to 14.
Always keep in mind that if an AP with hidden SSID and station is set to a passive scan channel, the passive scan will not find it. In other words, if the application hopes to find the AP with hidden SSID in every channel, the policy of country info should be configured to WIFI_COUNTRY_POLICY_MANUAL.
@@ -1381,24 +1375,25 @@ The following table depicts which country info is used in different Wi-Fi modes
For scan:
- - If schan+nchan-1 >11 :
-
- Use active scan from schan to 11 and use passive scan from 12 to schan+nchan-1.
-
- - If schan+nchan-1 <= 11 :
-
Use active scan from schan to schan+nchan-1.
+
* - AP
- WIFI_COUNTRY_POLICY_AUTO
- Always use the configured country info.
+
* - AP
- WIFI_COUNTRY_POLICY_MANUAL
- Always use the configured country info.
+
* - Station/AP-coexistence
- WIFI_COUNTRY_POLICY_AUTO
- - If the station does not connect to any external AP, the AP uses the configured country info. If the station connects to an external AP, the AP has the same country info as the station.
+ - Station: Same as station mode with policy WIFI_COUNTRY_POLICY_AUTO.
+ AP: If the station does not connect to any external AP, the AP uses the configured country info. If the station connects to an external AP, the AP has the same country info as the station.
- Same as station mode with policy WIFI_COUNTRY_POLICY_AUTO.
+ * - Station/AP-coexistence
+ - WIFI_COUNTRY_POLICY_MANUAL
+ - Station: Same as station mode with policy WIFI_COUNTRY_POLICY_MANUAL.
+ AP: Same as AP mode with policy WIFI_COUNTRY_POLICY_MANUAL.
Home Channel
|
add debsource as a requirement to deb | @@ -775,7 +775,7 @@ ubuntu-bionic-source: distclean preparedeb
@rm debian/oidc-agent-desktop.triggers
.PHONY: deb
-deb: cleanapi create_obj_dir_structure preparedeb
+deb: cleanapi create_obj_dir_structure preparedeb debsource
debuild -i -b -uc -us
@echo "Success: DEBs are in parent directory"
|
Use PSA_BLOCK_CIPHER_BLOCK_LENGTH instead of PSA_BLOCK_CIPHER_BLOCK_MAX_SIZE in ssl_tls12_populate_transform() | @@ -7103,7 +7103,7 @@ static int ssl_tls12_populate_transform( mbedtls_ssl_transform *transform,
ssl_mode == MBEDTLS_SSL_MODE_CBC_ETM )
{
#if defined(MBEDTLS_USE_PSA_CRYPTO)
- size_t block_size = PSA_BLOCK_CIPHER_BLOCK_MAX_SIZE;
+ size_t block_size = PSA_BLOCK_CIPHER_BLOCK_LENGTH( key_type );
#else
size_t block_size = cipher_info->block_size;
#endif /* MBEDTLS_USE_PSA_CRYPTO */
|
vlib: always replace bootstrap vlib_mains
Fixes an obscure issue when spinning up pure pthreads which don't have
vlib_main_t clones.
VLIB_REGISTER_THREAD () = {
<snip>
.use_pthreads = 1,
.no_data_structure_clone = 1,
}
Type: improvement | @@ -707,16 +707,15 @@ start_workers (vlib_main_t * vm)
clib_callback_data_init (&vm->vlib_node_runtime_perf_callbacks,
&vm->worker_thread_main_loop_callback_lock);
- if (n_vlib_mains > 1)
- {
/* Replace hand-crafted length-1 vector with a real vector */
vlib_mains = 0;
- vec_validate_aligned (vlib_mains, tm->n_vlib_mains - 1,
- CLIB_CACHE_LINE_BYTES);
+ vec_validate_aligned (vlib_mains, n_vlib_mains - 1, CLIB_CACHE_LINE_BYTES);
_vec_len (vlib_mains) = 0;
vec_add1_aligned (vlib_mains, vm, CLIB_CACHE_LINE_BYTES);
+ if (n_vlib_mains > 1)
+ {
vlib_worker_threads->wait_at_barrier =
clib_mem_alloc_aligned (sizeof (u32), CLIB_CACHE_LINE_BYTES);
vlib_worker_threads->workers_at_barrier =
|
type: small fix in docu | @@ -21,7 +21,7 @@ provides to other programming languages.
The type checker plugin supports all basic CORBA types:
`short`, `unsigned_short`, `long`, `unsigned_long`, `long_long`,
`unsigned_long_long`, `float`, `double`, `char`, `boolean`, `any` and
-`octet`. In Elektra `octet` is the same as `char`. When checking any it
+`octet`. In Elektra `octet` is the same as `char`. When checking `any` it
will always be successful, regardless of the content. Elektra also added
other types. `empty` will only yield true if there is no value. `string`
allows any non-empty sequence of octets.
|
Jenkinsfile: fix link to jenkins | * A Snippet generator is available to the public at
* https://qa.nuxeo.org/jenkins/pipeline-syntax/.
* A list of available commands on the build server can be found after a login at
- * https://build.libelektra.org/job/elektra-jenkinsfile/pipeline-syntax/.
+ * https://build.libelektra.org/job/libelektra/pipeline-syntax/.
*/
// TODO have a per plugin/binding deps in Dockerfile for easier maintenance
|
Add 2/3/4 parition index negative tests | @@ -1777,9 +1777,9 @@ class CLINTest(CLITestBase):
# Run the command, incrementally omitting arguments
self.exec_with_omit(command, 7)
- def test_cl_partitionlimit_missing_args(self):
+ def test_cl_2partitionlimit_missing_args(self):
"""
- Test -cl with -partitionindexlimit and missing arguments.
+ Test -cl with -2partitionindexlimit and missing arguments.
"""
# Build a valid command
command = [
@@ -1787,7 +1787,37 @@ class CLINTest(CLITestBase):
self.get_ref_image_path("LDR", "input", "A"),
self.get_tmp_image_path("LDR", "comp"),
"4x4", "-fast",
- "-partitionindexlimit", "3"]
+ "-2partitionindexlimit", "3"]
+
+ # Run the command, incrementally omitting arguments
+ self.exec_with_omit(command, 7)
+
+ def test_cl_3partitionlimit_missing_args(self):
+ """
+ Test -cl with -3partitionindexlimit and missing arguments.
+ """
+ # Build a valid command
+ command = [
+ self.binary, "-cl",
+ self.get_ref_image_path("LDR", "input", "A"),
+ self.get_tmp_image_path("LDR", "comp"),
+ "4x4", "-fast",
+ "-3partitionindexlimit", "3"]
+
+ # Run the command, incrementally omitting arguments
+ self.exec_with_omit(command, 7)
+
+ def test_cl_4partitionlimit_missing_args(self):
+ """
+ Test -cl with -4partitionindexlimit and missing arguments.
+ """
+ # Build a valid command
+ command = [
+ self.binary, "-cl",
+ self.get_ref_image_path("LDR", "input", "A"),
+ self.get_tmp_image_path("LDR", "comp"),
+ "4x4", "-fast",
+ "-4partitionindexlimit", "3"]
# Run the command, incrementally omitting arguments
self.exec_with_omit(command, 7)
|
driver/ioexpander/pca9675.c: Format with clang-format
BRANCH=none
TEST=none | @@ -21,8 +21,8 @@ static struct pca9675_ioexpander pca9675_iox[CONFIG_IO_EXPANDER_PORT_COUNT];
static int pca9675_read16(int ioex, uint16_t *data)
{
return i2c_xfer(ioex_config[ioex].i2c_host_port,
- ioex_config[ioex].i2c_addr_flags,
- NULL, 0, (uint8_t *)data, 2);
+ ioex_config[ioex].i2c_addr_flags, NULL, 0,
+ (uint8_t *)data, 2);
}
static int pca9675_write16(int ioex, uint16_t data)
@@ -35,22 +35,21 @@ static int pca9675_write16(int ioex, uint16_t data)
data |= pca9675_iox[ioex].io_direction;
return i2c_xfer(ioex_config[ioex].i2c_host_port,
- ioex_config[ioex].i2c_addr_flags,
- (uint8_t *)&data, 2, NULL, 0);
+ ioex_config[ioex].i2c_addr_flags, (uint8_t *)&data, 2,
+ NULL, 0);
}
static int pca9675_reset(int ioex)
{
uint8_t reset = PCA9675_RESET_SEQ_DATA;
- return i2c_xfer(ioex_config[ioex].i2c_host_port,
- 0, &reset, 1, NULL, 0);
+ return i2c_xfer(ioex_config[ioex].i2c_host_port, 0, &reset, 1, NULL, 0);
}
static int pca9675_get_flags_by_mask(int ioex, int port, int mask, int *flags)
{
- *flags = mask & pca9675_iox[ioex].io_direction ?
- GPIO_INPUT : GPIO_OUTPUT;
+ *flags = mask & pca9675_iox[ioex].io_direction ? GPIO_INPUT :
+ GPIO_OUTPUT;
return EC_SUCCESS;
}
|
reverted there is a better solution | @@ -211,7 +211,6 @@ public:
, Func_(func)
, Arg_(arg)
, Name_(name)
- , SwitchCount_(0)
, Cancelled_(false)
, Scheduled_(false)
{
@@ -226,10 +225,7 @@ public:
inline void SwitchTo(TCont* next) noexcept {
DBGOUT(PCORO(this) << " switch to " << PCORO(next));
- if (next == this) {
- return;
- }
- SwitchCount_ += 1;
+
Context()->SwitchTo(next->Context());
}
@@ -265,10 +261,6 @@ public:
return Name_;
}
- inline ui64 GetSwitchCount() const noexcept {
- return SwitchCount_;
- }
-
void PrintMe(IOutputStream& out) const noexcept;
inline void Yield() noexcept {
@@ -487,7 +479,6 @@ private:
void* Arg_;
const char* Name_;
TIntrusiveList<TJoinWait> Waiters_;
- ui64 SwitchCount_;
bool Cancelled_;
bool Scheduled_;
};
|
Update WebTransport requirements | @@ -221,9 +221,9 @@ static X509Certificate2 CreatePkcs12FromPem(string certPem, string keyPem)
## Using a self-signed certificate for HTTP/3
-Chromium-based browsers requires the server certificate to be trusted by a default CA for QUIC (e.g. HTTP/3 and WebTransport), even though the same certificate may already be trusted for HTTP/1.1 and HTTP/2. To use a self-signed certificate or a certificate that is not ultimately issued by one of the default CAs, you need to white list its fingerprint (or that of any certificate in the chain) via the `--ignore-certificate-errors-spki-list` switch.
+Chromium-based browsers requires the server certificate to be trusted by a default CA for QUIC (e.g. HTTP/3 and WebTransport), even though the same certificate may already be trusted for HTTP/1.1 and HTTP/2. To use a self-signed certificate or a certificate that is not ultimately issued by one of the default CAs, you need to whitelist its SHA-256 hash via the [serverCertificateHashes](https://w3c.github.io/webtransport/#dom-webtransportoptions-servercertificatehashes) option and follow stricter [requirements](https://w3c.github.io/webtransport/#custom-certificate-requirements).
-See [Chromium network switches](https://source.chromium.org/chromium/chromium/src/+/main:services/network/public/cpp/network_switches.cc;l=36;drc=f8c933c2bd17344ce7ac61be2ac7725ed840b19f)
+See [FlyByWireless.CustomCertificate.Generate()](https://github.com/wegylexy/webtransport/blob/c55d9cc5a11f3a8b8dfd2a12c8d02ad462dc693d/Server/CustomCertificate.cs#L8-L33) on how to generate such a certificate.
## Collecting a Packet Capture
|
Samples:Host Exerciser Enhancement
Fixed the parse error with unit test case | @@ -159,7 +159,7 @@ TEST_P(fpgaperf_counter_c_p, fpgaperf_3) {
FILE *f = stdout;
EXPECT_EQ(fpgaPerfCounterPrint(f, fpga_perf), FPGA_OK);
- EXPECT_EQ(fpgaPerfCounterPrint(NULL), FPGA_INVALID_PARAM);
+ EXPECT_EQ(fpgaPerfCounterPrint(f, NULL), FPGA_INVALID_PARAM);
}
/**
|
bootloader: dont print assert msg if CONFIG_OPTIMIZATION_ASSERTIONS_SILENT is set
Closes | void __assert_func(const char *file, int line, const char *func, const char *expr)
{
+
+#if !CONFIG_OPTIMIZATION_ASSERTIONS_SILENT
esp_rom_printf("Assert failed in %s, %s:%d (%s)\r\n", func, file, line, expr);
+#endif
+
while (1) {
}
}
|
Fix incorrect SO_RCVLOWAT usage
SO_RCVLOWAT is SOL_SOCKET option, not IPPROTO_TCP [1].
[1] | @@ -84,7 +84,7 @@ int s2n_socket_read_snapshot(struct s2n_connection *conn)
struct s2n_socket_read_io_context *r_io_ctx = (struct s2n_socket_read_io_context *) conn->recv_io_context;
notnull_check(r_io_ctx);
- getsockopt(r_io_ctx->fd, IPPROTO_TCP, SO_RCVLOWAT, &r_io_ctx->original_rcvlowat_val, &watlen);
+ getsockopt(r_io_ctx->fd, SOL_SOCKET, SO_RCVLOWAT, &r_io_ctx->original_rcvlowat_val, &watlen);
eq_check(watlen, sizeof(int));
r_io_ctx->original_rcvlowat_is_set = 1;
#endif
@@ -117,7 +117,7 @@ int s2n_socket_read_restore(struct s2n_connection *conn)
if (!r_io_ctx->original_rcvlowat_is_set) {
return 0;
}
- setsockopt(r_io_ctx->fd, IPPROTO_TCP, SO_RCVLOWAT, &r_io_ctx->original_rcvlowat_val, sizeof(r_io_ctx->original_rcvlowat_val));
+ setsockopt(r_io_ctx->fd, SOL_SOCKET, SO_RCVLOWAT, &r_io_ctx->original_rcvlowat_val, sizeof(r_io_ctx->original_rcvlowat_val));
r_io_ctx->original_rcvlowat_is_set = 0;
#endif
@@ -173,7 +173,7 @@ int s2n_socket_set_read_size(struct s2n_connection *conn, int size)
struct s2n_socket_read_io_context *r_io_ctx = (struct s2n_socket_read_io_context *) conn->recv_io_context;
notnull_check(r_io_ctx);
- setsockopt(r_io_ctx->fd, IPPROTO_TCP, SO_RCVLOWAT, &size, sizeof(size));
+ setsockopt(r_io_ctx->fd, SOL_SOCKET, SO_RCVLOWAT, &size, sizeof(size));
#endif
return 0;
|
[CUDA] Implement update_event to fix timer race condition | @@ -142,6 +142,7 @@ pocl_cuda_init_device_ops (struct pocl_device_ops *ops)
ops->submit = pocl_cuda_submit;
ops->notify = pocl_cuda_notify;
ops->wait_event = pocl_cuda_wait_event;
+ ops->update_event = pocl_cuda_update_event;
ops->free_event_data = pocl_cuda_free_event_data;
ops->join = pocl_cuda_join;
ops->flush = pocl_cuda_flush;
@@ -158,7 +159,6 @@ pocl_cuda_init_device_ops (struct pocl_device_ops *ops)
/* TODO: implement remaining ops functions if needed: */
/* broadcast */
/* get_timer_value */
- /* update_event */
}
cl_int
@@ -1327,14 +1327,31 @@ pocl_cuda_finalize_command (cl_device_id device, cl_event event)
return;
}
- /* Retain this event temporarily, otherwise the call to
- * POCL_UPDATE_EVENT_COMPLETE might release the event before we've
- * dealt with the timing.
- * TODO: Maybe better to handle update_event so we don't have to do this? */
- POname (clRetainEvent) (event);
-
POCL_UPDATE_EVENT_RUNNING (&event);
POCL_UPDATE_EVENT_COMPLETE (&event);
+}
+
+void
+pocl_cuda_update_event (cl_device_id device, cl_event event, cl_int status)
+{
+ switch (status)
+ {
+ case CL_QUEUED:
+ if (event->queue->properties & CL_QUEUE_PROFILING_ENABLE)
+ event->time_queue = device->ops->get_timer_value (device->data);
+ event->status = status;
+ break;
+ case CL_SUBMITTED:
+ if (event->queue->properties & CL_QUEUE_PROFILING_ENABLE)
+ event->time_submit = device->ops->get_timer_value (device->data);
+ event->status = status;
+ break;
+ case CL_RUNNING:
+ /* Wait until complete to get the timing from the CUDA events */
+ event->status = status;
+ break;
+ case CL_COMPLETE:
+ pocl_mem_objs_cleanup (event);
/* Update timing info with CUDA event timers if profiling enabled */
if (event->queue->properties & CL_QUEUE_PROFILING_ENABLE)
@@ -1350,6 +1367,9 @@ pocl_cuda_finalize_command (cl_device_id device, cl_event event)
* sane. */
float diff;
+ CUresult result;
+ pocl_cuda_event_data_t *event_data
+ = (pocl_cuda_event_data_t *)event->data;
cl_ulong epoch = ((pocl_cuda_device_data_t *)device->data)->epoch;
result = cuEventElapsedTime (
@@ -1367,7 +1387,18 @@ pocl_cuda_finalize_command (cl_device_id device, cl_event event)
event->time_end = max (event->time_end, event->time_start + 1);
}
- POname (clReleaseEvent) (event);
+ POCL_LOCK_OBJ (event);
+ event->status = CL_COMPLETE;
+ device->ops->broadcast (event);
+ POCL_UNLOCK_OBJ (event);
+
+ pocl_update_command_queue (event);
+
+ break;
+ default:
+ assert ("Invalid event status\n");
+ break;
+ }
}
void
|
Fix broken CLANG build. No module clang/3.6.2 | @@ -5,4 +5,6 @@ if [[ ! $(command -v module > /dev/null 2>&1) ]]; then
. /usr/local/apps/module/init/bash
fi
module unload grib_api
-module switch gnu clang/3.6.2
+#module switch gnu clang/3.6.2
+#module switch gnu clang/3.9.1
+module switch gnu clang
|
Missing reserve | @@ -121,6 +121,7 @@ namespace carto {
std::size_t edgeVertexCount = std::accumulate(ringsInternalPoses.begin(), ringsInternalPoses.end(), std::size_t(0), [](std::size_t count, const std::vector<MapPos>& ringInternalPoses) { return count + ringInternalPoses.size(); });
_coords.reserve(roofIndices.size() + edgeVertexCount * 6);
_normals.reserve(roofIndices.size() + edgeVertexCount * 6);
+ _attribs.reserve(roofIndices.size() + edgeVertexCount * 6);
// Convert triangulator output to coord array
for (std::size_t i = 0; i < roofIndices.size(); i++) {
|
Fix runas dialog default username | @@ -795,12 +795,24 @@ INT_PTR CALLBACK PhpRunAsDlgProc(
if (!context->ProcessId)
{
- SetWindowText(
+ PPH_STRING runAsUserName = PhaGetStringSetting(L"RunAsUserName");
+ INT runAsUserNameIndex = CB_ERR;
+
+ // Fire the user name changed event so we can fix the logon type.
+ if (!PhIsNullOrEmptyString(runAsUserName))
+ {
+ runAsUserNameIndex = ComboBox_FindString(
context->UserComboBoxWindowHandle,
- PH_AUTO_T(PH_STRING, PhGetStringSetting(L"RunAsUserName"))->Buffer
+ 0,
+ PhGetString(runAsUserName)
);
+ }
+
+ if (runAsUserNameIndex != CB_ERR)
+ ComboBox_SetCurSel(context->UserComboBoxWindowHandle, runAsUserNameIndex);
+ else
+ ComboBox_SetCurSel(context->UserComboBoxWindowHandle, 0);
- // Fire the user name changed event so we can fix the logon type.
SendMessage(hwndDlg, WM_COMMAND, MAKEWPARAM(IDC_USERNAME, CBN_EDITCHANGE), 0);
}
else
|
unlink recorded file end of test to improve readi test's performance | @@ -764,12 +764,9 @@ static int audio_tc_launcher(int argc, char **args)
utc_audio_pcm_readi_n();
utc_audio_pcm_writei_p();
utc_audio_pcm_writei_n();
-#if 0
- utc_audio_pcm_start_p();
- utc_audio_pcm_start_n();
- utc_audio_pcm_stop_p();
- utc_audio_pcm_stop_n();
-#endif
+
+ unlink(AUDIO_TEST_FILE);
+
printf("#########################################\n");
printf(" FS TC Result \n");
printf(" PASS : %d FAIL : %d \n", total_pass, total_fail);
|
Cellular, LARA-R6, revert temporary change, i.e. bring back RoT.
Commit temporarily removed RoT from LARA-R6's features; RoT is now back. | @@ -278,7 +278,7 @@ const uCellPrivateModule_t gUCellPrivateModuleList[] = {
(1ULL << (int32_t) U_CELL_NET_RAT_UTRAN)) /* RATs */,
((1ULL << (int32_t) U_CELL_PRIVATE_FEATURE_MNO_PROFILE) |
(1ULL << (int32_t) U_CELL_PRIVATE_FEATURE_CSCON) |
- //(1ULL << (int32_t) U_CELL_PRIVATE_FEATURE_ROOT_OF_TRUST) |
+ (1ULL << (int32_t) U_CELL_PRIVATE_FEATURE_ROOT_OF_TRUST) |
(1ULL << (int32_t) U_CELL_PRIVATE_FEATURE_SECURITY_TLS_IANA_NUMBERING) |
(1ULL << (int32_t) U_CELL_PRIVATE_FEATURE_SECURITY_TLS_SERVER_NAME_INDICATION) |
(1ULL << (int32_t) U_CELL_PRIVATE_FEATURE_MQTT) |
|
test: add ethereum test case
test_001CreateWallet_0008CreateOneTimeWalletFailureProtocolUnknown | @@ -214,6 +214,23 @@ START_TEST(test_001CreateWallet_0007CreateOneTimeWalletSuccessLongSize)
}
END_TEST
+START_TEST(test_001CreateWallet_0008CreateOneTimeWalletFailureProtocolUnknown)
+{
+ BSINT32 rtnVal;
+ BoatEthWallet *g_ethereum_wallet_ptr;
+ BoatEthWalletConfig wallet_config = get_ethereum_wallet_settings();
+ extern BoatIotSdkContext g_boat_iot_sdk_context;
+ /* 1. execute unit test */
+ rtnVal = BoatWalletCreate(BOAT_PROTOCOL_UNKNOWN, NULL, &wallet_config, sizeof(BoatEthWalletConfig));
+ /* 2. verify test result */
+ /* 2-1. verify the return value */
+ ck_assert_int_eq(rtnVal, BOAT_ERROR);
+
+ /* 2-2. verify the global variables that be affected */
+ ck_assert(g_boat_iot_sdk_context.wallet_list[0].is_used == false);
+}
+END_TEST
+
Suite *make_wallet_suite(void)
{
/* Create Suite */
@@ -232,6 +249,7 @@ Suite *make_wallet_suite(void)
tcase_add_test(tc_wallet_api, test_001CreateWallet_0005CreateLoadWalletFailureNoExist);
tcase_add_test(tc_wallet_api, test_001CreateWallet_0006CreateOneTimeWalletFailureShortSize);
tcase_add_test(tc_wallet_api, test_001CreateWallet_0007CreateOneTimeWalletSuccessLongSize);
+ tcase_add_test(tc_wallet_api, test_001CreateWallet_0008CreateOneTimeWalletFailureProtocolUnknown);
return s_wallet;
}
|
docs: add sphinx warnings in format_idf_target
format_idf_target will now log sphinx warnings for any
{IDF_TARGET_X} that were not replaced. | @@ -5,6 +5,7 @@ from docutils import io, nodes, statemachine, utils
from docutils.utils.error_reporting import SafeString, ErrorString
from docutils.parsers.rst import directives
from sphinx.directives.other import Include as BaseInclude
+from sphinx.util import logging
def setup(app):
@@ -21,6 +22,16 @@ def setup(app):
return {'parallel_read_safe': True, 'parallel_write_safe': True, 'version': '0.2'}
+def check_content(content, docname):
+ # Log warnings for any {IDF_TARGET} expressions that haven't been replaced
+ logger = logging.getLogger(__name__)
+
+ errors = re.findall(r'{IDF_TARGET.*?}', content)
+
+ for err in errors:
+ logger.warning('Badly formated string substitution: {}'.format(err), location=docname)
+
+
class StringSubstituter:
""" Allows for string substitution of target related strings
before any markup is parsed
@@ -113,6 +124,8 @@ class StringSubstituter:
def substitute_source_read_cb(self, app, docname, source):
source[0] = self.substitute(source[0])
+ check_content(source[0], docname)
+
class FormatedInclude(BaseInclude):
|
schema compile CHANGE making container implicitely presence
clarify the warning message and, in addition, make the container presence in case of
having when statement. | @@ -3384,18 +3384,22 @@ lys_compile_node_container(struct lysc_ctx *ctx, struct lysp_node *node_p, struc
cont->flags |= LYS_PRESENCE;
} else if (cont_p->musts) {
/* container with a must condition */
- LOGWRN(ctx->ctx, "Container \"%s\" changed to presence because it has a \"must\" condition.", cont_p->name);
+ LOGWRN(ctx->ctx, "Container \"%s\" changed to presence because it has a meaning from its \"must\" condition.", cont_p->name);
+ cont->flags |= LYS_PRESENCE;
+ } else if (cont_p->when) {
+ /* container with a when condition */
+ LOGWRN(ctx->ctx, "Container \"%s\" changed to presence because it has a meaning from its \"when\" condition.", cont_p->name);
cont->flags |= LYS_PRESENCE;
} else if (cont_p->parent) {
if (cont_p->parent->nodetype == LYS_CHOICE) {
/* container is an implicit case, so its existence decides the existence of the whole case */
- LOGWRN(ctx->ctx, "Container \"%s\" changed to presence because it is an implicit case of choice \"%s\".",
+ LOGWRN(ctx->ctx, "Container \"%s\" changed to presence because it has a meaning as a case of choice \"%s\".",
cont_p->name, cont_p->parent->name);
cont->flags |= LYS_PRESENCE;
} else if ((cont_p->parent->nodetype == LYS_CASE)
&& (((struct lysp_node_case *)cont_p->parent)->child == node_p) && !cont_p->next) {
/* container is the only node in a case, so its existence decides the existence of the whole case */
- LOGWRN(ctx->ctx, "Container \"%s\" changed to presence because it is the only data node of case \"%s\".",
+ LOGWRN(ctx->ctx, "Container \"%s\" changed to presence because it has a meaning as a case of choice \"%s\".",
cont_p->name, cont_p->parent->name);
cont->flags |= LYS_PRESENCE;
}
|
Add Equal to compare arrayRef & vector (simplify code to avoid vector to arrayRef cast). | @@ -69,3 +69,8 @@ void Assign(TConstArrayRef<T> arrayRef, TVector<T>* v) {
v->assign(arrayRef.begin(), arrayRef.end());
}
+template<class T>
+bool Equal(TConstArrayRef<T> arrayRef, const TVector<T>& v) {
+ return arrayRef == TConstArrayRef<T>(v);
+}
+
|
world transform was still being connected for history assets | @@ -73,8 +73,13 @@ createInputGeometryNode(string $transformAttr, string $geometryAttr)
string $inputNode = `createNode -skipSelect houdiniInputGeometry`;
- if($transformAttr != "")
- connectAttr $transformAttr ($inputNode + ".inputTransform");
+ // don't connect the input world transform for a history asset
+ // Since the output is still parented under that same hierarchy
+ // but if you really need the input transform, uncomment this code
+
+ //if($transformAttr != "")
+ //connectAttr $transformAttr ($inputNode + ".inputTransform");
+
connectAttr $geometryAttr ($inputNode + ".inputGeometry");
return ($inputNode + ".outputNodeId");
|
parallel build for appveyor | @@ -8,19 +8,20 @@ environment:
- COMPILER: MinGW-w64
BUILD_SERVER:
secure: LONrmrMRv+m7++Sxtki7FpSe5kDZY/ak0CJmAknnLFGUJ55JVsIwBBOhIdijhMcF
+ BUILD_TYPE: MinSizeRel
install:
- git submodule update --init --recursive
- if [%COMPILER%]==[MinGW-w64] set PATH=C:\mingw-w64\x86_64-7.3.0-posix-seh-rt_v5-rev0\mingw64\bin;%PATH%
build_script:
- - if [%COMPILER%]==[MinGW-w64] cmake -G "MinGW Makefiles" -DCMAKE_BUILD_TYPE=MinSizeRel -DCMAKE_SH=CMAKE_SH-NOTFOUND .
- - if [%COMPILER%]==[MSVC] cmake -G "Visual Studio 14 2015" -DCMAKE_BUILD_TYPE=MinSizeRel .
- - if [%COMPILER%]==[MSVC64] cmake -G "Visual Studio 14 2015 Win64" -DCMAKE_BUILD_TYPE=MinSizeRel .
- - cmake --build .
+ - if [%COMPILER%]==[MinGW-w64] cmake -G "MinGW Makefiles" -DCMAKE_BUILD_TYPE=%BUILD_TYPE% -DCMAKE_SH=CMAKE_SH-NOTFOUND .
+ - if [%COMPILER%]==[MSVC] cmake -G "Visual Studio 14 2015" -DCMAKE_BUILD_TYPE=%BUILD_TYPE% .
+ - if [%COMPILER%]==[MSVC64] cmake -G "Visual Studio 14 2015 Win64" -DCMAKE_BUILD_TYPE=%BUILD_TYPE% .
+ - cmake --build . --parallel --config %BUILD_TYPE%
after_build:
- 7z a tic80.zip %APPVEYOR_BUILD_FOLDER%\bin\*.exe
+ 7z a tic80.zip %APPVEYOR_BUILD_FOLDER%\bin\tic80.exe
on_success:
- ps: $env:DATE=[TimeZoneInfo]::ConvertTimeBySystemTimeZoneId([DateTime]::UtcNow, "Belarus Standard Time").ToString('yyyy.MM.dd')
|
Allow NULL path in HRN_STORAGE_PATH_CREATE() macro. | @@ -485,7 +485,7 @@ hrnStoragePathCreate(const Storage *const storage, const char *const path, HrnSt
ASSERT(storage != NULL);
- const String *const pathFull = storagePathP(storage, STR(path));
+ const String *const pathFull = storagePathP(storage, path == NULL ? NULL : STR(path));
printf("create path '%s'", strZ(pathFull));
|
Avoid overflows in left shifts
Left-shifting uint8_t types kinda works because of integer promotion
rules. However, this is incorrect on platforms with 16-bit int types.
We should explicitly use 32-bit unsigned integers to get the correct
behavior on all platforms (and avoid warnings by undefined behavior
sanitizer). | @@ -79,7 +79,7 @@ void soter_crc32_update(soter_crc32_t* crc, const void* buf, size_t len)
uint32_t soter_crc32_final(soter_crc32_t* crc)
{
uint32_t result = ~(*crc);
- uint8_t byte0, byte1, byte2, byte3;
+ uint32_t byte0, byte1, byte2, byte3;
/* result now holds the negated polynomial remainder;
* since the table and algorithm is "reflected" [williams95].
|
fix wording per Eng correction | @@ -25,7 +25,7 @@ AppScope 1.2.0 introduces substantial new functionality:
- Support for a **filter list** where you specify which processes to scope and optionally, which not to scope. The filter list is an argument to the `scope start` command. In its allowlist section, you specify the configuration for AppScope to use when scoping particular processes. Apart from the `scope start` command, `LD_PRELOAD` looks for a filter list, but the `scope run` and `scope attach` commands do not. Related issues: [#987](https://github.com/criblio/appscope/issues/987),[#1027](https://github.com/criblio/appscope/issues/1027),[#1038](https://github.com/criblio/appscope/issues/1038),[#1069](https://github.com/criblio/appscope/issues/1069).
-- The new `scope detach` command, which enables you to stop scoping a process without terminating the process. Note: constraints inherent in the kernel require that the `scope detach` command leave the library loaded. Related issues: [#370](https://github.com/criblio/appscope/issues/370),[#1039](https://github.com/criblio/appscope/issues/1039),[#1085](https://github.com/criblio/appscope/issues/1085),[#1087](https://github.com/criblio/appscope/issues/1087),[#1088](https://github.com/criblio/appscope/issues/1088),[#1089](https://github.com/criblio/appscope/issues/1089),[#1094](https://github.com/criblio/appscope/issues/1094),[#1107](https://github.com/criblio/appscope/issues/1107),[#1129](https://github.com/criblio/appscope/issues/1129).
+- The new `scope detach` command, which enables you to stop scoping a process without terminating the process. Note: constraints inherent in the dynamic loader require that the `scope detach` command leave the library loaded. Related issues: [#370](https://github.com/criblio/appscope/issues/370),[#1039](https://github.com/criblio/appscope/issues/1039),[#1085](https://github.com/criblio/appscope/issues/1085),[#1087](https://github.com/criblio/appscope/issues/1087),[#1088](https://github.com/criblio/appscope/issues/1088),[#1089](https://github.com/criblio/appscope/issues/1089),[#1094](https://github.com/criblio/appscope/issues/1094),[#1107](https://github.com/criblio/appscope/issues/1107),[#1129](https://github.com/criblio/appscope/issues/1129).
- To attach to or detach from a process running in a container, it is no longer necessary to run AppScope within the container. You can just scope the container processes from the host. Related issues: [#1061](https://github.com/criblio/appscope/issues/1061),
[#1119](https://github.com/criblio/appscope/issues/1119).
|
Improve multiarch condition in coredump access | @@ -57,7 +57,7 @@ _UCD_access_reg (unw_addr_space_t as,
#elif defined(UNW_TARGET_S390X)
if (regnum > UNW_S390X_R15)
goto badreg;
-#elif defined(UNW_TARGET_IA64)
+#elif defined(UNW_TARGET_IA64) || defined(UNW_TARGET_HPPA) || defined(UNW_TARGET_PPC32) || defined(UNW_TARGET_PPC64)
if (regnum >= ARRAY_SIZE(ui->prstatus->pr_reg))
goto badreg;
#else
|
added missing implementation for lv_fragment_get_manager | @@ -55,6 +55,13 @@ void lv_fragment_del(lv_fragment_t * fragment)
lv_mem_free(fragment);
}
+lv_fragment_manager_t * lv_fragment_get_manager(lv_fragment_t * fragment)
+{
+ LV_ASSERT_NULL(fragment);
+ LV_ASSERT_NULL(fragment->managed);
+ return fragment->managed->manager;
+}
+
lv_obj_t * const * lv_fragment_get_container(lv_fragment_t * fragment)
{
LV_ASSERT_NULL(fragment);
|
added j2534 troubleshooting instructions | @@ -137,5 +137,14 @@ More details here:
relaxed to allow serialization of messages based on their address
(making multiple queues, effectively one queue per address).
+# Troubleshooting:
+troubleshooting:
+1. Install DrewTech J2534-1 Bus Analysis Tool
+http://www.drewtech.com/downloads/tools/Drew%20Technologies%20Tool%20for%20J2534-1%20API%20v1.07.msi
+2. Open DrewTech tool and make sure it shows "panda" as a device listed (this means registry settings are correct)
+3. When DrewTech tool attempts to load the driver it will show an error if it fails
+4. To figure out why the driver fails to load install Process Monitor and filter by the appropriate process name
+https://docs.microsoft.com/en-us/sysinternals/downloads/procmon
+
# Other:
Panda head ASCII art by dcau
\ No newline at end of file
|
Update sample (cache name string as array) | @@ -91,7 +91,7 @@ static void simple_complete(ocf_cache_t cache, void *priv, int error)
*/
int initialize_cache(ocf_ctx_t ctx, ocf_cache_t *cache)
{
- struct ocf_mngt_cache_config cache_cfg = { };
+ struct ocf_mngt_cache_config cache_cfg = { .name = "cache1" };
struct ocf_mngt_cache_device_config device_cfg = { };
struct cache_priv *cache_priv;
struct simple_context context;
@@ -106,7 +106,6 @@ int initialize_cache(ocf_ctx_t ctx, ocf_cache_t *cache)
/* Cache configuration */
ocf_mngt_cache_config_set_default(&cache_cfg);
cache_cfg.metadata_volatile = true;
- cache_cfg.name = "cache1";
/* Cache deivce (volume) configuration */
ocf_mngt_cache_device_config_set_default(&device_cfg);
|
Add python3 to AppVeyor. | @@ -46,6 +46,7 @@ $CMAKE_FLAGS = "$CMAKE_FLAGS -G ""$GENERATOR"""
###############################################################################
choco install -y -r swig --version 3.0.9
choco install -y -r lua53
+choco install -y -r python3
if ($Env:COMPILER -eq "mingw") {
if ($Env:PLATFORM -eq "Win64") {
choco install -y -r mingw
@@ -62,6 +63,7 @@ $CMAKE_FLAGS = "$CMAKE_FLAGS -DCMAKE_BUILD_TYPE=Release"
$CMAKE_FLAGS = "$CMAKE_FLAGS -DTINYSPLINE_ENABLE_CSHARP=True"
$CMAKE_FLAGS = "$CMAKE_FLAGS -DTINYSPLINE_ENABLE_D=True"
$CMAKE_FLAGS = "$CMAKE_FLAGS -DTINYSPLINE_ENABLE_JAVA=True"
+$CMAKE_FLAGS = "$CMAKE_FLAGS -DTINYSPLINE_ENABLE_PYTHON=True"
|
Added libfontconfig wrapped function FcFontSetCreate | @@ -81,7 +81,7 @@ GO(FcFontList, pFppp)
GO(FcFontMatch, pFppp)
GO(FcFontRenderPrepare, pFppp)
GO(FcFontSetAdd, iFpp)
-//GO(FcFontSetCreate,
+GO(FcFontSetCreate, pFv)
GO(FcFontSetDestroy, vFp)
//GO(FcFontSetList,
//GO(FcFontSetMatch,
|
Add closing of main menu after using New file item | @@ -1897,6 +1897,13 @@ void push_main_menu(void) {
qmenu_push_to_nav(qm);
}
+void pop_qnav_if_main_menu(void) {
+ Qblock* qb = qnav_top_block();
+ if (qb && qb->tag == Qblock_type_qmenu &&
+ qmenu_id(qmenu_of(qb)) == Main_menu_id)
+ qnav_stack_pop();
+}
+
void push_autofit_menu(void) {
Qmenu* qm = qmenu_create(Autofit_menu_id);
qmenu_set_title(qm, "Auto-fit Grid");
@@ -2810,6 +2817,7 @@ int main(int argc, char** argv) {
ged_state.is_draw_dirty = true;
heapstr_set_cstr(&file_name, "");
ged_state.filename = "unnamed"; // slightly redundant
+ pop_qnav_if_main_menu();
}
} break;
case Main_menu_open:
@@ -2863,10 +2871,7 @@ int main(int argc, char** argv) {
ged_make_cursor_visible(&ged_state);
}
qnav_stack_pop();
- qb = qnav_top_block();
- if (qb && qb->tag == Qblock_type_qmenu &&
- qmenu_id(qmenu_of(qb)) == Main_menu_id)
- qnav_stack_pop();
+ pop_qnav_if_main_menu();
}
} break;
}
@@ -2905,11 +2910,7 @@ int main(int argc, char** argv) {
ged_make_cursor_visible(&ged_state);
ged_state.needs_remarking = true;
ged_state.is_draw_dirty = true;
- // Pop main menu if it's open, too
- qb = qnav_top_block();
- if (qb && qb->tag == Qblock_type_qmenu &&
- qmenu_id(qmenu_of(qb)) == Main_menu_id)
- qnav_stack_pop();
+ pop_qnav_if_main_menu();
} else {
undo_history_pop(&ged_state.undo_hist, &ged_state.field,
&ged_state.tick_num);
|
Fix package rac's examples for godoc convention | @@ -28,14 +28,14 @@ import (
"github.com/google/wuffs/lib/rac"
)
-// ExampleILAEnd demonstrates using the low level "rac" package to encode a
-// RAC+Zlib formatted file with IndexLocationAtEnd.
+// Example_indexLocationAtEnd demonstrates using the low level "rac" package to
+// encode a RAC+Zlib formatted file with IndexLocationAtEnd.
//
// The sibling "raczlib" package provides a higher level API that is easier to
// use.
//
// See the RAC specification for an explanation of the file format.
-func ExampleILAEnd() {
+func Example_indexLocationAtEnd() {
// Manually construct a zlib encoding of "More!\n", one that uses a literal
// block (that's easy to see in a hex dump) instead of a more compressible
// Huffman block.
@@ -90,14 +90,15 @@ func ExampleILAEnd() {
// 00000030 00 00 00 01 01 |.....|
}
-// ExampleILAStart demonstrates using the low level "rac" package to encode and
-// then decode a RAC+Zlib formatted file with IndexLocationAtStart.
+// Example_indexLocationAtStart demonstrates using the low level "rac" package
+// to encode and then decode a RAC+Zlib formatted file with
+// IndexLocationAtStart.
//
// The sibling "raczlib" package provides a higher level API that is easier to
// use.
//
// See the RAC specification for an explanation of the file format.
-func ExampleILAStart() {
+func Example_indexLocationAtStart() {
buf := &bytes.Buffer{}
w := &rac.Writer{
Writer: buf,
|
Update barstools.rst
Added info about adding our own sram-macros cache | @@ -24,7 +24,7 @@ A list of unique SRAM configurations is output to a ``.conf`` file by FIRRTL, wh
Without this transform, FIRRTL will map all ``SeqMem`` s to flip-flop arrays with equivalent behavior, which may lead to a design that is difficult to route.
The ``.conf`` file is consumed by a tool called MacroCompiler, which is part of the :ref:`Tools/Barstools:Barstools` scala package.
-MacroCompiler is also passed an ``.mdf`` file that describes the available list of technology SRAMs or the capabilities of the SRAM compiler, if one is provided by the foundry.
+MacroCompiler is also passed an ``.mdf`` file that describes the available list of technology SRAMs or the capabilities of the SRAM compiler, if one is provided by the foundry. You may wish to create a cache of your available SRAM macros either manually, or via a script. An reference script for creating a JSON of your SRAM macros is in the `asap7 technology library folder <https://github.com/ucb-bar/hammer/blob/8fd1486499b875d56f09b060f03a62775f0a6aa7/src/hammer-vlsi/technology/asap7/sram-cache-gen.py>`__; the expected JSON format is `here <https://github.com/ucb-bar/plsi-mdf/blob/4be9b173647c77f990a542f4eb5f69af01d77316/macro_format.json>`_.
Typically a foundry SRAM compiler will be able to generate a set of different SRAMs collateral based on some requirements on size, aspect ratio, etc. (see :ref:`Tools/Barstools:SRAM MDF Fields`).
Using a user-customizable cost function, MacroCompiler will select the SRAMs that are the best fit for each dimensionality in the ``.conf`` file.
This may include over provisioning (e.g. using a 64x1024 SRAM for a requested 60x1024, if the latter is not available) or arraying.
|
Fix incomplete deprecation guard in test/sslapitest.c
OPENSSL_NO_DEPRECATED_3_0 should be used rather than OPENSSL_NO_DEPRECATED,
as the latter doesn't take the configuration option '--api=' in account.
Fixes | @@ -8157,7 +8157,7 @@ static EVP_PKEY *get_tmp_dh_params(void)
return tmp_dh_params;
}
-# ifndef OPENSSL_NO_DEPRECATED
+# ifndef OPENSSL_NO_DEPRECATED_3_0
/* Callback used by test_set_tmp_dh() */
static DH *tmp_dh_callback(SSL *s, int is_export, int keylen)
{
|
[MUSL] Modify `CURRENT_UTF8`
drop support for `pthread_self` structure
use value stored in global struct | @@ -39,7 +39,7 @@ hidden char *__gettextdomain(void);
#define CURRENT_LOCALE (&libc.global_locale)
-#define CURRENT_UTF8 (!!__pthread_self()->locale->cat[LC_CTYPE])
+#define CURRENT_UTF8 (!!&libc.global_locale.cat[LC_CTYPE])
#undef MB_CUR_MAX
#define MB_CUR_MAX (CURRENT_UTF8 ? 4 : 1)
|
Remove hack fix for server hybrid issue | @@ -961,20 +961,6 @@ static int ssl_conf_check(const mbedtls_ssl_context *ssl)
return( 0 );
}
-#if defined(MBEDTLS_SSL_PROTO_TLS1_2) && \
- defined(MBEDTLS_SSL_PROTO_TLS1_3) && \
- defined(MBEDTLS_SSL_SRV_C) && \
- defined(MBEDTLS_KEY_EXCHANGE_WITH_CERT_ENABLED)
-/* Remove below lines if server side hybrid mode implemented.
- * To fix wrong default signature algorithm setting when both
- * TLS1.2 and TLS1.3 enabled.
- */
-static void ssl_fix_server_side_negotiation_fail( mbedtls_ssl_context *ssl );
-#endif /* MBEDTLS_SSL_PROTO_TLS1_2 &&
- MBEDTLS_SSL_PROTO_TLS1_3 &&
- MBEDTLS_SSL_SRV_C &&
- MBEDTLS_KEY_EXCHANGE_WITH_CERT_ENABLED */
-
/*
* Setup an SSL context
*/
@@ -2998,20 +2984,8 @@ int mbedtls_ssl_handshake_step( mbedtls_ssl_context *ssl )
if( ssl->conf->endpoint == MBEDTLS_SSL_IS_SERVER )
{
#if defined(MBEDTLS_SSL_PROTO_TLS1_3)
-
if( mbedtls_ssl_conf_is_tls13_only( ssl->conf ) )
- {
-#if defined(MBEDTLS_SSL_PROTO_TLS1_2) && \
- defined(MBEDTLS_KEY_EXCHANGE_WITH_CERT_ENABLED)
- /* Remove below lines if server side hybrid mode implemented. */
- if( ssl->state == MBEDTLS_SSL_HELLO_REQUEST )
- {
- ssl_fix_server_side_negotiation_fail( ssl );
- }
-#endif /* MBEDTLS_SSL_PROTO_TLS1_2 &&
- MBEDTLS_KEY_EXCHANGE_WITH_CERT_ENABLED */
ret = mbedtls_ssl_tls13_handshake_server_step( ssl );
- }
#endif /* MBEDTLS_SSL_PROTO_TLS1_3 */
#if defined(MBEDTLS_SSL_PROTO_TLS1_2)
|
improves cli for oidc-gen | #include <stdio.h>
#include <string.h>
#include <syslog.h>
+#include <argp.h>
#include "oidc-gen.h"
#include "../../src/provider.h"
#define OIDC_SOCK_ENV_NAME "OIDC_SOCK"
+const char *argp_program_version = "oidc-gen 0.1.0";
+
+const char *argp_program_bug_address = "<[email protected]>";
+
+struct arguments {
+};
+
+static struct argp_option options[] = {
+ {0}
+};
+
+static error_t parse_opt (int key, char *arg, struct argp_state *state) {
+ struct arguments *arguments = state->input;
+
+ switch (key)
+ {
+ case ARGP_KEY_ARG:
+ argp_usage(state);
+ break;
+ case ARGP_KEY_END:
+ break;
+ default:
+ return ARGP_ERR_UNKNOWN;
+ }
+ return 0;
+}
+
+static char args_doc[] = "";
+
+static char doc[] = "oidc-gen -- A tool for generating oidc provider configuration which can be used by oidc-gen";
+
+static struct argp argp = {options, parse_opt, args_doc, doc};
+
+
static struct oidc_provider* provider = NULL;
-int main(/* int argc, char** argv */) {
+
+int main(int argc, char** argv) {
openlog("oidc-gen", LOG_CONS|LOG_PID, LOG_AUTHPRIV);
// setlogmask(LOG_UPTO(LOG_DEBUG));
setlogmask(LOG_UPTO(LOG_NOTICE));
+
+ struct arguments arguments;
+ argp_parse (&argp, argc, argv, 0, 0, &arguments);
+
provider = genNewProvider();
char* json = providerToJSON(*provider);
struct connection con = {0,0,0};
|
Fix aesni_xts compile error
Block copy bug.. | @@ -84,7 +84,7 @@ static int cipher_hw_aesni_xts_initkey(PROV_CIPHER_CTX *ctx,
PROV_AES_XTS_CTX *xctx = (PROV_AES_XTS_CTX *)ctx;
XTS_SET_KEY_FN(aesni_set_encrypt_key, aesni_set_decrypt_key,
- aesni_xts_encrypt, aesni_decrypt,
+ aesni_encrypt, aesni_decrypt,
aesni_xts_encrypt, aesni_xts_decrypt);
return 1;
}
|
Fix error code for superblock checksum mismatch
Fix error code for superblock checksum mismatch.
Superblock validation now returns a proper error on checksum check fail. | @@ -127,7 +127,7 @@ int ocf_metadata_validate_superblock(ocf_ctx_t ctx,
if (crc != superblock->checksum[metadata_segment_sb_config]) {
ocf_log_invalid_superblock("checksum");
- return -OCF_ERR_INVAL;
+ return -OCF_ERR_CRC_MISMATCH;
}
if (superblock->clean_shutdown > ocf_metadata_clean_shutdown) {
|
Pacify the compiler for "maybe uninitialized" variables in the socket module. | @@ -55,7 +55,7 @@ static Value bindSocket(VM *vm, int argCount, Value *args) {
server.sin_port = htons(port);
if (bind(sock->socket, (struct sockaddr *)&server , sizeof(server)) < 0) {
- Value module;
+ Value module = 0;
tableGet(&vm->modules, copyString(vm, "Socket", 6), &module);
SET_ERRNO(AS_MODULE(module));
return NIL_VAL;
@@ -83,7 +83,7 @@ static Value listenSocket(VM *vm, int argCount, Value *args) {
ObjSocket *sock = AS_SOCKET(args[0]);
if (listen(sock->socket, backlog) == -1) {
- Value module;
+ Value module = 0;
tableGet(&vm->modules, copyString(vm, "Socket", 6), &module);
SET_ERRNO(AS_MODULE(module));
return NIL_VAL;
@@ -125,7 +125,7 @@ static Value writeSocket(VM *vm, int argCount, Value *args) {
int writeRet = write(sock->socket , message->chars, message->length);
if (writeRet == -1) {
- Value module;
+ Value module = 0;
tableGet(&vm->modules, copyString(vm, "Socket", 6), &module);
SET_ERRNO(AS_MODULE(module));
return NIL_VAL;
@@ -152,7 +152,7 @@ static Value recvSocket(VM *vm, int argCount, Value *args) {
int read_size = recv(sock->socket, buffer, bufferSize, 0);
if (read_size == -1) {
- Value module;
+ Value module = 0;
tableGet(&vm->modules, copyString(vm, "Socket", 6), &module);
SET_ERRNO(AS_MODULE(module));
FREE_ARRAY(vm, char, buffer, bufferSize);
@@ -194,7 +194,7 @@ static Value setSocketOpt(VM *vm, int argCount, Value *args) {
int option = AS_NUMBER(args[2]);
if (setsockopt(sock->socket, level, option, &(int){1}, sizeof(int)) == -1) {
- Value module;
+ Value module = 0;
tableGet(&vm->modules, copyString(vm, "Socket", 6), &module);
SET_ERRNO(AS_MODULE(module));
return NIL_VAL;
|
classify: Fix 2 coverity errors
Validate two tainted scalars, filter_sw_if_index, that
came from an API message.
Type: fix | @@ -237,6 +237,10 @@ vl_api_policer_classify_dump_t_handler (vl_api_policer_classify_dump_t * mp)
return;
filter_sw_if_index = ntohl (mp->sw_if_index);
+ if (filter_sw_if_index
+ >= vec_len (pcm->classify_table_index_by_sw_if_index[mp->type]))
+ return;
+
if (filter_sw_if_index != ~0)
vec_tbl =
&pcm->classify_table_index_by_sw_if_index[mp->type][filter_sw_if_index];
@@ -519,6 +523,10 @@ vl_api_flow_classify_dump_t_handler (vl_api_flow_classify_dump_t * mp)
return;
filter_sw_if_index = ntohl (mp->sw_if_index);
+ if (filter_sw_if_index
+ >= vec_len (pcm->classify_table_index_by_sw_if_index[mp->type]))
+ return;
+
if (filter_sw_if_index != ~0)
vec_tbl =
&pcm->classify_table_index_by_sw_if_index[mp->type][filter_sw_if_index];
|
Improve checking process SIDs with sandboxie | @@ -2390,14 +2390,14 @@ VOID PhProcessProviderUpdate(
// User
if (NT_SUCCESS(PhGetTokenUser(tokenHandle, &tokenUser)))
{
- if (!RtlEqualSid(processItem->Sid, tokenUser->User.Sid))
+ if (!processItem->Sid || !RtlEqualSid(processItem->Sid, tokenUser->User.Sid))
{
PSID processSid;
// HACK (dmex)
processSid = processItem->Sid;
processItem->Sid = PhAllocateCopy(tokenUser->User.Sid, RtlLengthSid(tokenUser->User.Sid));
- PhFree(processSid);
+ if (processSid) PhFree(processSid);
PhMoveReference(&processItem->UserName, PhpGetSidFullNameCachedSlow(processItem->Sid));
|
[ci] Disable CCACHE everywhere
It seems the env flag does not work in some instances, use a custom path
for those cases | @@ -21,6 +21,7 @@ variables:
RUST_LOG: 'memora=debug'
VERILATOR_ROOT: '$CI_PROJECT_DIR/install/verilator'
PATH: '/home/gitlabci/.cargo/bin:/usr/local/bin:/usr/bin:/usr/sbin:/sbin:/usr/local/condor/bin:/usr/sepp/bin:$VERILATOR_ROOT/bin'
+ OBJCACHE: ''
workflow:
rules:
@@ -74,14 +75,11 @@ verilator:
- |
if ! $CI_PROJECT_DIR/.gitlab-ci.d/memora_retry.sh lookup verilator; then
git submodule update --init --recursive -- toolchain/verilator
- export CCACHE_DISABLE=1
export PYTHON3=$(which python3-3.6.5)
export VERILATOR_CI="--disable-defenv"
make verilator
$CI_PROJECT_DIR/.gitlab-ci.d/memora_retry.sh insert verilator
fi
- variables:
- CCACHE_DISABLE: "1"
halide:
stage: build
|
Cache hash values for small strings as well | @@ -39,12 +39,12 @@ String Type
typedef struct {
fiobj_object_header_s head;
+ uint64_t hash;
uint8_t is_small;
uint8_t frozen;
uint8_t slen;
intptr_t len;
uintptr_t capa;
- uint64_t hash;
char *str;
} fiobj_str_s;
@@ -500,12 +500,17 @@ size_t fiobj_str_join(FIOBJ dest, FIOBJ obj) {
*/
uint64_t fiobj_str_hash(FIOBJ o) {
assert(FIOBJ_TYPE_IS(o, FIOBJ_T_STRING));
- if (obj2str(o)->is_small) {
- return fio_siphash(STR_INTENAL_STR(o), STR_INTENAL_LEN(o));
- } else if (obj2str(o)->hash) {
+ // if (obj2str(o)->is_small) {
+ // return fio_siphash(STR_INTENAL_STR(o), STR_INTENAL_LEN(o));
+ // } else
+ if (obj2str(o)->hash) {
return obj2str(o)->hash;
}
+ if (obj2str(o)->is_small) {
+ obj2str(o)->hash = fio_siphash(STR_INTENAL_STR(o), STR_INTENAL_LEN(o));
+ } else {
obj2str(o)->hash = fio_siphash(obj2str(o)->str, obj2str(o)->len);
+ }
return obj2str(o)->hash;
}
|
FireSim generator should also produce hwacha test suite makefrag rules | @@ -18,6 +18,8 @@ import freechips.rocketchip.tile.XLen
import firesim.util.{GeneratorArgs, HasTargetAgnosticUtilites, HasFireSimGeneratorUtilities}
+import scala.util.Try
+
import utilities.TestSuiteHelper
trait HasTestSuites {
@@ -28,6 +30,14 @@ trait HasTestSuites {
TestGeneration.addSuite(SlowBlockdevTests)
if (!targetName.contains("NoNIC"))
TestGeneration.addSuite(NICLoopbackTests)
+
+ import hwacha.HwachaTestSuites._
+ if (Try(params(hwacha.HwachaNLanes)).getOrElse(0) > 0) {
+ TestGeneration.addSuites(rv64uv.map(_("p")))
+ TestGeneration.addSuites(rv64uv.map(_("vp")))
+ TestGeneration.addSuite(rv64sv("p"))
+ TestGeneration.addSuite(hwachaBmarks)
+ }
}
}
|
server_shutdown: make access to running synchronised | @@ -59,7 +59,7 @@ struct _server {
size_t secondariescnt;
char failover:1;
char failure; /* full byte for atomic access */
- char running:1;
+ char running; /* full byte for atomic access */
char keep_running; /* full byte for atomic access */
unsigned char stallseq; /* full byte for atomic access */
size_t metrics;
@@ -538,7 +538,7 @@ server_queuereader(void *d)
idle = 0;
}
- self->running = 0;
+ __sync_and_and_fetch(&(self->running), 0);
if (self->fd >= 0)
close(self->fd);
@@ -774,7 +774,7 @@ server_shutdown(server *s)
for (i = 0; i < s->secondariescnt; i++) {
if (__sync_add_and_fetch(&(s->secondaries[i]->failure), 0))
failures++;
- if (s->secondaries[i]->running)
+ if (__sync_add_and_fetch(&(s->secondaries[i]->running), 0))
inqueue += queue_len(s->secondaries[i]->queue);
}
/* loop until we all failed, or nothing is in the queues */
@@ -794,7 +794,7 @@ server_shutdown(server *s)
/* wait for the secondaries to be stopped so we surely don't get
* invalid reads when server_free is called */
for (i = 0; i < s->secondariescnt; i++) {
- while (s->secondaries[i]->running)
+ while (__sync_add_and_fetch(&(s->secondaries[i]->running), 0))
usleep((200 + (rand() % 100)) * 1000);
}
}
|
Leaf: Add function to decrease array indices | @@ -227,6 +227,32 @@ CppKey changeArrayIndexByOne (CppKey const & parent, CppKey const & element, boo
return elementNewIndex;
}
+/**
+ * @brief Decrease the array index of array elements by one.
+ *
+ * @param parents This parameter contains the array parents for which this function decrease the index by one.
+ * @param arrays This variable stores the arrays elements (and parents) this function updates.
+ *
+ * @return A copy of `arrays`, where all indices specified by `parents` are decreased by one
+ */
+CppKeySet decreaseArrayIndices (CppKeySet const & parents, CppKeySet const & arrays)
+{
+ CppKeySet arraysIndexDecreased = arrays.dup ();
+ CppKeySet arrayParents = parents.dup ();
+
+ while (CppKey parent = arrayParents.pop ())
+ {
+ arraysIndexDecreased = accumulate (arraysIndexDecreased.begin (), arraysIndexDecreased.end (), CppKeySet{},
+ [&parent](CppKeySet collected, CppKey key) {
+ if (key.isBelow (parent)) changeArrayIndexByOne (parent, key, false);
+ collected.append (key);
+ return collected;
+ });
+ }
+
+ return arraysIndexDecreased;
+}
+
/**
* @brief Increase the array index of array elements by one.
*
|
Fix nvmxml parsing for show goal | @@ -226,7 +226,7 @@ ShowGoalPrintDetailedView(
ZeroMem(DimmStr, sizeof(DimmStr));
- SetDisplayInfo(L"ConfigGoal", ListView, NULL);
+ SetDisplayInfo(L"ConfigGoal", ListView, L"=");
if (pRegionConfigsInfo == NULL || (DisplayOptionSet && pDisplayValues == NULL)) {
goto Finish;
|
OcDevicePathLib: Add missing dependencies | DevicePathLib
MemoryAllocationLib
OcFileLib
+ OcStringLib
UefiBootServicesTableLib
[Guids]
gEfiDevicePathToTextProtocolGuid
gEfiSimpleFileSystemProtocolGuid
gEfiBlockIoProtocolGuid
+ gEfiPciIoProtocolGuid
gEfiUsbIoProtocolGuid
gEfiLoadFileProtocolGuid
|
Support less 3D model file formats;
Support OBJ, FBX, and COLLADA, as these seem to be the most common
formats. This is done to reduce the size of the assimp dll. More
formats can be added as needed. | @@ -67,8 +67,15 @@ else()
endif()
# assimp
-option(ASSIMP_BUILD_ASSIMP_TOOLS OFF)
-option(ASSIMP_BUILD_TESTS OFF)
+set(ASSIMP_BUILD_ASSIMP_TOOLS OFF CACHE BOOL "")
+set(ASSIMP_BUILD_TESTS OFF CACHE BOOL "")
+set(ASSIMP_NO_EXPORT ON OFF CACHE BOOL "")
+
+set(ASSIMP_BUILD_ALL_IMPORTERS_BY_DEFAULT OFF CACHE BOOL "")
+set(ASSIMP_BUILD_COLLADA_IMPORTER ON CACHE BOOL "")
+set(ASSIMP_BUILD_OBJ_IMPORTER ON CACHE BOOL "")
+set(ASSIMP_BUILD_FBX_IMPORTER ON CACHE BOOL "")
+
if(WIN32)
add_subdirectory(deps/assimp assimp)
include_directories(deps/assimp/include)
|
[hardware] Use correct variable name | @@ -25,7 +25,7 @@ module axi_rab_wrap #(
parameter int unsigned AxiUserWidth = 0,
// AXI types
parameter type axi_req_t = logic,
- parameter type axi_resp_t = logic
+ parameter type axi_resp_t = logic,
parameter type axi_lite_req_t = logic,
parameter type axi_lite_resp_t = logic
@@ -71,7 +71,7 @@ module axi_rab_wrap #(
.AXI_M_ADDR_WIDTH (AxiAddrWidth ),
.AXI_LITE_DATA_WIDTH (64/* TODO */ ),
.AXI_LITE_ADDR_WIDTH (32/* TODO */ ),
- .AXI_ID_WIDTH (AxiMempoolIdWidth ),
+ .AXI_ID_WIDTH (AxiIdWidth ),
.AXI_USER_WIDTH (AxiUserWidth ),
.MH_FIFO_DEPTH (MhFifoDepth )
) i_rab (
|
Silent compiler warnings in kiwi | @@ -169,6 +169,7 @@ kiwi_read_startup_size(char *data, uint32_t data_size)
KIWI_API static inline int
kiwi_validate_startup_header(char *data, uint32_t data_size, uint32_t *size)
{
+ (void)data_size; // Silent Compiler warnings
assert(data_size >= sizeof(uint32_t));
*size = kiwi_read_startup_size(data, sizeof(uint32_t));
/* do not expect big startup messages */
@@ -180,6 +181,7 @@ kiwi_validate_startup_header(char *data, uint32_t data_size, uint32_t *size)
KIWI_API static inline int
kiwi_validate_header(char *data, uint32_t data_size, uint32_t *size)
{
+ (void)data_size; // Silent Compiler warnings
assert(data_size >= sizeof(kiwi_header_t));
*size = kiwi_read_size(data, sizeof(kiwi_header_t));
|
python: fix crash in EndInterpreter | @@ -159,6 +159,9 @@ static unsigned open_cnt = 0;
static void Python_Shutdown (moduleData * data)
{
+ if (!data->shutdown)
+ return;
+
/* destroy python if plugin isn't used anymore */
if (Py_IsInitialized ())
{
|
media: add missing state for paused case
The condition of resume is needed. | @@ -228,9 +228,10 @@ void MediaRecorderImpl::startRecorder()
{
medvdbg("MediaRecorderImpl::startRecorder()\n");
- if (mCurState != RECORDER_STATE_READY) {
+ if (mCurState != RECORDER_STATE_READY && mCurState != RECORDER_STATE_PAUSED) {
notifyObserver(OBSERVER_COMMAND_ERROR);
- meddbg("MediaRecorderImpl::startRecorder() - (mCurState != RECORDER_STATE_READY)\n");
+ meddbg("MediaRecorderImpl::startRecorder() - " \
+ "mCurState != RECORDER_STATE_READY && mCurState != RECORDER_STATE_PAUSED\n");
return;
}
|
Fix: removed unnecessary buffer reversal | @@ -72,9 +72,6 @@ for table in hlir.tables:
#[ prefix_length += ${get_key_byte_width(k)};
if k.matchType.path.name == "lpm": # TODO: LS Check!
#[ prefix_length += ${target_name}_prefix_length;
- #[ int c, d;
- #[ for(c = ${byte_idx-1}, d = 0; c >= 0; c--, d++) *(reverse_buffer+d) = *(key+c);
- #[ for(c = 0; c < ${byte_idx}; c++) *(key+c) = *(reverse_buffer+c);
#[ lpm_add_promote(TABLE_${table.name}, (uint8_t*)key, prefix_length, (uint8_t*)&action, false, has_fields);
if table.matchType.name == "exact":
|
Allow the MQTT client example for platform native | @@ -8,6 +8,6 @@ CONTIKI = ../..
MODULES_REL += arch/platform/$(TARGET)
-PLATFORMS_ONLY = srf06-cc26xx cc2538dk openmote-cc2538 zoul
+PLATFORMS_ONLY = srf06-cc26xx cc2538dk openmote-cc2538 zoul native
include $(CONTIKI)/Makefile.include
|
Check for absence of categorical features correctly | @@ -58,7 +58,7 @@ void TPlainFoldBuilder::DoMap(NPar::IUserContext* ctx, int hostId, TInput* /*unu
void TApproxReconstructor::DoMap(NPar::IUserContext* ctx, int hostId, TInput* valuedForest, TOutput* /*unused*/) const {
NPar::TCtxPtr<TTrainData> trainData(ctx, SHARED_ID_TRAIN_DATA, hostId);
- Y_ASSERT(IsCategoricalFeaturesEmpty(trainData->TrainData.AllFeatures));
+ Y_ASSERT(AllOf(trainData->TrainData.AllFeatures.CatFeaturesRemapped, [](const auto& catFeature) { return catFeature.empty(); }));
auto& localData = TLocalTensorSearchData::GetRef();
Y_ASSERT(IsPlainMode(localData.Params.BoostingOptions->BoostingType));
|
fix(underglow): Fix error on release of rgb_ug | @@ -48,8 +48,14 @@ static int on_keymap_binding_pressed(struct zmk_behavior_binding *binding,
return -ENOTSUP;
}
+static int on_keymap_binding_released(struct zmk_behavior_binding *binding,
+ struct zmk_behavior_binding_event event) {
+ return 0;
+}
+
static const struct behavior_driver_api behavior_rgb_underglow_driver_api = {
.binding_pressed = on_keymap_binding_pressed,
+ .binding_released = on_keymap_binding_released,
};
DEVICE_AND_API_INIT(behavior_rgb_underglow, DT_INST_LABEL(0), behavior_rgb_underglow_init, NULL,
|
sysdeps/linux: add ioctl, access, tcgetattr, tcsetattr sysdeps | #define NR_sigaction 13
#define NR_rt_sigprocmask 14
#define NR_ioctl 16
+#define NR_access 21
#define NR_pipe 22
#define NR_select 23
#define NR_nanosleep 35
@@ -296,6 +297,15 @@ int sys_isatty(int fd) {
return 1;
}
+int sys_ioctl(int fd, unsigned long request, void *arg, int *result) {
+ auto ret = do_syscall(NR_ioctl, fd, request, arg);
+ if (int e = sc_error(ret); e)
+ return e;
+ if (result)
+ *result = sc_int_result<unsigned long>(ret);
+ return 0;
+}
+
int sys_connect(int sockfd, const struct sockaddr *addr, socklen_t addrlen) {
auto ret = do_cp_syscall(NR_connect, sockfd, addr, addrlen);
if (int e = sc_error(ret); e)
@@ -398,6 +408,36 @@ int sys_tgkill(int tgid, int tid, int sig) {
return 0;
}
+int sys_tcgetattr(int fd, struct termios *attr) {
+ auto ret = do_syscall(NR_ioctl, fd, TCGETS, attr);
+ if (int e = sc_error(ret); e)
+ return e;
+ return 0;
+}
+
+int sys_tcsetattr(int fd, int optional_action, const struct termios *attr) {
+ int req;
+
+ switch (optional_action) {
+ case TCSANOW: req = TCSETS; break;
+ case TCSADRAIN: req = TCSETSW; break;
+ case TCSAFLUSH: req = TCSETSF; break;
+ default: return EINVAL;
+ }
+
+ auto ret = do_syscall(NR_ioctl, fd, req, attr);
+ if (int e = sc_error(ret); e)
+ return e;
+ return 0;
+}
+
+int sys_access(const char *path, int mode) {
+ auto ret = do_syscall(NR_access, path, mode);
+ if (int e = sc_error(ret); e)
+ return e;
+ return 0;
+}
+
#endif // __MLIBC_POSIX_OPTION
pid_t sys_getpid() {
|
Fix complex ABI for 32bit SolarisStudio builds | @@ -78,6 +78,12 @@ static __inline unsigned long rpcc(void){
#define __BIG_ENDIAN__
#endif
+#ifdef C_SUN
+#ifndef __64BIT
+#define RETURN_BY_STACK
+#endif
+#endif
+
#ifdef DOUBLE
#define GET_IMAGE(res) __asm__ __volatile__("fmovd %%f2, %0" : "=f"(res) : : "memory")
#else
|
tweak tar assembly options | @@ -84,9 +84,6 @@ my $tmp_dir = File::Temp::tempdir(CLEANUP => 0, DIR=> $tmp_path) || MYERROR("Una
print "\nStaging dist creation in $tmp_dir...\n";
-#make_path($tmp_dir) || die("Unable to create $tmp_dir: $!\n");
-#die "Unable to create $tmp_dir: $!\n" unless(mkdir $tmp_dir);
-
if( !$skipCopy) {
# copy base repo contents
@@ -122,7 +119,7 @@ foreach my $distro (@distros) {
foreach my $arch (@arches) {
my $dist_filename = "$dest_dir/OpenHPC-$release.$distro\_$arch.tar";
my $tar_args = "-cvf $dist_filename \\\n";
- $tar_args .= "--cd $tmp_dir \\\n";
+
# exclude other arches, iso and src dirs
my @n_arches = grep { $_ ne $arch } @arches;
@@ -139,6 +136,9 @@ foreach my $distro (@distros) {
}
}
+ # cd to $tmp_dir
+ $tar_args .= "-C $tmp_dir \\\n";
+
# add .repo file
my $repoFile="$tmp_dir/$distro/OpenHPC.local.repo";
@@ -150,7 +150,7 @@ foreach my $distro (@distros) {
File::Copy::copy("repo.base.in","$repoFile") || die "Unable to copy repo.base.in\n";
if($micro) {
- print "Adding [update] repofile contents\n";
+ print "Appending [update] repo\n";
open (BASE, ">>","$repoFile") || die "Unable to append to repo file\n";
open (UPDATE,"<","repo.update.in") || die "Unable to access repo.update.in\n";
print BASE <UPDATE>;
@@ -166,8 +166,10 @@ foreach my $distro (@distros) {
$contents =~ s/\@VERSION_MICRO\@/$release/g;
$file_update->spew_utf8( $contents );
- $tar_args .= " $tmp_dir/$distro";
- print "\nCreating dist for $distro:$arch -- \n";
+ # Assemble final tarball
+# $tar_args .= " $tmp_dir/$distro";
+ $tar_args .= " /$distro";
+ print "\nCreating dist tarball for $distro:$arch -- \n";
print "tar command -> tar $tar_args\n";
system("tar $tar_args");
|
packaging: handle change in location of JSON schema for release | @@ -183,19 +183,9 @@ fi
# Sign YUM repo meta-data
find "/var/www/apt.fluentbit.io" -name repomd.xml -exec gpg --detach-sign --armor --yes -u "[email protected]" {} \;
-# Handle the JSON schema by copying in the new versions (if they exist) and then updating the symlinks that point at the latest.
-if compgen -G "$SOURCE_DIR/fluent-bit-schema*.json" > /dev/null; then
+# Handle the JSON schema by copying in the new versions (if they exist).
echo "Updating JSON schema"
- cp -vf "$SOURCE_DIR"/fluent-bit-schema*.json /var/www/releases.fluentbit.io/releases/"$MAJOR_VERSION/"
-
- # Simpler than 'ln --relative --target-directory=/var/www/releases.fluentbit.io/releases/"$MAJOR_VERSION"'
- pushd /var/www/releases.fluentbit.io/releases/"$MAJOR_VERSION"
- ln -sf "fluent-bit-schema-$VERSION.json" fluent-bit-schema.json
- ln -sf "fluent-bit-schema-pretty-$VERSION.json" fluent-bit-schema-pretty.json
- popd
-else
- echo "Missing JSON schema"
-fi
+find "$SOURCE_DIR/" -iname "fluent-bit-schema*$VERSION*.json" -exec cp -vf "{}" /var/www/releases.fluentbit.io/releases/"$MAJOR_VERSION/" \;
# Windows - we do want word splitting and ensure some files exist
if compgen -G "$SOURCE_DIR/windows/*$VERSION*" > /dev/null; then
|
Fix build failure on gcc-11
Function prototypes changed to use array parameters instead of
pointers. | @@ -680,20 +680,20 @@ static void ssl_calc_finished_ssl( mbedtls_ssl_context *, unsigned char *, int )
#endif
#if defined(MBEDTLS_SSL_PROTO_TLS1) || defined(MBEDTLS_SSL_PROTO_TLS1_1)
-static void ssl_calc_verify_tls( const mbedtls_ssl_context *, unsigned char *, size_t * );
+static void ssl_calc_verify_tls( const mbedtls_ssl_context *, unsigned char [36], size_t * );
static void ssl_calc_finished_tls( mbedtls_ssl_context *, unsigned char *, int );
#endif
#if defined(MBEDTLS_SSL_PROTO_TLS1_2)
#if defined(MBEDTLS_SHA256_C)
static void ssl_update_checksum_sha256( mbedtls_ssl_context *, const unsigned char *, size_t );
-static void ssl_calc_verify_tls_sha256( const mbedtls_ssl_context *,unsigned char *, size_t * );
+static void ssl_calc_verify_tls_sha256( const mbedtls_ssl_context *,unsigned char [32], size_t * );
static void ssl_calc_finished_tls_sha256( mbedtls_ssl_context *,unsigned char *, int );
#endif
#if defined(MBEDTLS_SHA512_C)
static void ssl_update_checksum_sha384( mbedtls_ssl_context *, const unsigned char *, size_t );
-static void ssl_calc_verify_tls_sha384( const mbedtls_ssl_context *, unsigned char *, size_t * );
+static void ssl_calc_verify_tls_sha384( const mbedtls_ssl_context *, unsigned char [48], size_t * );
static void ssl_calc_finished_tls_sha384( mbedtls_ssl_context *, unsigned char *, int );
#endif
#endif /* MBEDTLS_SSL_PROTO_TLS1_2 */
|
server: stream: check if input IOV is empty before write | @@ -184,6 +184,10 @@ int mk_channel_write(struct mk_channel *channel, size_t *count)
}
else if (input->type == MK_STREAM_IOV) {
iov = input->buffer;
+ if (!iov) {
+ return MK_CHANNEL_EMPTY;
+ }
+
bytes = mk_sched_conn_writev(channel, iov);
MK_TRACE("[CH %i] STREAM_IOV, wrote %d bytes",
|
firdespm: replacing assert calls with explicit checks | @@ -706,7 +706,8 @@ int firdespm_iext_search(firdespm _q)
if ( ((_q->E[i]>=0.0) && (_q->E[i-1]<=_q->E[i]) && (_q->E[i+1]<=_q->E[i]) ) ||
((_q->E[i]< 0.0) && (_q->E[i-1]>=_q->E[i]) && (_q->E[i+1]>=_q->E[i]) ) )
{
- assert(num_found < nmax);
+ //assert(num_found < nmax);
+ if (num_found < nmax)
found_iext[num_found++] = i;
#if LIQUID_FIRDESPM_DEBUG_PRINT
printf("num_found : %4u [%4u / %4u]\n", num_found, i, _q->grid_size);
@@ -720,7 +721,8 @@ int firdespm_iext_search(firdespm _q)
found_iext[num_found++] = _q->grid_size-1;
#else
// force f=0.5 into candidate set
- assert(num_found < nmax);
+ //assert(num_found < nmax);
+ if (num_found < nmax)
found_iext[num_found++] = _q->grid_size-1;
//printf("num_found : %4u [%4u / %4u]\n", num_found, _q->grid_size-1, _q->grid_size);
#endif
|
Minor bug solved from previous commit in MacOS. | @@ -74,7 +74,7 @@ int portability_executable_path(portability_executable_path_str path, portabilit
strncpy(path, path_ptr, *length);
#endif
- if (*length == -1 || *length == 0 || *length == path_max_length)
+ if (*length <= 0 || *length == path_max_length)
{
return 1;
}
|
read hostname for blacklist command | @@ -171,7 +171,7 @@ static void cmd_exec(FILE *fp, const char request[], int allow_debug)
} else if (sscanf( request, "announce %255[^: ]:%d %d %c", hostname, &port, &minutes, &d) == 3) {
cmd_announce(fp, hostname, port, minutes );
} else if (match(request, "list %*s %n") && allow_debug) {
- if (match(request, "blacklist %255[^: ]%n")) {
+ if (sscanf(request, "blacklist %255[^: ]", hostname) == 1) {
cmd_blacklist(fp, hostname);
} else if (gconf->is_daemon == 1) {
fprintf(fp ,"The 'list' command is not available while KadNode runs as daemon.\n" );
@@ -206,7 +206,6 @@ static void cmd_exec(FILE *fp, const char request[], int allow_debug)
} else {
fprintf(fp, "Unknown command.\n");
}
- fprintf(fp ,"\nOutput send to console.\n" );
} else {
// Print usage
fprintf(fp, g_server_usage);
|
SOVERSION bump to version 5.2.4 | @@ -39,7 +39,7 @@ set(SYSREPO_VERSION ${SYSREPO_MAJOR_VERSION}.${SYSREPO_MINOR_VERSION}.${SYSREPO_
# with backward compatible change and micro version is connected with any internal change of the library.
set(SYSREPO_MAJOR_SOVERSION 5)
set(SYSREPO_MINOR_SOVERSION 2)
-set(SYSREPO_MICRO_SOVERSION 3)
+set(SYSREPO_MICRO_SOVERSION 4)
set(SYSREPO_SOVERSION_FULL ${SYSREPO_MAJOR_SOVERSION}.${SYSREPO_MINOR_SOVERSION}.${SYSREPO_MICRO_SOVERSION})
set(SYSREPO_SOVERSION ${SYSREPO_MAJOR_SOVERSION})
|
Fix Error: src/cscaledbitmap.cxx:456:26: error: result of comparison of constant 256 with expression of type 'uint8_t' (aka 'unsigned char') is always true | @@ -433,7 +433,7 @@ bool CScaledBitmap::scaleColor(FAR const struct rgbcolor_s &incolor1,
FAR const struct rgbcolor_s &incolor2,
b16_t fraction, FAR struct rgbcolor_s &outcolor)
{
- uint8_t component;
+ uint32_t component;
b16_t red;
b16_t green;
b16_t blue;
|
dbginfo.sh: correct location of qemu guest agent command
qemu-ga command was accidently moved to z/VM commands | @@ -525,6 +525,7 @@ CMDS="${CMDS}\
# product / distro specific commands
CMDS="${CMDS}\
:java -version\
+ :qemu-ga -V\
:SPident # SLES service package\
"
@@ -659,7 +660,6 @@ VM_CMDS="q userid\
:ind load\
:ind sp\
:ind user\
- :qemu-ga -V\
"
###############################################################################
|
Fix lpc17_40_sdcard.c:614:24: error: unused function 'lpc17_40_getpwrctrl' | @@ -299,7 +299,6 @@ static void lpc17_40_configwaitints(struct lpc17_40_dev_s *priv,
static void lpc17_40_configxfrints(struct lpc17_40_dev_s *priv,
uint32_t xfrmask);
static void lpc17_40_setpwrctrl(uint32_t pwrctrl);
-static inline uint32_t lpc17_40_getpwrctrl(void);
/* DMA Helpers **************************************************************/
@@ -595,29 +594,6 @@ static void lpc17_40_setpwrctrl(uint32_t pwrctrl)
putreg32(regval, LPC17_40_SDCARD_PWR);
}
-/****************************************************************************
- * Name: lpc17_40_getpwrctrl
- *
- * Description:
- * Return the current value of the the PWRCTRL field of the SD card P
- * register. This function can be used to see if the SD card is powered ON
- * or OFF
- *
- * Input Parameters:
- * None
- *
- * Returned Value:
- * The current value of the the PWRCTRL field of the SD card PWR register.
- *
- ****************************************************************************/
-
-static inline uint32_t lpc17_40_getpwrctrl(void)
-{
- /* Extract and return the PWRCTRL field of the PWR register. */
-
- return getreg32(LPC17_40_SDCARD_PWR) & SDCARD_PWR_CTRL_MASK;
-}
-
/****************************************************************************
* DMA Helpers
****************************************************************************/
|
check-lxdialog.sh: Support libintl on FreeBSD
To unbreak "make menuconfig."
Merges | @@ -7,6 +7,10 @@ ldflags()
if [ $(uname -s) == "Darwin" ]; then
#OSX seems to need ncurses too
echo -n "-lncurses "
+ elif [ $(uname -s) == "FreeBSD" ]; then
+ # On FreeBSD, the linker needs to know to search port libs for
+ # libintl
+ echo -n "-L/usr/local/lib -lintl "
fi
pkg-config --libs ncursesw 2>/dev/null && exit
pkg-config --libs ncurses 2>/dev/null && exit
@@ -44,6 +48,10 @@ ccflags()
if [ $(uname -s) == "Darwin" ]; then
#OSX doesn't have libintl
echo -n "-DKBUILD_NO_NLS -Wno-format-security "
+ elif [ $(uname -s) == "FreeBSD" ]; then
+ # FreeBSD gettext port has libintl.h, but the compiler needs
+ # to know to search port includes
+ echo -n "-I/usr/local/include "
fi
}
|
OpenCanopy: Fix broken selector click event | @@ -905,7 +905,8 @@ InternalBootPickerSelectorPtrEvent (
if (IsHit) {
if (Event == GuiPointerPrimaryUp) {
ASSERT (mBootPicker.SelectedEntry != NULL);
- Context->BootEntry = mBootPicker.SelectedEntry->Context;
+ ASSERT (Context->BootEntry == mBootPicker.SelectedEntry->Context);
+ Context->ReadyToBoot = TRUE;
} else {
ButtonImage = &Context->Icons[ICON_SELECTOR][ICON_TYPE_HELD];
}
|
ps8xxx: define PS8815 USB EQ register
BRANCH=none
TEST=make -j BOARD=gimble | /* Vendor defined registers */
#define PS8815_P1_REG_HW_REVISION 0xF0
+/* Vendor defined registers */
+#define PS8815_REG_APTX_EQ_AT_10G 0x20
+#define PS8815_REG_RX_EQ_AT_10G 0x22
+#define PS8815_REG_APTX_EQ_AT_5G 0x24
+#define PS8815_REG_RX_EQ_AT_5G 0x26
+
/*
* Below register is defined from Parade PS8815 Register Table,
* See b:189587527 for more detail.
|
Add `DIAG_LEVEL_REQUIRED` to fix build break in modules. | @@ -86,6 +86,7 @@ constexpr std::uint8_t PDL_NONE {0u}; /* Logg
#define DIAG_LEVEL_FULL 3 /* Extra activity and enhanced reporting */
/* Microsoft NGP diagnostic level classification */
+constexpr std::uint8_t DIAG_LEVEL_REQUIRED {1u}; /* Data that we need to collect in order to keep the product secure, up to date, and performing as expected */
constexpr std::uint8_t PDL_REQUIRED {1u}; /* Data that we need to collect in order to keep the product secure, up to date, and performing as expected */
constexpr std::uint8_t PDL_OPTIONAL {2u}; /* Additional optional data */
constexpr std::uint8_t PDL_REQUIREDSERVICEDATA {110u}; /* Data required for services to be able to function properly */
|
Fix error handling/cleanup | @@ -473,19 +473,11 @@ static int i2d_name_canon(STACK_OF(STACK_OF_X509_NAME_ENTRY) * _intname,
int X509_NAME_set(X509_NAME **xn, X509_NAME *name)
{
- X509_NAME *in;
-
- if (!xn || !name)
- return (0);
-
- if (*xn != name) {
- in = X509_NAME_dup(name);
- if (in != NULL) {
+ if ((name = X509_NAME_dup(name)) == NULL)
+ return 0;
X509_NAME_free(*xn);
- *xn = in;
- }
- }
- return (*xn != NULL);
+ *xn = name;
+ return 1;
}
int X509_NAME_print(BIO *bp, const X509_NAME *name, int obase)
|
Add a format flag to tar to prevent sparse files.
Turns out `bsdtar` will by default create an archive which 7-zip can't
properly read. That make these archives very confusing to check on. | @@ -543,7 +543,7 @@ package_files = (
if sys.platform == "win32":
zip = env.Zip("${DIST_NAME}.zip", package_files)
else:
- env.Append(TARFLAGS="-c -z")
+ env.Append(TARFLAGS="-z --format=ustar")
zip = env.Tar("${DIST_NAME}.tar.gz", package_files)
Alias("dist", zip)
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.