id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
1624493576
HDDS-8069. [Snapshot] Compaction DAG reconstruction logic potentially crashing OM on startup What changes were proposed in this pull request? Problem: As part of jira-HDDS-7873, we added an optimization to early prune out SST files that won't be needed for compaction DAG base snap diff. More details in PR-4235. Files are removed from back dir but compaction logs don't get updated because those are needed for DAG traversal based diffing. On OM restart, we count number of keys in the file present in compaction log. It is possible that file has been deleted due to above optimization which is causing OM crash. This change is to log the exception instead throwing exception on OM restart. It will not affect Compaction DAG based snap diff because If file is not present during diff, it will throw the FileNotFoundException. What is the link to the Apache JIRA https://issues.apache.org/jira/browse/HDDS-8069 How was this patch tested? Unit tests. Thanks @hemantk-12 for the patch.
gharchive/pull-request
2023-03-15T00:13:07
2025-04-01T04:55:59.633994
{ "authors": [ "hemantk-12", "smengcl" ], "repo": "apache/ozone", "url": "https://github.com/apache/ozone/pull/4401", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2225632311
HDDS-5569. recipe/BotoClient.md translation What changes were proposed in this pull request? HDDS-5569. recipe/BotoClient.md translation to Chinese What is the link to the Apache JIRA https://issues.apache.org/jira/browse/HDDS-5569 How was this patch tested? hugo serve thx for review @smengcl Thanks @will-sh for the PR. I think there is a Github glitch that prevents it from marking this PR as Merged. It is already merged in master: https://github.com/apache/ozone/commit/5d999d2068618b9e1e8ed66188b3ea22dc225d1d I will leave this PR here for a few hours to see if it fixes itself. If not I will close this manually. Thanks @will-sh for the translation, @smengcl for the review.
gharchive/pull-request
2024-04-04T14:12:41
2025-04-01T04:55:59.637221
{ "authors": [ "adoroszlai", "smengcl", "will-sh" ], "repo": "apache/ozone", "url": "https://github.com/apache/ozone/pull/6480", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2231264966
[compaction] support multi table AppendOnlyCompactionTask Purpose related https://github.com/apache/paimon/issues/2670 Tests API and Format Documentation @leaves12138 PTAL finished at #2858
gharchive/pull-request
2024-04-08T13:42:42
2025-04-01T04:55:59.639295
{ "authors": [ "wg1026688210" ], "repo": "apache/paimon", "url": "https://github.com/apache/paimon/pull/3174", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2368056669
build failed on AArch64, Fedora 33 Apache Thrift 0.12.0 is required. Building it reports unsupported .NET, etc. Installing 0.13.0 using yum results in an error on mvn package. Reporter: Lutz Weischer Note: This issue was originally created as PARQUET-2000. Please see the migration documentation for further details. Lutz Weischer: I restarted from scratch and get this: LC_ALL=C mvn install ... Testing PARQUET_1_0/FIXED_LEN_BYTE_ARRAY/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384 Testing PARQUET_2_0/FIXED_LEN_BYTE_ARRAY/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384 Tests run: 16, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1,100.501 sec Results : Tests in error: testInteropReadEncryptedParquetFiles(org.apache.parquet.hadoop.ITTestEncryptionOptions): github.com: Name or service not known Tests run: 17, Failures: 0, Errors: 1, Skipped: 0 [INFO] [INFO] — maven-failsafe-plugin:2.10:verify (default) @ parquet-hadoop — [INFO] Failsafe report directory: /data/jw/parquet/parquet-hadoop/target/failsafe-reports [INFO] ------------------------------------------------------------------------ [INFO] Reactor Summary for Apache Parquet MR 1.12.0-SNAPSHOT: [INFO] [INFO] Apache Parquet MR .................................. SUCCESS [ 8.688 s] [INFO] Apache Parquet Format Structures ................... SUCCESS [ 8.891 s] [INFO] Apache Parquet Generator ........................... SUCCESS [ 3.289 s] [INFO] Apache Parquet Common .............................. SUCCESS [ 10.041 s] [INFO] Apache Parquet Encodings ........................... SUCCESS [ 33.835 s] [INFO] Apache Parquet Column .............................. SUCCESS [01:32 min] [INFO] Apache Parquet Arrow ............................... SUCCESS [ 3.288 s] [INFO] Apache Parquet Jackson ............................. SUCCESS [ 4.596 s] [INFO] Apache Parquet Hadoop .............................. FAILURE [22:59 min] [INFO] Apache Parquet Avro ................................ SKIPPED [INFO] Apache Parquet Benchmarks .......................... SKIPPED [INFO] Apache Parquet Pig ................................. SKIPPED [INFO] Apache Parquet Thrift .............................. SKIPPED [INFO] Apache Parquet Cascading [Deprecated, will be removed after 1.12] SKIPPED [INFO] Apache Parquet Cascading (for Cascading 3.0 onwards) [Deprecated, will be removed after 1.12] SKIPPED [INFO] Apache Parquet Command-line ........................ SKIPPED [INFO] Apache Parquet Pig Bundle .......................... SKIPPED [INFO] Apache Parquet Protobuf ............................ SKIPPED [INFO] Apache Parquet Scala ............................... SKIPPED [INFO] Apache Parquet Scrooge [Deprecated, will be removed after 1.12] SKIPPED [INFO] Apache Parquet Hadoop Bundle ....................... SKIPPED [INFO] Apache Parquet Tools; [Deprecated, will be removed after 1.12; Use parquet-cli instead] SKIPPED [INFO] ------------------------------------------------------------------------ [INFO] BUILD FAILURE [INFO] ------------------------------------------------------------------------ [INFO] Total time: 25:45 min [INFO] Finished at: 2021-03-24T15:06:55+01:00 [INFO] ------------------------------------------------------------------------ [ERROR] Failed to execute goal org.apache.maven.plugins:maven-failsafe-plugin:2.10:verify (default) on project parquet-hadoop: There are test failures. [ERROR] [ERROR] Please refer to /data/jw/parquet/parquet-hadoop/target/failsafe-reports for the individual test results. [ERROR] -> [Help 1] [ERROR] [ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch. [ERROR] Re-run Maven using the -X switch to enable full debug logging. [ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles: [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException [ERROR] [ERROR] After correcting the problems, you can resume the build with the command [ERROR] mvn -rf :parquet-hadoop (base) [jw@cn06 parquet]$ I can access github.com, because I git clone'd from it. (base) [jw@cn06 parquet]$ ls /data/jw/parquet/parquet-hadoop/target/failsafe-reports TEST-org.apache.parquet.encodings.FileEncodingsIT.xml org.apache.parquet.encodings.FileEncodingsIT.txt TEST-org.apache.parquet.hadoop.ITTestEncryptionOptions.xml org.apache.parquet.hadoop.ITTestEncryptionOptions.txt failsafe-summary.xml (base) [jw@cn06 parquet]$ nano /data/jw/parquet/parquet-hadoop/target/failsafe-reports/failsafe-summary.xml shows Gabor Szadovszky / @gszadovszky: Could you try to get the file https://github.com/apache/parquet-testing/raw/40379b3/data/encrypt_columns_and_footer.parquet.encrypted manually (e.g. with wget or curl)? It would also help if you could dig up the whole stack trace from failsafe-reports. @andersonm-ibm, do you have an idea why it is happening and why we get the error message github.com: Name or service not known? Lutz Weischer: Switched to parquet-mr. [jw@cn05 parquet]$ LC_ALL=C mvn install ... [INFO] [INFO] — maven-jar-plugin:2.4:jar (default-jar) @ parquet-hadoop — [INFO] Building jar: /home/jw/apache/parquet/parquet-hadoop/target/parquet-hadoop-1.12.0-SNAPSHOT.jar [INFO] [INFO] — maven-site-plugin:3.4:attach-descriptor (attach-descriptor) @ parquet-hadoop — [INFO] [INFO] — maven-jar-plugin:2.4:test-jar (default) @ parquet-hadoop — [INFO] Building jar: /home/jw/apache/parquet/parquet-hadoop/target/parquet-hadoop-1.12.0-SNAPSHOT-tests.jar [INFO] [INFO] — maven-shade-plugin:3.2.1:shade (default) @ parquet-hadoop — [INFO] Excluding org.apache.parquet:parquet-column:jar:1.12.0-SNAPSHOT from the shaded jar. [INFO] Excluding org.apache.parquet:parquet-common:jar:1.12.0-SNAPSHOT from the shaded jar. [INFO] Excluding org.apache.yetus:audience-annotations:jar:0.12.0 from the shaded jar. [INFO] Excluding org.apache.parquet:parquet-encoding:jar:1.12.0-SNAPSHOT from the shaded jar. [INFO] Including net.openhft:zero-allocation-hashing:jar:0.9 in the shaded jar. [INFO] Excluding org.apache.parquet:parquet-format-structures:jar:1.12.0-SNAPSHOT from the shaded jar. [INFO] Excluding org.slf4j:slf4j-api:jar:1.7.22 from the shaded jar. [INFO] Excluding javax.annotation:javax.annotation-api:jar:1.3.2 from the shaded jar. [INFO] Excluding org.apache.parquet:parquet-jackson:jar:1.12.0-SNAPSHOT from the shaded jar. [INFO] Including com.fasterxml.jackson.core:jackson-core:jar:2.12.2 in the shaded jar. [INFO] Including com.fasterxml.jackson.core:jackson-databind:jar:2.12.2 in the shaded jar. [INFO] Including com.fasterxml.jackson.core:jackson-annotations:jar:2.12.2 in the shaded jar. [INFO] Excluding org.xerial.snappy:snappy-java:jar:1.1.8 from the shaded jar. [INFO] Excluding commons-pool:commons-pool:jar:1.6 from the shaded jar. [INFO] Including it.unimi.dsi:fastutil:jar:8.4.2 in the shaded jar. [INFO] Excluding com.github.luben:zstd-jni:jar:1.4.9-1 from the shaded jar. [INFO] Minimizing jar org.apache.parquet:parquet-hadoop:jar:1.12.0-SNAPSHOT [INFO] Minimized 11128 -> 491 (4%) [INFO] Replacing original artifact with shaded artifact. [INFO] Replacing /home/jw/apache/parquet/parquet-hadoop/target/parquet-hadoop-1.12.0-SNAPSHOT.jar with /home/jw/apache/parquet/parquet-hadoop/target/parquet-hadoop-1.12.0-SNAPSHOT-shaded.jar [INFO] Dependency-reduced POM written at: /home/jw/apache/parquet/parquet-hadoop/dependency-reduced-pom.xml [INFO] [INFO] — maven-failsafe-plugin:2.10:integration-test (default) @ parquet-hadoop — [INFO] Failsafe report directory: /home/jw/apache/parquet/parquet-hadoop/target/failsafe-reports T E S T S Running org.apache.parquet.hadoop.ITTestEncryptionOptions [main] INFO org.apache.parquet.hadoop.TestEncryptionOptions - ======== testInteropReadEncryptedParquetFiles target/parquet-testing/data ======== [main] INFO org.apache.parquet.hadoop.TestEncryptionOptions - Download interop files if needed WARNING: An illegal reflective access operation has occurred WARNING: Illegal reflective access by org.apache.hadoop.security.authentication.util.KerberosUtil (file:/home/jw/.m2/repository/org/apache/hadoop/hadoop-auth/2.10.1/hadoop-auth-2.10.1.jar) to method sun.security.krb5.Config.getInstance() WARNING: Please consider reporting this to the maintainers of org.apache.hadoop.security.authentication.util.KerberosUtil WARNING: Use --illegal-access=warn to enable warnings of further illegal reflective access operations WARNING: All illegal access operations will be denied in a future release [main] WARN org.apache.hadoop.util.NativeCodeLoader - Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 10:20:52,298 DEBUG Tracer: 106 - sampler.classes = ; loaded no samplers 10:20:52,318 DEBUG Tracer: 128 - span.receiver.classes = ; loaded no span receivers [main] INFO org.apache.parquet.hadoop.TestEncryptionOptions - target/parquet-testing/data exists?: true [main] INFO org.apache.parquet.hadoop.TestEncryptionOptions - Download interop file: https://github.com/apache/parquet-testing/raw/40379b3/data/uniform_encryption.parquet.encrypted Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 8.554 sec <<< FAILURE! Running org.apache.parquet.encodings.FileEncodingsIT Testing codecs: [UNCOMPRESSED] Testing PARQUET_1_0/BOOLEAN/UNCOMPRESSED encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384 Testing PARQUET_2_0/BOOLEAN/UNCOMPRESSED encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384 Testing PARQUET_1_0/BOOLEAN/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384 Testing PARQUET_2_0/BOOLEAN/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384 Testing PARQUET_1_0/INT32/UNCOMPRESSED encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384 Testing PARQUET_2_0/INT32/UNCOMPRESSED encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384 Testing PARQUET_1_0/INT32/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384 Testing PARQUET_2_0/INT32/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384 Testing PARQUET_1_0/INT64/UNCOMPRESSED encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384 Testing PARQUET_2_0/INT64/UNCOMPRESSED encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384 Testing PARQUET_1_0/INT64/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384 Testing PARQUET_2_0/INT64/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384 Testing PARQUET_1_0/INT96/UNCOMPRESSED encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384 Testing PARQUET_2_0/INT96/UNCOMPRESSED encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384 Testing PARQUET_1_0/INT96/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384 Testing PARQUET_2_0/INT96/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384 Testing PARQUET_1_0/FLOAT/UNCOMPRESSED encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384 Testing PARQUET_2_0/FLOAT/UNCOMPRESSED encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384 Testing PARQUET_1_0/FLOAT/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384 Testing PARQUET_2_0/FLOAT/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384 Testing PARQUET_1_0/DOUBLE/UNCOMPRESSED encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384 Testing PARQUET_2_0/DOUBLE/UNCOMPRESSED encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384 Testing PARQUET_1_0/DOUBLE/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384 Testing PARQUET_2_0/DOUBLE/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384 Testing PARQUET_1_0/BINARY/UNCOMPRESSED encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384 Testing PARQUET_2_0/BINARY/UNCOMPRESSED encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384 Testing PARQUET_1_0/BINARY/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384 Testing PARQUET_2_0/BINARY/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384 Testing PARQUET_1_0/FIXED_LEN_BYTE_ARRAY/UNCOMPRESSED encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384 Testing PARQUET_2_0/FIXED_LEN_BYTE_ARRAY/UNCOMPRESSED encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384 Testing PARQUET_1_0/FIXED_LEN_BYTE_ARRAY/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384 Testing PARQUET_2_0/FIXED_LEN_BYTE_ARRAY/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384 Tests run: 16, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1,108.336 sec Results : Tests in error: testInteropReadEncryptedParquetFiles(org.apache.parquet.hadoop.ITTestEncryptionOptions): github.com: Name or service not known Tests run: 17, Failures: 0, Errors: 1, Skipped: 0 [INFO] [INFO] — maven-failsafe-plugin:2.10:verify (default) @ parquet-hadoop — [INFO] Failsafe report directory: /home/jw/apache/parquet/parquet-hadoop/target/failsafe-reports [INFO] ------------------------------------------------------------------------ [INFO] Reactor Summary for Apache Parquet MR 1.12.0-SNAPSHOT: [INFO] [INFO] Apache Parquet MR .................................. SUCCESS [ 10.510 s] [INFO] Apache Parquet Format Structures ................... SUCCESS [ 9.290 s] [INFO] Apache Parquet Generator ........................... SUCCESS [ 3.472 s] [INFO] Apache Parquet Common .............................. SUCCESS [ 10.361 s] [INFO] Apache Parquet Encodings ........................... SUCCESS [ 37.394 s] [INFO] Apache Parquet Column .............................. SUCCESS [01:41 min] [INFO] Apache Parquet Arrow ............................... SUCCESS [ 3.497 s] [INFO] Apache Parquet Jackson ............................. SUCCESS [ 5.068 s] [INFO] Apache Parquet Hadoop .............................. FAILURE [23:20 min] [INFO] Apache Parquet Avro ................................ SKIPPED [INFO] Apache Parquet Benchmarks .......................... SKIPPED [INFO] Apache Parquet Pig ................................. SKIPPED [INFO] Apache Parquet Thrift .............................. SKIPPED [INFO] Apache Parquet Cascading [Deprecated, will be removed after 1.12] SKIPPED [INFO] Apache Parquet Cascading (for Cascading 3.0 onwards) [Deprecated, will be removed after 1.12] SKIPPED [INFO] Apache Parquet Command-line ........................ SKIPPED [INFO] Apache Parquet Pig Bundle .......................... SKIPPED [INFO] Apache Parquet Protobuf ............................ SKIPPED [INFO] Apache Parquet Scala ............................... SKIPPED [INFO] Apache Parquet Scrooge [Deprecated, will be removed after 1.12] SKIPPED [INFO] Apache Parquet Hadoop Bundle ....................... SKIPPED [INFO] Apache Parquet Tools; [Deprecated, will be removed after 1.12; Use parquet-cli instead] SKIPPED [INFO] ------------------------------------------------------------------------ [INFO] BUILD FAILURE [INFO] ------------------------------------------------------------------------ [INFO] Total time: 26:21 min [INFO] Finished at: 2021-03-22T10:39:29+01:00 [INFO] ------------------------------------------------------------------------ [ERROR] Failed to execute goal org.apache.maven.plugins:maven-failsafe-plugin:2.10:verify (default) on project parquet-hadoop: There are test failures. [ERROR] [ERROR] Please refer to /home/jw/apache/parquet/parquet-hadoop/target/failsafe-reports for the individual test results. [ERROR] -> [Help 1] [ERROR] [ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch. [ERROR] Re-run Maven using the -X switch to enable full debug logging. [ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles: [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException [ERROR] [ERROR] After correcting the problems, you can resume the build with the command [ERROR] mvn -rf :parquet-hadoop [jw@cn05 parquet]$ Gabor Szadovszky / @gszadovszky: The current master release requires 0.13.0. The build would fail with any other version of the thrift binary. Could you please explain what mvn error do you see when you use 0.13.0? Gabor Szadovszky / @gszadovszky: The failing test tries to download a file from github directly. The exact link is listed in the log. It seems github.com is not accessible from your environment. Lutz Weischer: [jw@cn06 parquet]$ mvn package [INFO] Scanning for projects... [INFO] [INFO] -----------------< org.apache.parquet:parquet-format >------------------ [INFO] Building Apache Parquet Format 2.9.0-SNAPSHOT [INFO] -------------------------------[ jar ]-------------------------------- [INFO] [INFO] — thrift-maven-plugin:0.10.0:compile (thrift-sources) @ parquet-format — [INFO] [INFO] — exec-maven-plugin:1.2.1:exec (check-thrift-version) @ parquet-format — ========== [FATAL] Build is configured to require Thrift version 0.12.0 ========== ========== Currently installed: Thrift version 0.13.0 [INFO] ------------------------------------------------------------------------ [INFO] BUILD FAILURE [INFO] ------------------------------------------------------------------------ [INFO] Total time: 2.682 s [INFO] Finished at: 2021-03-17T08:54:58+01:00 [INFO] ------------------------------------------------------------------------ [ERROR] Failed to execute goal org.codehaus.mojo:exec-maven-plugin:1.2.1:exec (check-thrift-version) on project parquet-format: Command execution failed.: Process exited with an error: 1 (Exit value: 1) -> [Help 1] [ERROR] [ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch. [ERROR] Re-run Maven using the -X switch to enable full debug logging. [ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles: [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException [jw@cn06 parquet]$ thrift -version Thrift version 0.13.0 [jw@cn06 parquet]$ mvn install -DskipTests [INFO] Scanning for projects... [INFO] [INFO] -----------------< org.apache.parquet:parquet-format >------------------ [INFO] Building Apache Parquet Format 2.9.0-SNAPSHOT [INFO] -------------------------------[ jar ]-------------------------------- [INFO] [INFO] — thrift-maven-plugin:0.10.0:compile (thrift-sources) @ parquet-format — [INFO] [INFO] — exec-maven-plugin:1.2.1:exec (check-thrift-version) @ parquet-format — ========== [FATAL] Build is configured to require Thrift version 0.12.0 ========== ========== Currently installed: Thrift version 0.13.0 [INFO] ------------------------------------------------------------------------ [INFO] BUILD FAILURE [INFO] ------------------------------------------------------------------------ [INFO] Total time: 2.428 s [INFO] Finished at: 2021-03-17T08:56:36+01:00 [INFO] ------------------------------------------------------------------------ [ERROR] Failed to execute goal org.codehaus.mojo:exec-maven-plugin:1.2.1:exec (check-thrift-version) on project parquet-format: Command execution failed.: Process exited with an error: 1 (Exit value: 1) -> [Help 1] [ERROR] [ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch. [ERROR] Re-run Maven using the -X switch to enable full debug logging. [ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles: [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException [jw@cn06 parquet]$ Gabor Szadovszky / @gszadovszky: So, you are trying to build parquet-format (and not parquet-mr). parquet-format still configured for thrift 0.12.0 while it does not really means anything because the parquet-format thrift compile is only for testing purposes. The production code for parquet-format is generated inside parquet-mr (in the module parquet-format-structures). Why are you trying to build parquet-format? If you are interested in the java implementation of parquet I would suggest looking into parquet-mr. Lutz Weischer: I got a fresh installation and have now this result: (base) [jw@cn05 parquet]$ LC_ALL=C mvn install -DskipTests ... [INFO] Apache Parquet Pig ................................. SUCCESS [ 9.340 s] [INFO] Apache Parquet Thrift .............................. SUCCESS [ 11.328 s] [INFO] Apache Parquet Cascading [Deprecated, will be removed after 1.12] FAILURE [ 1.496 s] [INFO] Apache Parquet Cascading (for Cascading 3.0 onwards) [Deprecated, will be removed after 1.12] SKIPPED [INFO] Apache Parquet Command-line ........................ SKIPPED [INFO] Apache Parquet Pig Bundle .......................... SKIPPED [INFO] Apache Parquet Protobuf ............................ SKIPPED [INFO] Apache Parquet Scala ............................... SKIPPED [INFO] Apache Parquet Scrooge [Deprecated, will be removed after 1.12] SKIPPED [INFO] Apache Parquet Hadoop Bundle ....................... SKIPPED [INFO] Apache Parquet Tools; [Deprecated, will be removed after 1.12; Use parquet-cli instead] SKIPPED [INFO] ------------------------------------------------------------------------ [INFO] BUILD FAILURE [INFO] ------------------------------------------------------------------------ [INFO] Total time: 03:21 min [INFO] Finished at: 2021-03-26T14:08:18+01:00 [INFO] ------------------------------------------------------------------------ [ERROR] Failed to execute goal com.github.siom79.japicmp:japicmp-maven-plugin:0.14.2:cmp (default) on project parquet-cascading-deprecated: Could not resolve org.apache.parquet:parquet-cascading:jar:1.12.0 -> [Help 1] [ERROR] [ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch. [ERROR] Re-run Maven using the -X switch to enable full debug logging. [ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles: [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException [ERROR] [ERROR] After correcting the problems, you can resume the build with the command [ERROR] mvn -rf :parquet-cascading-deprecated (base) [jw@cn05 parquet]$ Trying differently: (base) [jw@cn05 parquet]$ mvn install -DskipTests ... [INFO] Apache Parquet Pig ................................. SUCCESS [ 8.357 s] [INFO] Apache Parquet Thrift .............................. SUCCESS [ 10.033 s] [INFO] Apache Parquet Cascading [Deprecated, will be removed after 1.12] FAILURE [ 0.686 s] [INFO] Apache Parquet Cascading (for Cascading 3.0 onwards) [Deprecated, will be removed after 1.12] SKIPPED [INFO] Apache Parquet Command-line ........................ SKIPPED [INFO] Apache Parquet Pig Bundle .......................... SKIPPED [INFO] Apache Parquet Protobuf ............................ SKIPPED [INFO] Apache Parquet Scala ............................... SKIPPED [INFO] Apache Parquet Scrooge [Deprecated, will be removed after 1.12] SKIPPED [INFO] Apache Parquet Hadoop Bundle ....................... SKIPPED [INFO] Apache Parquet Tools; [Deprecated, will be removed after 1.12; Use parquet-cli instead] SKIPPED [INFO] ------------------------------------------------------------------------ [INFO] BUILD FAILURE [INFO] ------------------------------------------------------------------------ [INFO] Total time: 03:06 min [INFO] Finished at: 2021-03-26T14:33:17+01:00 [INFO] ------------------------------------------------------------------------ [ERROR] Failed to execute goal org.apache.rat:apache-rat-plugin:0.12:check (default) on project parquet-cascading-deprecated: Too many files with unapproved license: 1 See RAT report in: /data/jw/apache/metron/parquet/parquet-cascading-deprecated/target/rat.txt -> [Help 1] [ERROR] [ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch. [ERROR] Re-run Maven using the -X switch to enable full debug logging. [ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles: [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException [ERROR] [ERROR] After correcting the problems, you can resume the build with the command [ERROR] mvn -rf :parquet-cascading-deprecated (base) [jw@cn05 parquet]$ And: (base) [jw@cn05 parquet]$ mvn install -DskipTests -Drat.skip=true ... [INFO] Apache Parquet Pig ................................. SUCCESS [ 8.106 s] [INFO] Apache Parquet Thrift .............................. SUCCESS [ 9.405 s] [INFO] Apache Parquet Cascading [Deprecated, will be removed after 1.12] FAILURE [ 0.566 s] [INFO] Apache Parquet Cascading (for Cascading 3.0 onwards) [Deprecated, will be removed after 1.12] SKIPPED [INFO] Apache Parquet Command-line ........................ SKIPPED [INFO] Apache Parquet Pig Bundle .......................... SKIPPED [INFO] Apache Parquet Protobuf ............................ SKIPPED [INFO] Apache Parquet Scala ............................... SKIPPED [INFO] Apache Parquet Scrooge [Deprecated, will be removed after 1.12] SKIPPED [INFO] Apache Parquet Hadoop Bundle ....................... SKIPPED [INFO] Apache Parquet Tools; [Deprecated, will be removed after 1.12; Use parquet-cli instead] SKIPPED [INFO] ------------------------------------------------------------------------ [INFO] BUILD FAILURE [INFO] ------------------------------------------------------------------------ [INFO] Total time: 03:03 min [INFO] Finished at: 2021-03-26T14:42:35+01:00 [INFO] ------------------------------------------------------------------------ [ERROR] Failed to execute goal com.github.siom79.japicmp:japicmp-maven-plugin:0.14.2:cmp (default) on project parquet-cascading-deprecated: Could not resolve org.apache.parquet:parquet-cascading:jar:1.12.0 -> [Help 1] [ERROR] [ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch. [ERROR] Re-run Maven using the -X switch to enable full debug logging. [ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles: [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException [ERROR] [ERROR] After correcting the problems, you can resume the build with the command [ERROR] mvn -rf :parquet-cascading-deprecated (base) [jw@cn05 parquet]$ Gabor Szadovszky / @gszadovszky: [~LutzWeischerFujitsu], this is the previous issue you've already posted. Please, check my [previous comment|#comment-17307872]. Lutz Weischer: Looks like we've progress but we're not there yet, due to failing tests. Using Java 11, setting ANT_OPTS with proxy: (base) [jw@cn06 parquet]$ LC_ALL=C mvn install ... Testing PARQUET_2_0/FIXED_LEN_BYTE_ARRAY/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384 Tests run: 16, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1,062.679 sec Results : Tests in error: testInteropReadEncryptedParquetFiles(org.apache.parquet.hadoop.ITTestEncryptionOptions): github.com: Name or service not known Tests run: 17, Failures: 0, Errors: 1, Skipped: 0 [INFO] [INFO] — maven-failsafe-plugin:2.10:verify (default) @ parquet-hadoop — [INFO] Failsafe report directory: /data/jw/apache/parquet/parquet-hadoop/target/failsafe-reports [INFO] ------------------------------------------------------------------------ [INFO] Reactor Summary for Apache Parquet MR 1.13.0-SNAPSHOT: [INFO] [INFO] Apache Parquet MR .................................. SUCCESS [ 8.110 s] [INFO] Apache Parquet Format Structures ................... SUCCESS [ 8.766 s] [INFO] Apache Parquet Generator ........................... SUCCESS [ 3.380 s] [INFO] Apache Parquet Common .............................. SUCCESS [ 9.466 s] [INFO] Apache Parquet Encodings ........................... SUCCESS [ 35.689 s] [INFO] Apache Parquet Column .............................. SUCCESS [01:33 min] [INFO] Apache Parquet Arrow ............................... SUCCESS [ 3.252 s] [INFO] Apache Parquet Jackson ............................. SUCCESS [ 4.634 s] [INFO] Apache Parquet Hadoop .............................. FAILURE [22:20 min] [INFO] Apache Parquet Avro ................................ SKIPPED [INFO] Apache Parquet Benchmarks .......................... SKIPPED [INFO] Apache Parquet Pig ................................. SKIPPED [INFO] Apache Parquet Thrift .............................. SKIPPED [INFO] Apache Parquet Cascading [Deprecated, will be removed after 1.12] SKIPPED [INFO] Apache Parquet Cascading (for Cascading 3.0 onwards) [Deprecated, will be removed after 1.12] SKIPPED [INFO] Apache Parquet Command-line ........................ SKIPPED [INFO] Apache Parquet Pig Bundle .......................... SKIPPED [INFO] Apache Parquet Protobuf ............................ SKIPPED [INFO] Apache Parquet Scala ............................... SKIPPED [INFO] Apache Parquet Scrooge [Deprecated, will be removed after 1.12] SKIPPED [INFO] Apache Parquet Hadoop Bundle ....................... SKIPPED [INFO] Apache Parquet Tools; [Deprecated, will be removed after 1.12; Use parquet-cli instead] SKIPPED [INFO] ------------------------------------------------------------------------ [INFO] BUILD FAILURE [INFO] ------------------------------------------------------------------------ [INFO] Total time: 25:08 min [INFO] Finished at: 2021-04-01T12:04:30+02:00 [INFO] ------------------------------------------------------------------------ [ERROR] Failed to execute goal org.apache.maven.plugins:maven-failsafe-plugin:2.10:verify (default) on project parquet-hadoop: There are test failures. [ERROR] [ERROR] Please refer to /data/jw/apache/parquet/parquet-hadoop/target/failsafe-reports for the individual test results. [ERROR] -> [Help 1] [ERROR] [ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch. [ERROR] Re-run Maven using the -X switch to enable full debug logging. [ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles: [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException [ERROR] [ERROR] After correcting the problems, you can resume the build with the command [ERROR] mvn -rf :parquet-hadoop (base) [jw@cn06 parquet]$ Lutz Weischer: wget https://github.com/apache/parquet-testing/raw/40379b3/data/encrypt_columns_and_footer.parquet.encrypted works. LC_ALL=C mvn clean install which includes the tests, works, using Java 16. Maya Anderson / @andersonm-ibm: @gszadovszky, just like you suggested, I expect the same error when trying to get the file [https://github.com/apache/parquet-testing/raw/40379b3/data/encrypt_columns_and_footer.parquet.encrypted] manually (e.g. with wget or curl), since this looks like a DNS resolution issue. [~LutzWeischerFujitsu]  , to isolate the problem - does adding github.com directly to /etc/hosts  make any difference? Gabor Szadovszky / @gszadovszky: [~LutzWeischerFujitsu], this is a known issue. See PARQUET-2010 for details. You may either cherry-pick my change in the related PR or step back to the previous release in the history to go forward. I hope my change will get in soon so everyone would be able to continue working on the new release. Gabor Szadovszky / @gszadovszky: [~LutzWeischerFujitsu], PARQUET-2010 is already in master so it should work now. Lutz Weischer: Please be so kind to inform when the new (master) release is available and should not have that issue any more. I will than re-try. Thanks.
gharchive/issue
2021-03-16T12:18:35
2025-04-01T04:55:59.750068
{ "authors": [ "asfimport" ], "repo": "apache/parquet-java", "url": "https://github.com/apache/parquet-java/issues/2587", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1716956130
Extensible PinotDataBuffer SPI This is a PEP-Request. I propose to create a extensible SPI that can be used to change the PinotDataBuffer implementation used at runtime. Treat this issue as a clone of https://github.com/apache/pinot/issues/9162, but more formal. What needs to be done? Right now PinotDataBuffer instances are obtained by calling static methods: PinotDataBuffer allocateDirect(long size, ByteOrder byteOrder, String description) PinotDataBuffer loadFile(File file, long offset, long size, ByteOrder byteOrder, String description) PinotDataBuffer mapFile(File file, boolean readOnly, long offset, long size, ByteOrder byteOrder, String description) These methods have a static implementation that return a PinotDataBuffer that is backed by a ByteBuffer or by a LArray, depending on whether the requested size is greater than 2GBs or not. ByteBuffers are faster and more reliable than LArray, but they cannot be larger than Integer.MAX_INT (aka 2GB - 1byte). Here I propose to change that implementation. Instead of having a static implementation, the algorithm used to instantiate the PinotDataBuffer will be delegated to an interface. The specific implementation will be set at Pinot startup time by reading the property pinot.offheap.buffer.factory, which will be an optional property whose value is a String. In case it exists, it should the qualified class name of a Class that implements PinotDataBufferFactory. There will be also another optional property called pinot.offheap.prioritize.bytebuffer which will be a boolean. If it is true, then the ByteBuffer implementation will be used when less than 2GB buffers are requested. Why the feature is needed The reason to be able to have new PinotDataBuffer implementations is that LArray is not compatible with Java >= 16, as detected in #8529. It also seems that our LArray implementation may have some bugs (see https://github.com/apache/pinot/pull/10774) and it doesn't seem that LArray will be updated (See https://github.com/xerial/larray/issues/75), so we need to move on. Initial idea/proposal Adding a SPI is something we have done several times in the last months/years. Most of the times we use one or more configuration properties to decide which instance we need to use and set that instance in a static attribute. That system works, but it is problematic when we want to change the implementation in an isolated way (for example, when doing tests). That is why I propose to extend this system with a thread local on top of that. Something like: /** * The default {@link PinotBufferFactory} used by all threads that do not define their own factory. */ private static PinotBufferFactory _defaultFactory = createDefaultFactory(); /** * A thread local variable that can be used to customize the {@link PinotBufferFactory} used on tests. This is mostly * useful in tests. */ private static final ThreadLocal<PinotBufferFactory> _FACTORY = new ThreadLocal<>(); /** * Creates the default factory depending on the JVM version */ public static PinotBufferFactory createDefaultFactory() {...} /** * Changes the default factory */ public static void setDefaultFactory(PinotBufferFactory) {...} /** * Change the {@link PinotBufferFactory} used by the current thread. * * @see #loadDefaultFactory(PinotConfiguration) */ public static void useFactory(PinotBufferFactory factory) { _FACTORY.set(factory); } /** * Returns the factory the current thread should use. */ public static PinotBufferFactory getFactory() { PinotBufferFactory pinotBufferFactory = _FACTORY.get(); if (pinotBufferFactory == null) { pinotBufferFactory = _defaultFactory; } return pinotBufferFactory; } Then the static methods PinotDataBuffer.allocateDirect (and others) should delegate on PinotDataBuffer.getFactory(). By doing that, tests can call PinotDataBuffer.useFactory() in order to use, on that thread, the buffer library they want to test. Exploratory draft https://github.com/apache/pinot/pull/10528 implements all of these and modifies tests to use this API. Actual buffer implementations This is an optional part of the PEP. What we request is to have the PinotBufferFactory SPI. This is what we would like to do with it, but it could be part of another PEP if necessary. Given that we cannot use LArray in modern versions of Java, we have three alternatives : Use other third party libraries Implement our own library on top of Unsafe Use Foreign Memory API I tried to use Chronicle Bytes in the past with partial success, see https://github.com/apache/pinot/pull/9842. Given that I don't know more third party buffer libraries that use long offsets (Netty and Agrona use ints), I explored the last two alternatives in https://github.com/apache/pinot/pull/10528. Foreign Memory API is clearly the future, as it provides exactly what we need: A high performance buffer API that is maintained (given that will be included in the JVM) and uses longs as offsets and sizes. But it has two issues: It is not included in Java 17 It is still in preview mode and it will continue like that in Java 21 (see this post where one of the author explains the reasons and the changes since Java 20). Therefore we have three options: Wait until it is stable (possible in Java 22 or 23, which won't be LTS, so maybe we need to wait 2 years until Java 25). Use it as a preview in Java 21. This would imply that we could either run with Java 11+LArray or Java 21+Foreign, but do not support Java 17. Also, in Java 21 we would need to compile and start with --enable-preview. Create our own library. Previously mentioned https://github.com/apache/pinot/pull/10528 does create two buffer implementations: One on top of Unsafe and another on top of Foreign Memory API. The latter is trivial to implement, but it is very difficult to include in the current Apache Pinot CI. The problems I found are: Some Maven plugins don't work in Java 21-ea (at least spotless doesn't work) GitHub Actions cannot be configured to use Java 21-ea (see https://github.com/actions/setup-java/issues/492#issuecomment-1551508335). In general, it is difficult to maintain in the same compilation unit code that has to be compiled with Java 11 and code that has to be compiled with Java 21 Presto still requires to use Java 8, which makes it even more difficult. In case we decide to use my draft in the actual implementation of the PEP, I would suggest to remove the Foreign Memory API implementation from the branch and optionally move it to another github repo. I've also modified BenchmarkPinotDataBuffer in order to run with the Unsafe based implementation. The modifications are in the draft. I've run the benchmark in a M1 Pro and in a Ryzen 9 3900X with Ubuntu 22.10. The exact results of the benchmark can be found here but the following chars should be good enough: Please note that LArray only runs in Java 11, so there are no data with LArray in Java 17 and 21. This one shows the cost of executing batch writes (aka call PinotDataBuffer.readFrom) and batch read (aka call PinotDataBuffer.copyTo) with a byte array of 1024 elements (units are ns/op in Linux): It is important to note that the implementation used in both LArray and Unsafe when dealing with batch reads and writes is to create a temporal direct byte buffer that points to the same address of the current buffer, so we can assume that the performance difference is due to the new instance creation. This one shows the cost of executing non batch writes (calling PinotDataBuffer.putByte()) and reads (calling PinotDataBuffer.getByte) in a loop with 1024 consecutive offsets starting from a random one (units are ns/op in Linux): In my opinion the latest is the most interesting, as it shows the improvements introduced in Java 17 and 21 JIT when dealing with loops. I've also tried to add the Foreign Memory API implementation in the benchmark, but I found several problems so I gave up. This is amazing @gortiz . I am +1 +1 from me as well. +1, awesome work! +1. Thanks @gortiz So I am guessing the next plan of action is to: Review the SPI and Unsafe based implementation in https://github.com/apache/pinot/pull/10528 Also, Foreign Memory is not ruled out right. Later on we should move to Foreign Memory API when stable support is available ? Review the SPI and Unsafe based implementation in https://github.com/apache/pinot/pull/10528 I have some issues there. Some tests always get out of memory in GitHub Actions, while my amd pc and my M1 laptop run them without problems, even when running with mvn -P github-actions. Given that I cannot reproduce the problem locally, it is being very difficult to fix the problem. Any help there would be helpful. Later on we should move to Foreign Memory when stable support is available ? I guess we should decide what we want to do with that. We can either: Wait until Foreign Memory API is production ready and stay in Java 11 meanwhile. Versions on which we could run Pinot: [11, 15] and [22, ...). Assuming that Foreign Memory API is released in Java 22. Advantage: we will only use proven and stable apis. Disadvantages: But it implies that we would need to either run with Java 11 until Java 23 or we would need to use non LTS versions. Support preview of Foreign Memory API: Versions on which we could run Pinot: [11, 15] and [21, ...). Advantage: We would support Java 21 if running with --enable-preview Disadvantage: If OpenJDK decides to add more breaking changes in Java 22, we would need to maintain different versions of the code. We would need to run with --preview-enable, which is not intended to be executed in production. Use our own library on top of Unsafe Versions on which we could run Pinot: [11, ...). Advantages: In Java [11, 15] we can choose to use LArray or Unsafe In Java [16, 20] we can only use Unsafe In Java 21 we can use Unsafe or run with --enable-preview and use Foreign Memory API Once Foreign Memory API is released, we could always use Foreign Memory API Disadvantages: We need to maintain the Unsafe library. Errors there may kill the JVM with segmentation faults. Details may depend on CPU architectures and OS. For example, mmap needs to be page aligned in Linux (otherwise segmentation fault kills the JVM). Unsafe methods do change between major JVM versions. I was able to find the reason why some tests were failing in https://github.com/apache/pinot/pull/10528. It was a Maven config problem related on how I was configuring the surefire argLine property. Now I have green lights in Java 11 and Java 20. Last execution failed in Java 17 and I'm trying to verify if it is a flaky test or something actually related to these changes. I'm thinking on removing the draft tag from the PR. We should discuss whether to include the Foreign Memory API or not in the PR. I think it is better to just drop that code and add them later once is it stable (or if we want to invest resources into trying with --preview-enabled) We may want to close this issue
gharchive/issue
2023-05-19T09:45:08
2025-04-01T04:55:59.776906
{ "authors": [ "gortiz", "kishoreg", "mayankshriv", "siddharthteotia", "xiangfu0" ], "repo": "apache/pinot", "url": "https://github.com/apache/pinot/issues/10783", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1085236740
Add instance id in response Description This PR adds instance id at the error message in the response. Currently the response is like below, which is hard to identify the target server: "errors": [ { "message": "ServerTableMissing:\nFailed to find table: testTable_OFFLINE", "type": "UNKNOWN" } ] Upgrade Notes Does this PR prevent a zero down-time upgrade? (Assume upgrade order: Controller, Broker, Server, Minion) [ ] Yes (Please label as backward-incompat, and complete the section below on Release Notes) Does this PR fix a zero-downtime upgrade introduced earlier? [ ] Yes (Please label this as backward-incompat, and complete the section below on Release Notes) Does this PR otherwise need attention when creating release notes? Things to consider: New configuration options Deprecation of configurations Signature changes to public methods/interfaces New plugins added or old plugins removed [ ] Yes (Please label this PR as release-notes and complete the section on Release Notes) Release Notes Documentation Codecov Report Merging #7937 (dfaf9e9) into master (aa2da07) will decrease coverage by 57.00%. The diff coverage is 0.00%. @@ Coverage Diff @@ ## master #7937 +/- ## ============================================= - Coverage 71.32% 14.32% -57.01% + Complexity 4092 80 -4012 ============================================= Files 1589 1548 -41 Lines 82139 80502 -1637 Branches 12270 12067 -203 ============================================= - Hits 58589 11532 -47057 - Misses 19578 68109 +48531 + Partials 3972 861 -3111 Flag Coverage Δ integration1 ? integration2 ? unittests1 ? unittests2 14.32% <0.00%> (-0.04%) :arrow_down: Flags with carried forward coverage won't be shown. Click here to find out more. Impacted Files Coverage Δ ...core/query/executor/ServerQueryExecutorV1Impl.java 0.00% <0.00%> (-83.34%) :arrow_down: ...ain/java/org/apache/pinot/core/data/table/Key.java 0.00% <0.00%> (-100.00%) :arrow_down: .../java/org/apache/pinot/spi/utils/BooleanUtils.java 0.00% <0.00%> (-100.00%) :arrow_down: .../java/org/apache/pinot/core/data/table/Record.java 0.00% <0.00%> (-100.00%) :arrow_down: .../java/org/apache/pinot/core/util/GroupByUtils.java 0.00% <0.00%> (-100.00%) :arrow_down: ...ava/org/apache/pinot/spi/config/table/FSTType.java 0.00% <0.00%> (-100.00%) :arrow_down: ...ava/org/apache/pinot/spi/data/MetricFieldSpec.java 0.00% <0.00%> (-100.00%) :arrow_down: ...va/org/apache/pinot/spi/utils/BigDecimalUtils.java 0.00% <0.00%> (-100.00%) :arrow_down: ...java/org/apache/pinot/common/tier/TierFactory.java 0.00% <0.00%> (-100.00%) :arrow_down: ...java/org/apache/pinot/common/utils/StringUtil.java 0.00% <0.00%> (-100.00%) :arrow_down: ... and 1264 more Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update aa2da07...dfaf9e9. Read the comment docs.
gharchive/pull-request
2021-12-20T21:56:36
2025-04-01T04:55:59.801210
{ "authors": [ "codecov-commenter", "jackjlli" ], "repo": "apache/pinot", "url": "https://github.com/apache/pinot/pull/7937", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1357886992
[WIP] Adaptive Server Selection to improve Query Processing Resiliency label = feature OSS issue https://github.com/apache/pinot/issues/8618 Design Doc PR Description Based on the segment assignment strategies used, we use one of the following instance selectors today: BalancedInstanceSelector ReplicaGroupInstanceSelector StrictReplicaGroupInstanceSelector Irrespective of which instance selection mechanism is used, we use a round robin approach. The round-robin approach is not sensitive to changes in the system like server slowness, GC on servers, etc. Hence we could end up overloading some servers especially in cases where they are underperforming thereby leading to higher latencies for queries. This PR provides the implementation for the Adaptive Server Selection feature at brokers. When a query is received, we could use one of the implemented Adaptive Selectors (NumInFlightRequests, Latency, Hybrid) to efficiently route queries to the best server instead of using a naive round robin approach. Test Results TBD Codecov Report Merging #9311 (d975dbf) into master (1c9528c) will decrease coverage by 41.46%. The diff coverage is 15.98%. @@ Coverage Diff @@ ## master #9311 +/- ## ============================================= - Coverage 69.73% 28.27% -41.47% + Complexity 5017 53 -4964 ============================================= Files 1873 1868 -5 Lines 99598 99613 +15 Branches 15163 15169 +6 ============================================= - Hits 69453 28163 -41290 - Misses 25217 68713 +43496 + Partials 4928 2737 -2191 Flag Coverage Δ integration1 26.05% <12.69%> (-0.09%) :arrow_down: integration2 24.81% <14.97%> (+0.02%) :arrow_up: unittests1 ? unittests2 ? Flags with carried forward coverage won't be shown. Click here to find out more. Impacted Files Coverage Δ ...e/pinot/broker/api/resources/PinotBrokerDebug.java 82.75% <0.00%> (-2.96%) :arrow_down: ...roker/requesthandler/BaseBrokerRequestHandler.java 60.47% <ø> (-9.78%) :arrow_down: ...routing/adaptiveserverselector/HybridSelector.java 0.00% <0.00%> (ø) ...outing/adaptiveserverselector/LatencySelector.java 0.00% <0.00%> (ø) ...adaptiveserverselector/NumInFlightReqSelector.java 0.00% <0.00%> (ø) ...e/pinot/common/utils/ExponentialMovingAverage.java 0.00% <0.00%> (ø) ...org/apache/pinot/core/transport/QueryResponse.java 100.00% <ø> (ø) .../server/routing/stats/ServerRoutingStatsEntry.java 0.00% <0.00%> (ø) ...va/org/apache/pinot/spi/utils/CommonConstants.java 0.00% <0.00%> (-27.70%) :arrow_down: ...erver/routing/stats/ServerRoutingStatsManager.java 6.04% <6.04%> (ø) ... and 1299 more :mega: We’re building smart automated test selection to slash your CI/CD build times. Learn more @Jackie-Jiang Thanks for the comments. I've addressed them in #9462. Please review. I'll add the following soon. Release-note in PR description. Example of configs to use this feature. Pinot docs. Merged user docs - https://docs.pinot.apache.org/operators/operating-pinot/tuning/query-routing-using-adaptive-server-selection
gharchive/pull-request
2022-08-31T21:22:05
2025-04-01T04:55:59.822776
{ "authors": [ "codecov-commenter", "vvivekiyer" ], "repo": "apache/pinot", "url": "https://github.com/apache/pinot/pull/9311", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
985368535
Exclusive subscription mode for partitioned topics not seeing consistent order among different consumers Describe the bug In the official documentation of Apache Pulsar it says that an exclusive subscription sees a consistent order for a single consumer. Suppose we have multiple consumers each of them with its own exclusive subscription and they are reading from a partitioned topic. As presented in the docs: "Decisions about routing and subscription modes can be made separately in most cases. In general, throughput concerns should guide partitioning/routing decisions while subscription decisions should be guided by application semantics. There is no difference between partitioned topics and normal topics in terms of how subscription modes work, as partitioning only determines what happens between when a message is published by a producer and processed and acknowledged by a consumer." Those statements lead us to infer that the readers would get a consistent global ordering among topic partitions. But that is not the case in my tests so far: The code I have used to test it (Scala): Producer: package pulsar import org.apache.pulsar.client.api.PulsarClient import java.util.UUID object Producer { def main(args: Array[String]): Unit = { val client = PulsarClient.builder() .serviceUrl(SERVICE_URL) .allowTlsInsecureConnection(true) .build() val producer = client.newProducer() .topic(TOPIC) .enableBatching(true) //.accessMode(ProducerAccessMode.Exclusive) .create() for(i<-0 until 100){ val key = UUID.randomUUID().toString.getBytes() //val key = s"Hello-${i}".getBytes() producer.newMessage().orderingKey("k0".getBytes()).value(key).send() println(s"produced msg: ${i.toString}") } producer.flush() producer.close() client.close() } } Consumer: import org.apache.pulsar.client.api.{PulsarClient, SubscriptionInitialPosition, SubscriptionType} object Consumer { def main(args: Array[String]): Unit = { val client = PulsarClient.builder() .serviceUrl(SERVICE_URL) .allowTlsInsecureConnection(true) .build() var l1 = Seq.empty[String] var l2 = Seq.empty[String] val c1 = client.newConsumer() .topic(TOPIC) .subscriptionType(SubscriptionType.Exclusive) .subscriptionInitialPosition(SubscriptionInitialPosition.Earliest) .subscriptionName("c1") .subscribe() // To stop the consuption I put a limit (100) - this limit is known while(l1.length < 100){ val msg = c1.receive() val str = new String(msg.getData) println(s"${Console.MAGENTA_B}$str${Console.RESET}") l1 = l1 :+ str //c1.acknowledge(msg.getMessageId) } val c2 = client.newConsumer() .topic(TOPIC) .subscriptionType(SubscriptionType.Exclusive) .subscriptionInitialPosition(SubscriptionInitialPosition.Earliest) .subscriptionName("c2") .subscribe() while(l2.length < 100){ val msg = c2.receive() val str = new String(msg.getData) println(s"${Console.GREEN_B}$str${Console.RESET}") l2 = l2 :+ str //c2.acknowledge(msg.getMessageId) } println() println(l1) println() println() println(l2) println() try { assert(l1 == l2) } finally { c1.close() c2.close() client.close() } } } Am I wrong about it or Pulsar does not support the described behavior I expect? **OBS.: I've tried every configuration for that. I've set Retention policies for the namespace as infinite both in size and number of messages. It does not work :( I also tried: SinglePartitionRouting for producer (it does not matter tho) Setting an ordering key for the messages** @lucasrpb I think pulsar ensures the orderly consumer of one partition-topic. If multiple partition-topics are consumed in subscription mode, the a consistent global order cannot be guaranteed. Just to make sure @shibd : Using the combo of SinglePartition config for the producer and ExclusiveSubscription for the consumer would not guarantee ordering? Just to make sure @shibd : Using the combo of SinglePartition config for the producer and ExclusiveSubscription for the consumer would not guarantee ordering? This can guarantee ordering. Just to make sure @shibd : Using the combo of SinglePartition config for the producer and ExclusiveSubscription for the consumer would not guarantee ordering? This can guarantee ordering. https://stackoverflow.com/questions/68988929/does-pulsar-partitioned-topics-support-global-ordering-when-consuming-from-diffe/69032885#69032885 Thanks guys for the explanations! Unfortunately with the current implementation I cannot get the behavior I want! But it will be very easy to achieve it using an approach I have in mind already ! :) Maybe in the near future I can propose a pull request on it! :) The issue had no activity for 30 days, mark with Stale label.
gharchive/issue
2021-09-01T16:42:00
2025-04-01T04:55:59.839911
{ "authors": [ "codelipenghui", "lucasrpb", "miguelemosreverte", "shibd" ], "repo": "apache/pulsar", "url": "https://github.com/apache/pulsar/issues/11883", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
953730132
Add offload ledger info for admin topics stats Motivation Currently, we don't have offload stats when getting the topic stats bin/pulsar-admin topics stats. We should add metrics in topic stats on last offloaded ledger id, last successful offloaded timestamp, and last offload failure timestamp. Modifications Add lastOffloadedLedgerId, lastOffloadSuccessTimestamp, lastOffloadFailureTimestamp for ManagedLedgerImpl and TopicStatsImpl. Thanks for your contribution. For this PR, do we need to update docs? (The PR template contains info about doc, which helps others know more about the changes. Can you provide doc-related info in this and future PR descriptions? Thanks) Thanks for your contribution. For this PR, do we need to update docs? (The PR template contains info about doc, which helps others know more about the changes. Can you provide doc-related info in this and future PR descriptions? Thanks) Yes, we need the "REST APIs => PERSISTENT TOPIC => Get the stats for the topic" api return result updated.
gharchive/pull-request
2021-07-27T10:19:05
2025-04-01T04:55:59.843579
{ "authors": [ "Anonymitaet", "frankxieke" ], "repo": "apache/pulsar", "url": "https://github.com/apache/pulsar/pull/11465", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1094986110
bundle metrics would overwrite loadbalance metrics Motivation #12366 added bundle metrics, But, If we enabled expose these metrics, It would overwrite previous load balancing metrics. Modifications seperate bundle metrics and load balance metrics Documentation Check the box below or label this PR directly (if you have committer privilege). Need to update docs? [ ] doc-required (If you need help on updating docs, create a doc issue) [x] no-need-doc (Please explain why) [ ] doc (If this PR contains doc changes) /pulsarbot run-failure-checks /pulsarbot run-failure-checks /pulsarbot run-failure-checks @hangc0276 PTAL at this. Great catch. The code looks good to me. Before we merge this, can you please update the documentation? The doc is located here: site2/docs/reference-metrics.md. @michaeljmarshall Done. @michaeljmarshall docs are done in https://github.com/apache/pulsar/pull/12825/files, so I label this PR w/ doc-added and remove doc-required. @Anonymitaet - thank you. I forgot to switch labels once the docs were added.
gharchive/pull-request
2022-01-06T05:37:18
2025-04-01T04:55:59.849098
{ "authors": [ "Anonymitaet", "gaozhangmin", "michaeljmarshall" ], "repo": "apache/pulsar", "url": "https://github.com/apache/pulsar/pull/13641", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1345483676
[fix][doc] Migrate config docs to pulsar-site Motivation Pulsar configuration docs are going to be migrated to https://pulsar.apache.org/reference (Not available currently, will be after apache/pulsar-site#169 is merged. Current version is deployed here). This PR aims to adjust the docs so that everything looks fine in the new site. Modifications Remove config docs that are auto-generated with code. Manually maintained ones are uneffected. Adjust the format of admonition. Correct the invalid links. Adjust sidebar structure. Verifying this change [x] Make sure that the change passes the CI checks. This change is a trivial rework / code cleanup without any test coverage. Does this pull request potentially affect one of the following parts: Dependencies (does it add or upgrade a dependency): no The public API: no The schema: no The default values of configurations: no The wire protocol: no The rest endpoints: no The admin cli options: no Anything that affects deployment: no Documentation Check the box below or label this PR directly. Need to update docs? [ ] doc-required (Your PR needs to update docs and you will update later) [] doc-not-needed (Please explain why) [x] doc (Your PR contains doc changes) [ ] doc-complete (Docs have been already added) @Anonymitaet PTAL /pulsarbot run-failure-checks /pulsarbot run-failure-checks /pulsarbot run-failure-checks /pulsarbot run-failure-checks
gharchive/pull-request
2022-08-21T13:09:46
2025-04-01T04:55:59.856649
{ "authors": [ "Anonymitaet", "SignorMercurio" ], "repo": "apache/pulsar", "url": "https://github.com/apache/pulsar/pull/17197", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1643431457
[fix][build] Dump Jacoco coverage data to file with JMX interface in TestNG listener Fixes #19931 Motivation See #19931. Sometimes the default Jacoco shutdown hook doesn't run and there's no coverage data. This causes the Codecov upload to fail since there's no Jacoco coverage report available. Modifications enable Jacoco agent's JMX interface Dump Jacoco coverage to file with JMX interface in TestNG listener Documentation [ ] doc [ ] doc-required [x] doc-not-needed [ ] doc-complete Matching PR in forked repository PR in forked repository: https://github.com/lhotari/pulsar/pull/145 Yes, that's the case. Are your going to rerun the workflow in your fork and notify here once it passed, or we tag this PR as ready-to-test and rerun here? @tisonkun re-running it. in progress at https://github.com/lhotari/pulsar/actions/runs/4540821034/jobs/8006055251?pr=145 . This time build step completed properly. Waiting for the whole build job to complete. /pulsarbot rerun-failure-checks Hmmm...Now we can pass the flaky tests job, but it seems the integration tests upload job is still suffering: @lhotari the error info says "not_found" and I read it as file not created. But it seems different from the previous one. Do we now always dump the file, but the upload action itself is still unstable? IIRC there is an upstream issue for the latter one. @lhotari the error info says "not_found" and I read it as file not created. But it seems different from the previous one. Do we now always dump the file, but the upload action itself is still unstable? IIRC there is an upstream issue for the latter one. Exactly. For the upload issue, it seems that Codecov suggests to not consider the upload token as a secret: Public repositories that rely on PRs via forks will find that they cannot effectively use Codecov if the token is stored as a GitHub secret. The scope of the Codecov token is only to confirm that the coverage uploaded comes from a specific repository, not to pull down source code or make any code changes. For this reason, we recommend that teams with public repositories that rely on PRs via forks consider the security ramifications of making the Codecov token available as opposed to being in a secret.
gharchive/pull-request
2023-03-28T08:12:40
2025-04-01T04:55:59.864666
{ "authors": [ "lhotari", "tisonkun" ], "repo": "apache/pulsar", "url": "https://github.com/apache/pulsar/pull/19947", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
450897224
Fixed NPE when closing batch during a reconnection Motivation There is a race condition when closing a batch and the producer is getting disconnected. This has been seen in some of the tests run and can be reproduced by running this particular test many times in a loop. 2019-05-31\T\16:32:49.647 [ERROR] testProducerReconnect(org.apache.pulsar.client.api.ClientErrorsTest) Time elapsed: 0.034 s <<< FAILURE! org.apache.pulsar.client.api.PulsarClientException: java.lang.NullPointerException at org.apache.pulsar.client.impl.ProducerImpl.batchMessageAndSend(ProducerImpl.java:1354) at org.apache.pulsar.client.impl.ProducerImpl.triggerFlush(ProducerImpl.java:1282) at org.apache.pulsar.client.impl.ProducerBase.send(ProducerBase.java:82) at org.apache.pulsar.client.impl.TypedMessageBuilderImpl.send(TypedMessageBuilderImpl.java:61) at org.apache.pulsar.client.impl.ProducerBase.send(ProducerBase.java:52) at org.apache.pulsar.client.api.ClientErrorsTest.testProducerReconnect(ClientErrorsTest.java:600) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.testng.internal.MethodInvocationHelper.invokeMethod(MethodInvocationHelper.java:124) at org.testng.internal.InvokeMethodRunnable.runOne(InvokeMethodRunnable.java:54) at org.testng.internal.InvokeMethodRunnable.run(InvokeMethodRunnable.java:44) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Caused by: java.lang.NullPointerException at org.apache.pulsar.client.impl.ProducerImpl.batchMessageAndSend(ProducerImpl.java:1329) ... 17 more The problem is that we're checking isConnected() and then using the cnx() which can return null if the producer is disconnected. Rather, we should first cache the reference to the ClientCnx instance and then double check if we're connected. The same pattern is already applied for the non-batching scenario: https://github.com/apache/pulsar/blob/ba24d73b5f77c42f832cd9d2720dc0aba388f909/pulsar-client/src/main/java/org/apache/pulsar/client/impl/ProducerImpl.java#L399-L405 run java8 tests run java8 tests run java8 tests
gharchive/pull-request
2019-05-31T17:10:45
2025-04-01T04:55:59.868190
{ "authors": [ "merlimat" ], "repo": "apache/pulsar", "url": "https://github.com/apache/pulsar/pull/4427", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
668704909
[Doc] Add a note about sink-type and source-type value of the built-in connector. Fixes #6109 Motivation Tell user the available values of "--sink-type" "--source-type" parameter of the Cassandra built-in connector. Modifications Update the following docs: Pulsar IO > get started Connector Admin CLI @tuteng , please help review the doc update. @jiazhai @Jennifer88huang , cloud you please help review the PR? Thanks.
gharchive/pull-request
2020-07-30T13:08:56
2025-04-01T04:55:59.870466
{ "authors": [ "Huanli-Meng" ], "repo": "apache/pulsar", "url": "https://github.com/apache/pulsar/pull/7699", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
674878017
make pulsar executor pool size configurable Motivation The pulsar executor pool size number is hard code to 20 when pulsar service start, it should be configurable in broker.conf. private final ScheduledExecutorService executor = Executors.newScheduledThreadPool(20, new DefaultThreadFactory("pulsar")); Changes make the executor pool size configurable in broker.conf As a followup to these changes in thread pool sizes, it would be nice to reduce the size of these pools when running the unit tests As a followup to these changes in thread pool sizes, it would be nice to reduce the size of these pools when running the unit tests Yes, for the unit tests framework, i will reduce the pool size. As a followup to these changes in thread pool sizes, it would be nice to reduce the size of these pools when running the unit tests Yes, for the unit tests framework, i will reduce the pool size. @merlimat @jiazhai I have changed the default executor pool size to Runtime.getRuntime().availableProcessors() and update unit test executor pool size to 5, please take a look, thanks. /pulsarbot run-failure-checks
gharchive/pull-request
2020-08-07T09:09:40
2025-04-01T04:55:59.873804
{ "authors": [ "hangc0276", "merlimat" ], "repo": "apache/pulsar", "url": "https://github.com/apache/pulsar/pull/7782", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1205181157
can not run in Windows when i run mvn clean package -Dmaven.test.skip=true my os alert this software can not run on your platform it seems that the version of yarn is incorretly! Windows10 maven3.2.X jdk 8 Could you provide the Error message? By the way, If yarn doesn't work on your environment, you can remove the frontend-maven-plugin from pom.xml. link thank u for your reply this is the screenshot of the error could i use local yarn and node to build this project? It seems node and yarn were installed succeed, but no permission when running yarn install, maybe we should fix the permission issue. Another work around is that you can remove the frontend-maven-plugin from pom.xml as I mentioned before.
gharchive/issue
2022-04-15T02:15:08
2025-04-01T04:55:59.880222
{ "authors": [ "798284909", "StyleTang" ], "repo": "apache/rocketmq-dashboard", "url": "https://github.com/apache/rocketmq-dashboard/issues/85", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1676766636
Build Fails on ARM MAC When building on ARM MAC required to add -Dos.arch=x86_64 to bypass protobuff build issues When this is passed the next issue is related to this https://github.com/apache/rocketmq-mqtt/issues/160 having PostConstruct deps are missing from POM.xml of related projects. Any particular reason the javax.annotation deps are missing they are not part of ANY JDK it looks like these somehow on your build/classpath in project authors set-up and the required build instructions/set-up is then missing. <dependency> <groupId>javax.annotation</groupId> <artifactId>javax.annotation-api</artifactId> <version>1.3.2</version> </dependency> Also going back to item 2 pointing to issue is this project tied to https://github.com/apache/rocketmq-mqtt/issues/160#issuecomment-1329993294 RocketMQ 4.9.4 and must be build in Docker if yes any hint in setting this up is appreciated? Thank you The Java Common Annotations Module java.xml.ws.annotation was deprecated in Java version 9 and was removed in java version 11. If this leads to a problem you could try to add javax.annotation. No these are two different deps javax.annotation is not the same as java.xml.ws.annotation my point is that your project is either missing proper build instructions in your readme. or pom.xml needs to have correct dependency added since other folks were able build this only while in they have javax.annotation deps on classpath https://github.com/apache/rocketmq-mqtt/issues/160#issuecomment-1329993294
gharchive/issue
2023-04-20T13:47:59
2025-04-01T04:55:59.885396
{ "authors": [ "meshpaul", "pingww" ], "repo": "apache/rocketmq-mqtt", "url": "https://github.com/apache/rocketmq-mqtt/issues/193", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
642101872
LitePullConsumerSubscribe testcase not work [4.6.0] BUG REPORT Please describe the issue you observed: I run the example below, but it doesn't work when the consumer group subscribe the topic the first time(I mean the consumer group have never consumed the topic). The consumer can not consume any message, include old messages and new messages. testcase file: https://github.com/apache/rocketmq/blob/rocketmq-all-4.6.0/example/src/main/java/org/apache/rocketmq/example/simple/LitePullConsumerSubscribe.java If the consumer group have consumed some messages before , (may be consumed by DefaultPushConsumer), then the testcase works. Please tell us about your environment: rocketmq 4.6.0 Can you try if the same issue still exists in the latest version ? 4.7.0 still not work
gharchive/issue
2020-06-19T16:42:41
2025-04-01T04:55:59.888425
{ "authors": [ "RongtongJin", "henrypoter" ], "repo": "apache/rocketmq", "url": "https://github.com/apache/rocketmq/issues/2110", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
351098438
为什么源代码都没加注释 至少关键部分应该加 @safph The issue tracker is only used for bug report and feature request, you can subscribe mailing list for help , community friends will answer your questions about the design or even the code. Of course, we will also improve our documentation to help everyone become familiar with RocketMQ.
gharchive/issue
2018-08-16T08:06:25
2025-04-01T04:55:59.889799
{ "authors": [ "duhengforever", "safph" ], "repo": "apache/rocketmq", "url": "https://github.com/apache/rocketmq/issues/416", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
473185502
One Swc depended on another Swc Causes Index.html Blank I am compiling my own library DBIZLib.swc Console : https://paste.apache.org/h4c5d what should i do to remove these warnings? because if i ignore these warnings and compile any project index.html is blank Errors: Uncaught ReferenceError: applications_mx_managers_SystemManager is not defined at index.html:962 Uncaught Error: Bad dependency path or symbol: org.apache.royale.effects.IEffect at visit (base.js:2796) at visit (base.js:2808) at visit (base.js:2808) at goog.DebugLoader_.load_ (base.js:2815) at Object.goog.require (base.js:846) at index.html:957 https://user-images.githubusercontent.com/42200979/61864583-1ab30c00-aeeb-11e9-9a8f-690798fec1cd.png It looks like you're missing playerglobal.swc as an external dependency. It would help to see what build configuration you are using. Recently Josh made changes to how libraries are linked. I'm not using ANT, but I think ANT solves this with config.xml files in projects. Basically instead of listing all in library-path node, you must use "external-library-path" node in libraries to ensure all the code in the libraries linked are not compiled into that library. Then in Application, you must continue using library-path, since you want all code used be compiled in the application. In this part your application maybe is missing playerglobal or js swcs as Harbs comment.
gharchive/issue
2019-07-26T05:37:23
2025-04-01T04:55:59.893548
{ "authors": [ "Harbs", "carlosrovira", "pashminakazi" ], "repo": "apache/royale-asjs", "url": "https://github.com/apache/royale-asjs/issues/437", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1695527134
Add test cases for service.manager.impl Make sure that: [√ ] You have read the contribution guidelines. [√ ] You submit test cases (unit or integration tests) that back your changes. [ √] Your local test passed ./mvnw clean install -Dmaven.javadoc.skip=true. hi, pls checked checkstyle and build in you local env. so much checkstyle error hi, pls checked checkstyle and build in you local env. so much checkstyle error thx,After this inspection, I learned a lot and found my own shortcomings. At present, I have made a modification Codecov Report Merging #4610 (e6a4129) into master (d6067f4) will increase coverage by 0.56%. The diff coverage is 62.75%. @@ Coverage Diff @@ ## master #4610 +/- ## ============================================ + Coverage 67.62% 68.18% +0.56% - Complexity 7845 7919 +74 ============================================ Files 1070 1078 +8 Lines 30650 30775 +125 Branches 2790 2807 +17 ============================================ + Hits 20726 20983 +257 + Misses 8303 8139 -164 - Partials 1621 1653 +32 Impacted Files Coverage Δ ...e/shenyu/plugin/base/fallback/FallbackHandler.java 36.84% <0.00%> (-17.01%) :arrow_down: .../apache/shenyu/plugin/hystrix/command/Command.java 7.69% <0.00%> (+2.69%) :arrow_up: .../shenyu/plugin/hystrix/command/HystrixCommand.java 55.55% <0.00%> (ø) ...plugin/hystrix/command/HystrixCommandOnThread.java 61.11% <0.00%> (ø) .../springmvc/ShenyuSpringMvcClientConfiguration.java 100.00% <ø> (ø) ...henyuSpringMvcClientInfoRegisterConfiguration.java 0.00% <0.00%> (ø) ...pringboot/starter/gateway/ShenyuConfiguration.java 88.00% <0.00%> (-3.67%) :arrow_down: ...shenyu/web/fallback/DefaultFallbackController.java 75.00% <0.00%> (-25.00%) :arrow_down: ...pache/shenyu/web/filter/CollapseSlashesFilter.java 0.00% <0.00%> (ø) ...ava/org/apache/shenyu/web/filter/HealthFilter.java 62.50% <60.00%> (-37.50%) :arrow_down: ... and 6 more ... and 7 files with indirect coverage changes :mega: We’re building smart automated test selection to slash your CI/CD build times. Learn more @lijay7674 hi, can you updata master to you dev,then summit pull request ? not merged sorry,Can I initiate a new pull request @lijay7674 you can add my wechat: xixy199195, i wiil invite you join contributor group
gharchive/pull-request
2023-05-04T08:35:57
2025-04-01T04:55:59.956304
{ "authors": [ "codecov-commenter", "lijay7674", "yu199195" ], "repo": "apache/shenyu", "url": "https://github.com/apache/shenyu/pull/4610", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
705231042
login interface Please answer these questions before submitting your issue. Why do you submit this issue? Question Question What do you want to know? Hi Mr.Wu , I was wondering if skywalking8.1.0 has a login interface . If there is , then which config file is to set up the user and password ? Thanks. Requirement or improvement I just upgrade the skw 6.1.0 to 8.1.0 and I can not find my log in interface . Neither could I set up in webapp.yml No, there isn't. The old login is not real. Use APISIX, Nginx or other proxy tech to set up a login page. roger that
gharchive/issue
2020-09-21T01:31:15
2025-04-01T04:55:59.959698
{ "authors": [ "sodared", "wu-sheng" ], "repo": "apache/skywalking", "url": "https://github.com/apache/skywalking/issues/5525", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
936986406
持久化到es后,造成es占用内存持续增加 jdk11环境,docker部署es7,使用SkyWalking8.6持久化到es后,会让es的内存逐渐增大,直到100%造成OOM,请问如何解决? 持久化到mysql8,也会使mysql8占用内存持续增长到100%造成oom Github discussion is required to be English for global users. Your description is totally wrong. We push data into a storage, then database fails, try to make the database more powerful.
gharchive/issue
2021-07-05T11:17:58
2025-04-01T04:55:59.961470
{ "authors": [ "ZaneGeng", "wu-sheng" ], "repo": "apache/skywalking", "url": "https://github.com/apache/skywalking/issues/7246", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
512309027
[CI] Run build jobs on different nodes parallelly In this patch, I've made it to run multiple jobs parallelly on different Jenkins slave nodes, thus, our build jobs can be distributed more evenly to our VM nodes, and the VM with lower resources can also be scheduled fairly. In this PR, I also rewrote the Jenkisfile with scripted Pipeline syntax, our previous syntax is DSL, it's much more human readable, but lack of flexibility in our complex scenarios, now we're using scripted syntax, based on Groovy, it provides a programmable ability, and flexibility, to simplify the Jenkinsfile, (from ~200 lines to ~100 lines) We don't need to group the jobs anymore, because the jobs are scheduled on different nodes in a smaller granularity (test case), meaning that every test case will take an executor, if there is no more available executor, the test case should be in queue automatically Please move this as a record in WIKI. We may need this someday.
gharchive/pull-request
2019-10-25T05:22:26
2025-04-01T04:55:59.963452
{ "authors": [ "kezhenxu94", "wu-sheng" ], "repo": "apache/skywalking", "url": "https://github.com/apache/skywalking/pull/3714", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
948222777
doc: Add Self Observability service discovery (k8s) [X] If this pull request closes/resolves/fixes an existing issue, replace the issue number. Closes #7259. [X] Update the CHANGES log. And for docs updates, I think it's better to loop @wingwong-knh in to review when the docs are updated in PR rather than review the whole docs after they're merged I am not sure whether @wingwong-knh is available. Wing, if you are, please take a look. And for docs updates, I think it's better to loop @wingwong-knh in to review when the docs are updated in PR rather than review the whole docs after they're merged I am not sure whether @wingwong-knh is available. Wing, if you are, please take a look. Indeed, for now I'm not sure if I can keep up with the daily updates! In the future, once all docs have been gone through once, I can start reviewing the updates in PR. For now, I'm only halfway through the backend doc (will up my pace!) For this particular doc, I will need to review from scratch. And for docs updates, I think it's better to loop @wingwong-knh in to review when the docs are updated in PR rather than review the whole docs after they're merged I am not sure whether @wingwong-knh is available. Wing, if you are, please take a look. Indeed, for now I'm not sure if I can keep up with the daily updates! In the future, once all docs have been gone through once, I can start reviewing the updates in PR. For now, I'm only halfway through the backend doc (will up my pace!) For this particular doc, I will need to review from scratch. Take your time. I know you are busy. We are going to continue, you could be back and polish the doc.
gharchive/pull-request
2021-07-20T03:53:31
2025-04-01T04:55:59.967855
{ "authors": [ "wankai123", "wingwong-knh", "wu-sheng" ], "repo": "apache/skywalking", "url": "https://github.com/apache/skywalking/pull/7339", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
130406687
[SPARK-10605][SQL] Add structs to collect_list/collect_set We currently use the Hive implementations for the collect_list/collect_set aggregate functions. This has a few major drawbacks, the use of HiveUDAF (which has quite a bit of overhead) and the lack of support for struct datatypes. This PR adds native implementation of these functions to Spark. The size of the collected list/set vary, this means we cannot use the fast, Tungsten, aggregation path to perform the aggregation, and that we fallback to the slower sort based path. Another big problem with these operators is that when the size of the collected list/set grows too large, we can start experiencing large GC pauzes and OOMEs. This PR implements these UDAFs as ImperativeAggregates and uses a custom mutable ArrayData structure in order to efficiently collect the items. The current PR allows for partial aggregates, but I am not sure yet if we should support this. Comments and feedback are appreciated. cc @yhuai Test build #50490 has started for PR 11004 at commit 8247d8e. Test build #50490 has finished for PR 11004 at commit 8247d8e. This patch fails Spark unit tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test FAILed. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/50490/ Test FAILed. Test build #50493 has started for PR 11004 at commit 326a213. Test build #50493 has finished for PR 11004 at commit 326a213. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/50493/ Test PASSed. Merged build finished. Test PASSed. closing this one for a better approach.
gharchive/pull-request
2016-02-01T16:14:37
2025-04-01T04:55:59.977163
{ "authors": [ "AmplabJenkins", "SparkQA", "hvanhovell" ], "repo": "apache/spark", "url": "https://github.com/apache/spark/pull/11004", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
144458293
[SPARK-14262] correct appStatus after master leader changed What changes were proposed in this pull request? Suppose master leadership changes, even if recovery is completed, those privious apps in RUNNING state are now in WAITING state from master's WebUI, although they are really at work. Now correct the state. How was this patch tested? (Please explain how this patch was tested. E.g. unit tests, integration tests, manual tests) Can one of the admins verify this patch? I think I have a PR #10506 which already covered this issue. Can one of the admins verify this patch?
gharchive/pull-request
2016-03-30T04:07:41
2025-04-01T04:55:59.979217
{ "authors": [ "AmplabJenkins", "GavinGavinNo1", "jerryshao" ], "repo": "apache/spark", "url": "https://github.com/apache/spark/pull/12054", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
153120354
[SPARK-11249][launcher] Throw error if app resource is not provided. Without this, the code would build an invalid spark-submit command line, and a more cryptic error would be presented to the user. Also, expose a constant that allows users to set a dummy resource in cases where they don't need an actual resource file; for backwards compatibility, that uses the same "spark-internal" resource that Spark itself uses. Tested via unit tests, run-example, spark-shell, and running the thrift server with mixed spark and hive command line arguments. Test build #57807 has started for PR 12909 at commit 0ab4a28. Test build #57807 has finished for PR 12909 at commit 0ab4a28. This patch fails from timeout after a configured wait of `250m`. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test FAILed. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/57807/ Test FAILed. retest this please Test build #57900 has started for PR 12909 at commit 0ab4a28. Test build #57900 has finished for PR 12909 at commit 0ab4a28. This patch fails from timeout after a configured wait of `250m`. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test FAILed. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/57900/ Test FAILed. Hmm, pyspark seems to be getting stuck, let me see if I broke it... Test build #58030 has started for PR 12909 at commit 46664ba. Test build #58030 has finished for PR 12909 at commit 46664ba. This patch fails Spark unit tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test FAILed. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/58030/ Test FAILed. retest this please Test build #58039 has started for PR 12909 at commit 46664ba. Test build #58039 has finished for PR 12909 at commit 46664ba. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/58039/ Test PASSed. @andrewor14 Merging to master / 2.0.
gharchive/pull-request
2016-05-04T21:33:33
2025-04-01T04:55:59.996684
{ "authors": [ "AmplabJenkins", "SparkQA", "vanzin" ], "repo": "apache/spark", "url": "https://github.com/apache/spark/pull/12909", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
176435374
SPARK-17424: Fix unsound substitution bug in ScalaReflection. What changes were proposed in this pull request? This method gets a type's primary constructor and fills in type parameters with concrete types. For example, MapPartitions[T, U] -> MapPartitions[Int, String]. This Substitution fails when the actual type args are empty because they are still unknown. Instead, when there are no resolved types to subsitute, this returns the original args with unresolved type parameters. How was this patch tested? This doesn't affect substitutions where the type args are determined. This fixes our case where the actual type args are empty and our job runs successfully. Test build #65269 has started for PR 15062 at commit 931f156. Test build #65269 has finished for PR 15062 at commit 931f156. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/65269/ Test PASSed. Could you add a regression test? @hvanhovell, I'm not sure what's causing this to build the test. I can step through with a debugger to see that the problem is caused by an empty actualTypeArgs list, but I don't know why it is that way. Maybe someone with more knowledge of Scala types can help fill in the details. Until then, I'm not sure how to construct a test case for this. I don't quite understand what use-case this patch fixes. Can you provide an example (in the form of a test) that reproduces the issue? Oh, I was too quick to comment, I see a complete description is in the JIRA. It would still be good if you could add a test though. @jodersky, as I said above, I'm not really sure how to build a test for this because I'm not too familiar with the Scala internals that are misused here. However, since I can confirm that it works in practice, I think it's reasonable to commit it anyway, before committers that know this code well can have a look to make sure this method is tested properly. Another way to go is to revert PR #10970. That is a work-around because another test was already hitting this problem and the solution was to avoid the parameter substitution. Jenkins retest this please Test build #66190 has started for PR 15062 at commit 931f156. I'm just unclear of why this is needed . According to the linked issue, some tests were failing and this fixed them. Everything just seems very vague to me. It would be very helpful if you could provide an example (it doesn't have to be a unit-test) that reproduces the error. It would greatly help me understand what's going on and also distill things down to a regression test. I think the best way to reproduce this is to revert #10970. I don't have a case that reproduces it that I can share since this was in a fairly large job. Test build #66190 has finished for PR 15062 at commit 931f156. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/66190/ Test PASSed. Hi @rdblue can you revert #10970 in this PR? Then other reviewers can test this PR locally and verify if it works. I tried to reproduce with #10970 reverted, but I didn't hit the issue in testing. I still think it's fine to move forward on this, even if it is hard to reproduce because we know the code is wrong and this fixes it. retest this please I'm fine with this change, cc @yhuai @liancheng Test build #68939 has started for PR 15062 at commit 931f156. Test build #68939 has finished for PR 15062 at commit 931f156. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/68939/ Test PASSed. any chance that is the same issue as https://issues.apache.org/jira/browse/SPARK-17109? @rdblue When you were debugging this issue, which version of scala did you use? Scala 2.10 or Scala 2.11? If you were using scala 2.10, is it possible to try scala 2.11? Thanks! We were seeing the problem when using the datasets API in our 1.6.1 build, which is based on Scala 2.10. I recently tried to reproduce this on master with 2.11 and #10970 reverted, but I didn't get a case that failed. Either way, I think the fix here makes sense: if there are no types to substitute, don't do it. This does look the same as SPARK-17109. Does this fix that issue? retest this please Test build #76863 has started for PR 15062 at commit 931f156. Test build #76863 has finished for PR 15062 at commit 931f156. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/76863/ Test PASSed. Merged build finished. Test PASSed. the change looks reasonable, merging to master/2.2/2.1/2.0!
gharchive/pull-request
2016-09-12T17:30:46
2025-04-01T04:56:00.019501
{ "authors": [ "AmplabJenkins", "SparkQA", "cloud-fan", "hvanhovell", "jodersky", "rdblue", "srowen", "yhuai" ], "repo": "apache/spark", "url": "https://github.com/apache/spark/pull/15062", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
179575985
[SPARK-17618] Guard against invalid comparisons between UnsafeRow and other formats This patch ports changes from #15185 to Spark 2.x. In that patch, a correctness bug in Spark 1.6.x which was caused by an invalid equals() comparison between an UnsafeRow and another row of a different format. Spark 2.x is not affected by that specific correctness bug but it can still reap the error-prevention benefits of that patch's changes, which modify ``UnsafeRow.equals()to throw an IllegalArgumentException if it is called with an object that is not anUnsafeRow`. Test build #65991 has started for PR 15265 at commit f4de93f. LGTM - pending jenkins. Test build #65991 has finished for PR 15265 at commit f4de93f. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/65991/ Test PASSed. Merging to master and branch-2.0. Thanks!
gharchive/pull-request
2016-09-27T18:32:59
2025-04-01T04:56:00.025315
{ "authors": [ "AmplabJenkins", "JoshRosen", "SparkQA", "hvanhovell" ], "repo": "apache/spark", "url": "https://github.com/apache/spark/pull/15265", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
200965481
[SPARK-19182][DStream] Optimize the lock in StreamingJobProgressListener to not block UI when generating Streaming jobs What changes were proposed in this pull request? When DStreamGraph is generating a job, it will hold a lock and block other APIs. Because StreamingJobProgressListener (numInactiveReceivers, streamName(streamId: Int), streamIds) needs to call DStreamGraph's methods to access some information, the UI may hang if generating a job is very slow (e.g., talking to the slow Kafka cluster to fetch metadata). It's better to optimize the locks in DStreamGraph and StreamingJobProgressListener to make the UI not block by job generation. How was this patch tested? existing ut cc @zsxwing Test build #71441 has started for PR 16601 at commit 46036bf. Test build #71441 has finished for PR 16601 at commit 46036bf. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/71441/ Test PASSed. also cc @tdas Test build #71492 has started for PR 16601 at commit eaa7b15. Test build #71492 has finished for PR 16601 at commit eaa7b15. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/71492/ Test PASSed. @zsxwing Take a review please! Test build #71553 has started for PR 16601 at commit e51623c. Test build #71553 has finished for PR 16601 at commit e51623c. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/71553/ Test PASSed. LGTM. Merging to master. Thanks!
gharchive/pull-request
2017-01-16T09:17:13
2025-04-01T04:56:00.038272
{ "authors": [ "AmplabJenkins", "SparkQA", "uncleGen", "zsxwing" ], "repo": "apache/spark", "url": "https://github.com/apache/spark/pull/16601", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
201042342
[SPARK-19246][SQL]CataLogTable's partitionSchema order and exist check What changes were proposed in this pull request? CataLogTable's partitionSchema should check if each column name in partitionColumnNames must match one and only one field in schema, if not we should throw an exception and CataLogTable's partitionSchema should keep order with partitionColumnNames How was this patch tested? Test build #71451 has started for PR 16606 at commit eaf18ce. Test build #71451 has finished for PR 16606 at commit eaf18ce. This patch fails Scala style tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test FAILed. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/71451/ Test FAILed. Test build #71453 has started for PR 16606 at commit 9296624. Test build #71453 has finished for PR 16606 at commit 9296624. This patch fails Spark unit tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test FAILed. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/71453/ Test FAILed. Test build #71483 has started for PR 16606 at commit 4260f84. Merged build finished. Test FAILed. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/71483/ Test FAILed. retest this please Test build #71494 has started for PR 16606 at commit 4260f84. Test build #71497 has started for PR 16606 at commit 6f2816e. Test build #71494 has finished for PR 16606 at commit 4260f84. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/71494/ Test PASSed. Test build #71497 has finished for PR 16606 at commit 6f2816e. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/71497/ Test PASSed. Test build #71519 has started for PR 16606 at commit c08e1c9. Test build #71519 has finished for PR 16606 at commit c08e1c9. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/71519/ Test PASSed. Test build #71552 has started for PR 16606 at commit 8cbee32. Test build #71552 has finished for PR 16606 at commit 8cbee32. This patch fails Scala style tests. This patch merges cleanly. This patch adds no public classes. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/71552/ Test FAILed. Merged build finished. Test FAILed. Test build #71554 has started for PR 16606 at commit 79e2e3f. Test build #71554 has finished for PR 16606 at commit 79e2e3f. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/71554/ Test PASSed. Test build #71580 has started for PR 16606 at commit 5e60f14. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/71580/ Test FAILed. Merged build finished. Test FAILed. Test build #71853 has started for PR 16606 at commit 206b232. how about we just add an assert? assert(schema.takeRight(partitionColumnNames.length) == partitionColumnNames) Test build #71853 has finished for PR 16606 at commit 206b232. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/71853/ Test PASSed. Merged build finished. Test PASSed. Test build #71889 has started for PR 16606 at commit 7e30cc7. Test build #71889 has finished for PR 16606 at commit 7e30cc7. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/71889/ Test PASSed. Test build #71902 has started for PR 16606 at commit 04d3940. LGTM pending test Test build #71902 has finished for PR 16606 at commit 04d3940. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/71902/ Test PASSed. Test build #71912 has started for PR 16606 at commit 72164eb. LGTM, pending tests Merged build finished. Test FAILed. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/71912/ Test FAILed. retest this please Test build #71922 has started for PR 16606 at commit 72164eb. Test build #71922 has finished for PR 16606 at commit 72164eb. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/71922/ Test PASSed. thanks, merging to master!
gharchive/pull-request
2017-01-16T15:05:15
2025-04-01T04:56:00.082541
{ "authors": [ "AmplabJenkins", "SparkQA", "cloud-fan", "gatorsmile", "windpiger" ], "repo": "apache/spark", "url": "https://github.com/apache/spark/pull/16606", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
225391068
[SPARK-20543][SPARKR][WIP] skip tests when running on CRAN What changes were proposed in this pull request? General rule on skip or not: skip if RDD tests tests could run long or complicated (streaming, hivecontext) tests on error conditions tests won't likely change/break How was this patch tested? unit tests, R CMD check --as-cran, R CMD check Test build #76342 has started for PR 17817 at commit fc4965a. Test build #76342 has finished for PR 17817 at commit fc4965a. This patch fails R style tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test FAILed. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/76342/ Test FAILed. Test build #76343 has started for PR 17817 at commit 2c5e267. Test build #76343 has finished for PR 17817 at commit 2c5e267. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/76343/ Test PASSed. Merged build finished. Test PASSed. Test build #76346 has started for PR 17817 at commit bd3295a. Test build #76346 has finished for PR 17817 at commit bd3295a. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/76346/ Test PASSed. Skipping 153 tests ON CRAN Jenkins, all tests: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/76346/console CRAN: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/76343/console merged to master/2.2
gharchive/pull-request
2017-05-01T06:06:41
2025-04-01T04:56:00.096137
{ "authors": [ "AmplabJenkins", "SparkQA", "felixcheung" ], "repo": "apache/spark", "url": "https://github.com/apache/spark/pull/17817", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
243562870
[SPARK-21444] Be more defensive when removing broadcasts in MapOutputTracker What changes were proposed in this pull request? In SPARK-21444, @sitalkedia reported an issue where the Broadcast.destroy() call in MapOutputTracker's ShuffleStatus.invalidateSerializedMapOutputStatusCache() was failing with an IOException, causing the DAGScheduler to crash and bring down the entire driver. This is a bug introduced by #17955. In the old code, we removed a broadcast variable by calling BroadcastManager.unbroadcast with blocking=false, but the new code simply calls Broadcast.destroy() which is capable of failing with an IOException in case certain blocking RPCs time out. The fix implemented here is to replace this with a call to destroy(blocking = false) and to wrap the entire operation in Utils.tryLogNonFatalError. How was this patch tested? I haven't written regression tests for this because it's really hard to inject mocks to simulate RPC failures here. Instead, this class of issue is probably best uncovered with more generalized error injection / network unreliability / fuzz testing tools. Test build #79691 has started for PR 18662 at commit a5ebcac. cc @jiangxb1987 @JoshRosen - Thanks for the lightning fast response. The change looks reasonable to me. LGTM Test build #79691 has finished for PR 18662 at commit a5ebcac. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/79691/ Test PASSed. Merged build finished. Test PASSed. Merged to master. Thanks for the quick reviews.
gharchive/pull-request
2017-07-18T00:05:29
2025-04-01T04:56:00.102827
{ "authors": [ "AmplabJenkins", "JoshRosen", "SparkQA", "cloud-fan", "sameeragarwal", "sitalkedia" ], "repo": "apache/spark", "url": "https://github.com/apache/spark/pull/18662", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
259698011
[WIP][SPARK-22096][ML] use aggregateByKeyLocally in feature frequency calc… What changes were proposed in this pull request? NaiveBayes currently takes aggreateByKey followed by a collect to calculate frequency for each feature/label. We can implement a new function 'aggregateByKeyLocally' in RDD that merges locally on each mapper before sending results to a reducer to save one stage. We tested on NaiveBayes and see ~16% performance gain on training with these changes. Signed-off-by: Vincent Xie vincent.xie@intel.com How was this patch tested? existing test Test build #82059 has started for PR 19318 at commit efb0fe9. Test build #82059 has finished for PR 19318 at commit efb0fe9. This patch fails to build. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test FAILed. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/82059/ Test FAILed. The aggregateByKeyLocally haven't been implemented. Please wait until #19317 finished or merge that PR contents into this, otherwise this PR cannot use. We should not open a PR which cannot build. @VinceShieh can you please mark this PR's title as "[WIP]"? thanks :) Build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/testing-k8s-prb-make-spark-distribution-unified/4201/ Test PASSed. Test build #97740 has started for PR 19318 at commit efb0fe9. Build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/testing-k8s-prb-make-spark-distribution-unified/4223/ Test PASSed. Test build #97740 has finished for PR 19318 at commit efb0fe9. This patch fails due to an unknown error code, -9. This patch does not merge cleanly. This patch adds no public classes. Build finished. Test FAILed. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/97740/ Test FAILed. Build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/testing-k8s-prb-make-spark-distribution-unified/4239/ Test PASSed. Test build #97752 has started for PR 19318 at commit efb0fe9. Test build #97752 has finished for PR 19318 at commit efb0fe9. This patch fails to build. This patch does not merge cleanly. This patch adds no public classes. Build finished. Test FAILed. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/97752/ Test FAILed. Test build #97768 has started for PR 19318 at commit efb0fe9. Test build #97768 has finished for PR 19318 at commit efb0fe9. This patch fails to build. This patch does not merge cleanly. This patch adds no public classes. Build finished. Test FAILed. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/97768/ Test FAILed. Test build #97818 has started for PR 19318 at commit efb0fe9. Build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/testing-k8s-prb-make-spark-distribution-unified/4273/ Test PASSed. Test build #97818 has finished for PR 19318 at commit efb0fe9. This patch fails build dependency tests. This patch does not merge cleanly. This patch adds no public classes. Build finished. Test FAILed. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/97818/ Test FAILed. @VinceShieh if you'd rebase this, we can consider it now that aggregateByKeyLocally is implemented OK. Thanks. @srowen it seems 'aggregateByKeyLocally' is not implemented Oops yeah https://github.com/apache/spark/pull/19317 was not merged. I think this can't continue.
gharchive/pull-request
2017-09-22T04:11:53
2025-04-01T04:56:00.128724
{ "authors": [ "AmplabJenkins", "SparkQA", "VinceShieh", "WeichenXu123", "chenghao-intel", "srowen" ], "repo": "apache/spark", "url": "https://github.com/apache/spark/pull/19318", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
349582280
[SPARK-25088][CORE][MESOS][DOCS] Update Rest Server docs & defaults. What changes were proposed in this pull request? (a) disabled rest submission server by default in standalone mode (b) fails the standalone master if rest server enabled & authentication secret set (c) fails the mesos cluster dispatcher if authentication secret set (d) doc updates (e) when submitting a standalone app, only try the rest submission first if spark.master.rest.enabled=true otherwise you'd see a 10 second pause like 18/08/09 08:13:22 INFO RestSubmissionClient: Submitting a request to launch an application in spark://... 18/08/09 08:13:33 WARN RestSubmissionClient: Unable to connect to server spark://... I also made sure the mesos cluster dispatcher failed with the secret enabled, though I had to do that on slightly different code as I don't have mesos native libs around. How was this patch tested? I ran the tests in the mesos module & in core for org.apache.spark.deploy.* I ran a test on a cluster with standalone master to make sure I could still start with the right configs, and would fail the right way too. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/testing-k8s-prb-make-spark-distribution-unified/2049/ Test PASSed. Test build #94572 has started for PR 22071 at commit b4ca224. Test build #94572 has finished for PR 22071 at commit b4ca224. This patch fails Spark unit tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test FAILed. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/94572/ Test FAILed. Test build #4241 has started for PR 22071 at commit b4ca224. @tnachen Test build #4244 has started for PR 22071 at commit b4ca224. Test build #4244 has finished for PR 22071 at commit b4ca224. This patch fails Spark unit tests. This patch does not merge cleanly. This patch adds no public classes. Test build #4245 has started for PR 22071 at commit b4ca224. Test build #4245 has finished for PR 22071 at commit b4ca224. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Test build #94706 has started for PR 22071 at commit 897b587. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/testing-k8s-prb-make-spark-distribution-unified/2149/ Test PASSed. Test build #94706 has finished for PR 22071 at commit 897b587. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/94706/ Test PASSed. LGTM as well +1 Merged to master any objections about putting this in prior branches as well? in this case maybe ok. perhaps just rel note this iff there's another 2.2.x or 2.1.x releases? Hi @squito, if I may, I have a few comments to share regarding this change and probably you can help me to understand it in better details. From Spark Security Doc page it looks like spark.master.rest.enabled property belongs to Spark Standalone Master only. IMO, It makes sense, because it's the server side who is responsible for enabling and disabling REST endpoint according to the source code. Now, from the [comment in the code] (https://github.com/apache/spark/blob/v2.4.0/core/src/main/scala/org/apache/spark/deploy/SparkSubmit.scala#L176-L180) of SparkSubmit.scala it looks like REST-based gateway introduced in Spark 1.3 is the default way of communication with Spark Standalone Master with a fallback to legacy RPC. And here's the problem: now spark.master.rest.enabled became a client-side property, not Spark Standalone Master property. So in order to communicate with Master, a user have to provide spark.master.rest.enabled with each job she submits because without it, the submission will fail. This is a pretty breaking change. What is the reason to fail the start of MesosClusterDispatcher if spark.authenticate.secret is set? Mesos uses different mechanisms for authentication and there's no need to fail the Dispatcher - this property can be just ignored or logged at most. TL;DR It looks like spark.master.rest.enabled should reside only in Spark Standalone Master code and shouldn't leak into the client side of SparkSubmit. Also, it can be completely ignored with Mesos Dispatcher, because REST interface is the only endpoint it exposes. Please let me know what you think. Thanks.
gharchive/pull-request
2018-08-10T16:35:57
2025-04-01T04:56:00.151144
{ "authors": [ "AmplabJenkins", "SparkQA", "akirillov", "felixcheung", "squito", "srowen", "tgravescs", "tnachen" ], "repo": "apache/spark", "url": "https://github.com/apache/spark/pull/22071", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
551525835
[SPARK-22590][SQL] Copy sparkContext.localproperties to child thread in BroadcastExchangeExec.executionContext What changes were proposed in this pull request? In org.apache.spark.sql.execution.exchange.BroadcastExchangeExec#relationFuture make a copy of org.apache.spark.SparkContext#localProperties and pass it to the broadcast execution thread in org.apache.spark.sql.execution.exchange.BroadcastExchangeExec#executionContext Why are the changes needed? When executing BroadcastExchangeExec, the relationFuture is evaluated via a separate thread. The threads inherit the localProperties from sparkContext as they are the child threads. These threads are created in the executionContext (thread pools). Each Thread pool has a default keepAliveSeconds of 60 seconds for idle threads. Scenarios where the thread pool has threads which are idle and reused for a subsequent new query, the thread local properties will not be inherited from spark context (thread properties are inherited only on thread creation) hence end up having old or no properties set. This will cause taskset properties to be missing when properties are transferred by child thread via sparkContext.runJob/submitJob Does this PR introduce any user-facing change? No How was this patch tested? Added UT @ajithme are you still working on it? I'd like to get this in before 3.0 release and take over this one if you don't have time to complete it. @ajithme are you still working on it? I'd like to get this in before 3.0 release and take over this one if you don't have time to complete it. Yes. I will update the PR shortly. Sorry for delay Test build #118581 has started for PR 27266 at commit 0f39043. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/23335/ Test PASSed. @cloud-fan @hvanhovell @HyukjinKwon Updated as per all the comments, please review Test build #118582 has started for PR 27266 at commit 742d322. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/23336/ Test PASSed. Test build #118583 has started for PR 27266 at commit 11fffca. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/23338/ Test PASSed. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/23339/ Test PASSed. Test build #118585 has started for PR 27266 at commit 763d1bc. Test build #118586 has started for PR 27266 at commit 2ed76c3. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/23340/ Test PASSed. Test build #118583 has finished for PR 27266 at commit 11fffca. This patch fails Spark unit tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test FAILed. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/118583/ Test FAILed. Test build #118581 has finished for PR 27266 at commit 0f39043. This patch fails Spark unit tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test FAILed. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/118581/ Test FAILed. Test build #118585 has finished for PR 27266 at commit 763d1bc. This patch fails Spark unit tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test FAILed. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/118585/ Test FAILed. Test build #118586 has finished for PR 27266 at commit 2ed76c3. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/118586/ Test PASSed. thanks, merging to master/3.0! Test build #118582 has finished for PR 27266 at commit 742d322. This patch fails from timeout after a configured wait of 400m. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test FAILed. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/118582/ Test FAILed. +1 from me too. Thanks @ajithme.
gharchive/pull-request
2020-01-17T16:55:25
2025-04-01T04:56:00.177998
{ "authors": [ "AmplabJenkins", "HyukjinKwon", "SparkQA", "ajithme", "cloud-fan" ], "repo": "apache/spark", "url": "https://github.com/apache/spark/pull/27266", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
610022891
[SPARK-24266][k8s] Restart the watcher when we receive a version changed from k8s What changes were proposed in this pull request? Restart the watcher when it failed with a HTTP_GONE code from the kubernetes api. Which means a resource version has changed. For more relevant information see here: https://github.com/fabric8io/kubernetes-client/issues/1075 Why are the changes needed? Does this PR introduce any user-facing change? No How was this patch tested? Running spark-submit to a k8s cluster. Not sure how to make an automated test for this. If someone can help me out that would be great. Can one of the admins verify this patch? Can one of the admins verify this patch? ok to test Thank you for your contribution, @stijndehaes . Test build #122141 has started for PR 28423 at commit a9ce548. Test build #122141 has finished for PR 28423 at commit a9ce548. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/122141/ Test PASSed. Kubernetes integration test starting URL: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/26813/ Kubernetes integration test status success URL: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/26813/ Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/26813/ Test PASSed. How do we feel about backporting this to Spark 2.4.6? How do we feel about backporting this to Spark 2.4.6? I would very much like that, we ran into this using spark 2.4.x. Do you think we can have a unit test case for this, @stijndehaes ? Test build #122248 has started for PR 28423 at commit 2a4cbb6. Test build #122248 has finished for PR 28423 at commit 2a4cbb6. This patch fails Scala style tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test FAILed. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/122248/ Test FAILed. Test build #122249 has started for PR 28423 at commit f832acf. Test build #122249 has finished for PR 28423 at commit f832acf. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/122249/ Test PASSed. Kubernetes integration test starting URL: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/26919/ Kubernetes integration test starting URL: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/26920/ Kubernetes integration test status success URL: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/26919/ Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/26919/ Test PASSed. Kubernetes integration test status success URL: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/26920/ Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/26920/ Test PASSed. Do you think we can have a unit test case for this, @stijndehaes ? The current tests completely mock out this behavior, see org.apache.spark.deploy.k8s.submit.ClientSuite, I think writing a test for this where we manipulate and fake the HTTP gone with a mock is not that useful. Maybe I can look into an integration test, but then I have to be able to trigger that a resource version changes. Not sure if I will be able to @holdenk Maybe we should refactor this behavior using the sharedinformers. See the comment made here: https://github.com/fabric8io/kubernetes-client/issues/1075#issuecomment-623354623 I can make an example implementation of this, maybe best to do that in another PR. What do you think? Ok I have tested this in production, there is something wrong with the code, went ahead and tried the sharedinformers approach. Will try that in production today. You can see the code here: https://github.com/stijndehaes/spark/tree/test/shared-informers Ok reverting back to the old approach found the missing piece I think testing that out. Shared informers have the problem that you have to watch every pod in the whole cluster atm. Test build #122321 has started for PR 28423 at commit dad7ea2. Test build #122321 has finished for PR 28423 at commit dad7ea2. This patch fails Spark unit tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test FAILed. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/122321/ Test FAILed. Test build #122322 has started for PR 28423 at commit f05db8f. Test build #122322 has finished for PR 28423 at commit f05db8f. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/122322/ Test PASSed. Kubernetes integration test starting URL: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/26991/ Kubernetes integration test status failure URL: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/26991/ Merged build finished. Test FAILed. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/26991/ Test FAILed. Kubernetes integration test starting URL: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/26992/ Kubernetes integration test status success URL: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/26992/ Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/26992/ Test PASSed. @holdenk @dongjoon-hyun I have tested this code in production and it works. I have a couple of jobs that take roughly 4 hours to finish, these all failed without the fix and are now succeeding. Could you take the time to review the code again? +1 for this. Hit this in GKE today. Retest this please. Hi, @stijndehaes . Could you describe how to manually verify this PR? Test build #124237 has started for PR 28423 at commit f05db8f. Merged build finished. Test FAILed. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/28856/ Test FAILed. Test build #124237 has finished for PR 28423 at commit f05db8f. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/124237/ Test PASSed. Test build #124283 has started for PR 28423 at commit 6ef9e12. Test build #124283 has finished for PR 28423 at commit 6ef9e12. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/124283/ Test PASSed. Kubernetes integration test starting URL: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/28902/ Kubernetes integration test status success URL: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/28902/ Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/28902/ Test PASSed. Test build #124323 has started for PR 28423 at commit 581b6f9. Test build #124323 has finished for PR 28423 at commit 581b6f9. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/124323/ Test PASSed. Kubernetes integration test starting URL: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/28942/ Kubernetes integration test status failure URL: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/28942/ Merged build finished. Test FAILed. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/28942/ Test FAILed. Test build #124636 has started for PR 28423 at commit 9da57c9. Test build #124636 has finished for PR 28423 at commit 9da57c9. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test FAILed. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/29254/ Test FAILed. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/124636/ Test PASSed. Retest this please Test build #125529 has started for PR 28423 at commit 9da57c9. Test build #125529 has finished for PR 28423 at commit 9da57c9. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/125529/ Test PASSed. Merged build finished. Test FAILed. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/30144/ Test FAILed. Hi Folks @dongjoon-hyun and @holdenk, do we plan to include this patch in 2.4.7 release? @ScrapCodes the code in the 2.4.x is significantly different from the code here. But we can reuse the same idea as here. I guess it has to be a new PR BTW, when do we receive a version changed from K8s? Test build #125828 has started for PR 28423 at commit 51dc743. @ScrapCodes the code in the 2.4.x is significantly different from the code here. But we can reuse the same idea as here. I guess it has to be a new PR Currently the JIRA has target version set as "2.4.7", so I wanted to clarify. BTW, when do we receive a version changed from K8s? It happens when etcd compaction kicks in for example. On aws EKS I never saw this happening on EKS 1.14, but it happens roughly every hour on EKS 1.15. Meaning that longer spark application continued working but the watch failed. Test build #125828 has finished for PR 28423 at commit 51dc743. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/125828/ Test PASSed. Merged build finished. Test FAILed. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/30439/ Test FAILed. BTW, when do we receive a version changed from K8s? It happens when etcd compaction kicks in for example. On aws EKS I never saw this happening on EKS 1.14, but it happens roughly every hour on EKS 1.15. Meaning that longer spark application continued working but the watch failed. Okay makes sense, thank you for explaining. I am wondering how useful is the watch for longer running jobs, for example all it reports is that jobs is running, completed or errored. Probably, spark ui is more informative. BTW, when do we receive a version changed from K8s? It happens when etcd compaction kicks in for example. On aws EKS I never saw this happening on EKS 1.14, but it happens roughly every hour on EKS 1.15. Meaning that longer spark application continued working but the watch failed. Okay makes sense, thank you for explaining. I am wondering how useful is the watch for longer running jobs, for example all it reports is that jobs is running, completed or errored. Probably, spark ui is more informative. Our use case is that airflow does a spark submit. And thus airflow follows it up that way :) Will there be a backport of this to branch-2.4? test this please (btw i'm testing the freshly upgraded R installation on the minikube workers) Test build #126271 has started for PR 28423 at commit 51dc743. Test build #126271 has finished for PR 28423 at commit 51dc743. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/126271/ Test PASSed. Kubernetes integration test starting URL: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/30887/ Kubernetes integration test status success URL: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/30887/ Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/30887/ Test PASSed. Lets focus on 3.1 and then explore backporting after. LGTM pending Jenkins Oh wait it has passed Jenkins, excellent. If @dongjoon-hyun is ok with this PR I'll merge it by the end of the week. @holdenk . Thank you for pinging me. Feel free to merge if you think it's okay. I don't want to be a blocker for the community PR. ;) Whats your JIRA username @stijndehaes ? @holdenk my JIRA username if sdehaes I took the commits from master and made a partial attempt to rebase this onto branch-2.4 [1]. However, the k8s api has evolved from 2.4 quite a bit so the watchOrStop function needs to be backported [2]. You can see the error message in this gitlab build [3]. Would it be useful to make a WIP pull request from [1] ? [1] https://github.com/jkleckner/spark/tree/SPARK-24266-on-branch2.4 [2] https://github.com/jkleckner/spark/blob/SPARK-24266-on-branch2.4/resource-managers/kubernetes/core/src/main/scala/org/apache/spark/deploy/k8s/submit/LoggingPodStatusWatcher.scala#L193 [3] https://gitlab.com/jkleckner/spark/-/jobs/651515950 @stijndehaes In private discussions about the hang we are seeing, there appears to be another watcher [1] for the driver watching executors that also may lose notifications. Have you run into any situations like this? [1] https://github.com/apache/spark/blob/master/resource-managers/kubernetes/core/src/main/scala/org/apache/spark/scheduler/cluster/k8s/ExecutorPodsWatchSnapshotSource.scala @jkleckner I have never had a problem with the driver watching the executors. I think there was already a fallback mechanism there, but I never looked into the code for that one. @liyinan926 Do you think there is an adequate existing fallback mechanism or do you still believe that there is a need to create a similar patch for ExecutorPodsWatchSnapshotSource ? I see this error a lot in the batch jobs: {"level":"WARN","timestamp":"2020-08-10 19:17:35,985","thread":"OkHttp https://kubernetes.default.svc/...","source":"io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager", "line":"209","message":"Exec Failure"} java.io.EOFException at okio.RealBufferedSource.require(RealBufferedSource.java:61) at okio.RealBufferedSource.readByte(RealBufferedSource.java:74) at okhttp3.internal.ws.WebSocketReader.readHeader(WebSocketReader.java:117) at okhttp3.internal.ws.WebSocketReader.processNextFrame(WebSocketReader.java:101) at okhttp3.internal.ws.RealWebSocket.loopReader(RealWebSocket.java:274) at okhttp3.internal.ws.RealWebSocket$2.onResponse(RealWebSocket.java:214) at okhttp3.RealCall$AsyncCall.execute(RealCall.java:206) at okhttp3.internal.NamedRunnable.run(NamedRunnable.java:32) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) ^C I do think its related to the above issue. The batch job starts, Driver is able to spin up new executors, communicate with them and get the job done, but cannot clean them up. This is with Spark 2.4.5 and Kubernetes Version: 1.15 and 1.16 with Multi Kubernetes Masters. The above message repeats every 10 seconds. Let me know if its not related It looks a bit different from what I see. For me, it appears to get stuck at the very end of writing data to Bigtable in the very last task of a job. Our partner is working to back port the fix I mentioned and I will let you know if that addresses the hang.
gharchive/pull-request
2020-04-30T13:54:37
2025-04-01T04:56:00.262979
{ "authors": [ "AmplabJenkins", "ScrapCodes", "SparkQA", "dongjoon-hyun", "holdenk", "jkleckner", "puneetloya", "shaneknapp", "stijndehaes" ], "repo": "apache/spark", "url": "https://github.com/apache/spark/pull/28423", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
663863005
[do not review][testing master PR build][discard] comment change What changes were proposed in this pull request? Why are the changes needed? Does this PR introduce any user-facing change? How was this patch tested? Test build #126344 has started for PR 29190 at commit 4b88a16. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/30960/ Test PASSed. Test build #126344 has finished for PR 29190 at commit 4b88a16. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/126344/ Test PASSed. Test build #126360 has started for PR 29190 at commit 13951c8. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/30976/ Test PASSed. Test build #126360 has finished for PR 29190 at commit 13951c8. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/126360/ Test PASSed. Merged build finished. Test PASSed. Test build #126387 has started for PR 29190 at commit fa41817. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/31002/ Test PASSed. Test build #126387 has finished for PR 29190 at commit fa41817. This patch fails due to an unknown error code, -9. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test FAILed. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/126387/ Test FAILed. Test build #126455 has started for PR 29190 at commit 54f8574. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/31069/ Test PASSed. Test build #126455 has finished for PR 29190 at commit 54f8574. This patch fails Spark unit tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test FAILed. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/126455/ Test FAILed. Test build #126508 has started for PR 29190 at commit 2f6264d. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/31127/ Test PASSed. Test build #126508 has finished for PR 29190 at commit 2f6264d. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/126508/ Test PASSed. Superceded by https://github.com/apache/spark/pull/29226
gharchive/pull-request
2020-07-22T15:44:52
2025-04-01T04:56:00.288882
{ "authors": [ "AmplabJenkins", "SparkQA", "agrawaldevesh" ], "repo": "apache/spark", "url": "https://github.com/apache/spark/pull/29190", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
771097732
[SPARK-33841][CORE] Fix issue with jobs disappearing intermittently from the SHS under high load What changes were proposed in this pull request? Mark SHS event log entries that were processing at the beginning of the checkForLogs run as not stale and check for this mark before deleting an event log. This fixes the issue when a particular job was displayed in the SHS and disappeared after some time, but then, in several minutes showed up again. Why are the changes needed? The issue is caused by SPARK-29043, which is designated to improve the concurrent performance of the History Server. The change breaks the "app deletion" logic because of missing proper synchronization for processing event log entries. Since SHS now filters out all processing event log entries, such entries do not have a chance to be updated with the new lastProcessed time and thus any entity that completes processing right after filtering and before the check for stale entities will be identified as stale and will be deleted from the UI until the next checkForLogs run. This is because updated lastProcessed time is used as criteria, and event log entries that missed to be updated with a new time, will match that criteria. The issue can be reproduced by generating a big number of event logs and uploading them to the SHS event log directory on S3. Essentially, around 236(26.7 MB) copies of an event log directory were created using shs-monitor script. Strange behavior of SHS counting the total number of applications was noticed - at first, the number was increasing as expected, but with the next page refresh, the total number of applications decreased. No errors were logged by SHS. 58 entities are displayed at 17:35:35: 25 entities are displayed at 17:36:40: Does this PR introduce any user-facing change? Yes, SHS users won't face the behavior when the number of displayed applications decreases periodically. How was this patch tested? Tested using shs-monitor script: Build SHS with the proposed change Download Hadoop AWS and AWS Java SDK Prepare S3 bucket and user for programmatic access, grant required roles to the user. Get access key and secret key Configure SHS to read event logs from S3 Start monitor script to query SHS API Run 5 producers for ~5 mins, create 125(14.2 MB) event log directory copies Wait for SHS to load all the applications Verify that the number of loaded applications increases continuously over time For more details, please refer to the shs-monitor repository. This version of the reproduction uses event log directories instead of single files, since recent optimization SPARK-33790 makes it hard to reproduce the issue with single event log files. Can one of the admins verify this patch? cc @HeartSaVioR The PR has been rebased. Thank you, @dongjoon-hyun! add to whitelist retest this, please I don't think this PR has an issue, but we need to ping @tgravescs to get his approval technically. Ping, @tgravescs . +1 and I'd consider @tgravescs approved this PR as only target branch is different. Test build #133038 has started for PR 30845 at commit e9f13a5. Kubernetes integration test starting URL: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/37637/ Kubernetes integration test status failure URL: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/37637/ Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/37637/ Thank you for your first contribution, @vladhlinsky . I added you to the Apache Spark contributor group and assigned SPARK-33841 to you. Welcome to the Apache Spark community. Thank you, @dongjoon-hyun! Test build #133038 has finished for PR 30845 at commit e9f13a5. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/133038/
gharchive/pull-request
2020-12-18T18:50:18
2025-04-01T04:56:00.307733
{ "authors": [ "AmplabJenkins", "HeartSaVioR", "SparkQA", "dongjoon-hyun", "vladhlinsky" ], "repo": "apache/spark", "url": "https://github.com/apache/spark/pull/30845", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1088316703
[MINOR][DOCS] Update pandas_pyspark.rst What changes were proposed in this pull request? 'to' --> 'the' Why are the changes needed? Grammar fix Does this PR introduce any user-facing change? Documentation fix How was this patch tested? No test needed, same as yesterday's PR Can one of the admins verify this patch? @kamelCased while we're here, can we find all these instances, and fix them in one PR? @HyukjinKwon I checked it's the last one in this file. I will check the rest of the PySpark docs. The test fails because your bunch is not synced to the latest matster branch. do you mind rebasing to the latest master branch? Also please update PR description to reflect the changes. ok to test @HyukjinKwon fetched upstream and PR description updated @HyukjinKwon fetched upstream and PR description updated @HyukjinKwon fetched upstream and PR description updated Merged to master
gharchive/pull-request
2021-12-24T10:59:40
2025-04-01T04:56:00.313014
{ "authors": [ "AmplabJenkins", "HyukjinKwon", "kamelCased", "srowen" ], "repo": "apache/spark", "url": "https://github.com/apache/spark/pull/35016", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1243995818
[SPARK-39249][SQL] Improve subexpression elimination for conditional expressions What changes were proposed in this pull request? Currently we can do subexpression elimination for conditional expressions when the subexpression is common across all branchGroups. In fact, we can farther improve this when there are common expressions between alwaysEvaluatedInputs and branchGroups. Why are the changes needed? Take the following case as an example IF(IsNull(a), b, KnowNotNull(a)) a may miss subexpression elimination chances since it is not the common expression between all branchGroups, but it's safe to evaluate a as common subexpression and eagerly execute it since it's part of the prediction, which will always be executed. If a is a time-expensive expression, we may waste time on running it. This kind of expressions are common when we do sum on decimal type because of https://github.com/apache/spark/pull/29026 https://github.com/apache/spark/blob/291d155b3c514f8b590a6b078f7efd42a30e67f0/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/aggregate/Sum.scala#L125 Performance result on TPC-DS 10T Query With this PR Without this PR Speed up 4 310.862 635.299 104.37% 80 36.723 46.006 25.28% Does this PR introduce any user-facing change? No How was this patch tested? add more UT. @viirya @cloud-fan Could you please help review this? Can one of the admins verify this patch? FYI I created https://github.com/apache/spark/pull/32987 a while ago to address this in a much more general way. I've tried to keep it up to date, but there seemed to be concerns about creating a subexpression for something that might only execute once I guess? Even though that's already happening in certain cases @viirya @cloud-fan Could you please help review this?
gharchive/pull-request
2022-05-21T14:55:19
2025-04-01T04:56:00.320067
{ "authors": [ "AmplabJenkins", "Kimahriman", "WangGuangxin" ], "repo": "apache/spark", "url": "https://github.com/apache/spark/pull/36626", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
54532482
[SPARK-4874] [CORE] Collect record count metrics Collects record counts for both Input/Output and Shuffle Metrics. For the input/output metrics, it just appends the counter every time the iterators get accessed. For shuffle on the write side, we count the metrics post aggregation (after a map side combine) and on the read side we count the metrics pre aggregation. This allows both the bytes read/written metrics and the records read/written to line up. For backwards compatibility, if we deserialize an older event that doesn't have record metrics, we set the metric to -1. Test build #25628 has started for PR 4067 at commit 571cb69. This patch merges cleanly. Test build #25628 has finished for PR 4067 at commit 571cb69. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/25628/ Test PASSed. Can you also paste some screenshots on what the UI changes look like? Thanks. Test build #25638 has started for PR 4067 at commit 1572054. This patch does not merge cleanly. This change was dependent on https://github.com/apache/spark/pull/3120, that just got merged and now there are some merge conflicts. I need to fix those first and will update the pr. Hey @ksakellis - Thanks for working on this. I took a very quick look at the patch. Overall I feel the patch should be fairly straightforward, but the specific implementations might've gone a bit over board with Scala language features (a lot of Options, orElse, closures, etc) and design (too many new classes introduced). If we can reduce those, the pr would be a lot easier to understand. @rxin I updated the PR after doing a rebase and also incorporated some of your feedback. You made two general comments: specific implementations might've gone a bit over board with Scala language features too many new classes introduced Can you please be more specific here pointing to the specific code. These comments are not very actionable as is. Hi again - can't find my previous comment since the line is no longer in the diff due to the other pr being merged. Can you still add comment for that one (the part with Option and orElse and set ...)? Want to make sure if we read that code one year from now, we can still understand what's going on. Test build #25641 has started for PR 4067 at commit 3c2d021. This patch merges cleanly. The Scala stuff was mostly about the previous PR that got merged (and now no longer showing up as part of this diff). So is this code you were referring to in HadoopRDD? // Find a function that will return the FileSystem bytes read by this thread. Do this before // creating RecordReader, because RecordReader's constructor might read some bytes val bytesReadCallback = inputMetrics.bytesReadCallback.orElse( split.inputSplit.value match { case split: FileSplit => SparkHadoopUtil.get.getFSBytesReadOnThreadCallback(split.getPath, jobConf) case _ => None } ) Shows a stage that has Input Metrics (reading from a file) and writes data for next stage. Shows a stage that has both shuffle reading and writing - no input or output metrics. Shows a stage that has outputting to a file. Test build #25638 timed out for PR 4067 at commit 1572054 after a configured wait of 120m. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/25638/ Test FAILed. Test build #25641 has finished for PR 4067 at commit 3c2d021. This patch passes all tests. This patch merges cleanly. This patch adds the following public classes (experimental): class AfterNextInterceptingIterator[A](sub: Iterator[A]) extends Iterator[A] Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/25641/ Test PASSed. What about combining the input size and records in the same column. Overall this will help with the expansion in the number of columns. The title could be "Input Size / Records" If we do that we wouldn't be able to sort on num records and bytes independently. Yes - you'd only be able to sort on bytes. Wouldn't that be okay? These would likely track closely in most cases. A big motivation to add recordsRead/Written was to detect data skew. In these cases bytes and records might not track very closely. Thinking more about this, I suspect that having an Avg. record Size column (bytesRead/recordsRead) would be what you'd want to sort on. We could add this metric to the UI, make it sortable and then combine the bytesRead and recordsRead metrics into one column. Thoughts? New Screenshot that correspond to CR feedback. Shows a stage that has Input Metrics (reading from a file) and writes data for next stage. Shows a stage that has both shuffle reading and writing - no input or output metrics. Shows a stage that has outputting to a file. @pwendell can you please re-review this? I'd like to get it in to 1.3. Some of our customers have been asking for metrics to help them determine data skew. Thx. Test build #26441 has started for PR 4067 at commit fe3f715. This patch merges cleanly. Test build #26441 has finished for PR 4067 at commit fe3f715. This patch passes all tests. This patch merges cleanly. This patch adds the following public classes (experimental): class AfterNextInterceptingIterator[A](sub: Iterator[A]) extends Iterator[A] Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/26441/ Test PASSed. Having a count does seem like a good idea, but I think the current approach of adding function calls for each iteration is too expensive. What about using a counting iterator and just asking it at the end? BTW we can also just add counting directly to InterruptibleIterator @rxin yeah that seems good. @ksakellis one other thing I realized which is a little confusing, right now we report the bytes as "read" from a cached RDD as soon as they are fetched/present on the executor even if they have not been consumed by the task. Tracking consumption incrementally (in bytes) will be really hard, so maybe this is the best answer for now. It's a bit weird though - I think this logic was written before we sent incremental updates back. In terms of getting the total records read, we might need to make the assumption that the iterator is consumed in its entirety. @pwendell I'm not sure how we can do what you propose without having an O(n) loop through all the records before passing the InterruptableIterator? We could do something fancy like counting incrementally and when we finish the task, if there are more records left, then do the loop to count the rest of the unread records. I don't think the complication is worth it. Also, reporting the accurate records read i think is better. Alternatively, we can fix the bytesRead to be more accurate. Right now they are computed in ShuffleBlockFetcherIterator and calculated based on the blocks fetched. Since we do the flatMap on that iterator in BlockStoreShuffleFetcher we report that we read all the bytes even if we didn't. We can move the bytesRead collection out of ShuffleBlockFetcherIterator and move it into the same iterator that computes the # records read. So they line up and are more accurate. If it would save on perf, I would be in favor of adding the counting to InterruptibleIterator. In core loops like these, I think efficiency is more important than elegance. Test build #26530 has started for PR 4067 at commit 1aaa980. This patch merges cleanly. Test build #26530 has finished for PR 4067 at commit 1aaa980. This patch passes all tests. This patch merges cleanly. This patch adds the following public classes (experimental): class ChiSqSelectorModel (val selectedFeatures: Array[Int]) extends VectorTransformer class ChiSqSelector (val numTopFeatures: Int) Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/26530/ Test PASSed. Test build #26678 has started for PR 4067 at commit 9ecf912. This patch merges cleanly. Test build #26678 has finished for PR 4067 at commit 9ecf912. This patch fails MiMa tests. This patch merges cleanly. This patch adds the following public classes (experimental): class SimpleFunctionRegistry(val caseSensitive: Boolean) extends FunctionRegistry class StringKeyHashMap[T](normalizer: (String) => String) case class MultiAlias(child: Expression, names: Seq[String]) Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/26678/ Test FAILed. Hey @ksakellis - I did a pretty thorough review here, any chance you could take a look? Most of the things were small. The only major thing is that I think if we are going to always show the output metrics for Hadoop (even for older versions) we need to make sure that we at least do a "best effort" presentation of the output size for older versions. I think it would be good if we could sneak that in, since otherwise that output metrics thing is not so useful for many people running with older versions. However, I'd also be okay just not showing either bytes or count in that case (which I think was the behavior before). A second issue I realized (let's open a new JIRA for this) is that we never track output metrics (size or bytes) when data is written to cache, which is confusing because we track input metrics when data is read from cache. New screenshots with the irrelevant columns invisible: Shows a stage that has Input Metrics (reading from a file) and writes data for next stage. Shows a stage that has both shuffle reading and writing - no input or output metrics. Shows a stage outputting to an HDFS file. Test build #26876 has started for PR 4067 at commit e156560. This patch merges cleanly. Test build #26876 has finished for PR 4067 at commit e156560. This patch fails Spark unit tests. This patch merges cleanly. This patch adds no public classes. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/26876/ Test FAILed. Jenkins, retest this please Test build #26896 has started for PR 4067 at commit e156560. This patch merges cleanly. Test build #26896 has finished for PR 4067 at commit e156560. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/26896/ Test PASSed. Test build #26904 has started for PR 4067 at commit dad4d57. This patch merges cleanly. Test build #26906 has started for PR 4067 at commit bd919be. This patch merges cleanly. Test build #26904 has finished for PR 4067 at commit dad4d57. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/26904/ Test PASSed. Test build #26906 has finished for PR 4067 at commit bd919be. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/26906/ Test PASSed. Jenkins, test this please. This LGTM pending tests. Test build #26936 has started for PR 4067 at commit bd919be. This patch merges cleanly. Test build #26936 has finished for PR 4067 at commit bd919be. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/26936/ Test PASSed. Merging this, thanks Kos. It looks like the "InputOutputMetricsSuite input metrics with mixed read methods" and "InputOutputMetricsSuite input metrics with interleaved reads" test may have started failing in the hadoop-2.2 build since this patch: https://amplab.cs.berkeley.edu/jenkins/view/Spark/job/Spark-1.3-SBT/AMPLAB_JENKINS_BUILD_PROFILE=hadoop2.2,label=centos/29/testReport/ Yikes, @JoshRosen i'm looking into this. @ksakellis How do I collect these metrics on a console right after the task is done.
gharchive/pull-request
2015-01-16T02:18:58
2025-04-01T04:56:00.375397
{ "authors": [ "AmplabJenkins", "JoshRosen", "SparkQA", "aguyyala", "ksakellis", "pwendell", "rxin", "sryza" ], "repo": "apache/spark", "url": "https://github.com/apache/spark/pull/4067", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1960156183
[SPARK-45655][SQL][SS] Allow non-deterministic expressions inside AggregateFunctions in CollectMetrics What changes were proposed in this pull request? This PR allows non-deterministic expressions wrapped inside an AggregateFunction such as count inside CollectMetrics node. CollectMetrics is used to collect arbitrary metrics from the query, in certain scenarios user would like to collect metrics for filtering based on non-deterministic expressions (see query example below). Currently, Analyzer does not allow non-deterministic expressions inside a AggregateFunction for CollectMetrics. This constraint is relaxed to allow collection of such metrics. Note that the metrics are relevant for a completed batch, and can change if the batch is replayed (because non-deterministic expression can behave differently for different runs). While working on this feature, I found a issue with checkMetric logic to validate non-deterministic expressions inside an AggregateExpression. An expression is determined as non-deterministic if any of its children is non-deterministic, hence we need to match the case for !e.deterministic && !seenAggregate after we have matched if the current expression is a AggregateExpression. If the current expression is a AggregateExpression, we should validate further down in the tree recursively - otherwise we will fail for any non-deterministic expression. val inputData = MemoryStream[Timestamp] inputData.toDF()       .filter("value < current_date()")       .observe("metrics", count(expr("value >= current_date()")).alias("dropped"))       .writeStream       .queryName("ts_metrics_test")       .format("memory")       .outputMode("append")       .start() Why are the changes needed? Added a testcase to calculate dropped rows (by CurrentBatchTimestamp) and ensure the query is successful. As an example, the query below fails (without this change) due to observe call on the DataFrame. val inputData = MemoryStream[Timestamp] inputData.toDF()       .filter("value < current_date()")       .observe("metrics", count(expr("value >= current_date()")).alias("dropped"))       .writeStream       .queryName("ts_metrics_test")       .format("memory")       .outputMode("append")       .start() Added testing in AnalysisSuite for non-deterministic expressions inside a AggregateFunction. Does this PR introduce any user-facing change? No How was this patch tested? Unit test cases added. [warn] 20 warnings found WARNING: Using incubator modules: jdk.incubator.vector, jdk.incubator.foreign [info] StreamingQueryStatusAndProgressSuite: 09:14:39.684 WARN org.apache.hadoop.util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable [info] Passed: Total 0, Failed 0, Errors 0, Passed 0 [info] No tests to run for hive / Test / testOnly [info] - StreamingQueryProgress - prettyJson (436 milliseconds) [info] - StreamingQueryProgress - json (3 milliseconds) [info] - StreamingQueryProgress - toString (5 milliseconds) [info] - StreamingQueryProgress - jsonString and fromJson (163 milliseconds) [info] - StreamingQueryStatus - prettyJson (1 millisecond) [info] - StreamingQueryStatus - json (1 millisecond) [info] - StreamingQueryStatus - toString (2 milliseconds) 09:14:41.674 WARN org.apache.spark.sql.execution.streaming.ResolveWriteToStream: Temporary checkpoint location created which is deleted normally when the query didn't fail: /Users/bhuwan.sahni/workspace/spark/target/tmp/temporary-34d2749f-f4d0 -46d8-bc51-29da6411e1c5. If it's required to delete it under any circumstances, please set spark.sql.streaming.forceDeleteTempCheckpointLocation to true. Important to know deleting temp checkpoint folder is best effort. 09:14:41.710 WARN org.apache.spark.sql.execution.streaming.ResolveWriteToStream: spark.sql.adaptive.enabled is not supported in streaming DataFrames/Datasets and will be disabled. [info] - progress classes should be Serializable (5 seconds, 552 milliseconds) 09:14:46.345 WARN org.apache.spark.sql.execution.streaming.ResolveWriteToStream: Temporary checkpoint location created which is deleted normally when the query didn't fail: /Users/bhuwan.sahni/workspace/spark/target/tmp/temporary-3a41d397-c3c1 -490b-9cc7-d775b0c42208. If it's required to delete it under any circumstances, please set spark.sql.streaming.forceDeleteTempCheckpointLocation to true. Important to know deleting temp checkpoint folder is best effort. 09:14:46.345 WARN org.apache.spark.sql.execution.streaming.ResolveWriteToStream: spark.sql.adaptive.enabled is not supported in streaming DataFrames/Datasets and will be disabled. [info] - SPARK-19378: Continue reporting stateOp metrics even if there is no active trigger (1 second, 337 milliseconds) 09:14:47.677 WARN org.apache.spark.sql.execution.streaming.ResolveWriteToStream: spark.sql.adaptive.enabled is not supported in streaming DataFrames/Datasets and will be disabled. [info] - SPARK-29973: Make `processedRowsPerSecond` calculated more accurately and meaningfully (455 milliseconds) 09:14:48.174 WARN org.apache.spark.sql.execution.streaming.ResolveWriteToStream: Temporary checkpoint location created which is deleted normally when the query didn't fail: /Users/bhuwan.sahni/workspace/spark/target/tmp/temporary-360fc3b9-a2c5 -430c-a892-c9869f1f8339. If it's required to delete it under any circumstances, please set spark.sql.streaming.forceDeleteTempCheckpointLocation to true. Important to know deleting temp checkpoint folder is best effort. 09:14:48.174 WARN org.apache.spark.sql.execution.streaming.ResolveWriteToStream: spark.sql.adaptive.enabled is not supported in streaming DataFrames/Datasets and will be disabled. [info] - SPARK-45655: Use current batch timestamp in observe API (587 milliseconds) 09:14:48.768 WARN org.apache.spark.sql.streaming.StreamingQueryStatusAndProgressSuite: Was this patch authored or co-authored using generative AI tooling? No FYI, the previous test failure was relevant to this change, and @sahnib is looking into it. (I'm sorry to miss leaving a note in prior.) Please check SPARK-45655: Use current batch timestamp in observe API *** FAILED *** (241 milliseconds) Please check SPARK-45655: Use current batch timestamp in observe API *** FAILED *** (241 milliseconds) This test was failing because CI machine has milliseconds precision of 8 decimal points, I have fixed the testcase. Thanks! Merging to master. @sahnib Thanks for your first contribution to Apache Spark! I merged this to master.
gharchive/pull-request
2023-10-24T22:05:48
2025-04-01T04:56:00.385935
{ "authors": [ "HeartSaVioR", "beliefer", "sahnib" ], "repo": "apache/spark", "url": "https://github.com/apache/spark/pull/43517", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2373990818
[MINOR] Fix CodeStyle Error for ExpressionImplUtilsSuite What changes were proposed in this pull request? The pr is following up https://github.com/apache/spark/pull/46845, to fix dev/scalastyle check error. Why are the changes needed? Make sh dev/scalastyle hadppy. Does this PR introduce any user-facing change? No. How was this patch tested? Manually test sh dev/scakastyle Before: (base) ➜ spark-community git:(master) ✗ sh dev/scalastyle -e Scalastyle checks failed at following occurrences: [error] /Users/panbingkun/Developer/spark/spark-community/sql/catalyst/src/test/java/org/apache/spark/sql/catalyst/expressions/ExpressionImplUtilsSuite.scala:394:28: nonascii.message [error] /Users/panbingkun/Developer/spark/spark-community/sql/catalyst/src/test/java/org/apache/spark/sql/catalyst/expressions/ExpressionImplUtilsSuite.scala:396:28: nonascii.message [error] Total time: 23 s, completed Jun 26, 2024, 9:49:06 AM After: (base) ➜ spark-community git:(fix_scalastyle) ✗ sh dev/scalastyle -e Scalastyle checks passed. Was this patch authored or co-authored using generative AI tooling? No. cc @HyukjinKwon @cloud-fan @uros-db let me merge it now to unblock the CI.
gharchive/pull-request
2024-06-26T01:43:18
2025-04-01T04:56:00.390023
{ "authors": [ "cloud-fan", "panbingkun" ], "repo": "apache/spark", "url": "https://github.com/apache/spark/pull/47094", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
67102377
[SQL] Faster Scala row conversion This is a follow-up of #5279 and #5398. ScalaReflection.convertRowToScala is on a critical path, but was implemented in a rather inefficient way. Test build #29856 has started for PR 5419 at commit d698d03. Test build #29856 has finished for PR 5419 at commit d698d03. This patch passes all tests. This patch merges cleanly. This patch adds no public classes. This patch does not change any dependencies. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/29856/ Test PASSed. Actually the convertRowToScala part is already handled in #5279, and the lazy val schema part is handled in #5398, so I'm closing this.
gharchive/pull-request
2015-04-08T10:48:06
2025-04-01T04:56:00.395135
{ "authors": [ "AmplabJenkins", "SparkQA", "liancheng" ], "repo": "apache/spark", "url": "https://github.com/apache/spark/pull/5419", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
112479556
[SPARK-11221] [SPARKR] fix R doc for lit and add examples Currently the documentation for lit is inconsistent with doc format, references "Scala symbol" and has no example. Fixing that. @shivaram Can one of the admins verify this patch? Jenkins, ok to test Jenkins, add to whitelist Merged build triggered. Merged build started. Test build #44013 has started for PR 9187 at commit 1ca6612. Test build #44013 has finished for PR 9187 at commit 1ca6612. This patch passes all tests. This patch merges cleanly. This patch adds the following public classes (experimental):\n * class BinaryClassificationEvaluator @Since(\"1.4.0\") (@Since(\"1.4.0\") override val uid: String)\n * class MulticlassClassificationEvaluator @Since(\"1.5.0\") (@Since(\"1.5.0\") override val uid: String)\n * final class RegressionEvaluator @Since(\"1.4.0\") (@Since(\"1.4.0\") override val uid: String)\n * abstract class ColumnarIterator extends Iterator[InternalRow] \n * class SpecificColumnarIterator extends $\n Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/44013/ Test PASSed. LGTM. Thanks @felixcheung Merging this
gharchive/pull-request
2015-10-20T23:19:56
2025-04-01T04:56:00.401412
{ "authors": [ "AmplabJenkins", "SparkQA", "felixcheung", "shivaram" ], "repo": "apache/spark", "url": "https://github.com/apache/spark/pull/9187", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
246847633
WIP: STORM-2648/STORM-2357: Add storm-kafka-client support for at-most-onc… …e processing and a toggle for whether messages should be emitted with a message id when not using at-least-once See https://issues.apache.org/jira/browse/STORM-2357 and https://issues.apache.org/jira/browse/STORM-2648. I'd like to get some opinions on whether this approach is a good idea, or whether I've overlooked a better option, before finishing this up with some tests. I don't love that we'll end up with 3 different committing behaviors. In 2357 it was noted that the spout doesn't currently support true at-most-once, because using Kafka's auto commit option leaves the possibility that the spout receives a tuple, emits it to the topology, crashes and recovers, and then receives and emits the same tuple. The linked issue suggests solving this by committing polled offsets before emitting them to the topology, which is an option added here. 2648 notes that there is currently no way to make Storm track messages when using auto commit with this spout. This prevents Storm UI from showing the complete latency for the spout, and I would assume also prevents max spout pending from having an effect. I've added a toggle to KafkaSpoutConfig to force the spout to emit messages with message ids, even when using auto commit or the at-most-once option. The spout does nothing on ack or fail when not doing at-least-once. I'd like to keep the spout config simple for the user, so I've added a processing guarantee setting corresponding to the standard at-least-once code path, the path that uses auto commit, and the path that commits offsets before emitting any tuples. @srdo can you please assign JIRA's to you and mark them as in progress as you work on them and/or submit a pull request. Thanks. Yes, I forgot. Conceptually the changes look good to me. I have not dug into it in great detail yet, but I do like the direction of the change. I would also like to see the documentation and examples updated to reflect the new change. +1 @srdo reviewing it Added some tests and updated the docs. @hmcl Are you reviewing this, or are you satisfied with it? Still +1 @hmcl I don't mean to try to rush you, but please let me know if you're still reviewing. If not I'll probably merge in the next few days. @srdo apologies for the delay. I will finish today. Thanks :) @hmcl Are you still reviewing this? It's fine if you don't have time to look at this, but please say so. I'd like to not keep holding this up. This PR has been waiting for about 2 months, and once it gets +1 and no -1, it can be merged. I'm +1 and will just merge. @hmcl Please vote -1 later and rollback the merge if you have concern about the patch and would want to vote -1. Merged via https://github.com/apache/storm/commit/48f6969027e7b02a5b9220577189d3911aa2226d @srdo Sorry I forgot to add auto close message while squashing commits. Could you close this? And please craft the patch for 1.x branch since it doesn't looks like a clean cherry-pick. Thanks in advance! Thanks for reviews. Opened the 1.x version here https://github.com/apache/storm/pull/2353.
gharchive/pull-request
2017-07-31T18:46:34
2025-04-01T04:56:00.410160
{ "authors": [ "HeartSaVioR", "hmcl", "revans2", "srdo" ], "repo": "apache/storm", "url": "https://github.com/apache/storm/pull/2249", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
479285601
STORM-3071: change checkstyle plugin setting logViolationsToConsole t… …o true Logging checkstyle violations to console which is now possible since all violations have been fixed Remove unnecessary maxAllowedViolations from configurations of projects Note: The removal of maxAllowedViolations in the server project is covered in #3050 which needs to be merged before. I took a look at #3050 . We can come back to this once the former's changes are merged. It looks nice. Will come back to this once #3050 is merged #3050 is now merged. Could you please rebase this so we can get this in to? Thanks again for the efforts The failed test doesn't seem related [INFO] Running org.apache.storm.daemon.drpc.DRPCTest [ERROR] Tests run: 6, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 1.038 s <<< FAILURE! - in org.apache.storm.daemon.drpc.DRPCTest [ERROR] testGoodBlocking Time elapsed: 0.012 s <<< ERROR! java.util.concurrent.TimeoutException
gharchive/pull-request
2019-08-10T16:47:16
2025-04-01T04:56:00.412958
{ "authors": [ "Ethanlm", "d2r", "krichter722" ], "repo": "apache/storm", "url": "https://github.com/apache/storm/pull/3100", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
611290066
SUBMARINE-487. JNA does not work with maven-shade-plugin What is this PR for? submarine-spark-security fails with java.lang.UnsatisfiedLinkError What type of PR is it? [Bug Fix] Todos [ ] - Task What is the Jira issue? https://issues.apache.org/jira/browse/SUBMARINE-487 How should this be tested? Screenshots (if appropriate) Questions: Does the licenses files need update? No Is there breaking changes for older versions? No Does this needs documentation? No @gumartinm Thank you for your interest in the submarine project. :-) @yaooqinn Please help review this PR, Thanks! Thanks for the fix, LGTM @gumartinm feel free to send follow-ups with your no-uber-jar ideas, I use uber-jar here because of jar conflict between spark and ranger project. @liuxunorg I am ok to merge this. @gumartinm feel free to send follow-ups with your no-uber-jar ideas, I use uber-jar here because of jar conflict between spark and ranger project. @liuxunorg I am ok to merge this. @yaooqinn I just wanted to say that you are doing a great job with this project. I've been following it since it was called spark-authorizer. Hopefully, one day, it will be integrated in the Spark framework.
gharchive/pull-request
2020-05-02T22:48:02
2025-04-01T04:56:00.418095
{ "authors": [ "gumartinm", "liuxunorg", "yaooqinn" ], "repo": "apache/submarine", "url": "https://github.com/apache/submarine/pull/273", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
913343921
Determination of the select queries in BaseEngineSpec Hello! We are now developing a new SQLAlchemy dialect for Kusto (Azure Data Explorer) and want add support for Kusto database in Superset in nearly future. (This also covers https://github.com/apache/superset/issues/10646.) Kusto language has two dialects: SQL and KQL. While there is no problem with SQL dialect, we also want to integrate KQL in Superset. Problem KQL select queries do not contain the word "SELECT" so they look like MyTable | take 10 (this is equivalent of Select * from MyTable Limit 10). The absence of the "SELECT" keyword breaks the limit behavior in SQLLab because it relies on the sqlparse package and we can't add a limit to queries from UI. Proposed solution The solution is to make the determination of the select queries in BaseEngineSpec so that you can override it in your own spec. This is the same way how base dialect determines read-only queries with is_readonly_query. We will provide PR to implement this small feature. Thank you! Closed by https://github.com/apache/superset/pull/15013
gharchive/issue
2021-06-07T09:34:20
2025-04-01T04:56:00.421585
{ "authors": [ "xneg" ], "repo": "apache/superset", "url": "https://github.com/apache/superset/issues/15011", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2755104947
Edit CSS Live can not @import google fonts? Bug description I try to @import google font in Edit CSS Live section. It not effect ??? @import url('https://fonts.googleapis.com/css2?family=Prompt:ital,wght@0,100;0,200;0,300;0,400;0,500;0,600;0,700;0,800;0,900;1,100;1,200;1,300;1,400;1,500;1,600;1,700;1,800;1,900&display=swap'); body{ font-family:'Prompt'; } Screenshots/recordings No response Superset version 4.1.1 Python version 3.9 Node version 16 Browser Chrome Additional context No response Checklist [ ] I have searched Superset docs and Slack and didn't find a solution to my problem. [X] I have searched the GitHub issue tracker and didn't find a similar bug report. [ ] I have checked Superset's logs for errors and if I found a relevant Python stacktrace, I included it here as text in the "additional context" section. You can, but dosubot is rigjt... you will have to punch a hole in TALISMAN_CONFIG for it. Closing this since it's not a bug, but a question that's now been answered. Holler if you have any further troubles with it :)
gharchive/issue
2024-12-23T03:02:54
2025-04-01T04:56:00.426641
{ "authors": [ "psongyoo", "rusackas" ], "repo": "apache/superset", "url": "https://github.com/apache/superset/issues/31597", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
942367461
refactor: Move is_user_admin() to security_manager SUMMARY Having is_user_admin is a design flaw, but in any case should always be handled within the security manager so it can be overridden. BEFORE/AFTER SCREENSHOTS OR ANIMATED GIF TESTING INSTRUCTIONS Test admin user works as befre ADDITIONAL INFORMATION [ ] Has associated issue: [ ] Changes UI [ ] Includes DB Migration (follow approval process in SIP-59) [ ] Migration is atomic, supports rollback & is backwards-compatible [ ] Confirm DB migration upgrade and downgrade tested [ ] Runtime estimates and downtime expectations provided [ ] Introduces new feature or API [ ] Removes existing feature or API @bolkedebruin are you still hoping to get this PR merged? There are some review comments that would need to be addressed to get this merged (and a rebase). Yes, I'll take a look at it again.
gharchive/pull-request
2021-07-12T19:17:29
2025-04-01T04:56:00.430813
{ "authors": [ "bolkedebruin", "villebro" ], "repo": "apache/superset", "url": "https://github.com/apache/superset/pull/15636", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1338283845
fix(install): set SUPERSET_VERSION_RC at the right time fixes #21082 SUMMARY Fixes problem as described in the issue #21082 TESTING INSTRUCTIONS execute command below in a environment which doesn't contain any of the superset env variables. [YOUR_SUPERSET_PATH]/RELEASING/make_tarball.sh [VERSION] [RC] [PGP_KEY_NAME] this should run without failures. ADDITIONAL INFORMATION fixes #21082 [ ] Has associated issue: [ ] Required feature flags: [ ] Changes UI [ ] Includes DB Migration (follow approval process in SIP-59) [ ] Migration is atomic, supports rollback & is backwards-compatible [ ] Confirm DB migration upgrade and downgrade tested [ ] Runtime estimates and downtime expectations provided [ ] Introduces new feature or API [ ] Removes existing feature or API @Joel-Haeberli — Sorry this slipped though the cracks for so long! Thanks for reporting both the issue and opening the PR. Please feel free to tag committers here or ping us on Superset Slack if things get stuck like this. @eschutho this seems like a straightforward fix to a problem I haven't encountered... curious your take :)
gharchive/pull-request
2022-08-14T15:51:23
2025-04-01T04:56:00.435736
{ "authors": [ "Joel-Haeberli", "rusackas" ], "repo": "apache/superset", "url": "https://github.com/apache/superset/pull/21083", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2287450475
fix: tests on database, dataset apis SUMMARY BEFORE/AFTER SCREENSHOTS OR ANIMATED GIF TESTING INSTRUCTIONS ADDITIONAL INFORMATION [ ] Has associated issue: [ ] Required feature flags: [ ] Changes UI [ ] Includes DB Migration (follow approval process in SIP-59) [ ] Migration is atomic, supports rollback & is backwards-compatible [ ] Confirm DB migration upgrade and downgrade tested [ ] Runtime estimates and downtime expectations provided [ ] Introduces new feature or API [ ] Removes existing feature or API Codecov Report All modified and coverable lines are covered by tests :white_check_mark: Project coverage is 64.13%. Comparing base (2e5f3ed) to head (2a94fab). Report is 57 commits behind head on master. Additional details and impacted files @@ Coverage Diff @@ ## master #28397 +/- ## ========================================== + Coverage 60.49% 64.13% +3.64% ========================================== Files 1931 521 -1410 Lines 76241 37255 -38986 Branches 8566 0 -8566 ========================================== - Hits 46122 23895 -22227 + Misses 28015 13360 -14655 + Partials 2104 0 -2104 Flag Coverage Δ hive 49.08% <ø> (-0.09%) :arrow_down: javascript ? presto 53.66% <ø> (-0.14%) :arrow_down: python 64.13% <ø> (+0.65%) :arrow_up: unit 58.35% <ø> (+0.72%) :arrow_up: Flags with carried forward coverage won't be shown. Click here to find out more. :umbrella: View full report in Codecov by Sentry. :loudspeaker: Have feedback on the report? Share it here.
gharchive/pull-request
2024-05-09T11:03:40
2025-04-01T04:56:00.447563
{ "authors": [ "codecov-commenter", "dpgaspar" ], "repo": "apache/superset", "url": "https://github.com/apache/superset/pull/28397", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1181456820
TOMEE-3871 - TomEE Plume is missing BatchEE / JCS Cache batchee and jcs added in Plume to reflect Plus flavor removed xml api from from Plume to reflect Plus flavor remains some jar refs in jar scan exclusions (xerces, xml-apis) AFAIK plume is a webprofile based distro so batchee and jcs should belong to a new "plume-all" distro probably? AFAIK plume is a webprofile based distro so batchee and jcs should belong to a new "plume-all" distro probably? This originates from this list thread and @dblevins comment there, that: "Those distributions are supposed to be the same minus the JPA and JSF providers.". There is also a blog post about the differences but I am too new to judge the original thoughts regarding "plume" vs "plus". AFAIK plume is a webprofile based distro if this is correct then Plume should not contains jars for MicroProfile nor JMS ActiveMQ Yep, think they shouldnt but as Im not a big user of that I dont care much of what it becomes while webpro and plus stays aligned on their goal ;) webpro and plus have a clear focus to me as well, even though "Plus" branding could be changed to better tell its "Full EE" 🤔 or i can just describe that in the flavors page. You could drop Plume but you would lose the benefits from its "ready to try out" capabilities. i am thinking about the AutoConfiguration for Jakarta Persistence (JPA) which facilitates learning Tomee. I think it is good to have this discussion to clarify what the intended distributions should look like in 2022 considering the history of Plus / Plume and the incosistent current status of the distros / website. So thanks for the additional thoughts @rmannibucau and @sultan does this need a mailing list vote from committers ? Here's the list message on the name and motivation for Plume: https://lists.apache.org/thread/tkpfjlzv70y8kzz5f905qj9vs95b34rw "a version of TomEE Plus with Mojarra and EclipseLink. Tentatively called it "TomEE Plume" which stands for TomEE PLUs, Mojarra, Eclipselink (PLU.M.E)." nice to read ! i learned that a setup installer for TomEE "like eclipse oomph" (cherry pick modules) is not possible, as it breaks TCK rules. unless we want a parallel non TCK certified TomEE (meh) i thought like @rmannibucau at first with "per functionalities" LIB folders (like others EE servers) but this would make TomEE harder to configure / extend. (where do user put their jars !?) there could be "per functionalities" uber or shaded jars to ease remove and replace implementations by others. but i wonder if i did read that current project policy is to avoid shading/uber. there could be a common POM between plus and plume to help fill the gaps/ease maintenance. there could be an external tool to modify a dezipped certified TomEE to change its implementation jars. the only one i think i can currently help on is n°4 (common plus-plume pom) there could be an external tool to modify a dezipped certified TomEE to change its implementation jars. If you look tomee-maven-plugin it enables to do any distro from plain tomcat to advanced customized tomee and you can use it to build a tomee plume from tomee webprofile or any distro ;). Installer is legacy since years :) @sultan Can you check your rebase with tomee-8.x. It does not look sane to me (at least judging from the changeset now). i think i was able to clean up my mess by reverting to my previous draft. commits seems legit now. can someone launch checks ?
gharchive/pull-request
2022-03-26T01:47:48
2025-04-01T04:56:00.475226
{ "authors": [ "dblevins", "rmannibucau", "rzo1", "sultan" ], "repo": "apache/tomee", "url": "https://github.com/apache/tomee/pull/828", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1448526684
Revert Traffic Monitor renaming to rascal Which Traffic Control components are affected by this PR? Documentation Traffic Control Cache Config (t3c, formerly ORT) Traffic Control Health Client (tc-health-client) Traffic Control Client Traffic Monitor Traffic Ops Traffic Portal Traffic Router Traffic Stats Grove CDN in a Box Automation What is the best way to verify this PR? Run integration, unit test. Ensure CIAB build correctly TO API endpoints work as expected. Clients (eg: T3C) work as expected. If this is a bugfix, which Traffic Control versions contained the bug? PR submission checklist [x] This PR has tests [x] This PR has documentation [x] This PR has a CHANGELOG.md entry [x] This PR DOES NOT FIX A SERIOUS SECURITY VULNERABILITY (see the Apache Software Foundation's security guidelines for details) This PR reverts #7068
gharchive/pull-request
2022-11-14T18:15:54
2025-04-01T04:56:00.482504
{ "authors": [ "rimashah25", "zrhoffman" ], "repo": "apache/trafficcontrol", "url": "https://github.com/apache/trafficcontrol/pull/7190", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1569966862
Still load legacy config if it exists instead of failing Everything is broken and I'm not prepared for records.yaml Thanks, @cmcfarlen - swapping from a format that's been in use for years to yaml necessitates a change of a ton of tooling. The prior change would have been breaking for just about every org that uses ATS, with a significant investment in trying to change every part of build, validation, and config management. Even if you believe that swapping to yaml would be better, you cannot expect that to be a minor change. We have already talked about this on slack but just for the record: As discussed(dev and user mailing list) before, we agreed to have single parsing and fail if the legacy file is found. As also requested, a tool is provided for the translation with a guide on how to move on. I have no issues with having it dual but just making sure that we are all on the same page. If there is anything I can do to make the process easy just let me know and I'll do it. I've created an issue which I will be implementing to track a solution for this. https://github.com/apache/trafficserver/issues/9384 We talked about this in the PR/issue scrub and would like to have a discussion at next weeks PR/issue scrub. A final decision should go to the mailing list. Marking it as a draft, so it doesn't get merged in before we discuss. I'm going to close this and we can work on the yaml support around tooling instead.
gharchive/pull-request
2023-02-03T14:54:12
2025-04-01T04:56:00.485791
{ "authors": [ "DoingTheNeedful", "brbzull0", "bryancall", "cmcfarlen" ], "repo": "apache/trafficserver", "url": "https://github.com/apache/trafficserver/pull/9377", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1174113276
[skip ci] Skip test_conv2d in Hexagon https://github.com/apache/tvm/pull/10666 marked this xfail but it still runs the test causing segfault and CI failure in PR jobs and main. @driazati @areusch @junrushao1994 But this test already has https://github.com/apache/tvm/blob/871a83f9719d74d2eb15786d8b8878229b194c53/tests/python/contrib/test_hexagon/test_conv2d_conv2d.py#L165-L167 So I wonder why i386 test is still failing... Existing @pytest.mark.skipif( platform.processor() == "i686", reason="Test known to be flaky on i386 machines" ) doesn't seem to help actually skipping test. I'm trying if platform.processor() == "i386" does make things skip. As the log https://ci.tlcpack.ai/blue/rest/organizations/jenkins/pipelines/tvm/branches/PR-10687/runs/7/nodes/315/steps/583/log/?start=0 shows, platform.machine() prints x86_64 despite being supposed to run on the i386 image. I don't know have a good way to conditionally skip this test on i386, so I'm going to skip it completely.
gharchive/pull-request
2022-03-18T23:48:48
2025-04-01T04:56:00.489498
{ "authors": [ "masahi" ], "repo": "apache/tvm", "url": "https://github.com/apache/tvm/pull/10687", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1625007663
[Fix][TIR] Fix tvm::arith::UnionLowerBound The UnionLowerBound function does not take into account the condition that the empty set has a special representation [+inf, -inf]. cc @wrongtest-intellif cc @wrongtest-intellif Could you please add a regression test? We could add a testcase to tests/python/unittest/test_arith_intset.py::test_union_lower_bound Could you please add a regression test? Already added, thanks for your suggestion. We could add a testcase to tests/python/unittest/test_arith_intset.py::test_union_lower_bound Already added, thanks for your suggestion @tvm-bot rerun Please make the CI green and we can get it in :-) Hi, CI is finished.
gharchive/pull-request
2023-03-15T08:30:25
2025-04-01T04:56:00.493334
{ "authors": [ "Biubiubiu12", "Hzfengsy", "wrongtest-intellif" ], "repo": "apache/tvm", "url": "https://github.com/apache/tvm/pull/14304", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2264251604
[YUNIKORN-2587] Core: Convert AllocationID to AllocationKey What is this PR for? Having both AllocationID and AllocationKey in the core is redundant. Replace all references to AllocationID with AllocationKey and remove the duplicate variable in Allocation. What type of PR is it? [ ] - Bug Fix [ ] - Improvement [ ] - Feature [ ] - Documentation [ ] - Hot Fix [ ] - Refactoring Todos [x] - Task What is the Jira issue? https://issues.apache.org/jira/browse/YUNIKORN-2587 How should this be tested? Unit tests updated as necessary. Screenshots (if appropriate) Questions: [ ] - The licenses files need update. [ ] - There is breaking changes for older versions. [ ] - It needs documentation. Codecov Report Attention: Patch coverage is 79.16667% with 15 lines in your changes are missing coverage. Please review. :exclamation: No coverage uploaded for pull request base (master@e17eafa). Click here to learn what that means. Files Patch % Lines pkg/scheduler/partition.go 75.75% 7 Missing and 1 partial :warning: pkg/examples/simple_example.go 0.00% 2 Missing :warning: pkg/scheduler/objects/application.go 83.33% 2 Missing :warning: pkg/scheduler/objects/node.go 86.66% 2 Missing :warning: pkg/scheduler/context.go 0.00% 1 Missing :warning: Additional details and impacted files @@ Coverage Diff @@ ## master #856 +/- ## ========================================= Coverage ? 77.02% ========================================= Files ? 97 Lines ? 12002 Branches ? 0 ========================================= Hits ? 9244 Misses ? 2421 Partials ? 337 :umbrella: View full report in Codecov by Sentry. :loudspeaker: Have feedback on the report? Share it here.
gharchive/pull-request
2024-04-25T18:34:02
2025-04-01T04:56:00.506783
{ "authors": [ "codecov-commenter", "craigcondit" ], "repo": "apache/yunikorn-core", "url": "https://github.com/apache/yunikorn-core/pull/856", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
165943464
[ZEPPELIN-1196] Fix for bug ZEPPELIN-1196 What is this PR for? Fixed issue related to connecting to remote running interpreter process with multiple interpreters in interpreter group throws illegal thread state exception What type of PR is it? Bug Fix Todos What is the Jira issue? [ZEPPELIN-1196] https://issues.apache.org/jira/browse/ZEPPELIN-1196 Questions: Does the licenses files need update? No Is there breaking changes for older versions? No Does this needs documentation? No One of the travis check is failing but its not related to my changes Can someone please review this PR @SachinJanani Hi, can you add a test case for it? It would help understand and test this PR. @jongyoul Thanks for reviewing.As suggested I have added a test case for this PR @SachinJanani Thanks for the quick fix, but if it fails to launch RemoteInterpreterServer with 10 seconds, this test always fails, isn't it? how to deal with that case? I think we need to make it more concrete. How about you? @jongyoul Yes you are right if RemoteInterpreterServer fails to launch within 10 seconds then this test will fail,but we need to have some threshold for that because if say there is some bug in RemoteInterpreterServer which causes it to hang then our test will never complete.Should I increase it to 20 seconds.What you think? @jongyoul One more thing I found that as per ZEPPELIN_INTERPRETER_CONNECT_TIMEOUT property default timeout that we consider is 30 seconds.So I think setting it to 30 seconds will make more sense. @SachinJanani 30 seconds would be realistic. And could you please leave memo that it may occurs a potential bug when it fails launching it? Thanks @jongyoul. Sure will make the change also will add a comment in the code about about the 30 seconds timeout LGTM. Merging if there's no more discussion. Can someone please merge this PR as I don't have permission I'll merge it after 24 hours. @SachinJanani I tried to merge it but this is based on branch-0.6. Could you please rebase from current master? I've tried to do it but I occurs error about merge conflict. Can you handle it? And for the next time, you'd better contribute the codes to avoid this situation. @jongyoul This changes are only for branch-0.6 as master does not have this files (even this code) anymore due to helium changes.Also i think issue might not occur for zeppelin-0.7.So this is fix will only be for zeppelin-0.6.1.I think it should be only be merged in branch-0.6. @jongyoul @SachinJanani I managed it to be merged only to branch-0.6. Will close this PR manually. Thanks @minahlee
gharchive/pull-request
2016-07-16T19:15:41
2025-04-01T04:56:00.516235
{ "authors": [ "SachinJanani", "jongyoul", "minahlee" ], "repo": "apache/zeppelin", "url": "https://github.com/apache/zeppelin/pull/1197", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
410195180
Finished genReadOp() description. #20 Answered all the questions of part 1 of the assignment. Should fix #20. Codecov Report Merging #43 into lab3 will not change coverage. The diff coverage is n/a. @@ Coverage Diff @@ ## lab3 #43 +/- ## ======================================= Coverage 68.31% 68.31% ======================================= Files 107 107 Lines 7334 7334 Branches 1388 1388 ======================================= Hits 5010 5010 Misses 1868 1868 Partials 456 456 Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update a8c3c77...82b8b44. Read the comment docs. Codecov Report Merging #43 into lab3 will not change coverage. The diff coverage is n/a. @@ Coverage Diff @@ ## lab3 #43 +/- ## ======================================= Coverage 68.31% 68.31% ======================================= Files 107 107 Lines 7334 7334 Branches 1388 1388 ======================================= Hits 5010 5010 Misses 1868 1868 Partials 456 456 Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update a8c3c77...82b8b44. Read the comment docs.
gharchive/pull-request
2019-02-14T09:03:35
2025-04-01T04:56:00.549293
{ "authors": [ "codecov-io", "emilelmarsson" ], "repo": "apeinot/java", "url": "https://github.com/apeinot/java/pull/43", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
470912710
Fix saving uuid as item title This bug happens in some cases when hitting Save button the XML of an item is still not available . Not sure this is the correct fix. I mean, it may work, but I think the problem comes from the XML doc being out of date because the client script getState promise only runs once. There is also still the problem of validation hooks not being unloaded, but that's a different issue :) Not sure this is the correct fix. I mean, it may work, but I think the problem comes from the XML doc being out of date because the client script getState promise only runs once. There is also still the problem of validation hooks not being unloaded, but that's a different issue :) Yes, getState promise only runs once. So I try to get it running again inside the "presubmit". And then I found the real reason is the item's XML is not really ready when calling getState. If you go to line 106 of WizardApi, you would see 'itemPack.getXml ' actually does not have the XML we need yet. Something I would test is: Put a repeater on the first page and add some repeater items. Now remove a repeater item. Now save + publish. Check the item you removed is not in the XML And: Put some optional controls on page 2 Go to page 2 and fill in some values Go back to page 1 Go back to page 2 and put in some new values Now save + publish. Check that the first set of values you put on page 2 are not in the XML. Something I would test is: Put a repeater on the first page and add some repeater items. Now remove a repeater item. Now save + publish. Check the item you removed is not in the XML And: Put some optional controls on page 2 Go to page 2 and fill in some values Go back to page 1 Go back to page 2 and put in some new values Now save + publish. Check that the first set of values you put on page 2 are not in the XML. Thanks Aaron. Let me test these two cases now. For the second test, looks good. For the Repeater test :
gharchive/pull-request
2019-07-22T06:18:32
2025-04-01T04:56:00.559078
{ "authors": [ "PenghaiZhang", "abidingotter" ], "repo": "apereo/openEQUELLA", "url": "https://github.com/apereo/openEQUELLA/pull/1127", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1879118953
hysteria2 ERROR info in openwrt 1.system: run hysteria2 in OpenWrt with transparent proxy 2.client config { "tls": { "insecure": false, "sni": "mysite.com" }, "lazy": true, "tcpTProxy": { "listen": "0.0.0.0:1234" }, "socks5":{ "listen": "0.0.0.0:1080" }, "fast_open": true, "bandwidth": { "down": "300 mbps", "up": "100 mbps" }, "disable_mtu_discovery": false, "auth": "Strong_password", "server": "mysite.com:443" } 3.ERROR info 2023-09-03T14:11:20Z INFO TCP transparent proxy listening {"addr": "0.0.0.0:1234"} 2023-09-03T14:11:23Z INFO connected to server {"count": 1} 2023-09-03T14:11:23Z ERROR TCP transparent proxy error {"addr": "123.96.20.227:40526", "reqAddr": "127.0.0.1:1234", "error": "dial error: dial tcp4 127.0.0.1: 1234: connect: connection refused"} 2023-09-03T14:11:23Z ERROR TCP transparent proxy error {"addr": "123.96.20.227:40534", "reqAddr": "127.0.0.1:1234", "error": "dial error: dial tcp4 127.0.0.1: 1234: connect: connection refused"} 2023-09-03T14:11:23Z ERROR TCP transparent proxy error {"addr": "123.96.20.227:40542", "reqAddr": "127.0.0.1:1234", "error": "dial error: dial tcp4 127.0.0.1: 1234: connect: connection refused"} 2023-09-03T14:11:24Z ERROR TCP transparent proxy error {"addr": "123.96.20.227:40558", "reqAddr": "127.0.0.1:1234", "error": "dial error: dial tcp4 127.0.0.1: 1234: connect: connection refused"} server's ACL causes the issue.solved 服务器的 ACL 导致该问题。已解决 你是怎么解决的?我也有同样的错误~
gharchive/issue
2023-09-03T14:28:01
2025-04-01T04:56:00.591176
{ "authors": [ "AmberisMyShiba", "sjkhsl" ], "repo": "apernet/hysteria", "url": "https://github.com/apernet/hysteria/issues/656", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
835756783
Not able to use toggleSeries with next js application I want to toggleSeries data of a chart in my next js application. I am trying to use that method in 2 different ways, but none of them is working. I will show you the error I am getting with both these methods differently. Please help me with a solution. Method 1: import ApexCharts from 'apexcharts'; toggleSeries = (seriesName) => { ApexCharts.exec('myChartId', 'toggleSeries', 'seriesName'); } In method 1 there are 2 problems Most of the time I get window is not defined for ApexCharts Second, sometimes I get: apexcharts__WEBPACK_IMPORTED_MODULE_9___default.a.exec is not a function In order to resolve error: window is not defined I use dynamic import as I use dynamic import for react-apexcharts see below, even checking from typeof window !== undefined is not working here Method 2: const ApexCharts = dynamic(() => import('apexcharts'), { ssr: false }); toggleSeries = (seriesName) => { ApexCharts.exec('myChartId', 'toggleSeries', 'seriesName'); } In method 2 it says ApexCharts.exec is not a function. Please help me with the solution for this Basically exec function is coming undefined Can anybody help ??
gharchive/issue
2021-03-19T09:34:31
2025-04-01T04:56:00.626690
{ "authors": [ "hbole" ], "repo": "apexcharts/apexcharts.js", "url": "https://github.com/apexcharts/apexcharts.js/issues/2299", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
451440384
Does apexcharts support this kind of chart, see image in the link? Does apexcharts support this kind of chart, see image in the link, I'm more interested on the top green label (+$6.31) (need to show the difference in amount between this year and last year in numbers) https://imgur.com/a/moLi1Q4 No, ApexCharts directly doesn't support this functionality. Although, you can use annotations and do the calculations to put such custom label anywhere on the chart. https://apexcharts.com/javascript-chart-demos/line-charts/line-chart-annotations/ Docs
gharchive/issue
2019-06-03T11:52:58
2025-04-01T04:56:00.629280
{ "authors": [ "junedchhipa", "visaru" ], "repo": "apexcharts/apexcharts.js", "url": "https://github.com/apexcharts/apexcharts.js/issues/644", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
923992392
'sys' is not defined In linux ubuntu I got the error showed in the image below: Solution: I fixed it adding the following import line in the beginning of the ETL_coletar_dados_e_gravar_BD.py file: import sys Thank you @jeff-pal ! Solved by this commit: https://github.com/aphonsoar/Receita_Federal_do_Brasil_-_Dados_Publicos_CNPJ/commit/95fbb3bbb2d474053793caaa08dc4629d7c28871
gharchive/issue
2021-06-17T14:28:24
2025-04-01T04:56:00.631536
{ "authors": [ "aphonsoar", "jeff-pal" ], "repo": "aphonsoar/Receita_Federal_do_Brasil_-_Dados_Publicos_CNPJ", "url": "https://github.com/aphonsoar/Receita_Federal_do_Brasil_-_Dados_Publicos_CNPJ/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2015241145
Remove stopDataFetcher Closes https://github.com/api3dao/airseeker-v2/issues/129 If you remove the stop function you probably don't need the interval stored in state either, but again, super minor Yeah, I removed that in some later PR but I could have done it here.
gharchive/pull-request
2023-11-28T20:24:45
2025-04-01T04:56:00.647164
{ "authors": [ "Siegrift" ], "repo": "api3dao/airseeker-v2", "url": "https://github.com/api3dao/airseeker-v2/pull/133", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
281035263
Request headers not rendered This is my code: ### Set/Update customer info [PUT /xxx/customer/{id}] Update the personal info of the customer. None of the fields are mandatory. The fields not specified will not be overwritten in the database. + Parameters + id (string) - Id of the customer + Request + Headers Authentication: Bearer SomeTokenValue + Attributes (object) + email (string, optional) - email address of the customer + password (string, optional) - email address of the customer + firstName (string, optional) - first name + lastName (string, optional) - last name + phoneNumber (string, optional) - phone number + receiveOffers (number, optional) - 1 if the customer wants to receive offers, 0 otherwise + Body { "email":"example@domain.com", "firstName":"Jos", "lastName":"Bos", "phoneNumber":"+32488697059", "receiveOffers":1 } This is what the result looks like Why is the request header not shown? The request header is shown inside your code example (on line 7). If you don't want to view a code example you can switch to the default (Raw) to see the raw request information (headers and request body). But why is it separate from the parameters and the attributes? That makes no sense to me. You can also say that the request parameters are shown to inside the code example, but doesn't it make more sense to have all variables grouped together? Why put url parameters at the top, but http headers in another section?
gharchive/issue
2017-12-11T14:27:55
2025-04-01T04:56:00.650168
{ "authors": [ "ir-fuel", "kylef" ], "repo": "apiaryio/api-blueprint", "url": "https://github.com/apiaryio/api-blueprint/issues/410", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
219690424
Updated NPM_VERSION for nodejs 4 and 6 Please do not merge it yet. I want to properly test it in core app and update readme in coreapp. @freaz fine by me; if you want to have both, you can add an extra folder to create a new tag (e.g. nodejs/6-npm4-4 or something less ugly...) No I want have just one version, but the npm update was side effect and I wanted to test it at least locally.
gharchive/pull-request
2017-04-05T19:30:27
2025-04-01T04:56:00.653261
{ "authors": [ "freaz", "nadade" ], "repo": "apiaryio/docker-base-images", "url": "https://github.com/apiaryio/docker-base-images/pull/108", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
24336903
Report possible keywords For example in # Resource [/] + Model + Body (application/json) { ... } The + Body (application/json) is not recognized as a body (content type should be defined on the model level) and thus the whole list item is considered to be model's description. Migrated to apiaryio/drafter#645
gharchive/issue
2013-12-16T09:05:44
2025-04-01T04:56:00.658068
{ "authors": [ "abtris", "zdne" ], "repo": "apiaryio/snowcrash", "url": "https://github.com/apiaryio/snowcrash/issues/65", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
495411657
Integrate adblocker functionality Interesting tip from HN (for Dashblock): Maybe you already do it, but I think integrating adblocker functionality when loading JS sites would be desirable to reduce load time. And if ads are what the API user is interested in, perhaps add a flag for whether or not one wants ads to load. Recommendation: https://github.com/cliqz-oss/adblocker Should be the fastest adblocker library (used by Ghostery, Cliqz and Brave) This could be integrated into Apify.launchPuppeteer() function as useAdBlock: true option. https://sdk.apify.com/docs/api/apify#module_Apify.launchPuppeteer Greetings. So the thing would be to implement ad blocker to increase the speed of the scrap/crawl? I could work on this 🙏 Yes exactly, it could boost the speed especially for some websites that are heavy on ads (news sites). But it would be great to first test this assumption. Would you be interested also in trying this out? Use Apify SDK to run scraper with and without ad blocker against some websites? Sure! I can set up a test and run it to check this first with some timing debug, I'll create it and run it, then attach it here for you to see, thank you 🚀 interesting. I manually block all the common ad networks using blockRequests, this would offload the task to the extension Makes sense for a lot of users I guess but fyi it's an explicit anti-feature with usecase-killing effect for me. I'd need this off with zero sideeffects on current behavior. Makes sense for a lot of users I guess but fyi it's an explicit anti-feature with usecase-killing effect for me. I'd need this off with zero sideeffects on current behavior. In the small POC I proposed a while ago https://github.com/apify/apify-js/pull/600, the feature is completely disabled by default and only does some work when blocking is enabled by the user. Yeah, sorry @remusao . We still have not figured out if the performance will improve or not. I apologize. Yeah, sorry @remusao . We still have not figured out if the performance will improve or not. I apologize. Of course, no worries at all, I just wanted to make clear to @matjaeck that there should be a way to integrate such a feature without any overhead when it's disabled.
gharchive/issue
2019-09-18T19:25:24
2025-04-01T04:56:00.665036
{ "authors": [ "Darking360", "jakubbalada", "matjaeck", "mnmkng", "mtrunkat", "pocesar", "remusao" ], "repo": "apify/apify-js", "url": "https://github.com/apify/apify-js/issues/456", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1376867102
🛑 Harbor is down In dbf72db, Harbor (https://registry.openstorage.xyz) was down: HTTP code: 521 Response time: 270 ms Resolved: Harbor is back up in e07c713.
gharchive/issue
2022-09-17T20:31:36
2025-04-01T04:56:00.687098
{ "authors": [ "apinter" ], "repo": "apinter/OSC-mon", "url": "https://github.com/apinter/OSC-mon/issues/28", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
451777372
optimizations for generated target code the generated target code is rather verbose right now. for example, here is the spec program and output for the "dangerous" millionaires game: spec: /* the classic MPC problem */ a : {A}; c : {C}; c_richer : {(A|C)-> & (A&C)<-}; c_richer := declassify(a <= c, {(A|C)-> & (A&C)<-}); /* the dangerous version: if you have less money, you lose all of it */ if (c_richer) { a := 0; } else { c := 0; } generated target code: MPC({alice,chuck})_1 { c_1 <- recv chuck; a_1 <- recv alice; downgrade_1_1 := (a_1 <= c_1); send downgrade_1_1 to alice; send downgrade_1_1 to chuck; } alice { a : {⊥<-}; c_richer : {⊥<-}; send a to MPC({alice,chuck})_1; downgrade_1_2 <- recv MPC({alice,chuck})_1; assgn_1_1 := downgrade_1_2; c_richer := assgn_1_1; c_richer_1 := c_richer; guard_1_1 := c_richer_1; guard_1_4 := guard_1_1; if (guard_1_4) { assgn_2_1 := 0; a := assgn_2_1; a <- recv chuck; } else { assgn_3_1 := 0; send assgn_3_1 to chuck; } } chuck { c : {⊥<-}; c_richer : {⊥<-}; send c to MPC({alice,chuck})_1; downgrade_1_3 <- recv MPC({alice,chuck})_1; assgn_1_2 := downgrade_1_3; c_richer := assgn_1_2; c_richer_2 := c_richer; guard_1_2 := c_richer_2; guard_1_3 := guard_1_2; if (guard_1_3) { assgn_2_2 := 0; send assgn_2_2 to alice; } else { c <- recv alice; assgn_3_2 := 0; c := assgn_3_2; } } as you can see, there are a lot of unnecessary variables. we can run copy propagation, dead code elimination, and constant propagation to remove most of these and make the target a lot readable. copy prop implemented in https://github.com/apl-cornell/viaduct/commit/347d4cac7362e8f16c3bd132486accea58f27348 constant prop not yet implemented constant prop implemented in e3ccfd719801c6924b899335289b35280e6d2e6f
gharchive/issue
2019-06-04T03:12:52
2025-04-01T04:56:00.697614
{ "authors": [ "rolph-recto" ], "repo": "apl-cornell/viaduct", "url": "https://github.com/apl-cornell/viaduct/issues/27", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
305475457
Correct license name Usually "Apache License" means version 1, and according to other files, this library actually uses version 2. Looks good, thanks.
gharchive/pull-request
2018-03-15T09:35:42
2025-04-01T04:56:00.698827
{ "authors": [ "dlackty", "paolorotolo" ], "repo": "apl-devs/AppIntro", "url": "https://github.com/apl-devs/AppIntro/pull/524", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
367906094
Deprecating NGRX Cache If someone is interested to maintain it, please let me know. The whole idea behind that package and using ngrx with Apollo in general, makes no sense to me. If anyone has any objections please comment here. The reason to use ngrx is same as for using any other redux implementation. In particular using ngrx-store with angular-apollomakes sense in case of some part of an app would need to use ngrx/store and to avoid having two different redux implementation in the app (to my understanding that was general idea behind making new apollo client independent on redux implementation). As for me for now I would like to relay fully on Apollo and hope to have no need to access directly underlying redux implementation. What having a normalized cache inside of ngrx gives you? Do you use it inside ngrx as part of selectors? Having one single store that is a source of truth makes sense only if one data relies on another, if they are totally independent you can have multiple of them. Main reason would be the size of an app and avoid pulling in one more dependency (well set of deps) Also, when you're using ngrx/router-module in combination with apollo graphql that package is rather useful. I was just exploring this library, here's what I thought the library was doing, maybe that'll be helpful. Typically, most usages of ngrx store will look something like: Ask to load some data using an action Have an effect do the actual loading, and dispatch an action when the data has been loaded Have a reducer update the store with the loaded data (and more specifically, resolve updating v.s. adding to lists, e.g. using ngrx entity) Use a selector to get the updated data, if it was updated Since with Apollo and GraphQL it's possible to identify entities, it would be possible to normalize the returned data in a store any time a query is made. Then I can simply use a selector to get it from the "cache". It would remove a lot of boilerplate for most use cases, and for any use cases where the store has more complex logic, the apollo ngrx store would only be used as a cache. I agree that using ngrx store as a simple cache does not make a lot of sense by itself, but making it easier to update complex models in a store automatically could be very useful, at least in simpler apps. This however would require "breaking down" the response in separate Entity lists, and to be useful you'd have to map these entities to a type (e.g. using graphql-code-generator). Then it would look something like: Same Same, except my service would use a graphqlquery, AND it would update the apollo store. No reducer required (except if you need to do something fancy, you can alway use the response and do something with it but it's not necessary) Use a selector to get the data (and it could be possible to generate those too for most cases, like "selected entry in a list", "all items in a list", etc.) Hope that helps! If someone wants to take care of apollo-angular-cache-ngrx then feel free to message me, do pull requests, I will merge all of that, right away. Since I don't use it, I see no point of me maintaining it. @kamilkisiela what's your suggested apollo, angular local state management stack? Just a few thoughts: if you use just apollo client as ngrx replacement, how do you implement effects? If you use apollo-client and ngrx than you have decent router and effects support but you have to transform the graphql responses into the ngrx store. Is there any approach I am missing which solves these disadvantages? IMO you should just use apollo for state management. What would be the case where you may need to access store directly? effect in Apollo are just GraphQL queries/mutations and if you need i.e. access geolocation then you still can do this with @client resolvers @ciekawy interesting approach. But isn't the idea of redux/ngrx to separate pure reducers from effects with side effects. With the approach you mentioned, it feels a bit like these two concepts are not clearly separated anymore. Apollo is supposed to fully take over the state management. I moved to Apollo with similar concerns initially and now I am using only Apollo. Biggest benefits for me: no more need for verbose boilerplate of reducers and effect. Initially I really liked the concept of redux for state management. Now I love Apollo even more :) there are various integrations available for non-graphql data sources @client for anything client-side, @rest for REST endpoints you cannot move to GraphQL, also you can define your own as well. here you can read more https://blog.apollographql.com/the-future-of-state-management-dd410864cae2 another important value is that Apollo do also data normalization when keeping it in local store Thx @ciekawy I will dig into that one more time. Actually I thought that loonajs will be the solution to fill the gap between the redux approach and apollojs/graphql. But atm it seems like that the guild also lost its confidence in the loonajs approach. I would be really interested what the guild (@kamilkisiela, @urigo,...) is suggesting as best approach. A usecase I have come across is about synchronisation of cache/store. I have a dumb component that emits two events in certain cases on a single browser UI event. The handler for first event mutates the client state, then second handler picks that state and performs server mutation. With ngrx store, the store update by first handler by action dispatch appears synchronous, and the effect for action from second handler gets the updated state. However, with apollo, the client only mutation is not synchronous, possibly because it uses promise api internally. Therefore, the mutation from second handler does not get updated data from cache. If someone wants to maintain NGRX cache for Apollo I would be happy to see a fork so I could give an access to the package on npm.
gharchive/issue
2018-10-08T18:35:11
2025-04-01T04:56:00.721749
{ "authors": [ "ceelian", "christianrondeau", "ciekawy", "kamilkisiela", "wesselvdv", "wolfvik" ], "repo": "apollographql/apollo-angular", "url": "https://github.com/apollographql/apollo-angular/issues/886", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1502859269
Add test for sdl field This test will check that the sdl field inside the _service object type has the right type (non-nullable string). Apollo Studio and potentially other Apollo tools check this field to enable/disable federation features. TODO: [ ] Add directives @dariuszkuc I wonder if we should make a new table for directives (which I haven't added just yet), my main issue with this test is that it is hard to see what fails, since it stops at the first failure @patrick91 if this is still an issue feel free to re-open this PR
gharchive/pull-request
2022-12-19T12:18:33
2025-04-01T04:56:00.736524
{ "authors": [ "dariuszkuc", "patrick91" ], "repo": "apollographql/apollo-federation-subgraph-compatibility", "url": "https://github.com/apollographql/apollo-federation-subgraph-compatibility/pull/302", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
328946917
Can schema stitching be implemented in Python ? Sorry, that wouldn't be in this library.
gharchive/issue
2018-06-04T07:58:41
2025-04-01T04:56:00.752223
{ "authors": [ "s1h1ikha", "stubailo" ], "repo": "apollographql/graphql-tools", "url": "https://github.com/apollographql/graphql-tools/issues/831", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
193114768
make afterware errors bubble up This makes errors thrown from afterwares bubble up so that apollo don't try to parse bad responses. TODO: [ ] If this PR is a new feature, reference an issue where a consensus about the design was reached (not necessary for small changes) [ ] Make sure all of the significant new logic is covered by tests [x] Rebase your changes on master so that they can be merged easily [ ] Make sure all tests and linter rules pass [ ] Update CHANGELOG.md with your change [x] Add your name and email to the AUTHORS file (optional) [ ] If this was a change that affects the external API, update the docs and post a link to the PR in the discussion @edvinerikson Thanks for the PR! Can you explain what problem this solves and add a test case if applicable? Right now it doesn't look like your PR would change any behavior. Looks like the change makes it so that the promise returns the return value of the afterware. However this should definitely come with a test that ensures the behavior continues to work, especially since it's so easy to miss. Yeah I will add a test. I just need to fix my local setup first. Can't run it locally due to TS errors. On Fri, 2 Dec 2016 at 23:55, Sashko Stubailo notifications@github.com wrote: Looks like the change makes it so that the promise returns the return value of the afterware. However this should definitely come with a test that ensures the behavior continues to work, especially since it's so easy to miss. — You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub https://github.com/apollostack/apollo-client/pull/982#issuecomment-264584646, or mute the thread https://github.com/notifications/unsubscribe-auth/ADtdjIkVxD0gvTIVN8hR7lwS3Y2vvHcNks5rEKHHgaJpZM4LCi3H . @helfer the reason for the change is because I have a afterware which I check the response status and decide if Apollo will be able to parse the response. in some cases Apollo won't be able to do that and when that happens I throw a error which I can catch in my error handling but without this change I am not able to catch the error and instead Apollo will fail with when trying to parse the json body. e.g getDataFromTree(app).catch(error => doSomethingWithErrorHere()) but currently the error I get there will be something like unexpected token < at 1:1 (due to a html response) instead of my own error. maybe a improvement to the network layer would be to check response.ok before trying to parse / handle the response as well. will probably help fix some of the cryptic error messages that I get sometimes. @edvinerikson Okay. Please add a test, and then we can merge it. I think checking response.ok won't work, because some people return error codes that are not in the 200 range for responses that should be parsed and processed. It's unfortunate, but I think we have to live with that for now. I think checking response.ok won't work, because some people return error codes that are not in the 200 range for responses that should be parsed and processed. It's unfortunate, but I think we have to live with that for now. Yeah I will create my own network layer to handle my cases instead. Still gonna try to add a test and get this merged. test is added Sweet, thanks a lot!
gharchive/pull-request
2016-12-02T13:21:12
2025-04-01T04:56:00.788120
{ "authors": [ "edvinerikson", "helfer", "stubailo" ], "repo": "apollostack/apollo-client", "url": "https://github.com/apollostack/apollo-client/pull/982", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1781998268
Add Whatsapp Icon Description Add a WhatsApp icon to connect with the admin on WhatsApp Screenshots No response Additional information No response @apoorvaron Kindly assign me this issue @Nishitbaria assigned Now start solving issues😂 @apoorvaron I have done this one already i was just uploading video https://github.com/apoorvaron/Shorty/assets/85815172/34d7e252-f599-4aed-a0cc-082196ee704e @apoorvaron Since its Friday so last working day in my job and very less load so trying to create issue and finding some raw solution then full working from tomorrow https://github.com/apoorvaron/Shorty/assets/85815172/c4a3c7ff-9f81-462d-85c5-ae6009b4aa65 @apoorvaron Thanks Man
gharchive/issue
2023-06-30T06:52:46
2025-04-01T04:56:00.796114
{ "authors": [ "Nishitbaria", "apoorvaron" ], "repo": "apoorvaron/Shorty", "url": "https://github.com/apoorvaron/Shorty/issues/190", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1051664363
feat: apply stopwords by language What does this do / why do we need it? https://www.notion.so/appbase/Arc-Suggestions-API-enhancements-2a4c26df77d44a77b751d9e13cf4cb0f What should your reviewer look out for in this PR? Which issue(s) does this PR fix? If this PR affects any API reference documentation, please share the updated endpoint references @siddharthlatest Docs PR https://github.com/appbaseio/Docs/pull/246 If searchLanguage isn't explicitly passed in RS API, will this take the search relevance's language for stopwords? It should be this way. Yes, review this PR https://github.com/appbaseio-confidential/arc-noss/pull/230
gharchive/pull-request
2021-11-12T07:32:38
2025-04-01T04:56:00.811613
{ "authors": [ "bietkul" ], "repo": "appbaseio/reactivesearch-api", "url": "https://github.com/appbaseio/reactivesearch-api/pull/213", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
173249897
Kafka is not reachable from amplifier Amplifier is connecting to kafka through localhost:9092. However, this hostname is used only for the first connection to kafka. On this first call, kafka gives its own hostname to the client (amplifier). The problem is that the hostname given by kafka (which is actually the kafka container short id) is not reachable from amplifier because it's not part of amp-swarm network. Possible fixes: Running amplifier as a container and make it part of the amp-swarm network using docker run , however services and containers cannot be part of the same network. docker run --network amp-swarm appcelerator/amp docker: Error response from daemon: swarm-scoped network (amp-swarm) is not compatible with `docker create` or `docker run`. This network can only be used by a docker service. See 'docker run --help'. Giving the service a hostname, however, I don't think it's possible: docker service create --help Usage: docker service create [OPTIONS] IMAGE [COMMAND] [ARG...] Create a new service Options: --constraint value Placement constraints (default []) --container-label value Container labels (default []) --endpoint-mode string Endpoint mode (vip or dnsrr) -e, --env value Set environment variables (default []) --help Print usage -l, --label value Service labels (default []) --limit-cpu value Limit CPUs (default 0.000) --limit-memory value Limit Memory (default 0 B) --log-driver string Logging driver for service --log-opt value Logging driver options (default []) --mode string Service mode (replicated or global) (default "replicated") --mount value Attach a mount to the service --name string Service name --network value Network attachments (default []) -p, --publish value Publish a port as a node port (default []) --replicas value Number of tasks (default none) --reserve-cpu value Reserve CPUs (default 0.000) --reserve-memory value Reserve Memory (default 0 B) --restart-condition string Restart when condition is met (none, on-failure, or any) --restart-delay value Delay between restart attempts (default none) --restart-max-attempts value Maximum number of restarts before giving up (default none) --restart-window value Window used to evaluate the restart policy (default none) --stop-grace-period value Time to wait before force killing a container (default none) --update-delay duration Delay between updates --update-failure-action string Action on update failure (pause|continue) (default "pause") --update-parallelism uint Maximum number of tasks updated simultaneously (0 to update all at once) (default 1) -u, --user string Username or UID --with-registry-auth Send registry authentication details to swarm agents -w, --workdir string Working directory inside the container Running amplifier as a service: It's definitely possible, however, it's not compatible with development where you need to be able to quickly launch, fix, and repeat. Indeed, docker services are not using local images but instead pull images from the hub (or a private registry) on each start (check this issue https://github.com/docker/swarmkit/issues/1246). It would require to push, build and publish the image every time which defeats the purpose. The only workaround I found so far is to add kafka container hostname to /etc/host: Getting kafka hostname: docker ps docker inspect kafka_container_id | grep hostname sudo cat /var/lib/docker/containers/a4424bcafbab110b67922994243aca15fc57ab3422fdec21f7460fe16b024456/hostname Update /etc/hosts in order to assign 127.0.0.1 to kafka hostname: 127.0.0.1 localhost a4424bcafbab Unfortunately the following doesn't work for Docker for Mac: sudo cat /var/lib/docker/containers/a4424bcafbab110b67922994243aca15fc57ab3422fdec21f7460fe16b024456/hostname On Docker for Mac these files are stored in the VM located here: $HOME/Library/Containers/com.docker.docker/Data/com.docker.driver.amd64-linux/Docker.qcow2 But since it looks like the hostname is just the short-id anyway, I think this will work (at least for our temporary workaround right now): docker ps --filter "name=kafka\." --format '{{ .ID }}' (The funky filter is to eliminate kafka-manager from the result.) I'll test that out and report my results back here shortly. We need a better interim strategy; asking everyone to help brainstorm on this. It seems like advertised.listeners might be what we're looking for in the Broker Configs docs... @bertrand-quenin @ndegory Can you two investigate further? Regarding my last comment, I don't think the issue is advertising the hostname, it's making sure the hostname is available outside the swarm's network when we're developing ... I suppose this is where kafka updates and then gets the server address for subsequent requests on the client broker: https://github.com/Shopify/sarama/blob/master/broker.go#L83 https://github.com/Shopify/sarama/blob/master/broker.go#L161 @subfuzion Actually, I already tried this solution. It effectively makes kafka reachable from amplifier by advertising localhost. The problem is that other components (telegraf, amp-log-worker) are trying to reach kafka on localhost as well ... making them fail miserably. @bertrand-quenin Unfortunately, that makes perfect sense. We may have to create a script that can take the output of docker ps --filter "name=kafka\." --format '{{ .ID }}' and update the entry for 127.0.0.1 in /etc/hosts. We can call the script from the swarm script when we start up the swarm. @ndegory, we're all slammed preparing for our deadline on Monday; is there any way you can take this on for us? I'll take a look @bertrand-quenin Now that I think about it, can't we override the address in the client after it gets updated with the server's metadata, just to ensure we're using the address we want to use? After all, in the real swarm we would want to use the actual DNS name (kafka) anyway since swarm is load balancing for us. There isn't some special requirement for affinity to a particular server in the cluster, is there? Thanks for the PR, @ndegory. It's working well for now. I'll leave this issue open while we consider our long term strategy for this and similar/related issues. There is another solution which do not imply modify /etc/host. If we have amplifier inside swarm network it works as explains above, but then the issue that we can't build locally an amplifier image to test. Except, for the test need, we modify the amplifier image name 'appcelerator/amp' to 'appcelerator/amp_fr (for instance for me) and modify the swarm file to point on this image name instead of the normal. Then, all work, docker see that appcelerator/amp_fr doesn't exist in repo and use the local image @bertrand-quenin Can you respond to https://github.com/appcelerator/amp/issues/69#issuecomment-242868643? Yes, it's an alternate solution to this issue. It can work as well. @freignat91 @bertrand-quenin @ndegory Since I prefer that we switch to NATS, I'm not particularly concerned about the long term resolution for this particular issue at this point. However, it is definitely the case that amplifier will eventually expose its services on an overlay network (for example, amp-swarm-public), and it will also overlap with the backbone network (amp-swarm) to communicate with our infrastructure services. @freignat91, we should start moving toward this configuration this sprint, and also start incorporating haproxy for mapping exposed IPs to consumers (eg, the CLI). It's interesting to be aware of how clusters and node discovery impact other services for client communication as well. For example, the etcd client when etcd is behind a load-balanced cluster: https://github.com/hashicorp/vault/pull/921. Since we have a working solution in place for now thanks to @ndegory and our priority has shifted to transitioning to NATS (#113), I'm going to close this.
gharchive/issue
2016-08-25T16:27:14
2025-04-01T04:56:00.827026
{ "authors": [ "bertrand-quenin", "freignat91", "ndegory", "subfuzion" ], "repo": "appcelerator/amp", "url": "https://github.com/appcelerator/amp/issues/69", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
177344009
Default retention policy for InfluxDB 1.0.0 Hi, I found an issue and came up with a way to fix it. The issue: InfluxDB changed the name of the default retention policy in 1.0.0 from "default" to "autogen". The telegraf.conf.tpl file needs to be updated. Proposed fix: To make this work and be backwards compatible I created an env variable for the policy name. I have it defaulting to "default". It could default to "", which is what the telegraf 1.0.0 telegraf.conf file has. To make it clear how to use it I added it to the custom example at the top of the readme. My pleasure.
gharchive/pull-request
2016-09-16T04:41:14
2025-04-01T04:56:00.830089
{ "authors": [ "atribe" ], "repo": "appcelerator/docker-telegraf", "url": "https://github.com/appcelerator/docker-telegraf/pull/16", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
294630606
How to test that a button is disabled On Android, I have disabled a button being able to get clicked until a certain condition is met within the input. (Python client) The object currently has clickable="false" in the XML but the appium driver tells me different, because - button.is_enabled() returns True In the case, you can get an arbitrary element with el.get_attribute('clickable'). https://github.com/appium/python-client/blob/f37733dbd3dd79be97452a9ad98eb4a9325755ce/test/functional/ios/find_by_ios_class_chain_tests.py#L32 If your xml says the element is disable, you can see false by el.get_attribute('clickable'). BTW, enable and clickable are https://stackoverflow.com/questions/15615823/setenabled-vs-setclickable-what-is-the-difference If you have any issue, the page source and logs are helpful for us.
gharchive/issue
2018-02-06T04:42:56
2025-04-01T04:56:00.969169
{ "authors": [ "Enether", "KazuCocoa" ], "repo": "appium/sample-code", "url": "https://github.com/appium/sample-code/issues/149", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2303061671
Generalize the removal of compiler argument options during indexing The existing ad-hoc logic was not quite correct because it didn’t eg. remove -MT/depfile because it assumed that -MT was followed by a space. It also didn’t take into account that serialize-diagnostics can be spelled with a single dash or two dashes. Create a CompilerCommandLineOption type that forces decisions to be made about the dash spelling and argument styles, which should help avoid problems like this in the future. @swift-ci Please test @swift-ci Please test @swift-ci Please test Windows @swift-ci Please test Windows @swift-ci Please test Windows
gharchive/pull-request
2024-05-17T15:46:28
2025-04-01T04:56:01.018500
{ "authors": [ "ahoppen" ], "repo": "apple/sourcekit-lsp", "url": "https://github.com/apple/sourcekit-lsp/pull/1314", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
593757420
TestURLSession: Re-enable tests Re-enable the TestURLSession tests which should no longer be flaky however specifc tests are still disabled due to known issues. Disable test_concurrentRequests() as this fails about 4% of the time. Disable test_getAllTasks() as this causes other tests to fail. @swift-ci test linux @swift-ci test linux @swift-ci test linux
gharchive/pull-request
2020-04-04T06:34:03
2025-04-01T04:56:01.020835
{ "authors": [ "spevans" ], "repo": "apple/swift-corelibs-foundation", "url": "https://github.com/apple/swift-corelibs-foundation/pull/2757", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1260346553
Convert DowningStrategyShell into a distributed actor Resolves https://github.com/apple/swift-distributed-actors/issues/935 Only one thing about the logger: https://github.com/apple/swift-distributed-actors/pull/938#discussion_r889430590 otherwise ready to go, thank you!
gharchive/pull-request
2022-06-03T20:29:02
2025-04-01T04:56:01.022376
{ "authors": [ "ktoso", "yim-lee" ], "repo": "apple/swift-distributed-actors", "url": "https://github.com/apple/swift-distributed-actors/pull/938", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1648317524
Prevent scrolling on navigation event if router meta field preventScrolling is true Bug/issue #, if applicable: rdar://103146273 Summary If preventScrolling in the meta field is set to true, then we should not change the scroller location when navigating to a new page. Dependencies rdar://103119754 Testing Steps: set preventScrolling to true for a page (i.e. '/foo') Scroll down a little bit on any page (i.e. '/bar'). ('/foo' should be longer than '/bar') trigger a navigation event to '/foo' Verify that the scroller did not move up/down nor left/right Checklist Make sure you check off the following items. If they cannot be completed, provide a reason. [x] Added tests [x] Ran npm test, and it succeeded [ ] Updated documentation if necessary - NA @swift-ci test
gharchive/pull-request
2023-03-30T21:03:49
2025-04-01T04:56:01.026686
{ "authors": [ "hqhhuang" ], "repo": "apple/swift-docc-render", "url": "https://github.com/apple/swift-docc-render/pull/563", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
786444435
[Collections] add should only add to config if fetch is successful Motivation: The code currently adds a collection source to the config even when initial fetch fails. Modifications: We should fetch collection first and make sure that's successful before adding to config. Result: Bad sources are not added to the config. @swift-ci please smoke test @swift-ci please smoke test @swift-ci please smoke test @swift-ci please smoke test @swift-ci please smoke test @swift-ci please smoke test @swift-ci please smoke test @swift-ci please smoke test
gharchive/pull-request
2021-01-15T00:41:58
2025-04-01T04:56:01.029839
{ "authors": [ "abertelrud", "yim-lee" ], "repo": "apple/swift-package-manager", "url": "https://github.com/apple/swift-package-manager/pull/3195", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1518715665
Remove reference to AsTypePattern Companion of https://github.com/apple/swift-syntax/pull/1178. https://github.com/apple/swift-syntax/pull/1178 @swift-ci Please test
gharchive/pull-request
2023-01-04T10:26:23
2025-04-01T04:56:01.031556
{ "authors": [ "ahoppen" ], "repo": "apple/swift-stress-tester", "url": "https://github.com/apple/swift-stress-tester/pull/219", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
506920895
[SyntaxParse] Revert syntax parsing changes from the master branch Update for https://github.com/apple/swift/pull/27675 @swift-ci Please test
gharchive/pull-request
2019-10-14T23:25:21
2025-04-01T04:56:01.032635
{ "authors": [ "rintaro" ], "repo": "apple/swift-syntax", "url": "https://github.com/apple/swift-syntax/pull/156", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
358221711
Fix RawSyntax's leading/trailingTrivia computation to return nil if the outermost children don't have leading/trailing trivia Previously they would skip over outer children that didn't have leading/trailing trivia until they found one that did. This was causing those trivia to be included twice when computing the totalLength/byteSize of the node; once in leading/trailingTrivia and again in contentLength, which only skipped the trivia of the outermost children. @swift-ci please test Good catch. Thanks. We should also cherry-pick this over to swift-4.2-branch once #6 is merged.
gharchive/pull-request
2018-09-07T21:48:49
2025-04-01T04:56:01.034343
{ "authors": [ "ahoppen", "nathawes" ], "repo": "apple/swift-syntax", "url": "https://github.com/apple/swift-syntax/pull/8", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
675802806
build: ignore deprecation warnings on Windows Adjust the compile flags for TSCBasic on Windows. Avoid the separate checks and addition of flags in favour of generator expressions. CC: @tomerd @neonichu @abertelrud @swift-ci please test do we have windows CI for this? Not yet, but the nighters do build t-s-c, as we bring up more pieces we can enable them. e.g. https://dev.azure.com/compnerd/swift-build/_build/results?buildId=35571&view=logs&s=4a582161-ef84-5ffa-1ff0-506a7473f086&j=73ce2acc-4df7-56ec-1894-6d2e45aae46f
gharchive/pull-request
2020-08-10T00:30:34
2025-04-01T04:56:01.036788
{ "authors": [ "compnerd", "tomerd" ], "repo": "apple/swift-tools-support-core", "url": "https://github.com/apple/swift-tools-support-core/pull/102", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1214380541
[SR-15154] ObjC projects using SPM fail when we expose our models into another module Previous ID SR-15154 Radar rdar://problem/82827693 Original Reporter taquitos (JIRA User) Type Bug Attachment: Download Additional Detail from JIRA Votes 3 Component/s Labels Bug Assignee None Priority Medium md5: 4659038e109b6bed99ec521e272addf0 Issue Description: Setup: We create an ObjC App that uses our SPM package "RevenueCat". In our app, we create an object that uses an object from the "RevenueCat" package as part of its api. We expose this object's header (which exposes the method containing the RevenueCat object as a parameter) as a public header. We create a Swift test target that depends on the ObjC App and has a bridging header for that object we created. We attempt to use the API in a test. API that doesn't include RevenueCat objects is fine, but any API that uses RevenueCat objects in the parameter is not found. Environment: Xcode 12.5.1 (Build version 12E507) & Xcode 13 Beta 5 M1 Mac as well as Intel mac. Swift Package Manager - Swift 5.4.0 Swift Package Manager - Swift 5.5.0 Component: Package Manager Reproduction project: https://github.com/taquitos/SPMBug Attempt to build the test target. You'll get `Value of type 'TaquitosAreTheBest' has no member 'purchaserInfo'` I have faced similar issue. This bug is also reproduces in mixed Swift/Objective-C project (not only in Test target). Case: header file has forward class declaration from SPM and this type used for declare properties or parameters in this header. Those properties and parameters won't be exposed into Swift. You will receive build error: 👋 We have encountered this same issue while converting the Braintree SDK from Obj-C to Swift. Our SDK is comprised of several sub-modules, and while we thought we could convert to Swift module-by-module, this SPM bug has us blocked and forced to convert our entire SDK at once. Our team has created a Swift Forum post on the issue here, for those that want to follow along / try to gain traction. cc @tomerd @tomerd Please help me transfer this if it is indeed a package manager issue. cc @abertelrud @neonichu Sounds like an Xcode issue to me.
gharchive/issue
2021-09-04T00:29:02
2025-04-01T04:56:01.047041
{ "authors": [ "AnthonyLatsis", "neonichu", "oleksandryevdokymov", "scannillo", "swift-ci", "tomerd" ], "repo": "apple/swift", "url": "https://github.com/apple/swift/issues/57480", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1529777034
Partial block result builder fails to pick correct overload and generates compiler error Description Partial block result builder fails to pick correct overload and generates a compiler error Ambiguous use of 'parser(of:)' Steps to reproduce Paste the following code into a new project (This issue is a reduction of trying to compile an example (final 4 lines of the code below) of the swift-parsing package) //MARK: Parsers @rethrows public protocol Parser<Input> { associatedtype Input } extension String: Parser { public typealias Input = Substring } extension Int { static func parser( of inputType: Substring.Type = Substring.self ) -> FromSubstringToUTF8<IntParser<Substring.UTF8View>> { FromSubstringToUTF8 { IntParser<Substring.UTF8View>() } } static func parser( of inputType: Substring.UTF8View.Type = Substring.UTF8View.self ) -> IntParser<Substring.UTF8View> { .init() } } struct FromSubstringToUTF8<P: Parser>: Parser where P.Input == Substring.UTF8View { typealias Input = Substring let parser: P init(@ParserBuilder _ build: () -> P) { self.parser = build() } } struct IntParser<Input: Collection>: Parser where Input.SubSequence == Input, Input.Element == UTF8.CodeUnit { public init() { } } //MARK: ParserBuilder struct Parse<Parsers: Parser>: Parser { typealias Input = Parsers.Input let parsers: Parsers init(@ParserBuilder with build: () -> Parsers) { self.parsers = build() } } @resultBuilder struct ParserBuilder { public static func buildPartialBlock<P: Parser>(first: P) -> P { first } public static func buildPartialBlock<P0, P1>(accumulated: P0, next: P1) -> SkipFirst<P0, P1> { .init(p0: accumulated, p1: next) } } struct SkipFirst<P0: Parser, P1: Parser>: Parser where P0.Input == P1.Input { typealias Input = P0.Input let p0: P0 let p1: P1 } //MARK: main let first = ParserBuilder.buildPartialBlock(first: ",") let second = ParserBuilder.buildPartialBlock(accumulated: first, next: Int.parser()) let parser = Parse { "," Int.parser() // Ambiguous use of 'parser(of:)' } Expected behavior This compiles for toolchain 2023-01-02 but breaks from 2023-01-07. Expected to compile and choose the correct overload (in this case static func parser(of inputType: Substring.Type = Substring.self)) Since the first parser passed to the resultBuilder closure is of type String therefore the Parser associated type Input is of type Substring. Therefore the compiler should be able to infer that the parser returned by Int.parser() should have an Input of Substring as well. Calling the individual partialBuildBlock functions does compile works: let first = ParserBuilder.buildPartialBlock(first: ",") let second = ParserBuilder.buildPartialBlock(accumulated: first, next: Int.parser()) fails: let parser = Parse { "," Int.parser() // Ambiguous use of 'parser(of:)' } Environment Swift compiler version info Swift Development Snapshot 2023-01-09 Xcode version info Xcode 13.4.1 Build version 13F100 Deployment target: M1 running macOS 12.5 Ran using the 2023-01-06 snapshot as well and that fails to compile too. @JaapWijnen We have changed implementation to be more strict about how the inference works (aligned with https://github.com/apple/swift-evolution/blob/main/proposals/0348-buildpartialblock.md) and as per proposal result builder is a source transform that works this way: let parser = Parse { let __builder0 = "," let __builder1 = Int.parser() let __builder2 = ParserBuilder.buildPartialBlock(first: __builder0) let __builder3 = ParserBuilder.buildPartialBlock(accumulated: __builder2, next: __builder1) return __builder3 } which means that buildPartialBlock cannot in any way affect type-checking behavior of Int.parser(). The fix here would be to use static func buildExpression<T: Parser>(_ expression: T) -> T { expression } @Xedin Hm I'm not sure I 100% follow. In the process of creating a reproduces I might have minimised my example a bit too much to convey what I'm actually trying to do. Let me know if your answer still applies after my explanation in that case I'm missing how I would have to implement your suggested fix. What is actually happening in my use case is that on the Parser protocol there's a function that takes an inout input, produces an output and in the process modifies that input. The modified input is then passed to the next parser (In this case the Int parser). Therefore there's this dependency that both parsers have the same Input associatedtype. This restriction in this example is implemented by the type constraints on the SkipFirst parser. The first parser is a string and it's defined to have an Input of Substring therefore SkipFirst<String, X> is required to have an Input of Substring which in turn requires X (in this case an IntParser) would be required to have Substring as input. To describe, the full call hierarchy would be as follows: (This writing might be obvious and/or I'm not sure what's the best way of communicating this is) I'll add an updated version of my code below as well.) I'd expect the Parse { } call to return a parser of type: Parse<SkipFirst<String, FromSubstringToUTF8<IntParser<Substring.UTF8View>>>> we can call parse(&input) on that which just passes it on to SkipFirst since that contains a single parser This in turn will call two parsers first p0.parse(&input) this runs removes the "," input and returns Void second p1.parse(&input) // FromSubstringToUTF8<IntParser<Substring.UTF8View>> this second call in turn calls a parser that wraps a parser of a different input. Since the IntParser does not work for an input of SubString as it's requiring Input: Collection where where Input.SubSequence == Input, Input.Element == UTF8.CodeUnit So this parser, FromSubstringToUTF8, transforms the input, calls parse(&transformedInput), modifying it and then transforms the modified input back to the original Input type and overwrites the input variable. Finally propagating the returned value of the IntParser to the FromSubstringToUTF8 to the SkipFirst and finally returning the value. I really hope this is helpful. Let me know if there's information missing/unclear or how I can help in any other way! code: struct ParsingError: Error { } //MARK: Parsers @rethrows public protocol Parser<Input, Output> { associatedtype Input associatedtype Output func parse(input: inout Input) throws -> Output } extension String: Parser { public typealias Input = Substring public func parse(input: inout Substring) throws { guard input.starts(with: self) else { throw ParsingError() } input.removeFirst(self.count) } } extension Int { static func parser( of inputType: Substring.Type = Substring.self ) -> FromSubstringToUTF8<IntParser<Substring.UTF8View>> { FromSubstringToUTF8 { IntParser<Substring.UTF8View>() } } static func parser( of inputType: Substring.UTF8View.Type = Substring.UTF8View.self ) -> IntParser<Substring.UTF8View> { .init() } } struct FromSubstringToUTF8<P: Parser>: Parser where P.Input == Substring.UTF8View { typealias Input = Substring let parser: P init(@ParserBuilder _ build: () -> P) { self.parser = build() } func parse(input: inout Substring) throws -> P.Output { var transformedInput = input.utf8 let result = try parser.parse(input: &transformedInput) input = Substring(transformedInput) return result } } struct IntParser<Input: Collection>: Parser where Input.SubSequence == Input, Input.Element == UTF8.CodeUnit { public init() { } func parse(input: inout Input) throws -> Int { // some int parsing logic return 4 } } //MARK: ParserBuilder struct Parse<Parsers: Parser>: Parser { let parsers: Parsers init(@ParserBuilder with build: () -> Parsers) { self.parsers = build() } func parse(input: inout Parsers.Input) throws -> Parsers.Output { try parsers.parse(input: &input) } } @resultBuilder struct ParserBuilder { public static func buildPartialBlock<P: Parser>(first: P) -> P { first } public static func buildPartialBlock<P0, P1>(accumulated: P0, next: P1) -> SkipFirst<P0, P1> { .init(p0: accumulated, p1: next) } } struct SkipFirst<P0: Parser, P1: Parser>: Parser where P0.Input == P1.Input { let p0: P0 let p1: P1 func parse(input: inout P0.Input) throws -> P1.Output { let _ = try p0.parse(input: &input) return try p1.parse(input: &input) } } //MARK: main let first = ParserBuilder.buildPartialBlock(first: ",") let second = ParserBuilder.buildPartialBlock(accumulated: first, next: Int.parser()) let parser = Parse { "," Int.parser() // Ambiguous use of 'parser(of:)' } var input = ",4"[...] let result = try parser.parse(input: &input) print(result) // 4 I understand what you mean, this kind of workflow is not what is going to be supported by result builder transform implementation going forward. The result builder transform semantics are such that each element in the body is type-checked independently from others and the resulting value is then passed to a final buildBlock or a series of buildPartialBlock calls and returned just like I outlined in my example, the old implementation failed to enforce the "solved independently" bit which caused all sorts of diagnostics and performance issues. In your example there are two overloads of parser(of:) method, both have argument that accepts a default value which means that the type-checker won't be able to disambiguate between them without buildExpression or buildPartialBlock providing more context (via generic requirements) just like if you wrote _ = Int.parser() without using result builders. So does this mean that code previously relying on this feature will fail to compile in a future release? I've seen multiple uses of this in libraries that use resultBuilders to create sort of pipelines which might be an easier example than my above parser example. // a pipeline has an input and output let pipeline = Pipeline { Pipeline1 // input: A, output: B Pipeline2 // input: B, output: C Pipeline3 // input: C, output: D } These types Pipeline1, 2 and 3 are then no longer able to be defined using generics and type restrictions? Is that correct? Or is there a fix for my example that I'm missing at the moment? It’s not about generics or type restrictions, the problem is that implementation did not adhere to the proposed semantics and allowed invalid code to type-check, which would not longer be the same. There are ways to change the result builder itself to archive portion of what was allowed before but cross-element inference would not be supported. Just to clarify you can still use buildParticalBlock to produce a chain you described A -> B -> C -> D but it would not operate on the original elements, the builder transform in your example is going to be: var __builder0 = Pipeline1() var __builder1 = Pipeline2() var __builder2 = Pipeline3() var r0 = Pipeline.buildPartialBlock(first: __builder0) var r1 = Pipeline.buildPartialBlock(accumulated: r0, next: __builder1) var r2 = Pipeline.buildPartialBlock(accumulated: r1, next: __builder2) return r2 not: var r0 = Pipeline.buildPartialBlock(first: Pipeline1()) var r1 = Pipeline.buildPartialBlock(accumulated: r0, next: Pipeline2()) var r2 = Pipeline.buildPartialBlock(accumulated: r1, next: Pipeline3()) return r2 because that violates the code principle behind result builders - type-checking of Pipeline{1, 2, 3} initializers would be affected by buildPartialBlock. If either Pipeline1, Pipeline2 or Pipeline3 initializers are ambiguous all you can do is to use buildExpression to provide additional context. This is exactly how the transform is described in the proposal that I've mentioned. Aha that is very helpful! What I then finally perhaps fail to understand is how to use 'buildExpression' to add context could you elaborate a little bit on that? Sure! static func buildExpression when declared is going to be called for every element, back to your previous example: var __builder0 = Pipeline.buildExpression(Pipeline1()) var __builder1 = Pipeline.buildExpression(Pipeline2()) var __builder2 = Pipeline.buildExpression(Pipeline3()) var r0 = Pipeline.buildPartialBlock(first: __builder0) var r1 = Pipeline.buildPartialBlock(accumulated: r0, next: __builder1) var r2 = Pipeline.buildPartialBlock(accumulated: r1, next: __builder2) return r2 So you could declare that method in Pipeline to have any kind of generic parameters and return type, for example: static func buildExpression<T: SomeProtoocol>(_ expression: T) -> T { expression } This would make sure that Pipeline{1, 2, 3} all conform to SomeProtocol. Got it! Thanks @xedin. One final question in that case. Is there then a possibility to also restrict associated types of that protocol? As in I'd like for all types supplied to the result builder that A they conform to SomeProtocol and B that their associated types SomeType (contained in SomeProtocol) are equal? Or is that not possible? Yes, you can place whatever generic requirements are necessary i.e. <T: SomeProtocol>(...) where T.SomeType == ..., T.SomeOtherMember : OtherProtocol etc. Hmm I don't fully follow. I can't make the requirement for the following example: protocol: SomeProtocol { associatedtype Input associatedtype Output } struct P1<Input>: SomeProtocol { typealias Output = Int } struct P2<Input>: SomeProtocol { typealias Output = String } //When using these in a result builder: let object = Builder { P1() P2() } Can I force the Input associated type to be the same for all entries given to the builder? I don't see how to do that with restrictions on buildExpression that seems to me to only put restrictions on the individual entries is that correct? There are a couple of ways depending on what you want. You can make your builder generic like this: @resultBuilder struct Builder<T> { static func buildExpression<U: SomeProtocol>(_ e: U) -> U where U.Input == T { e } ... } This would make sure that Input is the same for all the elements eagerly. You could also enforce that requirement in a buildBlock: @resultBuilder struct Builder<T> { static func buildBlock<Input, T0: SomeProtocol, T1: SomeProtocol, ...>(_ t0: T0, _ t1: T1) -> <#Result#> where T0.Input == Input, T1.Input == Intput, T<N> == Input { e } ... } The caveat here is that element type-checking is free to pick what ever overload for actual elements. @xedin ah yes Nr 1 is exactly what I've been trying the last few minutes! It seems to work for my minimum example fine. I've also applied it to the library as a whole but unfortunately haven't had success of getting it to fully work as before. Thanks for all the help! No problem! I'm going to resolve this one, feel free to reach out on forums as well if you need more help with result builders. @xedin Do you happen to remember which recent PRs were involved in the change? Would be interested in reading up on that context. The implementation is https://github.com/apple/swift/pull/60065 and there are a few smaller follow-up PRs. Enablement is https://github.com/apple/swift/pull/62734. I’m also working on a forums post to announce the change and describe some of the situations that were allowed before but wouldn’t be now. @xedin Thanks! Will be interested to see how it improves result builder compile times generally :smile: @xedin @stephencelis Might be interesting: I've worked on making the result builder generic over Input for swift-parsing But am running into long build times for some parsers and some even times out with The compiler is unable to type-check this expression in reasonable time; try breaking up the expression into distinct sub-expressions Do you have a branch where you are making these changes? I can probably take a look and give you suggestions. @xedin that would be great! My changes can be found here: https://github.com/jaapwijnen/swift-parsing/tree/generic-parser-builder I've commented out some tests (which you can see in the changes) these were the ones that were mostly acting up. No worries, I'll try to take a look tomorrow. If you could reduce a problematic case together with the builder and post here that would be very helpful as well. Hmm it's hard to isolate the builder from the library in a singular case. Which is what makes it so hard for me to debug why it's having trouble compiling the parser. Is there any way to identify which resultBuilder methods are used under the hood? Here's an example of a parser that's not compiling at the moment however (taken from one of the tests I commented out in the PR I linked): var parser = Parse { "Hello" if parseComma { "," } " " Prefix { $0 != "!" } "!" } The Prefix parser is the one causing trouble in this case I believe, although not sure why. It's something like: var parser = Parse { var v0 = ParserBuilder.buildExpression("Hello") var v1: ? if parseComma { var v2 = ParserBuilder.buildExpression(",") v1 = ParserBuilder.buildBlock(v2) } else { v1 = ParserBuilder.buidBlock(Optional<???>.none) } var v3 = ParserBuilder.buildExpression(" ") var v4 = ParserBuilder.buildExpression(Prefix { $0 != "!" }) var v5 = ParserBuilder.buildExpression("!") var r0 = ParserBuilder.buildPartialBlock(first: v0) var r1 = ParserBuilder.buildPartialBlock(accumulated: r0, next: v1) var r2 = ParserBuilder.buildPartialBlock(accumulated: r1, next: v3) var r3 = ParserBuilder.buildPartialBlock(accumulated: r2, next: v4) var r4 = ParserBuilder.buildPartialBlock(accumulated: r3, next: v5) return r4 } Not sure if Prefix { $0 != "!" } is transformed or not. The place marked with ??? is the type of v2, this interaction cannot be expressed in the surface language as stated by the proposal. Hm interestingly it seems to be the combination of the parser wrapped in an if statement and the Prefix parser. Since when I delete either of them the parser does compile. What is $0 supposed to be? I see that Input associated with Prefix supposed to conform to Collection protocol. Substring in this case I believe It's Input.Element: https://github.com/pointfreeco/swift-parsing/blob/a3d52c666982f7cc9f243b910151cdb9f6283db1/Sources/Parsing/ParserPrinters/Prefix.swift#L55 Which is Substring.Element, or Character. Ah my bad thanks @stephencelis The good news is that the performance issues are not due to result builder implementation, old one is too complex too on that example. Just need to figure out what exactly is causing this, I suspect it's multiple overloads of buildIf. Huh that is strange, @stephencelis this used to compile fine right? It's part of the library's tests. I think you might be building with old compiler where transform is disabled actually because with transform I get the following diagnostic: error: generic parameter 'Parsers' could not be inferred let parser = Parse { ^ swift-parsing/Sources/Parsing/ParserPrinters/Parse.swift:23:21: note: 'Parsers' declared as parameter to type 'Parse' public struct Parse<Parsers: Parser>: Parser { ^ swift-parsing/Sources/Parsing/ParserPrinters/Prefix.swift:225:18: note: explicitly specify the generic arguments to fix this issue let parser = Parse { ^ <<#Parsers: Parser#>> @JaapWijnen Looks like all you might need to do is to fix init declaration in Parse.swift line 46 to: @inlinable public init<Input>(@ParserBuilder<Input> with build: () -> Parsers) where Parsers.Input == Input { self.parsers = build() } Parsers.Input cannot be used directly because the concrete type for Parsers is not known until the whole body of build: is type-checked, instead you want the Input to inferred from the body and passed up to the Parsers.Input if that generic parameter is not specified explicitly. Also make sure that the toolchain you are using actually has result builder transform enabled :) That did help :) Now I'm running into the ambiguity issue again unfortunately haha. Must be a different example then since the one you pasted builds just fine for me after that init change. Hmm not entirely for me. Made some changes to the parsers in the library to all accept an Input generic. Now I'm having trouble to get it to compile on the 2022-10-09 toolchain however: This one times out again: var parser = Parse { "Hello" if parseComma { "," } " " Prefix { $0 != "!" } "!" } And for this one I get the following error let p = ParsePrint { Prefix { $0 != "\n" } // Ambiguous use of 'init(while:)' First() } And another very similar one but slightly different: let p = ParsePrint { Prefix(3) { $0.isNumber } First() //Ambiguous use of 'init()' } My latest changes can be found here if you're interested! : https://github.com/JaapWijnen/swift-parsing/commit/edcdf2d98c00dd4b210c7e904d05b9e151ff6b4e I think you might want to pick up a newer toolchain maybe because I pulled your branch and couldn't reproduce anything except to: func testSkipSuccess() { var input = "42 Hello, world!"[...].utf8 XCTAssert(try () == XCTUnwrap(Skip { Int.parser() }.parse(&input))) // Ambiguous use of 'parser(of:radix:)' XCTAssertEqual(" Hello, world!", Substring(input)) } This is indeed ambiguous because both overloads that match are marked as @_disfavoredOverload, have all their parameters defaulted and produce a type that conforms to Parser - From conforms to Parser via ParserPrinter, IntParser conforms to parser directly. Downloading the newest now! Ah ok interesting, I thought because the input is included which is of type Substring.UTF8View the Skip parser would know it's input generic is of that same type therefore the result builder will use that as it's Input generic which in turn will pick that specific Int.parser() overload. But I guess it doesn't work like that. Could you point me towards where my logic breaks down/how that type inference actually works in this case? Thanks for all the help! The ranking of overloads would check the number of defaulted parameters (which is the same) and whether any of the parameters have subtyping relationship, none of this is applicable in this case. Ah ok thanks @xedin! Then indeed some of the ergonomics of doing it like this are simply lost due to this regression (at least in their current form) Thanks for helping me understand! Hope this is useful to you as well @stephencelis. Not sure what useful next steps are in terms of getting swift-parsing to compile for new toolchains (probably not an immediate priority to you?) But I'd love to help if you do have ideas about that! We're using swift-parsing with snapshots so I'd love for it to work as it did before but right now at least we can get around the problem by just annotating all the parsers with a lot more type information. This is definitely helpful! While we knew a lot of changes we needed to make based off more recent result builder projects we've taken on, this thread provided a lot of additional food for thought. Thanks, @xedin! Looks like all you might need to do is to fix init declaration in Parse.swift line 46 to: @inlinable public init<Input>(@ParserBuilder<Input> with build: () -> Parsers) where Parsers.Input == Input { self.parsers = build() } Parsers.Input cannot be used directly because the concrete type for Parsers is not known until the whole body of build: is type-checked, instead you want the Input to inferred from the body and passed up to the Parsers.Input if that generic parameter is not specified explicitly. Also make sure that the toolchain you are using actually has result builder transform enabled :) @xedin I tried to apply this technique (on Swift Development Snapshot 2023-01-23) to a similar generic builder, but get "Generic parameter 'X' could not be inferred" where 'X' is the builder's generic. Here's a basic example demonstrating the problem: @resultBuilder enum Builder<A> { static func buildExpression<T: P>(_ expression: T) -> T where T.A == A { expression } static func buildBlock() -> E<A> { E() } static func buildBlock<T: P>(_ component: T) -> T where T.A == A { component } static func buildPartialBlock<T: P>(first: T) -> T where T.A == A { first } static func buildPartialBlock<T0: P, T1: P>(accumulated: T0, next: T1) -> T1 where T0.A == A, T1.A == A { next } } protocol P<A> { associatedtype A associatedtype Body @Builder<A> var body: Body { get } } extension P where Body == Never { var body: Body { fatalError() } } struct E<A>: P {} struct C<T: P>: P { typealias A = T.A let t: T init<A>(@Builder<A> build: () -> T) where T.A == A { self.t = build() } } struct S: P { struct A {} @Builder<A> var body: some P<A> { C { // 🛑 Generic parameter 'A' could not be inferred E() } C {} // 🛑 Generic parameter 'A' could not be inferred } } Am I maybe missing something? Is there an -Xfrontend setting (or otherwise) I need to enabled on this snapshot? Or is this a potential bug? There is no where to infer A from in C { E() } or C {} though right? @xedin Shouldn't the A generic in @Builder<A> flow through the body context and hydrate C's init<A>, which in turn will know that its @Builder is generic over A and flow into the E? Or is this going in the wrong direction? I don't quite know how to precisely translate builder code to the equivalent buildBlock (I can approximate it, but I'm not sure if I'm missing anything). @stephencelis No, this is actually the whole point. The body is not solved as a unit, it's solved piece-by-piece just like a regular multi-statement closure so @Builder<A> affects only final return build{Partial}Block(...): struct S: P { struct A {} @Builder<A> var body: some P<A> { var $__builder0 = Builder.buildExpression(C { var $__builder0 = Builder.buildExpression(E()) return Builder.buildBlock($__builder0) }) var $__builder1 = Builder.buildExpression(C { return Builder.buildBlock() }) let r1 = Builder.buildPartialBlock(first: $__builder0) let r2 = Builder.buildPartialBlock(accumulated: r1, next: $__builder1) return r2 } } In here Builder in Builder.buildExpression(E()) is different from Builder<A> attached to var body, there is no connection between them. Maybe this is going to be more helpful: buildExpression states that T.A == A where T: P - to be able to solve that we need to first solve the expression C.init { E() }, C.init again tries to propagate A "up" (to the context) from the body of the result builder because in where T.A == A the right-hand side is not yet resolved - it's waiting for the { E() } to get resolved (as stated at the very beginning). Note that buildExpression requirement T.A == A cannot propagate information down from @Builder<A> until T is resolved and that is only resolved when the expression is solved and part of that expression is solved in isolation because argument is a result builder. @xedin I might be missing some nuance here, but doesn't this place a pretty big limitation on generic result builders? I would hope that the generic Builder<A> would propagate into the context and be used to anchor all of the buildExpressions, etc., and allow a generic type like E() to resolve without specifying generics explicitly. Why wouldn't the following be preferable? struct S: P { struct A {} @Builder<A> var body: some P<A> { var $__builder0 = Builder<A>.buildExpression(C { @Builder<A> in var $__builder0 = Builder<A>.buildExpression(E()) return Builder<A>.buildBlock($__builder0) }) var $__builder1 = Builder<A>.buildExpression(C { @Builder<A> in return Builder<A>.buildBlock() }) let r1 = Builder<A>.buildPartialBlock(first: $__builder0) let r2 = Builder<A>.buildPartialBlock(accumulated: r1, next: $__builder1) return r2 } } Wouldn't dragging the <A> into things make this code compile just fine and without taking a big toll on the type checker? would propagate into the context and be used to anchor all of the buildExpression It does propagate into the context but that context is top-level things and the closure associated with C.init` because that is a completely different builder. Is there a reason why the A couldn't trickle into the inner @Builder<???> context to resolve? All the type information seems to be there... Is there any way to make generic builders compose nicely? For it to get into C.init closure it needs to get connected to something which would resolve it without relying on bi-directional inference which I don't think is possible because the language cannot pick up relations from context. Also it's very important to note that the example you've posted is not accepted by the original implementation either so this is not a regression. @xedin So it's just impossible for C.init's builder to be influenced by the parent buildExpression? While I understand that result builder type inference is limited for performance reasons, I thought buildExpression was specifically designed to provide context to work around some of these limitations. Is there any reason why buildExpression couldn't still use bidirectional inference while the rest of the result builder APIs remain limited? This limitation is quite a bummer if there's no workaround, and it prevents us from shipping an API that isn't cumbersome to use and difficult to troubleshoot or needlessly verbose, e.g.: C(for: A.self) { // Worse when multiple generics are involved // ... } I thought buildExpression was specifically designed to provide context to work around some of these limitations. Is there any reason why buildExpression couldn't still use bidirectional inference while the rest of the result builder APIs remain limited? buildExpression can and does provide context, but the way you have defined it, there is no context it could pass down. The problem here is not the luck on bi-directional inference inside of buildExpression argument, the problem is that the builder associated with C.init has Builder<???>.buildExpression(E()) expression where inference fails because { E() } is treated as a transformed multi-statement closure (as per result builder proposal). The generic parameter cannot be transferred into the body of { E() }, that didn't work before and it won't work with new transform. Also it's very important to note that the example you've posted is not accepted by the original implementation either so this is not a regression. Do you mean existing code should continue to compile? Or do you mean it's not a regression as far as the original evolution proposal conveyed? Given the swift-parsing code that used to compile, as well as some swift-composable-architecture code that used to compile, these changes definitely introduce source breaking changes and regressions, no? While I get a compiler crash with my example above when I run with Xcode 14.2's built-in toolchain, I have some very similar library code that compiles fine in Xcode 14.2 but does not on the 5.8 (or development) toolchain. If you clone https://github.com/pointfreeco/swift-composable-architecture (main branch), and paste the following code somewhere into the project: struct R: ReducerProtocol { struct State {} enum Action {} @ReducerBuilder<State, Action> var body: some ReducerProtocol<State, Action> { CombineReducers {} CombineReducers { EmptyReducer() } } } ...it compiles just fine. However, if I switch to a 5.8 or development toolchain, it doesn't. These are pretty unfortunate source breaking changes to introduce to library users, especially if there's nothing we can do as library maintainers to make the upgrade to Swift 5.8 painless: I also began to sketch out another branch with some adjustments for some of the changes coming: https://github.com/pointfreeco/swift-composable-architecture/compare/builder-updates Here, too, of course, the above won't compile (and still compiles fine in Xcode 14.2), but I was hoping the changes would at least address some problems with the existing implementation. This example is problematic because ReducerBuilderOf is declared as: public typealias ReducerBuilderOf<R: ReducerProtocol> = ReducerBuilder<R.State, R.Action> and CombineReducers.init is declared as: public init( @ReducerBuilderOf<Reducers> _ build: () -> Reducers ) which means that: CombineReducers { var $__builder0 = ReducerBuilderOf<$T>.builderExpression(EmptyReducer()) return ReducerBuilderOf<$T>.buildBlock($__builder0) } $T inference depends on the result type of buildBlock. The only way to infer that would be to attempt to solve whole body as a unit. @xedin See the second link, where I've deprecated ReducerBuilderOf and refactored away usage. @stephencelis Even without that explicit R: ReducerProtocol the type of ReducerBuilder still depends on the type produced by the body, although it's unpacked directly now. @xedin I think I understand that, but what I'm trying to show you is that this statements you made don't appear to be true (as I'm understanding them): Also it's very important to note that the example you've posted is not accepted by the original implementation either so this is not a regression. The generic parameter cannot be transferred into the body of { E() }, that didn't work before and it won't work with new transform. (Emphasis mine.) The code does currently compile, so this does seem like a regression. See my earlier comment again for more context. The code does currently compile, so this does seem like a regression. @stephencelis I was referring to the code in https://github.com/apple/swift/issues/62978#issuecomment-1402356383 @xedin Hm, OK. So seeing as this is a regression with no current workaround, should I start a forum discussion to get wider visibility? Currently, these changes could break a lot of userland code. This is currently being discussed by the language group. There is a workaround here though - specify generic arguments explicitly. I guess I meant workaround that wouldn't be passed on to library users and become a new support vector for us 😅 And ideally one that doesn't make every nested generic builder more verbose than folks expect in Swift. As a side note, I guess better compiler diagnostics and support would help here. If the compiler could help users migrate (or we could provide such migration APIs as fix-its) that'd make this change less painful. Well, that the problem, compiler cannot determine what the type should be here so we cannot suggest one, it really is very unfortunate that declarations like that used to be accepted by the compiler because they result in performance issues as bodies grow and bad diagnostics because whole body is type-checked together. @xedin Since the compiler can figure it out today, one would hope that the diagnostic could do this work when compilation fails in a post-compilation phase, at least during a migratory period... Unfortunately no, otherwise we wouldn’t be spending time re-implementing the transform. The principle issue here is that type-checking everything together is too expensive especially for invalid code. So seeing as this is a regression with no current workaround, should I start a forum discussion to get wider visibility? Currently, these changes could break a lot of userland code. @stephencelis I'm going to take a look at the result builders in TCA to see whether I can figure out a tweak to the builder to enable the same call-site ergonomics. I think leaving this issue open is the best course of action for now. Pavel and I are working on a post for the Swift Forums to widely communicate the motivation, compile-time improvements, and the source compatibility impact of this change. Thank you for your patience and for reporting the impact on your project! Thanks @hborla and @xedin for the communication, in general! We do appreciate the need and reasoning behind the changes, and we hope that the majority of TCA result builder usage will remain unaffected. We just wanted to do what we could to signal potential breakage to the core team and minimize frustration for end users of the library that hit what we hope are edge cases. We also hope there's a clear way to communicate to users when they need to do: Reduce { state, action in // or CombineReducers { vs. Reduce<State, Action> { state, action in Or if we should just enforce specifying via argument wherever there may be ambiguity, which is more verbose than needed a lot of the time, but at least avoids the problematic case: Reduce(into: State.self, action: Action.self) { state, action in @stephencelis I actually found a solution for CombineReducers problem (based on builder-updates branch): diff --git a/Sources/ComposableArchitecture/Reducer/Reducers/CombineReducers.swift b/Sources/ComposableArchitecture/Reducer/Reducers/CombineReducers.swift index bb569e8dc..031a8dc51 100644 --- a/Sources/ComposableArchitecture/Reducer/Reducers/CombineReducers.swift +++ b/Sources/ComposableArchitecture/Reducer/Reducers/CombineReducers.swift @@ -15,7 +15,8 @@ /// .ifLet(\.child, action: /Action.child) /// } /// ``` -public struct CombineReducers<Reducers: ReducerProtocol>: ReducerProtocol { +public struct CombineReducers<Action, State, Reducers: ReducerProtocol>: ReducerProtocol + where Action == Reducers.Action, State == Reducers.State { @usableFromInline let reducers: Reducers @@ -23,9 +24,9 @@ public struct CombineReducers<Reducers: ReducerProtocol>: ReducerProtocol { /// /// - Parameter build: A reducer builder. @inlinable - public init<State, Action>( + public init( @ReducerBuilder<State, Action> _ build: () -> Reducers - ) where State == Reducers.State, Action == Reducers.Action { + ) { self.init(internal: build()) } Adding this indirection for State/Action generic parameter makes sure that when CombineReducers() is wrapped in buildExpression declared as: static func buildExpression(_: some ReducerProtocol<State, Action>) -> some ReducerProtocol<State, Action> { return expression } State Action could be infer for it based on the "builder self": ReducerBuilder<State, Action>.builderExpression(CombineReducers()) where State and Action are struct and enum from your example. Example you have posted previously type-checks with these changes. @xedin Wow! Thanks! That actually gets the one regression in our test suite building again without needing to be explicit with \State.path and allowing abbreviation in \.path again! https://github.com/pointfreeco/swift-composable-architecture/commit/feae99aa459f47962a01289682ee8468e2dc969a @stephencelis No problem! Any other issues I can help with? @xedin Will let you know if anything else comes up! Really appreciate the help along the way! No worries, happy to help! @xedin One more question... :smile: We have this generic Reduce<State, Action> type: https://github.com/pointfreeco/swift-composable-architecture/blob/main/Sources/ComposableArchitecture/Reducer/Reducers/Reduce.swift When used from a builder with a single build block, e.g. here: https://github.com/pointfreeco/swift-composable-architecture/blob/5e4caa89ee7e1add70557f77c7c35cfad73ff024/Examples/TicTacToe/tic-tac-toe/Sources/LoginCore/LoginCore.swift#L33 The closure seems to be type-checked efficiently, and Xcode autocomplete works just great when trying to autocomplete state.<ESC> parameter names and case .<ESC> action names. However as soon as we add another reducer to the build block (like EmptyReducer()), above or below this Reduce, things still compile just fine, but we lose all autocompletion help in the closure. state.<ESC> and case .<ESC> are of no help till we explicitly specify Reduce<State, Action>. Do you happen to have a tip we can employ to keep the compiler happy without this extra busy work? Cc @ahoppen Hi @stephencelis, Thanks for sharing the code completion issue. I hope I can find time to take a look at it sometime next week. @ahoppen Thanks! I have managed to reduce the code completion issue down to the following. Checking with Xcode 14, the issue already existed in that release, so it’s not a regression. The problem here appears to be that code completion is getting confused which closures it needs to type check. Specifically, the result builder with multiple statements the closure is not in the ReadyForTypeCheck body state here and thus we don’t continue. https://github.com/apple/swift/blob/4ceec83cd87acc140e0e5d93956036861de88589/lib/Sema/TypeCheckStmt.cpp#L2419-L2420 I think this should resolve itself after we Get rid of LeaveClosuresUnchecked (https://github.com/apple/swift/pull/59944) Migrate the remaining code completion types to solver-based (https://github.com/apple/swift/pull/42280 and https://github.com/apple/swift/pull/42156) Get rid of the the code path mentioned above entirely There’s still some way to go on that journey but it’s something I am doing right now anyway. // RUN: %swift-ide-test -code-completion -source-filename %s -code-completion-token COMPLETE @resultBuilder enum ReducerBuilder2<Action> { static func buildBlock(_ r: Reduce2<Action>) -> Reduce2<Action> { r } static func buildBlock(_ r0: Reduce2<Action>, _ r1: Reduce2<Action>) -> Reduce2<Action> { r0 } static func buildExpression(_ r: Reduce2<Action>) -> Reduce2<Action> { r } } enum Action { case alertDismissed } struct Reduce2<Action> { init() {} init(_ reduce: (Action) -> Int) {} } struct Login2 { @ReducerBuilder2<Action> var body: Reduce2<Action> { Reduce2() // Remove this to receive code completion results get code completion working again Reduce2 { action in switch action { case .#^COMPLETE^# alertDismissed: return 0 } } } } @ahoppen That's great news! Thanks for looking into this :smile: I stumbled across a similar issue today and this will be fixed by https://github.com/apple/swift/pull/59944. The test case I mentioned above is included in that PR.
gharchive/issue
2023-01-11T22:10:52
2025-04-01T04:56:01.128471
{ "authors": [ "JaapWijnen", "ahoppen", "hborla", "stephencelis", "xedin" ], "repo": "apple/swift", "url": "https://github.com/apple/swift/issues/62978", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1955180614
Methods on parameter packed generic types specialized with zero types cannot be used without crashing the compiler Description You can't create a generic type with a parameter pack, specialize it with zero types, and then call a method that uses that parameter pack. It crashes the compiler. Steps to reproduce Stack trace: 0. Program arguments: /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/swift-frontend -frontend -c /Users/fnord/Documents/Swift/TestMacOS/TestMacOSCore/Sources/TestMacOSCore/Dependencies.swift -primary-file /Users/fnord/Documents/Swift/TestMacOS/TestMacOSCore/Sources/TestMacOSCore/TestMacOSCore.swift -emit-dependencies-path /Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Build/Intermediates.noindex/TestMacOSCore.build/Debug/TestMacOSCore.build/Objects-normal/arm64/TestMacOSCore.d -emit-const-values-path /Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Build/Intermediates.noindex/TestMacOSCore.build/Debug/TestMacOSCore.build/Objects-normal/arm64/TestMacOSCore.swiftconstvalues -emit-reference-dependencies-path /Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Build/Intermediates.noindex/TestMacOSCore.build/Debug/TestMacOSCore.build/Objects-normal/arm64/TestMacOSCore.swiftdeps -serialize-diagnostics-path /Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Build/Intermediates.noindex/TestMacOSCore.build/Debug/TestMacOSCore.build/Objects-normal/arm64/TestMacOSCore.dia -target arm64-apple-macos14.0 -Xllvm -aarch64-use-tbi -enable-objc-interop -stack-check -sdk /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX14.0.sdk -I /Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Build/Products/Debug -I /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/usr/lib -F /Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Build/Products/Debug/PackageFrameworks -F /Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Build/Products/Debug/PackageFrameworks -F /Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Build/Products/Debug/PackageFrameworks -F /Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Build/Products/Debug/PackageFrameworks -F /Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Build/Products/Debug/PackageFrameworks -F /Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Build/Products/Debug -F /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/Library/Frameworks -no-color-diagnostics -enable-testing -g -module-cache-path /Users/fnord/Library/Developer/Xcode/DerivedData/ModuleCache.noindex -profile-generate -profile-coverage-mapping -swift-version 5 -enforce-exclusivity=checked -Onone -D SWIFT_PACKAGE -D DEBUG -D Xcode -serialize-debugging-options -package-name testmacoscore -const-gather-protocols-file /Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Build/Intermediates.noindex/TestMacOSCore.build/Debug/TestMacOSCore.build/Objects-normal/arm64/TestMacOSCore_const_extract_protocols.json -empty-abi-descriptor -validate-clang-modules-once -clang-build-session-file /Users/fnord/Library/Developer/Xcode/DerivedData/ModuleCache.noindex/Session.modulevalidation -Xcc -working-directory -Xcc /Users/fnord/Documents/Swift/TestMacOS/TestMacOSCore -resource-dir /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/lib/swift -enable-anonymous-context-mangled-names -Xcc -ivfsstatcache -Xcc /Users/fnord/Library/Developer/Xcode/DerivedData/SDKStatCaches.noindex/macosx14.0-23A334-4a76ee93ef69e462914ffe9883de22b8.sdkstatcache -Xcc -I/Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Build/Intermediates.noindex/TestMacOSCore.build/Debug/TestMacOSCore.build/swift-overrides.hmap -Xcc -I/Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Build/Products/Debug/include -Xcc -I/Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Build/Intermediates.noindex/TestMacOSCore.build/Debug/TestMacOSCore.build/DerivedSources-normal/arm64 -Xcc -I/Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Build/Intermediates.noindex/TestMacOSCore.build/Debug/TestMacOSCore.build/DerivedSources/arm64 -Xcc -I/Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Build/Intermediates.noindex/TestMacOSCore.build/Debug/TestMacOSCore.build/DerivedSources -Xcc -DSWIFT_PACKAGE -Xcc -DDEBUG=1 -module-name TestMacOSCore -frontend-parseable-output -disable-clang-spi -target-sdk-version 14.0 -target-sdk-name macosx14.0 -external-plugin-path /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX14.0.sdk/usr/lib/swift/host/plugins#/Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX14.0.sdk/usr/bin/swift-plugin-server -external-plugin-path /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX14.0.sdk/usr/local/lib/swift/host/plugins#/Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX14.0.sdk/usr/bin/swift-plugin-server -external-plugin-path /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/usr/lib/swift/host/plugins#/Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/usr/bin/swift-plugin-server -external-plugin-path /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/usr/local/lib/swift/host/plugins#/Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/usr/bin/swift-plugin-server -plugin-path /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/lib/swift/host/plugins -plugin-path /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/local/lib/swift/host/plugins -o /Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Build/Intermediates.noindex/TestMacOSCore.build/Debug/TestMacOSCore.build/Objects-normal/arm64/TestMacOSCore.o -index-unit-output-path /TestMacOSCore.build/Debug/TestMacOSCore.build/Objects-normal/arm64/TestMacOSCore.o -index-store-path /Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Index.noindex/DataStore -index-system-modules 1. Apple Swift version 5.9 (swiftlang-5.9.0.128.108 clang-1500.0.40.1) 2. Compiling with the current language version 3. While evaluating request ASTLoweringRequest(Lowering AST to SIL for file "/Users/fnord/Documents/Swift/TestMacOS/TestMacOSCore/Sources/TestMacOSCore/TestMacOSCore.swift") 4. While silgen emitFunction SIL function "@$s13TestMacOSCore3fooyyF". for 'foo()' (at /Users/fnord/Documents/Swift/TestMacOS/TestMacOSCore/Sources/TestMacOSCore/TestMacOSCore.swift:23:1) Stack dump without symbol names (ensure you have llvm-symbolizer in your PATH or set the environment var `LLVM_SYMBOLIZER_PATH` to point to it): 0 swift-frontend 0x00000001054f314c llvm::sys::PrintStackTrace(llvm::raw_ostream&, int) + 56 1 swift-frontend 0x0000000108071f2c llvm::sys::RunSignalHandlers() + 112 2 swift-frontend 0x0000000103db310c SignalHandler(int) + 352 3 libsystem_platform.dylib 0x000000018a961a24 _sigtramp + 56 4 swift-frontend 0x00000001058b51d0 (anonymous namespace)::ArgEmitter::emitPackArg(llvm::MutableArrayRef<swift::Lowering::ArgumentSource>, swift::Lowering::AbstractionPattern) + 164 5 swift-frontend 0x00000001058a2658 (anonymous namespace)::ArgEmitter::emitPreparedArgs(swift::Lowering::PreparedArguments&&, swift::Lowering::AbstractionPattern) + 512 6 swift-frontend 0x000000010591dea0 (anonymous namespace)::CallSite::emit(swift::Lowering::SILGenFunction&, swift::Lowering::AbstractionPattern, swift::CanTypeWrapper<swift::SILFunctionType>, (anonymous namespace)::ParamLowering&, llvm::SmallVectorImpl<swift::Lowering::ManagedValue>&, llvm::SmallVectorImpl<(anonymous namespace)::DelayedArgument>&, swift::ForeignInfo const&) && + 672 7 swift-frontend 0x000000010591cab4 (anonymous namespace)::CallEmission::emitArgumentsForNormalApply(swift::Lowering::AbstractionPattern, swift::CanTypeWrapper<swift::SILFunctionType>, swift::ForeignInfo const&, llvm::SmallVectorImpl<swift::Lowering::ManagedValue>&, llvm::Optional<swift::SILLocation>&) + 1000 8 swift-frontend 0x0000000102edb34c (anonymous namespace)::CallEmission::apply(swift::Lowering::SGFContext) + 4044 9 swift-frontend 0x000000010585574c swift::Lowering::SILGenFunction::emitApplyExpr(swift::ApplyExpr*, swift::Lowering::SGFContext) + 3072 10 swift-frontend 0x0000000105ab8468 swift::Lowering::SILGenFunction::emitIgnoredExpr(swift::Expr*) + 892 11 swift-frontend 0x000000010368f570 swift::ASTVisitor<(anonymous namespace)::StmtEmitter, void, void, void, void, void, void>::visit(swift::Stmt*) (.llvm.5384354941888427540) + 5512 12 swift-frontend 0x0000000105be5864 swift::Lowering::SILGenFunction::emitFunction(swift::FuncDecl*) + 632 13 swift-frontend 0x0000000102ec8a4c swift::Lowering::SILGenModule::emitFunctionDefinition(swift::SILDeclRef, swift::SILFunction*) + 8004 14 swift-frontend 0x00000001057d563c emitOrDelayFunction(swift::Lowering::SILGenModule&, swift::SILDeclRef) (.llvm.12369091251732681984) + 168 15 swift-frontend 0x00000001057bcbe8 swift::Lowering::SILGenModule::emitFunction(swift::FuncDecl*) + 292 16 swift-frontend 0x00000001057e4608 swift::ASTLoweringRequest::evaluate(swift::Evaluator&, swift::ASTLoweringDescriptor) const + 2908 17 swift-frontend 0x000000010440c4fc swift::SimpleRequest<swift::ASTLoweringRequest, std::__1::unique_ptr<swift::SILModule, std::__1::default_delete<swift::SILModule>> (swift::ASTLoweringDescriptor), (swift::RequestFlags)9>::evaluateRequest(swift::ASTLoweringRequest const&, swift::Evaluator&) + 200 18 swift-frontend 0x0000000105802948 llvm::Expected<swift::ASTLoweringRequest::OutputType> swift::Evaluator::getResultUncached<swift::ASTLoweringRequest>(swift::ASTLoweringRequest const&) + 584 19 swift-frontend 0x00000001080298d8 performCompile(swift::CompilerInstance&, int&, swift::FrontendObserver*) + 1680 20 swift-frontend 0x000000010802d474 swift::performFrontend(llvm::ArrayRef<char const*>, char const*, void*, swift::FrontendObserver*) + 4568 21 swift-frontend 0x00000001079734e0 swift::mainEntry(int, char const**) + 4116 22 dyld 0x000000018a5b9058 start + 2224 This causes the compiler to crash. class CounterBox<each Param> { func recordCall(_ params: repeat each Param) {} } func foo() { CounterBox< >().recordCall() // crash } Expected behavior The above should compile. Additionally, the parameter pack syntax is slightly different from what's described in the proposal. CounterBox<>() doesn't compile. Environment Swift compiler version info: swift-driver version: 1.87.1 Apple Swift version 5.9 (swiftlang-5.9.0.128.108 clang-1500.0.40.1) Target: arm64-apple-macosx14.0 Xcode version info Xcode 15.0 Build version 15A240d Deployment target: macOS 14.0 Additional context: Development is being done in a Swift Package using Xcode. Assertion failed: (nextSubstParamIndex == substParams.size()), function claimParams, file SILGenApply.cpp, line 4556. Stack dump: 0. Program arguments: bin/swift-frontend /Users/mac/Desktop/test.swift -debug-diagnostic-names -interpret 1. Swift version 5.11-dev (LLVM 5601e23279f85d8, Swift 140caf63be3f0fe) 2. Compiling with the current language version 3. While evaluating request ASTLoweringRequest(Lowering AST to SIL for module test) 4. While silgen emitFunction SIL function "@$s4test3fooyyF". for 'foo()' (at /Users/mac/Desktop/test.swift:85:1) Stack dump without symbol names (ensure you have llvm-symbolizer in your PATH or set the environment var `LLVM_SYMBOLIZER_PATH` to point to it): 0 swift-frontend 0x0000000112e6b51b llvm::sys::PrintStackTrace(llvm::raw_ostream&, int) + 43 1 swift-frontend 0x0000000112e6a755 llvm::sys::RunSignalHandlers() + 85 2 swift-frontend 0x0000000112e6bb4e SignalHandler(int) + 270 3 libsystem_platform.dylib 0x00007ff81913037d _sigtramp + 29 4 swift-frontend 0x000000010b27d24a swift::CanType::CanType(swift::TypeBase*) + 42 5 libsystem_c.dylib 0x00007ff819021a4d abort + 126 6 libsystem_c.dylib 0x00007ff819020d60 err + 0 7 swift-frontend 0x000000010bd223f6 (anonymous namespace)::ParamLowering::claimParams(swift::Lowering::AbstractionPattern, llvm::ArrayRef<swift::AnyFunctionType::Param>, swift::ForeignInfo const&) + 566 8 swift-frontend 0x000000010bd2146a (anonymous namespace)::CallSite::emit(swift::Lowering::SILGenFunction&, swift::Lowering::AbstractionPattern, swift::CanTypeWrapper<swift::SILFunctionType>, (anonymous namespace)::ParamLowering&, llvm::SmallVectorImpl<swift::Lowering::ManagedValue>&, llvm::SmallVectorImpl<(anonymous namespace)::DelayedArgument>&, swift::ForeignInfo const&) && + 170 9 swift-frontend 0x000000010bd2054e (anonymous namespace)::CallEmission::emitArgumentsForNormalApply(swift::Lowering::AbstractionPattern, swift::CanTypeWrapper<swift::SILFunctionType>, swift::ForeignInfo const&, llvm::SmallVectorImpl<swift::Lowering::ManagedValue>&, llvm::Optional<swift::SILLocation>&) + 1630 10 swift-frontend 0x000000010bd1f890 (anonymous namespace)::CallEmission::applyNormalCall(swift::Lowering::SGFContext) + 1120 11 swift-frontend 0x000000010bd1ddcd (anonymous namespace)::CallEmission::applyFirstLevelCallee(swift::Lowering::SGFContext) + 205 12 swift-frontend 0x000000010bcee740 (anonymous namespace)::CallEmission::apply(swift::Lowering::SGFContext) + 80 13 swift-frontend 0x000000010bcee0c8 swift::Lowering::SILGenFunction::emitApplyExpr(swift::ApplyExpr*, swift::Lowering::SGFContext) + 136 14 swift-frontend 0x000000010bdf09ab (anonymous namespace)::RValueEmitter::visitApplyExpr(swift::ApplyExpr*, swift::Lowering::SGFContext) + 59 15 swift-frontend 0x000000010bdde0e8 swift::ASTVisitor<(anonymous namespace)::RValueEmitter, swift::Lowering::RValue, void, void, void, void, void, swift::Lowering::SGFContext>::visitCallExpr(swift::CallExpr*, swift::Lowering::SGFContext) + 56 16 swift-frontend 0x000000010bdc78bc swift::ASTVisitor<(anonymous namespace)::RValueEmitter, swift::Lowering::RValue, void, void, void, void, void, swift::Lowering::SGFContext>::visit(swift::Expr*, swift::Lowering::SGFContext) + 2956 17 swift-frontend 0x000000010bdb8d23 swift::Lowering::SILGenFunction::emitRValue(swift::Expr*, swift::Lowering::SGFContext) + 195 18 swift-frontend 0x000000010bdbd9e6 swift::Lowering::SILGenFunction::emitIgnoredExpr(swift::Expr*) + 1990 19 swift-frontend 0x000000010bef76d9 (anonymous namespace)::StmtEmitter::visitBraceStmt(swift::BraceStmt*) + 2233 20 swift-frontend 0x000000010bef3181 swift::ASTVisitor<(anonymous namespace)::StmtEmitter, void, void, void, void, void, void>::visit(swift::Stmt*) + 81 21 swift-frontend 0x000000010bef30fa swift::Lowering::SILGenFunction::emitStmt(swift::Stmt*) + 42 22 swift-frontend 0x000000010be0f55a swift::Lowering::SILGenFunction::emitFunction(swift::FuncDecl*) + 842 23 swift-frontend 0x000000010bcc452b swift::Lowering::SILGenModule::emitFunctionDefinition(swift::SILDeclRef, swift::SILFunction*) + 4411 24 swift-frontend 0x000000010bcc7a75 swift::Lowering::SILGenModule::emitOrDelayFunction(swift::SILDeclRef) + 501 25 swift-frontend 0x000000010bcc33e5 swift::Lowering::SILGenModule::emitFunction(swift::FuncDecl*) + 309 26 swift-frontend 0x000000010bcc329d swift::Lowering::SILGenModule::visitFuncDecl(swift::FuncDecl*) + 29 27 swift-frontend 0x000000010bcc2f64 swift::ASTVisitor<swift::Lowering::SILGenModule, void, void, void, void, void, void>::visit(swift::Decl*) + 996 28 swift-frontend 0x000000010bcc2b6b swift::Lowering::SILGenModule::visit(swift::Decl*) + 59 29 swift-frontend 0x000000010bccc72a (anonymous namespace)::SILGenModuleRAII::emitSourceFile(swift::SourceFile*) + 314 30 swift-frontend 0x000000010bccc2e4 swift::ASTLoweringRequest::evaluate(swift::Evaluator&, swift::ASTLoweringDescriptor) const + 916 31 swift-frontend 0x000000010bef22de std::__1::unique_ptr<swift::SILModule, std::__1::default_delete<swift::SILModule>> swift::SimpleRequest<swift::ASTLoweringRequest, std::__1::unique_ptr<swift::SILModule, std::__1::default_delete<swift::SILModule>> (swift::ASTLoweringDescriptor), (swift::RequestFlags)9>::callDerived<0ul>(swift::Evaluator&, std::__1::integer_sequence<unsigned long, 0ul>) const + 142 32 swift-frontend 0x000000010bef21f8 swift::SimpleRequest<swift::ASTLoweringRequest, std::__1::unique_ptr<swift::SILModule, std::__1::default_delete<swift::SILModule>> (swift::ASTLoweringDescriptor), (swift::RequestFlags)9>::evaluateRequest(swift::ASTLoweringRequest const&, swift::Evaluator&) + 40 33 swift-frontend 0x000000010bcdeddc llvm::Expected<swift::ASTLoweringRequest::OutputType> swift::Evaluator::getResultUncached<swift::ASTLoweringRequest>(swift::ASTLoweringRequest const&) + 316 34 swift-frontend 0x000000010bcccee8 llvm::Expected<swift::ASTLoweringRequest::OutputType> swift::Evaluator::operator()<swift::ASTLoweringRequest, (void*)0>(swift::ASTLoweringRequest const&) + 40 35 swift-frontend 0x000000010bcccde0 swift::performASTLowering(swift::ModuleDecl*, swift::Lowering::TypeConverter&, swift::SILOptions const&, swift::IRGenOptions const*) + 256 36 swift-frontend 0x000000010b010cfb swift::performCompileStepsPostSema(swift::CompilerInstance&, int&, swift::FrontendObserver*) + 283 37 swift-frontend 0x000000010b040003 performAction(swift::CompilerInstance&, int&, swift::FrontendObserver*)::$_28::operator()(swift::CompilerInstance&) const + 147 38 swift-frontend 0x000000010b03ff5d bool llvm::function_ref<bool (swift::CompilerInstance&)>::callback_fn<performAction(swift::CompilerInstance&, int&, swift::FrontendObserver*)::$_28>(long, swift::CompilerInstance&) + 29 39 swift-frontend 0x000000010b03f421 llvm::function_ref<bool (swift::CompilerInstance&)>::operator()(swift::CompilerInstance&) const + 33 40 swift-frontend 0x000000010b03e438 withSemanticAnalysis(swift::CompilerInstance&, swift::FrontendObserver*, llvm::function_ref<bool (swift::CompilerInstance&)>, bool) + 392 41 swift-frontend 0x000000010b036f26 performAction(swift::CompilerInstance&, int&, swift::FrontendObserver*) + 1222 42 swift-frontend 0x000000010b014186 performCompile(swift::CompilerInstance&, int&, swift::FrontendObserver*) + 262 43 swift-frontend 0x000000010b012b60 swift::performFrontend(llvm::ArrayRef<char const*>, char const*, void*, swift::FrontendObserver*) + 2816 44 swift-frontend 0x000000010aee53df run_driver(llvm::StringRef, llvm::ArrayRef<char const*>, llvm::ArrayRef<char const*>) + 2431 45 swift-frontend 0x000000010aee4057 swift::mainEntry(int, char const**) + 1543 46 swift-frontend 0x000000010aee36a2 main + 34 47 dyld 0x00007ff818d7a3a6 start + 1942 Abort trap: 6 This no longer crashes on top of tree; please verify. This regressed at some point, and is reproducible with the Swift 5.10 toolchain included with Xcode 15.3 (15E204a). swift-driver version: 1.90.11.1 Apple Swift version 5.10 (swiftlang-5.10.0.13 clang-1500.3.9.4) Target: arm64-apple-macosx14.0 Stack dumpStack dump: 0. Program arguments: /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/swift-frontend -frontend -interpret Swift69313.swift -Xllvm -aarch64-use-tbi -enable-objc-interop -stack-check -sdk /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX14.4.sdk -color-diagnostics -new-driver-path /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/swift-driver -empty-abi-descriptor -resource-dir /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/lib/swift -module-name Swift69313 -disable-clang-spi -target-sdk-version 14.4 -target-sdk-name macosx14.4 -external-plugin-path /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX14.4.sdk/usr/lib/swift/host/plugins#/Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX14.4.sdk/usr/bin/swift-plugin-server -external-plugin-path /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX14.4.sdk/usr/local/lib/swift/host/plugins#/Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX14.4.sdk/usr/bin/swift-plugin-server -external-plugin-path /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/usr/lib/swift/host/plugins#/Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/usr/bin/swift-plugin-server -external-plugin-path /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/usr/local/lib/swift/host/plugins#/Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/usr/bin/swift-plugin-server -plugin-path /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/lib/swift/host/plugins -plugin-path /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/local/lib/swift/host/plugins 1. Apple Swift version 5.10 (swiftlang-5.10.0.13 clang-1500.3.9.4) 2. Compiling with the current language version 3. While evaluating request ASTLoweringRequest(Lowering AST to SIL for module Swift69313) 4. While silgen emitFunction SIL function "@$s10Swift693133fooyyF". for 'foo()' (at Swift69313.swift:5:1) Stack dump without symbol names (ensure you have llvm-symbolizer in your PATH or set the environment var `LLVM_SYMBOLIZER_PATH` to point to it): 0 swift-frontend 0x00000001063bbf3c llvm::sys::PrintStackTrace(llvm::raw_ostream&, int) + 56 1 swift-frontend 0x00000001063bb0f8 llvm::sys::RunSignalHandlers() + 112 2 swift-frontend 0x00000001063bc544 SignalHandler(int) + 360 3 libsystem_platform.dylib 0x000000019ded3584 _sigtramp + 56 4 swift-frontend 0x00000001012fbe9c (anonymous namespace)::ArgEmitter::emitPackArg(llvm::MutableArrayRef, swift::Lowering::AbstractionPattern) + 168 5 swift-frontend 0x00000001012f979c (anonymous namespace)::ArgEmitter::emitPreparedArgs(swift::Lowering::PreparedArguments&&, swift::Lowering::AbstractionPattern) + 512 6 swift-frontend 0x0000000101305fb4 (anonymous namespace)::CallSite::emit(swift::Lowering::SILGenFunction&, swift::Lowering::AbstractionPattern, swift::CanTypeWrapper, (anonymous namespace)::ParamLowering&, llvm::SmallVectorImpl&, llvm::SmallVectorImpl<(anonymous namespace)::DelayedArgument>&, swift::ForeignInfo const&) && + 672 7 swift-frontend 0x00000001013059cc (anonymous namespace)::CallEmission::emitArgumentsForNormalApply(swift::Lowering::AbstractionPattern, swift::CanTypeWrapper, swift::ForeignInfo const&, llvm::SmallVectorImpl&, llvm::Optional&) + 1000 8 swift-frontend 0x00000001012edd68 (anonymous namespace)::CallEmission::apply(swift::Lowering::SGFContext) + 4064 9 swift-frontend 0x00000001012eb7bc swift::Lowering::SILGenFunction::emitApplyExpr(swift::ApplyExpr*, swift::Lowering::SGFContext) + 3060 10 swift-frontend 0x0000000101357618 swift::Lowering::SILGenFunction::emitIgnoredExpr(swift::Expr*) + 892 11 swift-frontend 0x00000001013ef684 swift::ASTVisitor<(anonymous namespace)::StmtEmitter, void, void, void, void, void, void>::visit(swift::Stmt*) + 5032 12 swift-frontend 0x0000000101386538 swift::Lowering::SILGenFunction::emitFunction(swift::FuncDecl*) + 632 13 swift-frontend 0x00000001012dae48 swift::Lowering::SILGenModule::emitFunctionDefinition(swift::SILDeclRef, swift::SILFunction*) + 8344 14 swift-frontend 0x00000001012db5bc swift::Lowering::SILGenModule::emitOrDelayFunction(swift::SILDeclRef) + 172 15 swift-frontend 0x00000001012d8d9c swift::Lowering::SILGenModule::emitFunction(swift::FuncDecl*) + 292 16 swift-frontend 0x00000001012deb54 swift::ASTLoweringRequest::evaluate(swift::Evaluator&, swift::ASTLoweringDescriptor) const + 1668 17 swift-frontend 0x00000001013edf1c swift::SimpleRequest> (swift::ASTLoweringDescriptor), (swift::RequestFlags)9>::evaluateRequest(swift::ASTLoweringRequest const&, swift::Evaluator&) + 196 18 swift-frontend 0x00000001012e1f44 llvm::Expected swift::Evaluator::getResultUncached(swift::ASTLoweringRequest const&) + 584 19 swift-frontend 0x00000001012df854 swift::performASTLowering(swift::ModuleDecl*, swift::Lowering::TypeConverter&, swift::SILOptions const&, swift::IRGenOptions const*) + 152 20 swift-frontend 0x0000000100cb99f8 swift::performCompileStepsPostSema(swift::CompilerInstance&, int&, swift::FrontendObserver*) + 2068 21 swift-frontend 0x0000000100cbdf0c performCompile(swift::CompilerInstance&, int&, swift::FrontendObserver*) + 3616 22 swift-frontend 0x0000000100cbb6d0 swift::performFrontend(llvm::ArrayRef, char const*, void*, swift::FrontendObserver*) + 4968 23 swift-frontend 0x0000000100c4ae8c swift::mainEntry(int, char const**) + 2612 24 dyld 0x000000019db1a0e0 start + 2360 fish: Job 1, 'xcrun swift Swift69313.swift' terminated by signal SIGSEGV (Address boundary error) @rjmccall What Swift version was at the top of the tree at this point? I'm seeing this the Swift toolchain that shipped with the most recent stable Xcode release. (5.10) Ah makes sense. Thanks!
gharchive/issue
2023-10-21T01:01:47
2025-04-01T04:56:01.145974
{ "authors": [ "AnthonyLatsis", "rjmccall", "slice", "twof" ], "repo": "apple/swift", "url": "https://github.com/apple/swift/issues/69313", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
302408091
Add a cmake flag to the benchmarks: SWIFT_BENCHMARK_EXTRA_FLAGS. We used to have this cmake flag, but it seems to have disappeared. Building the benchmarks with different swiftc flags is central to performance analysis, so I'm not sure how people were getting this done. @swift-ci smoke test. @swift-ci smoke benchmark Build comment file: Optimized (O) Regression (3) TEST OLD NEW DELTA SPEEDUP StaticArray 9 11 +22.2% 0.82x DropLastCountableRangeLazy 11 12 +9.1% 0.92x ObjectiveCBridgeStubFromArrayOfNSString 27825 29699 +6.7% 0.94x (?) Improvement (3) TEST OLD NEW DELTA SPEEDUP ObjectiveCBridgeFromNSArrayAnyObject 22371 20673 -7.6% 1.08x (?) StringBuilderLong 1124 1053 -6.3% 1.07x (?) DictionaryBridge 2127 2019 -5.1% 1.05x (?) No Changes (372) TEST OLD NEW DELTA SPEEDUP AngryPhonebook 3645 3706 +1.7% 0.98x (?) AnyHashableWithAClass 73443 73565 +0.2% 1.00x (?) Array2D 2555 2556 +0.0% 1.00x (?) ArrayAppend 1023 1016 -0.7% 1.01x (?) ArrayAppendArrayOfInt 797 799 +0.3% 1.00x (?) ArrayAppendAscii 13239 13247 +0.1% 1.00x (?) ArrayAppendFromGeneric 797 800 +0.4% 1.00x (?) ArrayAppendGenericStructs 1427 1443 +1.1% 0.99x ArrayAppendLatin1 38072 38111 +0.1% 1.00x ArrayAppendLazyMap 1341 1343 +0.1% 1.00x (?) ArrayAppendOptionals 1429 1431 +0.1% 1.00x (?) ArrayAppendRepeatCol 1337 1339 +0.1% 1.00x (?) ArrayAppendReserved 756 756 +0.0% 1.00x ArrayAppendSequence 1130 1134 +0.4% 1.00x (?) ArrayAppendStrings 15243 15253 +0.1% 1.00x (?) ArrayAppendToFromGeneric 806 799 -0.9% 1.01x (?) ArrayAppendToGeneric 805 802 -0.4% 1.00x (?) ArrayAppendUTF16 37154 37111 -0.1% 1.00x (?) ArrayInClass 86 85 -1.2% 1.01x ArrayLiteral 0 0 +0.0% 1.00x ArrayOfGenericPOD2 151 151 +0.0% 1.00x ArrayOfGenericRef 4438 4444 +0.1% 1.00x (?) ArrayOfPOD 183 183 +0.0% 1.00x ArrayOfRef 4386 4443 +1.3% 0.99x (?) ArrayPlusEqualArrayOfInt 797 798 +0.1% 1.00x (?) ArrayPlusEqualFiveElementCollection 5530 5518 -0.2% 1.00x (?) ArrayPlusEqualSingleElementCollection 1023 1021 -0.2% 1.00x (?) ArrayPlusEqualThreeElements 1657 1654 -0.2% 1.00x (?) ArraySubscript 1536 1533 -0.2% 1.00x (?) ArrayValueProp 8 8 +0.0% 1.00x ArrayValueProp2 8 8 +0.0% 1.00x ArrayValueProp3 8 8 +0.0% 1.00x ArrayValueProp4 8 8 +0.0% 1.00x BinaryFloatingPointConversionFromBinaryInteger 41 41 +0.0% 1.00x BitCount 202 202 +0.0% 1.00x ByteSwap 99 99 +0.0% 1.00x COWTree 5506 5560 +1.0% 0.99x (?) CSVParsing 756043 762200 +0.8% 0.99x (?) CSVParsingAlt 731897 738938 +1.0% 0.99x CSVParsingAltIndices 368011 358621 -2.6% 1.03x CStringLongAscii 4774 4897 +2.6% 0.97x CStringLongNonAscii 2166 2193 +1.2% 0.99x CStringShortAscii 4740 4746 +0.1% 1.00x (?) Calculator 649 647 -0.3% 1.00x (?) CaptureProp 3914 3920 +0.2% 1.00x CharIndexing_ascii_unicodeScalars 14738 14733 -0.0% 1.00x (?) CharIndexing_ascii_unicodeScalars_Backwards 14251 14246 -0.0% 1.00x (?) CharIndexing_chinese_unicodeScalars 11172 11163 -0.1% 1.00x (?) CharIndexing_chinese_unicodeScalars_Backwards 10793 10793 +0.0% 1.00x CharIndexing_japanese_unicodeScalars 17649 17652 +0.0% 1.00x (?) CharIndexing_japanese_unicodeScalars_Backwards 17071 17059 -0.1% 1.00x (?) CharIndexing_korean_unicodeScalars 14293 14284 -0.1% 1.00x (?) CharIndexing_korean_unicodeScalars_Backwards 13814 13817 +0.0% 1.00x (?) CharIndexing_punctuatedJapanese_unicodeScalars 2680 2678 -0.1% 1.00x (?) CharIndexing_punctuatedJapanese_unicodeScalars_Backwards 2592 2589 -0.1% 1.00x CharIndexing_punctuated_unicodeScalars 3349 3348 -0.0% 1.00x (?) CharIndexing_punctuated_unicodeScalars_Backwards 3239 3237 -0.1% 1.00x (?) CharIndexing_russian_unicodeScalars 12289 12285 -0.0% 1.00x (?) CharIndexing_russian_unicodeScalars_Backwards 11883 11875 -0.1% 1.00x (?) CharIndexing_tweet_unicodeScalars 29086 29072 -0.0% 1.00x (?) CharIndexing_tweet_unicodeScalars_Backwards 27623 27632 +0.0% 1.00x (?) CharIndexing_utf16_unicodeScalars 20210 20221 +0.1% 1.00x (?) CharIndexing_utf16_unicodeScalars_Backwards 20844 20845 +0.0% 1.00x (?) CharIteration_ascii_unicodeScalars 20420 20436 +0.1% 1.00x (?) CharIteration_ascii_unicodeScalars_Backwards 12706 12684 -0.2% 1.00x CharIteration_chinese_unicodeScalars 15440 15453 +0.1% 1.00x CharIteration_chinese_unicodeScalars_Backwards 9638 9641 +0.0% 1.00x (?) CharIteration_japanese_unicodeScalars 24439 24463 +0.1% 1.00x (?) CharIteration_japanese_unicodeScalars_Backwards 15179 15180 +0.0% 1.00x (?) CharIteration_korean_unicodeScalars 19818 19811 -0.0% 1.00x (?) CharIteration_korean_unicodeScalars_Backwards 12319 12319 +0.0% 1.00x CharIteration_punctuatedJapanese_unicodeScalars 3644 3644 +0.0% 1.00x CharIteration_punctuatedJapanese_unicodeScalars_Backwards 2341 2340 -0.0% 1.00x (?) CharIteration_punctuated_unicodeScalars 4579 4577 -0.0% 1.00x (?) CharIteration_punctuated_unicodeScalars_Backwards 2916 2915 -0.0% 1.00x (?) CharIteration_russian_unicodeScalars 17031 17017 -0.1% 1.00x CharIteration_russian_unicodeScalars_Backwards 10605 10600 -0.0% 1.00x CharIteration_tweet_unicodeScalars 40358 40352 -0.0% 1.00x (?) CharIteration_tweet_unicodeScalars_Backwards 25055 25040 -0.1% 1.00x (?) CharIteration_utf16_unicodeScalars 27072 27053 -0.1% 1.00x (?) CharIteration_utf16_unicodeScalars_Backwards 15869 15877 +0.1% 1.00x (?) CharacterLiteralsLarge 6008 6038 +0.5% 1.00x (?) CharacterLiteralsSmall 410 411 +0.2% 1.00x (?) CharacterPropertiesFetch 4682 4697 +0.3% 1.00x (?) CharacterPropertiesPrecomputed 961 973 +1.2% 0.99x CharacterPropertiesStashed 1548 1566 +1.2% 0.99x CharacterPropertiesStashedMemo 1517 1533 +1.1% 0.99x Chars 969 970 +0.1% 1.00x (?) ClassArrayGetter 15 15 +0.0% 1.00x Combos 463 463 +0.0% 1.00x DictOfArraysToArrayOfDicts 791 803 +1.5% 0.99x (?) Dictionary 507 503 -0.8% 1.01x Dictionary2 1618 1614 -0.2% 1.00x (?) Dictionary2OfObjects 2977 2980 +0.1% 1.00x (?) Dictionary3 234 234 +0.0% 1.00x Dictionary3OfObjects 682 681 -0.1% 1.00x (?) Dictionary4 147 147 +0.0% 1.00x Dictionary4OfObjects 200 200 +0.0% 1.00x DictionaryGroup 133 132 -0.8% 1.01x DictionaryGroupOfObjects 1957 1958 +0.1% 1.00x (?) DictionaryLiteral 1590 1586 -0.3% 1.00x (?) DictionaryOfObjects 2182 2171 -0.5% 1.01x (?) DictionaryRemove 2927 2990 +2.2% 0.98x (?) DictionaryRemoveOfObjects 24034 24002 -0.1% 1.00x (?) DictionarySubscriptDefaultMutation 143 143 +0.0% 1.00x DictionarySubscriptDefaultMutationArray 493 494 +0.2% 1.00x (?) DictionarySubscriptDefaultMutationArrayOfObjects 3812 3817 +0.1% 1.00x (?) DictionarySubscriptDefaultMutationOfObjects 1397 1400 +0.2% 1.00x (?) DictionarySwap 471 471 +0.0% 1.00x DictionarySwapOfObjects 7468 7462 -0.1% 1.00x (?) DoubleWidthDivision 0 0 +0.0% 1.00x DropFirstAnyCollection 102 101 -1.0% 1.01x DropFirstAnyCollectionLazy 67046 67032 -0.0% 1.00x (?) DropFirstAnySeqCRangeIter 19580 19583 +0.0% 1.00x (?) DropFirstAnySeqCRangeIterLazy 19588 19558 -0.2% 1.00x (?) DropFirstAnySeqCntRange 94 94 +0.0% 1.00x DropFirstAnySeqCntRangeLazy 94 94 +0.0% 1.00x DropFirstAnySequence 5082 5082 +0.0% 1.00x DropFirstAnySequenceLazy 5084 5081 -0.1% 1.00x (?) DropFirstArray 35 35 +0.0% 1.00x DropFirstArrayLazy 35 35 +0.0% 1.00x DropFirstCountableRange 35 35 +0.0% 1.00x DropFirstCountableRangeLazy 35 35 +0.0% 1.00x DropFirstSequence 2699 2698 -0.0% 1.00x (?) DropFirstSequenceLazy 2785 2784 -0.0% 1.00x (?) DropLastAnyCollection 37 37 +0.0% 1.00x DropLastAnyCollectionLazy 22445 22381 -0.3% 1.00x (?) DropLastAnySeqCRangeIter 3914 3914 +0.0% 1.00x DropLastAnySeqCRangeIterLazy 3916 3908 -0.2% 1.00x (?) DropLastAnySeqCntRange 30 29 -3.3% 1.03x DropLastAnySeqCntRangeLazy 30 30 +0.0% 1.00x DropLastAnySequence 5218 5210 -0.2% 1.00x DropLastAnySequenceLazy 5321 5322 +0.0% 1.00x (?) DropLastCountableRange 11 11 +0.0% 1.00x DropLastSequence 642 640 -0.3% 1.00x (?) DropLastSequenceLazy 642 642 +0.0% 1.00x DropWhileAnyCollection 125 125 +0.0% 1.00x DropWhileAnyCollectionLazy 166 166 +0.0% 1.00x DropWhileAnySeqCRangeIter 16093 16069 -0.1% 1.00x (?) DropWhileAnySeqCRangeIterLazy 166 166 +0.0% 1.00x DropWhileAnySeqCntRange 118 118 +0.0% 1.00x DropWhileAnySeqCntRangeLazy 166 166 +0.0% 1.00x DropWhileAnySequence 5787 5788 +0.0% 1.00x (?) DropWhileAnySequenceLazy 1858 1856 -0.1% 1.00x DropWhileArrayLazy 129 129 +0.0% 1.00x DropWhileCountableRange 36 36 +0.0% 1.00x DropWhileCountableRangeLazy 111 111 +0.0% 1.00x DropWhileSequence 1344 1343 -0.1% 1.00x (?) DropWhileSequenceLazy 88 88 +0.0% 1.00x EqualStringSubstring 48 48 +0.0% 1.00x EqualSubstringString 48 48 +0.0% 1.00x EqualSubstringSubstring 48 48 +0.0% 1.00x EqualSubstringSubstringGenericEquatable 48 48 +0.0% 1.00x ErrorHandling 2341 2330 -0.5% 1.00x (?) ExclusivityGlobal 5 5 +0.0% 1.00x ExclusivityIndependent 2 2 +0.0% 1.00x FilterEvenUsingReduce 1333 1329 -0.3% 1.00x (?) FilterEvenUsingReduceInto 151 150 -0.7% 1.01x (?) FrequenciesUsingReduce 6653 6689 +0.5% 0.99x (?) FrequenciesUsingReduceInto 3113 3106 -0.2% 1.00x (?) Hanoi 3334 3328 -0.2% 1.00x (?) HashTest 1765 1763 -0.1% 1.00x (?) Histogram 321 321 +0.0% 1.00x Integrate 334 335 +0.3% 1.00x IterateData 1357 1353 -0.3% 1.00x Join 372 371 -0.3% 1.00x LazilyFilteredArrayContains 36273 36259 -0.0% 1.00x (?) LazilyFilteredArrays 66106 66016 -0.1% 1.00x (?) LazilyFilteredRange 3919 3852 -1.7% 1.02x (?) LessSubstringSubstring 47 48 +2.1% 0.98x LessSubstringSubstringGenericComparable 47 47 +0.0% 1.00x LinkedList 7521 7522 +0.0% 1.00x (?) LuhnAlgoEager 610 610 +0.0% 1.00x LuhnAlgoLazy 611 614 +0.5% 1.00x (?) MapReduce 400 398 -0.5% 1.01x MapReduceAnyCollection 430 429 -0.2% 1.00x MapReduceAnyCollectionShort 2243 2245 +0.1% 1.00x (?) MapReduceClass 3049 3045 -0.1% 1.00x MapReduceClassShort 4582 4595 +0.3% 1.00x MapReduceLazyCollection 13 13 +0.0% 1.00x MapReduceLazyCollectionShort 37 37 +0.0% 1.00x MapReduceLazySequence 86 86 +0.0% 1.00x MapReduceSequence 460 456 -0.9% 1.01x MapReduceShort 2030 2027 -0.1% 1.00x (?) MapReduceShortString 21 21 +0.0% 1.00x MapReduceString 83 83 +0.0% 1.00x Memset 222 215 -3.2% 1.03x (?) MonteCarloE 10290 10267 -0.2% 1.00x MonteCarloPi 43095 42670 -1.0% 1.01x NSDictionaryCastToSwift 5606 5633 +0.5% 1.00x (?) NSError 303 304 +0.3% 1.00x (?) NSStringConversion 306 306 +0.0% 1.00x NibbleSort 3712 3710 -0.1% 1.00x (?) NopDeinit 21597 21576 -0.1% 1.00x (?) ObjectAllocation 185 184 -0.5% 1.01x (?) ObjectiveCBridgeFromNSArrayAnyObjectForced 4344 4534 +4.4% 0.96x (?) ObjectiveCBridgeFromNSArrayAnyObjectToString 38750 40417 +4.3% 0.96x (?) ObjectiveCBridgeFromNSArrayAnyObjectToStringForced 36634 37599 +2.6% 0.97x (?) ObjectiveCBridgeFromNSDictionaryAnyObject 124775 125322 +0.4% 1.00x (?) ObjectiveCBridgeFromNSSetAnyObject 58638 60157 +2.6% 0.97x (?) ObjectiveCBridgeFromNSSetAnyObjectForced 4484 4675 +4.3% 0.96x (?) ObjectiveCBridgeFromNSSetAnyObjectToString 69796 67683 -3.0% 1.03x (?) ObjectiveCBridgeFromNSString 1311 1316 +0.4% 1.00x (?) ObjectiveCBridgeFromNSStringForced 2441 2474 +1.4% 0.99x (?) ObjectiveCBridgeStubDataAppend 4123 3942 -4.4% 1.05x (?) ObjectiveCBridgeStubDateMutation 258 257 -0.4% 1.00x ObjectiveCBridgeStubFromNSDate 4132 4131 -0.0% 1.00x ObjectiveCBridgeStubFromNSString 861 864 +0.3% 1.00x (?) ObjectiveCBridgeStubFromNSStringRef 161 161 +0.0% 1.00x ObjectiveCBridgeStubNSDataAppend 2521 2511 -0.4% 1.00x (?) ObjectiveCBridgeStubNSDateMutationRef 13132 13205 +0.6% 0.99x (?) ObjectiveCBridgeStubToArrayOfNSString 28821 29526 +2.4% 0.98x (?) ObjectiveCBridgeStubToNSDate 15324 15568 +1.6% 0.98x (?) ObjectiveCBridgeStubToNSDateRef 3478 3492 +0.4% 1.00x (?) ObjectiveCBridgeStubToNSString 1519 1519 +0.0% 1.00x ObjectiveCBridgeStubToNSStringRef 115 115 +0.0% 1.00x ObjectiveCBridgeStubURLAppendPath 334340 334017 -0.1% 1.00x (?) ObjectiveCBridgeStubURLAppendPathRef 340455 332243 -2.4% 1.02x ObjectiveCBridgeToNSArray 29360 29455 +0.3% 1.00x (?) ObjectiveCBridgeToNSDictionary 49334 48748 -1.2% 1.01x (?) ObjectiveCBridgeToNSSet 42089 42048 -0.1% 1.00x (?) ObjectiveCBridgeToNSString 1291 1291 +0.0% 1.00x ObserverClosure 2155 2169 +0.6% 0.99x (?) ObserverForwarderStruct 1008 998 -1.0% 1.01x (?) ObserverPartiallyAppliedMethod 3761 3739 -0.6% 1.01x (?) ObserverUnappliedMethod 2366 2372 +0.3% 1.00x (?) OpenClose 123 123 +0.0% 1.00x PartialApplyDynamicType 0 0 +0.0% 1.00x Phonebook 3881 3882 +0.0% 1.00x PointerArithmetics 31517 31497 -0.1% 1.00x PolymorphicCalls 25 25 +0.0% 1.00x PopFrontArray 1945 1946 +0.1% 1.00x (?) PopFrontArrayGeneric 1970 1961 -0.5% 1.00x (?) PopFrontUnsafePointer 9486 9484 -0.0% 1.00x (?) PrefixAnyCollection 102 101 -1.0% 1.01x PrefixAnyCollectionLazy 67014 67024 +0.0% 1.00x (?) PrefixAnySeqCRangeIter 15252 15245 -0.0% 1.00x (?) PrefixAnySeqCRangeIterLazy 15259 15246 -0.1% 1.00x (?) PrefixAnySeqCntRange 94 94 +0.0% 1.00x PrefixAnySeqCntRangeLazy 94 94 +0.0% 1.00x PrefixAnySequence 4497 4496 -0.0% 1.00x (?) PrefixAnySequenceLazy 4483 4491 +0.2% 1.00x PrefixArray 35 35 +0.0% 1.00x PrefixArrayLazy 35 35 +0.0% 1.00x PrefixCountableRange 35 35 +0.0% 1.00x PrefixCountableRangeLazy 35 35 +0.0% 1.00x PrefixSequence 1324 1326 +0.2% 1.00x (?) PrefixSequenceLazy 1413 1414 +0.1% 1.00x (?) PrefixWhileAnyCollection 172 173 +0.6% 0.99x PrefixWhileAnyCollectionLazy 108 108 +0.0% 1.00x PrefixWhileAnySeqCRangeIter 9644 9626 -0.2% 1.00x (?) PrefixWhileAnySeqCRangeIterLazy 90 90 +0.0% 1.00x PrefixWhileAnySeqCntRange 165 165 +0.0% 1.00x PrefixWhileAnySeqCntRangeLazy 108 108 +0.0% 1.00x PrefixWhileAnySequence 10801 10779 -0.2% 1.00x (?) PrefixWhileAnySequenceLazy 1394 1394 +0.0% 1.00x PrefixWhileArray 105 105 +0.0% 1.00x PrefixWhileArrayLazy 70 70 +0.0% 1.00x PrefixWhileCountableRange 46 46 +0.0% 1.00x PrefixWhileCountableRangeLazy 35 35 +0.0% 1.00x PrefixWhileSequence 397 395 -0.5% 1.01x (?) PrefixWhileSequenceLazy 52 52 +0.0% 1.00x Prims 775 770 -0.6% 1.01x (?) PrimsSplit 767 767 +0.0% 1.00x QueueConcrete 1305 1307 +0.2% 1.00x (?) QueueGeneric 1158 1157 -0.1% 1.00x RC4 171 171 +0.0% 1.00x RGBHistogram 2214 2204 -0.5% 1.00x (?) RGBHistogramOfObjects 23815 23773 -0.2% 1.00x (?) RangeAssignment 356 356 +0.0% 1.00x RangeIterationSigned 171 171 +0.0% 1.00x RangeIterationSigned64 200 200 +0.0% 1.00x RangeIterationUnsigned 200 200 +0.0% 1.00x RangeReplaceableCollectionPlusDefault 977 988 +1.1% 0.99x (?) RecursiveOwnedParameter 2347 2348 +0.0% 1.00x (?) RemoveWhereFilterInts 46 46 +0.0% 1.00x RemoveWhereFilterString 294 294 +0.0% 1.00x RemoveWhereFilterStrings 432 432 +0.0% 1.00x RemoveWhereMoveInts 17 17 +0.0% 1.00x RemoveWhereMoveStrings 519 519 +0.0% 1.00x RemoveWhereQuadraticInts 1289 1286 -0.2% 1.00x (?) RemoveWhereQuadraticString 206 206 +0.0% 1.00x RemoveWhereQuadraticStrings 2758 2760 +0.1% 1.00x (?) RemoveWhereSwapInts 19 19 +0.0% 1.00x RemoveWhereSwapStrings 848 849 +0.1% 1.00x (?) ReversedArray 57 57 +0.0% 1.00x ReversedBidirectional 17391 17366 -0.1% 1.00x (?) ReversedDictionary 134 135 +0.7% 0.99x (?) RomanNumbers 114190 119878 +5.0% 0.95x (?) SequenceAlgosAnySequence 10988 11027 +0.4% 1.00x (?) SequenceAlgosArray 1579 1579 +0.0% 1.00x SequenceAlgosContiguousArray 1584 1588 +0.3% 1.00x (?) SequenceAlgosList 1351 1351 +0.0% 1.00x SequenceAlgosRange 4297 4294 -0.1% 1.00x SequenceAlgosUnfoldSequence 1076 1077 +0.1% 1.00x SetExclusiveOr 3452 3440 -0.3% 1.00x (?) SetExclusiveOr_OfObjects 8915 8892 -0.3% 1.00x (?) SetIntersect 319 319 +0.0% 1.00x SetIntersect_OfObjects 1762 1766 +0.2% 1.00x (?) SetIsSubsetOf 298 298 +0.0% 1.00x SetIsSubsetOf_OfObjects 363 363 +0.0% 1.00x SetUnion 3121 3149 +0.9% 0.99x SetUnion_OfObjects 7252 7283 +0.4% 1.00x (?) SevenBoom 1517 1517 +0.0% 1.00x Sim2DArray 599 599 +0.0% 1.00x SortLargeExistentials 6667 6678 +0.2% 1.00x SortLettersInPlace 1124 1124 +0.0% 1.00x SortSortedStrings 1015 1014 -0.1% 1.00x (?) SortStrings 1820 1820 +0.0% 1.00x SortStringsUnicode 2515 2519 +0.2% 1.00x (?) StackPromo 22257 22275 +0.1% 1.00x (?) StrComplexWalk 1559 1559 +0.0% 1.00x StrToInt 2113 2121 +0.4% 1.00x (?) StringAdder 4195 4190 -0.1% 1.00x (?) StringBuilder 1395 1397 +0.1% 1.00x (?) StringComparison_abnormal 741 741 +0.0% 1.00x StringComparison_ascii 991 991 +0.0% 1.00x StringComparison_emoji 794 777 -2.1% 1.02x (?) StringComparison_fastPrenormal 732 731 -0.1% 1.00x (?) StringComparison_latin1 553 552 -0.2% 1.00x StringComparison_longSharedPrefix 899 895 -0.4% 1.00x (?) StringComparison_nonBMPSlowestPrenormal 1560 1531 -1.9% 1.02x StringComparison_slowerPrenormal 1668 1618 -3.0% 1.03x StringComparison_zalgo 113739 113421 -0.3% 1.00x StringEdits 118166 117547 -0.5% 1.01x (?) StringEnumRawValueInitialization 965 958 -0.7% 1.01x (?) StringEqualPointerComparison 286 286 +0.0% 1.00x StringFromLongWholeSubstring 21 21 +0.0% 1.00x StringFromLongWholeSubstringGeneric 10 10 +0.0% 1.00x StringHasPrefixAscii 1288 1288 +0.0% 1.00x StringHasPrefixUnicode 27676 27640 -0.1% 1.00x (?) StringHasSuffixAscii 1431 1433 +0.1% 1.00x (?) StringHasSuffixUnicode 84885 84865 -0.0% 1.00x (?) StringInterpolation 10863 10962 +0.9% 0.99x (?) StringMatch 7287 7286 -0.0% 1.00x (?) StringRemoveDupes 1211 1202 -0.7% 1.01x (?) StringUTF16Builder 2400 2413 +0.5% 0.99x (?) StringWalk 1437 1437 +0.0% 1.00x StringWithCString 43179 43165 -0.0% 1.00x (?) StringWordBuilder 1703 1707 +0.2% 1.00x (?) StringWordBuilderReservingCapacity 1270 1268 -0.2% 1.00x (?) SubstringComparable 41 41 +0.0% 1.00x SubstringEqualString 533 531 -0.4% 1.00x (?) SubstringEquatable 1366 1372 +0.4% 1.00x (?) SubstringFromLongString 10 10 +0.0% 1.00x SubstringFromLongStringGeneric 69 69 +0.0% 1.00x SuffixAnyCollection 43 43 +0.0% 1.00x SuffixAnyCollectionLazy 22378 22416 +0.2% 1.00x (?) SuffixAnySeqCRangeIter 4099 4111 +0.3% 1.00x (?) SuffixAnySeqCRangeIterLazy 4109 4102 -0.2% 1.00x (?) SuffixAnySeqCntRange 35 35 +0.0% 1.00x SuffixAnySeqCntRangeLazy 35 35 +0.0% 1.00x SuffixAnySequence 5215 5213 -0.0% 1.00x (?) SuffixAnySequenceLazy 5298 5300 +0.0% 1.00x (?) SuffixCountableRange 12 12 +0.0% 1.00x SuffixCountableRangeLazy 12 12 +0.0% 1.00x SuffixSequence 3897 3897 +0.0% 1.00x SuffixSequenceLazy 3902 3897 -0.1% 1.00x (?) SumUsingReduce 96 96 +0.0% 1.00x SumUsingReduceInto 96 96 +0.0% 1.00x SuperChars 41486 41321 -0.4% 1.00x (?) TwoSum 986 991 +0.5% 0.99x TypeFlood 0 0 +0.0% 1.00x UTF8Decode 327 327 +0.0% 1.00x Walsh 419 419 +0.0% 1.00x WordCountHistogramASCII 7271 7290 +0.3% 1.00x (?) WordCountHistogramUTF16 24377 25342 +4.0% 0.96x WordCountUniqueASCII 1748 1818 +4.0% 0.96x WordCountUniqueUTF16 16538 16611 +0.4% 1.00x (?) WordSplitASCII 21693 21464 -1.1% 1.01x (?) WordSplitUTF16 23161 23149 -0.1% 1.00x (?) XorLoop 394 395 +0.3% 1.00x (?) Unoptimized (Onone) Regression (9) TEST OLD NEW DELTA SPEEDUP WordCountUniqueUTF16 26909 30099 +11.9% 0.89x (?) ObjectiveCBridgeFromNSSetAnyObjectForced 7158 7887 +10.2% 0.91x (?) CharacterPropertiesFetch 5927 6484 +9.4% 0.91x (?) CharIndexing_chinese_unicodeScalars 275390 299901 +8.9% 0.92x (?) ObjectiveCBridgeFromNSSetAnyObjectToString 75097 81138 +8.0% 0.93x (?) DictionaryBridge 2085 2244 +7.6% 0.93x CharIteration_tweet_unicodeScalars_Backwards 606174 648211 +6.9% 0.94x (?) CharIteration_utf16_unicodeScalars 132997 140780 +5.9% 0.94x StringUTF16Builder 7757 8208 +5.8% 0.95x Improvement (6) TEST OLD NEW DELTA SPEEDUP CharIndexing_russian_unicodeScalars 338061 311345 -7.9% 1.09x (?) ObjectiveCBridgeFromNSArrayAnyObjectForced 9440 8799 -6.8% 1.07x (?) DropFirstAnyCollectionLazy 109972 102762 -6.6% 1.07x (?) StringEdits 362663 339297 -6.4% 1.07x (?) ObjectiveCBridgeStubFromArrayOfNSString 29611 27892 -5.8% 1.06x (?) CharIndexing_chinese_unicodeScalars_Backwards 343159 326791 -4.8% 1.05x (?) No Changes (363) TEST OLD NEW DELTA SPEEDUP AngryPhonebook 5152 5156 +0.1% 1.00x (?) AnyHashableWithAClass 90360 90383 +0.0% 1.00x (?) Array2D 631258 631335 +0.0% 1.00x (?) ArrayAppend 4594 4586 -0.2% 1.00x (?) ArrayAppendArrayOfInt 868 869 +0.1% 1.00x (?) ArrayAppendAscii 40575 40594 +0.0% 1.00x (?) ArrayAppendFromGeneric 878 876 -0.2% 1.00x ArrayAppendGenericStructs 1509 1514 +0.3% 1.00x (?) ArrayAppendLatin1 64174 64138 -0.1% 1.00x (?) ArrayAppendLazyMap 178042 180063 +1.1% 0.99x (?) ArrayAppendOptionals 1510 1514 +0.3% 1.00x (?) ArrayAppendRepeatCol 187835 187719 -0.1% 1.00x (?) ArrayAppendReserved 4280 4279 -0.0% 1.00x (?) ArrayAppendSequence 157196 157016 -0.1% 1.00x (?) ArrayAppendStrings 15365 15379 +0.1% 1.00x (?) ArrayAppendToFromGeneric 872 874 +0.2% 1.00x (?) ArrayAppendToGeneric 879 876 -0.3% 1.00x (?) ArrayAppendUTF16 63630 63648 +0.0% 1.00x (?) ArrayInClass 6256 6254 -0.0% 1.00x (?) ArrayLiteral 1799 1801 +0.1% 1.00x (?) ArrayOfGenericPOD2 1132 1135 +0.3% 1.00x (?) ArrayOfGenericRef 10277 10291 +0.1% 1.00x (?) ArrayOfPOD 842 842 +0.0% 1.00x ArrayOfRef 9467 9510 +0.5% 1.00x (?) ArrayPlusEqualArrayOfInt 873 871 -0.2% 1.00x (?) ArrayPlusEqualFiveElementCollection 240989 241217 +0.1% 1.00x (?) ArrayPlusEqualSingleElementCollection 236901 235678 -0.5% 1.01x (?) ArrayPlusEqualThreeElements 9528 9503 -0.3% 1.00x (?) ArraySubscript 116652 116829 +0.2% 1.00x (?) ArrayValueProp 3697 3698 +0.0% 1.00x (?) ArrayValueProp2 16783 16895 +0.7% 0.99x (?) ArrayValueProp3 4174 4226 +1.2% 0.99x (?) ArrayValueProp4 4162 4189 +0.6% 0.99x (?) BinaryFloatingPointConversionFromBinaryInteger 6310 6240 -1.1% 1.01x (?) BitCount 7874 7877 +0.0% 1.00x (?) ByteSwap 9217 9215 -0.0% 1.00x (?) COWTree 12282 11754 -4.3% 1.04x CSVParsing 2500667 2491196 -0.4% 1.00x (?) CSVParsingAlt 1316982 1324118 +0.5% 0.99x (?) CSVParsingAltIndices 2277002 2235821 -1.8% 1.02x (?) CStringLongAscii 4857 4797 -1.2% 1.01x (?) CStringLongNonAscii 2136 2140 +0.2% 1.00x CStringShortAscii 7546 7528 -0.2% 1.00x (?) Calculator 1506 1505 -0.1% 1.00x (?) CaptureProp 265070 264978 -0.0% 1.00x (?) CharIndexing_ascii_unicodeScalars 376720 373393 -0.9% 1.01x (?) CharIndexing_ascii_unicodeScalars_Backwards 428576 426625 -0.5% 1.00x (?) CharIndexing_japanese_unicodeScalars 441226 444417 +0.7% 0.99x (?) CharIndexing_japanese_unicodeScalars_Backwards 517606 543780 +5.1% 0.95x (?) CharIndexing_korean_unicodeScalars 361857 377805 +4.4% 0.96x (?) CharIndexing_korean_unicodeScalars_Backwards 414745 416511 +0.4% 1.00x (?) CharIndexing_punctuatedJapanese_unicodeScalars 65240 64980 -0.4% 1.00x (?) CharIndexing_punctuatedJapanese_unicodeScalars_Backwards 74478 74179 -0.4% 1.00x (?) CharIndexing_punctuated_unicodeScalars 81357 80515 -1.0% 1.01x (?) CharIndexing_punctuated_unicodeScalars_Backwards 93634 95052 +1.5% 0.99x (?) CharIndexing_russian_unicodeScalars_Backwards 362398 365302 +0.8% 0.99x (?) CharIndexing_tweet_unicodeScalars 739021 720819 -2.5% 1.03x (?) CharIndexing_tweet_unicodeScalars_Backwards 857012 885568 +3.3% 0.97x (?) CharIndexing_utf16_unicodeScalars 312631 310874 -0.6% 1.01x (?) CharIndexing_utf16_unicodeScalars_Backwards 359985 358599 -0.4% 1.00x (?) CharIteration_ascii_unicodeScalars 156699 156979 +0.2% 1.00x (?) CharIteration_ascii_unicodeScalars_Backwards 304533 302870 -0.5% 1.01x (?) CharIteration_chinese_unicodeScalars 118735 118599 -0.1% 1.00x (?) CharIteration_chinese_unicodeScalars_Backwards 231432 232490 +0.5% 1.00x (?) CharIteration_japanese_unicodeScalars 188116 188136 +0.0% 1.00x (?) CharIteration_japanese_unicodeScalars_Backwards 368006 366874 -0.3% 1.00x (?) CharIteration_korean_unicodeScalars 151369 151413 +0.0% 1.00x (?) CharIteration_korean_unicodeScalars_Backwards 295666 295118 -0.2% 1.00x (?) CharIteration_punctuatedJapanese_unicodeScalars 28157 28147 -0.0% 1.00x (?) CharIteration_punctuatedJapanese_unicodeScalars_Backwards 52896 53372 +0.9% 0.99x (?) CharIteration_punctuated_unicodeScalars 35204 35209 +0.0% 1.00x (?) CharIteration_punctuated_unicodeScalars_Backwards 67342 67628 +0.4% 1.00x (?) CharIteration_russian_unicodeScalars 130353 130418 +0.0% 1.00x (?) CharIteration_russian_unicodeScalars_Backwards 254336 255629 +0.5% 0.99x (?) CharIteration_tweet_unicodeScalars 308803 308602 -0.1% 1.00x (?) CharIteration_utf16_unicodeScalars_Backwards 265635 271270 +2.1% 0.98x (?) CharacterLiteralsLarge 6019 6013 -0.1% 1.00x (?) CharacterLiteralsSmall 689 688 -0.1% 1.00x CharacterPropertiesPrecomputed 4438 4364 -1.7% 1.02x (?) CharacterPropertiesStashed 2451 2370 -3.3% 1.03x (?) CharacterPropertiesStashedMemo 5103 5278 +3.4% 0.97x (?) Chars 40482 40571 +0.2% 1.00x (?) ClassArrayGetter 987 988 +0.1% 1.00x (?) Combos 2227 2212 -0.7% 1.01x (?) DictOfArraysToArrayOfDicts 3479 3450 -0.8% 1.01x (?) Dictionary 2767 2767 +0.0% 1.00x Dictionary2 3048 3051 +0.1% 1.00x (?) Dictionary2OfObjects 5746 5890 +2.5% 0.98x (?) Dictionary3 1374 1370 -0.3% 1.00x (?) Dictionary3OfObjects 2333 2406 +3.1% 0.97x (?) Dictionary4 1254 1249 -0.4% 1.00x (?) Dictionary4OfObjects 1948 1869 -4.1% 1.04x DictionaryGroup 4307 4305 -0.0% 1.00x (?) DictionaryGroupOfObjects 7439 7441 +0.0% 1.00x (?) DictionaryLiteral 8813 8788 -0.3% 1.00x (?) DictionaryOfObjects 6429 6628 +3.1% 0.97x DictionaryRemove 19225 19218 -0.0% 1.00x (?) DictionaryRemoveOfObjects 51929 51788 -0.3% 1.00x (?) DictionarySubscriptDefaultMutation 2096 2078 -0.9% 1.01x (?) DictionarySubscriptDefaultMutationArray 2330 2331 +0.0% 1.00x (?) DictionarySubscriptDefaultMutationArrayOfObjects 9171 9134 -0.4% 1.00x (?) DictionarySubscriptDefaultMutationOfObjects 5818 5656 -2.8% 1.03x (?) DictionarySwap 5696 5689 -0.1% 1.00x (?) DictionarySwapOfObjects 22877 22840 -0.2% 1.00x (?) DoubleWidthDivision 0 0 +0.0% 1.00x DropFirstAnyCollection 14494 14534 +0.3% 1.00x (?) DropFirstAnySeqCRangeIter 21585 21636 +0.2% 1.00x (?) DropFirstAnySeqCRangeIterLazy 21629 21618 -0.1% 1.00x (?) DropFirstAnySeqCntRange 14616 14615 -0.0% 1.00x (?) DropFirstAnySeqCntRangeLazy 14552 14590 +0.3% 1.00x (?) DropFirstAnySequence 11767 11753 -0.1% 1.00x (?) DropFirstAnySequenceLazy 11670 11672 +0.0% 1.00x (?) DropFirstArray 4049 4056 +0.2% 1.00x DropFirstArrayLazy 31819 31807 -0.0% 1.00x (?) DropFirstCountableRange 326 326 +0.0% 1.00x DropFirstCountableRangeLazy 28082 28081 -0.0% 1.00x (?) DropFirstSequence 11225 11234 +0.1% 1.00x (?) DropFirstSequenceLazy 11253 11283 +0.3% 1.00x DropLastAnyCollection 4903 4862 -0.8% 1.01x DropLastAnyCollectionLazy 34471 34682 +0.6% 0.99x (?) DropLastAnySeqCRangeIter 38902 38866 -0.1% 1.00x DropLastAnySeqCRangeIterLazy 39012 39049 +0.1% 1.00x (?) DropLastAnySeqCntRange 4879 4881 +0.0% 1.00x (?) DropLastAnySeqCntRangeLazy 4827 4823 -0.1% 1.00x (?) DropLastAnySequence 29676 29651 -0.1% 1.00x (?) DropLastAnySequenceLazy 29652 29578 -0.2% 1.00x DropLastCountableRange 112 113 +0.9% 0.99x DropLastCountableRangeLazy 9310 9373 +0.7% 0.99x (?) DropLastSequence 29486 29474 -0.0% 1.00x (?) DropLastSequenceLazy 29545 29588 +0.1% 1.00x (?) DropWhileAnyCollection 18633 18627 -0.0% 1.00x (?) DropWhileAnyCollectionLazy 20800 20823 +0.1% 1.00x (?) DropWhileAnySeqCRangeIter 23328 23266 -0.3% 1.00x (?) DropWhileAnySeqCRangeIterLazy 20913 20825 -0.4% 1.00x (?) DropWhileAnySeqCntRange 18242 18253 +0.1% 1.00x (?) DropWhileAnySeqCntRangeLazy 20785 20769 -0.1% 1.00x (?) DropWhileAnySequence 13443 13403 -0.3% 1.00x DropWhileAnySequenceLazy 11655 11649 -0.1% 1.00x (?) DropWhileArrayLazy 14617 14607 -0.1% 1.00x (?) DropWhileCountableRange 4183 4179 -0.1% 1.00x DropWhileCountableRangeLazy 20135 20115 -0.1% 1.00x (?) DropWhileSequence 13102 13134 +0.2% 1.00x (?) DropWhileSequenceLazy 10925 10923 -0.0% 1.00x (?) EqualStringSubstring 91 92 +1.1% 0.99x EqualSubstringString 92 92 +0.0% 1.00x EqualSubstringSubstring 92 92 +0.0% 1.00x EqualSubstringSubstringGenericEquatable 97 97 +0.0% 1.00x ErrorHandling 7035 7332 +4.2% 0.96x (?) ExclusivityGlobal 186 186 +0.0% 1.00x ExclusivityIndependent 70 71 +1.4% 0.99x FilterEvenUsingReduce 3671 3684 +0.4% 1.00x FilterEvenUsingReduceInto 1957 1947 -0.5% 1.01x (?) FrequenciesUsingReduce 10711 10407 -2.8% 1.03x (?) FrequenciesUsingReduceInto 5689 5594 -1.7% 1.02x (?) Hanoi 19959 20055 +0.5% 1.00x HashTest 21718 21657 -0.3% 1.00x (?) Histogram 6810 6802 -0.1% 1.00x (?) Integrate 817 817 +0.0% 1.00x IterateData 6083 6081 -0.0% 1.00x (?) Join 1122 1120 -0.2% 1.00x (?) LazilyFilteredArrayContains 802440 799120 -0.4% 1.00x (?) LazilyFilteredArrays 1482250 1481903 -0.0% 1.00x (?) LazilyFilteredRange 496874 496737 -0.0% 1.00x (?) LessSubstringSubstring 92 92 +0.0% 1.00x LessSubstringSubstringGenericComparable 96 96 +0.0% 1.00x LinkedList 32147 32152 +0.0% 1.00x (?) LuhnAlgoEager 4944 5029 +1.7% 0.98x (?) LuhnAlgoLazy 5090 5106 +0.3% 1.00x (?) MapReduce 25935 25934 -0.0% 1.00x (?) MapReduceAnyCollection 25868 25809 -0.2% 1.00x (?) MapReduceAnyCollectionShort 38151 37465 -1.8% 1.02x (?) MapReduceClass 30429 30446 +0.1% 1.00x (?) MapReduceClassShort 41715 41547 -0.4% 1.00x (?) MapReduceLazyCollection 24259 24257 -0.0% 1.00x (?) MapReduceLazyCollectionShort 34827 34829 +0.0% 1.00x (?) MapReduceLazySequence 18938 18937 -0.0% 1.00x (?) MapReduceSequence 30165 30209 +0.1% 1.00x (?) MapReduceShort 37627 38058 +1.1% 0.99x (?) MapReduceShortString 236 239 +1.3% 0.99x (?) MapReduceString 1859 1915 +3.0% 0.97x Memset 44715 44734 +0.0% 1.00x (?) MonteCarloE 1244837 1248003 +0.3% 1.00x (?) MonteCarloPi 5624224 5621708 -0.0% 1.00x (?) NSDictionaryCastToSwift 6764 6839 +1.1% 0.99x (?) NSError 741 720 -2.8% 1.03x (?) NSStringConversion 352 352 +0.0% 1.00x NibbleSort 436348 448658 +2.8% 0.97x NopDeinit 169344 169167 -0.1% 1.00x (?) ObjectAllocation 1386 1386 +0.0% 1.00x ObjectiveCBridgeFromNSArrayAnyObject 22989 24104 +4.9% 0.95x (?) ObjectiveCBridgeFromNSArrayAnyObjectToString 42888 44994 +4.9% 0.95x (?) ObjectiveCBridgeFromNSArrayAnyObjectToStringForced 36538 36709 +0.5% 1.00x (?) ObjectiveCBridgeFromNSDictionaryAnyObject 127888 127193 -0.5% 1.01x (?) ObjectiveCBridgeFromNSSetAnyObject 63363 65927 +4.0% 0.96x (?) ObjectiveCBridgeFromNSString 3172 3157 -0.5% 1.00x (?) ObjectiveCBridgeFromNSStringForced 2779 2784 +0.2% 1.00x (?) ObjectiveCBridgeStubDataAppend 4384 4506 +2.8% 0.97x (?) ObjectiveCBridgeStubDateMutation 632 634 +0.3% 1.00x (?) ObjectiveCBridgeStubFromNSDate 4789 4794 +0.1% 1.00x (?) ObjectiveCBridgeStubFromNSString 924 924 +0.0% 1.00x ObjectiveCBridgeStubFromNSStringRef 199 200 +0.5% 1.00x ObjectiveCBridgeStubNSDataAppend 2984 3004 +0.7% 0.99x (?) ObjectiveCBridgeStubNSDateMutationRef 16802 16452 -2.1% 1.02x (?) ObjectiveCBridgeStubToArrayOfNSString 29040 28927 -0.4% 1.00x (?) ObjectiveCBridgeStubToNSDate 16551 16650 +0.6% 0.99x (?) ObjectiveCBridgeStubToNSDateRef 4151 4090 -1.5% 1.01x ObjectiveCBridgeStubToNSString 1590 1590 +0.0% 1.00x ObjectiveCBridgeStubToNSStringRef 153 153 +0.0% 1.00x ObjectiveCBridgeStubURLAppendPath 344933 334843 -2.9% 1.03x (?) ObjectiveCBridgeStubURLAppendPathRef 356033 353326 -0.8% 1.01x (?) ObjectiveCBridgeToNSArray 29426 29948 +1.8% 0.98x (?) ObjectiveCBridgeToNSDictionary 51893 51222 -1.3% 1.01x (?) ObjectiveCBridgeToNSSet 43779 42952 -1.9% 1.02x (?) ObjectiveCBridgeToNSString 1362 1362 +0.0% 1.00x ObserverClosure 6469 6508 +0.6% 0.99x (?) ObserverForwarderStruct 4568 4570 +0.0% 1.00x (?) ObserverPartiallyAppliedMethod 8180 8138 -0.5% 1.01x (?) ObserverUnappliedMethod 8219 8225 +0.1% 1.00x (?) OpenClose 541 543 +0.4% 1.00x (?) PartialApplyDynamicType 41361 41277 -0.2% 1.00x (?) Phonebook 19959 20064 +0.5% 0.99x PointerArithmetics 114533 114584 +0.0% 1.00x (?) PolymorphicCalls 6106 6097 -0.1% 1.00x (?) PopFrontArray 4713 4696 -0.4% 1.00x (?) PopFrontArrayGeneric 5735 5719 -0.3% 1.00x (?) PopFrontUnsafePointer 11251 11343 +0.8% 0.99x (?) PrefixAnyCollection 14527 14691 +1.1% 0.99x (?) PrefixAnyCollectionLazy 105963 107133 +1.1% 0.99x (?) PrefixAnySeqCRangeIter 17398 17417 +0.1% 1.00x (?) PrefixAnySeqCRangeIterLazy 17437 17408 -0.2% 1.00x (?) PrefixAnySeqCntRange 14543 14606 +0.4% 1.00x (?) PrefixAnySeqCntRangeLazy 14421 14438 +0.1% 1.00x (?) PrefixAnySequence 9833 9764 -0.7% 1.01x PrefixAnySequenceLazy 9816 9777 -0.4% 1.00x (?) PrefixArray 4054 4050 -0.1% 1.00x (?) PrefixArrayLazy 31930 31753 -0.6% 1.01x (?) PrefixCountableRange 326 326 +0.0% 1.00x PrefixCountableRangeLazy 27955 27925 -0.1% 1.00x (?) PrefixSequence 9635 9602 -0.3% 1.00x (?) PrefixSequenceLazy 9541 9583 +0.4% 1.00x (?) PrefixWhileAnyCollection 25918 25926 +0.0% 1.00x (?) PrefixWhileAnyCollectionLazy 17010 17131 +0.7% 0.99x (?) PrefixWhileAnySeqCRangeIter 33980 33947 -0.1% 1.00x (?) PrefixWhileAnySeqCRangeIterLazy 17073 17086 +0.1% 1.00x (?) PrefixWhileAnySeqCntRange 25854 25874 +0.1% 1.00x PrefixWhileAnySeqCntRangeLazy 17232 17234 +0.0% 1.00x (?) PrefixWhileAnySequence 27034 26956 -0.3% 1.00x (?) PrefixWhileAnySequenceLazy 10461 10440 -0.2% 1.00x (?) PrefixWhileArray 10919 10935 +0.1% 1.00x (?) PrefixWhileArrayLazy 12933 12926 -0.1% 1.00x (?) PrefixWhileCountableRange 11761 11762 +0.0% 1.00x (?) PrefixWhileCountableRangeLazy 17004 17031 +0.2% 1.00x PrefixWhileSequence 26608 26673 +0.2% 1.00x PrefixWhileSequenceLazy 10126 10137 +0.1% 1.00x (?) Prims 9815 9752 -0.6% 1.01x (?) PrimsSplit 9689 9822 +1.4% 0.99x (?) QueueConcrete 15272 15270 -0.0% 1.00x (?) QueueGeneric 19849 19598 -1.3% 1.01x (?) RC4 16685 16808 +0.7% 0.99x (?) RGBHistogram 26680 26560 -0.4% 1.00x (?) RGBHistogramOfObjects 96151 96077 -0.1% 1.00x (?) RangeAssignment 2198 2211 +0.6% 0.99x RangeIterationSigned 16754 16734 -0.1% 1.00x (?) RangeIterationSigned64 39903 40180 +0.7% 0.99x (?) RangeIterationUnsigned 35643 35616 -0.1% 1.00x (?) RangeReplaceableCollectionPlusDefault 8195 8414 +2.7% 0.97x RecursiveOwnedParameter 7939 7944 +0.1% 1.00x (?) RemoveWhereFilterInts 2119 2117 -0.1% 1.00x RemoveWhereFilterString 1516 1515 -0.1% 1.00x (?) RemoveWhereFilterStrings 2982 2984 +0.1% 1.00x RemoveWhereMoveInts 3338 3336 -0.1% 1.00x RemoveWhereMoveStrings 4037 4151 +2.8% 0.97x (?) RemoveWhereQuadraticInts 7926 7924 -0.0% 1.00x (?) RemoveWhereQuadraticString 2338 2337 -0.0% 1.00x (?) RemoveWhereQuadraticStrings 9602 9593 -0.1% 1.00x (?) RemoveWhereSwapInts 5982 6033 +0.9% 0.99x (?) RemoveWhereSwapStrings 6792 6798 +0.1% 1.00x (?) ReversedArray 13917 13909 -0.1% 1.00x (?) ReversedBidirectional 44528 44118 -0.9% 1.01x (?) ReversedDictionary 23602 23541 -0.3% 1.00x (?) RomanNumbers 1352351 1374041 +1.6% 0.98x (?) SequenceAlgosAnySequence 11767 11725 -0.4% 1.00x (?) SequenceAlgosArray 800057 799940 -0.0% 1.00x (?) SequenceAlgosContiguousArray 341107 338947 -0.6% 1.01x (?) SequenceAlgosList 8395 8397 +0.0% 1.00x (?) SequenceAlgosRange 1144611 1143780 -0.1% 1.00x (?) SequenceAlgosUnfoldSequence 5965 5960 -0.1% 1.00x (?) SetExclusiveOr 18219 18169 -0.3% 1.00x (?) SetExclusiveOr_OfObjects 42394 42377 -0.0% 1.00x (?) SetIntersect 9053 9040 -0.1% 1.00x (?) SetIntersect_OfObjects 12951 12946 -0.0% 1.00x (?) SetIsSubsetOf 1459 1458 -0.1% 1.00x (?) SetIsSubsetOf_OfObjects 1648 1652 +0.2% 1.00x (?) SetUnion 12247 12261 +0.1% 1.00x (?) SetUnion_OfObjects 29508 29898 +1.3% 0.99x SevenBoom 1663 1673 +0.6% 0.99x (?) Sim2DArray 43679 43653 -0.1% 1.00x (?) SortLargeExistentials 15458 15441 -0.1% 1.00x (?) SortLettersInPlace 2736 2711 -0.9% 1.01x SortSortedStrings 1072 1070 -0.2% 1.00x (?) SortStrings 2129 2125 -0.2% 1.00x SortStringsUnicode 2755 2760 +0.2% 1.00x (?) StackPromo 93173 92716 -0.5% 1.00x (?) StaticArray 2615 2617 +0.1% 1.00x (?) StrComplexWalk 6434 6440 +0.1% 1.00x (?) StrToInt 58251 58350 +0.2% 1.00x (?) StringAdder 4785 4791 +0.1% 1.00x (?) StringBuilder 6923 6990 +1.0% 0.99x StringBuilderLong 2726 2733 +0.3% 1.00x (?) StringComparison_abnormal 1342 1340 -0.1% 1.00x (?) StringComparison_ascii 9790 9765 -0.3% 1.00x (?) StringComparison_emoji 2125 2122 -0.1% 1.00x (?) StringComparison_fastPrenormal 5374 5317 -1.1% 1.01x (?) StringComparison_latin1 4180 4183 +0.1% 1.00x (?) StringComparison_longSharedPrefix 2501 2500 -0.0% 1.00x (?) StringComparison_nonBMPSlowestPrenormal 3923 3928 +0.1% 1.00x (?) StringComparison_slowerPrenormal 4405 4404 -0.0% 1.00x (?) StringComparison_zalgo 115299 115627 +0.3% 1.00x StringEnumRawValueInitialization 17405 17488 +0.5% 1.00x (?) StringEqualPointerComparison 3557 3560 +0.1% 1.00x (?) StringFromLongWholeSubstring 24 24 +0.0% 1.00x StringFromLongWholeSubstringGeneric 205 207 +1.0% 0.99x (?) StringHasPrefixAscii 3068 3066 -0.1% 1.00x (?) StringHasPrefixUnicode 29206 29538 +1.1% 0.99x (?) StringHasSuffixAscii 3143 3143 +0.0% 1.00x StringHasSuffixUnicode 86907 86628 -0.3% 1.00x (?) StringInterpolation 13309 13337 +0.2% 1.00x (?) StringMatch 34432 34131 -0.9% 1.01x (?) StringRemoveDupes 1417 1431 +1.0% 0.99x (?) StringWalk 13196 13150 -0.3% 1.00x (?) StringWithCString 39447 39427 -0.1% 1.00x (?) StringWordBuilder 2733 2724 -0.3% 1.00x (?) StringWordBuilderReservingCapacity 2318 2320 +0.1% 1.00x (?) SubstringComparable 2108 2103 -0.2% 1.00x (?) SubstringEqualString 1912 1916 +0.2% 1.00x (?) SubstringEquatable 6139 6107 -0.5% 1.01x (?) SubstringFromLongString 24 25 +4.2% 0.96x SubstringFromLongStringGeneric 105 105 +0.0% 1.00x SuffixAnyCollection 4849 4855 +0.1% 1.00x (?) SuffixAnyCollectionLazy 34297 35831 +4.5% 0.96x (?) SuffixAnySeqCRangeIter 34749 34706 -0.1% 1.00x (?) SuffixAnySeqCRangeIterLazy 34722 34827 +0.3% 1.00x (?) SuffixAnySeqCntRange 4904 4908 +0.1% 1.00x (?) SuffixAnySeqCntRangeLazy 4831 4832 +0.0% 1.00x (?) SuffixAnySequence 25551 25527 -0.1% 1.00x (?) SuffixAnySequenceLazy 25475 25498 +0.1% 1.00x (?) SuffixCountableRange 112 112 +0.0% 1.00x SuffixCountableRangeLazy 9316 9319 +0.0% 1.00x (?) SuffixSequence 25672 25630 -0.2% 1.00x (?) SuffixSequenceLazy 25425 25407 -0.1% 1.00x SumUsingReduce 167649 167787 +0.1% 1.00x (?) SumUsingReduceInto 162345 162348 +0.0% 1.00x (?) SuperChars 125659 124863 -0.6% 1.01x (?) TwoSum 4255 4280 +0.6% 0.99x (?) TypeFlood 163 162 -0.6% 1.01x (?) UTF8Decode 31010 31085 +0.2% 1.00x Walsh 12069 12059 -0.1% 1.00x (?) WordCountHistogramASCII 46795 46884 +0.2% 1.00x (?) WordCountHistogramUTF16 69554 72099 +3.7% 0.96x (?) WordCountUniqueASCII 8194 8206 +0.1% 1.00x (?) WordSplitASCII 26057 27249 +4.6% 0.96x WordSplitUTF16 27513 28045 +1.9% 0.98x (?) XorLoop 24847 24841 -0.0% 1.00x (?) Hardware Overview Model Name: Mac Pro Model Identifier: MacPro6,1 Processor Name: 12-Core Intel Xeon E5 Processor Speed: 2.7 GHz Number of Processors: 1 Total Number of Cores: 12 L2 Cache (per Core): 256 KB L3 Cache: 30 MB Memory: 64 GB
gharchive/pull-request
2018-03-05T18:20:56
2025-04-01T04:56:01.618752
{ "authors": [ "atrick", "swift-ci" ], "repo": "apple/swift", "url": "https://github.com/apple/swift/pull/14985", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
306998181
[benchmark] Add a benchmark for .lazy.filter,map chaining... ... and functionally equivalent single call to .lazy.compactMap. See https://forums.swift.org/t/introduce-lazy-version-of-compactmap/9835 for more info. @swift-ci Please smoke test @swift-ci Please smoke benchmark /cc @tellowkrinkle Related to https://github.com/apple/swift/pull/14841 Build comment file: Optimized (O) Regression (16) TEST OLD NEW DELTA SPEEDUP MapReduceLazyCollectionShort 34 38 +11.8% 0.89x StringBuilderLong 1082 1199 +10.8% 0.90x (?) StringEqualPointerComparison 286 315 +10.1% 0.91x CharIteration_chinese_unicodeScalars 15456 16850 +9.0% 0.92x CharIteration_tweet_unicodeScalars 40659 44106 +8.5% 0.92x RemoveWhereMoveInts 14 15 +7.1% 0.93x CharIteration_punctuated_unicodeScalars 4603 4931 +7.1% 0.93x CharIteration_korean_unicodeScalars 19791 21185 +7.0% 0.93x CharIteration_russian_unicodeScalars 17030 18221 +7.0% 0.93x NopDeinit 30188 32262 +6.9% 0.94x CharIteration_ascii_unicodeScalars 20438 21831 +6.8% 0.94x CharIteration_japanese_unicodeScalars 24478 26120 +6.7% 0.94x Chars 975 1036 +6.3% 0.94x CharIteration_punctuatedJapanese_unicodeScalars 3679 3905 +6.1% 0.94x ObjectiveCBridgeFromNSArrayAnyObject 23862 25211 +5.7% 0.95x (?) ObjectiveCBridgeStubToNSStringRef 112 118 +5.4% 0.95x Improvement (32) TEST OLD NEW DELTA SPEEDUP Sim2DArray 599 417 -30.4% 1.44x StringBuilderWithLongSubstring 1881 1346 -28.4% 1.40x (?) RangeIterationSigned 200 171 -14.5% 1.17x DataSubscript 259 237 -8.5% 1.09x CharIndexing_japanese_unicodeScalars 20338 18627 -8.4% 1.09x PointerArithmetics 34391 31516 -8.4% 1.09x CharIndexing_korean_unicodeScalars 16463 15091 -8.3% 1.09x DropLastCountableRange 12 11 -8.3% 1.09x CharIndexing_russian_unicodeScalars 14141 12963 -8.3% 1.09x CharIndexing_ascii_unicodeScalars 16968 15567 -8.3% 1.09x CharIndexing_chinese_unicodeScalars 12856 11797 -8.2% 1.09x OpenClose 137 126 -8.0% 1.09x CharIteration_utf16_unicodeScalars_Backwards 19435 17881 -8.0% 1.09x CharIndexing_punctuated_unicodeScalars 3833 3550 -7.4% 1.08x CharIndexing_tweet_unicodeScalars 33080 30651 -7.3% 1.08x DataReplaceMediumBuffer 13133 12176 -7.3% 1.08x (?) DropLastAnySeqCntRange 14 13 -7.1% 1.08x CharIndexing_punctuatedJapanese_unicodeScalars 3061 2844 -7.1% 1.08x DataReplaceMedium 11812 11024 -6.7% 1.07x (?) DataAppendArray 6585 6181 -6.1% 1.07x (?) DataReplaceLarge 42464 39903 -6.0% 1.06x (?) MapReduceAnyCollection 429 404 -5.8% 1.06x CharIteration_tweet_unicodeScalars_Backwards 30767 29144 -5.3% 1.06x CharIteration_korean_unicodeScalars_Backwards 15108 14314 -5.3% 1.06x CharIteration_ascii_unicodeScalars_Backwards 15579 14765 -5.2% 1.06x CharIteration_japanese_unicodeScalars_Backwards 18644 17674 -5.2% 1.05x CharIteration_russian_unicodeScalars_Backwards 12980 12308 -5.2% 1.05x CharIteration_chinese_unicodeScalars_Backwards 11789 11185 -5.1% 1.05x StringComparison_ascii 929 882 -5.1% 1.05x CharIteration_punctuated_unicodeScalars_Backwards 3530 3353 -5.0% 1.05x SumUsingReduceInto 102 97 -4.9% 1.05x CharIteration_punctuatedJapanese_unicodeScalars_Backwards 2821 2685 -4.8% 1.05x No Changes (375) TEST OLD NEW DELTA SPEEDUP AngryPhonebook 3523 3554 +0.9% 0.99x (?) AnyHashableWithAClass 81035 81069 +0.0% 1.00x (?) Array2D 2550 2551 +0.0% 1.00x (?) ArrayAppend 1109 1112 +0.3% 1.00x (?) ArrayAppendArrayOfInt 795 795 +0.0% 1.00x ArrayAppendAscii 13249 13248 -0.0% 1.00x (?) ArrayAppendFromGeneric 797 796 -0.1% 1.00x (?) ArrayAppendGenericStructs 1428 1423 -0.4% 1.00x ArrayAppendLatin1 38001 38031 +0.1% 1.00x (?) ArrayAppendLazyMap 1345 1339 -0.4% 1.00x (?) ArrayAppendOptionals 1426 1423 -0.2% 1.00x ArrayAppendRepeatCol 1357 1340 -1.3% 1.01x ArrayAppendReserved 839 842 +0.4% 1.00x (?) ArrayAppendSequence 1120 1121 +0.1% 1.00x (?) ArrayAppendStrings 15176 15145 -0.2% 1.00x (?) ArrayAppendToFromGeneric 799 797 -0.3% 1.00x ArrayAppendToGeneric 787 803 +2.0% 0.98x (?) ArrayAppendUTF16 37689 37735 +0.1% 1.00x (?) ArrayInClass 85 85 +0.0% 1.00x ArrayLiteral 0 0 +0.0% 1.00x ArrayOfGenericPOD2 150 151 +0.7% 0.99x (?) ArrayOfGenericRef 4398 4397 -0.0% 1.00x (?) ArrayOfPOD 183 183 +0.0% 1.00x ArrayOfRef 4398 4396 -0.0% 1.00x (?) ArrayPlusEqualArrayOfInt 797 798 +0.1% 1.00x (?) ArrayPlusEqualFiveElementCollection 6558 6686 +2.0% 0.98x ArrayPlusEqualSingleElementCollection 5005 5111 +2.1% 0.98x (?) ArrayPlusEqualThreeElements 2009 2010 +0.0% 1.00x (?) ArraySubscript 1528 1539 +0.7% 0.99x (?) ArrayValueProp 8 8 +0.0% 1.00x ArrayValueProp2 8 8 +0.0% 1.00x ArrayValueProp3 8 8 +0.0% 1.00x ArrayValueProp4 8 8 +0.0% 1.00x BinaryFloatingPointConversionFromBinaryInteger 40 40 +0.0% 1.00x BinaryFloatingPointPropertiesBinade 31 31 +0.0% 1.00x BinaryFloatingPointPropertiesNextUp 28 28 +0.0% 1.00x BinaryFloatingPointPropertiesUlp 37 37 +0.0% 1.00x BitCount 211 211 +0.0% 1.00x ByteSwap 106 107 +0.9% 0.99x (?) COWTree 5681 5513 -3.0% 1.03x (?) CSVParsing 755612 760650 +0.7% 0.99x (?) CSVParsingAlt 725460 730518 +0.7% 0.99x (?) CSVParsingAltIndices 340516 340298 -0.1% 1.00x (?) CStringLongAscii 4745 4737 -0.2% 1.00x (?) CStringLongNonAscii 2222 2213 -0.4% 1.00x (?) CStringShortAscii 5530 5521 -0.2% 1.00x (?) Calculator 666 651 -2.3% 1.02x CaptureProp 4111 4114 +0.1% 1.00x (?) CharIndexing_ascii_unicodeScalars_Backwards 15714 15833 +0.8% 0.99x (?) CharIndexing_chinese_unicodeScalars_Backwards 11890 12008 +1.0% 0.99x CharIndexing_japanese_unicodeScalars_Backwards 18774 18944 +0.9% 0.99x CharIndexing_korean_unicodeScalars_Backwards 15215 15346 +0.9% 0.99x CharIndexing_punctuatedJapanese_unicodeScalars_Backwards 2859 2862 +0.1% 1.00x (?) CharIndexing_punctuated_unicodeScalars_Backwards 3577 3589 +0.3% 1.00x CharIndexing_russian_unicodeScalars_Backwards 13074 13191 +0.9% 0.99x CharIndexing_tweet_unicodeScalars_Backwards 30876 31314 +1.4% 0.99x CharIndexing_utf16_unicodeScalars 21818 21091 -3.3% 1.03x CharIndexing_utf16_unicodeScalars_Backwards 21657 21583 -0.3% 1.00x CharIteration_utf16_unicodeScalars 27877 27983 +0.4% 1.00x CharacterLiteralsLarge 5752 5760 +0.1% 1.00x (?) CharacterLiteralsSmall 217 220 +1.4% 0.99x CharacterPropertiesFetch 4576 4610 +0.7% 0.99x (?) CharacterPropertiesPrecomputed 1251 1252 +0.1% 1.00x (?) CharacterPropertiesStashed 1450 1438 -0.8% 1.01x (?) CharacterPropertiesStashedMemo 1681 1697 +1.0% 0.99x (?) ClassArrayGetter 15 15 +0.0% 1.00x Combos 460 460 +0.0% 1.00x DataAccessBytes 1139 1145 +0.5% 0.99x DataAppendBytes 6311 6120 -3.0% 1.03x (?) DataAppendDataLargeToLarge 69622 69282 -0.5% 1.00x (?) DataAppendDataLargeToMedium 36518 36492 -0.1% 1.00x (?) DataAppendDataLargeToSmall 35615 35078 -1.5% 1.02x (?) DataAppendDataMediumToLarge 39126 38568 -1.4% 1.01x (?) DataAppendDataMediumToMedium 7712 7819 +1.4% 0.99x (?) DataAppendDataMediumToSmall 7043 6862 -2.6% 1.03x (?) DataAppendDataSmallToLarge 37556 37549 -0.0% 1.00x (?) DataAppendDataSmallToMedium 7288 7213 -1.0% 1.01x (?) DataAppendDataSmallToSmall 6832 6642 -2.8% 1.03x (?) DataAppendSequence 18990 19732 +3.9% 0.96x (?) DataCopyBytes 2540 2508 -1.3% 1.01x (?) DataCount 38 37 -2.6% 1.03x DataMutateBytes 4221 4242 +0.5% 1.00x (?) DataReplaceLargeBuffer 57826 57919 +0.2% 1.00x (?) DataReplaceSmall 8295 8713 +5.0% 0.95x (?) DataReplaceSmallBuffer 8738 9058 +3.7% 0.96x (?) DataReset 3154 3194 +1.3% 0.99x (?) DataSetCount 836 856 +2.4% 0.98x DictOfArraysToArrayOfDicts 824 829 +0.6% 0.99x (?) Dictionary 721 716 -0.7% 1.01x Dictionary2 1818 1816 -0.1% 1.00x (?) Dictionary2OfObjects 3355 3342 -0.4% 1.00x (?) Dictionary3 427 427 +0.0% 1.00x Dictionary3OfObjects 890 890 +0.0% 1.00x Dictionary4 852 851 -0.1% 1.00x (?) Dictionary4OfObjects 937 946 +1.0% 0.99x (?) DictionaryBridge 2016 2084 +3.4% 0.97x DictionaryCopy 109879 107660 -2.0% 1.02x (?) DictionaryFilter 108294 106601 -1.6% 1.02x DictionaryGroup 267 279 +4.5% 0.96x DictionaryGroupOfObjects 2124 2097 -1.3% 1.01x DictionaryLiteral 2034 2000 -1.7% 1.02x DictionaryOfObjects 2636 2644 +0.3% 1.00x DictionaryRemove 5333 5276 -1.1% 1.01x DictionaryRemoveOfObjects 28878 28610 -0.9% 1.01x (?) DictionarySubscriptDefaultMutation 301 292 -3.0% 1.03x DictionarySubscriptDefaultMutationArray 642 642 +0.0% 1.00x DictionarySubscriptDefaultMutationArrayOfObjects 4194 4201 +0.2% 1.00x (?) DictionarySubscriptDefaultMutationOfObjects 1973 1959 -0.7% 1.01x (?) DictionarySwap 1370 1356 -1.0% 1.01x DictionarySwapAt 6970 7196 +3.2% 0.97x DictionarySwapAtOfObjects 52840 52964 +0.2% 1.00x (?) DictionarySwapOfObjects 9546 9762 +2.3% 0.98x DoubleWidthDivision 0 0 +0.0% 1.00x DropFirstAnyCollection 84 84 +0.0% 1.00x DropFirstAnyCollectionLazy 60652 60934 +0.5% 1.00x (?) DropFirstAnySeqCRangeIter 17032 17054 +0.1% 1.00x (?) DropFirstAnySeqCRangeIterLazy 17046 17042 -0.0% 1.00x (?) DropFirstAnySeqCntRange 29 29 +0.0% 1.00x DropFirstAnySeqCntRangeLazy 28 29 +3.6% 0.97x DropFirstAnySequence 4883 4887 +0.1% 1.00x (?) DropFirstAnySequenceLazy 4877 4901 +0.5% 1.00x DropFirstArray 35 35 +0.0% 1.00x DropFirstArrayLazy 35 35 +0.0% 1.00x DropFirstCountableRange 35 35 +0.0% 1.00x DropFirstCountableRangeLazy 35 35 +0.0% 1.00x DropFirstSequence 2681 2682 +0.0% 1.00x (?) DropFirstSequenceLazy 2778 2772 -0.2% 1.00x (?) DropLastAnyCollection 31 31 +0.0% 1.00x DropLastAnyCollectionLazy 20297 20226 -0.3% 1.00x (?) DropLastAnySeqCRangeIter 3595 3579 -0.4% 1.00x DropLastAnySeqCRangeIterLazy 3595 3583 -0.3% 1.00x DropLastAnySeqCntRangeLazy 14 14 +0.0% 1.00x DropLastAnySequence 4977 4938 -0.8% 1.01x (?) DropLastAnySequenceLazy 5070 4997 -1.4% 1.01x (?) DropLastCountableRangeLazy 11 11 +0.0% 1.00x DropLastSequence 653 629 -3.7% 1.04x DropLastSequenceLazy 653 629 -3.7% 1.04x DropWhileAnyCollection 107 107 +0.0% 1.00x DropWhileAnyCollectionLazy 125 125 +0.0% 1.00x DropWhileAnySeqCRangeIter 14447 14484 +0.3% 1.00x (?) DropWhileAnySeqCRangeIterLazy 90 90 +0.0% 1.00x DropWhileAnySeqCntRange 36 36 +0.0% 1.00x DropWhileAnySeqCntRangeLazy 90 90 +0.0% 1.00x DropWhileAnySequence 5581 5580 -0.0% 1.00x (?) DropWhileAnySequenceLazy 1858 1857 -0.1% 1.00x (?) DropWhileArrayLazy 88 88 +0.0% 1.00x DropWhileCountableRange 35 35 +0.0% 1.00x DropWhileCountableRangeLazy 105 105 +0.0% 1.00x DropWhileSequence 1326 1326 +0.0% 1.00x DropWhileSequenceLazy 88 88 +0.0% 1.00x EqualStringSubstring 49 49 +0.0% 1.00x EqualSubstringString 49 49 +0.0% 1.00x EqualSubstringSubstring 49 48 -2.0% 1.02x EqualSubstringSubstringGenericEquatable 49 49 +0.0% 1.00x ErrorHandling 2524 2461 -2.5% 1.03x (?) ExclusivityGlobal 5 5 +0.0% 1.00x ExclusivityIndependent 2 2 +0.0% 1.00x FilterEvenUsingReduce 1323 1311 -0.9% 1.01x (?) FilterEvenUsingReduceInto 149 148 -0.7% 1.01x FloatingPointPrinting_Double_description_small 80967 79552 -1.7% 1.02x FloatingPointPrinting_Double_description_uniform 137015 135303 -1.2% 1.01x FloatingPointPrinting_Double_interpolated 184605 182004 -1.4% 1.01x (?) FloatingPointPrinting_Float80_description_small 93205 91699 -1.6% 1.02x FloatingPointPrinting_Float80_description_uniform 1705051 1706668 +0.1% 1.00x FloatingPointPrinting_Float80_interpolated 1750653 1752467 +0.1% 1.00x (?) FloatingPointPrinting_Float_description_small 42343 43848 +3.6% 0.97x FloatingPointPrinting_Float_description_uniform 44675 44573 -0.2% 1.00x (?) FloatingPointPrinting_Float_interpolated 85074 85142 +0.1% 1.00x (?) FrequenciesUsingReduce 6963 7079 +1.7% 0.98x (?) FrequenciesUsingReduceInto 3650 3657 +0.2% 1.00x (?) Hanoi 2139 2207 +3.2% 0.97x HashTest 1780 1776 -0.2% 1.00x (?) Histogram 814 814 +0.0% 1.00x Integrate 338 350 +3.6% 0.97x IterateData 1931 1977 +2.4% 0.98x Join 352 352 +0.0% 1.00x LazilyFilteredArrayContains 38685 37860 -2.1% 1.02x LazilyFilteredArrays 67404 65146 -3.3% 1.03x (?) LazilyFilteredRange 3881 3840 -1.1% 1.01x LessSubstringSubstring 49 49 +0.0% 1.00x LessSubstringSubstringGenericComparable 51 53 +3.9% 0.96x LinkedList 7610 7539 -0.9% 1.01x LuhnAlgoEager 538 536 -0.4% 1.00x (?) LuhnAlgoLazy 538 538 +0.0% 1.00x MapReduce 383 398 +3.9% 0.96x MapReduceAnyCollectionShort 2277 2280 +0.1% 1.00x (?) MapReduceClass 3029 3013 -0.5% 1.01x (?) MapReduceClassShort 4548 4606 +1.3% 0.99x (?) MapReduceLazyCollection 13 13 +0.0% 1.00x MapReduceLazySequence 86 86 +0.0% 1.00x MapReduceSequence 449 450 +0.2% 1.00x (?) MapReduceShort 2023 2007 -0.8% 1.01x (?) MapReduceShortString 21 21 +0.0% 1.00x MapReduceString 77 79 +2.6% 0.97x Memset 219 220 +0.5% 1.00x (?) MonteCarloE 10253 10269 +0.2% 1.00x MonteCarloPi 42726 42659 -0.2% 1.00x NSDictionaryCastToSwift 5581 5535 -0.8% 1.01x (?) NSError 304 302 -0.7% 1.01x (?) NSStringConversion 401 401 +0.0% 1.00x NibbleSort 3686 3680 -0.2% 1.00x ObjectAllocation 135 136 +0.7% 0.99x ObjectiveCBridgeFromNSArrayAnyObjectForced 4532 4483 -1.1% 1.01x (?) ObjectiveCBridgeFromNSArrayAnyObjectToString 42740 44687 +4.6% 0.96x (?) ObjectiveCBridgeFromNSArrayAnyObjectToStringForced 39780 41623 +4.6% 0.96x (?) ObjectiveCBridgeFromNSDictionaryAnyObject 106079 111458 +5.1% 0.95x (?) ObjectiveCBridgeFromNSSetAnyObject 48927 49153 +0.5% 1.00x (?) ObjectiveCBridgeFromNSSetAnyObjectForced 4209 4124 -2.0% 1.02x (?) ObjectiveCBridgeFromNSSetAnyObjectToString 78137 78266 +0.2% 1.00x (?) ObjectiveCBridgeFromNSString 1345 1291 -4.0% 1.04x (?) ObjectiveCBridgeFromNSStringForced 2460 2445 -0.6% 1.01x (?) ObjectiveCBridgeStubDataAppend 10660 10863 +1.9% 0.98x ObjectiveCBridgeStubDateMutation 401 400 -0.2% 1.00x ObjectiveCBridgeStubFromArrayOfNSString 31950 32671 +2.3% 0.98x (?) ObjectiveCBridgeStubFromNSDate 6040 5998 -0.7% 1.01x (?) ObjectiveCBridgeStubFromNSString 953 963 +1.0% 0.99x ObjectiveCBridgeStubFromNSStringRef 159 159 +0.0% 1.00x ObjectiveCBridgeStubNSDataAppend 2614 2623 +0.3% 1.00x (?) ObjectiveCBridgeStubNSDateMutationRef 13129 13333 +1.6% 0.98x (?) ObjectiveCBridgeStubToArrayOfNSString 27580 27962 +1.4% 0.99x (?) ObjectiveCBridgeStubToNSDate 15380 16074 +4.5% 0.96x (?) ObjectiveCBridgeStubToNSDateRef 3456 3449 -0.2% 1.00x (?) ObjectiveCBridgeStubToNSString 1560 1623 +4.0% 0.96x ObjectiveCBridgeStubURLAppendPath 336757 353786 +5.1% 0.95x (?) ObjectiveCBridgeStubURLAppendPathRef 357378 346134 -3.1% 1.03x (?) ObjectiveCBridgeToNSArray 28030 28039 +0.0% 1.00x (?) ObjectiveCBridgeToNSDictionary 43402 42123 -2.9% 1.03x (?) ObjectiveCBridgeToNSSet 33380 34346 +2.9% 0.97x (?) ObjectiveCBridgeToNSString 1285 1300 +1.2% 0.99x ObserverClosure 2167 2154 -0.6% 1.01x (?) ObserverForwarderStruct 1235 1296 +4.9% 0.95x ObserverPartiallyAppliedMethod 3732 3748 +0.4% 1.00x ObserverUnappliedMethod 2567 2665 +3.8% 0.96x PartialApplyDynamicType 0 0 +0.0% 1.00x Phonebook 3924 3859 -1.7% 1.02x PolymorphicCalls 25 25 +0.0% 1.00x PopFrontArray 1887 1880 -0.4% 1.00x (?) PopFrontArrayGeneric 1884 1887 +0.2% 1.00x (?) PopFrontUnsafePointer 8674 9046 +4.3% 0.96x (?) PrefixAnyCollection 84 84 +0.0% 1.00x PrefixAnyCollectionLazy 60627 61648 +1.7% 0.98x (?) PrefixAnySeqCRangeIter 13267 13293 +0.2% 1.00x (?) PrefixAnySeqCRangeIterLazy 13307 13322 +0.1% 1.00x PrefixAnySeqCntRange 29 29 +0.0% 1.00x PrefixAnySeqCntRangeLazy 28 29 +3.6% 0.97x PrefixAnySequence 4214 4167 -1.1% 1.01x PrefixAnySequenceLazy 4224 4214 -0.2% 1.00x (?) PrefixArray 35 35 +0.0% 1.00x PrefixArrayLazy 35 35 +0.0% 1.00x PrefixCountableRange 35 35 +0.0% 1.00x PrefixCountableRangeLazy 35 35 +0.0% 1.00x PrefixSequence 1325 1328 +0.2% 1.00x PrefixSequenceLazy 1413 1413 +0.0% 1.00x PrefixWhileAnyCollection 155 155 +0.0% 1.00x PrefixWhileAnyCollectionLazy 90 90 +0.0% 1.00x PrefixWhileAnySeqCRangeIter 8822 8728 -1.1% 1.01x (?) PrefixWhileAnySeqCRangeIterLazy 24 24 +0.0% 1.00x PrefixWhileAnySeqCntRange 60 60 +0.0% 1.00x PrefixWhileAnySeqCntRangeLazy 24 24 +0.0% 1.00x PrefixWhileAnySequence 9936 9933 -0.0% 1.00x (?) PrefixWhileAnySequenceLazy 1394 1393 -0.1% 1.00x PrefixWhileArray 88 88 +0.0% 1.00x PrefixWhileArrayLazy 70 70 +0.0% 1.00x PrefixWhileCountableRange 35 35 +0.0% 1.00x PrefixWhileCountableRangeLazy 35 35 +0.0% 1.00x PrefixWhileSequence 364 362 -0.5% 1.01x PrefixWhileSequenceLazy 52 52 +0.0% 1.00x Prims 1129 1129 +0.0% 1.00x PrimsSplit 1141 1161 +1.8% 0.98x (?) QueueConcrete 1133 1128 -0.4% 1.00x (?) QueueGeneric 1136 1135 -0.1% 1.00x (?) RC4 164 162 -1.2% 1.01x RGBHistogram 3699 3678 -0.6% 1.01x (?) RGBHistogramOfObjects 25326 25180 -0.6% 1.01x (?) RangeAssignment 351 338 -3.7% 1.04x RangeIterationSigned64 200 200 +0.0% 1.00x RangeIterationUnsigned 200 200 +0.0% 1.00x RangeReplaceableCollectionPlusDefault 908 905 -0.3% 1.00x (?) RecursiveOwnedParameter 115 115 +0.0% 1.00x RemoveWhereFilterInts 45 43 -4.4% 1.05x RemoveWhereFilterString 375 375 +0.0% 1.00x RemoveWhereFilterStrings 432 432 +0.0% 1.00x RemoveWhereMoveStrings 701 703 +0.3% 1.00x RemoveWhereQuadraticInts 1288 1290 +0.2% 1.00x (?) RemoveWhereQuadraticString 492 489 -0.6% 1.01x (?) RemoveWhereQuadraticStrings 2752 2752 +0.0% 1.00x RemoveWhereSwapInts 19 19 +0.0% 1.00x RemoveWhereSwapStrings 850 848 -0.2% 1.00x ReversedArray 57 57 +0.0% 1.00x ReversedBidirectional 15869 16222 +2.2% 0.98x ReversedDictionary 301 293 -2.7% 1.03x (?) RomanNumbers 134373 134491 +0.1% 1.00x (?) SequenceAlgosAnySequence 9393 9439 +0.5% 1.00x SequenceAlgosArray 1573 1579 +0.4% 1.00x (?) SequenceAlgosContiguousArray 1579 1578 -0.1% 1.00x (?) SequenceAlgosList 1351 1356 +0.4% 1.00x SequenceAlgosRange 2578 2577 -0.0% 1.00x (?) SequenceAlgosUnfoldSequence 1080 1079 -0.1% 1.00x (?) SetExclusiveOr 6141 6137 -0.1% 1.00x (?) SetExclusiveOr_OfObjects 13161 12970 -1.5% 1.01x SetIntersect 1000 982 -1.8% 1.02x SetIntersect_OfObjects 2313 2331 +0.8% 0.99x (?) SetIsSubsetOf 382 382 +0.0% 1.00x SetIsSubsetOf_OfObjects 523 527 +0.8% 0.99x (?) SetUnion 5154 5115 -0.8% 1.01x (?) SetUnion_OfObjects 11602 11365 -2.0% 1.02x SevenBoom 1530 1531 +0.1% 1.00x (?) SortLargeExistentials 5430 5557 +2.3% 0.98x SortLettersInPlace 990 995 +0.5% 0.99x (?) SortSortedStrings 841 829 -1.4% 1.01x SortStrings 1657 1630 -1.6% 1.02x SortStringsUnicode 2337 2335 -0.1% 1.00x (?) StackPromo 21945 21938 -0.0% 1.00x (?) StaticArray 10 10 +0.0% 1.00x StrComplexWalk 1563 1560 -0.2% 1.00x StrToInt 1451 1456 +0.3% 1.00x StringAdder 4262 4254 -0.2% 1.00x (?) StringBuilder 1596 1614 +1.1% 0.99x (?) StringComparison_abnormal 789 802 +1.6% 0.98x (?) StringComparison_emoji 780 777 -0.4% 1.00x StringComparison_fastPrenormal 715 708 -1.0% 1.01x StringComparison_latin1 553 547 -1.1% 1.01x StringComparison_longSharedPrefix 898 894 -0.4% 1.00x StringComparison_nonBMPSlowestPrenormal 1536 1528 -0.5% 1.01x (?) StringComparison_slowerPrenormal 1629 1652 +1.4% 0.99x StringComparison_zalgo 122689 122890 +0.2% 1.00x (?) StringEdits 145215 144639 -0.4% 1.00x (?) StringEnumRawValueInitialization 1073 1065 -0.7% 1.01x (?) StringFromLongWholeSubstring 21 21 +0.0% 1.00x StringFromLongWholeSubstringGeneric 97 98 +1.0% 0.99x (?) StringHasPrefixAscii 1805 1775 -1.7% 1.02x StringHasPrefixUnicode 111806 110203 -1.4% 1.01x StringHasSuffixAscii 1947 1891 -2.9% 1.03x StringHasSuffixUnicode 112843 113585 +0.7% 0.99x (?) StringInterpolation 10139 10219 +0.8% 0.99x (?) StringInterpolationManySmallSegments 21197 21809 +2.9% 0.97x StringInterpolationSmall 8864 8803 -0.7% 1.01x StringMatch 9523 9530 +0.1% 1.00x (?) StringRemoveDupes 1355 1343 -0.9% 1.01x (?) StringUTF16Builder 2535 2528 -0.3% 1.00x (?) StringUTF16SubstringBuilder 5463 5466 +0.1% 1.00x (?) StringWalk 1417 1467 +3.5% 0.97x StringWithCString 43171 43184 +0.0% 1.00x (?) StringWordBuilder 2409 2522 +4.7% 0.96x StringWordBuilderReservingCapacity 2091 2166 +3.6% 0.97x SubstringComparable 49 49 +0.0% 1.00x SubstringEqualString 541 536 -0.9% 1.01x SubstringEquatable 1359 1358 -0.1% 1.00x (?) SubstringFromLongString 10 10 +0.0% 1.00x SubstringFromLongStringGeneric 74 74 +0.0% 1.00x SuffixAnyCollection 31 31 +0.0% 1.00x SuffixAnyCollectionLazy 20261 20259 -0.0% 1.00x (?) SuffixAnySeqCRangeIter 3769 3782 +0.3% 1.00x SuffixAnySeqCRangeIterLazy 3776 3773 -0.1% 1.00x (?) SuffixAnySeqCntRange 21 21 +0.0% 1.00x SuffixAnySeqCntRangeLazy 21 21 +0.0% 1.00x SuffixAnySequence 4873 4881 +0.2% 1.00x (?) SuffixAnySequenceLazy 5008 4968 -0.8% 1.01x (?) SuffixCountableRange 11 11 +0.0% 1.00x SuffixCountableRangeLazy 11 11 +0.0% 1.00x SuffixSequence 3644 3610 -0.9% 1.01x SuffixSequenceLazy 3632 3620 -0.3% 1.00x (?) SumUsingReduce 97 101 +4.1% 0.96x SuperChars 41258 40432 -2.0% 1.02x (?) TwoSum 1594 1614 +1.3% 0.99x TypeFlood 0 0 +0.0% 1.00x UTF8Decode 291 295 +1.4% 0.99x Walsh 405 400 -1.2% 1.01x WordCountHistogramASCII 7498 7297 -2.7% 1.03x WordCountHistogramUTF16 13123 12954 -1.3% 1.01x (?) WordCountUniqueASCII 2403 2401 -0.1% 1.00x (?) WordCountUniqueUTF16 6565 6647 +1.2% 0.99x (?) WordSplitASCII 19086 18805 -1.5% 1.01x (?) WordSplitUTF16 21287 20674 -2.9% 1.03x (?) XorLoop 402 396 -1.5% 1.02x Added (2) TEST MIN MAX MEAN MAX_RSS ChainedFilterMap 1248 1249 1249 8323072 FatCompactMap 195537 195668 195599 8437760 Unoptimized (Onone) Regression (10) TEST OLD NEW DELTA SPEEDUP StackPromo 95158 108646 +14.2% 0.88x ReversedDictionary 22358 24617 +10.1% 0.91x FilterEvenUsingReduceInto 1827 1993 +9.1% 0.92x ObjectiveCBridgeStubToNSDate 15312 16688 +9.0% 0.92x (?) StringInterpolationSmall 11994 13063 +8.9% 0.92x (?) FloatingPointPrinting_Float_interpolated 135830 147581 +8.7% 0.92x (?) DataAppendArray 5373 5734 +6.7% 0.94x NSDictionaryCastToSwift 6075 6482 +6.7% 0.94x (?) Array2D 633100 674530 +6.5% 0.94x ArrayAppendReserved 4061 4282 +5.4% 0.95x Improvement (9) TEST OLD NEW DELTA SPEEDUP CharacterPropertiesFetch 7408 5453 -26.4% 1.36x ObjectiveCBridgeFromNSSetAnyObjectForced 5988 5345 -10.7% 1.12x ObjectiveCBridgeStubFromArrayOfNSString 34853 31232 -10.4% 1.12x (?) Combos 2190 2020 -7.8% 1.08x DataSetCount 633 592 -6.5% 1.07x ErrorHandling 7234 6807 -5.9% 1.06x (?) StringBuilder 6687 6318 -5.5% 1.06x (?) RemoveWhereFilterInts 2024 1914 -5.4% 1.06x WordSplitUTF16 25214 24004 -4.8% 1.05x (?) No Changes (404) TEST OLD NEW DELTA SPEEDUP AngryPhonebook 5091 5167 +1.5% 0.99x AnyHashableWithAClass 98620 99296 +0.7% 0.99x (?) ArrayAppend 4444 4570 +2.8% 0.97x ArrayAppendArrayOfInt 859 862 +0.3% 1.00x ArrayAppendAscii 39165 39319 +0.4% 1.00x (?) ArrayAppendFromGeneric 865 866 +0.1% 1.00x (?) ArrayAppendGenericStructs 1503 1501 -0.1% 1.00x (?) ArrayAppendLatin1 63519 63444 -0.1% 1.00x ArrayAppendLazyMap 162368 163090 +0.4% 1.00x ArrayAppendOptionals 1496 1499 +0.2% 1.00x (?) ArrayAppendRepeatCol 154463 154497 +0.0% 1.00x (?) ArrayAppendSequence 99425 99428 +0.0% 1.00x (?) ArrayAppendStrings 15393 15403 +0.1% 1.00x (?) ArrayAppendToFromGeneric 865 861 -0.5% 1.00x (?) ArrayAppendToGeneric 866 871 +0.6% 0.99x (?) ArrayAppendUTF16 62351 62146 -0.3% 1.00x (?) ArrayInClass 6144 6254 +1.8% 0.98x ArrayLiteral 1799 1799 +0.0% 1.00x ArrayOfGenericPOD2 1073 1073 +0.0% 1.00x ArrayOfGenericRef 10811 10822 +0.1% 1.00x (?) ArrayOfPOD 760 762 +0.3% 1.00x ArrayOfRef 9924 9951 +0.3% 1.00x (?) ArrayPlusEqualArrayOfInt 864 862 -0.2% 1.00x ArrayPlusEqualFiveElementCollection 232069 233477 +0.6% 0.99x (?) ArrayPlusEqualSingleElementCollection 230289 228396 -0.8% 1.01x (?) ArrayPlusEqualThreeElements 9178 9223 +0.5% 1.00x ArraySubscript 106664 107752 +1.0% 0.99x ArrayValueProp 3690 3686 -0.1% 1.00x (?) ArrayValueProp2 14802 14860 +0.4% 1.00x (?) ArrayValueProp3 4175 4127 -1.1% 1.01x (?) ArrayValueProp4 4123 4115 -0.2% 1.00x (?) BinaryFloatingPointConversionFromBinaryInteger 5312 5293 -0.4% 1.00x (?) BinaryFloatingPointPropertiesBinade 85 83 -2.4% 1.02x BinaryFloatingPointPropertiesNextUp 131 125 -4.6% 1.05x BinaryFloatingPointPropertiesUlp 129 125 -3.1% 1.03x BitCount 8999 8615 -4.3% 1.04x ByteSwap 9247 9465 +2.4% 0.98x COWTree 11708 11514 -1.7% 1.02x (?) CSVParsing 2415550 2390892 -1.0% 1.01x (?) CSVParsingAlt 1370485 1350302 -1.5% 1.01x (?) CSVParsingAltIndices 2235437 2139120 -4.3% 1.05x CStringLongAscii 4447 4445 -0.0% 1.00x (?) CStringLongNonAscii 2259 2273 +0.6% 0.99x (?) CStringShortAscii 7680 7673 -0.1% 1.00x (?) Calculator 1269 1275 +0.5% 1.00x CaptureProp 235006 235368 +0.2% 1.00x CharIndexing_ascii_unicodeScalars 299867 302100 +0.7% 0.99x (?) CharIndexing_ascii_unicodeScalars_Backwards 339038 338131 -0.3% 1.00x (?) CharIndexing_chinese_unicodeScalars 226342 228379 +0.9% 0.99x (?) CharIndexing_chinese_unicodeScalars_Backwards 257882 256235 -0.6% 1.01x (?) CharIndexing_japanese_unicodeScalars 362695 361798 -0.2% 1.00x (?) CharIndexing_japanese_unicodeScalars_Backwards 407105 404974 -0.5% 1.01x (?) CharIndexing_korean_unicodeScalars 290658 295485 +1.7% 0.98x (?) CharIndexing_korean_unicodeScalars_Backwards 332845 330171 -0.8% 1.01x (?) CharIndexing_punctuatedJapanese_unicodeScalars 53636 53461 -0.3% 1.00x (?) CharIndexing_punctuatedJapanese_unicodeScalars_Backwards 58576 59832 +2.1% 0.98x (?) CharIndexing_punctuated_unicodeScalars 66528 67112 +0.9% 0.99x CharIndexing_punctuated_unicodeScalars_Backwards 74277 74054 -0.3% 1.00x (?) CharIndexing_russian_unicodeScalars 248690 252445 +1.5% 0.99x CharIndexing_russian_unicodeScalars_Backwards 285172 283483 -0.6% 1.01x (?) CharIndexing_tweet_unicodeScalars 597571 599734 +0.4% 1.00x CharIndexing_tweet_unicodeScalars_Backwards 670848 675748 +0.7% 0.99x (?) CharIndexing_utf16_unicodeScalars 265853 263382 -0.9% 1.01x (?) CharIndexing_utf16_unicodeScalars_Backwards 290887 289293 -0.5% 1.01x (?) CharIteration_ascii_unicodeScalars 144830 145341 +0.4% 1.00x (?) CharIteration_ascii_unicodeScalars_Backwards 249431 246595 -1.1% 1.01x (?) CharIteration_chinese_unicodeScalars 109708 110011 +0.3% 1.00x (?) CharIteration_chinese_unicodeScalars_Backwards 188092 187521 -0.3% 1.00x (?) CharIteration_japanese_unicodeScalars 174308 174365 +0.0% 1.00x (?) CharIteration_japanese_unicodeScalars_Backwards 300679 296543 -1.4% 1.01x CharIteration_korean_unicodeScalars 140987 141092 +0.1% 1.00x (?) CharIteration_korean_unicodeScalars_Backwards 241259 240327 -0.4% 1.00x CharIteration_punctuatedJapanese_unicodeScalars 25783 25914 +0.5% 0.99x (?) CharIteration_punctuatedJapanese_unicodeScalars_Backwards 43217 42916 -0.7% 1.01x (?) CharIteration_punctuated_unicodeScalars 32426 32601 +0.5% 0.99x (?) CharIteration_punctuated_unicodeScalars_Backwards 54950 54527 -0.8% 1.01x (?) CharIteration_russian_unicodeScalars 120795 121264 +0.4% 1.00x (?) CharIteration_russian_unicodeScalars_Backwards 207276 205268 -1.0% 1.01x (?) CharIteration_tweet_unicodeScalars 286514 287674 +0.4% 1.00x (?) CharIteration_tweet_unicodeScalars_Backwards 501309 491156 -2.0% 1.02x (?) CharIteration_utf16_unicodeScalars 124410 125329 +0.7% 0.99x (?) CharIteration_utf16_unicodeScalars_Backwards 215567 215160 -0.2% 1.00x (?) CharacterLiteralsLarge 5782 5802 +0.3% 1.00x (?) CharacterLiteralsSmall 658 659 +0.2% 1.00x (?) CharacterPropertiesPrecomputed 3829 3794 -0.9% 1.01x (?) CharacterPropertiesStashed 2335 2319 -0.7% 1.01x (?) CharacterPropertiesStashedMemo 4626 4586 -0.9% 1.01x (?) Chars 34757 34730 -0.1% 1.00x (?) ClassArrayGetter 985 986 +0.1% 1.00x DataAccessBytes 2343 2336 -0.3% 1.00x (?) DataAppendBytes 5160 5098 -1.2% 1.01x (?) DataAppendDataLargeToLarge 67998 67961 -0.1% 1.00x (?) DataAppendDataLargeToMedium 35889 35671 -0.6% 1.01x (?) DataAppendDataLargeToSmall 34639 34783 +0.4% 1.00x (?) DataAppendDataMediumToLarge 38226 37703 -1.4% 1.01x DataAppendDataMediumToMedium 6980 6844 -1.9% 1.02x (?) DataAppendDataMediumToSmall 5977 6071 +1.6% 0.98x (?) DataAppendDataSmallToLarge 36563 36843 +0.8% 0.99x DataAppendDataSmallToMedium 6190 6315 +2.0% 0.98x (?) DataAppendDataSmallToSmall 6099 5894 -3.4% 1.03x (?) DataAppendSequence 1593337 1588844 -0.3% 1.00x (?) DataCopyBytes 2463 2477 +0.6% 0.99x (?) DataCount 223 223 +0.0% 1.00x DataMutateBytes 5425 5435 +0.2% 1.00x (?) DataReplaceLarge 40434 40017 -1.0% 1.01x (?) DataReplaceLargeBuffer 57801 57675 -0.2% 1.00x (?) DataReplaceMedium 9172 9246 +0.8% 0.99x (?) DataReplaceMediumBuffer 11857 12082 +1.9% 0.98x (?) DataReplaceSmall 6857 6862 +0.1% 1.00x (?) DataReplaceSmallBuffer 8479 8542 +0.7% 0.99x (?) DataReset 2854 2825 -1.0% 1.01x (?) DataSubscript 441 441 +0.0% 1.00x DictOfArraysToArrayOfDicts 3394 3377 -0.5% 1.01x (?) Dictionary 2279 2289 +0.4% 1.00x (?) Dictionary2 2561 2561 +0.0% 1.00x Dictionary2OfObjects 5816 5871 +0.9% 0.99x (?) Dictionary3 1159 1155 -0.3% 1.00x (?) Dictionary3OfObjects 2379 2329 -2.1% 1.02x (?) Dictionary4 1526 1539 +0.9% 0.99x (?) Dictionary4OfObjects 2219 2142 -3.5% 1.04x DictionaryBridge 2120 2113 -0.3% 1.00x (?) DictionaryCopy 307589 311704 +1.3% 0.99x (?) DictionaryFilter 304174 312245 +2.7% 0.97x DictionaryGroup 4008 4005 -0.1% 1.00x (?) DictionaryGroupOfObjects 7328 7323 -0.1% 1.00x (?) DictionaryLiteral 8751 8688 -0.7% 1.01x (?) DictionaryOfObjects 6113 6044 -1.1% 1.01x (?) DictionaryRemove 17873 17542 -1.9% 1.02x DictionaryRemoveOfObjects 53953 54583 +1.2% 0.99x (?) DictionarySubscriptDefaultMutation 2008 2019 +0.5% 0.99x (?) DictionarySubscriptDefaultMutationArray 2278 2273 -0.2% 1.00x (?) DictionarySubscriptDefaultMutationArrayOfObjects 9359 9403 +0.5% 1.00x (?) DictionarySubscriptDefaultMutationOfObjects 5406 5363 -0.8% 1.01x (?) DictionarySwap 5252 5261 +0.2% 1.00x DictionarySwapAt 35422 35352 -0.2% 1.00x (?) DictionarySwapAtOfObjects 114386 113496 -0.8% 1.01x (?) DictionarySwapOfObjects 19801 19967 +0.8% 0.99x (?) DoubleWidthDivision 0 0 +0.0% 1.00x DropFirstAnyCollection 12549 12440 -0.9% 1.01x DropFirstAnyCollectionLazy 90028 89536 -0.5% 1.01x (?) DropFirstAnySeqCRangeIter 19688 19764 +0.4% 1.00x DropFirstAnySeqCRangeIterLazy 19618 19663 +0.2% 1.00x DropFirstAnySeqCntRange 12557 12456 -0.8% 1.01x DropFirstAnySeqCntRangeLazy 12520 12441 -0.6% 1.01x DropFirstAnySequence 11331 11319 -0.1% 1.00x (?) DropFirstAnySequenceLazy 11349 11276 -0.6% 1.01x (?) DropFirstArray 3129 3104 -0.8% 1.01x DropFirstArrayLazy 22080 22218 +0.6% 0.99x (?) DropFirstCountableRange 309 309 +0.0% 1.00x DropFirstCountableRangeLazy 23585 23409 -0.7% 1.01x DropFirstSequence 10757 10813 +0.5% 0.99x (?) DropFirstSequenceLazy 10822 10800 -0.2% 1.00x (?) DropLastAnyCollection 4199 4160 -0.9% 1.01x DropLastAnyCollectionLazy 29018 28784 -0.8% 1.01x (?) DropLastAnySeqCRangeIter 37083 37023 -0.2% 1.00x (?) DropLastAnySeqCRangeIterLazy 37358 37059 -0.8% 1.01x (?) DropLastAnySeqCntRange 4185 4149 -0.9% 1.01x DropLastAnySeqCntRangeLazy 4188 4169 -0.5% 1.00x DropLastAnySequence 29384 29417 +0.1% 1.00x (?) DropLastAnySequenceLazy 29826 29779 -0.2% 1.00x DropLastCountableRange 110 107 -2.7% 1.03x DropLastCountableRangeLazy 7859 7854 -0.1% 1.00x (?) DropLastSequence 29384 29258 -0.4% 1.00x DropLastSequenceLazy 29638 29446 -0.6% 1.01x DropWhileAnyCollection 16004 15926 -0.5% 1.00x DropWhileAnyCollectionLazy 18068 18100 +0.2% 1.00x DropWhileAnySeqCRangeIter 21200 21295 +0.4% 1.00x DropWhileAnySeqCRangeIterLazy 17980 18044 +0.4% 1.00x (?) DropWhileAnySeqCntRange 16045 16009 -0.2% 1.00x DropWhileAnySeqCntRangeLazy 18148 18199 +0.3% 1.00x DropWhileAnySequence 12950 12841 -0.8% 1.01x (?) DropWhileAnySequenceLazy 10831 10863 +0.3% 1.00x (?) DropWhileArrayLazy 13570 13722 +1.1% 0.99x DropWhileCountableRange 3929 3930 +0.0% 1.00x (?) DropWhileCountableRangeLazy 17602 17625 +0.1% 1.00x (?) DropWhileSequence 12386 12474 +0.7% 0.99x DropWhileSequenceLazy 10433 10349 -0.8% 1.01x (?) EqualStringSubstring 73 73 +0.0% 1.00x EqualSubstringString 73 73 +0.0% 1.00x EqualSubstringSubstring 74 74 +0.0% 1.00x EqualSubstringSubstringGenericEquatable 56 56 +0.0% 1.00x ExclusivityGlobal 183 183 +0.0% 1.00x ExclusivityIndependent 71 72 +1.4% 0.99x (?) FilterEvenUsingReduce 3553 3691 +3.9% 0.96x FloatingPointPrinting_Double_description_small 81088 81010 -0.1% 1.00x (?) FloatingPointPrinting_Double_description_uniform 154426 148719 -3.7% 1.04x (?) FloatingPointPrinting_Double_interpolated 252590 251935 -0.3% 1.00x (?) FloatingPointPrinting_Float80_description_small 94282 93067 -1.3% 1.01x FloatingPointPrinting_Float80_description_uniform 1736455 1747659 +0.6% 0.99x FloatingPointPrinting_Float80_interpolated 1838540 1838723 +0.0% 1.00x (?) FloatingPointPrinting_Float_description_small 44464 43591 -2.0% 1.02x (?) FloatingPointPrinting_Float_description_uniform 59783 57970 -3.0% 1.03x (?) FrequenciesUsingReduce 13596 13498 -0.7% 1.01x (?) FrequenciesUsingReduceInto 5593 5674 +1.4% 0.99x (?) Hanoi 20185 20631 +2.2% 0.98x HashTest 22111 21622 -2.2% 1.02x Histogram 6382 6391 +0.1% 1.00x Integrate 458 460 +0.4% 1.00x IterateData 5682 5778 +1.7% 0.98x Join 1118 1076 -3.8% 1.04x LazilyFilteredArrayContains 730238 733046 +0.4% 1.00x LazilyFilteredArrays 1376021 1375433 -0.0% 1.00x (?) LazilyFilteredRange 438384 438011 -0.1% 1.00x LessSubstringSubstring 74 73 -1.4% 1.01x LessSubstringSubstringGenericComparable 55 55 +0.0% 1.00x LinkedList 32077 32294 +0.7% 0.99x LuhnAlgoEager 5739 5722 -0.3% 1.00x (?) LuhnAlgoLazy 5865 5860 -0.1% 1.00x (?) MapReduce 25425 25430 +0.0% 1.00x (?) MapReduceAnyCollection 25540 25459 -0.3% 1.00x (?) MapReduceAnyCollectionShort 35585 36104 +1.5% 0.99x (?) MapReduceClass 29670 29689 +0.1% 1.00x (?) MapReduceClassShort 39296 40068 +2.0% 0.98x (?) MapReduceLazyCollection 21784 21867 +0.4% 1.00x (?) MapReduceLazyCollectionShort 33721 32887 -2.5% 1.03x (?) MapReduceLazySequence 18553 18545 -0.0% 1.00x (?) MapReduceSequence 29100 29162 +0.2% 1.00x (?) MapReduceShort 37268 36373 -2.4% 1.02x (?) MapReduceShortString 221 228 +3.2% 0.97x MapReduceString 1758 1762 +0.2% 1.00x (?) Memset 44131 44126 -0.0% 1.00x (?) MonteCarloE 1129454 1148399 +1.7% 0.98x MonteCarloPi 5136177 5208437 +1.4% 0.99x NSError 723 729 +0.8% 0.99x (?) NSStringConversion 444 443 -0.2% 1.00x NibbleSort 427361 430132 +0.6% 0.99x NopDeinit 188551 190924 +1.3% 0.99x (?) ObjectAllocation 1253 1254 +0.1% 1.00x (?) ObjectiveCBridgeFromNSArrayAnyObject 27542 27169 -1.4% 1.01x (?) ObjectiveCBridgeFromNSArrayAnyObjectForced 9664 9648 -0.2% 1.00x (?) ObjectiveCBridgeFromNSArrayAnyObjectToString 45131 46984 +4.1% 0.96x (?) ObjectiveCBridgeFromNSArrayAnyObjectToStringForced 44179 43924 -0.6% 1.01x (?) ObjectiveCBridgeFromNSDictionaryAnyObject 105375 106913 +1.5% 0.99x (?) ObjectiveCBridgeFromNSSetAnyObject 52411 51272 -2.2% 1.02x (?) ObjectiveCBridgeFromNSSetAnyObjectToString 79881 82141 +2.8% 0.97x (?) ObjectiveCBridgeFromNSString 2669 2658 -0.4% 1.00x (?) ObjectiveCBridgeFromNSStringForced 2640 2584 -2.1% 1.02x (?) ObjectiveCBridgeStubDataAppend 5564 5522 -0.8% 1.01x (?) ObjectiveCBridgeStubDateMutation 716 716 +0.0% 1.00x ObjectiveCBridgeStubFromNSDate 6684 6595 -1.3% 1.01x (?) ObjectiveCBridgeStubFromNSString 973 980 +0.7% 0.99x (?) ObjectiveCBridgeStubFromNSStringRef 193 196 +1.6% 0.98x ObjectiveCBridgeStubNSDataAppend 2911 2959 +1.6% 0.98x ObjectiveCBridgeStubNSDateMutationRef 16139 16264 +0.8% 0.99x (?) ObjectiveCBridgeStubToArrayOfNSString 27851 27966 +0.4% 1.00x (?) ObjectiveCBridgeStubToNSDateRef 3457 3482 +0.7% 0.99x (?) ObjectiveCBridgeStubToNSString 1614 1584 -1.9% 1.02x ObjectiveCBridgeStubToNSStringRef 153 153 +0.0% 1.00x ObjectiveCBridgeStubURLAppendPath 373502 363286 -2.7% 1.03x (?) ObjectiveCBridgeStubURLAppendPathRef 368089 374099 +1.6% 0.98x (?) ObjectiveCBridgeToNSArray 28806 28375 -1.5% 1.02x (?) ObjectiveCBridgeToNSDictionary 43170 44958 +4.1% 0.96x (?) ObjectiveCBridgeToNSSet 34370 34535 +0.5% 1.00x (?) ObjectiveCBridgeToNSString 1353 1348 -0.4% 1.00x (?) ObserverClosure 6411 6408 -0.0% 1.00x (?) ObserverForwarderStruct 4250 4256 +0.1% 1.00x (?) ObserverPartiallyAppliedMethod 8000 7931 -0.9% 1.01x ObserverUnappliedMethod 8177 8053 -1.5% 1.02x OpenClose 530 512 -3.4% 1.04x PartialApplyDynamicType 38951 39374 +1.1% 0.99x (?) Phonebook 17655 17798 +0.8% 0.99x PointerArithmetics 126076 125999 -0.1% 1.00x (?) PolymorphicCalls 2439 2450 +0.5% 1.00x PopFrontArray 4641 4654 +0.3% 1.00x (?) PopFrontArrayGeneric 5494 5469 -0.5% 1.00x (?) PopFrontUnsafePointer 10649 10712 +0.6% 0.99x (?) PrefixAnyCollection 12541 12416 -1.0% 1.01x PrefixAnyCollectionLazy 89301 89184 -0.1% 1.00x (?) PrefixAnySeqCRangeIter 15985 15945 -0.3% 1.00x (?) PrefixAnySeqCRangeIterLazy 15928 15934 +0.0% 1.00x (?) PrefixAnySeqCntRange 12542 12415 -1.0% 1.01x PrefixAnySeqCntRangeLazy 12561 12458 -0.8% 1.01x PrefixAnySequence 9518 9399 -1.3% 1.01x PrefixAnySequenceLazy 9393 9418 +0.3% 1.00x (?) PrefixArray 3143 3099 -1.4% 1.01x PrefixArrayLazy 22029 22233 +0.9% 0.99x PrefixCountableRange 317 309 -2.5% 1.03x PrefixCountableRangeLazy 23530 23381 -0.6% 1.01x PrefixSequence 8940 8792 -1.7% 1.02x PrefixSequenceLazy 8749 8718 -0.4% 1.00x (?) PrefixWhileAnyCollection 23139 23092 -0.2% 1.00x PrefixWhileAnyCollectionLazy 15170 15009 -1.1% 1.01x PrefixWhileAnySeqCRangeIter 31600 32133 +1.7% 0.98x PrefixWhileAnySeqCRangeIterLazy 15006 15169 +1.1% 0.99x PrefixWhileAnySeqCntRange 23231 23197 -0.1% 1.00x (?) PrefixWhileAnySeqCntRangeLazy 15109 15106 -0.0% 1.00x (?) PrefixWhileAnySequence 25561 25734 +0.7% 0.99x PrefixWhileAnySequenceLazy 9684 9766 +0.8% 0.99x (?) PrefixWhileArray 10227 10290 +0.6% 0.99x (?) PrefixWhileArrayLazy 11779 11748 -0.3% 1.00x (?) PrefixWhileCountableRange 11062 11039 -0.2% 1.00x (?) PrefixWhileCountableRangeLazy 14679 14684 +0.0% 1.00x (?) PrefixWhileSequence 25285 25280 -0.0% 1.00x (?) PrefixWhileSequenceLazy 9476 9433 -0.5% 1.00x (?) Prims 9576 9623 +0.5% 1.00x (?) PrimsSplit 9592 9597 +0.1% 1.00x (?) QueueConcrete 14035 14322 +2.0% 0.98x QueueGeneric 18101 18135 +0.2% 1.00x RC4 16406 16598 +1.2% 0.99x RGBHistogram 24526 24559 +0.1% 1.00x RGBHistogramOfObjects 77817 77960 +0.2% 1.00x (?) RangeAssignment 2645 2659 +0.5% 0.99x RangeIterationSigned 14387 14589 +1.4% 0.99x (?) RangeIterationSigned64 35170 35053 -0.3% 1.00x (?) RangeIterationUnsigned 31810 31772 -0.1% 1.00x (?) RangeReplaceableCollectionPlusDefault 10006 9976 -0.3% 1.00x (?) RecursiveOwnedParameter 5771 5773 +0.0% 1.00x (?) RemoveWhereFilterString 1281 1279 -0.2% 1.00x (?) RemoveWhereFilterStrings 2432 2475 +1.8% 0.98x RemoveWhereMoveInts 3362 3357 -0.1% 1.00x (?) RemoveWhereMoveStrings 3884 3887 +0.1% 1.00x RemoveWhereQuadraticInts 7595 7571 -0.3% 1.00x RemoveWhereQuadraticString 2243 2234 -0.4% 1.00x (?) RemoveWhereQuadraticStrings 9289 9230 -0.6% 1.01x (?) RemoveWhereSwapInts 6026 6042 +0.3% 1.00x RemoveWhereSwapStrings 6731 6753 +0.3% 1.00x ReversedArray 12422 12445 +0.2% 1.00x (?) ReversedBidirectional 42502 42323 -0.4% 1.00x (?) RomanNumbers 1111536 1116798 +0.5% 1.00x (?) SequenceAlgosAnySequence 10177 10199 +0.2% 1.00x SequenceAlgosArray 760575 734662 -3.4% 1.04x SequenceAlgosContiguousArray 282808 274793 -2.8% 1.03x SequenceAlgosList 8315 8243 -0.9% 1.01x SequenceAlgosRange 999990 1000278 +0.0% 1.00x (?) SequenceAlgosUnfoldSequence 5747 5738 -0.2% 1.00x (?) SetExclusiveOr 17413 17426 +0.1% 1.00x (?) SetExclusiveOr_OfObjects 45768 45460 -0.7% 1.01x (?) SetIntersect 7169 7165 -0.1% 1.00x (?) SetIntersect_OfObjects 11123 11322 +1.8% 0.98x (?) SetIsSubsetOf 1263 1272 +0.7% 0.99x SetIsSubsetOf_OfObjects 1793 1818 +1.4% 0.99x SetUnion 12379 12322 -0.5% 1.00x (?) SetUnion_OfObjects 32210 32035 -0.5% 1.01x (?) SevenBoom 1647 1663 +1.0% 0.99x Sim2DArray 43441 43483 +0.1% 1.00x (?) SortLargeExistentials 10958 10976 +0.2% 1.00x (?) SortLettersInPlace 1805 1806 +0.1% 1.00x (?) SortSortedStrings 935 916 -2.0% 1.02x SortStrings 1832 1856 +1.3% 0.99x SortStringsUnicode 2622 2600 -0.8% 1.01x StaticArray 2604 2611 +0.3% 1.00x (?) StrComplexWalk 6204 6199 -0.1% 1.00x StrToInt 72696 72867 +0.2% 1.00x (?) StringAdder 4668 4679 +0.2% 1.00x (?) StringBuilderLong 2754 2748 -0.2% 1.00x (?) StringBuilderWithLongSubstring 4488 4470 -0.4% 1.00x (?) StringComparison_abnormal 1352 1317 -2.6% 1.03x StringComparison_ascii 8598 8583 -0.2% 1.00x StringComparison_emoji 1927 1924 -0.2% 1.00x (?) StringComparison_fastPrenormal 4678 4705 +0.6% 0.99x StringComparison_latin1 3648 3633 -0.4% 1.00x (?) StringComparison_longSharedPrefix 2297 2289 -0.3% 1.00x StringComparison_nonBMPSlowestPrenormal 3559 3552 -0.2% 1.00x (?) StringComparison_slowerPrenormal 4022 4012 -0.2% 1.00x (?) StringComparison_zalgo 125468 125447 -0.0% 1.00x (?) StringEdits 338816 343958 +1.5% 0.99x (?) StringEnumRawValueInitialization 17082 17019 -0.4% 1.00x (?) StringEqualPointerComparison 1547 1517 -1.9% 1.02x StringFromLongWholeSubstring 22 22 +0.0% 1.00x StringFromLongWholeSubstringGeneric 314 318 +1.3% 0.99x (?) StringHasPrefixAscii 3199 3162 -1.2% 1.01x StringHasPrefixUnicode 114193 113165 -0.9% 1.01x (?) StringHasSuffixAscii 3318 3245 -2.2% 1.02x StringHasSuffixUnicode 106116 107408 +1.2% 0.99x (?) StringInterpolation 14264 14509 +1.7% 0.98x (?) StringInterpolationManySmallSegments 22752 22731 -0.1% 1.00x (?) StringMatch 33021 32334 -2.1% 1.02x StringRemoveDupes 1387 1418 +2.2% 0.98x (?) StringUTF16Builder 7214 7311 +1.3% 0.99x (?) StringUTF16SubstringBuilder 21255 21121 -0.6% 1.01x (?) StringWalk 12594 12742 +1.2% 0.99x (?) StringWithCString 37562 37562 +0.0% 1.00x StringWordBuilder 2446 2439 -0.3% 1.00x StringWordBuilderReservingCapacity 2121 2118 -0.1% 1.00x (?) SubstringComparable 1589 1572 -1.1% 1.01x SubstringEqualString 1675 1676 +0.1% 1.00x (?) SubstringEquatable 5274 5379 +2.0% 0.98x SubstringFromLongString 24 24 +0.0% 1.00x SubstringFromLongStringGeneric 101 101 +0.0% 1.00x SuffixAnyCollection 4217 4166 -1.2% 1.01x SuffixAnyCollectionLazy 29970 29059 -3.0% 1.03x (?) SuffixAnySeqCRangeIter 32616 32536 -0.2% 1.00x (?) SuffixAnySeqCRangeIterLazy 32663 32637 -0.1% 1.00x (?) SuffixAnySeqCntRange 4183 4164 -0.5% 1.00x SuffixAnySeqCntRangeLazy 4199 4163 -0.9% 1.01x SuffixAnySequence 25050 25079 +0.1% 1.00x (?) SuffixAnySequenceLazy 25242 25279 +0.1% 1.00x (?) SuffixCountableRange 105 105 +0.0% 1.00x SuffixCountableRangeLazy 7851 7854 +0.0% 1.00x (?) SuffixSequence 24922 25030 +0.4% 1.00x SuffixSequenceLazy 24937 25033 +0.4% 1.00x (?) SumUsingReduce 159243 158381 -0.5% 1.01x (?) SumUsingReduceInto 149800 151236 +1.0% 0.99x SuperChars 123894 122950 -0.8% 1.01x (?) TwoSum 3643 3666 +0.6% 0.99x (?) TypeFlood 191 194 +1.6% 0.98x (?) UTF8Decode 29181 29330 +0.5% 0.99x Walsh 12097 12022 -0.6% 1.01x (?) WordCountHistogramASCII 34825 34809 -0.0% 1.00x (?) WordCountHistogramUTF16 43334 43534 +0.5% 1.00x (?) WordCountUniqueASCII 7411 7411 +0.0% 1.00x WordCountUniqueUTF16 11813 11835 +0.2% 1.00x (?) WordSplitASCII 21896 22552 +3.0% 0.97x (?) XorLoop 23227 23228 +0.0% 1.00x (?) Added (2) TEST MIN MAX MEAN MAX_RSS ChainedFilterMap 224542 240224 229985 8867840 FatCompactMap 294530 298007 295731 8863744 Hardware Overview Model Name: Mac Pro Model Identifier: MacPro6,1 Processor Name: 12-Core Intel Xeon E5 Processor Speed: 2.7 GHz Number of Processors: 1 Total Number of Cores: 12 L2 Cache (per Core): 256 KB L3 Cache: 30 MB Memory: 64 GB
gharchive/pull-request
2018-03-20T19:01:44
2025-04-01T04:56:02.166051
{ "authors": [ "moiseev", "swift-ci" ], "repo": "apple/swift", "url": "https://github.com/apple/swift/pull/15376", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
343768469
Don’t dive into null parts of AST. When compiling code with certain errors, it is possible to get null pointers in the AST. The compiler indexes even when there are errors, and the indexing code then tries to walk into the null pointers, causing a crash. As a result, the frontend crashes, and the developer does not get to see the errors. This PR adds a null check for this specific case. It does not solve the general problem. Resolves rdar://42314665 @swift-ci please smoke test @akyrtzi Did you want to look at this? Thanks! Generally it would be ideal if you could add a regression test but I can understand if it is too difficult to reduce the original failing test case. @swift-ci please smoke test I agree a regression test would be great. If I can come up with one, I'll do another PR.
gharchive/pull-request
2018-07-23T19:41:21
2025-04-01T04:56:02.175237
{ "authors": [ "akyrtzi", "davidungar" ], "repo": "apple/swift", "url": "https://github.com/apple/swift/pull/18153", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }