id
stringlengths 4
10
| text
stringlengths 4
2.14M
| source
stringclasses 2
values | created
timestamp[s]date 2001-05-16 21:05:09
2025-01-01 03:38:30
| added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| metadata
dict |
---|---|---|---|---|---|
1624493576
|
HDDS-8069. [Snapshot] Compaction DAG reconstruction logic potentially crashing OM on startup
What changes were proposed in this pull request?
Problem: As part of jira-HDDS-7873, we added an optimization to early prune out SST files that won't be needed for compaction DAG base snap diff. More details in PR-4235.
Files are removed from back dir but compaction logs don't get updated because those are needed for DAG traversal based diffing. On OM restart, we count number of keys in the file present in compaction log. It is possible that file has been deleted due to above optimization which is causing OM crash.
This change is to log the exception instead throwing exception on OM restart. It will not affect Compaction DAG based snap diff because If file is not present during diff, it will throw the FileNotFoundException.
What is the link to the Apache JIRA
https://issues.apache.org/jira/browse/HDDS-8069
How was this patch tested?
Unit tests.
Thanks @hemantk-12 for the patch.
|
gharchive/pull-request
| 2023-03-15T00:13:07 |
2025-04-01T04:55:59.633994
|
{
"authors": [
"hemantk-12",
"smengcl"
],
"repo": "apache/ozone",
"url": "https://github.com/apache/ozone/pull/4401",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2225632311
|
HDDS-5569. recipe/BotoClient.md translation
What changes were proposed in this pull request?
HDDS-5569. recipe/BotoClient.md translation to Chinese
What is the link to the Apache JIRA
https://issues.apache.org/jira/browse/HDDS-5569
How was this patch tested?
hugo serve
thx for review @smengcl
Thanks @will-sh for the PR.
I think there is a Github glitch that prevents it from marking this PR as Merged. It is already merged in master:
https://github.com/apache/ozone/commit/5d999d2068618b9e1e8ed66188b3ea22dc225d1d
I will leave this PR here for a few hours to see if it fixes itself. If not I will close this manually.
Thanks @will-sh for the translation, @smengcl for the review.
|
gharchive/pull-request
| 2024-04-04T14:12:41 |
2025-04-01T04:55:59.637221
|
{
"authors": [
"adoroszlai",
"smengcl",
"will-sh"
],
"repo": "apache/ozone",
"url": "https://github.com/apache/ozone/pull/6480",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2231264966
|
[compaction] support multi table AppendOnlyCompactionTask
Purpose
related https://github.com/apache/paimon/issues/2670
Tests
API and Format
Documentation
@leaves12138 PTAL
finished at #2858
|
gharchive/pull-request
| 2024-04-08T13:42:42 |
2025-04-01T04:55:59.639295
|
{
"authors": [
"wg1026688210"
],
"repo": "apache/paimon",
"url": "https://github.com/apache/paimon/pull/3174",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2368056669
|
build failed on AArch64, Fedora 33
Apache Thrift 0.12.0 is required. Building it reports unsupported .NET, etc.
Installing 0.13.0 using yum results in an error on mvn package.
Reporter: Lutz Weischer
Note: This issue was originally created as PARQUET-2000. Please see the migration documentation for further details.
Lutz Weischer:
I restarted from scratch and get this:
LC_ALL=C mvn install
...
Testing PARQUET_1_0/FIXED_LEN_BYTE_ARRAY/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384
Testing PARQUET_2_0/FIXED_LEN_BYTE_ARRAY/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384
Tests run: 16, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1,100.501 sec
Results :
Tests in error:
testInteropReadEncryptedParquetFiles(org.apache.parquet.hadoop.ITTestEncryptionOptions): github.com: Name or service not known
Tests run: 17, Failures: 0, Errors: 1, Skipped: 0
[INFO]
[INFO] — maven-failsafe-plugin:2.10:verify (default) @ parquet-hadoop —
[INFO] Failsafe report directory: /data/jw/parquet/parquet-hadoop/target/failsafe-reports
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary for Apache Parquet MR 1.12.0-SNAPSHOT:
[INFO]
[INFO] Apache Parquet MR .................................. SUCCESS [ 8.688 s]
[INFO] Apache Parquet Format Structures ................... SUCCESS [ 8.891 s]
[INFO] Apache Parquet Generator ........................... SUCCESS [ 3.289 s]
[INFO] Apache Parquet Common .............................. SUCCESS [ 10.041 s]
[INFO] Apache Parquet Encodings ........................... SUCCESS [ 33.835 s]
[INFO] Apache Parquet Column .............................. SUCCESS [01:32 min]
[INFO] Apache Parquet Arrow ............................... SUCCESS [ 3.288 s]
[INFO] Apache Parquet Jackson ............................. SUCCESS [ 4.596 s]
[INFO] Apache Parquet Hadoop .............................. FAILURE [22:59 min]
[INFO] Apache Parquet Avro ................................ SKIPPED
[INFO] Apache Parquet Benchmarks .......................... SKIPPED
[INFO] Apache Parquet Pig ................................. SKIPPED
[INFO] Apache Parquet Thrift .............................. SKIPPED
[INFO] Apache Parquet Cascading [Deprecated, will be removed after 1.12] SKIPPED
[INFO] Apache Parquet Cascading (for Cascading 3.0 onwards) [Deprecated, will be removed after 1.12] SKIPPED
[INFO] Apache Parquet Command-line ........................ SKIPPED
[INFO] Apache Parquet Pig Bundle .......................... SKIPPED
[INFO] Apache Parquet Protobuf ............................ SKIPPED
[INFO] Apache Parquet Scala ............................... SKIPPED
[INFO] Apache Parquet Scrooge [Deprecated, will be removed after 1.12] SKIPPED
[INFO] Apache Parquet Hadoop Bundle ....................... SKIPPED
[INFO] Apache Parquet Tools; [Deprecated, will be removed after 1.12; Use parquet-cli instead] SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 25:45 min
[INFO] Finished at: 2021-03-24T15:06:55+01:00
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-failsafe-plugin:2.10:verify (default) on project parquet-hadoop: There are test failures.
[ERROR]
[ERROR] Please refer to /data/jw/parquet/parquet-hadoop/target/failsafe-reports for the individual test results.
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn -rf :parquet-hadoop
(base) [jw@cn06 parquet]$
I can access github.com, because I git clone'd from it.
(base) [jw@cn06 parquet]$ ls /data/jw/parquet/parquet-hadoop/target/failsafe-reports
TEST-org.apache.parquet.encodings.FileEncodingsIT.xml org.apache.parquet.encodings.FileEncodingsIT.txt
TEST-org.apache.parquet.hadoop.ITTestEncryptionOptions.xml org.apache.parquet.hadoop.ITTestEncryptionOptions.txt
failsafe-summary.xml
(base) [jw@cn06 parquet]$ nano /data/jw/parquet/parquet-hadoop/target/failsafe-reports/failsafe-summary.xml
shows
Gabor Szadovszky / @gszadovszky:
Could you try to get the file https://github.com/apache/parquet-testing/raw/40379b3/data/encrypt_columns_and_footer.parquet.encrypted manually (e.g. with wget or curl)?
It would also help if you could dig up the whole stack trace from failsafe-reports.
@andersonm-ibm, do you have an idea why it is happening and why we get the error message github.com: Name or service not known?
Lutz Weischer:
Switched to parquet-mr.
[jw@cn05 parquet]$ LC_ALL=C mvn install
...
[INFO]
[INFO] — maven-jar-plugin:2.4:jar (default-jar) @ parquet-hadoop —
[INFO] Building jar: /home/jw/apache/parquet/parquet-hadoop/target/parquet-hadoop-1.12.0-SNAPSHOT.jar
[INFO]
[INFO] — maven-site-plugin:3.4:attach-descriptor (attach-descriptor) @ parquet-hadoop —
[INFO]
[INFO] — maven-jar-plugin:2.4:test-jar (default) @ parquet-hadoop —
[INFO] Building jar: /home/jw/apache/parquet/parquet-hadoop/target/parquet-hadoop-1.12.0-SNAPSHOT-tests.jar
[INFO]
[INFO] — maven-shade-plugin:3.2.1:shade (default) @ parquet-hadoop —
[INFO] Excluding org.apache.parquet:parquet-column:jar:1.12.0-SNAPSHOT from the shaded jar.
[INFO] Excluding org.apache.parquet:parquet-common:jar:1.12.0-SNAPSHOT from the shaded jar.
[INFO] Excluding org.apache.yetus:audience-annotations:jar:0.12.0 from the shaded jar.
[INFO] Excluding org.apache.parquet:parquet-encoding:jar:1.12.0-SNAPSHOT from the shaded jar.
[INFO] Including net.openhft:zero-allocation-hashing:jar:0.9 in the shaded jar.
[INFO] Excluding org.apache.parquet:parquet-format-structures:jar:1.12.0-SNAPSHOT from the shaded jar.
[INFO] Excluding org.slf4j:slf4j-api:jar:1.7.22 from the shaded jar.
[INFO] Excluding javax.annotation:javax.annotation-api:jar:1.3.2 from the shaded jar.
[INFO] Excluding org.apache.parquet:parquet-jackson:jar:1.12.0-SNAPSHOT from the shaded jar.
[INFO] Including com.fasterxml.jackson.core:jackson-core:jar:2.12.2 in the shaded jar.
[INFO] Including com.fasterxml.jackson.core:jackson-databind:jar:2.12.2 in the shaded jar.
[INFO] Including com.fasterxml.jackson.core:jackson-annotations:jar:2.12.2 in the shaded jar.
[INFO] Excluding org.xerial.snappy:snappy-java:jar:1.1.8 from the shaded jar.
[INFO] Excluding commons-pool:commons-pool:jar:1.6 from the shaded jar.
[INFO] Including it.unimi.dsi:fastutil:jar:8.4.2 in the shaded jar.
[INFO] Excluding com.github.luben:zstd-jni:jar:1.4.9-1 from the shaded jar.
[INFO] Minimizing jar org.apache.parquet:parquet-hadoop:jar:1.12.0-SNAPSHOT
[INFO] Minimized 11128 -> 491 (4%)
[INFO] Replacing original artifact with shaded artifact.
[INFO] Replacing /home/jw/apache/parquet/parquet-hadoop/target/parquet-hadoop-1.12.0-SNAPSHOT.jar with /home/jw/apache/parquet/parquet-hadoop/target/parquet-hadoop-1.12.0-SNAPSHOT-shaded.jar
[INFO] Dependency-reduced POM written at: /home/jw/apache/parquet/parquet-hadoop/dependency-reduced-pom.xml
[INFO]
[INFO] — maven-failsafe-plugin:2.10:integration-test (default) @ parquet-hadoop —
[INFO] Failsafe report directory: /home/jw/apache/parquet/parquet-hadoop/target/failsafe-reports
T E S T S
Running org.apache.parquet.hadoop.ITTestEncryptionOptions
[main] INFO org.apache.parquet.hadoop.TestEncryptionOptions - ======== testInteropReadEncryptedParquetFiles target/parquet-testing/data ========
[main] INFO org.apache.parquet.hadoop.TestEncryptionOptions - Download interop files if needed
WARNING: An illegal reflective access operation has occurred
WARNING: Illegal reflective access by org.apache.hadoop.security.authentication.util.KerberosUtil (file:/home/jw/.m2/repository/org/apache/hadoop/hadoop-auth/2.10.1/hadoop-auth-2.10.1.jar) to method sun.security.krb5.Config.getInstance()
WARNING: Please consider reporting this to the maintainers of org.apache.hadoop.security.authentication.util.KerberosUtil
WARNING: Use --illegal-access=warn to enable warnings of further illegal reflective access operations
WARNING: All illegal access operations will be denied in a future release
[main] WARN org.apache.hadoop.util.NativeCodeLoader - Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
10:20:52,298 DEBUG Tracer: 106 - sampler.classes = ; loaded no samplers
10:20:52,318 DEBUG Tracer: 128 - span.receiver.classes = ; loaded no span receivers
[main] INFO org.apache.parquet.hadoop.TestEncryptionOptions - target/parquet-testing/data exists?: true
[main] INFO org.apache.parquet.hadoop.TestEncryptionOptions - Download interop file: https://github.com/apache/parquet-testing/raw/40379b3/data/uniform_encryption.parquet.encrypted
Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 8.554 sec <<< FAILURE!
Running org.apache.parquet.encodings.FileEncodingsIT
Testing codecs: [UNCOMPRESSED]
Testing PARQUET_1_0/BOOLEAN/UNCOMPRESSED encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384
Testing PARQUET_2_0/BOOLEAN/UNCOMPRESSED encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384
Testing PARQUET_1_0/BOOLEAN/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384
Testing PARQUET_2_0/BOOLEAN/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384
Testing PARQUET_1_0/INT32/UNCOMPRESSED encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384
Testing PARQUET_2_0/INT32/UNCOMPRESSED encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384
Testing PARQUET_1_0/INT32/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384
Testing PARQUET_2_0/INT32/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384
Testing PARQUET_1_0/INT64/UNCOMPRESSED encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384
Testing PARQUET_2_0/INT64/UNCOMPRESSED encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384
Testing PARQUET_1_0/INT64/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384
Testing PARQUET_2_0/INT64/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384
Testing PARQUET_1_0/INT96/UNCOMPRESSED encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384
Testing PARQUET_2_0/INT96/UNCOMPRESSED encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384
Testing PARQUET_1_0/INT96/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384
Testing PARQUET_2_0/INT96/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384
Testing PARQUET_1_0/FLOAT/UNCOMPRESSED encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384
Testing PARQUET_2_0/FLOAT/UNCOMPRESSED encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384
Testing PARQUET_1_0/FLOAT/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384
Testing PARQUET_2_0/FLOAT/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384
Testing PARQUET_1_0/DOUBLE/UNCOMPRESSED encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384
Testing PARQUET_2_0/DOUBLE/UNCOMPRESSED encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384
Testing PARQUET_1_0/DOUBLE/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384
Testing PARQUET_2_0/DOUBLE/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384
Testing PARQUET_1_0/BINARY/UNCOMPRESSED encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384
Testing PARQUET_2_0/BINARY/UNCOMPRESSED encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384
Testing PARQUET_1_0/BINARY/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384
Testing PARQUET_2_0/BINARY/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384
Testing PARQUET_1_0/FIXED_LEN_BYTE_ARRAY/UNCOMPRESSED encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384
Testing PARQUET_2_0/FIXED_LEN_BYTE_ARRAY/UNCOMPRESSED encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384
Testing PARQUET_1_0/FIXED_LEN_BYTE_ARRAY/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384
Testing PARQUET_2_0/FIXED_LEN_BYTE_ARRAY/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384
Tests run: 16, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1,108.336 sec
Results :
Tests in error:
testInteropReadEncryptedParquetFiles(org.apache.parquet.hadoop.ITTestEncryptionOptions): github.com: Name or service not known
Tests run: 17, Failures: 0, Errors: 1, Skipped: 0
[INFO]
[INFO] — maven-failsafe-plugin:2.10:verify (default) @ parquet-hadoop —
[INFO] Failsafe report directory: /home/jw/apache/parquet/parquet-hadoop/target/failsafe-reports
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary for Apache Parquet MR 1.12.0-SNAPSHOT:
[INFO]
[INFO] Apache Parquet MR .................................. SUCCESS [ 10.510 s]
[INFO] Apache Parquet Format Structures ................... SUCCESS [ 9.290 s]
[INFO] Apache Parquet Generator ........................... SUCCESS [ 3.472 s]
[INFO] Apache Parquet Common .............................. SUCCESS [ 10.361 s]
[INFO] Apache Parquet Encodings ........................... SUCCESS [ 37.394 s]
[INFO] Apache Parquet Column .............................. SUCCESS [01:41 min]
[INFO] Apache Parquet Arrow ............................... SUCCESS [ 3.497 s]
[INFO] Apache Parquet Jackson ............................. SUCCESS [ 5.068 s]
[INFO] Apache Parquet Hadoop .............................. FAILURE [23:20 min]
[INFO] Apache Parquet Avro ................................ SKIPPED
[INFO] Apache Parquet Benchmarks .......................... SKIPPED
[INFO] Apache Parquet Pig ................................. SKIPPED
[INFO] Apache Parquet Thrift .............................. SKIPPED
[INFO] Apache Parquet Cascading [Deprecated, will be removed after 1.12] SKIPPED
[INFO] Apache Parquet Cascading (for Cascading 3.0 onwards) [Deprecated, will be removed after 1.12] SKIPPED
[INFO] Apache Parquet Command-line ........................ SKIPPED
[INFO] Apache Parquet Pig Bundle .......................... SKIPPED
[INFO] Apache Parquet Protobuf ............................ SKIPPED
[INFO] Apache Parquet Scala ............................... SKIPPED
[INFO] Apache Parquet Scrooge [Deprecated, will be removed after 1.12] SKIPPED
[INFO] Apache Parquet Hadoop Bundle ....................... SKIPPED
[INFO] Apache Parquet Tools; [Deprecated, will be removed after 1.12; Use parquet-cli instead] SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 26:21 min
[INFO] Finished at: 2021-03-22T10:39:29+01:00
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-failsafe-plugin:2.10:verify (default) on project parquet-hadoop: There are test failures.
[ERROR]
[ERROR] Please refer to /home/jw/apache/parquet/parquet-hadoop/target/failsafe-reports for the individual test results.
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn -rf :parquet-hadoop
[jw@cn05 parquet]$
Gabor Szadovszky / @gszadovszky:
The current master release requires 0.13.0. The build would fail with any other version of the thrift binary. Could you please explain what mvn error do you see when you use 0.13.0?
Gabor Szadovszky / @gszadovszky:
The failing test tries to download a file from github directly. The exact link is listed in the log. It seems github.com is not accessible from your environment.
Lutz Weischer:
[jw@cn06 parquet]$ mvn package
[INFO] Scanning for projects...
[INFO]
[INFO] -----------------< org.apache.parquet:parquet-format >------------------
[INFO] Building Apache Parquet Format 2.9.0-SNAPSHOT
[INFO] -------------------------------[ jar ]--------------------------------
[INFO]
[INFO] — thrift-maven-plugin:0.10.0:compile (thrift-sources) @ parquet-format —
[INFO]
[INFO] — exec-maven-plugin:1.2.1:exec (check-thrift-version) @ parquet-format —
========== [FATAL] Build is configured to require Thrift version 0.12.0 ==========
========== Currently installed: Thrift version 0.13.0
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 2.682 s
[INFO] Finished at: 2021-03-17T08:54:58+01:00
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.codehaus.mojo:exec-maven-plugin:1.2.1:exec (check-thrift-version) on project parquet-format: Command execution failed.: Process exited with an error: 1 (Exit value: 1) -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException
[jw@cn06 parquet]$ thrift -version
Thrift version 0.13.0
[jw@cn06 parquet]$ mvn install -DskipTests
[INFO] Scanning for projects...
[INFO]
[INFO] -----------------< org.apache.parquet:parquet-format >------------------
[INFO] Building Apache Parquet Format 2.9.0-SNAPSHOT
[INFO] -------------------------------[ jar ]--------------------------------
[INFO]
[INFO] — thrift-maven-plugin:0.10.0:compile (thrift-sources) @ parquet-format —
[INFO]
[INFO] — exec-maven-plugin:1.2.1:exec (check-thrift-version) @ parquet-format —
========== [FATAL] Build is configured to require Thrift version 0.12.0 ==========
========== Currently installed: Thrift version 0.13.0
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 2.428 s
[INFO] Finished at: 2021-03-17T08:56:36+01:00
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.codehaus.mojo:exec-maven-plugin:1.2.1:exec (check-thrift-version) on project parquet-format: Command execution failed.: Process exited with an error: 1 (Exit value: 1) -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException
[jw@cn06 parquet]$
Gabor Szadovszky / @gszadovszky:
So, you are trying to build parquet-format (and not parquet-mr). parquet-format still configured for thrift 0.12.0 while it does not really means anything because the parquet-format thrift compile is only for testing purposes. The production code for parquet-format is generated inside parquet-mr (in the module parquet-format-structures).
Why are you trying to build parquet-format? If you are interested in the java implementation of parquet I would suggest looking into parquet-mr.
Lutz Weischer:
I got a fresh installation and have now this result:
(base) [jw@cn05 parquet]$ LC_ALL=C mvn install -DskipTests
...
[INFO] Apache Parquet Pig ................................. SUCCESS [ 9.340 s]
[INFO] Apache Parquet Thrift .............................. SUCCESS [ 11.328 s]
[INFO] Apache Parquet Cascading [Deprecated, will be removed after 1.12] FAILURE [ 1.496 s]
[INFO] Apache Parquet Cascading (for Cascading 3.0 onwards) [Deprecated, will be removed after 1.12] SKIPPED
[INFO] Apache Parquet Command-line ........................ SKIPPED
[INFO] Apache Parquet Pig Bundle .......................... SKIPPED
[INFO] Apache Parquet Protobuf ............................ SKIPPED
[INFO] Apache Parquet Scala ............................... SKIPPED
[INFO] Apache Parquet Scrooge [Deprecated, will be removed after 1.12] SKIPPED
[INFO] Apache Parquet Hadoop Bundle ....................... SKIPPED
[INFO] Apache Parquet Tools; [Deprecated, will be removed after 1.12; Use parquet-cli instead] SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 03:21 min
[INFO] Finished at: 2021-03-26T14:08:18+01:00
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal com.github.siom79.japicmp:japicmp-maven-plugin:0.14.2:cmp (default) on project parquet-cascading-deprecated: Could not resolve org.apache.parquet:parquet-cascading:jar:1.12.0 -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn -rf :parquet-cascading-deprecated
(base) [jw@cn05 parquet]$
Trying differently:
(base) [jw@cn05 parquet]$ mvn install -DskipTests
...
[INFO] Apache Parquet Pig ................................. SUCCESS [ 8.357 s]
[INFO] Apache Parquet Thrift .............................. SUCCESS [ 10.033 s]
[INFO] Apache Parquet Cascading [Deprecated, will be removed after 1.12] FAILURE [ 0.686 s]
[INFO] Apache Parquet Cascading (for Cascading 3.0 onwards) [Deprecated, will be removed after 1.12] SKIPPED
[INFO] Apache Parquet Command-line ........................ SKIPPED
[INFO] Apache Parquet Pig Bundle .......................... SKIPPED
[INFO] Apache Parquet Protobuf ............................ SKIPPED
[INFO] Apache Parquet Scala ............................... SKIPPED
[INFO] Apache Parquet Scrooge [Deprecated, will be removed after 1.12] SKIPPED
[INFO] Apache Parquet Hadoop Bundle ....................... SKIPPED
[INFO] Apache Parquet Tools; [Deprecated, will be removed after 1.12; Use parquet-cli instead] SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 03:06 min
[INFO] Finished at: 2021-03-26T14:33:17+01:00
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.rat:apache-rat-plugin:0.12:check (default) on project parquet-cascading-deprecated: Too many files with unapproved license: 1 See RAT report in: /data/jw/apache/metron/parquet/parquet-cascading-deprecated/target/rat.txt -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn -rf :parquet-cascading-deprecated
(base) [jw@cn05 parquet]$
And:
(base) [jw@cn05 parquet]$ mvn install -DskipTests -Drat.skip=true
...
[INFO] Apache Parquet Pig ................................. SUCCESS [ 8.106 s]
[INFO] Apache Parquet Thrift .............................. SUCCESS [ 9.405 s]
[INFO] Apache Parquet Cascading [Deprecated, will be removed after 1.12] FAILURE [ 0.566 s]
[INFO] Apache Parquet Cascading (for Cascading 3.0 onwards) [Deprecated, will be removed after 1.12] SKIPPED
[INFO] Apache Parquet Command-line ........................ SKIPPED
[INFO] Apache Parquet Pig Bundle .......................... SKIPPED
[INFO] Apache Parquet Protobuf ............................ SKIPPED
[INFO] Apache Parquet Scala ............................... SKIPPED
[INFO] Apache Parquet Scrooge [Deprecated, will be removed after 1.12] SKIPPED
[INFO] Apache Parquet Hadoop Bundle ....................... SKIPPED
[INFO] Apache Parquet Tools; [Deprecated, will be removed after 1.12; Use parquet-cli instead] SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 03:03 min
[INFO] Finished at: 2021-03-26T14:42:35+01:00
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal com.github.siom79.japicmp:japicmp-maven-plugin:0.14.2:cmp (default) on project parquet-cascading-deprecated: Could not resolve org.apache.parquet:parquet-cascading:jar:1.12.0 -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn -rf :parquet-cascading-deprecated
(base) [jw@cn05 parquet]$
Gabor Szadovszky / @gszadovszky:
[~LutzWeischerFujitsu], this is the previous issue you've already posted. Please, check my [previous comment|#comment-17307872].
Lutz Weischer:
Looks like we've progress but we're not there yet, due to failing tests. Using Java 11, setting ANT_OPTS with proxy:
(base) [jw@cn06 parquet]$ LC_ALL=C mvn install
...
Testing PARQUET_2_0/FIXED_LEN_BYTE_ARRAY/UNCOMPRESSED + DICTIONARY encodings using ROW_GROUP_SIZE=131072 PAGE_SIZE=16384
Tests run: 16, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1,062.679 sec
Results :
Tests in error:
testInteropReadEncryptedParquetFiles(org.apache.parquet.hadoop.ITTestEncryptionOptions): github.com: Name or service not known
Tests run: 17, Failures: 0, Errors: 1, Skipped: 0
[INFO]
[INFO] — maven-failsafe-plugin:2.10:verify (default) @ parquet-hadoop —
[INFO] Failsafe report directory: /data/jw/apache/parquet/parquet-hadoop/target/failsafe-reports
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary for Apache Parquet MR 1.13.0-SNAPSHOT:
[INFO]
[INFO] Apache Parquet MR .................................. SUCCESS [ 8.110 s]
[INFO] Apache Parquet Format Structures ................... SUCCESS [ 8.766 s]
[INFO] Apache Parquet Generator ........................... SUCCESS [ 3.380 s]
[INFO] Apache Parquet Common .............................. SUCCESS [ 9.466 s]
[INFO] Apache Parquet Encodings ........................... SUCCESS [ 35.689 s]
[INFO] Apache Parquet Column .............................. SUCCESS [01:33 min]
[INFO] Apache Parquet Arrow ............................... SUCCESS [ 3.252 s]
[INFO] Apache Parquet Jackson ............................. SUCCESS [ 4.634 s]
[INFO] Apache Parquet Hadoop .............................. FAILURE [22:20 min]
[INFO] Apache Parquet Avro ................................ SKIPPED
[INFO] Apache Parquet Benchmarks .......................... SKIPPED
[INFO] Apache Parquet Pig ................................. SKIPPED
[INFO] Apache Parquet Thrift .............................. SKIPPED
[INFO] Apache Parquet Cascading [Deprecated, will be removed after 1.12] SKIPPED
[INFO] Apache Parquet Cascading (for Cascading 3.0 onwards) [Deprecated, will be removed after 1.12] SKIPPED
[INFO] Apache Parquet Command-line ........................ SKIPPED
[INFO] Apache Parquet Pig Bundle .......................... SKIPPED
[INFO] Apache Parquet Protobuf ............................ SKIPPED
[INFO] Apache Parquet Scala ............................... SKIPPED
[INFO] Apache Parquet Scrooge [Deprecated, will be removed after 1.12] SKIPPED
[INFO] Apache Parquet Hadoop Bundle ....................... SKIPPED
[INFO] Apache Parquet Tools; [Deprecated, will be removed after 1.12; Use parquet-cli instead] SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 25:08 min
[INFO] Finished at: 2021-04-01T12:04:30+02:00
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-failsafe-plugin:2.10:verify (default) on project parquet-hadoop: There are test failures.
[ERROR]
[ERROR] Please refer to /data/jw/apache/parquet/parquet-hadoop/target/failsafe-reports for the individual test results.
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn -rf :parquet-hadoop
(base) [jw@cn06 parquet]$
Lutz Weischer:
wget https://github.com/apache/parquet-testing/raw/40379b3/data/encrypt_columns_and_footer.parquet.encrypted
works.
LC_ALL=C mvn clean install
which includes the tests, works, using Java 16.
Maya Anderson / @andersonm-ibm:
@gszadovszky, just like you suggested, I expect the same error when trying to get the file [https://github.com/apache/parquet-testing/raw/40379b3/data/encrypt_columns_and_footer.parquet.encrypted] manually (e.g. with wget or curl), since this looks like a DNS resolution issue.
[~LutzWeischerFujitsu] , to isolate the problem - does adding github.com directly to /etc/hosts make any difference?
Gabor Szadovszky / @gszadovszky:
[~LutzWeischerFujitsu], this is a known issue. See PARQUET-2010 for details. You may either cherry-pick my change in the related PR or step back to the previous release in the history to go forward. I hope my change will get in soon so everyone would be able to continue working on the new release.
Gabor Szadovszky / @gszadovszky:
[~LutzWeischerFujitsu], PARQUET-2010 is already in master so it should work now.
Lutz Weischer:
Please be so kind to inform when the new (master) release is available and should not have that issue any more. I will than re-try. Thanks.
|
gharchive/issue
| 2021-03-16T12:18:35 |
2025-04-01T04:55:59.750068
|
{
"authors": [
"asfimport"
],
"repo": "apache/parquet-java",
"url": "https://github.com/apache/parquet-java/issues/2587",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1716956130
|
Extensible PinotDataBuffer SPI
This is a PEP-Request. I propose to create a extensible SPI that can be used to change the PinotDataBuffer implementation used at runtime. Treat this issue as a clone of https://github.com/apache/pinot/issues/9162, but more formal.
What needs to be done?
Right now PinotDataBuffer instances are obtained by calling static methods:
PinotDataBuffer allocateDirect(long size, ByteOrder byteOrder, String description)
PinotDataBuffer loadFile(File file, long offset, long size, ByteOrder byteOrder, String description)
PinotDataBuffer mapFile(File file, boolean readOnly, long offset, long size, ByteOrder byteOrder, String description)
These methods have a static implementation that return a PinotDataBuffer that is backed by a ByteBuffer or by a LArray, depending on whether the requested size is greater than 2GBs or not. ByteBuffers are faster and more reliable than LArray, but they cannot be larger than Integer.MAX_INT (aka 2GB - 1byte).
Here I propose to change that implementation. Instead of having a static implementation, the algorithm used to instantiate the PinotDataBuffer will be delegated to an interface. The specific implementation will be set at Pinot startup time by reading the property pinot.offheap.buffer.factory, which will be an optional property whose value is a String. In case it exists, it should the qualified class name of a Class that implements PinotDataBufferFactory. There will be also another optional property called pinot.offheap.prioritize.bytebuffer which will be a boolean. If it is true, then the ByteBuffer implementation will be used when less than 2GB buffers are requested.
Why the feature is needed
The reason to be able to have new PinotDataBuffer implementations is that LArray is not compatible with Java >= 16, as detected in #8529. It also seems that our LArray implementation may have some bugs (see https://github.com/apache/pinot/pull/10774) and it doesn't seem that LArray will be updated (See https://github.com/xerial/larray/issues/75), so we need to move on.
Initial idea/proposal
Adding a SPI is something we have done several times in the last months/years. Most of the times we use one or more configuration properties to decide which instance we need to use and set that instance in a static attribute. That system works, but it is problematic when we want to change the implementation in an isolated way (for example, when doing tests).
That is why I propose to extend this system with a thread local on top of that. Something like:
/**
* The default {@link PinotBufferFactory} used by all threads that do not define their own factory.
*/
private static PinotBufferFactory _defaultFactory = createDefaultFactory();
/**
* A thread local variable that can be used to customize the {@link PinotBufferFactory} used on tests. This is mostly
* useful in tests.
*/
private static final ThreadLocal<PinotBufferFactory> _FACTORY = new ThreadLocal<>();
/**
* Creates the default factory depending on the JVM version
*/
public static PinotBufferFactory createDefaultFactory() {...}
/**
* Changes the default factory
*/
public static void setDefaultFactory(PinotBufferFactory) {...}
/**
* Change the {@link PinotBufferFactory} used by the current thread.
*
* @see #loadDefaultFactory(PinotConfiguration)
*/
public static void useFactory(PinotBufferFactory factory) {
_FACTORY.set(factory);
}
/**
* Returns the factory the current thread should use.
*/
public static PinotBufferFactory getFactory() {
PinotBufferFactory pinotBufferFactory = _FACTORY.get();
if (pinotBufferFactory == null) {
pinotBufferFactory = _defaultFactory;
}
return pinotBufferFactory;
}
Then the static methods PinotDataBuffer.allocateDirect (and others) should delegate on PinotDataBuffer.getFactory(). By doing that, tests can call PinotDataBuffer.useFactory() in order to use, on that thread, the buffer library they want to test. Exploratory draft https://github.com/apache/pinot/pull/10528 implements all of these and modifies tests to use this API.
Actual buffer implementations
This is an optional part of the PEP. What we request is to have the PinotBufferFactory SPI. This is what we would like to do with it, but it could be part of another PEP if necessary.
Given that we cannot use LArray in modern versions of Java, we have three alternatives :
Use other third party libraries
Implement our own library on top of Unsafe
Use Foreign Memory API
I tried to use Chronicle Bytes in the past with partial success, see https://github.com/apache/pinot/pull/9842. Given that I don't know more third party buffer libraries that use long offsets (Netty and Agrona use ints), I explored the last two alternatives in https://github.com/apache/pinot/pull/10528.
Foreign Memory API is clearly the future, as it provides exactly what we need: A high performance buffer API that is maintained (given that will be included in the JVM) and uses longs as offsets and sizes. But it has two issues:
It is not included in Java 17
It is still in preview mode and it will continue like that in Java 21 (see this post where one of the author explains the reasons and the changes since Java 20).
Therefore we have three options:
Wait until it is stable (possible in Java 22 or 23, which won't be LTS, so maybe we need to wait 2 years until Java 25).
Use it as a preview in Java 21. This would imply that we could either run with Java 11+LArray or Java 21+Foreign, but do not support Java 17. Also, in Java 21 we would need to compile and start with --enable-preview.
Create our own library.
Previously mentioned https://github.com/apache/pinot/pull/10528 does create two buffer implementations: One on top of Unsafe and another on top of Foreign Memory API. The latter is trivial to implement, but it is very difficult to include in the current Apache Pinot CI. The problems I found are:
Some Maven plugins don't work in Java 21-ea (at least spotless doesn't work)
GitHub Actions cannot be configured to use Java 21-ea (see https://github.com/actions/setup-java/issues/492#issuecomment-1551508335).
In general, it is difficult to maintain in the same compilation unit code that has to be compiled with Java 11 and code that has to be compiled with Java 21
Presto still requires to use Java 8, which makes it even more difficult.
In case we decide to use my draft in the actual implementation of the PEP, I would suggest to remove the Foreign Memory API implementation from the branch and optionally move it to another github repo.
I've also modified BenchmarkPinotDataBuffer in order to run with the Unsafe based implementation. The modifications are in the draft.
I've run the benchmark in a M1 Pro and in a Ryzen 9 3900X with Ubuntu 22.10. The exact results of the benchmark can be found here but the following chars should be good enough:
Please note that LArray only runs in Java 11, so there are no data with LArray in Java 17 and 21.
This one shows the cost of executing batch writes (aka call PinotDataBuffer.readFrom) and batch read (aka call PinotDataBuffer.copyTo) with a byte array of 1024 elements (units are ns/op in Linux):
It is important to note that the implementation used in both LArray and Unsafe when dealing with batch reads and writes is to create a temporal direct byte buffer that points to the same address of the current buffer, so we can assume that the performance difference is due to the new instance creation.
This one shows the cost of executing non batch writes (calling PinotDataBuffer.putByte()) and reads (calling PinotDataBuffer.getByte) in a loop with 1024 consecutive offsets starting from a random one (units are ns/op in Linux):
In my opinion the latest is the most interesting, as it shows the improvements introduced in Java 17 and 21 JIT when dealing with loops.
I've also tried to add the Foreign Memory API implementation in the benchmark, but I found several problems so I gave up.
This is amazing @gortiz . I am +1
+1 from me as well.
+1, awesome work!
+1. Thanks @gortiz
So I am guessing the next plan of action is to:
Review the SPI and Unsafe based implementation in https://github.com/apache/pinot/pull/10528
Also, Foreign Memory is not ruled out right. Later on we should move to Foreign Memory API when stable support is available ?
Review the SPI and Unsafe based implementation in https://github.com/apache/pinot/pull/10528
I have some issues there. Some tests always get out of memory in GitHub Actions, while my amd pc and my M1 laptop run them without problems, even when running with mvn -P github-actions. Given that I cannot reproduce the problem locally, it is being very difficult to fix the problem. Any help there would be helpful.
Later on we should move to Foreign Memory when stable support is available ?
I guess we should decide what we want to do with that. We can either:
Wait until Foreign Memory API is production ready and stay in Java 11 meanwhile.
Versions on which we could run Pinot: [11, 15] and [22, ...).
Assuming that Foreign Memory API is released in Java 22.
Advantage: we will only use proven and stable apis.
Disadvantages: But it implies that we would need to either run with Java 11 until Java 23 or we would need to use non LTS versions.
Support preview of Foreign Memory API:
Versions on which we could run Pinot: [11, 15] and [21, ...).
Advantage: We would support Java 21 if running with --enable-preview
Disadvantage:
If OpenJDK decides to add more breaking changes in Java 22, we would need to maintain different versions of the code.
We would need to run with --preview-enable, which is not intended to be executed in production.
Use our own library on top of Unsafe
Versions on which we could run Pinot: [11, ...).
Advantages:
In Java [11, 15] we can choose to use LArray or Unsafe
In Java [16, 20] we can only use Unsafe
In Java 21 we can use Unsafe or run with --enable-preview and use Foreign Memory API
Once Foreign Memory API is released, we could always use Foreign Memory API
Disadvantages:
We need to maintain the Unsafe library.
Errors there may kill the JVM with segmentation faults. Details may depend on CPU architectures and OS. For example, mmap needs to be page aligned in Linux (otherwise segmentation fault kills the JVM).
Unsafe methods do change between major JVM versions.
I was able to find the reason why some tests were failing in https://github.com/apache/pinot/pull/10528. It was a Maven config problem related on how I was configuring the surefire argLine property.
Now I have green lights in Java 11 and Java 20. Last execution failed in Java 17 and I'm trying to verify if it is a flaky test or something actually related to these changes.
I'm thinking on removing the draft tag from the PR. We should discuss whether to include the Foreign Memory API or not in the PR. I think it is better to just drop that code and add them later once is it stable (or if we want to invest resources into trying with --preview-enabled)
We may want to close this issue
|
gharchive/issue
| 2023-05-19T09:45:08 |
2025-04-01T04:55:59.776906
|
{
"authors": [
"gortiz",
"kishoreg",
"mayankshriv",
"siddharthteotia",
"xiangfu0"
],
"repo": "apache/pinot",
"url": "https://github.com/apache/pinot/issues/10783",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1085236740
|
Add instance id in response
Description
This PR adds instance id at the error message in the response. Currently the response is like below, which is hard to identify the target server:
"errors": [
{
"message": "ServerTableMissing:\nFailed to find table: testTable_OFFLINE",
"type": "UNKNOWN"
}
]
Upgrade Notes
Does this PR prevent a zero down-time upgrade? (Assume upgrade order: Controller, Broker, Server, Minion)
[ ] Yes (Please label as backward-incompat, and complete the section below on Release Notes)
Does this PR fix a zero-downtime upgrade introduced earlier?
[ ] Yes (Please label this as backward-incompat, and complete the section below on Release Notes)
Does this PR otherwise need attention when creating release notes? Things to consider:
New configuration options
Deprecation of configurations
Signature changes to public methods/interfaces
New plugins added or old plugins removed
[ ] Yes (Please label this PR as release-notes and complete the section on Release Notes)
Release Notes
Documentation
Codecov Report
Merging #7937 (dfaf9e9) into master (aa2da07) will decrease coverage by 57.00%.
The diff coverage is 0.00%.
@@ Coverage Diff @@
## master #7937 +/- ##
=============================================
- Coverage 71.32% 14.32% -57.01%
+ Complexity 4092 80 -4012
=============================================
Files 1589 1548 -41
Lines 82139 80502 -1637
Branches 12270 12067 -203
=============================================
- Hits 58589 11532 -47057
- Misses 19578 68109 +48531
+ Partials 3972 861 -3111
Flag
Coverage Δ
integration1
?
integration2
?
unittests1
?
unittests2
14.32% <0.00%> (-0.04%)
:arrow_down:
Flags with carried forward coverage won't be shown. Click here to find out more.
Impacted Files
Coverage Δ
...core/query/executor/ServerQueryExecutorV1Impl.java
0.00% <0.00%> (-83.34%)
:arrow_down:
...ain/java/org/apache/pinot/core/data/table/Key.java
0.00% <0.00%> (-100.00%)
:arrow_down:
.../java/org/apache/pinot/spi/utils/BooleanUtils.java
0.00% <0.00%> (-100.00%)
:arrow_down:
.../java/org/apache/pinot/core/data/table/Record.java
0.00% <0.00%> (-100.00%)
:arrow_down:
.../java/org/apache/pinot/core/util/GroupByUtils.java
0.00% <0.00%> (-100.00%)
:arrow_down:
...ava/org/apache/pinot/spi/config/table/FSTType.java
0.00% <0.00%> (-100.00%)
:arrow_down:
...ava/org/apache/pinot/spi/data/MetricFieldSpec.java
0.00% <0.00%> (-100.00%)
:arrow_down:
...va/org/apache/pinot/spi/utils/BigDecimalUtils.java
0.00% <0.00%> (-100.00%)
:arrow_down:
...java/org/apache/pinot/common/tier/TierFactory.java
0.00% <0.00%> (-100.00%)
:arrow_down:
...java/org/apache/pinot/common/utils/StringUtil.java
0.00% <0.00%> (-100.00%)
:arrow_down:
... and 1264 more
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update aa2da07...dfaf9e9. Read the comment docs.
|
gharchive/pull-request
| 2021-12-20T21:56:36 |
2025-04-01T04:55:59.801210
|
{
"authors": [
"codecov-commenter",
"jackjlli"
],
"repo": "apache/pinot",
"url": "https://github.com/apache/pinot/pull/7937",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1357886992
|
[WIP] Adaptive Server Selection to improve Query Processing Resiliency
label = feature
OSS issue https://github.com/apache/pinot/issues/8618
Design Doc
PR Description
Based on the segment assignment strategies used, we use one of the following instance selectors today:
BalancedInstanceSelector
ReplicaGroupInstanceSelector
StrictReplicaGroupInstanceSelector
Irrespective of which instance selection mechanism is used, we use a round robin approach. The round-robin approach is not sensitive to changes in the system like server slowness, GC on servers, etc. Hence we could end up overloading some servers especially in cases where they are underperforming thereby leading to higher latencies for queries.
This PR provides the implementation for the Adaptive Server Selection feature at brokers. When a query is received, we could use one of the implemented Adaptive Selectors (NumInFlightRequests, Latency, Hybrid) to efficiently route queries to the best server instead of using a naive round robin approach.
Test Results
TBD
Codecov Report
Merging #9311 (d975dbf) into master (1c9528c) will decrease coverage by 41.46%.
The diff coverage is 15.98%.
@@ Coverage Diff @@
## master #9311 +/- ##
=============================================
- Coverage 69.73% 28.27% -41.47%
+ Complexity 5017 53 -4964
=============================================
Files 1873 1868 -5
Lines 99598 99613 +15
Branches 15163 15169 +6
=============================================
- Hits 69453 28163 -41290
- Misses 25217 68713 +43496
+ Partials 4928 2737 -2191
Flag
Coverage Δ
integration1
26.05% <12.69%> (-0.09%)
:arrow_down:
integration2
24.81% <14.97%> (+0.02%)
:arrow_up:
unittests1
?
unittests2
?
Flags with carried forward coverage won't be shown. Click here to find out more.
Impacted Files
Coverage Δ
...e/pinot/broker/api/resources/PinotBrokerDebug.java
82.75% <0.00%> (-2.96%)
:arrow_down:
...roker/requesthandler/BaseBrokerRequestHandler.java
60.47% <ø> (-9.78%)
:arrow_down:
...routing/adaptiveserverselector/HybridSelector.java
0.00% <0.00%> (ø)
...outing/adaptiveserverselector/LatencySelector.java
0.00% <0.00%> (ø)
...adaptiveserverselector/NumInFlightReqSelector.java
0.00% <0.00%> (ø)
...e/pinot/common/utils/ExponentialMovingAverage.java
0.00% <0.00%> (ø)
...org/apache/pinot/core/transport/QueryResponse.java
100.00% <ø> (ø)
.../server/routing/stats/ServerRoutingStatsEntry.java
0.00% <0.00%> (ø)
...va/org/apache/pinot/spi/utils/CommonConstants.java
0.00% <0.00%> (-27.70%)
:arrow_down:
...erver/routing/stats/ServerRoutingStatsManager.java
6.04% <6.04%> (ø)
... and 1299 more
:mega: We’re building smart automated test selection to slash your CI/CD build times. Learn more
@Jackie-Jiang Thanks for the comments. I've addressed them in #9462. Please review.
I'll add the following soon.
Release-note in PR description.
Example of configs to use this feature.
Pinot docs.
Merged user docs - https://docs.pinot.apache.org/operators/operating-pinot/tuning/query-routing-using-adaptive-server-selection
|
gharchive/pull-request
| 2022-08-31T21:22:05 |
2025-04-01T04:55:59.822776
|
{
"authors": [
"codecov-commenter",
"vvivekiyer"
],
"repo": "apache/pinot",
"url": "https://github.com/apache/pinot/pull/9311",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
985368535
|
Exclusive subscription mode for partitioned topics not seeing consistent order among different consumers
Describe the bug
In the official documentation of Apache Pulsar it says that an exclusive subscription sees a consistent order for a single consumer. Suppose we have multiple consumers each of them with its own exclusive subscription and they are reading from a partitioned topic. As presented in the docs:
"Decisions about routing and subscription modes can be made separately in most cases. In general, throughput concerns should guide partitioning/routing decisions while subscription decisions should be guided by application semantics.
There is no difference between partitioned topics and normal topics in terms of how subscription modes work, as partitioning only determines what happens between when a message is published by a producer and processed and acknowledged by a consumer."
Those statements lead us to infer that the readers would get a consistent global ordering among topic partitions. But that is not the case in my tests so far:
The code I have used to test it (Scala):
Producer:
package pulsar
import org.apache.pulsar.client.api.PulsarClient
import java.util.UUID
object Producer {
def main(args: Array[String]): Unit = {
val client = PulsarClient.builder()
.serviceUrl(SERVICE_URL)
.allowTlsInsecureConnection(true)
.build()
val producer = client.newProducer()
.topic(TOPIC)
.enableBatching(true)
//.accessMode(ProducerAccessMode.Exclusive)
.create()
for(i<-0 until 100){
val key = UUID.randomUUID().toString.getBytes()
//val key = s"Hello-${i}".getBytes()
producer.newMessage().orderingKey("k0".getBytes()).value(key).send()
println(s"produced msg: ${i.toString}")
}
producer.flush()
producer.close()
client.close()
}
}
Consumer:
import org.apache.pulsar.client.api.{PulsarClient, SubscriptionInitialPosition, SubscriptionType}
object Consumer {
def main(args: Array[String]): Unit = {
val client = PulsarClient.builder()
.serviceUrl(SERVICE_URL)
.allowTlsInsecureConnection(true)
.build()
var l1 = Seq.empty[String]
var l2 = Seq.empty[String]
val c1 = client.newConsumer()
.topic(TOPIC)
.subscriptionType(SubscriptionType.Exclusive)
.subscriptionInitialPosition(SubscriptionInitialPosition.Earliest)
.subscriptionName("c1")
.subscribe()
// To stop the consuption I put a limit (100) - this limit is known
while(l1.length < 100){
val msg = c1.receive()
val str = new String(msg.getData)
println(s"${Console.MAGENTA_B}$str${Console.RESET}")
l1 = l1 :+ str
//c1.acknowledge(msg.getMessageId)
}
val c2 = client.newConsumer()
.topic(TOPIC)
.subscriptionType(SubscriptionType.Exclusive)
.subscriptionInitialPosition(SubscriptionInitialPosition.Earliest)
.subscriptionName("c2")
.subscribe()
while(l2.length < 100){
val msg = c2.receive()
val str = new String(msg.getData)
println(s"${Console.GREEN_B}$str${Console.RESET}")
l2 = l2 :+ str
//c2.acknowledge(msg.getMessageId)
}
println()
println(l1)
println()
println()
println(l2)
println()
try {
assert(l1 == l2)
} finally {
c1.close()
c2.close()
client.close()
}
}
}
Am I wrong about it or Pulsar does not support the described behavior I expect?
**OBS.: I've tried every configuration for that. I've set Retention policies for the namespace as infinite both in size and number of messages. It does not work :(
I also tried:
SinglePartitionRouting for producer (it does not matter tho)
Setting an ordering key for the messages**
@lucasrpb
I think pulsar ensures the orderly consumer of one partition-topic. If multiple partition-topics are consumed in subscription mode, the a consistent global order cannot be guaranteed.
Just to make sure @shibd : Using the combo of SinglePartition config for the producer and ExclusiveSubscription for the consumer would not guarantee ordering?
Just to make sure @shibd : Using the combo of SinglePartition config for the producer and ExclusiveSubscription for the consumer would not guarantee ordering?
This can guarantee ordering.
Just to make sure @shibd : Using the combo of SinglePartition config for the producer and ExclusiveSubscription for the consumer would not guarantee ordering?
This can guarantee ordering.
https://stackoverflow.com/questions/68988929/does-pulsar-partitioned-topics-support-global-ordering-when-consuming-from-diffe/69032885#69032885
Thanks guys for the explanations! Unfortunately with the current implementation I cannot get the behavior I want! But it will be very easy to achieve it using an approach I have in mind already ! :) Maybe in the near future I can propose a pull request on it! :)
The issue had no activity for 30 days, mark with Stale label.
|
gharchive/issue
| 2021-09-01T16:42:00 |
2025-04-01T04:55:59.839911
|
{
"authors": [
"codelipenghui",
"lucasrpb",
"miguelemosreverte",
"shibd"
],
"repo": "apache/pulsar",
"url": "https://github.com/apache/pulsar/issues/11883",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
953730132
|
Add offload ledger info for admin topics stats
Motivation
Currently, we don't have offload stats when getting the topic stats bin/pulsar-admin topics stats.
We should add metrics in topic stats on last offloaded ledger id, last successful offloaded timestamp, and last offload failure timestamp.
Modifications
Add lastOffloadedLedgerId, lastOffloadSuccessTimestamp, lastOffloadFailureTimestamp for ManagedLedgerImpl and TopicStatsImpl.
Thanks for your contribution. For this PR, do we need to update docs?
(The PR template contains info about doc, which helps others know more about the changes. Can you provide doc-related info in this and future PR descriptions? Thanks)
Thanks for your contribution. For this PR, do we need to update docs?
(The PR template contains info about doc, which helps others know more about the changes. Can you provide doc-related info in this and future PR descriptions? Thanks)
Yes, we need the "REST APIs => PERSISTENT TOPIC => Get the stats for the topic" api return result updated.
|
gharchive/pull-request
| 2021-07-27T10:19:05 |
2025-04-01T04:55:59.843579
|
{
"authors": [
"Anonymitaet",
"frankxieke"
],
"repo": "apache/pulsar",
"url": "https://github.com/apache/pulsar/pull/11465",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1094986110
|
bundle metrics would overwrite loadbalance metrics
Motivation
#12366 added bundle metrics, But, If we enabled expose these metrics, It would overwrite previous load balancing metrics.
Modifications
seperate bundle metrics and load balance metrics
Documentation
Check the box below or label this PR directly (if you have committer privilege).
Need to update docs?
[ ] doc-required
(If you need help on updating docs, create a doc issue)
[x] no-need-doc
(Please explain why)
[ ] doc
(If this PR contains doc changes)
/pulsarbot run-failure-checks
/pulsarbot run-failure-checks
/pulsarbot run-failure-checks
@hangc0276 PTAL at this.
Great catch. The code looks good to me. Before we merge this, can you please update the documentation? The doc is located here: site2/docs/reference-metrics.md.
@michaeljmarshall Done.
@michaeljmarshall docs are done in https://github.com/apache/pulsar/pull/12825/files, so I label this PR w/ doc-added and remove doc-required.
@Anonymitaet - thank you. I forgot to switch labels once the docs were added.
|
gharchive/pull-request
| 2022-01-06T05:37:18 |
2025-04-01T04:55:59.849098
|
{
"authors": [
"Anonymitaet",
"gaozhangmin",
"michaeljmarshall"
],
"repo": "apache/pulsar",
"url": "https://github.com/apache/pulsar/pull/13641",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1345483676
|
[fix][doc] Migrate config docs to pulsar-site
Motivation
Pulsar configuration docs are going to be migrated to https://pulsar.apache.org/reference (Not available currently, will be after apache/pulsar-site#169 is merged. Current version is deployed here). This PR aims to adjust the docs so that everything looks fine in the new site.
Modifications
Remove config docs that are auto-generated with code. Manually maintained ones are uneffected.
Adjust the format of admonition.
Correct the invalid links.
Adjust sidebar structure.
Verifying this change
[x] Make sure that the change passes the CI checks.
This change is a trivial rework / code cleanup without any test coverage.
Does this pull request potentially affect one of the following parts:
Dependencies (does it add or upgrade a dependency): no
The public API: no
The schema: no
The default values of configurations: no
The wire protocol: no
The rest endpoints: no
The admin cli options: no
Anything that affects deployment: no
Documentation
Check the box below or label this PR directly.
Need to update docs?
[ ] doc-required
(Your PR needs to update docs and you will update later)
[] doc-not-needed
(Please explain why)
[x] doc
(Your PR contains doc changes)
[ ] doc-complete
(Docs have been already added)
@Anonymitaet PTAL
/pulsarbot run-failure-checks
/pulsarbot run-failure-checks
/pulsarbot run-failure-checks
/pulsarbot run-failure-checks
|
gharchive/pull-request
| 2022-08-21T13:09:46 |
2025-04-01T04:55:59.856649
|
{
"authors": [
"Anonymitaet",
"SignorMercurio"
],
"repo": "apache/pulsar",
"url": "https://github.com/apache/pulsar/pull/17197",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1643431457
|
[fix][build] Dump Jacoco coverage data to file with JMX interface in TestNG listener
Fixes #19931
Motivation
See #19931. Sometimes the default Jacoco shutdown hook doesn't run and there's no coverage data. This causes the Codecov upload to fail since there's no Jacoco coverage report available.
Modifications
enable Jacoco agent's JMX interface
Dump Jacoco coverage to file with JMX interface in TestNG listener
Documentation
[ ] doc
[ ] doc-required
[x] doc-not-needed
[ ] doc-complete
Matching PR in forked repository
PR in forked repository: https://github.com/lhotari/pulsar/pull/145
Yes, that's the case. Are your going to rerun the workflow in your fork and notify here once it passed, or we tag this PR as ready-to-test and rerun here?
@tisonkun re-running it. in progress at https://github.com/lhotari/pulsar/actions/runs/4540821034/jobs/8006055251?pr=145 . This time build step completed properly. Waiting for the whole build job to complete.
/pulsarbot rerun-failure-checks
Hmmm...Now we can pass the flaky tests job, but it seems the integration tests upload job is still suffering:
@lhotari the error info says "not_found" and I read it as file not created. But it seems different from the previous one.
Do we now always dump the file, but the upload action itself is still unstable? IIRC there is an upstream issue for the latter one.
@lhotari the error info says "not_found" and I read it as file not created. But it seems different from the previous one.
Do we now always dump the file, but the upload action itself is still unstable? IIRC there is an upstream issue for the latter one.
Exactly. For the upload issue, it seems that Codecov suggests to not consider the upload token as a secret:
Public repositories that rely on PRs via forks will find that they cannot effectively use Codecov if the token is stored as a GitHub secret. The scope of the Codecov token is only to confirm that the coverage uploaded comes from a specific repository, not to pull down source code or make any code changes.
For this reason, we recommend that teams with public repositories that rely on PRs via forks consider the security ramifications of making the Codecov token available as opposed to being in a secret.
|
gharchive/pull-request
| 2023-03-28T08:12:40 |
2025-04-01T04:55:59.864666
|
{
"authors": [
"lhotari",
"tisonkun"
],
"repo": "apache/pulsar",
"url": "https://github.com/apache/pulsar/pull/19947",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
450897224
|
Fixed NPE when closing batch during a reconnection
Motivation
There is a race condition when closing a batch and the producer is getting disconnected. This has been seen in some of the tests run and can be reproduced by running this particular test many times in a loop.
2019-05-31\T\16:32:49.647 [ERROR] testProducerReconnect(org.apache.pulsar.client.api.ClientErrorsTest) Time elapsed: 0.034 s <<< FAILURE!
org.apache.pulsar.client.api.PulsarClientException: java.lang.NullPointerException
at org.apache.pulsar.client.impl.ProducerImpl.batchMessageAndSend(ProducerImpl.java:1354)
at org.apache.pulsar.client.impl.ProducerImpl.triggerFlush(ProducerImpl.java:1282)
at org.apache.pulsar.client.impl.ProducerBase.send(ProducerBase.java:82)
at org.apache.pulsar.client.impl.TypedMessageBuilderImpl.send(TypedMessageBuilderImpl.java:61)
at org.apache.pulsar.client.impl.ProducerBase.send(ProducerBase.java:52)
at org.apache.pulsar.client.api.ClientErrorsTest.testProducerReconnect(ClientErrorsTest.java:600)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.testng.internal.MethodInvocationHelper.invokeMethod(MethodInvocationHelper.java:124)
at org.testng.internal.InvokeMethodRunnable.runOne(InvokeMethodRunnable.java:54)
at org.testng.internal.InvokeMethodRunnable.run(InvokeMethodRunnable.java:44)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Caused by: java.lang.NullPointerException
at org.apache.pulsar.client.impl.ProducerImpl.batchMessageAndSend(ProducerImpl.java:1329)
... 17 more
The problem is that we're checking isConnected() and then using the cnx() which can return null if the producer is disconnected. Rather, we should first cache the reference to the ClientCnx instance and then double check if we're connected.
The same pattern is already applied for the non-batching scenario:
https://github.com/apache/pulsar/blob/ba24d73b5f77c42f832cd9d2720dc0aba388f909/pulsar-client/src/main/java/org/apache/pulsar/client/impl/ProducerImpl.java#L399-L405
run java8 tests
run java8 tests
run java8 tests
|
gharchive/pull-request
| 2019-05-31T17:10:45 |
2025-04-01T04:55:59.868190
|
{
"authors": [
"merlimat"
],
"repo": "apache/pulsar",
"url": "https://github.com/apache/pulsar/pull/4427",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
668704909
|
[Doc] Add a note about sink-type and source-type value of the built-in connector.
Fixes #6109
Motivation
Tell user the available values of "--sink-type" "--source-type" parameter of the Cassandra built-in connector.
Modifications
Update the following docs:
Pulsar IO > get started
Connector Admin CLI
@tuteng , please help review the doc update.
@jiazhai @Jennifer88huang , cloud you please help review the PR? Thanks.
|
gharchive/pull-request
| 2020-07-30T13:08:56 |
2025-04-01T04:55:59.870466
|
{
"authors": [
"Huanli-Meng"
],
"repo": "apache/pulsar",
"url": "https://github.com/apache/pulsar/pull/7699",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
674878017
|
make pulsar executor pool size configurable
Motivation
The pulsar executor pool size number is hard code to 20 when pulsar service start, it should be configurable in broker.conf.
private final ScheduledExecutorService executor = Executors.newScheduledThreadPool(20,
new DefaultThreadFactory("pulsar"));
Changes
make the executor pool size configurable in broker.conf
As a followup to these changes in thread pool sizes, it would be nice to reduce the size of these pools when running the unit tests
As a followup to these changes in thread pool sizes, it would be nice to reduce the size of these pools when running the unit tests
Yes, for the unit tests framework, i will reduce the pool size.
As a followup to these changes in thread pool sizes, it would be nice to reduce the size of these pools when running the unit tests
Yes, for the unit tests framework, i will reduce the pool size.
@merlimat @jiazhai I have changed the default executor pool size to Runtime.getRuntime().availableProcessors() and update unit test executor pool size to 5, please take a look, thanks.
/pulsarbot run-failure-checks
|
gharchive/pull-request
| 2020-08-07T09:09:40 |
2025-04-01T04:55:59.873804
|
{
"authors": [
"hangc0276",
"merlimat"
],
"repo": "apache/pulsar",
"url": "https://github.com/apache/pulsar/pull/7782",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1205181157
|
can not run in Windows
when i run mvn clean package -Dmaven.test.skip=true
my os alert this software can not run on your platform
it seems that the version of yarn is incorretly!
Windows10
maven3.2.X
jdk 8
Could you provide the Error message?
By the way, If yarn doesn't work on your environment, you can remove the frontend-maven-plugin from pom.xml.
link
thank u for your reply
this is the screenshot of the error
could i use local yarn and node to build this project?
It seems node and yarn were installed succeed, but no permission when running yarn install, maybe we should fix the permission issue.
Another work around is that you can remove the frontend-maven-plugin from pom.xml as I mentioned before.
|
gharchive/issue
| 2022-04-15T02:15:08 |
2025-04-01T04:55:59.880222
|
{
"authors": [
"798284909",
"StyleTang"
],
"repo": "apache/rocketmq-dashboard",
"url": "https://github.com/apache/rocketmq-dashboard/issues/85",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1676766636
|
Build Fails on ARM MAC
When building on ARM MAC required to add -Dos.arch=x86_64 to bypass protobuff build issues
When this is passed the next issue is related to this https://github.com/apache/rocketmq-mqtt/issues/160 having PostConstruct deps are missing from POM.xml of related projects.
Any particular reason the javax.annotation deps are missing they are not part of ANY JDK it looks like these somehow on your build/classpath in project authors set-up and the required build instructions/set-up is then missing.
<dependency>
<groupId>javax.annotation</groupId>
<artifactId>javax.annotation-api</artifactId>
<version>1.3.2</version>
</dependency>
Also going back to item 2 pointing to issue is this project tied to https://github.com/apache/rocketmq-mqtt/issues/160#issuecomment-1329993294 RocketMQ 4.9.4 and must be build in Docker if yes any hint in setting this up is appreciated?
Thank you
The Java Common Annotations Module java.xml.ws.annotation was deprecated in Java version 9 and was removed in java version 11. If this leads to a problem you could try to add javax.annotation.
No these are two different deps javax.annotation is not the same as java.xml.ws.annotation my point is that your project is either missing proper build instructions in your readme. or pom.xml needs to have correct dependency added since other folks were able build this only while in they have javax.annotation deps on classpath https://github.com/apache/rocketmq-mqtt/issues/160#issuecomment-1329993294
|
gharchive/issue
| 2023-04-20T13:47:59 |
2025-04-01T04:55:59.885396
|
{
"authors": [
"meshpaul",
"pingww"
],
"repo": "apache/rocketmq-mqtt",
"url": "https://github.com/apache/rocketmq-mqtt/issues/193",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
642101872
|
LitePullConsumerSubscribe testcase not work [4.6.0]
BUG REPORT
Please describe the issue you observed:
I run the example below, but it doesn't work when the consumer group subscribe the topic the first time(I mean the consumer group have never consumed the topic). The consumer can not consume any message, include old messages and new messages.
testcase file:
https://github.com/apache/rocketmq/blob/rocketmq-all-4.6.0/example/src/main/java/org/apache/rocketmq/example/simple/LitePullConsumerSubscribe.java
If the consumer group have consumed some messages before , (may be consumed by DefaultPushConsumer), then the testcase works.
Please tell us about your environment:
rocketmq 4.6.0
Can you try if the same issue still exists in the latest version ?
4.7.0 still not work
|
gharchive/issue
| 2020-06-19T16:42:41 |
2025-04-01T04:55:59.888425
|
{
"authors": [
"RongtongJin",
"henrypoter"
],
"repo": "apache/rocketmq",
"url": "https://github.com/apache/rocketmq/issues/2110",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
351098438
|
为什么源代码都没加注释
至少关键部分应该加
@safph The issue tracker is only used for bug report and feature request, you can subscribe mailing list for help , community friends will answer your questions about the design or even the code.
Of course, we will also improve our documentation to help everyone become familiar with RocketMQ.
|
gharchive/issue
| 2018-08-16T08:06:25 |
2025-04-01T04:55:59.889799
|
{
"authors": [
"duhengforever",
"safph"
],
"repo": "apache/rocketmq",
"url": "https://github.com/apache/rocketmq/issues/416",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
473185502
|
One Swc depended on another Swc Causes Index.html Blank
I am compiling my own library DBIZLib.swc
Console : https://paste.apache.org/h4c5d
what should i do to remove these warnings?
because if i ignore these warnings and compile any project index.html is blank
Errors:
Uncaught ReferenceError: applications_mx_managers_SystemManager is not defined at index.html:962
Uncaught Error: Bad dependency path or symbol: org.apache.royale.effects.IEffect
at visit (base.js:2796)
at visit (base.js:2808)
at visit (base.js:2808)
at goog.DebugLoader_.load_ (base.js:2815)
at Object.goog.require (base.js:846)
at index.html:957
https://user-images.githubusercontent.com/42200979/61864583-1ab30c00-aeeb-11e9-9a8f-690798fec1cd.png
It looks like you're missing playerglobal.swc as an external dependency.
It would help to see what build configuration you are using.
Recently Josh made changes to how libraries are linked. I'm not using ANT, but I think ANT solves this with config.xml files in projects. Basically instead of listing all in library-path node, you must use "external-library-path" node in libraries to ensure all the code in the libraries linked are not compiled into that library. Then in Application, you must continue using library-path, since you want all code used be compiled in the application. In this part your application maybe is missing playerglobal or js swcs as Harbs comment.
|
gharchive/issue
| 2019-07-26T05:37:23 |
2025-04-01T04:55:59.893548
|
{
"authors": [
"Harbs",
"carlosrovira",
"pashminakazi"
],
"repo": "apache/royale-asjs",
"url": "https://github.com/apache/royale-asjs/issues/437",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1695527134
|
Add test cases for service.manager.impl
Make sure that:
[√ ] You have read the contribution guidelines.
[√ ] You submit test cases (unit or integration tests) that back your changes.
[ √] Your local test passed ./mvnw clean install -Dmaven.javadoc.skip=true.
hi, pls checked checkstyle and build in you local env.
so much checkstyle error
hi, pls checked checkstyle and build in you local env.
so much checkstyle error
thx,After this inspection, I learned a lot and found my own shortcomings. At present, I have made a modification
Codecov Report
Merging #4610 (e6a4129) into master (d6067f4) will increase coverage by 0.56%.
The diff coverage is 62.75%.
@@ Coverage Diff @@
## master #4610 +/- ##
============================================
+ Coverage 67.62% 68.18% +0.56%
- Complexity 7845 7919 +74
============================================
Files 1070 1078 +8
Lines 30650 30775 +125
Branches 2790 2807 +17
============================================
+ Hits 20726 20983 +257
+ Misses 8303 8139 -164
- Partials 1621 1653 +32
Impacted Files
Coverage Δ
...e/shenyu/plugin/base/fallback/FallbackHandler.java
36.84% <0.00%> (-17.01%)
:arrow_down:
.../apache/shenyu/plugin/hystrix/command/Command.java
7.69% <0.00%> (+2.69%)
:arrow_up:
.../shenyu/plugin/hystrix/command/HystrixCommand.java
55.55% <0.00%> (ø)
...plugin/hystrix/command/HystrixCommandOnThread.java
61.11% <0.00%> (ø)
.../springmvc/ShenyuSpringMvcClientConfiguration.java
100.00% <ø> (ø)
...henyuSpringMvcClientInfoRegisterConfiguration.java
0.00% <0.00%> (ø)
...pringboot/starter/gateway/ShenyuConfiguration.java
88.00% <0.00%> (-3.67%)
:arrow_down:
...shenyu/web/fallback/DefaultFallbackController.java
75.00% <0.00%> (-25.00%)
:arrow_down:
...pache/shenyu/web/filter/CollapseSlashesFilter.java
0.00% <0.00%> (ø)
...ava/org/apache/shenyu/web/filter/HealthFilter.java
62.50% <60.00%> (-37.50%)
:arrow_down:
... and 6 more
... and 7 files with indirect coverage changes
:mega: We’re building smart automated test selection to slash your CI/CD build times. Learn more
@lijay7674 hi, can you updata master to you dev,then summit pull request ?
not merged
sorry,Can I initiate a new pull request
@lijay7674 you can add my wechat: xixy199195, i wiil invite you join contributor group
|
gharchive/pull-request
| 2023-05-04T08:35:57 |
2025-04-01T04:55:59.956304
|
{
"authors": [
"codecov-commenter",
"lijay7674",
"yu199195"
],
"repo": "apache/shenyu",
"url": "https://github.com/apache/shenyu/pull/4610",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
705231042
|
login interface
Please answer these questions before submitting your issue.
Why do you submit this issue?
Question
Question
What do you want to know?
Hi Mr.Wu , I was wondering if skywalking8.1.0 has a login interface . If there is , then which config file is to set up the user and password ? Thanks.
Requirement or improvement
I just upgrade the skw 6.1.0 to 8.1.0 and I can not find my log in interface . Neither could I set up in webapp.yml
No, there isn't. The old login is not real.
Use APISIX, Nginx or other proxy tech to set up a login page.
roger that
|
gharchive/issue
| 2020-09-21T01:31:15 |
2025-04-01T04:55:59.959698
|
{
"authors": [
"sodared",
"wu-sheng"
],
"repo": "apache/skywalking",
"url": "https://github.com/apache/skywalking/issues/5525",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
936986406
|
持久化到es后,造成es占用内存持续增加
jdk11环境,docker部署es7,使用SkyWalking8.6持久化到es后,会让es的内存逐渐增大,直到100%造成OOM,请问如何解决?
持久化到mysql8,也会使mysql8占用内存持续增长到100%造成oom
Github discussion is required to be English for global users.
Your description is totally wrong. We push data into a storage, then database fails, try to make the database more powerful.
|
gharchive/issue
| 2021-07-05T11:17:58 |
2025-04-01T04:55:59.961470
|
{
"authors": [
"ZaneGeng",
"wu-sheng"
],
"repo": "apache/skywalking",
"url": "https://github.com/apache/skywalking/issues/7246",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
512309027
|
[CI] Run build jobs on different nodes parallelly
In this patch, I've made it to run multiple jobs parallelly on different Jenkins slave nodes, thus, our build jobs can be distributed more evenly to our VM nodes, and the VM with lower resources can also be scheduled fairly.
In this PR, I also rewrote the Jenkisfile with scripted Pipeline syntax, our previous syntax is DSL, it's much more human readable, but lack of flexibility in our complex scenarios, now we're using scripted syntax, based on Groovy, it provides a programmable ability, and flexibility, to simplify the Jenkinsfile, (from ~200 lines to ~100 lines)
We don't need to group the jobs anymore, because the jobs are scheduled on different nodes in a smaller granularity (test case), meaning that every test case will take an executor, if there is no more available executor, the test case should be in queue automatically
Please move this as a record in WIKI. We may need this someday.
|
gharchive/pull-request
| 2019-10-25T05:22:26 |
2025-04-01T04:55:59.963452
|
{
"authors": [
"kezhenxu94",
"wu-sheng"
],
"repo": "apache/skywalking",
"url": "https://github.com/apache/skywalking/pull/3714",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
948222777
|
doc: Add Self Observability service discovery (k8s)
[X] If this pull request closes/resolves/fixes an existing issue, replace the issue number. Closes #7259.
[X] Update the CHANGES log.
And for docs updates, I think it's better to loop @wingwong-knh in to review when the docs are updated in PR rather than review the whole docs after they're merged
I am not sure whether @wingwong-knh is available. Wing, if you are, please take a look.
And for docs updates, I think it's better to loop @wingwong-knh in to review when the docs are updated in PR rather than review the whole docs after they're merged
I am not sure whether @wingwong-knh is available. Wing, if you are, please take a look.
Indeed, for now I'm not sure if I can keep up with the daily updates! In the future, once all docs have been gone through once, I can start reviewing the updates in PR. For now, I'm only halfway through the backend doc (will up my pace!) For this particular doc, I will need to review from scratch.
And for docs updates, I think it's better to loop @wingwong-knh in to review when the docs are updated in PR rather than review the whole docs after they're merged
I am not sure whether @wingwong-knh is available. Wing, if you are, please take a look.
Indeed, for now I'm not sure if I can keep up with the daily updates! In the future, once all docs have been gone through once, I can start reviewing the updates in PR. For now, I'm only halfway through the backend doc (will up my pace!) For this particular doc, I will need to review from scratch.
Take your time. I know you are busy. We are going to continue, you could be back and polish the doc.
|
gharchive/pull-request
| 2021-07-20T03:53:31 |
2025-04-01T04:55:59.967855
|
{
"authors": [
"wankai123",
"wingwong-knh",
"wu-sheng"
],
"repo": "apache/skywalking",
"url": "https://github.com/apache/skywalking/pull/7339",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
130406687
|
[SPARK-10605][SQL] Add structs to collect_list/collect_set
We currently use the Hive implementations for the collect_list/collect_set aggregate functions. This has a few major drawbacks, the use of HiveUDAF (which has quite a bit of overhead) and the lack of support for struct datatypes. This PR adds native implementation of these functions to Spark.
The size of the collected list/set vary, this means we cannot use the fast, Tungsten, aggregation path to perform the aggregation, and that we fallback to the slower sort based path. Another big problem with these operators is that when the size of the collected list/set grows too large, we can start experiencing large GC pauzes and OOMEs.
This PR implements these UDAFs as ImperativeAggregates and uses a custom mutable ArrayData structure in order to efficiently collect the items. The current PR allows for partial aggregates, but I am not sure yet if we should support this.
Comments and feedback are appreciated.
cc @yhuai
Test build #50490 has started for PR 11004 at commit 8247d8e.
Test build #50490 has finished for PR 11004 at commit 8247d8e.
This patch fails Spark unit tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test FAILed.
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/50490/
Test FAILed.
Test build #50493 has started for PR 11004 at commit 326a213.
Test build #50493 has finished for PR 11004 at commit 326a213.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/50493/
Test PASSed.
Merged build finished. Test PASSed.
closing this one for a better approach.
|
gharchive/pull-request
| 2016-02-01T16:14:37 |
2025-04-01T04:55:59.977163
|
{
"authors": [
"AmplabJenkins",
"SparkQA",
"hvanhovell"
],
"repo": "apache/spark",
"url": "https://github.com/apache/spark/pull/11004",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
144458293
|
[SPARK-14262] correct appStatus after master leader changed
What changes were proposed in this pull request?
Suppose master leadership changes, even if recovery is completed, those privious apps in RUNNING state are now in WAITING state from master's WebUI, although they are really at work. Now correct the state.
How was this patch tested?
(Please explain how this patch was tested. E.g. unit tests, integration tests, manual tests)
Can one of the admins verify this patch?
I think I have a PR #10506 which already covered this issue.
Can one of the admins verify this patch?
|
gharchive/pull-request
| 2016-03-30T04:07:41 |
2025-04-01T04:55:59.979217
|
{
"authors": [
"AmplabJenkins",
"GavinGavinNo1",
"jerryshao"
],
"repo": "apache/spark",
"url": "https://github.com/apache/spark/pull/12054",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
153120354
|
[SPARK-11249][launcher] Throw error if app resource is not provided.
Without this, the code would build an invalid spark-submit command line,
and a more cryptic error would be presented to the user. Also, expose
a constant that allows users to set a dummy resource in cases where
they don't need an actual resource file; for backwards compatibility,
that uses the same "spark-internal" resource that Spark itself uses.
Tested via unit tests, run-example, spark-shell, and running the
thrift server with mixed spark and hive command line arguments.
Test build #57807 has started for PR 12909 at commit 0ab4a28.
Test build #57807 has finished for PR 12909 at commit 0ab4a28.
This patch fails from timeout after a configured wait of `250m`.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test FAILed.
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/57807/
Test FAILed.
retest this please
Test build #57900 has started for PR 12909 at commit 0ab4a28.
Test build #57900 has finished for PR 12909 at commit 0ab4a28.
This patch fails from timeout after a configured wait of `250m`.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test FAILed.
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/57900/
Test FAILed.
Hmm, pyspark seems to be getting stuck, let me see if I broke it...
Test build #58030 has started for PR 12909 at commit 46664ba.
Test build #58030 has finished for PR 12909 at commit 46664ba.
This patch fails Spark unit tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test FAILed.
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/58030/
Test FAILed.
retest this please
Test build #58039 has started for PR 12909 at commit 46664ba.
Test build #58039 has finished for PR 12909 at commit 46664ba.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/58039/
Test PASSed.
@andrewor14
Merging to master / 2.0.
|
gharchive/pull-request
| 2016-05-04T21:33:33 |
2025-04-01T04:55:59.996684
|
{
"authors": [
"AmplabJenkins",
"SparkQA",
"vanzin"
],
"repo": "apache/spark",
"url": "https://github.com/apache/spark/pull/12909",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
176435374
|
SPARK-17424: Fix unsound substitution bug in ScalaReflection.
What changes were proposed in this pull request?
This method gets a type's primary constructor and fills in type parameters with concrete types. For example, MapPartitions[T, U] -> MapPartitions[Int, String]. This Substitution fails when the actual type args are empty because they are still unknown. Instead, when there are no resolved types to subsitute, this returns the original args with unresolved type parameters.
How was this patch tested?
This doesn't affect substitutions where the type args are determined. This fixes our case where the actual type args are empty and our job runs successfully.
Test build #65269 has started for PR 15062 at commit 931f156.
Test build #65269 has finished for PR 15062 at commit 931f156.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/65269/
Test PASSed.
Could you add a regression test?
@hvanhovell, I'm not sure what's causing this to build the test. I can step through with a debugger to see that the problem is caused by an empty actualTypeArgs list, but I don't know why it is that way. Maybe someone with more knowledge of Scala types can help fill in the details. Until then, I'm not sure how to construct a test case for this.
I don't quite understand what use-case this patch fixes. Can you provide an example (in the form of a test) that reproduces the issue?
Oh, I was too quick to comment, I see a complete description is in the JIRA.
It would still be good if you could add a test though.
@jodersky, as I said above, I'm not really sure how to build a test for this because I'm not too familiar with the Scala internals that are misused here. However, since I can confirm that it works in practice, I think it's reasonable to commit it anyway, before committers that know this code well can have a look to make sure this method is tested properly.
Another way to go is to revert PR #10970. That is a work-around because another test was already hitting this problem and the solution was to avoid the parameter substitution.
Jenkins retest this please
Test build #66190 has started for PR 15062 at commit 931f156.
I'm just unclear of why this is needed . According to the linked issue, some tests were failing and this fixed them. Everything just seems very vague to me. It would be very helpful if you could provide an example (it doesn't have to be a unit-test) that reproduces the error.
It would greatly help me understand what's going on and also distill things down to a regression test.
I think the best way to reproduce this is to revert #10970. I don't have a case that reproduces it that I can share since this was in a fairly large job.
Test build #66190 has finished for PR 15062 at commit 931f156.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/66190/
Test PASSed.
Hi @rdblue can you revert #10970 in this PR? Then other reviewers can test this PR locally and verify if it works.
I tried to reproduce with #10970 reverted, but I didn't hit the issue in testing. I still think it's fine to move forward on this, even if it is hard to reproduce because we know the code is wrong and this fixes it.
retest this please
I'm fine with this change, cc @yhuai @liancheng
Test build #68939 has started for PR 15062 at commit 931f156.
Test build #68939 has finished for PR 15062 at commit 931f156.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/68939/
Test PASSed.
any chance that is the same issue as https://issues.apache.org/jira/browse/SPARK-17109?
@rdblue When you were debugging this issue, which version of scala did you use? Scala 2.10 or Scala 2.11? If you were using scala 2.10, is it possible to try scala 2.11? Thanks!
We were seeing the problem when using the datasets API in our 1.6.1 build, which is based on Scala 2.10. I recently tried to reproduce this on master with 2.11 and #10970 reverted, but I didn't get a case that failed. Either way, I think the fix here makes sense: if there are no types to substitute, don't do it.
This does look the same as SPARK-17109. Does this fix that issue?
retest this please
Test build #76863 has started for PR 15062 at commit 931f156.
Test build #76863 has finished for PR 15062 at commit 931f156.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/76863/
Test PASSed.
Merged build finished. Test PASSed.
the change looks reasonable, merging to master/2.2/2.1/2.0!
|
gharchive/pull-request
| 2016-09-12T17:30:46 |
2025-04-01T04:56:00.019501
|
{
"authors": [
"AmplabJenkins",
"SparkQA",
"cloud-fan",
"hvanhovell",
"jodersky",
"rdblue",
"srowen",
"yhuai"
],
"repo": "apache/spark",
"url": "https://github.com/apache/spark/pull/15062",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
179575985
|
[SPARK-17618] Guard against invalid comparisons between UnsafeRow and other formats
This patch ports changes from #15185 to Spark 2.x. In that patch, a correctness bug in Spark 1.6.x which was caused by an invalid equals() comparison between an UnsafeRow and another row of a different format. Spark 2.x is not affected by that specific correctness bug but it can still reap the error-prevention benefits of that patch's changes, which modify ``UnsafeRow.equals()to throw an IllegalArgumentException if it is called with an object that is not anUnsafeRow`.
Test build #65991 has started for PR 15265 at commit f4de93f.
LGTM - pending jenkins.
Test build #65991 has finished for PR 15265 at commit f4de93f.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/65991/
Test PASSed.
Merging to master and branch-2.0. Thanks!
|
gharchive/pull-request
| 2016-09-27T18:32:59 |
2025-04-01T04:56:00.025315
|
{
"authors": [
"AmplabJenkins",
"JoshRosen",
"SparkQA",
"hvanhovell"
],
"repo": "apache/spark",
"url": "https://github.com/apache/spark/pull/15265",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
200965481
|
[SPARK-19182][DStream] Optimize the lock in StreamingJobProgressListener to not block UI when generating Streaming jobs
What changes were proposed in this pull request?
When DStreamGraph is generating a job, it will hold a lock and block other APIs. Because StreamingJobProgressListener (numInactiveReceivers, streamName(streamId: Int), streamIds) needs to call DStreamGraph's methods to access some information, the UI may hang if generating a job is very slow (e.g., talking to the slow Kafka cluster to fetch metadata).
It's better to optimize the locks in DStreamGraph and StreamingJobProgressListener to make the UI not block by job generation.
How was this patch tested?
existing ut
cc @zsxwing
Test build #71441 has started for PR 16601 at commit 46036bf.
Test build #71441 has finished for PR 16601 at commit 46036bf.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/71441/
Test PASSed.
also cc @tdas
Test build #71492 has started for PR 16601 at commit eaa7b15.
Test build #71492 has finished for PR 16601 at commit eaa7b15.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/71492/
Test PASSed.
@zsxwing Take a review please!
Test build #71553 has started for PR 16601 at commit e51623c.
Test build #71553 has finished for PR 16601 at commit e51623c.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/71553/
Test PASSed.
LGTM. Merging to master. Thanks!
|
gharchive/pull-request
| 2017-01-16T09:17:13 |
2025-04-01T04:56:00.038272
|
{
"authors": [
"AmplabJenkins",
"SparkQA",
"uncleGen",
"zsxwing"
],
"repo": "apache/spark",
"url": "https://github.com/apache/spark/pull/16601",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
201042342
|
[SPARK-19246][SQL]CataLogTable's partitionSchema order and exist check
What changes were proposed in this pull request?
CataLogTable's partitionSchema should check if each column name in partitionColumnNames must match one and only one field in schema, if not we should throw an exception
and CataLogTable's partitionSchema should keep order with partitionColumnNames
How was this patch tested?
Test build #71451 has started for PR 16606 at commit eaf18ce.
Test build #71451 has finished for PR 16606 at commit eaf18ce.
This patch fails Scala style tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test FAILed.
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/71451/
Test FAILed.
Test build #71453 has started for PR 16606 at commit 9296624.
Test build #71453 has finished for PR 16606 at commit 9296624.
This patch fails Spark unit tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test FAILed.
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/71453/
Test FAILed.
Test build #71483 has started for PR 16606 at commit 4260f84.
Merged build finished. Test FAILed.
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/71483/
Test FAILed.
retest this please
Test build #71494 has started for PR 16606 at commit 4260f84.
Test build #71497 has started for PR 16606 at commit 6f2816e.
Test build #71494 has finished for PR 16606 at commit 4260f84.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/71494/
Test PASSed.
Test build #71497 has finished for PR 16606 at commit 6f2816e.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/71497/
Test PASSed.
Test build #71519 has started for PR 16606 at commit c08e1c9.
Test build #71519 has finished for PR 16606 at commit c08e1c9.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/71519/
Test PASSed.
Test build #71552 has started for PR 16606 at commit 8cbee32.
Test build #71552 has finished for PR 16606 at commit 8cbee32.
This patch fails Scala style tests.
This patch merges cleanly.
This patch adds no public classes.
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/71552/
Test FAILed.
Merged build finished. Test FAILed.
Test build #71554 has started for PR 16606 at commit 79e2e3f.
Test build #71554 has finished for PR 16606 at commit 79e2e3f.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/71554/
Test PASSed.
Test build #71580 has started for PR 16606 at commit 5e60f14.
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/71580/
Test FAILed.
Merged build finished. Test FAILed.
Test build #71853 has started for PR 16606 at commit 206b232.
how about we just add an assert? assert(schema.takeRight(partitionColumnNames.length) == partitionColumnNames)
Test build #71853 has finished for PR 16606 at commit 206b232.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/71853/
Test PASSed.
Merged build finished. Test PASSed.
Test build #71889 has started for PR 16606 at commit 7e30cc7.
Test build #71889 has finished for PR 16606 at commit 7e30cc7.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/71889/
Test PASSed.
Test build #71902 has started for PR 16606 at commit 04d3940.
LGTM pending test
Test build #71902 has finished for PR 16606 at commit 04d3940.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/71902/
Test PASSed.
Test build #71912 has started for PR 16606 at commit 72164eb.
LGTM, pending tests
Merged build finished. Test FAILed.
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/71912/
Test FAILed.
retest this please
Test build #71922 has started for PR 16606 at commit 72164eb.
Test build #71922 has finished for PR 16606 at commit 72164eb.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/71922/
Test PASSed.
thanks, merging to master!
|
gharchive/pull-request
| 2017-01-16T15:05:15 |
2025-04-01T04:56:00.082541
|
{
"authors": [
"AmplabJenkins",
"SparkQA",
"cloud-fan",
"gatorsmile",
"windpiger"
],
"repo": "apache/spark",
"url": "https://github.com/apache/spark/pull/16606",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
225391068
|
[SPARK-20543][SPARKR][WIP] skip tests when running on CRAN
What changes were proposed in this pull request?
General rule on skip or not:
skip if
RDD tests
tests could run long or complicated (streaming, hivecontext)
tests on error conditions
tests won't likely change/break
How was this patch tested?
unit tests, R CMD check --as-cran, R CMD check
Test build #76342 has started for PR 17817 at commit fc4965a.
Test build #76342 has finished for PR 17817 at commit fc4965a.
This patch fails R style tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test FAILed.
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/76342/
Test FAILed.
Test build #76343 has started for PR 17817 at commit 2c5e267.
Test build #76343 has finished for PR 17817 at commit 2c5e267.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/76343/
Test PASSed.
Merged build finished. Test PASSed.
Test build #76346 has started for PR 17817 at commit bd3295a.
Test build #76346 has finished for PR 17817 at commit bd3295a.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/76346/
Test PASSed.
Skipping 153 tests ON CRAN
Jenkins, all tests: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/76346/console
CRAN: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/76343/console
merged to master/2.2
|
gharchive/pull-request
| 2017-05-01T06:06:41 |
2025-04-01T04:56:00.096137
|
{
"authors": [
"AmplabJenkins",
"SparkQA",
"felixcheung"
],
"repo": "apache/spark",
"url": "https://github.com/apache/spark/pull/17817",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
243562870
|
[SPARK-21444] Be more defensive when removing broadcasts in MapOutputTracker
What changes were proposed in this pull request?
In SPARK-21444, @sitalkedia reported an issue where the Broadcast.destroy() call in MapOutputTracker's ShuffleStatus.invalidateSerializedMapOutputStatusCache() was failing with an IOException, causing the DAGScheduler to crash and bring down the entire driver.
This is a bug introduced by #17955. In the old code, we removed a broadcast variable by calling BroadcastManager.unbroadcast with blocking=false, but the new code simply calls Broadcast.destroy() which is capable of failing with an IOException in case certain blocking RPCs time out.
The fix implemented here is to replace this with a call to destroy(blocking = false) and to wrap the entire operation in Utils.tryLogNonFatalError.
How was this patch tested?
I haven't written regression tests for this because it's really hard to inject mocks to simulate RPC failures here. Instead, this class of issue is probably best uncovered with more generalized error injection / network unreliability / fuzz testing tools.
Test build #79691 has started for PR 18662 at commit a5ebcac.
cc @jiangxb1987
@JoshRosen - Thanks for the lightning fast response. The change looks reasonable to me.
LGTM
Test build #79691 has finished for PR 18662 at commit a5ebcac.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/79691/
Test PASSed.
Merged build finished. Test PASSed.
Merged to master. Thanks for the quick reviews.
|
gharchive/pull-request
| 2017-07-18T00:05:29 |
2025-04-01T04:56:00.102827
|
{
"authors": [
"AmplabJenkins",
"JoshRosen",
"SparkQA",
"cloud-fan",
"sameeragarwal",
"sitalkedia"
],
"repo": "apache/spark",
"url": "https://github.com/apache/spark/pull/18662",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
259698011
|
[WIP][SPARK-22096][ML] use aggregateByKeyLocally in feature frequency calc…
What changes were proposed in this pull request?
NaiveBayes currently takes aggreateByKey followed by a collect to calculate frequency for each feature/label. We can implement a new function 'aggregateByKeyLocally' in RDD that merges locally on each mapper before sending results to a reducer to save one stage.
We tested on NaiveBayes and see ~16% performance gain on training with these changes.
Signed-off-by: Vincent Xie vincent.xie@intel.com
How was this patch tested?
existing test
Test build #82059 has started for PR 19318 at commit efb0fe9.
Test build #82059 has finished for PR 19318 at commit efb0fe9.
This patch fails to build.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test FAILed.
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/82059/
Test FAILed.
The aggregateByKeyLocally haven't been implemented. Please wait until #19317 finished or merge that PR contents into this, otherwise this PR cannot use. We should not open a PR which cannot build.
@VinceShieh can you please mark this PR's title as "[WIP]"?
thanks :)
Build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/testing-k8s-prb-make-spark-distribution-unified/4201/
Test PASSed.
Test build #97740 has started for PR 19318 at commit efb0fe9.
Build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/testing-k8s-prb-make-spark-distribution-unified/4223/
Test PASSed.
Test build #97740 has finished for PR 19318 at commit efb0fe9.
This patch fails due to an unknown error code, -9.
This patch does not merge cleanly.
This patch adds no public classes.
Build finished. Test FAILed.
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/97740/
Test FAILed.
Build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/testing-k8s-prb-make-spark-distribution-unified/4239/
Test PASSed.
Test build #97752 has started for PR 19318 at commit efb0fe9.
Test build #97752 has finished for PR 19318 at commit efb0fe9.
This patch fails to build.
This patch does not merge cleanly.
This patch adds no public classes.
Build finished. Test FAILed.
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/97752/
Test FAILed.
Test build #97768 has started for PR 19318 at commit efb0fe9.
Test build #97768 has finished for PR 19318 at commit efb0fe9.
This patch fails to build.
This patch does not merge cleanly.
This patch adds no public classes.
Build finished. Test FAILed.
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/97768/
Test FAILed.
Test build #97818 has started for PR 19318 at commit efb0fe9.
Build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/testing-k8s-prb-make-spark-distribution-unified/4273/
Test PASSed.
Test build #97818 has finished for PR 19318 at commit efb0fe9.
This patch fails build dependency tests.
This patch does not merge cleanly.
This patch adds no public classes.
Build finished. Test FAILed.
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/97818/
Test FAILed.
@VinceShieh if you'd rebase this, we can consider it now that aggregateByKeyLocally is implemented
OK. Thanks. @srowen
it seems 'aggregateByKeyLocally' is not implemented
Oops yeah https://github.com/apache/spark/pull/19317 was not merged. I think this can't continue.
|
gharchive/pull-request
| 2017-09-22T04:11:53 |
2025-04-01T04:56:00.128724
|
{
"authors": [
"AmplabJenkins",
"SparkQA",
"VinceShieh",
"WeichenXu123",
"chenghao-intel",
"srowen"
],
"repo": "apache/spark",
"url": "https://github.com/apache/spark/pull/19318",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
349582280
|
[SPARK-25088][CORE][MESOS][DOCS] Update Rest Server docs & defaults.
What changes were proposed in this pull request?
(a) disabled rest submission server by default in standalone mode
(b) fails the standalone master if rest server enabled & authentication secret set
(c) fails the mesos cluster dispatcher if authentication secret set
(d) doc updates
(e) when submitting a standalone app, only try the rest submission first if spark.master.rest.enabled=true
otherwise you'd see a 10 second pause like
18/08/09 08:13:22 INFO RestSubmissionClient: Submitting a request to launch an application in spark://...
18/08/09 08:13:33 WARN RestSubmissionClient: Unable to connect to server spark://...
I also made sure the mesos cluster dispatcher failed with the secret enabled, though I had to do that on slightly different code as I don't have mesos native libs around.
How was this patch tested?
I ran the tests in the mesos module & in core for org.apache.spark.deploy.*
I ran a test on a cluster with standalone master to make sure I could still start with the right configs, and would fail the right way too.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/testing-k8s-prb-make-spark-distribution-unified/2049/
Test PASSed.
Test build #94572 has started for PR 22071 at commit b4ca224.
Test build #94572 has finished for PR 22071 at commit b4ca224.
This patch fails Spark unit tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test FAILed.
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/94572/
Test FAILed.
Test build #4241 has started for PR 22071 at commit b4ca224.
@tnachen
Test build #4244 has started for PR 22071 at commit b4ca224.
Test build #4244 has finished for PR 22071 at commit b4ca224.
This patch fails Spark unit tests.
This patch does not merge cleanly.
This patch adds no public classes.
Test build #4245 has started for PR 22071 at commit b4ca224.
Test build #4245 has finished for PR 22071 at commit b4ca224.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Test build #94706 has started for PR 22071 at commit 897b587.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/testing-k8s-prb-make-spark-distribution-unified/2149/
Test PASSed.
Test build #94706 has finished for PR 22071 at commit 897b587.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/94706/
Test PASSed.
LGTM as well
+1
Merged to master
any objections about putting this in prior branches as well?
in this case maybe ok. perhaps just rel note this iff there's another 2.2.x or 2.1.x releases?
Hi @squito, if I may, I have a few comments to share regarding this change and probably you can help me to understand it in better details.
From Spark Security Doc page it looks like spark.master.rest.enabled property belongs to Spark Standalone Master only. IMO, It makes sense, because it's the server side who is responsible for enabling and disabling REST endpoint according to the source code.
Now, from the [comment in the code] (https://github.com/apache/spark/blob/v2.4.0/core/src/main/scala/org/apache/spark/deploy/SparkSubmit.scala#L176-L180) of SparkSubmit.scala it looks like REST-based gateway introduced in Spark 1.3 is the default way of communication with Spark Standalone Master with a fallback to legacy RPC.
And here's the problem: now spark.master.rest.enabled became a client-side property, not Spark Standalone Master property. So in order to communicate with Master, a user have to provide spark.master.rest.enabled with each job she submits because without it, the submission will fail. This is a pretty breaking change.
What is the reason to fail the start of MesosClusterDispatcher if spark.authenticate.secret is set? Mesos uses different mechanisms for authentication and there's no need to fail the Dispatcher - this property can be just ignored or logged at most.
TL;DR It looks like spark.master.rest.enabled should reside only in Spark Standalone Master code and shouldn't leak into the client side of SparkSubmit. Also, it can be completely ignored with Mesos Dispatcher, because REST interface is the only endpoint it exposes.
Please let me know what you think. Thanks.
|
gharchive/pull-request
| 2018-08-10T16:35:57 |
2025-04-01T04:56:00.151144
|
{
"authors": [
"AmplabJenkins",
"SparkQA",
"akirillov",
"felixcheung",
"squito",
"srowen",
"tgravescs",
"tnachen"
],
"repo": "apache/spark",
"url": "https://github.com/apache/spark/pull/22071",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
551525835
|
[SPARK-22590][SQL] Copy sparkContext.localproperties to child thread in BroadcastExchangeExec.executionContext
What changes were proposed in this pull request?
In org.apache.spark.sql.execution.exchange.BroadcastExchangeExec#relationFuture make a copy of org.apache.spark.SparkContext#localProperties and pass it to the broadcast execution thread in org.apache.spark.sql.execution.exchange.BroadcastExchangeExec#executionContext
Why are the changes needed?
When executing BroadcastExchangeExec, the relationFuture is evaluated via a separate thread. The threads inherit the localProperties from sparkContext as they are the child threads.
These threads are created in the executionContext (thread pools). Each Thread pool has a default keepAliveSeconds of 60 seconds for idle threads.
Scenarios where the thread pool has threads which are idle and reused for a subsequent new query, the thread local properties will not be inherited from spark context (thread properties are inherited only on thread creation) hence end up having old or no properties set. This will cause taskset properties to be missing when properties are transferred by child thread via sparkContext.runJob/submitJob
Does this PR introduce any user-facing change?
No
How was this patch tested?
Added UT
@ajithme are you still working on it? I'd like to get this in before 3.0 release and take over this one if you don't have time to complete it.
@ajithme are you still working on it? I'd like to get this in before 3.0 release and take over this one if you don't have time to complete it.
Yes. I will update the PR shortly. Sorry for delay
Test build #118581 has started for PR 27266 at commit 0f39043.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/23335/
Test PASSed.
@cloud-fan @hvanhovell @HyukjinKwon Updated as per all the comments, please review
Test build #118582 has started for PR 27266 at commit 742d322.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/23336/
Test PASSed.
Test build #118583 has started for PR 27266 at commit 11fffca.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/23338/
Test PASSed.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/23339/
Test PASSed.
Test build #118585 has started for PR 27266 at commit 763d1bc.
Test build #118586 has started for PR 27266 at commit 2ed76c3.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/23340/
Test PASSed.
Test build #118583 has finished for PR 27266 at commit 11fffca.
This patch fails Spark unit tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test FAILed.
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/118583/
Test FAILed.
Test build #118581 has finished for PR 27266 at commit 0f39043.
This patch fails Spark unit tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test FAILed.
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/118581/
Test FAILed.
Test build #118585 has finished for PR 27266 at commit 763d1bc.
This patch fails Spark unit tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test FAILed.
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/118585/
Test FAILed.
Test build #118586 has finished for PR 27266 at commit 2ed76c3.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/118586/
Test PASSed.
thanks, merging to master/3.0!
Test build #118582 has finished for PR 27266 at commit 742d322.
This patch fails from timeout after a configured wait of 400m.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test FAILed.
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/118582/
Test FAILed.
+1 from me too. Thanks @ajithme.
|
gharchive/pull-request
| 2020-01-17T16:55:25 |
2025-04-01T04:56:00.177998
|
{
"authors": [
"AmplabJenkins",
"HyukjinKwon",
"SparkQA",
"ajithme",
"cloud-fan"
],
"repo": "apache/spark",
"url": "https://github.com/apache/spark/pull/27266",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
610022891
|
[SPARK-24266][k8s] Restart the watcher when we receive a version changed from k8s
What changes were proposed in this pull request?
Restart the watcher when it failed with a HTTP_GONE code from the kubernetes api. Which means a resource version has changed.
For more relevant information see here: https://github.com/fabric8io/kubernetes-client/issues/1075
Why are the changes needed?
Does this PR introduce any user-facing change?
No
How was this patch tested?
Running spark-submit to a k8s cluster.
Not sure how to make an automated test for this. If someone can help me out that would be great.
Can one of the admins verify this patch?
Can one of the admins verify this patch?
ok to test
Thank you for your contribution, @stijndehaes .
Test build #122141 has started for PR 28423 at commit a9ce548.
Test build #122141 has finished for PR 28423 at commit a9ce548.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/122141/
Test PASSed.
Kubernetes integration test starting
URL: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/26813/
Kubernetes integration test status success
URL: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/26813/
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/26813/
Test PASSed.
How do we feel about backporting this to Spark 2.4.6?
How do we feel about backporting this to Spark 2.4.6?
I would very much like that, we ran into this using spark 2.4.x.
Do you think we can have a unit test case for this, @stijndehaes ?
Test build #122248 has started for PR 28423 at commit 2a4cbb6.
Test build #122248 has finished for PR 28423 at commit 2a4cbb6.
This patch fails Scala style tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test FAILed.
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/122248/
Test FAILed.
Test build #122249 has started for PR 28423 at commit f832acf.
Test build #122249 has finished for PR 28423 at commit f832acf.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/122249/
Test PASSed.
Kubernetes integration test starting
URL: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/26919/
Kubernetes integration test starting
URL: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/26920/
Kubernetes integration test status success
URL: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/26919/
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/26919/
Test PASSed.
Kubernetes integration test status success
URL: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/26920/
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/26920/
Test PASSed.
Do you think we can have a unit test case for this, @stijndehaes ?
The current tests completely mock out this behavior, see org.apache.spark.deploy.k8s.submit.ClientSuite, I think writing a test for this where we manipulate and fake the HTTP gone with a mock is not that useful. Maybe I can look into an integration test, but then I have to be able to trigger that a resource version changes. Not sure if I will be able to
@holdenk Maybe we should refactor this behavior using the sharedinformers.
See the comment made here: https://github.com/fabric8io/kubernetes-client/issues/1075#issuecomment-623354623
I can make an example implementation of this, maybe best to do that in another PR. What do you think?
Ok I have tested this in production, there is something wrong with the code, went ahead and tried the sharedinformers approach. Will try that in production today. You can see the code here: https://github.com/stijndehaes/spark/tree/test/shared-informers
Ok reverting back to the old approach found the missing piece I think testing that out.
Shared informers have the problem that you have to watch every pod in the whole cluster atm.
Test build #122321 has started for PR 28423 at commit dad7ea2.
Test build #122321 has finished for PR 28423 at commit dad7ea2.
This patch fails Spark unit tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test FAILed.
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/122321/
Test FAILed.
Test build #122322 has started for PR 28423 at commit f05db8f.
Test build #122322 has finished for PR 28423 at commit f05db8f.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/122322/
Test PASSed.
Kubernetes integration test starting
URL: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/26991/
Kubernetes integration test status failure
URL: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/26991/
Merged build finished. Test FAILed.
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/26991/
Test FAILed.
Kubernetes integration test starting
URL: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/26992/
Kubernetes integration test status success
URL: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/26992/
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/26992/
Test PASSed.
@holdenk @dongjoon-hyun I have tested this code in production and it works. I have a couple of jobs that take roughly 4 hours to finish, these all failed without the fix and are now succeeding.
Could you take the time to review the code again?
+1 for this. Hit this in GKE today.
Retest this please.
Hi, @stijndehaes . Could you describe how to manually verify this PR?
Test build #124237 has started for PR 28423 at commit f05db8f.
Merged build finished. Test FAILed.
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/28856/
Test FAILed.
Test build #124237 has finished for PR 28423 at commit f05db8f.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/124237/
Test PASSed.
Test build #124283 has started for PR 28423 at commit 6ef9e12.
Test build #124283 has finished for PR 28423 at commit 6ef9e12.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/124283/
Test PASSed.
Kubernetes integration test starting
URL: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/28902/
Kubernetes integration test status success
URL: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/28902/
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/28902/
Test PASSed.
Test build #124323 has started for PR 28423 at commit 581b6f9.
Test build #124323 has finished for PR 28423 at commit 581b6f9.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/124323/
Test PASSed.
Kubernetes integration test starting
URL: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/28942/
Kubernetes integration test status failure
URL: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/28942/
Merged build finished. Test FAILed.
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/28942/
Test FAILed.
Test build #124636 has started for PR 28423 at commit 9da57c9.
Test build #124636 has finished for PR 28423 at commit 9da57c9.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test FAILed.
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/29254/
Test FAILed.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/124636/
Test PASSed.
Retest this please
Test build #125529 has started for PR 28423 at commit 9da57c9.
Test build #125529 has finished for PR 28423 at commit 9da57c9.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/125529/
Test PASSed.
Merged build finished. Test FAILed.
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/30144/
Test FAILed.
Hi Folks @dongjoon-hyun and @holdenk, do we plan to include this patch in 2.4.7 release?
@ScrapCodes the code in the 2.4.x is significantly different from the code here. But we can reuse the same idea as here. I guess it has to be a new PR
BTW, when do we receive a version changed from K8s?
Test build #125828 has started for PR 28423 at commit 51dc743.
@ScrapCodes the code in the 2.4.x is significantly different from the code here. But we can reuse the same idea as here. I guess it has to be a new PR
Currently the JIRA has target version set as "2.4.7", so I wanted to clarify.
BTW, when do we receive a version changed from K8s?
It happens when etcd compaction kicks in for example. On aws EKS I never saw this happening on EKS 1.14, but it happens roughly every hour on EKS 1.15. Meaning that longer spark application continued working but the watch failed.
Test build #125828 has finished for PR 28423 at commit 51dc743.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/125828/
Test PASSed.
Merged build finished. Test FAILed.
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/30439/
Test FAILed.
BTW, when do we receive a version changed from K8s?
It happens when etcd compaction kicks in for example. On aws EKS I never saw this happening on EKS 1.14, but it happens roughly every hour on EKS 1.15. Meaning that longer spark application continued working but the watch failed.
Okay makes sense, thank you for explaining. I am wondering how useful is the watch for longer running jobs, for example all it reports is that jobs is running, completed or errored. Probably, spark ui is more informative.
BTW, when do we receive a version changed from K8s?
It happens when etcd compaction kicks in for example. On aws EKS I never saw this happening on EKS 1.14, but it happens roughly every hour on EKS 1.15. Meaning that longer spark application continued working but the watch failed.
Okay makes sense, thank you for explaining. I am wondering how useful is the watch for longer running jobs, for example all it reports is that jobs is running, completed or errored. Probably, spark ui is more informative.
Our use case is that airflow does a spark submit. And thus airflow follows it up that way :)
Will there be a backport of this to branch-2.4?
test this please
(btw i'm testing the freshly upgraded R installation on the minikube workers)
Test build #126271 has started for PR 28423 at commit 51dc743.
Test build #126271 has finished for PR 28423 at commit 51dc743.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/126271/
Test PASSed.
Kubernetes integration test starting
URL: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/30887/
Kubernetes integration test status success
URL: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/30887/
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/30887/
Test PASSed.
Lets focus on 3.1 and then explore backporting after.
LGTM pending Jenkins
Oh wait it has passed Jenkins, excellent. If @dongjoon-hyun is ok with this PR I'll merge it by the end of the week.
@holdenk . Thank you for pinging me. Feel free to merge if you think it's okay. I don't want to be a blocker for the community PR. ;)
Whats your JIRA username @stijndehaes ?
@holdenk my JIRA username if sdehaes
I took the commits from master and made a partial attempt to rebase this onto branch-2.4 [1].
However, the k8s api has evolved from 2.4 quite a bit so the watchOrStop function needs to be backported [2].
You can see the error message in this gitlab build [3].
Would it be useful to make a WIP pull request from [1] ?
[1] https://github.com/jkleckner/spark/tree/SPARK-24266-on-branch2.4
[2] https://github.com/jkleckner/spark/blob/SPARK-24266-on-branch2.4/resource-managers/kubernetes/core/src/main/scala/org/apache/spark/deploy/k8s/submit/LoggingPodStatusWatcher.scala#L193
[3] https://gitlab.com/jkleckner/spark/-/jobs/651515950
@stijndehaes In private discussions about the hang we are seeing, there appears to be another watcher [1] for the driver watching executors that also may lose notifications.
Have you run into any situations like this?
[1] https://github.com/apache/spark/blob/master/resource-managers/kubernetes/core/src/main/scala/org/apache/spark/scheduler/cluster/k8s/ExecutorPodsWatchSnapshotSource.scala
@jkleckner I have never had a problem with the driver watching the executors. I think there was already a fallback mechanism there, but I never looked into the code for that one.
@liyinan926 Do you think there is an adequate existing fallback mechanism or do you still believe that there is a need to create a similar patch for ExecutorPodsWatchSnapshotSource ?
I see this error a lot in the batch jobs:
{"level":"WARN","timestamp":"2020-08-10 19:17:35,985","thread":"OkHttp https://kubernetes.default.svc/...","source":"io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager", "line":"209","message":"Exec Failure"}
java.io.EOFException
at okio.RealBufferedSource.require(RealBufferedSource.java:61)
at okio.RealBufferedSource.readByte(RealBufferedSource.java:74)
at okhttp3.internal.ws.WebSocketReader.readHeader(WebSocketReader.java:117)
at okhttp3.internal.ws.WebSocketReader.processNextFrame(WebSocketReader.java:101)
at okhttp3.internal.ws.RealWebSocket.loopReader(RealWebSocket.java:274)
at okhttp3.internal.ws.RealWebSocket$2.onResponse(RealWebSocket.java:214)
at okhttp3.RealCall$AsyncCall.execute(RealCall.java:206)
at okhttp3.internal.NamedRunnable.run(NamedRunnable.java:32)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
^C
I do think its related to the above issue. The batch job starts, Driver is able to spin up new executors, communicate with them and get the job done, but cannot clean them up.
This is with Spark 2.4.5 and Kubernetes Version: 1.15 and 1.16 with Multi Kubernetes Masters.
The above message repeats every 10 seconds. Let me know if its not related
It looks a bit different from what I see. For me, it appears to get stuck at the very end of writing data to Bigtable in the very last task of a job. Our partner is working to back port the fix I mentioned and I will let you know if that addresses the hang.
|
gharchive/pull-request
| 2020-04-30T13:54:37 |
2025-04-01T04:56:00.262979
|
{
"authors": [
"AmplabJenkins",
"ScrapCodes",
"SparkQA",
"dongjoon-hyun",
"holdenk",
"jkleckner",
"puneetloya",
"shaneknapp",
"stijndehaes"
],
"repo": "apache/spark",
"url": "https://github.com/apache/spark/pull/28423",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
663863005
|
[do not review][testing master PR build][discard] comment change
What changes were proposed in this pull request?
Why are the changes needed?
Does this PR introduce any user-facing change?
How was this patch tested?
Test build #126344 has started for PR 29190 at commit 4b88a16.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/30960/
Test PASSed.
Test build #126344 has finished for PR 29190 at commit 4b88a16.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/126344/
Test PASSed.
Test build #126360 has started for PR 29190 at commit 13951c8.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/30976/
Test PASSed.
Test build #126360 has finished for PR 29190 at commit 13951c8.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/126360/
Test PASSed.
Merged build finished. Test PASSed.
Test build #126387 has started for PR 29190 at commit fa41817.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/31002/
Test PASSed.
Test build #126387 has finished for PR 29190 at commit fa41817.
This patch fails due to an unknown error code, -9.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test FAILed.
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/126387/
Test FAILed.
Test build #126455 has started for PR 29190 at commit 54f8574.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/31069/
Test PASSed.
Test build #126455 has finished for PR 29190 at commit 54f8574.
This patch fails Spark unit tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test FAILed.
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/126455/
Test FAILed.
Test build #126508 has started for PR 29190 at commit 2f6264d.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/31127/
Test PASSed.
Test build #126508 has finished for PR 29190 at commit 2f6264d.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/126508/
Test PASSed.
Superceded by https://github.com/apache/spark/pull/29226
|
gharchive/pull-request
| 2020-07-22T15:44:52 |
2025-04-01T04:56:00.288882
|
{
"authors": [
"AmplabJenkins",
"SparkQA",
"agrawaldevesh"
],
"repo": "apache/spark",
"url": "https://github.com/apache/spark/pull/29190",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
771097732
|
[SPARK-33841][CORE] Fix issue with jobs disappearing intermittently from the SHS under high load
What changes were proposed in this pull request?
Mark SHS event log entries that were processing at the beginning of the checkForLogs run as not stale and check for this mark before deleting an event log. This fixes the issue when a particular job was displayed in the SHS and disappeared after some time, but then, in several minutes showed up again.
Why are the changes needed?
The issue is caused by SPARK-29043, which is designated to improve the concurrent performance of the History Server. The change breaks the "app deletion" logic because of missing proper synchronization for processing event log entries. Since SHS now filters out all processing event log entries, such entries do not have a chance to be updated with the new lastProcessed time and thus any entity that completes processing right after filtering and before the check for stale entities will be identified as stale and will be deleted from the UI until the next checkForLogs run. This is because updated lastProcessed time is used as criteria, and event log entries that missed to be updated with a new time, will match that criteria.
The issue can be reproduced by generating a big number of event logs and uploading them to the SHS event log directory on S3. Essentially, around 236(26.7 MB) copies of an event log directory were created using shs-monitor script. Strange behavior of SHS counting the total number of applications was noticed - at first, the number was increasing as expected, but with the next page refresh, the total number of applications decreased. No errors were logged by SHS.
58 entities are displayed at 17:35:35:
25 entities are displayed at 17:36:40:
Does this PR introduce any user-facing change?
Yes, SHS users won't face the behavior when the number of displayed applications decreases periodically.
How was this patch tested?
Tested using shs-monitor script:
Build SHS with the proposed change
Download Hadoop AWS and AWS Java SDK
Prepare S3 bucket and user for programmatic access, grant required roles to the user. Get access key and secret key
Configure SHS to read event logs from S3
Start monitor script to query SHS API
Run 5 producers for ~5 mins, create 125(14.2 MB) event log directory copies
Wait for SHS to load all the applications
Verify that the number of loaded applications increases continuously over time
For more details, please refer to the shs-monitor repository.
This version of the reproduction uses event log directories instead of single files, since recent optimization
SPARK-33790 makes it hard to reproduce the issue with single event log files.
Can one of the admins verify this patch?
cc @HeartSaVioR
The PR has been rebased.
Thank you, @dongjoon-hyun!
add to whitelist
retest this, please
I don't think this PR has an issue, but we need to ping @tgravescs to get his approval technically. Ping, @tgravescs .
+1 and I'd consider @tgravescs approved this PR as only target branch is different.
Test build #133038 has started for PR 30845 at commit e9f13a5.
Kubernetes integration test starting
URL: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/37637/
Kubernetes integration test status failure
URL: https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/37637/
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder-K8s/37637/
Thank you for your first contribution, @vladhlinsky .
I added you to the Apache Spark contributor group and assigned SPARK-33841 to you.
Welcome to the Apache Spark community.
Thank you, @dongjoon-hyun!
Test build #133038 has finished for PR 30845 at commit e9f13a5.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/133038/
|
gharchive/pull-request
| 2020-12-18T18:50:18 |
2025-04-01T04:56:00.307733
|
{
"authors": [
"AmplabJenkins",
"HeartSaVioR",
"SparkQA",
"dongjoon-hyun",
"vladhlinsky"
],
"repo": "apache/spark",
"url": "https://github.com/apache/spark/pull/30845",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1088316703
|
[MINOR][DOCS] Update pandas_pyspark.rst
What changes were proposed in this pull request?
'to' --> 'the'
Why are the changes needed?
Grammar fix
Does this PR introduce any user-facing change?
Documentation fix
How was this patch tested?
No test needed, same as yesterday's PR
Can one of the admins verify this patch?
@kamelCased while we're here, can we find all these instances, and fix them in one PR?
@HyukjinKwon I checked it's the last one in this file. I will check the rest of the PySpark docs.
The test fails because your bunch is not synced to the latest matster branch. do you mind rebasing to the latest master branch?
Also please update PR description to reflect the changes.
ok to test
@HyukjinKwon fetched upstream and PR description updated
@HyukjinKwon fetched upstream and PR description updated
@HyukjinKwon fetched upstream and PR description updated
Merged to master
|
gharchive/pull-request
| 2021-12-24T10:59:40 |
2025-04-01T04:56:00.313014
|
{
"authors": [
"AmplabJenkins",
"HyukjinKwon",
"kamelCased",
"srowen"
],
"repo": "apache/spark",
"url": "https://github.com/apache/spark/pull/35016",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1243995818
|
[SPARK-39249][SQL] Improve subexpression elimination for conditional expressions
What changes were proposed in this pull request?
Currently we can do subexpression elimination for conditional expressions when the subexpression is common across all branchGroups. In fact, we can farther improve this when there are common expressions between alwaysEvaluatedInputs and branchGroups.
Why are the changes needed?
Take the following case as an example
IF(IsNull(a), b, KnowNotNull(a))
a may miss subexpression elimination chances since it is not the common expression between all branchGroups, but it's safe to evaluate a as common subexpression and eagerly execute it since it's part of the prediction, which will always be executed. If a is a time-expensive expression, we may waste time on running it.
This kind of expressions are common when we do sum on decimal type because of https://github.com/apache/spark/pull/29026
https://github.com/apache/spark/blob/291d155b3c514f8b590a6b078f7efd42a30e67f0/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/aggregate/Sum.scala#L125
Performance result on TPC-DS 10T
Query
With this PR
Without this PR
Speed up
4
310.862
635.299
104.37%
80
36.723
46.006
25.28%
Does this PR introduce any user-facing change?
No
How was this patch tested?
add more UT.
@viirya @cloud-fan Could you please help review this?
Can one of the admins verify this patch?
FYI I created https://github.com/apache/spark/pull/32987 a while ago to address this in a much more general way. I've tried to keep it up to date, but there seemed to be concerns about creating a subexpression for something that might only execute once I guess? Even though that's already happening in certain cases
@viirya @cloud-fan Could you please help review this?
|
gharchive/pull-request
| 2022-05-21T14:55:19 |
2025-04-01T04:56:00.320067
|
{
"authors": [
"AmplabJenkins",
"Kimahriman",
"WangGuangxin"
],
"repo": "apache/spark",
"url": "https://github.com/apache/spark/pull/36626",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
54532482
|
[SPARK-4874] [CORE] Collect record count metrics
Collects record counts for both Input/Output and Shuffle Metrics. For the input/output metrics, it just appends the counter every time the iterators get accessed.
For shuffle on the write side, we count the metrics post aggregation (after a map side combine) and on the read side we count the metrics pre aggregation. This allows both the bytes read/written metrics and the records read/written to line up.
For backwards compatibility, if we deserialize an older event that doesn't have record metrics, we set the metric to -1.
Test build #25628 has started for PR 4067 at commit 571cb69.
This patch merges cleanly.
Test build #25628 has finished for PR 4067 at commit 571cb69.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/25628/
Test PASSed.
Can you also paste some screenshots on what the UI changes look like? Thanks.
Test build #25638 has started for PR 4067 at commit 1572054.
This patch does not merge cleanly.
This change was dependent on https://github.com/apache/spark/pull/3120, that just got merged and now there are some merge conflicts. I need to fix those first and will update the pr.
Hey @ksakellis - Thanks for working on this.
I took a very quick look at the patch. Overall I feel the patch should be fairly straightforward, but the specific implementations might've gone a bit over board with Scala language features (a lot of Options, orElse, closures, etc) and design (too many new classes introduced). If we can reduce those, the pr would be a lot easier to understand.
@rxin I updated the PR after doing a rebase and also incorporated some of your feedback. You made two general comments:
specific implementations might've gone a bit over board with Scala language features
too many new classes introduced
Can you please be more specific here pointing to the specific code. These comments are not very actionable as is.
Hi again - can't find my previous comment since the line is no longer in the diff due to the other pr being merged. Can you still add comment for that one (the part with Option and orElse and set ...)? Want to make sure if we read that code one year from now, we can still understand what's going on.
Test build #25641 has started for PR 4067 at commit 3c2d021.
This patch merges cleanly.
The Scala stuff was mostly about the previous PR that got merged (and now no longer showing up as part of this diff).
So is this code you were referring to in HadoopRDD?
// Find a function that will return the FileSystem bytes read by this thread. Do this before
// creating RecordReader, because RecordReader's constructor might read some bytes
val bytesReadCallback = inputMetrics.bytesReadCallback.orElse(
split.inputSplit.value match {
case split: FileSplit =>
SparkHadoopUtil.get.getFSBytesReadOnThreadCallback(split.getPath, jobConf)
case _ => None
}
)
Shows a stage that has Input Metrics (reading from a file) and writes data for next stage.
Shows a stage that has both shuffle reading and writing - no input or output metrics.
Shows a stage that has outputting to a file.
Test build #25638 timed out for PR 4067 at commit 1572054 after a configured wait of 120m.
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/25638/
Test FAILed.
Test build #25641 has finished for PR 4067 at commit 3c2d021.
This patch passes all tests.
This patch merges cleanly.
This patch adds the following public classes (experimental):
class AfterNextInterceptingIterator[A](sub: Iterator[A]) extends Iterator[A]
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/25641/
Test PASSed.
What about combining the input size and records in the same column. Overall this will help with the expansion in the number of columns. The title could be "Input Size / Records"
If we do that we wouldn't be able to sort on num records and bytes independently.
Yes - you'd only be able to sort on bytes. Wouldn't that be okay? These would likely track closely in most cases.
A big motivation to add recordsRead/Written was to detect data skew. In these cases bytes and records might not track very closely.
Thinking more about this, I suspect that having an Avg. record Size column (bytesRead/recordsRead) would be what you'd want to sort on. We could add this metric to the UI, make it sortable and then combine the bytesRead and recordsRead metrics into one column. Thoughts?
New Screenshot that correspond to CR feedback.
Shows a stage that has Input Metrics (reading from a file) and writes data for next stage.
Shows a stage that has both shuffle reading and writing - no input or output metrics.
Shows a stage that has outputting to a file.
@pwendell can you please re-review this? I'd like to get it in to 1.3. Some of our customers have been asking for metrics to help them determine data skew. Thx.
Test build #26441 has started for PR 4067 at commit fe3f715.
This patch merges cleanly.
Test build #26441 has finished for PR 4067 at commit fe3f715.
This patch passes all tests.
This patch merges cleanly.
This patch adds the following public classes (experimental):
class AfterNextInterceptingIterator[A](sub: Iterator[A]) extends Iterator[A]
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/26441/
Test PASSed.
Having a count does seem like a good idea, but I think the current approach of adding function calls for each iteration is too expensive. What about using a counting iterator and just asking it at the end?
BTW we can also just add counting directly to InterruptibleIterator
@rxin yeah that seems good.
@ksakellis one other thing I realized which is a little confusing, right now we report the bytes as "read" from a cached RDD as soon as they are fetched/present on the executor even if they have not been consumed by the task. Tracking consumption incrementally (in bytes) will be really hard, so maybe this is the best answer for now. It's a bit weird though - I think this logic was written before we sent incremental updates back. In terms of getting the total records read, we might need to make the assumption that the iterator is consumed in its entirety.
@pwendell I'm not sure how we can do what you propose without having an O(n) loop through all the records before passing the InterruptableIterator? We could do something fancy like counting incrementally and when we finish the task, if there are more records left, then do the loop to count the rest of the unread records. I don't think the complication is worth it. Also, reporting the accurate records read i think is better.
Alternatively, we can fix the bytesRead to be more accurate. Right now they are computed in ShuffleBlockFetcherIterator and calculated based on the blocks fetched. Since we do the flatMap on that iterator in BlockStoreShuffleFetcher we report that we read all the bytes even if we didn't. We can move the bytesRead collection out of ShuffleBlockFetcherIterator and move it into the same iterator that computes the # records read. So they line up and are more accurate.
If it would save on perf, I would be in favor of adding the counting to InterruptibleIterator. In core loops like these, I think efficiency is more important than elegance.
Test build #26530 has started for PR 4067 at commit 1aaa980.
This patch merges cleanly.
Test build #26530 has finished for PR 4067 at commit 1aaa980.
This patch passes all tests.
This patch merges cleanly.
This patch adds the following public classes (experimental):
class ChiSqSelectorModel (val selectedFeatures: Array[Int]) extends VectorTransformer
class ChiSqSelector (val numTopFeatures: Int)
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/26530/
Test PASSed.
Test build #26678 has started for PR 4067 at commit 9ecf912.
This patch merges cleanly.
Test build #26678 has finished for PR 4067 at commit 9ecf912.
This patch fails MiMa tests.
This patch merges cleanly.
This patch adds the following public classes (experimental):
class SimpleFunctionRegistry(val caseSensitive: Boolean) extends FunctionRegistry
class StringKeyHashMap[T](normalizer: (String) => String)
case class MultiAlias(child: Expression, names: Seq[String])
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/26678/
Test FAILed.
Hey @ksakellis - I did a pretty thorough review here, any chance you could take a look? Most of the things were small. The only major thing is that I think if we are going to always show the output metrics for Hadoop (even for older versions) we need to make sure that we at least do a "best effort" presentation of the output size for older versions. I think it would be good if we could sneak that in, since otherwise that output metrics thing is not so useful for many people running with older versions. However, I'd also be okay just not showing either bytes or count in that case (which I think was the behavior before).
A second issue I realized (let's open a new JIRA for this) is that we never track output metrics (size or bytes) when data is written to cache, which is confusing because we track input metrics when data is read from cache.
New screenshots with the irrelevant columns invisible:
Shows a stage that has Input Metrics (reading from a file) and writes data for next stage.
Shows a stage that has both shuffle reading and writing - no input or output metrics.
Shows a stage outputting to an HDFS file.
Test build #26876 has started for PR 4067 at commit e156560.
This patch merges cleanly.
Test build #26876 has finished for PR 4067 at commit e156560.
This patch fails Spark unit tests.
This patch merges cleanly.
This patch adds no public classes.
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/26876/
Test FAILed.
Jenkins, retest this please
Test build #26896 has started for PR 4067 at commit e156560.
This patch merges cleanly.
Test build #26896 has finished for PR 4067 at commit e156560.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/26896/
Test PASSed.
Test build #26904 has started for PR 4067 at commit dad4d57.
This patch merges cleanly.
Test build #26906 has started for PR 4067 at commit bd919be.
This patch merges cleanly.
Test build #26904 has finished for PR 4067 at commit dad4d57.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/26904/
Test PASSed.
Test build #26906 has finished for PR 4067 at commit bd919be.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/26906/
Test PASSed.
Jenkins, test this please. This LGTM pending tests.
Test build #26936 has started for PR 4067 at commit bd919be.
This patch merges cleanly.
Test build #26936 has finished for PR 4067 at commit bd919be.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/26936/
Test PASSed.
Merging this, thanks Kos.
It looks like the "InputOutputMetricsSuite input metrics with mixed read methods" and "InputOutputMetricsSuite input metrics with interleaved reads" test may have started failing in the hadoop-2.2 build since this patch:
https://amplab.cs.berkeley.edu/jenkins/view/Spark/job/Spark-1.3-SBT/AMPLAB_JENKINS_BUILD_PROFILE=hadoop2.2,label=centos/29/testReport/
Yikes, @JoshRosen i'm looking into this.
@ksakellis How do I collect these metrics on a console right after the task is done.
|
gharchive/pull-request
| 2015-01-16T02:18:58 |
2025-04-01T04:56:00.375397
|
{
"authors": [
"AmplabJenkins",
"JoshRosen",
"SparkQA",
"aguyyala",
"ksakellis",
"pwendell",
"rxin",
"sryza"
],
"repo": "apache/spark",
"url": "https://github.com/apache/spark/pull/4067",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1960156183
|
[SPARK-45655][SQL][SS] Allow non-deterministic expressions inside AggregateFunctions in CollectMetrics
What changes were proposed in this pull request?
This PR allows non-deterministic expressions wrapped inside an AggregateFunction such as count inside CollectMetrics node. CollectMetrics is used to collect arbitrary metrics from the query, in certain scenarios user would like to collect metrics for filtering based on non-deterministic expressions (see query example below).
Currently, Analyzer does not allow non-deterministic expressions inside a AggregateFunction for CollectMetrics. This constraint is relaxed to allow collection of such metrics. Note that the metrics are relevant for a completed batch, and can change if the batch is replayed (because non-deterministic expression can behave differently for different runs).
While working on this feature, I found a issue with checkMetric logic to validate non-deterministic expressions inside an AggregateExpression. An expression is determined as non-deterministic if any of its children is non-deterministic, hence we need to match the case for !e.deterministic && !seenAggregate after we have matched if the current expression is a AggregateExpression. If the current expression is a AggregateExpression, we should validate further down in the tree recursively - otherwise we will fail for any non-deterministic expression.
val inputData = MemoryStream[Timestamp]
inputData.toDF()
.filter("value < current_date()")
.observe("metrics", count(expr("value >= current_date()")).alias("dropped"))
.writeStream
.queryName("ts_metrics_test")
.format("memory")
.outputMode("append")
.start()
Why are the changes needed?
Added a testcase to calculate dropped rows (by CurrentBatchTimestamp) and ensure the query is successful.
As an example, the query below fails (without this change) due to observe call on the DataFrame.
val inputData = MemoryStream[Timestamp]
inputData.toDF()
.filter("value < current_date()")
.observe("metrics", count(expr("value >= current_date()")).alias("dropped"))
.writeStream
.queryName("ts_metrics_test")
.format("memory")
.outputMode("append")
.start()
Added testing in AnalysisSuite for non-deterministic expressions inside a AggregateFunction.
Does this PR introduce any user-facing change?
No
How was this patch tested?
Unit test cases added.
[warn] 20 warnings found
WARNING: Using incubator modules: jdk.incubator.vector, jdk.incubator.foreign
[info] StreamingQueryStatusAndProgressSuite:
09:14:39.684 WARN org.apache.hadoop.util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
[info] Passed: Total 0, Failed 0, Errors 0, Passed 0
[info] No tests to run for hive / Test / testOnly
[info] - StreamingQueryProgress - prettyJson (436 milliseconds)
[info] - StreamingQueryProgress - json (3 milliseconds)
[info] - StreamingQueryProgress - toString (5 milliseconds)
[info] - StreamingQueryProgress - jsonString and fromJson (163 milliseconds)
[info] - StreamingQueryStatus - prettyJson (1 millisecond)
[info] - StreamingQueryStatus - json (1 millisecond)
[info] - StreamingQueryStatus - toString (2 milliseconds)
09:14:41.674 WARN org.apache.spark.sql.execution.streaming.ResolveWriteToStream: Temporary checkpoint location created which is deleted normally when the query didn't fail: /Users/bhuwan.sahni/workspace/spark/target/tmp/temporary-34d2749f-f4d0
-46d8-bc51-29da6411e1c5. If it's required to delete it under any circumstances, please set spark.sql.streaming.forceDeleteTempCheckpointLocation to true. Important to know deleting temp checkpoint folder is best effort.
09:14:41.710 WARN org.apache.spark.sql.execution.streaming.ResolveWriteToStream: spark.sql.adaptive.enabled is not supported in streaming DataFrames/Datasets and will be disabled.
[info] - progress classes should be Serializable (5 seconds, 552 milliseconds)
09:14:46.345 WARN org.apache.spark.sql.execution.streaming.ResolveWriteToStream: Temporary checkpoint location created which is deleted normally when the query didn't fail: /Users/bhuwan.sahni/workspace/spark/target/tmp/temporary-3a41d397-c3c1
-490b-9cc7-d775b0c42208. If it's required to delete it under any circumstances, please set spark.sql.streaming.forceDeleteTempCheckpointLocation to true. Important to know deleting temp checkpoint folder is best effort.
09:14:46.345 WARN org.apache.spark.sql.execution.streaming.ResolveWriteToStream: spark.sql.adaptive.enabled is not supported in streaming DataFrames/Datasets and will be disabled.
[info] - SPARK-19378: Continue reporting stateOp metrics even if there is no active trigger (1 second, 337 milliseconds)
09:14:47.677 WARN org.apache.spark.sql.execution.streaming.ResolveWriteToStream: spark.sql.adaptive.enabled is not supported in streaming DataFrames/Datasets and will be disabled.
[info] - SPARK-29973: Make `processedRowsPerSecond` calculated more accurately and meaningfully (455 milliseconds)
09:14:48.174 WARN org.apache.spark.sql.execution.streaming.ResolveWriteToStream: Temporary checkpoint location created which is deleted normally when the query didn't fail: /Users/bhuwan.sahni/workspace/spark/target/tmp/temporary-360fc3b9-a2c5
-430c-a892-c9869f1f8339. If it's required to delete it under any circumstances, please set spark.sql.streaming.forceDeleteTempCheckpointLocation to true. Important to know deleting temp checkpoint folder is best effort.
09:14:48.174 WARN org.apache.spark.sql.execution.streaming.ResolveWriteToStream: spark.sql.adaptive.enabled is not supported in streaming DataFrames/Datasets and will be disabled.
[info] - SPARK-45655: Use current batch timestamp in observe API (587 milliseconds)
09:14:48.768 WARN org.apache.spark.sql.streaming.StreamingQueryStatusAndProgressSuite:
Was this patch authored or co-authored using generative AI tooling?
No
FYI, the previous test failure was relevant to this change, and @sahnib is looking into it.
(I'm sorry to miss leaving a note in prior.)
Please check SPARK-45655: Use current batch timestamp in observe API *** FAILED *** (241 milliseconds)
Please check SPARK-45655: Use current batch timestamp in observe API *** FAILED *** (241 milliseconds)
This test was failing because CI machine has milliseconds precision of 8 decimal points, I have fixed the testcase.
Thanks! Merging to master.
@sahnib Thanks for your first contribution to Apache Spark! I merged this to master.
|
gharchive/pull-request
| 2023-10-24T22:05:48 |
2025-04-01T04:56:00.385935
|
{
"authors": [
"HeartSaVioR",
"beliefer",
"sahnib"
],
"repo": "apache/spark",
"url": "https://github.com/apache/spark/pull/43517",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2373990818
|
[MINOR] Fix CodeStyle Error for ExpressionImplUtilsSuite
What changes were proposed in this pull request?
The pr is following up https://github.com/apache/spark/pull/46845, to fix dev/scalastyle check error.
Why are the changes needed?
Make sh dev/scalastyle hadppy.
Does this PR introduce any user-facing change?
No.
How was this patch tested?
Manually test
sh dev/scakastyle
Before:
(base) ➜ spark-community git:(master) ✗ sh dev/scalastyle
-e Scalastyle checks failed at following occurrences:
[error] /Users/panbingkun/Developer/spark/spark-community/sql/catalyst/src/test/java/org/apache/spark/sql/catalyst/expressions/ExpressionImplUtilsSuite.scala:394:28: nonascii.message
[error] /Users/panbingkun/Developer/spark/spark-community/sql/catalyst/src/test/java/org/apache/spark/sql/catalyst/expressions/ExpressionImplUtilsSuite.scala:396:28: nonascii.message
[error] Total time: 23 s, completed Jun 26, 2024, 9:49:06 AM
After:
(base) ➜ spark-community git:(fix_scalastyle) ✗ sh dev/scalastyle
-e Scalastyle checks passed.
Was this patch authored or co-authored using generative AI tooling?
No.
cc @HyukjinKwon @cloud-fan @uros-db
let me merge it now to unblock the CI.
|
gharchive/pull-request
| 2024-06-26T01:43:18 |
2025-04-01T04:56:00.390023
|
{
"authors": [
"cloud-fan",
"panbingkun"
],
"repo": "apache/spark",
"url": "https://github.com/apache/spark/pull/47094",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
67102377
|
[SQL] Faster Scala row conversion
This is a follow-up of #5279 and #5398. ScalaReflection.convertRowToScala is on a critical path, but was implemented in a rather inefficient way.
Test build #29856 has started for PR 5419 at commit d698d03.
Test build #29856 has finished for PR 5419 at commit d698d03.
This patch passes all tests.
This patch merges cleanly.
This patch adds no public classes.
This patch does not change any dependencies.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/29856/
Test PASSed.
Actually the convertRowToScala part is already handled in #5279, and the lazy val schema part is handled in #5398, so I'm closing this.
|
gharchive/pull-request
| 2015-04-08T10:48:06 |
2025-04-01T04:56:00.395135
|
{
"authors": [
"AmplabJenkins",
"SparkQA",
"liancheng"
],
"repo": "apache/spark",
"url": "https://github.com/apache/spark/pull/5419",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
112479556
|
[SPARK-11221] [SPARKR] fix R doc for lit and add examples
Currently the documentation for lit is inconsistent with doc format, references "Scala symbol" and has no example. Fixing that.
@shivaram
Can one of the admins verify this patch?
Jenkins, ok to test
Jenkins, add to whitelist
Merged build triggered.
Merged build started.
Test build #44013 has started for PR 9187 at commit 1ca6612.
Test build #44013 has finished for PR 9187 at commit 1ca6612.
This patch passes all tests.
This patch merges cleanly.
This patch adds the following public classes (experimental):\n * class BinaryClassificationEvaluator @Since(\"1.4.0\") (@Since(\"1.4.0\") override val uid: String)\n * class MulticlassClassificationEvaluator @Since(\"1.5.0\") (@Since(\"1.5.0\") override val uid: String)\n * final class RegressionEvaluator @Since(\"1.4.0\") (@Since(\"1.4.0\") override val uid: String)\n * abstract class ColumnarIterator extends Iterator[InternalRow] \n * class SpecificColumnarIterator extends $\n
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/44013/
Test PASSed.
LGTM. Thanks @felixcheung Merging this
|
gharchive/pull-request
| 2015-10-20T23:19:56 |
2025-04-01T04:56:00.401412
|
{
"authors": [
"AmplabJenkins",
"SparkQA",
"felixcheung",
"shivaram"
],
"repo": "apache/spark",
"url": "https://github.com/apache/spark/pull/9187",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
246847633
|
WIP: STORM-2648/STORM-2357: Add storm-kafka-client support for at-most-onc…
…e processing and a toggle for whether messages should be emitted with a message id when not using at-least-once
See https://issues.apache.org/jira/browse/STORM-2357 and https://issues.apache.org/jira/browse/STORM-2648.
I'd like to get some opinions on whether this approach is a good idea, or whether I've overlooked a better option, before finishing this up with some tests. I don't love that we'll end up with 3 different committing behaviors.
In 2357 it was noted that the spout doesn't currently support true at-most-once, because using Kafka's auto commit option leaves the possibility that the spout receives a tuple, emits it to the topology, crashes and recovers, and then receives and emits the same tuple. The linked issue suggests solving this by committing polled offsets before emitting them to the topology, which is an option added here.
2648 notes that there is currently no way to make Storm track messages when using auto commit with this spout. This prevents Storm UI from showing the complete latency for the spout, and I would assume also prevents max spout pending from having an effect. I've added a toggle to KafkaSpoutConfig to force the spout to emit messages with message ids, even when using auto commit or the at-most-once option. The spout does nothing on ack or fail when not doing at-least-once.
I'd like to keep the spout config simple for the user, so I've added a processing guarantee setting corresponding to the standard at-least-once code path, the path that uses auto commit, and the path that commits offsets before emitting any tuples.
@srdo can you please assign JIRA's to you and mark them as in progress as you work on them and/or submit a pull request. Thanks.
Yes, I forgot.
Conceptually the changes look good to me. I have not dug into it in great detail yet, but I do like the direction of the change.
I would also like to see the documentation and examples updated to reflect the new change.
+1
@srdo reviewing it
Added some tests and updated the docs.
@hmcl Are you reviewing this, or are you satisfied with it?
Still +1
@hmcl I don't mean to try to rush you, but please let me know if you're still reviewing. If not I'll probably merge in the next few days.
@srdo apologies for the delay. I will finish today.
Thanks :)
@hmcl Are you still reviewing this? It's fine if you don't have time to look at this, but please say so. I'd like to not keep holding this up.
This PR has been waiting for about 2 months, and once it gets +1 and no -1, it can be merged.
I'm +1 and will just merge.
@hmcl Please vote -1 later and rollback the merge if you have concern about the patch and would want to vote -1.
Merged via https://github.com/apache/storm/commit/48f6969027e7b02a5b9220577189d3911aa2226d
@srdo
Sorry I forgot to add auto close message while squashing commits. Could you close this?
And please craft the patch for 1.x branch since it doesn't looks like a clean cherry-pick.
Thanks in advance!
Thanks for reviews. Opened the 1.x version here https://github.com/apache/storm/pull/2353.
|
gharchive/pull-request
| 2017-07-31T18:46:34 |
2025-04-01T04:56:00.410160
|
{
"authors": [
"HeartSaVioR",
"hmcl",
"revans2",
"srdo"
],
"repo": "apache/storm",
"url": "https://github.com/apache/storm/pull/2249",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
479285601
|
STORM-3071: change checkstyle plugin setting logViolationsToConsole t…
…o true
Logging checkstyle violations to console which is now possible since all violations have been fixed
Remove unnecessary maxAllowedViolations from configurations of projects
Note: The removal of maxAllowedViolations in the server project is covered in #3050 which needs to be merged before.
I took a look at #3050 . We can come back to this once the former's changes are merged.
It looks nice. Will come back to this once #3050 is merged
#3050 is now merged. Could you please rebase this so we can get this in to? Thanks again for the efforts
The failed test doesn't seem related
[INFO] Running org.apache.storm.daemon.drpc.DRPCTest
[ERROR] Tests run: 6, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 1.038 s <<< FAILURE! - in org.apache.storm.daemon.drpc.DRPCTest
[ERROR] testGoodBlocking Time elapsed: 0.012 s <<< ERROR!
java.util.concurrent.TimeoutException
|
gharchive/pull-request
| 2019-08-10T16:47:16 |
2025-04-01T04:56:00.412958
|
{
"authors": [
"Ethanlm",
"d2r",
"krichter722"
],
"repo": "apache/storm",
"url": "https://github.com/apache/storm/pull/3100",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
611290066
|
SUBMARINE-487. JNA does not work with maven-shade-plugin
What is this PR for?
submarine-spark-security fails with java.lang.UnsatisfiedLinkError
What type of PR is it?
[Bug Fix]
Todos
[ ] - Task
What is the Jira issue?
https://issues.apache.org/jira/browse/SUBMARINE-487
How should this be tested?
Screenshots (if appropriate)
Questions:
Does the licenses files need update? No
Is there breaking changes for older versions? No
Does this needs documentation? No
@gumartinm Thank you for your interest in the submarine project. :-)
@yaooqinn Please help review this PR, Thanks!
Thanks for the fix, LGTM
@gumartinm feel free to send follow-ups with your no-uber-jar ideas, I use uber-jar here because of jar conflict between spark and ranger project.
@liuxunorg I am ok to merge this.
@gumartinm feel free to send follow-ups with your no-uber-jar ideas, I use uber-jar here because of jar conflict between spark and ranger project.
@liuxunorg I am ok to merge this.
@yaooqinn I just wanted to say that you are doing a great job with this project. I've been following it since it was called spark-authorizer. Hopefully, one day, it will be integrated in the Spark framework.
|
gharchive/pull-request
| 2020-05-02T22:48:02 |
2025-04-01T04:56:00.418095
|
{
"authors": [
"gumartinm",
"liuxunorg",
"yaooqinn"
],
"repo": "apache/submarine",
"url": "https://github.com/apache/submarine/pull/273",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
913343921
|
Determination of the select queries in BaseEngineSpec
Hello!
We are now developing a new SQLAlchemy dialect for Kusto (Azure Data Explorer) and want add support for Kusto database in Superset in nearly future. (This also covers https://github.com/apache/superset/issues/10646.)
Kusto language has two dialects: SQL and KQL. While there is no problem with SQL dialect, we also want to integrate KQL in Superset.
Problem
KQL select queries do not contain the word "SELECT" so they look like
MyTable | take 10 (this is equivalent of Select * from MyTable Limit 10).
The absence of the "SELECT" keyword breaks the limit behavior in SQLLab because it relies on the sqlparse package and we can't add a limit to queries from UI.
Proposed solution
The solution is to make the determination of the select queries in BaseEngineSpec so that you can override it in your own spec. This is the same way how base dialect determines read-only queries with is_readonly_query.
We will provide PR to implement this small feature.
Thank you!
Closed by https://github.com/apache/superset/pull/15013
|
gharchive/issue
| 2021-06-07T09:34:20 |
2025-04-01T04:56:00.421585
|
{
"authors": [
"xneg"
],
"repo": "apache/superset",
"url": "https://github.com/apache/superset/issues/15011",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2755104947
|
Edit CSS Live can not @import google fonts?
Bug description
I try to @import google font in Edit CSS Live section. It not effect ???
@import url('https://fonts.googleapis.com/css2?family=Prompt:ital,wght@0,100;0,200;0,300;0,400;0,500;0,600;0,700;0,800;0,900;1,100;1,200;1,300;1,400;1,500;1,600;1,700;1,800;1,900&display=swap');
body{
font-family:'Prompt';
}
Screenshots/recordings
No response
Superset version
4.1.1
Python version
3.9
Node version
16
Browser
Chrome
Additional context
No response
Checklist
[ ] I have searched Superset docs and Slack and didn't find a solution to my problem.
[X] I have searched the GitHub issue tracker and didn't find a similar bug report.
[ ] I have checked Superset's logs for errors and if I found a relevant Python stacktrace, I included it here as text in the "additional context" section.
You can, but dosubot is rigjt... you will have to punch a hole in TALISMAN_CONFIG for it.
Closing this since it's not a bug, but a question that's now been answered. Holler if you have any further troubles with it :)
|
gharchive/issue
| 2024-12-23T03:02:54 |
2025-04-01T04:56:00.426641
|
{
"authors": [
"psongyoo",
"rusackas"
],
"repo": "apache/superset",
"url": "https://github.com/apache/superset/issues/31597",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
942367461
|
refactor: Move is_user_admin() to security_manager
SUMMARY
Having is_user_admin is a design flaw, but in any
case should always be handled within the security
manager so it can be overridden.
BEFORE/AFTER SCREENSHOTS OR ANIMATED GIF
TESTING INSTRUCTIONS
Test admin user works as befre
ADDITIONAL INFORMATION
[ ] Has associated issue:
[ ] Changes UI
[ ] Includes DB Migration (follow approval process in SIP-59)
[ ] Migration is atomic, supports rollback & is backwards-compatible
[ ] Confirm DB migration upgrade and downgrade tested
[ ] Runtime estimates and downtime expectations provided
[ ] Introduces new feature or API
[ ] Removes existing feature or API
@bolkedebruin are you still hoping to get this PR merged? There are some review comments that would need to be addressed to get this merged (and a rebase).
Yes, I'll take a look at it again.
|
gharchive/pull-request
| 2021-07-12T19:17:29 |
2025-04-01T04:56:00.430813
|
{
"authors": [
"bolkedebruin",
"villebro"
],
"repo": "apache/superset",
"url": "https://github.com/apache/superset/pull/15636",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1338283845
|
fix(install): set SUPERSET_VERSION_RC at the right time
fixes #21082
SUMMARY
Fixes problem as described in the issue #21082
TESTING INSTRUCTIONS
execute command below in a environment which doesn't contain any of the superset env variables.
[YOUR_SUPERSET_PATH]/RELEASING/make_tarball.sh [VERSION] [RC] [PGP_KEY_NAME]
this should run without failures.
ADDITIONAL INFORMATION
fixes #21082
[ ] Has associated issue:
[ ] Required feature flags:
[ ] Changes UI
[ ] Includes DB Migration (follow approval process in SIP-59)
[ ] Migration is atomic, supports rollback & is backwards-compatible
[ ] Confirm DB migration upgrade and downgrade tested
[ ] Runtime estimates and downtime expectations provided
[ ] Introduces new feature or API
[ ] Removes existing feature or API
@Joel-Haeberli — Sorry this slipped though the cracks for so long! Thanks for reporting both the issue and opening the PR. Please feel free to tag committers here or ping us on Superset Slack if things get stuck like this.
@eschutho this seems like a straightforward fix to a problem I haven't encountered... curious your take :)
|
gharchive/pull-request
| 2022-08-14T15:51:23 |
2025-04-01T04:56:00.435736
|
{
"authors": [
"Joel-Haeberli",
"rusackas"
],
"repo": "apache/superset",
"url": "https://github.com/apache/superset/pull/21083",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2287450475
|
fix: tests on database, dataset apis
SUMMARY
BEFORE/AFTER SCREENSHOTS OR ANIMATED GIF
TESTING INSTRUCTIONS
ADDITIONAL INFORMATION
[ ] Has associated issue:
[ ] Required feature flags:
[ ] Changes UI
[ ] Includes DB Migration (follow approval process in SIP-59)
[ ] Migration is atomic, supports rollback & is backwards-compatible
[ ] Confirm DB migration upgrade and downgrade tested
[ ] Runtime estimates and downtime expectations provided
[ ] Introduces new feature or API
[ ] Removes existing feature or API
Codecov Report
All modified and coverable lines are covered by tests :white_check_mark:
Project coverage is 64.13%. Comparing base (2e5f3ed) to head (2a94fab).
Report is 57 commits behind head on master.
Additional details and impacted files
@@ Coverage Diff @@
## master #28397 +/- ##
==========================================
+ Coverage 60.49% 64.13% +3.64%
==========================================
Files 1931 521 -1410
Lines 76241 37255 -38986
Branches 8566 0 -8566
==========================================
- Hits 46122 23895 -22227
+ Misses 28015 13360 -14655
+ Partials 2104 0 -2104
Flag
Coverage Δ
hive
49.08% <ø> (-0.09%)
:arrow_down:
javascript
?
presto
53.66% <ø> (-0.14%)
:arrow_down:
python
64.13% <ø> (+0.65%)
:arrow_up:
unit
58.35% <ø> (+0.72%)
:arrow_up:
Flags with carried forward coverage won't be shown. Click here to find out more.
:umbrella: View full report in Codecov by Sentry.
:loudspeaker: Have feedback on the report? Share it here.
|
gharchive/pull-request
| 2024-05-09T11:03:40 |
2025-04-01T04:56:00.447563
|
{
"authors": [
"codecov-commenter",
"dpgaspar"
],
"repo": "apache/superset",
"url": "https://github.com/apache/superset/pull/28397",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1181456820
|
TOMEE-3871 - TomEE Plume is missing BatchEE / JCS Cache
batchee and jcs added in Plume to reflect Plus flavor
removed xml api from from Plume to reflect Plus flavor
remains some jar refs in jar scan exclusions (xerces, xml-apis)
AFAIK plume is a webprofile based distro so batchee and jcs should belong to a new "plume-all" distro probably?
AFAIK plume is a webprofile based distro so batchee and jcs should belong to a new "plume-all" distro probably?
This originates from this list thread and @dblevins comment there, that: "Those distributions are supposed to be the same minus the JPA and JSF providers.". There is also a blog post about the differences but I am too new to judge the original thoughts regarding "plume" vs "plus".
AFAIK plume is a webprofile based distro
if this is correct then Plume should not contains jars for MicroProfile nor JMS ActiveMQ
Yep, think they shouldnt but as Im not a big user of that I dont care much of what it becomes while webpro and plus stays aligned on their goal ;)
webpro and plus have a clear focus to me as well, even though "Plus" branding could be changed to better tell its "Full EE" 🤔 or i can just describe that in the flavors page.
You could drop Plume but you would lose the benefits from its "ready to try out" capabilities.
i am thinking about the AutoConfiguration for Jakarta Persistence (JPA) which facilitates learning Tomee.
I think it is good to have this discussion to clarify what the intended distributions should look like in 2022 considering the history of Plus / Plume and the incosistent current status of the distros / website. So thanks for the additional thoughts @rmannibucau and @sultan
does this need a mailing list vote from committers ?
Here's the list message on the name and motivation for Plume:
https://lists.apache.org/thread/tkpfjlzv70y8kzz5f905qj9vs95b34rw
"a version of TomEE Plus with Mojarra and EclipseLink. Tentatively called it "TomEE Plume" which stands for TomEE PLUs, Mojarra, Eclipselink (PLU.M.E)."
nice to read !
i learned that a setup installer for TomEE "like eclipse oomph" (cherry pick modules) is not possible, as it breaks TCK rules. unless we want a parallel non TCK certified TomEE (meh)
i thought like @rmannibucau at first with "per functionalities" LIB folders (like others EE servers) but this would make TomEE harder to configure / extend. (where do user put their jars !?)
there could be "per functionalities" uber or shaded jars to ease remove and replace implementations by others. but i wonder if i did read that current project policy is to avoid shading/uber.
there could be a common POM between plus and plume to help fill the gaps/ease maintenance.
there could be an external tool to modify a dezipped certified TomEE to change its implementation jars.
the only one i think i can currently help on is n°4 (common plus-plume pom)
there could be an external tool to modify a dezipped certified TomEE to change its implementation jars.
If you look tomee-maven-plugin it enables to do any distro from plain tomcat to advanced customized tomee and you can use it to build a tomee plume from tomee webprofile or any distro ;). Installer is legacy since years :)
@sultan Can you check your rebase with tomee-8.x. It does not look sane to me (at least judging from the changeset now).
i think i was able to clean up my mess by reverting to my previous draft. commits seems legit now. can someone launch checks ?
|
gharchive/pull-request
| 2022-03-26T01:47:48 |
2025-04-01T04:56:00.475226
|
{
"authors": [
"dblevins",
"rmannibucau",
"rzo1",
"sultan"
],
"repo": "apache/tomee",
"url": "https://github.com/apache/tomee/pull/828",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1448526684
|
Revert Traffic Monitor renaming to rascal
Which Traffic Control components are affected by this PR?
Documentation
Traffic Control Cache Config (t3c, formerly ORT)
Traffic Control Health Client (tc-health-client)
Traffic Control Client
Traffic Monitor
Traffic Ops
Traffic Portal
Traffic Router
Traffic Stats
Grove
CDN in a Box
Automation
What is the best way to verify this PR?
Run integration, unit test.
Ensure CIAB build correctly
TO API endpoints work as expected.
Clients (eg: T3C) work as expected.
If this is a bugfix, which Traffic Control versions contained the bug?
PR submission checklist
[x] This PR has tests
[x] This PR has documentation
[x] This PR has a CHANGELOG.md entry
[x] This PR DOES NOT FIX A SERIOUS SECURITY VULNERABILITY (see the Apache Software Foundation's security guidelines for details)
This PR reverts #7068
|
gharchive/pull-request
| 2022-11-14T18:15:54 |
2025-04-01T04:56:00.482504
|
{
"authors": [
"rimashah25",
"zrhoffman"
],
"repo": "apache/trafficcontrol",
"url": "https://github.com/apache/trafficcontrol/pull/7190",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1569966862
|
Still load legacy config if it exists instead of failing
Everything is broken and I'm not prepared for records.yaml
Thanks, @cmcfarlen - swapping from a format that's been in use for years to yaml necessitates a change of a ton of tooling. The prior change would have been breaking for just about every org that uses ATS, with a significant investment in trying to change every part of build, validation, and config management. Even if you believe that swapping to yaml would be better, you cannot expect that to be a minor change.
We have already talked about this on slack but just for the record:
As discussed(dev and user mailing list) before, we agreed to have single parsing and fail if the legacy file is found. As also requested, a tool is provided for the translation with a guide on how to move on.
I have no issues with having it dual but just making sure that we are all on the same page.
If there is anything I can do to make the process easy just let me know and I'll do it.
I've created an issue which I will be implementing to track a solution for this.
https://github.com/apache/trafficserver/issues/9384
We talked about this in the PR/issue scrub and would like to have a discussion at next weeks PR/issue scrub. A final decision should go to the mailing list. Marking it as a draft, so it doesn't get merged in before we discuss.
I'm going to close this and we can work on the yaml support around tooling instead.
|
gharchive/pull-request
| 2023-02-03T14:54:12 |
2025-04-01T04:56:00.485791
|
{
"authors": [
"DoingTheNeedful",
"brbzull0",
"bryancall",
"cmcfarlen"
],
"repo": "apache/trafficserver",
"url": "https://github.com/apache/trafficserver/pull/9377",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1174113276
|
[skip ci] Skip test_conv2d in Hexagon
https://github.com/apache/tvm/pull/10666 marked this xfail but it still runs the test causing segfault and CI failure in PR jobs and main.
@driazati @areusch @junrushao1994
But this test already has
https://github.com/apache/tvm/blob/871a83f9719d74d2eb15786d8b8878229b194c53/tests/python/contrib/test_hexagon/test_conv2d_conv2d.py#L165-L167
So I wonder why i386 test is still failing...
Existing
@pytest.mark.skipif(
platform.processor() == "i686", reason="Test known to be flaky on i386 machines"
)
doesn't seem to help actually skipping test. I'm trying if platform.processor() == "i386" does make things skip.
As the log https://ci.tlcpack.ai/blue/rest/organizations/jenkins/pipelines/tvm/branches/PR-10687/runs/7/nodes/315/steps/583/log/?start=0 shows, platform.machine() prints x86_64 despite being supposed to run on the i386 image. I don't know have a good way to conditionally skip this test on i386, so I'm going to skip it completely.
|
gharchive/pull-request
| 2022-03-18T23:48:48 |
2025-04-01T04:56:00.489498
|
{
"authors": [
"masahi"
],
"repo": "apache/tvm",
"url": "https://github.com/apache/tvm/pull/10687",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1625007663
|
[Fix][TIR] Fix tvm::arith::UnionLowerBound
The UnionLowerBound function does not take into account the condition that the empty set has a special representation [+inf, -inf].
cc @wrongtest-intellif
cc @wrongtest-intellif
Could you please add a regression test?
We could add a testcase to tests/python/unittest/test_arith_intset.py::test_union_lower_bound
Could you please add a regression test?
Already added, thanks for your suggestion.
We could add a testcase to tests/python/unittest/test_arith_intset.py::test_union_lower_bound
Already added, thanks for your suggestion
@tvm-bot rerun
Please make the CI green and we can get it in :-)
Hi, CI is finished.
|
gharchive/pull-request
| 2023-03-15T08:30:25 |
2025-04-01T04:56:00.493334
|
{
"authors": [
"Biubiubiu12",
"Hzfengsy",
"wrongtest-intellif"
],
"repo": "apache/tvm",
"url": "https://github.com/apache/tvm/pull/14304",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2264251604
|
[YUNIKORN-2587] Core: Convert AllocationID to AllocationKey
What is this PR for?
Having both AllocationID and AllocationKey in the core is redundant. Replace all references to AllocationID with AllocationKey and remove the duplicate variable in Allocation.
What type of PR is it?
[ ] - Bug Fix
[ ] - Improvement
[ ] - Feature
[ ] - Documentation
[ ] - Hot Fix
[ ] - Refactoring
Todos
[x] - Task
What is the Jira issue?
https://issues.apache.org/jira/browse/YUNIKORN-2587
How should this be tested?
Unit tests updated as necessary.
Screenshots (if appropriate)
Questions:
[ ] - The licenses files need update.
[ ] - There is breaking changes for older versions.
[ ] - It needs documentation.
Codecov Report
Attention: Patch coverage is 79.16667% with 15 lines in your changes are missing coverage. Please review.
:exclamation: No coverage uploaded for pull request base (master@e17eafa). Click here to learn what that means.
Files
Patch %
Lines
pkg/scheduler/partition.go
75.75%
7 Missing and 1 partial :warning:
pkg/examples/simple_example.go
0.00%
2 Missing :warning:
pkg/scheduler/objects/application.go
83.33%
2 Missing :warning:
pkg/scheduler/objects/node.go
86.66%
2 Missing :warning:
pkg/scheduler/context.go
0.00%
1 Missing :warning:
Additional details and impacted files
@@ Coverage Diff @@
## master #856 +/- ##
=========================================
Coverage ? 77.02%
=========================================
Files ? 97
Lines ? 12002
Branches ? 0
=========================================
Hits ? 9244
Misses ? 2421
Partials ? 337
:umbrella: View full report in Codecov by Sentry.
:loudspeaker: Have feedback on the report? Share it here.
|
gharchive/pull-request
| 2024-04-25T18:34:02 |
2025-04-01T04:56:00.506783
|
{
"authors": [
"codecov-commenter",
"craigcondit"
],
"repo": "apache/yunikorn-core",
"url": "https://github.com/apache/yunikorn-core/pull/856",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
165943464
|
[ZEPPELIN-1196] Fix for bug ZEPPELIN-1196
What is this PR for?
Fixed issue related to connecting to remote running interpreter process with multiple interpreters in interpreter group throws illegal thread state exception
What type of PR is it?
Bug Fix
Todos
What is the Jira issue?
[ZEPPELIN-1196] https://issues.apache.org/jira/browse/ZEPPELIN-1196
Questions:
Does the licenses files need update? No
Is there breaking changes for older versions? No
Does this needs documentation? No
One of the travis check is failing but its not related to my changes
Can someone please review this PR
@SachinJanani Hi, can you add a test case for it? It would help understand and test this PR.
@jongyoul Thanks for reviewing.As suggested I have added a test case for this PR
@SachinJanani Thanks for the quick fix, but if it fails to launch RemoteInterpreterServer with 10 seconds, this test always fails, isn't it? how to deal with that case? I think we need to make it more concrete. How about you?
@jongyoul Yes you are right if RemoteInterpreterServer fails to launch within 10 seconds then this test will fail,but we need to have some threshold for that because if say there is some bug in RemoteInterpreterServer which causes it to hang then our test will never complete.Should I increase it to 20 seconds.What you think?
@jongyoul One more thing I found that as per ZEPPELIN_INTERPRETER_CONNECT_TIMEOUT property default timeout that we consider is 30 seconds.So I think setting it to 30 seconds will make more sense.
@SachinJanani 30 seconds would be realistic. And could you please leave memo that it may occurs a potential bug when it fails launching it?
Thanks @jongyoul. Sure will make the change also will add a comment in the code about about the 30 seconds timeout
LGTM. Merging if there's no more discussion.
Can someone please merge this PR as I don't have permission
I'll merge it after 24 hours.
@SachinJanani I tried to merge it but this is based on branch-0.6. Could you please rebase from current master? I've tried to do it but I occurs error about merge conflict. Can you handle it? And for the next time, you'd better contribute the codes to avoid this situation.
@jongyoul This changes are only for branch-0.6 as master does not have this files (even this code) anymore due to helium changes.Also i think issue might not occur for zeppelin-0.7.So this is fix will only be for zeppelin-0.6.1.I think it should be only be merged in branch-0.6.
@jongyoul @SachinJanani I managed it to be merged only to branch-0.6. Will close this PR manually.
Thanks @minahlee
|
gharchive/pull-request
| 2016-07-16T19:15:41 |
2025-04-01T04:56:00.516235
|
{
"authors": [
"SachinJanani",
"jongyoul",
"minahlee"
],
"repo": "apache/zeppelin",
"url": "https://github.com/apache/zeppelin/pull/1197",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
410195180
|
Finished genReadOp() description. #20
Answered all the questions of part 1 of the assignment.
Should fix #20.
Codecov Report
Merging #43 into lab3 will not change coverage.
The diff coverage is n/a.
@@ Coverage Diff @@
## lab3 #43 +/- ##
=======================================
Coverage 68.31% 68.31%
=======================================
Files 107 107
Lines 7334 7334
Branches 1388 1388
=======================================
Hits 5010 5010
Misses 1868 1868
Partials 456 456
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update a8c3c77...82b8b44. Read the comment docs.
Codecov Report
Merging #43 into lab3 will not change coverage.
The diff coverage is n/a.
@@ Coverage Diff @@
## lab3 #43 +/- ##
=======================================
Coverage 68.31% 68.31%
=======================================
Files 107 107
Lines 7334 7334
Branches 1388 1388
=======================================
Hits 5010 5010
Misses 1868 1868
Partials 456 456
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update a8c3c77...82b8b44. Read the comment docs.
|
gharchive/pull-request
| 2019-02-14T09:03:35 |
2025-04-01T04:56:00.549293
|
{
"authors": [
"codecov-io",
"emilelmarsson"
],
"repo": "apeinot/java",
"url": "https://github.com/apeinot/java/pull/43",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
470912710
|
Fix saving uuid as item title
This bug happens in some cases when hitting Save button the XML of an item is still not available .
Not sure this is the correct fix. I mean, it may work, but I think the problem comes from the XML doc being out of date because the client script getState promise only runs once.
There is also still the problem of validation hooks not being unloaded, but that's a different issue :)
Not sure this is the correct fix. I mean, it may work, but I think the problem comes from the XML doc being out of date because the client script getState promise only runs once.
There is also still the problem of validation hooks not being unloaded, but that's a different issue :)
Yes, getState promise only runs once. So I try to get it running again inside the "presubmit". And then I found the real reason is the item's XML is not really ready when calling getState.
If you go to line 106 of WizardApi, you would see 'itemPack.getXml ' actually does not have the XML we need yet.
Something I would test is:
Put a repeater on the first page and add some repeater items.
Now remove a repeater item.
Now save + publish.
Check the item you removed is not in the XML
And:
Put some optional controls on page 2
Go to page 2 and fill in some values
Go back to page 1
Go back to page 2 and put in some new values
Now save + publish.
Check that the first set of values you put on page 2 are not in the XML.
Something I would test is:
Put a repeater on the first page and add some repeater items.
Now remove a repeater item.
Now save + publish.
Check the item you removed is not in the XML
And:
Put some optional controls on page 2
Go to page 2 and fill in some values
Go back to page 1
Go back to page 2 and put in some new values
Now save + publish.
Check that the first set of values you put on page 2 are not in the XML.
Thanks Aaron. Let me test these two cases now.
For the second test, looks good.
For the Repeater test :
|
gharchive/pull-request
| 2019-07-22T06:18:32 |
2025-04-01T04:56:00.559078
|
{
"authors": [
"PenghaiZhang",
"abidingotter"
],
"repo": "apereo/openEQUELLA",
"url": "https://github.com/apereo/openEQUELLA/pull/1127",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1879118953
|
hysteria2 ERROR info in openwrt
1.system:
run hysteria2 in OpenWrt with transparent proxy
2.client config
{
"tls": {
"insecure": false,
"sni": "mysite.com"
},
"lazy": true,
"tcpTProxy": {
"listen": "0.0.0.0:1234"
},
"socks5":{
"listen": "0.0.0.0:1080"
},
"fast_open": true,
"bandwidth": {
"down": "300 mbps",
"up": "100 mbps"
},
"disable_mtu_discovery": false,
"auth": "Strong_password",
"server": "mysite.com:443"
}
3.ERROR info
2023-09-03T14:11:20Z INFO TCP transparent proxy listening {"addr": "0.0.0.0:1234"}
2023-09-03T14:11:23Z INFO connected to server {"count": 1}
2023-09-03T14:11:23Z ERROR TCP transparent proxy error {"addr": "123.96.20.227:40526", "reqAddr": "127.0.0.1:1234", "error": "dial error: dial tcp4 127.0.0.1: 1234: connect: connection refused"}
2023-09-03T14:11:23Z ERROR TCP transparent proxy error {"addr": "123.96.20.227:40534", "reqAddr": "127.0.0.1:1234", "error": "dial error: dial tcp4 127.0.0.1: 1234: connect: connection refused"}
2023-09-03T14:11:23Z ERROR TCP transparent proxy error {"addr": "123.96.20.227:40542", "reqAddr": "127.0.0.1:1234", "error": "dial error: dial tcp4 127.0.0.1: 1234: connect: connection refused"}
2023-09-03T14:11:24Z ERROR TCP transparent proxy error {"addr": "123.96.20.227:40558", "reqAddr": "127.0.0.1:1234", "error": "dial error: dial tcp4 127.0.0.1: 1234: connect: connection refused"}
server's ACL causes the issue.solved
服务器的 ACL 导致该问题。已解决
你是怎么解决的?我也有同样的错误~
|
gharchive/issue
| 2023-09-03T14:28:01 |
2025-04-01T04:56:00.591176
|
{
"authors": [
"AmberisMyShiba",
"sjkhsl"
],
"repo": "apernet/hysteria",
"url": "https://github.com/apernet/hysteria/issues/656",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
835756783
|
Not able to use toggleSeries with next js application
I want to toggleSeries data of a chart in my next js application. I am trying to use that method in 2 different ways, but none of them is working. I will show you the error I am getting with both these methods differently. Please help me with a solution.
Method 1:
import ApexCharts from 'apexcharts';
toggleSeries = (seriesName) => {
ApexCharts.exec('myChartId', 'toggleSeries', 'seriesName');
}
In method 1 there are 2 problems
Most of the time I get window is not defined for ApexCharts
Second, sometimes I get: apexcharts__WEBPACK_IMPORTED_MODULE_9___default.a.exec is not a function
In order to resolve error: window is not defined I use dynamic import as I use dynamic import for react-apexcharts see below, even checking from typeof window !== undefined is not working here
Method 2:
const ApexCharts = dynamic(() => import('apexcharts'), { ssr: false });
toggleSeries = (seriesName) => {
ApexCharts.exec('myChartId', 'toggleSeries', 'seriesName');
}
In method 2 it says ApexCharts.exec is not a function.
Please help me with the solution for this
Basically exec function is coming undefined
Can anybody help ??
|
gharchive/issue
| 2021-03-19T09:34:31 |
2025-04-01T04:56:00.626690
|
{
"authors": [
"hbole"
],
"repo": "apexcharts/apexcharts.js",
"url": "https://github.com/apexcharts/apexcharts.js/issues/2299",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
451440384
|
Does apexcharts support this kind of chart, see image in the link?
Does apexcharts support this kind of chart, see image in the link, I'm more interested on the top green label (+$6.31) (need to show the difference in amount between this year and last year in numbers)
https://imgur.com/a/moLi1Q4
No, ApexCharts directly doesn't support this functionality.
Although, you can use annotations and do the calculations to put such custom label anywhere on the chart.
https://apexcharts.com/javascript-chart-demos/line-charts/line-chart-annotations/
Docs
|
gharchive/issue
| 2019-06-03T11:52:58 |
2025-04-01T04:56:00.629280
|
{
"authors": [
"junedchhipa",
"visaru"
],
"repo": "apexcharts/apexcharts.js",
"url": "https://github.com/apexcharts/apexcharts.js/issues/644",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
923992392
|
'sys' is not defined
In linux ubuntu I got the error showed in the image below:
Solution: I fixed it adding the following import line in the beginning of the ETL_coletar_dados_e_gravar_BD.py file:
import sys
Thank you @jeff-pal ! Solved by this commit: https://github.com/aphonsoar/Receita_Federal_do_Brasil_-_Dados_Publicos_CNPJ/commit/95fbb3bbb2d474053793caaa08dc4629d7c28871
|
gharchive/issue
| 2021-06-17T14:28:24 |
2025-04-01T04:56:00.631536
|
{
"authors": [
"aphonsoar",
"jeff-pal"
],
"repo": "aphonsoar/Receita_Federal_do_Brasil_-_Dados_Publicos_CNPJ",
"url": "https://github.com/aphonsoar/Receita_Federal_do_Brasil_-_Dados_Publicos_CNPJ/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2015241145
|
Remove stopDataFetcher
Closes https://github.com/api3dao/airseeker-v2/issues/129
If you remove the stop function you probably don't need the interval stored in state either, but again, super minor
Yeah, I removed that in some later PR but I could have done it here.
|
gharchive/pull-request
| 2023-11-28T20:24:45 |
2025-04-01T04:56:00.647164
|
{
"authors": [
"Siegrift"
],
"repo": "api3dao/airseeker-v2",
"url": "https://github.com/api3dao/airseeker-v2/pull/133",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
281035263
|
Request headers not rendered
This is my code:
### Set/Update customer info [PUT /xxx/customer/{id}]
Update the personal info of the customer. None of the fields are mandatory. The fields not specified will not be overwritten in the database.
+ Parameters
+ id (string) - Id of the customer
+ Request
+ Headers
Authentication: Bearer SomeTokenValue
+ Attributes (object)
+ email (string, optional) - email address of the customer
+ password (string, optional) - email address of the customer
+ firstName (string, optional) - first name
+ lastName (string, optional) - last name
+ phoneNumber (string, optional) - phone number
+ receiveOffers (number, optional) - 1 if the customer wants to receive offers, 0 otherwise
+ Body
{
"email":"example@domain.com",
"firstName":"Jos",
"lastName":"Bos",
"phoneNumber":"+32488697059",
"receiveOffers":1
}
This is what the result looks like
Why is the request header not shown?
The request header is shown inside your code example (on line 7). If you don't want to view a code example you can switch to the default (Raw) to see the raw request information (headers and request body).
But why is it separate from the parameters and the attributes?
That makes no sense to me.
You can also say that the request parameters are shown to inside the code example, but doesn't it make more sense to have all variables grouped together? Why put url parameters at the top, but http headers in another section?
|
gharchive/issue
| 2017-12-11T14:27:55 |
2025-04-01T04:56:00.650168
|
{
"authors": [
"ir-fuel",
"kylef"
],
"repo": "apiaryio/api-blueprint",
"url": "https://github.com/apiaryio/api-blueprint/issues/410",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
219690424
|
Updated NPM_VERSION for nodejs 4 and 6
Please do not merge it yet. I want to properly test it in core app and update readme in coreapp.
@freaz fine by me; if you want to have both, you can add an extra folder to create a new tag (e.g. nodejs/6-npm4-4 or something less ugly...)
No I want have just one version, but the npm update was side effect and I wanted to test it at least locally.
|
gharchive/pull-request
| 2017-04-05T19:30:27 |
2025-04-01T04:56:00.653261
|
{
"authors": [
"freaz",
"nadade"
],
"repo": "apiaryio/docker-base-images",
"url": "https://github.com/apiaryio/docker-base-images/pull/108",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
24336903
|
Report possible keywords
For example in
# Resource [/]
+ Model
+ Body (application/json)
{ ... }
The + Body (application/json) is not recognized as a body (content type should be defined on the model level) and thus the whole list item is considered to be model's description.
Migrated to apiaryio/drafter#645
|
gharchive/issue
| 2013-12-16T09:05:44 |
2025-04-01T04:56:00.658068
|
{
"authors": [
"abtris",
"zdne"
],
"repo": "apiaryio/snowcrash",
"url": "https://github.com/apiaryio/snowcrash/issues/65",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
495411657
|
Integrate adblocker functionality
Interesting tip from HN (for Dashblock):
Maybe you already do it, but I think integrating adblocker functionality when loading JS sites would be desirable to reduce load time. And if ads are what the API user is interested in, perhaps add a flag for whether or not one wants ads to load. Recommendation: https://github.com/cliqz-oss/adblocker Should be the fastest adblocker library (used by Ghostery, Cliqz and Brave)
This could be integrated into Apify.launchPuppeteer() function as useAdBlock: true option.
https://sdk.apify.com/docs/api/apify#module_Apify.launchPuppeteer
Greetings. So the thing would be to implement ad blocker to increase the speed of the scrap/crawl? I could work on this 🙏
Yes exactly, it could boost the speed especially for some websites that are heavy on ads (news sites). But it would be great to first test this assumption. Would you be interested also in trying this out? Use Apify SDK to run scraper with and without ad blocker against some websites?
Sure! I can set up a test and run it to check this first with some timing debug, I'll create it and run it, then attach it here for you to see, thank you 🚀
interesting. I manually block all the common ad networks using blockRequests, this would offload the task to the extension
Makes sense for a lot of users I guess but fyi it's an explicit anti-feature with usecase-killing effect for me. I'd need this off with zero sideeffects on current behavior.
Makes sense for a lot of users I guess but fyi it's an explicit anti-feature with usecase-killing effect for me. I'd need this off with zero sideeffects on current behavior.
In the small POC I proposed a while ago https://github.com/apify/apify-js/pull/600, the feature is completely disabled by default and only does some work when blocking is enabled by the user.
Yeah, sorry @remusao . We still have not figured out if the performance will improve or not. I apologize.
Yeah, sorry @remusao . We still have not figured out if the performance will improve or not. I apologize.
Of course, no worries at all, I just wanted to make clear to @matjaeck that there should be a way to integrate such a feature without any overhead when it's disabled.
|
gharchive/issue
| 2019-09-18T19:25:24 |
2025-04-01T04:56:00.665036
|
{
"authors": [
"Darking360",
"jakubbalada",
"matjaeck",
"mnmkng",
"mtrunkat",
"pocesar",
"remusao"
],
"repo": "apify/apify-js",
"url": "https://github.com/apify/apify-js/issues/456",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1376867102
|
🛑 Harbor is down
In dbf72db, Harbor (https://registry.openstorage.xyz) was down:
HTTP code: 521
Response time: 270 ms
Resolved: Harbor is back up in e07c713.
|
gharchive/issue
| 2022-09-17T20:31:36 |
2025-04-01T04:56:00.687098
|
{
"authors": [
"apinter"
],
"repo": "apinter/OSC-mon",
"url": "https://github.com/apinter/OSC-mon/issues/28",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
451777372
|
optimizations for generated target code
the generated target code is rather verbose right now. for example, here is the spec program and output for the "dangerous" millionaires game:
spec:
/* the classic MPC problem */
a : {A};
c : {C};
c_richer : {(A|C)-> & (A&C)<-};
c_richer := declassify(a <= c, {(A|C)-> & (A&C)<-});
/* the dangerous version: if you have less money, you lose all of it */
if (c_richer) {
a := 0;
} else {
c := 0;
}
generated target code:
MPC({alice,chuck})_1 {
c_1 <- recv chuck;
a_1 <- recv alice;
downgrade_1_1 := (a_1 <= c_1);
send downgrade_1_1 to alice;
send downgrade_1_1 to chuck;
}
alice {
a : {⊥<-};
c_richer : {⊥<-};
send a to MPC({alice,chuck})_1;
downgrade_1_2 <- recv MPC({alice,chuck})_1;
assgn_1_1 := downgrade_1_2;
c_richer := assgn_1_1;
c_richer_1 := c_richer;
guard_1_1 := c_richer_1;
guard_1_4 := guard_1_1;
if (guard_1_4) {
assgn_2_1 := 0;
a := assgn_2_1;
a <- recv chuck;
} else {
assgn_3_1 := 0;
send assgn_3_1 to chuck;
}
}
chuck {
c : {⊥<-};
c_richer : {⊥<-};
send c to MPC({alice,chuck})_1;
downgrade_1_3 <- recv MPC({alice,chuck})_1;
assgn_1_2 := downgrade_1_3;
c_richer := assgn_1_2;
c_richer_2 := c_richer;
guard_1_2 := c_richer_2;
guard_1_3 := guard_1_2;
if (guard_1_3) {
assgn_2_2 := 0;
send assgn_2_2 to alice;
} else {
c <- recv alice;
assgn_3_2 := 0;
c := assgn_3_2;
}
}
as you can see, there are a lot of unnecessary variables. we can run copy propagation, dead code elimination, and constant propagation to remove most of these and make the target a lot readable.
copy prop implemented in https://github.com/apl-cornell/viaduct/commit/347d4cac7362e8f16c3bd132486accea58f27348
constant prop not yet implemented
constant prop implemented in e3ccfd719801c6924b899335289b35280e6d2e6f
|
gharchive/issue
| 2019-06-04T03:12:52 |
2025-04-01T04:56:00.697614
|
{
"authors": [
"rolph-recto"
],
"repo": "apl-cornell/viaduct",
"url": "https://github.com/apl-cornell/viaduct/issues/27",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
305475457
|
Correct license name
Usually "Apache License" means version 1, and according to other files, this library actually uses version 2.
Looks good, thanks.
|
gharchive/pull-request
| 2018-03-15T09:35:42 |
2025-04-01T04:56:00.698827
|
{
"authors": [
"dlackty",
"paolorotolo"
],
"repo": "apl-devs/AppIntro",
"url": "https://github.com/apl-devs/AppIntro/pull/524",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
367906094
|
Deprecating NGRX Cache
If someone is interested to maintain it, please let me know.
The whole idea behind that package and using ngrx with Apollo in general, makes no sense to me.
If anyone has any objections please comment here.
The reason to use ngrx is same as for using any other redux implementation. In particular using ngrx-store with angular-apollomakes sense in case of some part of an app would need to use ngrx/store and to avoid having two different redux implementation in the app (to my understanding that was general idea behind making new apollo client independent on redux implementation).
As for me for now I would like to relay fully on Apollo and hope to have no need to access directly underlying redux implementation.
What having a normalized cache inside of ngrx gives you? Do you use it inside ngrx as part of selectors? Having one single store that is a source of truth makes sense only if one data relies on another, if they are totally independent you can have multiple of them.
Main reason would be the size of an app and avoid pulling in one more dependency (well set of deps)
Also, when you're using ngrx/router-module in combination with apollo graphql that package is rather useful.
I was just exploring this library, here's what I thought the library was doing, maybe that'll be helpful.
Typically, most usages of ngrx store will look something like:
Ask to load some data using an action
Have an effect do the actual loading, and dispatch an action when the data has been loaded
Have a reducer update the store with the loaded data (and more specifically, resolve updating v.s. adding to lists, e.g. using ngrx entity)
Use a selector to get the updated data, if it was updated
Since with Apollo and GraphQL it's possible to identify entities, it would be possible to normalize the returned data in a store any time a query is made. Then I can simply use a selector to get it from the "cache". It would remove a lot of boilerplate for most use cases, and for any use cases where the store has more complex logic, the apollo ngrx store would only be used as a cache.
I agree that using ngrx store as a simple cache does not make a lot of sense by itself, but making it easier to update complex models in a store automatically could be very useful, at least in simpler apps.
This however would require "breaking down" the response in separate Entity lists, and to be useful you'd have to map these entities to a type (e.g. using graphql-code-generator). Then it would look something like:
Same
Same, except my service would use a graphqlquery, AND it would update the apollo store.
No reducer required (except if you need to do something fancy, you can alway use the response and do something with it but it's not necessary)
Use a selector to get the data (and it could be possible to generate those too for most cases, like "selected entry in a list", "all items in a list", etc.)
Hope that helps!
If someone wants to take care of apollo-angular-cache-ngrx then feel free to message me, do pull requests, I will merge all of that, right away.
Since I don't use it, I see no point of me maintaining it.
@kamilkisiela what's your suggested apollo, angular local state management stack?
Just a few thoughts: if you use just apollo client as ngrx replacement, how do you implement effects?
If you use apollo-client and ngrx than you have decent router and effects support but you have to transform the graphql responses into the ngrx store.
Is there any approach I am missing which solves these disadvantages?
IMO you should just use apollo for state management. What would be the case where you may need to access store directly?
effect in Apollo are just GraphQL queries/mutations and if you need i.e. access geolocation then you still can do this with @client resolvers
@ciekawy interesting approach. But isn't the idea of redux/ngrx to separate pure reducers from effects with side effects. With the approach you mentioned, it feels a bit like these two concepts are not clearly separated anymore.
Apollo is supposed to fully take over the state management. I moved to Apollo with similar concerns initially and now I am using only Apollo. Biggest benefits for me:
no more need for verbose boilerplate of reducers and effect. Initially I really liked the concept of redux for state management. Now I love Apollo even more :)
there are various integrations available for non-graphql data sources @client for anything client-side, @rest for REST endpoints you cannot move to GraphQL, also you can define your own as well.
here you can read more
https://blog.apollographql.com/the-future-of-state-management-dd410864cae2
another important value is that Apollo do also data normalization when keeping it in local store
Thx @ciekawy I will dig into that one more time.
Actually I thought that loonajs will be the solution to fill the gap between the redux approach and apollojs/graphql. But atm it seems like that the guild also lost its confidence in the loonajs approach. I would be really interested what the guild (@kamilkisiela, @urigo,...) is suggesting as best approach.
A usecase I have come across is about synchronisation of cache/store. I have a dumb component that emits two events in certain cases on a single browser UI event. The handler for first event mutates the client state, then second handler picks that state and performs server mutation. With ngrx store, the store update by first handler by action dispatch appears synchronous, and the effect for action from second handler gets the updated state. However, with apollo, the client only mutation is not synchronous, possibly because it uses promise api internally. Therefore, the mutation from second handler does not get updated data from cache.
If someone wants to maintain NGRX cache for Apollo I would be happy to see a fork so I could give an access to the package on npm.
|
gharchive/issue
| 2018-10-08T18:35:11 |
2025-04-01T04:56:00.721749
|
{
"authors": [
"ceelian",
"christianrondeau",
"ciekawy",
"kamilkisiela",
"wesselvdv",
"wolfvik"
],
"repo": "apollographql/apollo-angular",
"url": "https://github.com/apollographql/apollo-angular/issues/886",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1502859269
|
Add test for sdl field
This test will check that the sdl field inside the _service object type has the right type (non-nullable string).
Apollo Studio and potentially other Apollo tools check this field to enable/disable federation features.
TODO:
[ ] Add directives
@dariuszkuc I wonder if we should make a new table for directives (which I haven't added just yet), my main issue with this test is that it is hard to see what fails, since it stops at the first failure
@patrick91 if this is still an issue feel free to re-open this PR
|
gharchive/pull-request
| 2022-12-19T12:18:33 |
2025-04-01T04:56:00.736524
|
{
"authors": [
"dariuszkuc",
"patrick91"
],
"repo": "apollographql/apollo-federation-subgraph-compatibility",
"url": "https://github.com/apollographql/apollo-federation-subgraph-compatibility/pull/302",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
328946917
|
Can schema stitching be implemented in Python ?
Sorry, that wouldn't be in this library.
|
gharchive/issue
| 2018-06-04T07:58:41 |
2025-04-01T04:56:00.752223
|
{
"authors": [
"s1h1ikha",
"stubailo"
],
"repo": "apollographql/graphql-tools",
"url": "https://github.com/apollographql/graphql-tools/issues/831",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
193114768
|
make afterware errors bubble up
This makes errors thrown from afterwares bubble up so that apollo don't try to parse bad responses.
TODO:
[ ] If this PR is a new feature, reference an issue where a consensus about the design was reached (not necessary for small changes)
[ ] Make sure all of the significant new logic is covered by tests
[x] Rebase your changes on master so that they can be merged easily
[ ] Make sure all tests and linter rules pass
[ ] Update CHANGELOG.md with your change
[x] Add your name and email to the AUTHORS file (optional)
[ ] If this was a change that affects the external API, update the docs and post a link to the PR in the discussion
@edvinerikson Thanks for the PR! Can you explain what problem this solves and add a test case if applicable? Right now it doesn't look like your PR would change any behavior.
Looks like the change makes it so that the promise returns the return value of the afterware. However this should definitely come with a test that ensures the behavior continues to work, especially since it's so easy to miss.
Yeah I will add a test. I just need to fix my local setup first. Can't run
it locally due to TS errors.
On Fri, 2 Dec 2016 at 23:55, Sashko Stubailo notifications@github.com
wrote:
Looks like the change makes it so that the promise returns the return
value of the afterware. However this should definitely come with a test
that ensures the behavior continues to work, especially since it's so easy
to miss.
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
https://github.com/apollostack/apollo-client/pull/982#issuecomment-264584646,
or mute the thread
https://github.com/notifications/unsubscribe-auth/ADtdjIkVxD0gvTIVN8hR7lwS3Y2vvHcNks5rEKHHgaJpZM4LCi3H
.
@helfer the reason for the change is because I have a afterware which I check the response status and decide if Apollo will be able to parse the response. in some cases Apollo won't be able to do that and when that happens I throw a error which I can catch in my error handling but without this change I am not able to catch the error and instead Apollo will fail with when trying to parse the json body.
e.g getDataFromTree(app).catch(error => doSomethingWithErrorHere()) but currently the error I get there will be something like unexpected token < at 1:1 (due to a html response) instead of my own error.
maybe a improvement to the network layer would be to check response.ok before trying to parse / handle the response as well. will probably help fix some of the cryptic error messages that I get sometimes.
@edvinerikson Okay. Please add a test, and then we can merge it.
I think checking response.ok won't work, because some people return error codes that are not in the 200 range for responses that should be parsed and processed. It's unfortunate, but I think we have to live with that for now.
I think checking response.ok won't work, because some people return error codes that are not in the 200 range for responses that should be parsed and processed. It's unfortunate, but I think we have to live with that for now.
Yeah I will create my own network layer to handle my cases instead.
Still gonna try to add a test and get this merged.
test is added
Sweet, thanks a lot!
|
gharchive/pull-request
| 2016-12-02T13:21:12 |
2025-04-01T04:56:00.788120
|
{
"authors": [
"edvinerikson",
"helfer",
"stubailo"
],
"repo": "apollostack/apollo-client",
"url": "https://github.com/apollostack/apollo-client/pull/982",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1781998268
|
Add Whatsapp Icon
Description
Add a WhatsApp icon to connect with the admin on WhatsApp
Screenshots
No response
Additional information
No response
@apoorvaron Kindly assign me this issue
@Nishitbaria assigned
Now start solving issues😂
@apoorvaron I have done this one already i was just uploading video
https://github.com/apoorvaron/Shorty/assets/85815172/34d7e252-f599-4aed-a0cc-082196ee704e
@apoorvaron Since its Friday so last working day in my job and very less load so trying to create issue and finding some raw solution then full working from tomorrow
https://github.com/apoorvaron/Shorty/assets/85815172/c4a3c7ff-9f81-462d-85c5-ae6009b4aa65
@apoorvaron Thanks Man
|
gharchive/issue
| 2023-06-30T06:52:46 |
2025-04-01T04:56:00.796114
|
{
"authors": [
"Nishitbaria",
"apoorvaron"
],
"repo": "apoorvaron/Shorty",
"url": "https://github.com/apoorvaron/Shorty/issues/190",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1051664363
|
feat: apply stopwords by language
What does this do / why do we need it?
https://www.notion.so/appbase/Arc-Suggestions-API-enhancements-2a4c26df77d44a77b751d9e13cf4cb0f
What should your reviewer look out for in this PR?
Which issue(s) does this PR fix?
If this PR affects any API reference documentation, please share the updated endpoint references
@siddharthlatest Docs PR https://github.com/appbaseio/Docs/pull/246
If searchLanguage isn't explicitly passed in RS API, will this take the search relevance's language for stopwords? It should be this way.
Yes, review this PR https://github.com/appbaseio-confidential/arc-noss/pull/230
|
gharchive/pull-request
| 2021-11-12T07:32:38 |
2025-04-01T04:56:00.811613
|
{
"authors": [
"bietkul"
],
"repo": "appbaseio/reactivesearch-api",
"url": "https://github.com/appbaseio/reactivesearch-api/pull/213",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
173249897
|
Kafka is not reachable from amplifier
Amplifier is connecting to kafka through localhost:9092.
However, this hostname is used only for the first connection to kafka. On this first call, kafka gives its own hostname to the client (amplifier).
The problem is that the hostname given by kafka (which is actually the kafka container short id) is not reachable from amplifier because it's not part of amp-swarm network.
Possible fixes:
Running amplifier as a container and make it part of the amp-swarm network using docker run , however services and containers cannot be part of the same network.
docker run --network amp-swarm appcelerator/amp
docker: Error response from daemon: swarm-scoped network (amp-swarm) is not compatible with `docker create` or `docker run`. This network can only be used by a docker service.
See 'docker run --help'.
Giving the service a hostname, however, I don't think it's possible:
docker service create --help
Usage: docker service create [OPTIONS] IMAGE [COMMAND] [ARG...]
Create a new service
Options:
--constraint value Placement constraints (default [])
--container-label value Container labels (default [])
--endpoint-mode string Endpoint mode (vip or dnsrr)
-e, --env value Set environment variables (default [])
--help Print usage
-l, --label value Service labels (default [])
--limit-cpu value Limit CPUs (default 0.000)
--limit-memory value Limit Memory (default 0 B)
--log-driver string Logging driver for service
--log-opt value Logging driver options (default [])
--mode string Service mode (replicated or global) (default "replicated")
--mount value Attach a mount to the service
--name string Service name
--network value Network attachments (default [])
-p, --publish value Publish a port as a node port (default [])
--replicas value Number of tasks (default none)
--reserve-cpu value Reserve CPUs (default 0.000)
--reserve-memory value Reserve Memory (default 0 B)
--restart-condition string Restart when condition is met (none, on-failure, or any)
--restart-delay value Delay between restart attempts (default none)
--restart-max-attempts value Maximum number of restarts before giving up (default none)
--restart-window value Window used to evaluate the restart policy (default none)
--stop-grace-period value Time to wait before force killing a container (default none)
--update-delay duration Delay between updates
--update-failure-action string Action on update failure (pause|continue) (default "pause")
--update-parallelism uint Maximum number of tasks updated simultaneously (0 to update all at once) (default 1)
-u, --user string Username or UID
--with-registry-auth Send registry authentication details to swarm agents
-w, --workdir string Working directory inside the container
Running amplifier as a service: It's definitely possible, however, it's not compatible with development where you need to be able to quickly launch, fix, and repeat. Indeed, docker services are not using local images but instead pull images from the hub (or a private registry) on each start (check this issue https://github.com/docker/swarmkit/issues/1246). It would require to push, build and publish the image every time which defeats the purpose.
The only workaround I found so far is to add kafka container hostname to /etc/host:
Getting kafka hostname:
docker ps
docker inspect kafka_container_id | grep hostname
sudo cat /var/lib/docker/containers/a4424bcafbab110b67922994243aca15fc57ab3422fdec21f7460fe16b024456/hostname
Update /etc/hosts in order to assign 127.0.0.1 to kafka hostname:
127.0.0.1 localhost a4424bcafbab
Unfortunately the following doesn't work for Docker for Mac:
sudo cat /var/lib/docker/containers/a4424bcafbab110b67922994243aca15fc57ab3422fdec21f7460fe16b024456/hostname
On Docker for Mac these files are stored in the VM located here:
$HOME/Library/Containers/com.docker.docker/Data/com.docker.driver.amd64-linux/Docker.qcow2
But since it looks like the hostname is just the short-id anyway, I think this will work (at least for our temporary workaround right now):
docker ps --filter "name=kafka\." --format '{{ .ID }}'
(The funky filter is to eliminate kafka-manager from the result.)
I'll test that out and report my results back here shortly.
We need a better interim strategy; asking everyone to help brainstorm on this.
It seems like advertised.listeners might be what we're looking for in the Broker Configs docs...
@bertrand-quenin @ndegory Can you two investigate further?
Regarding my last comment, I don't think the issue is advertising the hostname, it's making sure the hostname is available outside the swarm's network when we're developing ...
I suppose this is where kafka updates and then gets the server address for subsequent requests on the client broker:
https://github.com/Shopify/sarama/blob/master/broker.go#L83
https://github.com/Shopify/sarama/blob/master/broker.go#L161
@subfuzion Actually, I already tried this solution. It effectively makes kafka reachable from amplifier by advertising localhost. The problem is that other components (telegraf, amp-log-worker) are trying to reach kafka on localhost as well ... making them fail miserably.
@bertrand-quenin Unfortunately, that makes perfect sense.
We may have to create a script that can take the output of docker ps --filter "name=kafka\." --format '{{ .ID }}' and update the entry for 127.0.0.1 in /etc/hosts. We can call the script from the swarm script when we start up the swarm. @ndegory, we're all slammed preparing for our deadline on Monday; is there any way you can take this on for us?
I'll take a look
@bertrand-quenin Now that I think about it, can't we override the address in the client after it gets updated with the server's metadata, just to ensure we're using the address we want to use? After all, in the real swarm we would want to use the actual DNS name (kafka) anyway since swarm is load balancing for us. There isn't some special requirement for affinity to a particular server in the cluster, is there?
Thanks for the PR, @ndegory. It's working well for now. I'll leave this issue open while we consider our long term strategy for this and similar/related issues.
There is another solution which do not imply modify /etc/host.
If we have amplifier inside swarm network it works as explains above, but then the issue that we can't build locally an amplifier image to test.
Except, for the test need, we modify the amplifier image name 'appcelerator/amp' to 'appcelerator/amp_fr (for instance for me) and modify the swarm file to point on this image name instead of the normal.
Then, all work, docker see that appcelerator/amp_fr doesn't exist in repo and use the local image
@bertrand-quenin Can you respond to https://github.com/appcelerator/amp/issues/69#issuecomment-242868643?
Yes, it's an alternate solution to this issue. It can work as well.
@freignat91 @bertrand-quenin @ndegory Since I prefer that we switch to NATS, I'm not particularly concerned about the long term resolution for this particular issue at this point. However, it is definitely the case that amplifier will eventually expose its services on an overlay network (for example, amp-swarm-public), and it will also overlap with the backbone network (amp-swarm) to communicate with our infrastructure services. @freignat91, we should start moving toward this configuration this sprint, and also start incorporating haproxy for mapping exposed IPs to consumers (eg, the CLI).
It's interesting to be aware of how clusters and node discovery impact other services for client communication as well. For example, the etcd client when etcd is behind a load-balanced cluster: https://github.com/hashicorp/vault/pull/921.
Since we have a working solution in place for now thanks to @ndegory and our priority has shifted to transitioning to NATS (#113), I'm going to close this.
|
gharchive/issue
| 2016-08-25T16:27:14 |
2025-04-01T04:56:00.827026
|
{
"authors": [
"bertrand-quenin",
"freignat91",
"ndegory",
"subfuzion"
],
"repo": "appcelerator/amp",
"url": "https://github.com/appcelerator/amp/issues/69",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
177344009
|
Default retention policy for InfluxDB 1.0.0
Hi, I found an issue and came up with a way to fix it.
The issue: InfluxDB changed the name of the default retention policy in 1.0.0 from "default" to "autogen". The telegraf.conf.tpl file needs to be updated.
Proposed fix: To make this work and be backwards compatible I created an env variable for the policy name. I have it defaulting to "default". It could default to "", which is what the telegraf 1.0.0 telegraf.conf file has.
To make it clear how to use it I added it to the custom example at the top of the readme.
My pleasure.
|
gharchive/pull-request
| 2016-09-16T04:41:14 |
2025-04-01T04:56:00.830089
|
{
"authors": [
"atribe"
],
"repo": "appcelerator/docker-telegraf",
"url": "https://github.com/appcelerator/docker-telegraf/pull/16",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
294630606
|
How to test that a button is disabled
On Android, I have disabled a button being able to get clicked until a certain condition is met within the input.
(Python client)
The object currently has clickable="false" in the XML but the appium driver tells me different, because - button.is_enabled() returns True
In the case, you can get an arbitrary element with el.get_attribute('clickable').
https://github.com/appium/python-client/blob/f37733dbd3dd79be97452a9ad98eb4a9325755ce/test/functional/ios/find_by_ios_class_chain_tests.py#L32
If your xml says the element is disable, you can see false by el.get_attribute('clickable').
BTW, enable and clickable are https://stackoverflow.com/questions/15615823/setenabled-vs-setclickable-what-is-the-difference
If you have any issue, the page source and logs are helpful for us.
|
gharchive/issue
| 2018-02-06T04:42:56 |
2025-04-01T04:56:00.969169
|
{
"authors": [
"Enether",
"KazuCocoa"
],
"repo": "appium/sample-code",
"url": "https://github.com/appium/sample-code/issues/149",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2303061671
|
Generalize the removal of compiler argument options during indexing
The existing ad-hoc logic was not quite correct because it didn’t eg. remove -MT/depfile because it assumed that -MT was followed by a space. It also didn’t take into account that serialize-diagnostics can be spelled with a single dash or two dashes.
Create a CompilerCommandLineOption type that forces decisions to be made about the dash spelling and argument styles, which should help avoid problems like this in the future.
@swift-ci Please test
@swift-ci Please test
@swift-ci Please test Windows
@swift-ci Please test Windows
@swift-ci Please test Windows
|
gharchive/pull-request
| 2024-05-17T15:46:28 |
2025-04-01T04:56:01.018500
|
{
"authors": [
"ahoppen"
],
"repo": "apple/sourcekit-lsp",
"url": "https://github.com/apple/sourcekit-lsp/pull/1314",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
593757420
|
TestURLSession: Re-enable tests
Re-enable the TestURLSession tests which should no longer be flaky
however specifc tests are still disabled due to known issues.
Disable test_concurrentRequests() as this fails about 4% of the time.
Disable test_getAllTasks() as this causes other tests to fail.
@swift-ci test linux
@swift-ci test linux
@swift-ci test linux
|
gharchive/pull-request
| 2020-04-04T06:34:03 |
2025-04-01T04:56:01.020835
|
{
"authors": [
"spevans"
],
"repo": "apple/swift-corelibs-foundation",
"url": "https://github.com/apple/swift-corelibs-foundation/pull/2757",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1260346553
|
Convert DowningStrategyShell into a distributed actor
Resolves https://github.com/apple/swift-distributed-actors/issues/935
Only one thing about the logger: https://github.com/apple/swift-distributed-actors/pull/938#discussion_r889430590 otherwise ready to go, thank you!
|
gharchive/pull-request
| 2022-06-03T20:29:02 |
2025-04-01T04:56:01.022376
|
{
"authors": [
"ktoso",
"yim-lee"
],
"repo": "apple/swift-distributed-actors",
"url": "https://github.com/apple/swift-distributed-actors/pull/938",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1648317524
|
Prevent scrolling on navigation event if router meta field preventScrolling is true
Bug/issue #, if applicable: rdar://103146273
Summary
If preventScrolling in the meta field is set to true, then we should not change the scroller location when navigating to a new page.
Dependencies
rdar://103119754
Testing
Steps:
set preventScrolling to true for a page (i.e. '/foo')
Scroll down a little bit on any page (i.e. '/bar'). ('/foo' should be longer than '/bar')
trigger a navigation event to '/foo'
Verify that the scroller did not move up/down nor left/right
Checklist
Make sure you check off the following items. If they cannot be completed, provide a reason.
[x] Added tests
[x] Ran npm test, and it succeeded
[ ] Updated documentation if necessary - NA
@swift-ci test
|
gharchive/pull-request
| 2023-03-30T21:03:49 |
2025-04-01T04:56:01.026686
|
{
"authors": [
"hqhhuang"
],
"repo": "apple/swift-docc-render",
"url": "https://github.com/apple/swift-docc-render/pull/563",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
786444435
|
[Collections] add should only add to config if fetch is successful
Motivation:
The code currently adds a collection source to the config even when initial fetch fails.
Modifications:
We should fetch collection first and make sure that's successful before adding to config.
Result:
Bad sources are not added to the config.
@swift-ci please smoke test
@swift-ci please smoke test
@swift-ci please smoke test
@swift-ci please smoke test
@swift-ci please smoke test
@swift-ci please smoke test
@swift-ci please smoke test
@swift-ci please smoke test
|
gharchive/pull-request
| 2021-01-15T00:41:58 |
2025-04-01T04:56:01.029839
|
{
"authors": [
"abertelrud",
"yim-lee"
],
"repo": "apple/swift-package-manager",
"url": "https://github.com/apple/swift-package-manager/pull/3195",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1518715665
|
Remove reference to AsTypePattern
Companion of https://github.com/apple/swift-syntax/pull/1178.
https://github.com/apple/swift-syntax/pull/1178
@swift-ci Please test
|
gharchive/pull-request
| 2023-01-04T10:26:23 |
2025-04-01T04:56:01.031556
|
{
"authors": [
"ahoppen"
],
"repo": "apple/swift-stress-tester",
"url": "https://github.com/apple/swift-stress-tester/pull/219",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
506920895
|
[SyntaxParse] Revert syntax parsing changes from the master branch
Update for https://github.com/apple/swift/pull/27675
@swift-ci Please test
|
gharchive/pull-request
| 2019-10-14T23:25:21 |
2025-04-01T04:56:01.032635
|
{
"authors": [
"rintaro"
],
"repo": "apple/swift-syntax",
"url": "https://github.com/apple/swift-syntax/pull/156",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
358221711
|
Fix RawSyntax's leading/trailingTrivia computation to return nil if the outermost children don't have leading/trailing trivia
Previously they would skip over outer children that didn't have leading/trailing trivia until they found one that did. This was causing those trivia to be included twice when computing the totalLength/byteSize of the node; once in leading/trailingTrivia and again in contentLength, which only skipped the trivia of the outermost children.
@swift-ci please test
Good catch. Thanks. We should also cherry-pick this over to swift-4.2-branch once #6 is merged.
|
gharchive/pull-request
| 2018-09-07T21:48:49 |
2025-04-01T04:56:01.034343
|
{
"authors": [
"ahoppen",
"nathawes"
],
"repo": "apple/swift-syntax",
"url": "https://github.com/apple/swift-syntax/pull/8",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
675802806
|
build: ignore deprecation warnings on Windows
Adjust the compile flags for TSCBasic on Windows. Avoid the separate
checks and addition of flags in favour of generator expressions.
CC: @tomerd @neonichu @abertelrud
@swift-ci please test
do we have windows CI for this?
Not yet, but the nighters do build t-s-c, as we bring up more pieces we can enable them.
e.g. https://dev.azure.com/compnerd/swift-build/_build/results?buildId=35571&view=logs&s=4a582161-ef84-5ffa-1ff0-506a7473f086&j=73ce2acc-4df7-56ec-1894-6d2e45aae46f
|
gharchive/pull-request
| 2020-08-10T00:30:34 |
2025-04-01T04:56:01.036788
|
{
"authors": [
"compnerd",
"tomerd"
],
"repo": "apple/swift-tools-support-core",
"url": "https://github.com/apple/swift-tools-support-core/pull/102",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1214380541
|
[SR-15154] ObjC projects using SPM fail when we expose our models into another module
Previous ID
SR-15154
Radar
rdar://problem/82827693
Original Reporter
taquitos (JIRA User)
Type
Bug
Attachment: Download
Additional Detail from JIRA
Votes
3
Component/s
Labels
Bug
Assignee
None
Priority
Medium
md5: 4659038e109b6bed99ec521e272addf0
Issue Description:
Setup:
We create an ObjC App that uses our SPM package "RevenueCat".
In our app, we create an object that uses an object from the "RevenueCat" package as part of its api.
We expose this object's header (which exposes the method containing the RevenueCat object as a parameter) as a public header.
We create a Swift test target that depends on the ObjC App and has a bridging header for that object we created.
We attempt to use the API in a test. API that doesn't include RevenueCat objects is fine, but any API that uses RevenueCat objects in the parameter is not found.
Environment:
Xcode 12.5.1 (Build version 12E507) & Xcode 13 Beta 5
M1 Mac as well as Intel mac.
Swift Package Manager - Swift 5.4.0
Swift Package Manager - Swift 5.5.0
Component: Package Manager
Reproduction project: https://github.com/taquitos/SPMBug
Attempt to build the test target.
You'll get
`Value of type 'TaquitosAreTheBest' has no member 'purchaserInfo'`
I have faced similar issue. This bug is also reproduces in mixed Swift/Objective-C project (not only in Test target).
Case: header file has forward class declaration from SPM and this type used for declare properties or parameters in this header.
Those properties and parameters won't be exposed into Swift.
You will receive build error:
👋 We have encountered this same issue while converting the Braintree SDK from Obj-C to Swift. Our SDK is comprised of several sub-modules, and while we thought we could convert to Swift module-by-module, this SPM bug has us blocked and forced to convert our entire SDK at once.
Our team has created a Swift Forum post on the issue here, for those that want to follow along / try to gain traction.
cc @tomerd
@tomerd Please help me transfer this if it is indeed a package manager issue.
cc @abertelrud @neonichu
Sounds like an Xcode issue to me.
|
gharchive/issue
| 2021-09-04T00:29:02 |
2025-04-01T04:56:01.047041
|
{
"authors": [
"AnthonyLatsis",
"neonichu",
"oleksandryevdokymov",
"scannillo",
"swift-ci",
"tomerd"
],
"repo": "apple/swift",
"url": "https://github.com/apple/swift/issues/57480",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1529777034
|
Partial block result builder fails to pick correct overload and generates compiler error
Description
Partial block result builder fails to pick correct overload and generates a compiler error Ambiguous use of 'parser(of:)'
Steps to reproduce
Paste the following code into a new project
(This issue is a reduction of trying to compile an example (final 4 lines of the code below) of the swift-parsing package)
//MARK: Parsers
@rethrows public protocol Parser<Input> {
associatedtype Input
}
extension String: Parser {
public typealias Input = Substring
}
extension Int {
static func parser(
of inputType: Substring.Type = Substring.self
) -> FromSubstringToUTF8<IntParser<Substring.UTF8View>> {
FromSubstringToUTF8 { IntParser<Substring.UTF8View>() }
}
static func parser(
of inputType: Substring.UTF8View.Type = Substring.UTF8View.self
) -> IntParser<Substring.UTF8View> {
.init()
}
}
struct FromSubstringToUTF8<P: Parser>: Parser where P.Input == Substring.UTF8View {
typealias Input = Substring
let parser: P
init(@ParserBuilder _ build: () -> P) { self.parser = build() }
}
struct IntParser<Input: Collection>: Parser where Input.SubSequence == Input, Input.Element == UTF8.CodeUnit {
public init() { }
}
//MARK: ParserBuilder
struct Parse<Parsers: Parser>: Parser {
typealias Input = Parsers.Input
let parsers: Parsers
init(@ParserBuilder with build: () -> Parsers) { self.parsers = build() }
}
@resultBuilder
struct ParserBuilder {
public static func buildPartialBlock<P: Parser>(first: P) -> P { first }
public static func buildPartialBlock<P0, P1>(accumulated: P0, next: P1) -> SkipFirst<P0, P1> {
.init(p0: accumulated, p1: next)
}
}
struct SkipFirst<P0: Parser, P1: Parser>: Parser where P0.Input == P1.Input {
typealias Input = P0.Input
let p0: P0
let p1: P1
}
//MARK: main
let first = ParserBuilder.buildPartialBlock(first: ",")
let second = ParserBuilder.buildPartialBlock(accumulated: first, next: Int.parser())
let parser = Parse {
","
Int.parser() // Ambiguous use of 'parser(of:)'
}
Expected behavior
This compiles for toolchain 2023-01-02 but breaks from 2023-01-07.
Expected to compile and choose the correct overload (in this case static func parser(of inputType: Substring.Type = Substring.self))
Since the first parser passed to the resultBuilder closure is of type String therefore the Parser associated type Input is of type Substring. Therefore the compiler should be able to infer that the parser returned by Int.parser() should have an Input of Substring as well.
Calling the individual partialBuildBlock functions does compile
works:
let first = ParserBuilder.buildPartialBlock(first: ",")
let second = ParserBuilder.buildPartialBlock(accumulated: first, next: Int.parser())
fails:
let parser = Parse {
","
Int.parser() // Ambiguous use of 'parser(of:)'
}
Environment
Swift compiler version info Swift Development Snapshot 2023-01-09
Xcode version info Xcode 13.4.1 Build version 13F100
Deployment target: M1 running macOS 12.5
Ran using the 2023-01-06 snapshot as well and that fails to compile too.
@JaapWijnen We have changed implementation to be more strict about how the inference works (aligned with https://github.com/apple/swift-evolution/blob/main/proposals/0348-buildpartialblock.md) and as per proposal result builder is a source transform that works this way:
let parser = Parse {
let __builder0 = ","
let __builder1 = Int.parser()
let __builder2 = ParserBuilder.buildPartialBlock(first: __builder0)
let __builder3 = ParserBuilder.buildPartialBlock(accumulated: __builder2, next: __builder1)
return __builder3
}
which means that buildPartialBlock cannot in any way affect type-checking behavior of Int.parser().
The fix here would be to use static func buildExpression<T: Parser>(_ expression: T) -> T { expression }
@Xedin
Hm I'm not sure I 100% follow. In the process of creating a reproduces I might have minimised my example a bit too much to convey what I'm actually trying to do. Let me know if your answer still applies after my explanation in that case I'm missing how I would have to implement your suggested fix.
What is actually happening in my use case is that on the Parser protocol there's a function that takes an inout input, produces an output and in the process modifies that input. The modified input is then passed to the next parser (In this case the Int parser). Therefore there's this dependency that both parsers have the same Input associatedtype. This restriction in this example is implemented by the type constraints on the SkipFirst parser.
The first parser is a string and it's defined to have an Input of Substring therefore SkipFirst<String, X> is required to have an Input of Substring which in turn requires X (in this case an IntParser) would be required to have Substring as input.
To describe, the full call hierarchy would be as follows: (This writing might be obvious and/or I'm not sure what's the best way of communicating this is) I'll add an updated version of my code below as well.)
I'd expect the Parse { } call to return a parser of type:
Parse<SkipFirst<String, FromSubstringToUTF8<IntParser<Substring.UTF8View>>>>
we can call parse(&input) on that which just passes it on to SkipFirst since that contains a single parser
This in turn will call two parsers
first p0.parse(&input) this runs removes the "," input and returns Void
second p1.parse(&input) // FromSubstringToUTF8<IntParser<Substring.UTF8View>>
this second call in turn calls a parser that wraps a parser of a different input. Since the IntParser does not work for an input of SubString as it's requiring Input: Collection where where Input.SubSequence == Input, Input.Element == UTF8.CodeUnit
So this parser, FromSubstringToUTF8, transforms the input, calls parse(&transformedInput), modifying it and then transforms the modified input back to the original Input type and overwrites the input variable.
Finally propagating the returned value of the IntParser to the FromSubstringToUTF8 to the SkipFirst and finally returning the value.
I really hope this is helpful. Let me know if there's information missing/unclear or how I can help in any other way!
code:
struct ParsingError: Error { }
//MARK: Parsers
@rethrows public protocol Parser<Input, Output> {
associatedtype Input
associatedtype Output
func parse(input: inout Input) throws -> Output
}
extension String: Parser {
public typealias Input = Substring
public func parse(input: inout Substring) throws {
guard input.starts(with: self) else {
throw ParsingError()
}
input.removeFirst(self.count)
}
}
extension Int {
static func parser(
of inputType: Substring.Type = Substring.self
) -> FromSubstringToUTF8<IntParser<Substring.UTF8View>> {
FromSubstringToUTF8 { IntParser<Substring.UTF8View>() }
}
static func parser(
of inputType: Substring.UTF8View.Type = Substring.UTF8View.self
) -> IntParser<Substring.UTF8View> {
.init()
}
}
struct FromSubstringToUTF8<P: Parser>: Parser where P.Input == Substring.UTF8View {
typealias Input = Substring
let parser: P
init(@ParserBuilder _ build: () -> P) { self.parser = build() }
func parse(input: inout Substring) throws -> P.Output {
var transformedInput = input.utf8
let result = try parser.parse(input: &transformedInput)
input = Substring(transformedInput)
return result
}
}
struct IntParser<Input: Collection>: Parser where Input.SubSequence == Input, Input.Element == UTF8.CodeUnit {
public init() { }
func parse(input: inout Input) throws -> Int {
// some int parsing logic
return 4
}
}
//MARK: ParserBuilder
struct Parse<Parsers: Parser>: Parser {
let parsers: Parsers
init(@ParserBuilder with build: () -> Parsers) { self.parsers = build() }
func parse(input: inout Parsers.Input) throws -> Parsers.Output {
try parsers.parse(input: &input)
}
}
@resultBuilder
struct ParserBuilder {
public static func buildPartialBlock<P: Parser>(first: P) -> P { first }
public static func buildPartialBlock<P0, P1>(accumulated: P0, next: P1) -> SkipFirst<P0, P1> {
.init(p0: accumulated, p1: next)
}
}
struct SkipFirst<P0: Parser, P1: Parser>: Parser where P0.Input == P1.Input {
let p0: P0
let p1: P1
func parse(input: inout P0.Input) throws -> P1.Output {
let _ = try p0.parse(input: &input)
return try p1.parse(input: &input)
}
}
//MARK: main
let first = ParserBuilder.buildPartialBlock(first: ",")
let second = ParserBuilder.buildPartialBlock(accumulated: first, next: Int.parser())
let parser = Parse {
","
Int.parser() // Ambiguous use of 'parser(of:)'
}
var input = ",4"[...]
let result = try parser.parse(input: &input)
print(result) // 4
I understand what you mean, this kind of workflow is not what is going to be supported by result builder transform implementation going forward. The result builder transform semantics are such that each element in the body is type-checked independently from others and the resulting value is then passed to a final buildBlock or a series of buildPartialBlock calls and returned just like I outlined in my example, the old implementation failed to enforce the "solved independently" bit which caused all sorts of diagnostics and performance issues.
In your example there are two overloads of parser(of:) method, both have argument that accepts a default value which means that the type-checker won't be able to disambiguate between them without buildExpression or buildPartialBlock providing more context (via generic requirements) just like if you wrote _ = Int.parser() without using result builders.
So does this mean that code previously relying on this feature will fail to compile in a future release?
I've seen multiple uses of this in libraries that use resultBuilders to create sort of pipelines which might be an easier example than my above parser example.
// a pipeline has an input and output
let pipeline = Pipeline {
Pipeline1 // input: A, output: B
Pipeline2 // input: B, output: C
Pipeline3 // input: C, output: D
}
These types Pipeline1, 2 and 3 are then no longer able to be defined using generics and type restrictions? Is that correct?
Or is there a fix for my example that I'm missing at the moment?
It’s not about generics or type restrictions, the problem is that implementation did not adhere to the proposed semantics and allowed invalid code to type-check, which would not longer be the same. There are ways to change the result builder itself to archive portion of what was allowed before but cross-element inference would not be supported.
Just to clarify you can still use buildParticalBlock to produce a chain you described A -> B -> C -> D but it would not operate on the original elements, the builder transform in your example is going to be:
var __builder0 = Pipeline1()
var __builder1 = Pipeline2()
var __builder2 = Pipeline3()
var r0 = Pipeline.buildPartialBlock(first: __builder0)
var r1 = Pipeline.buildPartialBlock(accumulated: r0, next: __builder1)
var r2 = Pipeline.buildPartialBlock(accumulated: r1, next: __builder2)
return r2
not:
var r0 = Pipeline.buildPartialBlock(first: Pipeline1())
var r1 = Pipeline.buildPartialBlock(accumulated: r0, next: Pipeline2())
var r2 = Pipeline.buildPartialBlock(accumulated: r1, next: Pipeline3())
return r2
because that violates the code principle behind result builders - type-checking of Pipeline{1, 2, 3} initializers would be affected by buildPartialBlock.
If either Pipeline1, Pipeline2 or Pipeline3 initializers are ambiguous all you can do is to use buildExpression to provide additional context.
This is exactly how the transform is described in the proposal that I've mentioned.
Aha that is very helpful! What I then finally perhaps fail to understand is how to use 'buildExpression' to add context could you elaborate a little bit on that?
Sure! static func buildExpression when declared is going to be called for every element, back to your previous example:
var __builder0 = Pipeline.buildExpression(Pipeline1())
var __builder1 = Pipeline.buildExpression(Pipeline2())
var __builder2 = Pipeline.buildExpression(Pipeline3())
var r0 = Pipeline.buildPartialBlock(first: __builder0)
var r1 = Pipeline.buildPartialBlock(accumulated: r0, next: __builder1)
var r2 = Pipeline.buildPartialBlock(accumulated: r1, next: __builder2)
return r2
So you could declare that method in Pipeline to have any kind of generic parameters and return type, for example:
static func buildExpression<T: SomeProtoocol>(_ expression: T) -> T { expression }
This would make sure that Pipeline{1, 2, 3} all conform to SomeProtocol.
Got it! Thanks @xedin. One final question in that case. Is there then a possibility to also restrict associated types of that protocol? As in I'd like for all types supplied to the result builder that A they conform to SomeProtocol and B that their associated types SomeType (contained in SomeProtocol) are equal? Or is that not possible?
Yes, you can place whatever generic requirements are necessary i.e. <T: SomeProtocol>(...) where T.SomeType == ..., T.SomeOtherMember : OtherProtocol etc.
Hmm I don't fully follow. I can't make the requirement for the following example:
protocol: SomeProtocol {
associatedtype Input
associatedtype Output
}
struct P1<Input>: SomeProtocol {
typealias Output = Int
}
struct P2<Input>: SomeProtocol {
typealias Output = String
}
//When using these in a result builder:
let object = Builder {
P1()
P2()
}
Can I force the Input associated type to be the same for all entries given to the builder? I don't see how to do that with restrictions on buildExpression that seems to me to only put restrictions on the individual entries is that correct?
There are a couple of ways depending on what you want.
You can make your builder generic like this:
@resultBuilder
struct Builder<T> {
static func buildExpression<U: SomeProtocol>(_ e: U) -> U where U.Input == T { e }
...
}
This would make sure that Input is the same for all the elements eagerly.
You could also enforce that requirement in a buildBlock:
@resultBuilder
struct Builder<T> {
static func buildBlock<Input, T0: SomeProtocol, T1: SomeProtocol, ...>(_ t0: T0, _ t1: T1) -> <#Result#> where T0.Input == Input, T1.Input == Intput, T<N> == Input { e }
...
}
The caveat here is that element type-checking is free to pick what ever overload for actual elements.
@xedin ah yes Nr 1 is exactly what I've been trying the last few minutes! It seems to work for my minimum example fine. I've also applied it to the library as a whole but unfortunately haven't had success of getting it to fully work as before.
Thanks for all the help!
No problem! I'm going to resolve this one, feel free to reach out on forums as well if you need more help with result builders.
@xedin Do you happen to remember which recent PRs were involved in the change? Would be interested in reading up on that context.
The implementation is https://github.com/apple/swift/pull/60065 and there are a few smaller follow-up PRs. Enablement is https://github.com/apple/swift/pull/62734. I’m also working on a forums post to announce the change and describe some of the situations that were allowed before but wouldn’t be now.
@xedin Thanks! Will be interested to see how it improves result builder compile times generally :smile:
@xedin @stephencelis Might be interesting: I've worked on making the result builder generic over Input for swift-parsing But am running into long build times for some parsers and some even times out with The compiler is unable to type-check this expression in reasonable time; try breaking up the expression into distinct sub-expressions
Do you have a branch where you are making these changes? I can probably take a look and give you suggestions.
@xedin that would be great! My changes can be found here: https://github.com/jaapwijnen/swift-parsing/tree/generic-parser-builder
I've commented out some tests (which you can see in the changes) these were the ones that were mostly acting up.
No worries, I'll try to take a look tomorrow. If you could reduce a problematic case together with the builder and post here that would be very helpful as well.
Hmm it's hard to isolate the builder from the library in a singular case. Which is what makes it so hard for me to debug why it's having trouble compiling the parser. Is there any way to identify which resultBuilder methods are used under the hood?
Here's an example of a parser that's not compiling at the moment however (taken from one of the tests I commented out in the PR I linked):
var parser = Parse {
"Hello"
if parseComma {
","
}
" "
Prefix { $0 != "!" }
"!"
}
The Prefix parser is the one causing trouble in this case I believe, although not sure why.
It's something like:
var parser = Parse {
var v0 = ParserBuilder.buildExpression("Hello")
var v1: ?
if parseComma {
var v2 = ParserBuilder.buildExpression(",")
v1 = ParserBuilder.buildBlock(v2)
} else {
v1 = ParserBuilder.buidBlock(Optional<???>.none)
}
var v3 = ParserBuilder.buildExpression(" ")
var v4 = ParserBuilder.buildExpression(Prefix { $0 != "!" })
var v5 = ParserBuilder.buildExpression("!")
var r0 = ParserBuilder.buildPartialBlock(first: v0)
var r1 = ParserBuilder.buildPartialBlock(accumulated: r0, next: v1)
var r2 = ParserBuilder.buildPartialBlock(accumulated: r1, next: v3)
var r3 = ParserBuilder.buildPartialBlock(accumulated: r2, next: v4)
var r4 = ParserBuilder.buildPartialBlock(accumulated: r3, next: v5)
return r4
}
Not sure if Prefix { $0 != "!" } is transformed or not. The place marked with ??? is the type of v2, this interaction cannot be expressed in the surface language as stated by the proposal.
Hm interestingly it seems to be the combination of the parser wrapped in an if statement and the Prefix parser. Since when I delete either of them the parser does compile.
What is $0 supposed to be? I see that Input associated with Prefix supposed to conform to Collection protocol.
Substring in this case I believe
It's Input.Element:
https://github.com/pointfreeco/swift-parsing/blob/a3d52c666982f7cc9f243b910151cdb9f6283db1/Sources/Parsing/ParserPrinters/Prefix.swift#L55
Which is Substring.Element, or Character.
Ah my bad thanks @stephencelis
The good news is that the performance issues are not due to result builder implementation, old one is too complex too on that example. Just need to figure out what exactly is causing this, I suspect it's multiple overloads of buildIf.
Huh that is strange, @stephencelis this used to compile fine right? It's part of the library's tests.
I think you might be building with old compiler where transform is disabled actually because with transform I get the following diagnostic:
error: generic parameter 'Parsers' could not be inferred
let parser = Parse {
^
swift-parsing/Sources/Parsing/ParserPrinters/Parse.swift:23:21: note: 'Parsers' declared as parameter to type 'Parse'
public struct Parse<Parsers: Parser>: Parser {
^
swift-parsing/Sources/Parsing/ParserPrinters/Prefix.swift:225:18: note: explicitly specify the generic arguments to fix this issue
let parser = Parse {
^
<<#Parsers: Parser#>>
@JaapWijnen Looks like all you might need to do is to fix init declaration in Parse.swift line 46 to:
@inlinable
public init<Input>(@ParserBuilder<Input> with build: () -> Parsers) where Parsers.Input == Input {
self.parsers = build()
}
Parsers.Input cannot be used directly because the concrete type for Parsers is not known until the whole body of build: is type-checked, instead you want the Input to inferred from the body and passed up to the Parsers.Input if that generic parameter is not specified explicitly.
Also make sure that the toolchain you are using actually has result builder transform enabled :)
That did help :) Now I'm running into the ambiguity issue again unfortunately haha.
Must be a different example then since the one you pasted builds just fine for me after that init change.
Hmm not entirely for me. Made some changes to the parsers in the library to all accept an Input generic. Now I'm having trouble to get it to compile on the 2022-10-09 toolchain however:
This one times out again:
var parser = Parse {
"Hello"
if parseComma {
","
}
" "
Prefix { $0 != "!" }
"!"
}
And for this one I get the following error
let p = ParsePrint {
Prefix { $0 != "\n" } // Ambiguous use of 'init(while:)'
First()
}
And another very similar one but slightly different:
let p = ParsePrint {
Prefix(3) { $0.isNumber }
First() //Ambiguous use of 'init()'
}
My latest changes can be found here if you're interested! : https://github.com/JaapWijnen/swift-parsing/commit/edcdf2d98c00dd4b210c7e904d05b9e151ff6b4e
I think you might want to pick up a newer toolchain maybe because I pulled your branch and couldn't reproduce anything except to:
func testSkipSuccess() {
var input = "42 Hello, world!"[...].utf8
XCTAssert(try () == XCTUnwrap(Skip { Int.parser() }.parse(&input))) // Ambiguous use of 'parser(of:radix:)'
XCTAssertEqual(" Hello, world!", Substring(input))
}
This is indeed ambiguous because both overloads that match are marked as @_disfavoredOverload, have all their parameters defaulted and produce a type that conforms to Parser - From conforms to Parser via ParserPrinter, IntParser conforms to parser directly.
Downloading the newest now!
Ah ok interesting, I thought because the input is included which is of type Substring.UTF8View the Skip parser would know it's input generic is of that same type therefore the result builder will use that as it's Input generic which in turn will pick that specific Int.parser() overload. But I guess it doesn't work like that. Could you point me towards where my logic breaks down/how that type inference actually works in this case?
Thanks for all the help!
The ranking of overloads would check the number of defaulted parameters (which is the same) and whether any of the parameters have subtyping relationship, none of this is applicable in this case.
Ah ok thanks @xedin! Then indeed some of the ergonomics of doing it like this are simply lost due to this regression (at least in their current form) Thanks for helping me understand!
Hope this is useful to you as well @stephencelis. Not sure what useful next steps are in terms of getting swift-parsing to compile for new toolchains (probably not an immediate priority to you?) But I'd love to help if you do have ideas about that!
We're using swift-parsing with snapshots so I'd love for it to work as it did before but right now at least we can get around the problem by just annotating all the parsers with a lot more type information.
This is definitely helpful! While we knew a lot of changes we needed to make based off more recent result builder projects we've taken on, this thread provided a lot of additional food for thought. Thanks, @xedin!
Looks like all you might need to do is to fix init declaration in Parse.swift line 46 to:
@inlinable
public init<Input>(@ParserBuilder<Input> with build: () -> Parsers) where Parsers.Input == Input {
self.parsers = build()
}
Parsers.Input cannot be used directly because the concrete type for Parsers is not known until the whole body of build: is type-checked, instead you want the Input to inferred from the body and passed up to the Parsers.Input if that generic parameter is not specified explicitly.
Also make sure that the toolchain you are using actually has result builder transform enabled :)
@xedin I tried to apply this technique (on Swift Development Snapshot 2023-01-23) to a similar generic builder, but get "Generic parameter 'X' could not be inferred" where 'X' is the builder's generic.
Here's a basic example demonstrating the problem:
@resultBuilder
enum Builder<A> {
static func buildExpression<T: P>(_ expression: T) -> T where T.A == A {
expression
}
static func buildBlock() -> E<A> {
E()
}
static func buildBlock<T: P>(_ component: T) -> T where T.A == A {
component
}
static func buildPartialBlock<T: P>(first: T) -> T where T.A == A {
first
}
static func buildPartialBlock<T0: P, T1: P>(accumulated: T0, next: T1) -> T1
where T0.A == A, T1.A == A {
next
}
}
protocol P<A> {
associatedtype A
associatedtype Body
@Builder<A>
var body: Body { get }
}
extension P where Body == Never {
var body: Body { fatalError() }
}
struct E<A>: P {}
struct C<T: P>: P {
typealias A = T.A
let t: T
init<A>(@Builder<A> build: () -> T) where T.A == A {
self.t = build()
}
}
struct S: P {
struct A {}
@Builder<A>
var body: some P<A> {
C { // 🛑 Generic parameter 'A' could not be inferred
E()
}
C {} // 🛑 Generic parameter 'A' could not be inferred
}
}
Am I maybe missing something? Is there an -Xfrontend setting (or otherwise) I need to enabled on this snapshot? Or is this a potential bug?
There is no where to infer A from in C { E() } or C {} though right?
@xedin Shouldn't the A generic in @Builder<A> flow through the body context and hydrate C's init<A>, which in turn will know that its @Builder is generic over A and flow into the E? Or is this going in the wrong direction?
I don't quite know how to precisely translate builder code to the equivalent buildBlock (I can approximate it, but I'm not sure if I'm missing anything).
@stephencelis No, this is actually the whole point. The body is not solved as a unit, it's solved piece-by-piece just like a regular multi-statement closure so @Builder<A> affects only final return build{Partial}Block(...):
struct S: P {
struct A {}
@Builder<A>
var body: some P<A> {
var $__builder0 = Builder.buildExpression(C {
var $__builder0 = Builder.buildExpression(E())
return Builder.buildBlock($__builder0)
})
var $__builder1 = Builder.buildExpression(C {
return Builder.buildBlock()
})
let r1 = Builder.buildPartialBlock(first: $__builder0)
let r2 = Builder.buildPartialBlock(accumulated: r1, next: $__builder1)
return r2
}
}
In here Builder in Builder.buildExpression(E()) is different from Builder<A> attached to var body, there is no connection between them.
Maybe this is going to be more helpful: buildExpression states that T.A == A where T: P - to be able to solve that we need to first solve the expression C.init { E() }, C.init again tries to propagate A "up" (to the context) from the body of the result builder because in where T.A == A the right-hand side is not yet resolved - it's waiting for the { E() } to get resolved (as stated at the very beginning).
Note that buildExpression requirement T.A == A cannot propagate information down from @Builder<A> until T is resolved and that is only resolved when the expression is solved and part of that expression is solved in isolation because argument is a result builder.
@xedin I might be missing some nuance here, but doesn't this place a pretty big limitation on generic result builders? I would hope that the generic Builder<A> would propagate into the context and be used to anchor all of the buildExpressions, etc., and allow a generic type like E() to resolve without specifying generics explicitly. Why wouldn't the following be preferable?
struct S: P {
struct A {}
@Builder<A>
var body: some P<A> {
var $__builder0 = Builder<A>.buildExpression(C { @Builder<A> in
var $__builder0 = Builder<A>.buildExpression(E())
return Builder<A>.buildBlock($__builder0)
})
var $__builder1 = Builder<A>.buildExpression(C { @Builder<A> in
return Builder<A>.buildBlock()
})
let r1 = Builder<A>.buildPartialBlock(first: $__builder0)
let r2 = Builder<A>.buildPartialBlock(accumulated: r1, next: $__builder1)
return r2
}
}
Wouldn't dragging the <A> into things make this code compile just fine and without taking a big toll on the type checker?
would propagate into the context and be used to anchor all of the buildExpression
It does propagate into the context but that context is top-level things and the closure associated with C.init` because that is a completely different builder.
Is there a reason why the A couldn't trickle into the inner @Builder<???> context to resolve? All the type information seems to be there... Is there any way to make generic builders compose nicely?
For it to get into C.init closure it needs to get connected to something which would resolve it without relying on bi-directional inference which I don't think is possible because the language cannot pick up relations from context. Also it's very important to note that the example you've posted is not accepted by the original implementation either so this is not a regression.
@xedin So it's just impossible for C.init's builder to be influenced by the parent buildExpression? While I understand that result builder type inference is limited for performance reasons, I thought buildExpression was specifically designed to provide context to work around some of these limitations. Is there any reason why buildExpression couldn't still use bidirectional inference while the rest of the result builder APIs remain limited?
This limitation is quite a bummer if there's no workaround, and it prevents us from shipping an API that isn't cumbersome to use and difficult to troubleshoot or needlessly verbose, e.g.:
C(for: A.self) { // Worse when multiple generics are involved
// ...
}
I thought buildExpression was specifically designed to provide context to work around some of these limitations. Is there any reason why buildExpression couldn't still use bidirectional inference while the rest of the result builder APIs remain limited?
buildExpression can and does provide context, but the way you have defined it, there is no context it could pass down. The problem here is not the luck on bi-directional inference inside of buildExpression argument, the problem is that the builder associated with C.init has Builder<???>.buildExpression(E()) expression where inference fails because { E() } is treated as a transformed multi-statement closure (as per result builder proposal). The generic parameter cannot be transferred into the body of { E() }, that didn't work before and it won't work with new transform.
Also it's very important to note that the example you've posted is not accepted by the original implementation either so this is not a regression.
Do you mean existing code should continue to compile? Or do you mean it's not a regression as far as the original evolution proposal conveyed? Given the swift-parsing code that used to compile, as well as some swift-composable-architecture code that used to compile, these changes definitely introduce source breaking changes and regressions, no?
While I get a compiler crash with my example above when I run with Xcode 14.2's built-in toolchain, I have some very similar library code that compiles fine in Xcode 14.2 but does not on the 5.8 (or development) toolchain.
If you clone https://github.com/pointfreeco/swift-composable-architecture (main branch), and paste the following code somewhere into the project:
struct R: ReducerProtocol {
struct State {}
enum Action {}
@ReducerBuilder<State, Action>
var body: some ReducerProtocol<State, Action> {
CombineReducers {}
CombineReducers {
EmptyReducer()
}
}
}
...it compiles just fine.
However, if I switch to a 5.8 or development toolchain, it doesn't. These are pretty unfortunate source breaking changes to introduce to library users, especially if there's nothing we can do as library maintainers to make the upgrade to Swift 5.8 painless:
I also began to sketch out another branch with some adjustments for some of the changes coming:
https://github.com/pointfreeco/swift-composable-architecture/compare/builder-updates
Here, too, of course, the above won't compile (and still compiles fine in Xcode 14.2), but I was hoping the changes would at least address some problems with the existing implementation.
This example is problematic because ReducerBuilderOf is declared as:
public typealias ReducerBuilderOf<R: ReducerProtocol> = ReducerBuilder<R.State, R.Action>
and CombineReducers.init is declared as:
public init(
@ReducerBuilderOf<Reducers> _ build: () -> Reducers
)
which means that:
CombineReducers {
var $__builder0 = ReducerBuilderOf<$T>.builderExpression(EmptyReducer())
return ReducerBuilderOf<$T>.buildBlock($__builder0)
}
$T inference depends on the result type of buildBlock. The only way to infer that would be to attempt to solve whole body as a unit.
@xedin See the second link, where I've deprecated ReducerBuilderOf and refactored away usage.
@stephencelis Even without that explicit R: ReducerProtocol the type of ReducerBuilder still depends on the type produced by the body, although it's unpacked directly now.
@xedin I think I understand that, but what I'm trying to show you is that this statements you made don't appear to be true (as I'm understanding them):
Also it's very important to note that the example you've posted is not accepted by the original implementation either so this is not a regression.
The generic parameter cannot be transferred into the body of { E() }, that didn't work before and it won't work with new transform.
(Emphasis mine.) The code does currently compile, so this does seem like a regression.
See my earlier comment again for more context.
The code does currently compile, so this does seem like a regression.
@stephencelis I was referring to the code in https://github.com/apple/swift/issues/62978#issuecomment-1402356383
@xedin Hm, OK. So seeing as this is a regression with no current workaround, should I start a forum discussion to get wider visibility? Currently, these changes could break a lot of userland code.
This is currently being discussed by the language group. There is a workaround here though - specify generic arguments explicitly.
I guess I meant workaround that wouldn't be passed on to library users and become a new support vector for us 😅
And ideally one that doesn't make every nested generic builder more verbose than folks expect in Swift.
As a side note, I guess better compiler diagnostics and support would help here. If the compiler could help users migrate (or we could provide such migration APIs as fix-its) that'd make this change less painful.
Well, that the problem, compiler cannot determine what the type should be here so we cannot suggest one, it really is very unfortunate that declarations like that used to be accepted by the compiler because they result in performance issues as bodies grow and bad diagnostics because whole body is type-checked together.
@xedin Since the compiler can figure it out today, one would hope that the diagnostic could do this work when compilation fails in a post-compilation phase, at least during a migratory period...
Unfortunately no, otherwise we wouldn’t be spending time re-implementing the transform. The principle issue here is that type-checking everything together is too expensive especially for invalid code.
So seeing as this is a regression with no current workaround, should I start a forum discussion to get wider visibility? Currently, these changes could break a lot of userland code.
@stephencelis I'm going to take a look at the result builders in TCA to see whether I can figure out a tweak to the builder to enable the same call-site ergonomics. I think leaving this issue open is the best course of action for now. Pavel and I are working on a post for the Swift Forums to widely communicate the motivation, compile-time improvements, and the source compatibility impact of this change. Thank you for your patience and for reporting the impact on your project!
Thanks @hborla and @xedin for the communication, in general! We do appreciate the need and reasoning behind the changes, and we hope that the majority of TCA result builder usage will remain unaffected. We just wanted to do what we could to signal potential breakage to the core team and minimize frustration for end users of the library that hit what we hope are edge cases. We also hope there's a clear way to communicate to users when they need to do:
Reduce { state, action in
// or
CombineReducers {
vs.
Reduce<State, Action> { state, action in
Or if we should just enforce specifying via argument wherever there may be ambiguity, which is more verbose than needed a lot of the time, but at least avoids the problematic case:
Reduce(into: State.self, action: Action.self) { state, action in
@stephencelis I actually found a solution for CombineReducers problem (based on builder-updates branch):
diff --git a/Sources/ComposableArchitecture/Reducer/Reducers/CombineReducers.swift b/Sources/ComposableArchitecture/Reducer/Reducers/CombineReducers.swift
index bb569e8dc..031a8dc51 100644
--- a/Sources/ComposableArchitecture/Reducer/Reducers/CombineReducers.swift
+++ b/Sources/ComposableArchitecture/Reducer/Reducers/CombineReducers.swift
@@ -15,7 +15,8 @@
/// .ifLet(\.child, action: /Action.child)
/// }
/// ```
-public struct CombineReducers<Reducers: ReducerProtocol>: ReducerProtocol {
+public struct CombineReducers<Action, State, Reducers: ReducerProtocol>: ReducerProtocol
+ where Action == Reducers.Action, State == Reducers.State {
@usableFromInline
let reducers: Reducers
@@ -23,9 +24,9 @@ public struct CombineReducers<Reducers: ReducerProtocol>: ReducerProtocol {
///
/// - Parameter build: A reducer builder.
@inlinable
- public init<State, Action>(
+ public init(
@ReducerBuilder<State, Action> _ build: () -> Reducers
- ) where State == Reducers.State, Action == Reducers.Action {
+ ) {
self.init(internal: build())
}
Adding this indirection for State/Action generic parameter makes sure that when CombineReducers() is wrapped in buildExpression declared as:
static func buildExpression(_: some ReducerProtocol<State, Action>) -> some ReducerProtocol<State, Action> {
return expression
}
State Action could be infer for it based on the "builder self": ReducerBuilder<State, Action>.builderExpression(CombineReducers()) where State and Action are struct and enum from your example.
Example you have posted previously type-checks with these changes.
@xedin Wow! Thanks! That actually gets the one regression in our test suite building again without needing to be explicit with \State.path and allowing abbreviation in \.path again!
https://github.com/pointfreeco/swift-composable-architecture/commit/feae99aa459f47962a01289682ee8468e2dc969a
@stephencelis No problem! Any other issues I can help with?
@xedin Will let you know if anything else comes up! Really appreciate the help along the way!
No worries, happy to help!
@xedin One more question... :smile:
We have this generic Reduce<State, Action> type:
https://github.com/pointfreeco/swift-composable-architecture/blob/main/Sources/ComposableArchitecture/Reducer/Reducers/Reduce.swift
When used from a builder with a single build block, e.g. here:
https://github.com/pointfreeco/swift-composable-architecture/blob/5e4caa89ee7e1add70557f77c7c35cfad73ff024/Examples/TicTacToe/tic-tac-toe/Sources/LoginCore/LoginCore.swift#L33
The closure seems to be type-checked efficiently, and Xcode autocomplete works just great when trying to autocomplete state.<ESC> parameter names and case .<ESC> action names.
However as soon as we add another reducer to the build block (like EmptyReducer()), above or below this Reduce, things still compile just fine, but we lose all autocompletion help in the closure. state.<ESC> and case .<ESC> are of no help till we explicitly specify Reduce<State, Action>.
Do you happen to have a tip we can employ to keep the compiler happy without this extra busy work?
Cc @ahoppen
Hi @stephencelis,
Thanks for sharing the code completion issue. I hope I can find time to take a look at it sometime next week.
@ahoppen Thanks!
I have managed to reduce the code completion issue down to the following. Checking with Xcode 14, the issue already existed in that release, so it’s not a regression.
The problem here appears to be that code completion is getting confused which closures it needs to type check. Specifically, the result builder with multiple statements the closure is not in the ReadyForTypeCheck body state here and thus we don’t continue.
https://github.com/apple/swift/blob/4ceec83cd87acc140e0e5d93956036861de88589/lib/Sema/TypeCheckStmt.cpp#L2419-L2420
I think this should resolve itself after we
Get rid of LeaveClosuresUnchecked (https://github.com/apple/swift/pull/59944)
Migrate the remaining code completion types to solver-based (https://github.com/apple/swift/pull/42280 and https://github.com/apple/swift/pull/42156)
Get rid of the the code path mentioned above entirely
There’s still some way to go on that journey but it’s something I am doing right now anyway.
// RUN: %swift-ide-test -code-completion -source-filename %s -code-completion-token COMPLETE
@resultBuilder
enum ReducerBuilder2<Action> {
static func buildBlock(_ r: Reduce2<Action>) -> Reduce2<Action> { r }
static func buildBlock(_ r0: Reduce2<Action>, _ r1: Reduce2<Action>) -> Reduce2<Action> { r0 }
static func buildExpression(_ r: Reduce2<Action>) -> Reduce2<Action> { r }
}
enum Action {
case alertDismissed
}
struct Reduce2<Action> {
init() {}
init(_ reduce: (Action) -> Int) {}
}
struct Login2 {
@ReducerBuilder2<Action>
var body: Reduce2<Action> {
Reduce2() // Remove this to receive code completion results get code completion working again
Reduce2 { action in
switch action {
case .#^COMPLETE^# alertDismissed:
return 0
}
}
}
}
@ahoppen That's great news! Thanks for looking into this :smile:
I stumbled across a similar issue today and this will be fixed by https://github.com/apple/swift/pull/59944. The test case I mentioned above is included in that PR.
|
gharchive/issue
| 2023-01-11T22:10:52 |
2025-04-01T04:56:01.128471
|
{
"authors": [
"JaapWijnen",
"ahoppen",
"hborla",
"stephencelis",
"xedin"
],
"repo": "apple/swift",
"url": "https://github.com/apple/swift/issues/62978",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1955180614
|
Methods on parameter packed generic types specialized with zero types cannot be used without crashing the compiler
Description
You can't create a generic type with a parameter pack, specialize it with zero types, and then call a method that uses that parameter pack. It crashes the compiler.
Steps to reproduce
Stack trace:
0. Program arguments: /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/swift-frontend -frontend -c /Users/fnord/Documents/Swift/TestMacOS/TestMacOSCore/Sources/TestMacOSCore/Dependencies.swift -primary-file /Users/fnord/Documents/Swift/TestMacOS/TestMacOSCore/Sources/TestMacOSCore/TestMacOSCore.swift -emit-dependencies-path /Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Build/Intermediates.noindex/TestMacOSCore.build/Debug/TestMacOSCore.build/Objects-normal/arm64/TestMacOSCore.d -emit-const-values-path /Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Build/Intermediates.noindex/TestMacOSCore.build/Debug/TestMacOSCore.build/Objects-normal/arm64/TestMacOSCore.swiftconstvalues -emit-reference-dependencies-path /Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Build/Intermediates.noindex/TestMacOSCore.build/Debug/TestMacOSCore.build/Objects-normal/arm64/TestMacOSCore.swiftdeps -serialize-diagnostics-path /Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Build/Intermediates.noindex/TestMacOSCore.build/Debug/TestMacOSCore.build/Objects-normal/arm64/TestMacOSCore.dia -target arm64-apple-macos14.0 -Xllvm -aarch64-use-tbi -enable-objc-interop -stack-check -sdk /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX14.0.sdk -I /Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Build/Products/Debug -I /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/usr/lib -F /Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Build/Products/Debug/PackageFrameworks -F /Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Build/Products/Debug/PackageFrameworks -F /Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Build/Products/Debug/PackageFrameworks -F /Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Build/Products/Debug/PackageFrameworks -F /Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Build/Products/Debug/PackageFrameworks -F /Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Build/Products/Debug -F /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/Library/Frameworks -no-color-diagnostics -enable-testing -g -module-cache-path /Users/fnord/Library/Developer/Xcode/DerivedData/ModuleCache.noindex -profile-generate -profile-coverage-mapping -swift-version 5 -enforce-exclusivity=checked -Onone -D SWIFT_PACKAGE -D DEBUG -D Xcode -serialize-debugging-options -package-name testmacoscore -const-gather-protocols-file /Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Build/Intermediates.noindex/TestMacOSCore.build/Debug/TestMacOSCore.build/Objects-normal/arm64/TestMacOSCore_const_extract_protocols.json -empty-abi-descriptor -validate-clang-modules-once -clang-build-session-file /Users/fnord/Library/Developer/Xcode/DerivedData/ModuleCache.noindex/Session.modulevalidation -Xcc -working-directory -Xcc /Users/fnord/Documents/Swift/TestMacOS/TestMacOSCore -resource-dir /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/lib/swift -enable-anonymous-context-mangled-names -Xcc -ivfsstatcache -Xcc /Users/fnord/Library/Developer/Xcode/DerivedData/SDKStatCaches.noindex/macosx14.0-23A334-4a76ee93ef69e462914ffe9883de22b8.sdkstatcache -Xcc -I/Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Build/Intermediates.noindex/TestMacOSCore.build/Debug/TestMacOSCore.build/swift-overrides.hmap -Xcc -I/Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Build/Products/Debug/include -Xcc -I/Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Build/Intermediates.noindex/TestMacOSCore.build/Debug/TestMacOSCore.build/DerivedSources-normal/arm64 -Xcc -I/Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Build/Intermediates.noindex/TestMacOSCore.build/Debug/TestMacOSCore.build/DerivedSources/arm64 -Xcc -I/Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Build/Intermediates.noindex/TestMacOSCore.build/Debug/TestMacOSCore.build/DerivedSources -Xcc -DSWIFT_PACKAGE -Xcc -DDEBUG=1 -module-name TestMacOSCore -frontend-parseable-output -disable-clang-spi -target-sdk-version 14.0 -target-sdk-name macosx14.0 -external-plugin-path /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX14.0.sdk/usr/lib/swift/host/plugins#/Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX14.0.sdk/usr/bin/swift-plugin-server -external-plugin-path /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX14.0.sdk/usr/local/lib/swift/host/plugins#/Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX14.0.sdk/usr/bin/swift-plugin-server -external-plugin-path /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/usr/lib/swift/host/plugins#/Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/usr/bin/swift-plugin-server -external-plugin-path /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/usr/local/lib/swift/host/plugins#/Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/usr/bin/swift-plugin-server -plugin-path /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/lib/swift/host/plugins -plugin-path /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/local/lib/swift/host/plugins -o /Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Build/Intermediates.noindex/TestMacOSCore.build/Debug/TestMacOSCore.build/Objects-normal/arm64/TestMacOSCore.o -index-unit-output-path /TestMacOSCore.build/Debug/TestMacOSCore.build/Objects-normal/arm64/TestMacOSCore.o -index-store-path /Users/fnord/Library/Developer/Xcode/DerivedData/TestMacOSCore-ahgngaarecknxsfwsewfijgzqaot/Index.noindex/DataStore -index-system-modules
1. Apple Swift version 5.9 (swiftlang-5.9.0.128.108 clang-1500.0.40.1)
2. Compiling with the current language version
3. While evaluating request ASTLoweringRequest(Lowering AST to SIL for file "/Users/fnord/Documents/Swift/TestMacOS/TestMacOSCore/Sources/TestMacOSCore/TestMacOSCore.swift")
4. While silgen emitFunction SIL function "@$s13TestMacOSCore3fooyyF".
for 'foo()' (at /Users/fnord/Documents/Swift/TestMacOS/TestMacOSCore/Sources/TestMacOSCore/TestMacOSCore.swift:23:1)
Stack dump without symbol names (ensure you have llvm-symbolizer in your PATH or set the environment var `LLVM_SYMBOLIZER_PATH` to point to it):
0 swift-frontend 0x00000001054f314c llvm::sys::PrintStackTrace(llvm::raw_ostream&, int) + 56
1 swift-frontend 0x0000000108071f2c llvm::sys::RunSignalHandlers() + 112
2 swift-frontend 0x0000000103db310c SignalHandler(int) + 352
3 libsystem_platform.dylib 0x000000018a961a24 _sigtramp + 56
4 swift-frontend 0x00000001058b51d0 (anonymous namespace)::ArgEmitter::emitPackArg(llvm::MutableArrayRef<swift::Lowering::ArgumentSource>, swift::Lowering::AbstractionPattern) + 164
5 swift-frontend 0x00000001058a2658 (anonymous namespace)::ArgEmitter::emitPreparedArgs(swift::Lowering::PreparedArguments&&, swift::Lowering::AbstractionPattern) + 512
6 swift-frontend 0x000000010591dea0 (anonymous namespace)::CallSite::emit(swift::Lowering::SILGenFunction&, swift::Lowering::AbstractionPattern, swift::CanTypeWrapper<swift::SILFunctionType>, (anonymous namespace)::ParamLowering&, llvm::SmallVectorImpl<swift::Lowering::ManagedValue>&, llvm::SmallVectorImpl<(anonymous namespace)::DelayedArgument>&, swift::ForeignInfo const&) && + 672
7 swift-frontend 0x000000010591cab4 (anonymous namespace)::CallEmission::emitArgumentsForNormalApply(swift::Lowering::AbstractionPattern, swift::CanTypeWrapper<swift::SILFunctionType>, swift::ForeignInfo const&, llvm::SmallVectorImpl<swift::Lowering::ManagedValue>&, llvm::Optional<swift::SILLocation>&) + 1000
8 swift-frontend 0x0000000102edb34c (anonymous namespace)::CallEmission::apply(swift::Lowering::SGFContext) + 4044
9 swift-frontend 0x000000010585574c swift::Lowering::SILGenFunction::emitApplyExpr(swift::ApplyExpr*, swift::Lowering::SGFContext) + 3072
10 swift-frontend 0x0000000105ab8468 swift::Lowering::SILGenFunction::emitIgnoredExpr(swift::Expr*) + 892
11 swift-frontend 0x000000010368f570 swift::ASTVisitor<(anonymous namespace)::StmtEmitter, void, void, void, void, void, void>::visit(swift::Stmt*) (.llvm.5384354941888427540) + 5512
12 swift-frontend 0x0000000105be5864 swift::Lowering::SILGenFunction::emitFunction(swift::FuncDecl*) + 632
13 swift-frontend 0x0000000102ec8a4c swift::Lowering::SILGenModule::emitFunctionDefinition(swift::SILDeclRef, swift::SILFunction*) + 8004
14 swift-frontend 0x00000001057d563c emitOrDelayFunction(swift::Lowering::SILGenModule&, swift::SILDeclRef) (.llvm.12369091251732681984) + 168
15 swift-frontend 0x00000001057bcbe8 swift::Lowering::SILGenModule::emitFunction(swift::FuncDecl*) + 292
16 swift-frontend 0x00000001057e4608 swift::ASTLoweringRequest::evaluate(swift::Evaluator&, swift::ASTLoweringDescriptor) const + 2908
17 swift-frontend 0x000000010440c4fc swift::SimpleRequest<swift::ASTLoweringRequest, std::__1::unique_ptr<swift::SILModule, std::__1::default_delete<swift::SILModule>> (swift::ASTLoweringDescriptor), (swift::RequestFlags)9>::evaluateRequest(swift::ASTLoweringRequest const&, swift::Evaluator&) + 200
18 swift-frontend 0x0000000105802948 llvm::Expected<swift::ASTLoweringRequest::OutputType> swift::Evaluator::getResultUncached<swift::ASTLoweringRequest>(swift::ASTLoweringRequest const&) + 584
19 swift-frontend 0x00000001080298d8 performCompile(swift::CompilerInstance&, int&, swift::FrontendObserver*) + 1680
20 swift-frontend 0x000000010802d474 swift::performFrontend(llvm::ArrayRef<char const*>, char const*, void*, swift::FrontendObserver*) + 4568
21 swift-frontend 0x00000001079734e0 swift::mainEntry(int, char const**) + 4116
22 dyld 0x000000018a5b9058 start + 2224
This causes the compiler to crash.
class CounterBox<each Param> {
func recordCall(_ params: repeat each Param) {}
}
func foo() {
CounterBox< >().recordCall() // crash
}
Expected behavior
The above should compile. Additionally, the parameter pack syntax is slightly different from what's described in the proposal. CounterBox<>() doesn't compile.
Environment
Swift compiler version info:
swift-driver version: 1.87.1 Apple Swift version 5.9 (swiftlang-5.9.0.128.108 clang-1500.0.40.1)
Target: arm64-apple-macosx14.0
Xcode version info
Xcode 15.0
Build version 15A240d
Deployment target: macOS 14.0
Additional context: Development is being done in a Swift Package using Xcode.
Assertion failed: (nextSubstParamIndex == substParams.size()), function claimParams, file SILGenApply.cpp, line 4556.
Stack dump:
0. Program arguments: bin/swift-frontend /Users/mac/Desktop/test.swift -debug-diagnostic-names -interpret
1. Swift version 5.11-dev (LLVM 5601e23279f85d8, Swift 140caf63be3f0fe)
2. Compiling with the current language version
3. While evaluating request ASTLoweringRequest(Lowering AST to SIL for module test)
4. While silgen emitFunction SIL function "@$s4test3fooyyF".
for 'foo()' (at /Users/mac/Desktop/test.swift:85:1)
Stack dump without symbol names (ensure you have llvm-symbolizer in your PATH or set the environment var `LLVM_SYMBOLIZER_PATH` to point to it):
0 swift-frontend 0x0000000112e6b51b llvm::sys::PrintStackTrace(llvm::raw_ostream&, int) + 43
1 swift-frontend 0x0000000112e6a755 llvm::sys::RunSignalHandlers() + 85
2 swift-frontend 0x0000000112e6bb4e SignalHandler(int) + 270
3 libsystem_platform.dylib 0x00007ff81913037d _sigtramp + 29
4 swift-frontend 0x000000010b27d24a swift::CanType::CanType(swift::TypeBase*) + 42
5 libsystem_c.dylib 0x00007ff819021a4d abort + 126
6 libsystem_c.dylib 0x00007ff819020d60 err + 0
7 swift-frontend 0x000000010bd223f6 (anonymous namespace)::ParamLowering::claimParams(swift::Lowering::AbstractionPattern, llvm::ArrayRef<swift::AnyFunctionType::Param>, swift::ForeignInfo const&) + 566
8 swift-frontend 0x000000010bd2146a (anonymous namespace)::CallSite::emit(swift::Lowering::SILGenFunction&, swift::Lowering::AbstractionPattern, swift::CanTypeWrapper<swift::SILFunctionType>, (anonymous namespace)::ParamLowering&, llvm::SmallVectorImpl<swift::Lowering::ManagedValue>&, llvm::SmallVectorImpl<(anonymous namespace)::DelayedArgument>&, swift::ForeignInfo const&) && + 170
9 swift-frontend 0x000000010bd2054e (anonymous namespace)::CallEmission::emitArgumentsForNormalApply(swift::Lowering::AbstractionPattern, swift::CanTypeWrapper<swift::SILFunctionType>, swift::ForeignInfo const&, llvm::SmallVectorImpl<swift::Lowering::ManagedValue>&, llvm::Optional<swift::SILLocation>&) + 1630
10 swift-frontend 0x000000010bd1f890 (anonymous namespace)::CallEmission::applyNormalCall(swift::Lowering::SGFContext) + 1120
11 swift-frontend 0x000000010bd1ddcd (anonymous namespace)::CallEmission::applyFirstLevelCallee(swift::Lowering::SGFContext) + 205
12 swift-frontend 0x000000010bcee740 (anonymous namespace)::CallEmission::apply(swift::Lowering::SGFContext) + 80
13 swift-frontend 0x000000010bcee0c8 swift::Lowering::SILGenFunction::emitApplyExpr(swift::ApplyExpr*, swift::Lowering::SGFContext) + 136
14 swift-frontend 0x000000010bdf09ab (anonymous namespace)::RValueEmitter::visitApplyExpr(swift::ApplyExpr*, swift::Lowering::SGFContext) + 59
15 swift-frontend 0x000000010bdde0e8 swift::ASTVisitor<(anonymous namespace)::RValueEmitter, swift::Lowering::RValue, void, void, void, void, void, swift::Lowering::SGFContext>::visitCallExpr(swift::CallExpr*, swift::Lowering::SGFContext) + 56
16 swift-frontend 0x000000010bdc78bc swift::ASTVisitor<(anonymous namespace)::RValueEmitter, swift::Lowering::RValue, void, void, void, void, void, swift::Lowering::SGFContext>::visit(swift::Expr*, swift::Lowering::SGFContext) + 2956
17 swift-frontend 0x000000010bdb8d23 swift::Lowering::SILGenFunction::emitRValue(swift::Expr*, swift::Lowering::SGFContext) + 195
18 swift-frontend 0x000000010bdbd9e6 swift::Lowering::SILGenFunction::emitIgnoredExpr(swift::Expr*) + 1990
19 swift-frontend 0x000000010bef76d9 (anonymous namespace)::StmtEmitter::visitBraceStmt(swift::BraceStmt*) + 2233
20 swift-frontend 0x000000010bef3181 swift::ASTVisitor<(anonymous namespace)::StmtEmitter, void, void, void, void, void, void>::visit(swift::Stmt*) + 81
21 swift-frontend 0x000000010bef30fa swift::Lowering::SILGenFunction::emitStmt(swift::Stmt*) + 42
22 swift-frontend 0x000000010be0f55a swift::Lowering::SILGenFunction::emitFunction(swift::FuncDecl*) + 842
23 swift-frontend 0x000000010bcc452b swift::Lowering::SILGenModule::emitFunctionDefinition(swift::SILDeclRef, swift::SILFunction*) + 4411
24 swift-frontend 0x000000010bcc7a75 swift::Lowering::SILGenModule::emitOrDelayFunction(swift::SILDeclRef) + 501
25 swift-frontend 0x000000010bcc33e5 swift::Lowering::SILGenModule::emitFunction(swift::FuncDecl*) + 309
26 swift-frontend 0x000000010bcc329d swift::Lowering::SILGenModule::visitFuncDecl(swift::FuncDecl*) + 29
27 swift-frontend 0x000000010bcc2f64 swift::ASTVisitor<swift::Lowering::SILGenModule, void, void, void, void, void, void>::visit(swift::Decl*) + 996
28 swift-frontend 0x000000010bcc2b6b swift::Lowering::SILGenModule::visit(swift::Decl*) + 59
29 swift-frontend 0x000000010bccc72a (anonymous namespace)::SILGenModuleRAII::emitSourceFile(swift::SourceFile*) + 314
30 swift-frontend 0x000000010bccc2e4 swift::ASTLoweringRequest::evaluate(swift::Evaluator&, swift::ASTLoweringDescriptor) const + 916
31 swift-frontend 0x000000010bef22de std::__1::unique_ptr<swift::SILModule, std::__1::default_delete<swift::SILModule>> swift::SimpleRequest<swift::ASTLoweringRequest, std::__1::unique_ptr<swift::SILModule, std::__1::default_delete<swift::SILModule>> (swift::ASTLoweringDescriptor), (swift::RequestFlags)9>::callDerived<0ul>(swift::Evaluator&, std::__1::integer_sequence<unsigned long, 0ul>) const + 142
32 swift-frontend 0x000000010bef21f8 swift::SimpleRequest<swift::ASTLoweringRequest, std::__1::unique_ptr<swift::SILModule, std::__1::default_delete<swift::SILModule>> (swift::ASTLoweringDescriptor), (swift::RequestFlags)9>::evaluateRequest(swift::ASTLoweringRequest const&, swift::Evaluator&) + 40
33 swift-frontend 0x000000010bcdeddc llvm::Expected<swift::ASTLoweringRequest::OutputType> swift::Evaluator::getResultUncached<swift::ASTLoweringRequest>(swift::ASTLoweringRequest const&) + 316
34 swift-frontend 0x000000010bcccee8 llvm::Expected<swift::ASTLoweringRequest::OutputType> swift::Evaluator::operator()<swift::ASTLoweringRequest, (void*)0>(swift::ASTLoweringRequest const&) + 40
35 swift-frontend 0x000000010bcccde0 swift::performASTLowering(swift::ModuleDecl*, swift::Lowering::TypeConverter&, swift::SILOptions const&, swift::IRGenOptions const*) + 256
36 swift-frontend 0x000000010b010cfb swift::performCompileStepsPostSema(swift::CompilerInstance&, int&, swift::FrontendObserver*) + 283
37 swift-frontend 0x000000010b040003 performAction(swift::CompilerInstance&, int&, swift::FrontendObserver*)::$_28::operator()(swift::CompilerInstance&) const + 147
38 swift-frontend 0x000000010b03ff5d bool llvm::function_ref<bool (swift::CompilerInstance&)>::callback_fn<performAction(swift::CompilerInstance&, int&, swift::FrontendObserver*)::$_28>(long, swift::CompilerInstance&) + 29
39 swift-frontend 0x000000010b03f421 llvm::function_ref<bool (swift::CompilerInstance&)>::operator()(swift::CompilerInstance&) const + 33
40 swift-frontend 0x000000010b03e438 withSemanticAnalysis(swift::CompilerInstance&, swift::FrontendObserver*, llvm::function_ref<bool (swift::CompilerInstance&)>, bool) + 392
41 swift-frontend 0x000000010b036f26 performAction(swift::CompilerInstance&, int&, swift::FrontendObserver*) + 1222
42 swift-frontend 0x000000010b014186 performCompile(swift::CompilerInstance&, int&, swift::FrontendObserver*) + 262
43 swift-frontend 0x000000010b012b60 swift::performFrontend(llvm::ArrayRef<char const*>, char const*, void*, swift::FrontendObserver*) + 2816
44 swift-frontend 0x000000010aee53df run_driver(llvm::StringRef, llvm::ArrayRef<char const*>, llvm::ArrayRef<char const*>) + 2431
45 swift-frontend 0x000000010aee4057 swift::mainEntry(int, char const**) + 1543
46 swift-frontend 0x000000010aee36a2 main + 34
47 dyld 0x00007ff818d7a3a6 start + 1942
Abort trap: 6
This no longer crashes on top of tree; please verify.
This regressed at some point, and is reproducible with the Swift 5.10 toolchain included with Xcode 15.3 (15E204a).
swift-driver version: 1.90.11.1 Apple Swift version 5.10 (swiftlang-5.10.0.13 clang-1500.3.9.4)
Target: arm64-apple-macosx14.0
Stack dumpStack dump:
0. Program arguments: /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/swift-frontend -frontend -interpret Swift69313.swift -Xllvm -aarch64-use-tbi -enable-objc-interop -stack-check -sdk /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX14.4.sdk -color-diagnostics -new-driver-path /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/swift-driver -empty-abi-descriptor -resource-dir /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/lib/swift -module-name Swift69313 -disable-clang-spi -target-sdk-version 14.4 -target-sdk-name macosx14.4 -external-plugin-path /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX14.4.sdk/usr/lib/swift/host/plugins#/Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX14.4.sdk/usr/bin/swift-plugin-server -external-plugin-path /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX14.4.sdk/usr/local/lib/swift/host/plugins#/Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX14.4.sdk/usr/bin/swift-plugin-server -external-plugin-path /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/usr/lib/swift/host/plugins#/Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/usr/bin/swift-plugin-server -external-plugin-path /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/usr/local/lib/swift/host/plugins#/Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/usr/bin/swift-plugin-server -plugin-path /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/lib/swift/host/plugins -plugin-path /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/local/lib/swift/host/plugins
1. Apple Swift version 5.10 (swiftlang-5.10.0.13 clang-1500.3.9.4)
2. Compiling with the current language version
3. While evaluating request ASTLoweringRequest(Lowering AST to SIL for module Swift69313)
4. While silgen emitFunction SIL function "@$s10Swift693133fooyyF".
for 'foo()' (at Swift69313.swift:5:1)
Stack dump without symbol names (ensure you have llvm-symbolizer in your PATH or set the environment var `LLVM_SYMBOLIZER_PATH` to point to it):
0 swift-frontend 0x00000001063bbf3c llvm::sys::PrintStackTrace(llvm::raw_ostream&, int) + 56
1 swift-frontend 0x00000001063bb0f8 llvm::sys::RunSignalHandlers() + 112
2 swift-frontend 0x00000001063bc544 SignalHandler(int) + 360
3 libsystem_platform.dylib 0x000000019ded3584 _sigtramp + 56
4 swift-frontend 0x00000001012fbe9c (anonymous namespace)::ArgEmitter::emitPackArg(llvm::MutableArrayRef, swift::Lowering::AbstractionPattern) + 168
5 swift-frontend 0x00000001012f979c (anonymous namespace)::ArgEmitter::emitPreparedArgs(swift::Lowering::PreparedArguments&&, swift::Lowering::AbstractionPattern) + 512
6 swift-frontend 0x0000000101305fb4 (anonymous namespace)::CallSite::emit(swift::Lowering::SILGenFunction&, swift::Lowering::AbstractionPattern, swift::CanTypeWrapper, (anonymous namespace)::ParamLowering&, llvm::SmallVectorImpl&, llvm::SmallVectorImpl<(anonymous namespace)::DelayedArgument>&, swift::ForeignInfo const&) && + 672
7 swift-frontend 0x00000001013059cc (anonymous namespace)::CallEmission::emitArgumentsForNormalApply(swift::Lowering::AbstractionPattern, swift::CanTypeWrapper, swift::ForeignInfo const&, llvm::SmallVectorImpl&, llvm::Optional&) + 1000
8 swift-frontend 0x00000001012edd68 (anonymous namespace)::CallEmission::apply(swift::Lowering::SGFContext) + 4064
9 swift-frontend 0x00000001012eb7bc swift::Lowering::SILGenFunction::emitApplyExpr(swift::ApplyExpr*, swift::Lowering::SGFContext) + 3060
10 swift-frontend 0x0000000101357618 swift::Lowering::SILGenFunction::emitIgnoredExpr(swift::Expr*) + 892
11 swift-frontend 0x00000001013ef684 swift::ASTVisitor<(anonymous namespace)::StmtEmitter, void, void, void, void, void, void>::visit(swift::Stmt*) + 5032
12 swift-frontend 0x0000000101386538 swift::Lowering::SILGenFunction::emitFunction(swift::FuncDecl*) + 632
13 swift-frontend 0x00000001012dae48 swift::Lowering::SILGenModule::emitFunctionDefinition(swift::SILDeclRef, swift::SILFunction*) + 8344
14 swift-frontend 0x00000001012db5bc swift::Lowering::SILGenModule::emitOrDelayFunction(swift::SILDeclRef) + 172
15 swift-frontend 0x00000001012d8d9c swift::Lowering::SILGenModule::emitFunction(swift::FuncDecl*) + 292
16 swift-frontend 0x00000001012deb54 swift::ASTLoweringRequest::evaluate(swift::Evaluator&, swift::ASTLoweringDescriptor) const + 1668
17 swift-frontend 0x00000001013edf1c swift::SimpleRequest> (swift::ASTLoweringDescriptor), (swift::RequestFlags)9>::evaluateRequest(swift::ASTLoweringRequest const&, swift::Evaluator&) + 196
18 swift-frontend 0x00000001012e1f44 llvm::Expected swift::Evaluator::getResultUncached(swift::ASTLoweringRequest const&) + 584
19 swift-frontend 0x00000001012df854 swift::performASTLowering(swift::ModuleDecl*, swift::Lowering::TypeConverter&, swift::SILOptions const&, swift::IRGenOptions const*) + 152
20 swift-frontend 0x0000000100cb99f8 swift::performCompileStepsPostSema(swift::CompilerInstance&, int&, swift::FrontendObserver*) + 2068
21 swift-frontend 0x0000000100cbdf0c performCompile(swift::CompilerInstance&, int&, swift::FrontendObserver*) + 3616
22 swift-frontend 0x0000000100cbb6d0 swift::performFrontend(llvm::ArrayRef, char const*, void*, swift::FrontendObserver*) + 4968
23 swift-frontend 0x0000000100c4ae8c swift::mainEntry(int, char const**) + 2612
24 dyld 0x000000019db1a0e0 start + 2360
fish: Job 1, 'xcrun swift Swift69313.swift' terminated by signal SIGSEGV (Address boundary error)
@rjmccall What Swift version was at the top of the tree at this point? I'm seeing this the Swift toolchain that shipped with the most recent stable Xcode release. (5.10)
Ah makes sense. Thanks!
|
gharchive/issue
| 2023-10-21T01:01:47 |
2025-04-01T04:56:01.145974
|
{
"authors": [
"AnthonyLatsis",
"rjmccall",
"slice",
"twof"
],
"repo": "apple/swift",
"url": "https://github.com/apple/swift/issues/69313",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
302408091
|
Add a cmake flag to the benchmarks: SWIFT_BENCHMARK_EXTRA_FLAGS.
We used to have this cmake flag, but it seems to have disappeared. Building the
benchmarks with different swiftc flags is central to performance analysis, so
I'm not sure how people were getting this done.
@swift-ci smoke test.
@swift-ci smoke benchmark
Build comment file:
Optimized (O)
Regression (3)
TEST
OLD
NEW
DELTA
SPEEDUP
StaticArray
9
11
+22.2%
0.82x
DropLastCountableRangeLazy
11
12
+9.1%
0.92x
ObjectiveCBridgeStubFromArrayOfNSString
27825
29699
+6.7%
0.94x (?)
Improvement (3)
TEST
OLD
NEW
DELTA
SPEEDUP
ObjectiveCBridgeFromNSArrayAnyObject
22371
20673
-7.6%
1.08x (?)
StringBuilderLong
1124
1053
-6.3%
1.07x (?)
DictionaryBridge
2127
2019
-5.1%
1.05x (?)
No Changes (372)
TEST
OLD
NEW
DELTA
SPEEDUP
AngryPhonebook
3645
3706
+1.7%
0.98x (?)
AnyHashableWithAClass
73443
73565
+0.2%
1.00x (?)
Array2D
2555
2556
+0.0%
1.00x (?)
ArrayAppend
1023
1016
-0.7%
1.01x (?)
ArrayAppendArrayOfInt
797
799
+0.3%
1.00x (?)
ArrayAppendAscii
13239
13247
+0.1%
1.00x (?)
ArrayAppendFromGeneric
797
800
+0.4%
1.00x (?)
ArrayAppendGenericStructs
1427
1443
+1.1%
0.99x
ArrayAppendLatin1
38072
38111
+0.1%
1.00x
ArrayAppendLazyMap
1341
1343
+0.1%
1.00x (?)
ArrayAppendOptionals
1429
1431
+0.1%
1.00x (?)
ArrayAppendRepeatCol
1337
1339
+0.1%
1.00x (?)
ArrayAppendReserved
756
756
+0.0%
1.00x
ArrayAppendSequence
1130
1134
+0.4%
1.00x (?)
ArrayAppendStrings
15243
15253
+0.1%
1.00x (?)
ArrayAppendToFromGeneric
806
799
-0.9%
1.01x (?)
ArrayAppendToGeneric
805
802
-0.4%
1.00x (?)
ArrayAppendUTF16
37154
37111
-0.1%
1.00x (?)
ArrayInClass
86
85
-1.2%
1.01x
ArrayLiteral
0
0
+0.0%
1.00x
ArrayOfGenericPOD2
151
151
+0.0%
1.00x
ArrayOfGenericRef
4438
4444
+0.1%
1.00x (?)
ArrayOfPOD
183
183
+0.0%
1.00x
ArrayOfRef
4386
4443
+1.3%
0.99x (?)
ArrayPlusEqualArrayOfInt
797
798
+0.1%
1.00x (?)
ArrayPlusEqualFiveElementCollection
5530
5518
-0.2%
1.00x (?)
ArrayPlusEqualSingleElementCollection
1023
1021
-0.2%
1.00x (?)
ArrayPlusEqualThreeElements
1657
1654
-0.2%
1.00x (?)
ArraySubscript
1536
1533
-0.2%
1.00x (?)
ArrayValueProp
8
8
+0.0%
1.00x
ArrayValueProp2
8
8
+0.0%
1.00x
ArrayValueProp3
8
8
+0.0%
1.00x
ArrayValueProp4
8
8
+0.0%
1.00x
BinaryFloatingPointConversionFromBinaryInteger
41
41
+0.0%
1.00x
BitCount
202
202
+0.0%
1.00x
ByteSwap
99
99
+0.0%
1.00x
COWTree
5506
5560
+1.0%
0.99x (?)
CSVParsing
756043
762200
+0.8%
0.99x (?)
CSVParsingAlt
731897
738938
+1.0%
0.99x
CSVParsingAltIndices
368011
358621
-2.6%
1.03x
CStringLongAscii
4774
4897
+2.6%
0.97x
CStringLongNonAscii
2166
2193
+1.2%
0.99x
CStringShortAscii
4740
4746
+0.1%
1.00x (?)
Calculator
649
647
-0.3%
1.00x (?)
CaptureProp
3914
3920
+0.2%
1.00x
CharIndexing_ascii_unicodeScalars
14738
14733
-0.0%
1.00x (?)
CharIndexing_ascii_unicodeScalars_Backwards
14251
14246
-0.0%
1.00x (?)
CharIndexing_chinese_unicodeScalars
11172
11163
-0.1%
1.00x (?)
CharIndexing_chinese_unicodeScalars_Backwards
10793
10793
+0.0%
1.00x
CharIndexing_japanese_unicodeScalars
17649
17652
+0.0%
1.00x (?)
CharIndexing_japanese_unicodeScalars_Backwards
17071
17059
-0.1%
1.00x (?)
CharIndexing_korean_unicodeScalars
14293
14284
-0.1%
1.00x (?)
CharIndexing_korean_unicodeScalars_Backwards
13814
13817
+0.0%
1.00x (?)
CharIndexing_punctuatedJapanese_unicodeScalars
2680
2678
-0.1%
1.00x (?)
CharIndexing_punctuatedJapanese_unicodeScalars_Backwards
2592
2589
-0.1%
1.00x
CharIndexing_punctuated_unicodeScalars
3349
3348
-0.0%
1.00x (?)
CharIndexing_punctuated_unicodeScalars_Backwards
3239
3237
-0.1%
1.00x (?)
CharIndexing_russian_unicodeScalars
12289
12285
-0.0%
1.00x (?)
CharIndexing_russian_unicodeScalars_Backwards
11883
11875
-0.1%
1.00x (?)
CharIndexing_tweet_unicodeScalars
29086
29072
-0.0%
1.00x (?)
CharIndexing_tweet_unicodeScalars_Backwards
27623
27632
+0.0%
1.00x (?)
CharIndexing_utf16_unicodeScalars
20210
20221
+0.1%
1.00x (?)
CharIndexing_utf16_unicodeScalars_Backwards
20844
20845
+0.0%
1.00x (?)
CharIteration_ascii_unicodeScalars
20420
20436
+0.1%
1.00x (?)
CharIteration_ascii_unicodeScalars_Backwards
12706
12684
-0.2%
1.00x
CharIteration_chinese_unicodeScalars
15440
15453
+0.1%
1.00x
CharIteration_chinese_unicodeScalars_Backwards
9638
9641
+0.0%
1.00x (?)
CharIteration_japanese_unicodeScalars
24439
24463
+0.1%
1.00x (?)
CharIteration_japanese_unicodeScalars_Backwards
15179
15180
+0.0%
1.00x (?)
CharIteration_korean_unicodeScalars
19818
19811
-0.0%
1.00x (?)
CharIteration_korean_unicodeScalars_Backwards
12319
12319
+0.0%
1.00x
CharIteration_punctuatedJapanese_unicodeScalars
3644
3644
+0.0%
1.00x
CharIteration_punctuatedJapanese_unicodeScalars_Backwards
2341
2340
-0.0%
1.00x (?)
CharIteration_punctuated_unicodeScalars
4579
4577
-0.0%
1.00x (?)
CharIteration_punctuated_unicodeScalars_Backwards
2916
2915
-0.0%
1.00x (?)
CharIteration_russian_unicodeScalars
17031
17017
-0.1%
1.00x
CharIteration_russian_unicodeScalars_Backwards
10605
10600
-0.0%
1.00x
CharIteration_tweet_unicodeScalars
40358
40352
-0.0%
1.00x (?)
CharIteration_tweet_unicodeScalars_Backwards
25055
25040
-0.1%
1.00x (?)
CharIteration_utf16_unicodeScalars
27072
27053
-0.1%
1.00x (?)
CharIteration_utf16_unicodeScalars_Backwards
15869
15877
+0.1%
1.00x (?)
CharacterLiteralsLarge
6008
6038
+0.5%
1.00x (?)
CharacterLiteralsSmall
410
411
+0.2%
1.00x (?)
CharacterPropertiesFetch
4682
4697
+0.3%
1.00x (?)
CharacterPropertiesPrecomputed
961
973
+1.2%
0.99x
CharacterPropertiesStashed
1548
1566
+1.2%
0.99x
CharacterPropertiesStashedMemo
1517
1533
+1.1%
0.99x
Chars
969
970
+0.1%
1.00x (?)
ClassArrayGetter
15
15
+0.0%
1.00x
Combos
463
463
+0.0%
1.00x
DictOfArraysToArrayOfDicts
791
803
+1.5%
0.99x (?)
Dictionary
507
503
-0.8%
1.01x
Dictionary2
1618
1614
-0.2%
1.00x (?)
Dictionary2OfObjects
2977
2980
+0.1%
1.00x (?)
Dictionary3
234
234
+0.0%
1.00x
Dictionary3OfObjects
682
681
-0.1%
1.00x (?)
Dictionary4
147
147
+0.0%
1.00x
Dictionary4OfObjects
200
200
+0.0%
1.00x
DictionaryGroup
133
132
-0.8%
1.01x
DictionaryGroupOfObjects
1957
1958
+0.1%
1.00x (?)
DictionaryLiteral
1590
1586
-0.3%
1.00x (?)
DictionaryOfObjects
2182
2171
-0.5%
1.01x (?)
DictionaryRemove
2927
2990
+2.2%
0.98x (?)
DictionaryRemoveOfObjects
24034
24002
-0.1%
1.00x (?)
DictionarySubscriptDefaultMutation
143
143
+0.0%
1.00x
DictionarySubscriptDefaultMutationArray
493
494
+0.2%
1.00x (?)
DictionarySubscriptDefaultMutationArrayOfObjects
3812
3817
+0.1%
1.00x (?)
DictionarySubscriptDefaultMutationOfObjects
1397
1400
+0.2%
1.00x (?)
DictionarySwap
471
471
+0.0%
1.00x
DictionarySwapOfObjects
7468
7462
-0.1%
1.00x (?)
DoubleWidthDivision
0
0
+0.0%
1.00x
DropFirstAnyCollection
102
101
-1.0%
1.01x
DropFirstAnyCollectionLazy
67046
67032
-0.0%
1.00x (?)
DropFirstAnySeqCRangeIter
19580
19583
+0.0%
1.00x (?)
DropFirstAnySeqCRangeIterLazy
19588
19558
-0.2%
1.00x (?)
DropFirstAnySeqCntRange
94
94
+0.0%
1.00x
DropFirstAnySeqCntRangeLazy
94
94
+0.0%
1.00x
DropFirstAnySequence
5082
5082
+0.0%
1.00x
DropFirstAnySequenceLazy
5084
5081
-0.1%
1.00x (?)
DropFirstArray
35
35
+0.0%
1.00x
DropFirstArrayLazy
35
35
+0.0%
1.00x
DropFirstCountableRange
35
35
+0.0%
1.00x
DropFirstCountableRangeLazy
35
35
+0.0%
1.00x
DropFirstSequence
2699
2698
-0.0%
1.00x (?)
DropFirstSequenceLazy
2785
2784
-0.0%
1.00x (?)
DropLastAnyCollection
37
37
+0.0%
1.00x
DropLastAnyCollectionLazy
22445
22381
-0.3%
1.00x (?)
DropLastAnySeqCRangeIter
3914
3914
+0.0%
1.00x
DropLastAnySeqCRangeIterLazy
3916
3908
-0.2%
1.00x (?)
DropLastAnySeqCntRange
30
29
-3.3%
1.03x
DropLastAnySeqCntRangeLazy
30
30
+0.0%
1.00x
DropLastAnySequence
5218
5210
-0.2%
1.00x
DropLastAnySequenceLazy
5321
5322
+0.0%
1.00x (?)
DropLastCountableRange
11
11
+0.0%
1.00x
DropLastSequence
642
640
-0.3%
1.00x (?)
DropLastSequenceLazy
642
642
+0.0%
1.00x
DropWhileAnyCollection
125
125
+0.0%
1.00x
DropWhileAnyCollectionLazy
166
166
+0.0%
1.00x
DropWhileAnySeqCRangeIter
16093
16069
-0.1%
1.00x (?)
DropWhileAnySeqCRangeIterLazy
166
166
+0.0%
1.00x
DropWhileAnySeqCntRange
118
118
+0.0%
1.00x
DropWhileAnySeqCntRangeLazy
166
166
+0.0%
1.00x
DropWhileAnySequence
5787
5788
+0.0%
1.00x (?)
DropWhileAnySequenceLazy
1858
1856
-0.1%
1.00x
DropWhileArrayLazy
129
129
+0.0%
1.00x
DropWhileCountableRange
36
36
+0.0%
1.00x
DropWhileCountableRangeLazy
111
111
+0.0%
1.00x
DropWhileSequence
1344
1343
-0.1%
1.00x (?)
DropWhileSequenceLazy
88
88
+0.0%
1.00x
EqualStringSubstring
48
48
+0.0%
1.00x
EqualSubstringString
48
48
+0.0%
1.00x
EqualSubstringSubstring
48
48
+0.0%
1.00x
EqualSubstringSubstringGenericEquatable
48
48
+0.0%
1.00x
ErrorHandling
2341
2330
-0.5%
1.00x (?)
ExclusivityGlobal
5
5
+0.0%
1.00x
ExclusivityIndependent
2
2
+0.0%
1.00x
FilterEvenUsingReduce
1333
1329
-0.3%
1.00x (?)
FilterEvenUsingReduceInto
151
150
-0.7%
1.01x (?)
FrequenciesUsingReduce
6653
6689
+0.5%
0.99x (?)
FrequenciesUsingReduceInto
3113
3106
-0.2%
1.00x (?)
Hanoi
3334
3328
-0.2%
1.00x (?)
HashTest
1765
1763
-0.1%
1.00x (?)
Histogram
321
321
+0.0%
1.00x
Integrate
334
335
+0.3%
1.00x
IterateData
1357
1353
-0.3%
1.00x
Join
372
371
-0.3%
1.00x
LazilyFilteredArrayContains
36273
36259
-0.0%
1.00x (?)
LazilyFilteredArrays
66106
66016
-0.1%
1.00x (?)
LazilyFilteredRange
3919
3852
-1.7%
1.02x (?)
LessSubstringSubstring
47
48
+2.1%
0.98x
LessSubstringSubstringGenericComparable
47
47
+0.0%
1.00x
LinkedList
7521
7522
+0.0%
1.00x (?)
LuhnAlgoEager
610
610
+0.0%
1.00x
LuhnAlgoLazy
611
614
+0.5%
1.00x (?)
MapReduce
400
398
-0.5%
1.01x
MapReduceAnyCollection
430
429
-0.2%
1.00x
MapReduceAnyCollectionShort
2243
2245
+0.1%
1.00x (?)
MapReduceClass
3049
3045
-0.1%
1.00x
MapReduceClassShort
4582
4595
+0.3%
1.00x
MapReduceLazyCollection
13
13
+0.0%
1.00x
MapReduceLazyCollectionShort
37
37
+0.0%
1.00x
MapReduceLazySequence
86
86
+0.0%
1.00x
MapReduceSequence
460
456
-0.9%
1.01x
MapReduceShort
2030
2027
-0.1%
1.00x (?)
MapReduceShortString
21
21
+0.0%
1.00x
MapReduceString
83
83
+0.0%
1.00x
Memset
222
215
-3.2%
1.03x (?)
MonteCarloE
10290
10267
-0.2%
1.00x
MonteCarloPi
43095
42670
-1.0%
1.01x
NSDictionaryCastToSwift
5606
5633
+0.5%
1.00x (?)
NSError
303
304
+0.3%
1.00x (?)
NSStringConversion
306
306
+0.0%
1.00x
NibbleSort
3712
3710
-0.1%
1.00x (?)
NopDeinit
21597
21576
-0.1%
1.00x (?)
ObjectAllocation
185
184
-0.5%
1.01x (?)
ObjectiveCBridgeFromNSArrayAnyObjectForced
4344
4534
+4.4%
0.96x (?)
ObjectiveCBridgeFromNSArrayAnyObjectToString
38750
40417
+4.3%
0.96x (?)
ObjectiveCBridgeFromNSArrayAnyObjectToStringForced
36634
37599
+2.6%
0.97x (?)
ObjectiveCBridgeFromNSDictionaryAnyObject
124775
125322
+0.4%
1.00x (?)
ObjectiveCBridgeFromNSSetAnyObject
58638
60157
+2.6%
0.97x (?)
ObjectiveCBridgeFromNSSetAnyObjectForced
4484
4675
+4.3%
0.96x (?)
ObjectiveCBridgeFromNSSetAnyObjectToString
69796
67683
-3.0%
1.03x (?)
ObjectiveCBridgeFromNSString
1311
1316
+0.4%
1.00x (?)
ObjectiveCBridgeFromNSStringForced
2441
2474
+1.4%
0.99x (?)
ObjectiveCBridgeStubDataAppend
4123
3942
-4.4%
1.05x (?)
ObjectiveCBridgeStubDateMutation
258
257
-0.4%
1.00x
ObjectiveCBridgeStubFromNSDate
4132
4131
-0.0%
1.00x
ObjectiveCBridgeStubFromNSString
861
864
+0.3%
1.00x (?)
ObjectiveCBridgeStubFromNSStringRef
161
161
+0.0%
1.00x
ObjectiveCBridgeStubNSDataAppend
2521
2511
-0.4%
1.00x (?)
ObjectiveCBridgeStubNSDateMutationRef
13132
13205
+0.6%
0.99x (?)
ObjectiveCBridgeStubToArrayOfNSString
28821
29526
+2.4%
0.98x (?)
ObjectiveCBridgeStubToNSDate
15324
15568
+1.6%
0.98x (?)
ObjectiveCBridgeStubToNSDateRef
3478
3492
+0.4%
1.00x (?)
ObjectiveCBridgeStubToNSString
1519
1519
+0.0%
1.00x
ObjectiveCBridgeStubToNSStringRef
115
115
+0.0%
1.00x
ObjectiveCBridgeStubURLAppendPath
334340
334017
-0.1%
1.00x (?)
ObjectiveCBridgeStubURLAppendPathRef
340455
332243
-2.4%
1.02x
ObjectiveCBridgeToNSArray
29360
29455
+0.3%
1.00x (?)
ObjectiveCBridgeToNSDictionary
49334
48748
-1.2%
1.01x (?)
ObjectiveCBridgeToNSSet
42089
42048
-0.1%
1.00x (?)
ObjectiveCBridgeToNSString
1291
1291
+0.0%
1.00x
ObserverClosure
2155
2169
+0.6%
0.99x (?)
ObserverForwarderStruct
1008
998
-1.0%
1.01x (?)
ObserverPartiallyAppliedMethod
3761
3739
-0.6%
1.01x (?)
ObserverUnappliedMethod
2366
2372
+0.3%
1.00x (?)
OpenClose
123
123
+0.0%
1.00x
PartialApplyDynamicType
0
0
+0.0%
1.00x
Phonebook
3881
3882
+0.0%
1.00x
PointerArithmetics
31517
31497
-0.1%
1.00x
PolymorphicCalls
25
25
+0.0%
1.00x
PopFrontArray
1945
1946
+0.1%
1.00x (?)
PopFrontArrayGeneric
1970
1961
-0.5%
1.00x (?)
PopFrontUnsafePointer
9486
9484
-0.0%
1.00x (?)
PrefixAnyCollection
102
101
-1.0%
1.01x
PrefixAnyCollectionLazy
67014
67024
+0.0%
1.00x (?)
PrefixAnySeqCRangeIter
15252
15245
-0.0%
1.00x (?)
PrefixAnySeqCRangeIterLazy
15259
15246
-0.1%
1.00x (?)
PrefixAnySeqCntRange
94
94
+0.0%
1.00x
PrefixAnySeqCntRangeLazy
94
94
+0.0%
1.00x
PrefixAnySequence
4497
4496
-0.0%
1.00x (?)
PrefixAnySequenceLazy
4483
4491
+0.2%
1.00x
PrefixArray
35
35
+0.0%
1.00x
PrefixArrayLazy
35
35
+0.0%
1.00x
PrefixCountableRange
35
35
+0.0%
1.00x
PrefixCountableRangeLazy
35
35
+0.0%
1.00x
PrefixSequence
1324
1326
+0.2%
1.00x (?)
PrefixSequenceLazy
1413
1414
+0.1%
1.00x (?)
PrefixWhileAnyCollection
172
173
+0.6%
0.99x
PrefixWhileAnyCollectionLazy
108
108
+0.0%
1.00x
PrefixWhileAnySeqCRangeIter
9644
9626
-0.2%
1.00x (?)
PrefixWhileAnySeqCRangeIterLazy
90
90
+0.0%
1.00x
PrefixWhileAnySeqCntRange
165
165
+0.0%
1.00x
PrefixWhileAnySeqCntRangeLazy
108
108
+0.0%
1.00x
PrefixWhileAnySequence
10801
10779
-0.2%
1.00x (?)
PrefixWhileAnySequenceLazy
1394
1394
+0.0%
1.00x
PrefixWhileArray
105
105
+0.0%
1.00x
PrefixWhileArrayLazy
70
70
+0.0%
1.00x
PrefixWhileCountableRange
46
46
+0.0%
1.00x
PrefixWhileCountableRangeLazy
35
35
+0.0%
1.00x
PrefixWhileSequence
397
395
-0.5%
1.01x (?)
PrefixWhileSequenceLazy
52
52
+0.0%
1.00x
Prims
775
770
-0.6%
1.01x (?)
PrimsSplit
767
767
+0.0%
1.00x
QueueConcrete
1305
1307
+0.2%
1.00x (?)
QueueGeneric
1158
1157
-0.1%
1.00x
RC4
171
171
+0.0%
1.00x
RGBHistogram
2214
2204
-0.5%
1.00x (?)
RGBHistogramOfObjects
23815
23773
-0.2%
1.00x (?)
RangeAssignment
356
356
+0.0%
1.00x
RangeIterationSigned
171
171
+0.0%
1.00x
RangeIterationSigned64
200
200
+0.0%
1.00x
RangeIterationUnsigned
200
200
+0.0%
1.00x
RangeReplaceableCollectionPlusDefault
977
988
+1.1%
0.99x (?)
RecursiveOwnedParameter
2347
2348
+0.0%
1.00x (?)
RemoveWhereFilterInts
46
46
+0.0%
1.00x
RemoveWhereFilterString
294
294
+0.0%
1.00x
RemoveWhereFilterStrings
432
432
+0.0%
1.00x
RemoveWhereMoveInts
17
17
+0.0%
1.00x
RemoveWhereMoveStrings
519
519
+0.0%
1.00x
RemoveWhereQuadraticInts
1289
1286
-0.2%
1.00x (?)
RemoveWhereQuadraticString
206
206
+0.0%
1.00x
RemoveWhereQuadraticStrings
2758
2760
+0.1%
1.00x (?)
RemoveWhereSwapInts
19
19
+0.0%
1.00x
RemoveWhereSwapStrings
848
849
+0.1%
1.00x (?)
ReversedArray
57
57
+0.0%
1.00x
ReversedBidirectional
17391
17366
-0.1%
1.00x (?)
ReversedDictionary
134
135
+0.7%
0.99x (?)
RomanNumbers
114190
119878
+5.0%
0.95x (?)
SequenceAlgosAnySequence
10988
11027
+0.4%
1.00x (?)
SequenceAlgosArray
1579
1579
+0.0%
1.00x
SequenceAlgosContiguousArray
1584
1588
+0.3%
1.00x (?)
SequenceAlgosList
1351
1351
+0.0%
1.00x
SequenceAlgosRange
4297
4294
-0.1%
1.00x
SequenceAlgosUnfoldSequence
1076
1077
+0.1%
1.00x
SetExclusiveOr
3452
3440
-0.3%
1.00x (?)
SetExclusiveOr_OfObjects
8915
8892
-0.3%
1.00x (?)
SetIntersect
319
319
+0.0%
1.00x
SetIntersect_OfObjects
1762
1766
+0.2%
1.00x (?)
SetIsSubsetOf
298
298
+0.0%
1.00x
SetIsSubsetOf_OfObjects
363
363
+0.0%
1.00x
SetUnion
3121
3149
+0.9%
0.99x
SetUnion_OfObjects
7252
7283
+0.4%
1.00x (?)
SevenBoom
1517
1517
+0.0%
1.00x
Sim2DArray
599
599
+0.0%
1.00x
SortLargeExistentials
6667
6678
+0.2%
1.00x
SortLettersInPlace
1124
1124
+0.0%
1.00x
SortSortedStrings
1015
1014
-0.1%
1.00x (?)
SortStrings
1820
1820
+0.0%
1.00x
SortStringsUnicode
2515
2519
+0.2%
1.00x (?)
StackPromo
22257
22275
+0.1%
1.00x (?)
StrComplexWalk
1559
1559
+0.0%
1.00x
StrToInt
2113
2121
+0.4%
1.00x (?)
StringAdder
4195
4190
-0.1%
1.00x (?)
StringBuilder
1395
1397
+0.1%
1.00x (?)
StringComparison_abnormal
741
741
+0.0%
1.00x
StringComparison_ascii
991
991
+0.0%
1.00x
StringComparison_emoji
794
777
-2.1%
1.02x (?)
StringComparison_fastPrenormal
732
731
-0.1%
1.00x (?)
StringComparison_latin1
553
552
-0.2%
1.00x
StringComparison_longSharedPrefix
899
895
-0.4%
1.00x (?)
StringComparison_nonBMPSlowestPrenormal
1560
1531
-1.9%
1.02x
StringComparison_slowerPrenormal
1668
1618
-3.0%
1.03x
StringComparison_zalgo
113739
113421
-0.3%
1.00x
StringEdits
118166
117547
-0.5%
1.01x (?)
StringEnumRawValueInitialization
965
958
-0.7%
1.01x (?)
StringEqualPointerComparison
286
286
+0.0%
1.00x
StringFromLongWholeSubstring
21
21
+0.0%
1.00x
StringFromLongWholeSubstringGeneric
10
10
+0.0%
1.00x
StringHasPrefixAscii
1288
1288
+0.0%
1.00x
StringHasPrefixUnicode
27676
27640
-0.1%
1.00x (?)
StringHasSuffixAscii
1431
1433
+0.1%
1.00x (?)
StringHasSuffixUnicode
84885
84865
-0.0%
1.00x (?)
StringInterpolation
10863
10962
+0.9%
0.99x (?)
StringMatch
7287
7286
-0.0%
1.00x (?)
StringRemoveDupes
1211
1202
-0.7%
1.01x (?)
StringUTF16Builder
2400
2413
+0.5%
0.99x (?)
StringWalk
1437
1437
+0.0%
1.00x
StringWithCString
43179
43165
-0.0%
1.00x (?)
StringWordBuilder
1703
1707
+0.2%
1.00x (?)
StringWordBuilderReservingCapacity
1270
1268
-0.2%
1.00x (?)
SubstringComparable
41
41
+0.0%
1.00x
SubstringEqualString
533
531
-0.4%
1.00x (?)
SubstringEquatable
1366
1372
+0.4%
1.00x (?)
SubstringFromLongString
10
10
+0.0%
1.00x
SubstringFromLongStringGeneric
69
69
+0.0%
1.00x
SuffixAnyCollection
43
43
+0.0%
1.00x
SuffixAnyCollectionLazy
22378
22416
+0.2%
1.00x (?)
SuffixAnySeqCRangeIter
4099
4111
+0.3%
1.00x (?)
SuffixAnySeqCRangeIterLazy
4109
4102
-0.2%
1.00x (?)
SuffixAnySeqCntRange
35
35
+0.0%
1.00x
SuffixAnySeqCntRangeLazy
35
35
+0.0%
1.00x
SuffixAnySequence
5215
5213
-0.0%
1.00x (?)
SuffixAnySequenceLazy
5298
5300
+0.0%
1.00x (?)
SuffixCountableRange
12
12
+0.0%
1.00x
SuffixCountableRangeLazy
12
12
+0.0%
1.00x
SuffixSequence
3897
3897
+0.0%
1.00x
SuffixSequenceLazy
3902
3897
-0.1%
1.00x (?)
SumUsingReduce
96
96
+0.0%
1.00x
SumUsingReduceInto
96
96
+0.0%
1.00x
SuperChars
41486
41321
-0.4%
1.00x (?)
TwoSum
986
991
+0.5%
0.99x
TypeFlood
0
0
+0.0%
1.00x
UTF8Decode
327
327
+0.0%
1.00x
Walsh
419
419
+0.0%
1.00x
WordCountHistogramASCII
7271
7290
+0.3%
1.00x (?)
WordCountHistogramUTF16
24377
25342
+4.0%
0.96x
WordCountUniqueASCII
1748
1818
+4.0%
0.96x
WordCountUniqueUTF16
16538
16611
+0.4%
1.00x (?)
WordSplitASCII
21693
21464
-1.1%
1.01x (?)
WordSplitUTF16
23161
23149
-0.1%
1.00x (?)
XorLoop
394
395
+0.3%
1.00x (?)
Unoptimized (Onone)
Regression (9)
TEST
OLD
NEW
DELTA
SPEEDUP
WordCountUniqueUTF16
26909
30099
+11.9%
0.89x (?)
ObjectiveCBridgeFromNSSetAnyObjectForced
7158
7887
+10.2%
0.91x (?)
CharacterPropertiesFetch
5927
6484
+9.4%
0.91x (?)
CharIndexing_chinese_unicodeScalars
275390
299901
+8.9%
0.92x (?)
ObjectiveCBridgeFromNSSetAnyObjectToString
75097
81138
+8.0%
0.93x (?)
DictionaryBridge
2085
2244
+7.6%
0.93x
CharIteration_tweet_unicodeScalars_Backwards
606174
648211
+6.9%
0.94x (?)
CharIteration_utf16_unicodeScalars
132997
140780
+5.9%
0.94x
StringUTF16Builder
7757
8208
+5.8%
0.95x
Improvement (6)
TEST
OLD
NEW
DELTA
SPEEDUP
CharIndexing_russian_unicodeScalars
338061
311345
-7.9%
1.09x (?)
ObjectiveCBridgeFromNSArrayAnyObjectForced
9440
8799
-6.8%
1.07x (?)
DropFirstAnyCollectionLazy
109972
102762
-6.6%
1.07x (?)
StringEdits
362663
339297
-6.4%
1.07x (?)
ObjectiveCBridgeStubFromArrayOfNSString
29611
27892
-5.8%
1.06x (?)
CharIndexing_chinese_unicodeScalars_Backwards
343159
326791
-4.8%
1.05x (?)
No Changes (363)
TEST
OLD
NEW
DELTA
SPEEDUP
AngryPhonebook
5152
5156
+0.1%
1.00x (?)
AnyHashableWithAClass
90360
90383
+0.0%
1.00x (?)
Array2D
631258
631335
+0.0%
1.00x (?)
ArrayAppend
4594
4586
-0.2%
1.00x (?)
ArrayAppendArrayOfInt
868
869
+0.1%
1.00x (?)
ArrayAppendAscii
40575
40594
+0.0%
1.00x (?)
ArrayAppendFromGeneric
878
876
-0.2%
1.00x
ArrayAppendGenericStructs
1509
1514
+0.3%
1.00x (?)
ArrayAppendLatin1
64174
64138
-0.1%
1.00x (?)
ArrayAppendLazyMap
178042
180063
+1.1%
0.99x (?)
ArrayAppendOptionals
1510
1514
+0.3%
1.00x (?)
ArrayAppendRepeatCol
187835
187719
-0.1%
1.00x (?)
ArrayAppendReserved
4280
4279
-0.0%
1.00x (?)
ArrayAppendSequence
157196
157016
-0.1%
1.00x (?)
ArrayAppendStrings
15365
15379
+0.1%
1.00x (?)
ArrayAppendToFromGeneric
872
874
+0.2%
1.00x (?)
ArrayAppendToGeneric
879
876
-0.3%
1.00x (?)
ArrayAppendUTF16
63630
63648
+0.0%
1.00x (?)
ArrayInClass
6256
6254
-0.0%
1.00x (?)
ArrayLiteral
1799
1801
+0.1%
1.00x (?)
ArrayOfGenericPOD2
1132
1135
+0.3%
1.00x (?)
ArrayOfGenericRef
10277
10291
+0.1%
1.00x (?)
ArrayOfPOD
842
842
+0.0%
1.00x
ArrayOfRef
9467
9510
+0.5%
1.00x (?)
ArrayPlusEqualArrayOfInt
873
871
-0.2%
1.00x (?)
ArrayPlusEqualFiveElementCollection
240989
241217
+0.1%
1.00x (?)
ArrayPlusEqualSingleElementCollection
236901
235678
-0.5%
1.01x (?)
ArrayPlusEqualThreeElements
9528
9503
-0.3%
1.00x (?)
ArraySubscript
116652
116829
+0.2%
1.00x (?)
ArrayValueProp
3697
3698
+0.0%
1.00x (?)
ArrayValueProp2
16783
16895
+0.7%
0.99x (?)
ArrayValueProp3
4174
4226
+1.2%
0.99x (?)
ArrayValueProp4
4162
4189
+0.6%
0.99x (?)
BinaryFloatingPointConversionFromBinaryInteger
6310
6240
-1.1%
1.01x (?)
BitCount
7874
7877
+0.0%
1.00x (?)
ByteSwap
9217
9215
-0.0%
1.00x (?)
COWTree
12282
11754
-4.3%
1.04x
CSVParsing
2500667
2491196
-0.4%
1.00x (?)
CSVParsingAlt
1316982
1324118
+0.5%
0.99x (?)
CSVParsingAltIndices
2277002
2235821
-1.8%
1.02x (?)
CStringLongAscii
4857
4797
-1.2%
1.01x (?)
CStringLongNonAscii
2136
2140
+0.2%
1.00x
CStringShortAscii
7546
7528
-0.2%
1.00x (?)
Calculator
1506
1505
-0.1%
1.00x (?)
CaptureProp
265070
264978
-0.0%
1.00x (?)
CharIndexing_ascii_unicodeScalars
376720
373393
-0.9%
1.01x (?)
CharIndexing_ascii_unicodeScalars_Backwards
428576
426625
-0.5%
1.00x (?)
CharIndexing_japanese_unicodeScalars
441226
444417
+0.7%
0.99x (?)
CharIndexing_japanese_unicodeScalars_Backwards
517606
543780
+5.1%
0.95x (?)
CharIndexing_korean_unicodeScalars
361857
377805
+4.4%
0.96x (?)
CharIndexing_korean_unicodeScalars_Backwards
414745
416511
+0.4%
1.00x (?)
CharIndexing_punctuatedJapanese_unicodeScalars
65240
64980
-0.4%
1.00x (?)
CharIndexing_punctuatedJapanese_unicodeScalars_Backwards
74478
74179
-0.4%
1.00x (?)
CharIndexing_punctuated_unicodeScalars
81357
80515
-1.0%
1.01x (?)
CharIndexing_punctuated_unicodeScalars_Backwards
93634
95052
+1.5%
0.99x (?)
CharIndexing_russian_unicodeScalars_Backwards
362398
365302
+0.8%
0.99x (?)
CharIndexing_tweet_unicodeScalars
739021
720819
-2.5%
1.03x (?)
CharIndexing_tweet_unicodeScalars_Backwards
857012
885568
+3.3%
0.97x (?)
CharIndexing_utf16_unicodeScalars
312631
310874
-0.6%
1.01x (?)
CharIndexing_utf16_unicodeScalars_Backwards
359985
358599
-0.4%
1.00x (?)
CharIteration_ascii_unicodeScalars
156699
156979
+0.2%
1.00x (?)
CharIteration_ascii_unicodeScalars_Backwards
304533
302870
-0.5%
1.01x (?)
CharIteration_chinese_unicodeScalars
118735
118599
-0.1%
1.00x (?)
CharIteration_chinese_unicodeScalars_Backwards
231432
232490
+0.5%
1.00x (?)
CharIteration_japanese_unicodeScalars
188116
188136
+0.0%
1.00x (?)
CharIteration_japanese_unicodeScalars_Backwards
368006
366874
-0.3%
1.00x (?)
CharIteration_korean_unicodeScalars
151369
151413
+0.0%
1.00x (?)
CharIteration_korean_unicodeScalars_Backwards
295666
295118
-0.2%
1.00x (?)
CharIteration_punctuatedJapanese_unicodeScalars
28157
28147
-0.0%
1.00x (?)
CharIteration_punctuatedJapanese_unicodeScalars_Backwards
52896
53372
+0.9%
0.99x (?)
CharIteration_punctuated_unicodeScalars
35204
35209
+0.0%
1.00x (?)
CharIteration_punctuated_unicodeScalars_Backwards
67342
67628
+0.4%
1.00x (?)
CharIteration_russian_unicodeScalars
130353
130418
+0.0%
1.00x (?)
CharIteration_russian_unicodeScalars_Backwards
254336
255629
+0.5%
0.99x (?)
CharIteration_tweet_unicodeScalars
308803
308602
-0.1%
1.00x (?)
CharIteration_utf16_unicodeScalars_Backwards
265635
271270
+2.1%
0.98x (?)
CharacterLiteralsLarge
6019
6013
-0.1%
1.00x (?)
CharacterLiteralsSmall
689
688
-0.1%
1.00x
CharacterPropertiesPrecomputed
4438
4364
-1.7%
1.02x (?)
CharacterPropertiesStashed
2451
2370
-3.3%
1.03x (?)
CharacterPropertiesStashedMemo
5103
5278
+3.4%
0.97x (?)
Chars
40482
40571
+0.2%
1.00x (?)
ClassArrayGetter
987
988
+0.1%
1.00x (?)
Combos
2227
2212
-0.7%
1.01x (?)
DictOfArraysToArrayOfDicts
3479
3450
-0.8%
1.01x (?)
Dictionary
2767
2767
+0.0%
1.00x
Dictionary2
3048
3051
+0.1%
1.00x (?)
Dictionary2OfObjects
5746
5890
+2.5%
0.98x (?)
Dictionary3
1374
1370
-0.3%
1.00x (?)
Dictionary3OfObjects
2333
2406
+3.1%
0.97x (?)
Dictionary4
1254
1249
-0.4%
1.00x (?)
Dictionary4OfObjects
1948
1869
-4.1%
1.04x
DictionaryGroup
4307
4305
-0.0%
1.00x (?)
DictionaryGroupOfObjects
7439
7441
+0.0%
1.00x (?)
DictionaryLiteral
8813
8788
-0.3%
1.00x (?)
DictionaryOfObjects
6429
6628
+3.1%
0.97x
DictionaryRemove
19225
19218
-0.0%
1.00x (?)
DictionaryRemoveOfObjects
51929
51788
-0.3%
1.00x (?)
DictionarySubscriptDefaultMutation
2096
2078
-0.9%
1.01x (?)
DictionarySubscriptDefaultMutationArray
2330
2331
+0.0%
1.00x (?)
DictionarySubscriptDefaultMutationArrayOfObjects
9171
9134
-0.4%
1.00x (?)
DictionarySubscriptDefaultMutationOfObjects
5818
5656
-2.8%
1.03x (?)
DictionarySwap
5696
5689
-0.1%
1.00x (?)
DictionarySwapOfObjects
22877
22840
-0.2%
1.00x (?)
DoubleWidthDivision
0
0
+0.0%
1.00x
DropFirstAnyCollection
14494
14534
+0.3%
1.00x (?)
DropFirstAnySeqCRangeIter
21585
21636
+0.2%
1.00x (?)
DropFirstAnySeqCRangeIterLazy
21629
21618
-0.1%
1.00x (?)
DropFirstAnySeqCntRange
14616
14615
-0.0%
1.00x (?)
DropFirstAnySeqCntRangeLazy
14552
14590
+0.3%
1.00x (?)
DropFirstAnySequence
11767
11753
-0.1%
1.00x (?)
DropFirstAnySequenceLazy
11670
11672
+0.0%
1.00x (?)
DropFirstArray
4049
4056
+0.2%
1.00x
DropFirstArrayLazy
31819
31807
-0.0%
1.00x (?)
DropFirstCountableRange
326
326
+0.0%
1.00x
DropFirstCountableRangeLazy
28082
28081
-0.0%
1.00x (?)
DropFirstSequence
11225
11234
+0.1%
1.00x (?)
DropFirstSequenceLazy
11253
11283
+0.3%
1.00x
DropLastAnyCollection
4903
4862
-0.8%
1.01x
DropLastAnyCollectionLazy
34471
34682
+0.6%
0.99x (?)
DropLastAnySeqCRangeIter
38902
38866
-0.1%
1.00x
DropLastAnySeqCRangeIterLazy
39012
39049
+0.1%
1.00x (?)
DropLastAnySeqCntRange
4879
4881
+0.0%
1.00x (?)
DropLastAnySeqCntRangeLazy
4827
4823
-0.1%
1.00x (?)
DropLastAnySequence
29676
29651
-0.1%
1.00x (?)
DropLastAnySequenceLazy
29652
29578
-0.2%
1.00x
DropLastCountableRange
112
113
+0.9%
0.99x
DropLastCountableRangeLazy
9310
9373
+0.7%
0.99x (?)
DropLastSequence
29486
29474
-0.0%
1.00x (?)
DropLastSequenceLazy
29545
29588
+0.1%
1.00x (?)
DropWhileAnyCollection
18633
18627
-0.0%
1.00x (?)
DropWhileAnyCollectionLazy
20800
20823
+0.1%
1.00x (?)
DropWhileAnySeqCRangeIter
23328
23266
-0.3%
1.00x (?)
DropWhileAnySeqCRangeIterLazy
20913
20825
-0.4%
1.00x (?)
DropWhileAnySeqCntRange
18242
18253
+0.1%
1.00x (?)
DropWhileAnySeqCntRangeLazy
20785
20769
-0.1%
1.00x (?)
DropWhileAnySequence
13443
13403
-0.3%
1.00x
DropWhileAnySequenceLazy
11655
11649
-0.1%
1.00x (?)
DropWhileArrayLazy
14617
14607
-0.1%
1.00x (?)
DropWhileCountableRange
4183
4179
-0.1%
1.00x
DropWhileCountableRangeLazy
20135
20115
-0.1%
1.00x (?)
DropWhileSequence
13102
13134
+0.2%
1.00x (?)
DropWhileSequenceLazy
10925
10923
-0.0%
1.00x (?)
EqualStringSubstring
91
92
+1.1%
0.99x
EqualSubstringString
92
92
+0.0%
1.00x
EqualSubstringSubstring
92
92
+0.0%
1.00x
EqualSubstringSubstringGenericEquatable
97
97
+0.0%
1.00x
ErrorHandling
7035
7332
+4.2%
0.96x (?)
ExclusivityGlobal
186
186
+0.0%
1.00x
ExclusivityIndependent
70
71
+1.4%
0.99x
FilterEvenUsingReduce
3671
3684
+0.4%
1.00x
FilterEvenUsingReduceInto
1957
1947
-0.5%
1.01x (?)
FrequenciesUsingReduce
10711
10407
-2.8%
1.03x (?)
FrequenciesUsingReduceInto
5689
5594
-1.7%
1.02x (?)
Hanoi
19959
20055
+0.5%
1.00x
HashTest
21718
21657
-0.3%
1.00x (?)
Histogram
6810
6802
-0.1%
1.00x (?)
Integrate
817
817
+0.0%
1.00x
IterateData
6083
6081
-0.0%
1.00x (?)
Join
1122
1120
-0.2%
1.00x (?)
LazilyFilteredArrayContains
802440
799120
-0.4%
1.00x (?)
LazilyFilteredArrays
1482250
1481903
-0.0%
1.00x (?)
LazilyFilteredRange
496874
496737
-0.0%
1.00x (?)
LessSubstringSubstring
92
92
+0.0%
1.00x
LessSubstringSubstringGenericComparable
96
96
+0.0%
1.00x
LinkedList
32147
32152
+0.0%
1.00x (?)
LuhnAlgoEager
4944
5029
+1.7%
0.98x (?)
LuhnAlgoLazy
5090
5106
+0.3%
1.00x (?)
MapReduce
25935
25934
-0.0%
1.00x (?)
MapReduceAnyCollection
25868
25809
-0.2%
1.00x (?)
MapReduceAnyCollectionShort
38151
37465
-1.8%
1.02x (?)
MapReduceClass
30429
30446
+0.1%
1.00x (?)
MapReduceClassShort
41715
41547
-0.4%
1.00x (?)
MapReduceLazyCollection
24259
24257
-0.0%
1.00x (?)
MapReduceLazyCollectionShort
34827
34829
+0.0%
1.00x (?)
MapReduceLazySequence
18938
18937
-0.0%
1.00x (?)
MapReduceSequence
30165
30209
+0.1%
1.00x (?)
MapReduceShort
37627
38058
+1.1%
0.99x (?)
MapReduceShortString
236
239
+1.3%
0.99x (?)
MapReduceString
1859
1915
+3.0%
0.97x
Memset
44715
44734
+0.0%
1.00x (?)
MonteCarloE
1244837
1248003
+0.3%
1.00x (?)
MonteCarloPi
5624224
5621708
-0.0%
1.00x (?)
NSDictionaryCastToSwift
6764
6839
+1.1%
0.99x (?)
NSError
741
720
-2.8%
1.03x (?)
NSStringConversion
352
352
+0.0%
1.00x
NibbleSort
436348
448658
+2.8%
0.97x
NopDeinit
169344
169167
-0.1%
1.00x (?)
ObjectAllocation
1386
1386
+0.0%
1.00x
ObjectiveCBridgeFromNSArrayAnyObject
22989
24104
+4.9%
0.95x (?)
ObjectiveCBridgeFromNSArrayAnyObjectToString
42888
44994
+4.9%
0.95x (?)
ObjectiveCBridgeFromNSArrayAnyObjectToStringForced
36538
36709
+0.5%
1.00x (?)
ObjectiveCBridgeFromNSDictionaryAnyObject
127888
127193
-0.5%
1.01x (?)
ObjectiveCBridgeFromNSSetAnyObject
63363
65927
+4.0%
0.96x (?)
ObjectiveCBridgeFromNSString
3172
3157
-0.5%
1.00x (?)
ObjectiveCBridgeFromNSStringForced
2779
2784
+0.2%
1.00x (?)
ObjectiveCBridgeStubDataAppend
4384
4506
+2.8%
0.97x (?)
ObjectiveCBridgeStubDateMutation
632
634
+0.3%
1.00x (?)
ObjectiveCBridgeStubFromNSDate
4789
4794
+0.1%
1.00x (?)
ObjectiveCBridgeStubFromNSString
924
924
+0.0%
1.00x
ObjectiveCBridgeStubFromNSStringRef
199
200
+0.5%
1.00x
ObjectiveCBridgeStubNSDataAppend
2984
3004
+0.7%
0.99x (?)
ObjectiveCBridgeStubNSDateMutationRef
16802
16452
-2.1%
1.02x (?)
ObjectiveCBridgeStubToArrayOfNSString
29040
28927
-0.4%
1.00x (?)
ObjectiveCBridgeStubToNSDate
16551
16650
+0.6%
0.99x (?)
ObjectiveCBridgeStubToNSDateRef
4151
4090
-1.5%
1.01x
ObjectiveCBridgeStubToNSString
1590
1590
+0.0%
1.00x
ObjectiveCBridgeStubToNSStringRef
153
153
+0.0%
1.00x
ObjectiveCBridgeStubURLAppendPath
344933
334843
-2.9%
1.03x (?)
ObjectiveCBridgeStubURLAppendPathRef
356033
353326
-0.8%
1.01x (?)
ObjectiveCBridgeToNSArray
29426
29948
+1.8%
0.98x (?)
ObjectiveCBridgeToNSDictionary
51893
51222
-1.3%
1.01x (?)
ObjectiveCBridgeToNSSet
43779
42952
-1.9%
1.02x (?)
ObjectiveCBridgeToNSString
1362
1362
+0.0%
1.00x
ObserverClosure
6469
6508
+0.6%
0.99x (?)
ObserverForwarderStruct
4568
4570
+0.0%
1.00x (?)
ObserverPartiallyAppliedMethod
8180
8138
-0.5%
1.01x (?)
ObserverUnappliedMethod
8219
8225
+0.1%
1.00x (?)
OpenClose
541
543
+0.4%
1.00x (?)
PartialApplyDynamicType
41361
41277
-0.2%
1.00x (?)
Phonebook
19959
20064
+0.5%
0.99x
PointerArithmetics
114533
114584
+0.0%
1.00x (?)
PolymorphicCalls
6106
6097
-0.1%
1.00x (?)
PopFrontArray
4713
4696
-0.4%
1.00x (?)
PopFrontArrayGeneric
5735
5719
-0.3%
1.00x (?)
PopFrontUnsafePointer
11251
11343
+0.8%
0.99x (?)
PrefixAnyCollection
14527
14691
+1.1%
0.99x (?)
PrefixAnyCollectionLazy
105963
107133
+1.1%
0.99x (?)
PrefixAnySeqCRangeIter
17398
17417
+0.1%
1.00x (?)
PrefixAnySeqCRangeIterLazy
17437
17408
-0.2%
1.00x (?)
PrefixAnySeqCntRange
14543
14606
+0.4%
1.00x (?)
PrefixAnySeqCntRangeLazy
14421
14438
+0.1%
1.00x (?)
PrefixAnySequence
9833
9764
-0.7%
1.01x
PrefixAnySequenceLazy
9816
9777
-0.4%
1.00x (?)
PrefixArray
4054
4050
-0.1%
1.00x (?)
PrefixArrayLazy
31930
31753
-0.6%
1.01x (?)
PrefixCountableRange
326
326
+0.0%
1.00x
PrefixCountableRangeLazy
27955
27925
-0.1%
1.00x (?)
PrefixSequence
9635
9602
-0.3%
1.00x (?)
PrefixSequenceLazy
9541
9583
+0.4%
1.00x (?)
PrefixWhileAnyCollection
25918
25926
+0.0%
1.00x (?)
PrefixWhileAnyCollectionLazy
17010
17131
+0.7%
0.99x (?)
PrefixWhileAnySeqCRangeIter
33980
33947
-0.1%
1.00x (?)
PrefixWhileAnySeqCRangeIterLazy
17073
17086
+0.1%
1.00x (?)
PrefixWhileAnySeqCntRange
25854
25874
+0.1%
1.00x
PrefixWhileAnySeqCntRangeLazy
17232
17234
+0.0%
1.00x (?)
PrefixWhileAnySequence
27034
26956
-0.3%
1.00x (?)
PrefixWhileAnySequenceLazy
10461
10440
-0.2%
1.00x (?)
PrefixWhileArray
10919
10935
+0.1%
1.00x (?)
PrefixWhileArrayLazy
12933
12926
-0.1%
1.00x (?)
PrefixWhileCountableRange
11761
11762
+0.0%
1.00x (?)
PrefixWhileCountableRangeLazy
17004
17031
+0.2%
1.00x
PrefixWhileSequence
26608
26673
+0.2%
1.00x
PrefixWhileSequenceLazy
10126
10137
+0.1%
1.00x (?)
Prims
9815
9752
-0.6%
1.01x (?)
PrimsSplit
9689
9822
+1.4%
0.99x (?)
QueueConcrete
15272
15270
-0.0%
1.00x (?)
QueueGeneric
19849
19598
-1.3%
1.01x (?)
RC4
16685
16808
+0.7%
0.99x (?)
RGBHistogram
26680
26560
-0.4%
1.00x (?)
RGBHistogramOfObjects
96151
96077
-0.1%
1.00x (?)
RangeAssignment
2198
2211
+0.6%
0.99x
RangeIterationSigned
16754
16734
-0.1%
1.00x (?)
RangeIterationSigned64
39903
40180
+0.7%
0.99x (?)
RangeIterationUnsigned
35643
35616
-0.1%
1.00x (?)
RangeReplaceableCollectionPlusDefault
8195
8414
+2.7%
0.97x
RecursiveOwnedParameter
7939
7944
+0.1%
1.00x (?)
RemoveWhereFilterInts
2119
2117
-0.1%
1.00x
RemoveWhereFilterString
1516
1515
-0.1%
1.00x (?)
RemoveWhereFilterStrings
2982
2984
+0.1%
1.00x
RemoveWhereMoveInts
3338
3336
-0.1%
1.00x
RemoveWhereMoveStrings
4037
4151
+2.8%
0.97x (?)
RemoveWhereQuadraticInts
7926
7924
-0.0%
1.00x (?)
RemoveWhereQuadraticString
2338
2337
-0.0%
1.00x (?)
RemoveWhereQuadraticStrings
9602
9593
-0.1%
1.00x (?)
RemoveWhereSwapInts
5982
6033
+0.9%
0.99x (?)
RemoveWhereSwapStrings
6792
6798
+0.1%
1.00x (?)
ReversedArray
13917
13909
-0.1%
1.00x (?)
ReversedBidirectional
44528
44118
-0.9%
1.01x (?)
ReversedDictionary
23602
23541
-0.3%
1.00x (?)
RomanNumbers
1352351
1374041
+1.6%
0.98x (?)
SequenceAlgosAnySequence
11767
11725
-0.4%
1.00x (?)
SequenceAlgosArray
800057
799940
-0.0%
1.00x (?)
SequenceAlgosContiguousArray
341107
338947
-0.6%
1.01x (?)
SequenceAlgosList
8395
8397
+0.0%
1.00x (?)
SequenceAlgosRange
1144611
1143780
-0.1%
1.00x (?)
SequenceAlgosUnfoldSequence
5965
5960
-0.1%
1.00x (?)
SetExclusiveOr
18219
18169
-0.3%
1.00x (?)
SetExclusiveOr_OfObjects
42394
42377
-0.0%
1.00x (?)
SetIntersect
9053
9040
-0.1%
1.00x (?)
SetIntersect_OfObjects
12951
12946
-0.0%
1.00x (?)
SetIsSubsetOf
1459
1458
-0.1%
1.00x (?)
SetIsSubsetOf_OfObjects
1648
1652
+0.2%
1.00x (?)
SetUnion
12247
12261
+0.1%
1.00x (?)
SetUnion_OfObjects
29508
29898
+1.3%
0.99x
SevenBoom
1663
1673
+0.6%
0.99x (?)
Sim2DArray
43679
43653
-0.1%
1.00x (?)
SortLargeExistentials
15458
15441
-0.1%
1.00x (?)
SortLettersInPlace
2736
2711
-0.9%
1.01x
SortSortedStrings
1072
1070
-0.2%
1.00x (?)
SortStrings
2129
2125
-0.2%
1.00x
SortStringsUnicode
2755
2760
+0.2%
1.00x (?)
StackPromo
93173
92716
-0.5%
1.00x (?)
StaticArray
2615
2617
+0.1%
1.00x (?)
StrComplexWalk
6434
6440
+0.1%
1.00x (?)
StrToInt
58251
58350
+0.2%
1.00x (?)
StringAdder
4785
4791
+0.1%
1.00x (?)
StringBuilder
6923
6990
+1.0%
0.99x
StringBuilderLong
2726
2733
+0.3%
1.00x (?)
StringComparison_abnormal
1342
1340
-0.1%
1.00x (?)
StringComparison_ascii
9790
9765
-0.3%
1.00x (?)
StringComparison_emoji
2125
2122
-0.1%
1.00x (?)
StringComparison_fastPrenormal
5374
5317
-1.1%
1.01x (?)
StringComparison_latin1
4180
4183
+0.1%
1.00x (?)
StringComparison_longSharedPrefix
2501
2500
-0.0%
1.00x (?)
StringComparison_nonBMPSlowestPrenormal
3923
3928
+0.1%
1.00x (?)
StringComparison_slowerPrenormal
4405
4404
-0.0%
1.00x (?)
StringComparison_zalgo
115299
115627
+0.3%
1.00x
StringEnumRawValueInitialization
17405
17488
+0.5%
1.00x (?)
StringEqualPointerComparison
3557
3560
+0.1%
1.00x (?)
StringFromLongWholeSubstring
24
24
+0.0%
1.00x
StringFromLongWholeSubstringGeneric
205
207
+1.0%
0.99x (?)
StringHasPrefixAscii
3068
3066
-0.1%
1.00x (?)
StringHasPrefixUnicode
29206
29538
+1.1%
0.99x (?)
StringHasSuffixAscii
3143
3143
+0.0%
1.00x
StringHasSuffixUnicode
86907
86628
-0.3%
1.00x (?)
StringInterpolation
13309
13337
+0.2%
1.00x (?)
StringMatch
34432
34131
-0.9%
1.01x (?)
StringRemoveDupes
1417
1431
+1.0%
0.99x (?)
StringWalk
13196
13150
-0.3%
1.00x (?)
StringWithCString
39447
39427
-0.1%
1.00x (?)
StringWordBuilder
2733
2724
-0.3%
1.00x (?)
StringWordBuilderReservingCapacity
2318
2320
+0.1%
1.00x (?)
SubstringComparable
2108
2103
-0.2%
1.00x (?)
SubstringEqualString
1912
1916
+0.2%
1.00x (?)
SubstringEquatable
6139
6107
-0.5%
1.01x (?)
SubstringFromLongString
24
25
+4.2%
0.96x
SubstringFromLongStringGeneric
105
105
+0.0%
1.00x
SuffixAnyCollection
4849
4855
+0.1%
1.00x (?)
SuffixAnyCollectionLazy
34297
35831
+4.5%
0.96x (?)
SuffixAnySeqCRangeIter
34749
34706
-0.1%
1.00x (?)
SuffixAnySeqCRangeIterLazy
34722
34827
+0.3%
1.00x (?)
SuffixAnySeqCntRange
4904
4908
+0.1%
1.00x (?)
SuffixAnySeqCntRangeLazy
4831
4832
+0.0%
1.00x (?)
SuffixAnySequence
25551
25527
-0.1%
1.00x (?)
SuffixAnySequenceLazy
25475
25498
+0.1%
1.00x (?)
SuffixCountableRange
112
112
+0.0%
1.00x
SuffixCountableRangeLazy
9316
9319
+0.0%
1.00x (?)
SuffixSequence
25672
25630
-0.2%
1.00x (?)
SuffixSequenceLazy
25425
25407
-0.1%
1.00x
SumUsingReduce
167649
167787
+0.1%
1.00x (?)
SumUsingReduceInto
162345
162348
+0.0%
1.00x (?)
SuperChars
125659
124863
-0.6%
1.01x (?)
TwoSum
4255
4280
+0.6%
0.99x (?)
TypeFlood
163
162
-0.6%
1.01x (?)
UTF8Decode
31010
31085
+0.2%
1.00x
Walsh
12069
12059
-0.1%
1.00x (?)
WordCountHistogramASCII
46795
46884
+0.2%
1.00x (?)
WordCountHistogramUTF16
69554
72099
+3.7%
0.96x (?)
WordCountUniqueASCII
8194
8206
+0.1%
1.00x (?)
WordSplitASCII
26057
27249
+4.6%
0.96x
WordSplitUTF16
27513
28045
+1.9%
0.98x (?)
XorLoop
24847
24841
-0.0%
1.00x (?)
Hardware Overview
Model Name: Mac Pro
Model Identifier: MacPro6,1
Processor Name: 12-Core Intel Xeon E5
Processor Speed: 2.7 GHz
Number of Processors: 1
Total Number of Cores: 12
L2 Cache (per Core): 256 KB
L3 Cache: 30 MB
Memory: 64 GB
|
gharchive/pull-request
| 2018-03-05T18:20:56 |
2025-04-01T04:56:01.618752
|
{
"authors": [
"atrick",
"swift-ci"
],
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/14985",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
306998181
|
[benchmark] Add a benchmark for .lazy.filter,map chaining...
... and functionally equivalent single call to .lazy.compactMap.
See https://forums.swift.org/t/introduce-lazy-version-of-compactmap/9835
for more info.
@swift-ci Please smoke test
@swift-ci Please smoke benchmark
/cc @tellowkrinkle
Related to https://github.com/apple/swift/pull/14841
Build comment file:
Optimized (O)
Regression (16)
TEST
OLD
NEW
DELTA
SPEEDUP
MapReduceLazyCollectionShort
34
38
+11.8%
0.89x
StringBuilderLong
1082
1199
+10.8%
0.90x (?)
StringEqualPointerComparison
286
315
+10.1%
0.91x
CharIteration_chinese_unicodeScalars
15456
16850
+9.0%
0.92x
CharIteration_tweet_unicodeScalars
40659
44106
+8.5%
0.92x
RemoveWhereMoveInts
14
15
+7.1%
0.93x
CharIteration_punctuated_unicodeScalars
4603
4931
+7.1%
0.93x
CharIteration_korean_unicodeScalars
19791
21185
+7.0%
0.93x
CharIteration_russian_unicodeScalars
17030
18221
+7.0%
0.93x
NopDeinit
30188
32262
+6.9%
0.94x
CharIteration_ascii_unicodeScalars
20438
21831
+6.8%
0.94x
CharIteration_japanese_unicodeScalars
24478
26120
+6.7%
0.94x
Chars
975
1036
+6.3%
0.94x
CharIteration_punctuatedJapanese_unicodeScalars
3679
3905
+6.1%
0.94x
ObjectiveCBridgeFromNSArrayAnyObject
23862
25211
+5.7%
0.95x (?)
ObjectiveCBridgeStubToNSStringRef
112
118
+5.4%
0.95x
Improvement (32)
TEST
OLD
NEW
DELTA
SPEEDUP
Sim2DArray
599
417
-30.4%
1.44x
StringBuilderWithLongSubstring
1881
1346
-28.4%
1.40x (?)
RangeIterationSigned
200
171
-14.5%
1.17x
DataSubscript
259
237
-8.5%
1.09x
CharIndexing_japanese_unicodeScalars
20338
18627
-8.4%
1.09x
PointerArithmetics
34391
31516
-8.4%
1.09x
CharIndexing_korean_unicodeScalars
16463
15091
-8.3%
1.09x
DropLastCountableRange
12
11
-8.3%
1.09x
CharIndexing_russian_unicodeScalars
14141
12963
-8.3%
1.09x
CharIndexing_ascii_unicodeScalars
16968
15567
-8.3%
1.09x
CharIndexing_chinese_unicodeScalars
12856
11797
-8.2%
1.09x
OpenClose
137
126
-8.0%
1.09x
CharIteration_utf16_unicodeScalars_Backwards
19435
17881
-8.0%
1.09x
CharIndexing_punctuated_unicodeScalars
3833
3550
-7.4%
1.08x
CharIndexing_tweet_unicodeScalars
33080
30651
-7.3%
1.08x
DataReplaceMediumBuffer
13133
12176
-7.3%
1.08x (?)
DropLastAnySeqCntRange
14
13
-7.1%
1.08x
CharIndexing_punctuatedJapanese_unicodeScalars
3061
2844
-7.1%
1.08x
DataReplaceMedium
11812
11024
-6.7%
1.07x (?)
DataAppendArray
6585
6181
-6.1%
1.07x (?)
DataReplaceLarge
42464
39903
-6.0%
1.06x (?)
MapReduceAnyCollection
429
404
-5.8%
1.06x
CharIteration_tweet_unicodeScalars_Backwards
30767
29144
-5.3%
1.06x
CharIteration_korean_unicodeScalars_Backwards
15108
14314
-5.3%
1.06x
CharIteration_ascii_unicodeScalars_Backwards
15579
14765
-5.2%
1.06x
CharIteration_japanese_unicodeScalars_Backwards
18644
17674
-5.2%
1.05x
CharIteration_russian_unicodeScalars_Backwards
12980
12308
-5.2%
1.05x
CharIteration_chinese_unicodeScalars_Backwards
11789
11185
-5.1%
1.05x
StringComparison_ascii
929
882
-5.1%
1.05x
CharIteration_punctuated_unicodeScalars_Backwards
3530
3353
-5.0%
1.05x
SumUsingReduceInto
102
97
-4.9%
1.05x
CharIteration_punctuatedJapanese_unicodeScalars_Backwards
2821
2685
-4.8%
1.05x
No Changes (375)
TEST
OLD
NEW
DELTA
SPEEDUP
AngryPhonebook
3523
3554
+0.9%
0.99x (?)
AnyHashableWithAClass
81035
81069
+0.0%
1.00x (?)
Array2D
2550
2551
+0.0%
1.00x (?)
ArrayAppend
1109
1112
+0.3%
1.00x (?)
ArrayAppendArrayOfInt
795
795
+0.0%
1.00x
ArrayAppendAscii
13249
13248
-0.0%
1.00x (?)
ArrayAppendFromGeneric
797
796
-0.1%
1.00x (?)
ArrayAppendGenericStructs
1428
1423
-0.4%
1.00x
ArrayAppendLatin1
38001
38031
+0.1%
1.00x (?)
ArrayAppendLazyMap
1345
1339
-0.4%
1.00x (?)
ArrayAppendOptionals
1426
1423
-0.2%
1.00x
ArrayAppendRepeatCol
1357
1340
-1.3%
1.01x
ArrayAppendReserved
839
842
+0.4%
1.00x (?)
ArrayAppendSequence
1120
1121
+0.1%
1.00x (?)
ArrayAppendStrings
15176
15145
-0.2%
1.00x (?)
ArrayAppendToFromGeneric
799
797
-0.3%
1.00x
ArrayAppendToGeneric
787
803
+2.0%
0.98x (?)
ArrayAppendUTF16
37689
37735
+0.1%
1.00x (?)
ArrayInClass
85
85
+0.0%
1.00x
ArrayLiteral
0
0
+0.0%
1.00x
ArrayOfGenericPOD2
150
151
+0.7%
0.99x (?)
ArrayOfGenericRef
4398
4397
-0.0%
1.00x (?)
ArrayOfPOD
183
183
+0.0%
1.00x
ArrayOfRef
4398
4396
-0.0%
1.00x (?)
ArrayPlusEqualArrayOfInt
797
798
+0.1%
1.00x (?)
ArrayPlusEqualFiveElementCollection
6558
6686
+2.0%
0.98x
ArrayPlusEqualSingleElementCollection
5005
5111
+2.1%
0.98x (?)
ArrayPlusEqualThreeElements
2009
2010
+0.0%
1.00x (?)
ArraySubscript
1528
1539
+0.7%
0.99x (?)
ArrayValueProp
8
8
+0.0%
1.00x
ArrayValueProp2
8
8
+0.0%
1.00x
ArrayValueProp3
8
8
+0.0%
1.00x
ArrayValueProp4
8
8
+0.0%
1.00x
BinaryFloatingPointConversionFromBinaryInteger
40
40
+0.0%
1.00x
BinaryFloatingPointPropertiesBinade
31
31
+0.0%
1.00x
BinaryFloatingPointPropertiesNextUp
28
28
+0.0%
1.00x
BinaryFloatingPointPropertiesUlp
37
37
+0.0%
1.00x
BitCount
211
211
+0.0%
1.00x
ByteSwap
106
107
+0.9%
0.99x (?)
COWTree
5681
5513
-3.0%
1.03x (?)
CSVParsing
755612
760650
+0.7%
0.99x (?)
CSVParsingAlt
725460
730518
+0.7%
0.99x (?)
CSVParsingAltIndices
340516
340298
-0.1%
1.00x (?)
CStringLongAscii
4745
4737
-0.2%
1.00x (?)
CStringLongNonAscii
2222
2213
-0.4%
1.00x (?)
CStringShortAscii
5530
5521
-0.2%
1.00x (?)
Calculator
666
651
-2.3%
1.02x
CaptureProp
4111
4114
+0.1%
1.00x (?)
CharIndexing_ascii_unicodeScalars_Backwards
15714
15833
+0.8%
0.99x (?)
CharIndexing_chinese_unicodeScalars_Backwards
11890
12008
+1.0%
0.99x
CharIndexing_japanese_unicodeScalars_Backwards
18774
18944
+0.9%
0.99x
CharIndexing_korean_unicodeScalars_Backwards
15215
15346
+0.9%
0.99x
CharIndexing_punctuatedJapanese_unicodeScalars_Backwards
2859
2862
+0.1%
1.00x (?)
CharIndexing_punctuated_unicodeScalars_Backwards
3577
3589
+0.3%
1.00x
CharIndexing_russian_unicodeScalars_Backwards
13074
13191
+0.9%
0.99x
CharIndexing_tweet_unicodeScalars_Backwards
30876
31314
+1.4%
0.99x
CharIndexing_utf16_unicodeScalars
21818
21091
-3.3%
1.03x
CharIndexing_utf16_unicodeScalars_Backwards
21657
21583
-0.3%
1.00x
CharIteration_utf16_unicodeScalars
27877
27983
+0.4%
1.00x
CharacterLiteralsLarge
5752
5760
+0.1%
1.00x (?)
CharacterLiteralsSmall
217
220
+1.4%
0.99x
CharacterPropertiesFetch
4576
4610
+0.7%
0.99x (?)
CharacterPropertiesPrecomputed
1251
1252
+0.1%
1.00x (?)
CharacterPropertiesStashed
1450
1438
-0.8%
1.01x (?)
CharacterPropertiesStashedMemo
1681
1697
+1.0%
0.99x (?)
ClassArrayGetter
15
15
+0.0%
1.00x
Combos
460
460
+0.0%
1.00x
DataAccessBytes
1139
1145
+0.5%
0.99x
DataAppendBytes
6311
6120
-3.0%
1.03x (?)
DataAppendDataLargeToLarge
69622
69282
-0.5%
1.00x (?)
DataAppendDataLargeToMedium
36518
36492
-0.1%
1.00x (?)
DataAppendDataLargeToSmall
35615
35078
-1.5%
1.02x (?)
DataAppendDataMediumToLarge
39126
38568
-1.4%
1.01x (?)
DataAppendDataMediumToMedium
7712
7819
+1.4%
0.99x (?)
DataAppendDataMediumToSmall
7043
6862
-2.6%
1.03x (?)
DataAppendDataSmallToLarge
37556
37549
-0.0%
1.00x (?)
DataAppendDataSmallToMedium
7288
7213
-1.0%
1.01x (?)
DataAppendDataSmallToSmall
6832
6642
-2.8%
1.03x (?)
DataAppendSequence
18990
19732
+3.9%
0.96x (?)
DataCopyBytes
2540
2508
-1.3%
1.01x (?)
DataCount
38
37
-2.6%
1.03x
DataMutateBytes
4221
4242
+0.5%
1.00x (?)
DataReplaceLargeBuffer
57826
57919
+0.2%
1.00x (?)
DataReplaceSmall
8295
8713
+5.0%
0.95x (?)
DataReplaceSmallBuffer
8738
9058
+3.7%
0.96x (?)
DataReset
3154
3194
+1.3%
0.99x (?)
DataSetCount
836
856
+2.4%
0.98x
DictOfArraysToArrayOfDicts
824
829
+0.6%
0.99x (?)
Dictionary
721
716
-0.7%
1.01x
Dictionary2
1818
1816
-0.1%
1.00x (?)
Dictionary2OfObjects
3355
3342
-0.4%
1.00x (?)
Dictionary3
427
427
+0.0%
1.00x
Dictionary3OfObjects
890
890
+0.0%
1.00x
Dictionary4
852
851
-0.1%
1.00x (?)
Dictionary4OfObjects
937
946
+1.0%
0.99x (?)
DictionaryBridge
2016
2084
+3.4%
0.97x
DictionaryCopy
109879
107660
-2.0%
1.02x (?)
DictionaryFilter
108294
106601
-1.6%
1.02x
DictionaryGroup
267
279
+4.5%
0.96x
DictionaryGroupOfObjects
2124
2097
-1.3%
1.01x
DictionaryLiteral
2034
2000
-1.7%
1.02x
DictionaryOfObjects
2636
2644
+0.3%
1.00x
DictionaryRemove
5333
5276
-1.1%
1.01x
DictionaryRemoveOfObjects
28878
28610
-0.9%
1.01x (?)
DictionarySubscriptDefaultMutation
301
292
-3.0%
1.03x
DictionarySubscriptDefaultMutationArray
642
642
+0.0%
1.00x
DictionarySubscriptDefaultMutationArrayOfObjects
4194
4201
+0.2%
1.00x (?)
DictionarySubscriptDefaultMutationOfObjects
1973
1959
-0.7%
1.01x (?)
DictionarySwap
1370
1356
-1.0%
1.01x
DictionarySwapAt
6970
7196
+3.2%
0.97x
DictionarySwapAtOfObjects
52840
52964
+0.2%
1.00x (?)
DictionarySwapOfObjects
9546
9762
+2.3%
0.98x
DoubleWidthDivision
0
0
+0.0%
1.00x
DropFirstAnyCollection
84
84
+0.0%
1.00x
DropFirstAnyCollectionLazy
60652
60934
+0.5%
1.00x (?)
DropFirstAnySeqCRangeIter
17032
17054
+0.1%
1.00x (?)
DropFirstAnySeqCRangeIterLazy
17046
17042
-0.0%
1.00x (?)
DropFirstAnySeqCntRange
29
29
+0.0%
1.00x
DropFirstAnySeqCntRangeLazy
28
29
+3.6%
0.97x
DropFirstAnySequence
4883
4887
+0.1%
1.00x (?)
DropFirstAnySequenceLazy
4877
4901
+0.5%
1.00x
DropFirstArray
35
35
+0.0%
1.00x
DropFirstArrayLazy
35
35
+0.0%
1.00x
DropFirstCountableRange
35
35
+0.0%
1.00x
DropFirstCountableRangeLazy
35
35
+0.0%
1.00x
DropFirstSequence
2681
2682
+0.0%
1.00x (?)
DropFirstSequenceLazy
2778
2772
-0.2%
1.00x (?)
DropLastAnyCollection
31
31
+0.0%
1.00x
DropLastAnyCollectionLazy
20297
20226
-0.3%
1.00x (?)
DropLastAnySeqCRangeIter
3595
3579
-0.4%
1.00x
DropLastAnySeqCRangeIterLazy
3595
3583
-0.3%
1.00x
DropLastAnySeqCntRangeLazy
14
14
+0.0%
1.00x
DropLastAnySequence
4977
4938
-0.8%
1.01x (?)
DropLastAnySequenceLazy
5070
4997
-1.4%
1.01x (?)
DropLastCountableRangeLazy
11
11
+0.0%
1.00x
DropLastSequence
653
629
-3.7%
1.04x
DropLastSequenceLazy
653
629
-3.7%
1.04x
DropWhileAnyCollection
107
107
+0.0%
1.00x
DropWhileAnyCollectionLazy
125
125
+0.0%
1.00x
DropWhileAnySeqCRangeIter
14447
14484
+0.3%
1.00x (?)
DropWhileAnySeqCRangeIterLazy
90
90
+0.0%
1.00x
DropWhileAnySeqCntRange
36
36
+0.0%
1.00x
DropWhileAnySeqCntRangeLazy
90
90
+0.0%
1.00x
DropWhileAnySequence
5581
5580
-0.0%
1.00x (?)
DropWhileAnySequenceLazy
1858
1857
-0.1%
1.00x (?)
DropWhileArrayLazy
88
88
+0.0%
1.00x
DropWhileCountableRange
35
35
+0.0%
1.00x
DropWhileCountableRangeLazy
105
105
+0.0%
1.00x
DropWhileSequence
1326
1326
+0.0%
1.00x
DropWhileSequenceLazy
88
88
+0.0%
1.00x
EqualStringSubstring
49
49
+0.0%
1.00x
EqualSubstringString
49
49
+0.0%
1.00x
EqualSubstringSubstring
49
48
-2.0%
1.02x
EqualSubstringSubstringGenericEquatable
49
49
+0.0%
1.00x
ErrorHandling
2524
2461
-2.5%
1.03x (?)
ExclusivityGlobal
5
5
+0.0%
1.00x
ExclusivityIndependent
2
2
+0.0%
1.00x
FilterEvenUsingReduce
1323
1311
-0.9%
1.01x (?)
FilterEvenUsingReduceInto
149
148
-0.7%
1.01x
FloatingPointPrinting_Double_description_small
80967
79552
-1.7%
1.02x
FloatingPointPrinting_Double_description_uniform
137015
135303
-1.2%
1.01x
FloatingPointPrinting_Double_interpolated
184605
182004
-1.4%
1.01x (?)
FloatingPointPrinting_Float80_description_small
93205
91699
-1.6%
1.02x
FloatingPointPrinting_Float80_description_uniform
1705051
1706668
+0.1%
1.00x
FloatingPointPrinting_Float80_interpolated
1750653
1752467
+0.1%
1.00x (?)
FloatingPointPrinting_Float_description_small
42343
43848
+3.6%
0.97x
FloatingPointPrinting_Float_description_uniform
44675
44573
-0.2%
1.00x (?)
FloatingPointPrinting_Float_interpolated
85074
85142
+0.1%
1.00x (?)
FrequenciesUsingReduce
6963
7079
+1.7%
0.98x (?)
FrequenciesUsingReduceInto
3650
3657
+0.2%
1.00x (?)
Hanoi
2139
2207
+3.2%
0.97x
HashTest
1780
1776
-0.2%
1.00x (?)
Histogram
814
814
+0.0%
1.00x
Integrate
338
350
+3.6%
0.97x
IterateData
1931
1977
+2.4%
0.98x
Join
352
352
+0.0%
1.00x
LazilyFilteredArrayContains
38685
37860
-2.1%
1.02x
LazilyFilteredArrays
67404
65146
-3.3%
1.03x (?)
LazilyFilteredRange
3881
3840
-1.1%
1.01x
LessSubstringSubstring
49
49
+0.0%
1.00x
LessSubstringSubstringGenericComparable
51
53
+3.9%
0.96x
LinkedList
7610
7539
-0.9%
1.01x
LuhnAlgoEager
538
536
-0.4%
1.00x (?)
LuhnAlgoLazy
538
538
+0.0%
1.00x
MapReduce
383
398
+3.9%
0.96x
MapReduceAnyCollectionShort
2277
2280
+0.1%
1.00x (?)
MapReduceClass
3029
3013
-0.5%
1.01x (?)
MapReduceClassShort
4548
4606
+1.3%
0.99x (?)
MapReduceLazyCollection
13
13
+0.0%
1.00x
MapReduceLazySequence
86
86
+0.0%
1.00x
MapReduceSequence
449
450
+0.2%
1.00x (?)
MapReduceShort
2023
2007
-0.8%
1.01x (?)
MapReduceShortString
21
21
+0.0%
1.00x
MapReduceString
77
79
+2.6%
0.97x
Memset
219
220
+0.5%
1.00x (?)
MonteCarloE
10253
10269
+0.2%
1.00x
MonteCarloPi
42726
42659
-0.2%
1.00x
NSDictionaryCastToSwift
5581
5535
-0.8%
1.01x (?)
NSError
304
302
-0.7%
1.01x (?)
NSStringConversion
401
401
+0.0%
1.00x
NibbleSort
3686
3680
-0.2%
1.00x
ObjectAllocation
135
136
+0.7%
0.99x
ObjectiveCBridgeFromNSArrayAnyObjectForced
4532
4483
-1.1%
1.01x (?)
ObjectiveCBridgeFromNSArrayAnyObjectToString
42740
44687
+4.6%
0.96x (?)
ObjectiveCBridgeFromNSArrayAnyObjectToStringForced
39780
41623
+4.6%
0.96x (?)
ObjectiveCBridgeFromNSDictionaryAnyObject
106079
111458
+5.1%
0.95x (?)
ObjectiveCBridgeFromNSSetAnyObject
48927
49153
+0.5%
1.00x (?)
ObjectiveCBridgeFromNSSetAnyObjectForced
4209
4124
-2.0%
1.02x (?)
ObjectiveCBridgeFromNSSetAnyObjectToString
78137
78266
+0.2%
1.00x (?)
ObjectiveCBridgeFromNSString
1345
1291
-4.0%
1.04x (?)
ObjectiveCBridgeFromNSStringForced
2460
2445
-0.6%
1.01x (?)
ObjectiveCBridgeStubDataAppend
10660
10863
+1.9%
0.98x
ObjectiveCBridgeStubDateMutation
401
400
-0.2%
1.00x
ObjectiveCBridgeStubFromArrayOfNSString
31950
32671
+2.3%
0.98x (?)
ObjectiveCBridgeStubFromNSDate
6040
5998
-0.7%
1.01x (?)
ObjectiveCBridgeStubFromNSString
953
963
+1.0%
0.99x
ObjectiveCBridgeStubFromNSStringRef
159
159
+0.0%
1.00x
ObjectiveCBridgeStubNSDataAppend
2614
2623
+0.3%
1.00x (?)
ObjectiveCBridgeStubNSDateMutationRef
13129
13333
+1.6%
0.98x (?)
ObjectiveCBridgeStubToArrayOfNSString
27580
27962
+1.4%
0.99x (?)
ObjectiveCBridgeStubToNSDate
15380
16074
+4.5%
0.96x (?)
ObjectiveCBridgeStubToNSDateRef
3456
3449
-0.2%
1.00x (?)
ObjectiveCBridgeStubToNSString
1560
1623
+4.0%
0.96x
ObjectiveCBridgeStubURLAppendPath
336757
353786
+5.1%
0.95x (?)
ObjectiveCBridgeStubURLAppendPathRef
357378
346134
-3.1%
1.03x (?)
ObjectiveCBridgeToNSArray
28030
28039
+0.0%
1.00x (?)
ObjectiveCBridgeToNSDictionary
43402
42123
-2.9%
1.03x (?)
ObjectiveCBridgeToNSSet
33380
34346
+2.9%
0.97x (?)
ObjectiveCBridgeToNSString
1285
1300
+1.2%
0.99x
ObserverClosure
2167
2154
-0.6%
1.01x (?)
ObserverForwarderStruct
1235
1296
+4.9%
0.95x
ObserverPartiallyAppliedMethod
3732
3748
+0.4%
1.00x
ObserverUnappliedMethod
2567
2665
+3.8%
0.96x
PartialApplyDynamicType
0
0
+0.0%
1.00x
Phonebook
3924
3859
-1.7%
1.02x
PolymorphicCalls
25
25
+0.0%
1.00x
PopFrontArray
1887
1880
-0.4%
1.00x (?)
PopFrontArrayGeneric
1884
1887
+0.2%
1.00x (?)
PopFrontUnsafePointer
8674
9046
+4.3%
0.96x (?)
PrefixAnyCollection
84
84
+0.0%
1.00x
PrefixAnyCollectionLazy
60627
61648
+1.7%
0.98x (?)
PrefixAnySeqCRangeIter
13267
13293
+0.2%
1.00x (?)
PrefixAnySeqCRangeIterLazy
13307
13322
+0.1%
1.00x
PrefixAnySeqCntRange
29
29
+0.0%
1.00x
PrefixAnySeqCntRangeLazy
28
29
+3.6%
0.97x
PrefixAnySequence
4214
4167
-1.1%
1.01x
PrefixAnySequenceLazy
4224
4214
-0.2%
1.00x (?)
PrefixArray
35
35
+0.0%
1.00x
PrefixArrayLazy
35
35
+0.0%
1.00x
PrefixCountableRange
35
35
+0.0%
1.00x
PrefixCountableRangeLazy
35
35
+0.0%
1.00x
PrefixSequence
1325
1328
+0.2%
1.00x
PrefixSequenceLazy
1413
1413
+0.0%
1.00x
PrefixWhileAnyCollection
155
155
+0.0%
1.00x
PrefixWhileAnyCollectionLazy
90
90
+0.0%
1.00x
PrefixWhileAnySeqCRangeIter
8822
8728
-1.1%
1.01x (?)
PrefixWhileAnySeqCRangeIterLazy
24
24
+0.0%
1.00x
PrefixWhileAnySeqCntRange
60
60
+0.0%
1.00x
PrefixWhileAnySeqCntRangeLazy
24
24
+0.0%
1.00x
PrefixWhileAnySequence
9936
9933
-0.0%
1.00x (?)
PrefixWhileAnySequenceLazy
1394
1393
-0.1%
1.00x
PrefixWhileArray
88
88
+0.0%
1.00x
PrefixWhileArrayLazy
70
70
+0.0%
1.00x
PrefixWhileCountableRange
35
35
+0.0%
1.00x
PrefixWhileCountableRangeLazy
35
35
+0.0%
1.00x
PrefixWhileSequence
364
362
-0.5%
1.01x
PrefixWhileSequenceLazy
52
52
+0.0%
1.00x
Prims
1129
1129
+0.0%
1.00x
PrimsSplit
1141
1161
+1.8%
0.98x (?)
QueueConcrete
1133
1128
-0.4%
1.00x (?)
QueueGeneric
1136
1135
-0.1%
1.00x (?)
RC4
164
162
-1.2%
1.01x
RGBHistogram
3699
3678
-0.6%
1.01x (?)
RGBHistogramOfObjects
25326
25180
-0.6%
1.01x (?)
RangeAssignment
351
338
-3.7%
1.04x
RangeIterationSigned64
200
200
+0.0%
1.00x
RangeIterationUnsigned
200
200
+0.0%
1.00x
RangeReplaceableCollectionPlusDefault
908
905
-0.3%
1.00x (?)
RecursiveOwnedParameter
115
115
+0.0%
1.00x
RemoveWhereFilterInts
45
43
-4.4%
1.05x
RemoveWhereFilterString
375
375
+0.0%
1.00x
RemoveWhereFilterStrings
432
432
+0.0%
1.00x
RemoveWhereMoveStrings
701
703
+0.3%
1.00x
RemoveWhereQuadraticInts
1288
1290
+0.2%
1.00x (?)
RemoveWhereQuadraticString
492
489
-0.6%
1.01x (?)
RemoveWhereQuadraticStrings
2752
2752
+0.0%
1.00x
RemoveWhereSwapInts
19
19
+0.0%
1.00x
RemoveWhereSwapStrings
850
848
-0.2%
1.00x
ReversedArray
57
57
+0.0%
1.00x
ReversedBidirectional
15869
16222
+2.2%
0.98x
ReversedDictionary
301
293
-2.7%
1.03x (?)
RomanNumbers
134373
134491
+0.1%
1.00x (?)
SequenceAlgosAnySequence
9393
9439
+0.5%
1.00x
SequenceAlgosArray
1573
1579
+0.4%
1.00x (?)
SequenceAlgosContiguousArray
1579
1578
-0.1%
1.00x (?)
SequenceAlgosList
1351
1356
+0.4%
1.00x
SequenceAlgosRange
2578
2577
-0.0%
1.00x (?)
SequenceAlgosUnfoldSequence
1080
1079
-0.1%
1.00x (?)
SetExclusiveOr
6141
6137
-0.1%
1.00x (?)
SetExclusiveOr_OfObjects
13161
12970
-1.5%
1.01x
SetIntersect
1000
982
-1.8%
1.02x
SetIntersect_OfObjects
2313
2331
+0.8%
0.99x (?)
SetIsSubsetOf
382
382
+0.0%
1.00x
SetIsSubsetOf_OfObjects
523
527
+0.8%
0.99x (?)
SetUnion
5154
5115
-0.8%
1.01x (?)
SetUnion_OfObjects
11602
11365
-2.0%
1.02x
SevenBoom
1530
1531
+0.1%
1.00x (?)
SortLargeExistentials
5430
5557
+2.3%
0.98x
SortLettersInPlace
990
995
+0.5%
0.99x (?)
SortSortedStrings
841
829
-1.4%
1.01x
SortStrings
1657
1630
-1.6%
1.02x
SortStringsUnicode
2337
2335
-0.1%
1.00x (?)
StackPromo
21945
21938
-0.0%
1.00x (?)
StaticArray
10
10
+0.0%
1.00x
StrComplexWalk
1563
1560
-0.2%
1.00x
StrToInt
1451
1456
+0.3%
1.00x
StringAdder
4262
4254
-0.2%
1.00x (?)
StringBuilder
1596
1614
+1.1%
0.99x (?)
StringComparison_abnormal
789
802
+1.6%
0.98x (?)
StringComparison_emoji
780
777
-0.4%
1.00x
StringComparison_fastPrenormal
715
708
-1.0%
1.01x
StringComparison_latin1
553
547
-1.1%
1.01x
StringComparison_longSharedPrefix
898
894
-0.4%
1.00x
StringComparison_nonBMPSlowestPrenormal
1536
1528
-0.5%
1.01x (?)
StringComparison_slowerPrenormal
1629
1652
+1.4%
0.99x
StringComparison_zalgo
122689
122890
+0.2%
1.00x (?)
StringEdits
145215
144639
-0.4%
1.00x (?)
StringEnumRawValueInitialization
1073
1065
-0.7%
1.01x (?)
StringFromLongWholeSubstring
21
21
+0.0%
1.00x
StringFromLongWholeSubstringGeneric
97
98
+1.0%
0.99x (?)
StringHasPrefixAscii
1805
1775
-1.7%
1.02x
StringHasPrefixUnicode
111806
110203
-1.4%
1.01x
StringHasSuffixAscii
1947
1891
-2.9%
1.03x
StringHasSuffixUnicode
112843
113585
+0.7%
0.99x (?)
StringInterpolation
10139
10219
+0.8%
0.99x (?)
StringInterpolationManySmallSegments
21197
21809
+2.9%
0.97x
StringInterpolationSmall
8864
8803
-0.7%
1.01x
StringMatch
9523
9530
+0.1%
1.00x (?)
StringRemoveDupes
1355
1343
-0.9%
1.01x (?)
StringUTF16Builder
2535
2528
-0.3%
1.00x (?)
StringUTF16SubstringBuilder
5463
5466
+0.1%
1.00x (?)
StringWalk
1417
1467
+3.5%
0.97x
StringWithCString
43171
43184
+0.0%
1.00x (?)
StringWordBuilder
2409
2522
+4.7%
0.96x
StringWordBuilderReservingCapacity
2091
2166
+3.6%
0.97x
SubstringComparable
49
49
+0.0%
1.00x
SubstringEqualString
541
536
-0.9%
1.01x
SubstringEquatable
1359
1358
-0.1%
1.00x (?)
SubstringFromLongString
10
10
+0.0%
1.00x
SubstringFromLongStringGeneric
74
74
+0.0%
1.00x
SuffixAnyCollection
31
31
+0.0%
1.00x
SuffixAnyCollectionLazy
20261
20259
-0.0%
1.00x (?)
SuffixAnySeqCRangeIter
3769
3782
+0.3%
1.00x
SuffixAnySeqCRangeIterLazy
3776
3773
-0.1%
1.00x (?)
SuffixAnySeqCntRange
21
21
+0.0%
1.00x
SuffixAnySeqCntRangeLazy
21
21
+0.0%
1.00x
SuffixAnySequence
4873
4881
+0.2%
1.00x (?)
SuffixAnySequenceLazy
5008
4968
-0.8%
1.01x (?)
SuffixCountableRange
11
11
+0.0%
1.00x
SuffixCountableRangeLazy
11
11
+0.0%
1.00x
SuffixSequence
3644
3610
-0.9%
1.01x
SuffixSequenceLazy
3632
3620
-0.3%
1.00x (?)
SumUsingReduce
97
101
+4.1%
0.96x
SuperChars
41258
40432
-2.0%
1.02x (?)
TwoSum
1594
1614
+1.3%
0.99x
TypeFlood
0
0
+0.0%
1.00x
UTF8Decode
291
295
+1.4%
0.99x
Walsh
405
400
-1.2%
1.01x
WordCountHistogramASCII
7498
7297
-2.7%
1.03x
WordCountHistogramUTF16
13123
12954
-1.3%
1.01x (?)
WordCountUniqueASCII
2403
2401
-0.1%
1.00x (?)
WordCountUniqueUTF16
6565
6647
+1.2%
0.99x (?)
WordSplitASCII
19086
18805
-1.5%
1.01x (?)
WordSplitUTF16
21287
20674
-2.9%
1.03x (?)
XorLoop
402
396
-1.5%
1.02x
Added (2)
TEST
MIN
MAX
MEAN
MAX_RSS
ChainedFilterMap
1248
1249
1249
8323072
FatCompactMap
195537
195668
195599
8437760
Unoptimized (Onone)
Regression (10)
TEST
OLD
NEW
DELTA
SPEEDUP
StackPromo
95158
108646
+14.2%
0.88x
ReversedDictionary
22358
24617
+10.1%
0.91x
FilterEvenUsingReduceInto
1827
1993
+9.1%
0.92x
ObjectiveCBridgeStubToNSDate
15312
16688
+9.0%
0.92x (?)
StringInterpolationSmall
11994
13063
+8.9%
0.92x (?)
FloatingPointPrinting_Float_interpolated
135830
147581
+8.7%
0.92x (?)
DataAppendArray
5373
5734
+6.7%
0.94x
NSDictionaryCastToSwift
6075
6482
+6.7%
0.94x (?)
Array2D
633100
674530
+6.5%
0.94x
ArrayAppendReserved
4061
4282
+5.4%
0.95x
Improvement (9)
TEST
OLD
NEW
DELTA
SPEEDUP
CharacterPropertiesFetch
7408
5453
-26.4%
1.36x
ObjectiveCBridgeFromNSSetAnyObjectForced
5988
5345
-10.7%
1.12x
ObjectiveCBridgeStubFromArrayOfNSString
34853
31232
-10.4%
1.12x (?)
Combos
2190
2020
-7.8%
1.08x
DataSetCount
633
592
-6.5%
1.07x
ErrorHandling
7234
6807
-5.9%
1.06x (?)
StringBuilder
6687
6318
-5.5%
1.06x (?)
RemoveWhereFilterInts
2024
1914
-5.4%
1.06x
WordSplitUTF16
25214
24004
-4.8%
1.05x (?)
No Changes (404)
TEST
OLD
NEW
DELTA
SPEEDUP
AngryPhonebook
5091
5167
+1.5%
0.99x
AnyHashableWithAClass
98620
99296
+0.7%
0.99x (?)
ArrayAppend
4444
4570
+2.8%
0.97x
ArrayAppendArrayOfInt
859
862
+0.3%
1.00x
ArrayAppendAscii
39165
39319
+0.4%
1.00x (?)
ArrayAppendFromGeneric
865
866
+0.1%
1.00x (?)
ArrayAppendGenericStructs
1503
1501
-0.1%
1.00x (?)
ArrayAppendLatin1
63519
63444
-0.1%
1.00x
ArrayAppendLazyMap
162368
163090
+0.4%
1.00x
ArrayAppendOptionals
1496
1499
+0.2%
1.00x (?)
ArrayAppendRepeatCol
154463
154497
+0.0%
1.00x (?)
ArrayAppendSequence
99425
99428
+0.0%
1.00x (?)
ArrayAppendStrings
15393
15403
+0.1%
1.00x (?)
ArrayAppendToFromGeneric
865
861
-0.5%
1.00x (?)
ArrayAppendToGeneric
866
871
+0.6%
0.99x (?)
ArrayAppendUTF16
62351
62146
-0.3%
1.00x (?)
ArrayInClass
6144
6254
+1.8%
0.98x
ArrayLiteral
1799
1799
+0.0%
1.00x
ArrayOfGenericPOD2
1073
1073
+0.0%
1.00x
ArrayOfGenericRef
10811
10822
+0.1%
1.00x (?)
ArrayOfPOD
760
762
+0.3%
1.00x
ArrayOfRef
9924
9951
+0.3%
1.00x (?)
ArrayPlusEqualArrayOfInt
864
862
-0.2%
1.00x
ArrayPlusEqualFiveElementCollection
232069
233477
+0.6%
0.99x (?)
ArrayPlusEqualSingleElementCollection
230289
228396
-0.8%
1.01x (?)
ArrayPlusEqualThreeElements
9178
9223
+0.5%
1.00x
ArraySubscript
106664
107752
+1.0%
0.99x
ArrayValueProp
3690
3686
-0.1%
1.00x (?)
ArrayValueProp2
14802
14860
+0.4%
1.00x (?)
ArrayValueProp3
4175
4127
-1.1%
1.01x (?)
ArrayValueProp4
4123
4115
-0.2%
1.00x (?)
BinaryFloatingPointConversionFromBinaryInteger
5312
5293
-0.4%
1.00x (?)
BinaryFloatingPointPropertiesBinade
85
83
-2.4%
1.02x
BinaryFloatingPointPropertiesNextUp
131
125
-4.6%
1.05x
BinaryFloatingPointPropertiesUlp
129
125
-3.1%
1.03x
BitCount
8999
8615
-4.3%
1.04x
ByteSwap
9247
9465
+2.4%
0.98x
COWTree
11708
11514
-1.7%
1.02x (?)
CSVParsing
2415550
2390892
-1.0%
1.01x (?)
CSVParsingAlt
1370485
1350302
-1.5%
1.01x (?)
CSVParsingAltIndices
2235437
2139120
-4.3%
1.05x
CStringLongAscii
4447
4445
-0.0%
1.00x (?)
CStringLongNonAscii
2259
2273
+0.6%
0.99x (?)
CStringShortAscii
7680
7673
-0.1%
1.00x (?)
Calculator
1269
1275
+0.5%
1.00x
CaptureProp
235006
235368
+0.2%
1.00x
CharIndexing_ascii_unicodeScalars
299867
302100
+0.7%
0.99x (?)
CharIndexing_ascii_unicodeScalars_Backwards
339038
338131
-0.3%
1.00x (?)
CharIndexing_chinese_unicodeScalars
226342
228379
+0.9%
0.99x (?)
CharIndexing_chinese_unicodeScalars_Backwards
257882
256235
-0.6%
1.01x (?)
CharIndexing_japanese_unicodeScalars
362695
361798
-0.2%
1.00x (?)
CharIndexing_japanese_unicodeScalars_Backwards
407105
404974
-0.5%
1.01x (?)
CharIndexing_korean_unicodeScalars
290658
295485
+1.7%
0.98x (?)
CharIndexing_korean_unicodeScalars_Backwards
332845
330171
-0.8%
1.01x (?)
CharIndexing_punctuatedJapanese_unicodeScalars
53636
53461
-0.3%
1.00x (?)
CharIndexing_punctuatedJapanese_unicodeScalars_Backwards
58576
59832
+2.1%
0.98x (?)
CharIndexing_punctuated_unicodeScalars
66528
67112
+0.9%
0.99x
CharIndexing_punctuated_unicodeScalars_Backwards
74277
74054
-0.3%
1.00x (?)
CharIndexing_russian_unicodeScalars
248690
252445
+1.5%
0.99x
CharIndexing_russian_unicodeScalars_Backwards
285172
283483
-0.6%
1.01x (?)
CharIndexing_tweet_unicodeScalars
597571
599734
+0.4%
1.00x
CharIndexing_tweet_unicodeScalars_Backwards
670848
675748
+0.7%
0.99x (?)
CharIndexing_utf16_unicodeScalars
265853
263382
-0.9%
1.01x (?)
CharIndexing_utf16_unicodeScalars_Backwards
290887
289293
-0.5%
1.01x (?)
CharIteration_ascii_unicodeScalars
144830
145341
+0.4%
1.00x (?)
CharIteration_ascii_unicodeScalars_Backwards
249431
246595
-1.1%
1.01x (?)
CharIteration_chinese_unicodeScalars
109708
110011
+0.3%
1.00x (?)
CharIteration_chinese_unicodeScalars_Backwards
188092
187521
-0.3%
1.00x (?)
CharIteration_japanese_unicodeScalars
174308
174365
+0.0%
1.00x (?)
CharIteration_japanese_unicodeScalars_Backwards
300679
296543
-1.4%
1.01x
CharIteration_korean_unicodeScalars
140987
141092
+0.1%
1.00x (?)
CharIteration_korean_unicodeScalars_Backwards
241259
240327
-0.4%
1.00x
CharIteration_punctuatedJapanese_unicodeScalars
25783
25914
+0.5%
0.99x (?)
CharIteration_punctuatedJapanese_unicodeScalars_Backwards
43217
42916
-0.7%
1.01x (?)
CharIteration_punctuated_unicodeScalars
32426
32601
+0.5%
0.99x (?)
CharIteration_punctuated_unicodeScalars_Backwards
54950
54527
-0.8%
1.01x (?)
CharIteration_russian_unicodeScalars
120795
121264
+0.4%
1.00x (?)
CharIteration_russian_unicodeScalars_Backwards
207276
205268
-1.0%
1.01x (?)
CharIteration_tweet_unicodeScalars
286514
287674
+0.4%
1.00x (?)
CharIteration_tweet_unicodeScalars_Backwards
501309
491156
-2.0%
1.02x (?)
CharIteration_utf16_unicodeScalars
124410
125329
+0.7%
0.99x (?)
CharIteration_utf16_unicodeScalars_Backwards
215567
215160
-0.2%
1.00x (?)
CharacterLiteralsLarge
5782
5802
+0.3%
1.00x (?)
CharacterLiteralsSmall
658
659
+0.2%
1.00x (?)
CharacterPropertiesPrecomputed
3829
3794
-0.9%
1.01x (?)
CharacterPropertiesStashed
2335
2319
-0.7%
1.01x (?)
CharacterPropertiesStashedMemo
4626
4586
-0.9%
1.01x (?)
Chars
34757
34730
-0.1%
1.00x (?)
ClassArrayGetter
985
986
+0.1%
1.00x
DataAccessBytes
2343
2336
-0.3%
1.00x (?)
DataAppendBytes
5160
5098
-1.2%
1.01x (?)
DataAppendDataLargeToLarge
67998
67961
-0.1%
1.00x (?)
DataAppendDataLargeToMedium
35889
35671
-0.6%
1.01x (?)
DataAppendDataLargeToSmall
34639
34783
+0.4%
1.00x (?)
DataAppendDataMediumToLarge
38226
37703
-1.4%
1.01x
DataAppendDataMediumToMedium
6980
6844
-1.9%
1.02x (?)
DataAppendDataMediumToSmall
5977
6071
+1.6%
0.98x (?)
DataAppendDataSmallToLarge
36563
36843
+0.8%
0.99x
DataAppendDataSmallToMedium
6190
6315
+2.0%
0.98x (?)
DataAppendDataSmallToSmall
6099
5894
-3.4%
1.03x (?)
DataAppendSequence
1593337
1588844
-0.3%
1.00x (?)
DataCopyBytes
2463
2477
+0.6%
0.99x (?)
DataCount
223
223
+0.0%
1.00x
DataMutateBytes
5425
5435
+0.2%
1.00x (?)
DataReplaceLarge
40434
40017
-1.0%
1.01x (?)
DataReplaceLargeBuffer
57801
57675
-0.2%
1.00x (?)
DataReplaceMedium
9172
9246
+0.8%
0.99x (?)
DataReplaceMediumBuffer
11857
12082
+1.9%
0.98x (?)
DataReplaceSmall
6857
6862
+0.1%
1.00x (?)
DataReplaceSmallBuffer
8479
8542
+0.7%
0.99x (?)
DataReset
2854
2825
-1.0%
1.01x (?)
DataSubscript
441
441
+0.0%
1.00x
DictOfArraysToArrayOfDicts
3394
3377
-0.5%
1.01x (?)
Dictionary
2279
2289
+0.4%
1.00x (?)
Dictionary2
2561
2561
+0.0%
1.00x
Dictionary2OfObjects
5816
5871
+0.9%
0.99x (?)
Dictionary3
1159
1155
-0.3%
1.00x (?)
Dictionary3OfObjects
2379
2329
-2.1%
1.02x (?)
Dictionary4
1526
1539
+0.9%
0.99x (?)
Dictionary4OfObjects
2219
2142
-3.5%
1.04x
DictionaryBridge
2120
2113
-0.3%
1.00x (?)
DictionaryCopy
307589
311704
+1.3%
0.99x (?)
DictionaryFilter
304174
312245
+2.7%
0.97x
DictionaryGroup
4008
4005
-0.1%
1.00x (?)
DictionaryGroupOfObjects
7328
7323
-0.1%
1.00x (?)
DictionaryLiteral
8751
8688
-0.7%
1.01x (?)
DictionaryOfObjects
6113
6044
-1.1%
1.01x (?)
DictionaryRemove
17873
17542
-1.9%
1.02x
DictionaryRemoveOfObjects
53953
54583
+1.2%
0.99x (?)
DictionarySubscriptDefaultMutation
2008
2019
+0.5%
0.99x (?)
DictionarySubscriptDefaultMutationArray
2278
2273
-0.2%
1.00x (?)
DictionarySubscriptDefaultMutationArrayOfObjects
9359
9403
+0.5%
1.00x (?)
DictionarySubscriptDefaultMutationOfObjects
5406
5363
-0.8%
1.01x (?)
DictionarySwap
5252
5261
+0.2%
1.00x
DictionarySwapAt
35422
35352
-0.2%
1.00x (?)
DictionarySwapAtOfObjects
114386
113496
-0.8%
1.01x (?)
DictionarySwapOfObjects
19801
19967
+0.8%
0.99x (?)
DoubleWidthDivision
0
0
+0.0%
1.00x
DropFirstAnyCollection
12549
12440
-0.9%
1.01x
DropFirstAnyCollectionLazy
90028
89536
-0.5%
1.01x (?)
DropFirstAnySeqCRangeIter
19688
19764
+0.4%
1.00x
DropFirstAnySeqCRangeIterLazy
19618
19663
+0.2%
1.00x
DropFirstAnySeqCntRange
12557
12456
-0.8%
1.01x
DropFirstAnySeqCntRangeLazy
12520
12441
-0.6%
1.01x
DropFirstAnySequence
11331
11319
-0.1%
1.00x (?)
DropFirstAnySequenceLazy
11349
11276
-0.6%
1.01x (?)
DropFirstArray
3129
3104
-0.8%
1.01x
DropFirstArrayLazy
22080
22218
+0.6%
0.99x (?)
DropFirstCountableRange
309
309
+0.0%
1.00x
DropFirstCountableRangeLazy
23585
23409
-0.7%
1.01x
DropFirstSequence
10757
10813
+0.5%
0.99x (?)
DropFirstSequenceLazy
10822
10800
-0.2%
1.00x (?)
DropLastAnyCollection
4199
4160
-0.9%
1.01x
DropLastAnyCollectionLazy
29018
28784
-0.8%
1.01x (?)
DropLastAnySeqCRangeIter
37083
37023
-0.2%
1.00x (?)
DropLastAnySeqCRangeIterLazy
37358
37059
-0.8%
1.01x (?)
DropLastAnySeqCntRange
4185
4149
-0.9%
1.01x
DropLastAnySeqCntRangeLazy
4188
4169
-0.5%
1.00x
DropLastAnySequence
29384
29417
+0.1%
1.00x (?)
DropLastAnySequenceLazy
29826
29779
-0.2%
1.00x
DropLastCountableRange
110
107
-2.7%
1.03x
DropLastCountableRangeLazy
7859
7854
-0.1%
1.00x (?)
DropLastSequence
29384
29258
-0.4%
1.00x
DropLastSequenceLazy
29638
29446
-0.6%
1.01x
DropWhileAnyCollection
16004
15926
-0.5%
1.00x
DropWhileAnyCollectionLazy
18068
18100
+0.2%
1.00x
DropWhileAnySeqCRangeIter
21200
21295
+0.4%
1.00x
DropWhileAnySeqCRangeIterLazy
17980
18044
+0.4%
1.00x (?)
DropWhileAnySeqCntRange
16045
16009
-0.2%
1.00x
DropWhileAnySeqCntRangeLazy
18148
18199
+0.3%
1.00x
DropWhileAnySequence
12950
12841
-0.8%
1.01x (?)
DropWhileAnySequenceLazy
10831
10863
+0.3%
1.00x (?)
DropWhileArrayLazy
13570
13722
+1.1%
0.99x
DropWhileCountableRange
3929
3930
+0.0%
1.00x (?)
DropWhileCountableRangeLazy
17602
17625
+0.1%
1.00x (?)
DropWhileSequence
12386
12474
+0.7%
0.99x
DropWhileSequenceLazy
10433
10349
-0.8%
1.01x (?)
EqualStringSubstring
73
73
+0.0%
1.00x
EqualSubstringString
73
73
+0.0%
1.00x
EqualSubstringSubstring
74
74
+0.0%
1.00x
EqualSubstringSubstringGenericEquatable
56
56
+0.0%
1.00x
ExclusivityGlobal
183
183
+0.0%
1.00x
ExclusivityIndependent
71
72
+1.4%
0.99x (?)
FilterEvenUsingReduce
3553
3691
+3.9%
0.96x
FloatingPointPrinting_Double_description_small
81088
81010
-0.1%
1.00x (?)
FloatingPointPrinting_Double_description_uniform
154426
148719
-3.7%
1.04x (?)
FloatingPointPrinting_Double_interpolated
252590
251935
-0.3%
1.00x (?)
FloatingPointPrinting_Float80_description_small
94282
93067
-1.3%
1.01x
FloatingPointPrinting_Float80_description_uniform
1736455
1747659
+0.6%
0.99x
FloatingPointPrinting_Float80_interpolated
1838540
1838723
+0.0%
1.00x (?)
FloatingPointPrinting_Float_description_small
44464
43591
-2.0%
1.02x (?)
FloatingPointPrinting_Float_description_uniform
59783
57970
-3.0%
1.03x (?)
FrequenciesUsingReduce
13596
13498
-0.7%
1.01x (?)
FrequenciesUsingReduceInto
5593
5674
+1.4%
0.99x (?)
Hanoi
20185
20631
+2.2%
0.98x
HashTest
22111
21622
-2.2%
1.02x
Histogram
6382
6391
+0.1%
1.00x
Integrate
458
460
+0.4%
1.00x
IterateData
5682
5778
+1.7%
0.98x
Join
1118
1076
-3.8%
1.04x
LazilyFilteredArrayContains
730238
733046
+0.4%
1.00x
LazilyFilteredArrays
1376021
1375433
-0.0%
1.00x (?)
LazilyFilteredRange
438384
438011
-0.1%
1.00x
LessSubstringSubstring
74
73
-1.4%
1.01x
LessSubstringSubstringGenericComparable
55
55
+0.0%
1.00x
LinkedList
32077
32294
+0.7%
0.99x
LuhnAlgoEager
5739
5722
-0.3%
1.00x (?)
LuhnAlgoLazy
5865
5860
-0.1%
1.00x (?)
MapReduce
25425
25430
+0.0%
1.00x (?)
MapReduceAnyCollection
25540
25459
-0.3%
1.00x (?)
MapReduceAnyCollectionShort
35585
36104
+1.5%
0.99x (?)
MapReduceClass
29670
29689
+0.1%
1.00x (?)
MapReduceClassShort
39296
40068
+2.0%
0.98x (?)
MapReduceLazyCollection
21784
21867
+0.4%
1.00x (?)
MapReduceLazyCollectionShort
33721
32887
-2.5%
1.03x (?)
MapReduceLazySequence
18553
18545
-0.0%
1.00x (?)
MapReduceSequence
29100
29162
+0.2%
1.00x (?)
MapReduceShort
37268
36373
-2.4%
1.02x (?)
MapReduceShortString
221
228
+3.2%
0.97x
MapReduceString
1758
1762
+0.2%
1.00x (?)
Memset
44131
44126
-0.0%
1.00x (?)
MonteCarloE
1129454
1148399
+1.7%
0.98x
MonteCarloPi
5136177
5208437
+1.4%
0.99x
NSError
723
729
+0.8%
0.99x (?)
NSStringConversion
444
443
-0.2%
1.00x
NibbleSort
427361
430132
+0.6%
0.99x
NopDeinit
188551
190924
+1.3%
0.99x (?)
ObjectAllocation
1253
1254
+0.1%
1.00x (?)
ObjectiveCBridgeFromNSArrayAnyObject
27542
27169
-1.4%
1.01x (?)
ObjectiveCBridgeFromNSArrayAnyObjectForced
9664
9648
-0.2%
1.00x (?)
ObjectiveCBridgeFromNSArrayAnyObjectToString
45131
46984
+4.1%
0.96x (?)
ObjectiveCBridgeFromNSArrayAnyObjectToStringForced
44179
43924
-0.6%
1.01x (?)
ObjectiveCBridgeFromNSDictionaryAnyObject
105375
106913
+1.5%
0.99x (?)
ObjectiveCBridgeFromNSSetAnyObject
52411
51272
-2.2%
1.02x (?)
ObjectiveCBridgeFromNSSetAnyObjectToString
79881
82141
+2.8%
0.97x (?)
ObjectiveCBridgeFromNSString
2669
2658
-0.4%
1.00x (?)
ObjectiveCBridgeFromNSStringForced
2640
2584
-2.1%
1.02x (?)
ObjectiveCBridgeStubDataAppend
5564
5522
-0.8%
1.01x (?)
ObjectiveCBridgeStubDateMutation
716
716
+0.0%
1.00x
ObjectiveCBridgeStubFromNSDate
6684
6595
-1.3%
1.01x (?)
ObjectiveCBridgeStubFromNSString
973
980
+0.7%
0.99x (?)
ObjectiveCBridgeStubFromNSStringRef
193
196
+1.6%
0.98x
ObjectiveCBridgeStubNSDataAppend
2911
2959
+1.6%
0.98x
ObjectiveCBridgeStubNSDateMutationRef
16139
16264
+0.8%
0.99x (?)
ObjectiveCBridgeStubToArrayOfNSString
27851
27966
+0.4%
1.00x (?)
ObjectiveCBridgeStubToNSDateRef
3457
3482
+0.7%
0.99x (?)
ObjectiveCBridgeStubToNSString
1614
1584
-1.9%
1.02x
ObjectiveCBridgeStubToNSStringRef
153
153
+0.0%
1.00x
ObjectiveCBridgeStubURLAppendPath
373502
363286
-2.7%
1.03x (?)
ObjectiveCBridgeStubURLAppendPathRef
368089
374099
+1.6%
0.98x (?)
ObjectiveCBridgeToNSArray
28806
28375
-1.5%
1.02x (?)
ObjectiveCBridgeToNSDictionary
43170
44958
+4.1%
0.96x (?)
ObjectiveCBridgeToNSSet
34370
34535
+0.5%
1.00x (?)
ObjectiveCBridgeToNSString
1353
1348
-0.4%
1.00x (?)
ObserverClosure
6411
6408
-0.0%
1.00x (?)
ObserverForwarderStruct
4250
4256
+0.1%
1.00x (?)
ObserverPartiallyAppliedMethod
8000
7931
-0.9%
1.01x
ObserverUnappliedMethod
8177
8053
-1.5%
1.02x
OpenClose
530
512
-3.4%
1.04x
PartialApplyDynamicType
38951
39374
+1.1%
0.99x (?)
Phonebook
17655
17798
+0.8%
0.99x
PointerArithmetics
126076
125999
-0.1%
1.00x (?)
PolymorphicCalls
2439
2450
+0.5%
1.00x
PopFrontArray
4641
4654
+0.3%
1.00x (?)
PopFrontArrayGeneric
5494
5469
-0.5%
1.00x (?)
PopFrontUnsafePointer
10649
10712
+0.6%
0.99x (?)
PrefixAnyCollection
12541
12416
-1.0%
1.01x
PrefixAnyCollectionLazy
89301
89184
-0.1%
1.00x (?)
PrefixAnySeqCRangeIter
15985
15945
-0.3%
1.00x (?)
PrefixAnySeqCRangeIterLazy
15928
15934
+0.0%
1.00x (?)
PrefixAnySeqCntRange
12542
12415
-1.0%
1.01x
PrefixAnySeqCntRangeLazy
12561
12458
-0.8%
1.01x
PrefixAnySequence
9518
9399
-1.3%
1.01x
PrefixAnySequenceLazy
9393
9418
+0.3%
1.00x (?)
PrefixArray
3143
3099
-1.4%
1.01x
PrefixArrayLazy
22029
22233
+0.9%
0.99x
PrefixCountableRange
317
309
-2.5%
1.03x
PrefixCountableRangeLazy
23530
23381
-0.6%
1.01x
PrefixSequence
8940
8792
-1.7%
1.02x
PrefixSequenceLazy
8749
8718
-0.4%
1.00x (?)
PrefixWhileAnyCollection
23139
23092
-0.2%
1.00x
PrefixWhileAnyCollectionLazy
15170
15009
-1.1%
1.01x
PrefixWhileAnySeqCRangeIter
31600
32133
+1.7%
0.98x
PrefixWhileAnySeqCRangeIterLazy
15006
15169
+1.1%
0.99x
PrefixWhileAnySeqCntRange
23231
23197
-0.1%
1.00x (?)
PrefixWhileAnySeqCntRangeLazy
15109
15106
-0.0%
1.00x (?)
PrefixWhileAnySequence
25561
25734
+0.7%
0.99x
PrefixWhileAnySequenceLazy
9684
9766
+0.8%
0.99x (?)
PrefixWhileArray
10227
10290
+0.6%
0.99x (?)
PrefixWhileArrayLazy
11779
11748
-0.3%
1.00x (?)
PrefixWhileCountableRange
11062
11039
-0.2%
1.00x (?)
PrefixWhileCountableRangeLazy
14679
14684
+0.0%
1.00x (?)
PrefixWhileSequence
25285
25280
-0.0%
1.00x (?)
PrefixWhileSequenceLazy
9476
9433
-0.5%
1.00x (?)
Prims
9576
9623
+0.5%
1.00x (?)
PrimsSplit
9592
9597
+0.1%
1.00x (?)
QueueConcrete
14035
14322
+2.0%
0.98x
QueueGeneric
18101
18135
+0.2%
1.00x
RC4
16406
16598
+1.2%
0.99x
RGBHistogram
24526
24559
+0.1%
1.00x
RGBHistogramOfObjects
77817
77960
+0.2%
1.00x (?)
RangeAssignment
2645
2659
+0.5%
0.99x
RangeIterationSigned
14387
14589
+1.4%
0.99x (?)
RangeIterationSigned64
35170
35053
-0.3%
1.00x (?)
RangeIterationUnsigned
31810
31772
-0.1%
1.00x (?)
RangeReplaceableCollectionPlusDefault
10006
9976
-0.3%
1.00x (?)
RecursiveOwnedParameter
5771
5773
+0.0%
1.00x (?)
RemoveWhereFilterString
1281
1279
-0.2%
1.00x (?)
RemoveWhereFilterStrings
2432
2475
+1.8%
0.98x
RemoveWhereMoveInts
3362
3357
-0.1%
1.00x (?)
RemoveWhereMoveStrings
3884
3887
+0.1%
1.00x
RemoveWhereQuadraticInts
7595
7571
-0.3%
1.00x
RemoveWhereQuadraticString
2243
2234
-0.4%
1.00x (?)
RemoveWhereQuadraticStrings
9289
9230
-0.6%
1.01x (?)
RemoveWhereSwapInts
6026
6042
+0.3%
1.00x
RemoveWhereSwapStrings
6731
6753
+0.3%
1.00x
ReversedArray
12422
12445
+0.2%
1.00x (?)
ReversedBidirectional
42502
42323
-0.4%
1.00x (?)
RomanNumbers
1111536
1116798
+0.5%
1.00x (?)
SequenceAlgosAnySequence
10177
10199
+0.2%
1.00x
SequenceAlgosArray
760575
734662
-3.4%
1.04x
SequenceAlgosContiguousArray
282808
274793
-2.8%
1.03x
SequenceAlgosList
8315
8243
-0.9%
1.01x
SequenceAlgosRange
999990
1000278
+0.0%
1.00x (?)
SequenceAlgosUnfoldSequence
5747
5738
-0.2%
1.00x (?)
SetExclusiveOr
17413
17426
+0.1%
1.00x (?)
SetExclusiveOr_OfObjects
45768
45460
-0.7%
1.01x (?)
SetIntersect
7169
7165
-0.1%
1.00x (?)
SetIntersect_OfObjects
11123
11322
+1.8%
0.98x (?)
SetIsSubsetOf
1263
1272
+0.7%
0.99x
SetIsSubsetOf_OfObjects
1793
1818
+1.4%
0.99x
SetUnion
12379
12322
-0.5%
1.00x (?)
SetUnion_OfObjects
32210
32035
-0.5%
1.01x (?)
SevenBoom
1647
1663
+1.0%
0.99x
Sim2DArray
43441
43483
+0.1%
1.00x (?)
SortLargeExistentials
10958
10976
+0.2%
1.00x (?)
SortLettersInPlace
1805
1806
+0.1%
1.00x (?)
SortSortedStrings
935
916
-2.0%
1.02x
SortStrings
1832
1856
+1.3%
0.99x
SortStringsUnicode
2622
2600
-0.8%
1.01x
StaticArray
2604
2611
+0.3%
1.00x (?)
StrComplexWalk
6204
6199
-0.1%
1.00x
StrToInt
72696
72867
+0.2%
1.00x (?)
StringAdder
4668
4679
+0.2%
1.00x (?)
StringBuilderLong
2754
2748
-0.2%
1.00x (?)
StringBuilderWithLongSubstring
4488
4470
-0.4%
1.00x (?)
StringComparison_abnormal
1352
1317
-2.6%
1.03x
StringComparison_ascii
8598
8583
-0.2%
1.00x
StringComparison_emoji
1927
1924
-0.2%
1.00x (?)
StringComparison_fastPrenormal
4678
4705
+0.6%
0.99x
StringComparison_latin1
3648
3633
-0.4%
1.00x (?)
StringComparison_longSharedPrefix
2297
2289
-0.3%
1.00x
StringComparison_nonBMPSlowestPrenormal
3559
3552
-0.2%
1.00x (?)
StringComparison_slowerPrenormal
4022
4012
-0.2%
1.00x (?)
StringComparison_zalgo
125468
125447
-0.0%
1.00x (?)
StringEdits
338816
343958
+1.5%
0.99x (?)
StringEnumRawValueInitialization
17082
17019
-0.4%
1.00x (?)
StringEqualPointerComparison
1547
1517
-1.9%
1.02x
StringFromLongWholeSubstring
22
22
+0.0%
1.00x
StringFromLongWholeSubstringGeneric
314
318
+1.3%
0.99x (?)
StringHasPrefixAscii
3199
3162
-1.2%
1.01x
StringHasPrefixUnicode
114193
113165
-0.9%
1.01x (?)
StringHasSuffixAscii
3318
3245
-2.2%
1.02x
StringHasSuffixUnicode
106116
107408
+1.2%
0.99x (?)
StringInterpolation
14264
14509
+1.7%
0.98x (?)
StringInterpolationManySmallSegments
22752
22731
-0.1%
1.00x (?)
StringMatch
33021
32334
-2.1%
1.02x
StringRemoveDupes
1387
1418
+2.2%
0.98x (?)
StringUTF16Builder
7214
7311
+1.3%
0.99x (?)
StringUTF16SubstringBuilder
21255
21121
-0.6%
1.01x (?)
StringWalk
12594
12742
+1.2%
0.99x (?)
StringWithCString
37562
37562
+0.0%
1.00x
StringWordBuilder
2446
2439
-0.3%
1.00x
StringWordBuilderReservingCapacity
2121
2118
-0.1%
1.00x (?)
SubstringComparable
1589
1572
-1.1%
1.01x
SubstringEqualString
1675
1676
+0.1%
1.00x (?)
SubstringEquatable
5274
5379
+2.0%
0.98x
SubstringFromLongString
24
24
+0.0%
1.00x
SubstringFromLongStringGeneric
101
101
+0.0%
1.00x
SuffixAnyCollection
4217
4166
-1.2%
1.01x
SuffixAnyCollectionLazy
29970
29059
-3.0%
1.03x (?)
SuffixAnySeqCRangeIter
32616
32536
-0.2%
1.00x (?)
SuffixAnySeqCRangeIterLazy
32663
32637
-0.1%
1.00x (?)
SuffixAnySeqCntRange
4183
4164
-0.5%
1.00x
SuffixAnySeqCntRangeLazy
4199
4163
-0.9%
1.01x
SuffixAnySequence
25050
25079
+0.1%
1.00x (?)
SuffixAnySequenceLazy
25242
25279
+0.1%
1.00x (?)
SuffixCountableRange
105
105
+0.0%
1.00x
SuffixCountableRangeLazy
7851
7854
+0.0%
1.00x (?)
SuffixSequence
24922
25030
+0.4%
1.00x
SuffixSequenceLazy
24937
25033
+0.4%
1.00x (?)
SumUsingReduce
159243
158381
-0.5%
1.01x (?)
SumUsingReduceInto
149800
151236
+1.0%
0.99x
SuperChars
123894
122950
-0.8%
1.01x (?)
TwoSum
3643
3666
+0.6%
0.99x (?)
TypeFlood
191
194
+1.6%
0.98x (?)
UTF8Decode
29181
29330
+0.5%
0.99x
Walsh
12097
12022
-0.6%
1.01x (?)
WordCountHistogramASCII
34825
34809
-0.0%
1.00x (?)
WordCountHistogramUTF16
43334
43534
+0.5%
1.00x (?)
WordCountUniqueASCII
7411
7411
+0.0%
1.00x
WordCountUniqueUTF16
11813
11835
+0.2%
1.00x (?)
WordSplitASCII
21896
22552
+3.0%
0.97x (?)
XorLoop
23227
23228
+0.0%
1.00x (?)
Added (2)
TEST
MIN
MAX
MEAN
MAX_RSS
ChainedFilterMap
224542
240224
229985
8867840
FatCompactMap
294530
298007
295731
8863744
Hardware Overview
Model Name: Mac Pro
Model Identifier: MacPro6,1
Processor Name: 12-Core Intel Xeon E5
Processor Speed: 2.7 GHz
Number of Processors: 1
Total Number of Cores: 12
L2 Cache (per Core): 256 KB
L3 Cache: 30 MB
Memory: 64 GB
|
gharchive/pull-request
| 2018-03-20T19:01:44 |
2025-04-01T04:56:02.166051
|
{
"authors": [
"moiseev",
"swift-ci"
],
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/15376",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
343768469
|
Don’t dive into null parts of AST.
When compiling code with certain errors, it is possible to get null pointers in the AST.
The compiler indexes even when there are errors, and the indexing code then tries to walk into the null pointers, causing a crash. As a result, the frontend crashes, and the developer does not get to see the errors. This PR adds a null check for this specific case. It does not solve the general problem.
Resolves rdar://42314665
@swift-ci please smoke test
@akyrtzi Did you want to look at this? Thanks!
Generally it would be ideal if you could add a regression test but I can understand if it is too difficult to reduce the original failing test case.
@swift-ci please smoke test
I agree a regression test would be great. If I can come up with one, I'll do another PR.
|
gharchive/pull-request
| 2018-07-23T19:41:21 |
2025-04-01T04:56:02.175237
|
{
"authors": [
"akyrtzi",
"davidungar"
],
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/18153",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.