id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
2300738361
Feat: Add SignIn with google option Describe the feature We can add google authentication to directly sign in using google firebase. Add ScreenShots Record [X] I agree to follow this project's Code of Conduct [X] I'm a GSSOC'24 contributor [X] I want to work on this issue @anuragverma108 Please look at this issue raised
gharchive/issue
2024-05-16T15:37:54
2025-04-01T04:55:56.755449
{ "authors": [ "aryansharma220" ], "repo": "anuragverma108/SwapReads", "url": "https://github.com/anuragverma108/SwapReads/issues/549", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2579170082
enhanced the login page issue no :- #728 This pull request focuses on enhancing the user interface of the login page by adjusting the layout of the social media login buttons. The goal is to create a more user-friendly and visually appealing experience for users attempting to log in via social media accounts. Key Changes: Button Alignment: Adjusted the alignment of the social media buttons to be centrally aligned, providing a more balanced look. Added consistent spacing between buttons to prevent a cluttered appearance. Button Styling: Updated the button styles to match the overall theme of the login page, ensuring that they are visually cohesive with other UI elements. Implemented hover effects for buttons to enhance interactivity and provide feedback to users. Responsive Design: Ensured that the layout is responsive, making it accessible on various devices, including mobile and tablet screens. Utilized media queries to adjust button sizes and spacing on smaller screens for optimal usability. Attach screenshots of the changed sections After and before. @Kedar-sonavani @Kedar-sonavani check this PR #863 , don't you think those social media buttons without text are more attractive than yours with text ? Just wander to know. @Kedar-sonavani check this PR #863 , don't you think those social media buttons without text are more attractive than yours with text ? Just wander to know. Yeah but I think so my blur effect is better than that @Kedar-sonavani I appreciate your design bro !
gharchive/pull-request
2024-10-10T15:19:51
2025-04-01T04:55:56.759944
{ "authors": [ "Kedar-sonavani", "akash70629" ], "repo": "anuragverma108/WildGuard", "url": "https://github.com/anuragverma108/WildGuard/pull/869", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
111720225
gexf file from gephi to 3d graph? First of all: amazing work! here you say that the ngraph generated from a .gexf file could also easily be rendered by using e.g. ngraph.three as the renderer. I changed ngraph.pixi to ngraph.three, but it doesn't seem to work. Could you please help? Here is some raw code doing what you're (were) looking for module.exports.main = async function() { let query = require('query-string').parse(window.location.search.substring(1)); let graph = await getGraphFromDisk(query); let createThree = require('ngraph.three'); let graphics = createThree(graph, { interactive: true }); graphics.run(); // begin animation loop: graphics.camera.position.z = getNumber(query.z, 2000); }; function getGraphFromDisk(query) { let gexf = require('ngraph.gexf'); let filename = query.filename || 'graph.gexf'; return new Promise(function(resolve) { let req = new XMLHttpRequest(); req.open('GET', filename); req.onload = function() { let content = this.responseText; let graph = gexf.load(content); resolve(graph); }; req.send(); }); } function getNumber(string, defaultValue) { let number = parseFloat(string); return (typeof number === 'number') && !isNaN(number) ? number : (defaultValue || 10); } Starting a local web server like python3 -m http.server and browsing to http://localhost:8000
gharchive/issue
2015-10-15T22:22:59
2025-04-01T04:55:56.766569
{ "authors": [ "sandboxdp", "ticapix" ], "repo": "anvaka/ngraph", "url": "https://github.com/anvaka/ngraph/issues/21", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2588803665
Add popcnt target feature Problem Without popcnt feature the bitwise operations will use the standard library implementation instead of the hardware supported popcnt instruction. Instances where it will affect: bloom/src/bloom.rs:223: let num_bits_set = bits.iter().map(|x| x.count_ones() as u64).sum(); core/src/banking_stage/transaction_scheduler/thread_aware_account_locks.rs:394: self.0.count_ones() core/src/banking_stage/transaction_scheduler/thread_aware_account_locks.rs:399: (self.num_threads() == 1).then_some(self.0.trailing_zeros() as ThreadId) On x86 popcnt is a single instruction but without this target-feature rustc will generate something like this: (https://godbolt.org/z/69oc98chW) ; Without the patch example::count1::h62041e16c115ac6a: mov eax, edi shr eax and eax, 1431655765 sub edi, eax mov eax, edi and eax, 858993459 shr edi, 2 and edi, 858993459 add edi, eax mov eax, edi shr eax, 4 add eax, edi and eax, 252645135 imul eax, eax, 16843009 shr eax, 24 ret ; With the patch example::count1: popcnt eax,edi ret Summary of Changes Add target-feature popcnt to generate single instruction for popcnt, count leading zeros etc. I’m happy to help integrate this into the ci process, including tesing and releasing. however, I’ll defer to another domain expert for the reviewing. The std library implementation uses the hardware instruction if it is available. User can always specify the -C target-cpu=native or their specifici cpu architecture to give the compiler knowledge of instruction set. I'm guess I'm not sure I understand why we'd specifically enable this feature, only for x86. There are other instruction features I'd expect to have much more significant impact than popcnt, so why this one? And since popcnt is not part of the x86 baseline (see https://en.wikipedia.org/wiki/X86-64#Microarchitecture_levels) I'm not sure how rustc handles if this is not present. Does it just fail to compile? I don't think anyone is running/compiling on such old targets, so maybe that's fine? I don't think we should enable popcnt explicitly. If we wanted to do something, we could bump to a more recent target-cpu family (that among other things will include popcnt). But I'm not sure it's worth doing, I don't think that many people are using our binaries, most people compile from source with -C target-cpu=native which will automatically select the best depending on the host machine. Agree with this. Additionally target-feature is considered unsafe according to rustc, so probably don't want to use that directly anyway. From rustc -C help -C target-feature=val -- target specific attributes. (`rustc --print target-features` for details). This feature is unsafe. If this is the case, then all good. But how do we ensure they are doing it? That's the best part, we don't need to ensure they are doing it. That's why there is a fallback
gharchive/pull-request
2024-10-15T13:34:41
2025-04-01T04:55:56.779237
{ "authors": [ "apfitzge", "ksolana", "yihau" ], "repo": "anza-xyz/agave", "url": "https://github.com/anza-xyz/agave/pull/3176", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2276668485
Firebase auth The signin dynamic function seems like a good place to start but I am at a loss to where to start with firebase authentication with this wallet. Is there anybody who implemented this solution? Can we use firebase authentication with message signing so users are logged in to firebase also? Wallet-adapter is not a wallet. You'd need to create a wallet that works with Firebase auth. See https://github.com/anza-xyz/wallet-adapter/blob/master/WALLET.md for information on creating a wallet that is compatible with wallet-adapter.
gharchive/issue
2024-05-02T23:38:46
2025-04-01T04:55:56.781596
{ "authors": [ "agolho", "mcintyre94" ], "repo": "anza-xyz/wallet-adapter", "url": "https://github.com/anza-xyz/wallet-adapter/issues/958", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
84963351
Would Moloch work with an existing ES cluster that already has some indices in it? I already have an ES cluster, holding some other indices. There's plenty of capacity, so I'm thinking of using it for moloch capture as well. Would the additional indices cause a problem, say with the viewer or the API queries? yes it will
gharchive/issue
2015-06-04T07:39:22
2025-04-01T04:55:56.790412
{ "authors": [ "awick", "thongsia" ], "repo": "aol/moloch", "url": "https://github.com/aol/moloch/issues/394", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
869798525
Failed with Drone: Please provide compiled classes of your project with sonar.java.binaries property ERROR: Error during SonarScanner execution 64 | org.sonar.java.AnalysisException: Please provide compiled classes of your project with sonar.java.binaries property 65 | at org.sonar.java.JavaClasspath.init(JavaClasspath.java:64) 66 | at org.sonar.java.AbstractJavaClasspath.getElements(AbstractJavaClasspath.java:280) 67 | at I use the lastest plugin image, and my .drone.yml just like this: - name: sonar image: aosapps/drone-sonar-plugin pull: if-not-exists settings: sonar_host: from_secret: sonar_host sonar_token: from_secret: sonar_token when: event: - push Please help. OK, I fix this by adding a new environment parameter JAVA_BINARIES and build my own image. Check it here: mailbyms/drone-sonar-plugin
gharchive/issue
2021-04-28T10:36:23
2025-04-01T04:55:56.821986
{ "authors": [ "mailbyms" ], "repo": "aosapps/drone-sonar-plugin", "url": "https://github.com/aosapps/drone-sonar-plugin/issues/30", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2375434564
🛑 Main Site is down In 37fea88, Main Site (https://aosp.app) was down: HTTP code: 502 Response time: 617 ms Resolved: Main Site is back up in f665d50 after 17 minutes.
gharchive/issue
2024-06-26T13:55:37
2025-04-01T04:55:56.831184
{ "authors": [ "Ylarod" ], "repo": "aospapp/upptime", "url": "https://github.com/aospapp/upptime/issues/452", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1684233940
Bump plugins and build-related deps Bump plugin versions to latest Bump build related deps and plugin deps to latest (errorprone, easymock, checkstyle, slf4j) Ignore license-maven-plugin execution when using Eclipse M2E (broken by https://github.com/eclipse-m2e/m2e-core/issues/1332) Update source with modernizer recommendations to use Optional.orElseThrow instead of Optional.get (this is also recommended in the Optional javadoc) Add comment for why exec-maven-plugin can't be updated (they broke -p again) The updated errorprone found a legitimate bug in the Precondition pattern message added in https://github.com/apache/accumulo/commit/57ed0ec503159e23d608628aa856ca8a3f164465#diff-23a22b999a2787a6c8a0be675465f31b71067f68de3ba5f633c0a73da8780b2fR328-R330 for #1389
gharchive/pull-request
2023-04-26T04:10:58
2025-04-01T04:55:56.835814
{ "authors": [ "ctubbsii" ], "repo": "apache/accumulo", "url": "https://github.com/apache/accumulo/pull/3344", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1009564955
ARTEMIS-3501 Added exception handling on #handleAddMessage to not sto… …p broker from starting with currupted messages I've been unable to write a proper test for this as I cannot simulate a journal broken in the "correct" way. This did however solve the issue I was seeing with an actual broken journal. It will discard and log the corrupted message and proceed with loading the rest of the journal I don't think it will discard.. the message would still be in the journal until the next restart... right? At least in my case I end up in #handleNoMessageReferences (from AbstractJournalStorageManager) and the message is removed by: storageManager.deleteMessage(msg.getMessageID()); And the following log: AMQ221019: Deleting unreferenced message id=$MESSAGEID from the journal
gharchive/pull-request
2021-09-28T10:22:47
2025-04-01T04:55:56.839019
{ "authors": [ "AntonRoskvist", "clebertsuconic" ], "repo": "apache/activemq-artemis", "url": "https://github.com/apache/activemq-artemis/pull/3776", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
178812528
Make start script work on busybox The current version of the activemq start script uses options to ps which are unavailable in busybox. Suggested change refactors the start script to remove duplicated code to find a java process given a pidfile and makes it work on busybox. I have limited possibilities of verifying the change on different platforms but it is confirmed to work on at least Mac OS X in addition to busybox and should work on platforms where previous "ps -p" command worked. Can one of the admins verify this patch?
gharchive/pull-request
2016-09-23T08:13:31
2025-04-01T04:55:56.841107
{ "authors": [ "fjollberg", "lucastetreault" ], "repo": "apache/activemq", "url": "https://github.com/apache/activemq/pull/199", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1339302956
AMBARI-25705. Bump spring to 5.3.22 to address the CVE What changes were proposed in this pull request? Bump Spring to the latest versions How was this patch tested? Verified it was built ok. Existing UT should cover it enough. @jojochuang any chance test after this upgrade..? Build succeed in trunk branch, I'm assuming there are no runtime errors, will be fully tested after we add BIGTOP stack. Thanks @jojochuang Build succeed in trunk branch, I'm assuming there are no runtime errors, will be fully tested after we add BIGTOP stack. It will be good, once it's tested we can merge. when are you planning test against BIGTOP stack ? and once can you please update the test results..? Thank you @kevinw66!! Just to understand better, do you have some docker based script that can bring up ambari server and few agents that we can test locally on dev machine itself? Unless such script already exists somewhere (with mpacks) that I might not be aware of. Otherwise building rpms and deploying to test cluster might be perhaps too much efforts even for small changes. Curious what your thoughts are. Thanks Thank you @kevinw66!! Just to understand better, do you have some docker based script that can bring up ambari server and few agents that we can test locally on dev machine itself? Unless such script already exists somewhere (with mpacks) that I might not be aware of. Otherwise building rpms and deploying to test cluster might be perhaps too much efforts even for small changes. Curious what your thoughts are. Thanks Hi @virajjasani For develop and test Bigtop Mpack Please take a look at this PR: https://github.com/apache/bigtop/pull/940 And set CentOS7 repo to this: https://bigtop-snapshot.s3.amazonaws.com/centos-7/$basearch (which is using BIGTOP 3.2.0 snapshot packages, you can find discussion details in here: https://issues.apache.org/jira/browse/BIGTOP-3745) BTW, MPack it's now using no-prefix version of components For Ambari itself I'll work on it sooner or later, I'm not sure if previous scripts still works, I haven't try it, you can take a look at it if you got time, and give some feedback if possible, thanks!! https://github.com/kevinw66/ambari/blob/trunk/dev-support/docker/README.md Thanks a lot @kevinw66, you have provided many refs, let me go through it. Having dockerized setup would greatly benefit new devs as well as those devs who are not using exact latest versions of branch-2.7 for instance and are still using HDP stack backed deployment in their test/prod clusters. Thanks Thanks a lot @kevinw66, you have provided many refs, let me go through it. Having dockerized setup (specifically with mpack installation) would greatly benefit new devs as well as those devs who are still using HDP stack backed deployment in their test/prod clusters. Thanks Thanks a lot @kevinw66, you have provided many refs, let me go through it. Having dockerized setup (specifically with mpack installation) would greatly benefit new devs as well as those devs who are still using HDP stack backed deployment in their test/prod clusters. Thanks My pleasure, thanks for your contributions on Ambari, ping me anytime if you need help!
gharchive/pull-request
2022-08-15T18:12:17
2025-04-01T04:55:56.958479
{ "authors": [ "brahmareddybattula", "jojochuang", "kevinw66", "virajjasani" ], "repo": "apache/ambari", "url": "https://github.com/apache/ambari/pull/3336", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1485085708
AMBARI-25802: Add group functionality is not working for Mysql8 Verified using mysql-8.0.30. Able to create group from Ambari web UI. @kevinw66 Kindly review. lgtm. @jojochuang can you also check once..? Looks good. @bhavikpatel9977 Why do we have to use quotes with escape. Is it because of keyword group? +1. going to commit, as there is no concerns on this.
gharchive/pull-request
2022-12-08T17:25:24
2025-04-01T04:55:56.960686
{ "authors": [ "bhavikpatel9977", "brahmareddybattula", "mnpoonia" ], "repo": "apache/ambari", "url": "https://github.com/apache/ambari/pull/3600", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
224924908
APEXCORE-711 create a new attribute CUSTOM_SSL_SERVER_CONFIG and use its value to set custom ssl server config @PramodSSImmaneni pls review and merge as appropriate also requesting @devtagare to review and merge ... @PramodSSImmaneni incorporated your feedback. Pls review and if okay I will squash @PramodSSImmaneni if there are no more comments can you merge now? @vrozov are you fine with @sanjaypujare's comments on the JIRA. Also YARN-6457 has been resolved. YARN-4562 is also resolved and both YARN-4562 and YARN-6457 have been merged into 2.7 and later branches by the Hadoop committers based on our requests - I was impressed by the speed. It is time this PR gets merged now. @sanjaypujare can you rebase See contributor guidelines for git setup: Sanjay Pujare sanjaypujare@Sanjay-DT-Mac2.local
gharchive/pull-request
2017-04-27T21:57:59
2025-04-01T04:55:56.963859
{ "authors": [ "PramodSSImmaneni", "sanjaypujare", "tweise" ], "repo": "apache/apex-core", "url": "https://github.com/apache/apex-core/pull/520", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
184781176
APEXMALHAR-2314 Improper functioning in partitioning for sequentialFileRead for FSRecord Fix the StreamCodec for FSRecordReader, initially it was hashcode of blockId's mostly always unique. Hence unable to satisfy the sequentialFileRead property. Now the StreamCodec is modified to work with hashcode of filePath. So all blocks related to a file would be partitioned on same operator. Tested with recordReader and verified for sequentialFileRead that all blocks related to a file are partitioned to single operator. @yogidevendra: Could you please review ? Let us reuse SequentialFileBlockMetadataCodec from FSInputModule instead of defining separate one. @yogidevendra : Thanks for suggestion. Makes sense. Only thing that worried me about creating unnecessary dependency of FSInputModule's StreamCodec on FSRecordReaderModule. So if someone changes FSInput StreamCodec he must consider same in FSRecoredReaderModule. This dependency should be OK. FSRecordReader is closely aligned with FSInputModule. Incorporated Yogi's suggestion. Also please fix the spelling mistake in the JIRA title and commit message. sequencialFileRead =>sequentialFileRead. Check for spellings in the code, javadoc as well. Done. Thanks ! On Wed, Oct 26, 2016 at 11:15 AM, yogidevendra notifications@github.com wrote: Also please fix the spelling mistake in the JIRA title and commit message. sequencialFileRead =>sequentialFileRead. Check for spellings in the code, javadoc as well. — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/apache/apex-malhar/pull/468#issuecomment-256255096, or mute the thread https://github.com/notifications/unsubscribe-auth/ASvTtiq6elsdBijUCNmb4Zeci-sfmslBks5q3uj8gaJpZM4Keh-S . -- Thanks & Regards Deepak Narkhede [ERROR] src/main/java/org/apache/apex/malhar/lib/fs/FSRecordReaderModule.java:[32,8] (imports) UnusedImports: Unused import - com.datatorrent.common.partitioner.StatelessPartitioner. [ERROR] src/main/java/org/apache/apex/malhar/lib/fs/FSRecordReaderModule.java:[33,8] (imports) UnusedImports: Unused import - com.datatorrent.lib.io.block.FSSliceReader. please fix above checkstyle error. Sorry my bad. After rebase it occurred. Will resolve the conflicts. On Nov 7, 2016 3:33 PM, "Tushar R. Gosavi" notifications@github.com wrote: [ERROR] src/main/java/org/apache/apex/malhar/lib/fs/ FSRecordReaderModule.java:32,8 http://imports UnusedImports: Unused import - com.datatorrent.common.partitioner.StatelessPartitioner. [ERROR] src/main/java/org/apache/apex/malhar/lib/fs/ FSRecordReaderModule.java:33,8 http://imports UnusedImports: Unused import - com.datatorrent.lib.io.block.FSSliceReader. please fix above checkstyle error. — You are receiving this because you modified the open/close state. Reply to this email directly, view it on GitHub https://github.com/apache/apex-malhar/pull/468#issuecomment-258793883, or mute the thread https://github.com/notifications/unsubscribe-auth/ASvTtqvJp4rc28gSu76eARinF0EAytHpks5q7vdpgaJpZM4Keh-S .
gharchive/pull-request
2016-10-24T08:43:07
2025-04-01T04:55:56.974442
{ "authors": [ "deepak-narkhede", "tushargosavi", "yogidevendra" ], "repo": "apache/apex-malhar", "url": "https://github.com/apache/apex-malhar/pull/468", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2478179304
Support SHOW command What feature or improvement would you like to see? In postgres, the show command returns server/session configuration attributes: SHOW all The show command doesn't appear to be recognized by abdc: >>> cur.execute("show all") Traceback (most recent call last): File "<stdin>", line 1, in <module> File "C:\Projects\cqn-bur\.venv_cqn\Lib\site-packages\adbc_driver_manager\dbapi.py", line 698, in execute handle, self._rowcount = _blocking_call( ^^^^^^^^^^^^^^^ File "adbc_driver_manager\\_lib.pyx", line 1590, in adbc_driver_manager._lib._blocking_call File "adbc_driver_manager\\_lib.pyx", line 1213, in adbc_driver_manager._lib.AdbcStatement.execute_query File "adbc_driver_manager\\_lib.pyx", line 260, in adbc_driver_manager._lib.check_error adbc_driver_manager.ProgrammingError: INVALID_ARGUMENT: [libpq] Failed to execute query: could not begin COPY: ERROR: syntax error at or near "show" LINE 1: COPY (show all) TO STDOUT (FORMAT binary) Thank you!
gharchive/issue
2024-08-21T14:24:38
2025-04-01T04:55:57.017389
{ "authors": [ "mcrumiller" ], "repo": "apache/arrow-adbc", "url": "https://github.com/apache/arrow-adbc/issues/2093", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2143353626
feat: Support Emit::First for SumDecimalGroupsAccumulator Which issue does this PR close? Closes #46. Rationale for this change What changes are included in this PR? How are these changes tested? CI tests are passed. The failure is due to WARNING: /Users/runner/hostedtoolcache/Java_Adopt_jdk/17.0.10-7/x64/Contents/Home/bin/java is loading libcrypto in an unsafe way @viirya I think you need to rebase this PR Hmm, I think I already rebased on latest main. Hmm not sure. The same issue happened to https://github.com/apache/arrow-datafusion-comet/pull/40 and it is OK now after the rebasing. I re-triggered the pipeline. Hmm, it is actually flaky. Retriggered run is okay. Merged and added details in the PR description. Thanks.
gharchive/pull-request
2024-02-20T01:09:02
2025-04-01T04:55:57.021426
{ "authors": [ "sunchao", "viirya" ], "repo": "apache/arrow-datafusion-comet", "url": "https://github.com/apache/arrow-datafusion-comet/pull/47", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
873703054
Update repository url Which issue does this PR close? Closes #16. Rationale for this change We moved to a new repo and need to update the URL. What changes are included in this PR? Update repo URL in Cargo.toml Are there any user-facing changes? No Codecov Report Merging #233 (57cdc95) into master (c945b03) will not change coverage. The diff coverage is n/a. :exclamation: Current head 57cdc95 differs from pull request most recent head c832aa5. Consider uploading reports for the commit c832aa5 to get more accurate results @@ Coverage Diff @@ ## master #233 +/- ## ======================================= Coverage 76.46% 76.46% ======================================= Files 135 135 Lines 23250 23250 ======================================= Hits 17777 17777 Misses 5473 5473 Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update c945b03...c832aa5. Read the comment docs. There's one more in 'ballista-ui' => https://github.com/apache/arrow-datafusion/blob/master/ballista/ui/scheduler/src/components/Header.tsx#L70 Please check 🙏
gharchive/pull-request
2021-05-01T15:33:15
2025-04-01T04:55:57.029841
{ "authors": [ "andygrove", "codecov-commenter", "djKooks" ], "repo": "apache/arrow-datafusion", "url": "https://github.com/apache/arrow-datafusion/pull/233", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1366344980
Fix the test failure of size_of_scalar failure on M1/M2 Mac Signed-off-by: remzi 13716567376yh@gmail.com Which issue does this PR close? Closes #3383. Rationale for this change What changes are included in this PR? Are there any user-facing changes? Codecov Report Merging #3398 (4090903) into master (81addf7) will decrease coverage by 0.00%. The diff coverage is 100.00%. @@ Coverage Diff @@ ## master #3398 +/- ## ========================================== - Coverage 85.68% 85.68% -0.01% ========================================== Files 298 298 Lines 54654 54653 -1 ========================================== - Hits 46833 46832 -1 Misses 7821 7821 Impacted Files Coverage Δ datafusion/common/src/scalar.rs 85.11% <100.00%> (-0.01%) :arrow_down: datafusion/expr/src/logical_plan/plan.rs 77.02% <0.00%> (ø) :mega: We’re building smart automated test selection to slash your CI/CD build times. Learn more I think we need to merge this before releasing 12.0.0. @alamb @Dandandan any objection to merging this? Go ahead :+1:
gharchive/pull-request
2022-09-08T13:33:38
2025-04-01T04:55:57.038145
{ "authors": [ "Dandandan", "HaoYang670", "andygrove", "codecov-commenter" ], "repo": "apache/arrow-datafusion", "url": "https://github.com/apache/arrow-datafusion/pull/3398", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2359788733
Make ObjectStoreScheme in the object_store crate public Is your feature request related to a problem or challenge? Please describe what you are trying to do. I want to parse an object store URL into a type. Using parse_url and parse_url_opts casts the result into a Box<dyn ...> which makes it impossible to get the underlying type from. Describe the solution you'd like Making ObjectStoreScheme public will allow me to do: let (scheme, path) = ObjectStoreScheme::parse("s3://foo/bar") +1, I want to support all the different cloud providers in my app, but I need more fine-grained control over how the ObjectStore is constructed, so I can't just use parse_url -- having this public would be very useful, I literally just came here to submit a PR to do the same thing :) label_issue.py automatically added labels {'object-store'} from #5912
gharchive/issue
2024-06-18T12:49:36
2025-04-01T04:55:57.040938
{ "authors": [ "alamb", "drmorr0", "orf" ], "repo": "apache/arrow-rs", "url": "https://github.com/apache/arrow-rs/issues/5911", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2494005839
Wrap all tokio::io::AsyncWrite types to a specific type which is implemented AsyncFileWriter In current, parquet defines all AsyncWrite implement AsyncFileWriter: impl<T: AsyncWrite + Unpin + Send> AsyncFileWriter for T { ... ... } which would in conflict with: impl<T: AsyncFileWriter + ?Sized + Unpin + Send> AsyncFileWriter for &mut T { ... ... } this definition is not included right now, however, it should be helpful in more general cases, numerous of similar cases choose to implement this, such as: tokio::io::AsyncWrite futures::io::AsyncWrite Describe the solution you'd like Wrap all tokio::io::AsyncWrite into a proxy struct, then we can define &mut AsyncFileWriter implements AsyncFileWriter: pub struct TokioFileWriter<F: AsyncWrite> { file: F, } impl<F: AsyncWrite> From<F> for TokioFileWriter { ... ... } impl<F: AsyncWrite + Unpin + Send> for TokioFileWriter { ... ... } Describe alternatives you've considered I don't think there is an alternatives. Additional context This is a breaking change, but it values to be considered. Hi, could you share the use case for &mut AsyncFileWriter? But AsyncArrowWriter is created from AsyncFileWriter, creating an AsyncArrowWriter from &mut AsyncFileWriter is exactly what I want. struct CustomFile { // ... } impl AsyncFileWriter for CustomFile { fn write( &mut self, bs: bytes::Bytes, ) -> futures::future::BoxFuture<'_, parquet::errors::Result<()>> { todo!() } fn complete(&mut self) -> futures::future::BoxFuture<'_, parquet::errors::Result<()>> { todo!() } } fn test() { let mut file = CustomFile {}; { let mut writer = AsyncArrowWriter::try_new(&mut file, SchemaRef::from(Schema::empty()), None).unwrap(); // write record batch writer.close() } // do something with file after writes } Also it is helpful to make boxed AsyncFileWriter general, compare with impl AsyncFileWriter for Box<dyn AsyncFileWriter> currently: impl<T: AsyncFileWriter + ?Sized + Unpin + Send> AsyncFileWriter for Box<T> { fn write(&mut self, bs: Bytes) -> BoxFuture<'_, Result<()>> { self.as_mut().write(bs) } fn complete(&mut self) -> BoxFuture<'_, Result<()>> { self.as_mut().complete() } } // do something with file after writes From my current understanding, you should not do anything to this file after closing it. Can you provide a detailed list of actions you plan to take? It is my mistake, actually I just want to flush the AsyncArrowWriter, then seek and read this file as native file rather than a AsyncFileWriter, I have fixed the example code. Understood. I believe you need AsyncArrowWriter::into_inner. First of all, allowing users to manipulate the writer after AsyncArrowWriter has written the header and data could be error-prone. Additionally, it doesn't make sense given our context could involve S3, Azure Blob, or GCS storage services. In these contexts, the file doesn't exist before close has been called, so you can't perform seek/read or other actions. If you simply want to use this API in the local filesystem context, would you consider using try_clone instead? Additionally, it doesn't make sense given our context could involve S3, Azure Blob, or GCS storage services. It does make sense in local filesystems, which allows to implements AsyncFileWriter. First of all, allowing users to manipulate the writer after AsyncArrowWriter has written the header and data could be error-prone However, &mut tokio::io::AsyncWriter is a valid AsyncFileWriter in the current, if you think parquet should not allow &mut AsyncFileWriter be an AsyncFileWriter, then I think &mut tokio::io::AsyncWriter should be fixed. What ever happends, the behavior of tokio::io::AsyncWriter and non-AsyncWriter should be aligned. It does make sense in local filesystems, which allows to be a AsyncFileWriter, don't give up eating for fear of choking. I agree with you that it useful in the local fs context. Therefore, I propose a new API called AsyncArrowWriter::into_inner. I believe this can address your needs in a compatible manner. Understood. I believe you need AsyncArrowWriter::into_inner. I think into_inner has different behaviors and takes different overheads, it would move the AsyncFileWriter but &mut AsyncFileWriter is in-placed. But I think it should be nice to have AsyncArrowWriter::into_inner method.
gharchive/issue
2024-08-29T10:04:28
2025-04-01T04:55:57.052863
{ "authors": [ "Xuanwo", "ethe" ], "repo": "apache/arrow-rs", "url": "https://github.com/apache/arrow-rs/issues/6327", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1631165461
feat: Add Commands enum to decode prost messages to strong type Which issue does this PR close? Closes #3874 Rationale for this change Per the issue, this PR introduces a Commands enum to allow client to match on all supported messages. The following minor improvements were made to the prost_message_ext macro: 1. Generate a constant for the type URLs: const ACTION_CLOSE_PREPARED_STATEMENT_REQUEST_TYPE_URL: &'static str = concat!( "type.googleapis.com/arrow.flight.protocol.sql.", stringify!( ActionClosePreparedStatementRequest ) ); const ACTION_CREATE_PREPARED_STATEMENT_REQUEST_TYPE_URL: &'static str = concat!( "type.googleapis.com/arrow.flight.protocol.sql.", stringify!( ActionCreatePreparedStatementRequest ) ); // Remainder omitted for brevity 2. Commands enum to unpack any of the supported FlightSQL commands: pub enum Commands { ActionClosePreparedStatementRequest(ActionClosePreparedStatementRequest), ActionCreatePreparedStatementRequest(ActionCreatePreparedStatementRequest), // Remainder omitted for brevity } with an associated unpack function: impl Commands { pub fn unpack(any: Any) -> Result<Commands, ArrowError> { match any.type_url.as_str() { ACTION_CLOSE_PREPARED_STATEMENT_REQUEST_TYPE_URL => { let m: ActionClosePreparedStatementRequest = Message::decode(&*any.value).map_err(|err| { ArrowError::ParseError({ let res = ::alloc::fmt::format(format_args!("Unable to decode Any value: {err}", err = err)); res }) })?; Ok(Self::ActionClosePreparedStatementRequest(m)) } ACTION_CREATE_PREPARED_STATEMENT_REQUEST_TYPE_URL => { let m: ActionCreatePreparedStatementRequest = Message::decode(&*any.value).map_err(|err| { ArrowError::ParseError({ let res = ::alloc::fmt::format(format_args!("Unable to decode Any value: {err}", err = err)); res }) })?; Ok(Self::ActionCreatePreparedStatementRequest(m)) } // Remainder omitted for brevity _ => Err(ArrowError::ParseError({ let res = ::alloc::fmt::format(IntellijRustDollarCrate::__export::format_args!("Unable to decode Any value: {}", any.type_url)); res })) } } } 3. Update other generated code to use the shared constant for the type URL: impl ProstMessageExt for ActionClosePreparedStatementRequest { fn type_url() -> &'static str { ACTION_CLOSE_PREPARED_STATEMENT_REQUEST_TYPE_URL } fn as_any(&self) -> Any { Any { type_url: <ActionClosePreparedStatementRequest>::type_url().to_string(), value: self.encode_to_vec().into(), } } } impl ProstMessageExt for ActionCreatePreparedStatementRequest { fn type_url() -> &'static str { ACTION_CREATE_PREPARED_STATEMENT_REQUEST_TYPE_URL } fn as_any(&self) -> Any { Any { type_url: <ActionCreatePreparedStatementRequest>::type_url().to_string(), value: self.encode_to_vec().into(), } } } // Remainder omitted for brevity What changes are included in this PR? Are there any user-facing changes? API improvements for users of the FlightSQL module. I plan to review this shortly Here are some suggested improvements: https://github.com/stuartcarnie/arrow-rs/pull/2# (docs + revamp some of the server.rs code to use the new dispatch logic 👍 ) I pushed a few minor doc updates directly to this branch. @alamb see ea18fa9697d1face7945e40dc6542b59c3890614 for the addition of the Unknown(Any) variant, which was a great suggestion. I've also updated do_get and do_put, and the simple test now passes ✅ Thanks -- I plan to come back to this PR over the weekend but got overloaded with other things this week I am sorry for the delay -- I will get this done / merged this week Github seems to have some issue -- restarting I propose some additional documentation for this feature here https://github.com/apache/arrow-rs/pull/4012
gharchive/pull-request
2023-03-19T23:51:10
2025-04-01T04:55:57.061488
{ "authors": [ "alamb", "stuartcarnie" ], "repo": "apache/arrow-rs", "url": "https://github.com/apache/arrow-rs/pull/3887", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1528268182
[C++][Python] Support start == stop in list_slice kernel [GitHub PR 14395 | https://github.com/apache/arrow/pull/14395] adds the list_slice kernel, but does not implement the case where stop == stop, which should return empty lists. Reporter: Miles Granger / @milesgranger Assignee: Miles Granger / @milesgranger Related issues: [C++] Add kernel for slicing list values (relates to) PRs and other links: GitHub Pull Request #14836 Note: This issue was originally created as ARROW-18281. Please see the migration documentation for further details. take
gharchive/issue
2022-11-08T04:21:41
2025-04-01T04:55:57.066295
{ "authors": [ "asfimport", "felipecrv" ], "repo": "apache/arrow", "url": "https://github.com/apache/arrow/issues/33459", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2342332325
[Java] Update Unit Tests for Compression Module Describe the enhancement requested This is a sub-issue of the issue below. #41680 The unit tests should use the org.junit.jupiter.api package. This issue deals with the compression module. Component(s) Java Issue resolved by pull request 42044 https://github.com/apache/arrow/pull/42044
gharchive/issue
2024-06-09T15:17:42
2025-04-01T04:55:57.068639
{ "authors": [ "lidavidm", "llama90" ], "repo": "apache/arrow", "url": "https://github.com/apache/arrow/issues/42042", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
993727327
ARROW-13976: [C++] Add path to libjvm.so in ARM CPU resolve issue ARROW-13976, add path in hdfs_internal.cc to find libjvm.so in ARM CPU. @cyb70289 Would you know if this looks good? MinGW failures tracked at https://issues.apache.org/jira/browse/ARROW-13999 @pitrou, do you have comments? No :-) Thanks @wuzhuoming !
gharchive/pull-request
2021-09-11T03:04:42
2025-04-01T04:55:57.071345
{ "authors": [ "cyb70289", "pitrou", "wuzhuoming" ], "repo": "apache/arrow", "url": "https://github.com/apache/arrow/pull/11135", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1282511787
ARROW-16918: [Gandiva][C++] Adding UTC-local timezone conversion functions Adding functions to_utc_timezone : Converts a timestamp from local timezone to UTC time from_utc_timezone : Converts a timestamp from UTC time to local time We can fix style by cmake --build ${BUILD_DIR} --target format but you need to install clang-format-12. If it's difficult to prepare clang-format-12, I can fix style and push the fix to this branch.
gharchive/pull-request
2022-06-23T14:35:50
2025-04-01T04:55:57.073344
{ "authors": [ "kou", "palak-9202" ], "repo": "apache/arrow", "url": "https://github.com/apache/arrow/pull/13428", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
342129054
ARROW-2868: [Packaging] Fix Apache Arrow ORC GLib related problems .pc file name was wrong: arrow-glib-orc.pc -> arrow-orc-glib.pc We use arrow-XXX-glib.pc rule such as arrow-gpu-glib.pc. .spec didn't include arrow-orc-glib.pc. .spec included duplicated -DARROW_ORC=ON https://github.com/apache/arrow-dist/commit/6580a6b7f7fd3d2e5f06b68fe606767db72faf9e was backported. Codecov Report Merging #2280 into master will increase coverage by <.01%. The diff coverage is n/a. @@ Coverage Diff @@ ## master #2280 +/- ## ========================================== + Coverage 84.24% 84.25% +<.01% ========================================== Files 290 290 Lines 44297 44297 ========================================== + Hits 37320 37321 +1 + Misses 6946 6945 -1 Partials 31 31 Impacted Files Coverage Δ cpp/src/arrow/util/thread-pool-test.cc 99.45% <0%> (+0.54%) :arrow_up: Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 4ba8769...73cd564. Read the comment docs. Testing it in build: https://travis-ci.org/kszucs/crossbow/builds/405266989 The build passed, successfully uploaded the artifacts too: https://github.com/kszucs/crossbow/releases/tag/build-232-centos-7 Thanks @kou!
gharchive/pull-request
2018-07-17T23:57:36
2025-04-01T04:55:57.083088
{ "authors": [ "codecov-io", "kou", "kszucs" ], "repo": "apache/arrow", "url": "https://github.com/apache/arrow/pull/2280", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1744954796
GH-35926: [C++][Parquet] Testing disable ColumnIndex by disable statistics Rationale for this change Allow only build Offset Index for some column by disable statistics What changes are included in this PR? Fix a tiny bug Are these changes tested? Yes Are there any user-facing changes? Actually no change? Closes: #35926 cc @wgtmac @pitrou @pitrou Mind take a look? @pitrou @wgtmac Do you still have any comments? The remaining CI failures are unrelated, will merge.
gharchive/pull-request
2023-06-07T03:17:16
2025-04-01T04:55:57.087472
{ "authors": [ "mapleFU", "pitrou" ], "repo": "apache/arrow", "url": "https://github.com/apache/arrow/pull/35958", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2392866223
GH-43228: [C++] Fix Abseil compile error on GCC 13 Rationale for this change When trying to compile Arrow with GCC 13, it fails due to ABSEIL missing a <cstdint> include, this PR addresses the issue by adding the missing include. There have been past reports for this issue too: https://github.com/apache/arrow/issues/36969 This is a more minimal fix that tries to avoid the complexity of previous attempts like https://github.com/apache/arrow/pull/43147 and https://github.com/apache/arrow/pull/37066 which involved updating Abseil and facing additional issues to fix. What changes are included in this PR? Add the missing include when GCC>=13 Are these changes tested? They are tested by the existing compile infrastructure and testsuite. We don't seem to have GCC 13 in our testsuite, I could only verify on my own PC that it compiled successfully Are there any user-facing changes? No, all behaviours should remain the same GitHub Issue: #43228 @github-actions crossbow submit -g cpp -g r @raulcd @assignUser this should be ready for review, I tried a more conservative approach as upgrading Abseil introduced more problems than it fixed even on the most conservative version upgrade. Could you open a new issue for this instead of reusing existing closed issue? Could you open a new issue for this instead of reusing existing closed issue? Given it's the same exact error message shouldn't we preserve a consolidated history by keeping the original issue? Happy to open a new issue if that's the preference The problem is that the original issue is in the 13.0.0 milestone and all our current release scripts take that into account to generate release notes, etcetera. If we reuse the same issue it would be as if this PR was released on 13.0.0 and if we require to back port this issue it gets messy. It has it's pros and cons but the current process doesn't account for two PRs pointing to the same GH issue. The problem is that the original issue is in the 13.0.0 milestone Makes sense, I didn't check and thought it was closed without being done. I'll open a new one. @github-actions crossbow submit test-r-gcc-13 @github-actions crossbow submit test-ubuntu-24.04-cpp-gcc-13 @github-actions crossbow submit test-ubuntu-24.04-cpp-gcc-13 @github-actions crossbow submit test-ubuntu-24.04-cpp-gcc-13 @github-actions crossbow submit test-ubuntu-24.04-cpp-gcc-13 @github-actions crossbow submit test-ubuntu-24.04-cpp-gcc-13-bundled @github-actions crossbow submit test-ubuntu-24.04-cpp-gcc-13-bundled @github-actions crossbow submit test-ubuntu-24.04-cpp-gcc-13-bundled @github-actions crossbow submit test-ubuntu-24.04-cpp-gcc-13-bundled Hmm. Could you also add -e ARROW_VERBOSE_THIRDPARTY_BUILD=ON? @github-actions crossbow submit test-ubuntu-24.04-cpp-gcc-13-bundled @github-actions crossbow submit test-ubuntu-24.04-cpp-gcc-13-bundled @github-actions crossbow submit test-ubuntu-24.04-cpp-gcc-13-bundled @kou test-ubuntu-24.04-cpp-gcc-13-bundled now compiles correctly. It fails the test, but it seems unrelated as they are timezone related tests. I think that this is ready for final review. @github-actions crossbow submit test-ubuntu-20.04-cpp-bundled @github-actions crossbow submit test-ubuntu-24.04-cpp-gcc-13-bundled The timestamp tests now pass too by installing tzdata-legacy @github-actions crossbow submit test-r-library-r-base-latest test-r-depsource-bundled @github-actions crossbow submit test-r-depsource-bundled Looks good, thanks! Could you also remove the workaround we added to r/configure? Done and tested on test-r-depsource-bundled @github-actions crossbow submit r-binary-packages Nice, thanks. I'll run this as it has a build on 24.04 which comes with the new gcc. I'll merge after Ugh, turns out ubuntu-latest is still 22.04, I modified the crossbow workflow to re-run on 24.04
gharchive/pull-request
2024-07-05T16:15:00
2025-04-01T04:55:57.100972
{ "authors": [ "amol-", "assignUser", "kou", "raulcd" ], "repo": "apache/arrow", "url": "https://github.com/apache/arrow/pull/43157", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
473160370
ARROW-6031: [Java] Support iterating a vector by ArrowBufPointer Provide the functionality to traverse a vector (fixed-width vector & variable-width vector) by an iterator. This is convenient for scenarios when accessing vector elements in sequence. Codecov Report Merging #4950 into master will decrease coverage by 23.5%. The diff coverage is n/a. @@ Coverage Diff @@ ## master #4950 +/- ## =========================================== - Coverage 88.53% 65.02% -23.51% =========================================== Files 912 488 -424 Lines 116620 64921 -51699 Branches 1418 0 -1418 =========================================== - Hits 103255 42218 -61037 - Misses 13003 22703 +9700 + Partials 362 0 -362 Impacted Files Coverage Δ cpp/src/arrow/util/memory.h 0% <0%> (-100%) :arrow_down: cpp/src/gandiva/date_utils.h 0% <0%> (-100%) :arrow_down: cpp/src/arrow/util/memory.cc 0% <0%> (-100%) :arrow_down: cpp/src/arrow/filesystem/util-internal.cc 0% <0%> (-100%) :arrow_down: cpp/src/arrow/util/sse-util.h 0% <0%> (-100%) :arrow_down: cpp/src/gandiva/decimal_type_util.h 0% <0%> (-100%) :arrow_down: cpp/src/arrow/compute/logical_type.h 0% <0%> (-100%) :arrow_down: cpp/src/parquet/hasher.h 0% <0%> (-100%) :arrow_down: cpp/src/gandiva/basic_decimal_scalar.h 0% <0%> (-100%) :arrow_down: cpp/src/arrow/compute/kernels/boolean.cc 0% <0%> (-100%) :arrow_down: ... and 664 more Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 141a213...29f63d2. Read the comment docs. @jacques-n thanks a lot for your comments. I am convinced that these APIs should not be added to the core APIs. In fact, they are not intended to be the primary way of accessing vector data elements. The iterator can be something nice to have, but not essential. I can think of two applications of this iterator: quickly calculating/combining hash code & determining element equality. for traversing some "vectors" that does not support random access, like compressed vectors. @jacques-n @emkornfield I have removed the methods out of the core APIs. Please check if it looks good. Thank yo so much. @jacques-n any more concerns if the APIs are not directly on Vector classes? I think @jacques-n is out this week, if there are no more comments next week we can probably merge this. I think @jacques-n is out this week, if there are no more comments next week we can probably merge this. @emkornfield Thanks for your comments. I have revised the PR a little, since now we have a common super interface for fixed width vectors and the variable width vectors. @jacques-n 's comments make sense. However, I think this is at least useful for variable width vectors. +1.
gharchive/pull-request
2019-07-26T03:31:32
2025-04-01T04:55:57.118888
{ "authors": [ "codecov-io", "emkornfield", "liyafan82" ], "repo": "apache/arrow", "url": "https://github.com/apache/arrow/pull/4950", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
643422123
ARROW-9210: [C++] Use BitBlockCounter in array/visitor_inline.h This significantly speeds up processing of mostly-not-null or mostly-null data, while having almost no overhead for the other scenarios where you rarely have a word-sized run of all-not-null or all-null-data. Because BitUtil::GetBit is used for bit-checking in the scenario where you need to check every bit in the whole array individually I show slight but inconclusive perf regression similar with the perf difference we've seen comparing BitmapReader with the naive approach calling GetBit inside a loop. This small perf degradation seems to be present mostly with gcc and not meaningfully with clang on Linux. For data with null_count 0, data is processed in blocks of INT16_MAX values at a time, so this adds no meaningful overhead for this case either. I modified the hash benchmarks where this code is used to exhibit both the cases that benefit from this optimization as well as the ones that don't. Here's a benchmark run with gcc-8 --------------------------------------------------------------- Benchmark Time CPU Iterations --------------------------------------------------------------- BuildDictionary 3219443 ns 3219440 ns 218 1.21215GB/s BuildStringDictionary 3692881 ns 3692881 ns 192 81.7532MB/s UniqueInt64/0 14413456 ns 14413251 ns 48 null_percent=0 2.16814GB/s UniqueInt64/1 15516052 ns 15515737 ns 45 null_percent=0.1 2.01408GB/s UniqueInt64/2 17031282 ns 17031266 ns 41 null_percent=1 1.83486GB/s UniqueInt64/3 20680114 ns 20680064 ns 34 null_percent=10 1.51112GB/s UniqueInt64/4 12018069 ns 12017844 ns 57 null_percent=99 2.6003GB/s UniqueInt64/5 9179953 ns 9179946 ns 77 null_percent=100 3.40416GB/s UniqueInt64/6 15501523 ns 15501496 ns 45 null_percent=0 2.01593GB/s UniqueInt64/7 16482935 ns 16482300 ns 41 null_percent=0.1 1.89597GB/s UniqueInt64/8 18349988 ns 18349317 ns 38 null_percent=1 1.70306GB/s UniqueInt64/9 21439268 ns 21439244 ns 32 null_percent=10 1.45761GB/s UniqueInt64/10 12530067 ns 12529871 ns 55 null_percent=99 2.49404GB/s UniqueInt64/11 9167314 ns 9167365 ns 75 null_percent=100 3.40883GB/s UniqueString10bytes/0 43535899 ns 43535846 ns 16 null_percent=0 918.783MB/s UniqueString10bytes/1 45130595 ns 45129634 ns 16 null_percent=0.1 886.336MB/s UniqueString10bytes/2 45249034 ns 45247983 ns 15 null_percent=1 884.017MB/s UniqueString10bytes/3 45101533 ns 45100209 ns 16 null_percent=10 886.914MB/s UniqueString10bytes/4 4316048 ns 4316019 ns 163 null_percent=99 9.05059GB/s UniqueString10bytes/5 1435781 ns 1435763 ns 485 null_percent=100 27.2068GB/s UniqueString10bytes/6 59100344 ns 59098817 ns 12 null_percent=0 676.832MB/s UniqueString10bytes/7 59797544 ns 59795857 ns 12 null_percent=0.1 668.943MB/s UniqueString10bytes/8 61024697 ns 61023090 ns 11 null_percent=1 655.49MB/s UniqueString10bytes/9 59817211 ns 59816339 ns 12 null_percent=10 668.714MB/s UniqueString10bytes/10 4950387 ns 4950242 ns 134 null_percent=99 7.89103GB/s UniqueString10bytes/11 1443482 ns 1443434 ns 446 null_percent=100 27.0622GB/s UniqueString100bytes/0 95609006 ns 95606132 ns 7 null_percent=0 4.08577GB/s UniqueString100bytes/1 96850582 ns 96849441 ns 7 null_percent=0.1 4.03332GB/s UniqueString100bytes/2 95404742 ns 95404634 ns 7 null_percent=1 4.0944GB/s UniqueString100bytes/3 89401775 ns 89401006 ns 8 null_percent=10 4.36936GB/s UniqueString100bytes/4 4705868 ns 4705746 ns 148 null_percent=99 83.0102GB/s UniqueString100bytes/5 1434077 ns 1434055 ns 486 null_percent=100 272.392GB/s UniqueString100bytes/6 206155133 ns 206148425 ns 3 null_percent=0 1.89487GB/s UniqueString100bytes/7 204661287 ns 204653659 ns 3 null_percent=0.1 1.90871GB/s UniqueString100bytes/8 205941884 ns 205941271 ns 3 null_percent=1 1.89678GB/s UniqueString100bytes/9 192074501 ns 192073431 ns 4 null_percent=10 2.03373GB/s UniqueString100bytes/10 6180349 ns 6180227 ns 111 null_percent=99 63.2056GB/s UniqueString100bytes/11 1474565 ns 1474564 ns 482 null_percent=100 264.909GB/s UniqueUInt8/0 1990025 ns 1990023 ns 348 null_percent=0 1.96292GB/s UniqueUInt8/1 2594146 ns 2594089 ns 272 null_percent=0.1 1.50583GB/s UniqueUInt8/2 4726027 ns 4726053 ns 145 null_percent=1 846.372MB/s UniqueUInt8/3 9465222 ns 9465126 ns 75 null_percent=10 422.604MB/s UniqueUInt8/4 3557141 ns 3557135 ns 195 null_percent=99 1124.5MB/s UniqueUInt8/5 2259664 ns 2259664 ns 314 null_percent=100 1.72869GB/s Here is the % diff versus the baseline. Cases 1 and 7 are the mostly-not-null cases. This shows a 15-20% perf improvement Cases 5 and 11 are the all-null cases. Case 4 is the 99% null case The "BuildDictionary" case at the bottom with the perf regression is one of the "worst case scenarios". 89% of the values are null and so we almost never observe an all-null or all-not-null block. The use of BitUtil::GetBit over BitmapReader causes this slightly regression since nearly every validity bit must be checked separately. I don't think it's worth optimizing for this case since the others are more empirically representative of real world data benchmark baseline contender change % regression 8 UniqueString100bytes/5 40.668 GiB/sec 272.392 GiB/sec 569.787 False 37 UniqueString10bytes/5 4.064 GiB/sec 27.207 GiB/sec 569.456 False 33 UniqueString10bytes/11 4.065 GiB/sec 27.062 GiB/sec 565.751 False 12 UniqueString100bytes/11 40.578 GiB/sec 264.909 GiB/sec 552.841 False 0 UniqueString10bytes/4 3.568 GiB/sec 9.051 GiB/sec 153.692 False 36 UniqueString100bytes/4 34.408 GiB/sec 83.010 GiB/sec 141.252 False 19 UniqueString10bytes/10 3.375 GiB/sec 7.891 GiB/sec 133.794 False 24 UniqueUInt8/1 677.981 MiB/sec 1.506 GiB/sec 127.435 False 5 UniqueString100bytes/10 30.775 GiB/sec 63.206 GiB/sec 105.381 False 27 UniqueUInt8/5 1000.163 MiB/sec 1.729 GiB/sec 76.989 False 13 UniqueUInt8/2 650.819 MiB/sec 846.372 MiB/sec 30.047 False 29 UniqueInt64/11 2.703 GiB/sec 3.409 GiB/sec 26.126 False 7 UniqueInt64/5 2.704 GiB/sec 3.404 GiB/sec 25.903 False 18 UniqueUInt8/4 932.926 MiB/sec 1.098 GiB/sec 20.535 False 23 UniqueInt64/1 1.681 GiB/sec 2.014 GiB/sec 19.840 False 21 UniqueInt64/7 1.628 GiB/sec 1.896 GiB/sec 16.476 False 31 UniqueInt64/2 1.658 GiB/sec 1.835 GiB/sec 10.651 False 20 UniqueString10bytes/7 612.647 MiB/sec 668.943 MiB/sec 9.189 False 16 UniqueInt64/3 1.386 GiB/sec 1.511 GiB/sec 9.053 False 38 UniqueString10bytes/8 601.259 MiB/sec 655.490 MiB/sec 9.019 False 1 UniqueUInt8/0 1.808 GiB/sec 1.963 GiB/sec 8.588 False 41 UniqueInt64/9 1.355 GiB/sec 1.458 GiB/sec 7.562 False 14 UniqueString10bytes/1 830.614 MiB/sec 886.336 MiB/sec 6.709 False 4 UniqueInt64/8 1.603 GiB/sec 1.703 GiB/sec 6.260 False 32 UniqueString10bytes/2 847.018 MiB/sec 884.017 MiB/sec 4.368 False 42 UniqueInt64/4 2.508 GiB/sec 2.600 GiB/sec 3.701 False 39 UniqueString10bytes/3 855.985 MiB/sec 886.914 MiB/sec 3.613 False 28 UniqueInt64/10 2.413 GiB/sec 2.494 GiB/sec 3.360 False 34 UniqueString100bytes/3 4.254 GiB/sec 4.369 GiB/sec 2.722 False 11 UniqueString100bytes/2 3.993 GiB/sec 4.094 GiB/sec 2.544 False 9 UniqueString10bytes/9 654.257 MiB/sec 668.714 MiB/sec 2.210 False 35 UniqueString10bytes/6 662.915 MiB/sec 676.832 MiB/sec 2.099 False 6 BuildStringDictionary 80.971 MiB/sec 81.753 MiB/sec 0.966 False 22 UniqueString100bytes/1 4.002 GiB/sec 4.033 GiB/sec 0.783 False 25 UniqueInt64/0 2.153 GiB/sec 2.168 GiB/sec 0.697 False 17 UniqueString10bytes/0 917.726 MiB/sec 918.783 MiB/sec 0.115 False 43 UniqueInt64/6 2.017 GiB/sec 2.016 GiB/sec -0.071 False 40 UniqueString100bytes/0 4.091 GiB/sec 4.086 GiB/sec -0.130 False 3 UniqueString100bytes/7 1.938 GiB/sec 1.909 GiB/sec -1.519 False 26 UniqueString100bytes/8 1.954 GiB/sec 1.897 GiB/sec -2.935 False 2 UniqueString100bytes/9 2.114 GiB/sec 2.034 GiB/sec -3.782 False 30 UniqueString100bytes/6 2.008 GiB/sec 1.895 GiB/sec -5.649 True 10 UniqueUInt8/3 474.468 MiB/sec 422.604 MiB/sec -10.931 True 15 BuildDictionary 1.776 GiB/sec 1.212 GiB/sec -31.742 True Also on the binary size, these changes add about 75KB to libarrow.so. My guess is the difference is mostly coming from code inlining for the all-null case (which wasn't split out before) FWIW on the "gcc/clang perf discussion", clang also shows performance benefits and limited downside benchmark baseline contender change % regression 2 UniqueInt64/11 6.444 GiB/sec 18.511 GiB/sec 187.240 False 31 UniqueInt64/5 6.470 GiB/sec 18.390 GiB/sec 184.244 False 39 UniqueUInt8/5 810.180 MiB/sec 1.747 GiB/sec 120.867 False 26 UniqueUInt8/1 683.475 MiB/sec 1.430 GiB/sec 114.196 False 42 UniqueInt64/4 5.424 GiB/sec 6.965 GiB/sec 28.397 False 18 UniqueInt64/1 2.672 GiB/sec 3.411 GiB/sec 27.627 False 40 UniqueUInt8/2 654.320 MiB/sec 826.916 MiB/sec 26.378 False 33 UniqueUInt8/4 758.115 MiB/sec 947.360 MiB/sec 24.962 False 25 UniqueInt64/10 5.248 GiB/sec 6.426 GiB/sec 22.460 False 9 UniqueString100bytes/5 26.923 GiB/sec 32.142 GiB/sec 19.384 False 35 UniqueString10bytes/11 2.691 GiB/sec 3.207 GiB/sec 19.173 False 3 UniqueString10bytes/5 2.695 GiB/sec 3.200 GiB/sec 18.731 False 20 UniqueString100bytes/11 26.909 GiB/sec 31.831 GiB/sec 18.291 False 30 UniqueInt64/7 2.514 GiB/sec 2.890 GiB/sec 14.960 False 37 UniqueInt64/2 2.619 GiB/sec 2.975 GiB/sec 13.578 False 11 UniqueString10bytes/4 2.487 GiB/sec 2.700 GiB/sec 8.596 False 32 UniqueString10bytes/10 2.386 GiB/sec 2.589 GiB/sec 8.481 False 0 UniqueString100bytes/4 24.419 GiB/sec 26.365 GiB/sec 7.966 False 38 UniqueString100bytes/10 22.463 GiB/sec 24.128 GiB/sec 7.411 False 34 UniqueInt64/8 2.392 GiB/sec 2.563 GiB/sec 7.157 False 19 UniqueString10bytes/1 781.817 MiB/sec 835.760 MiB/sec 6.900 False 43 UniqueInt64/3 2.184 GiB/sec 2.331 GiB/sec 6.721 False 24 UniqueString10bytes/7 583.523 MiB/sec 621.007 MiB/sec 6.424 False 15 UniqueString100bytes/7 1.936 GiB/sec 2.024 GiB/sec 4.538 False 6 UniqueString10bytes/2 780.337 MiB/sec 805.686 MiB/sec 3.248 False 27 UniqueString100bytes/2 3.934 GiB/sec 4.059 GiB/sec 3.197 False 13 UniqueString100bytes/1 3.898 GiB/sec 3.995 GiB/sec 2.485 False 7 UniqueString10bytes/8 592.115 MiB/sec 604.865 MiB/sec 2.153 False 29 UniqueString100bytes/8 1.969 GiB/sec 2.011 GiB/sec 2.111 False 21 UniqueInt64/9 2.034 GiB/sec 2.048 GiB/sec 0.676 False 1 BuildStringDictionary 85.937 MiB/sec 85.928 MiB/sec -0.010 False 41 UniqueUInt8/3 449.171 MiB/sec 448.844 MiB/sec -0.073 False 28 UniqueString100bytes/0 4.084 GiB/sec 4.077 GiB/sec -0.161 False 4 UniqueString100bytes/3 4.255 GiB/sec 4.235 GiB/sec -0.450 False 5 UniqueString100bytes/6 2.054 GiB/sec 2.033 GiB/sec -1.041 False 14 UniqueString100bytes/9 2.138 GiB/sec 2.107 GiB/sec -1.449 False 8 UniqueUInt8/0 1.777 GiB/sec 1.750 GiB/sec -1.487 False 23 UniqueInt64/0 3.860 GiB/sec 3.799 GiB/sec -1.560 False 10 UniqueString10bytes/9 616.458 MiB/sec 605.470 MiB/sec -1.782 False 22 UniqueString10bytes/3 799.494 MiB/sec 783.825 MiB/sec -1.960 False 17 UniqueString10bytes/6 647.921 MiB/sec 631.631 MiB/sec -2.514 False 36 BuildDictionary 1.539 GiB/sec 1.498 GiB/sec -2.694 False 16 UniqueInt64/6 3.193 GiB/sec 3.077 GiB/sec -3.634 False 12 UniqueString10bytes/0 881.975 MiB/sec 839.487 MiB/sec -4.817 False @ursabot benchmark --suite-filter=arrow-compute-vector-sort-benchmark I see big performance drop from some counting sort cases, also tested on my local machine. Should be related to these visitor code: https://github.com/apache/arrow/blob/master/cpp/src/arrow/compute/kernels/vector_sort.cc#L133-L155 Sorting seems too important to leave it to these relatively complex templates, I would suggest implementing the counting sort without using VisitArrayDataInline Also, I don't really understand the use of util::optional in these templates. The user should pass separate lambdas for the not-null and null cases I'm refactoring to nix util::optional. FWIW the performance issue seems to be more pronounced on gcc than clang, here is the benchmark comparison on my machine with clang-8 benchmark baseline contender change % counters 1 SortToIndicesInt64Count/32768/10000/min_time:1.000 1.560 GiB/sec 2.000 GiB/sec 28.163 {'iterations': 70030, 'null_percent': 0.01} 15 SortToIndicesInt64Compare/32768/10000/min_time:1.000 145.735 MiB/sec 158.918 MiB/sec 9.046 {'iterations': 6654, 'null_percent': 0.01} 5 SortToIndicesInt64Compare/32768/100/min_time:1.000 149.117 MiB/sec 159.609 MiB/sec 7.036 {'iterations': 6545, 'null_percent': 1.0} 7 SortToIndicesInt64Compare/32768/0/min_time:1.000 153.027 MiB/sec 162.227 MiB/sec 6.012 {'iterations': 6862, 'null_percent': 0.0} 4 SortToIndicesInt64Compare/32768/10/min_time:1.000 160.419 MiB/sec 167.725 MiB/sec 4.554 {'iterations': 6934, 'null_percent': 10.0} 2 SortToIndicesInt64Compare/32768/2/min_time:1.000 255.024 MiB/sec 260.284 MiB/sec 2.063 {'iterations': 11390, 'null_percent': 50.0} 9 SortToIndicesInt64Count/32768/100/min_time:1.000 1.486 GiB/sec 1.458 GiB/sec -1.912 {'iterations': 66757, 'null_percent': 1.0} 10 SortToIndicesInt64Count/32768/0/min_time:1.000 2.143 GiB/sec 2.067 GiB/sec -3.568 {'iterations': 98191, 'null_percent': 0.0} 13 SortToIndicesInt64Count/8388608/1/min_time:1.000 4.215 GiB/sec 3.813 GiB/sec -9.531 {'iterations': 762, 'null_percent': 100.0} 11 SortToIndicesInt64Count/32768/2/min_time:1.000 679.023 MiB/sec 609.379 MiB/sec -10.256 {'iterations': 29602, 'null_percent': 50.0} 0 SortToIndicesInt64Count/1048576/1/min_time:1.000 4.487 GiB/sec 4.021 GiB/sec -10.400 {'iterations': 6550, 'null_percent': 100.0} 12 SortToIndicesInt64Compare/8388608/1/min_time:1.000 4.250 GiB/sec 3.762 GiB/sec -11.476 {'iterations': 766, 'null_percent': 100.0} 6 SortToIndicesInt64Count/32768/1/min_time:1.000 4.758 GiB/sec 4.185 GiB/sec -12.040 {'iterations': 217705, 'null_percent': 100.0} 8 SortToIndicesInt64Compare/32768/1/min_time:1.000 4.730 GiB/sec 4.125 GiB/sec -12.780 {'iterations': 213908, 'null_percent': 100.0} 3 SortToIndicesInt64Compare/1048576/1/min_time:1.000 4.556 GiB/sec 3.953 GiB/sec -13.228 {'iterations': 6539, 'null_percent': 100.0} 14 SortToIndicesInt64Count/32768/10/min_time:1.000 1.316 GiB/sec 1.051 GiB/sec -20.108 {'iterations': 59539, 'null_percent': 10.0} I'm refactoring to nix util::optional. I'm too tired to finish it tonight so I'll work on it tomorrow morning. If the perf regression isn't gone I'll rewrite the sort kernels. Very likely I'm wrong. I remember util::optional is added due to CI failure https://github.com/apache/arrow/pull/6495#issuecomment-593732821 I think this patch is okay. Sorting regression can be fixed (maybe improved). I'm okay to do the follow up changes. Let's leave sorting optimizations for another PR. I'll review this one. thanks @pitrou and @cyb70289 -- I will spend a little time on the count-sort implementation and post a new patch
gharchive/pull-request
2020-06-22T23:28:49
2025-04-01T04:55:57.131475
{ "authors": [ "cyb70289", "pitrou", "wesm" ], "repo": "apache/arrow", "url": "https://github.com/apache/arrow/pull/7521", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1167338890
AVRO-2211: SchemaBuilder equivalent or other means of schema creation Make sure you have checked all steps below. Jira [x] My PR addresses the following Avro Jira https://issues.apache.org/jira/browse/AVRO-2211 In case you are adding a dependency, check if the license complies with the ASF 3rd Party License Policy. Tests [x] My PR adds the following unit tests OR does not need testing for this extremely good reason: Add a new test case to the Avro.Test.CodeGenTest.TestCodeGen unit test Enable the tests to run under .NET Core by adding compilation support with the Microsoft.CodeAnalysis.CSharp package Commits [x] My commits all reference Jira issues in their subject lines. In addition, my commits follow the guidelines from "How to write a good git commit message": Subject is separated from body by a blank line Subject is limited to 50 characters (not including Jira issue reference) Subject does not end with a period Subject uses the imperative mood ("add", not "adding") Body wraps at 72 characters Body explains "what" and "why", not "how" Documentation [x] In case of new functionality, my PR adds documentation that describes how to use it. All the public functions and the classes in the PR contain Javadoc that explain what it does @yanivru Can you start by updating the description to follow the standards for the pull request? Long awaited feature. Thanks for doing it! I assume https://issues.apache.org/jira/browse/AVRO-3003 is s typo in the PR description. The PR title has the proper tickjet number. @KyleSchoonover Fixed all issues. I changed the constructors to factory methods. If you think it's ok I will leave it. If not I can revert it. @zcsizmadia Added more unit tests @yanivru Plz review the CodeQL warnings or suggestions and I think it is ready to be approved. Thanks for all the changes! @zcsizmadia I think I fixed all CodeQL issues Thank you, @yanivru !
gharchive/pull-request
2022-03-12T16:52:07
2025-04-01T04:55:57.142931
{ "authors": [ "KyleSchoonover", "martin-g", "yanivru", "zcsizmadia" ], "repo": "apache/avro", "url": "https://github.com/apache/avro/pull/1597", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
523077370
[BEAM-8666] Remove dependency between DataflowRunner and PortableRunner introduced by PR#9811 PR#9811 creates a new dependency between DataflowRunner and PortableRunner. DataflowRunner uses DockerEnvironment but there was no need for DataflowRunner to depend on the PortableRunner module before. We will refactor out default_docker_image() method to live in apache_beam.transforms.Environment. JIRA issue: https://issues.apache.org/jira/browse/BEAM-8666 Thank you for your contribution! Follow this checklist to help us incorporate your contribution quickly and easily: [x] Choose reviewer(s) and mention them in a comment (R: @username). [x] Format the pull request title like [BEAM-XXX] Fixes bug in ApproximateQuantiles, where you replace BEAM-XXX with the appropriate JIRA issue, if applicable. This will automatically link the pull request to the issue. [ ] If this contribution is large, please file an Apache Individual Contributor License Agreement. See the Contributor Guide for more tips on how to make review process smoother. Post-Commit Tests Status (on master branch) Lang SDK Apex Dataflow Flink Gearpump Samza Spark Go --- --- --- --- Java Python --- --- --- XLang --- --- --- --- --- --- Pre-Commit Tests Status (on master branch) --- Java Python Go Website Non-portable Portable --- --- --- See .test-infra/jenkins/README for trigger phrase, status and link of all Jenkins jobs. R: @mxm @chamikaramj @chadrik LGTM. Thanks for the fix. Will merge after tests pass.
gharchive/pull-request
2019-11-14T20:07:42
2025-04-01T04:55:57.170033
{ "authors": [ "chamikaramj", "violalyu" ], "repo": "apache/beam", "url": "https://github.com/apache/beam/pull/10112", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
583399208
[BEAM-9468] Hl7v2 io Add HL7v2IO connector for the Google Cloud Healthcare API. Initial work for these connectors is being prioritized based to support various customer use cases for HL7v2 -> FHIR ETL pipelines. Healthcare clients are extremely concerned with not losing data in their pipelines and want dead letter queues for these sources / sinks. With this in mind the HL7v2IO.Read is designed as a mini pipeline that: Starts with an arbitrary PTransform to populate a PCollection of message IDs. This was chosen to give flexibility between reading the HL7v2 pubsub notifications, reading an entire HL7v2 store (with optional filter) or using some other method (e.g. a hand prepared subset of message IDs uploaded to GCS and read w/ TextIO). In the future we can gain efficiency here when the Healthcare API exposes a batch read method where we can batch up requests rather than making a separate call per element. Then a DoFn that attempts to fetch the actual message contents from the HL7v2 store. Successfully fetched messages will be written to HL7v2IO.Read.OUT tag and Failures will be written to a HLv2IO.Read.DEAD_LETTER tag. The "expected" failure case would be attempting to fetch a message ID that doesn't exist. Similarly, the HL7v2IO.Write is designed as a mini pipeline that starts with a PTransform that attempts to ingest a PCollection of Messages then applies the this.getDeadLetterPTransform() to the PCollection of messages that failed to ingest. This gives the customer the ability to capture dead letters in a system of their chosing (e.g. gcs, bq, pubsub) based on their needs. HL7v2: Unbounded Read: Uses PubsubIO to read notification subscription Bounded Read: DoFn to get message IDs using the Messages.List REST API method Future Work will include a very similar IO transform for FHIR store: FhirIO Write use the executeBundle REST API method to execute transactions on the FHIR Store Read with read / search / from notification subscription Thank you for your contribution! Follow this checklist to help us incorporate your contribution quickly and easily: [x] Choose reviewer(s) and mention them in a comment (R: @username). [x] Format the pull request title like [BEAM-XXX] Fixes bug in ApproximateQuantiles, where you replace BEAM-XXX with the appropriate JIRA issue, if applicable. This will automatically link the pull request to the issue. [x] Update CHANGES.md with noteworthy changes. [ ] If this contribution is large, please file an Apache Individual Contributor License Agreement. See the Contributor Guide for more tips on how to make review process smoother. Post-Commit Tests Status (on master branch) Lang SDK Apex Dataflow Flink Gearpump Samza Spark Go --- --- --- --- Java Python --- --- --- XLang --- --- --- --- --- Pre-Commit Tests Status (on master branch) --- Java Python Go Website Non-portable Portable --- --- --- See .test-infra/jenkins/README for trigger phrase, status and link of all Jenkins jobs. @lastomato FYI Moved this to it's own PR. Once we get feedback and iterate on this with beam reviewer, I can open a PR for FHIRIO. R: @pabloem retest this please ("retest this please" is a code phrase for jenkins to run precommits on the PR) since your change is a Java change, you don't need to worry about the Python errors for now. They seem to come from upstream. @lastomato Thoughts on how we could rework the HCAPI client to get around this failing CI Check? https://builds.apache.org/job/beam_PreCommit_Java_Commit/10418/testReport/org.apache.beam.sdk.io.gcp/GcpApiSurfaceTest/testGcpApiSurface/ let me know if this is ready for review or not retest this please retest this please @pabloem sorry for the rage commits today. This is now ready for review. Open Questions: Should we remove adaptive throttling? Seems that we're using retries in the client request initializer and right now a "bad record" will slow down the Read / Write (even though the error has nothing to do with the HL7v2 store being overwhelmed). Originally we wanted to be safe with overwhelming QPS on the HL7v2 store in batch scenarios. Should we add more to the HealthcareIOError? Add (processing time) Timestamp? Add a convenience DoFn HealthcareIOErrrorToTableRowFn to ease writing deadletter queue to BigQuery. Would it me more useful to expose an error rate metric than an error count? retest this please retest this please retest this please retest this please Future improvements: Currently this uses the alpha API because the motivating use case of HL7v2 -> FHIR mapping which requires the schematizedData field (not yet available in beta API). For backwards compatibility the ListHL7v2MessageIDs always uses Messages.List to get message IDs and relies on the the HL7v2IO.Read to fetch the message contents. This provides a consistent flow for real-time and batch and provides backwards compatibility. However, in the alpha API Messages.List returns the actual message contents (rather than just the message IDs). This leads to us reading the messages contents twice for alpha HL7v2 stores. I suggest we address optimizing this "double fetch" in a future PR as the alpha API stabilizes. retest this please Ok an updates here from an internal thread w/ API team. Message.List returning message contents is available in beta API with the view parameter. Schematized Data should be in next beta release roughly in ~2 weeks. right now the sink is outputting schematized data json wrapped in "{data=<actual_valid_json>}" In light of these I will do the following refactors: how we batch read from to always avoid the double get. This will make it a completely parallel code path than the real-time path but I think that's ok. refactor to use beta client library I'll strip out that {data=} wrapper to make this easier for users. retest this please retest this please Sorry for the delay. Please see my comments inline. Open Questions: Should we remove adaptive throttling? I think it is fine to remove it since the API has quota enabled by default, and retry logic is in place. Seems that we're using retries in the client request initializer and right now a "bad record" will slow down the Read / Write (even though the error has nothing to do with the HL7v2 store being overwhelmed). Originally we wanted to be safe with overwhelming QPS on the HL7v2 store in batch scenarios. Should we add more to the HealthcareIOError? The second point below would be very helpful, but I am fine with adding it to next PR (if that's easier). Add (processing time) Timestamp? Add a convenience DoFn HealthcareIOErrrorToTableRowFn to ease writing deadletter queue to BigQuery. Would it be more useful to expose an error rate metric than an error count? This functionality is probably already provided by services like stackdriver, we might want to wait until there is a concrete use case. retest this please I'm starting to take a good look at this now. in case it's not obvious: feel free to ignore failures in unrelated test suites (e.g. communitymetrics) Next Steps (based on offline feed): [x] Improve API for users: [x] Add static methods for common patterns with ListHL7v2Messages [x] Add ValueProvider<String> support to ease use in the DataflowTemplates [x] ListHL7v2Messages (hl7v2Store and filter) [x] Write (hl7v2Store) [ ] "standardize" integration tests [ ] Remove hard coding of my HL7v2Store / project in integration tests. [ ] Add HL7v2 store to Beam GCP IT project @pabloem PTAL. I've addressed all outstanding feedback and setup integration testing infrastructure Note on integration tests: Integration tests seem to be passing. Created the following resources gs://temp-storage-for-healthcare-io-tests will house temp files for FHIR Import in upcoming FhirIO PR projects/apache-beam-testing/locations/us-central1/datasets/apache-beam-integration-testing which houses ephemeral HL7v2 store per test class (and later ephemeral FHIR store per class.) retest this please retest this please @pabloem Questions: How could / should we add this to this page for built in connectors? Are there other documentation things that I might have missed? Should we file jira issues or buganizer for future work, notably: Using Batch import / export for read when that api is available Updating to newer version of client library (coming soon, which will simplify the hack-y handling of schematized data in this PR) What is the feasibility / process for back-porting this to prior beam releases for expediting use by users who want to use an already released version e.g. 2.19 ? Run Java PostCommit Run Java PostCommit Java PostCommit is probably our slowest suite of all, but once it runs we should be good to go 3. What is the feasibility / process for back-porting this to prior beam releases for expediting use by users who want to use an already released version e.g. 2.19 ? bumping this question. would like to know timing on when I can see it in a released version... either backport, or a new one. What is the feasibility / process for back-porting this to prior beam releases for expediting use by users who want to use an already released version e.g. 2.19 ? bumping this question. would like to know timing on when I can see it in a released version... either backport, or a new one. You would need to get support from the community to do the backport and release which seems unlikely based upon this being a new feature (vs a critical bug/security issue) and people are constrained time wise due to COVID-19. Once this is merged it would be part of the subsequent release, the branch cut dates are here: https://calendar.google.com/calendar/embed?src=0p73sl034k80oob7seouanigd0%40group.calendar.google.com (with 2.22.0 scheduled on May 20th) retest this please retest this please @lukecwik I ran the linkeage check "as-is" (log) and IIUC this has no issues. Is this acceptable or is it a hard requirement that I update google_clients_version? @lukecwik I ran the linkeage check "as-is" (log) and IIUC this has no issues. Is this acceptable or is it a hard requirement that I update google_clients_version? This is a hard requirement. Either we are using the google_clients_version of shared deps or google_healthcare_clients_version for deps. We can't be using both. Note that your dependency report seems to be showing that you should be able to update to the version needed by the healthcare lib. @lukecwik that's what I thought to but when I do that I get the output in this comment @lukecwik that's what I thought to but when I do that I get the output in this comment A lot of google libraries have three parts to the version string (API major version, API minor version, shared lib version) You have to find and upgrade the client libraries to versions that contain the 1.30.9 shared lib version like: google-api-services-cloudresourcemanager:v1-rev20191206-1.30.4 becomes google-api-services-cloudresourcemanager:v1-rev20200311-1.30.9 Best way to find matching versions is to look at the index of the maven repo for each artifact (e.g.): https://repo1.maven.org/maven2/com/google/apis/google-api-services-cloudresourcemanager/ Find ones that have the same API major version and any newer API minor version with the shard lib version being 1.30.9. Major version changes are breaking API changes, minor version increases are meant to be safe to upgrade and shared lib version just needs to be the same across all versions. Thanks for the pointer! I will give that a try. @lukecwik some progress I was able to update storage / cloudresourcemanager that way. However, in https://repo1.maven.org/maven2/com/google/oauth-client/google-oauth-client/ https://repo1.maven.org/maven2/com/google/oauth-client/google-oauth-client-java6/ They stop at 1.30.6 @lukecwik would it be a reasonable compromise to introduce a google_oauth_clients_verion? If not I believe we are at an impass. Reasoning / Research / Background: features in healthcare API making this PR possible require >=1.30.9 (healthcare api absolute minimum version is 1.30.8) all deps under com.google.apis are updatable to 1.30.9 the two deps under com.google.oauth-client have versions that stop at 1.30.6 I have a suspicion that this 1.30.9 for com.google.apis is not even really related to the 1.30.6 for com.google.oauth-clients but an unfortunate naming convention similarity see that 1.30.6 oauth client updated fairly recently on 2020-03-05 I've prepared this in the latest commit to make this a concrete proposal to review / run pre/post commits on. retest this please retest this please retest this please now linkage checker is giving me no output: jferriero@shadow-gallery:~/VersionControl/beam$ sdks/java/build-tools/beam-linkage-check.sh Fri 10 Apr 2020 04:42:32 PM PDT: Installing artifacts of HL7v2IO(785a5937) to Maven local repository. jferriero@shadow-gallery:~/VersionControl/beam$ Run Java PostCommit Run Java PreCommit Run Java PreCommit Run Java PreCommit Run Java PostCommit Run Java PostCommit retest this please Run Java PostCommit Run Java PostCommit retest this please retest this please Run Java PostCommit oops. looks like a merge issue 13:11:33 /home/jenkins/jenkins-slave/workspace/beam_PreCommit_Java_Commit/src/buildSrc/src/main/groovy/org/apache/beam/gradle/BeamModulePlugin.groovy: 386: The current scope already contains a variable of the name google_oauth_clients_version 13:11:33 @ line 386, column 9. 13:11:33 def google_oauth_clients_version = "1.30.6" 13:11:33 ^ 13:11:33 retest this please retest this please Run Java PostCommit Run Java PostCommit this looks fine to me as long as the dependency changes look fine to @lukecwik @lukecwik ptal so @jaketf won't have to rebase if changes look fine @lukecwik please let me know if there's anything else I can do to give confidence in the dependency changes. @lukecwik please let me know if there's anything else I can do to give confidence in the dependency changes. I have a lot of reviews on my radar and will get to this one as I can. If you want, you can reach out to the community for another reviewer to double check the dependency work and/or have Pablo take it up. @suztomo - Tomo are you able to review the dependency changes for this PR? @suztomo would you be able take a look at the dependency changes? otherwise I will drop a note to the dev list. retest this please retest this please retest this please retest this please retest this please retest this please retest this please Run Java PostCommit retest this please retest this please Run Java PreCommit Run Java PostCommit Yeah, this can be a performance bottleneck and this whole operation will be limited to a single machine. Usually sources need an additional level of parallelism due to being high fanout. BTW it might sense to add a Reshuffle at the end here just to allow any subsequent steps to parallelize. @chamikaramj that sounds like a good idea. However, I get deprecation working on org.apache.beam.sdk.transforms.Reshuffle is there a new blessed way of doing this? You can ignore the deprecation warning for Reshuffle.viaRandomKey(). The deprecation was just because the behavior across runners for Reshuffle is not well defined. Transform is not going away. Many runners add a fusion break after GBK. So this will allow subsequent steps to parallelize better. You can also consider adding an option to not add the Reshuffle to avoid adding any additional shuffle for anyone who already will have a GBK downstream. See here: https://github.com/apache/beam/blob/master/sdks/java/io/jdbc/src/main/java/org/apache/beam/sdk/io/jdbc/JdbcIO.java#L823 @pabloem please do not retest this until I say so. reshuffle is messing up something to do w/ coders in my ITs. will investigate and let you know when it's safe to re run tests @pabloem reshuffle added and ITs passing locally as of c50df5f retest this please retest this please Run Java PostCommit Run Java PostCommit And liftoff Seems like one of the new tests is flaky. https://builds.apache.org/job/beam_PostCommit_Java/5947/ https://builds.apache.org/job/beam_PostCommit_Java/5943/ https://builds.apache.org/job/beam_PostCommit_Java/5942/ @chamikaramj looking into this now. Struggling to reproduce locally.
gharchive/pull-request
2020-03-18T01:28:29
2025-04-01T04:55:57.251892
{ "authors": [ "brianlucier", "chamikaramj", "jaketf", "lastomato", "lukecwik", "pabloem" ], "repo": "apache/beam", "url": "https://github.com/apache/beam/pull/11151", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
885501904
[BEAM-12246] Fix ib.collect(dataframe) indexing Indexing wasn't being properly propagated. This also preserves any hierarchical indexing from the dataframes. There is one problem with passing dataframes, however, which is that the indexing isn't kept and is reset per bundle. This solves this by resetting the index once the DataFrame is collected but only if it is an unnamed index. This behavior can be overridden with the reset_unnamed_indexes=False flag. Thank you for your contribution! Follow this checklist to help us incorporate your contribution quickly and easily: [ ] Choose reviewer(s) and mention them in a comment (R: @username). [ ] Format the pull request title like [BEAM-XXX] Fixes bug in ApproximateQuantiles, where you replace BEAM-XXX with the appropriate JIRA issue, if applicable. This will automatically link the pull request to the issue. [ ] Update CHANGES.md with noteworthy changes. [ ] If this contribution is large, please file an Apache Individual Contributor License Agreement. See the Contributor Guide for more tips on how to make review process smoother. ValidatesRunner compliance status (on master branch) Lang ULR Dataflow Flink Samza Spark Twister2 Go --- --- --- Java Python --- --- --- XLang --- --- Examples testing status on various runners Lang ULR Dataflow Flink Samza Spark Twister2 Go --- --- --- --- --- --- --- Java --- --- --- --- --- --- Python --- --- --- --- --- --- --- XLang --- --- --- --- --- --- --- Post-Commit SDK/Transform Integration Tests Status (on master branch) Go Java Python Pre-Commit Tests Status (on master branch) --- Java Python Go Website Whitespace Typescript Non-portable Portable --- --- --- --- See .test-infra/jenkins/README for trigger phrase, status and link of all Jenkins jobs. GitHub Actions Tests Status (on master branch) See CI.md for more information about GitHub Actions CI. Run Portable_Python PreCommit Added TODOs, thanks Brian!
gharchive/pull-request
2021-05-11T00:51:18
2025-04-01T04:55:57.278304
{ "authors": [ "rohdesamuel" ], "repo": "apache/beam", "url": "https://github.com/apache/beam/pull/14778", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
988161511
[BEAM-9482] Disable Kafka perf tests. Disable a perma-red suite https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO. Thank you for your contribution! Follow this checklist to help us incorporate your contribution quickly and easily: [ ] Choose reviewer(s) and mention them in a comment (R: @username). [ ] Format the pull request title like [BEAM-XXX] Fixes bug in ApproximateQuantiles, where you replace BEAM-XXX with the appropriate JIRA issue, if applicable. This will automatically link the pull request to the issue. [ ] Update CHANGES.md with noteworthy changes. [ ] If this contribution is large, please file an Apache Individual Contributor License Agreement. See the Contributor Guide for more tips on how to make review process smoother. ValidatesRunner compliance status (on master branch) Lang ULR Dataflow Flink Samza Spark Twister2 Go --- --- Java Python --- --- XLang --- Examples testing status on various runners Lang ULR Dataflow Flink Samza Spark Twister2 Go --- --- --- --- --- --- --- Java --- --- --- --- --- --- Python --- --- --- --- --- --- --- XLang --- --- --- --- --- --- --- Post-Commit SDK/Transform Integration Tests Status (on master branch) Go Java Python Pre-Commit Tests Status (on master branch) --- Java Python Go Website Whitespace Typescript Non-portable Portable --- --- --- --- See .test-infra/jenkins/README for trigger phrase, status and link of all Jenkins jobs. GitHub Actions Tests Status (on master branch) See CI.md for more information about GitHub Actions CI. R: @chamikaramj or @ibzib Run Seed Job Run Seed Job
gharchive/pull-request
2021-09-03T23:52:12
2025-04-01T04:55:57.304614
{ "authors": [ "tvalentyn" ], "repo": "apache/beam", "url": "https://github.com/apache/beam/pull/15459", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1430664207
[DO NOT MERGE] Test failures Please add a meaningful description for your change here Thank you for your contribution! Follow this checklist to help us incorporate your contribution quickly and easily: [ ] Choose reviewer(s) and mention them in a comment (R: @username). [ ] Mention the appropriate issue in your description (for example: addresses #123), if applicable. This will automatically add a link to the pull request in the issue. If you would like the issue to automatically close on merging the pull request, comment fixes #<ISSUE NUMBER> instead. [ ] Update CHANGES.md with noteworthy changes. [ ] If this contribution is large, please file an Apache Individual Contributor License Agreement. See the Contributor Guide for more tips on how to make review process smoother. To check the build health, please visit https://github.com/apache/beam/blob/master/.test-infra/BUILD_STATUS.md GitHub Actions Tests Status (on master branch) See CI.md for more information about GitHub Actions CI. Run Seed Job Run Jpms Dataflow Java 11 PostCommit Run Jpms Dataflow Java 17 PostCommit
gharchive/pull-request
2022-11-01T00:22:09
2025-04-01T04:55:57.312428
{ "authors": [ "kileys" ], "repo": "apache/beam", "url": "https://github.com/apache/beam/pull/23912", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1529567861
DONT MERGE Please add a meaningful description for your change here Thank you for your contribution! Follow this checklist to help us incorporate your contribution quickly and easily: [ ] Mention the appropriate issue in your description (for example: addresses #123), if applicable. This will automatically add a link to the pull request in the issue. If you would like the issue to automatically close on merging the pull request, comment fixes #<ISSUE NUMBER> instead. [ ] Update CHANGES.md with noteworthy changes. [ ] If this contribution is large, please file an Apache Individual Contributor License Agreement. See the Contributor Guide for more tips on how to make review process smoother. To check the build health, please visit https://github.com/apache/beam/blob/master/.test-infra/BUILD_STATUS.md GitHub Actions Tests Status (on master branch) See CI.md for more information about GitHub Actions CI. Run Inference Benchmarks
gharchive/pull-request
2023-01-11T19:29:51
2025-04-01T04:55:57.319460
{ "authors": [ "AnandInguva" ], "repo": "apache/beam", "url": "https://github.com/apache/beam/pull/24978", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1930564290
Add support to use side inputs with Combine.PerKeyWithHotKeyFanout Addresses #20637 This PR fixes the bug that you cannot use side inputs with Combine.PerKeyWithHotKeyFanout by passing down the side inputs from Combine.PerKey Thank you for your contribution! Follow this checklist to help us incorporate your contribution quickly and easily: [x] Mention the appropriate issue in your description (for example: addresses #123), if applicable. This will automatically add a link to the pull request in the issue. If you would like the issue to automatically close on merging the pull request, comment fixes #<ISSUE NUMBER> instead. [ ] Update CHANGES.md with noteworthy changes. [ ] If this contribution is large, please file an Apache Individual Contributor License Agreement. See the Contributor Guide for more tips on how to make review process smoother. To check the build health, please visit https://github.com/apache/beam/blob/master/.test-infra/BUILD_STATUS.md GitHub Actions Tests Status (on master branch) See CI.md for more information about GitHub Actions CI or the workflows README to see a list of phrases to trigger workflows. assign set of reviewers Run Java PreCommit @damondouglas what are next steps here? @damccorm How do we rerun the two failing checks? Run Java PreCommit @marc7806 Thank you again for submitting this PR. I'm running ./gradlew :runners:google-cloud-dataflow-java:validatesRunner to check this PR as the regular ./gradlew :sdks:java:core:check does not trigger relevant tests.
gharchive/pull-request
2023-10-06T16:26:33
2025-04-01T04:55:57.327884
{ "authors": [ "damccorm", "damondouglas", "marc7806" ], "repo": "apache/beam", "url": "https://github.com/apache/beam/pull/28867", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
281996035
Fix javadoc in FnDataReceiver Follow this checklist to help us incorporate your contribution quickly and easily: [ ] Make sure there is a JIRA issue filed for the change (usually before you start working on it). Trivial changes like typos do not require a JIRA issue. Your pull request should address just this issue, without pulling in other changes. [ ] Each commit in the pull request should have a meaningful subject line and body. [ ] Format the pull request title like [BEAM-XXX] Fixes bug in ApproximateQuantiles, where you replace BEAM-XXX with the appropriate JIRA issue. [ ] Write a pull request description that is detailed enough to understand what the pull request does, how, and why. [ ] Run mvn clean verify to make sure basic checks pass. A more thorough check will be performed on your pull request automatically. [ ] If this contribution is large, please file an Apache Individual Contributor License Agreement. retest this please This LGTM.
gharchive/pull-request
2017-12-14T06:35:33
2025-04-01T04:55:57.331361
{ "authors": [ "jbonofre", "lukecwik" ], "repo": "apache/beam", "url": "https://github.com/apache/beam/pull/4259", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
426564454
[BEAM-4610] Add SSL support for RedisIO Hi all, Here is the PR for SSL connection support to RedisIO. The idea is to just add a SSL flag to allow the Redis client to open a SSL connection with its default SSL parameters. Thanks in advance for the review. (R: @iemejia ) Regards, Guobao Post-Commit Tests Status (on master branch) Lang SDK Apex Dataflow Flink Gearpump Samza Spark Go --- --- --- --- --- --- Java Python --- --- --- --- Pre-Commit Tests Status (on master branch) --- Java Python Go Website Non-portable Portable --- --- --- See .test-infra/jenkins/README for trigger phrase, status and link of all Jenkins jobs. Thanks @iemejia for the code review! FYI, the modification has been done and it is ready to get merged. Run Java PreCommit Run Java PreCommit
gharchive/pull-request
2019-03-28T15:39:44
2025-04-01T04:55:57.350976
{ "authors": [ "EdgarLGB", "iemejia" ], "repo": "apache/beam", "url": "https://github.com/apache/beam/pull/8161", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
777164744
Check whether progress reporting could be incorrect in certain cases (technical debt) See original issue on GitLab In GitLab by [Gitlab user @jonathanmaw] on Jul 26, 2019, 14:31 Background !1482 adds a way to report progress when loading elements, but there are points where an element might be loaded or resolved that happen outside the scope of a SimpleTask. Tests have been added to check that the tallies are correct when loading across a junction in a simple case, but the full spread of possibilities where elements might be loaded or resolved has not been investigated. Task description [x] Trace through loader code to see which points Loader._collect_element are called to see which ones will and won't have a Task passed in. [x] Assess whether a Task should be passed in for those cases. [x] Write tests to cover any edge cases that are not already covered Acceptance Criteria In GitLab by [Gitlab user @LaurenceUrhegyi] on Jul 30, 2019, 16:03 mentioned in merge request !1358 In GitLab by [Gitlab user @LaurenceUrhegyi] on Jul 30, 2019, 16:03 mentioned in merge request !1358 In GitLab by [Gitlab user @tlater] on Sep 10, 2019, 11:31 assigned to [Gitlab user @tlater] In GitLab by [Gitlab user @tlater] on Sep 10, 2019, 11:31 assigned to [Gitlab user @tlater] In GitLab by [Gitlab user @tlater] on Sep 13, 2019, 17:42 This seems to only occur here: https://gitlab.com/BuildStream/buildstream/blob/master/src/buildstream/_loader/loader.py#L575 Which, given that this only loads a junction element and creates a Loader from it, which in turn is always given a Task when it loads elements, should be fine. In GitLab by [Gitlab user @tlater] on Sep 13, 2019, 17:42 This seems to only occur here: https://gitlab.com/BuildStream/buildstream/blob/master/src/buildstream/_loader/loader.py#L575 Which, given that this only loads a junction element and creates a Loader from it, which in turn is always given a Task when it loads elements, should be fine. In GitLab by [Gitlab user @tlater] on Sep 18, 2019, 18:33 mentioned in merge request !1608 In GitLab by [Gitlab user @tlater] on Sep 18, 2019, 18:33 mentioned in merge request !1608 In GitLab by [Gitlab user @tlater] on Oct 8, 2019, 14:26 marked the task Write tests to cover any edge cases that are not already covered as completed In GitLab by [Gitlab user @tlater] on Oct 8, 2019, 14:26 marked the task Write tests to cover any edge cases that are not already covered as completed In GitLab by [Gitlab user @tlater] on Oct 8, 2019, 14:26 marked the task Write tests to cover any edge cases that are not already covered as incomplete In GitLab by [Gitlab user @tlater] on Oct 8, 2019, 14:26 marked the task Write tests to cover any edge cases that are not already covered as incomplete In GitLab by [Gitlab user @tlater] on Oct 8, 2019, 14:26 marked the task Trace through loader code to see which points Loader._collect_element are called to see which ones will and won't have a Task passed in. as completed In GitLab by [Gitlab user @tlater] on Oct 8, 2019, 14:26 marked the task Trace through loader code to see which points Loader._collect_element are called to see which ones will and won't have a Task passed in. as completed In GitLab by [Gitlab user @tlater] on Oct 8, 2019, 14:26 marked the task Assess whether a Task should be passed in for those cases. as completed In GitLab by [Gitlab user @tlater] on Oct 8, 2019, 14:26 marked the task Assess whether a Task should be passed in for those cases. as completed In GitLab by [Gitlab user @tlater] on Oct 8, 2019, 14:27 marked the task Write tests to cover any edge cases that are not already covered as completed In GitLab by [Gitlab user @tlater] on Oct 8, 2019, 14:27 marked the task Write tests to cover any edge cases that are not already covered as completed In GitLab by [Gitlab user @marge-bot123] on Oct 10, 2019, 13:45 closed via merge request !1608 In GitLab by [Gitlab user @marge-bot123] on Oct 10, 2019, 13:45 closed via merge request !1608 In GitLab by [Gitlab user @marge-bot123] on Oct 10, 2019, 13:45 mentioned in commit 6c6c581162ed7d87bad680330610e29d828d0f25 In GitLab by [Gitlab user @marge-bot123] on Oct 10, 2019, 13:45 mentioned in commit 6c6c581162ed7d87bad680330610e29d828d0f25
gharchive/issue
2021-01-01T00:20:59
2025-04-01T04:55:57.403492
{ "authors": [ "BuildStream-Migration-Bot" ], "repo": "apache/buildstream", "url": "https://github.com/apache/buildstream/issues/1093", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1708674579
"Too many open files" when running with remote execution When building with remote execution, I get the following error. I don't know if it's consistent, but retrying once gives the same error. [16:35:19][00:11:56][e8a2758a][ build:bootstrap/build/gcc-stage1.bst] BUG Build An unhandled exception occured: Traceback (most recent call last): File "/home/abderrahimkitouni/.local/pipx/venvs/buildstream/lib/python3.11/site-packages/buildstream/_scheduler/jobs/job.py", line 438, in child_action result = self.child_process() # pylint: disable=assignment-from-no-return ^^^^^^^^^^^^^^^^^^^^ File "/home/abderrahimkitouni/.local/pipx/venvs/buildstream/lib/python3.11/site-packages/buildstream/_scheduler/jobs/elementjob.py", line 91, in child_process return self._action_cb(self._element) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/home/abderrahimkitouni/.local/pipx/venvs/buildstream/lib/python3.11/site-packages/buildstream/_scheduler/queues/buildqueue.py", line 55, in _assemble_element element._assemble() File "/home/abderrahimkitouni/.local/pipx/venvs/buildstream/lib/python3.11/site-packages/buildstream/element.py", line 1689, in _assemble collect = self.assemble(sandbox) # pylint: disable=assignment-from-no-return ^^^^^^^^^^^^^^^^^^^^^^ File "/home/abderrahimkitouni/.local/pipx/venvs/buildstream/lib/python3.11/site-packages/buildstream/buildelement.py", line 315, in assemble with sandbox.batch(root_read_only=True, label="Running commands"): File "/usr/lib/python3.11/contextlib.py", line 144, in __exit__ next(self.gen) File "/home/abderrahimkitouni/.local/pipx/venvs/buildstream/lib/python3.11/site-packages/buildstream/sandbox/sandbox.py", line 265, in batch batch.execute() File "/home/abderrahimkitouni/.local/pipx/venvs/buildstream/lib/python3.11/site-packages/buildstream/sandbox/_sandboxreapi.py", line 234, in execute self.sandbox._run_with_flags( File "/home/abderrahimkitouni/.local/pipx/venvs/buildstream/lib/python3.11/site-packages/buildstream/sandbox/sandbox.py", line 374, in _run_with_flags return self._run(command, flags=flags, cwd=cwd, env=env) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/home/abderrahimkitouni/.local/pipx/venvs/buildstream/lib/python3.11/site-packages/buildstream/sandbox/_sandboxreapi.py", line 101, in _run action_result = self._execute_action(action, flags) # pylint: disable=assignment-from-no-return ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/home/abderrahimkitouni/.local/pipx/venvs/buildstream/lib/python3.11/site-packages/buildstream/sandbox/_sandboxremote.py", line 221, in _execute_action cascache.pull_tree(casremote, tree_digest) File "/home/abderrahimkitouni/.local/pipx/venvs/buildstream/lib/python3.11/site-packages/buildstream/_cas/cascache.py", line 278, in pull_tree digest = self._fetch_tree(remote, digest) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/home/abderrahimkitouni/.local/pipx/venvs/buildstream/lib/python3.11/site-packages/buildstream/_cas/cascache.py", line 644, in _fetch_tree dirdigests = self.add_objects(buffers=dirbuffers) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/home/abderrahimkitouni/.local/pipx/venvs/buildstream/lib/python3.11/site-packages/buildstream/_cas/cascache.py", line 370, in add_objects tmp = stack.enter_context(self._temporary_object()) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/lib/python3.11/contextlib.py", line 505, in enter_context result = _enter(cm) ^^^^^^^^^^ File "/usr/lib/python3.11/contextlib.py", line 137, in __enter__ return next(self.gen) ^^^^^^^^^^^^^^ File "/home/abderrahimkitouni/.local/pipx/venvs/buildstream/lib/python3.11/site-packages/buildstream/_cas/cascache.py", line 595, in _temporary_object with utils._tempnamedfile(dir=self.tmpdir) as f: File "/usr/lib/python3.11/contextlib.py", line 137, in __enter__ return next(self.gen) ^^^^^^^^^^^^^^ File "/home/abderrahimkitouni/.local/pipx/venvs/buildstream/lib/python3.11/site-packages/buildstream/utils.py", line 1268, in _tempnamedfile with _signals.terminator(close_tempfile), tempfile.NamedTemporaryFile( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/lib/python3.11/tempfile.py", line 702, in NamedTemporaryFile file = _io.open(dir, mode, buffering=buffering, ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/lib/python3.11/tempfile.py", line 699, in opener fd, name = _mkstemp_inner(dir, prefix, suffix, flags, output_type) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/lib/python3.11/tempfile.py", line 395, in _mkstemp_inner fd = _os.open(file, flags, 0o600) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ OSError: [Errno 24] Too many open files: '/home/abderrahimkitouni/.cache/buildstream/tmp/tmpr9a6fdwb' @abderrahim did you ever figure out what this was about? Is there some fd leak? Nope, I just increased the limit with ulimit and got my build going.
gharchive/issue
2023-05-13T16:24:47
2025-04-01T04:55:57.406792
{ "authors": [ "abderrahim", "nanonyme" ], "repo": "apache/buildstream", "url": "https://github.com/apache/buildstream/issues/1842", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
799535968
Support for incremental builds for workspaces See original issue on GitLab In GitLab by [Gitlab user @tristanvb] on Jan 16, 2018, 08:09 Currently BuildStream never caches any build trees after a build, as noted in #21. However, when a workspace is active we have the opportunity to reuse the same workspace directory to build inside of - and workspace builds are where incremental builds will benefit the developer a lot by reducing the edit/compile/test cycles. This is partly discussed in this mail thread in October 2017, since then a lot of work and discussion has taken place in MR !126. In GitLab by [Gitlab user @tristanvb] on Jan 16, 2018, 08:16 mentioned in merge request !126 In GitLab by [Gitlab user @cs-shadow] on Jan 25, 2018, 11:58 mentioned in commit 5f1be604603843bc266252cf5a3a91939d2c09f5 In GitLab by [Gitlab user @cs-shadow] on Jan 25, 2018, 13:16 closed via commit 5f1be604603843bc266252cf5a3a91939d2c09f5
gharchive/issue
2021-02-02T18:06:41
2025-04-01T04:55:57.411239
{ "authors": [ "BuildStream-Migration-Bot" ], "repo": "apache/buildstream", "url": "https://github.com/apache/buildstream/issues/192", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
801005603
WIP: bst fmt: Add basic functionality See original merge request on GitLab In GitLab by [Gitlab user @coldtom] on Nov 15, 2018, 16:55 Description Adds a bst fmt command, which modifies the format of a projects element files into a canonical yaml format. This has the main benefit of automating files into the format bst track will dump them into when run. Currently the functionality is somewhat stinted by #767, and there are a couple of other features I would like to add to this - namely an option which doesn't modify the yaml, but creates a diff, and a way for a user to define the order of top-level nodes in a .bst file, into which bst fmt sorts the nodes. Changes proposed in this merge request: Add a bst fmt command which formats elements into a canonical format This merge request, when approved, will close #485 In GitLab by [Gitlab user @jjardon] on Nov 15, 2018, 17:14 changed the description In GitLab by [Gitlab user @coldtom] on Nov 15, 2018, 17:59 added 1 commit 1d2d0de4 - bst fmt: Add basic functionality Compare with previous version In GitLab by [Gitlab user @knownexus] on Nov 19, 2018, 14:21 Commented on buildstream/_frontend/cli.py line 374 Shouldn't there be a requirement for --all to be enabled before we allow --except? I suppose it doesn't break anything if you do it without, but it seems to have no value in other cases from what i can see In GitLab by [Gitlab user @knownexus] on Nov 19, 2018, 14:25 Commented on buildstream/_scheduler/queues/formatqueue.py line 12 Is this called anywhere? I can't see it but i may be missing it In GitLab by [Gitlab user @coldtom] on Nov 21, 2018, 12:56 Commented on buildstream/_scheduler/queues/formatqueue.py line 12 As I understand it this is where one implements the process performed by the queue, and is called by the scheduler. It's an abstract method in the Queue class. However I was wrong to give a return value for the feature implemented here, I'll fix that presently. In GitLab by [Gitlab user @coldtom] on Nov 21, 2018, 13:47 Commented on buildstream/_frontend/cli.py line 374 Other elements with similar functionality use the same method of implementing --except, personally I'm of the opinion that adding a guard against it would make the code more awkward for little benefit, but I'm not opposed to it. In GitLab by [Gitlab user @tlater] on Nov 21, 2018, 16:16 Commented on buildstream/_scheduler/queues/formatqueue.py line 12 [Gitlab user @knownexus] it's part of the Queue API here. It's what keeps us from creating new job types for every queue type. In GitLab by [Gitlab user @tlater] on Nov 21, 2018, 16:19 Commented on buildstream/_stream.py line 297 While I understand that you're just reusing code here, I'd prefer if you changed the name of the kwarg to do so. Perhaps modify_selection or somesuch? In GitLab by [Gitlab user @tlater] on Nov 21, 2018, 16:26 Commented on buildstream/_frontend/cli.py line 374 I think that the help string should at least mention that it has no effect unelss --all is specified. Otherwise this is probably fine. In GitLab by [Gitlab user @tlater] on Nov 21, 2018, 16:27 Commented on buildstream/_scheduler/queues/formatqueue.py line 15 I'd be pretty impressed if you could do this. Any suggestions on how this would be implemented? Surely you'd run into the halting problem here. In GitLab by [Gitlab user @tlater] on Nov 21, 2018, 16:38 So, the ML was pretty happy with this, but as much as I see it being useful, I worry about feature creep. We have been adding a lot of new commands to BuildStream in this cycle, but they have mostly been there to expose missing bits of API and clean up inconsistencies. This adds an entirely new feature (tracking partially does this, but that's by accident, not design). I would be unhappy landing this without having [Gitlab user @tristan] or [Gitlab user @juergbi] at least approve the feature (neither of them seems to have replied to the ML topic). Before we land this we should probably also define what "canonical" means, document it, and ensure that our implementation here actually applies that format through tests. I doubt that it currently is particularly stable, or matches what the larger yaml community thinks should be "canonical". In my mind this aims to be an opinionated formatter, and if we're writing one of those, we should do it right. Hence I also think that this entire feature probably should be a separate project. bst-lint is something that has been discussed before, and probably more suitable. Also, on that note, please add some tests beyond ensuring that completion works. In GitLab by [Gitlab user @coldtom] on Nov 22, 2018, 09:48 I understand concerns about feature creep, I don't think BuildStream would need this feature if not for bst track changing the formatting on being run. As this is due to ruamel.yaml, there isn't much we can do but mitigate the damage. Currently canonical format is "whatever the ruamel.yaml dumper spits out by default" but this could be configured to be something defined and consistent. The format we consider "canonical" will have to be consistent with what the dumper does when run, otherwise we will have the same issue when running bst track. This may limit how opinionated we can be. I'm not opposed to the formatter being a separate tool, but if it is we need to document that bst track will modify your yaml format and we strongly recommend using the tool. In GitLab by [Gitlab user @coldtom] on Nov 22, 2018, 09:54 Commented on buildstream/_scheduler/queues/formatqueue.py line 15 While considering the implementation details I realised that this would offer little to no optimisation given the feature as it stands. The quickest way I can think of to check if an element is formatted correctly is to dump the node and compare the files, but obviously this is slower than just dumping the node. In GitLab by [Gitlab user @coldtom] on Nov 22, 2018, 16:44 Commented on buildstream/_stream.py line 297 changed this line in version 3 of the diff In GitLab by [Gitlab user @coldtom] on Nov 22, 2018, 16:44 Commented on buildstream/_scheduler/queues/formatqueue.py line 15 changed this line in version 3 of the diff In GitLab by [Gitlab user @coldtom] on Nov 22, 2018, 16:44 added 74 commits 1d2d0de4...abef70fe - 72 commits from branch master3bf4ed98 - bst fmt: Add basic functionality8d77677e - bst fmt: Add tests for core functionality Compare with previous version In GitLab by [Gitlab user @tristanvb] on Dec 3, 2018, 09:45 Commented on buildstream/element.py line 1342 This probably needs to have a lot more control over the ordering of keys in the output. I.e. see https://gitlab.com/BuildStream/buildstream/issues/485#note_121782763 In GitLab by [Gitlab user @coldtom] on Dec 13, 2018, 11:31 marked as a Work In Progress In GitLab by [Gitlab user @coldtom] on Dec 13, 2018, 13:46 added 157 commits 8d77677e...224aa4c2 - 154 commits from branch master4d3c032a - bst fmt: Add basic functionality6b0a2e1a - bst fmt: Add tests for core functionality778f1db9 - bst-fmt: Allow greater control over order Compare with previous version In GitLab by [Gitlab user @coldtom] on Dec 17, 2018, 12:37 added 15 commits 778f1db9...b23bec55 - 12 commits from branch master95307bb4 - bst fmt: Add basic functionality574e5a78 - bst fmt: Add tests for core functionalityc1cef0cc - bst-fmt: Allow greater control over order Compare with previous version In GitLab by [Gitlab user @coldtom] on Dec 20, 2018, 11:41 added 12 commits c1cef0cc...aae5e4b3 - 9 commits from branch master67d8d0a1 - bst fmt: Add basic functionality9b5e2133 - bst fmt: Add tests for core functionalityecaf433f - bst-fmt: Allow greater control over node order Compare with previous version In GitLab by [Gitlab user @coldtom] on Dec 20, 2018, 11:43 Commented on buildstream/element.py line 1342 [Gitlab user @tristanvb] I have reworked this to have greater control over the order of dumping, but it has introduced an issue in stripping comments in some situations. I still need to add tests but want feedback on the approach used first. In GitLab by [Gitlab user @LaurenceUrhegyi] on Mar 26, 2019, 18:15 Hi, thanks for the contribution! Unfortunately this WIP MR is over a month without an update from the author, so as per our policy I'm going to close this to keep our queue tidy. Please do re-open the MR if you come back to work on it. Cheers! In GitLab by [Gitlab user @LaurenceUrhegyi] on Mar 26, 2019, 18:15 closed In GitLab by [Gitlab user @jjardon] on Mar 26, 2019, 19:32 This MR is waiting for feedback from the maintainers; reopening In GitLab by [Gitlab user @jjardon] on Mar 26, 2019, 19:32 reopened In GitLab by [Gitlab user @jjardon] on May 4, 2019, 03:54 [Gitlab user @coldtom] can you please rebase this? For the comments seems It's (almost) ready to merge In GitLab by [Gitlab user @danielsilverstone-ct] on Jul 4, 2019, 09:44 Hi, thanks for the contribution! Unfortunately this WIP MR is over a month without an update from the author, so as per our policy I'm going to close this to keep our queue tidy. [Gitlab user @coldtom] please re-open the MR if you come back to work on it. Cheers! In GitLab by [Gitlab user @danielsilverstone-ct] on Jul 4, 2019, 09:44 closed In GitLab by [Gitlab user @jjardon] on Dec 9, 2019, 09:15 Again, This MR was waiting for feedback from the maintainers; reopening In GitLab by [Gitlab user @jjardon] on Dec 9, 2019, 09:15 reopened In GitLab by [Gitlab user @BenjaminSchubert] on Dec 9, 2019, 10:13 I'm not sure what we would gain with that command: The ruamel formatter in theory should not change the format of the file at all. In practice, if it does, that's a bug that should be reported upstream and fixed. Other tools (like yamllint) are very good at linting your yaml, and I believe that having them as a precommit or part of the pipeline might be easier and less work to maintain in BuildStream itself. In GitLab by [Gitlab user @jjardon] on Dec 12, 2019, 04:34 [Gitlab user @BenjaminSchubert] does yamllint convert to a canonical format or only errors/warns if something is not following the format? This is about the former In GitLab by [Gitlab user @BenjaminSchubert] on May 11, 2020, 20:46 [Gitlab user @jjardon] woops seems I missed that one. No, yamllint doesn't convert to a canonical format. Seems like we'd still need that one then.
gharchive/pull-request
2021-02-04T07:16:39
2025-04-01T04:55:57.453642
{ "authors": [ "BuildStream-Migration-Bot" ], "repo": "apache/buildstream", "url": "https://github.com/apache/buildstream/pull/1415", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
636839944
kamel run directory (self contained) It would also in the future be nice to be able to run by just specifying a directory, then the CLI figures out what files are in the directory and what kind they are myrepo/foo - MyRoute.java - application.properties - myawesomelib.jar kamel run github:user/myrepo/foo And it gets translated into that example above from Luca in #1522 (incl adding the -d for the JAR). In fact this is "anti modeline", as you dont use the //came-k: comments but I love when end users can just keep all their stuff in the same folder, and also seperated a bit. properties goes into real properties files. And extra JARs are just downloaded JARs (no funky long maven dependencies). I think this can be made compatible with modeline. Thinking to a local directory, the CLI can look for all compatible files with a // camel-k: language=x (or another flag that marks it as runnable) and run each runnable file using existing modeline options. It's common e.g. to write different .properties files that should be picked by some integrations, but not all. Modeline now also contain dependencies that should be read by the CLI. The only problem that I see in a remote (github.com/user/repo/) is that the CLI would require to list the files or check them out, which is not a lightweight operation. Ah yeah the github is not as important (just cool for demos and new users to Camel K). Good point about shared properties files etc or allowing to specify which properties file are to be used by which runnable file. So you could have myrepo/foo - MyRoute.java - application.properties - Funky.groovy - func.properties - myawesomelib.jar Then in MyRoute.java and Funky.groovy you can use the modeline to specify which properties files and JARs to include. And then you can run it as 2 integrations by kamel run foo
gharchive/issue
2020-06-11T08:49:54
2025-04-01T04:55:57.458740
{ "authors": [ "davsclaus", "nicolaferraro" ], "repo": "apache/camel-k", "url": "https://github.com/apache/camel-k/issues/1523", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1227715024
1.9.x TestHealthTrait/Readiness_condition_with_stopped_route failure Right now, only this test is failing: --- FAIL: TestHealthTrait/Readiness_condition_with_stopped_route (416.44s) Originally posted by @squakez in https://github.com/apache/camel-k/issues/3256#issuecomment-1119494943 Should be fixed already.
gharchive/issue
2022-05-06T10:58:51
2025-04-01T04:55:57.460581
{ "authors": [ "squakez", "tadayosi" ], "repo": "apache/camel-k", "url": "https://github.com/apache/camel-k/issues/3258", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2409271131
Bitcoin Source fails with It seems like there is a dependency missing for the bitcoin-source. Note: All the following resources were neated and stripped of namespace for ease of use during debugging. When running: apiVersion: camel.apache.org/v1 kind: Pipe metadata: annotations: camel.apache.org/operator.id: camel-k name: bitcoin-rate spec: sink: ref: apiVersion: camel.apache.org/v1 kind: Kamelet name: log-sink source: ref: apiVersion: camel.apache.org/v1 kind: Kamelet name: bitcoin-source it is built correctly, but throws a NoClassDefFoundError during startup for javax/ws/rs/Path. Operator and client version are both 2.2.0. The integration platform is configured as follows: apiVersion: camel.apache.org/v1 kind: IntegrationPlatform metadata: annotations: camel.apache.org/operator.id: camel-k labels: app: camel-k name: cs spec: build: maven: settings: secretKeyRef: key: settings.xml name: maven-settings settingsSecurity: secretKeyRef: key: settings-security.xml name: maven-settings registry: address: harbor.192-168-178-58.sslip.io insecure: true organization: myorg secret: harbor runtimeVersion: 3.6.0 cluster: Kubernetes traits: camel: runtimeVersion: 3.6.0 Full stacktrace as logged: 2024-07-15 17:31:12,191 ERROR [io.qua.run.Application] (main) Failed to start application (with profile [prod]): java.lang.RuntimeException: Failed to start quarkus at io.quarkus.runner.ApplicationImpl.doStart(Unknown Source) at io.quarkus.runtime.Application.start(Application.java:101) at io.quarkus.runtime.ApplicationLifecycleManager.run(ApplicationLifecycleManager.java:111) at io.quarkus.runtime.Quarkus.run(Quarkus.java:71) at io.quarkus.runtime.Quarkus.run(Quarkus.java:44) at io.quarkus.runtime.Quarkus.run(Quarkus.java:124) at io.quarkus.runner.GeneratedMain.main(Unknown Source) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at io.quarkus.bootstrap.runner.QuarkusEntryPoint.doRun(QuarkusEntryPoint.java:61) at io.quarkus.bootstrap.runner.QuarkusEntryPoint.main(QuarkusEntryPoint.java:32) Caused by: java.lang.NoClassDefFoundError: javax/ws/rs/Path at si.mazi.rescu.RestInvocationHandler.<init>(RestInvocationHandler.java:59) at si.mazi.rescu.RestProxyFactory.createProxy(RestProxyFactory.java:47) at si.mazi.rescu.RestProxyFactoryImpl.createProxy(RestProxyFactoryImpl.java:9) at org.knowm.xchange.client.ExchangeRestProxyBuilder.build(ExchangeRestProxyBuilder.java:80) at org.knowm.xchange.binance.service.BinanceBaseService.<init>(BinanceBaseService.java:37) at org.knowm.xchange.binance.service.BinanceMarketDataServiceRaw.<init>(BinanceMarketDataServiceRaw.java:23) at org.knowm.xchange.binance.service.BinanceMarketDataService.<init>(BinanceMarketDataService.java:28) at org.knowm.xchange.binance.BinanceExchange.initServices(BinanceExchange.java:29) at org.knowm.xchange.BaseExchange.applySpecification(BaseExchange.java:108) at org.knowm.xchange.binance.BinanceExchange.applySpecification(BinanceExchange.java:72) at org.knowm.xchange.ExchangeFactory.createExchange(ExchangeFactory.java:110) at org.knowm.xchange.ExchangeFactory.createExchange(ExchangeFactory.java:53) at org.apache.camel.component.xchange.XChangeComponent.createExchange(XChangeComponent.java:64) at org.apache.camel.component.xchange.XChangeComponent.getOrCreateXChange(XChangeComponent.java:72) at org.apache.camel.component.xchange.XChangeComponent.createEndpoint(XChangeComponent.java:43) at org.apache.camel.support.DefaultComponent.createEndpoint(DefaultComponent.java:170) at org.apache.camel.impl.engine.AbstractCamelContext.doGetEndpoint(AbstractCamelContext.java:804) at org.apache.camel.impl.engine.AbstractCamelContext.getEndpoint(AbstractCamelContext.java:738) at org.apache.camel.support.CamelContextHelper.resolveEndpoint(CamelContextHelper.java:123) at org.apache.camel.reifier.SendReifier.resolveEndpoint(SendReifier.java:45) at org.apache.camel.reifier.SendReifier.createProcessor(SendReifier.java:37) at org.apache.camel.reifier.ProcessorReifier.makeProcessor(ProcessorReifier.java:864) at org.apache.camel.reifier.ProcessorReifier.addRoutes(ProcessorReifier.java:604) at org.apache.camel.reifier.RouteReifier.doCreateRoute(RouteReifier.java:213) at org.apache.camel.reifier.RouteReifier.createRoute(RouteReifier.java:76) at org.apache.camel.impl.DefaultModelReifierFactory.createRoute(DefaultModelReifierFactory.java:49) at org.apache.camel.impl.DefaultCamelContext.startRouteDefinitions(DefaultCamelContext.java:724) at org.apache.camel.component.kamelet.KameletComponent$LifecycleHandler.createRouteForEndpoint(KameletComponent.java:416) at org.apache.camel.component.kamelet.KameletComponent$LifecycleHandler.onContextInitialized(KameletComponent.java:430) at org.apache.camel.impl.engine.AbstractCamelContext.doInit(AbstractCamelContext.java:2360) at org.apache.camel.quarkus.core.FastCamelContext.doInit(FastCamelContext.java:176) at org.apache.camel.support.service.BaseService.init(BaseService.java:78) at org.apache.camel.impl.engine.AbstractCamelContext.init(AbstractCamelContext.java:1983) at org.apache.camel.support.service.BaseService.start(BaseService.java:105) at org.apache.camel.impl.engine.AbstractCamelContext.start(AbstractCamelContext.java:2002) at org.apache.camel.impl.DefaultCamelContext.start(DefaultCamelContext.java:208) at org.apache.camel.quarkus.main.CamelMain.doStart(CamelMain.java:94) at org.apache.camel.support.service.BaseService.start(BaseService.java:113) at org.apache.camel.quarkus.main.CamelMain.startEngine(CamelMain.java:140) at org.apache.camel.quarkus.main.CamelMainRuntime.start(CamelMainRuntime.java:49) at org.apache.camel.quarkus.core.CamelBootstrapRecorder.start(CamelBootstrapRecorder.java:45) at io.quarkus.deployment.steps.CamelBootstrapProcessor$boot173480958.deploy_0(Unknown Source) at io.quarkus.deployment.steps.CamelBootstrapProcessor$boot173480958.deploy(Unknown Source) ... 13 more Caused by: java.lang.ClassNotFoundException: javax.ws.rs.Path at java.base/jdk.internal.loader.BuiltinClassLoader.loadClass(BuiltinClassLoader.java:641) at java.base/jdk.internal.loader.ClassLoaders$AppClassLoader.loadClass(ClassLoaders.java:188) at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:525) at io.quarkus.bootstrap.runner.RunnerClassLoader.loadClass(RunnerClassLoader.java:115) at io.quarkus.bootstrap.runner.RunnerClassLoader.loadClass(RunnerClassLoader.java:65) ... 56 more Please let me know in case you need additional information. This is a problem in the Camel K runtime. If you run the Kamelet in pure Camel with Jbang it will work out of the box. Eventually this should go in Camel K/Camel Quarkus issues. Thanks for reporting.
gharchive/issue
2024-07-15T17:36:04
2025-04-01T04:55:57.465686
{ "authors": [ "mwmahlberg", "oscerd" ], "repo": "apache/camel-kamelets", "url": "https://github.com/apache/camel-kamelets/issues/2112", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2120235034
Extend the pod specs for the dev mode and build Is there a way we could extend / customize the pod / deployment yaml file of the run in devmode and build? Basically, we need to push some traits / tolerances to the pods running, and could not find a way to update it. Thanks! Add deployment.jkube.yaml to your project with deployment spec fragment, ex: spec: replicas: 1 template: spec: containers: - volumeMounts: - name: xxx mountPath: /xxxxx serviceAccount: karavan imagePullSecrets: - name: karavan-registry works for build, not for devmode Traits are high level named features of Camel K Karavan does not use Camel-K but Camel+Jib+JKube
gharchive/issue
2024-02-06T08:35:58
2025-04-01T04:55:57.468477
{ "authors": [ "arheom", "mgubaidullin" ], "repo": "apache/camel-karavan", "url": "https://github.com/apache/camel-karavan/issues/1106", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1937644990
Upgrade Netty to version 4.1.100.Final (#11690) Description Target [x] I checked that the commit is targeting the correct branch (note that Camel 3 uses camel-3.x, whereas Camel 4 uses the main branch) Tracking [ ] If this is a large change, bug fix, or code improvement, I checked there is a JIRA issue filed for the change (usually before you start working on it). Apache Camel coding standards and style [x] I checked that each commit in the pull request has a meaningful subject line and body. [x] I have run mvn clean install -DskipTests locally and I have committed all auto-generated changes /component-test camel-amqp camel-aws/camel-aws2-kinesis camel-coap camel-grpc camel-hl7 camel-iec60870 camel-lumberjack camel-netty-http camel-netty camel-opentelemetry camel-platform-http-main camel-rest-openapi camel-salesforce/camel-salesforce-component camel-stitch camel-syslog camel-telegram camel-undertow camel-webhook camel-whatsapp
gharchive/pull-request
2023-10-11T12:22:35
2025-04-01T04:55:57.475179
{ "authors": [ "oscerd" ], "repo": "apache/camel", "url": "https://github.com/apache/camel/pull/11694", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
276929955
CAMEL-12030: Add CoAP response code header https://issues.apache.org/jira/browse/CAMEL-12030 Thanks the PR has been merged.
gharchive/pull-request
2017-11-27T08:00:50
2025-04-01T04:55:57.476629
{ "authors": [ "jamesnetherton", "oscerd" ], "repo": "apache/camel", "url": "https://github.com/apache/camel/pull/2121", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
745580600
[UI] UI Changes to support new storage plugin for PowerFlex/ScaleIO storage pool Generate URL from the UI inputs when adding "PowerFlex" Primary Storage Allow VM Snapshot for stopped VM on KVM hypervisor and PowerFlex/ScaleIO storage pool @blueorangutan package @shwstppr a Jenkins job has been kicked to build primate packages. I'll keep you posted as I make progress. Packaging result: :heavy_check_mark:centos :heavy_check_mark:debian :heavy_check_mark:archive. QA: http://primate-qa.cloudstack.cloud:8080/client/pr/863 (JID-3673) @sureshanaparti is this ready to review/test? @sureshanaparti is this ready to review/test? not yet @GabrielBrascher @blueorangutan package @blueorangutan package @borisstoyanov a Jenkins job has been kicked to build primate packages. I'll keep you posted as I make progress. @borisstoyanov a Jenkins job has been kicked to build primate packages. I'll keep you posted as I make progress. Packaging result: :heavy_check_mark:centos :heavy_check_mark:debian :heavy_check_mark:archive. QA: http://primate-qa.cloudstack.cloud:8080/client/pr/863 (JID-3816) Packaging result: :heavy_check_mark:centos :heavy_check_mark:debian :heavy_check_mark:archive. QA: http://primate-qa.cloudstack.cloud:8080/client/pr/863 (JID-3816)
gharchive/pull-request
2020-11-18T11:32:09
2025-04-01T04:55:57.482670
{ "authors": [ "GabrielBrascher", "blueorangutan", "borisstoyanov", "shwstppr", "sureshanaparti" ], "repo": "apache/cloudstack-primate", "url": "https://github.com/apache/cloudstack-primate/pull/863", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2550789718
Multiple secondary storage in one zone; problem with volume snapshots ISSUE TYPE Bug Report CLOUDSTACK VERSION 4.19.1.0, xcp-ng hypervisor SUMMARY We are planning to add multiple NFS secondary storages for backup purpose. And I don't understand the principle of distribution of storage snapshots among secstors. I created 10 Intsances in one Account and all their volume snapshots exept one were copied to the first secstore. Only one volume snapshot was copied to the second NFS secstore, but I can't create addiditinal snapshots of this volume. I attach logs in comments. STEPS TO REPRODUCE Add multiple NFS secondary storages, try to create volume snapshots of different instances EXPECTED RESULTS Snapshots are copied based on free space allocation mode to all secstores ACTUAL RESULTS All snapshotes are copied to the first secstore Example of successful snapshot 2024-09-26 17:21:41,283 DEBUG [o.a.c.f.j.i.AsyncJobManagerImpl] (API-Job-Executor-63:ctx-968ed109 job-72464) (logid:c8e45246) Executing AsyncJobVO: {id:72464, userId: 298, accountId: 225, instanceType: Snapshot, instanceId: 870, cmd: org.apache.cloudstack.api.command.user.snapshot.CreateSnapshotCmd, cmdInfo: {"asyncbackup":"false","response":"json","ctxUserId":"298","volumeid":"5e964ddb-3d14-4037-a241-12972ebc5354","httpmethod":"GET","ctxStartEventId":"1811491","id":"870","ctxDetails":"{\"interface com.cloud.storage.Volume\":\"5e964ddb-3d14-4037-a241-12972ebc5354\",\"interface com.cloud.storage.Snapshot\":\"7bdc207e-974e-41ed-a51f-f104bac1102c\"}","ctxAccountId":"225","uuid":"7bdc207e-974e-41ed-a51f-f104bac1102c","cmdEventType":"SNAPSHOT.CREATE"}, cmdVersion: 0, status: IN_PROGRESS, processStatus: 0, resultCode: 0, result: null, initMsid: 108597816265214, completeMsid: null, lastUpdated: null, lastPolled: null, created: null, removed: null} 2024-09-26 17:21:41,293 DEBUG [c.c.u.AccountManagerImpl] (API-Job-Executor-63:ctx-968ed109 job-72464 ctx-7ca9e6a7) (logid:c8e45246) Access to Account [{"accountName":"karasev-cluster5","id":225,"uuid":"fd32d818-5d3b-4d3c-9084-34704beefaf5"}] granted to Account [{"accountName":"karasev-cluster5","id":225,"uuid":"fd32d818-5d3b-4d3c-9084-34704beefaf5"}] by DomainChecker 2024-09-26 17:21:41,304 DEBUG [c.c.a.ApiServlet] (qtp501107890-4821895:ctx-285c7273) (logid:dbfd31f0) ===START=== 172.16.11.61 -- GET jobId=c8e45246-57d3-4238-8cf1-6b7fa887c144&command=queryAsyncJobResult&response=json 2024-09-26 17:21:41,329 DEBUG [c.c.a.ApiServlet] (qtp501107890-4821895:ctx-285c7273 ctx-655df114) (logid:dbfd31f0) ===END=== 172.16.11.61 -- GET jobId=c8e45246-57d3-4238-8cf1-6b7fa887c144&command=queryAsyncJobResult&response=json 2024-09-26 17:21:41,333 DEBUG [c.c.u.AccountManagerImpl] (API-Job-Executor-63:ctx-968ed109 job-72464 ctx-7ca9e6a7) (logid:c8e45246) Access to org.apache.cloudstack.storage.volume.VolumeObject@46645c52 granted to Account [{"accountName":"karasev-cluster5","id":225,"uuid":"fd32d818-5d3b-4d3c-9084-34704beefaf5"}] by DomainChecker 2024-09-26 17:21:41,341 DEBUG [c.c.u.AccountManagerImpl] (API-Job-Executor-63:ctx-968ed109 job-72464 ctx-7ca9e6a7) (logid:c8e45246) Access to VM instance {"id":5474,"instanceName":"i-225-5474-VM","type":"User","uuid":"64c1d590-88a5-4dea-aede-76e0afa0bf32"} granted to Account [{"accountName":"karasev-cluster5","id":225,"uuid":"fd32d818-5d3b-4d3c-9084-34704beefaf5"}] by DomainChecker 2024-09-26 17:21:41,350 DEBUG [o.a.c.f.j.i.AsyncJobManagerImpl] (API-Job-Executor-63:ctx-968ed109 job-72464 ctx-7ca9e6a7) (logid:c8e45246) Sync job-72465 execution on object VmWorkJobQueue.5474 2024-09-26 17:21:42,578 DEBUG [o.a.c.f.j.i.AsyncJobManagerImpl] (Work-Job-Executor-152:ctx-2271fcd8 job-72464/job-72465) (logid:c8e45246) Executing AsyncJobVO: {id:72465, userId: 298, accountId: 225, instanceType: null, instanceId: null, cmd: com.cloud.vm.VmWorkTakeVolumeSnapshot, cmdInfo: rO0ABXNyACVjb20uY2xvdWQudm0uVm1Xb3JrVGFrZVZvbHVtZVNuYXBzaG90BL5gG4Li1c8CAAdaAAthc3luY0JhY2t1cFoACXF1aWVzY2VWbUwADGxvY2F0aW9uVHlwZXQAKUxjb20vY2xvdWQvc3RvcmFnZS9TbmFwc2hvdCRMb2NhdGlvblR5cGU7TAAIcG9saWN5SWR0ABBMamF2YS9sYW5nL0xvbmc7TAAKc25hcHNob3RJZHEAfgACTAAIdm9sdW1lSWRxAH4AAkwAB3pvbmVJZHN0ABBMamF2YS91dGlsL0xpc3Q7eHIAE2NvbS5jbG91ZC52bS5WbVdvcmufmbZW8CVnawIABEoACWFjY291bnRJZEoABnVzZXJJZEoABHZtSWRMAAtoYW5kbGVyTmFtZXQAEkxqYXZhL2xhbmcvU3RyaW5nO3hwAAAAAAAAAOEAAAAAAAABKgAAAAAAABVidAAUVm9sdW1lQXBpU2VydmljZUltcGwAAHBzcgAOamF2YS5sYW5nLkxvbmc7i-SQzI8j3wIAAUoABXZhbHVleHIAEGphdmEubGFuZy5OdW1iZXKGrJUdC5TgiwIAAHhwAAAAAAAAAABzcQB-AAgAAAAAAAADZnNxAH4ACAAAAAAAABgTcA, cmdVersion: 0, status: IN_PROGRESS, processStatus: 0, resultCode: 0, result: null, initMsid: 108597816265214, completeMsid: null, lastUpdated: null, lastPolled: null, created: Thu Sep 26 17:21:41 MSK 2024, removed: null} 2024-09-26 17:21:42,579 DEBUG [c.c.v.VmWorkJobDispatcher] (Work-Job-Executor-152:ctx-2271fcd8 job-72464/job-72465) (logid:c8e45246) Run VM work job: com.cloud.vm.VmWorkTakeVolumeSnapshot for VM 5474, job origin: 72464 2024-09-26 17:21:42,581 DEBUG [c.c.v.VmWorkJobHandlerProxy] (Work-Job-Executor-152:ctx-2271fcd8 job-72464/job-72465 ctx-75e57019) (logid:c8e45246) Execute VM work job: com.cloud.vm.VmWorkTakeVolumeSnapshot{"volumeId":6163,"policyId":0,"snapshotId":870,"quiesceVm":false,"asyncBackup":false,"userId":298,"accountId":225,"vmId":5474,"handlerName":"VolumeApiServiceImpl"} 2024-09-26 17:21:42,594 DEBUG [o.a.c.s.s.StorPoolSnapshotStrategy] (Work-Job-Executor-152:ctx-2271fcd8 job-72464/job-72465 ctx-75e57019) (logid:c8e45246) StorpoolSnapshotStrategy.canHandle: snapshot=backup-test-karasev-9_ROOT-5474_20240926142141, uuid=7bdc207e-974e-41ed-a51f-f104bac1102c, op=TAKE 2024-09-26 17:21:42,632 DEBUG [o.a.c.s.d.d.CloudStackPrimaryDataStoreDriverImpl] (Work-Job-Executor-152:ctx-2271fcd8 job-72464/job-72465 ctx-75e57019) (logid:c8e45246) Taking snapshot of org.apache.cloudstack.storage.snapshot.SnapshotObject@4bd9bc84 2024-09-26 17:21:42,666 DEBUG [o.a.c.s.d.d.CloudStackPrimaryDataStoreDriverImpl] (Work-Job-Executor-152:ctx-2271fcd8 job-72464/job-72465 ctx-75e57019) (logid:c8e45246) Taking snapshot of org.apache.cloudstack.storage.snapshot.SnapshotObject@4bd9bc84 and encryption required is false 2024-09-26 17:21:42,666 DEBUG [c.c.h.o.r.Ovm3HypervisorGuru] (Work-Job-Executor-152:ctx-2271fcd8 job-72464/job-72465 ctx-75e57019) (logid:c8e45246) getCommandHostDelegation: class org.apache.cloudstack.storage.command.CreateObjectCommand 2024-09-26 17:21:42,667 DEBUG [c.c.h.XenServerGuru] (Work-Job-Executor-152:ctx-2271fcd8 job-72464/job-72465 ctx-75e57019) (logid:c8e45246) We are returning the default host to execute commands because the command is not of Copy type. 2024-09-26 17:21:42,670 DEBUG [c.c.a.t.Request] (Work-Job-Executor-152:ctx-2271fcd8 job-72464/job-72465 ctx-75e57019) (logid:c8e45246) Seq 77360-7409547286931430185: Sending { Cmd , MgmtId: 108597816265214, via: 77360(xcp-sr246-u0708.aisrp.local), Ver: v1, Flags: 100011, [{"org.apache.cloudstack.storage.command.CreateObjectCommand":{"data":{"org.apache.cloudstack.storage.to.SnapshotObjectTO":{"volume":{"uuid":"5e964ddb-3d14-4037-a241-12972ebc5354","volumeType":"ROOT","dataStore":{"org.apache.cloudstack.storage.to.PrimaryDataStoreTO":{"uuid":"DM7100_LUN4","name":"cluster5-dm7100_lun4","id":"54","poolType":"PreSetup","host":"localhost","path":"/DM7100_LUN4","port":"0","url":"PreSetup://localhost/DM7100_LUN4/?ROLE=Primary&STOREUUID=DM7100_LUN4","isManaged":"false"}},"name":"ROOT-5474","size":"(10.00 GB) 10737418240","path":"9c99b264-d618-4030-916b-1502ec5fadba","volumeId":"6163","vmName":"i-225-5474-VM","accountId":"225","format":"VHD","provisioningType":"THIN","poolId":"54","id":"6163","deviceId":"0","cacheMode":"NONE","hypervisorType":"XenServer","directDownload":"false","deployAsIs":"false","followRedirects":"true"},"dataStore":{"org.apache.cloudstack.storage.to.PrimaryDataStoreTO":{"uuid":"DM7100_LUN4","name":"cluster5-dm7100_lun4","id":"54","poolType":"PreSetup","host":"localhost","path":"/DM7100_LUN4","port":"0","url":"PreSetup://localhost/DM7100_LUN4/?ROLE=Primary&STOREUUID=DM7100_LUN4","isManaged":"false"}},"vmName":"i-225-5474-VM","name":"backup-test-karasev-9_ROOT-5474_20240926142141","hypervisorType":"XenServer","id":"870","quiescevm":"false","physicalSize":"0","accountId":"225","followRedirects":"false"}},"wait":"0","bypassHostMaintenance":"false"}}] } 2024-09-26 17:21:42,671 DEBUG [c.c.a.t.Request] (Work-Job-Executor-152:ctx-2271fcd8 job-72464/job-72465 ctx-75e57019) (logid:c8e45246) Seq 77360-7409547286931430185: Executing: { Cmd , MgmtId: 108597816265214, via: 77360(xcp-sr246-u0708.aisrp.local), Ver: v1, Flags: 100011, [{"org.apache.cloudstack.storage.command.CreateObjectCommand":{"data":{"org.apache.cloudstack.storage.to.SnapshotObjectTO":{"volume":{"uuid":"5e964ddb-3d14-4037-a241-12972ebc5354","volumeType":"ROOT","dataStore":{"org.apache.cloudstack.storage.to.PrimaryDataStoreTO":{"uuid":"DM7100_LUN4","name":"cluster5-dm7100_lun4","id":"54","poolType":"PreSetup","host":"localhost","path":"/DM7100_LUN4","port":"0","url":"PreSetup://localhost/DM7100_LUN4/?ROLE=Primary&STOREUUID=DM7100_LUN4","isManaged":"false"}},"name":"ROOT-5474","size":"(10.00 GB) 10737418240","path":"9c99b264-d618-4030-916b-1502ec5fadba","volumeId":"6163","vmName":"i-225-5474-VM","accountId":"225","format":"VHD","provisioningType":"THIN","poolId":"54","id":"6163","deviceId":"0","cacheMode":"NONE","hypervisorType":"XenServer","directDownload":"false","deployAsIs":"false","followRedirects":"true"},"dataStore":{"org.apache.cloudstack.storage.to.PrimaryDataStoreTO":{"uuid":"DM7100_LUN4","name":"cluster5-dm7100_lun4","id":"54","poolType":"PreSetup","host":"localhost","path":"/DM7100_LUN4","port":"0","url":"PreSetup://localhost/DM7100_LUN4/?ROLE=Primary&STOREUUID=DM7100_LUN4","isManaged":"false"}},"vmName":"i-225-5474-VM","name":"backup-test-karasev-9_ROOT-5474_20240926142141","hypervisorType":"XenServer","id":"870","quiescevm":"false","physicalSize":"0","accountId":"225","followRedirects":"false"}},"wait":"0","bypassHostMaintenance":"false"}}] } 2024-09-26 17:21:42,672 DEBUG [c.c.s.r.StorageSubsystemCommandHandlerBase] (DirectAgent-287:ctx-df71784d) (logid:c8e45246) Executing command CreateObjectCommand: [{"data":{"org.apache.cloudstack.storage.to.SnapshotObjectTO":{"volume":{"uuid":"5e964ddb-3d14-4037-a241-12972ebc5354","volumeType":"ROOT","dataStore":{"org.apache.cloudstack.storage.to.PrimaryDataStoreTO":{"uuid":"DM7100_LUN4","name":"cluster5-dm7100_lun4","id":54,"poolType":"PreSetup","host":"localhost","path":"/DM7100_LUN4","port":0,"url":"PreSetup://localhost/DM7100_LUN4/?ROLE=Primary&STOREUUID=DM7100_LUN4","isManaged":false}},"name":"ROOT-5474","size":10737418240,"path":"9c99b264-d618-4030-916b-1502ec5fadba","volumeId":6163,"vmName":"i-225-5474-VM","accountId":225,"format":"VHD","provisioningType":"THIN","poolId":54,"id":6163,"deviceId":0,"cacheMode":"NONE","hypervisorType":"XenServer","directDownload":false,"deployAsIs":false,"followRedirects":true},"dataStore":{"org.apache.cloudstack.storage.to.PrimaryDataStoreTO":{"uuid":"DM7100_LUN4","name":"cluster5-dm7100_lun4","id":54,"poolType":"PreSetup","host":"localhost","path":"/DM7100_LUN4","port":0,"url":"PreSetup://localhost/DM7100_LUN4/?ROLE=Primary&STOREUUID=DM7100_LUN4","isManaged":false}},"vmName":"i-225-5474-VM","name":"backup-test-karasev-9_ROOT-5474_20240926142141","hypervisorType":"XenServer","id":870,"quiescevm":false,"physicalSize":0,"accountId":225,"followRedirects":false}},"wait":0,"bypassHostMaintenance":false}]. 2024-09-26 17:21:44,421 DEBUG [c.c.a.ApiServlet] (qtp501107890-4834338:ctx-38b579a2) (logid:2740bef9) ===START=== 172.16.11.61 -- GET jobId=c8e45246-57d3-4238-8cf1-6b7fa887c144&command=queryAsyncJobResult&response=json 2024-09-26 17:21:44,446 DEBUG [c.c.a.ApiServlet] (qtp501107890-4834338:ctx-38b579a2 ctx-9d7b2d8d) (logid:2740bef9) ===END=== 172.16.11.61 -- GET jobId=c8e45246-57d3-4238-8cf1-6b7fa887c144&command=queryAsyncJobResult&response=json 2024-09-26 17:21:47,537 DEBUG [c.c.a.ApiServlet] (qtp501107890-4834338:ctx-eb108eb7) (logid:25800034) ===START=== 172.16.11.61 -- GET jobId=c8e45246-57d3-4238-8cf1-6b7fa887c144&command=queryAsyncJobResult&response=json 2024-09-26 17:21:47,561 DEBUG [c.c.a.ApiServlet] (qtp501107890-4834338:ctx-eb108eb7 ctx-d3852c77) (logid:25800034) ===END=== 172.16.11.61 -- GET jobId=c8e45246-57d3-4238-8cf1-6b7fa887c144&command=queryAsyncJobResult&response=json 2024-09-26 17:21:50,671 DEBUG [c.c.a.ApiServlet] (qtp501107890-4824845:ctx-95e72378) (logid:4e24b10a) ===START=== 172.16.11.61 -- GET jobId=c8e45246-57d3-4238-8cf1-6b7fa887c144&command=queryAsyncJobResult&response=json 2024-09-26 17:21:50,693 DEBUG [c.c.a.ApiServlet] (qtp501107890-4824845:ctx-95e72378 ctx-9321eab1) (logid:4e24b10a) ===END=== 172.16.11.61 -- GET jobId=c8e45246-57d3-4238-8cf1-6b7fa887c144&command=queryAsyncJobResult&response=json 2024-09-26 17:21:53,779 DEBUG [c.c.a.ApiServlet] (qtp501107890-4821895:ctx-581ef07b) (logid:dca0df5b) ===START=== 172.16.11.61 -- GET jobId=c8e45246-57d3-4238-8cf1-6b7fa887c144&command=queryAsyncJobResult&response=json 2024-09-26 17:21:53,811 DEBUG [c.c.a.ApiServlet] (qtp501107890-4821895:ctx-581ef07b ctx-69923e87) (logid:dca0df5b) ===END=== 172.16.11.61 -- GET jobId=c8e45246-57d3-4238-8cf1-6b7fa887c144&command=queryAsyncJobResult&response=json 2024-09-26 17:21:56,898 DEBUG [c.c.a.ApiServlet] (qtp501107890-4824845:ctx-cc466224) (logid:d706fc8d) ===START=== 172.16.11.61 -- GET jobId=c8e45246-57d3-4238-8cf1-6b7fa887c144&command=queryAsyncJobResult&response=json 2024-09-26 17:21:56,926 DEBUG [c.c.a.ApiServlet] (qtp501107890-4824845:ctx-cc466224 ctx-f28715e9) (logid:d706fc8d) ===END=== 172.16.11.61 -- GET jobId=c8e45246-57d3-4238-8cf1-6b7fa887c144&command=queryAsyncJobResult&response=json 2024-09-26 17:21:57,319 DEBUG [c.c.a.m.DirectAgentAttache] (DirectAgent-287:ctx-df71784d) (logid:c8e45246) Seq 77360-7409547286931430185: Response Received: 2024-09-26 17:21:57,319 DEBUG [c.c.a.t.Request] (DirectAgent-287:ctx-df71784d) (logid:c8e45246) Seq 77360-7409547286931430185: Processing: { Ans: , MgmtId: 108597816265214, via: 77360(xcp-sr246-u0708.aisrp.local), Ver: v1, Flags: 10, [{"org.apache.cloudstack.storage.command.CreateObjectAnswer":{"data":{"org.apache.cloudstack.storage.to.SnapshotObjectTO":{"path":"d8f3d271-4b1d-4276-b4ea-390c3f82df57","id":"0","quiescevm":"false","physicalSize":"0","accountId":"0","followRedirects":"false"}},"result":"true","wait":"0","bypassHostMaintenance":"false"}}] } 2024-09-26 17:21:57,319 DEBUG [c.c.a.t.Request] (Work-Job-Executor-152:ctx-2271fcd8 job-72464/job-72465 ctx-75e57019) (logid:c8e45246) Seq 77360-7409547286931430185: Received: { Ans: , MgmtId: 108597816265214, via: 77360(xcp-sr246-u0708.aisrp.local), Ver: v1, Flags: 10, { CreateObjectAnswer } } 2024-09-26 17:21:57,377 DEBUG [o.a.c.s.h.HeuristicRuleHelper] (Work-Job-Executor-152:ctx-2271fcd8 job-72464/job-72465 ctx-75e57019) (logid:c8e45246) No heuristic rules found for zone with ID [7] and heuristic type [SNAPSHOT]. Returning null. 2024-09-26 17:21:57,381 DEBUG [c.c.s.StatsCollector] (Work-Job-Executor-152:ctx-2271fcd8 job-72464/job-72465 ctx-75e57019) (logid:c8e45246) Verifying image storage [7]. Capacity: total=[785 GB], used=[243 GB], threshold=[89.99999761581421%]. 2024-09-26 17:21:57,381 DEBUG [c.c.s.StatsCollector] (Work-Job-Executor-152:ctx-2271fcd8 job-72464/job-72465 ctx-75e57019) (logid:c8e45246) Verifying image storage [9]. Capacity: total=[785 GB], used=[72 GB], threshold=[89.99999761581421%]. 2024-09-26 17:21:57,382 DEBUG [c.c.s.StatsCollector] (Work-Job-Executor-152:ctx-2271fcd8 job-72464/job-72465 ctx-75e57019) (logid:c8e45246) Verifying image storage [7]. Capacity: total=[785 GB], used=[243 GB], threshold=[89.99999761581421%]. 2024-09-26 17:21:57,419 DEBUG [o.a.c.s.m.AncientDataMotionStrategy] (Work-Job-Executor-152:ctx-2271fcd8 job-72464/job-72465 ctx-75e57019) (logid:c8e45246) copyAsync inspecting src type SNAPSHOT copyAsync inspecting dest type SNAPSHOT 2024-09-26 17:21:57,529 ERROR [o.a.c.s.e.DefaultEndPointSelector] (Work-Job-Executor-152:ctx-2271fcd8 job-72464/job-72465 ctx-75e57019) (logid:c8e45246) IR24 select BACKUPSNAPSHOT from primary to secondary 870 dest=870 2024-09-26 17:21:57,535 DEBUG [c.c.h.o.r.Ovm3HypervisorGuru] (Work-Job-Executor-152:ctx-2271fcd8 job-72464/job-72465 ctx-75e57019) (logid:c8e45246) getCommandHostDelegation: class org.apache.cloudstack.storage.command.CopyCommand 2024-09-26 17:21:57,537 DEBUG [c.c.h.XenServerGuru] (Work-Job-Executor-152:ctx-2271fcd8 job-72464/job-72465 ctx-75e57019) (logid:c8e45246) We are returning the default host to execute commands because the source and destination objects are not NFS type. 2024-09-26 17:21:57,540 DEBUG [c.c.a.t.Request] (Work-Job-Executor-152:ctx-2271fcd8 job-72464/job-72465 ctx-75e57019) (logid:c8e45246) Seq 77360-7409547286931430189: Sending { Cmd , MgmtId: 108597816265214, via: 77360(xcp-sr246-u0708.aisrp.local), Ver: v1, Flags: 100111, [{"org.apache.cloudstack.storage.command.CopyCommand":{"srcTO":{"org.apache.cloudstack.storage.to.SnapshotObjectTO":{"path":"d8f3d271-4b1d-4276-b4ea-390c3f82df57","volume":{"uuid":"5e964ddb-3d14-4037-a241-12972ebc5354","volumeType":"ROOT","dataStore":{"org.apache.cloudstack.storage.to.PrimaryDataStoreTO":{"uuid":"DM7100_LUN4","name":"cluster5-dm7100_lun4","id":"54","poolType":"PreSetup","host":"localhost","path":"/DM7100_LUN4","port":"0","url":"PreSetup://localhost/DM7100_LUN4/?ROLE=Primary&STOREUUID=DM7100_LUN4","isManaged":"false"}},"name":"ROOT-5474","size":"(10.00 GB) 10737418240","path":"9c99b264-d618-4030-916b-1502ec5fadba","volumeId":"6163","vmName":"i-225-5474-VM","accountId":"225","format":"VHD","provisioningType":"THIN","poolId":"54","id":"6163","deviceId":"0","cacheMode":"NONE","hypervisorType":"XenServer","directDownload":"false","deployAsIs":"false","followRedirects":"true"},"dataStore":{"org.apache.cloudstack.storage.to.PrimaryDataStoreTO":{"uuid":"DM7100_LUN4","name":"cluster5-dm7100_lun4","id":"54","poolType":"PreSetup","host":"localhost","path":"/DM7100_LUN4","port":"0","url":"PreSetup://localhost/DM7100_LUN4/?ROLE=Primary&STOREUUID=DM7100_LUN4","isManaged":"false"}},"vmName":"i-225-5474-VM","name":"backup-test-karasev-9_ROOT-5474_20240926142141","hypervisorType":"XenServer","id":"870","quiescevm":"false","physicalSize":"0","accountId":"225","followRedirects":"false"}},"destTO":{"org.apache.cloudstack.storage.to.SnapshotObjectTO":{"path":"snapshots/225/6163","volume":{"uuid":"5e964ddb-3d14-4037-a241-12972ebc5354","volumeType":"ROOT","dataStore":{"org.apache.cloudstack.storage.to.PrimaryDataStoreTO":{"uuid":"DM7100_LUN4","name":"cluster5-dm7100_lun4","id":"54","poolType":"PreSetup","host":"localhost","path":"/DM7100_LUN4","port":"0","url":"PreSetup://localhost/DM7100_LUN4/?ROLE=Primary&STOREUUID=DM7100_LUN4","isManaged":"false"}},"name":"ROOT-5474","size":"(10.00 GB) 10737418240","path":"9c99b264-d618-4030-916b-1502ec5fadba","volumeId":"6163","vmName":"i-225-5474-VM","accountId":"225","format":"VHD","provisioningType":"THIN","poolId":"54","id":"6163","deviceId":"0","cacheMode":"NONE","hypervisorType":"XenServer","directDownload":"false","deployAsIs":"false","followRedirects":"true"},"dataStore":{"com.cloud.agent.api.to.NfsTO":{"_url":"nfs://10.69.105.254/var/NFS","_role":"Image"}},"vmName":"i-225-5474-VM","name":"backup-test-karasev-9_ROOT-5474_20240926142141","hypervisorType":"XenServer","id":"870","quiescevm":"false","physicalSize":"0","accountId":"225","followRedirects":"false"}},"executeInSequence":"true","options":{"snapshot.backup.to.secondary":"true","fullSnapshot":"true"},"options2":{},"wait":"21600","bypassHostMaintenance":"false"}}] } 2024-09-26 17:21:57,562 DEBUG [c.c.a.t.Request] (Work-Job-Executor-152:ctx-2271fcd8 job-72464/job-72465 ctx-75e57019) (logid:c8e45246) Seq 77360-7409547286931430189: Executing: { Cmd , MgmtId: 108597816265214, via: 77360(xcp-sr246-u0708.aisrp.local), Ver: v1, Flags: 100111, [{"org.apache.cloudstack.storage.command.CopyCommand":{"srcTO":{"org.apache.cloudstack.storage.to.SnapshotObjectTO":{"path":"d8f3d271-4b1d-4276-b4ea-390c3f82df57","volume":{"uuid":"5e964ddb-3d14-4037-a241-12972ebc5354","volumeType":"ROOT","dataStore":{"org.apache.cloudstack.storage.to.PrimaryDataStoreTO":{"uuid":"DM7100_LUN4","name":"cluster5-dm7100_lun4","id":"54","poolType":"PreSetup","host":"localhost","path":"/DM7100_LUN4","port":"0","url":"PreSetup://localhost/DM7100_LUN4/?ROLE=Primary&STOREUUID=DM7100_LUN4","isManaged":"false"}},"name":"ROOT-5474","size":"(10.00 GB) 10737418240","path":"9c99b264-d618-4030-916b-1502ec5fadba","volumeId":"6163","vmName":"i-225-5474-VM","accountId":"225","format":"VHD","provisioningType":"THIN","poolId":"54","id":"6163","deviceId":"0","cacheMode":"NONE","hypervisorType":"XenServer","directDownload":"false","deployAsIs":"false","followRedirects":"true"},"dataStore":{"org.apache.cloudstack.storage.to.PrimaryDataStoreTO":{"uuid":"DM7100_LUN4","name":"cluster5-dm7100_lun4","id":"54","poolType":"PreSetup","host":"localhost","path":"/DM7100_LUN4","port":"0","url":"PreSetup://localhost/DM7100_LUN4/?ROLE=Primary&STOREUUID=DM7100_LUN4","isManaged":"false"}},"vmName":"i-225-5474-VM","name":"backup-test-karasev-9_ROOT-5474_20240926142141","hypervisorType":"XenServer","id":"870","quiescevm":"false","physicalSize":"0","accountId":"225","followRedirects":"false"}},"destTO":{"org.apache.cloudstack.storage.to.SnapshotObjectTO":{"path":"snapshots/225/6163","volume":{"uuid":"5e964ddb-3d14-4037-a241-12972ebc5354","volumeType":"ROOT","dataStore":{"org.apache.cloudstack.storage.to.PrimaryDataStoreTO":{"uuid":"DM7100_LUN4","name":"cluster5-dm7100_lun4","id":"54","poolType":"PreSetup","host":"localhost","path":"/DM7100_LUN4","port":"0","url":"PreSetup://localhost/DM7100_LUN4/?ROLE=Primary&STOREUUID=DM7100_LUN4","isManaged":"false"}},"name":"ROOT-5474","size":"(10.00 GB) 10737418240","path":"9c99b264-d618-4030-916b-1502ec5fadba","volumeId":"6163","vmName":"i-225-5474-VM","accountId":"225","format":"VHD","provisioningType":"THIN","poolId":"54","id":"6163","deviceId":"0","cacheMode":"NONE","hypervisorType":"XenServer","directDownload":"false","deployAsIs":"false","followRedirects":"true"},"dataStore":{"com.cloud.agent.api.to.NfsTO":{"_url":"nfs://10.69.105.254/var/NFS","_role":"Image"}},"vmName":"i-225-5474-VM","name":"backup-test-karasev-9_ROOT-5474_20240926142141","hypervisorType":"XenServer","id":"870","quiescevm":"false","physicalSize":"0","accountId":"225","followRedirects":"false"}},"executeInSequence":"true","options":{"snapshot.backup.to.secondary":"true","fullSnapshot":"true"},"options2":{},"wait":"21600","bypassHostMaintenance":"false"}}] } 2024-09-26 17:21:57,562 DEBUG [c.c.s.r.StorageSubsystemCommandHandlerBase] (DirectAgent-366:ctx-f42b18d7) (logid:c8e45246) Executing command CopyCommand: [{"srcTO":{"org.apache.cloudstack.storage.to.SnapshotObjectTO":{"path":"d8f3d271-4b1d-4276-b4ea-390c3f82df57","volume":{"uuid":"5e964ddb-3d14-4037-a241-12972ebc5354","volumeType":"ROOT","dataStore":{"org.apache.cloudstack.storage.to.PrimaryDataStoreTO":{"uuid":"DM7100_LUN4","name":"cluster5-dm7100_lun4","id":54,"poolType":"PreSetup","host":"localhost","path":"/DM7100_LUN4","port":0,"url":"PreSetup://localhost/DM7100_LUN4/?ROLE=Primary&STOREUUID=DM7100_LUN4","isManaged":false}},"name":"ROOT-5474","size":10737418240,"path":"9c99b264-d618-4030-916b-1502ec5fadba","volumeId":6163,"vmName":"i-225-5474-VM","accountId":225,"format":"VHD","provisioningType":"THIN","poolId":54,"id":6163,"deviceId":0,"cacheMode":"NONE","hypervisorType":"XenServer","directDownload":false,"deployAsIs":false,"followRedirects":true},"dataStore":{"org.apache.cloudstack.storage.to.PrimaryDataStoreTO":{"uuid":"DM7100_LUN4","name":"cluster5-dm7100_lun4","id":54,"poolType":"PreSetup","host":"localhost","path":"/DM7100_LUN4","port":0,"url":"PreSetup://localhost/DM7100_LUN4/?ROLE=Primary&STOREUUID=DM7100_LUN4","isManaged":false}},"vmName":"i-225-5474-VM","name":"backup-test-karasev-9_ROOT-5474_20240926142141","hypervisorType":"XenServer","id":870,"quiescevm":false,"physicalSize":0,"accountId":225,"followRedirects":false}},"destTO":{"org.apache.cloudstack.storage.to.SnapshotObjectTO":{"path":"snapshots/225/6163","volume":{"uuid":"5e964ddb-3d14-4037-a241-12972ebc5354","volumeType":"ROOT","dataStore":{"org.apache.cloudstack.storage.to.PrimaryDataStoreTO":{"uuid":"DM7100_LUN4","name":"cluster5-dm7100_lun4","id":54,"poolType":"PreSetup","host":"localhost","path":"/DM7100_LUN4","port":0,"url":"PreSetup://localhost/DM7100_LUN4/?ROLE=Primary&STOREUUID=DM7100_LUN4","isManaged":false}},"name":"ROOT-5474","size":10737418240,"path":"9c99b264-d618-4030-916b-1502ec5fadba","volumeId":6163,"vmName":"i-225-5474-VM","accountId":225,"format":"VHD","provisioningType":"THIN","poolId":54,"id":6163,"deviceId":0,"cacheMode":"NONE","hypervisorType":"XenServer","directDownload":false,"deployAsIs":false,"followRedirects":true},"dataStore":{"com.cloud.agent.api.to.NfsTO":{"_url":"nfs://10.69.105.254/var/NFS","_role":"Image"}},"vmName":"i-225-5474-VM","name":"backup-test-karasev-9_ROOT-5474_20240926142141","hypervisorType":"XenServer","id":870,"quiescevm":false,"physicalSize":0,"accountId":225,"followRedirects":false}},"executeInSequence":true,"options":{"snapshot.backup.to.secondary":"true","fullSnapshot":"true"},"options2":{},"wait":21600,"bypassHostMaintenance":false}]. 2024-09-26 17:21:59,970 DEBUG [c.c.h.x.r.XenServerStorageProcessor] (DirectAgent-366:ctx-f42b18d7) (logid:c8e45246) No file SR found for path: /var/cloud_mount/440acca8-622d-3dfb-825d-b3e62d7bba87/snapshots/225/6163 2024-09-26 17:21:59,970 DEBUG [c.c.h.x.r.XenServerStorageProcessor] (DirectAgent-366:ctx-f42b18d7) (logid:c8e45246) Creating file SR for path [/var/cloud_mount/440acca8-622d-3dfb-825d-b3e62d7bba87/snapshots/225/6163] on host [7fe64564-c05c-4425-8436-ec2da15f740a] 2024-09-26 17:22:00,006 DEBUG [c.c.a.ApiServlet] (qtp501107890-4832815:ctx-29b0404b) (logid:360fad58) ===START=== 172.16.11.61 -- GET jobId=c8e45246-57d3-4238-8cf1-6b7fa887c144&command=queryAsyncJobResult&response=json 2024-09-26 17:22:00,032 DEBUG [c.c.a.ApiServlet] (qtp501107890-4832815:ctx-29b0404b ctx-d7854f91) (logid:360fad58) ===END=== 172.16.11.61 -- GET jobId=c8e45246-57d3-4238-8cf1-6b7fa887c144&command=queryAsyncJobResult&response=json 2024-09-26 17:22:03,119 DEBUG [c.c.a.ApiServlet] (qtp501107890-4834338:ctx-c04fc898) (logid:2219e417) ===START=== 172.16.11.61 -- GET jobId=c8e45246-57d3-4238-8cf1-6b7fa887c144&command=queryAsyncJobResult&response=json 2024-09-26 17:22:03,142 DEBUG [c.c.a.ApiServlet] (qtp501107890-4834338:ctx-c04fc898 ctx-fec4282e) (logid:2219e417) ===END=== 172.16.11.61 -- GET jobId=c8e45246-57d3-4238-8cf1-6b7fa887c144&command=queryAsyncJobResult&response=json 2024-09-26 17:22:06,257 DEBUG [c.c.a.ApiServlet] (qtp501107890-4824845:ctx-63120037) (logid:8a9b8304) ===START=== 172.16.11.61 -- GET jobId=c8e45246-57d3-4238-8cf1-6b7fa887c144&command=queryAsyncJobResult&response=json 2024-09-26 17:22:06,280 DEBUG [c.c.a.ApiServlet] (qtp501107890-4824845:ctx-63120037 ctx-6adff946) (logid:8a9b8304) ===END=== 172.16.11.61 -- GET jobId=c8e45246-57d3-4238-8cf1-6b7fa887c144&command=queryAsyncJobResult&response=json 2024-09-26 17:22:09,402 DEBUG [c.c.a.ApiServlet] (qtp501107890-4832815:ctx-b3166a90) (logid:b7878727) ===START=== 172.16.11.61 -- GET jobId=c8e45246-57d3-4238-8cf1-6b7fa887c144&command=queryAsyncJobResult&response=json 2024-09-26 17:22:09,424 DEBUG [c.c.a.ApiServlet] (qtp501107890-4832815:ctx-b3166a90 ctx-af252453) (logid:b7878727) ===END=== 172.16.11.61 -- GET jobId=c8e45246-57d3-4238-8cf1-6b7fa887c144&command=queryAsyncJobResult&response=json 2024-09-26 17:22:12,536 DEBUG [c.c.a.ApiServlet] (qtp501107890-4824845:ctx-5e16f60d) (logid:ed82faf7) ===START=== 172.16.11.61 -- GET jobId=c8e45246-57d3-4238-8cf1-6b7fa887c144&command=queryAsyncJobResult&response=json 2024-09-26 17:22:12,559 DEBUG [c.c.a.ApiServlet] (qtp501107890-4824845:ctx-5e16f60d ctx-021e66d4) (logid:ed82faf7) ===END=== 172.16.11.61 -- GET jobId=c8e45246-57d3-4238-8cf1-6b7fa887c144&command=queryAsyncJobResult&response=json 2024-09-26 17:22:15,663 DEBUG [c.c.a.ApiServlet] (qtp501107890-4821895:ctx-516e5765) (logid:a4ec32b9) ===START=== 172.16.11.61 -- GET jobId=c8e45246-57d3-4238-8cf1-6b7fa887c144&command=queryAsyncJobResult&response=json 2024-09-26 17:22:15,686 DEBUG [c.c.a.ApiServlet] (qtp501107890-4821895:ctx-516e5765 ctx-f115c344) (logid:a4ec32b9) ===END=== 172.16.11.61 -- GET jobId=c8e45246-57d3-4238-8cf1-6b7fa887c144&command=queryAsyncJobResult&response=json 2024-09-26 17:22:18,799 DEBUG [c.c.a.ApiServlet] (qtp501107890-4824844:ctx-b60f4585) (logid:253c7fb4) ===START=== 172.16.11.61 -- GET jobId=c8e45246-57d3-4238-8cf1-6b7fa887c144&command=queryAsyncJobResult&response=json 2024-09-26 17:22:18,822 DEBUG [c.c.a.ApiServlet] (qtp501107890-4824844:ctx-b60f4585 ctx-cd2c67ed) (logid:253c7fb4) ===END=== 172.16.11.61 -- GET jobId=c8e45246-57d3-4238-8cf1-6b7fa887c144&command=queryAsyncJobResult&response=json 2024-09-26 17:22:21,936 DEBUG [c.c.a.ApiServlet] (qtp501107890-4821895:ctx-863d8a59) (logid:08b2e49f) ===START=== 172.16.11.61 -- GET jobId=c8e45246-57d3-4238-8cf1-6b7fa887c144&command=queryAsyncJobResult&response=json 2024-09-26 17:22:21,959 DEBUG [c.c.a.ApiServlet] (qtp501107890-4821895:ctx-863d8a59 ctx-95f1425b) (logid:08b2e49f) ===END=== 172.16.11.61 -- GET jobId=c8e45246-57d3-4238-8cf1-6b7fa887c144&command=queryAsyncJobResult&response=json 2024-09-26 17:22:25,071 DEBUG [c.c.a.ApiServlet] (qtp501107890-4821895:ctx-6e2f29f1) (logid:724bb944) ===START=== 172.16.11.61 -- GET jobId=c8e45246-57d3-4238-8cf1-6b7fa887c144&command=queryAsyncJobResult&response=json 2024-09-26 17:22:25,094 DEBUG [c.c.a.ApiServlet] (qtp501107890-4821895:ctx-6e2f29f1 ctx-fbc38980) (logid:724bb944) ===END=== 172.16.11.61 -- GET jobId=c8e45246-57d3-4238-8cf1-6b7fa887c144&command=queryAsyncJobResult&response=json 2024-09-26 17:22:27,025 DEBUG [c.c.h.x.r.CitrixResourceBase] (DirectAgent-366:ctx-f42b18d7) (logid:c8e45246) Host 10.69.105.21 OpaqueRef:ff865e44-ce8f-4670-b990-855700769647: Removing SR 2024-09-26 17:22:27,038 DEBUG [c.c.h.x.r.CitrixResourceBase] (DirectAgent-366:ctx-f42b18d7) (logid:c8e45246) Host 10.69.105.21 OpaqueRef:d5a56d3b-f5af-461e-bee4-b387938e18fe: Unplugging pbd 2024-09-26 17:22:27,449 DEBUG [c.c.h.x.r.CitrixResourceBase] (DirectAgent-366:ctx-f42b18d7) (logid:c8e45246) Host 10.69.105.21 OpaqueRef:ff865e44-ce8f-4670-b990-855700769647: Forgetting 2024-09-26 17:22:27,454 DEBUG [c.c.h.x.r.XenServerStorageProcessor] (DirectAgent-366:ctx-f42b18d7) (logid:c8e45246) Successfully destroyed snapshot on volume: 9c99b264-d618-4030-916b-1502ec5fadba execept this current snapshot d8f3d271-4b1d-4276-b4ea-390c3f82df57 2024-09-26 17:22:27,454 INFO [c.c.h.x.r.XenServerStorageProcessor] (DirectAgent-366:ctx-f42b18d7) (logid:c8e45246) New snapshot details: SnapshotTO[datastore=null|volume=null|pathsnapshots/225/6163/07116d77-9286-4eaa-9e40-f4b754def19f.vhd] 2024-09-26 17:22:27,455 INFO [c.c.h.x.r.XenServerStorageProcessor] (DirectAgent-366:ctx-f42b18d7) (logid:c8e45246) New snapshot physical utilization: (5.46 GB) 5866705408 2024-09-26 17:22:27,455 DEBUG [c.c.a.m.DirectAgentAttache] (DirectAgent-366:ctx-f42b18d7) (logid:c8e45246) Seq 77360-7409547286931430189: Response Received: 2024-09-26 17:22:27,455 DEBUG [c.c.a.t.Request] (DirectAgent-366:ctx-f42b18d7) (logid:c8e45246) Seq 77360-7409547286931430189: Processing: { Ans: , MgmtId: 108597816265214, via: 77360(xcp-sr246-u0708.aisrp.local), Ver: v1, Flags: 110, [{"org.apache.cloudstack.storage.command.CopyCmdAnswer":{"newData":{"org.apache.cloudstack.storage.to.SnapshotObjectTO":{"path":"snapshots/225/6163/07116d77-9286-4eaa-9e40-f4b754def19f.vhd","id":"0","quiescevm":"false","physicalSize":"5866705408","accountId":"0","followRedirects":"false"}},"result":"true","wait":"0","bypassHostMaintenance":"false"}}] } 2024-09-26 17:22:27,455 DEBUG [c.c.a.m.AgentAttache] (DirectAgent-366:ctx-f42b18d7) (logid:c8e45246) Seq 77360-7409547286931430189: No more commands found 2024-09-26 17:22:27,455 DEBUG [c.c.a.t.Request] (Work-Job-Executor-152:ctx-2271fcd8 job-72464/job-72465 ctx-75e57019) (logid:c8e45246) Seq 77360-7409547286931430189: Received: { Ans: , MgmtId: 108597816265214, via: 77360(xcp-sr246-u0708.aisrp.local), Ver: v1, Flags: 110, { CopyCmdAnswer } } 2024-09-26 17:22:27,491 DEBUG [c.c.r.ResourceLimitManagerImpl] (Work-Job-Executor-152:ctx-2271fcd8 job-72464/job-72465 ctx-75e57019) (logid:c8e45246) Updating resource Type = secondary_storage count for Account = 225 Operation = decreasing Amount = (4.54 GB) 4870712832 2024-09-26 17:22:27,499 DEBUG [c.c.v.VmWorkJobHandlerProxy] (Work-Job-Executor-152:ctx-2271fcd8 job-72464/job-72465 ctx-75e57019) (logid:c8e45246) Done executing VM work job: com.cloud.vm.VmWorkTakeVolumeSnapshot{"volumeId":6163,"policyId":0,"snapshotId":870,"quiesceVm":false,"asyncBackup":false,"userId":298,"accountId":225,"vmId":5474,"handlerName":"VolumeApiServiceImpl"} 2024-09-26 17:22:27,500 DEBUG [o.a.c.f.j.i.AsyncJobManagerImpl] (Work-Job-Executor-152:ctx-2271fcd8 job-72464/job-72465 ctx-75e57019) (logid:c8e45246) Complete async job-72465, jobStatus: SUCCEEDED, resultCode: 0, result: rO0ABXNyAA5qYXZhLmxhbmcuTG9uZzuL5JDMjyPfAgABSgAFdmFsdWV4cgAQamF2YS5sYW5nLk51bWJlcoaslR0LlOCLAgAAeHAAAAAAAAADZg 2024-09-26 17:22:27,501 DEBUG [o.a.c.f.j.i.AsyncJobManagerImpl] (Work-Job-Executor-152:ctx-2271fcd8 job-72464/job-72465 ctx-75e57019) (logid:c8e45246) Publish async job-72465 complete on message bus 2024-09-26 17:22:27,501 DEBUG [o.a.c.f.j.i.AsyncJobManagerImpl] (Work-Job-Executor-152:ctx-2271fcd8 job-72464/job-72465 ctx-75e57019) (logid:c8e45246) Wake up jobs related to job-72465 2024-09-26 17:22:27,501 DEBUG [o.a.c.f.j.i.AsyncJobManagerImpl] (Work-Job-Executor-152:ctx-2271fcd8 job-72464/job-72465 ctx-75e57019) (logid:c8e45246) Update db status for job-72465 2024-09-26 17:22:27,502 DEBUG [o.a.c.f.j.i.AsyncJobManagerImpl] (Work-Job-Executor-152:ctx-2271fcd8 job-72464/job-72465 ctx-75e57019) (logid:c8e45246) Wake up jobs joined with job-72465 and disjoin all subjobs created from job- 72465 2024-09-26 17:22:27,509 DEBUG [c.c.v.VmWorkJobDispatcher] (Work-Job-Executor-152:ctx-2271fcd8 job-72464/job-72465) (logid:c8e45246) Done with run of VM work job: com.cloud.vm.VmWorkTakeVolumeSnapshot for VM 5474, job origin: 72464 2024-09-26 17:22:27,509 DEBUG [o.a.c.f.j.i.AsyncJobManagerImpl] (Work-Job-Executor-152:ctx-2271fcd8 job-72464/job-72465) (logid:c8e45246) Done executing com.cloud.vm.VmWorkTakeVolumeSnapshot for job-72465 2024-09-26 17:22:27,511 INFO [o.a.c.f.j.i.AsyncJobMonitor] (Work-Job-Executor-152:ctx-2271fcd8 job-72464/job-72465) (logid:c8e45246) Remove job-72465 from job monitoring 2024-09-26 17:22:27,545 DEBUG [o.a.c.s.s.StorPoolSnapshotStrategy] (API-Job-Executor-63:ctx-968ed109 job-72464 ctx-7ca9e6a7) (logid:c8e45246) StorpoolSnapshotStrategy.canHandle: snapshot=backup-test-karasev-9_ROOT-5474_20240926142141, uuid=7bdc207e-974e-41ed-a51f-f104bac1102c, op=REVERT 2024-09-26 17:22:27,549 DEBUG [o.a.c.f.j.i.AsyncJobManagerImpl] (API-Job-Executor-63:ctx-968ed109 job-72464 ctx-7ca9e6a7) (logid:c8e45246) Complete async job-72464, jobStatus: SUCCEEDED, resultCode: 0, result: org.apache.cloudstack.api.response.SnapshotResponse/snapshot/{"id":"7bdc207e-974e-41ed-a51f-f104bac1102c","account":"karasev-cluster5","domainid":"ce7e4a4a-75b6-4c96-938c-fe3158da9a47","domain":"cluster5","snapshottype":"MANUAL","volumeid":"5e964ddb-3d14-4037-a241-12972ebc5354","volumename":"ROOT-5474","volumetype":"ROOT","created":"2024-09-26T17:21:41+0300","name":"backup-test-karasev-9_ROOT-5474_20240926142141","intervaltype":"MANUAL","state":"BackedUp","physicalsize":"5866705408","zoneid":"9505445a-7164-4675-876d-1a2fee1f7e46","zonename":"xcp-zone-02","revertable":"false","ostypeid":"7b53b54f-d3d2-409f-b9b0-acd39122368a","osdisplayname":"Ubuntu 22.04 LTS","virtualsize":"10737418240","tags":[],"hasannotations":"false"} 2024-09-26 17:22:27,550 DEBUG [o.a.c.f.j.i.AsyncJobManagerImpl] (API-Job-Executor-63:ctx-968ed109 job-72464 ctx-7ca9e6a7) (logid:c8e45246) Publish async job-72464 complete on message bus 2024-09-26 17:22:27,550 DEBUG [o.a.c.f.j.i.AsyncJobManagerImpl] (API-Job-Executor-63:ctx-968ed109 job-72464 ctx-7ca9e6a7) (logid:c8e45246) Wake up jobs related to job-72464 2024-09-26 17:22:27,550 DEBUG [o.a.c.f.j.i.AsyncJobManagerImpl] (API-Job-Executor-63:ctx-968ed109 job-72464 ctx-7ca9e6a7) (logid:c8e45246) Update db status for job-72464 2024-09-26 17:22:27,551 DEBUG [o.a.c.f.j.i.AsyncJobManagerImpl] (API-Job-Executor-63:ctx-968ed109 job-72464 ctx-7ca9e6a7) (logid:c8e45246) Wake up jobs joined with job-72464 and disjoin all subjobs created from job- 72464 2024-09-26 17:22:27,555 DEBUG [o.a.c.f.j.i.AsyncJobManagerImpl] (API-Job-Executor-63:ctx-968ed109 job-72464) (logid:c8e45246) Done executing org.apache.cloudstack.api.command.user.snapshot.CreateSnapshotCmd for job-72464 2024-09-26 17:22:27,555 INFO [o.a.c.f.j.i.AsyncJobMonitor] (API-Job-Executor-63:ctx-968ed109 job-72464) (logid:c8e45246) Remove job-72464 from job monitoring 2024-09-26 17:22:28,208 DEBUG [c.c.a.ApiServlet] (qtp501107890-4834306:ctx-41f1ef09) (logid:5789c854) ===START=== 172.16.11.61 -- GET jobId=c8e45246-57d3-4238-8cf1-6b7fa887c144&command=queryAsyncJobResult&response=json 2024-09-26 17:22:28,238 DEBUG [c.c.a.ApiServlet] (qtp501107890-4834306:ctx-41f1ef09 ctx-be04a227) (logid:5789c854) ===END=== 172.16.11.61 -- GET jobId=c8e45246-57d3-4238-8cf1-6b7fa887c144&command=queryAsyncJobResult&response=json Example when first volume snapshot stores on the second secstore and I create the next volume snapshot 2024-09-26 16:58:32,135 DEBUG [o.a.c.f.j.i.AsyncJobManagerImpl] (API-Job-Executor-58:ctx-c9d433a7 job-72454) (logid:f365dd96) Executing AsyncJobVO: {id:72454, userId: 298, accountId: 225, instanceType: Snapshot, instanceId: 867, cmd: org.apache.cloudstack.api.command.user.snapshot.CreateSnapshotCmd, cmdInfo: {"asyncBackup":"false","quiescevm":"false","response":"json","ctxUserId":"298","volumeId":"da420806-0ff3-43b8-9ee0-4aa2adfd9b99","httpmethod":"GET","ctxStartEventId":"1811467","id":"867","ctxDetails":"{\"interface com.cloud.storage.Volume\":\"da420806-0ff3-43b8-9ee0-4aa2adfd9b99\",\"interface com.cloud.storage.Snapshot\":\"c62b0aa0-5a1c-475c-b7f6-6c9fab17e325\"}","ctxAccountId":"225","uuid":"c62b0aa0-5a1c-475c-b7f6-6c9fab17e325","cmdEventType":"SNAPSHOT.CREATE"}, cmdVersion: 0, status: IN_PROGRESS, processStatus: 0, resultCode: 0, result: null, initMsid: 108597816265214, completeMsid: null, lastUpdated: null, lastPolled: null, created: null, removed: null} 2024-09-26 16:58:32,147 DEBUG [c.c.u.AccountManagerImpl] (API-Job-Executor-58:ctx-c9d433a7 job-72454 ctx-829ed64e) (logid:f365dd96) Access to Account [{"accountName":"karasev-cluster5","id":225,"uuid":"fd32d818-5d3b-4d3c-9084-34704beefaf5"}] granted to Account [{"accountName":"karasev-cluster5","id":225,"uuid":"fd32d818-5d3b-4d3c-9084-34704beefaf5"}] by DomainChecker 2024-09-26 16:58:32,167 DEBUG [c.c.a.ApiServlet] (qtp501107890-4830812:ctx-4bacce10) (logid:11a55db0) ===START=== 172.16.11.61 -- GET jobId=f365dd96-c018-454a-874c-37dce9396d96&command=queryAsyncJobResult&response=json 2024-09-26 16:58:32,183 DEBUG [c.c.u.AccountManagerImpl] (API-Job-Executor-58:ctx-c9d433a7 job-72454 ctx-829ed64e) (logid:f365dd96) Access to org.apache.cloudstack.storage.volume.VolumeObject@f1e18b7 granted to Account [{"accountName":"karasev-cluster5","id":225,"uuid":"fd32d818-5d3b-4d3c-9084-34704beefaf5"}] by DomainChecker 2024-09-26 16:58:32,190 DEBUG [c.c.a.ApiServlet] (qtp501107890-4830812:ctx-4bacce10 ctx-aa1d9072) (logid:11a55db0) ===END=== 172.16.11.61 -- GET jobId=f365dd96-c018-454a-874c-37dce9396d96&command=queryAsyncJobResult&response=json 2024-09-26 16:58:32,192 DEBUG [c.c.u.AccountManagerImpl] (API-Job-Executor-58:ctx-c9d433a7 job-72454 ctx-829ed64e) (logid:f365dd96) Access to VM instance {"id":5468,"instanceName":"i-225-5468-VM","type":"User","uuid":"d1e26871-d01f-44cc-ac61-f19f46ed741a"} granted to Account [{"accountName":"karasev-cluster5","id":225,"uuid":"fd32d818-5d3b-4d3c-9084-34704beefaf5"}] by DomainChecker 2024-09-26 16:58:32,206 DEBUG [o.a.c.f.j.i.AsyncJobManagerImpl] (API-Job-Executor-58:ctx-c9d433a7 job-72454 ctx-829ed64e) (logid:f365dd96) Sync job-72455 execution on object VmWorkJobQueue.5468 2024-09-26 16:58:32,575 DEBUG [o.a.c.f.j.i.AsyncJobManagerImpl] (Work-Job-Executor-148:ctx-ef42787d job-72454/job-72455) (logid:f365dd96) Executing AsyncJobVO: {id:72455, userId: 298, accountId: 225, instanceType: null, instanceId: null, cmd: com.cloud.vm.VmWorkTakeVolumeSnapshot, cmdInfo: rO0ABXNyACVjb20uY2xvdWQudm0uVm1Xb3JrVGFrZVZvbHVtZVNuYXBzaG90BL5gG4Li1c8CAAdaAAthc3luY0JhY2t1cFoACXF1aWVzY2VWbUwADGxvY2F0aW9uVHlwZXQAKUxjb20vY2xvdWQvc3RvcmFnZS9TbmFwc2hvdCRMb2NhdGlvblR5cGU7TAAIcG9saWN5SWR0ABBMamF2YS9sYW5nL0xvbmc7TAAKc25hcHNob3RJZHEAfgACTAAIdm9sdW1lSWRxAH4AAkwAB3pvbmVJZHN0ABBMamF2YS91dGlsL0xpc3Q7eHIAE2NvbS5jbG91ZC52bS5WbVdvcmufmbZW8CVnawIABEoACWFjY291bnRJZEoABnVzZXJJZEoABHZtSWRMAAtoYW5kbGVyTmFtZXQAEkxqYXZhL2xhbmcvU3RyaW5nO3hwAAAAAAAAAOEAAAAAAAABKgAAAAAAABVcdAAUVm9sdW1lQXBpU2VydmljZUltcGwAAHBzcgAOamF2YS5sYW5nLkxvbmc7i-SQzI8j3wIAAUoABXZhbHVleHIAEGphdmEubGFuZy5OdW1iZXKGrJUdC5TgiwIAAHhwAAAAAAAAAABzcQB-AAgAAAAAAAADY3NxAH4ACAAAAAAAABgMcA, cmdVersion: 0, status: IN_PROGRESS, processStatus: 0, resultCode: 0, result: null, initMsid: 108597816265214, completeMsid: null, lastUpdated: null, lastPolled: null, created: Thu Sep 26 16:58:32 MSK 2024, removed: null} 2024-09-26 16:58:32,575 DEBUG [c.c.v.VmWorkJobDispatcher] (Work-Job-Executor-148:ctx-ef42787d job-72454/job-72455) (logid:f365dd96) Run VM work job: com.cloud.vm.VmWorkTakeVolumeSnapshot for VM 5468, job origin: 72454 2024-09-26 16:58:32,576 DEBUG [c.c.v.VmWorkJobHandlerProxy] (Work-Job-Executor-148:ctx-ef42787d job-72454/job-72455 ctx-d1890501) (logid:f365dd96) Execute VM work job: com.cloud.vm.VmWorkTakeVolumeSnapshot{"volumeId":6156,"policyId":0,"snapshotId":867,"quiesceVm":false,"asyncBackup":false,"userId":298,"accountId":225,"vmId":5468,"handlerName":"VolumeApiServiceImpl"} 2024-09-26 16:58:32,588 DEBUG [o.a.c.s.s.StorPoolSnapshotStrategy] (Work-Job-Executor-148:ctx-ef42787d job-72454/job-72455 ctx-d1890501) (logid:f365dd96) StorpoolSnapshotStrategy.canHandle: snapshot=backup-test-karasev-3_ROOT-5468_20240926135832, uuid=c62b0aa0-5a1c-475c-b7f6-6c9fab17e325, op=TAKE 2024-09-26 16:58:32,622 DEBUG [o.a.c.s.d.d.CloudStackPrimaryDataStoreDriverImpl] (Work-Job-Executor-148:ctx-ef42787d job-72454/job-72455 ctx-d1890501) (logid:f365dd96) Taking snapshot of org.apache.cloudstack.storage.snapshot.SnapshotObject@18f701c2 2024-09-26 16:58:32,653 DEBUG [o.a.c.s.d.d.CloudStackPrimaryDataStoreDriverImpl] (Work-Job-Executor-148:ctx-ef42787d job-72454/job-72455 ctx-d1890501) (logid:f365dd96) Taking snapshot of org.apache.cloudstack.storage.snapshot.SnapshotObject@18f701c2 and encryption required is false 2024-09-26 16:58:32,653 DEBUG [c.c.h.o.r.Ovm3HypervisorGuru] (Work-Job-Executor-148:ctx-ef42787d job-72454/job-72455 ctx-d1890501) (logid:f365dd96) getCommandHostDelegation: class org.apache.cloudstack.storage.command.CreateObjectCommand 2024-09-26 16:58:32,655 DEBUG [c.c.h.XenServerGuru] (Work-Job-Executor-148:ctx-ef42787d job-72454/job-72455 ctx-d1890501) (logid:f365dd96) We are returning the default host to execute commands because the command is not of Copy type. 2024-09-26 16:58:32,657 DEBUG [c.c.a.t.Request] (Work-Job-Executor-148:ctx-ef42787d job-72454/job-72455 ctx-d1890501) (logid:f365dd96) Seq 77353-4517673376206130891: Sending { Cmd , MgmtId: 108597816265214, via: 77353(xcp-sr246-u1314.aisrp.local), Ver: v1, Flags: 100011, [{"org.apache.cloudstack.storage.command.CreateObjectCommand":{"data":{"org.apache.cloudstack.storage.to.SnapshotObjectTO":{"volume":{"uuid":"da420806-0ff3-43b8-9ee0-4aa2adfd9b99","volumeType":"ROOT","dataStore":{"org.apache.cloudstack.storage.to.PrimaryDataStoreTO":{"uuid":"DM7100_LUN3","name":"cluster5-dm7100_lun3","id":"53","poolType":"PreSetup","host":"localhost","path":"/DM7100_LUN3","port":"0","url":"PreSetup://localhost/DM7100_LUN3/?ROLE=Primary&STOREUUID=DM7100_LUN3","isManaged":"false"}},"name":"ROOT-5468","size":"(20.00 GB) 21474836480","path":"9e2a4d6e-3ab8-46f4-85cf-c7454fc9ea17","volumeId":"6156","vmName":"i-225-5468-VM","accountId":"225","format":"VHD","provisioningType":"THIN","poolId":"53","id":"6156","deviceId":"0","cacheMode":"NONE","hypervisorType":"XenServer","directDownload":"false","deployAsIs":"false","followRedirects":"true"},"parentSnapshotPath":"70424963-5933-4e73-b185-cfdc402175cd","dataStore":{"org.apache.cloudstack.storage.to.PrimaryDataStoreTO":{"uuid":"DM7100_LUN3","name":"cluster5-dm7100_lun3","id":"53","poolType":"PreSetup","host":"localhost","path":"/DM7100_LUN3","port":"0","url":"PreSetup://localhost/DM7100_LUN3/?ROLE=Primary&STOREUUID=DM7100_LUN3","isManaged":"false"}},"vmName":"i-225-5468-VM","name":"backup-test-karasev-3_ROOT-5468_20240926135832","hypervisorType":"XenServer","id":"867","quiescevm":"false","parents":["70424963-5933-4e73-b185-cfdc402175cd"],"physicalSize":"0","accountId":"225","followRedirects":"false"}},"wait":"0","bypassHostMaintenance":"false"}}] } 2024-09-26 16:58:32,659 DEBUG [c.c.a.t.Request] (Work-Job-Executor-148:ctx-ef42787d job-72454/job-72455 ctx-d1890501) (logid:f365dd96) Seq 77353-4517673376206130891: Executing: { Cmd , MgmtId: 108597816265214, via: 77353(xcp-sr246-u1314.aisrp.local), Ver: v1, Flags: 100011, [{"org.apache.cloudstack.storage.command.CreateObjectCommand":{"data":{"org.apache.cloudstack.storage.to.SnapshotObjectTO":{"volume":{"uuid":"da420806-0ff3-43b8-9ee0-4aa2adfd9b99","volumeType":"ROOT","dataStore":{"org.apache.cloudstack.storage.to.PrimaryDataStoreTO":{"uuid":"DM7100_LUN3","name":"cluster5-dm7100_lun3","id":"53","poolType":"PreSetup","host":"localhost","path":"/DM7100_LUN3","port":"0","url":"PreSetup://localhost/DM7100_LUN3/?ROLE=Primary&STOREUUID=DM7100_LUN3","isManaged":"false"}},"name":"ROOT-5468","size":"(20.00 GB) 21474836480","path":"9e2a4d6e-3ab8-46f4-85cf-c7454fc9ea17","volumeId":"6156","vmName":"i-225-5468-VM","accountId":"225","format":"VHD","provisioningType":"THIN","poolId":"53","id":"6156","deviceId":"0","cacheMode":"NONE","hypervisorType":"XenServer","directDownload":"false","deployAsIs":"false","followRedirects":"true"},"parentSnapshotPath":"70424963-5933-4e73-b185-cfdc402175cd","dataStore":{"org.apache.cloudstack.storage.to.PrimaryDataStoreTO":{"uuid":"DM7100_LUN3","name":"cluster5-dm7100_lun3","id":"53","poolType":"PreSetup","host":"localhost","path":"/DM7100_LUN3","port":"0","url":"PreSetup://localhost/DM7100_LUN3/?ROLE=Primary&STOREUUID=DM7100_LUN3","isManaged":"false"}},"vmName":"i-225-5468-VM","name":"backup-test-karasev-3_ROOT-5468_20240926135832","hypervisorType":"XenServer","id":"867","quiescevm":"false","parents":["70424963-5933-4e73-b185-cfdc402175cd"],"physicalSize":"0","accountId":"225","followRedirects":"false"}},"wait":"0","bypassHostMaintenance":"false"}}] } 2024-09-26 16:58:32,659 DEBUG [c.c.s.r.StorageSubsystemCommandHandlerBase] (DirectAgent-128:ctx-052ce805) (logid:f365dd96) Executing command CreateObjectCommand: [{"data":{"org.apache.cloudstack.storage.to.SnapshotObjectTO":{"volume":{"uuid":"da420806-0ff3-43b8-9ee0-4aa2adfd9b99","volumeType":"ROOT","dataStore":{"org.apache.cloudstack.storage.to.PrimaryDataStoreTO":{"uuid":"DM7100_LUN3","name":"cluster5-dm7100_lun3","id":53,"poolType":"PreSetup","host":"localhost","path":"/DM7100_LUN3","port":0,"url":"PreSetup://localhost/DM7100_LUN3/?ROLE=Primary&STOREUUID=DM7100_LUN3","isManaged":false}},"name":"ROOT-5468","size":21474836480,"path":"9e2a4d6e-3ab8-46f4-85cf-c7454fc9ea17","volumeId":6156,"vmName":"i-225-5468-VM","accountId":225,"format":"VHD","provisioningType":"THIN","poolId":53,"id":6156,"deviceId":0,"cacheMode":"NONE","hypervisorType":"XenServer","directDownload":false,"deployAsIs":false,"followRedirects":true},"parentSnapshotPath":"70424963-5933-4e73-b185-cfdc402175cd","dataStore":{"org.apache.cloudstack.storage.to.PrimaryDataStoreTO":{"uuid":"DM7100_LUN3","name":"cluster5-dm7100_lun3","id":53,"poolType":"PreSetup","host":"localhost","path":"/DM7100_LUN3","port":0,"url":"PreSetup://localhost/DM7100_LUN3/?ROLE=Primary&STOREUUID=DM7100_LUN3","isManaged":false}},"vmName":"i-225-5468-VM","name":"backup-test-karasev-3_ROOT-5468_20240926135832","hypervisorType":"XenServer","id":867,"quiescevm":false,"parents":["70424963-5933-4e73-b185-cfdc402175cd"],"physicalSize":0,"accountId":225,"followRedirects":false}},"wait":0,"bypassHostMaintenance":false}]. 2024-09-26 16:58:35,268 DEBUG [c.c.a.ApiServlet] (qtp501107890-4820302:ctx-f711cc5b) (logid:4e44467e) ===START=== 172.16.11.61 -- GET jobId=f365dd96-c018-454a-874c-37dce9396d96&command=queryAsyncJobResult&response=json 2024-09-26 16:58:35,290 DEBUG [c.c.a.ApiServlet] (qtp501107890-4820302:ctx-f711cc5b ctx-e8b6659b) (logid:4e44467e) ===END=== 172.16.11.61 -- GET jobId=f365dd96-c018-454a-874c-37dce9396d96&command=queryAsyncJobResult&response=json 2024-09-26 16:58:38,370 DEBUG [c.c.a.ApiServlet] (qtp501107890-4830812:ctx-fd303fe2) (logid:ef8914d8) ===START=== 172.16.11.61 -- GET jobId=f365dd96-c018-454a-874c-37dce9396d96&command=queryAsyncJobResult&response=json 2024-09-26 16:58:38,394 DEBUG [c.c.a.ApiServlet] (qtp501107890-4830812:ctx-fd303fe2 ctx-aec37e2b) (logid:ef8914d8) ===END=== 172.16.11.61 -- GET jobId=f365dd96-c018-454a-874c-37dce9396d96&command=queryAsyncJobResult&response=json 2024-09-26 16:58:39,977 DEBUG [c.c.a.m.DirectAgentAttache] (DirectAgent-128:ctx-052ce805) (logid:f365dd96) Seq 77353-4517673376206130891: Response Received: 2024-09-26 16:58:39,977 DEBUG [c.c.a.t.Request] (DirectAgent-128:ctx-052ce805) (logid:f365dd96) Seq 77353-4517673376206130891: Processing: { Ans: , MgmtId: 108597816265214, via: 77353(xcp-sr246-u1314.aisrp.local), Ver: v1, Flags: 10, [{"org.apache.cloudstack.storage.command.CreateObjectAnswer":{"data":{"org.apache.cloudstack.storage.to.SnapshotObjectTO":{"path":"e1971e58-5b6e-4f59-a256-154280cf4d16","id":"0","quiescevm":"false","physicalSize":"0","accountId":"0","followRedirects":"false"}},"result":"true","wait":"0","bypassHostMaintenance":"false"}}] } 2024-09-26 16:58:39,977 DEBUG [c.c.a.t.Request] (Work-Job-Executor-148:ctx-ef42787d job-72454/job-72455 ctx-d1890501) (logid:f365dd96) Seq 77353-4517673376206130891: Received: { Ans: , MgmtId: 108597816265214, via: 77353(xcp-sr246-u1314.aisrp.local), Ver: v1, Flags: 10, { CreateObjectAnswer } } 2024-09-26 16:58:40,045 DEBUG [o.a.c.s.h.HeuristicRuleHelper] (Work-Job-Executor-148:ctx-ef42787d job-72454/job-72455 ctx-d1890501) (logid:f365dd96) No heuristic rules found for zone with ID [7] and heuristic type [SNAPSHOT]. Returning null. 2024-09-26 16:58:40,049 DEBUG [c.c.s.StatsCollector] (Work-Job-Executor-148:ctx-ef42787d job-72454/job-72455 ctx-d1890501) (logid:f365dd96) Verifying image storage [7]. Capacity: total=[785 GB], used=[243 GB], threshold=[89.99999761581421%]. 2024-09-26 16:58:40,050 DEBUG [c.c.s.StatsCollector] (Work-Job-Executor-148:ctx-ef42787d job-72454/job-72455 ctx-d1890501) (logid:f365dd96) Verifying image storage [9]. Capacity: total=[785 GB], used=[70 GB], threshold=[89.99999761581421%]. 2024-09-26 16:58:40,051 DEBUG [c.c.s.StatsCollector] (Work-Job-Executor-148:ctx-ef42787d job-72454/job-72455 ctx-d1890501) (logid:f365dd96) Verifying image storage [7]. Capacity: total=[785 GB], used=[243 GB], threshold=[89.99999761581421%]. 2024-09-26 16:58:40,101 DEBUG [o.a.c.s.m.AncientDataMotionStrategy] (Work-Job-Executor-148:ctx-ef42787d job-72454/job-72455 ctx-d1890501) (logid:f365dd96) copyAsync inspecting src type SNAPSHOT copyAsync inspecting dest type SNAPSHOT 2024-09-26 16:58:40,224 ERROR [o.a.c.s.e.DefaultEndPointSelector] (Work-Job-Executor-148:ctx-ef42787d job-72454/job-72455 ctx-d1890501) (logid:f365dd96) IR24 select BACKUPSNAPSHOT from primary to secondary 867 dest=867 2024-09-26 16:58:40,236 DEBUG [c.c.h.o.r.Ovm3HypervisorGuru] (Work-Job-Executor-148:ctx-ef42787d job-72454/job-72455 ctx-d1890501) (logid:f365dd96) getCommandHostDelegation: class org.apache.cloudstack.storage.command.CopyCommand 2024-09-26 16:58:40,238 DEBUG [c.c.h.XenServerGuru] (Work-Job-Executor-148:ctx-ef42787d job-72454/job-72455 ctx-d1890501) (logid:f365dd96) We are returning the default host to execute commands because the source and destination objects are not NFS type. 2024-09-26 16:58:40,243 DEBUG [c.c.a.t.Request] (Work-Job-Executor-148:ctx-ef42787d job-72454/job-72455 ctx-d1890501) (logid:f365dd96) Seq 77358-1940207014466653252: Sending { Cmd , MgmtId: 108597816265214, via: 77358(xcp-sr246-u1112.aisrp.local), Ver: v1, Flags: 100111, [{"org.apache.cloudstack.storage.command.CopyCommand":{"srcTO":{"org.apache.cloudstack.storage.to.SnapshotObjectTO":{"path":"e1971e58-5b6e-4f59-a256-154280cf4d16","volume":{"uuid":"da420806-0ff3-43b8-9ee0-4aa2adfd9b99","volumeType":"ROOT","dataStore":{"org.apache.cloudstack.storage.to.PrimaryDataStoreTO":{"uuid":"DM7100_LUN3","name":"cluster5-dm7100_lun3","id":"53","poolType":"PreSetup","host":"localhost","path":"/DM7100_LUN3","port":"0","url":"PreSetup://localhost/DM7100_LUN3/?ROLE=Primary&STOREUUID=DM7100_LUN3","isManaged":"false"}},"name":"ROOT-5468","size":"(20.00 GB) 21474836480","path":"9e2a4d6e-3ab8-46f4-85cf-c7454fc9ea17","volumeId":"6156","vmName":"i-225-5468-VM","accountId":"225","format":"VHD","provisioningType":"THIN","poolId":"53","id":"6156","deviceId":"0","cacheMode":"NONE","hypervisorType":"XenServer","directDownload":"false","deployAsIs":"false","followRedirects":"true"},"parentSnapshotPath":"70424963-5933-4e73-b185-cfdc402175cd","dataStore":{"org.apache.cloudstack.storage.to.PrimaryDataStoreTO":{"uuid":"DM7100_LUN3","name":"cluster5-dm7100_lun3","id":"53","poolType":"PreSetup","host":"localhost","path":"/DM7100_LUN3","port":"0","url":"PreSetup://localhost/DM7100_LUN3/?ROLE=Primary&STOREUUID=DM7100_LUN3","isManaged":"false"}},"vmName":"i-225-5468-VM","name":"backup-test-karasev-3_ROOT-5468_20240926135832","hypervisorType":"XenServer","id":"867","quiescevm":"false","parents":["70424963-5933-4e73-b185-cfdc402175cd"],"physicalSize":"0","accountId":"225","followRedirects":"false"}},"destTO":{"org.apache.cloudstack.storage.to.SnapshotObjectTO":{"path":"snapshots/225/6156","volume":{"uuid":"da420806-0ff3-43b8-9ee0-4aa2adfd9b99","volumeType":"ROOT","dataStore":{"org.apache.cloudstack.storage.to.PrimaryDataStoreTO":{"uuid":"DM7100_LUN3","name":"cluster5-dm7100_lun3","id":"53","poolType":"PreSetup","host":"localhost","path":"/DM7100_LUN3","port":"0","url":"PreSetup://localhost/DM7100_LUN3/?ROLE=Primary&STOREUUID=DM7100_LUN3","isManaged":"false"}},"name":"ROOT-5468","size":"(20.00 GB) 21474836480","path":"9e2a4d6e-3ab8-46f4-85cf-c7454fc9ea17","volumeId":"6156","vmName":"i-225-5468-VM","accountId":"225","format":"VHD","provisioningType":"THIN","poolId":"53","id":"6156","deviceId":"0","cacheMode":"NONE","hypervisorType":"XenServer","directDownload":"false","deployAsIs":"false","followRedirects":"true"},"dataStore":{"com.cloud.agent.api.to.NfsTO":{"_url":"nfs://10.69.105.254/var/NFS","_role":"Image"}},"vmName":"i-225-5468-VM","name":"backup-test-karasev-3_ROOT-5468_20240926135832","hypervisorType":"XenServer","id":"867","quiescevm":"false","physicalSize":"0","accountId":"225","followRedirects":"false"}},"executeInSequence":"true","options":{"snapshot.backup.to.secondary":"true","fullSnapshot":"false"},"options2":{},"wait":"21600","bypassHostMaintenance":"false"}}] } 2024-09-26 16:58:40,245 DEBUG [c.c.a.t.Request] (Work-Job-Executor-148:ctx-ef42787d job-72454/job-72455 ctx-d1890501) (logid:f365dd96) Seq 77358-1940207014466653252: Executing: { Cmd , MgmtId: 108597816265214, via: 77358(xcp-sr246-u1112.aisrp.local), Ver: v1, Flags: 100111, [{"org.apache.cloudstack.storage.command.CopyCommand":{"srcTO":{"org.apache.cloudstack.storage.to.SnapshotObjectTO":{"path":"e1971e58-5b6e-4f59-a256-154280cf4d16","volume":{"uuid":"da420806-0ff3-43b8-9ee0-4aa2adfd9b99","volumeType":"ROOT","dataStore":{"org.apache.cloudstack.storage.to.PrimaryDataStoreTO":{"uuid":"DM7100_LUN3","name":"cluster5-dm7100_lun3","id":"53","poolType":"PreSetup","host":"localhost","path":"/DM7100_LUN3","port":"0","url":"PreSetup://localhost/DM7100_LUN3/?ROLE=Primary&STOREUUID=DM7100_LUN3","isManaged":"false"}},"name":"ROOT-5468","size":"(20.00 GB) 21474836480","path":"9e2a4d6e-3ab8-46f4-85cf-c7454fc9ea17","volumeId":"6156","vmName":"i-225-5468-VM","accountId":"225","format":"VHD","provisioningType":"THIN","poolId":"53","id":"6156","deviceId":"0","cacheMode":"NONE","hypervisorType":"XenServer","directDownload":"false","deployAsIs":"false","followRedirects":"true"},"parentSnapshotPath":"70424963-5933-4e73-b185-cfdc402175cd","dataStore":{"org.apache.cloudstack.storage.to.PrimaryDataStoreTO":{"uuid":"DM7100_LUN3","name":"cluster5-dm7100_lun3","id":"53","poolType":"PreSetup","host":"localhost","path":"/DM7100_LUN3","port":"0","url":"PreSetup://localhost/DM7100_LUN3/?ROLE=Primary&STOREUUID=DM7100_LUN3","isManaged":"false"}},"vmName":"i-225-5468-VM","name":"backup-test-karasev-3_ROOT-5468_20240926135832","hypervisorType":"XenServer","id":"867","quiescevm":"false","parents":["70424963-5933-4e73-b185-cfdc402175cd"],"physicalSize":"0","accountId":"225","followRedirects":"false"}},"destTO":{"org.apache.cloudstack.storage.to.SnapshotObjectTO":{"path":"snapshots/225/6156","volume":{"uuid":"da420806-0ff3-43b8-9ee0-4aa2adfd9b99","volumeType":"ROOT","dataStore":{"org.apache.cloudstack.storage.to.PrimaryDataStoreTO":{"uuid":"DM7100_LUN3","name":"cluster5-dm7100_lun3","id":"53","poolType":"PreSetup","host":"localhost","path":"/DM7100_LUN3","port":"0","url":"PreSetup://localhost/DM7100_LUN3/?ROLE=Primary&STOREUUID=DM7100_LUN3","isManaged":"false"}},"name":"ROOT-5468","size":"(20.00 GB) 21474836480","path":"9e2a4d6e-3ab8-46f4-85cf-c7454fc9ea17","volumeId":"6156","vmName":"i-225-5468-VM","accountId":"225","format":"VHD","provisioningType":"THIN","poolId":"53","id":"6156","deviceId":"0","cacheMode":"NONE","hypervisorType":"XenServer","directDownload":"false","deployAsIs":"false","followRedirects":"true"},"dataStore":{"com.cloud.agent.api.to.NfsTO":{"_url":"nfs://10.69.105.254/var/NFS","_role":"Image"}},"vmName":"i-225-5468-VM","name":"backup-test-karasev-3_ROOT-5468_20240926135832","hypervisorType":"XenServer","id":"867","quiescevm":"false","physicalSize":"0","accountId":"225","followRedirects":"false"}},"executeInSequence":"true","options":{"snapshot.backup.to.secondary":"true","fullSnapshot":"false"},"options2":{},"wait":"21600","bypassHostMaintenance":"false"}}] } 2024-09-26 16:58:40,246 DEBUG [c.c.s.r.StorageSubsystemCommandHandlerBase] (DirectAgent-322:ctx-abb19c96) (logid:f365dd96) Executing command CopyCommand: [{"srcTO":{"org.apache.cloudstack.storage.to.SnapshotObjectTO":{"path":"e1971e58-5b6e-4f59-a256-154280cf4d16","volume":{"uuid":"da420806-0ff3-43b8-9ee0-4aa2adfd9b99","volumeType":"ROOT","dataStore":{"org.apache.cloudstack.storage.to.PrimaryDataStoreTO":{"uuid":"DM7100_LUN3","name":"cluster5-dm7100_lun3","id":53,"poolType":"PreSetup","host":"localhost","path":"/DM7100_LUN3","port":0,"url":"PreSetup://localhost/DM7100_LUN3/?ROLE=Primary&STOREUUID=DM7100_LUN3","isManaged":false}},"name":"ROOT-5468","size":21474836480,"path":"9e2a4d6e-3ab8-46f4-85cf-c7454fc9ea17","volumeId":6156,"vmName":"i-225-5468-VM","accountId":225,"format":"VHD","provisioningType":"THIN","poolId":53,"id":6156,"deviceId":0,"cacheMode":"NONE","hypervisorType":"XenServer","directDownload":false,"deployAsIs":false,"followRedirects":true},"parentSnapshotPath":"70424963-5933-4e73-b185-cfdc402175cd","dataStore":{"org.apache.cloudstack.storage.to.PrimaryDataStoreTO":{"uuid":"DM7100_LUN3","name":"cluster5-dm7100_lun3","id":53,"poolType":"PreSetup","host":"localhost","path":"/DM7100_LUN3","port":0,"url":"PreSetup://localhost/DM7100_LUN3/?ROLE=Primary&STOREUUID=DM7100_LUN3","isManaged":false}},"vmName":"i-225-5468-VM","name":"backup-test-karasev-3_ROOT-5468_20240926135832","hypervisorType":"XenServer","id":867,"quiescevm":false,"parents":["70424963-5933-4e73-b185-cfdc402175cd"],"physicalSize":0,"accountId":225,"followRedirects":false}},"destTO":{"org.apache.cloudstack.storage.to.SnapshotObjectTO":{"path":"snapshots/225/6156","volume":{"uuid":"da420806-0ff3-43b8-9ee0-4aa2adfd9b99","volumeType":"ROOT","dataStore":{"org.apache.cloudstack.storage.to.PrimaryDataStoreTO":{"uuid":"DM7100_LUN3","name":"cluster5-dm7100_lun3","id":53,"poolType":"PreSetup","host":"localhost","path":"/DM7100_LUN3","port":0,"url":"PreSetup://localhost/DM7100_LUN3/?ROLE=Primary&STOREUUID=DM7100_LUN3","isManaged":false}},"name":"ROOT-5468","size":21474836480,"path":"9e2a4d6e-3ab8-46f4-85cf-c7454fc9ea17","volumeId":6156,"vmName":"i-225-5468-VM","accountId":225,"format":"VHD","provisioningType":"THIN","poolId":53,"id":6156,"deviceId":0,"cacheMode":"NONE","hypervisorType":"XenServer","directDownload":false,"deployAsIs":false,"followRedirects":true},"dataStore":{"com.cloud.agent.api.to.NfsTO":{"_url":"nfs://10.69.105.254/var/NFS","_role":"Image"}},"vmName":"i-225-5468-VM","name":"backup-test-karasev-3_ROOT-5468_20240926135832","hypervisorType":"XenServer","id":867,"quiescevm":false,"physicalSize":0,"accountId":225,"followRedirects":false}},"executeInSequence":true,"options":{"snapshot.backup.to.secondary":"true","fullSnapshot":"false"},"options2":{},"wait":21600,"bypassHostMaintenance":false}]. 2024-09-26 16:58:41,264 DEBUG [c.c.h.x.r.XenServerStorageProcessor] (DirectAgent-322:ctx-abb19c96) (logid:f365dd96) No file SR found for path: /var/cloud_mount/440acca8-622d-3dfb-825d-b3e62d7bba87/snapshots/225/6156 2024-09-26 16:58:41,264 DEBUG [c.c.h.x.r.XenServerStorageProcessor] (DirectAgent-322:ctx-abb19c96) (logid:f365dd96) Creating file SR for path [/var/cloud_mount/440acca8-622d-3dfb-825d-b3e62d7bba87/snapshots/225/6156] on host [4208d709-b553-47c0-aee2-7942d370b70c] 2024-09-26 16:58:41,495 DEBUG [c.c.a.ApiServlet] (qtp501107890-4824844:ctx-107e93ec) (logid:1bc5742a) ===START=== 172.16.11.61 -- GET jobId=f365dd96-c018-454a-874c-37dce9396d96&command=queryAsyncJobResult&response=json 2024-09-26 16:58:41,518 DEBUG [c.c.a.ApiServlet] (qtp501107890-4824844:ctx-107e93ec ctx-f58fcdf2) (logid:1bc5742a) ===END=== 172.16.11.61 -- GET jobId=f365dd96-c018-454a-874c-37dce9396d96&command=queryAsyncJobResult&response=json 2024-09-26 16:58:41,520 DEBUG [c.c.h.x.r.XenServerStorageProcessor] (DirectAgent-322:ctx-abb19c96) (logid:f365dd96) Unpluging PBD [c9ae6161-7db8-f308-db93-4fa060c94f9c] of SR [f6204e02-6efe-3005-95a3-095f19b458ac] as it is not working properly. 2024-09-26 16:58:41,536 DEBUG [c.c.h.x.r.XenServerStorageProcessor] (DirectAgent-322:ctx-abb19c96) (logid:f365dd96) Forgetting SR [f6204e02-6efe-3005-95a3-095f19b458ac] as it is not working properly. 2024-09-26 16:58:41,617 DEBUG [c.c.h.x.r.XenServerStorageProcessor] (DirectAgent-322:ctx-abb19c96) (logid:f365dd96) Could not create file SR [/var/cloud_mount/440acca8-622d-3dfb-825d-b3e62d7bba87/snapshots/225/6156] on host [4208d709-b553-47c0-aee2-7942d370b70c]. 2024-09-26 16:58:41,617 DEBUG [c.c.h.x.r.XenServerStorageProcessor] (DirectAgent-322:ctx-abb19c96) (logid:f365dd96) Exception in backupsnapshot stage due to com.cloud.utils.exception.CloudRuntimeException: Could not retrieve an already used file SR for path [/var/cloud_mount/440acca8-622d-3dfb-825d-b3e62d7bba87/snapshots/225/6156] or create a new file SR on host [4208d709-b553-47c0-aee2-7942d370b70c] 2024-09-26 16:58:41,617 WARN [c.c.h.x.r.XenServerStorageProcessor] (DirectAgent-322:ctx-abb19c96) (logid:f365dd96) BackupSnapshot Failed due to Exception in backupsnapshot stage due to com.cloud.utils.exception.CloudRuntimeException: Could not retrieve an already used file SR for path [/var/cloud_mount/440acca8-622d-3dfb-825d-b3e62d7bba87/snapshots/225/6156] or create a new file SR on host [4208d709-b553-47c0-aee2-7942d370b70c] 2024-09-26 16:58:44,625 DEBUG [c.c.a.ApiServlet] (qtp501107890-4825573:ctx-e72bd43c) (logid:d5a19d4e) ===START=== 172.16.11.61 -- GET jobId=f365dd96-c018-454a-874c-37dce9396d96&command=queryAsyncJobResult&response=json 2024-09-26 16:58:44,651 DEBUG [c.c.a.ApiServlet] (qtp501107890-4825573:ctx-e72bd43c ctx-707eaf3e) (logid:d5a19d4e) ===END=== 172.16.11.61 -- GET jobId=f365dd96-c018-454a-874c-37dce9396d96&command=queryAsyncJobResult&response=json 2024-09-26 16:58:47,762 DEBUG [c.c.a.ApiServlet] (qtp501107890-4824845:ctx-517f2c19) (logid:c6e8be8b) ===START=== 172.16.11.61 -- GET jobId=f365dd96-c018-454a-874c-37dce9396d96&command=queryAsyncJobResult&response=json 2024-09-26 16:58:47,784 DEBUG [c.c.a.ApiServlet] (qtp501107890-4824845:ctx-517f2c19 ctx-9a6894c4) (logid:c6e8be8b) ===END=== 172.16.11.61 -- GET jobId=f365dd96-c018-454a-874c-37dce9396d96&command=queryAsyncJobResult&response=json 2024-09-26 16:58:47,863 DEBUG [c.c.a.m.DirectAgentAttache] (DirectAgent-322:ctx-abb19c96) (logid:f365dd96) Seq 77358-1940207014466653252: Response Received: 2024-09-26 16:58:47,864 DEBUG [c.c.a.t.Request] (DirectAgent-322:ctx-abb19c96) (logid:f365dd96) Seq 77358-1940207014466653252: Processing: { Ans: , MgmtId: 108597816265214, via: 77358(xcp-sr246-u1112.aisrp.local), Ver: v1, Flags: 110, [{"org.apache.cloudstack.storage.command.CopyCmdAnswer":{"result":"false","details":"BackupSnapshot Failed due to Exception in backupsnapshot stage due to com.cloud.utils.exception.CloudRuntimeException: Could not retrieve an already used file SR for path [/var/cloud_mount/440acca8-622d-3dfb-825d-b3e62d7bba87/snapshots/225/6156] or create a new file SR on host [4208d709-b553-47c0-aee2-7942d370b70c]","wait":"0","bypassHostMaintenance":"false"}}] } 2024-09-26 16:58:47,864 DEBUG [c.c.a.m.AgentAttache] (DirectAgent-322:ctx-abb19c96) (logid:f365dd96) Seq 77358-1940207014466653252: No more commands found 2024-09-26 16:58:47,864 DEBUG [c.c.a.t.Request] (Work-Job-Executor-148:ctx-ef42787d job-72454/job-72455 ctx-d1890501) (logid:f365dd96) Seq 77358-1940207014466653252: Received: { Ans: , MgmtId: 108597816265214, via: 77358(xcp-sr246-u1112.aisrp.local), Ver: v1, Flags: 110, { CopyCmdAnswer } } 2024-09-26 16:58:47,885 DEBUG [c.c.s.s.SnapshotManagerImpl] (Work-Job-Executor-148:ctx-ef42787d job-72454/job-72455 ctx-d1890501) (logid:f365dd96) Failed to create snapshotBackupSnapshot Failed due to Exception in backupsnapshot stage due to com.cloud.utils.exception.CloudRuntimeException: Could not retrieve an already used file SR for path [/var/cloud_mount/440acca8-622d-3dfb-825d-b3e62d7bba87/snapshots/225/6156] or create a new file SR on host [4208d709-b553-47c0-aee2-7942d370b70c] 2024-09-26 16:58:47,885 DEBUG [c.c.r.ResourceLimitManagerImpl] (Work-Job-Executor-148:ctx-ef42787d job-72454/job-72455 ctx-d1890501) (logid:f365dd96) Updating resource Type = snapshot count for Account = 225 Operation = decreasing Amount = 1 2024-09-26 16:58:47,894 DEBUG [c.c.r.ResourceLimitManagerImpl] (Work-Job-Executor-148:ctx-ef42787d job-72454/job-72455 ctx-d1890501) (logid:f365dd96) Updating resource Type = secondary_storage count for Account = 225 Operation = decreasing Amount = (20.00 GB) 21474836480 2024-09-26 16:58:47,901 ERROR [o.a.c.s.v.VolumeServiceImpl] (Work-Job-Executor-148:ctx-ef42787d job-72454/job-72455 ctx-d1890501) (logid:f365dd96) Take snapshot: 6156 failed 2024-09-26 16:58:47,902 ERROR [c.c.v.VmWorkJobHandlerProxy] (Work-Job-Executor-148:ctx-ef42787d job-72454/job-72455 ctx-d1890501) (logid:f365dd96) Invocation exception, caused by: com.cloud.utils.exception.CloudRuntimeException: BackupSnapshot Failed due to Exception in backupsnapshot stage due to com.cloud.utils.exception.CloudRuntimeException: Could not retrieve an already used file SR for path [/var/cloud_mount/440acca8-622d-3dfb-825d-b3e62d7bba87/snapshots/225/6156] or create a new file SR on host [4208d709-b553-47c0-aee2-7942d370b70c] 2024-09-26 16:58:47,902 INFO [c.c.v.VmWorkJobHandlerProxy] (Work-Job-Executor-148:ctx-ef42787d job-72454/job-72455 ctx-d1890501) (logid:f365dd96) Rethrow exception com.cloud.utils.exception.CloudRuntimeException: BackupSnapshot Failed due to Exception in backupsnapshot stage due to com.cloud.utils.exception.CloudRuntimeException: Could not retrieve an already used file SR for path [/var/cloud_mount/440acca8-622d-3dfb-825d-b3e62d7bba87/snapshots/225/6156] or create a new file SR on host [4208d709-b553-47c0-aee2-7942d370b70c] 2024-09-26 16:58:47,902 DEBUG [c.c.v.VmWorkJobDispatcher] (Work-Job-Executor-148:ctx-ef42787d job-72454/job-72455) (logid:f365dd96) Done with run of VM work job: com.cloud.vm.VmWorkTakeVolumeSnapshot for VM 5468, job origin: 72454 2024-09-26 16:58:47,902 ERROR [c.c.v.VmWorkJobDispatcher] (Work-Job-Executor-148:ctx-ef42787d job-72454/job-72455) (logid:f365dd96) Unable to complete AsyncJobVO: {id:72455, userId: 298, accountId: 225, instanceType: null, instanceId: null, cmd: com.cloud.vm.VmWorkTakeVolumeSnapshot, cmdInfo: rO0ABXNyACVjb20uY2xvdWQudm0uVm1Xb3JrVGFrZVZvbHVtZVNuYXBzaG90BL5gG4Li1c8CAAdaAAthc3luY0JhY2t1cFoACXF1aWVzY2VWbUwADGxvY2F0aW9uVHlwZXQAKUxjb20vY2xvdWQvc3RvcmFnZS9TbmFwc2hvdCRMb2NhdGlvblR5cGU7TAAIcG9saWN5SWR0ABBMamF2YS9sYW5nL0xvbmc7TAAKc25hcHNob3RJZHEAfgACTAAIdm9sdW1lSWRxAH4AAkwAB3pvbmVJZHN0ABBMamF2YS91dGlsL0xpc3Q7eHIAE2NvbS5jbG91ZC52bS5WbVdvcmufmbZW8CVnawIABEoACWFjY291bnRJZEoABnVzZXJJZEoABHZtSWRMAAtoYW5kbGVyTmFtZXQAEkxqYXZhL2xhbmcvU3RyaW5nO3hwAAAAAAAAAOEAAAAAAAABKgAAAAAAABVcdAAUVm9sdW1lQXBpU2VydmljZUltcGwAAHBzcgAOamF2YS5sYW5nLkxvbmc7i-SQzI8j3wIAAUoABXZhbHVleHIAEGphdmEubGFuZy5OdW1iZXKGrJUdC5TgiwIAAHhwAAAAAAAAAABzcQB-AAgAAAAAAAADY3NxAH4ACAAAAAAAABgMcA, cmdVersion: 0, status: IN_PROGRESS, processStatus: 0, resultCode: 0, result: null, initMsid: 108597816265214, completeMsid: null, lastUpdated: null, lastPolled: null, created: Thu Sep 26 16:58:32 MSK 2024, removed: null}, job origin:72454 2024-09-26 16:58:47,904 DEBUG [o.a.c.f.j.i.AsyncJobManagerImpl] (Work-Job-Executor-148:ctx-ef42787d job-72454/job-72455) (logid:f365dd96) Complete async job-72455, jobStatus: FAILED, resultCode: 0, result: rO0ABXNyAC9jb20uY2xvdWQudXRpbHMuZXhjZXB0aW9uLkNsb3VkUnVudGltZUV4Y2VwdGlvblZNT3AAAAACAwABSQALY3NFcnJvckNvZGV4cgAaamF2YS5sYW5nLlJ1bnRpbWVFeGNlcHRpb26eXwZHCjSD5QIAAHhyABNqYXZhLmxhbmcuRXhjZXB0aW9u0P0fPho7HMQCAAB4cgATamF2YS5sYW5nLlRocm93YWJsZdXGNSc5d7jLAwAETAAFY2F1c2V0ABVMamF2YS9sYW5nL1Rocm93YWJsZTtMAA1kZXRhaWxNZXNzYWdldAASTGphdmEvbGFuZy9TdHJpbmc7WwAKc3RhY2tUcmFjZXQAHltMamF2YS9sYW5nL1N0YWNrVHJhY2VFbGVtZW50O0wAFHN1cHByZXNzZWRFeGNlcHRpb25zdAAQTGphdmEvdXRpbC9MaXN0O3hwcQB-AAh0ATxCYWNrdXBTbmFwc2hvdCBGYWlsZWQgZHVlIHRvIEV4Y2VwdGlvbiBpbiBiYWNrdXBzbmFwc2hvdCBzdGFnZSBkdWUgdG8gY29tLmNsb3VkLnV0aWxzLmV4Y2VwdGlvbi5DbG91ZFJ1bnRpbWVFeGNlcHRpb246IENvdWxkIG5vdCByZXRyaWV2ZSBhbiBhbHJlYWR5IHVzZWQgZmlsZSBTUiBmb3IgcGF0aCBbL3Zhci9jbG91ZF9tb3VudC80NDBhY2NhOC02MjJkLTNkZmItODI1ZC1iM2U2MmQ3YmJhODcvc25hcHNob3RzLzIyNS82MTU2XSBvciBjcmVhdGUgYSBuZXcgZmlsZSBTUiBvbiBob3N0IFs0MjA4ZDcwOS1iNTUzLTQ3YzAtYWVlMi03OTQyZDM3MGI3MGNddXIAHltMamF2YS5sYW5nLlN0YWNrVHJhY2VFbGVtZW50OwJGKjw8_SI5AgAAeHAAAAAtc3IAG2phdmEubGFuZy5TdGFja1RyYWNlRWxlbWVudGEJxZomNt2FAgAIQgAGZm9ybWF0SQAKbGluZU51bWJlckwAD2NsYXNzTG9hZGVyTmFtZXEAfgAFTAAOZGVjbGFyaW5nQ2xhc3NxAH4ABUwACGZpbGVOYW1lcQB-AAVMAAptZXRob2ROYW1lcQB-AAVMAAptb2R1bGVOYW1lcQB-AAVMAA1tb2R1bGVWZXJzaW9ucQB-AAV4cAEAAAF-dAADYXBwdAA6b3JnLmFwYWNoZS5jbG91ZHN0YWNrLnN0b3JhZ2Uuc25hcHNob3QuU25hcHNob3RTZXJ2aWNlSW1wbHQAGFNuYXBzaG90U2VydmljZUltcGwuamF2YXQADmJhY2t1cFNuYXBzaG90cHBzcQB-AAwBAAAAwXEAfgAOdAA-b3JnLmFwYWNoZS5jbG91ZHN0YWNrLnN0b3JhZ2Uuc25hcHNob3QuRGVmYXVsdFNuYXBzaG90U3RyYXRlZ3l0ABxEZWZhdWx0U25hcHNob3RTdHJhdGVneS5qYXZhcQB-ABFwcHNxAH4ADAEAAAWfcQB-AA50AC5jb20uY2xvdWQuc3RvcmFnZS5zbmFwc2hvdC5TbmFwc2hvdE1hbmFnZXJJbXBsdAAYU25hcHNob3RNYW5hZ2VySW1wbC5qYXZhdAAZYmFja3VwU25hcHNob3RUb1NlY29uZGFyeXBwc3EAfgAMAQAABWxxAH4ADnEAfgAWcQB-ABd0AAx0YWtlU25hcHNob3RwcHNxAH4ADAD_____cHQAMGpkay5pbnRlcm5hbC5yZWZsZWN0LkdlbmVyYXRlZE1ldGhvZEFjY2Vzc29yMTk1N3B0AAZpbnZva2VwcHNxAH4ADAIAAAArcHQAMWpkay5pbnRlcm5hbC5yZWZsZWN0LkRlbGVnYXRpbmdNZXRob2RBY2Nlc3NvckltcGx0ACFEZWxlZ2F0aW5nTWV0aG9kQWNjZXNzb3JJbXBsLmphdmFxAH4AHXQACWphdmEuYmFzZXQABzExLjAuMjRzcQB-AAwCAAACNnB0ABhqYXZhLmxhbmcucmVmbGVjdC5NZXRob2R0AAtNZXRob2QuamF2YXEAfgAdcQB-ACFxAH4AInNxAH4ADAEAAAFYcQB-AA50AChvcmcuc3ByaW5nZnJhbWV3b3JrLmFvcC5zdXBwb3J0LkFvcFV0aWxzdAANQW9wVXRpbHMuamF2YXQAHmludm9rZUpvaW5wb2ludFVzaW5nUmVmbGVjdGlvbnBwc3EAfgAMAQAAAMZxAH4ADnQAPG9yZy5zcHJpbmdmcmFtZXdvcmsuYW9wLmZyYW1ld29yay5SZWZsZWN0aXZlTWV0aG9kSW52b2NhdGlvbnQAH1JlZmxlY3RpdmVNZXRob2RJbnZvY2F0aW9uLmphdmF0AA9pbnZva2VKb2lucG9pbnRwcHNxAH4ADAEAAACjcQB-AA5xAH4AK3EAfgAsdAAHcHJvY2VlZHBwc3EAfgAMAQAAAGFxAH4ADnQAP29yZy5zcHJpbmdmcmFtZXdvcmsuYW9wLmludGVyY2VwdG9yLkV4cG9zZUludm9jYXRpb25JbnRlcmNlcHRvcnQAIEV4cG9zZUludm9jYXRpb25JbnRlcmNlcHRvci5qYXZhcQB-AB1wcHNxAH4ADAEAAAC6cQB-AA5xAH4AK3EAfgAscQB-AC9wcHNxAH4ADAEAAADXcQB-AA50ADRvcmcuc3ByaW5nZnJhbWV3b3JrLmFvcC5mcmFtZXdvcmsuSmRrRHluYW1pY0FvcFByb3h5dAAXSmRrRHluYW1pY0FvcFByb3h5LmphdmFxAH4AHXBwc3EAfgAMAP____9wdAAXY29tLnN1bi5wcm94eS4kUHJveHkyMzhwcQB-ABpwcHNxAH4ADAEAAArYcQB-AA50ADZvcmcuYXBhY2hlLmNsb3Vkc3RhY2suc3RvcmFnZS52b2x1bWUuVm9sdW1lU2VydmljZUltcGx0ABZWb2x1bWVTZXJ2aWNlSW1wbC5qYXZhcQB-ABpwcHNxAH4ADAEAAA68cQB-AA50ACZjb20uY2xvdWQuc3RvcmFnZS5Wb2x1bWVBcGlTZXJ2aWNlSW1wbHQAGVZvbHVtZUFwaVNlcnZpY2VJbXBsLmphdmF0AB1vcmNoZXN0cmF0ZVRha2VWb2x1bWVTbmFwc2hvdHBwc3EAfgAMAQAAE7lxAH4ADnEAfgA9cQB-AD5xAH4AP3Bwc3EAfgAMAP____9wdAAwamRrLmludGVybmFsLnJlZmxlY3QuR2VuZXJhdGVkTWV0aG9kQWNjZXNzb3IxOTU2cHEAfgAdcHBzcQB-AAwCAAAAK3BxAH4AH3EAfgAgcQB-AB1xAH4AIXEAfgAic3EAfgAMAgAAAjZwcQB-ACRxAH4AJXEAfgAdcQB-ACFxAH4AInNxAH4ADAEAAABpcQB-AA50ACJjb20uY2xvdWQudm0uVm1Xb3JrSm9iSGFuZGxlclByb3h5dAAaVm1Xb3JrSm9iSGFuZGxlclByb3h5LmphdmF0AA9oYW5kbGVWbVdvcmtKb2JwcHNxAH4ADAEAABPHcQB-AA5xAH4APXEAfgA-cQB-AEhwcHNxAH4ADAD_____cHQAMGpkay5pbnRlcm5hbC5yZWZsZWN0LkdlbmVyYXRlZE1ldGhvZEFjY2Vzc29yMTAzNnBxAH4AHXBwc3EAfgAMAgAAACtwcQB-AB9xAH4AIHEAfgAdcQB-ACFxAH4AInNxAH4ADAIAAAI2cHEAfgAkcQB-ACVxAH4AHXEAfgAhcQB-ACJzcQB-AAwBAAABWHEAfgAOcQB-ACdxAH4AKHEAfgApcHBzcQB-AAwBAAAAxnEAfgAOcQB-ACtxAH4ALHEAfgAtcHBzcQB-AAwBAAAAo3EAfgAOcQB-ACtxAH4ALHEAfgAvcHBzcQB-AAwBAAAAYXEAfgAOcQB-ADFxAH4AMnEAfgAdcHBzcQB-AAwBAAAAunEAfgAOcQB-ACtxAH4ALHEAfgAvcHBzcQB-AAwBAAAA13EAfgAOcQB-ADVxAH4ANnEAfgAdcHBzcQB-AAwA_____3B0ABdjb20uc3VuLnByb3h5LiRQcm94eTI0N3BxAH4ASHBwc3EAfgAMAQAAAGZxAH4ADnQAIGNvbS5jbG91ZC52bS5WbVdvcmtKb2JEaXNwYXRjaGVydAAYVm1Xb3JrSm9iRGlzcGF0Y2hlci5qYXZhdAAGcnVuSm9icHBzcQB-AAwBAAACjnEAfgAOdAA_b3JnLmFwYWNoZS5jbG91ZHN0YWNrLmZyYW1ld29yay5qb2JzLmltcGwuQXN5bmNKb2JNYW5hZ2VySW1wbCQ1dAAYQXN5bmNKb2JNYW5hZ2VySW1wbC5qYXZhdAAMcnVuSW5Db250ZXh0cHBzcQB-AAwBAAAAMHEAfgAOdAA-b3JnLmFwYWNoZS5jbG91ZHN0YWNrLm1hbmFnZWQuY29udGV4dC5NYW5hZ2VkQ29udGV4dFJ1bm5hYmxlJDF0ABtNYW5hZ2VkQ29udGV4dFJ1bm5hYmxlLmphdmF0AANydW5wcHNxAH4ADAEAAAA3cQB-AA50AEJvcmcuYXBhY2hlLmNsb3Vkc3RhY2subWFuYWdlZC5jb250ZXh0LmltcGwuRGVmYXVsdE1hbmFnZWRDb250ZXh0JDF0ABpEZWZhdWx0TWFuYWdlZENvbnRleHQuamF2YXQABGNhbGxwcHNxAH4ADAEAAABmcQB-AA50AEBvcmcuYXBhY2hlLmNsb3Vkc3RhY2subWFuYWdlZC5jb250ZXh0LmltcGwuRGVmYXVsdE1hbmFnZWRDb250ZXh0cQB-AGR0AA9jYWxsV2l0aENvbnRleHRwcHNxAH4ADAEAAAA0cQB-AA5xAH4AZ3EAfgBkdAAOcnVuV2l0aENvbnRleHRwcHNxAH4ADAEAAAAtcQB-AA50ADxvcmcuYXBhY2hlLmNsb3Vkc3RhY2subWFuYWdlZC5jb250ZXh0Lk1hbmFnZWRDb250ZXh0UnVubmFibGVxAH4AYHEAfgBhcHBzcQB-AAwBAAACWnEAfgAOcQB-AFtxAH4AXHEAfgBhcHBzcQB-AAwCAAACA3B0AC5qYXZhLnV0aWwuY29uY3VycmVudC5FeGVjdXRvcnMkUnVubmFibGVBZGFwdGVydAAORXhlY3V0b3JzLmphdmFxAH4AZXEAfgAhcQB-ACJzcQB-AAwCAAABCHB0AB9qYXZhLnV0aWwuY29uY3VycmVudC5GdXR1cmVUYXNrdAAPRnV0dXJlVGFzay5qYXZhcQB-AGFxAH4AIXEAfgAic3EAfgAMAgAABGhwdAAnamF2YS51dGlsLmNvbmN1cnJlbnQuVGhyZWFkUG9vbEV4ZWN1dG9ydAAXVGhyZWFkUG9vbEV4ZWN1dG9yLmphdmF0AAlydW5Xb3JrZXJxAH4AIXEAfgAic3EAfgAMAgAAAnRwdAAuamF2YS51dGlsLmNvbmN1cnJlbnQuVGhyZWFkUG9vbEV4ZWN1dG9yJFdvcmtlcnEAfgB2cQB-AGFxAH4AIXEAfgAic3EAfgAMAgAAAz1wdAAQamF2YS5sYW5nLlRocmVhZHQAC1RocmVhZC5qYXZhcQB-AGFxAH4AIXEAfgAic3IAH2phdmEudXRpbC5Db2xsZWN0aW9ucyRFbXB0eUxpc3R6uBe0PKee3gIAAHhweAAAEJp3CAAAAAAAAAAAeA 2024-09-26 16:58:47,905 DEBUG [o.a.c.f.j.i.AsyncJobManagerImpl] (Work-Job-Executor-148:ctx-ef42787d job-72454/job-72455) (logid:f365dd96) Publish async job-72455 complete on message bus 2024-09-26 16:58:47,905 DEBUG [o.a.c.f.j.i.AsyncJobManagerImpl] (Work-Job-Executor-148:ctx-ef42787d job-72454/job-72455) (logid:f365dd96) Wake up jobs related to job-72455 2024-09-26 16:58:47,905 DEBUG [o.a.c.f.j.i.AsyncJobManagerImpl] (Work-Job-Executor-148:ctx-ef42787d job-72454/job-72455) (logid:f365dd96) Update db status for job-72455 2024-09-26 16:58:47,906 DEBUG [o.a.c.f.j.i.AsyncJobManagerImpl] (Work-Job-Executor-148:ctx-ef42787d job-72454/job-72455) (logid:f365dd96) Wake up jobs joined with job-72455 and disjoin all subjobs created from job- 72455 2024-09-26 16:58:47,914 DEBUG [o.a.c.f.j.i.AsyncJobManagerImpl] (Work-Job-Executor-148:ctx-ef42787d job-72454/job-72455) (logid:f365dd96) Done executing com.cloud.vm.VmWorkTakeVolumeSnapshot for job-72455 2024-09-26 16:58:47,916 INFO [o.a.c.f.j.i.AsyncJobMonitor] (Work-Job-Executor-148:ctx-ef42787d job-72454/job-72455) (logid:f365dd96) Remove job-72455 from job monitoring 2024-09-26 16:58:47,930 ERROR [o.a.c.a.c.u.s.CreateSnapshotCmd] (API-Job-Executor-58:ctx-c9d433a7 job-72454 ctx-829ed64e) (logid:f365dd96) Failed to create snapshot due to an internal error creating snapshot for volume da420806-0ff3-43b8-9ee0-4aa2adfd9b99 2024-09-26 16:58:47,935 DEBUG [o.a.c.f.j.i.AsyncJobManagerImpl] (API-Job-Executor-58:ctx-c9d433a7 job-72454) (logid:f365dd96) Complete async job-72454, jobStatus: FAILED, resultCode: 530, result: org.apache.cloudstack.api.response.ExceptionResponse/null/{"uuidList":[],"errorcode":"530","errortext":"Failed to create snapshot due to an internal error creating snapshot for volume da420806-0ff3-43b8-9ee0-4aa2adfd9b99"} 2024-09-26 16:58:47,936 DEBUG [o.a.c.f.j.i.AsyncJobManagerImpl] (API-Job-Executor-58:ctx-c9d433a7 job-72454) (logid:f365dd96) Publish async job-72454 complete on message bus 2024-09-26 16:58:47,936 DEBUG [o.a.c.f.j.i.AsyncJobManagerImpl] (API-Job-Executor-58:ctx-c9d433a7 job-72454) (logid:f365dd96) Wake up jobs related to job-72454 2024-09-26 16:58:47,936 DEBUG [o.a.c.f.j.i.AsyncJobManagerImpl] (API-Job-Executor-58:ctx-c9d433a7 job-72454) (logid:f365dd96) Update db status for job-72454 2024-09-26 16:58:47,937 DEBUG [o.a.c.f.j.i.AsyncJobManagerImpl] (API-Job-Executor-58:ctx-c9d433a7 job-72454) (logid:f365dd96) Wake up jobs joined with job-72454 and disjoin all subjobs created from job- 72454 2024-09-26 16:58:47,941 DEBUG [o.a.c.f.j.i.AsyncJobManagerImpl] (API-Job-Executor-58:ctx-c9d433a7 job-72454) (logid:f365dd96) Done executing org.apache.cloudstack.api.command.user.snapshot.CreateSnapshotCmd for job-72454 2024-09-26 16:58:47,941 INFO [o.a.c.f.j.i.AsyncJobMonitor] (API-Job-Executor-58:ctx-c9d433a7 job-72454) (logid:f365dd96) Remove job-72454 from job monitoring 2024-09-26 16:58:50,891 DEBUG [c.c.a.ApiServlet] (qtp501107890-4824742:ctx-04767efc) (logid:00d3b60a) ===START=== 172.16.11.61 -- GET jobId=f365dd96-c018-454a-874c-37dce9396d96&command=queryAsyncJobResult&response=json 2024-09-26 16:58:50,913 DEBUG [c.c.a.ApiServlet] (qtp501107890-4824742:ctx-04767efc ctx-ff2eca28) (logid:00d3b60a) ===END=== 172.16.11.61 -- GET jobId=f365dd96-c018-454a-874c-37dce9396d96&command=queryAsyncJobResult&response=json Regarding the first issue of all/most snapshots going to one store, it would be helpful knowing the following: Are the 2 secondary stores of same or similar capacity? What is the value set for the global setting: image.store.allocation.algorithm ? @Pearl1594 1. the same capacity. Second is a copy of first 2. firstfitleastconsumed @Pearl1594 if I set first secstore Read-Only, snapshots are copied to the second secstore. But after I unset Read-Only all snapshots are copied to the first secstore again when creating the volume snaphots, were they are initiated at the same time i.e, either via API or through recurring snapshot schedule. If that was the case, this could possible happen because the stats collector would not have run to has updated the stats of the secondary store(s) to make the decision that store1 has capacity < than store2. Marking store1 to read only excludes it from the list of stores to which snapshot would be backed up to - expected behaviour.
gharchive/issue
2024-09-26T14:36:23
2025-04-01T04:55:57.525289
{ "authors": [ "Pearl1594", "top-secrett" ], "repo": "apache/cloudstack", "url": "https://github.com/apache/cloudstack/issues/9734", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2694036851
Projects listing breaks after deleting a user added to a project. ISSUE TYPE Bug Report COMPONENT NAME UI,API CLOUDSTACK VERSION 4.19.1.2 CONFIGURATION Projects SUMMARY Once you deleted a user in 'ROOT admin' role account, the projects can't be listed using UI/API. The deleted user was added to one of the projects. STEPS TO REPRODUCE 1. Add a user to the 'ROOT admin' Account. 2. Create a project as another account/user in the same domain. 3. Add the user from step 1 to this project as Project Admin. 4. As an 'admin' user delete the above user. EXPECTED RESULTS Deleting a user shouldn't break project access. ACTUAL RESULTS Deleting a user breaks projects access. Undo the user deletion. ( May not be safe for production) Workaround example: delete from project_account where user_id=<user_id> OR ;update user set state='disabled',removed=NULL where id=<id>; We could make the deleteUser API remove the user from all projects before deleting it (as we already do with accounts). To normalize environments that are affected, we could insert a query into the upgrade script. Something like: delete from project_account where user_id in (select id from user where removed); fixed in #10008
gharchive/issue
2024-11-26T10:13:57
2025-04-01T04:55:57.533194
{ "authors": [ "DaanHoogland", "rajujith", "winterhazel" ], "repo": "apache/cloudstack", "url": "https://github.com/apache/cloudstack/issues/9974", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
209752629
CLOUDSTACK-9727 Password reset discrepancy in RVR when one of the Rou… …ter is not in Running state. @bvbharatk can you format the code ? @ustcweizhou Reformatted the code. thanks. ACS CI BVT Run Sumarry: Build Number 394 Hypervisor xenserver NetworkType Advanced Passed=104 Failed=1 Skipped=7 Link to logs Folder (search by build_no): https://www.dropbox.com/sh/yj3wnzbceo9uef2/AAB6u-Iap-xztdm6jHX9SjPja?dl=0 Failed tests: test_routers_network_ops.py test_01_RVR_Network_FW_PF_SSH_default_routes_egress_true Failed Skipped tests: test_01_test_vm_volume_snapshot test_vm_nic_adapter_vmxnet3 test_static_role_account_acls test_11_ss_nfs_version_on_ssvm test_nested_virtualization_vmware test_3d_gpu_support test_deploy_vgpu_enabled_vm Passed test suits: test_deploy_vm_with_userdata.py test_affinity_groups_projects.py test_portable_publicip.py test_over_provisioning.py test_global_settings.py test_scale_vm.py test_service_offerings.py test_routers_iptables_default_policy.py test_loadbalance.py test_routers.py test_reset_vm_on_reboot.py test_deploy_vms_with_varied_deploymentplanners.py test_network.py test_router_dns.py test_non_contigiousvlan.py test_login.py test_deploy_vm_iso.py test_list_ids_parameter.py test_public_ip_range.py test_multipleips_per_nic.py test_regions.py test_affinity_groups.py test_network_acl.py test_pvlan.py test_volumes.py test_nic.py test_deploy_vm_root_resize.py test_resource_detail.py test_secondary_storage.py test_vm_life_cycle.py test_disk_offerings.py ACS CI BVT Run Sumarry: Build Number 400 Hypervisor xenserver NetworkType Advanced Passed=105 Failed=0 Skipped=7 Link to logs Folder (search by build_no): https://www.dropbox.com/sh/yj3wnzbceo9uef2/AAB6u-Iap-xztdm6jHX9SjPja?dl=0 Failed tests: Skipped tests: test_01_test_vm_volume_snapshot test_vm_nic_adapter_vmxnet3 test_static_role_account_acls test_11_ss_nfs_version_on_ssvm test_nested_virtualization_vmware test_3d_gpu_support test_deploy_vgpu_enabled_vm Passed test suits: test_deploy_vm_with_userdata.py test_affinity_groups_projects.py test_portable_publicip.py test_over_provisioning.py test_global_settings.py test_scale_vm.py test_service_offerings.py test_routers_iptables_default_policy.py test_loadbalance.py test_routers.py test_reset_vm_on_reboot.py test_deploy_vms_with_varied_deploymentplanners.py test_network.py test_router_dns.py test_non_contigiousvlan.py test_login.py test_deploy_vm_iso.py test_list_ids_parameter.py test_public_ip_range.py test_multipleips_per_nic.py test_regions.py test_affinity_groups.py test_network_acl.py test_pvlan.py test_volumes.py test_nic.py test_deploy_vm_root_resize.py test_resource_detail.py test_secondary_storage.py test_vm_life_cycle.py test_routers_network_ops.py test_disk_offerings.py Code changes LGTM @bvbharatk I think the new password should be saved to one of the running routers, not all running routers. I disagree with change line 797. ACS CI BVT Run Sumarry: Build Number 438 Hypervisor xenserver NetworkType Advanced Passed=104 Failed=1 Skipped=7 Link to logs Folder (search by build_no): https://www.dropbox.com/sh/yj3wnzbceo9uef2/AAB6u-Iap-xztdm6jHX9SjPja?dl=0 Failed tests: test_routers_network_ops.py test_03_RVR_Network_check_router_state Failed Skipped tests: test_01_test_vm_volume_snapshot test_vm_nic_adapter_vmxnet3 test_static_role_account_acls test_11_ss_nfs_version_on_ssvm test_nested_virtualization_vmware test_3d_gpu_support test_deploy_vgpu_enabled_vm Passed test suits: test_deploy_vm_with_userdata.py test_affinity_groups_projects.py test_portable_publicip.py test_over_provisioning.py test_global_settings.py test_scale_vm.py test_service_offerings.py test_routers_iptables_default_policy.py test_loadbalance.py test_routers.py test_reset_vm_on_reboot.py test_deploy_vms_with_varied_deploymentplanners.py test_network.py test_router_dns.py test_non_contigiousvlan.py test_login.py test_deploy_vm_iso.py test_list_ids_parameter.py test_public_ip_range.py test_multipleips_per_nic.py test_regions.py test_affinity_groups.py test_network_acl.py test_pvlan.py test_volumes.py test_nic.py test_deploy_vm_root_resize.py test_resource_detail.py test_secondary_storage.py test_vm_life_cycle.py test_disk_offerings.py @ustcweizhou Hi we are not saving the password to the router. we are saving the password in the VM details. When the VM starts we send the password to the router. Password reset cannot be called when the VM is running. We save the password in the VM details if one of the routers is not running in case of rvr network. Even if the master went down and Backup came up at the time of userVM start, we will serve the correct password as we are saving it. @bvbharatk Yes, you might know only the vm has sshkey attached will have the password in vm details. If the vm does not have ssh keypair, then the vm password will not saved into user_vm_details. Actually the vm password should be synced between master and backup. Saving to only one of them or saving to both of them are not working fine. for example, if we save password in master, but not save it in backup. Once the master is down, then vm cannot get password from backup vr. another example is, if we save password on both of master and backup, if vm get the password from master and reset it, once the master is down (or master->backup switch) and we reboot vm later, the vm will get the old password from backup again. @ustcweizhou We are saving the password in the user_vm_details explicitly. We are not checking if ssh key pair is set for this vm or not. I agree that ideally we should sync the password between the master and backup, For any kind of sync to work we need to know if the password was read from one of the VRs and In cases when one of the Vr is Stopped we will have to clear the password from db when it is read from the other one. These type of changes add complexity to the simple task of setting a password. The next best thing is to make sure we save the same password in both the routers. This will fix will at least solve the problem of sending the correct password even if the master and backup change state before the VM starts. Yes like you pointed out this will lead to the problem that the user might receive the old password when he stop starts the VM, In this case the user will get a notification in the UI that his password will be changed. So he at least knows what the password is and so he can log into the VM. Hi, Testing this PR end to end may not be possible now as there is some issue with the password server in case of RVRs.(CLOUDSTACK-9385). I have also encounter a similar problem when testing this manually, I saw that the password serve r is running on a different ip than expected and so the password script in the User VM failed to retrieve the password from the password server. However I was able to verify the intended behavior because of this PR, i.e. saving the password to the RVRs was successful. @bvbharatk for this issue, I have some ideas. (1) If nothing changes, the password will be applied to the first router. If the BACKUP VR is the first, then vm will not get new password. (2) If this PR is merged, then password will be applied to both router. In case of master<-> backup switch, the password will be stored to MASTER VR (=old BACKUP), then vm password will be reset after reboot/restart. (3) If password is only be applied to MASTER router, then there will be no password stored on BACKUP router. If vm password is reset (and stored to MASTER router), but vm is started after a MASTER<->BACKUP switch, then vm will not get new password. Considering these three options, I think option 3 is best. What do you think ? @ustcweizhou Even if we store the password only in master option 2 can happen. Suppose the routers change the state again before the user starts the VM, the current master will become backup and will have the password stored on it. At some point of time when the state change happens and the VM restarts the password on the user VM will be reset again, without the knowledge of the user. I think there is a design problem here and addressing this will need some fundamental changes, so rather than trying to force fit it, i think it is better if we leave it at this state and then document this behavior. The workaround for now would be to remove the password script from he userVM once he successfully logs in. Thanks, Bharat. @bvbharatk to be honest, we used option 2 on our production before and received some complain all vms' password are reset after reboot. Then we used option 3. if customer cannot get new password, they can stop vm and reset vm password again (this happens not often, because normally customer start vm sooner after vm password reset so there is no master<->backup switch in this period). Comparing these above 3 options, I do think option 3 has less impact and is more wise. VMpassword should be stored to only one place (yes I mean MASTER vr) if there is no sync. We are using password sync between master and backup in our fork now. However, I have no time to create a PR. @ustcweizhou Made changes as per your suggestion, Will send the password only to master router now. However we may still fail if router changes state while the password reset operation is in progress. ACS CI BVT Run Sumarry: Build Number 718 Hypervisor xenserver NetworkType Advanced Passed=105 Failed=8 Skipped=12 Link to logs Folder (search by build_no): https://www.dropbox.com/sh/r2si930m8xxzavs/AAAzNrnoF1fC3auFrvsKo_8-a?dl=0 Failed tests: test_loadbalance.py test_01_create_lb_rule_src_nat Failed test_01_create_lb_rule_src_nat Failing since 2 runs test_02_create_lb_rule_non_nat Failed test_assign_and_removal_lb Failed test_deploy_vm_iso.py test_deploy_vm_from_iso Failed test_volumes.py test_06_download_detached_volume Failed test_routers_network_ops.py test_01_RVR_Network_FW_PF_SSH_default_routes_egress_true Failed test_02_RVR_Network_FW_PF_SSH_default_routes_egress_false Failed Skipped tests: test_vm_nic_adapter_vmxnet3 test_01_verify_libvirt test_02_verify_libvirt_after_restart test_03_verify_libvirt_attach_disk test_04_verify_guest_lspci test_05_change_vm_ostype_restart test_06_verify_guest_lspci_again test_static_role_account_acls test_11_ss_nfs_version_on_ssvm test_nested_virtualization_vmware test_3d_gpu_support test_deploy_vgpu_enabled_vm Passed test suits: test_deploy_vm_with_userdata.py test_affinity_groups_projects.py test_portable_publicip.py test_vm_snapshots.py test_over_provisioning.py test_global_settings.py test_scale_vm.py test_service_offerings.py test_routers_iptables_default_policy.py test_routers.py test_reset_vm_on_reboot.py test_deploy_vms_with_varied_deploymentplanners.py test_network.py test_router_dns.py test_non_contigiousvlan.py test_login.py test_list_ids_parameter.py test_public_ip_range.py test_multipleips_per_nic.py test_metrics_api.py test_regions.py test_affinity_groups.py test_network_acl.py test_pvlan.py test_nic.py test_deploy_vm_root_resize.py test_resource_detail.py test_secondary_storage.py test_vm_life_cycle.py test_disk_offerings.py ACS CI BVT Run Sumarry: Build Number 736 Hypervisor xenserver NetworkType Advanced Passed=110 Failed=2 Skipped=12 Link to logs Folder (search by build_no): https://www.dropbox.com/sh/r2si930m8xxzavs/AAAzNrnoF1fC3auFrvsKo_8-a?dl=0 Failed tests: test_volumes.py test_06_download_detached_volume Failed test_routers_network_ops.py test_01_RVR_Network_FW_PF_SSH_default_routes_egress_true Failing since 2 runs Skipped tests: test_vm_nic_adapter_vmxnet3 test_01_verify_libvirt test_02_verify_libvirt_after_restart test_03_verify_libvirt_attach_disk test_04_verify_guest_lspci test_05_change_vm_ostype_restart test_06_verify_guest_lspci_again test_static_role_account_acls test_11_ss_nfs_version_on_ssvm test_nested_virtualization_vmware test_3d_gpu_support test_deploy_vgpu_enabled_vm Passed test suits: test_deploy_vm_with_userdata.py test_affinity_groups_projects.py test_portable_publicip.py test_vm_snapshots.py test_over_provisioning.py test_global_settings.py test_scale_vm.py test_service_offerings.py test_routers_iptables_default_policy.py test_loadbalance.py test_routers.py test_reset_vm_on_reboot.py test_deploy_vms_with_varied_deploymentplanners.py test_network.py test_router_dns.py test_non_contigiousvlan.py test_login.py test_deploy_vm_iso.py test_list_ids_parameter.py test_public_ip_range.py test_multipleips_per_nic.py test_metrics_api.py test_regions.py test_affinity_groups.py test_network_acl.py test_pvlan.py test_nic.py test_deploy_vm_root_resize.py test_resource_detail.py test_secondary_storage.py test_vm_life_cycle.py test_disk_offerings.py @ustcweizhou Hi, I have made the suggested changes, Can you please review?. @ustcweizhou Can you please review this, I have made the suggested changes. @blueorangutan package @rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. Packaging result: ✔centos6 ✔centos7 ✔debian. JID-1295 @weizhouapache is this still needed in your opinion? @weizhouapache is this fixed in recent version or is this still an issue? @rhtyd I think it has been fixed by #3903 Thanks @weizhouapache closing on your remark @bvbharatk pl raise a new PR if you still find the issue
gharchive/pull-request
2017-02-23T12:50:30
2025-04-01T04:55:57.591771
{ "authors": [ "DaanHoogland", "blueorangutan", "bvbharatk", "cloudmonger", "jayapalu", "rhtyd", "ustcweizhou", "weizhouapache" ], "repo": "apache/cloudstack", "url": "https://github.com/apache/cloudstack/pull/1965", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
259166668
CLOUDSTACK-10085 : Upload volume from local fails when global config max.account.seconday.storage is set to -1 ISSUE Upload volume from local fails when global config max.account.seconday.storage is set to -1 STEPS TO REPRODUCE Download any volume on your local machine. Change value of global configuration max.account.seconday.storage to -1. Restart Management server. Try to upload a volume from UI using "Upload from Local" option. Upload will fail. FIX Implementation of -1 option for unlimited secondary storage usage, which is absent in current code. Code LGTM @GabrielBrascher : I have rebase this PR against latest master and all checks are passing as well. Can you please merge this PR ? Thanks @niteshsarda but I am not the person you need now ;) LGTM we have the same fix in production for more than one year.
gharchive/pull-request
2017-09-20T13:38:54
2025-04-01T04:55:57.596022
{ "authors": [ "GabrielBrascher", "mrunalinikankariya", "niteshsarda", "ustcweizhou" ], "repo": "apache/cloudstack", "url": "https://github.com/apache/cloudstack/pull/2270", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
78160179
Cloudstack-8301 PR for configuring local storage for system VMs at zone level. FS @ https://cwiki.apache.org/confluence/display/CLOUDSTACK/Enable+configuring+local+storage+use+for+system+VMs+at+zone+level Jira @ https://issues.apache.org/jira/browse/CLOUDSTACK-8301 @rags22489664 contibuted UI changes (https://github.com/apache/cloudstack/pull/259) @nitt10prashant contributed tests (https://github.com/apache/cloudstack/pull/253) Thanks @koushik-das for your reply and the fix, can you share any test results? LGTM. List service offerings for systemvms and verify there should be two ... === TestName: test_01_list_system_offerngs_1_consoleproxy | Status : SUCCESS === ok List service offerings for systemvms and verify there should be two ... === TestName: test_01_list_system_offerngs_2_secondarystoragevm | Status : SUCCESS === ok List service offerings for systemvms and verify there should be two ... === TestName: test_01_list_system_offerngs_3_domainrouter | Status : SUCCESS === ok List service offerings for systemvms and verify there should be two ... === TestName: test_01_list_system_offerngs_4_internalloadbalancervm | Status : SUCCESS === ok Check if system vms are honouring zone level setting ... === TestName: test_02_system_vm_storage_1_consoleproxy | Status : SUCCESS === ok Check if system vms are honouring zone level setting ... === TestName: test_02_system_vm_storage_2_secondarystoragevm | Status : SUCCESS === ok update global setting with system offering and check if it is being ... === TestName: test_03_custom_so_1_consoleproxy | Status : SUCCESS === ok update global setting with system offering and check if it is being ... === TestName: test_03_custom_so_2_secondarystoragevm | Status : SUCCESS === ok Check if router vm is honouring zone level setting ... === TestName: test_04_router_vms | Status : SUCCESS === ok Ran 9 tests in 3201.200s OK
gharchive/pull-request
2015-05-19T16:33:31
2025-04-01T04:55:57.602681
{ "authors": [ "bhaisaab", "koushik-das", "nitt10prashant" ], "repo": "apache/cloudstack", "url": "https://github.com/apache/cloudstack/pull/263", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2166949548
Log4j2 refactor cloud engine orchestration module Description With the new version of Log4j there is some space for improving log readability using the new features. This PR refactors the logs from the cloud-orchestration module. Types of changes [ ] Breaking change (fix or feature that would cause existing functionality to change) [ ] New feature (non-breaking change which adds functionality) [ ] Bug fix (non-breaking change which fixes an issue) [ ] Enhancement (improves an existing feature and functionality) [X] Cleanup (Code refactoring and cleanup, that may add test cases) [ ] build/CI Feature/Enhancement Scale or Bug Severity Feature/Enhancement Scale [ ] Major [X] Minor Bug Severity [ ] BLOCKER [ ] Critical [ ] Major [ ] Minor [ ] Trivial How Has This Been Tested? I have done some operations on cloudstack to induce the generation of the logs No problems or changes to the log output were found Codecov Report Attention: Patch coverage is 27.91328% with 266 lines in your changes are missing coverage. Please review. Project coverage is 30.80%. Comparing base (a5508ac) to head (f323822). Report is 562 commits behind head on main. :exclamation: Current head f323822 differs from pull request most recent head b677fd4. Consider uploading reports for the commit b677fd4 to get more accurate results Files Patch % Lines ...cloud/agent/manager/ClusteredAgentManagerImpl.java 7.89% 105 Missing :warning: ...java/com/cloud/agent/manager/AgentManagerImpl.java 21.34% 70 Missing :warning: ...stack/engine/orchestration/VolumeOrchestrator.java 30.35% 39 Missing :warning: ...ain/java/com/cloud/agent/manager/AgentAttache.java 31.81% 14 Missing and 1 partial :warning: ...va/com/cloud/agent/manager/DirectAgentAttache.java 47.61% 11 Missing :warning: ...com/cloud/agent/manager/ClusteredAgentAttache.java 0.00% 9 Missing :warning: ...tack/engine/orchestration/NetworkOrchestrator.java 83.33% 5 Missing :warning: ...ack/engine/orchestration/DataMigrationUtility.java 0.00% 3 Missing :warning: ...tack/engine/orchestration/StorageOrchestrator.java 0.00% 3 Missing :warning: ...a/com/cloud/agent/manager/SynchronousListener.java 33.33% 1 Missing and 1 partial :warning: ... and 3 more Additional details and impacted files @@ Coverage Diff @@ ## main #8742 +/- ## ============================================= + Coverage 13.17% 30.80% +17.63% - Complexity 9214 33528 +24314 ============================================= Files 2725 5397 +2672 Lines 258235 379186 +120951 Branches 40249 55175 +14926 ============================================= + Hits 34013 116825 +82812 - Misses 219913 246935 +27022 - Partials 4309 15426 +11117 Flag Coverage Δ simulator-marvin-tests 24.27% <27.37%> (?) uitests 4.34% <ø> (?) unit-tests 16.89% <1.08%> (?) Flags with carried forward coverage won't be shown. Click here to find out more. :umbrella: View full report in Codecov by Sentry. :loudspeaker: Have feedback on the report? Share it here. @blueorangutan package @DaanHoogland a [SL] Jenkins job has been kicked to build packages. It will be bundled with KVM, XenServer and VMware SystemVM templates. I'll keep you posted as I make progress. CLGTM, hope I didn't miss anything 😅 that you'll have to test ;) the build fails @KlausDornsbach in engine according to https://github.com/apache/cloudstack/actions/runs/8757275560/job/24037204263?pr=8742#step:7:17606 , afraid you'll have to take another look :( Packaging result [SF]: ✖️ el7 ✖️ el8 ✖️ el9 ✖️ debian ✖️ suse15. SL-JID 9356 @blueorangutan package @DaanHoogland a [SL] Jenkins job has been kicked to build packages. It will be bundled with KVM, XenServer and VMware SystemVM templates. I'll keep you posted as I make progress. Packaging result [SF]: ✔️ el7 ✔️ el8 ✔️ el9 ✔️ debian ✔️ suse15. SL-JID 9370 @blueorangutan test @DaanHoogland a [SL] Trillian-Jenkins test job (centos7 mgmt + kvm-centos7) has been kicked to run smoke tests
gharchive/pull-request
2024-03-04T14:13:09
2025-04-01T04:55:57.632759
{ "authors": [ "DaanHoogland", "KlausDornsbach", "blueorangutan", "codecov-commenter" ], "repo": "apache/cloudstack", "url": "https://github.com/apache/cloudstack/pull/8742", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
770983888
Miscellaneous (cherry-picked) improvements Various improvements highlighted by the IDE. The improvements are cherry-picked where it makes sense. Each improvement category is grouped as a single commit. LGTM @arturobernalg I kept them separate so its remove/rebase a specific change. Will wait for #123 to avoid merge conflicts/duplicate work. @singhbaljit, can you rebase and fix conflicts here? We can then merge this in. @darkma773r its been rebased. @darkma773r its been rebased. Oops. Guess I missed that. Oops. Guess I missed that.
gharchive/pull-request
2020-12-18T15:42:48
2025-04-01T04:55:57.635848
{ "authors": [ "arturobernalg", "darkma773r", "singhbaljit" ], "repo": "apache/commons-geometry", "url": "https://github.com/apache/commons-geometry/pull/122", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
812538280
true && false String constant create true && false String constant in order to re use variable, and make the code more readable Coverage remained the same at 94.957% when pulling be5c63a4cdd0538ec72fe7cb15448c048603ea4b on arturobernalg:feature/boolean_string_const into 5689c91cf2a3a27f0a3a8a362857f2cf0919d4f6 on apache:master.
gharchive/pull-request
2021-02-20T07:13:03
2025-04-01T04:55:57.637752
{ "authors": [ "arturobernalg", "coveralls" ], "repo": "apache/commons-lang", "url": "https://github.com/apache/commons-lang/pull/714", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2628268558
refactor!: Cleanup & dependency modernization Motivation and Context coho is... unloved 😬 Description Update several dependencies and replace some with built-in NodeJS APIs Bump the minimum NodeJS version to 20.14.0+ (this is a dev tool not intended for public use, so that should be okay) Remove the code that tries to link Jira ticket IDs since we don't use Jira anymore Replace the code that fetches RAT with fetch() Replace q with standard Promises Get rid of almost all the shelljs uses, still a few to clean up around sed for versioning Fixes GH-239 There are some other quick wins that should be possible, but they are extensive enough that I didn't want to do them all in a single unreviewable PR. Things like getting rid of jira-client (see https://github.com/apache/cordova-coho/pull/236) and replacing the use of request in the github stats with fetch(). Need to investigate why newer versions of inquirer don't seem to work properly. Testing All existing unit tests pass (all very few of them) Successfully tested coho verify-archive Successfully tested coho audit-license-headers Successfully tested coho check-license Successfully tested coho verify-tags Successfully tested coho update-release-notes Successfully tested coho last-week Successfully tested coho list-pulls Successfully tested coho version Successfully tested coho repo-update Checklist [x] I've run the tests to see all new and existing tests pass [x] If this Pull Request resolves an issue, I linked to the issue in the text above (and used the correct keyword to close issues using keywords) Codecov Report Attention: Patch coverage is 71.42857% with 24 lines in your changes missing coverage. Please review. Project coverage is 58.21%. Comparing base (53ac7ad) to head (51e8856). Files with missing lines Patch % Lines src/superspawn.js 60.71% 22 Missing :warning: src/apputil.js 66.66% 1 Missing :warning: src/audit-license-headers.js 92.30% 1 Missing :warning: Additional details and impacted files @@ Coverage Diff @@ ## master #327 +/- ## ========================================== - Coverage 58.80% 58.21% -0.59% ========================================== Files 8 8 Lines 500 493 -7 ========================================== - Hits 294 287 -7 Misses 206 206 :umbrella: View full report in Codecov by Sentry. :loudspeaker: Have feedback on the report? Share it here.
gharchive/pull-request
2024-11-01T04:03:43
2025-04-01T04:55:57.656538
{ "authors": [ "codecov-commenter", "dpogue" ], "repo": "apache/cordova-coho", "url": "https://github.com/apache/cordova-coho/pull/327", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1699948368
Android API 32-33 Open NON Media Files Problem Hi everyone, i recently noticed that since the update to Android 12-13 / API 32-33, the readAsDataURL and readAsDataText methods are no longer working (return NULL) in case of reading PDF/TXT files, while they work no problem for pictures or videos. The folder I'm reading from is the Downloads folder of my Internal Storage. What is expected to happen? The result should return the Text or the Base64 Version (as Images) What does actually happen? The result is always null Command or Code ` Hi, this is my actual code for reading NON MEDIA File in MyFileService in Ionic 6; this code works fine when targetSDK is 29-30, but the Google Playstore no longer accepts this SDK. When i changed to target 32-33 it didn't work anymore: `//PROPERTY import { Injectable } from '@angular/core'; import { File } from '@ionic-native/file/ngx'; import { Device } from '@awesome-cordova-plugins/device/ngx'; import { FileChooser } from '@ionic-native/file-chooser/ngx'; import { FilePath } from '@awesome-cordova-plugins/file-path/ngx'; import { Platform } from '@ionic/angular'; import { HttpService } from './http.service'; import { Chooser } from '@awesome-cordova-plugins/chooser/ngx'; //CONSTRUCTOR public file: File, public device: Device, private platform: Platform, public fileChooser: FileChooser, private chooser: Chooser, public http: HttpService, private filePath: FilePath .... //SNIPPET WITH ERROR this.fileChooser .open() .then((uri) => { this.filePath .resolveNativePath(uri) .then((url) => { this.file .resolveLocalFilesystemUrl(url) .then((fileEntry: any) => { this.platform.ready().then(() => { debugger; this.file.checkFile(fileEntry['nativeURL'].replace(fileEntry['name'], ''), fileEntry['name']).then(response => { debugger; if (response === true) { this.file .readAsText( fileEntry['nativeURL'].replace(fileEntry['name'], ''), fileEntry['name'] ) .then((result) => { if (result) { ----> //RESULT NULL FOR PDF and TXT <--------- debugger; resolve(result); } else { reject('Errore'); } }) .catch((err) => { console.log('err-->' + JSON.stringify(err)); }); } }).catch(err => { debugger; }); }); }); }) .catch((err_3) => { reject(err_3); }); }) .catch((err_2) => { reject(err_2); });`` Environment, Platform, Device Ionic Cordova Android Platform 11 Version information Ionic: Ionic CLI : 6.20.3 (/usr/local/lib/node_modules/@ionic/cli) Ionic Framework : @ionic/angular 6.4.1 @angular-devkit/build-angular : 12.0.5 @angular-devkit/schematics : 12.2.18 @angular/cli : 12.0.5 @ionic/angular-toolkit : 4.0.0 Cordova: Cordova CLI : 11.0.0 Cordova Platforms : android 11.0.0, ios 6.2.0 Cordova Plugins : cordova-plugin-ionic-keyboard 2.2.0, cordova-plugin-ionic-webview 4.2.1, (and 27 other plugins) Utility: cordova-res : not installed globally native-run (update available: 1.7.2) : 1.7.1 System: Android SDK Tools : 26.1.1 (/Users/giuseppetaormina/Library/Android/sdk) ios-sim : 8.0.2 NodeJS : v18.12.0 (/usr/local/bin/node) npm : 8.19.2 OS : macOS Xcode : Xcode 14.3 Build version 14E222b The folder I'm reading from is the Downloads folder of my Internal Storage. Are you sure you don't mean external storage? Android, Internal Storage refers to to the storage medium guaranteed to be on the device, it usually contains the app install, and the data partition is private to the app only. External Storage on the other hand may or may not be on the device. If the device has a physical removable storage, like an sdcard, then external storage will be that storage medium, but android also emulates external storage for devices that either don't have an attached storage medium or just plainly doesn't support one. The Download/ directory is found on the External Storage medium. It may have a path such as /sdcard/Download/or /storage/emulated/0/Download. If the latter, then it's bit of a known issue unfortunately. As of API 30, Scoped Storage rules are fully enforced and there is a caveat on API 29 devices which I'll explain a bit later. What is Scoped Storage Scoped Storage is a privacy-focused storage system introduced by Android. It applies to External Storage medium and apps no longer have broad access to the external storage. WRITE_EXTERNAL_STORAGE no longer gives any permission, instead apps may freely read and write to the external storage. However they can only read and write to files that the app has created. If a file already exists but was created from from App A, then App B cannot read or write to that file using the Filesystem API. Natively to gain read & write access to these files, there is a non-filesystem API called the MediaStore API. The owner of the file must implement a file provider service so it can grant permission to a third-party app trying to read or write to it. I want to emphasize that this API isn't filesystem-like at all so it will be difficult to treat it as such which is why it hasn't really been addressed or resolved in this plugin. For example, you cannot progrommatically list or view the contents of a directory (you can but it will be filtered to only files owned/created by your app) and instead the MediaStore will open the system's file picker and the user may choose the file they want to open. Additionally, specificially on API 29 devices which is when Scoped Storage was first introduced (but while targeting API 29, it could be opted out, which is not possible today if you intend to deploy to Google Play store). API 29 devices does not have a File System bridge API into the scoped storage module. So using the Native File APIs to read/write files into external storage simply does not work and the only API available to do those actions is the previously mentioned MediaStore API. As a workaround for now, you'll need to use a plugin that implements/exposes the MediaStore API. Hi Norman, thanks a lot for the answer. Hi Norman, thank you so much for your quick reply. I had already seen similar Issues but I wanted to be sure about the presence of valid alternatives. I confirm that the path I access is file:///storage/emulated/0/Download/new.txt , so the problem is the one you explained. I just have a question about the implementation. You said that if the file is created by my app it is likely that it can be read, but if it is external and it is not MEDIA no. Can you confirm it? I'll explain the cases of my app: In my app it should be possible to export user profile file (with all the data that I have generated in the app) as a base64 with a format of my choice (*.vu) which was read directly with the readAsText function. This file would be used to carry my app's user data from one device to another. In my app it should be possible to read text files (txt, doc, pdf) to include as document attachments. Here I think you need to access totally external files coming from spaces like storage/emulated or google drive. Do you have any suggestions for me based on the above points? Many thanks in advance Giuseppe @gtaormina i think that latest code will fix this issue I'm going to shed some more light now that I've understand more about the Android scoped filesystem since the last post in May. Short answer is unfortunately I don't think the issue is resolvable by the plugin, and you'll likely need to use another plugin that interfaces MediaStore API rather than the File API. Long answer with history... History In API 29, Android introduces a system called Scoped Storage, however on API 29 devices for apps already published to the app store had the ability to opt out to the legacy system via requestLegacyExternalStorage flag. The legacy system gave apps free reign over external storage for as long as WRITE_EXTERNAL_STORAGE was granted. Scoped Storage (API 29-32) Scoped Storage makes the WRITE_EXTERNAL_STORAGE permission obsolete as apps have the ability to write to external storage without permission now, but they cannot write to a file that already exists if the file is owned by another app. Apps can also read from external storage without permission but visibility is limited to files only owned by your own app. READ_EXTERNAL_STORAGE will grant you to read media files, but not documents. Special Notes for API 29 On API 29 specifically, android does not have a File API bridge to Scoped Storage framework. This means if requestLegacyExternalStorage is not enabled, or if you have a brand new app that isn't published to the app store, all file APIs will fail to access external storage. In API 30 and onwards, Scoped Storage framework is forcefully enabled. Only in API 30 and onward, Android has implemented a File Bridge to allow File APIs to access content in external storage, allowing File APIs (and this file plugin) to somewhat work again. It also means this plugin will not work at all on API 29. It's important to note that even with these file based APIs enabled once again, they only have access to media files. To help your app work more smoothly with third-party media libraries, Android 11 allows you to use APIs other than the MediaStore API to access media files from shared storage using direct file paths. Paths like Documents or Downloads are not readable except for media files (images, videos and/or audio). In order to read non-media files, MediaStore API must be used. What is the MediaStore vs File APIs To the non-android devs that might be reading this, I'll write a quick explanation on the MediaStore API vs File API. The File API, is kinda what it sounds. It's a pretty standard filesystem oriented API where you operate on directories and files and you're able to stream data in and out of it. The API offers full programmatic control over the filesystem. MediaStore API is not a filesystem-oriented API. it's more like a database with a query system. The filesystem details is abstracted and you don't have programmatic control over the filesystem structure, and in some cases you have limited discoverability of files. It's a privacy-focused API so it comes with many restrictions. For example, you generally cannot programmatically read the list of files that may be present on the device. Rather you'll need to open an Intent (e.g. a file picker) where the user selects the file then, and only then your app can become aware that file exists. Attempting to use the MediaStore in a "file-like" API fashion as this plugin implements I don't think is feasible. Which is why I think a different plugin is required to properly handle accessing external storage on Android. Notes on API 33+ READ_EXTERNAL_STORAGE is now obsolete. It's still required to support API 32 and earlier, but on API 33, we have 3 new permissions: READ_MEDIA_AUDIO READ_MEDIA_IMAGES READ_MEDIA_VIDEO You'll notice we still don't have a permission for docs, so we cannot "discover" documentation. We can only open docs via a system file picker intent, like in previous android versions and only if the app that owns the document has a content provider implemented to grant third-party apps access to their document files. Notes on MANAGE_EXTERNAL_STORAGE Permission With the introduction of scoped storage, Android introduced MANAGE_EXTERNAL_STORAGE that I believe effectively gives you free reign over external storage similar to the legacy system. However this permission is protected and requires justification. Google will not allow any app with this permission be published to the app store unless if they have a very good reason to use it. An example of that the primary focus of your app is a file manager app, or an anti virus app. Therefore majority of users cannot take advantage of this permission. So with all that being said, I kinda foresee external storage support being stripped out in favour for a media store oriented plugin, however that is yet to been discussed at the Apache development level... https://www.npmjs.com/package/cordova-plugin-saf-mediastore is one plugin I've heard people had success with, but this isn't an endorsement and I'm not familiar with the NOPL license. So do your own research. I'm going to shed some more light now that I've understand more about the Android scoped filesystem since the last post in May. Short answer is unfortunately I don't think the issue is resolvable by the plugin, and you'll likely need to use another plugin that interfaces MediaStore API rather than the File API. Long answer with history... History In API 29, Android introduces a system called Scoped Storage, however on API 29 devices for apps already published to the app store had the ability to opt out to the legacy system via requestLegacyExternalStorage flag. The legacy system gave apps free reign over external storage for as long as WRITE_EXTERNAL_STORAGE was granted. Scoped Storage (API 29-32) Scoped Storage makes the WRITE_EXTERNAL_STORAGE permission obsolete as apps have the ability to write to external storage without permission now, but they cannot write to a file that already exists if the file is owned by another app. Apps can also read from external storage without permission but visibility is limited to files only owned by your own app. READ_EXTERNAL_STORAGE will grant you to read media files, but not documents. Special Notes for API 29 On API 29 specifically, android does not have a File API bridge to Scoped Storage framework. This means if requestLegacyExternalStorage is not enabled, or if you have a brand new app that isn't published to the app store, all file APIs will fail to access external storage. In API 30 and onwards, Scoped Storage framework is forcefully enabled. Only in API 30 and onward, Android has implemented a File Bridge to allow File APIs to access content in external storage, allowing File APIs (and this file plugin) to somewhat work again. It also means this plugin will not work at all on API 29. It's important to note that even with these file based APIs enabled once again, they only have access to media files. To help your app work more smoothly with third-party media libraries, Android 11 allows you to use APIs other than the MediaStore API to access media files from shared storage using direct file paths. Paths like Documents or Downloads are not readable except for media files (images, videos and/or audio). In order to read non-media files, MediaStore API must be used. What is the MediaStore vs File APIs To the non-android devs that might be reading this, I'll write a quick explanation on the MediaStore API vs File API. The File API, is kinda what it sounds. It's a pretty standard filesystem oriented API where you operate on directories and files and you're able to stream data in and out of it. The API offers full programmatic control over the filesystem. MediaStore API is not a filesystem-oriented API. it's more like a database with a query system. The filesystem details is abstracted and you don't have programmatic control over the filesystem structure, and in some cases you have limited discoverability of files. It's a privacy-focused API so it comes with many restrictions. For example, you generally cannot programmatically read the list of files that may be present on the device. Rather you'll need to open an Intent (e.g. a file picker) where the user selects the file then, and only then your app can become aware that file exists. Attempting to use the MediaStore in a "file-like" API fashion as this plugin implements I don't think is feasible. Which is why I think a different plugin is required to properly handle accessing external storage on Android. Notes on API 33+ READ_EXTERNAL_STORAGE is now obsolete. It's still required to support API 32 and earlier, but on API 33, we have 3 new permissions: READ_MEDIA_AUDIO READ_MEDIA_IMAGES READ_MEDIA_VIDEO You'll notice we still don't have a permission for docs, so we cannot "discover" documentation. We can only open docs via a system file picker intent, like in previous android versions and only if the app that owns the document has a content provider implemented to grant third-party apps access to their document files. Notes on MANAGE_EXTERNAL_STORAGE Permission With the introduction of scoped storage, Android introduced MANAGE_EXTERNAL_STORAGE that I believe effectively gives you free reign over external storage similar to the legacy system. However this permission is protected and requires justification. Google will not allow any app with this permission be published to the app store unless if they have a very good reason to use it. An example of that the primary focus of your app is a file manager app, or an anti virus app. Therefore majority of users cannot take advantage of this permission. So with all that being said, I kinda foresee external storage support being stripped out in favour for a media store oriented plugin, however that is yet to been discussed at the Apache development level... https://www.npmjs.com/package/cordova-plugin-saf-mediastore is one plugin I've heard people had success with, but this isn't an endorsement and I'm not familiar with the NOPL license. So do your own research. Has anyone a hint on how to use the cordova-plugin-saf-mediastore properly? I cannot get the function "getURI()" to work when I try passing "Pictures" as folder. I also tried the root-directory path and other combinations. My file-path I have is: "file:///storage/emulated/0/Pictures/cpcp_capture_65e7de6f.jpg" Goal: delete this file I appreciate any help! Thank you 💯 Closing because this is not actionable by Cordova in terms of file plugin API. Android simply doesn't allow the file plugin to read non-media files owned by other apps using the file API. MediaStore API is required. This is now noted at https://github.com/apache/cordova-plugin-file#androids-external-storage-quirks
gharchive/issue
2023-05-08T10:01:59
2025-04-01T04:55:57.699206
{ "authors": [ "EYALIN", "MauriceFrank", "breautek", "gtaormina" ], "repo": "apache/cordova-plugin-file", "url": "https://github.com/apache/cordova-plugin-file/issues/568", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
170179585
Fix reply filter for open_revs Current filter removing all not_found's from the final reply in presence of any {ok, Doc} in it, which is not always the correct behaviour, since we can have missing doc for one of the passed revisions. In general, in absence of latest attribute, we are expecting one reply per revision which could be either #doc or not found. This fix's making sure that we are removing not_found's only for the revisions that also have {ok, Doc} reply. COUCHDB-3097 +1 after adding a comment or tweaking that bit of logic to be a bit less subtle. @davisp I did both, does it look better now? 100% better. I kept staring at the {not_found, missing} tuple in both clauses of the fold and my brain just couldn't process what was going on for awhile. This is much better. +1 and good find on that.
gharchive/pull-request
2016-08-09T14:24:13
2025-04-01T04:55:57.710784
{ "authors": [ "davisp", "eiri" ], "repo": "apache/couchdb-fabric", "url": "https://github.com/apache/couchdb-fabric/pull/65", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
326723067
Add official support of Couchdb for Ubuntu 18.04 Hello, Please add this task in your 2.2.0 milestone. https://github.com/apache/couchdb/issues/1314 (Duplicate of this but it's closed and task is still open) @codeyash The convenience binaries for CouchDB are only semi-official. Whether that ticket is open or not, the work is ongoing. Any expected date please? Hello, I cannot install the couchdb 2.1.1 in Ubuntu 18.04. I use this url: https://apache.bintray.com/couchdb-deb/dists/bionic/main/ But it returns an error. Any suggestions? Thank you. Please see https://github.com/apache/couchdb/issues/1314
gharchive/issue
2018-05-26T09:03:40
2025-04-01T04:55:57.716594
{ "authors": [ "FgfdCBVCbnsebtgf", "codeyash", "wohali" ], "repo": "apache/couchdb", "url": "https://github.com/apache/couchdb/issues/1343", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
598482440
Broken unicode in CouchDB 3.0.0 Description After upgrading to CouchDB 3 we had a big problem with JS query server that now incorrectly works with unicode characters. Steps to Reproduce Create view with following map function: function (doc) { log(doc.string_field.length); } Put document with field string_field with any unicode characters. We're have text fields with Hebrew and Russian. Now you can see in console incorrect string length. Thus substring and slice also don't work. For example: {"string_field": "мама мыла раму"} Expected Behaviour Correct behaviour of length, substring and slice. Your Environment CouchDB 3.0 in single node mode or CouchDB 3.0 in cluster mode with 3 nodes. Installed from official docker image on 64bit host system. CouchDB version used: CouchDB 3.0, git_sha: 03a77db6c Operating system and version: Official docker image - couchdb:3 Duplicate of #2756 . Thanks for the additional background.
gharchive/issue
2020-04-12T13:16:32
2025-04-01T04:55:57.720591
{ "authors": [ "kimaiavd", "wohali" ], "repo": "apache/couchdb", "url": "https://github.com/apache/couchdb/issues/2779", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
385187074
Add Credo to Elixir test suite Overview This PR adds Credo as the static code analysis tool of choice for the Elixir test suite. Testing recommendations Just issue make elixir-credo first, then make elixir Related Issues or Pull Requests #1756 Checklist [X] Code is written and works correctly; [X] Changes are covered by tests; [X] Documentation reflects the changes; Here is a rationale of this PR. I added Credo to the Mix project, then I disabled some "not so useful" checks. They are: Credo.Check.Refactor.CyclomaticComplexity Credo.Check.Refactor.LongQuoteBlocks Credo.Check.Refactor.Nesting Then, I added Credo to the Makefile and Makefile.win with a proper goal, and I added that goal to make elixir. This way, make elixir formats the code according to #1767, then executes the elixir-credo goal erroring on basically everything. The only check I maintained but I changed its exit code to 0 is Credo.Check.Design.TagTODO, because I want to be warned about TODOs and monitor them, but I still want to be able to use them as reminders. @wohali I think your eyes could be useful here 🙂 I have a general question about max line length. This commit explicitly sets the max line length to 120, up from its default of 96. For Erlang code, the current consensus (and also @davisp's emelio) appears to be 80, for which there is ample precedent. Because it forces you to write more concise code, and it’s great being able to have two code windows open side by side (on a 15" laptop), I hope we also adopt 80 for Elixir code. But I'm curious to what others think it should be. @jaydoane also, as to your comment on "two code windows open side by side (on a 15" laptop)" is highly variable. I use a 15" laptop and often get 3 or 4 code windows open side by side, with >80 characters each. Resolution and font size (and your own visual acuity!) enter into it a lot :) @jaydoane Hrm, I don't mind it... Some comments on the above discussion: Wow, you were great guys, that was such a beautiful debate 😬 I enforced max line length to 120 because our formatter enforces the coding style to 120-chars lines. I'm gonna take it to 90 in Credo, and of course I'm updating the .formatter.exs as well. Also, I love the 120 limit, and the community does. But if we want to enforce a 90 columns limit, that's the beautiful point of having a tool that you can configure 😹 About disabling Credo.Check.Consistency.ParameterPatternMatching and the long quotes one: Consistency.ParameterPatternMatching is ok, but I think: 1) that we need a long quote block for our test helper 2) that long quote blocks are smells and we should limit that smell to a known place. Enforcing a 90 columns format will cause almost all of our tests to be reformatted. @wohali et al. do you want me to go further or do we want to do that in another subsequent PR? Reformatting all the lines that require it to pass a make check invocation is totally fine. Just keep it to an isolated commit and then make sure that make check fails after the fact so that we're not re-introducing breaking things after the fact. @davisp thanks! I'm going further then Hi all! Not long ago we also added checking elixir code format with mix format before running elixir test suite. Since line length is more code formatting issue rather than some logic error, like not used local var or something like that, I believe it's better to address this to mix format. Here is formatter options, :line_length exactly what we need. We just need to add this in our .formatter.exs wuth reasonable value. In that case make elixir if new changes have incorrect line length. And developer just need to run mix format to reformat his changes. In that case mix credo have to address only logic code errors. @dottorblaster sorry, looks like you are going to add line_length in formatter. @van-mronov yeah line_length is already inside the .formatter.exs and I'm gonna enforce that to 90 as I said before, no problem @wohali @davisp @jaydoane @van-mronov I added two more commits for that, gonna edit the Makefile soon, can you provide me feedbacks? @van-mronov I still prefer to maintain max length checks in Credo too because it's more developer friendly. I enforced the value formatter-side as well. +1 LGTM
gharchive/pull-request
2018-11-28T09:57:32
2025-04-01T04:55:57.733129
{ "authors": [ "davisp", "dottorblaster", "jaydoane", "van-mronov", "wohali" ], "repo": "apache/couchdb", "url": "https://github.com/apache/couchdb/pull/1769", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
237897071
Optimize ddoc cache Overview The previous version of ddoc_cache was written to rely on evicting entries after a maximum TTL. This leads to issues on clusters that have a large amount of load on databases with a large Q. What ends up happening is that when a design document is evicted we suddenly have a thundering herd scenario as every client attempts to reinsert it into the cache. This change instead relies on a monitor process for each cache entry that periodically attempts to refresh the cache. This way normal clients accessing a popular design document will never hit a point where it doesn't exist in cache. And we'll have at most one reader trying to write the value. Testing recommendations make eunit apps=ddoc_cache You'll find its got 100% test coverage if you enable cover. One note on the tests that's a bit different than other suites. I stumbled across a fun pattern of adding a function call to various locations in the code so that I can meck:wait on it. This is a compile time switch so its only affects the eunit runs. Production code has the calls compiled out and replaced by an atom which does nothing. Checklist [x] Code is written and works correctly; [x] Changes are covered by tests; Refreshing my memory on the mem3 PR I can definitely apply a lot of that to this as well so I'll update things for that. For the accessed messages there's not a lot I think we can do there with the existing structure. Although I could see changing things around a bit to make it a LFU cache instead of an LRU cache. It'd involve a table scan to evict but would remove a lot of the message passing to ddoc_cache_lru and friends. I'll make the mem3 inspired changes first and then will look at this approach. @chewbranca You managed to nerd snipe me for like four days trying to figure out the mem3 and concurrent LRU stuff. For the mem3 bits I think my conclusion is that its not appropriate here. For mem3_shards we wanted to make sure that there would always be progress so we load things outside of the gen_server and send them to cache. Given that its a local db read this isn't the most terrible thing in the world. Although, for ddoc_cache its a clustered fabric call, so if we followed the mem3_shards pattern we'd actually just be putting extra load on other parts of the system like rexi, and the couch_db_updaters at 3x the rate of client requests. So having them funnel through the opener actually makes more sense here. As to the distributed LRU approach for writes I've been playing with a couple different standalone tests to try and get a feel for our maximum throughput with that. So far (with some extremely synthetic tests) I think we're looking at topping out at around 200K updates a second (at least on my laptop). However, this test doesn't actually take into account actually evicting things. My plan for today is to write a basho_bench driver to try and gauge the relative throughput between what I've got written now and a second approach based on your direct ets write idea. I'll post results here when I get them. @chewbranca Epic nerd snipe. I've sprinkled more parallel on the ddoc_cache. Care to take another look. Currently measuring 1M+ ops/sec against a 1,000 item cache on my MBP. I have to do some work on the test suite tomorrow to finish up the rewritten rewrite work but I'm currently pretty happy with it. New style is one pid per entry which does all of the ets writes for the entry. the LRU process only exists to do evictions. I soaked 20,000 clients against a 500 item cache on my laptop for 30-45m and it sustained about 1.5M ops/sec without error and without blowing RAM up beyond 500M. Those numbers are for a custom module that does no loading but I think it shows that the cache itself won't be bound by the cache and instead it'll be fabric calls and or some other bottle neck before the cache. @davisp hehehehe happy to help with the sniping ;-) @davisp overall looks solid to me. @chewbranca I managed to figure out a decent way to avoid the unnecessary second fabric calls in the two cases you were worried about. Now the first client that opens an entry will kick off a process that will insert the corresponding cache entry if no entry existed before. I.e., if we call ddoc_cache:open(DbName, DDocId) and there's no entry for it, we'll take the result and insert it for the revid specific version as well (assuming there's not already an existing entry). And vice versa when opening a revid specific version. https://github.com/apache/couchdb/pull/610/commits/85b53855255be0a4166bbb14697c010ff4468d20 Also, for your comment on the revid specific entries I spent some time trying to figure out a decent way to be clever using latest=true or something but in the end I went entirely different. Now if a cache entry is not accessed within its refresh period it will voluntarily drop out of the cache. So this will apply for all entries which I think is good to help when someone wants to scan a bunch of views periodically. We'll end up dropping all of the unused entries which could help with memory pressure if someone has big design documents without much activity on more than a few. https://github.com/apache/couchdb/pull/610/commits/32cd254d7e7c678a10a9302e48e9c6f064701701 For the configurability, I added a refresh timeout config option and documented it in the default.ini as per usual. I also added notes for max_size as well. https://github.com/apache/couchdb/pull/610/commits/9151fb59a8f8553a95af23fa58ef67a76f069b80 And thanks for the good eye on clearing the waiters list. https://github.com/apache/couchdb/pull/610/commits/7c376bd6efc372cd7f107989fd14e40e8ab7a7e6 The only other thing you really noted was the comment update but I think that was just due to a misunderstanding so I'm assuming you're good my comment there. I ran the ddoc_cache_speed benchmark with the following diff: spawn_workers(WorkerCount) -> Self = self(), - WorkerDb = list_to_binary(integer_to_list(WorkerCount)), - spawn(fun() -> + WorkerId = WorkerCount, % rem ?RANGE, + WorkerDb = integer_to_binary(WorkerId), + spawn_link(fun() -> do_work(Self, WorkerDb, 0) end), spawn_workers(WorkerCount - 1). -do_work(Parent, WorkerDb, Count) when Count >= 25 -> +do_work(Parent, WorkerDb, Count) when Count >= 1000 -> Parent ! {done, Count}, do_work(Parent, WorkerDb, 0); case timer:now_diff(Now, Start) of N when N > 1000000 -> {_, MQL} = process_info(whereis(ddoc_cache_lru), message_queue_len), - io:format("~p ~p~n", [Count, MQL]), + CacheSize = ets:info(ddoc_cache_lru, size), + io:format("~p ~p ~p~n", [Count, MQL, CacheSize]), report(Now, 0); _ -> receive ddoc_cache_speed:go(2000). 1403000 1000 997 1441000 999 997 1416000 1001 998 1387000 1000 997 1403000 1001 996 1418000 1001 997 Left it running for 30 min or so and was seeing about 1M+ opens per second. Cache max size was left as the default at 1000. VM memory and number of processes stayed stable under 100MB and 3500 respectively. Cache ets size (the 3rd column) also stayed under 1000. Reducing the max_cache to 500 also reduced the table sizes accordingly: 1336000 1500 497 1252000 1499 498 1307000 1500 497 1240000 1500 497 The interesting observation is that the LRU message queue size is in a steady steady and is almost exactly equal to NumberOfWorkers-MaxSize. NumberOfWorkers=2000 and when MaxSize=1000, then message queue was 2000-1000=1000. Then when MaxSize was reduced to 500, queue size became 2000-500=1500. +1 (but first see a few questions about ets delete and other minor nits). Very nice work! Performance Benchmarking Update: New one is faster! And now for some graphs. For each of these graphs the max_dbs_open was set to 5,000. That's important to remember as I go through the parameter sets whether we're hitting that limit which will be an artificial limit on performance of the cache. Each of these tests is also just querying an empty view which leads to a ddoc_cache lookup for the coordinator, and then Q lookups for each RPC worker. So the requests per second numbers have to be multiplied out to get actual cache performance. Granted we really don't care given that its a constant factor. These first two runs are 1,000 workers hitting 1,000 different design documents in 1,000 different databases with two different Q values (i.e., 1,000 workers using their own ddoc and db). For the most part their ops/sec is basically identical while latencies for the new ddoc_cache are slightly better. This suggests that something else was bottlenecking performance. Possibly the basho_bench driver and possibly something else in CouchDB. Old ddoc_cache Q=4: New ddoc_cache Q=4: Old ddoc_cache Q=8 New ddoc_cache Q=8 So basically, trying to hit a bunch of different databases at once now has slightly better latencies but both are fast enough to not be the bottleneck. Next up was a shoot the moon test to try and test the limits of what sort of performance we could get trying to crush a single entry. These two graphs show 1,000 workers hitting a single design document in a single database. The Q for both of these is 128 which means that the ddoc_cache is going to have to try and maintain 129,000 lookups/sec. Old ddoc_cache: New ddoc_cache: The immediate thing to note here is that old ddoc_cache flat lines half way into its test. This was because couch_log_server spiked its message queue. The reason for this is because of the huge flood of fabric_rpc_worker timeout messages being logged. If you look at the graph for the old ddoc_cache you can see that it has two cliffs, one at 60s, and one at 120s when it then flat lines cause it pushed couch_log_server off a cliff. These drops are precisely what motivated this work in the first place when the old ddoc_cache would evict entries every 60s and the thundering herd would knock things off their rocker. For the new ddoc_cache you'll see that its still kinda crap performance even though it doesn't totally lose its mind. In the background this was because rexi was unable to keep up with the work load and spiked pretty bad. I'm gonna be investigating that area for more optimizations once this work is wrapped up. And finally, this last graph is a comparison between the old and new ddoc_caches sweeping through Q=8, 16, and 32 with the same 1,000 workers against a single db and ddoc. Red is old ddoc_cache, green is new ddoc_cache. Yes I know Tufte would kill me but that's the colors that get picked and I don't care enough to go fiddle with it. As you can see the new ddoc_cache is consistently faster as well as much less variable. Looking into the variance on the best Q=64 run for the old ddoc_cache vs the worst Q=64 run for new ddoc_cache shows a good example of the variance of the old approach. For these two runs I've also included what the rexi server message queue is doing so we can see its effect on performance. Old ddoc_cache: New ddoc_cache: Again we can see how badly that 60s eviction policy is when we have sustained load against a single design document. This leads to some fairly massive spikes in the system. For the old ddoc_cache on the third eviction at 180s we see it flat line again which is why those runs are so variable. For the new ddoc_cache we can see that db3's rex has sustained elevated message counts which are holding back the benchmark back from meeting some of the old ddoc_cache spikes when rexi had a chance to clear out. Now that I can duplicate that rexi issue easily enough though I'll be working on trying to figure out why its being slow and try and optimize around it. Hopefully this data is as convincing to everyone else as it is to me. If anyone wants me to check into anything else I certainly have the data and/or can design runs to try and run in some other configuration if requested. However poking both ends of the spectrum (lots of clients against separate design docs and lots of clients against a single design doc) I'm fairly confident that we're winning across the spectrum though most specifically for the single design doc case (which was the motivation for this work).
gharchive/pull-request
2017-06-22T16:08:40
2025-04-01T04:55:57.757978
{ "authors": [ "chewbranca", "davisp", "nickva" ], "repo": "apache/couchdb", "url": "https://github.com/apache/couchdb/pull/610", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
144847167
implemented AutoCloseable Client Implemented AutoCloseable interface which simple calls destroy() method This code does not seem to build. I got somee checkstyle errors. I also think setting the classloader on the bus is not correct as the bus may be for a complete bundle. So setting it when a DynamicClient is created and removing it when it is removed does not seem to be valid for all cases. Pelase also try to not reformat the code in commits that also change the logic as this makes it a lot harder to review. Ideally you should create one branch / pull request per jira issue and make sure the code builds on your machine using mvn clean install. Yes I've reformatted the code because I didn't knew about your formatter (then I found it). So probably you can fix the issue better if you know how. Sorry for bundling all commits to one PR, I'm not using git too long. No problem. I had also had a lot of issue with git when I started. I just updated the getting involved guide to show how you ideally structure your work in git. https://cxf.apache.org/getting-involved.html It does not show the git commands you need but how it should look like. Build triggered. sha1 is merged. Build started sha1 is merged. Build finished. 0 tests run, 0 skipped, 0 failed.
gharchive/pull-request
2016-03-31T09:50:41
2025-04-01T04:55:57.763006
{ "authors": [ "bilak", "cschneider", "tomitribe-dev" ], "repo": "apache/cxf", "url": "https://github.com/apache/cxf/pull/124", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1050499476
[Fix-#6783] switchVersion error (#6783) Purpose of the pull request https://github.com/apache/dolphinscheduler/issues/6783 Brief change log Verify this pull request This pull request is code cleanup without any test coverage. (or) This pull request is already covered by existing tests, such as (please describe tests). (or) This change added tests and can be verified as follows: @JinyLeeChina Hi @zwZjut , to target specific issue from PR, we recommend use keyword fix: #issue_id or close: #issue_id or closes: #issue_id in your PR describe. I would not only connect issue to PR but also close issue automatically when PR is be closed. Codecov Report Merging #6784 (4bbe66d) into dev (975131e) will decrease coverage by 0.04%. The diff coverage is 100.00%. @@ Coverage Diff @@ ## dev #6784 +/- ## ============================================ - Coverage 41.84% 41.80% -0.05% + Complexity 3616 3613 -3 ============================================ Files 641 641 Lines 25890 25891 +1 Branches 2795 2795 ============================================ - Hits 10833 10823 -10 - Misses 14073 14088 +15 + Partials 984 980 -4 Impacted Files Coverage Δ ...er/api/service/impl/TaskDefinitionServiceImpl.java 33.49% <100.00%> (+0.32%) :arrow_up: ...r/plugin/registry/zookeeper/ZookeeperRegistry.java 47.27% <0.00%> (-7.28%) :arrow_down: ...er/master/dispatch/host/assign/RandomSelector.java 77.77% <0.00%> (-5.56%) :arrow_down: ...org/apache/dolphinscheduler/remote/utils/Host.java 37.77% <0.00%> (-2.23%) :arrow_down: ...e/dolphinscheduler/remote/NettyRemotingClient.java 52.11% <0.00%> (-0.71%) :arrow_down: Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 975131e...4bbe66d. Read the comment docs.
gharchive/pull-request
2021-11-11T02:11:46
2025-04-01T04:55:57.778123
{ "authors": [ "codecov-commenter", "zhongjiajie", "zwZjut" ], "repo": "apache/dolphinscheduler", "url": "https://github.com/apache/dolphinscheduler/pull/6784", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1069391948
[Feature][Task] add seatunnel task plugin Purpose of the pull request add seatunnel task plugin Brief change log Verify this pull request This pull request is code cleanup without any test coverage. (or) This pull request is already covered by existing tests, such as (please describe tests). (or) This change added tests and can be verified as follows: Codecov Report Merging #7131 (62f3e18) into dev (12b46df) will decrease coverage by 0.01%. The diff coverage is 100.00%. @@ Coverage Diff @@ ## dev #7131 +/- ## ============================================ - Coverage 41.28% 41.26% -0.02% + Complexity 3665 3661 -4 ============================================ Files 634 634 Lines 26373 26373 Branches 2953 2953 ============================================ - Hits 10889 10884 -5 - Misses 14455 14460 +5 Partials 1029 1029 Impacted Files Coverage Δ ...hinscheduler/common/utils/TaskParametersUtils.java 57.14% <ø> (ø) ...apache/dolphinscheduler/common/enums/TaskType.java 95.65% <100.00%> (ø) ...er/master/dispatch/host/assign/RandomSelector.java 77.77% <0.00%> (-5.56%) :arrow_down: ...org/apache/dolphinscheduler/remote/utils/Host.java 37.77% <0.00%> (-2.23%) :arrow_down: ...dolphinscheduler/remote/future/ResponseFuture.java 81.96% <0.00%> (-1.64%) :arrow_down: ...e/dolphinscheduler/remote/NettyRemotingClient.java 52.11% <0.00%> (-1.41%) :arrow_down: Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 12b46df...62f3e18. Read the comment docs. I restart the failing CI
gharchive/pull-request
2021-12-02T10:40:06
2025-04-01T04:55:57.793343
{ "authors": [ "codecov-commenter", "zhongjiajie", "zhuangchong" ], "repo": "apache/dolphinscheduler", "url": "https://github.com/apache/dolphinscheduler/pull/7131", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2033881955
[Bug] (jdbc catalog) When the field value of some tables is null, the query will report an error. Search before asking [X] I had searched in the issues and found no similar issues. Version master What's Wrong? CAUSED BY: ClassCastException: [Ljava.lang.Object; cannot be cast to [Ljava.lang.Long; at org.apache.doris.qe.Coordinator.getNext(Coordinator.java:1278) ~[classes/:?] at org.apache.doris.qe.StmtExecutor.sendResult(StmtExecutor.java:1538) ~[classes/:?] at org.apache.doris.qe.StmtExecutor.handleQueryStmt(StmtExecutor.java:1473) ~[classes/:?] at org.apache.doris.qe.StmtExecutor.handleQueryWithRetry(StmtExecutor.java:653) ~[classes/:?] at org.apache.doris.qe.StmtExecutor.executeByNereids(StmtExecutor.java:586) ~[classes/:?] at org.apache.doris.qe.StmtExecutor.execute(StmtExecutor.java:444) ~[classes/:?] at org.apache.doris.qe.StmtExecutor.execute(StmtExecutor.java:431) ~[classes/:?] at org.apache.doris.qe.ConnectProcessor.handleQuery(ConnectProcessor.java:240) ~[classes/:?] at org.apache.doris.qe.MysqlConnectProcessor.handleQuery(MysqlConnectProcessor.java:160) ~[classes/:?] at org.apache.doris.qe.MysqlConnectProcessor.dispatch(MysqlConnectProcessor.java:187) ~[classes/:?] at org.apache.doris.qe.MysqlConnectProcessor.processOnce(MysqlConnectProcessor.java:240) ~[classes/:?] at org.apache.doris.mysql.ReadListener.lambda$handleEvent$0(ReadListener.java:52) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) ~[?:1.8.0_151] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) ~[?:1.8.0_151] at java.lang.Thread.run(Thread.java:748) ~[?:1.8.0_151] What You Expected? Results can be obtained. How to Reproduce? CREATE TABLE numeric_table ( id bigint(32), tinyint_col TINYINT, smallint_col SMALLINT, int_col INT, bigint_col BIGINT, decimal_col DECIMAL(10,2), float_col FLOAT(8,4), double_col DOUBLE(15,10), boolean_col BOOLEAN, date_col DATE, time_col TIME, datetime_col DATETIME, timestamp_col TIMESTAMP ); INSERT INTO numeric_table (id, tinyint_col, smallint_col, int_col, bigint_col, decimal_col, float_col, double_col, boolean_col, date_col, time_col, datetime_col, timestamp_col) VALUES (1, 1, 100, 1000, NULL, 10.50, 3.1415, 3.1415926535, TRUE, '2023-01-01', '12:34:56', '2023-01-01 12:34:56', CURRENT_TIMESTAMP), (2, 0, -50, -500, -NULL, -5.25, -2.7182, -2.7182818284, FALSE, '2023-02-02', '23:59:59', '2023-02-02 23:59:59', CURRENT_TIMESTAMP), (3, -1, 0, 0,NULL, 0.00, 0.0000, 0.0000000000, TRUE, '2023-03-03', '00:00:00', '2023-03-03 00:00:00', CURRENT_TIMESTAMP); Anything Else? Because the type of a certain column is null and the ColumnValueConverter is null. But the type is forced to change. Since the column values ​​are all null, the type cannot be recognized, so null is returned directly. Are you willing to submit PR? [X] Yes I am willing to submit a PR! Code of Conduct [X] I agree to follow this project's Code of Conduct Hello @liugddx I would like to work on this issue.
gharchive/issue
2023-12-09T14:58:01
2025-04-01T04:55:57.805470
{ "authors": [ "DevPJ9", "liugddx" ], "repo": "apache/doris", "url": "https://github.com/apache/doris/issues/28204", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1367462588
testadd test join case4 Proposed changes Issue Number: close #xxx Problem summary Describe your changes. Checklist(Required) Does it affect the original behavior: [ ] Yes [x] No [ ] I don't know Has unit tests been added: [x] Yes [ ] No [ ] No Need Has document been added or modified: [ ] Yes [x] No [ ] No Need Does it need to update dependencies: [ ] Yes [x] No Are there any changes that cannot be rolled back: [ ] Yes (If Yes, please explain WHY) [x] No Further comments If this is a relatively large or complex change, kick off the discussion at dev@doris.apache.org by explaining why you chose the solution you did and what alternatives you considered, etc... LGTM
gharchive/pull-request
2022-09-09T08:17:47
2025-04-01T04:55:57.811206
{ "authors": [ "dataalive", "zy-kkk" ], "repo": "apache/doris", "url": "https://github.com/apache/doris/pull/12508", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1389424140
regression Add regression tests for datev2 Proposed changes Issue Number: close #xxx Problem summary Describe your changes. Checklist(Required) Does it affect the original behavior: [ ] Yes [ ] No [ ] I don't know Has unit tests been added: [ ] Yes [ ] No [ ] No Need Has document been added or modified: [ ] Yes [ ] No [ ] No Need Does it need to update dependencies: [ ] Yes [ ] No Are there any changes that cannot be rolled back: [ ] Yes (If Yes, please explain WHY) [ ] No Further comments If this is a relatively large or complex change, kick off the discussion at dev@doris.apache.org by explaining why you chose the solution you did and what alternatives you considered, etc... I will use a 1MB dataset to replace cases for DateV2.
gharchive/pull-request
2022-09-28T14:03:27
2025-04-01T04:55:57.816684
{ "authors": [ "Gabriel39" ], "repo": "apache/doris", "url": "https://github.com/apache/doris/pull/13040", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1676280608
enhance: When requiredSlots is empty, prune child previously in ColumnPruning Proposed changes Issue Number: close #xxx Problem summary When requiredSlots is empty, prune child previously in ColumnPruning. A example: Agg count(*) - Filter age > 3 -- Scan id, age We should get Agg count(*) - Project age -- Filter age > 3 --- Scan id, age instead of Agg count(*) - Project age, id -- Filter age > 3 --- Scan id, age Checklist(Required) [ ] Does it affect the original behavior [ ] Has unit tests been added [ ] Has document been added or modified [ ] Does it need to update dependencies [ ] Is this PR support rollback (If NO, please explain WHY) Further comments If this is a relatively large or complex change, kick off the discussion at dev@doris.apache.org by explaining why you chose the solution you did and what alternatives you considered, etc... column pruning has lots of special cases, and can not image all of the cases, so I can not review this pr. but in my experience, this code more likely make the regression test, because I remember that I've write the code like you and failed. when you run into the bug, you can unify abstraction and modification based on all known cases.
gharchive/pull-request
2023-04-20T08:45:28
2025-04-01T04:55:57.820794
{ "authors": [ "924060929", "jackwener" ], "repo": "apache/doris", "url": "https://github.com/apache/doris/pull/18856", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1813445558
Fix Fix bugs in tombstone Bug: table ref has not decreased when replay by tombstone. table sync GC tombstone information can cause errors. Bug fix: fixed. fixed, db sync can be smoothly upgraded, but table sync requires GC of all binlogs run buildall (From new machine)TeamCity pipeline, clickbench performance test result: the sum of best hot time: 47.02 seconds stream load tsv: 509 seconds loaded 74807831229 Bytes, about 140 MB/s stream load json: 20 seconds loaded 2358488459 Bytes, about 112 MB/s stream load orc: 65 seconds loaded 1101869774 Bytes, about 16 MB/s stream load parquet: 32 seconds loaded 861443392 Bytes, about 25 MB/s insert into select: 28.9 seconds inserted 10000000 Rows, about 346K ops/s storage size: 17161222749 Bytes run buildall run buildall (From new machine)TeamCity pipeline, clickbench performance test result: the sum of best hot time: 45.54 seconds stream load tsv: 507 seconds loaded 74807831229 Bytes, about 140 MB/s stream load json: 19 seconds loaded 2358488459 Bytes, about 118 MB/s stream load orc: 65 seconds loaded 1101869774 Bytes, about 16 MB/s stream load parquet: 31 seconds loaded 861443392 Bytes, about 26 MB/s insert into select: 29.4 seconds inserted 10000000 Rows, about 340K ops/s storage size: 17163005615 Bytes
gharchive/pull-request
2023-07-20T08:23:25
2025-04-01T04:55:57.825710
{ "authors": [ "deadlinefen", "hello-stephen" ], "repo": "apache/doris", "url": "https://github.com/apache/doris/pull/22031", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2009446609
[branch-2.0] enable hive view by default Proposed changes Enable hive view support by default Further comments If this is a relatively large or complex change, kick off the discussion at dev@doris.apache.org by explaining why you chose the solution you did and what alternatives you considered, etc... run buildall
gharchive/pull-request
2023-11-24T10:04:08
2025-04-01T04:55:57.827551
{ "authors": [ "morningman" ], "repo": "apache/doris", "url": "https://github.com/apache/doris/pull/27550", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2120162445
fixpush more than one runtime filters into cte Proposed changes Issue Number: close #xxx Further comments If this is a relatively large or complex change, kick off the discussion at dev@doris.apache.org by explaining why you chose the solution you did and what alternatives you considered, etc... run buildall run buildall run p0 run pipelinex_p0 run pipelinex_p0 run p0 run buildall run buildall run p0
gharchive/pull-request
2024-02-06T07:48:25
2025-04-01T04:55:57.830686
{ "authors": [ "englefly" ], "repo": "apache/doris", "url": "https://github.com/apache/doris/pull/30901", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2122910216
[opt](ES catalog) Increase to 3 connect attempts per node (#30957) Proposed changes pick #30957 Further comments If this is a relatively large or complex change, kick off the discussion at dev@doris.apache.org by explaining why you chose the solution you did and what alternatives you considered, etc... run buildall
gharchive/pull-request
2024-02-07T12:20:05
2025-04-01T04:55:57.832275
{ "authors": [ "qidaye" ], "repo": "apache/doris", "url": "https://github.com/apache/doris/pull/30969", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2143865361
[Enhancement](jdbc catalog) Add security check on driver when creating Jdbc Catalog Proposed changes Issue Number: close #xxx Enhancements This PR introduces enhanced security measures for managing and verifying driver packages when creating Catalogs in Doris. Specifically, it adds path management and checksum verification to ensure the security of driver jars specified by the driver_url. The key highlights include: For driver packages specified by filename within the jdbc_drivers/ directory (configured via fe.conf and be.conf), Doris assumes these are secure and does not perform additional path checks. For driver packages specified through absolute local paths or HTTP URLs, Doris enforces checks against allowed paths configured via the jdbc_driver_secure_path FE configuration item. This setting supports multiple paths separated by semicolons. If the driver_url's path does not match any prefix in jdbc_driver_secure_path, the creation of the Catalog is denied. A checksum parameter can be specified when creating a Catalog, allowing Doris to verify the driver package's integrity post-load. Catalog creation is aborted if verification fails. Upgrade Impact The default configuration for jdbc_driver_secure_path is *, indicating that all driver package paths are allowed. This ensures backward compatibility and does not affect existing Catalogs upon upgrade. Since the verification is only performed during Catalog creation, existing Catalogs remain unaffected. Further comments If this is a relatively large or complex change, kick off the discussion at dev@doris.apache.org by explaining why you chose the solution you did and what alternatives you considered, etc... run buildall run buildall run buildall
gharchive/pull-request
2024-02-20T08:58:40
2025-04-01T04:55:57.837129
{ "authors": [ "zy-kkk" ], "repo": "apache/doris", "url": "https://github.com/apache/doris/pull/31153", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2144104632
fixFix hive p2 case (#31149) backport https://github.com/apache/doris/pull/31149 Further comments If this is a relatively large or complex change, kick off the discussion at dev@doris.apache.org by explaining why you chose the solution you did and what alternatives you considered, etc... run buildall
gharchive/pull-request
2024-02-20T10:57:58
2025-04-01T04:55:57.839514
{ "authors": [ "Jibing-Li" ], "repo": "apache/doris", "url": "https://github.com/apache/doris/pull/31165", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2302637415
opt handle oom exception in spill tasks Proposed changes Issue Number: close #xxx Further comments If this is a relatively large or complex change, kick off the discussion at dev@doris.apache.org by explaining why you chose the solution you did and what alternatives you considered, etc... run buildall run buildall run buildall run buildall run p0 run buildall run p0 run buildall
gharchive/pull-request
2024-05-17T12:34:37
2025-04-01T04:55:57.842332
{ "authors": [ "mrhhsg" ], "repo": "apache/doris", "url": "https://github.com/apache/doris/pull/35025", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2495000514
test find large memory allocation not catch exception Proposed changes Issue Number: close #xxx run buildall
gharchive/pull-request
2024-08-29T16:19:23
2025-04-01T04:55:57.843374
{ "authors": [ "yiguolei" ], "repo": "apache/doris", "url": "https://github.com/apache/doris/pull/40156", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2587528905
test based on https://github.com/apache/doris/pull/41784, resolve conflicts with the latest branch-2.1 to run branch-2.1 CI run buildall run buildall run buildall run buildall
gharchive/pull-request
2024-10-15T03:26:35
2025-04-01T04:55:57.845129
{ "authors": [ "bobhan1" ], "repo": "apache/doris", "url": "https://github.com/apache/doris/pull/41839", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2594184529
[Improvement](LDAP Auth)Enhance LDAP authentication with a configurable group filter Proposed changes This PR enhances LDAP authentication by adding an optional configurable filter for retrieving user groups, primarily to support Open Directory LDAP implementations. If the configurable property is left empty, the existing workflow will remain unchanged. run buildall run buildall @morningman @zddr Please could you assist in merging this PR
gharchive/pull-request
2024-10-17T09:40:21
2025-04-01T04:55:57.846892
{ "authors": [ "nsivarajan" ], "repo": "apache/doris", "url": "https://github.com/apache/doris/pull/42038", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2720234298
[test](index compaction)Add exception cases for index compaction What problem does this PR solve? Problem Summary: Add more exception cases for index compaction Refactor index compaction test code, add a util class for all cases Merge index_compaction_delete case to index compaction test Release note None Check List (For Author) Test [ ] Regression test [x] Unit Test [ ] Manual test (add detailed scripts or steps below) [ ] No need to test or manual test. Explain why: [ ] This is a refactor/code format and no logic has been changed. [ ] Previous test can cover this change. [ ] No code files have been changed. [ ] Other reason Behavior changed: [x] No. [ ] Yes. Does this need documentation? [x] No. [ ] Yes. Check List (For Reviewer who merge this PR) [ ] Confirm the release note [ ] Confirm test cases [ ] Confirm document [ ] Add branch pick label run buildall run buildall run buildall run buildall
gharchive/pull-request
2024-12-05T12:06:09
2025-04-01T04:55:57.853662
{ "authors": [ "qidaye" ], "repo": "apache/doris", "url": "https://github.com/apache/doris/pull/45056", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1772583361
DRILL-8443: Upgrade netty due to CVE Description https://github.com/apache/drill/security/dependabot/45 Testing CI build Some tests failing with this error - IO Running in secure mode, but config doesn't have a keytab Seems like a Hadoop issue. We may need to upgrade Hadoop or at least review the Hadoop config used in the tests. @pjfanning I think we may have a CI issue. @vvysotskyi @jnturton Any ideas here? Could this be related to that issue we encountered before with the CI and connections? Now tests for this PR fail for other reason (likely due to changes in the PR): Error: Failures: Error: TestResultSetLoaderOmittedValues.testOmittedValuesAtEndWithOverflow:264 Row 0 col d should be null Error: TestResultSetLoaderOverflow.testBatchSizeLimit:164 expected:<16385> but was:<8193> Error: TestOutputBatchSize.testSizerRepeatedRepeatedList:2922 expected:<1048576> but was:<1048560> Error: Errors: Error: TestResultSetLoaderOmittedValues>SubOperatorTest.classTeardown:39 » IllegalState Allocator[ROOT] closed with outstanding buffers allocated (9). @pjfanning @cgivre, @rymarm has made progress on this in #2857. Shall we close this one and try to get that one over the line? sure - I haven't been looking at this
gharchive/pull-request
2023-06-24T10:01:32
2025-04-01T04:55:57.856934
{ "authors": [ "cgivre", "jnturton", "pjfanning", "vvysotskyi" ], "repo": "apache/drill", "url": "https://github.com/apache/drill/pull/2813", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
484423426
Problem of dubbo.xsd file in dubbo2.5.6.jar Environment Dubbo version: 2.5.6 Operating System version: Windows10 Java version: 1.8.0_221 Steps to reproduce this issue Document description 2.5.6 supports netty4. I can't find the parameter client from dubbo.xsd in dubbo2.5.6.jar. I can find the parameter client from dubbo.xsd in dobbo2.5.7.jar. I hope: Please update the document, Describe it in more detail. Modify the dubbo.xsd file in dubbo2.5.6.jar, add parameter "client". Looking forward to official reply, Thank you very much. 2.5.x version is no longer supported, it is recommended to upgrade to the new version.
gharchive/issue
2019-08-23T09:19:05
2025-04-01T04:55:57.861650
{ "authors": [ "CrazyHZM", "alexander-wong-tech" ], "repo": "apache/dubbo", "url": "https://github.com/apache/dubbo/issues/4928", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1813555266
RpcContext proxy What is the purpose of the change Brief changelog Verifying this change Checklist [x] Make sure there is a GitHub_issue field for the change (usually before you start working on it). Trivial changes like typos do not require a GitHub issue. Your pull request should address just this issue, without pulling in other changes - one PR resolves one issue. [ ] Each commit in the pull request should have a meaningful subject line and body. [ ] Write a pull request description that is detailed enough to understand what the pull request does, how, and why. [ ] Check if is necessary to patch to Dubbo 3 if you are work on Dubbo 2.7 [ ] Write necessary unit-test to verify your logic correction, more mock a little better when cross module dependency exist. If the new feature or significant change is committed, please remember to add sample in dubbo samples project. [ ] Add some description to dubbo-website project if you are requesting to add a feature. [ ] GitHub Actions works fine on your own branch. [ ] If this contribution is large, please follow the Software Donation Guide. Codecov Report Merging #12766 (b383d21) into 3.2 (848a577) will increase coverage by 0.70%. The diff coverage is n/a. @@ Coverage Diff @@ ## 3.2 #12766 +/- ## ============================================= + Coverage 68.66% 69.36% +0.70% + Complexity 116 2 -114 ============================================= Files 3653 1650 -2003 Lines 170431 68384 -102047 Branches 28021 9999 -18022 ============================================= - Hits 117021 47437 -69584 + Misses 43284 16366 -26918 + Partials 10126 4581 -5545 see 2036 files with indirect coverage changes :mega: We’re building smart automated test selection to slash your CI/CD build times. Learn more @CrazyHZM PTAL
gharchive/pull-request
2023-07-20T09:26:28
2025-04-01T04:55:57.869072
{ "authors": [ "AlbumenJ", "codecov-commenter" ], "repo": "apache/dubbo", "url": "https://github.com/apache/dubbo/pull/12766", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2252830329
[Feature][3.3] Triple rest useTrailingSlashMatch useSuffixPatternMatch support #14036 What is the purpose of the change support useTrailingSlashMatch useSuffixPatternMatch #14036 Brief changelog Added constructor method and added support for useTrailingSlashMatch useSuffixPatternMatch Verifying this change No Checklist [x] Make sure there is a GitHub_issue field for the change (usually before you start working on it). Trivial changes like typos do not require a GitHub issue. Your pull request should address just this issue, without pulling in other changes - one PR resolves one issue. [ ] Each commit in the pull request should have a meaningful subject line and body. [ ] Write a pull request description that is detailed enough to understand what the pull request does, how, and why. [ ] Check if is necessary to patch to Dubbo 3 if you are work on Dubbo 2.7 [ ] Write necessary unit-test to verify your logic correction, more mock a little better when cross module dependency exist. If the new feature or significant change is committed, please remember to add sample in dubbo samples project. [ ] Add some description to dubbo-website project if you are requesting to add a feature. [ ] GitHub Actions works fine on your own branch. [ ] If this contribution is large, please follow the Software Donation Guide. changelog: support useTrailingSlashMatch useSuffixPatternMatch fix dependencies delete a test because it always failed even I do nothing.This test case has no relation to my modifications(I clone the resource code and run test ,it logged: Wanted but not invoked:errorTypeAwareLogger.info();-> at org.apache.dubbo.config.utils.ConfigValidationUtilsTest.testCheckQosInApplicationConfig(ConfigValidationUtilsTest.java:118) Actually, there were zero interactions with this mock. )
gharchive/pull-request
2024-04-19T12:16:15
2025-04-01T04:55:57.875419
{ "authors": [ "heliang666s" ], "repo": "apache/dubbo", "url": "https://github.com/apache/dubbo/pull/14105", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
733658784
Update dubbo-provider.xml What is the purpose of the change XXXXX Brief changelog XXXXX Verifying this change XXXXX Follow this checklist to help us incorporate your contribution quickly and easily: [x] Make sure there is a GITHUB_issue field for the change (usually before you start working on it). Trivial changes like typos do not require a GITHUB issue. Your pull request should address just this issue, without pulling in other changes - one PR resolves one issue. [ ] Format the pull request title like [Dubbo-XXX] Fix UnknownException when host config not exist #XXX. Each commit in the pull request should have a meaningful subject line and body. [ ] Write a pull request description that is detailed enough to understand what the pull request does, how, and why. [ ] Write necessary unit-test to verify your logic correction, more mock a little better when cross module dependency exist. If the new feature or significant change is committed, please remember to add sample in dubbo samples project. [ ] Run mvn clean install -DskipTests=false & mvn clean test-compile failsafe:integration-test to make sure unit-test and integration-test pass. [ ] If this contribution is large, please follow the Software Donation Guide. could you please provide more detail for this PR? could you please provide more detail for this PR? 不修改该地址zookeeper://127.0.0.1:2181?registry-type=service会报错, introduced by https://github.com/apache/dubbo/commit/13868b920d647d032d4c6bb70ddac069934fa47e#diff-1fceec52425b4ef669f6f4ef7a316637b20cb66a2782a800042f0aadd7858244R28 thanks for your contribution
gharchive/pull-request
2020-10-31T07:30:59
2025-04-01T04:55:57.881137
{ "authors": [ "baymaxxjf", "htynkn" ], "repo": "apache/dubbo", "url": "https://github.com/apache/dubbo/pull/6857", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2242912577
[Feature] 折线图series上的label支持和x轴一样的根据密度分布显示 What problem does this feature solve? 大数据量的时候想显示label,就会出现所有label都显示的请客。如果自己进行密度分布可能和x轴的数据显示不同步 What does the proposed API look like? 增加一个类似xAxis均匀显示的api xAxis: { data: xData, axisLabel: { interval: 'auto' } }, labelLayout { hideOverlap: true} ?
gharchive/issue
2024-04-15T07:25:10
2025-04-01T04:55:57.883200
{ "authors": [ "MatthiasMert", "SONASIDANTE" ], "repo": "apache/echarts", "url": "https://github.com/apache/echarts/issues/19832", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
215357137
Creditbureau configuration module This PR contains code for creditbureau configuration @nikpawar89 Please follow coding style defined by https://google.github.io/styleguide/javaguide.html @nikpawar89 Please let me know whether I understood correctly. With this implementation, I can define credit bureau, credit bureau products and map them to a loan product. So where is actual credit bureau high level implementation so that anybody can attach specific provider implementation? @nazeer1100126 This module is intended for configuring the credit bureau. Once the credit bureau is mapped to a specific loan product, i'll be using this configuration to map the credit check request with specific credit bureau.The code for mapping credit check with credit bureau is not included in this module, it will be included with each specific implementation of the credit bureau in CreditBureauConfigurationAPI.java reponening
gharchive/pull-request
2017-03-20T08:47:06
2025-04-01T04:55:57.886040
{ "authors": [ "nazeer1100126", "nikpawar89" ], "repo": "apache/fineract", "url": "https://github.com/apache/fineract/pull/320", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2488444617
[hotfix] Run schema coordinator logic asynchronously to avoid blocking the main thread This tweaks SchemaRegistry coordinator to let it handle requests / trigger checkpoints in background threads to avoid blocking the main thread. This also adds E2e test case running on various parallelisms. @loserwang1024 Would you like to review this PR when you have time ?
gharchive/pull-request
2024-08-27T06:47:15
2025-04-01T04:55:57.887301
{ "authors": [ "leonardBang", "yuxiqian" ], "repo": "apache/flink-cdc", "url": "https://github.com/apache/flink-cdc/pull/3577", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2716498944
[FLINK-36165][source-connector/mysql] Support capturing snapshot data with conditions Hey, this is an implementation designed to capture snapshot data with filtering conditions. For example, by specifying scan.snapshot.filters: db.user_table:id > 200;, we can synchronize only the user data where the id is greater than 200. issue link: https://issues.apache.org/jira/browse/FLINK-36165 Thanks for @uicosp's great work! It's indeed a long awaited feature. Seems Debezium has a similar option called snapshot.select.statement.overrides, which allows users to project out unwanted columns and filter rows based on custom predicates. As these options aren't available in incremental framework, it would be nice if we could support both row and column level pushdown with similar syntax, since they're both related to tweaking snapshot querying SQL. WDYT? Thanks for @uicosp's great work! It's indeed a long awaited feature. Seems Debezium has a similar option called snapshot.select.statement.overrides, which allows users to project out unwanted columns and filter rows based on custom predicates. As these options aren't available in incremental framework, it would be nice if we could support both row and column level pushdown with similar syntax, since they're both related to tweaking snapshot querying SQL. WDYT? it is also support table api and datastream api with incremental framework
gharchive/pull-request
2024-12-04T03:56:41
2025-04-01T04:55:57.891195
{ "authors": [ "Thorne-coder", "uicosp", "yuxiqian" ], "repo": "apache/flink-cdc", "url": "https://github.com/apache/flink-cdc/pull/3776", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1191428737
[hotfix] Change email/repository notifications to match with Flink Core settings Making sure that notification schema is setup like it's currently for Flink core. Details about this file can be found in https://cwiki.apache.org/confluence/display/INFRA/Git+-+.asf.yaml+features Dank. Dank. You're welcome :)
gharchive/pull-request
2022-04-04T08:21:15
2025-04-01T04:55:57.893120
{ "authors": [ "MartijnVisser", "mbalassi" ], "repo": "apache/flink-kubernetes-operator", "url": "https://github.com/apache/flink-kubernetes-operator/pull/154", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
658954231
[FLINK-16048][avro] Support read/write confluent schema registry avro… … data from Kafka What is the purpose of the change Supports read/write with SQL using schema registry avro format. The format details The factory identifier (or format id) There are 2 candidates now ~ avro-sr: the pattern borrowed from KSQL JSON_SR format [1] avro-confluent: the pattern borrowed from Clickhouse AvroConfluent [2] Personally i would prefer avro-sr because it is more concise and the confluent is a company name which i think is not that suitable for a format name. The format attributes Options required Remark schema-registry.url true URL to connect to schema registry service schema-registry.subject false Subject name to write to the Schema Registry service, required for sink Note: the avro schema string is always inferred from the DDL schema, so user should keep the nullability correct (DDL type default is nullable but avro default is non-nullable). Brief change log Add avro-sr read/write row data format Add tests Verifying this change Added tests. Does this pull request potentially affect one of the following parts: Dependencies (does it add or upgrade a dependency): no The public API, i.e., is any changed class annotated with @Public(Evolving): no The serializers: no The runtime per-record code paths (performance sensitive): no Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Kubernetes/Yarn/Mesos, ZooKeeper: no The S3 file system connector: (yes / no / don't know) no Documentation Does this pull request introduce a new feature? yes If yes, how is the feature documented? not documented +1 for SR and please tag me when you open a documentation PR for this feature +1 for SR and please tag me when you open a documentation PR for this feature Sure, thanks for taking care the document. After a quick glimpse. Could we unify the AvroRowDataDeserializationSchema with ConfluentRegistryAvroRowDataDeserializationSchema and RegistryAvroRowDataSerializationSchema? I really believe we need just a single AvroRowDeserializationSchema for avro for table API. I am quite sure sth like this would work: RowDataDeserializationSchema extends ResultTypeQueryable { private final DeserializationSchema<GenericRecord> nestedSchema; private final DeserializationRuntimeConverter runtimeConverter; private final TypeInformation<RowData> resultType; public AvroRowDataDeserializationSchema2( DeserializationSchema<GenericRecord> nestedSchema, DeserializationRuntimeConverter runtimeConverter, TypeInformation<RowData> resultType) { this.nestedSchema = nestedSchema; this.runtimeConverter = runtimeConverter; this.resultType = resultType; } @Override public void open(InitializationContext context) throws Exception { nestedSchema.open(context); } @Override public RowData deserialize(byte[] message) throws IOException { try { GenericRecord deserialize = nestedSchema.deserialize(message); return (RowData) runtimeConverter.convert(deserialize); } catch (Exception e) { throw new IOException("Failed to deserialize Avro record.", e); } } @Override public boolean isEndOfStream(RowData nextElement) { return false; } @Override public TypeInformation<RowData> getProducedType() { return resultType; } } and then you would use it like this: in AvroFormatFactory: new RowDataDeserializationSchema( AvroDeserializationSchema.forGeneric(AvroSchemaConverter.convertToSchema(rowType)), createRowConverter(rowType), // we would need to move this method to some utils class or to a common abstract class for factories rowDataTypeInfo ); in RegistryAvroFormatFactory: new RowDataDeserializationSchema( ConfluentRegistryAvroDeserializationSchema.forGeneric( AvroSchemaConverter.convertToSchema(rowType), schemaRegistryURL ), createRowConverter(rowType), rowDataTypeInfo ); Thanks for the nice review, i have addressed your comments. Would flink-avro-confluent be a better module name than flink-avro-confluent-registry? IMO registry has nothing to do with the format itself. Would flink-avro-confluent be a better module name than flink-avro-confluent-registry? IMO registry has nothing to do with the format itself. It depends on how we understand it, the confluent avro format is mainly designed for schema registry. Schema registry is a terminology and people always calls "schema registry url"[1], the same for "schema registry subject". [1] https://docs.confluent.io/current/schema-registry/index.html#high-availability-for-single-primary-setup @danny0405 Any reasons all the fields read and written by this format has prefix 'record_' ? @danny0405 @dawidwys Any reasons all the fields read and written by this format has prefix 'record_' ? (I'm using flink sql for this client) I found responsible code probably here but still have problem with this solution: https://github.com/apache/flink/blob/de87a2debde8546e6741390a81f43c032521c3c0/flink-formats/flink-avro/src/main/java/org/apache/flink/formats/avro/typeutils/AvroSchemaConverter.java#L365 It's because of the current strategy to infer the Avro schema is convert from the CREATE TABLE DDL, and there is no way to get the record name here. So we put a constant record as a prefix. The record write out all have explicit field name, but the type should be compatible. @danny0405 Problem is that with this strategy I'm unable to read anything from Kafka using Confluent Registry. Example: I have data in Kafka with following value schema: { "type": "record", "name": "myrecord", "fields": [ { "name": "f1", "type": "string" } ] } I'm creating table using this avro-confluent format: create table `test` ( `f1` STRING ) WITH ( 'connector' = 'kafka', 'topic' = 'test', 'properties.bootstrap.servers' = 'localhost:9092', 'properties.group.id' = 'test1234', 'scan.startup.mode' = 'earliest-offset', 'format' = 'avro-confluent' 'avro-confluent.schema-registry.url' = 'http://localhost:8081' ); When trying to select data I'm getting error: SELECT * FROM test; [ERROR] Could not execute SQL statement. Reason: org.apache.avro.AvroTypeException: Found myrecord, expecting record, missing required field record_f1 Thanks, i think this is a bug, i have logged an issue there. See https://issues.apache.org/jira/browse/FLINK-19779 I have fired a fix https://github.com/apache/flink/pull/13763/files, can you help check if possible @maver1ck :) I will check... mvn is running OK. It's working. I'm able to read data. @danny0405 I think we have one more problem. When Flink is creating schema in registry nullability is not properly set for logical types. Examples. Table: create table `test_logical_null` ( `string_field` STRING, `timestamp_field` TIMESTAMP(3) ) WITH ( 'connector' = 'kafka', 'topic' = 'test-logical-null', 'properties.bootstrap.servers' = 'localhost:9092', 'properties.group.id' = 'test12345', 'scan.startup.mode' = 'earliest-offset', 'format' = 'avro-confluent', -- Must be set to 'avro-confluent' to configure this format. 'avro-confluent.schema-registry.url' = 'http://localhost:8081', -- URL to connect to Confluent Schema Registry 'avro-confluent.schema-registry.subject' = 'test-logical-null' -- Subject name to write to the Schema Registry service; required for sinks ) Schema: { "type": "record", "name": "record", "fields": [ { "name": "string_field", "type": [ "string", "null" ] }, { "name": "timestamp_field", "type": { "type": "long", "logicalType": "timestamp-millis" } } ] } For not null fields: create table `test_logical_notnull` ( `string_field` STRING NOT NULL, `timestamp_field` TIMESTAMP(3) NOT NULL ) WITH ( 'connector' = 'kafka', 'topic' = 'test-logical-notnull', 'properties.bootstrap.servers' = 'localhost:9092', 'properties.group.id' = 'test12345', 'scan.startup.mode' = 'earliest-offset', 'format' = 'avro-confluent', -- Must be set to 'avro-confluent' to configure this format. 'avro-confluent.schema-registry.url' = 'http://localhost:8081', -- URL to connect to Confluent Schema Registry 'avro-confluent.schema-registry.subject' = 'test-logical-notnull-value' -- Subject name to write to the Schema Registry service; required for sinks ); Schema { "type": "record", "name": "record", "fields": [ { "name": "string_field", "type": "string" }, { "name": "timestamp_field", "type": { "type": "long", "logicalType": "timestamp-millis" } } ] } As you can see for string_field we have proper union with null (for nullable field). For timestamp_field in both examples union is missing. @maver1ck , you are right, we ignore the nullability of TIMESTAMP_WITHOUT_TIME_ZONE, DATE, and TIME_WITHOUT_TIME_ZONE and Decimal, would fix it altogether in this PR. Thanks @danny0405 I have updated the fix, @maver1ck , please check if you have time. @danny0405 I see code review is still in progress. Could you please let me know when this PR would be polished a little ?
gharchive/pull-request
2020-07-17T07:18:46
2025-04-01T04:55:57.914587
{ "authors": [ "KurtYoung", "danny0405", "maver1ck", "sjwiesman" ], "repo": "apache/flink", "url": "https://github.com/apache/flink/pull/12919", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
980068988
[FLINK-22002][tests] Let taskmanager.slot.timeout fall back to akka.ask.timeout This commit lets taskmanager.slot.timeout fall to akka.ask.timeout so that the MiniCluster will run by default with a 5 min taskmanager.slot.timeout. This should harden against CI pauses. Thanks for the review @zentol. Merging this PR now.
gharchive/pull-request
2021-08-26T10:09:30
2025-04-01T04:55:57.916618
{ "authors": [ "tillrohrmann" ], "repo": "apache/flink", "url": "https://github.com/apache/flink/pull/16995", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1748973167
[FLINK-30629][Client/Job Submission] Increase clientHeartbeatTimeout to 1 second What is the purpose of the change Increase clientHeartbeatTimeout to 1 second to avoid shutting down the job in ClientHeartbeatTest. Brief change log Increase clientHeartbeatTimeout to 1 second Verifying this change This change is a trivial rework / code cleanup without any test coverage. Does this pull request potentially affect one of the following parts: Dependencies (does it add or upgrade a dependency): (no) The public API, i.e., is any changed class annotated with @Public(Evolving): (no) The serializers: (no) The runtime per-record code paths (performance sensitive): (no) Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Kubernetes/Yarn, ZooKeeper: (no) The S3 file system connector: (no) Documentation Does this pull request introduce a new feature? (no) thanks @Myracle
gharchive/pull-request
2023-06-09T02:07:41
2025-04-01T04:55:57.920793
{ "authors": [ "Myracle", "snuyanzin" ], "repo": "apache/flink", "url": "https://github.com/apache/flink/pull/22742", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1939334554
[FLINK-32073][checkpoint] Implement file merging in snapshot What is the purpose of the change As one part of FLIP-306, this PR enables file merging in snapshot by implementing the corresponding output&input streams for FileMergingSnapshotManager Brief change log Add createCheckpointStateOutputStream in FileMergingSnapshotManager. Add output&input streams for file merging checkpoints Add SegmentFileStateHandle to represent a state that is written into a file segment. Verifying this change Please make sure both new and modified tests in this PR follows the conventions defined in our code quality guide: https://flink.apache.org/contributing/code-style-and-quality-common.html#testing This change added several UT tests in: FileMergingSnapshotManagerTest FileMergingCheckpointStateOutputStreamTest FsSegmentDataInputStreamTest. Does this pull request potentially affect one of the following parts: Dependencies (does it add or upgrade a dependency): (no) The public API, i.e., is any changed class annotated with @Public(Evolving): (no) The serializers: (no) The runtime per-record code paths (performance sensitive): (no) Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Kubernetes/Yarn, ZooKeeper: (yes) The S3 file system connector: (no) Documentation Does this pull request introduce a new feature? (no) If yes, how is the feature documented? (not applicable / docs / JavaDocs / not documented) Thanks @fredia and @Zakelly for your comments. Sorry for the late reply. I have addressed your comments. Would you please take another look? Thanks @fredia and @Zakelly for your comments. Sorry for the late reply. I have addressed your comments. Would you please take another look? Thanks for your effort. I have no comments but a minor one, PTAL, thanks. CI seems Failed with "99d7b97" @flinkbot run azure @curcur The CI failure has been addressed. (The file-merging checkpoint was unexpectedly enabled.) Would you please take another look? The PR looks GTM. I left a few comments. squash the commits. There are conflicts, resolve them and after turn green, ping me, and I will merge them. @curcur I have addressed your comments by removing the unused codes and squashing the commits. The PR is now free of conflicts and has passed the CI tests. PTAL.
gharchive/pull-request
2023-10-12T07:02:28
2025-04-01T04:55:57.930015
{ "authors": [ "AlexYinHan", "Zakelly", "curcur" ], "repo": "apache/flink", "url": "https://github.com/apache/flink/pull/23514", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1993147345
[FLINK-33410] Implement type inference for OVER function What is the purpose of the change It implements type inference for OVER, STREAM_RECORD and THROW_EXCEPTION functions. Verifying this change all available tests pass added tests for OVER type inference Does this pull request potentially affect one of the following parts: Dependencies (does it add or upgrade a dependency): (yes / no) The public API, i.e., is any changed class annotated with @Public(Evolving): (yes / no) The serializers: (yes / no / don't know) The runtime per-record code paths (performance sensitive): (yes / no / don't know) Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Kubernetes/Yarn, ZooKeeper: (yes / no / don't know) The S3 file system connector: (yes / no / don't know) Documentation Does this pull request introduce a new feature? (yes / no) If yes, how is the feature documented? (not applicable / docs / JavaDocs / not documented) LGTM, only left one minor comments
gharchive/pull-request
2023-11-14T16:52:10
2025-04-01T04:55:57.935569
{ "authors": [ "dawidwys", "liuyongvs" ], "repo": "apache/flink", "url": "https://github.com/apache/flink/pull/23713", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }