id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
1311859168
Suggestion - remove rust installation script I think automatic installation of rust toolchain should be completely removed from this tool, because - This is too low-level-system thing and it's always dangerous If someone wants to develop Rust, he should already have the toolchain Even if we have a dev that this is the first time he ever sees Rust, the best way to learn new language workflows and toolchains is from the actual Rust docs site, there he'll find explanation about what is rustup and targets and so on.. We shouldn't do any magic installations for him We only should have a message saying that you need wasm32-unknown-unknown target and link to the Rust docs page about it. @volovyk-s @gagdiez wdyt? I agree, lets simply detect if the user has Rust installed, and warn them if not. Ok, we can get rid of this magic. But let's make sure we point users to the Rust doc that will help them to install everything they need.
gharchive/issue
2022-07-20T20:27:25
2025-04-01T06:39:43.470378
{ "authors": [ "gagdiez", "maxhr", "volovyk-s" ], "repo": "near/create-near-app", "url": "https://github.com/near/create-near-app/issues/1851", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
894506343
SIGSEGV error when using Ledger Probably originated from Ledger itself. To reproduce: near send acc1.near acc2.near --useLedgerKey. Reproducable on both near-cli 1.6.0 and near-cli 2.0.0. The transaction is successfully executed, the error occurs after it. Also related to #758. I believe this and #758 are the same issue, so closing this as a duplicate of it.
gharchive/issue
2021-05-18T15:27:30
2025-04-01T06:39:43.472570
{ "authors": [ "gonta71", "nagisa", "volovyk-s" ], "repo": "near/near-cli", "url": "https://github.com/near/near-cli/issues/762", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1053069373
Allow custom HD path beyond single number Currently when recovering wallet, user can only change the last number in the derivation paths. Users might have keys on other derivation paths, Wallet should support that. This may be confusing for some users and is normally not offered by other wallets. Would it be acceptable if this can be managed via dev console, e.g. a function that takes as arguments the path integers?
gharchive/issue
2021-11-14T23:20:42
2025-04-01T06:39:43.473691
{ "authors": [ "ilblackdragon", "stefanopepe" ], "repo": "near/near-wallet", "url": "https://github.com/near/near-wallet/issues/2250", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1150460758
Feat farming validator UI for wallet testing Internal testing branch for the new stake farming feature as implemented on #2398 Updated LOCKUP_ACCOUNT_ID_SUFFIX on render preview to lockup.devnet. We can now use https://testnet.lockup.tech/ to deploy lockups for testing. Just update wallet.testnet.near.org to near-wallet-pr-2507.onrender.com when logging in and signing transactions. Validators for testing: zentriav2.factory.colorpalette.testnet domanodes.factory.colorpalette.testnet testnet preview url To test, the following farm validators can be used: zentriav2.factory.colorpalette.testnet domanodes.factory.colorpalette.testnet To test lockup staking, use https://testnet.lockup.tech/ to add a lockup to your account and update wallet.testnet.near.org to near-wallet-pr-2512.onrender.com when logging in and signing transactions.
gharchive/pull-request
2022-02-25T13:44:13
2025-04-01T06:39:43.478531
{ "authors": [ "esaminu" ], "repo": "near/near-wallet", "url": "https://github.com/near/near-wallet/pull/2507", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1935455423
Feature/migrate GitHub to existing framework Resolves #265 Still a WIP Testnet Preview Acceptance Criteria: [x] GitHub plugin aligns with the new framework (https://github.com/near/neardevhub-widgets/issues/253). So an admin can enable, configure, or remove the plugin in the community settings page. [x] If a community admin enables the GitHub plugin, the GitHub tab shows up on the community page navigation. [x] Admin can configure the plugin within the community settings page, not from the old GitHub tab. Remove any instances from the GitHub tab to configure the page from the tab. [x] If a community admin enables the GitHub plugin, they can click configure to customize the the tab name as well as the existing GitHub specific fields, including: [x] Title* [x] GitHub Repository URL*. Note: We will need to add a prefix here (similar to the about links section) that says: "https://github.com/" to make it clear what part of the GitHub URL the user needs to enter. [x] Ticket type* (Issue, Pull Request) [x] Ticket State* [x] Description* [x] New Column* (plus the supporting fields) [x] Community admin should have a way to preview GitHub board from the configure settings. [ ] If an admin enables the GitHub plugin, the GitHub field(s) with * above are required [ ] There is no disruption to old/existing GitHub plugins [ ] There is migration option provided for existing communities to transition to the new framework Notes: In order to edit the tab name issue #283 has to be finished implementing that functionality. I'm closing this draft, it is already handled by PR 344
gharchive/pull-request
2023-10-10T14:07:21
2025-04-01T06:39:43.489096
{ "authors": [ "Tguntenaar" ], "repo": "near/neardevhub-widgets", "url": "https://github.com/near/neardevhub-widgets/pull/291", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1600691397
[Bug] Pressing control in sign up address checking box causes re-checking Bug Description If you press the control key on your keyboard while focused on the address lookup box it will check again, even though the content has not changed Steps to Reproduce head to https://home.omg.lol/sign-up enter something in the box press ctrl observe it check again this is the most minor thing ever but i noticed it so youre getting an issue
gharchive/issue
2023-02-27T08:13:00
2025-04-01T06:39:43.513955
{ "authors": [ "vtheskeleton" ], "repo": "neatnik/omg.lol", "url": "https://github.com/neatnik/omg.lol/issues/616", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2324143406
Home page updates Updates to the home page, mainly stylistic changes, but also text updates and iconography. Reference Issues or PRs https://gitlab.jatic.net/jatic/team-metrostar/t-e-platform/-/issues/595 This update addresses the issues in the attached ticket. What does this implement/fix? This implementation fixes stylistic and design issues Put a x in the boxes that apply [ ] Bug fix (non-breaking change which fixes an issue) [x] New feature (non-breaking change which adds a feature) [ ] Breaking change (fix or feature that would cause existing features not to work as expected) [ ] Documentation Update [x] Code style update (formatting, renaming) [x] Refactoring (no functional changes, no API changes) [ ] Build related changes [ ] Other (please describe): Testing [x] Did you test the pull request locally? [x] Did you add new tests? Documentation Access-centered content checklist Text styling [ ] The content is written with plain language (where relevant). [ ] If there are headers, they use the proper header tags (with only one level-one header: H1 or # in markdown). [ ] All links describe where they link to (for example, check the Nebari website). [ ] This content adheres to the Nebari style guides. Non-text content [ ] All content is represented as text (for example, images need alt text, and videos need captions or descriptive transcripts). [ ] If there are emojis, there are not more than three in a row. [ ] Don't use flashing GIFs or videos. [ ] If the content were to be read as plain text, it still makes sense, and no information is missing. Any other comments? Code-wise looks great, just 1 thing to cleanup. Also, please include the following in this PR: Fix the header opacity Fix the row spacing between App rows Ensure the home navigation button can be themed √, √, √
gharchive/pull-request
2024-05-29T19:54:48
2025-04-01T06:39:43.523036
{ "authors": [ "kildre" ], "repo": "nebari-dev/jhub-apps", "url": "https://github.com/nebari-dev/jhub-apps/pull/311", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2184350542
Fix typo causing bad rendering (extra backtick) Before: After: Good catch! Thanks @krassowski 🚀
gharchive/pull-request
2024-03-13T15:51:17
2025-04-01T06:39:43.524911
{ "authors": [ "krassowski", "marcelovilla" ], "repo": "nebari-dev/nebari-docs", "url": "https://github.com/nebari-dev/nebari-docs/pull/419", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2010684136
feat: support specify space by param, fix #265 For example: Boolean spaceFromParam(@Param("specifySpace") String specifySpace); <select id="spaceFromParam" space="${specifySpace}" spaceFromParam="true"> RETURN true; </select> If the specifySpace is not current space, NgBatis will automatically switch. Can reduce the use of use space in ngql. Codecov Report Attention: 23 lines in your changes are missing coverage. Please review. Comparison is base (d781052) 0.00% compared to head (16da500) 0.00%. Report is 1 commits behind head on master. Files Patch % Lines ...ebula/contrib/ngbatis/io/MapperResourceLoader.java 0.00% 9 Missing :warning: .../org/nebula/contrib/ngbatis/proxy/MapperProxy.java 0.00% 7 Missing :warning: ...g/nebula/contrib/ngbatis/config/ParseCfgProps.java 0.00% 4 Missing :warning: ...org/nebula/contrib/ngbatis/models/MethodModel.java 0.00% 3 Missing :warning: :exclamation: Your organization needs to install the Codecov GitHub app to enable full functionality. Additional details and impacted files @@ Coverage Diff @@ ## master #267 +/- ## ====================================== Coverage 0.00% 0.00% ====================================== Files 75 75 Lines 2568 2583 +15 Branches 278 279 +1 ====================================== - Misses 2568 2583 +15 :umbrella: View full report in Codecov by Sentry. :loudspeaker: Have feedback on the report? Share it here.
gharchive/pull-request
2023-11-25T14:43:35
2025-04-01T06:39:43.535997
{ "authors": [ "CorvusYe", "codecov-commenter" ], "repo": "nebula-contrib/ngbatis", "url": "https://github.com/nebula-contrib/ngbatis/pull/267", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2391119937
Improve formatting of multiline ERB This PR does 2 things: Subtract the indentation level from Syntax Tree line width This could be even further improved: line length should take ERB tag start and end tag overhead in consideration Change the way multiline ERB is formatted. It will do the following: If formatted ruby yields more than 1 line, then: The ERB tag start and end will be on separate lines The entire Ruby code will be indented with an additional 2 spaces (Due to the fact that I was adding to repetitive logic, I chose to optimize the case statement. If this is undesirable and we prefer even more repetition, I'm OK with having that refactored back) Oh @elia I completely missed #7 and went my own way with this one.
gharchive/pull-request
2024-07-04T15:38:49
2025-04-01T06:39:43.538769
{ "authors": [ "pelletencate" ], "repo": "nebulab/erb-formatter", "url": "https://github.com/nebulab/erb-formatter/pull/56", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
335129095
there is no flexibility with different python versions. Originally reported by Abel Asefa (Bitbucket: abelandk, GitHub: abelandk) When the main process is running with python3. But some other subprocesses were running with python2, it could not get the correct result and produce a SyntaxError. #!python File ".../python/python3.5.2/python3.5/site-packages/coverage/parser.py", line 363, in __init__ self.code = compile_unicode(text, filename, "exec") File ".../python/python3.5.2/lib/python3.5/site-packages/coverage/phystokens.py", line 286, in compile_unicode code = compile(source, filename, mode) File ".../constructor.py", line 130 except TypeError, exc: ^ SyntaxError: invalid syntax During handling of the above exception, another exception occurred: File "/lab/python/python3.5.2/lib/python3.5/site-packages/coverage/control.py", line 1095, in html_report return reporter.report(morfs) File "/lab/python/python3.5.2/lib/python3.5/site-packages/coverage/html.py", line 139, in report self.report_files(self.html_file, morfs, self.config.html_dir) File "/lab/python/python3.5.2/lib/python3.5/site-packages/coverage/report.py", line 91, in report_files report_fn(fr, self.coverage._analyze(fr)) File "/lab/python/python3.5.2/lib/python3.5/site-packages/coverage/control.py", line 968, in _analyze return Analysis(self.data, it) File "/lab/python/python3.5.2/lib/python3.5/site-packages/coverage/results.py", line 19, in __init__ self.statements = self.file_reporter.lines() File "/lab/python/python3.5.2/lib/python3.5/site-packages/coverage/python.py", line 186, in lines return self.parser.statements File "/lab/python/python3.5.2/lib/python3.5/site-packages/coverage/python.py", line 181, in parser self._parser.parse_source() File "/lab/python/python3.5.2/lib/python3.5/site-packages/coverage/parser.py", line 237, in parse_source self._raw_parse() File "/lab/python/python3.5.2/lib/python3.5/site-packages/coverage/parser.py", line 206, in _raw_parse self.raw_statements.update(self.byte_parser._find_statements()) File "/lab/python/python3.5.2/lib/python3.5/site-packages/coverage/parser.py", line 96, in byte_parser self._byte_parser = ByteParser(self.text, filename=self.filename) File "/lab/python/python3.5.2/lib/python3.5/site-packages/coverage/parser.py", line 367, in __init__ filename, synerr.msg, synerr.lineno coverage.misc.NotPython: Couldn't parse '.../python2/yaml/constructor.py' as Python source: 'invalid syntax' at line 130 Bitbucket: https://bitbucket.org/ned/coveragepy/issue/655 @abelandk: any more information?
gharchive/issue
2018-04-25T07:41:00
2025-04-01T06:39:43.582354
{ "authors": [ "nedbat" ], "repo": "nedbat/coveragepy", "url": "https://github.com/nedbat/coveragepy/issues/655", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1367526018
Adding 678-Valid-Parenthesis-String.cpp File(s) Modified: 678-Valid-Parenthesis-String.cpp Language(s) Used: C++ Submission URL: _https://leetcode.com/submissions/detail/795416800/_ Thanks, @FahadulShadhin!
gharchive/pull-request
2022-09-09T09:10:48
2025-04-01T06:39:43.587120
{ "authors": [ "Ahmad-A0", "FahadulShadhin" ], "repo": "neetcode-gh/leetcode", "url": "https://github.com/neetcode-gh/leetcode/pull/1086", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2426229379
Actions don't show warning when station is removed from TrackAudio To reproduce: Connect TrackAudio Add SEA_GND Add a station state button that reflects SEA_GND, notice that it doesn't show the warning icon which is correct Delete SEA_GND from TrackAudio Result: SEA_GND station state button still looks ok Expected result: SEA_GND station state button should show the warning icon This will require changes on TrackAudio as well
gharchive/issue
2024-07-23T22:32:08
2025-04-01T06:39:43.600484
{ "authors": [ "neilenns" ], "repo": "neilenns/streamdeck-trackaudio", "url": "https://github.com/neilenns/streamdeck-trackaudio/issues/163", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
31778205
Sometimes incorrectly parsing output I have some code that is currently calling getMAC() about once a second. Every once in a while it will return 'eth0' instead of the MAC address. It doesn't happen very often compared to how often it's called, but it happens a few times a day. This is running on a Debian Linux machine, btw. We're hitting this problem too! No clue why it seems to be doing this. The first time i tired it returned the interface name every time, but on the second try it returns the correct mac address... If you have any insights into this let me know and i'll be happy to open a PR. We'll probably keep investigating. Figured it out! PR incoming ;) https://github.com/nekuz0r/node-arp/pull/8
gharchive/issue
2014-04-18T02:48:51
2025-04-01T06:39:43.619412
{ "authors": [ "ZackMattor", "jwarkentin" ], "repo": "nekuz0r/node-arp", "url": "https://github.com/nekuz0r/node-arp/issues/5", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
498693187
Suggestion: Rename of "Node" drop-down selection Suggest to change "Node" to "Network Type/ Network Name" or "Generationhash" to reflect that those nodes could be from different network. Maybe Network:Node? e.g. Mainnet:http://13.114.200.132:3000/ hey there @ivyfung1 ; I would agree that there could be some information about the network name in the list. But I think it is important to keep the node IPs there as this is the information that is relevant (which node is the explorer using to read the data I see.)
gharchive/issue
2019-09-26T06:48:23
2025-04-01T06:39:43.629691
{ "authors": [ "evias", "ivyfung1" ], "repo": "nemfoundation/nem2-explorer", "url": "https://github.com/nemfoundation/nem2-explorer/issues/103", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
756575454
Harvesting confirmation view password input Harvesting confirmation view password input is not visible and we need to scroll. For other transaction types input is always visible. Please watch video: https://share.getcloudapp.com/JruqbPel Same with harvesting activation. @NikolaiYurchenko could you check? @NikolaiYurchenko I think we have still inconsistent/bad looking transaction confirmation pop-ups. create aggregate pop-up -> https://share.getcloudapp.com/yAuZJbYn?utm_source=show create aggregate (with multisig) -> https://share.getcloudapp.com/d5uPgp2L (three scrools) how it looks on cosigner side (to sign) -> https://share.getcloudapp.com/p9urewoE (details cutted on top) how it looks on cosigner side (which already signed tx) -> https://share.getcloudapp.com/NQuKW19m Fixed.
gharchive/issue
2020-12-03T20:40:30
2025-04-01T06:39:43.633739
{ "authors": [ "cryptoBeliever" ], "repo": "nemgrouplimited/symbol-desktop-wallet", "url": "https://github.com/nemgrouplimited/symbol-desktop-wallet/issues/769", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
606338908
Mask not working as expected I have the following json { "tom": 3, "data1": { "first": "Hello", "second": {"third": "salude", "fourth": "nvm"} }, "data2": { "first": "Bye", "second": {"third": "Cheers", "fourth": "nvm"} } } And I'm applying the follwing mask on it. "data1(first,second/third),data2(first,second/third)" I expected to see the, { "data1": { "first": "Hello", "second": {"third": "salude"} }, "data2": { "first": "Bye", "second": {"third": "Cheers"} } } but I'm only getting { "data1": { "first": "Hello", "second": {"third": "salude"} } } Can you help me out? @MathiasHaudgaard thanks for letting me know! That appears to be a bug. I can reproduce it. Ok, please let me know when it's fixed. Thanks! :smiley: @nemtsov It seems like the error comes from the buildTree functions. It exits the scope in the wrong order. When it comes to ")" it pops "/" and now it cant exit the last scope because of if (peek && peek.tag === "/") doesn't match with "(" I'm bumping into the same error right now. Are there any updates here? any updates? Is the library abandonware? Hey folks, although I don’t have a ton of time to work on this library myself, I’m more than happy to review PRs and help guide people interested in making contributions to it. This the only bug i could find so far. Can you fix it? What i found is, if / is not at the end of the query, what is next is going to be ignored. Not working: data1(first,second/third),data2(first,second/third) Working: data1(first,second),data2(first,second/third) This is the only library that supports partial responses in plan JS. It is a shame just to abandon it. This is resolved in v1.0.4 thanks for reporting it @MathiasHaudgaard
gharchive/issue
2020-04-24T14:04:49
2025-04-01T06:39:43.645893
{ "authors": [ "MathiasHaudgaard", "alexrainman", "juan-restrepo", "mihanizm56", "nemtsov" ], "repo": "nemtsov/json-mask", "url": "https://github.com/nemtsov/json-mask/issues/72", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
606776947
CDC and Kafka Log Compaction Hi everyone !! This issue is somehow similar to this one #272 but the focus here is on Neo4j Streams Source for CDC in tandem with Kafka Log Compaction rather than on Neo4j Streams Procedures. When using Change Data Capture with Kafka, one of the most common use cases is managing Kafka partitions as logs. In this scenario, Log Compaction is somehow necessary for CDC: we need a complete dataset but we must also avoid unlimited growth in the size of partitions. From the Kafka documentation: Log compaction ensures that Kafka will always retain at least the last known value for each message key within the log of data for a single topic partition. It addresses use cases and scenarios such as restoring state after application crashes or system failure or reloading caches after application restarts during operational maintenance. On data retention: So far we have described only the simpler approach to data retention where old log data is discarded after a fixed period or when the log reaches some predetermined size. This works well for temporal event data such as logging where each record stands alone. However, an important class of data streams are the log of changes to keyed, mutable data (for example, the changes to a database table). The current implementation of Neo4j Streams prevents this use cases since the message key is a combination of the transaction id and the id of the event inside the transaction, as shown in the following code (KafkaEventRouter.kt- line 86 - sendEvent method): private fun sendEvent(partition: Int, topic: String, event: StreamsTransactionEvent) { if (log.isDebugEnabled) { log.debug("Trying to send a transaction event with txId ${event.meta.txId} and txEventId ${event.meta.txEventId} to kafka") } val producerRecord = ProducerRecord(topic, partition, System.currentTimeMillis(), "${event.meta.txId + event.meta.txEventId}-${event.meta.txEventId}", JSONUtils.writeValueAsBytes(event)) send(producerRecord) } As far as I know, this will generate a sequence of unique keys, preventing Log Compaction. Instead, choosing the node or relationship id as the message key ("${event.payload.id}") would enable Log Compaction on the Kafka side. However, I'm pretty sure there are specific reasons for creating the key that way: maybe the message keys are used to know where we are in the Neo4j log, to recover after a failure, but I'm not sure. Can you explain why the keys are generated in this way? Is it possible to enable Log Compaction in Neo4j Streams right now? If the answer is no, I think this is a must-have feature for Change Data Capture. Thanks !! Considered alternatives While writing this issue, I actually came up with a possible alternative solution (the same also applies to the issue #304): one can simply create a microservice using Kafka Streams to add custom logic and cope with order guarantees and log compaction. However, I am still convinced that these problems should be addressed within Neo4j Streams. @Solom00n thanks for the great analysis, we can do it via a configuration param in order to keep both options @mroiter-larus is this issue close-able? @moxious this had been put on hold that i know of Left open pending duplicate functionality PR agaist 3.5 branch. Issue is closeable when 3.5 branch is merged with same fix/improvement.
gharchive/issue
2020-04-25T14:07:23
2025-04-01T06:39:43.656714
{ "authors": [ "Solom00n", "conker84", "moxious", "mroiter-larus" ], "repo": "neo4j-contrib/neo4j-streams", "url": "https://github.com/neo4j-contrib/neo4j-streams/issues/305", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
142292257
Can't run integration unittests I have spend several hours trying to get the integration unittests to work. I'm running on a clean Windows 10 with Visual Studio 2015 Community and JRE 8. I'm running both VS, cosole and PS as administrator. I have tracked down the issue to when the PS scripts tries to start the neo4j servies by debugging the unittest. Then I tried running the PS scripts manually and got the same error. Here is the log: PS C:\Sources\neo4j\original\neo4j-dotnet-driver\Neo4j.Driver\Neo4j.Driver.IntegrationTests\bin\target\neo4j\neo4j-commu nity-3.0.0-RC1\bin> Invoke-Neo4j start -v VERBOSE: Neo4j Root is 'C:\Sources\neo4j\original\neo4j-dotnet-driver\Neo4j.Driver\Neo4j.Driver.IntegrationTests\bin\target\neo4j\neo4j-community-3.0.0-RC1' VERBOSE: Neo4j Server Type is 'Community' VERBOSE: Neo4j Version is '3.0.0-RC1' VERBOSE: Neo4j Database Mode is '' VERBOSE: Start command specified VERBOSE: Neo4j Windows Service Name is neo4j VERBOSE: Starting the service. This can take some time... Invoke-Neo4j : Failed to start service 'Neo4j Graph Database - neo4j (neo4j)'. At line:1 char:1 Invoke-Neo4j start -v + CategoryInfo : NotSpecified: (:) [Write-Error], WriteErrorException + FullyQualifiedErrorId : Microsoft.PowerShell.Commands.WriteErrorException,Invoke-Neo4j It works if I get it to run as console: PS C:\Sources\neo4j\original\neo4j-dotnet-driver\Neo4j.Driver\Neo4j.Driver.IntegrationTests\bin\target\neo4j\neo4j-community-3.0.0-RC1\bin> Invoke-Neo4j console -v VERBOSE: Neo4j Root is 'C:\Sources\neo4j\original\neo4j-dotnet-driver\Neo4j.Driver\Neo4j.Driver.IntegrationTests\bin\target\neo4j\neo4j-community-3.0.0-RC1' VERBOSE: Neo4j Server Type is 'Community' VERBOSE: Neo4j Version is '3.0.0-RC1' VERBOSE: Neo4j Database Mode is '' VERBOSE: Console command specified VERBOSE: Java detected at 'C:\Program Files\Java\jre1.8.0_73\bin\java.exe' VERBOSE: Java version detected as '1.8' VERBOSE: Starting Neo4j as a console with command line C:\Program Files\Java\jre1.8.0_73\bin\java.exe -cp "C:\Sources\neo4j\original\neo4j-dotnet-driver\Neo4j.Driver\Neo4j.Driver.IntegrationTests\bin\target\neo4j\neo4j-community-3.0.0-RC1/lib/;C:\Sources\neo4j\original\neo4j-dotnet-driver\Neo4j.D river\Neo4j.Driver.IntegrationTests\bin\target\neo4j\neo4j-community-3.0.0-RC1/plugins/" -server -Dorg.neo4j.config.file=conf/neo4j.conf -Dlog4j.configuration=file:conf/log4j.properties -Dneo4j.ext.udc.source=zip-powershell -Dorg.neo4j.cluster.logdirectory=data/log -Dorg.neo4j.config.file=conf/neo4j.conf -XX:+UseG1GC -XX:-OmitStackTraceInFastThrow -XX:hashCode=5 -XX:+AlwaysPreTouch -XX:+UnlockExperimentalVMOptions -XX:+TrustFinalNonStaticFields -XX:+DisableExplicitGC -Dunsupported.dbms.udc.source=zip -Dfile.encoding=UTF-8 org.neo4j.server.CommunityEntryPoint 2016-03-21 09:06:23.118+0000 INFO Starting... 2016-03-21 09:06:31.680+0000 INFO Started. 2016-03-21 09:06:34.633+0000 INFO Remote interface available at http://localhost:7474/ I have also tried starting the service manually from Windows Services, but got this message: The Neo4j Graph Database - neo4j service on Local Computer started and then stopped. Some services stop automatically if they are not in use by other services or programs. I would really like to get them running, so I can contribute to the project :) Any suggestions? Hello, I looked into your problem yesterday and it seems you have some problem with Write-Error command used in our powershell script. The build also have some problem to use the recent powershell script so I am also wondering if I should just switch to use the bat file instead to avoid these changes to disturb the normal driver code. I am working on this problem and I will come back to you once I got a solution or work around. Will ping you when the build is ready for your PR. Finally thanks a lot for the PRs. Thanks for you feed back! Using the BAT file might fix the issues I have seen :) As you might have seen by now, I have created a PR that does not use PS or the Windows Service for running the Neo4j server. I have tested this on 3 Window PCs and worked on them all - When using the original code, integration tests did not work on any of the 3 PCs. You can consider using this when not running on the build server (eg local testing) and use the original when running on the build server :) https://github.com/neo4j/neo4j-dotnet-driver/pull/33 As the pr is already merged in, I will close this issue now. Feel free to reopen it when you got problem again,
gharchive/issue
2016-03-21T09:14:45
2025-04-01T06:39:43.672909
{ "authors": [ "IngvarKofoed", "zhenlineo" ], "repo": "neo4j/neo4j-dotnet-driver", "url": "https://github.com/neo4j/neo4j-dotnet-driver/issues/28", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
332860657
Feature request: General async support Since there is a release of the async bolt driver. It would be very useful to have it in OGM too. https://neo4j.com/docs/developer-manual/current/drivers/sessions-transactions/#_asynchronous_programming Thanks for your input. Your options right now is using OGM through SDN and it's Spring based async support: Async query results. We have also spiked some ideas around full reactive support: https://github.com/michael-simons/neo4j-reactive-java-client I'm closing this issue for the time being. This doesn't mean we're not implementing async, we are tracking this already and having several such tickets around doesn't help. Thank you.
gharchive/issue
2018-06-15T17:41:59
2025-04-01T06:39:43.683560
{ "authors": [ "michael-simons", "wem" ], "repo": "neo4j/neo4j-ogm", "url": "https://github.com/neo4j/neo4j-ogm/issues/495", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
233571109
Applying each after variable length relationship ignores rel_length When each is called after an association with variable length the rel_length parameter is ignored. Additional information which could be helpful if relevant to your issue: Code example (inline, gist, or repo) [31] pry(main)> >> r.subordinates(rel_length: { min: 0 }).each.count Role#subordinates MATCH (previous:`Role`) WHERE (ID(previous) = {ID_previous}) OPTIONAL MATCH (previous)<-[rel1:`manager`]-(next:`Role`) RETURN ID(previous), collect(next) | {:ID_previous=>14930} HTTP REQUEST: 4ms POST http://localhost:7474/db/data/transaction/commit (1 bytes) => 6 [32] pry(main)> >> r.subordinates(rel_length: { min: 0 }).count Role#subordinates MATCH (role14930) WHERE (ID(role14930) = {ID_role14930}) MATCH (role14930)<-[rel1:`manager`*0..]-(result_subordinates:`Role`) RETURN count(result_subordinates) AS result_subordinates | {:ID_role14930=>14930} HTTP REQUEST: 7ms POST http://localhost:7474/db/data/transaction/commit (1 bytes) => 635 Runtime information: Neo4j database version: neo4j gem version: 8.0.17 neo4j-core gem version: 7.1.2 For reference a failing spec: describe 'bug' do before(:each) do clear_model_memory_caches delete_db stub_active_node_class('Person') do has_many :out, :knows, model_class: 'Person', type: nil end stub_active_node_class('Company') do has_one :out, :ceo, type: :ceo, model_class: 'Person' end end let(:ceo) { Person.create } let(:company) { Company.create(ceo: ceo) } it 'should not drop rel_length' do expect(ceo.knows(rel_length: { min: 0 }).count).to eq(1) expect(ceo.knows(rel_length: { min: 0 }).to_a.count).to eq(1) ceo_via_association = company.ceo expect(ceo_via_association).to eq(ceo) expect(ceo_via_association.knows(rel_length: { min: 0 }).count).to eq(1) expect(ceo_via_association.knows(rel_length: { min: 0 }).to_a.count).to eq(1) # fails returning 0 as the automated 1 + n prevention kicks in and ignores the parameters of knows end end
gharchive/issue
2017-06-05T12:21:26
2025-04-01T06:39:43.691152
{ "authors": [ "klobuczek" ], "repo": "neo4jrb/neo4j", "url": "https://github.com/neo4jrb/neo4j/issues/1391", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1003050579
Video Library Documentation The Documentation must contain: Detailed API Documentation Examples of API usage (Auth, etc. concepts) How to add custom Database How to alter the Mock Backend Develop using Docusaurus 2.0 Fixed #14
gharchive/issue
2021-09-21T18:38:07
2025-04-01T06:39:43.720919
{ "authors": [ "sohamsshah" ], "repo": "neogcamp/mockBee", "url": "https://github.com/neogcamp/mockBee/issues/10", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
194845942
Preserve line-breaks in error messages Neomake loses line breaks when displaying errors. Using Haskell as an example: Neomake w/ hdevtools: Neomake w/ ghcmod: Expected behavior ghcmod-vim preserves line breaks when displaying errors: History This seems to be a well-known problem. When Haskell Makers were first being added to Neomake, there was no attempt to preserve line breaks. However, there were problems capturing every line of a multi-line error, much like Syntastic experienced. @benekastah referenced #15, which was eventually resolved, but it appears the accepted solution was to join lines into one. In the ghcmod-vim project, @toonn expressed a preference for using that plugin with Syntastic for a few reasons which included preserving line breaks. @expipiplus1 suggested Neomake, but admitted it "still munges everything into a single line". He cited a Syntastic issue on the subject of preserving line breaks, in which @lcd047 imparted some crucial wisdom: What ghcmod-vim shows are valid error lines intermixed with invalid ones. You'll notice that going to, say, line 2 in the quickfix list and pressing Enter doesn't take you to the error line in the source. Whether this is better than syntastic's UI or not is up to debate. @expipiplus1 offered an alternative: ghcmod-vim could work around this by making each individual line from an error it's own proper error message The downside of this approach is that :lnext won't jump to the next error until after you get through all the lines of a multi-line message. The code to get ghcmod-vim's behavior exists, but it's a bit gnarly. Messages are first saved to a file, then read with readfile (which corrects null characters to newlines), and then lines after the first are added to the quickfix list without file or line number. These seem to be two non-ideal alternatives for multi-line messages (to use "invalid" lines or not), each with their own drawbacks but both, in my opinion, preferable to the status quo. A capable framework could provide all three as options to the user. If I wanted to pursue that route, where would I find the best spot in the code to hook a single-message-to-possibly-singleton-list-of-quickfix-items function? Steps to reproduce " Use one of the below settings and :Neomake. let g:neomake_haskell_enabled_makers = ['hdevtools'] let g:neomake_haskell_enabled_makers = ['ghcmod'] Output of the ":verb NeomakeInfo" command Neomake debug information Async support: 0 Current filetype: Enabled makers For the current filetype (with :Neomake): [] NOTE: the current buffer does not have a filetype. For the project (with :Neomake!): [] NOTE: you can define g:neomake_enabled_makers to configure it. Settings g:neomake_haskell_enabled_makers = ['ghcmod'] shell: /bin/bash shellcmdflag: -c Windows: 0 :version VIM - Vi IMproved 7.4 (2013 Aug 10, compiled Nov 24 2016 16:44:48) Included patches: 1-1689 Extra patches: 8.0.0056 Modified by pkg-vim-maintainers@lists.alioth.debian.org Compiled by pkg-vim-maintainers@lists.alioth.debian.org Huge version with GTK2-GNOME GUI. Features included (+) or not (-): +acl +conceal +file_in_path +linebreak -mouse_sysmouse +python3 +tcl +wildmenu +arabic +cryptv +find_in_path +lispindent +mouse_urxvt +quickfix +terminfo +windows +autocmd +cscope +float +listcmds +mouse_xterm +reltime +termresponse +writebackup +balloon_eval +cursorbind +folding +localmap +multi_byte +rightleft +textobjects +X11 +browse +cursorshape -footer +lua +multi_lang +ruby +timers -xfontset ++builtin_terms +dialog_con_gui +fork() +menu -mzscheme +scrollbind +title +xim +byte_offset +diff +gettext +mksession +netbeans_intg +signs +toolbar +xsmp_interact +channel +digraphs -hangul_input +modify_fname +packages +smartindent +user_commands +xterm_clipboard +cindent +dnd +iconv +mouse +path_extra +startuptime +vertsplit -xterm_save +clientserver -ebcdic +insert_expand +mouseshape +perl +statusline +virtualedit +xpm +clipboard +emacs_tags +job +mouse_dec +persistent_undo -sun_workshop +visual +cmdline_compl +eval +jumplist +mouse_gpm +postscript +syntax +visualextra +cmdline_hist +ex_extra +keymap -mouse_jsbterm +printer +tag_binary +viminfo +cmdline_info +extra_search +langmap +mouse_netterm +profile +tag_old_static +vreplace +comments +farsi +libcall +mouse_sgr -python -tag_any_white +wildignore system vimrc file: "$VIM/vimrc" user vimrc file: "$HOME/.vimrc" 2nd user vimrc file: "~/.vim/vimrc" user exrc file: "$HOME/.exrc" system gvimrc file: "$VIM/gvimrc" user gvimrc file: "$HOME/.gvimrc" 2nd user gvimrc file: "~/.vim/gvimrc" system menu file: "$VIMRUNTIME/menu.vim" fall-back for $VIM: "/usr/share/vim" Compilation: gcc -c -I. -Iproto -DHAVE_CONFIG_H -DFEAT_GUI_GTK -pthread -I/usr/include/gtk-2.0 -I/usr/lib/x86_64-linux-gnu/gtk-2.0/inclu de -I/usr/include/gio-unix-2.0/ -I/usr/include/cairo -I/usr/include/pango-1.0 -I/usr/include/atk-1.0 -I/usr/include/cairo -I/usr/include/ pixman-1 -I/usr/include/libpng12 -I/usr/include/gdk-pixbuf-2.0 -I/usr/include/libpng12 -I/usr/include/pango-1.0 -I/usr/include/harfbuzz - I/usr/include/pango-1.0 -I/usr/include/glib-2.0 -I/usr/lib/x86_64-linux-gnu/glib-2.0/include -I/usr/include/freetype2 -D_REENTRANT -DORBI T2=1 -pthread -I/usr/include/libgnomeui-2.0 -I/usr/include/gnome-keyring-1 -I/usr/include/libbonoboui-2.0 -I/usr/include/libxml2 -I/usr/i nclude/libgnome-2.0 -I/usr/include/libbonobo-2.0 -I/usr/include/bonobo-activation-2.0 -I/usr/include/orbit-2.0 -I/usr/include/libgnomecan vas-2.0 -I/usr/include/gail-1.0 -I/usr/include/libart-2.0 -I/usr/include/gtk-2.0 -I/usr/lib/x86_64-linux-gnu/gtk-2.0/include -I/usr/inclu de/gio-unix-2.0/ -I/usr/include/cairo -I/usr/include/pango-1.0 -I/usr/include/atk-1.0 -I/usr/include/cairo -I/usr/include/pixman-1 -I/usr /include/libpng12 -I/usr/include/pango-1.0 -I/usr/include/harfbuzz -I/usr/include/pango-1.0 -I/usr/include/freetype2 -I/usr/include/gdk-p ixbuf-2.0 -I/usr/include/libpng12 -I/usr/include/gnome-vfs-2.0 -I/usr/lib/x86_64-linux-gnu/gnome-vfs-2.0/include -I/usr/include/gconf/2 - I/usr/include/dbus-1.0 -I/usr/lib/x86_64-linux-gnu/dbus-1.0/include -I/usr/include/glib-2.0 -I/usr/lib/x86_64-linux-gnu/glib-2.0/include -Wdate-time -g -O2 -fPIE -fstack-protector-strong -Wformat -Werror=format-security -U_FORTIFY_SOURCE -D_FORTIFY_SOURCE=1 Linking: gcc -L. -Wl,-Bsymbolic-functions -Wl,-z,relro -fstack-protector -rdynamic -Wl,-export-dynamic -Wl,-E -Wl,-Bsymbolic-functions -fPIE -pie -Wl,-z,relro -Wl,-z,now -Wl,--as-needed -o vim -lgtk-x11-2.0 -lgdk-x11-2.0 -lpangocairo-1.0 -latk-1.0 -lcairo -lgdk_pixbuf- 2.0 -lgio-2.0 -lpangoft2-1.0 -lpango-1.0 -lgobject-2.0 -lglib-2.0 -lfontconfig -lfreetype -lgnomeui-2 -lSM -lICE -lbonoboui-2 -lgnome-2 -lpopt -lbonobo-2 -lbonobo-activation -lORBit-2 -lgnomecanvas-2 -lart_lgpl_2 -lgtk-x11-2.0 -lgdk-x11-2.0 -lpangocairo-1.0 -latk-1.0 -lcai ro -lgio-2.0 -lpangoft2-1.0 -lpango-1.0 -lfontconfig -lfreetype -lgdk_pixbuf-2.0 -lgnomevfs-2 -lgconf-2 -lgthread-2.0 -lgmodule-2.0 -lgob ject-2.0 -lglib-2.0 -lSM -lICE -lXpm -lXt -lX11 -lXdmcp -lSM -lICE -lm -ltinfo -lnsl -lselinux -lacl -lattr -lgpm -ldl -L/usr/lib -ll ua5.2 -Wl,-E -fstack-protector-strong -L/usr/local/lib -L/usr/lib/x86_64-linux-gnu/perl/5.22/CORE -lperl -ldl -lm -lpthread -lcrypt -L /usr/lib/python3.5/config-3.5m-x86_64-linux-gnu -lpython3.5m -lpthread -ldl -lutil -lm -L/usr/lib/x86_64-linux-gnu -ltcl8.6 -ldl -lz -lpt hread -lieee -lm -lruby-2.3 -lpthread -lgmp -ldl -lcrypt -lm I wonder if it'd be possible to replace lnext with a version which skips to the next qf line with a location. One potential solution here is to extend neomake with a new command that can pretty-print the current error. This seems like a good way to work around the fact that vim quickfix can only store a single line. Another idea could be to propose this as an enhancement to neovim - perhaps they would be willing to move quickfix in this direction.
gharchive/issue
2016-12-11T18:02:56
2025-04-01T06:39:43.746272
{ "authors": [ "dlants", "expipiplus1", "thejohnfreeman" ], "repo": "neomake/neomake", "url": "https://github.com/neomake/neomake/issues/832", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2173328288
pageserver: clean up ancestral layers after split, old index_part objects Per the "Cleaning up parent-shard layers" section in #6358 -- currently after a shard split, layers from the parent shards are not deleted until the whole tenant is eventually deleted. We should implement an occasional online scrub routine that checks which of these are referenced by children, and cleans them up. It likely makes sense to combine this work with cleaning up old-generation index_part.json objects, as these older objects will likely reference parent shard layers -- we should first define the criteria for cleaning up old indices, and then use the still-alive indices as the source of references for cleaning up parent layers. ### Tasks - [ ] https://github.com/neondatabase/neon/pull/7925 - [ ] https://github.com/neondatabase/neon/pull/8196 - [ ] https://github.com/neondatabase/cloud/issues/14024 This will be enabled in staging here: https://github.com/neondatabase/aws/pull/1654 Then we'll let it soak for at least a week before proceeding to prod.
gharchive/issue
2024-03-07T08:57:10
2025-04-01T06:39:43.755394
{ "authors": [ "jcsp" ], "repo": "neondatabase/neon", "url": "https://github.com/neondatabase/neon/issues/7043", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1509392862
Replace 'tar' crate with 'tokio-tar' The synchronous 'tar' crate has required us to use block_in_place and SyncIoBridge to work together with the async I/O in the client connection. Switch to 'async-tar' crate that uses async I/O natively. As part of this, move the CopyDataWriter implementation to postgres_backend_async.rs. Even though it's only used in one place currently, it's in principle generally applicable whenever you want to use COPY out. 'async-tar' is a fork of the 'tar' crate, just replacing sync functions with corresponding async ones. I'm not sure how well maintained it is, but a crate like this doesn't really need much change or maintenance. If we go with this, it might be good to take a close look at 'async-tar' and compare if it's missing any fixes that have been made on 'tar' crate. I'm surprised by how many dependencies the 'async-tar' crate pulls in. It depends on 'async-std'; is that really necessary, or could we update 'async-tar' to rely on the corresponding std functionality instead? It depends on 'async-std' Let's not pull that in then, it's yet another async runtime and we have tokio for that. Could something like https://crates.io/crates/tokio-tar be a better choice? It depends on 'async-std' Let's not pull that in then, it's yet another async runtime and we have tokio for that. I see. Could something like https://crates.io/crates/tokio-tar be a better choice? Tried that now. There's one problem with it: error[E0477]: the type `AbortableWrite<'a, W>` does not fulfill the required lifetime --> pageserver/src/basebackup.rs:44:9 | 44 | ar: Builder<AbortableWrite<'a, W>>, | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ | note: type must satisfy the static lifetime as required by this binding --> /home/heikki/.cargo/registry/src/github.com-1ecc6299db9ec823/tokio-tar-0.3.0/src/builder.rs:15:46 | 15 | pub struct Builder<W: Write + Unpin + Send + 'static> { | ^^^^^^^ tokio_tar::Builder requires the writer to be 'static. The reason for that is that it has a mechanism to write the tar EOF block (1024 bytes of zeros), if you just drop the Builder without calling finish. The plain tar crate has that too, and it's straightforward there: the Drop implementation writes the EOF marker. In the async_tar, the Drop implementation uses block_on to do the same. But in tokio_tar, it uses a channel and a separately spawned task to do it. That's hacky IMHO. The EOF-block-writing-task executes at some not-well defined time after the drop has happened. And it's super annoying for us, because we actually don't want the EOF marker to be written at all. We actually work hard to skip it, that's exactly why we have the AbortableWrite hack. So ideally tokio-tar just didn't have that Drop implementation, and then it wouldn't require 'static, and we could also remove AbortableWrite. I'm tempted to fork it and do that. Perhaps the upstream project would be interested in a PR for that too; IMHO it's a bad idea to write the EOF block on drop anyway. Switched to a modified version of 'tokio-tar' without the Drop implementation and the 'static requirement. Looking more, https://github.com/vorot93/tokio-tar/pull/9 does look quite important in case the work not completing on one poll, however that's only on the read path which we don't use. Looking at tokio-tar the AsyncWrite usage of slices is not cancellation safe. I doubt know if it was cancellation safe before, but now it definitely is not. I think however the only place it's cancelled is with the shutdown waiter, and then the writer is not restarted back again. Maybe a comment or two could be added about this near... I'll try to find that callsite. We don't use that Drop path indeed, and I've created a PR that should explicitly help us with that: https://github.com/neondatabase/tokio-tar/pull/1 , there are more details about the upstreaming activities, let's see if they ever get released. For now, out fork can do that: https://github.com/neondatabase/neon/pull/3239
gharchive/pull-request
2022-12-23T14:12:55
2025-04-01T06:39:43.764700
{ "authors": [ "SomeoneToIgnore", "hlinnaka", "koivunej" ], "repo": "neondatabase/neon", "url": "https://github.com/neondatabase/neon/pull/3202", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2315251768
fix: connection string uri scheme to postgresql Use postgresql scheme for connection string as postgres causes issue with some tools Validate database name passed to connection-string command :tada: This PR is included in version 1.29.4 :tada: The release is available on: npm package (@latest dist-tag) GitHub release Your semantic-release bot :package::rocket:
gharchive/pull-request
2024-05-24T12:10:19
2025-04-01T06:39:43.768132
{ "authors": [ "Shridhad", "duskpoet" ], "repo": "neondatabase/neonctl", "url": "https://github.com/neondatabase/neonctl/pull/218", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
107495250
Fatal error: Class 'TYPO3\SwiftMailer\Message' not found This line in the EmailFinisher class does not ensure that the required Swiftmailer class is actually available (and causes a fatal error if you don't have Swiftmailer installed): $mail = new \TYPO3\SwiftMailer\Message(); I suggest to either include Swiftmailer as dependency in composer.json or to check for installed Swiftmailer and throw an informative exception instead. Thanks, good point! I guess we should at least suggest the switftmailer package if we don't want to create a hard dependency. Fixed with #23
gharchive/issue
2015-09-21T11:53:13
2025-04-01T06:39:43.772229
{ "authors": [ "KatharinaSt", "kdambekalns", "kitsunet" ], "repo": "neos/form", "url": "https://github.com/neos/form/issues/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
178107977
BUGFIX: Prevent Fatal Error in Setup This is a hotfix for a regression introduced with https://github.com/neos/flow-development-collection/pull/499. Background: The latest Http\RequestHandler of Flow internally uses the ComponentChain to determine the current Request/Response. This package uses a specialized handler which doesn't rely on the ComponentChain leading to a Fatal Error. needs to go into the 2.0 branch right? needs to go into the 2.0 branch right? I'm not sure, 2.0 is used in Neos 2.0+ and the fix is only required for Neos 2.3.x. I guess we should probably create a 2.3 branch in the setup and apply it there!? The evils of inheritance... Makes sense, but be aware that the $request and $response properties will also be removed with 4.0 - so maybe already copy them over be aware that the $request and $response properties will also be removed with 4.0 That's news to me. Is that documented somewhere or is there a PR for this? Anyways, I rather merge this now to fix the fatal. I'd suggest to merge it into master and then create a branch 2.3 and require that in the affected distributions. Does that make sense? Is that documented somewhere or is there a PR for this? https://github.com/neos/flow-development-collection/pull/500 @bwaidelich: I'd rather avoid a separate version if it can be done in backwards compatible way, which this seems to be.. I messed up the rebase, see #28
gharchive/pull-request
2016-09-20T16:20:54
2025-04-01T06:39:43.798150
{ "authors": [ "aertmann", "albe", "bwaidelich" ], "repo": "neos/setup", "url": "https://github.com/neos/setup/pull/13", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1082856868
:LspRestart throws error when using non-supported language server Description When called without arguments in a buffer attached to a non-supported language server, :LspRestart throws the following error: Error executing vim.schedule lua callback: .../site/pack/packer/start/nvim-lspconfig/lua/lspconfig.lua:58: attempt to index a nil value stack traceback: .../site/pack/packer/start/nvim-lspconfig/lua/lspconfig.lua:58: in function '' vim.lua: in function '' vim.lua: in function <vim.lua:0> Neovim version NVIM v0.7.0-dev+732-g80210c189 Nvim-lspconfig version 64bd3b95345f2dbedc83a3d9a472662c48b3e940 Operating system and version macOS 12.1 Affected language servers All Steps to reproduce nvim --clean -u minimal_init.lua test.py :LspRestart Actual behavior The error from the description is thrown. Expected behavior No error is thrown. Minimal config local on_windows = vim.loop.os_uname().version:match("Windows") local function join_paths(...) local path_sep = on_windows and "\\" or "/" local result = table.concat({ ... }, path_sep) return result end vim.cmd([[set runtimepath=$VIMRUNTIME]]) local temp_dir if on_windows then temp_dir = vim.loop.os_getenv("TEMP") else temp_dir = "/tmp" end vim.cmd("set packpath=" .. join_paths(temp_dir, "nvim", "site")) local package_root = join_paths(temp_dir, "nvim", "site", "pack") local install_path = join_paths(package_root, "packer", "start", "packer.nvim") local compile_path = join_paths(install_path, "plugin", "packer_compiled.lua") local function load_plugins() require("packer").startup({ { "wbthomason/packer.nvim", "neovim/nvim-lspconfig", }, config = { package_root = package_root, compile_path = compile_path, }, }) end _G.attach_buffer = function() local client_id = vim.lsp.start_client({ cmd = { "pylsp" } }) -- cmd is irrelevant vim.lsp.buf_attach_client(vim.api.nvim_get_current_buf(), client_id) end _G.load_config = function() -- filetype is irrelevant but should be consistent with language server vim.cmd("autocmd FileType python lua attach_buffer()") end if vim.fn.isdirectory(install_path) == 0 then vim.fn.system({ "git", "clone", "https://github.com/wbthomason/packer.nvim", install_path }) load_plugins() require("packer").sync() vim.cmd([[autocmd User PackerComplete ++once lua load_config()]]) else load_plugins() require("packer").sync() _G.load_config() end Health check lspconfig: require("lspconfig.health").check() ======================================================================== - WARNING: Can't find any config. LSP log Probably not relevant The following trivial patch fixes the issue as well as a related issue where a config is defined but setup has not yet been called, meaning that launch is not yet defined. I didn't yet submit a PR in case I'm missing something but am happy to do so if this is the right fix. diff --git a/lua/lspconfig.lua b/lua/lspconfig.lua index 8404632..685b166 100644 --- a/lua/lspconfig.lua +++ b/lua/lspconfig.lua @@ -54,9 +54,11 @@ function M._root._setup() function(cmd_args) for _, client in ipairs(M.util.get_clients_from_cmd_args(cmd_args)) do client.stop() - vim.defer_fn(function() - configs[client.name].launch() - end, 500) + if configs[client.name] and configs[client.name].launch then + vim.defer_fn(function() + configs[client.name].launch() + end, 500) + end end end, '-nargs=? -complete=customlist,v:lua.lsp_get_active_client_ids', I think a cleaner "fix" would be to expose the table of client ids managed by lspconfig, and make these methods only restart those clients. Sure, that makes sense. Does this table already exist somewhere in the code, or is it something that would need to be created? I think a cleaner "fix" would be to expose the table of client ids managed by lspconfig, and make these methods only restart those clients. How about using the docs tag? or maybe add a special marker in the client config. Otherwise, how would you know which ones are managed? just because tsserver is active, doesn't mean it's managed by lspconfig. Because lspconfig tracks internally which clients it starts... Please read the code and evaluate the solution I suggested before theorycrafting alternatives. I took another look, and I can see that the active client IDs are available via manager.clients(). Does something like this make sense? local get_lspconfig_clients = function() local clients = {} for _, config in pairs(require("lspconfig.configs")) do if config.manager then vim.list_extend(clients, config.manager.clients()) end end return clients end If so, using that here instead of vim.lsp.get_active_clients() would solve this specific issue. Yup
gharchive/issue
2021-12-17T03:50:42
2025-04-01T06:39:43.948983
{ "authors": [ "jose-elias-alvarez", "kylo252", "mjlbach" ], "repo": "neovim/nvim-lspconfig", "url": "https://github.com/neovim/nvim-lspconfig/issues/1582", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1705066451
Error when decoding stdout from rust-analyzer Description Every time when I open a .rs file, the following warning occurs: [lspconfig] unhandled error: ...ig/lua/lspconfig/server_configurations/rust_analyzer.lua:41: Expected value but found T_END at character 1 and rust-analyzer fails to launch. The cause of the problem lies in lua/lspconfig/server_configurations/rust_analyzer.lua line 41 where we try to call vim.json.decode with parameter "" (empty string). How can we solve this? Neovim version NVIM v0.10.0-dev-311+g03ca36d1f8 Build type: Release LuaJIT 2.1.0-beta3 Nvim-lspconfig version 84717430fb6f51c9e9cc8d63da10cceafc91500c Operating system and version Linux 6.3.1-arch1-1 Affected language servers rust-analyzer Steps to reproduce nvim -nu minimal.lua Open an arbitrary .rs file Observe the warning Actual behavior rust-analyzer fails to launch. Expected behavior rust-analyzer launches and attaches successfully. Minimal config ---@diagnostic disable: undefined-global local on_windows = vim.loop.os_uname().version:match 'Windows' local function join_paths(...) local path_sep = on_windows and '\\' or '/' local result = table.concat({ ... }, path_sep) return result end vim.cmd [[set runtimepath=$VIMRUNTIME]] local temp_dir = vim.loop.os_getenv 'TEMP' or '/tmp' vim.cmd('set packpath=' .. join_paths(temp_dir, 'nvim', 'site')) local package_root = join_paths(temp_dir, 'nvim', 'site', 'pack') local lspconfig_path = join_paths(package_root, 'test', 'start', 'nvim-lspconfig') if vim.fn.isdirectory(lspconfig_path) ~= 1 then vim.fn.system { 'git', 'clone', 'https://github.com/neovim/nvim-lspconfig', lspconfig_path } end vim.lsp.set_log_level 'trace' require('vim.lsp.log').set_format_func(vim.inspect) local nvim_lsp = require 'lspconfig' local on_attach = function(_, bufnr) local function buf_set_option(...) vim.api.nvim_buf_set_option(bufnr, ...) end buf_set_option('omnifunc', 'v:lua.vim.lsp.omnifunc') -- Mappings. local opts = { buffer = bufnr, noremap = true, silent = true } vim.keymap.set('n', 'gD', vim.lsp.buf.declaration, opts) vim.keymap.set('n', 'gd', vim.lsp.buf.definition, opts) vim.keymap.set('n', 'K', vim.lsp.buf.hover, opts) vim.keymap.set('n', 'gi', vim.lsp.buf.implementation, opts) vim.keymap.set('n', '<C-k>', vim.lsp.buf.signature_help, opts) vim.keymap.set('n', '<space>wa', vim.lsp.buf.add_workspace_folder, opts) vim.keymap.set('n', '<space>wr', vim.lsp.buf.remove_workspace_folder, opts) vim.keymap.set('n', '<space>wl', function() print(vim.inspect(vim.lsp.buf.list_workspace_folders())) end, opts) vim.keymap.set('n', '<space>D', vim.lsp.buf.type_definition, opts) vim.keymap.set('n', '<space>rn', vim.lsp.buf.rename, opts) vim.keymap.set('n', 'gr', vim.lsp.buf.references, opts) vim.keymap.set('n', '<space>e', vim.diagnostic.open_float, opts) vim.keymap.set('n', '[d', vim.diagnostic.goto_prev, opts) vim.keymap.set('n', ']d', vim.diagnostic.goto_next, opts) vim.keymap.set('n', '<space>q', vim.diagnostic.setloclist, opts) end -- Add the server that troubles you here local name = 'rust_analyzer' local cmd = { 'rust_analyzer' } -- needed for elixirls, lua_ls, omnisharp if not name then print 'You have not defined a server name, please edit minimal_init.lua' end if not nvim_lsp[name].document_config.default_config.cmd and not cmd then print [[You have not defined a server default cmd for a server that requires it please edit minimal_init.lua]] end nvim_lsp[name].setup { cmd = cmd, on_attach = on_attach, } print [[You can find your log at $HOME/.cache/nvim/lsp.log. Please paste in a github issue under a details tag as described in the issue template.]] LSP log https://gist.github.com/Bekaboo/f727c89767529c9f532c723e6f1a06be Additional info: this warning only shows when opening bare .rs files (i.e. not in a rust project).
gharchive/issue
2023-05-11T04:57:58
2025-04-01T06:39:43.955656
{ "authors": [ "Bekaboo" ], "repo": "neovim/nvim-lspconfig", "url": "https://github.com/neovim/nvim-lspconfig/issues/2604", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
123177028
WebSocket support Create interfaces for client/Client and neptulon/Listener objects and separate implementation as: client/TCPClient & client/WSClient neptulon/TCPListener & neptulon/WSListener We can even skip this step and keep everything all in one if things are as easy as what we did with ListenTLS. We could also change project description after this to: TCP/WebSocket framework for Go. Minimalist web framework for Go. Done.
gharchive/issue
2015-12-20T18:43:34
2025-04-01T06:39:43.974423
{ "authors": [ "soygul" ], "repo": "neptulon/neptulon", "url": "https://github.com/neptulon/neptulon/issues/49", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
252842491
Worker unable to get token could not retreive token from any provider 2017/08/25 09:13:16 client.go:64: [DEBUG] HTTP Request: POST /v1/token/nerdalize-cloud/refresh/ HTTP/1.1 Host: local.dev.nlze.nl:8000 Accept: application/json Content-Type: application/json {"jwt":"eyJhbGciOiJFUzM4NCIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJhdXRoLm5lcmRhbGl6ZS5jb20iLCJzdWIiOiI0IiwiaWF0IjoxNTAzNjUyMDY0LCJobV9zZWNyZXQiOiJjZWU2MzM1ZDZkOGU4ZDc4NTZlMDU4NjBkNTA0YzNiMjhjNTAxZGY5ZmQ0MGQ0MzdkMTNhYTJmMjlmZTBlMTFjIiwibmJmIjoxNTAzNjUyMDA0LCJleHAiOjE1MDM2NTIzNjQsImFjY2VzcyI6W3sic2VydmljZSI6Im5jZS5uZXJkYWxpemUuY29tIiwidHlwZSI6Indvcmtsb2FkIiwiYWRkcmVzcyI6Indvcmtsb2FkLm5semUubmwiLCJyZXNvdXJjZV9pZGVudGlmaWVyIjoibmVyZGFsaXplLWNsb3VkIiwicmlnaHRzIjpbImRlbGV0ZSIsImxpc3QiLCJkZXNjcmliZSIsImNyZWF0ZSJdfSx7InNlcnZpY2UiOiJuY2UubmVyZGFsaXplLmNvbSIsInR5cGUiOiJ0b2tlbiIsImFkZHJlc3MiOiJzMy1ldS13ZXN0LTEuYW1hem9uYXdzLmNvbSIsInJlc291cmNlX2lkZW50aWZpZXIiOiJuZXJkYWxpemUtY2xvdWQiLCJyaWdodHMiOlsicmVhZCIsImRlbGV0ZSIsInVwZGF0ZSIsImNyZWF0ZSJdfSx7InNlcnZpY2UiOiJuY2UubmVyZGFsaXplLmNvbSIsInR5cGUiOiJ3b3JrZXIiLCJhZGRyZXNzIjoid29ya2VyLm5semUubmwiLCJyZXNvdXJjZV9pZGVudGlmaWVyIjoibmVyZGFsaXplLWNsb3VkIiwicmlnaHRzIjpbImRlbGV0ZSIsImxpc3QiLCJjcmVhdGUiXX0seyJzZXJ2aWNlIjoicmVnaXN0cnkubmVyZGFsaXplLmNvbSIsInR5cGUiOiJyZXBvc2l0b3J5IiwiYWRkcmVzcyI6InJlZ2lzdHJ5Lm5semUubmwiLCJyZXNvdXJjZV9pZGVudGlmaWVyIjoibmVyZGFsaXplLWNsb3VkIiwicmlnaHRzIjpbInB1bGwiLCJwdXNoIl19LHsic2VydmljZSI6Im5jZS5uZXJkYWxpemUuY29tIiwidHlwZSI6ImRhdGFzZXQiLCJhZGRyZXNzIjoiZGF0YXNldC5ubHplLm5sIiwicmVzb3VyY2VfaWRlbnRpZmllciI6Im5lcmRhbGl6ZS1jbG91ZCIsInJpZ2h0cyI6WyJ1cHJvZ3Jlc3MiLCJkZXNjcmliZSIsImNyZWF0ZSIsImRvd25sb2FkIiwibGlzdCIsInVwbG9hZCJdfSx7InNlcnZpY2UiOiJuY2UubmVyZGFsaXplLmNvbSIsInR5cGUiOiJ0YXNrIiwiYWRkcmVzcyI6InRhc2submx6ZS5ubCIsInJlc291cmNlX2lkZW50aWZpZXIiOiJuZXJkYWxpemUtY2xvdWQiLCJyaWdodHMiOlsiZGVzY3JpYmUiLCJzdG9wIiwic3RhcnQiLCJsaXN0IiwicnJlY2VpdmUiLCJycHJvZ3Jlc3MiXX1dfQ.QkbMaRD73ToBpUlCrCqEagIloiUmYwXrZrY_tO7Av-qhMelmGIHRo16nKvULS-x3xVyZ0Hm-ZimPS8rkro9-xIXsdWH-mZB3HdepMXtDrwys6nShKX7GZq81MSdzxZSB","secret":"PKoUFcPZwZlN99PqUv57XQUnA6xaYLzX"} 2017/08/25 09:13:16 working.go:255: [ERROR] Failed to receive task run: failed to describe workload: ErrTokenUnset failed to get nerd jwt github.com/nerdalize/nerd/nerd/jwt.(*AuthAPIProvider).Retrieve /go/src/github.com/nerdalize/nerd/nerd/jwt/authapi_provider.go:40 github.com/nerdalize/nerd/nerd/client/batch/v1.(*ChainedJWTProvider).Retrieve /go/src/github.com/nerdalize/nerd/nerd/client/batch/v1/jwt_provider.go:53 github.com/nerdalize/nerd/nerd/client/batch/v1.(*Client).getJWT /go/src/github.com/nerdalize/nerd/nerd/client/batch/v1/client.go:91 github.com/nerdalize/nerd/nerd/client/batch/v1.(*Client).doRequest /go/src/github.com/nerdalize/nerd/nerd/client/batch/v1/client.go:105 github.com/nerdalize/nerd/nerd/client/batch/v1.(*Client).DescribeWorkload /go/src/github.com/nerdalize/nerd/nerd/client/batch/v1/workload.go:61 github.com/nerdalize/nerd/nerd/client/batch/v1.(*Client).ReceiveTaskRuns /go/src/github.com/nerdalize/nerd/nerd/client/batch/v1/task.go:80 github.com/nerdalize/nerd/nerd/service/working/v1.(*Worker).startReceivingRuns.func1 /go/src/github.com/nerdalize/nerd/nerd/service/working/v1/working.go:229 runtime.goexit /usr/local/go/src/runtime/asm_amd64.s:2197 could not retreive token from any provider Closing this, we'll be moving to the new job running logic in the next milestone
gharchive/issue
2017-08-25T09:15:33
2025-04-01T06:39:43.979345
{ "authors": [ "advanderveer", "neoel" ], "repo": "nerdalize/nerd", "url": "https://github.com/nerdalize/nerd/issues/148", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
839614666
Fix on China AWS support if run Yace on china AWS, there is error “Couldn't describe resources for region cn-northwest-1: InvalidClientTokenId: The security token included in the request is invalid.\n\tstatus code: 403,” and china AWS is in seperate infra. Session new need to specified region, we fixed code based on Release 0.26.3-alpha, it has been tested successfully in China. and in aws_cloudwatch.go ln50: func createStsSession(roleArn string) this part of code seems new adding, and region is unnecessary here. It has been tested. Was fixed and merged the same issue ; {"arn":"","error":"InvalidClientTokenId: The security token included in the request is invalid.\n\tstatus code: 403, request id: 46cbd89f-1a5d-49b4-996f-c2b616f4479c","job_type":"alb","level":"error","msg":"Couldn't get account Id","region":"cn-northwest-1","time":"2023-10-24T11:43:05Z"}.
gharchive/issue
2021-03-24T11:13:28
2025-04-01T06:39:43.982032
{ "authors": [ "alibner", "insectme", "thomaspeitz" ], "repo": "nerdswords/yet-another-cloudwatch-exporter", "url": "https://github.com/nerdswords/yet-another-cloudwatch-exporter/issues/344", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1511329507
OSError: Unknown compute capability. Ensure PyTorch with CUDA support is installed. Installed nerfstudio with CUDA 11.3 as per setup page. Everything works fine. Followed the tutorial all the way to ns-train nerfacto --data data/nerfstudio/poster where I get this error: $ ns-train nerfacto --data data/nerfstudio/poster Traceback (most recent call last): File "C:\Users\glens\miniconda3\envs\ns\lib\runpy.py", line 194, in _run_module_as_main return _run_code(code, main_globals, None, File "C:\Users\glens\miniconda3\envs\ns\lib\runpy.py", line 87, in _run_code exec(code, run_globals) File "C:\Users\glens\miniconda3\envs\ns\Scripts\ns-train.exe\__main__.py", line 4, in <module> File "C:\Users\glens\miniconda3\envs\ns\lib\site-packages\scripts\train.py", line 50, in <module> from nerfstudio.configs.method_configs import AnnotatedBaseConfigUnion File "C:\Users\glens\miniconda3\envs\ns\lib\site-packages\nerfstudio\configs\method_configs.py", line 46, in <module> from nerfstudio.field_components.temporal_distortions import TemporalDistortionKind File "C:\Users\glens\miniconda3\envs\ns\lib\site-packages\nerfstudio\field_components\__init__.py", line 17, in <module> from .encodings import Encoding, ScalingAndOffset File "C:\Users\glens\miniconda3\envs\ns\lib\site-packages\nerfstudio\field_components\encodings.py", line 34, in <module> import tinycudann as tcnn File "C:\Users\glens\miniconda3\envs\ns\lib\site-packages\tinycudann\__init__.py", line 9, in <module> from tinycudann.modules import free_temporary_memory, NetworkWithInputEncoding, Network, Encoding File "C:\Users\glens\miniconda3\envs\ns\lib\site-packages\tinycudann\modules.py", line 16, in <module> raise EnvironmentError("Unknown compute capability. Ensure PyTorch with CUDA support is installed.") OSError: Unknown compute capability. Ensure PyTorch with CUDA support is installed. Removed CUDA 11.3, installed 11.7 and followed the install instructions on the Pytorch page, i.e. conda install pytorch torchvision torchaudio pytorch-cuda=11.7 -c pytorch -c nvidia. Everything works as per the pip install instructions. tinycudann even seemed to build a lot faster. Ended up with exactly the same error: OSError: Unknown compute capability. Ensure PyTorch with CUDA support is installed. CUDA variables are set as (not sure if this is relevant): I understand devs are busy and I don't want to bother anyone but I'm trying really hard to get this working and just keep getting pipped at the post. Let me know if there's a better forum. I ran into the same exact error today with conda installation of PyTorch with CUDA 11.7 (11.3 worked fine for me though). Seems like a problem with tinycuda rather than nerfstudio. I'm also using CUDA 11.3, did the tinycuda installation go entirely smoothly? @THE-COB I have both had tinycudann installation go smoothly and not work at all. I think @brawlawnzo is right, it's an issue with tinycuda. I just tried to install as per nerfstudio installation instructions on my work computer. Visual Studio 2019, CUDA 11.3, Python 3.8. Completely clean install. Got this error install tinycudann: pip install git+https://github.com/NVlabs/tiny-cuda-nn/#subdirectory=bindings/torch Collecting git+https://github.com/NVlabs/tiny-cuda-nn/#subdirectory=bindings/torch Cloning https://github.com/NVlabs/tiny-cuda-nn/ to c:\users\glens\appdata\local\temp\pip-req-build-s1rw0r85 Running command git clone --filter=blob:none --quiet https://github.com/NVlabs/tiny-cuda-nn/ 'C:\Users\glens\AppData\Local\Temp\pip-req-build-s1rw0r85' Resolved https://github.com/NVlabs/tiny-cuda-nn/ to commit 14053e9a87ebf449d32bda335c0363dd4f5667a4 Running command git submodule update --init --recursive -q Preparing metadata (setup.py) ... done Building wheels for collected packages: tinycudann Building wheel for tinycudann (setup.py) ... error error: subprocess-exited-with-error × python setup.py bdist_wheel did not run successfully. │ exit code: 1 ╰─> [1327 lines of output] Building PyTorch extension for tiny-cuda-nn version 1.7 Obtained compute capability 86 from PyTorch running bdist_wheel C:\Users\glens\miniconda3\envs\ns\lib\site-packages\torch\utils\cpp_extension.py:411: UserWarning: Attempted to use ninja as the BuildExtension backend but we could not find ninja.. Falling back to using the slow distutils backend. warnings.warn(msg.format('we could not find ninja.')) running build running build_py creating build creating build\lib.win-amd64-cpython-38 creating build\lib.win-amd64-cpython-38\tinycudann copying tinycudann\modules.py -> build\lib.win-amd64-cpython-38\tinycudann copying tinycudann\__init__.py -> build\lib.win-amd64-cpython-38\tinycudann running egg_info creating tinycudann.egg-info writing tinycudann.egg-info\PKG-INFO writing dependency_links to tinycudann.egg-info\dependency_links.txt writing top-level names to tinycudann.egg-info\top_level.txt writing manifest file 'tinycudann.egg-info\SOURCES.txt' reading manifest file 'tinycudann.egg-info\SOURCES.txt' writing manifest file 'tinycudann.egg-info\SOURCES.txt' copying tinycudann\bindings.cpp -> build\lib.win-amd64-cpython-38\tinycudann running build_ext building 'tinycudann_bindings_86._C' extension creating build\dependencies creating build\dependencies\fmt creating build\dependencies\fmt\src creating build\src creating build\temp.win-amd64-cpython-38 creating build\temp.win-amd64-cpython-38\Release creating build\temp.win-amd64-cpython-38\Release\tinycudann "C:\Program Files (x86)\Microsoft Visual Studio\2019\Professional\VC\Tools\MSVC\14.29.30133\bin\HostX86\x64\cl.exe" /c /nologo /O2 /W3 /GL /DNDEBUG /MD -IC:\Users\gle... followed by thousands of lines of errors, and then: Error limit reached. 100 errors detected in the compilation of "../../src/common.cu". Compilation terminated. common.cu error: command 'C:/Program Files/NVIDIA GPU Computing Toolkit/CUDA\\v11.3\\bin\\nvcc.exe' failed with exit code 1 [end of output] note: This error originates from a subprocess, and is likely not a problem with pip. error: legacy-install-failure × Encountered error while trying to install package. ╰─> tinycudann @grshennan I think I might have run into this as well! It seems to resemble the issue in here: https://github.com/NVlabs/tiny-cuda-nn/issues/169 One of the things in that thread that I think got me passed that (and then to the next error lol) was conda installing ninja. If that doesn't work maybe try reinstalling visual studio (2019 should be alright, but I'm currently using 2017) @THE-COB I have VS 2019 installed but same idea, yes. I think it's the same as this issue: https://github.com/NVlabs/tiny-cuda-nn/issues/147 Tried installing ninja and got a different error. Now it's running build_ext error: [WinError 2] The system cannot find the file specified as per the tinycudann issue. What is your current error? @grshennan $ pip install git+https://github.com/NVlabs/tiny-cuda-nn/#subdirectory=bindings/torch Collecting git+https://github.com/NVlabs/tiny-cuda-nn/#subdirectory=bindings/torch Cloning https://github.com/NVlabs/tiny-cuda-nn/ to c:\users\glens\appdata\local\temp\pip-req-build-6w2nnebl Running command git clone --filter=blob:none --quiet https://github.com/NVlabs/tiny-cuda-nn/ 'C:\Users\glens\AppData\Local\Temp\pip-req-build-6w2nnebl' Resolved https://github.com/NVlabs/tiny-cuda-nn/ to commit 14053e9a87ebf449d32bda335c0363dd4f5667a4 Running command git submodule update --init --recursive -q Preparing metadata (setup.py) ... done Building wheels for collected packages: tinycudann Building wheel for tinycudann (setup.py) ... error error: subprocess-exited-with-error × python setup.py bdist_wheel did not run successfully. │ exit code: 1 ╰─> [21 lines of output] Building PyTorch extension for tiny-cuda-nn version 1.7 Obtained compute capability 86 from PyTorch running bdist_wheel running build running build_py creating build creating build\lib.win-amd64-cpython-38 creating build\lib.win-amd64-cpython-38\tinycudann copying tinycudann\modules.py -> build\lib.win-amd64-cpython-38\tinycudann copying tinycudann\__init__.py -> build\lib.win-amd64-cpython-38\tinycudann running egg_info creating tinycudann.egg-info writing tinycudann.egg-info\PKG-INFO writing dependency_links to tinycudann.egg-info\dependency_links.txt writing top-level names to tinycudann.egg-info\top_level.txt writing manifest file 'tinycudann.egg-info\SOURCES.txt' reading manifest file 'tinycudann.egg-info\SOURCES.txt' writing manifest file 'tinycudann.egg-info\SOURCES.txt' copying tinycudann\bindings.cpp -> build\lib.win-amd64-cpython-38\tinycudann running build_ext error: [WinError 2] The system cannot find the file specified [end of output] note: This error originates from a subprocess, and is likely not a problem with pip. ERROR: Failed building wheel for tinycudann Running setup.py clean for tinycudann Failed to build tinycudann Installing collected packages: tinycudann Running setup.py install for tinycudann ... error error: subprocess-exited-with-error × Running setup.py install for tinycudann did not run successfully. │ exit code: 1 ╰─> [15 lines of output] Building PyTorch extension for tiny-cuda-nn version 1.7 Obtained compute capability 86 from PyTorch running install C:\Users\glens\miniconda3\envs\ns\lib\site-packages\setuptools\command\install.py:34: SetuptoolsDeprecationWarning: setup.py install is deprecated. Use build and pip and other standards-based tools. warnings.warn( running build running build_py running egg_info writing tinycudann.egg-info\PKG-INFO writing dependency_links to tinycudann.egg-info\dependency_links.txt writing top-level names to tinycudann.egg-info\top_level.txt reading manifest file 'tinycudann.egg-info\SOURCES.txt' writing manifest file 'tinycudann.egg-info\SOURCES.txt' running build_ext error: [WinError 2] The system cannot find the file specified [end of output] note: This error originates from a subprocess, and is likely not a problem with pip. error: legacy-install-failure × Encountered error while trying to install package. ╰─> tinycudann note: This is an issue with the package mentioned above, not pip. hint: See above for output from the failure. @grshennan that is tough, I think I might have run into something similar but I don't remember. If you want (this may or may not be worth it) try building tinycudann in the way suggested in this comment? https://github.com/NVlabs/tiny-cuda-nn/issues/169#issuecomment-1365984644 Sorry I wasn't able to give you concrete help I'm fairly new to this topic. Would you explain, where you have placed the tiny-cuda-nn files and in which directory you have started the cmake process? Is that necessary for everything to work? I am receiving this error: Traceback (most recent call last): File "F:\Programme\miniconda3\envs\nerfstudio\lib\runpy.py", line 194, in _run_module_as_main return _run_code(code, main_globals, None, File "F:\Programme\miniconda3\envs\nerfstudio\lib\runpy.py", line 87, in _run_code exec(code, run_globals) File "F:\Programme\miniconda3\envs\nerfstudio\Scripts\ns-train.exe\__main__.py", line 7, in <module> File "F:\Programme\miniconda3\envs\nerfstudio\lib\site-packages\scripts\train.py", line 247, in entrypoint main( File "F:\Programme\miniconda3\envs\nerfstudio\lib\site-packages\scripts\train.py", line 233, in main launch( File "F:\Programme\miniconda3\envs\nerfstudio\lib\site-packages\scripts\train.py", line 172, in launch main_func(local_rank=0, world_size=world_size, config=config) File "F:\Programme\miniconda3\envs\nerfstudio\lib\site-packages\scripts\train.py", line 86, in train_loop trainer.setup() File "F:\Programme\miniconda3\envs\nerfstudio\lib\site-packages\nerfstudio\engine\trainer.py", line 145, in setup self.pipeline = self.config.pipeline.setup( File "F:\Programme\miniconda3\envs\nerfstudio\lib\site-packages\nerfstudio\configs\base_config.py", line 57, in setup return self._target(self, **kwargs) File "F:\Programme\miniconda3\envs\nerfstudio\lib\site-packages\nerfstudio\pipelines\base_pipeline.py", line 229, in __init__ self._model = config.model.setup( File "F:\Programme\miniconda3\envs\nerfstudio\lib\site-packages\nerfstudio\configs\base_config.py", line 57, in setup return self._target(self, **kwargs) File "F:\Programme\miniconda3\envs\nerfstudio\lib\site-packages\nerfstudio\models\base_model.py", line 82, in __init__ self.populate_modules() # populate the modules File "F:\Programme\miniconda3\envs\nerfstudio\lib\site-packages\nerfstudio\models\nerfacto.py", line 135, in populate_modules self.field = TCNNNerfactoField( File "F:\Programme\miniconda3\envs\nerfstudio\lib\site-packages\nerfstudio\fields\nerfacto_field.py", line 134, in __init__ self.direction_encoding = tcnn.Encoding( NameError: name 'tcnn' is not defined It sounds like tiny-cuda-nn wasn't installed correctly. I would try running pip install git+https://github.com/NVlabs/tiny-cuda-nn/#subdirectory=bindings/torch again. Also try opening up a python prompt, python, then running import tinycudann to see what error you get. It sounds like tiny-cuda-nn wasn't installed correctly. I would try running pip install git+https://github.com/NVlabs/tiny-cuda-nn/#subdirectory=bindings/torch again. Also try opening up a python prompt, python, then running import tinycudann to see what error you get. Thank you very much! The reinstallation actually worked, although the pip installation method of Tiny CUDA Neural Networks has never worked before for me. To document what I have done: Installed CUDA 11.7 and VS 2019 and deinstalled all other versions Created a conda envoriment conda create --name nerfstudio -y python=3.8 conda activate nerfstudio python -m pip install --upgrade pip Installed pytorch pip install torch==1.12.1+cu113 torchvision==0.13.1+cu113 -f https://download.pytorch.org/whl/torch_stable.html Git cloned tiny-cuda-nn via the repo into miniconda3\envs\nerfstudio\Lib\site-packages git clone --recursive https://github.com/nvlabs/tiny-cuda-nn cd tiny-cuda-nn cmake . -B build cmake --build build --config RelWithDebInfo -j (cmake commands run in VS developer command prompt) Installed nerfstudio pip install nerfstudio Reinstalled tiny-cuda-nn via pip pip install git+https://github.com/NVlabs/tiny-cuda-nn/#subdirectory=bindings/torch I am not sure, why the installation via pip worked without errors during the last step, but nerfstudio is now fully functional.
gharchive/issue
2022-12-27T00:39:30
2025-04-01T06:39:44.010334
{ "authors": [ "THE-COB", "brawlawnzo", "dooonnis", "grshennan", "tancik" ], "repo": "nerfstudio-project/nerfstudio", "url": "https://github.com/nerfstudio-project/nerfstudio/issues/1177", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1765417031
rich MarkupError Hello there, I'm running nerfstudio on an Azure Maching Learning VM with Ubuntu 18.04. When I try to process data with ns-process-data images I get an error which is thrown by the rich library after downscaling the images: I already processed polycam data which worked without problems. So this error only appears in combination with colmap. Hope you can help me and thanks in advance! I'm a bit in a hurry to get this running, so if someone could help me out here, I would really appreciate it. :) Can you check that your terminal is set up to use UTF-8 Can you check that your terminal is set up to use UTF-8 Yes, it is but it’s still not working. Here is the error information ────────────────────────────────────────────── 💀 💀 💀 ERROR 💀 💀 💀 ─────────────────────────────────────────────── Error running command: colmap feature_extractor --database_path data/nerfstudio/render/colmap/database.db --image_path data/nerfstudio/render/images --ImageReader.single_camera 1 --ImageReader.camera_model OPENCV --SiftExtraction.use_gpu 1 ──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────── Traceback (most recent call last): File "/home/yangtongyu/.conda/envs/nerfstudio/bin/ns-process-data", line 8, in <module> sys.exit(entrypoint()) File "/home/yangtongyu/nerf/nerfstudio/nerfstudio/scripts/process_data.py", line 408, in entrypoint tyro.cli(Commands).main() File "/home/yangtongyu/nerf/nerfstudio/nerfstudio/process_data/images_to_nerfstudio_dataset.py", line 82, in main self._run_colmap() File "/home/yangtongyu/nerf/nerfstudio/nerfstudio/process_data/colmap_converter_to_nerfstudio_dataset.py", line 199, in _run_colmap colmap_utils.run_colmap( File "/home/yangtongyu/nerf/nerfstudio/nerfstudio/process_data/colmap_utils.py", line 129, in run_colmap run_command(feature_extractor_cmd, verbose=verbose) File "/home/yangtongyu/nerf/nerfstudio/nerfstudio/utils/scripts.py", line 37, in run_command CONSOLE.print(out.stderr.decode("utf-8")) File "/home/yangtongyu/.conda/envs/nerfstudio/lib/python3.8/site-packages/rich/console.py", line 1673, in print renderables = self._collect_renderables( File "/home/yangtongyu/.conda/envs/nerfstudio/lib/python3.8/site-packages/rich/console.py", line 1537, in _collect_renderables self.render_str( File "/home/yangtongyu/.conda/envs/nerfstudio/lib/python3.8/site-packages/rich/console.py", line 1429, in render_str rich_text = render_markup( File "/home/yangtongyu/.conda/envs/nerfstudio/lib/python3.8/site-packages/rich/markup.py", line 158, in render print(pop_style(style_name)) File "/home/yangtongyu/.conda/envs/nerfstudio/lib/python3.8/site-packages/rich/markup.py", line 145, in pop_style raise KeyError(style_name) KeyError: 'home/conda/feedstock_root/build_artifacts/colmap_1686935602614/work/src/base/database.cc, line 270' (nerfstudio) yangtongyu@amax21-1:~/nerf/nerfstudio$ ns-process-data images --data data/nerfstudio/render --output-dir data/nerfstudio/render [12:02:24] 🎉 Done copying images. process_data_utils.py:241 [12:02:27] 🎉 Done downscaling images. process_data_utils.py:364 ────────────────────────────────────────────── 💀 💀 💀 ERROR 💀 💀 💀 ─────────────────────────────────────────────── Error running command: colmap feature_extractor --database_path data/nerfstudio/render/colmap/database.db --image_path data/nerfstudio/render/images --ImageReader.single_camera 1 --ImageReader.camera_model OPENCV --SiftExtraction.use_gpu 1 ──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────── Traceback (most recent call last): File "/home/yangtongyu/.conda/envs/nerfstudio/bin/ns-process-data", line 8, in <module> sys.exit(entrypoint()) File "/home/yangtongyu/nerf/nerfstudio/nerfstudio/scripts/process_data.py", line 408, in entrypoint tyro.cli(Commands).main() File "/home/yangtongyu/nerf/nerfstudio/nerfstudio/process_data/images_to_nerfstudio_dataset.py", line 82, in main self._run_colmap() File "/home/yangtongyu/nerf/nerfstudio/nerfstudio/process_data/colmap_converter_to_nerfstudio_dataset.py", line 199, in _run_colmap colmap_utils.run_colmap( File "/home/yangtongyu/nerf/nerfstudio/nerfstudio/process_data/colmap_utils.py", line 129, in run_colmap run_command(feature_extractor_cmd, verbose=verbose) File "/home/yangtongyu/nerf/nerfstudio/nerfstudio/utils/scripts.py", line 37, in run_command CONSOLE.print(out.stderr.decode("utf-8")) File "/home/yangtongyu/.conda/envs/nerfstudio/lib/python3.8/site-packages/rich/console.py", line 1673, in print renderables = self._collect_renderables( File "/home/yangtongyu/.conda/envs/nerfstudio/lib/python3.8/site-packages/rich/console.py", line 1537, in _collect_renderables self.render_str( File "/home/yangtongyu/.conda/envs/nerfstudio/lib/python3.8/site-packages/rich/console.py", line 1429, in render_str rich_text = render_markup( File "/home/yangtongyu/.conda/envs/nerfstudio/lib/python3.8/site-packages/rich/markup.py", line 161, in render raise MarkupError( rich.errors.MarkupError: closing tag '[/home/conda/feedstock_root/build_artifacts/colmap_1686935602614/work/src/base/database.cc, line 270]' at position 13 doesn't match any open tag Hey, is there another solution? Because my terminal supports UTF-8 but it's still not working. :( Hi everyone, has anyone finished this problem? I found that if I use only cpu to process it will be fine (but this is too slow). Does colmap/ns-process-data can use gpu? I'm using the official docker image and encontered the same problem. My solution is to edit line 37 of scripts.py, using try and except with print() instead of rich console if anything goes wrong. So that it prints the correct error messages. Here is the error information ────────────────────────────────────────────── 💀 💀 💀 ERROR 💀 💀 💀 ─────────────────────────────────────────────── Error running command: colmap feature_extractor --database_path data/nerfstudio/render/colmap/database.db --image_path data/nerfstudio/render/images --ImageReader.single_camera 1 --ImageReader.camera_model OPENCV --SiftExtraction.use_gpu 1 ──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────── Traceback (most recent call last): File "/home/yangtongyu/.conda/envs/nerfstudio/bin/ns-process-data", line 8, in <module> sys.exit(entrypoint()) File "/home/yangtongyu/nerf/nerfstudio/nerfstudio/scripts/process_data.py", line 408, in entrypoint tyro.cli(Commands).main() File "/home/yangtongyu/nerf/nerfstudio/nerfstudio/process_data/images_to_nerfstudio_dataset.py", line 82, in main self._run_colmap() File "/home/yangtongyu/nerf/nerfstudio/nerfstudio/process_data/colmap_converter_to_nerfstudio_dataset.py", line 199, in _run_colmap colmap_utils.run_colmap( File "/home/yangtongyu/nerf/nerfstudio/nerfstudio/process_data/colmap_utils.py", line 129, in run_colmap run_command(feature_extractor_cmd, verbose=verbose) File "/home/yangtongyu/nerf/nerfstudio/nerfstudio/utils/scripts.py", line 37, in run_command CONSOLE.print(out.stderr.decode("utf-8")) File "/home/yangtongyu/.conda/envs/nerfstudio/lib/python3.8/site-packages/rich/console.py", line 1673, in print renderables = self._collect_renderables( File "/home/yangtongyu/.conda/envs/nerfstudio/lib/python3.8/site-packages/rich/console.py", line 1537, in _collect_renderables self.render_str( File "/home/yangtongyu/.conda/envs/nerfstudio/lib/python3.8/site-packages/rich/console.py", line 1429, in render_str rich_text = render_markup( File "/home/yangtongyu/.conda/envs/nerfstudio/lib/python3.8/site-packages/rich/markup.py", line 158, in render print(pop_style(style_name)) File "/home/yangtongyu/.conda/envs/nerfstudio/lib/python3.8/site-packages/rich/markup.py", line 145, in pop_style raise KeyError(style_name) KeyError: 'home/conda/feedstock_root/build_artifacts/colmap_1686935602614/work/src/base/database.cc, line 270' (nerfstudio) yangtongyu@amax21-1:~/nerf/nerfstudio$ ns-process-data images --data data/nerfstudio/render --output-dir data/nerfstudio/render [12:02:24] 🎉 Done copying images. process_data_utils.py:241 [12:02:27] 🎉 Done downscaling images. process_data_utils.py:364 ────────────────────────────────────────────── 💀 💀 💀 ERROR 💀 💀 💀 ─────────────────────────────────────────────── Error running command: colmap feature_extractor --database_path data/nerfstudio/render/colmap/database.db --image_path data/nerfstudio/render/images --ImageReader.single_camera 1 --ImageReader.camera_model OPENCV --SiftExtraction.use_gpu 1 ──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────── Traceback (most recent call last): File "/home/yangtongyu/.conda/envs/nerfstudio/bin/ns-process-data", line 8, in <module> sys.exit(entrypoint()) File "/home/yangtongyu/nerf/nerfstudio/nerfstudio/scripts/process_data.py", line 408, in entrypoint tyro.cli(Commands).main() File "/home/yangtongyu/nerf/nerfstudio/nerfstudio/process_data/images_to_nerfstudio_dataset.py", line 82, in main self._run_colmap() File "/home/yangtongyu/nerf/nerfstudio/nerfstudio/process_data/colmap_converter_to_nerfstudio_dataset.py", line 199, in _run_colmap colmap_utils.run_colmap( File "/home/yangtongyu/nerf/nerfstudio/nerfstudio/process_data/colmap_utils.py", line 129, in run_colmap run_command(feature_extractor_cmd, verbose=verbose) File "/home/yangtongyu/nerf/nerfstudio/nerfstudio/utils/scripts.py", line 37, in run_command CONSOLE.print(out.stderr.decode("utf-8")) File "/home/yangtongyu/.conda/envs/nerfstudio/lib/python3.8/site-packages/rich/console.py", line 1673, in print renderables = self._collect_renderables( File "/home/yangtongyu/.conda/envs/nerfstudio/lib/python3.8/site-packages/rich/console.py", line 1537, in _collect_renderables self.render_str( File "/home/yangtongyu/.conda/envs/nerfstudio/lib/python3.8/site-packages/rich/console.py", line 1429, in render_str rich_text = render_markup( File "/home/yangtongyu/.conda/envs/nerfstudio/lib/python3.8/site-packages/rich/markup.py", line 161, in render raise MarkupError( rich.errors.MarkupError: closing tag '[/home/conda/feedstock_root/build_artifacts/colmap_1686935602614/work/src/base/database.cc, line 270]' at position 13 doesn't match any open tag same problem, seems like my output dir is within my NAS, which may goes wrong with some linux file permissions. When I change the output dir to a folder under docker /workspace/ , everything goes well. Agree with @HugoFeng and @deephog, the issue might be coming from a database.db lock, which is based on it's location. I'm working on a remote Linux machine and ran into the issue. I moved the output folder somewhere up the directories and it worked as expected. hey guys I got a possible solution, If you're mounting the NAS on Linux using the mount command, make sure to toss in the nobrl option. This little trick helps avoid those annoying locking problems. Just run something like this: sudo mount -t cifs -o username=USERNAME,uid=$(id -u),gid=$(id -g),iocharset=utf8,nobrl //192.168.X.XX/Main /nas/Main This should sort out those pesky errors about file permissions and database locks.
gharchive/issue
2023-06-20T13:30:08
2025-04-01T06:39:44.022251
{ "authors": [ "HugoFeng", "Jordan-Pierce", "Terry10086", "forEachWhileTrue", "lan-creator", "qf-wei", "tancik" ], "repo": "nerfstudio-project/nerfstudio", "url": "https://github.com/nerfstudio-project/nerfstudio/issues/2105", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2314454244
Make templates resolutions the Coordinate's responsibility The puzzle 23-93ef7c09 from #23 has to be resolved: https://github.com/nergal-perm/java-decita/blob/18bd62953fdb5dfe351db2621e3a14e4e8db6144/java-decita/src/main/java/ru/ewc/commands/SimpleCommand.java#L68-L68 The puzzle was created by Евгений Терехов on 24-May-24. role: DEV. If you have any technical questions, don't ask me, submit new tickets instead. The task will be "done" when the problem is fixed and the text of the puzzle is removed from the source code. Here is more about PDD and about me. The puzzle 23-93ef7c09 has disappeared from the source code, that's why I closed this issue.
gharchive/issue
2024-05-24T05:53:07
2025-04-01T06:39:44.026004
{ "authors": [ "0pdd" ], "repo": "nergal-perm/java-decita", "url": "https://github.com/nergal-perm/java-decita/issues/120", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2544526323
Snakefile 1 - First Production level Snakefile for 06-slurm-array/array_create_db_files.sl Proposed Snakefile Following Snakefile defines a configuration with the Apptainer settings and excluded files. It gets a list of .apsimx files, excluding the specified files. The all rule defines the final output files we want to create. The process_apsimx rule defines how to process each .apsimx file: It requires both the .apsimx file and the corresponding .db ( this is the placeholder .db) file as input. It creates a .processed file as output to mark completion. Perhaps we can modify this later to save file quota sets the resources (memory and time) and number of threads. The shell command loads the Apptainer module, sets the bind path, and runs the Models command. import os from glob import glob # Configuration config = { "apptainer_bind": "/agr/scratch,/agr/persist", "apptainer_image": "/agr/persist/projects/2024_apsim_improvements/apsim-simulations/container/apsim-2024.09.7579.0.aimg", "excluded_files": ["2023-10-09_MasterSoilApsimLibrary.apsimx", "LargerExample.apsimx"] } # Get list of .apsimx files apsimx_files = [f for f in glob("*.apsimx") if f not in config["excluded_files"]] rule all: input: expand("{file}.processed", file=[os.path.splitext(f)[0] for f in apsimx_files]) rule process_apsimx: input: apsimx = "{file}.apsimx", db = "{file}.db" output: "{file}.processed" resources: mem_mb = 8000, time = "00:10:00" threads: 12 shell: """ module load Apptainer export APPTAINER_BIND="{config[apptainer_bind]}" apptainer exec {config[apptainer_image]} Models {input.apsimx} touch {output} """ Configure Snakemake - These instructions are for snakemake > 8 Create a configuration directory for Snakemake profiles: profile_dir="${HOME}/.config/snakemake" mkdir -p "$profile_dir" Use cookiecutter to create the Slurm profile template: If cookiecutter isn't installed, load snakemake module and run pip install --user cookiecutter template="gh:Snakemake-Profiles/slurm" cookiecutter --output-dir "$profile_dir" "$template" During the cookiecutter process, you'll be prompted to set values for your profile. For example: profile_name [slurm]: slurm.my_account sbatch_defaults []: account=my_account no-requeue exclusive cluster_sidecar_help: [Use cluster sidecar. NB! Requires snakemake >= 7.0! Enter to continue...] Select cluster_sidecar: 1 - yes 2 - no Choose from 1, 2 : cluster_name []: After completing the prompts, the profile scripts and configuration file will be installed in the $profile_dir as profile_name/[2]. You can then use this profile when running Snakemake by adding the --profile flag: snakemake --profile slurm.my_account ... This will configure Snakemake to submit jobs to Slurm using the settings you specified in the profile[2]. Remember that the profile typically includes scripts for job submission, status checking, and a configuration file (config.yaml) that defines default values for Snakemake command line arguments. You may need to adjust these files to match your specific Slurm environment and requirements. Working Snakefile for this rule : https://github.com/nesi/APSIM-HPC/pull/54
gharchive/issue
2024-09-24T07:01:10
2025-04-01T06:39:44.044323
{ "authors": [ "DininduSenanayake" ], "repo": "nesi/APSIM-HPC", "url": "https://github.com/nesi/APSIM-HPC/issues/53", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1996805575
Initial implementation of automatic MQTT config retrieval from HA Supervisor Breaking change Should be no breaking changes. This only touches the MQTT extension. Proposed change The MQTT configuration should be retrieved from HA Supervisor. I've explained the process in this issue: #1002 Type of change [ ] Dependency upgrade [ ] Bugfix (non-breaking change which fixes an issue) [x ] New feature (which adds functionality to an existing integration) [ ] Breaking change (fix/feature causing existing functionality to break) [ ] Code quality improvements to existing code or addition of tests Additional information This PR fixes or closes issue: fixes #1002 This PR is related to issue: #1002 Link to documentation pull request: Checklist [ ] The code change is tested and works locally. [ ] Local tests pass. Your PR cannot be merged unless tests pass [ ] There is no commented out code in this PR. [ ] I have followed the [development checklist][dev-checklist] [ ] The code compiles without warnings (code quality chek) [ ] Tests have been added to verify that the new code works. If user exposed functionality or configuration variables are added/changed: [ ] Documentation added/updated for www.home-assistant.io First draft, quickly thrown together. You may want to make some adjustments here: https://github.com/net-daemon/netdaemon/blob/main/Docker/rootfs/etc/services.d/netdaemon_addon/run#L14-L22 Also I couldn't find if/where you define a config.yaml file for the netdaemon. I think you would want to express that the app wants rather than needs the mqtt service. HA can only provide details for it if it's installed as an addon. In my case, I don't use run mosquitto as an addon (it runs on another machine). Ideally (I think) you'd want to collect optional MQTT config from users for people who use external brokers via config.yaml and pass it through to the right env vars if it's defined in the run command above. Consider rewrite, Mqtt env vars is already set på in run script.
gharchive/pull-request
2023-11-16T12:59:38
2025-04-01T06:39:44.151240
{ "authors": [ "LasseSkogland", "scottt732" ], "repo": "net-daemon/netdaemon", "url": "https://github.com/net-daemon/netdaemon/pull/1003", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1396547767
Dependency upgrades Breaking change Proposed change Type of change [x] Dependency upgrade [ ] Bugfix (non-breaking change which fixes an issue) [ ] New feature (which adds functionality to an existing integration) [ ] Breaking change (fix/feature causing existing functionality to break) [ ] Code quality improvements to existing code or addition of tests Additional information This PR fixes or closes issue: fixes # This PR is related to issue: Link to documentation pull request: Checklist [x] The code change is tested and works locally. [ ] Local tests pass. Your PR cannot be merged unless tests pass [ ] There is no commented out code in this PR. [ ] I have followed the [development checklist][dev-checklist] [ ] The code compiles without warnings (code quality chek) [ ] Tests have been added to verify that the new code works. If user exposed functionality or configuration variables are added/changed: [ ] Documentation added/updated for www.home-assistant.io Pull Request Test Coverage Report for Build 3183812732 0 of 0 changed or added relevant lines in 0 files are covered. 4 unchanged lines in 1 file lost coverage. Overall coverage decreased (-0.03%) to 63.208% Files with Coverage Reduction New Missed Lines % src/Daemon/NetDaemon.Daemon/Daemon/NetDaemonHost.cs 4 60.26% Totals Change from base Build 2943640611: -0.03% Covered Lines: 4505 Relevant Lines: 6666 💛 - Coveralls
gharchive/pull-request
2022-10-04T16:27:41
2025-04-01T06:39:44.162806
{ "authors": [ "coveralls", "helto4real" ], "repo": "net-daemon/netdaemon_v2", "url": "https://github.com/net-daemon/netdaemon_v2/pull/71", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1955255692
Ability to provide credentials When trying to download an Instgram video the following message appears: Provided the following link: https://www.instagram.com/reel/CybzFHyOcUo/?igshid=MzRlODBiNWFlZA== I'd like to add on to this issue. It appears that the frontend takes links from Gallery-dl before G-DL begins downloading, which causes it to download from sources without the credentials I've manually provided to G-DL, resulting in 403 forbidden errors being downloaded as html files instead. In my case, it was Pixiv that didn't work for all R18-tagged posts. And I don't think it would work on SW-DLT without major revision to how the frontend downloads galleries since Pixiv requires OAuth instead just providing username and password (G-DL uses an Android user-agent and Pixiv on Android browsers require OAuth afaik). For accounts that work with username & password, you can look in the [docs[(https://github.com/net00-1/SW-DLT/blob/master/Docs.md#authentication). There's a guide to setup a-shell to keep your credentials. I will keep this open to check for ways of other authentication types As a newbie, do I have this directory structure set up correctly? It keeps prompting "requires. authentication" [~group]$ ls -Ra . .. .com.apple.mobile_container_manager.metadata.plist .netrc Library SW_DLT.py gallery-dl yt-dlp ./Library: . Caches .. Preferences ./Library/Caches: . .. ./Library/Preferences: . .. group.AsheKube.a-Shell.plist ./gallery-dl: . config.json .. ./yt-dlp: . .. config [~group]$ cat yt-dlp/config --netrc [~group]$ cat gallery-dl/config.json { "extractor": { "twitter": { "username": "$$$$$$", "password": "$$$$$$" } }} Now much more convenient cookie-based authentication is available, you can check the updated docs about how to use it.
gharchive/issue
2023-10-21T03:00:34
2025-04-01T06:39:44.175907
{ "authors": [ "Cs0o", "kriodoxis", "nReus", "net00-1" ], "repo": "net00-1/SW-DLT", "url": "https://github.com/net00-1/SW-DLT/issues/38", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1049688029
Add standalone Seafile Docker tutorial This PR adds a tutorial on how to set up a standalone Seafile server including a valid LetsEncrypt certificate for TLS/SSL secured HTTPS connections. Please notice that the linked tutorial on how to install Docker and Docker-Compose is not merged yet (https://github.com/netcup-community/community-tutorials/pull/17). I have read and understood the Contributor's Certificate of Origin at the end of the template and I hereby certify that I meet the contribution criteria described in it. Signed-off-by: Max Kratz account@maxkratz.com Hi maxkratz, thank you for contributing! It's been a while since you opened this PR and it probably is an open secret we're currently overwhelmed with reviewing all the awesome tutorials we got. Please be patient - we're about to publish tutorials this and next week. Thank you! @netcupClaudiaM Thank you for the update! Take the time you need for all the reviews :). Hey @maxkratz, could you please upload all your pictures in english so that everything is consistent? Thanks! @raphi-0901 No problem, I've fixed the German screenshots. Thank you for the suggestion. Furthermore, I've added a configuration option for the services to automatically start in case the system had to be (re-)booted.
gharchive/pull-request
2021-11-10T11:06:38
2025-04-01T06:39:44.227350
{ "authors": [ "maxkratz", "netcupClaudiaM", "raphi-0901" ], "repo": "netcup-community/community-tutorials", "url": "https://github.com/netcup-community/community-tutorials/pull/55", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
175928089
android studio 2.2rc2 使用jack compiler 编译出现的问题 android studio 2.2rc2 使用jack compiler 编译出现的问题 09-09 12:51:41.657 18228-18228/com.netease.nim.demo E/AndroidRuntime: FATAL EXCEPTION: main Process: com.netease.nim.demo, PID: 18228 java.lang.IllegalAccessError: Illegal class access: 'com.netease.nimlib.sdk.NIMClient' attempting to access 'com.netease.nimlib.r.g' (declaration of 'com.netease.nimlib.sdk.NIMClient' appears in /data/app/com.netease.nim.demo-2/base.apk) at com.netease.nimlib.sdk.NIMClient.init(Unknown) at com.netease.nim.demo.NimApplication.onCreate(NimApplication.java:68) at android.app.Instrumentation.callApplicationOnCreate(Instrumentation.java:1015) at android.app.ActivityThread.handleBindApplication(ActivityThread.java:4815) at android.app.ActivityThread.access$1600(ActivityThread.java:168) at android.app.ActivityThread$H.handleMessage(ActivityThread.java:1440) at android.os.Handler.dispatchMessage(Handler.java:102) at android.os.Looper.loop(Looper.java:150) at android.app.ActivityThread.main(ActivityThread.java:5643) at java.lang.reflect.Method.invoke(Native Method) at com.android.internal.os.ZygoteInit$MethodAndArgsCaller.run(ZygoteInit.java:799) at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:689) 怎么样解决呢? 我现在也出现了这个,你那边解决了吗?! @zengbobo0805 用kotlin了
gharchive/issue
2016-09-09T04:54:00
2025-04-01T06:39:44.285006
{ "authors": [ "Ccixyj", "zengbobo0805" ], "repo": "netease-im/NIM_Android_UIKit", "url": "https://github.com/netease-im/NIM_Android_UIKit/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
616328041
鼠标进入 鼠标离开消息 如果这个控件的子控件 mouse属性都是false时, pVideoBox_->AttachMouseEnter([this](ui::EventArgs* param){ scale_menu_box_->SetVisible(true); return true; }); pVideoBox_->AttachMouseLeave([this](ui::EventArgs* param){ scale_menu_box_->SetVisible(true); return true; }); 要实现的,鼠标进入就显示某子控件 鼠标离开就隐藏某子控件 功能可以实现 但如果子控件mouse为true的 时候, 子控件截取了消息, 导致如上功能无法实现 我现在使用的 绘制时,通过判断鼠标位置 来动态隐藏与显示 还有更好的方式么 AttachBubbledEvent(ui::kEventMouseEnter AttachBubbledEvent(ui::kEventMouseLeave 使用bubbledEvent 恩恩 测试成功了 这个AttachBubbledEvent 是什么意思 这样虽然有子控件,但是也能收到鼠标进入消息 是不是,这个的意思是 控件收到了什么消息 不管有没有被子控件截取
gharchive/issue
2020-05-12T03:21:11
2025-04-01T06:39:44.288041
{ "authors": [ "shuilan0066", "smallevilbeast" ], "repo": "netease-im/NIM_Duilib_Framework", "url": "https://github.com/netease-im/NIM_Duilib_Framework/issues/206", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
629832355
怎么可以创建一个模态子窗口,然后主窗口等待子窗口的返回 模态窗口可以使用ShowModalFake函数,但如何能让主窗口阻塞在哪里,知道子窗口操作完成后,返回一个结果给主窗口呢;看了ShowMsgBox和文件对话框的实现 都没有找到好的方法 我是封装mfc对话框到dll,然后在导出的函数里doModal.... 我是封装mfc对话框到dll,然后在导出的函数里doModal.... 那弹出来就可以是模态对话框了吗? 我是封装mfc对话框到dll,然后在导出的函数里doModal.... 那弹出来就可以是模态对话框了吗? 是的,因为我的对话框是老项目里的对话框,封到dll里就能直接用了。本质上就是写一个新的消息循环,代码就在循环里了。
gharchive/issue
2020-06-03T09:13:35
2025-04-01T06:39:44.290568
{ "authors": [ "lsaejn", "yzxfz" ], "repo": "netease-im/NIM_Duilib_Framework", "url": "https://github.com/netease-im/NIM_Duilib_Framework/issues/215", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1318530620
Infrastructure: bump velero version and add selector Description This PR bumps the version of velero, and introduces a label selector to filter out the resources originated from CrownLabs instances, in order to prevent the partial failures due to the kubevirt/velero issue. Fixes # (issue) How Has This Been Tested? Please describe the tests that you ran to verify your changes. Please also note any relevant details for your test configuration. [x] Deploying the updated manifest in the cluster, and checking that the backup completed correctly /merge
gharchive/pull-request
2022-07-26T17:06:25
2025-04-01T06:39:44.292652
{ "authors": [ "giorio94" ], "repo": "netgroup-polito/CrownLabs", "url": "https://github.com/netgroup-polito/CrownLabs/pull/801", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
238822766
Fix values link on homepage Values link was pointing to /about. LGTM
gharchive/pull-request
2017-06-27T11:41:24
2025-04-01T06:39:44.328630
{ "authors": [ "calavera", "hdlopes" ], "repo": "netlify-templates/kaldi-hugo-cms-template", "url": "https://github.com/netlify-templates/kaldi-hugo-cms-template/pull/17", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
680134328
invlide update available messge When running "netlify dev function:invoke" I got a message saying a new version of the cli was available and I should run npm -i -g netlify-cli -i is invalid and should be just i I can't find this in the source code so guess it's a message from somewhere else. Thanks @SteveALee, it is probably coming from https://github.com/yeoman/update-notifier. Can you please share the output of npx envinfo --system --binaries --npmPackages netlify-cli --npmGlobalPackages netlify-cli so we can try and reproduce using the same environment? Sadly I reinstalled the cli - after having to delete some files (ntl.psi and ntl.cmd) from ...AppData/Roaming/npm/, even after npm rm -g netlify-cli But anyway C:\Users\steve>npx envinfo --system --binaries --npmPackages netlify-cli --npmGlobalPackages netlify-cli npx: installed 1 in 1.639s System: OS: Windows 10 10.0.19041 CPU: (8) x64 Intel(R) Core(TM) i7-8550U CPU @ 1.80GHz Memory: 4.08 GB / 15.86 GB Binaries: Node: 14.6.0 - C:\Program Files\nodejs\node.EXE npm: 6.14.6 - C:\Program Files\nodejs\npm.CMD Did you expect netlify-cli there? C:\Users\steve>npm ls -g netlify-cli C:\Users\steve\AppData\Roaming\npm `-- netlify-cli@2.59.1 Yeah I was expecting to see see the CLI, but maybe we have some installation issues with Windows machines. Thanks for the information. yes the update notifier text looks correct. And the version bumped .1 Perhaps I imaged the -i ? Closing as stale. Please re-open if needed
gharchive/issue
2020-08-17T10:41:38
2025-04-01T06:39:44.333574
{ "authors": [ "SteveALee", "erezrokah", "sarahetter" ], "repo": "netlify/cli", "url": "https://github.com/netlify/cli/issues/1116", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
753690520
add resource links to stages many stages refer to links and resources, so it might be a good idea to put those resources in a list Do you think we should separate this out in the UI, like make a text field Resource List in the stage data object? that's probably not the worst idea, yeah covered by #431 covered by #431
gharchive/issue
2020-11-30T18:07:57
2025-04-01T06:39:44.335250
{ "authors": [ "jlengstorf", "tzmanics" ], "repo": "netlify/explorers", "url": "https://github.com/netlify/explorers/issues/430", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
329371462
Hidden field in editor component is rendered - Do you want to request a feature or report a bug? bug - What is the current behavior? An editor component that declares a hidden field will have it rendered: - If the current behavior is a bug, please provide the steps to reproduce. Declare an editor component that has a field with widget: 'hidden': CMS.registerEditorComponent({ // Internal id of the component id: "youtube", // Visible label label: "Youtube", // Fields the user need to fill out when adding an instance of the component fields: [{ name: 'id', label: 'Youtube Video ID', widget: 'string' }, { name: 'myHiddenField', label: 'My hidden field', widget: 'hidden' }], // Pattern to identify a block as being an instance of this component pattern: /^youtube (\S+)$/, // Function to extract data elements from the regexp match fromBlock: function(match) { return { id: match[1] }; }, // Function to create a text block from an instance of this component toBlock: function(obj) { return 'youtube ' + obj.id; }, // Preview output for this component. Can either be a string or a React component // (component gives better render performance) toPreview: function(obj) { return ( '<img src="http://img.youtube.com/vi/' + obj.id + '/maxresdefault.jpg" alt="Youtube Video"/>' ); } }); Add an instance of the component to a markdown editor. Observe that the label for the hidden field is rendered with a No control for widget: 'hidden' message below it. The same thing happens if the hidden field has no label except that a small gray box is rendered instead of the label. - What is the expected behavior? That the field isn't rendered, similar to how hidden widgets work for collections. - Please mention your versions where applicable. Netlify CMS version: 1.8.4 Browser version: Chrome 67.0.3396.62/OSX Node.JS version: 9.5.0 Operating System: OSX - Please link or paste your config.yml below if applicable. Good catch!
gharchive/issue
2018-06-05T09:23:22
2025-04-01T06:39:44.341611
{ "authors": [ "erquhart", "papandreou" ], "repo": "netlify/netlify-cms", "url": "https://github.com/netlify/netlify-cms/issues/1405", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
726469152
config should have required property 'options' - Clearstatus Netlify I get this error post a successful build and deploy on netlify. https://youthful-swartz-68b931.netlify.app/ Error loading the CMS configuration Config Errors: config should have required property 'options' Check your config.yml file. Hi @thelynchpin, one of your select widgets is missing the options configuration: fields: - label: "Name" name: "name" widget: string - label: "Group" name: "group" widget: select multiple: false required: false - label: "Description" name: "description" widget: string required: false - label: "Link" name: "link" widget: string required: false I'm new to this and have deployed the Clearstatus app through the deploy to netlify button on https://github.com/weeblrpress/clearstatus. Please specify what needs to be modified on the list. Welcome :) Please see the docs for the select widget here: https://www.netlifycms.org/docs/widgets/#select I suggest you open an issue on that repo and ask where is the config.yml file is located @thelynchpin Can you give it a try now, I think we fixed the problem. @erezrokah What seems to have happened is that previously Netlify CMS would accept that select without options (which depends on the user particular site configuration) but at some point this started to trigger an error. I'm now removing the entire select if no "groups" option is configured by user. @erezrokah What seems to have happened is that previously Netlify CMS would accept that select without options (which depends on the user particular site configuration) but at some point this started to trigger an error. I'm now removing the entire select if no "groups" option is configured by user. Yes, we added schema validations for widgets configurations. The widget used output an error div prior to that: https://github.com/netlify/netlify-cms/blob/519cb2d4c2db729d2643c9116f93656b6a9dba23/packages/netlify-cms-widget-select/src/SelectControl.js#L114 Thanks. It's working now. Regards, Wilbur On Fri, Nov 6, 2020 at 3:13 PM clearstatus notifications@github.com wrote: @thelynchpin https://github.com/thelynchpin Can you give it a try now, I think we fixed the problem. @erezrokah https://github.com/erezrokah What seems to have happened is that previously Netlify CMS would accept that select without options (which depends on the user particular site configuration) but at some point this started to trigger an error. I'm now removing the entire select if no "groups" option is configured by user. — You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub https://github.com/netlify/netlify-cms/issues/4489#issuecomment-722983208, or unsubscribe https://github.com/notifications/unsubscribe-auth/ACEOACCGO2YD2AOUDIJEMOLSOPANZANCNFSM4SZYORPA .
gharchive/issue
2020-10-21T13:13:23
2025-04-01T06:39:44.351415
{ "authors": [ "clearstatus", "erezrokah", "thelynchpin" ], "repo": "netlify/netlify-cms", "url": "https://github.com/netlify/netlify-cms/issues/4489", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1433574428
[Bug]: fail build if output: 'standalone' config option is set Summary Those coming from the docker/k8s world are likely to follow this guidance from Next.js official docs, which instructs them to clone the with-docker example, whose config explicitly uses output: 'standalone': module.exports = { output: 'standalone', } This will create a folder at .next/standalone which can then be deployed on its own without installing node_modules. Additionally, a minimal server.js file is also output which can be used instead of next start. This minimal server does not copy the public or .next/static folders by default as these should ideally be handled by a CDN instead, although these folders can be copied to the standalone/public and standalone/.next/static folders manually, after which server.js file will serve these automatically. As this relies on an always-running node.js server, it's incompatible with our serverless lyfestyle. @ascorbic suggested that rather than changing the site's next.config.js file at build-time, we fail the build and output an error message: I think the most useful then would be for us to detect that setting and fail the build, with an appropriate message Having this error message will be immensely useful for those coming from Next.js wrapped in docker. Steps to reproduce Clone and deploy the with-docker example from Next.js examples A link to a reproduction repository No response Next Runtime version 4.28.3 More information about your build [ ] I am building using the CLI [ ] I am building using file-based configuration (netlify.toml) What OS are you using? No response Your netlify.toml file No response Your public/_redirects file No response Your next.config.js file `next.config.js` module.exports = { output: 'standalone', } Builds logs (or link to your logs) No response Function logs No response .next JSON files No response @ascorbic Is this still relevant with v5? It won't be, because it's forced to standalone anyway
gharchive/issue
2022-11-02T18:08:40
2025-04-01T06:39:44.359853
{ "authors": [ "ascorbic", "jasonbarry", "serhalp" ], "repo": "netlify/next-runtime", "url": "https://github.com/netlify/next-runtime/issues/1736", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1980316070
update README image with new logo Description Removes the old version of the README image and replaces it with one donning the new Netlify logo ✨ Documentation Tests Visit the README file See the change (updated Netlify logo) Relevant links (GitHub issues, etc.) or a picture of cute animal Superseded by v5. (I checked and it has the updated logo.)
gharchive/pull-request
2023-11-07T00:30:40
2025-04-01T06:39:44.362708
{ "authors": [ "iKristy", "serhalp" ], "repo": "netlify/next-runtime", "url": "https://github.com/netlify/next-runtime/pull/2355", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1633834592
NETOBSERV-857 After some time, it fails to retrieve flows Reload HOST token from file on every query /ok-to-test /lgtm up to you to do what I suggested above or not .. this isn't super critical /label qe-approved Verified no longer running into the cert issue with this change when authToken: HOST is set. @jpinsonneau are we good to merge? /approve
gharchive/pull-request
2023-03-21T12:40:49
2025-04-01T06:39:44.376200
{ "authors": [ "jotak", "jpinsonneau", "memodi" ], "repo": "netobserv/network-observability-console-plugin", "url": "https://github.com/netobserv/network-observability-console-plugin/pull/310", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
973208679
DatagramDnsQueryDecoder.newQuery() raises IndexOutOfBoundsException Actual behavior I tried to upgrade Netty 4.1.66 to 4.1.67, but our CI tests produced the following error when resolving a DNS query. io.netty.handler.codec.DecoderException: java.lang.IndexOutOfBoundsException: readerIndex(36) + length(2) exceeds writerIndex(36): PooledUnsafeDirectByteBuf(ridx: 36, widx: 36, cap: 2048) at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:98) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919) at io.netty.channel.nio.AbstractNioMessageChannel$NioMessageUnsafe.read(AbstractNioMessageChannel.java:97) at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:719) at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:655) at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:581) at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:493) at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:986) at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) at java.base/java.lang.Thread.run(Thread.java:831) Caused by: java.lang.IndexOutOfBoundsException: readerIndex(36) + length(2) exceeds writerIndex(36): PooledUnsafeDirectByteBuf(ridx: 36, widx: 36, cap: 2048) at io.netty.buffer.AbstractByteBuf.checkReadableBytes0(AbstractByteBuf.java:1442) at io.netty.buffer.AbstractByteBuf.readShort(AbstractByteBuf.java:749) at io.netty.buffer.AbstractByteBuf.readUnsignedShort(AbstractByteBuf.java:765) at io.netty.handler.codec.dns.DatagramDnsQueryDecoder.newQuery(DatagramDnsQueryDecoder.java:87) at io.netty.handler.codec.dns.DatagramDnsQueryDecoder.decode(DatagramDnsQueryDecoder.java:64) at io.netty.handler.codec.dns.DatagramDnsQueryDecoder.decode(DatagramDnsQueryDecoder.java:33) at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:88) ... 16 common frames omitted When DatagramDnsQueryDecoder.newQuery(packet, but) was about to called, the buf had been already fully consumed by DnsMessageUtil.decodeDnsQuery(). https://github.com/netty/netty/blob/794c45fd590ffc51b16bb4b8af87556c0d2025fb/codec-dns/src/main/java/io/netty/handler/codec/dns/DatagramDnsQueryDecoder.java#L57-L64 I guess the problem is caused by the changes in https://github.com/netty/netty/pull/11415/files#diff-41d08a26f08dcb6f92085ad9bb565c833bbe0e0d425e0fe7147fcc51b56d164aR64 Minimal yet complete reproducer code (or URL to code) https://github.com/line/armeria/blob/b5a2e1c264c264a4f504dfc1a5990866dea7b86f/core/src/test/java/com/linecorp/armeria/client/DnsMetricsTest.java#L67 Netty version 4.1.67.Final JVM version (e.g. java -version) OpenJDK 64-Bit Server VM AdoptOpenJDK (build 15.0.1+9, mixed mode, sharing) OS version (e.g. uname -a) Darwin 20.5.0 Darwin Kernel Version 20.5.0: Sat May 8 05:10:33 PDT 2021; root:xnu-7195.121.3~9/RELEASE_X86_64 x86_64 Doh! Yeah let me fix this @ikhoon PTAL https://github.com/netty/netty/pull/11592
gharchive/issue
2021-08-18T02:13:57
2025-04-01T06:39:44.407750
{ "authors": [ "ikhoon", "normanmaurer" ], "repo": "netty/netty", "url": "https://github.com/netty/netty/issues/11591", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
138919609
when i use netty-all-4.1.0.CR3 on my project ,happen this error.how about this happend,and how can i do it I make a ssl socket server base on netty-4.1.0.CR3,with a lot of connections (20k<). # # A fatal error has been detected by the Java Runtime Environment: # # SIGSEGV (0xb) at pc=0x00007f5e19fe8e17, pid=15117, tid=140042109077248 # # JRE version: Java(TM) SE Runtime Environment (7.0_65-b17) (build 1.7.0_65-b17) # Java VM: Java HotSpot(TM) 64-Bit Server VM (24.65-b04 mixed mode linux-amd64 compressed oops) # Problematic frame: # V [libjvm.so+0x885e17] oopDesc* PSPromotionManager::copy_to_survivor_space<false>(oopDesc*)+0x177 # # Core dump written. Default location: /root/netty_srv_dir/core or core.15117 # # If you would like to submit a bug report, please visit: # http://bugreport.sun.com/bugreport/crash.jsp # --------------- T H R E A D --------------- Current thread (0x00007f5e14021000): GCTaskThread [stack: 0x00007f5e181a4000,0x00007f5e182a5000] [id=15121] siginfo:si_signo=SIGSEGV: si_errno=0, si_code=2 (SEGV_ACCERR), si_addr=0x00007f568bdff018 Registers: RAX=0x00007f5e1a5839a0, RBX=0x00007f5704912834, RCX=0x0000000000000003, RDX=0x00007f568bdff000 RSP=0x00007f5e182a39a0, RBP=0x00007f5e182a3a10, RSI=0x00007f5bb0b26af8, RDI=0x00007f568bdff010 R8 =0x0000000000000001, R9 =0x00007f5e1a59f178, R10=0x00000000e6a82202, R11=0x00007f5e182a3950 R12=0x00007f5bb0b26af8, R13=0x0000000000000000, R14=0x0000000000000000, R15=0x00007f5e141b6680 RIP=0x00007f5e19fe8e17, EFLAGS=0x0000000000010206, CSGSFS=0x0000000000000033, ERR=0x0000000000000004 TRAPNO=0x000000000000000e Top of Stack: (sp=0x00007f5e182a39a0) 0x00007f5e182a39a0: 00007f5e182a3a20 00007f5e1a59f178 0x00007f5e182a39b0: 0000000000004230 00007f5e141b6610 0x00007f5e182a39c0: 00007f5704911320 0000000000000400 0x00007f5e182a39d0: 00007f5d79251d72 0000000000001000 0x00007f5e182a39e0: 00007f5e141b6640 00007f5704912834 0x00007f5e182a39f0: 00007f5e141b6610 00007f5e1a5839a0 0x00007f5e182a3a00: 0000000000000000 00007f5e141b6680 0x00007f5e182a3a10: 00007f5e182a3b00 00007f5e19fe85f0 0x00007f5e182a3a20: 00007f5e182a3aa0 00007f5e182a3a90 0x00007f5e182a3a30: 00007f5e182a3aa0 00007f5e182a3a60 0x00007f5e182a3a40: 00007f5e182a3a70 00007f5e182a3ac0 0x00007f5e182a3a50: 00007f5700468370 00007f5e141b66b0 0x00007f5e182a3a60: 00007f5e14022728 0000000000000000 0x00007f5e182a3a70: 00007f5e1ada6609 00007f57004697b4 0x00007f5e182a3a80: 00007f5e182a3b00 00007f5e19f5ebcf 0x00007f5e182a3a90: 00007f569104ee01 00007f5e19f5ebcf 0x00007f5e182a3aa0: 00007f568c608928 00007f569104ee00 0x00007f5e182a3ab0: 00007f5e182a3b00 00007f5e19d157d8 0x00007f5e182a3ac0: 00007f5704912835 00007f5e141b6678 0x00007f5e182a3ad0: 0000000000000000 00007f569104fa20 0x00007f5e182a3ae0: 00007f569104fa00 00007f568817f27d 0x00007f5e182a3af0: 00007f5e1401a890 00007f568817f29f 0x00007f5e182a3b00: 00007f5e182a3bc0 00007f5e19b1f976 0x00007f5e182a3b10: 00007f5e1401eea0 00007f568817f29e 0x00007f5e182a3b20: 00007f568817f20b 00007f5e1a59f178 0x00007f5e182a3b30: 00007f5e141b6610 00007f5700470370 0x00007f5e182a3b40: 00007f5e1401a620 00007f56884f9382 0x00007f5e182a3b50: 00007f569104dc08 0000000000000400 0x00007f5e182a3b60: 00007f5691053c70 00007f568817f26f 0x00007f5e182a3b70: 0000000000000240 00007f568817f400 0x00007f5e182a3b80: 00007f568817f600 00007f568817f27d 0x00007f5e182a3b90: 00007f568817f26f 00007f5e14055bf0 Instructions: (pc=0x00007f5e19fe8e17) 0x00007f5e19fe8df7: 83 f8 05 75 d3 31 db eb d6 48 8b 05 31 de 57 00 0x00007f5e19fe8e07: 8b 56 08 8b 48 08 48 d3 e2 48 03 10 48 8d 7a 10 0x00007f5e19fe8e17: 8b 77 08 83 fe 00 0f 8f db fe ff ff 0f 85 07 03 0x00007f5e19fe8e27: 00 00 0f 1f 80 00 00 00 00 48 8b 07 4c 89 e6 4c Register to memory mapping: RAX=0x00007f5e1a5839a0: <offset 0xe209a0> in /root/jdk1.7.0_65/jre/lib/amd64/server/libjvm.so at 0x00007f5e19763000 RBX=0x00007f5704912834 is pointing into object: 0x00007f5704912820 javax.crypto.Mac - klass: 'javax/crypto/Mac' RCX=0x0000000000000003 is an unknown value RDX=0x00007f568bdff000 is an unknown value RSP=0x00007f5e182a39a0 is an unknown value RBP=0x00007f5e182a3a10 is an unknown value RSI=0x00007f5bb0b26af8 is an unknown value RDI=0x00007f568bdff010 is an unknown value R8 =0x0000000000000001 is an unknown value R9 =0x00007f5e1a59f178: <offset 0xe3c178> in /root/jdk1.7.0_65/jre/lib/amd64/server/libjvm.so at 0x00007f5e19763000 R10=0x00000000e6a82202 is an unknown value R11=0x00007f5e182a3950 is an unknown value R12=0x00007f5bb0b26af8 is an unknown value R13=0x0000000000000000 is an unknown value R14=0x0000000000000000 is an unknown value R15=0x00007f5e141b6680 is an unknown value Stack: [0x00007f5e181a4000,0x00007f5e182a5000], sp=0x00007f5e182a39a0, free space=1022k Native frames: (J=compiled Java code, j=interpreted, Vv=VM code, C=native code) V [libjvm.so+0x885e17] oopDesc* PSPromotionManager::copy_to_survivor_space<false>(oopDesc*)+0x177 V [libjvm.so+0x8855f0] PSPromotionManager::drain_stacks_depth(bool)+0x7c0 V [libjvm.so+0x3bc976] CardTableExtension::scavenge_contents_parallel(ObjectStartArray*, MutableSpace*, HeapWord*, PSPromotionManager*, unsigned int, unsigned int)+0xb56 V [libjvm.so+0x888fbb] OldToYoungRootsTask::do_it(GCTaskManager*, unsigned int)+0x4b V [libjvm.so+0x55704f] GCTaskThread::run()+0x12f V [libjvm.so+0x81f988] java_start(Thread*)+0x108 --------------- P R O C E S S --------------- Java Threads: ( => current thread ) 0x00007f55f8029800 JavaThread "nioEventLoopGroup-3-16" [_thread_in_native, id=15157, stack(0x00007f565cefc000,0x00007f565cffd000)] 0x00007f55f8028000 JavaThread "nioEventLoopGroup-3-15" [_thread_in_native, id=15156, stack(0x00007f56640c6000,0x00007f56641c7000)] 0x00007f55f8026000 JavaThread "nioEventLoopGroup-3-14" [_thread_in_native, id=15155, stack(0x00007f56641c7000,0x00007f56642c8000)] 0x00007f55f8024800 JavaThread "nioEventLoopGroup-3-13" [_thread_blocked, id=15154, stack(0x00007f56642c8000,0x00007f56643c9000)] 0x00007f55f8023000 JavaThread "nioEventLoopGroup-3-12" [_thread_in_native, id=15153, stack(0x00007f56643c9000,0x00007f56644ca000)] 0x00007f55f8021800 JavaThread "nioEventLoopGroup-3-11" [_thread_in_native, id=15152, stack(0x00007f56644ca000,0x00007f56645cb000)] 0x00007f55f8020000 JavaThread "nioEventLoopGroup-3-10" [_thread_in_native, id=15151, stack(0x00007f56645cb000,0x00007f56646cc000)] 0x00007f55f801e800 JavaThread "nioEventLoopGroup-3-9" [_thread_in_native, id=15150, stack(0x00007f56646cc000,0x00007f56647cd000)] 0x00007f55f801c800 JavaThread "nioEventLoopGroup-3-8" [_thread_in_native, id=15149, stack(0x00007f56647cd000,0x00007f56648ce000)] 0x00007f55f801b000 JavaThread "nioEventLoopGroup-3-7" [_thread_in_native, id=15148, stack(0x00007f56648ce000,0x00007f56649cf000)] 0x00007f55f8019000 JavaThread "nioEventLoopGroup-3-6" [_thread_in_native, id=15147, stack(0x00007f56649cf000,0x00007f5664ad0000)] 0x00007f55f8017800 JavaThread "nioEventLoopGroup-3-5" [_thread_in_native, id=15146, stack(0x00007f5664ad0000,0x00007f5664bd1000)] 0x00007f55f8016000 JavaThread "nioEventLoopGroup-3-4" [_thread_in_native, id=15145, stack(0x00007f5664bd1000,0x00007f5664cd2000)] 0x00007f55f8014800 JavaThread "nioEventLoopGroup-3-3" [_thread_in_native, id=15144, stack(0x00007f5664cd2000,0x00007f5664dd3000)] 0x00007f55f8013000 JavaThread "nioEventLoopGroup-3-2" [_thread_in_native, id=15143, stack(0x00007f5664dd3000,0x00007f5664ed4000)] 0x00007f55ac023000 JavaThread "threadDeathWatcher-4-1" daemon [_thread_blocked, id=15142, stack(0x00007f5665ed5000,0x00007f5665fd6000)] 0x00007f55f8011000 JavaThread "nioEventLoopGroup-3-1" [_thread_in_native, id=15141, stack(0x00007f5665fd6000,0x00007f56660d7000)] 0x00007f5e146f8000 JavaThread "nioEventLoopGroup-2-2" [_thread_in_native, id=15140, stack(0x00007f56660d7000,0x00007f56661d8000)] 0x00007f5e146fc000 JavaThread "nioEventLoopGroup-2-1" [_thread_blocked, id=15139, stack(0x00007f5666c65000,0x00007f5666d66000)] 0x00007f5e141e5800 JavaThread "Service Thread" daemon [_thread_blocked, id=15134, stack(0x00007f5666e67000,0x00007f5666f68000)] 0x00007f5e141e3800 JavaThread "C2 CompilerThread1" daemon [_thread_blocked, id=15133, stack(0x00007f5666f68000,0x00007f5667069000)] 0x00007f5e141e0800 JavaThread "C2 CompilerThread0" daemon [_thread_blocked, id=15132, stack(0x00007f5667069000,0x00007f566716a000)] 0x00007f5e141de800 JavaThread "Signal Dispatcher" daemon [_thread_blocked, id=15131, stack(0x00007f566716a000,0x00007f566726b000)] 0x00007f5e141bf800 JavaThread "Finalizer" daemon [_thread_blocked, id=15130, stack(0x00007f56673f4000,0x00007f56674f5000)] 0x00007f5e141bd800 JavaThread "Reference Handler" daemon [_thread_blocked, id=15129, stack(0x00007f56674f5000,0x00007f56675f6000)] 0x00007f5e14009800 JavaThread "main" [_thread_blocked, id=15119, stack(0x00007f5e1b0bc000,0x00007f5e1b1bd000)] Other Threads: 0x00007f5e141b9000 VMThread [stack: 0x00007f56675f6000,0x00007f56676f7000] [id=15128] 0x00007f5e141f0800 WatcherThread [stack: 0x00007f5666d66000,0x00007f5666e67000] [id=15135] =>0x00007f5e14021000 (exited) GCTaskThread [stack: 0x00007f5e181a4000,0x00007f5e182a5000] [id=15121] VM state:at safepoint (normal execution) VM Mutex/Monitor currently owned by a thread: ([mutex/lock_event]) [0x00007f5e14005f20] Threads_lock - owner thread: 0x00007f5e141b9000 [0x00007f5e14006420] Heap_lock - owner thread: 0x00007f55f8024800 Heap PSYoungGen total 8978944K, used 8978943K [0x00007f5b91000000, 0x00007f5e11000000, 0x00007f5e11000000) eden space 7668224K, 100% used [0x00007f5b91000000,0x00007f5d65080000,0x00007f5d65080000) from space 1310720K, 99% used [0x00007f5d71000000,0x00007f5dc0fffee0,0x00007f5dc1000000) to space 1310720K, 99% used [0x00007f5dc1000000,0x00007f5e10ffbf40,0x00007f5e11000000) ParOldGen total 20971520K, used 2082433K [0x00007f5691000000, 0x00007f5b91000000, 0x00007f5b91000000) object space 20971520K, 9% used [0x00007f5691000000,0x00007f57101aa940,0x00007f5b91000000) PSPermGen total 21504K, used 12737K [0x00007f568be00000, 0x00007f568d300000, 0x00007f5691000000) object space 21504K, 59% used [0x00007f568be00000,0x00007f568ca705e0,0x00007f568d300000) Card table byte_map: [0x00007f5688156000,0x00007f568bd80000] byte_map_base: 0x00007f16dccf7000 Polling page: 0x00007f5e1b1d5000 Code Cache [0x00007f5e11000000, 0x00007f5e11520000, 0x00007f5e14000000) total_blobs=2150 nmethods=1831 adapters=271 free_code_cache=44036Kb largest_free_block=45004224 Compilation events (10 events): Event: 4080.570 Thread 0x00007f5e141e3800 1889 p2p_server.AppHandler$2::<init> (10 bytes) Event: 4080.570 Thread 0x00007f5e141e3800 nmethod 1889 0x00007f5e114ea250 code [0x00007f5e114ea3a0, 0x00007f5e114ea448] Event: 4099.029 Thread 0x00007f5e141e0800 1890 io.netty.channel.DefaultChannelId::asShortText (27 bytes) Event: 4099.030 Thread 0x00007f5e141e0800 nmethod 1890 0x00007f5e114e9f50 code [0x00007f5e114ea0a0, 0x00007f5e114ea188] Event: 4175.263 Thread 0x00007f5e141e3800 1891 ! p2p_server.AppHandler::channelRead (454 bytes) Event: 4175.297 Thread 0x00007f5e141e3800 nmethod 1891 0x00007f5e1150e0d0 code [0x00007f5e1150e660, 0x00007f5e115109d0] Event: 4175.309 Thread 0x00007f5e141e0800 1892 p2p_server.AppHandler$3::operationComplete (9 bytes) Event: 4175.310 Thread 0x00007f5e141e0800 nmethod 1892 0x00007f5e114e9ad0 code [0x00007f5e114e9c40, 0x00007f5e114e9d78] Event: 4448.098 Thread 0x00007f5e141e3800 1893 io.netty.buffer.WrappedByteBuf::skipBytes (11 bytes) Event: 4448.099 Thread 0x00007f5e141e3800 nmethod 1893 0x00007f5e114e9550 code [0x00007f5e114e96c0, 0x00007f5e114e9858] GC Heap History (10 events): Event: 545.137 GC heap after Heap after GC invocations=1 (full 0): PSYoungGen total 9175040K, used 931411K [0x00007f5b91000000, 0x00007f5e11000000, 0x00007f5e11000000) eden space 7864320K, 0% used [0x00007f5b91000000,0x00007f5b91000000,0x00007f5d71000000) from space 1310720K, 71% used [0x00007f5d71000000,0x00007f5da9d94ee8,0x00007f5dc1000000) to space 1310720K, 0% used [0x00007f5dc1000000,0x00007f5dc1000000,0x00007f5e11000000) ParOldGen total 20971520K, used 8K [0x00007f5691000000, 0x00007f5b91000000, 0x00007f5b91000000) object space 20971520K, 0% used [0x00007f5691000000,0x00007f5691002000,0x00007f5b91000000) PSPermGen total 21504K, used 12565K [0x00007f568be00000, 0x00007f568d300000, 0x00007f5691000000) object space 21504K, 58% used [0x00007f568be00000,0x00007f568ca45670,0x00007f568d300000) } Event: 1277.444 GC heap before {Heap before GC invocations=2 (full 0): PSYoungGen total 9175040K, used 8795731K [0x00007f5b91000000, 0x00007f5e11000000, 0x00007f5e11000000) eden space 7864320K, 100% used [0x00007f5b91000000,0x00007f5d71000000,0x00007f5d71000000) from space 1310720K, 71% used [0x00007f5d71000000,0x00007f5da9d94ee8,0x00007f5dc1000000) to space 1310720K, 0% used [0x00007f5dc1000000,0x00007f5dc1000000,0x00007f5e11000000) ParOldGen total 20971520K, used 8K [0x00007f5691000000, 0x00007f5b91000000, 0x00007f5b91000000) object space 20971520K, 0% used [0x00007f5691000000,0x00007f5691002000,0x00007f5b91000000) PSPermGen total 21504K, used 12566K [0x00007f568be00000, 0x00007f568d300000, 0x00007f5691000000) object space 21504K, 58% used [0x00007f568be00000,0x00007f568ca45b80,0x00007f568d300000) Event: 1277.830 GC heap after Heap after GC invocations=2 (full 0): PSYoungGen total 9175040K, used 1310716K [0x00007f5b91000000, 0x00007f5e11000000, 0x00007f5e11000000) eden space 7864320K, 0% used [0x00007f5b91000000,0x00007f5b91000000,0x00007f5d71000000) from space 1310720K, 99% used [0x00007f5dc1000000,0x00007f5e10fff2e0,0x00007f5e11000000) to space 1310720K, 0% used [0x00007f5d71000000,0x00007f5d71000000,0x00007f5dc1000000) ParOldGen total 20971520K, used 499466K [0x00007f5691000000, 0x00007f5b91000000, 0x00007f5b91000000) object space 20971520K, 2% used [0x00007f5691000000,0x00007f56af7c2930,0x00007f5b91000000) PSPermGen total 21504K, used 12566K [0x00007f568be00000, 0x00007f568d300000, 0x00007f5691000000) object space 21504K, 58% used [0x00007f568be00000,0x00007f568ca45b80,0x00007f568d300000) } Event: 2038.802 GC heap before {Heap before GC invocations=3 (full 0): PSYoungGen total 9175040K, used 9175036K [0x00007f5b91000000, 0x00007f5e11000000, 0x00007f5e11000000) eden space 7864320K, 100% used [0x00007f5b91000000,0x00007f5d71000000,0x00007f5d71000000) from space 1310720K, 99% used [0x00007f5dc1000000,0x00007f5e10fff2e0,0x00007f5e11000000) to space 1310720K, 0% used [0x00007f5d71000000,0x00007f5d71000000,0x00007f5dc1000000) ParOldGen total 20971520K, used 499466K [0x00007f5691000000, 0x00007f5b91000000, 0x00007f5b91000000) object space 20971520K, 2% used [0x00007f5691000000,0x00007f56af7c2930,0x00007f5b91000000) PSPermGen total 21504K, used 12733K [0x00007f568be00000, 0x00007f568d300000, 0x00007f5691000000) object space 21504K, 59% used [0x00007f568be00000,0x00007f568ca6f708,0x00007f568d300000) Event: 2039.081 GC heap after Heap after GC invocations=3 (full 0): PSYoungGen total 9175040K, used 1310712K [0x00007f5b91000000, 0x00007f5e11000000, 0x00007f5e11000000) eden space 7864320K, 0% used [0x00007f5b91000000,0x00007f5b91000000,0x00007f5d71000000) from space 1310720K, 99% used [0x00007f5d71000000,0x00007f5dc0ffe1c0,0x00007f5dc1000000) to space 1310720K, 0% used [0x00007f5dc1000000,0x00007f5dc1000000,0x00007f5e11000000) ParOldGen total 20971520K, used 639277K [0x00007f5691000000, 0x00007f5b91000000, 0x00007f5b91000000) object space 20971520K, 3% used [0x00007f5691000000,0x00007f56b804b440,0x00007f5b91000000) PSPermGen total 21504K, used 12733K [0x00007f568be00000, 0x00007f568d300000, 0x00007f5691000000) object space 21504K, 59% used [0x00007f568be00000,0x00007f568ca6f708,0x00007f568d300000) } Event: 2846.013 GC heap before {Heap before GC invocations=4 (full 0): PSYoungGen total 9175040K, used 9175032K [0x00007f5b91000000, 0x00007f5e11000000, 0x00007f5e11000000) eden space 7864320K, 100% used [0x00007f5b91000000,0x00007f5d71000000,0x00007f5d71000000) from space 1310720K, 99% used [0x00007f5d71000000,0x00007f5dc0ffe1c0,0x00007f5dc1000000) to space 1310720K, 0% used [0x00007f5dc1000000,0x00007f5dc1000000,0x00007f5e11000000) ParOldGen total 20971520K, used 639277K [0x00007f5691000000, 0x00007f5b91000000, 0x00007f5b91000000) object space 20971520K, 3% used [0x00007f5691000000,0x00007f56b804b440,0x00007f5b91000000) PSPermGen total 21504K, used 12735K [0x00007f568be00000, 0x00007f568d300000, 0x00007f5691000000) object space 21504K, 59% used [0x00007f568be00000,0x00007f568ca6ff20,0x00007f568d300000) Event: 2846.319 GC heap after Heap after GC invocations=4 (full 0): PSYoungGen total 9175040K, used 1310709K [0x00007f5b91000000, 0x00007f5e11000000, 0x00007f5e11000000) eden space 7864320K, 0% used [0x00007f5b91000000,0x00007f5b91000000,0x00007f5d71000000) from space 1310720K, 99% used [0x00007f5dc1000000,0x00007f5e10ffd470,0x00007f5e11000000) to space 1310720K, 0% used [0x00007f5d71000000,0x00007f5d71000000,0x00007f5dc1000000) ParOldGen total 20971520K, used 818909K [0x00007f5691000000, 0x00007f5b91000000, 0x00007f5b91000000) object space 20971520K, 3% used [0x00007f5691000000,0x00007f56c2fb76b0,0x00007f5b91000000) PSPermGen total 21504K, used 12735K [0x00007f568be00000, 0x00007f568d300000, 0x00007f5691000000) object space 21504K, 59% used [0x00007f568be00000,0x00007f568ca6ff20,0x00007f568d300000) } Event: 3666.166 GC heap before {Heap before GC invocations=5 (full 0): PSYoungGen total 9175040K, used 9175029K [0x00007f5b91000000, 0x00007f5e11000000, 0x00007f5e11000000) eden space 7864320K, 100% used [0x00007f5b91000000,0x00007f5d71000000,0x00007f5d71000000) from space 1310720K, 99% used [0x00007f5dc1000000,0x00007f5e10ffd470,0x00007f5e11000000) to space 1310720K, 0% used [0x00007f5d71000000,0x00007f5d71000000,0x00007f5dc1000000) ParOldGen total 20971520K, used 818909K [0x00007f5691000000, 0x00007f5b91000000, 0x00007f5b91000000) object space 20971520K, 3% used [0x00007f5691000000,0x00007f56c2fb76b0,0x00007f5b91000000) PSPermGen total 21504K, used 12737K [0x00007f568be00000, 0x00007f568d300000, 0x00007f5691000000) object space 21504K, 59% used [0x00007f568be00000,0x00007f568ca705e0,0x00007f568d300000) Event: 3666.670 GC heap after Heap after GC invocations=5 (full 0): PSYoungGen total 8978944K, used 1310719K [0x00007f5b91000000, 0x00007f5e11000000, 0x00007f5e11000000) eden space 7668224K, 0% used [0x00007f5b91000000,0x00007f5b91000000,0x00007f5d65080000) from space 1310720K, 99% used [0x00007f5d71000000,0x00007f5dc0fffee0,0x00007f5dc1000000) to space 1310720K, 0% used [0x00007f5dc1000000,0x00007f5dc1000000,0x00007f5e11000000) ParOldGen total 20971520K, used 1823168K [0x00007f5691000000, 0x00007f5b91000000, 0x00007f5b91000000) object space 20971520K, 8% used [0x00007f5691000000,0x00007f5700470370,0x00007f5b91000000) PSPermGen total 21504K, used 12737K [0x00007f568be00000, 0x00007f568d300000, 0x00007f5691000000) object space 21504K, 59% used [0x00007f568be00000,0x00007f568ca705e0,0x00007f568d300000) } Event: 4455.954 GC heap before {Heap before GC invocations=6 (full 0): PSYoungGen total 8978944K, used 8978943K [0x00007f5b91000000, 0x00007f5e11000000, 0x00007f5e11000000) eden space 7668224K, 100% used [0x00007f5b91000000,0x00007f5d65080000,0x00007f5d65080000) from space 1310720K, 99% used [0x00007f5d71000000,0x00007f5dc0fffee0,0x00007f5dc1000000) to space 1310720K, 0% used [0x00007f5dc1000000,0x00007f5dc1000000,0x00007f5e11000000) ParOldGen total 20971520K, used 1823168K [0x00007f5691000000, 0x00007f5b91000000, 0x00007f5b91000000) object space 20971520K, 8% used [0x00007f5691000000,0x00007f5700470370,0x00007f5b91000000) PSPermGen total 21504K, used 12737K [0x00007f568be00000, 0x00007f568d300000, 0x00007f5691000000) object space 21504K, 59% used [0x00007f568be00000,0x00007f568ca705e0,0x00007f568d300000) Deoptimization events (10 events): Event: 2833.496 Thread 0x00007f55f8024800 Uncommon trap: reason=bimorphic action=maybe_recompile pc=0x00007f5e110dc880 method=java.util.HashMap.put(Ljava/lang/Object;Ljava/lang/Object;)Ljava/lang/Object; @ 83 Event: 4056.850 Thread 0x00007f55f8029800 Uncommon trap: reason=class_check action=maybe_recompile pc=0x00007f5e111d85ec method=p2p_server.AppHandler.channelRead(Lio/netty/channel/ChannelHandlerContext;Ljava/lang/Object;)V @ 379 Event: 4056.850 Thread 0x00007f55f8029800 Uncommon trap: reason=class_check action=maybe_recompile pc=0x00007f5e114ef314 method=p2p_server.AppHandler$3.operationComplete(Lio/netty/channel/ChannelFuture;)V @ 4 Event: 4058.287 Thread 0x00007f55f8029800 Uncommon trap: reason=class_check action=maybe_recompile pc=0x00007f5e111d85ec method=p2p_server.AppHandler.channelRead(Lio/netty/channel/ChannelHandlerContext;Ljava/lang/Object;)V @ 379 Event: 4058.287 Thread 0x00007f55f8029800 Uncommon trap: reason=class_check action=maybe_recompile pc=0x00007f5e114ef314 method=p2p_server.AppHandler$3.operationComplete(Lio/netty/channel/ChannelFuture;)V @ 4 Event: 4059.930 Thread 0x00007f55f8029800 Uncommon trap: reason=class_check action=maybe_recompile pc=0x00007f5e111d85ec method=p2p_server.AppHandler.channelRead(Lio/netty/channel/ChannelHandlerContext;Ljava/lang/Object;)V @ 379 Event: 4059.930 Thread 0x00007f55f8029800 Uncommon trap: reason=class_check action=maybe_recompile pc=0x00007f5e114ef314 method=p2p_server.AppHandler$3.operationComplete(Lio/netty/channel/ChannelFuture;)V @ 4 Event: 4061.372 Thread 0x00007f55f8029800 Uncommon trap: reason=class_check action=maybe_recompile pc=0x00007f5e111d85ec method=p2p_server.AppHandler.channelRead(Lio/netty/channel/ChannelHandlerContext;Ljava/lang/Object;)V @ 379 Event: 4061.372 Thread 0x00007f55f8029800 Uncommon trap: reason=class_check action=maybe_recompile pc=0x00007f5e114ef314 method=p2p_server.AppHandler$3.operationComplete(Lio/netty/channel/ChannelFuture;)V @ 4 Event: 4062.815 Thread 0x00007f55f8029800 Uncommon trap: reason=class_check action=maybe_recompile pc=0x00007f5e113fd04c method=p2p_server.AppHandler$3.operationComplete(Lio/netty/channel/ChannelFuture;)V @ 4 Internal exceptions (10 events): Event: 1358.045 Thread 0x00007f55f8011000 Threw 0x00007f5bc4627ff8 at /HUDSON/workspace/7u-2-build-linux-amd64/jdk7u65/1224/hotspot/src/share/vm/prims/jvm.cpp:1244 Event: 1358.046 Thread 0x00007f55f8011000 Threw 0x00007f5bc462a320 at /HUDSON/workspace/7u-2-build-linux-amd64/jdk7u65/1224/hotspot/src/share/vm/prims/jvm.cpp:1244 Event: 1358.046 Thread 0x00007f55f8013000 Implicit null exception at 0x00007f5e11249334 to 0x00007f5e112493b2 Event: 1358.052 Thread 0x00007f55f8011000 Threw 0x00007f5bc462d5f0 at /HUDSON/workspace/7u-2-build-linux-amd64/jdk7u65/1224/hotspot/src/share/vm/prims/jvm.cpp:1244 Event: 1358.056 Thread 0x00007f55f8011000 Threw 0x00007f5bc4630a80 at /HUDSON/workspace/7u-2-build-linux-amd64/jdk7u65/1224/hotspot/src/share/vm/prims/jvm.cpp:1244 Event: 1358.057 Thread 0x00007f55f8011000 Threw 0x00007f5bc4634160 at /HUDSON/workspace/7u-2-build-linux-amd64/jdk7u65/1224/hotspot/src/share/vm/prims/jvm.cpp:1244 Event: 1358.058 Thread 0x00007f55f8011000 Threw 0x00007f5bc4637b78 at /HUDSON/workspace/7u-2-build-linux-amd64/jdk7u65/1224/hotspot/src/share/vm/prims/jvm.cpp:1244 Event: 1358.058 Thread 0x00007f55f8011000 Threw 0x00007f5bc463a2b8 at /HUDSON/workspace/7u-2-build-linux-amd64/jdk7u65/1224/hotspot/src/share/vm/prims/jvm.cpp:1244 Event: 1358.059 Thread 0x00007f55f8011000 Threw 0x00007f5bc4640638 at /HUDSON/workspace/7u-2-build-linux-amd64/jdk7u65/1224/hotspot/src/share/vm/prims/jvm.cpp:1244 Event: 1358.060 Thread 0x00007f55f8011000 Implicit null exception at 0x00007f5e112ab4fe to 0x00007f5e112ab805 Events (10 events): Event: 4061.372 Thread 0x00007f55f8029800 Uncommon trap: trap_request=0xffffffde fr.pc=0x00007f5e111d85ec Event: 4061.372 Thread 0x00007f55f8029800 DEOPT PACKING pc=0x00007f5e111d85ec sp=0x00007f565cffb120 Event: 4061.372 Thread 0x00007f55f8029800 DEOPT UNPACKING pc=0x00007f5e11039445 sp=0x00007f565cffb0c0 mode 2 Event: 4061.372 Thread 0x00007f55f8029800 Uncommon trap: trap_request=0xffffffde fr.pc=0x00007f5e114ef314 Event: 4061.372 Thread 0x00007f55f8029800 DEOPT PACKING pc=0x00007f5e114ef314 sp=0x00007f565cffaff0 Event: 4061.372 Thread 0x00007f55f8029800 DEOPT UNPACKING pc=0x00007f5e11039445 sp=0x00007f565cffaf48 mode 2 Event: 4062.815 Thread 0x00007f55f8029800 Uncommon trap: trap_request=0xffffffde fr.pc=0x00007f5e113fd04c Event: 4062.815 Thread 0x00007f55f8029800 DEOPT PACKING pc=0x00007f5e113fd04c sp=0x00007f565cffaf70 Event: 4062.815 Thread 0x00007f55f8029800 DEOPT UNPACKING pc=0x00007f5e11039445 sp=0x00007f565cffaf38 mode 2 Event: 4455.954 Executing VM operation: ParallelGCFailedAllocation Dynamic libraries: 00400000-00401000 r-xp 00000000 08:01 354231 /root/jdk1.7.0_65/bin/java 00600000-00601000 rw-p 00000000 08:01 354231 /root/jdk1.7.0_65/bin/java 0077f000-007a0000 rw-p 00000000 00:00 0 [heap] 7f5566fff000-7f5568000000 rw-p 00000000 00:00 0 7f5568000000-7f5568021000 rw-p 00000000 00:00 0 7f5568021000-7f556c000000 ---p 00000000 00:00 0 7f556c000000-7f556c021000 rw-p 00000000 00:00 0 7f556c021000-7f5570000000 ---p 00000000 00:00 0 7f5570000000-7f5570021000 rw-p 00000000 00:00 0 7f5570021000-7f5574000000 ---p 00000000 00:00 0 7f5574ffd000-7f5578000000 rw-p 00000000 00:00 0 7f5578000000-7f5578021000 rw-p 00000000 00:00 0 7f5578021000-7f557c000000 ---p 00000000 00:00 0 7f557c000000-7f557c021000 rw-p 00000000 00:00 0 7f557c021000-7f5580000000 ---p 00000000 00:00 0 7f5580000000-7f5580021000 rw-p 00000000 00:00 0 7f5580021000-7f5584000000 ---p 00000000 00:00 0 7f5584000000-7f558424d000 rw-p 00000000 00:00 0 7f558424d000-7f5588000000 ---p 00000000 00:00 0 7f5588ffd000-7f558c000000 rw-p 00000000 00:00 0 7f558c000000-7f558c021000 rw-p 00000000 00:00 0 7f558c021000-7f5590000000 ---p 00000000 00:00 0 7f5590000000-7f5590021000 rw-p 00000000 00:00 0 7f5590021000-7f5594000000 ---p 00000000 00:00 0 7f5594000000-7f559424d000 rw-p 00000000 00:00 0 7f559424d000-7f5598000000 ---p 00000000 00:00 0 7f5598000000-7f5598021000 rw-p 00000000 00:00 0 7f5598021000-7f559c000000 ---p 00000000 00:00 0 7f559c000000-7f559c021000 rw-p 00000000 00:00 0 7f559c021000-7f55a0000000 ---p 00000000 00:00 0 7f55a0000000-7f55a0021000 rw-p 00000000 00:00 0 7f55a0021000-7f55a4000000 ---p 00000000 00:00 0 7f55a4000000-7f55a4021000 rw-p 00000000 00:00 0 7f55a4021000-7f55a8000000 ---p 00000000 00:00 0 7f55a8000000-7f55a8021000 rw-p 00000000 00:00 0 7f55a8021000-7f55ac000000 ---p 00000000 00:00 0 7f55ac000000-7f55ac088000 rw-p 00000000 00:00 0 7f55ac088000-7f55b0000000 ---p 00000000 00:00 0 7f55b0000000-7f55b0021000 rw-p 00000000 00:00 0 7f55b0021000-7f55b4000000 ---p 00000000 00:00 0 7f55b4000000-7f55b4021000 rw-p 00000000 00:00 0 7f55b4021000-7f55b8000000 ---p 00000000 00:00 0 7f55b8c71000-7f55f4000000 rw-p 00000000 00:00 0 7f55f4000000-7f55f4021000 rw-p 00000000 00:00 0 7f55f4021000-7f55f8000000 ---p 00000000 00:00 0 7f55f8000000-7f55f802c000 rw-p 00000000 00:00 0 7f55f802c000-7f55fc000000 ---p 00000000 00:00 0 7f55fc000000-7f55fd3dd000 rw-p 00000000 00:00 0 7f55fd3dd000-7f5600000000 ---p 00000000 00:00 0 7f5600000000-7f5600021000 rw-p 00000000 00:00 0 7f5600021000-7f5604000000 ---p 00000000 00:00 0 7f5604000000-7f5604021000 rw-p 00000000 00:00 0 7f5604021000-7f5608000000 ---p 00000000 00:00 0 7f5608000000-7f5609705000 rw-p 00000000 00:00 0 7f5609705000-7f560c000000 ---p 00000000 00:00 0 7f560c000000-7f560c021000 rw-p 00000000 00:00 0 7f560c021000-7f5610000000 ---p 00000000 00:00 0 7f5610000000-7f5610021000 rw-p 00000000 00:00 0 7f5610021000-7f5614000000 ---p 00000000 00:00 0 7f5614000000-7f5614041000 rw-p 00000000 00:00 0 7f5614041000-7f5618000000 ---p 00000000 00:00 0 7f5618d6d000-7f5658000000 rw-p 00000000 00:00 0 7f5658000000-7f565824d000 rw-p 00000000 00:00 0 7f565824d000-7f565c000000 ---p 00000000 00:00 0 7f565cefc000-7f565ceff000 ---p 00000000 00:00 0 7f565ceff000-7f5660000000 rw-p 00000000 00:00 0 [stack:15157] 7f5660000000-7f5660021000 rw-p 00000000 00:00 0 7f5660021000-7f5664000000 ---p 00000000 00:00 0 7f56640c6000-7f56640c9000 ---p 00000000 00:00 0 7f56640c9000-7f56641c7000 rw-p 00000000 00:00 0 [stack:15156] 7f56641c7000-7f56641ca000 ---p 00000000 00:00 0 7f56641ca000-7f56642c8000 rw-p 00000000 00:00 0 [stack:15155] 7f56642c8000-7f56642cb000 ---p 00000000 00:00 0 7f56642cb000-7f56643c9000 rw-p 00000000 00:00 0 [stack:15154] 7f56643c9000-7f56643cc000 ---p 00000000 00:00 0 7f56643cc000-7f56644ca000 rw-p 00000000 00:00 0 [stack:15153] 7f56644ca000-7f56644cd000 ---p 00000000 00:00 0 7f56644cd000-7f56645cb000 rw-p 00000000 00:00 0 [stack:15152] 7f56645cb000-7f56645ce000 ---p 00000000 00:00 0 7f56645ce000-7f56646cc000 rw-p 00000000 00:00 0 [stack:15151] 7f56646cc000-7f56646cf000 ---p 00000000 00:00 0 7f56646cf000-7f56647cd000 rw-p 00000000 00:00 0 [stack:15150] 7f56647cd000-7f56647d0000 ---p 00000000 00:00 0 7f56647d0000-7f56648ce000 rw-p 00000000 00:00 0 [stack:15149] 7f56648ce000-7f56648d1000 ---p 00000000 00:00 0 7f56648d1000-7f56649cf000 rw-p 00000000 00:00 0 [stack:15148] 7f56649cf000-7f56649d2000 ---p 00000000 00:00 0 7f56649d2000-7f5664ad0000 rw-p 00000000 00:00 0 [stack:15147] 7f5664ad0000-7f5664ad3000 ---p 00000000 00:00 0 7f5664ad3000-7f5664bd1000 rw-p 00000000 00:00 0 [stack:15146] 7f5664bd1000-7f5664bd4000 ---p 00000000 00:00 0 7f5664bd4000-7f5664cd2000 rw-p 00000000 00:00 0 [stack:15145] 7f5664cd2000-7f5664cd5000 ---p 00000000 00:00 0 7f5664cd5000-7f5664dd3000 rw-p 00000000 00:00 0 [stack:15144] 7f5664dd3000-7f5664dd6000 ---p 00000000 00:00 0 7f5664dd6000-7f5665ed5000 rw-p 00000000 00:00 0 [stack:15143] 7f5665ed5000-7f5665ed8000 ---p 00000000 00:00 0 7f5665ed8000-7f5665fd6000 rw-p 00000000 00:00 0 [stack:15142] 7f5665fd6000-7f5665fd9000 ---p 00000000 00:00 0 7f5665fd9000-7f56660d7000 rw-p 00000000 00:00 0 [stack:15141] 7f56660d7000-7f56660da000 ---p 00000000 00:00 0 7f56660da000-7f56661d8000 rw-p 00000000 00:00 0 [stack:15140] 7f56661d8000-7f56661e0000 r-xp 00000000 08:01 354452 /root/jdk1.7.0_65/jre/lib/amd64/libmanagement.so 7f56661e0000-7f56663df000 ---p 00008000 08:01 354452 /root/jdk1.7.0_65/jre/lib/amd64/libmanagement.so 7f56663df000-7f56663e0000 rw-p 00007000 08:01 354452 /root/jdk1.7.0_65/jre/lib/amd64/libmanagement.so 7f56663e0000-7f56663f6000 r-xp 00000000 08:01 2098879 /lib/x86_64-linux-gnu/libgcc_s.so.1 7f56663f6000-7f56665f5000 ---p 00016000 08:01 2098879 /lib/x86_64-linux-gnu/libgcc_s.so.1 7f56665f5000-7f56665f6000 rw-p 00015000 08:01 2098879 /lib/x86_64-linux-gnu/libgcc_s.so.1 7f56665f6000-7f5666628000 r-xp 00000000 08:01 354462 /root/jdk1.7.0_65/jre/lib/amd64/libsunec.so 7f5666628000-7f5666827000 ---p 00032000 08:01 354462 /root/jdk1.7.0_65/jre/lib/amd64/libsunec.so 7f5666827000-7f566682d000 rw-p 00031000 08:01 354462 /root/jdk1.7.0_65/jre/lib/amd64/libsunec.so 7f566682d000-7f566683d000 rw-p 00000000 00:00 0 7f566683d000-7f566684d000 r-xp 00000000 08:01 354455 /root/jdk1.7.0_65/jre/lib/amd64/libnio.so 7f566684d000-7f5666a4d000 ---p 00010000 08:01 354455 /root/jdk1.7.0_65/jre/lib/amd64/libnio.so 7f5666a4d000-7f5666a4e000 rw-p 00010000 08:01 354455 /root/jdk1.7.0_65/jre/lib/amd64/libnio.so 7f5666a4e000-7f5666a64000 r-xp 00000000 08:01 354454 /root/jdk1.7.0_65/jre/lib/amd64/libnet.so 7f5666a64000-7f5666c64000 ---p 00016000 08:01 354454 /root/jdk1.7.0_65/jre/lib/amd64/libnet.so 7f5666c64000-7f5666c65000 rw-p 00016000 08:01 354454 /root/jdk1.7.0_65/jre/lib/amd64/libnet.so 7f5666c65000-7f5666c68000 ---p 00000000 00:00 0 7f5666c68000-7f5666d66000 rw-p 00000000 00:00 0 [stack:15139] 7f5666d66000-7f5666d67000 ---p 00000000 00:00 0 7f5666d67000-7f5666e67000 rw-p 00000000 00:00 0 [stack:15135] 7f5666e67000-7f5666e6a000 ---p 00000000 00:00 0 7f5666e6a000-7f5666f68000 rw-p 00000000 00:00 0 [stack:15134] 7f5666f68000-7f5666f6b000 ---p 00000000 00:00 0 7f5666f6b000-7f5667069000 rw-p 00000000 00:00 0 [stack:15133] 7f5667069000-7f566706c000 ---p 00000000 00:00 0 7f566706c000-7f566716a000 rw-p 00000000 00:00 0 [stack:15132] 7f566716a000-7f566716d000 ---p 00000000 00:00 0 7f566716d000-7f566726b000 rw-p 00000000 00:00 0 [stack:15131] 7f566726b000-7f56673f4000 r--p 00000000 08:01 526048 /usr/lib/locale/locale-archive 7f56673f4000-7f56673f7000 ---p 00000000 00:00 0 7f56673f7000-7f56674f5000 rw-p 00000000 00:00 0 [stack:15130] 7f56674f5000-7f56674f8000 ---p 00000000 00:00 0 7f56674f8000-7f56675f6000 rw-p 00000000 00:00 0 [stack:15129] 7f56675f6000-7f56675f7000 ---p 00000000 00:00 0 7f56675f7000-7f5668000000 rw-p 00000000 00:00 0 [stack:15128] 7f5668000000-7f5668021000 rw-p 00000000 00:00 0 7f5668021000-7f566c000000 ---p 00000000 00:00 0 7f566c0d4000-7f5670000000 rw-p 00000000 00:00 0 7f5670000000-7f5670021000 rw-p 00000000 00:00 0 7f5670021000-7f5674000000 ---p 00000000 00:00 0 7f5674000000-7f5674021000 rw-p 00000000 00:00 0 7f5674021000-7f5678000000 ---p 00000000 00:00 0 7f5678000000-7f5678021000 rw-p 00000000 00:00 0 7f5678021000-7f567c000000 ---p 00000000 00:00 0 7f567c000000-7f567c021000 rw-p 00000000 00:00 0 7f567c021000-7f5680000000 ---p 00000000 00:00 0 7f5680000000-7f5680021000 rw-p 00000000 00:00 0 7f5680021000-7f5684000000 ---p 00000000 00:00 0 7f5684009000-7f5684031000 r--s 00255000 08:01 356231 /root/netty_srv_dir/lib/netty-all-4.1.0.CR3-sources.jar 7f5684031000-7f5684068000 r--s 002d9000 08:01 356230 /root/netty_srv_dir/lib/netty-all-4.1.0.CR3.jar 7f5684068000-7f5684fc2000 rw-p 00000000 00:00 0 7f5684fc2000-7f5685181000 r--s 039e5000 08:01 354414 /root/jdk1.7.0_65/jre/lib/rt.jar 7f5685181000-7f5685451000 rw-p 00000000 00:00 0 7f5685451000-7f5685452000 ---p 00000000 00:00 0 7f5685452000-7f5685552000 rw-p 00000000 00:00 0 [stack:15127] 7f5685552000-7f5685553000 ---p 00000000 00:00 0 7f5685553000-7f5685653000 rw-p 00000000 00:00 0 [stack:15126] 7f5685653000-7f5685654000 ---p 00000000 00:00 0 7f5685654000-7f5685754000 rw-p 00000000 00:00 0 [stack:15125] 7f5685754000-7f5685755000 ---p 00000000 00:00 0 7f5685755000-7f5685855000 rw-p 00000000 00:00 0 [stack:15124] 7f5685855000-7f5685856000 ---p 00000000 00:00 0 7f5685856000-7f5688161000 rw-p 00000000 00:00 0 [stack:15123] 7f5688161000-7f568817f000 rw-p 00000000 00:00 0 7f568817f000-7f568bd80000 rw-p 00000000 00:00 0 7f568bd80000-7f568be00000 ---p 00000000 00:00 0 7f568be00000-7f568d300000 rw-p 00000000 00:00 0 7f568d300000-7f5691000000 rw-p 00000000 00:00 0 7f5691000000-7f5e11000000 rw-p 00000000 00:00 0 7f5e11000000-7f5e11520000 rwxp 00000000 00:00 0 7f5e11520000-7f5e14729000 rw-p 00000000 00:00 0 7f5e14729000-7f5e18000000 ---p 00000000 00:00 0 7f5e1801c000-7f5e18048000 r--s 00208000 08:01 356181 /root/netty_srv_dir/lib/guava-19.0.jar 7f5e18048000-7f5e180a3000 rw-p 00000000 00:00 0 7f5e180a3000-7f5e180a4000 ---p 00000000 00:00 0 7f5e180a4000-7f5e181a4000 rw-p 00000000 00:00 0 [stack:15122] 7f5e181a4000-7f5e181a5000 ---p 00000000 00:00 0 7f5e181a5000-7f5e182a5000 rw-p 00000000 00:00 0 [stack:15121] 7f5e182a5000-7f5e182a6000 ---p 00000000 00:00 0 7f5e182a6000-7f5e183b1000 rw-p 00000000 00:00 0 [stack:15120] 7f5e183b1000-7f5e183cf000 rw-p 00000000 00:00 0 7f5e183cf000-7f5e183e9000 r-xp 00000000 08:01 354466 /root/jdk1.7.0_65/jre/lib/amd64/libzip.so 7f5e183e9000-7f5e185e9000 ---p 0001a000 08:01 354466 /root/jdk1.7.0_65/jre/lib/amd64/libzip.so 7f5e185e9000-7f5e185ea000 rw-p 0001a000 08:01 354466 /root/jdk1.7.0_65/jre/lib/amd64/libzip.so 7f5e185ea000-7f5e185f5000 r-xp 00000000 08:01 2101767 /lib/x86_64-linux-gnu/libnss_files-2.19.so 7f5e185f5000-7f5e187f4000 ---p 0000b000 08:01 2101767 /lib/x86_64-linux-gnu/libnss_files-2.19.so 7f5e187f4000-7f5e187f5000 r--p 0000a000 08:01 2101767 /lib/x86_64-linux-gnu/libnss_files-2.19.so 7f5e187f5000-7f5e187f6000 rw-p 0000b000 08:01 2101767 /lib/x86_64-linux-gnu/libnss_files-2.19.so 7f5e187f6000-7f5e18800000 r-xp 00000000 08:01 2101769 /lib/x86_64-linux-gnu/libnss_nis-2.19.so 7f5e18800000-7f5e189ff000 ---p 0000a000 08:01 2101769 /lib/x86_64-linux-gnu/libnss_nis-2.19.so 7f5e189ff000-7f5e18a00000 r--p 00009000 08:01 2101769 /lib/x86_64-linux-gnu/libnss_nis-2.19.so 7f5e18a00000-7f5e18a01000 rw-p 0000a000 08:01 2101769 /lib/x86_64-linux-gnu/libnss_nis-2.19.so 7f5e18a01000-7f5e18a16000 r-xp 00000000 08:01 2101763 /lib/x86_64-linux-gnu/libnsl-2.19.so 7f5e18a16000-7f5e18c15000 ---p 00015000 08:01 2101763 /lib/x86_64-linux-gnu/libnsl-2.19.so 7f5e18c15000-7f5e18c16000 r--p 00014000 08:01 2101763 /lib/x86_64-linux-gnu/libnsl-2.19.so 7f5e18c16000-7f5e18c17000 rw-p 00015000 08:01 2101763 /lib/x86_64-linux-gnu/libnsl-2.19.so 7f5e18c17000-7f5e18c19000 rw-p 00000000 00:00 0 7f5e18c19000-7f5e18c20000 r-xp 00000000 08:01 2101765 /lib/x86_64-linux-gnu/libnss_compat-2.19.so 7f5e18c20000-7f5e18e1f000 ---p 00007000 08:01 2101765 /lib/x86_64-linux-gnu/libnss_compat-2.19.so 7f5e18e1f000-7f5e18e20000 r--p 00006000 08:01 2101765 /lib/x86_64-linux-gnu/libnss_compat-2.19.so 7f5e18e20000-7f5e18e21000 rw-p 00007000 08:01 2101765 /lib/x86_64-linux-gnu/libnss_compat-2.19.so 7f5e18e21000-7f5e18e4a000 r-xp 00000000 08:01 354436 /root/jdk1.7.0_65/jre/lib/amd64/libjava.so 7f5e18e4a000-7f5e1904a000 ---p 00029000 08:01 354436 /root/jdk1.7.0_65/jre/lib/amd64/libjava.so 7f5e1904a000-7f5e1904c000 rw-p 00029000 08:01 354436 /root/jdk1.7.0_65/jre/lib/amd64/libjava.so 7f5e1904c000-7f5e19059000 r-xp 00000000 08:01 354465 /root/jdk1.7.0_65/jre/lib/amd64/libverify.so 7f5e19059000-7f5e19258000 ---p 0000d000 08:01 354465 /root/jdk1.7.0_65/jre/lib/amd64/libverify.so 7f5e19258000-7f5e1925a000 rw-p 0000c000 08:01 354465 /root/jdk1.7.0_65/jre/lib/amd64/libverify.so 7f5e1925a000-7f5e19261000 r-xp 00000000 08:01 2101780 /lib/x86_64-linux-gnu/librt-2.19.so 7f5e19261000-7f5e19460000 ---p 00007000 08:01 2101780 /lib/x86_64-linux-gnu/librt-2.19.so 7f5e19460000-7f5e19461000 r--p 00006000 08:01 2101780 /lib/x86_64-linux-gnu/librt-2.19.so 7f5e19461000-7f5e19462000 rw-p 00007000 08:01 2101780 /lib/x86_64-linux-gnu/librt-2.19.so 7f5e19462000-7f5e19562000 r-xp 00000000 08:01 2101760 /lib/x86_64-linux-gnu/libm-2.19.so 7f5e19562000-7f5e19761000 ---p 00100000 08:01 2101760 /lib/x86_64-linux-gnu/libm-2.19.so 7f5e19761000-7f5e19762000 r--p 000ff000 08:01 2101760 /lib/x86_64-linux-gnu/libm-2.19.so 7f5e19762000-7f5e19763000 rw-p 00100000 08:01 2101760 /lib/x86_64-linux-gnu/libm-2.19.so 7f5e19763000-7f5e1a2d3000 r-xp 00000000 08:01 354470 /root/jdk1.7.0_65/jre/lib/amd64/server/libjvm.so 7f5e1a2d3000-7f5e1a4d2000 ---p 00b70000 08:01 354470 /root/jdk1.7.0_65/jre/lib/amd64/server/libjvm.so 7f5e1a4d2000-7f5e1a597000 rw-p 00b6f000 08:01 354470 /root/jdk1.7.0_65/jre/lib/amd64/server/libjvm.so 7f5e1a597000-7f5e1a5d7000 rw-p 00000000 00:00 0 7f5e1a5d7000-7f5e1a776000 r-xp 00000000 08:01 2101754 /lib/x86_64-linux-gnu/libc-2.19.so 7f5e1a776000-7f5e1a976000 ---p 0019f000 08:01 2101754 /lib/x86_64-linux-gnu/libc-2.19.so 7f5e1a976000-7f5e1a97a000 r--p 0019f000 08:01 2101754 /lib/x86_64-linux-gnu/libc-2.19.so 7f5e1a97a000-7f5e1a97c000 rw-p 001a3000 08:01 2101754 /lib/x86_64-linux-gnu/libc-2.19.so 7f5e1a97c000-7f5e1a980000 rw-p 00000000 00:00 0 7f5e1a980000-7f5e1a983000 r-xp 00000000 08:01 2101759 /lib/x86_64-linux-gnu/libdl-2.19.so 7f5e1a983000-7f5e1ab82000 ---p 00003000 08:01 2101759 /lib/x86_64-linux-gnu/libdl-2.19.so 7f5e1ab82000-7f5e1ab83000 r--p 00002000 08:01 2101759 /lib/x86_64-linux-gnu/libdl-2.19.so 7f5e1ab83000-7f5e1ab84000 rw-p 00003000 08:01 2101759 /lib/x86_64-linux-gnu/libdl-2.19.so 7f5e1ab84000-7f5e1ab9a000 r-xp 00000000 08:01 355067 /root/jdk1.7.0_65/lib/amd64/jli/libjli.so 7f5e1ab9a000-7f5e1ad9a000 ---p 00016000 08:01 355067 /root/jdk1.7.0_65/lib/amd64/jli/libjli.so 7f5e1ad9a000-7f5e1ad9b000 rw-p 00016000 08:01 355067 /root/jdk1.7.0_65/lib/amd64/jli/libjli.so 7f5e1ad9b000-7f5e1adb3000 r-xp 00000000 08:01 2101747 /lib/x86_64-linux-gnu/libpthread-2.19.so 7f5e1adb3000-7f5e1afb2000 ---p 00018000 08:01 2101747 /lib/x86_64-linux-gnu/libpthread-2.19.so 7f5e1afb2000-7f5e1afb3000 r--p 00017000 08:01 2101747 /lib/x86_64-linux-gnu/libpthread-2.19.so 7f5e1afb3000-7f5e1afb4000 rw-p 00018000 08:01 2101747 /lib/x86_64-linux-gnu/libpthread-2.19.so 7f5e1afb4000-7f5e1afb8000 rw-p 00000000 00:00 0 7f5e1afb8000-7f5e1afd8000 r-xp 00000000 08:01 2101748 /lib/x86_64-linux-gnu/ld-2.19.so 7f5e1afe9000-7f5e1afea000 rw-p 00000000 00:00 0 7f5e1afea000-7f5e1afee000 r--s 0002d000 08:01 354553 /root/jdk1.7.0_65/jre/lib/ext/sunjce_provider.jar 7f5e1afee000-7f5e1afef000 r--s 00003000 08:01 354552 /root/jdk1.7.0_65/jre/lib/ext/sunec.jar 7f5e1afef000-7f5e1aff3000 r--s 0008a000 08:01 354404 /root/jdk1.7.0_65/jre/lib/jsse.jar 7f5e1aff3000-7f5e1aff7000 r--s 0003d000 08:01 356179 /root/netty_srv_dir/lib/gson-2.6.1-javadoc.jar 7f5e1aff7000-7f5e1affc000 r--s 00034000 08:01 356180 /root/netty_srv_dir/lib/gson-2.6.1.jar 7f5e1affc000-7f5e1b011000 rw-p 00000000 00:00 0 7f5e1b011000-7f5e1b0bc000 rw-p 00000000 00:00 0 7f5e1b0bc000-7f5e1b0bf000 ---p 00000000 00:00 0 7f5e1b0bf000-7f5e1b1c1000 rw-p 00000000 00:00 0 [stack:15119] 7f5e1b1c1000-7f5e1b1c3000 r--s 00019000 08:01 354400 /root/jdk1.7.0_65/jre/lib/jce.jar 7f5e1b1c3000-7f5e1b1ca000 r--s 0003f000 08:01 356176 /root/netty_srv_dir/lib/commons-codec-1.10.jar 7f5e1b1ca000-7f5e1b1cc000 r--s 0000a000 08:01 354208 /root/netty_srv_dir/p2p_server.jar 7f5e1b1cc000-7f5e1b1d4000 rw-s 00000000 08:01 401099 /tmp/hsperfdata_root/15117 7f5e1b1d4000-7f5e1b1d5000 rw-p 00000000 00:00 0 7f5e1b1d5000-7f5e1b1d6000 ---p 00000000 00:00 0 7f5e1b1d6000-7f5e1b1d8000 rw-p 00000000 00:00 0 7f5e1b1d8000-7f5e1b1d9000 r--p 00020000 08:01 2101748 /lib/x86_64-linux-gnu/ld-2.19.so 7f5e1b1d9000-7f5e1b1da000 rw-p 00021000 08:01 2101748 /lib/x86_64-linux-gnu/ld-2.19.so 7f5e1b1da000-7f5e1b1db000 rw-p 00000000 00:00 0 7ffc6cd36000-7ffc6cd57000 rw-p 00000000 00:00 0 [stack] 7ffc6cd62000-7ffc6cd64000 r-xp 00000000 00:00 0 [vdso] 7ffc6cd64000-7ffc6cd66000 r--p 00000000 00:00 0 [vvar] ffffffffff600000-ffffffffff601000 r-xp 00000000 00:00 0 [vsyscall] VM Arguments: jvm_args: -Xmx30720M -Xms30720M -Xss1M -XX:HeapDumpPath=/root/netty_srv_dir/dumps/ -XX:+HeapDumpOnOutOfMemoryError java_command: p2p_server.jar Launcher Type: SUN_STANDARD Environment Variables: JAVA_HOME=/root/jdk1.7.0_65 CLASSPATH=.:/root/jdk1.7.0_65/lib/dt.jar PATH=/root/jdk1.7.0_65/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin SHELL=/bin/bash Signal Handlers: SIGSEGV: [libjvm.so+0x99f4a0], sa_mask[0]=0x7ffbfeff, sa_flags=0x10000004 SIGBUS: [libjvm.so+0x99f4a0], sa_mask[0]=0x7ffbfeff, sa_flags=0x10000004 SIGFPE: [libjvm.so+0x81a860], sa_mask[0]=0x7ffbfeff, sa_flags=0x10000004 SIGPIPE: [libjvm.so+0x81a860], sa_mask[0]=0x7ffbfeff, sa_flags=0x10000004 SIGXFSZ: [libjvm.so+0x81a860], sa_mask[0]=0x7ffbfeff, sa_flags=0x10000004 SIGILL: [libjvm.so+0x81a860], sa_mask[0]=0x7ffbfeff, sa_flags=0x10000004 SIGUSR1: SIG_DFL, sa_mask[0]=0x00000000, sa_flags=0x00000000 SIGUSR2: [libjvm.so+0x81c060], sa_mask[0]=0x00000000, sa_flags=0x10000004 SIGHUP: [libjvm.so+0x81d160], sa_mask[0]=0x7ffbfeff, sa_flags=0x10000004 SIGINT: SIG_IGN, sa_mask[0]=0x00000000, sa_flags=0x00000000 SIGTERM: [libjvm.so+0x81d160], sa_mask[0]=0x7ffbfeff, sa_flags=0x10000004 SIGQUIT: [libjvm.so+0x81d160], sa_mask[0]=0x7ffbfeff, sa_flags=0x10000004 --------------- S Y S T E M --------------- OS:8.2 uname:Linux 3.16.0-4-amd64 #1 SMP Debian 3.16.7-ckt20-1+deb8u2 (2016-01-02) x86_64 libc:glibc 2.19 NPTL 2.19 rlimit: STACK 8192k, CORE infinity, NPROC 128647, NOFILE 20000500, AS infinity load average:0.59 0.66 0.67 /proc/meminfo: MemTotal: 32966000 kB MemFree: 19330848 kB MemAvailable: 19319680 kB Buffers: 6548 kB Cached: 52364 kB SwapCached: 0 kB Active: 13048248 kB Inactive: 37956 kB Active(anon): 13034100 kB Inactive(anon): 17188 kB Active(file): 14148 kB Inactive(file): 20768 kB Unevictable: 7512 kB Mlocked: 7512 kB SwapTotal: 0 kB SwapFree: 0 kB Dirty: 52 kB Writeback: 0 kB AnonPages: 13035280 kB Mapped: 39640 kB Shmem: 17192 kB Slab: 435348 kB SReclaimable: 111668 kB SUnreclaim: 323680 kB KernelStack: 4080 kB PageTables: 28660 kB NFS_Unstable: 0 kB Bounce: 0 kB WritebackTmp: 0 kB CommitLimit: 16483000 kB Committed_AS: 34072488 kB VmallocTotal: 34359738367 kB VmallocUsed: 124264 kB VmallocChunk: 34359573760 kB HardwareCorrupted: 0 kB AnonHugePages: 0 kB HugePages_Total: 0 HugePages_Free: 0 HugePages_Rsvd: 0 HugePages_Surp: 0 Hugepagesize: 2048 kB DirectMap4k: 63648 kB DirectMap2M: 3031040 kB DirectMap1G: 30408704 kB CPU:total 8 (4 cores per cpu, 2 threads per core) family 6 model 60 stepping 3, cmov, cx8, fxsr, mmx, sse, sse2, sse3, ssse3, sse4.1, sse4.2, popcnt, avx, avx2, aes, erms, ht, tsc, tscinvbit /proc/cpuinfo: processor : 0 vendor_id : GenuineIntel cpu family : 6 model : 60 model name : Intel(R) Xeon(R) CPU E3-1231 v3 @ 3.40GHz stepping : 3 microcode : 0x12 cpu MHz : 3600.015 cache size : 8192 KB physical id : 0 siblings : 8 core id : 0 cpu cores : 4 apicid : 0 initial apicid : 0 fpu : yes fpu_exception : yes cpuid level : 13 wp : yes flags : fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc aperfmperf eagerfpu pni pclmulqdq dtes64 monitor ds_cpl vmx smx est tm2 ssse3 fma cx16 xtpr pdcm pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm ida arat xsaveopt pln pts dtherm tpr_shadow vnmi flexpriority ept vpid fsgsbase tsc_adjust bmi1 hle avx2 smep bmi2 erms invpcid rtm bogomips : 6784.38 clflush size : 64 cache_alignment : 64 address sizes : 39 bits physical, 48 bits virtual power management: processor : 1 vendor_id : GenuineIntel cpu family : 6 model : 60 model name : Intel(R) Xeon(R) CPU E3-1231 v3 @ 3.40GHz stepping : 3 microcode : 0x12 cpu MHz : 3600.015 cache size : 8192 KB physical id : 0 siblings : 8 core id : 1 cpu cores : 4 apicid : 2 initial apicid : 2 fpu : yes fpu_exception : yes cpuid level : 13 wp : yes flags : fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc aperfmperf eagerfpu pni pclmulqdq dtes64 monitor ds_cpl vmx smx est tm2 ssse3 fma cx16 xtpr pdcm pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm ida arat xsaveopt pln pts dtherm tpr_shadow vnmi flexpriority ept vpid fsgsbase tsc_adjust bmi1 hle avx2 smep bmi2 erms invpcid rtm bogomips : 6784.38 clflush size : 64 cache_alignment : 64 address sizes : 39 bits physical, 48 bits virtual power management: processor : 2 vendor_id : GenuineIntel cpu family : 6 model : 60 model name : Intel(R) Xeon(R) CPU E3-1231 v3 @ 3.40GHz stepping : 3 microcode : 0x12 cpu MHz : 3600.015 cache size : 8192 KB physical id : 0 siblings : 8 core id : 2 cpu cores : 4 apicid : 4 initial apicid : 4 fpu : yes fpu_exception : yes cpuid level : 13 wp : yes flags : fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc aperfmperf eagerfpu pni pclmulqdq dtes64 monitor ds_cpl vmx smx est tm2 ssse3 fma cx16 xtpr pdcm pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm ida arat xsaveopt pln pts dtherm tpr_shadow vnmi flexpriority ept vpid fsgsbase tsc_adjust bmi1 hle avx2 smep bmi2 erms invpcid rtm bogomips : 6784.38 clflush size : 64 cache_alignment : 64 address sizes : 39 bits physical, 48 bits virtual power management: processor : 3 vendor_id : GenuineIntel cpu family : 6 model : 60 model name : Intel(R) Xeon(R) CPU E3-1231 v3 @ 3.40GHz stepping : 3 microcode : 0x12 cpu MHz : 3600.015 cache size : 8192 KB physical id : 0 siblings : 8 core id : 3 cpu cores : 4 apicid : 6 initial apicid : 6 fpu : yes fpu_exception : yes cpuid level : 13 wp : yes flags : fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc aperfmperf eagerfpu pni pclmulqdq dtes64 monitor ds_cpl vmx smx est tm2 ssse3 fma cx16 xtpr pdcm pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm ida arat xsaveopt pln pts dtherm tpr_shadow vnmi flexpriority ept vpid fsgsbase tsc_adjust bmi1 hle avx2 smep bmi2 erms invpcid rtm bogomips : 6784.38 clflush size : 64 cache_alignment : 64 address sizes : 39 bits physical, 48 bits virtual power management: processor : 4 vendor_id : GenuineIntel cpu family : 6 model : 60 model name : Intel(R) Xeon(R) CPU E3-1231 v3 @ 3.40GHz stepping : 3 microcode : 0x12 cpu MHz : 3600.015 cache size : 8192 KB physical id : 0 siblings : 8 core id : 0 cpu cores : 4 apicid : 1 initial apicid : 1 fpu : yes fpu_exception : yes cpuid level : 13 wp : yes flags : fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc aperfmperf eagerfpu pni pclmulqdq dtes64 monitor ds_cpl vmx smx est tm2 ssse3 fma cx16 xtpr pdcm pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm ida arat xsaveopt pln pts dtherm tpr_shadow vnmi flexpriority ept vpid fsgsbase tsc_adjust bmi1 hle avx2 smep bmi2 erms invpcid rtm bogomips : 6784.38 clflush size : 64 cache_alignment : 64 address sizes : 39 bits physical, 48 bits virtual power management: processor : 5 vendor_id : GenuineIntel cpu family : 6 model : 60 model name : Intel(R) Xeon(R) CPU E3-1231 v3 @ 3.40GHz stepping : 3 microcode : 0x12 cpu MHz : 3600.015 cache size : 8192 KB physical id : 0 siblings : 8 core id : 1 cpu cores : 4 apicid : 3 initial apicid : 3 fpu : yes fpu_exception : yes cpuid level : 13 wp : yes flags : fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc aperfmperf eagerfpu pni pclmulqdq dtes64 monitor ds_cpl vmx smx est tm2 ssse3 fma cx16 xtpr pdcm pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm ida arat xsaveopt pln pts dtherm tpr_shadow vnmi flexpriority ept vpid fsgsbase tsc_adjust bmi1 hle avx2 smep bmi2 erms invpcid rtm bogomips : 6784.38 clflush size : 64 cache_alignment : 64 address sizes : 39 bits physical, 48 bits virtual power management: processor : 6 vendor_id : GenuineIntel cpu family : 6 model : 60 model name : Intel(R) Xeon(R) CPU E3-1231 v3 @ 3.40GHz stepping : 3 microcode : 0x12 cpu MHz : 3600.015 cache size : 8192 KB physical id : 0 siblings : 8 core id : 2 cpu cores : 4 apicid : 5 initial apicid : 5 fpu : yes fpu_exception : yes cpuid level : 13 wp : yes flags : fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc aperfmperf eagerfpu pni pclmulqdq dtes64 monitor ds_cpl vmx smx est tm2 ssse3 fma cx16 xtpr pdcm pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm ida arat xsaveopt pln pts dtherm tpr_shadow vnmi flexpriority ept vpid fsgsbase tsc_adjust bmi1 hle avx2 smep bmi2 erms invpcid rtm bogomips : 6784.38 clflush size : 64 cache_alignment : 64 address sizes : 39 bits physical, 48 bits virtual power management: processor : 7 vendor_id : GenuineIntel cpu family : 6 model : 60 model name : Intel(R) Xeon(R) CPU E3-1231 v3 @ 3.40GHz stepping : 3 microcode : 0x12 cpu MHz : 3600.015 cache size : 8192 KB physical id : 0 siblings : 8 core id : 3 cpu cores : 4 apicid : 7 initial apicid : 7 fpu : yes fpu_exception : yes cpuid level : 13 wp : yes flags : fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc aperfmperf eagerfpu pni pclmulqdq dtes64 monitor ds_cpl vmx smx est tm2 ssse3 fma cx16 xtpr pdcm pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm ida arat xsaveopt pln pts dtherm tpr_shadow vnmi flexpriority ept vpid fsgsbase tsc_adjust bmi1 hle avx2 smep bmi2 erms invpcid rtm bogomips : 6784.38 clflush size : 64 cache_alignment : 64 address sizes : 39 bits physical, 48 bits virtual power management: Memory: 4k page, physical 32966000k(19329856k free), swap 0k(0k free) vm_info: Java HotSpot(TM) 64-Bit Server VM (24.65-b04) for linux-amd64 JRE (1.7.0_65-b17), built on Jun 16 2014 20:55:34 by "java_re" with gcc 4.3.0 20080428 (Red Hat 4.3.0-8) time: Mon Mar 7 16:27:20 2016 elapsed time: 4456 seconds @yjdwbj this looks like a jdk bug. Try to update your java version and if you still see it open an issue in the openjdk project. Basically this gives the hind that the GC crashed the JVM. # V [libjvm.so+0x885e17] oopDesc* PSPromotionManager::copy_to_survivor_space<false>(oopDesc*)+0x177 ``
gharchive/issue
2016-03-07T08:37:54
2025-04-01T06:39:44.424216
{ "authors": [ "normanmaurer", "yjdwbj" ], "repo": "netty/netty", "url": "https://github.com/netty/netty/issues/4947", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
173016493
HTTP/2 Hpack Encoder Cleanup Motivation: The HTTP/2 HPACK Encoder class has some code which is only used for test purposes. This code can be removed to reduce complexity and member variable count. Modifications: Remove test code and update unit tests Other minor cleanup Result: Test code is removed from operational code. this is a minor cleanup ... we loose some ability to control the encoder in tests however I'm not sure if this provides too much value. @nmittler @normanmaurer @carl-mastrangelo - FYI No red flags from my pov. LGTM. LGTM 4.1 (208893a)
gharchive/pull-request
2016-08-24T17:50:21
2025-04-01T06:39:44.433520
{ "authors": [ "Scottmitch", "carl-mastrangelo", "nmittler" ], "repo": "netty/netty", "url": "https://github.com/netty/netty/pull/5746", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
319895025
Allow to disable native transport and native ssl support via system p… …roperty. Motivation: Sometimes it's useful to disable native transports / native ssl to debug a problem. We should allow to do so with a system property so people not need to adjust code for this. Modifications: Add system properties which allow to disable native transport and native ssl. Result: Easier to disable native code usage without code changes. @trustin done... will merge once ci pass. Awesome! This is, indeed, very useful for testing. Some notes for future readers who might also be interested in this change: This is dramatically easier to read with whitespace-only changes hidden The system property for disabling native SSL providers (i.e. netty-tcnative) is io.netty.handler.ssl.noOpenSsl (example: -Dio.netty.handler.ssl.noOpenSsl=true) The system property for disabling native transports (i.e. epoll and kqueue) is io.netty.transport.noNative (example: -Dio.netty.transport.noNative=true)
gharchive/pull-request
2018-05-03T12:13:24
2025-04-01T06:39:44.437335
{ "authors": [ "jchambers", "normanmaurer" ], "repo": "netty/netty", "url": "https://github.com/netty/netty/pull/7903", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2177681
Individual scenarios are no longer run if the entire feature is also sche Individual scenarios are no longer run if the entire feature is also scheduled to run for https://github.com/netzpirat/guard-cucumber/issues/22 Run Cucumber features features/password_client.feature -break test Run Cucumber features features/password_client.feature -fix test Run Cucumber features features/password_client.feature features/ password_client.feature:22 This is no longer the case Thanks a lot for this! I will release a bug fix version today.
gharchive/issue
2011-11-08T18:39:31
2025-04-01T06:39:44.472460
{ "authors": [ "netzpirat", "oreoshake" ], "repo": "netzpirat/guard-cucumber", "url": "https://github.com/netzpirat/guard-cucumber/issues/24", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
264010379
Use JsonUtf8Encoding : Encoding Escaping string character is hurt of performance of JSON serialization. It is possible to reduce escape cost by creating custom UTF8 Encoding that includes JSON encoding/decoding. for invoke internal FastAllocateString, it is necessary to inherit Encoding. public class JsonUtf8Encoding : Encoding { // called from GetString. // return CharCount is \" ... \" unescaped. public override int GetCharCount(byte[] bytes, int index, int count) { throw new NotImplementedException(); } public override int GetChars(byte[] bytes, int byteIndex, int byteCount, char[] chars, int charIndex) { throw new NotImplementedException(); } public override int GetMaxByteCount(int charCount) { throw new NotImplementedException(); } public override int GetMaxCharCount(int byteCount) { throw new NotImplementedException(); } public override int GetByteCount(char[] chars, int index, int count) { throw new NotImplementedException(); } public override int GetBytes(char[] chars, int charIndex, int charCount, byte[] bytes, int byteIndex) { throw new NotImplementedException(); } } Also, it is necessary to implement efficient UTF 8 encoding/decoding. I found this article. http://bjoern.hoehrmann.de/utf-8/decoder/dfa/ If there are any other good examples, please let me know. @itn3000 is trying fast utf8 <-> utf16 utilities. https://github.com/itn3000/unicode-convert-utilities @ufcpp is building custom UTF8 decoder. https://github.com/ufcpp/Utf8Utils NStack is golang like new encoding system. https://github.com/migueldeicaza/NStack System.Text.Utf8String is span based new primitive. https://github.com/dotnet/corefxlab/tree/master/src/System.Text.Utf8String/System/Text Regarding utf-8: http://nullprogram.com/blog/2017/10/06/ https://news.ycombinator.com/item?id=15423674 and related from https://github.com/dotnet/corefxlab/issues/1831 hi @neuecc we badly need your help, do you have an idea or workaround for our issue? https://github.com/neuecc/Utf8Json/issues/224
gharchive/issue
2017-10-09T20:04:31
2025-04-01T06:39:44.481910
{ "authors": [ "Tornhoof", "firephantomassasin", "neuecc" ], "repo": "neuecc/Utf8Json", "url": "https://github.com/neuecc/Utf8Json/issues/17", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2694072059
Feasibility and Capability of Inference on CPU Hi there, I have a usecase for your model on a spacecraft but the onboard system will only have a 1 GHz CPU and no dedicated GPU. I wanted to ask whether you believe it would be possible to implement your model on this level of hardware. I have attempted to and failed to carry out inference on the CPU also. I think you need to make changes to the model architecture code. Usually by replacing device variable with "cpu" instead of "0" or "cuda". I have tried this. There are challenges with regards to the half precision floating point data types. What exactly the problem is is not clear to me. Have you attempted to adjust the quantization settings of the model to prevent it from defaulting to half-precision (FP16)? I'll try to do the inference via a CPU and will let you know Hi leonbaiyu, you can change the datatype to FP32 instead of default FP16, by searching "amp" and make it False. Thank you Hi leonbaiyu, you can change the datatype to FP32 instead of default FP16, by searching "amp" and make it False. Thank you Thank you for this. This solved the problem and now I'm able to infer fairly quickly on CPU. Specifically, I removed any .cuda(), replaced 'cuda' device calls to 'cpu', replaced any .half() with float, and specified in model.init_bhwd for amp = False. No changes were required within the actual NeuFlow code, changes were only required in infer.py . The results are looking promising. We'll be in touch if we end up using your model in our applications.
gharchive/issue
2024-11-26T10:25:29
2025-04-01T06:39:44.486088
{ "authors": [ "ElseFaisal", "Study-is-happy", "leonbaiyu" ], "repo": "neufieldrobotics/NeuFlow_v2", "url": "https://github.com/neufieldrobotics/NeuFlow_v2/issues/10", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1566398607
Extractor pipeline 2.0 This is major update on the path to Generative Semantic Search The extractor pipeline was one of the first components in txtai, going all the way back to 1.0. Since then, much has changed both with txtai and externally. This pipeline has a lot of potential but it needs a couple updates. Make the following upgrades to the Extractor pipeline. Ability to run embeddings searches. Given that content is supported, text can be retrieved from the embeddings instance. In addition to extractive qa, support text generation models, sequence to sequence models and custom pipelines Better detection of when a tokenizer should be used (word vector models only) These changes will enable a prompt-driven approach to question-answering with LLMs. This includes Hugging Face models and external services like OpenAI/Cohere. Services can be called directly or with another library like langchain. Custom pipelines only require a __call__ interface. This recent notebook should help: https://github.com/neuml/txtai/blob/master/examples/62_RAG_with_llama_cpp_and_external_API_services.ipynb
gharchive/issue
2023-02-01T16:11:32
2025-04-01T06:39:44.489776
{ "authors": [ "davidmezzetti" ], "repo": "neuml/txtai", "url": "https://github.com/neuml/txtai/issues/417", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2008737790
[FIX] Remove --no-progress-meter Closes #213 Changes proposed in this pull request: replace the --no-progress-meter argument with -s -S which has the same effect but works on curl < 7.67.0 Checklist [ ] PR has an interpretable title with a prefix ([ENH], [FIX], [REF], [TST], [CI], [MNT], [INF], [MODEL], [DOC]) (see https://neurobagel.org/contributing/pull_requests for more info) [ ] PR links to GitHub issue with mention Closes #XXXX [ ] Tests pass [ ] Checks pass For new features: [ ] Tests have been added For bug fixes: [ ] There is at least one test that would fail under the original bug conditions. Pull Request Test Coverage Report for Build 6973817264 0 of 0 changed or added relevant lines in 0 files are covered. No unchanged relevant lines lost coverage. Overall coverage remained the same at 99.732% Totals Change from base Build 6882033445: 0.0% Covered Lines: 745 Relevant Lines: 747 💛 - Coveralls :crossed_fingers:
gharchive/pull-request
2023-11-23T19:57:39
2025-04-01T06:39:44.502411
{ "authors": [ "coveralls", "surchs" ], "repo": "neurobagel/api", "url": "https://github.com/neurobagel/api/pull/227", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
424485750
unimplemented DATA_PDU_TYPE_SUPPRESS_OUTPUT I want to use the API update_send_suppress_output in NeutrinoRDP,but xrdp not implement this. Somebody can help? xrdp_rdp_process_data(struct xrdp_rdp *self, struct stream *s) { int data_type; in_uint8s(s, 6); in_uint8s(s, 2); /* len */ in_uint8(s, data_type); in_uint8s(s, 1); /* ctype */ in_uint8s(s, 2); /* clen */ DEBUG(("xrdp_rdp_process_data code %d", data_type)); switch (data_type) { case RDP_DATA_PDU_POINTER: /* 27(0x1b) */ xrdp_rdp_process_data_pointer(self, s); break; case RDP_DATA_PDU_INPUT: /* 28(0x1c) */ xrdp_rdp_process_data_input(self, s); break; case RDP_DATA_PDU_CONTROL: /* 20(0x14) */ xrdp_rdp_process_data_control(self, s); break; case RDP_DATA_PDU_SYNCHRONISE: /* 31(0x1f) */ xrdp_rdp_process_data_sync(self); break; case 33: /* 33(0x21) ?? Invalidate an area I think */ xrdp_rdp_process_screen_update(self, s); break; case 35: /* 35(0x23) */ /* 35 ?? this comes when minimizing a full screen mstsc.exe 2600 */ /* I think this is saying the client no longer wants screen */ /* updates and it will issue a 33 above to catch up */ /* so minimized apps don't take bandwidth */ break; case 36: /* 36(0x24) ?? disconnect query? */ /* when this message comes, send a 37 back so the client */ /* is sure the connection is alive and it can ask if user */ /* really wants to disconnect */ xrdp_rdp_send_disconnect_query_response(self); /* send a 37 back */ break; case RDP_DATA_PDU_FONT2: /* 39(0x27) */ xrdp_rdp_process_data_font(self, s); break; case 56: /* PDUTYPE2_FRAME_ACKNOWLEDGE 0x38 */ xrdp_rdp_process_frame_ack(self, s); break; default: g_writeln("unknown in xrdp_rdp_process_data %d", data_type); break; } return 0; } Yes, I think you're right, xrdp is ignoring the message. I think we implemented invalidate(Client Refresh Rect PDU) but not suppress. https://docs.microsoft.com/en-us/openspecs/windows_protocols/ms-rdpbcgr/abebe6ac-ccdd-4547-89a8-d629831f5527 @ jsorg71 Can you help to implement "suppress"? @ jsorg71 thx for your help.Is there any progress? working on this in branch d6cc487b3dcdc5a51ba9f9749fdc64a5aa5cd867 I'm a bit confused on how this should work. We can easily stop surface commands and orders, but should we stop channels like audio data? I tested it on Win 10 to Win 10 RDPing. At least audio is not suppressed. Playback some YouTube video with high graphics movements (network bandwidth consumption is around 10Mbps) Minimize the mstsc Network bandwidth consumption become less (around 200kbps, looks like only audio is transfered) Thanks @metalefty I think that makes sense. They must only suppress orders, surface commands and GFX. I wonder if cursor updates happen. Docs don't really say. Anyway, the client does not seem to panic if suppressed things are still sent so we can start with surface commands and orders. After thinking about this a lot, I think we should handle this in the back-end. xrdp should pass the suppress info the the module so it can make the best use of it. In proxy mode, it can pass it to the upstream RDP server. When using Xorg, it can pass it to Xorg, It can get complicated with orders. This make that work nicely. VNC has this sort of thing built in so libvnc can handle it for the RDP to VNC proxy or Xvnc. @jsorg71 thx for your works. I have tested your branch. NeutrinoRDP send a suppress data PDU to xrdp for suppressing, but NeutrinoRDP still receive surface commands. log of xrdp: xrdp_rdp_process_suppress: allowDisplayUpdates 0 bytes left 3 xrdp_rdp_process_suppress: suppress_output 1 xrdp:xrdp_mm [4212885514]: xrdp_mm_suppress_output: suppress 1 left 0 top 0 right 0 bottom 0 xrdp_rdp_process_suppress: allowDisplayUpdates 1 bytes left 11 xrdp_rdp_process_suppress: suppress_output 0 left 0 top 0 right 1920 bottom 1080 xrdp:xrdp_mm [4212895430]: xrdp_mm_suppress_output: suppress 0 left 0 top 0 right 1920 bottom 1080 log of NeutrinoRDP: [2019-04-08 12:10:02] SuppressOutput@1850 update_send_suppress_output,allow=0 [2019-04-08 12:10:02] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:02] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:02] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:02] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:02] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:02] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:02] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:02] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:03] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:03] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:03] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:03] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:03] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:03] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:03] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:03] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:03] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:03] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:03] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:03] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:03] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:03] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:03] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:04] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:04] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:04] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:04] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:04] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:04] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:04] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:04] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:04] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:04] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:04] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:04] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:04] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:04] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:04] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:04] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:04] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:04] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:04] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:04] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:04] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:04] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:05] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:05] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:05] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:05] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:05] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:05] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:05] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:05] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:05] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:05] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:05] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:05] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:05] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:05] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:05] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:05] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:05] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:05] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:05] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:05] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:05] SuppressOutput@1850 update_send_suppress_output,allow=1 [2019-04-08 12:10:05] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:05] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:05] xf_gdi_surface_bits@666 get you!!!! [2019-04-08 12:10:05] xf_gdi_surface_bits@666 get you!!!! NeutrinoRDP/client/X11/xf_gdi.c void xf_gdi_surface_bits(rdpContext* context, SURFACE_BITS_COMMAND* surface_bits_command) { int i, tx, ty; XImage* image; RFX_MESSAGE* message; xfInfo* xfi = ((xfContext*) context)->xfi; RFX_CONTEXT* rfx_context = (RFX_CONTEXT*) xfi->rfx_context; NSC_CONTEXT* nsc_context = (NSC_CONTEXT*) xfi->nsc_context; LOG("get you!!!!"); if (surface_bits_command->codecID == CODEC_ID_H264) { ....... static void SuppressOutput(int allow) { RECTANGLE_16 updateAreas; updateAreas.left = 0; updateAreas.top = 0; updateAreas.right = screen_w; updateAreas.bottom = screen_h; LOG("update_send_suppress_output,allow=%d",allow) if(g_update != NULL ) { update_send_suppress_output(g_update->context, allow, &updateAreas); } } You should get xrdp from(same as above) https://github.com/jsorg71/xrdp/tree/suppress and xorgxrdp from https://github.com/jsorg71/xorgxrdp/tree/suppress If that works, I'll make a PR. @jsorg71 your xrdp and xorgxrdp works well for DATA_PDU_TYPE_SUPPRESS_OUTPUT. thx again. @jsorg71 ,update_send_suppress_output(allow=0) can suppress surface cmd update , but update_send_suppress_output(allow=1) cannot resume surface cmd update randomly. And at that moment,mouse can move, but graphic output not update any more. Another question:update_send_suppress_output(allow=1) called ,NeutrinoRDP receive a disconnect PDU from xrdp(randomly). log: [1970-01-01 21:28:34] SuppressOutput@1856 update_send_suppress_output,allow=1 [1970-01-01 21:28:34] rdp_read_header@253 rdp->disconnect = true,reason = 128 [1970-01-01 21:28:34] rdp_recv_tpkt_pdu@742 rdp_recv_tpkt_pdu: disconnect transport_check_fds: transport->recv_callback failed transport_check_fds: do_callback failed Failed to check FreeRDP file descriptor [1970-01-01 21:28:34] freerdp_context_free@217 state = 10, errorInfo = 0 static tbool rdp_recv_tpkt_pdu(rdpRdp* rdp, STREAM* s) { uint16 length; uint16 pduType; uint16 pduLength; uint16 pduSource; uint16 channelId; uint16 securityFlags; uint8* nextp; LLOGLN(10, ("rdp_recv_tpkt_pdu:")); if (!rdp_read_header(rdp, s, &length, &channelId)) { LLOGLN(0, ("Incorrect RDP header.")); return false; } LLOGLN(10, ("rdp_recv_tpkt_pdu: length %d", length)); if (rdp->disconnect) { //LLOGLN(0, ("rdp_recv_tpkt_pdu: disconnect")); LOG("rdp_recv_tpkt_pdu: disconnect") return false; } tbool rdp_read_header(rdpRdp* rdp, STREAM* s, uint16* length, uint16* channel_id) { uint8 reason; uint16 initiator; enum DomainMCSPDU MCSPDU; MCSPDU = (rdp->settings->server_mode) ? DomainMCSPDU_SendDataRequest : DomainMCSPDU_SendDataIndication; if (!mcs_read_domain_mcspdu_header(s, &MCSPDU, length)) { LLOGLN(0, ("rdp_read_header: mcs_read_domain_mcspdu_header failed")); return false; } if (*length - 8 > stream_get_left(s)) { LLOGLN(0, ("rdp_read_header: parse error")); return false; } if (MCSPDU == DomainMCSPDU_DisconnectProviderUltimatum) { if (!per_read_enumerated(s, &reason, 0)) { LLOGLN(0, ("rdp_read_header: per_read_enumerated failed")); return false; } rdp->disconnect = true; *channel_id = MCS_GLOBAL_CHANNEL_ID; LOG("rdp->disconnect = true,reason = %d",reason) return true; } @sherwinkh I've been testing, I can't reproduce. Are you using NeutrinoRDP from my branch. I ask because you said you were using NeutrinoRDP suppress output but it was not done. I figured you had your own version. Try my version and also make sure you make clean and make again to make sure you have everything built for all components xrdp, xorgxrdp, NeutrinoRDP. Also can you test the ms client? Does that disconnect too? I have merged your commits into xrdp, xorgxrdp,NeutrinoRDP(branch devel). NeutrinoRDP run on my arm board device,and jpeg codec is the only codec in xf_gdi_surface_bits,using hw jpeg codec. build setting args in xrdp :configure --enable-fuse --enable-tjpeg --enable-jpeg . Server :ubuntu 16.04.5 I ask because you said you were using NeutrinoRDP suppress output but it was not done ==> in most case ,suppress PDU(allow=0) can suppress graphic output ,PDU(allow=1) can resume graphic output. but sometimes,suppress PDU(allow=0) can suppress graphic output ,PDU(allow=1) cannot resume graphic output Another question:update_send_suppress_output(allow=1) called ,NeutrinoRDP receive a disconnect PDU from xrdp(randomly). log: [1970-01-01 21:28:34] SuppressOutput@1856 update_send_suppress_output,allow=1 [1970-01-01 21:28:34] rdp_read_header@253 rdp->disconnect = true,reason = 128 [1970-01-01 21:28:34] rdp_recv_tpkt_pdu@742 rdp_recv_tpkt_pdu: disconnect transport_check_fds: transport->recv_callback failed transport_check_fds: do_callback failed Failed to check FreeRDP file descriptor [1970-01-01 21:28:34] freerdp_context_free@217 state = 10, errorInfo = 0 static tbool rdp_recv_tpkt_pdu(rdpRdp* rdp, STREAM* s) { uint16 length; uint16 pduType; uint16 pduLength; uint16 pduSource; uint16 channelId; uint16 securityFlags; uint8* nextp; LLOGLN(10, ("rdp_recv_tpkt_pdu:")); if (!rdp_read_header(rdp, s, &length, &channelId)) { LLOGLN(0, ("Incorrect RDP header.")); return false; } LLOGLN(10, ("rdp_recv_tpkt_pdu: length %d", length)); if (rdp->disconnect) { //LLOGLN(0, ("rdp_recv_tpkt_pdu: disconnect")); LOG("rdp_recv_tpkt_pdu: disconnect") return false; } tbool rdp_read_header(rdpRdp* rdp, STREAM* s, uint16* length, uint16* channel_id) { uint8 reason; uint16 initiator; enum DomainMCSPDU MCSPDU; MCSPDU = (rdp->settings->server_mode) ? DomainMCSPDU_SendDataRequest : DomainMCSPDU_SendDataIndication; if (!mcs_read_domain_mcspdu_header(s, &MCSPDU, length)) { LLOGLN(0, ("rdp_read_header: mcs_read_domain_mcspdu_header failed")); return false; } if (*length - 8 > stream_get_left(s)) { LLOGLN(0, ("rdp_read_header: parse error")); return false; } if (MCSPDU == DomainMCSPDU_DisconnectProviderUltimatum) { if (!per_read_enumerated(s, &reason, 0)) { LLOGLN(0, ("rdp_read_header: per_read_enumerated failed")); return false; } rdp->disconnect = true; *channel_id = MCS_GLOBAL_CHANNEL_ID; LOG("rdp->disconnect = true,reason = %d",reason) return true; } I found some log in xorgxrdp.xx.log: [ 15465.685] rdpClientConRecv: g_sck_recv failed(returned 0) [ 15465.691] rdpClientConRecvMsg: error [ 15465.692] rdpClientConCheck: rdpClientConGotData failed [ 15465.692] rdpClientConDisconnect: [ 15465.692] rdpRemoveClientConFromDev: removing clientCon 0x5632c3596ed0 I looks like connect closed, but I don't know why static int rdpClientConRecv(rdpPtr dev, rdpClientCon *clientCon, char *data, int len) { int rcvd; if (!clientCon->connected) { return 1; } while (len > 0) { rcvd = g_sck_recv(clientCon->sck, data, len, 0); if (rcvd == -1) { if (g_sck_last_error_would_block(clientCon->sck)) { g_sleep(1); } else { LLOGLN(0, ("rdpClientConRecv: g_sck_recv failed(returned -1)")); clientCon->connected = FALSE; return 1; } } else if (rcvd == 0) { LLOGLN(0, ("rdpClientConRecv: g_sck_recv failed(returned 0)")); clientCon->connected = FALSE; return 1; } else { data += rcvd; len -= rcvd; } } return 0; } My guess is that you have some error in your suppress output PDU. Grab the suppress branch from NeutrinoRDP and test that on x86 or x64 first. I did a lot of minimize / restore and I don't see any problem. I test too with MSTSC. I grab the suppress branch from NeutrinoRDP,do the flowing tests: 8 arm board devices running NeutrinoRDP, sending SuppressOutput(0) or SuppressOutput(1) to ubuntu server in every 3 mins. After 4 hours , one arm board device can't restore output. static void SuppressOutput(int allow) { RECTANGLE_16 updateAreas; updateAreas.left = 0; updateAreas.top = 0; updateAreas.right = screen_w; updateAreas.bottom = screen_h; if(g_update != NULL ) { update_send_suppress_output(g_update->context, allow, &updateAreas); LOG("update_send_suppress_output,allow(%d)",allow) } } can you send me what your update_send_suppress_output() function looks like after 24 hours , 5 arm devices can't restore output.and one is disconnected.Only one device is OK. static void update_write_suppress_output(STREAM* s, uint8 allow, RECTANGLE_16* area) { stream_write_uint8(s, allow); /* allowDisplayUpdates (1 byte) */ stream_seek(s, 3); /* pad3Octets (3 bytes) */ if (allow > 0) freerdp_write_rectangle_16(s, area); } void update_send_suppress_output(rdpContext* context, uint8 allow, RECTANGLE_16* area) { STREAM* s; rdpRdp* rdp = context->rdp; s = rdp_data_pdu_init(rdp); update_write_suppress_output(s, allow, area); rdp_send_data_pdu(rdp, s, DATA_PDU_TYPE_SUPPRESS_OUTPUT, rdp->mcs->user_id); } I ran one session(MSTSC) in recorder all night, no problem. I'm doing RFX you are using a JPEG session I think. I can try JPEG. Shouldn't matter. How are you generating the event to send the suppress message? is it coming from a different thread? Any chance you can try my branch, I still think something is wrong with your client. The Xorg messages in your last post is normal when the client disconnects. [ 15465.685] rdpClientConRecv: g_sck_recv failed(returned 0) ... The Xorg messages in your last post is normal when the client disconnects. [ 15465.685] rdpClientConRecv: g_sck_recv failed(returned 0) ==> But when I disconnect the client ,I don't see that log in xorgrdp.xx.log I started a thread to watch a pipe,then another program will write "openSuppress" or "closeSuppress" to the pipe in every 3 minutes. I will grab your branch,modify nothing ,using the software jpeg codec,then test "suppress". My code like this: static void* control(void * none) { char buff[256]; char *argv[16]; int argc; int nRead; LOG("\t[%-16s] starting, the tid=%d",__func__,gettid()); while(1) { //nRead = read(fd_pipe_control, &flag, sizeof(int)); nRead = read(fd_pipe_control, buff, 256); if(nRead > 0){ argc = splitString(argv, buff, nRead); if(argc == 0) continue; if(strcmp(argv[0], "openSuppress") == 0) { SuppressOutput(0); } else if(strcmp(argv[0], "closeSuppress") == 0) { SuppressOutput(1); } } } LOG("control thread exit !..... " ) pthread_exit(NULL); } static int control_init(char *pipeName) { if( access(PIPENAME, R_OK) == -1 ){ PT_FUNC_PT_STR("", " pipe file not exist ") return -1; } int ret; fd_pipe_control = open(pipeName, O_RDWR/*O_RDONLY | O_NONBLOCK*/); if(-1 == fd_pipe_control) { PT_FUNC_PT_STR("", " open pipe error") return -1; } ret = pthread_create(&thread_broadcast_control, NULL, control, NULL); if(ret) { close(fd_pipe_control); return -1; } return 0; } int main(int argc, char* argv[]) { pthread_t thread; freerdp* instance; struct thread_data* data; freerdp_handle_signals(); setlocale(LC_ALL, ""); freerdp_channels_global_init(); g_sem = freerdp_sem_new(1); instance = freerdp_new(); jpegDecoder_init(); hky_show_init(); control_init(PIPENAME); instance->PreConnect = xf_pre_connect; instance->PostConnect = xf_post_connect; instance->Authenticate = xf_authenticate; instance->VerifyCertificate = xf_verify_certificate; instance->ReceiveChannelData = xf_receive_channel_data; instance->context_size = sizeof(xfContext); instance->ContextNew = (pContextNew) xf_context_new; instance->ContextFree = (pContextFree) xf_context_free; freerdp_context_new(instance); instance->context->argc = argc; instance->context->argv = argv; instance->settings->sw_gdi = false; data = (struct thread_data*) xzalloc(sizeof(struct thread_data)); data->instance = instance; g_thread_count++; pthread_create(&thread, 0, thread_func, data); It's the thread sending the suppress that making your session unstable. Only one thread can call into libfreerdp per session. You can have 2 sessions and then 2 threads, one for each but only one thread session. Get rid of the thread and move the pipe to the main loop and add it to the select in xfreerdp_run. Get rid of the thread and move the pipe to the main loop and add it to the select in xfreerdp_run ==》how to do this? I can't create my thread in rdp session? I edited your code a bit but I didn't compile. You'll have to debug. Thanks a lot ,testing is running. After about 24 hours,8 devices are running normally.I think It works. Thanks again~ Ok, great, I'll close this and we can open a new one if we find issues
gharchive/issue
2019-03-23T10:20:37
2025-04-01T06:39:44.581394
{ "authors": [ "jsorg71", "metalefty", "sherwinkh" ], "repo": "neutrinolabs/xrdp", "url": "https://github.com/neutrinolabs/xrdp/issues/1312", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1691452317
black screen xrdp in Mate It is not possible to connect to RDP in Mate environment on stock settings. Installed only xrdp and xorgxrdp packages. The connection goes through, but the session is not established. xrdp, version: 0.9.21 xorgxrdp, version: 0.9.19 xrdp.log: [20230501-16:16:08] [INFO ] address [0.0.0.0] port [3389] mode 1 [20230501-16:16:08] [INFO ] listening to port 3389 on 0.0.0.0 [20230501-16:16:08] [INFO ] xrdp_listen_pp done [20230501-16:16:08] [DEBUG] Closed socket 7 (AF_INET6 :: port 3389) [20230501-16:16:10] [INFO ] starting xrdp with pid 2455 [20230501-16:16:10] [INFO ] address [0.0.0.0] port [3389] mode 1 [20230501-16:16:10] [INFO ] listening to port 3389 on 0.0.0.0 [20230501-16:16:10] [INFO ] xrdp_listen_pp done [20230501-16:16:24] [INFO ] Socket 12: AF_INET6 connection received from ::ffff:192.168.0.82 port 49236 [20230501-16:16:24] [DEBUG] Closed socket 12 (AF_INET6 ::ffff:192.168.0.5 port 3389) [20230501-16:16:24] [DEBUG] Closed socket 11 (AF_INET6 :: port 3389) [20230501-16:16:24] [DEBUG] item ini_version, value 1 [20230501-16:16:24] [DEBUG] item fork, value true [20230501-16:16:24] [DEBUG] item port, value 3389 [20230501-16:16:24] [DEBUG] item use_vsock, value false [20230501-16:16:24] [DEBUG] item tcp_nodelay, value true [20230501-16:16:24] [DEBUG] item tcp_keepalive, value true [20230501-16:16:24] [DEBUG] item security_layer, value negotiate [20230501-16:16:24] [DEBUG] item crypt_level, value high [20230501-16:16:24] [DEBUG] item certificate, value [20230501-16:16:24] [INFO ] Using default X.509 certificate: /etc/xrdp/cert.pem [20230501-16:16:24] [DEBUG] item key_file, value [20230501-16:16:24] [INFO ] Using default X.509 key file: /etc/xrdp/key.pem [20230501-16:16:24] [DEBUG] item ssl_protocols, value TLSv1.2, TLSv1.3 [20230501-16:16:24] [DEBUG] TLSv1.3 enabled [20230501-16:16:24] [DEBUG] TLSv1.2 enabled [20230501-16:16:24] [DEBUG] item autorun, value [20230501-16:16:24] [DEBUG] item allow_channels, value true [20230501-16:16:24] [DEBUG] item allow_multimon, value true [20230501-16:16:24] [DEBUG] item bitmap_cache, value true [20230501-16:16:24] [DEBUG] item bitmap_compression, value true [20230501-16:16:24] [DEBUG] item bulk_compression, value true [20230501-16:16:24] [DEBUG] item max_bpp, value 16 [20230501-16:16:24] [DEBUG] item new_cursors, value true [20230501-16:16:24] [DEBUG] item use_fastpath, value both [20230501-16:16:24] [DEBUG] item blue, value 009cb5 [20230501-16:16:24] [DEBUG] item grey, value dedede [20230501-16:16:24] [DEBUG] item ls_top_window_bg_color, value 009cb5 [20230501-16:16:24] [DEBUG] item ls_width, value 350 [20230501-16:16:24] [DEBUG] item ls_height, value 430 [20230501-16:16:24] [DEBUG] item ls_bg_color, value dedede [20230501-16:16:24] [DEBUG] item ls_logo_filename, value [20230501-16:16:24] [DEBUG] item ls_logo_x_pos, value 55 [20230501-16:16:24] [DEBUG] item ls_logo_y_pos, value 50 [20230501-16:16:24] [DEBUG] item ls_label_x_pos, value 30 [20230501-16:16:24] [DEBUG] item ls_label_width, value 65 [20230501-16:16:24] [DEBUG] item ls_input_x_pos, value 110 [20230501-16:16:24] [DEBUG] item ls_input_width, value 210 [20230501-16:16:24] [DEBUG] item ls_input_y_pos, value 220 [20230501-16:16:24] [DEBUG] item ls_btn_ok_x_pos, value 142 [20230501-16:16:24] [DEBUG] item ls_btn_ok_y_pos, value 370 [20230501-16:16:24] [DEBUG] item ls_btn_ok_width, value 85 [20230501-16:16:24] [DEBUG] item ls_btn_ok_height, value 30 [20230501-16:16:24] [DEBUG] item ls_btn_cancel_x_pos, value 237 [20230501-16:16:24] [DEBUG] item ls_btn_cancel_y_pos, value 370 [20230501-16:16:24] [DEBUG] item ls_btn_cancel_width, value 85 [20230501-16:16:24] [DEBUG] item ls_btn_cancel_height, value 30 [20230501-16:16:24] [INFO ] Security protocol: configured [SSL|RDP], requested [SSL|HYBRID|HYBRID_EX|RDP], selected [SSL] [20230501-16:16:24] [DEBUG] Using TLS security, and setting RDP security crypto to LEVEL_NONE and METHOD_NONE [20230501-16:16:24] [DEBUG] [MCS Connection Sequence] receive connection request [20230501-16:16:24] [INFO ] Connected client computer name: iPhone (Але [20230501-16:16:24] [DEBUG] Client supports 40 bit encryption [20230501-16:16:24] [DEBUG] Client supports 128 bit encryption [20230501-16:16:24] [DEBUG] Client supports 56 bit encryption [20230501-16:16:24] [DEBUG] Client supports fips encryption [20230501-16:16:24] [DEBUG] The connection is using TLS, skipping RDP crypto negotiation [20230501-16:16:24] [DEBUG] Adding channel: name rdpdr, channel id 1004, flags 0x80800000 [20230501-16:16:24] [DEBUG] Adding channel: name rdpsnd, channel id 1005, flags 0xc0000000 [20230501-16:16:24] [DEBUG] Adding channel: name cliprdr, channel id 1006, flags 0xc0a00000 [20230501-16:16:24] [DEBUG] Adding channel: name drdynvc, channel id 1007, flags 0xc0800000 [20230501-16:16:24] [WARN ] Received [MS-RDPBCGR] TS_UD_HEADER type 0xc006 is unknown (ignored) [20230501-16:16:24] [WARN ] Received [MS-RDPBCGR] TS_UD_HEADER type 0xc00a is unknown (ignored) [20230501-16:16:24] [WARN ] Client requested 32 bpp color depth, but the server configuration is limited to 16 bpp. Downgrading the color depth to 16 bits-per-pixel. [20230501-16:16:24] [DEBUG] [MCS Connection Sequence] construct connection response [20230501-16:16:24] [DEBUG] using no security [20230501-16:16:24] [DEBUG] [MCS Connection Sequence] send connection response [20230501-16:16:24] [DEBUG] [MCS Connection Sequence] receive erect domain request [20230501-16:16:24] [DEBUG] [MCS Connection Sequence] receive attach user request [20230501-16:16:24] [DEBUG] [MCS Connection Sequence] send attach user confirm [20230501-16:16:24] [DEBUG] [MCS Connection Sequence (TLS)] receive channel join request [20230501-16:16:24] [DEBUG] [MCS Connection Sequence (TLS)] send channel join confirm [20230501-16:16:24] [DEBUG] [MCS Connection Sequence (TLS)] receive channel join request [20230501-16:16:24] [DEBUG] [MCS Connection Sequence (TLS)] send channel join confirm [20230501-16:16:24] [DEBUG] [MCS Connection Sequence (TLS)] receive channel join request [20230501-16:16:24] [DEBUG] [MCS Connection Sequence (TLS)] send channel join confirm [20230501-16:16:24] [DEBUG] [MCS Connection Sequence (TLS)] receive channel join request [20230501-16:16:24] [DEBUG] [MCS Connection Sequence (TLS)] send channel join confirm [20230501-16:16:24] [DEBUG] [MCS Connection Sequence (TLS)] receive channel join request [20230501-16:16:24] [DEBUG] [MCS Connection Sequence (TLS)] send channel join confirm [20230501-16:16:24] [DEBUG] [MCS Connection Sequence (TLS)] receive channel join request [20230501-16:16:24] [DEBUG] [MCS Connection Sequence (TLS)] send channel join confirm [20230501-16:16:24] [DEBUG] [MCS Connection Sequence (TLS)] completed [20230501-16:16:24] [INFO ] xrdp_load_keyboard_layout: Keyboard information sent by the RDP client, keyboard_type:[0x04], keyboard_subtype:[0x00], keylayout:[0x00000409] [20230501-16:16:24] [DEBUG] keyboard_cfg_file /etc/xrdp/xrdp_keyboard.ini [20230501-16:16:24] [DEBUG] xrdp_load_keyboard_layout: item rdp_layout_us value 0x00000409 [20230501-16:16:24] [DEBUG] xrdp_load_keyboard_layout: skipping configuration item - rdp_layout_us, continuing to next section [20230501-16:16:24] [DEBUG] xrdp_load_keyboard_layout: item rdp_layout_us value us [20230501-16:16:24] [DEBUG] xrdp_load_keyboard_layout: skipping configuration item - rdp_layout_us, continuing to next section [20230501-16:16:24] [DEBUG] xrdp_load_keyboard_layout: item keyboard_type value 4 [20230501-16:16:24] [DEBUG] xrdp_load_keyboard_layout: item keyboard_subtype value 3 [20230501-16:16:24] [DEBUG] xrdp_load_keyboard_layout: item keyboard_type value 7 [20230501-16:16:24] [DEBUG] xrdp_load_keyboard_layout: item keyboard_subtype value 2 [20230501-16:16:24] [DEBUG] xrdp_load_keyboard_layout: item model value pc105 [20230501-16:16:24] [DEBUG] xrdp_load_keyboard_layout: item rdp_layouts value default_rdp_layouts [20230501-16:16:24] [DEBUG] xrdp_load_keyboard_layout: item layouts_map value default_layouts_map [20230501-16:16:24] [DEBUG] xrdp_load_keyboard_layout: item rdp_layout_us value us [20230501-16:16:24] [DEBUG] xrdp_load_keyboard_layout: skipping configuration item - rdp_layout_us, continuing to next section [20230501-16:16:24] [INFO ] xrdp_load_keyboard_layout: model [] variant [] layout [us] options [] [20230501-16:16:24] [INFO ] TLS connection established from ::ffff:192.168.0.82 port 49236: TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 [20230501-16:16:24] [DEBUG] Client requested auto logon. [20230501-16:16:24] [DEBUG] Client requested compression enabled. [20230501-16:16:24] [DEBUG] Client supplied domain: [20230501-16:16:24] [DEBUG] Client supplied username: user [20230501-16:16:24] [DEBUG] Client supplied password: <omitted from log> [20230501-16:16:24] [DEBUG] Client supplied program: [20230501-16:16:24] [DEBUG] Client supplied directory: [20230501-16:16:24] [INFO ] xrdp_caps_process_pointer: client supports new(color) cursor [20230501-16:16:24] [INFO ] xrdp_process_offscreen_bmpcache: support level 0 cache size 0 MB cache entries 0 [20230501-16:16:24] [INFO ] xrdp_caps_process_codecs: nscodec, codec id 1, properties len 3 [20230501-16:16:24] [DEBUG] xrdp_000009a1_wm_login_state_event_00000001 [20230501-16:16:24] [INFO ] Loading keymap file /etc/xrdp/km-00000409.ini [20230501-16:16:24] [WARN ] local keymap file for 0x00000409 found and doesn't match built in keymap, using local keymap file [20230501-16:16:24] [DEBUG] Login state change request WMLS_RESET -> WMLS_RESET [20230501-16:16:24] [DEBUG] xrdp_wm_login_mode_changed: login_mode is 0 [20230501-16:16:24] [DEBUG] Login state change request WMLS_RESET -> WMLS_USER_PROMPT [20230501-16:16:24] [DEBUG] in xrdp_wm_init: [20230501-16:16:24] [DEBUG] ini_version: 1 [20230501-16:16:24] [DEBUG] use_bitmap_cache: 1 [20230501-16:16:24] [DEBUG] use_bitmap_compression: 1 [20230501-16:16:24] [DEBUG] port: 3389 [20230501-16:16:24] [DEBUG] crypt_level: 3 [20230501-16:16:24] [DEBUG] allow_channels: 1 [20230501-16:16:24] [DEBUG] max_bpp: 16 [20230501-16:16:24] [DEBUG] fork: 1 [20230501-16:16:24] [DEBUG] tcp_nodelay: 1 [20230501-16:16:24] [DEBUG] tcp_keepalive: 1 [20230501-16:16:24] [DEBUG] tcp_send_buffer_bytes: 0 [20230501-16:16:24] [DEBUG] tcp_recv_buffer_bytes: 0 [20230501-16:16:24] [DEBUG] new_cursors: 1 [20230501-16:16:24] [DEBUG] allow_multimon: 1 [20230501-16:16:24] [DEBUG] grey: 14606046 [20230501-16:16:24] [DEBUG] black: 0 [20230501-16:16:24] [DEBUG] dark_grey: 0 [20230501-16:16:24] [DEBUG] blue: 40117 [20230501-16:16:24] [DEBUG] dark_blue: 0 [20230501-16:16:24] [DEBUG] white: 0 [20230501-16:16:24] [DEBUG] red: 0 [20230501-16:16:24] [DEBUG] green: 0 [20230501-16:16:24] [DEBUG] background: 0 [20230501-16:16:24] [DEBUG] autorun: [20230501-16:16:24] [DEBUG] hidelogwindow: 0 [20230501-16:16:24] [DEBUG] require_credentials: 0 [20230501-16:16:24] [DEBUG] bulk_compression: 1 [20230501-16:16:24] [DEBUG] new_cursors: 1 [20230501-16:16:24] [DEBUG] nego_sec_layer: 0 [20230501-16:16:24] [DEBUG] allow_multimon: 1 [20230501-16:16:24] [DEBUG] enable_token_login: 0 [20230501-16:16:24] [DEBUG] ls_top_window_bg_color: 4f6 [20230501-16:16:24] [DEBUG] ls_width: 350 [20230501-16:16:24] [DEBUG] ls_height: 430 [20230501-16:16:24] [DEBUG] ls_bg_color: defb [20230501-16:16:24] [DEBUG] ls_title: [20230501-16:16:24] [DEBUG] ls_logo_filename: [20230501-16:16:24] [DEBUG] ls_logo_x_pos: 55 [20230501-16:16:24] [DEBUG] ls_logo_y_pos: 50 [20230501-16:16:24] [DEBUG] ls_label_x_pos: 30 [20230501-16:16:24] [DEBUG] ls_label_width: 65 [20230501-16:16:24] [DEBUG] ls_input_x_pos: 110 [20230501-16:16:24] [DEBUG] ls_input_width: 210 [20230501-16:16:24] [DEBUG] ls_input_y_pos: 220 [20230501-16:16:24] [DEBUG] ls_btn_ok_x_pos: 142 [20230501-16:16:24] [DEBUG] ls_btn_ok_y_pos: 370 [20230501-16:16:24] [DEBUG] ls_btn_ok_width: 85 [20230501-16:16:24] [DEBUG] ls_btn_ok_height: 30 [20230501-16:16:24] [DEBUG] ls_btn_cancel_x_pos: 237 [20230501-16:16:24] [DEBUG] ls_btn_cancel_y_pos: 370 [20230501-16:16:24] [DEBUG] ls_btn_cancel_width: 85 [20230501-16:16:24] [DEBUG] ls_btn_cancel_height: 30 [20230501-16:16:24] [DEBUG] libxrdp_query_channel - Channel 0 name rdpdr [20230501-16:16:24] [DEBUG] xrdp_wm_init: channel rdpdr channel id 0 is enabled [20230501-16:16:24] [DEBUG] Enabling channel 1004 (rdpdr) [20230501-16:16:24] [DEBUG] libxrdp_query_channel - Channel 1 name rdpsnd [20230501-16:16:24] [DEBUG] xrdp_wm_init: channel rdpsnd channel id 1 is enabled [20230501-16:16:24] [DEBUG] Enabling channel 1005 (rdpsnd) [20230501-16:16:24] [DEBUG] libxrdp_query_channel - Channel 2 name cliprdr [20230501-16:16:24] [DEBUG] xrdp_wm_init: channel cliprdr channel id 2 is enabled [20230501-16:16:24] [DEBUG] Enabling channel 1006 (cliprdr) [20230501-16:16:24] [DEBUG] libxrdp_query_channel - Channel 3 name drdynvc [20230501-16:16:24] [DEBUG] xrdp_wm_init: channel drdynvc channel id 3 is enabled [20230501-16:16:24] [DEBUG] Enabling channel 1007 (drdynvc) [20230501-16:16:24] [DEBUG] Login state change request WMLS_USER_PROMPT -> WMLS_START_CONNECT [20230501-16:16:24] [DEBUG] out xrdp_wm_init: [20230501-16:16:24] [DEBUG] xrdp_wm_login_mode_changed: login_mode is 2 [20230501-16:16:24] [DEBUG] Login state change request WMLS_START_CONNECT -> WMLS_CONNECT_IN_PROGRESS [20230501-16:16:24] [DEBUG] xrdp_wm_log_msg: connecting to sesman on 127.0.0.1:3350 [20230501-16:16:24] [INFO ] connecting to sesman on 127.0.0.1:3350 [20230501-16:16:24] [INFO ] xrdp_wm_log_msg: sesman connect ok [20230501-16:16:24] [INFO ] sesman connect ok [20230501-16:16:24] [DEBUG] xrdp_wm_log_msg: sending login info to session manager. Please wait... [20230501-16:16:24] [INFO ] sending login info to session manager. Please wait... [20230501-16:16:24] [DEBUG] xrdp_wm_login_mode_changed: login_mode is 3 [20230501-16:16:24] [INFO ] xrdp_wm_log_msg: login successful for user user on display 10 [20230501-16:16:24] [INFO ] login successful for user user on display 10 [20230501-16:16:24] [INFO ] loaded module 'libxup.so' ok, interface size 10296, version 4 [20230501-16:16:24] [DEBUG] xrdp_wm_log_msg: started connecting [20230501-16:16:24] [INFO ] started connecting [20230501-16:16:24] [INFO ] lib_mod_connect: connecting via UNIX socket [20230501-16:16:28] [DEBUG] Closed socket 19 (AF_UNIX) [20230501-16:16:31] [DEBUG] Closed socket 19 (AF_UNIX) [20230501-16:16:35] [DEBUG] Closed socket 19 (AF_UNIX) [20230501-16:16:38] [DEBUG] Closed socket 19 (AF_UNIX) [20230501-16:16:42] [DEBUG] Closed socket 19 (AF_UNIX) [20230501-16:16:45] [DEBUG] Closed socket 19 (AF_UNIX) [20230501-16:16:49] [DEBUG] Closed socket 19 (AF_UNIX) [20230501-16:16:52] [DEBUG] Closed socket 19 (AF_UNIX) [20230501-16:16:56] [DEBUG] Closed socket 19 (AF_UNIX) [20230501-16:16:59] [DEBUG] Closed socket 19 (AF_UNIX) [20230501-16:17:03] [DEBUG] Closed socket 19 (AF_UNIX) [20230501-16:17:06] [DEBUG] Closed socket 19 (AF_UNIX) [20230501-16:17:10] [DEBUG] Closed socket 19 (AF_UNIX) [20230501-16:17:13] [DEBUG] Closed socket 19 (AF_UNIX) [20230501-16:17:17] [DEBUG] Closed socket 19 (AF_UNIX) [20230501-16:17:20] [DEBUG] Closed socket 19 (AF_UNIX) [20230501-16:17:24] [DEBUG] Closed socket 19 (AF_UNIX) [20230501-16:17:27] [DEBUG] Closed socket 19 (AF_UNIX) xrdp-sesman.log: [20230501-16:16:08] [INFO ] starting xrdp-sesman with pid 2453 [20230501-16:16:24] [INFO ] Socket 12: AF_INET6 connection received from ::1 port 39962 [20230501-16:16:24] [INFO ] Terminal Server Users group is disabled, allowing authentication [20230501-16:16:24] [INFO ] ++ created session (access granted): username user, ip ::ffff:192.168.0.82:49236 - socket: 12 [20230501-16:16:24] [INFO ] starting Xorg session... [20230501-16:16:24] [INFO ] Starting session: session_pid 2466, display :10.0, width 1456, height 818, bpp 24, client ip ::ffff:192.168.0.82:49236 - socket: 12, user name user [20230501-16:16:24] [INFO ] [session start] (display 10): calling auth_start_session from pid 2466 [20230501-16:16:24] [ERROR] sesman_data_in: scp_process_msg failed [20230501-16:16:24] [ERROR] sesman_main_loop: trans_check_wait_objs failed, removing trans [20230501-16:16:24] [INFO ] Starting X server on display 10: Xorg :10 -auth .Xauthority -config xrdp/xorg.conf -noreset -nolisten tcp -logfile .xorgxrdp.%s.log [20230501-16:16:34] [WARN ] Timed out waiting for X server on display 10 to startup [20230501-16:16:34] [INFO ] Session started successfully for user user on display 10 [20230501-16:16:34] [INFO ] Session in progress on display 10, waiting until the window manager (pid 2467) exits to end the session [20230501-16:16:34] [INFO ] Starting the xrdp channel server for display 10 [20230501-16:16:34] [WARN ] Timed out waiting for X server on display 10 to startup [20230501-16:16:34] [ERROR] There is no X server active on display 10 [20230501-16:16:34] [ERROR] A fatal error has occurred attempting to start the window manager on display 10, aborting connection [20230501-16:16:34] [WARN ] Window manager (pid 2467, display 10) exited quickly (0 secs). This could indicate a window manager config problem [20230501-16:16:34] [INFO ] Calling auth_stop_session and auth_end from pid 2466 [20230501-16:16:34] [INFO ] Terminating X server (pid 2468) on display 10 [20230501-16:16:34] [INFO ] Terminating the xrdp channel server (pid 2476) on display 10 [20230501-16:16:34] [INFO ] X server on display 10 (pid 2468) returned exit code 1 and signal number 0 [20230501-16:16:34] [INFO ] xrdp channel server for display 10 (pid 2476) exit code 0 and signal number 0 [20230501-16:16:34] [INFO ] cleanup_sockets: [20230501-16:16:34] [INFO ] Process 2466 has exited [20230501-16:16:34] [INFO ] ++ terminated session: username user, display :10.0, session_pid 2466, ip ::ffff:192.168.0.82:49236 - socket: 12 Thanks for the logs. This line is interesting:- Starting X server on display 10: Xorg :10 -auth .Xauthority -config xrdp/xorg.conf -noreset -nolisten tcp -logfile .xorgxrdp.%s.log The Xorg command should be qualified with a PATH, or you will run the wrapper script in /usr/bin/Xorg. On Mint I believe the command should be /usr/lib/xorg/Xorg Have a look in /etc/xrdp/sesman.ini and find the [Xorg] section. Add a path to the Xorg command, restart xrdp-sesman and see what happens. Now the log file gives this [20230502-18:00:30] [INFO ] Socket 12: AF_INET6 connection received from ::1 port 53048 [20230502-18:00:30] [DEBUG] session_get_bydata: search policy 0 U user W 1456 H 818 bpp 24 T 3 IP ::ffff:192.168.0.82:49323 - socket: 12 [20230502-18:00:30] [INFO ] Terminal Server Users group is disabled, allowing authentication [20230502-18:00:30] [INFO ] ++ created session (access granted): username user, ip ::ffff:192.168.0.82:49323 - socket: 12 [20230502-18:00:30] [INFO ] starting Xorg session... [20230502-18:00:30] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:30] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:30] [DEBUG] Closed socket 13 (AF_INET6 :: port 5910) [20230502-18:00:30] [DEBUG] Did not find a running X server at 5910 [20230502-18:00:30] [DEBUG] Closed socket 13 (AF_INET6 :: port 6010) [20230502-18:00:30] [DEBUG] Did not find a running X server at 6010 [20230502-18:00:30] [DEBUG] Closed socket 13 (AF_INET6 :: port 6210) [20230502-18:00:30] [DEBUG] Did not find a running X server at 6210 [20230502-18:00:30] [DEBUG] Did not find a running X server at /tmp/.xrdp/xrdp_chansrv_socket_10 [20230502-18:00:30] [DEBUG] Did not find a running X server at /tmp/.xrdp/xrdp_chansrv_audio_out_socket_10 [20230502-18:00:30] [DEBUG] Did not find a running X server at /tmp/.xrdp/xrdp_chansrv_audio_in_socket_10 [20230502-18:00:30] [DEBUG] Did not find a running X server at /tmp/.xrdp/xrdpapi_10 [20230502-18:00:30] [INFO ] Starting session: session_pid 1022, display :10.0, width 1456, height 818, bpp 24, client ip ::ffff:192.168.0.82:49323 - socket: 12, user name user [20230502-18:00:30] [INFO ] [session start] (display 10): calling auth_start_session from pid 1022 [20230502-18:00:30] [ERROR] sesman_data_in: scp_process_msg failed [20230502-18:00:30] [ERROR] sesman_main_loop: trans_check_wait_objs failed, removing trans [20230502-18:00:30] [DEBUG] Closed socket 11 (AF_INET6 ::1 port 3350) [20230502-18:00:30] [DEBUG] Closed socket 12 (AF_INET6 ::1 port 3350) [20230502-18:00:30] [DEBUG] Closed socket 12 (AF_INET6 ::1 port 3350) [20230502-18:00:30] [DEBUG] Waiting for X server to start on display 10 [20230502-18:00:30] [DEBUG] Waiting for X server to start on display 10 [20230502-18:00:30] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:30] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:30] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:30] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:30] [INFO ] Starting X server on display 10: Xorg :10 -auth .Xauthority -config xrdp/xorg.conf -noreset -nolisten tcp -logfile .xorgxrdp.%s.log [20230502-18:00:30] [DEBUG] Calling exec (executable: Xorg, arguments: Xorg :10 -auth .Xauthority -config xrdp/xorg.conf -noreset -nolisten tcp -logfile .xorgxrdp.%s.log) [20230502-18:00:30] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:30] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:31] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:31] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:31] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:31] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:31] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:31] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:31] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:31] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:31] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:31] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:31] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:31] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:31] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:31] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:31] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:31] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:32] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:32] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:32] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:32] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:32] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:32] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:32] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:32] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:32] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:32] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:32] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:32] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:32] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:32] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:32] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:32] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:33] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:33] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:33] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:33] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:33] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:33] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:33] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:33] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:33] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:33] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:33] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:33] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:33] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:33] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:34] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:34] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:34] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:34] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:34] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:34] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:34] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:34] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:34] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:34] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:34] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:34] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:34] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:34] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:35] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:35] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:35] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:35] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:35] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:35] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:35] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:35] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:35] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:35] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:35] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:35] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:35] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:35] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:35] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:35] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:36] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:36] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:36] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:36] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:36] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:36] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:36] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:36] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:36] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:36] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:36] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:36] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:36] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:36] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:36] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:36] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:37] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:37] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:37] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:37] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:37] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:37] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:37] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:37] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:37] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:37] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:37] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:37] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:37] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:37] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:37] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:37] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:38] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:38] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:38] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:38] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:38] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:38] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:38] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:38] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:38] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:38] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:38] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:38] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:38] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:38] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:38] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:38] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:39] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:39] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:39] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:39] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:39] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:39] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:39] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:39] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:39] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:39] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:39] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:39] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:39] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:39] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:39] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:39] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:40] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:40] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:40] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:40] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:40] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:40] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:40] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:40] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:40] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:40] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:40] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:40] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:40] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:40] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:40] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:40] [WARN ] Timed out waiting for X server on display 10 to startup [20230502-18:00:40] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:40] [INFO ] Session started successfully for user user on display 10 [20230502-18:00:40] [INFO ] Starting the xrdp channel server for display 10 [20230502-18:00:40] [INFO ] Session in progress on display 10, waiting until the window manager (pid 1023) exits to end the session [20230502-18:00:40] [DEBUG] waiting for pid 1023 to exit [20230502-18:00:40] [DEBUG] Calling exec (executable: /usr/sbin/xrdp-chansrv, arguments: /usr/sbin/xrdp-chansrv) [20230502-18:00:41] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:41] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:41] [WARN ] Timed out waiting for X server on display 10 to startup [20230502-18:00:41] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:41] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:41] [ERROR] There is no X server active on display 10 [20230502-18:00:41] [ERROR] A fatal error has occurred attempting to start the window manager on display 10, aborting connection [20230502-18:00:41] [WARN ] Window manager (pid 1023, display 10) exited quickly (1 secs). This could indicate a window manager config problem [20230502-18:00:41] [INFO ] Calling auth_stop_session and auth_end from pid 1022 [20230502-18:00:41] [INFO ] Terminating X server (pid 1024) on display 10 [20230502-18:00:41] [INFO ] Terminating the xrdp channel server (pid 1026) on display 10 [20230502-18:00:41] [DEBUG] waiting for pid 1024 to exit [20230502-18:00:41] [INFO ] X server on display 10 (pid 1024) returned exit code 1 and signal number 0 [20230502-18:00:41] [DEBUG] waiting for pid 1026 to exit [20230502-18:00:41] [INFO ] xrdp channel server for display 10 (pid 1026) exit code 0 and signal number 0 [20230502-18:00:41] [INFO ] cleanup_sockets: [20230502-18:00:41] [DEBUG] receiving SIGCHLD [20230502-18:00:41] [INFO ] Process 1022 has exited [20230502-18:00:41] [INFO ] ++ terminated session: username user, display :10.0, session_pid 1022, ip ::ffff:192.168.0.82:49323 - socket: 12 I checked a few more options. Here is the full result of my attempts: /etc/xrdp/xrdp.ini [Globals] ; xrdp.ini file version number ini_version=1 ; fork a new process for each incoming connection fork=true ; ports to listen on, number alone means listen on all interfaces ; 0.0.0.0 or :: if ipv6 is configured ; space between multiple occurrences ; ALL specified interfaces must be UP when xrdp starts, otherwise xrdp will fail to start ; ; Examples: ; port=3389 ; port=unix://./tmp/xrdp.socket ; port=tcp://.:3389 127.0.0.1:3389 ; port=tcp://:3389 *:3389 ; port=tcp://<any ipv4 format addr>:3389 192.168.1.1:3389 ; port=tcp6://.:3389 ::1:3389 ; port=tcp6://:3389 *:3389 ; port=tcp6://{<any ipv6 format addr>}:3389 {FC00:0:0:0:0:0:0:1}:3389 ; port=vsock://<cid>:<port> port=3389 ; 'port' above should be connected to with vsock instead of tcp ; use this only with number alone in port above ; prefer use vsock://<cid>:<port> above use_vsock=false ; regulate if the listening socket use socket option tcp_nodelay ; no buffering will be performed in the TCP stack tcp_nodelay=true ; regulate if the listening socket use socket option keepalive ; if the network connection disappear without close messages the connection will be closed tcp_keepalive=true ; set tcp send/recv buffer (for experts) #tcp_send_buffer_bytes=32768 #tcp_recv_buffer_bytes=32768 ; security layer can be 'tls', 'rdp' or 'negotiate' ; for client compatible layer security_layer=negotiate ; minimum security level allowed for client for classic RDP encryption ; use tls_ciphers to configure TLS encryption ; can be 'none', 'low', 'medium', 'high', 'fips' crypt_level=high #crypt_level=none ; X.509 certificate and private key ; openssl req -x509 -newkey rsa:2048 -nodes -keyout key.pem -out cert.pem -days 365 certificate= key_file= ; set SSL protocols ; can be comma separated list of 'SSLv3', 'TLSv1', 'TLSv1.1', 'TLSv1.2', 'TLSv1.3' ssl_protocols=TLSv1.2, TLSv1.3 ; set TLS cipher suites #tls_ciphers=HIGH ; concats the domain name to the user if set for authentication with the separator ; for example when the server is multi homed with SSSd #domain_user_separator=@ ; The following options will override the keyboard layout settings. ; These options are for DEBUG and are not recommended for regular use. #xrdp.override_keyboard_type=0x04 #xrdp.override_keyboard_subtype=0x01 #xrdp.override_keylayout=0x00000409 ; Section name to use for automatic login if the client sends username ; and password. If empty, the domain name sent by the client is used. ; If empty and no domain name is given, the first suitable section in ; this file will be used. autorun= allow_channels=true allow_multimon=true bitmap_cache=true bitmap_compression=true bulk_compression=true #hidelogwindow=true max_bpp=32 new_cursors=true ; fastpath - can be 'input', 'output', 'both', 'none' use_fastpath=both ; when true, userid/password *must* be passed on cmd line #require_credentials=true ; when true, the userid will be used to try to authenticate #enable_token_login=true ; You can set the PAM error text in a gateway setup (MAX 256 chars) #pamerrortxt=change your password according to policy at http://url ; ; colors used by windows in RGB format ; blue=009cb5 grey=dedede #black=000000 #dark_grey=808080 #blue=08246b #dark_blue=08246b #white=ffffff #red=ff0000 #green=00ff00 #background=626c72 ; ; configure login screen ; ; Login Screen Window Title #ls_title=My Login Title ; top level window background color in RGB format ls_top_window_bg_color=009cb5 ; width and height of login screen ; ; The default height allows for about 5 fields to be comfortably displayed ; above the buttons at the bottom. To display more fields, make <ls_height> ; larger, and also increase <ls_btn_ok_y_pos> and <ls_btn_cancel_y_pos> ; below ; ls_width=350 ls_height=430 ; login screen background color in RGB format ls_bg_color=dedede ; optional background image filename. BMP format is always supported, ; but other formats will be supported if xrdp is build with imlib2 ; The transform can be one of the following:- ; none : No transformation. Image is placed in bottom-right corner ; of the screen. ; scale : Image is scaled to the screen size. The image aspect ; ratio is not preserved. ; zoom : Image is scaled to the screen size. The image aspect ; ratio is preserved by clipping the image. #ls_background_image= #ls_background_transform=none ; logo ; full path to file or file in shared folder. BMP format is always supported, ; but other formats will be supported if xrdp is build with imlib2 ; For transform values, see 'ls_background_transform'. The logo width and ; logo height are ignored for a transform of 'none'. ls_logo_filename= #ls_logo_transform=none #ls_logo_width=240 #ls_logo_height=140 ls_logo_x_pos=55 ls_logo_y_pos=50 ; for positioning labels such as username, password etc ls_label_x_pos=30 ls_label_width=65 ; for positioning text and combo boxes next to above labels ls_input_x_pos=110 ls_input_width=210 ; y pos for first label and combo box ls_input_y_pos=220 ; OK button ls_btn_ok_x_pos=142 ls_btn_ok_y_pos=370 ls_btn_ok_width=85 ls_btn_ok_height=30 ; Cancel button ls_btn_cancel_x_pos=237 ls_btn_cancel_y_pos=370 ls_btn_cancel_width=85 ls_btn_cancel_height=30 [Logging] ; Note: Log levels can be any of: core, error, warning, info, debug, or trace LogFile=xrdp.log LogLevel=debug EnableSyslog=true #SyslogLevel=INFO #EnableConsole=false #ConsoleLevel=INFO #EnableProcessId=false [LoggingPerLogger] ; Note: per logger configuration is only used if xrdp is built with ; --enable-devel-logging #xrdp.c=INFO #main()=INFO [Channels] ; Channel names not listed here will be blocked by XRDP. ; You can block any channel by setting its value to false. ; IMPORTANT! All channels are not supported in all use ; cases even if you set all values to true. ; You can override these settings on each session type ; These settings are only used if allow_channels=true rdpdr=true rdpsnd=true drdynvc=true cliprdr=true rail=true xrdpvr=true tcutils=true ; for debugging xrdp, in section xrdp1, change port=-1 to this: #port=/tmp/.xrdp/xrdp_display_10 ; ; Session types ; ; Some session types such as Xorg, X11rdp and Xvnc start a display server. ; Startup command-line parameters for the display server are configured ; in sesman.ini. See and configure also sesman.ini. [Xorg] name=Xorg lib=libxup.so username=ask password=ask ip=127.0.0.1 port=-1 code=20 [Xvnc] name=Xvnc lib=libvnc.so username=ask password=ask ip=127.0.0.1 port=-1 #xserverbpp=24 #delay_ms=2000 ; Disable requested encodings to support buggy VNC servers ; (1 = ExtendedDesktopSize) #disabled_encodings_mask=0 ; Use this to connect to a chansrv instance created outside of sesman ; (e.g. as part of an x11vnc console session). Replace '0' with the ; display number of the session #chansrvport=DISPLAY(0) ; Generic VNC Proxy ; Tailor this to specific hosts and VNC instances by specifying an ip ; and port and setting a suitable name. [vnc-any] name=vnc-any lib=libvnc.so ip=ask port=ask5900 username=na password=ask #pamusername=asksame #pampassword=asksame #pamsessionmng=127.0.0.1 #delay_ms=2000 ; Generic RDP proxy using NeutrinoRDP ; Tailor this to specific hosts by specifying an ip and port and setting ; a suitable name. [neutrinordp-any] name=neutrinordp-any ; To use this section, you should build xrdp with configure option ; --enable-neutrinordp. lib=libxrdpneutrinordp.so ip=ask port=ask3389 username=ask password=ask ; Uncomment the following lines to enable PAM authentication for proxy ; connections. #pamusername=ask #pampassword=ask #pamsessionmng=127.0.0.1 ; Currently NeutrinoRDP doesn't support dynamic resizing. Uncomment ; this line if you're using a client which does. #enable_dynamic_resizing=false ; By default, performance settings requested by the RDP client are ignored ; and chosen by NeutrinoRDP. Uncomment this line to allow the user to ; select performance settings in the RDP client. #perf.allow_client_experiencesettings=true ; Override any experience setting by uncommenting one or more of the ; following lines. #perf.wallpaper=false #perf.font_smoothing=false #perf.desktop_composition=false #perf.full_window_drag=false #perf.menu_anims=false #perf.themes=false #perf.cursor_blink=false ; By default NeutrinoRDP supports cursor shadows. If this is giving ; you problems (e.g. cursor is a black rectangle) try disabling cursor ; shadows by uncommenting the following line. #perf.cursor_shadow=false ; By default, NeutrinoRDP uses the keyboard layout of the remote RDP Server. ; If you want to tell the remote the keyboard layout of the RDP Client, ; by uncommenting the following line. #neutrinordp.allow_client_keyboardLayout=true ; The following options will override the remote keyboard layout settings. ; These options are for DEBUG and are not recommended for regular use. #neutrinordp.override_keyboardLayout_mask=0x0000FFFF #neutrinordp.override_kbd_type=0x04 #neutrinordp.override_kbd_subtype=0x01 #neutrinordp.override_kbd_fn_keys=12 #neutrinordp.override_kbd_layout=0x00000409 ; You can override the common channel settings for each session type #channel.rdpdr=true #channel.rdpsnd=true #channel.drdynvc=true #channel.cliprdr=true #channel.rail=true #channel.xrdpvr=true /etc/xrdp/sesman.ini ;; See `man 5 sesman.ini` for details [Globals] ListenAddress=127.0.0.1 ListenPort=3350 EnableUserWindowManager=true ; Give in relative path to user's home directory UserWindowManager=startwm.sh ; Give in full path or relative path to /etc/xrdp DefaultWindowManager=startwm.sh ; Give in full path or relative path to /etc/xrdp ReconnectScript=reconnectwm.sh [Security] AllowRootLogin=true MaxLoginRetry=4 TerminalServerUsers=tsusers TerminalServerAdmins=tsadmins ; When AlwaysGroupCheck=false access will be permitted ; if the group TerminalServerUsers is not defined. AlwaysGroupCheck=false ; When RestrictOutboundClipboard=all clipboard from the ; server is not pushed to the client. ; In addition, you can control text/file/image transfer restrictions ; respectively. It also accepts comma separated list such as text,file,image. ; To keep compatibility, some aliases are also available: ; true: an alias of all ; false: an alias of none ; yes: an alias of all RestrictOutboundClipboard=none ; When RestrictInboundClipboard=all clipboard from the ; client is not pushed to the server. ; In addition, you can control text/file/image transfer restrictions ; respectively. It also accepts comma separated list such as text,file,image. ; To keep compatibility, some aliases are also available: ; true: an alias of all ; false: an alias of none ; yes: an alias of all RestrictInboundClipboard=none [Sessions] ;; X11DisplayOffset - x11 display number offset ; Type: integer ; Default: 10 X11DisplayOffset=10 ;; MaxSessions - maximum number of connections to an xrdp server ; Type: integer ; Default: 0 MaxSessions=50 ;; KillDisconnected - kill disconnected sessions ; Type: boolean ; Default: false ; if 1, true, or yes, every session will be killed within DisconnectedTimeLimit ; seconds after the user disconnects KillDisconnected=false ;; DisconnectedTimeLimit (seconds) - wait before kill disconnected sessions ; Type: integer ; Default: 0 ; if KillDisconnected is set to false, this value is ignored DisconnectedTimeLimit=0 ;; IdleTimeLimit (seconds) - wait before disconnect idle sessions ; Type: integer ; Default: 0 ; Set to 0 to disable idle disconnection. IdleTimeLimit=0 ;; Policy - session allocation policy ; Type: enum [ "Default" | "UBD" | "UBI" | "UBC" | "UBDI" | "UBDC" ] ; "Default" session per <User,BitPerPixel> ; "UBD" session per <User,BitPerPixel,DisplaySize> ; "UBI" session per <User,BitPerPixel,IPAddr> ; "UBC" session per <User,BitPerPixel,Connection> ; "UBDI" session per <User,BitPerPixel,DisplaySize,IPAddr> ; "UBDC" session per <User,BitPerPixel,DisplaySize,Connection> Policy=Default [Logging] ; Note: Log levels can be any of: core, error, warning, info, debug, or trace LogFile=xrdp-sesman.log LogLevel=debug EnableSyslog=true #SyslogLevel=INFO #EnableConsole=false #ConsoleLevel=INFO #EnableProcessId=false [LoggingPerLogger] ; Note: per logger configuration is only used if xrdp is built with ; --enable-devel-logging #sesman.c=INFO #main()=INFO ; ; Session definitions - startup command-line parameters for each session type ; [Xorg] ; Specify the path of non-suid Xorg executable. It might differ depending ; on your distribution and version. Find out the appropriate path for your ; environment. The typical path is known as follows: ; ; Fedora 26 or later : param=/usr/libexec/Xorg ; Debian 9 or later : param=/usr/lib/xorg/Xorg ; Ubuntu 16.04 or later : param=/usr/lib/xorg/Xorg ; Arch Linux : param=/usr/lib/Xorg ; CentOS 7 : param=/usr/bin/Xorg or param=Xorg ; CentOS 8 : param=/usr/libexec/Xorg ; FreeBSD (from 2022Q4) : param=/usr/local/libexec/Xorg ; #param=Xorg param=/usr/bin/Xorg ; Leave the rest parameters as-is unless you understand what will happen. param=-config param=xrdp/xorg.conf param=-noreset param=-nolisten param=tcp param=-logfile param=.xorgxrdp.%s.log [Xvnc] param=Xvnc param=-bs param=-nolisten param=tcp param=-localhost param=-dpi param=96 [Chansrv] ; drive redirection ; See sesman.ini(5) for the format of this parameter #FuseMountName=/run/user/%u/thinclient_drives #FuseMountName=/media/thinclient_drives/%U/thinclient_drives FuseMountName=thinclient_drives ; this value allows only the user to access their own mapped drives. ; Make this more permissive (e.g. 022) if required. FileUmask=077 ; Can be used to disable FUSE functionality - see sesman.ini(5) #EnableFuseMount=false ; Uncomment this line only if you are using GNOME 3 versions 3.29.92 ; and up, and you wish to cut-paste files between Nautilus and Windows. Do ; not use this setting for GNOME 4, or other file managers #UseNautilus3FlistFormat=true [ChansrvLogging] ; Note: one log file is created per display and the LogFile config value ; is ignored. The channel server log file names follow the naming convention: ; xrdp-chansrv.${DISPLAY}.log ; ; Note: Log levels can be any of: core, error, warning, info, debug, or trace LogLevel=INFO EnableSyslog=true #SyslogLevel=INFO #EnableConsole=false #ConsoleLevel=INFO #EnableProcessId=false [ChansrvLoggingPerLogger] ; Note: per logger configuration is only used if xrdp is built with ; --enable-devel-logging #chansrv.c=INFO #main()=INFO [SessionVariables] PULSE_SCRIPT=/etc/xrdp/pulse/default.pa xrdp.log [20230502-22:32:22] [INFO ] Socket 12: AF_INET6 connection received from ::ffff:127.0.0.1 port 46378 [20230502-22:32:22] [DEBUG] Closed socket 12 (AF_INET6 ::ffff:127.0.0.1 port 3389) [20230502-22:32:22] [DEBUG] Closed socket 11 (AF_INET6 :: port 3389) [20230502-22:32:22] [DEBUG] item ini_version, value 1 [20230502-22:32:23] [DEBUG] item fork, value true [20230502-22:32:23] [DEBUG] item port, value 3389 [20230502-22:32:23] [DEBUG] item use_vsock, value false [20230502-22:32:23] [DEBUG] item tcp_nodelay, value true [20230502-22:32:23] [DEBUG] item tcp_keepalive, value true [20230502-22:32:23] [DEBUG] item security_layer, value negotiate [20230502-22:32:23] [DEBUG] item crypt_level, value high [20230502-22:32:23] [DEBUG] item certificate, value [20230502-22:32:23] [INFO ] Using default X.509 certificate: /etc/xrdp/cert.pem [20230502-22:32:23] [DEBUG] item key_file, value [20230502-22:32:23] [INFO ] Using default X.509 key file: /etc/xrdp/key.pem [20230502-22:32:23] [DEBUG] item ssl_protocols, value TLSv1.2, TLSv1.3 [20230502-22:32:23] [DEBUG] TLSv1.3 enabled [20230502-22:32:23] [DEBUG] TLSv1.2 enabled [20230502-22:32:23] [DEBUG] item autorun, value [20230502-22:32:23] [DEBUG] item allow_channels, value true [20230502-22:32:23] [DEBUG] item allow_multimon, value true [20230502-22:32:23] [DEBUG] item bitmap_cache, value true [20230502-22:32:23] [DEBUG] item bitmap_compression, value true [20230502-22:32:23] [DEBUG] item bulk_compression, value true [20230502-22:32:23] [DEBUG] item max_bpp, value 32 [20230502-22:32:23] [DEBUG] item new_cursors, value true [20230502-22:32:23] [DEBUG] item use_fastpath, value both [20230502-22:32:23] [DEBUG] item blue, value 009cb5 [20230502-22:32:23] [DEBUG] item grey, value dedede [20230502-22:32:23] [DEBUG] item ls_top_window_bg_color, value 009cb5 [20230502-22:32:23] [DEBUG] item ls_width, value 350 [20230502-22:32:23] [DEBUG] item ls_height, value 430 [20230502-22:32:23] [DEBUG] item ls_bg_color, value dedede [20230502-22:32:23] [DEBUG] item ls_logo_filename, value [20230502-22:32:23] [DEBUG] item ls_logo_x_pos, value 55 [20230502-22:32:23] [DEBUG] item ls_logo_y_pos, value 50 [20230502-22:32:23] [DEBUG] item ls_label_x_pos, value 30 [20230502-22:32:23] [DEBUG] item ls_label_width, value 65 [20230502-22:32:23] [DEBUG] item ls_input_x_pos, value 110 [20230502-22:32:23] [DEBUG] item ls_input_width, value 210 [20230502-22:32:23] [DEBUG] item ls_input_y_pos, value 220 [20230502-22:32:23] [DEBUG] item ls_btn_ok_x_pos, value 142 [20230502-22:32:23] [DEBUG] item ls_btn_ok_y_pos, value 370 [20230502-22:32:23] [DEBUG] item ls_btn_ok_width, value 85 [20230502-22:32:23] [DEBUG] item ls_btn_ok_height, value 30 [20230502-22:32:23] [DEBUG] item ls_btn_cancel_x_pos, value 237 [20230502-22:32:23] [DEBUG] item ls_btn_cancel_y_pos, value 370 [20230502-22:32:23] [DEBUG] item ls_btn_cancel_width, value 85 [20230502-22:32:23] [DEBUG] item ls_btn_cancel_height, value 30 [20230502-22:32:23] [INFO ] Security protocol: configured [SSL|RDP], requested [SSL|HYBRID|RDP], selected [SSL] [20230502-22:32:23] [DEBUG] Using TLS security, and setting RDP security crypto to LEVEL_NONE and METHOD_NONE [20230502-22:32:23] [DEBUG] [MCS Connection Sequence] receive connection request [20230502-22:32:23] [INFO ] Connected client computer name: solus [20230502-22:32:23] [DEBUG] The connection is using TLS, skipping RDP crypto negotiation [20230502-22:32:23] [DEBUG] Adding channel: name rdpdr, channel id 1004, flags 0xc0800000 [20230502-22:32:23] [DEBUG] Adding channel: name rdpsnd, channel id 1005, flags 0xc0000000 [20230502-22:32:23] [DEBUG] Adding channel: name cliprdr, channel id 1006, flags 0xc0a00000 [20230502-22:32:23] [DEBUG] Adding channel: name drdynvc, channel id 1007, flags 0xc0800000 [20230502-22:32:23] [WARN ] Received [MS-RDPBCGR] TS_UD_HEADER type 0xc006 is unknown (ignored) [20230502-22:32:23] [WARN ] Received [MS-RDPBCGR] TS_UD_HEADER type 0xc00a is unknown (ignored) [20230502-22:32:23] [DEBUG] [MCS Connection Sequence] construct connection response [20230502-22:32:23] [DEBUG] using no security [20230502-22:32:23] [DEBUG] [MCS Connection Sequence] send connection response [20230502-22:32:23] [DEBUG] [MCS Connection Sequence] receive erect domain request [20230502-22:32:23] [DEBUG] [MCS Connection Sequence] receive attach user request [20230502-22:32:23] [DEBUG] [MCS Connection Sequence] send attach user confirm [20230502-22:32:23] [DEBUG] [MCS Connection Sequence (TLS)] receive channel join request [20230502-22:32:23] [DEBUG] [MCS Connection Sequence (TLS)] send channel join confirm [20230502-22:32:23] [DEBUG] [MCS Connection Sequence (TLS)] receive channel join request [20230502-22:32:23] [DEBUG] [MCS Connection Sequence (TLS)] send channel join confirm [20230502-22:32:23] [DEBUG] [MCS Connection Sequence (TLS)] receive channel join request [20230502-22:32:23] [DEBUG] [MCS Connection Sequence (TLS)] send channel join confirm [20230502-22:32:23] [DEBUG] [MCS Connection Sequence (TLS)] receive channel join request [20230502-22:32:23] [DEBUG] [MCS Connection Sequence (TLS)] send channel join confirm [20230502-22:32:23] [DEBUG] [MCS Connection Sequence (TLS)] receive channel join request [20230502-22:32:23] [DEBUG] [MCS Connection Sequence (TLS)] send channel join confirm [20230502-22:32:23] [DEBUG] [MCS Connection Sequence (TLS)] receive channel join request [20230502-22:32:23] [DEBUG] [MCS Connection Sequence (TLS)] send channel join confirm [20230502-22:32:23] [DEBUG] [MCS Connection Sequence (TLS)] completed [20230502-22:32:23] [INFO ] xrdp_load_keyboard_layout: Keyboard information sent by the RDP client, keyboard_type:[0x04], keyboard_subtype:[0x00], keylayout:[0x00000409] [20230502-22:32:23] [DEBUG] keyboard_cfg_file /etc/xrdp/xrdp_keyboard.ini [20230502-22:32:23] [DEBUG] xrdp_load_keyboard_layout: item rdp_layout_us value 0x00000409 [20230502-22:32:23] [DEBUG] xrdp_load_keyboard_layout: skipping configuration item - rdp_layout_us, continuing to next section [20230502-22:32:23] [DEBUG] xrdp_load_keyboard_layout: item rdp_layout_us value us [20230502-22:32:23] [DEBUG] xrdp_load_keyboard_layout: skipping configuration item - rdp_layout_us, continuing to next section [20230502-22:32:23] [DEBUG] xrdp_load_keyboard_layout: item keyboard_type value 4 [20230502-22:32:23] [DEBUG] xrdp_load_keyboard_layout: item keyboard_subtype value 3 [20230502-22:32:23] [DEBUG] xrdp_load_keyboard_layout: item keyboard_type value 7 [20230502-22:32:23] [DEBUG] xrdp_load_keyboard_layout: item keyboard_subtype value 2 [20230502-22:32:23] [DEBUG] xrdp_load_keyboard_layout: item model value pc105 [20230502-22:32:23] [DEBUG] xrdp_load_keyboard_layout: item rdp_layouts value default_rdp_layouts [20230502-22:32:23] [DEBUG] xrdp_load_keyboard_layout: item layouts_map value default_layouts_map [20230502-22:32:23] [DEBUG] xrdp_load_keyboard_layout: item rdp_layout_us value us [20230502-22:32:23] [DEBUG] xrdp_load_keyboard_layout: skipping configuration item - rdp_layout_us, continuing to next section [20230502-22:32:23] [INFO ] xrdp_load_keyboard_layout: model [] variant [] layout [us] options [] [20230502-22:32:23] [INFO ] TLS connection established from ::ffff:127.0.0.1 port 46378: TLSv1.3 with cipher TLS_AES_256_GCM_SHA384 [20230502-22:32:23] [DEBUG] Client requested auto logon. [20230502-22:32:23] [DEBUG] Client requested compression enabled. [20230502-22:32:23] [DEBUG] Client supplied domain: [20230502-22:32:23] [DEBUG] Client supplied username: alexander [20230502-22:32:23] [DEBUG] Client supplied password: <omitted from log> [20230502-22:32:23] [DEBUG] Client supplied program: [20230502-22:32:23] [DEBUG] Client supplied directory: [20230502-22:32:24] [INFO ] xrdp_caps_process_pointer: client supports new(color) cursor [20230502-22:32:24] [INFO ] xrdp_process_offscreen_bmpcache: support level 1 cache size 7864320 MB cache entries 2000 [20230502-22:32:24] [INFO ] xrdp_caps_process_codecs: RemoteFX, codec id 3, properties len 49 [20230502-22:32:24] [WARN ] Client Capability: not enough orders supported by client, client wants off screen bitmap but offscreen bitmaps disabled [20230502-22:32:24] [DEBUG] xrdp_00000738_wm_login_state_event_00000001 [20230502-22:32:24] [INFO ] Loading keymap file /etc/xrdp/km-00000409.ini [20230502-22:32:24] [WARN ] local keymap file for 0x00000409 found and doesn't match built in keymap, using local keymap file [20230502-22:32:24] [DEBUG] Login state change request WMLS_RESET -> WMLS_RESET [20230502-22:32:24] [DEBUG] xrdp_wm_login_mode_changed: login_mode is 0 [20230502-22:32:24] [DEBUG] Login state change request WMLS_RESET -> WMLS_USER_PROMPT [20230502-22:32:24] [DEBUG] in xrdp_wm_init: [20230502-22:32:24] [DEBUG] ini_version: 1 [20230502-22:32:24] [DEBUG] use_bitmap_cache: 1 [20230502-22:32:24] [DEBUG] use_bitmap_compression: 1 [20230502-22:32:24] [DEBUG] port: 3389 [20230502-22:32:24] [DEBUG] crypt_level: 3 [20230502-22:32:24] [DEBUG] allow_channels: 1 [20230502-22:32:24] [DEBUG] max_bpp: 32 [20230502-22:32:24] [DEBUG] fork: 1 [20230502-22:32:24] [DEBUG] tcp_nodelay: 1 [20230502-22:32:24] [DEBUG] tcp_keepalive: 1 [20230502-22:32:24] [DEBUG] tcp_send_buffer_bytes: 0 [20230502-22:32:24] [DEBUG] tcp_recv_buffer_bytes: 0 [20230502-22:32:24] [DEBUG] new_cursors: 1 [20230502-22:32:24] [DEBUG] allow_multimon: 1 [20230502-22:32:24] [DEBUG] grey: 14606046 [20230502-22:32:24] [DEBUG] black: 0 [20230502-22:32:24] [DEBUG] dark_grey: 0 [20230502-22:32:24] [DEBUG] blue: 40117 [20230502-22:32:24] [DEBUG] dark_blue: 0 [20230502-22:32:24] [DEBUG] white: 0 [20230502-22:32:24] [DEBUG] red: 0 [20230502-22:32:24] [DEBUG] green: 0 [20230502-22:32:24] [DEBUG] background: 0 [20230502-22:32:24] [DEBUG] autorun: [20230502-22:32:24] [DEBUG] hidelogwindow: 0 [20230502-22:32:24] [DEBUG] require_credentials: 0 [20230502-22:32:24] [DEBUG] bulk_compression: 1 [20230502-22:32:24] [DEBUG] new_cursors: 1 [20230502-22:32:24] [DEBUG] nego_sec_layer: 0 [20230502-22:32:24] [DEBUG] allow_multimon: 1 [20230502-22:32:24] [DEBUG] enable_token_login: 0 [20230502-22:32:24] [DEBUG] ls_top_window_bg_color: b59c00 [20230502-22:32:24] [DEBUG] ls_width: 350 [20230502-22:32:24] [DEBUG] ls_height: 430 [20230502-22:32:24] [DEBUG] ls_bg_color: dedede [20230502-22:32:24] [DEBUG] ls_title: [20230502-22:32:24] [DEBUG] ls_logo_filename: [20230502-22:32:24] [DEBUG] ls_logo_x_pos: 55 [20230502-22:32:24] [DEBUG] ls_logo_y_pos: 50 [20230502-22:32:24] [DEBUG] ls_label_x_pos: 30 [20230502-22:32:24] [DEBUG] ls_label_width: 65 [20230502-22:32:24] [DEBUG] ls_input_x_pos: 110 [20230502-22:32:24] [DEBUG] ls_input_width: 210 [20230502-22:32:24] [DEBUG] ls_input_y_pos: 220 [20230502-22:32:24] [DEBUG] ls_btn_ok_x_pos: 142 [20230502-22:32:24] [DEBUG] ls_btn_ok_y_pos: 370 [20230502-22:32:24] [DEBUG] ls_btn_ok_width: 85 [20230502-22:32:24] [DEBUG] ls_btn_ok_height: 30 [20230502-22:32:24] [DEBUG] ls_btn_cancel_x_pos: 237 [20230502-22:32:24] [DEBUG] ls_btn_cancel_y_pos: 370 [20230502-22:32:24] [DEBUG] ls_btn_cancel_width: 85 [20230502-22:32:24] [DEBUG] ls_btn_cancel_height: 30 [20230502-22:32:24] [DEBUG] libxrdp_query_channel - Channel 0 name rdpdr [20230502-22:32:24] [DEBUG] xrdp_wm_init: channel rdpdr channel id 0 is enabled [20230502-22:32:24] [DEBUG] Enabling channel 1004 (rdpdr) [20230502-22:32:24] [DEBUG] libxrdp_query_channel - Channel 1 name rdpsnd [20230502-22:32:24] [DEBUG] xrdp_wm_init: channel rdpsnd channel id 1 is enabled [20230502-22:32:24] [DEBUG] Enabling channel 1005 (rdpsnd) [20230502-22:32:24] [DEBUG] libxrdp_query_channel - Channel 2 name cliprdr [20230502-22:32:24] [DEBUG] xrdp_wm_init: channel cliprdr channel id 2 is enabled [20230502-22:32:24] [DEBUG] Enabling channel 1006 (cliprdr) [20230502-22:32:24] [DEBUG] libxrdp_query_channel - Channel 3 name drdynvc [20230502-22:32:24] [DEBUG] xrdp_wm_init: channel drdynvc channel id 3 is enabled [20230502-22:32:24] [DEBUG] Enabling channel 1007 (drdynvc) [20230502-22:32:24] [DEBUG] Login state change request WMLS_USER_PROMPT -> WMLS_START_CONNECT [20230502-22:32:24] [DEBUG] out xrdp_wm_init: [20230502-22:32:24] [DEBUG] xrdp_wm_login_mode_changed: login_mode is 2 [20230502-22:32:24] [DEBUG] Login state change request WMLS_START_CONNECT -> WMLS_CONNECT_IN_PROGRESS [20230502-22:32:24] [DEBUG] xrdp_wm_log_msg: connecting to sesman on 127.0.0.1:3350 [20230502-22:32:24] [INFO ] connecting to sesman on 127.0.0.1:3350 [20230502-22:32:24] [INFO ] xrdp_wm_log_msg: sesman connect ok [20230502-22:32:24] [INFO ] sesman connect ok [20230502-22:32:24] [DEBUG] xrdp_wm_log_msg: sending login info to session manager. Please wait... [20230502-22:32:24] [INFO ] sending login info to session manager. Please wait... [20230502-22:32:24] [DEBUG] xrdp_wm_login_mode_changed: login_mode is 3 [20230502-22:32:24] [INFO ] xrdp_wm_log_msg: login successful for user alexander on display 10 [20230502-22:32:24] [INFO ] login successful for user alexander on display 10 [20230502-22:32:24] [INFO ] loaded module 'libxup.so' ok, interface size 10296, version 4 [20230502-22:32:24] [DEBUG] xrdp_wm_log_msg: started connecting [20230502-22:32:24] [INFO ] started connecting [20230502-22:32:24] [INFO ] lib_mod_connect: connecting via UNIX socket [20230502-22:32:27] [DEBUG] Closed socket 19 (AF_UNIX) [20230502-22:32:31] [DEBUG] Closed socket 19 (AF_UNIX) [20230502-22:32:34] [DEBUG] Closed socket 19 (AF_UNIX) [20230502-22:32:38] [DEBUG] Closed socket 19 (AF_UNIX) xrdp-sesman.log [20230502-22:32:24] [INFO ] Socket 12: AF_INET6 connection received from ::1 port 55930 [20230502-22:32:24] [DEBUG] session_get_bydata: search policy 0 U alexander W 2516 H 984 bpp 24 T 3 IP ::ffff:127.0.0.1:46378 - socket: 12 [20230502-22:32:24] [INFO ] Terminal Server Users group is disabled, allowing authentication [20230502-22:32:24] [INFO ] ++ created session (access granted): username alexander, ip ::ffff:127.0.0.1:46378 - socket: 12 [20230502-22:32:24] [INFO ] starting Xorg session... [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:24] [DEBUG] Closed socket 13 (AF_INET6 :: port 5910) [20230502-22:32:24] [DEBUG] Did not find a running X server at 5910 [20230502-22:32:24] [DEBUG] Closed socket 13 (AF_INET6 :: port 6010) [20230502-22:32:24] [DEBUG] Did not find a running X server at 6010 [20230502-22:32:24] [DEBUG] Closed socket 13 (AF_INET6 :: port 6210) [20230502-22:32:24] [DEBUG] Did not find a running X server at 6210 [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.xrdp/xrdp_chansrv_socket_10 [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.xrdp/xrdp_chansrv_audio_out_socket_10 [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.xrdp/xrdp_chansrv_audio_in_socket_10 [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.xrdp/xrdpapi_10 [20230502-22:32:24] [INFO ] Starting session: session_pid 1852, display :10.0, width 2516, height 984, bpp 24, client ip ::ffff:127.0.0.1:46378 - socket: 12, user name alexander [20230502-22:32:24] [INFO ] [session start] (display 10): calling auth_start_session from pid 1852 [20230502-22:32:24] [ERROR] sesman_data_in: scp_process_msg failed [20230502-22:32:24] [DEBUG] Closed socket 11 (AF_INET6 ::1 port 3350) [20230502-22:32:24] [ERROR] sesman_main_loop: trans_check_wait_objs failed, removing trans [20230502-22:32:24] [DEBUG] Closed socket 12 (AF_INET6 ::1 port 3350) [20230502-22:32:24] [DEBUG] Closed socket 12 (AF_INET6 ::1 port 3350) [20230502-22:32:24] [DEBUG] Waiting for X server to start on display 10 [20230502-22:32:24] [DEBUG] Waiting for X server to start on display 10 [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:24] [INFO ] Starting X server on display 10: /usr/bin/Xorg :10 -auth .Xauthority -config xrdp/xorg.conf -noreset -nolisten tcp -logfile .xorgxrdp.%s.log [20230502-22:32:24] [DEBUG] Calling exec (executable: /usr/bin/Xorg, arguments: /usr/bin/Xorg :10 -auth .Xauthority -config xrdp/xorg.conf -noreset -nolisten tcp -logfile .xorgxrdp.%s.log) [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:25] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:25] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:25] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:25] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:25] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:25] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:25] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:25] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:25] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:25] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:25] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:25] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:25] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:25] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:25] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:25] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:26] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:26] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:26] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:26] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:26] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:26] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:26] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:26] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:26] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:26] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:26] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:26] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:26] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:26] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:26] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:26] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:27] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:27] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:27] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:27] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:27] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:27] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:27] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:27] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:27] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:27] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:27] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:27] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:27] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:27] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:27] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:28] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:28] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:28] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:28] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:28] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:28] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:28] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:28] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:28] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:28] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:28] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:28] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:28] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:28] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:28] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:29] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:29] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:29] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:29] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:29] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:29] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:29] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:29] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:29] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:29] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:29] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:29] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:29] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:29] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:29] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:29] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:30] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:30] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:30] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:30] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:30] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:30] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:30] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:30] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:30] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:30] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:30] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:30] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:30] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:30] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:31] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:31] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:31] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:31] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:31] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:31] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:31] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:31] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:31] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:31] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:31] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:31] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:31] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:31] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:31] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:31] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:32] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:32] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:32] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:32] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:32] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:32] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:32] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:32] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:32] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:32] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:32] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:32] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:32] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:32] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:32] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:32] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:33] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:33] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:33] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:33] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:33] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:33] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:33] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:33] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:33] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:33] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:33] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:33] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:33] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:33] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:33] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:33] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:34] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:34] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:34] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:34] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:34] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:34] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:34] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:34] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:34] [WARN ] Timed out waiting for X server on display 10 to startup [20230502-22:32:34] [INFO ] Session started successfully for user alexander on display 10 [20230502-22:32:34] [INFO ] Session in progress on display 10, waiting until the window manager (pid 1853) exits to end the session [20230502-22:32:34] [INFO ] Starting the xrdp channel server for display 10 [20230502-22:32:34] [DEBUG] waiting for pid 1853 to exit [20230502-22:32:34] [DEBUG] Calling exec (executable: /usr/sbin/xrdp-chansrv, arguments: /usr/sbin/xrdp-chansrv) [20230502-22:32:34] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:34] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:34] [WARN ] Timed out waiting for X server on display 10 to startup [20230502-22:32:34] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:34] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:34] [ERROR] There is no X server active on display 10 [20230502-22:32:34] [ERROR] A fatal error has occurred attempting to start the window manager on display 10, aborting connection [20230502-22:32:34] [WARN ] Window manager (pid 1853, display 10) exited quickly (0 secs). This could indicate a window manager config problem [20230502-22:32:34] [INFO ] Calling auth_stop_session and auth_end from pid 1852 [20230502-22:32:34] [INFO ] Terminating X server (pid 1854) on display 10 [20230502-22:32:34] [INFO ] Terminating the xrdp channel server (pid 1879) on display 10 [20230502-22:32:34] [DEBUG] waiting for pid 1854 to exit [20230502-22:32:34] [INFO ] X server on display 10 (pid 1854) returned exit code 1 and signal number 0 [20230502-22:32:34] [DEBUG] waiting for pid 1879 to exit [20230502-22:32:34] [INFO ] xrdp channel server for display 10 (pid 1879) exit code 0 and signal number 0 [20230502-22:32:34] [INFO ] cleanup_sockets: [20230502-22:32:34] [DEBUG] receiving SIGCHLD [20230502-22:32:34] [INFO ] Process 1852 has exited [20230502-22:32:34] [INFO ] ++ terminated session: username alexander, display :10.0, session_pid 1852, ip ::ffff:127.0.0.1:46378 - socket: 12 Just in case, I made these settings: /etc/X11/xrdp/xrdp.conf Section "ServerLayout" Identifier "X11 Server" Screen "Screen (xrdpdev)" InputDevice "xrdpMouse" "CorePointer" InputDevice "xrdpKeyboard" "CoreKeyboard" EndSection Section "ServerFlags" # This line prevents "ServerLayout" sections in xorg.conf.d files # overriding the "X11 Server" layout (xrdp #1784) Option "DefaultServerLayout" "X11 Server" Option "DontVTSwitch" "on" Option "AutoAddDevices" "off" Option "AutoAddGPU" "off" EndSection Section "Module" Load "dbe" Load "ddc" Load "extmod" Load "glx" Load "int10" Load "record" Load "vbe" Load "xorgxrdp" Load "fb" EndSection Section "InputDevice" Identifier "xrdpKeyboard" Driver "xrdpkeyb" EndSection Section "InputDevice" Identifier "xrdpMouse" Driver "xrdpmouse" EndSection Section "Monitor" Identifier "Monitor" Option "DPMS" HorizSync 30-80 VertRefresh 60-75 ModeLine "1920x1080" 138.500 1920 1968 2000 2080 1080 1083 1088 1111 +hsync -vsync ModeLine "1280x720" 74.25 1280 1720 1760 1980 720 725 730 750 +HSync +VSync Modeline "1368x768" 72.25 1368 1416 1448 1528 768 771 781 790 +hsync -vsync Modeline "1600x900" 119.00 1600 1696 1864 2128 900 901 904 932 -hsync +vsync EndSection Section "Device" Identifier "Video Card (xrdpdev)" Driver "xrdpdev" Option "DRMDevice" "/dev/dri/renderD128" Option "DRI3" "1" EndSection Section "Screen" Identifier "Screen (xrdpdev)" Device "Video Card (xrdpdev)" GPUDevice "" Monitor "Monitor" DefaultDepth 24 SubSection "Display" Depth 24 Modes "640x480" "800x600" "1024x768" "1280x720" "1280x1024" "1600x900" "1920x1080" EndSubSection EndSection But for some reason the file is ~/.xorgxrdp.10.log is missing I checked a few more options. Here is the full result of my attempts: /etc/xrdp/xrdp.ini /etc/xrdp/sesman.ini xrdp.log xrdp-sesman.log Just in case, I made these settings: /etc/X11/xrdp/xrdp.conf /etc/xrdp/startwm.sh But for some reason the file is ~/.xorgxrdp.10.log is missing Hm... This solution helped me. For some reason, I didn't find any mention of this configuration file. How true is everything in such a decision? A couple of things:- In your second log (this post), you've still not got a qualified path to Xorg. This probably means you didn't successfully restart sesman at the time. The path in sesman.ini is incorrect. Where you have param=/usr/bin/Xorg, you should have param=/usr/lib/xorg/Xorg. Adding the Xorg in the standard PATH won't work. To see why, try cat /usr/bin/Xorg. This can be seen in your 3rd sesman.log above. You should get to a state where you can see this in your log:- [INFO ] Starting X server on display 10: /usr/lib/xorg/Xorg :10 -auth .Xauthority -config xrdp/xorg.conf -noreset -nolisten tcp -logfile .xorgxrdp.%s.log Your linked solution doesn't seem relevant to the problem as you've posted it. @matt335672 I apologize for the very late response. I have now checked all my settings and did as you indicated and the problem has been solved for me. > $ cat /usr/bin/Xorg #!/bin/sh # # Execute Xorg.wrap if it exists otherwise execute Xorg directly. # This allows distros to put the suid wrapper in a separate package. basedir="/usr/lib64/xorg-server" if [ -x "$basedir"/Xorg.wrap ]; then exec "$basedir"/Xorg.wrap "$@" else exec "$basedir"/Xorg "$@" fi > $ ls /usr/lib64/xorg-server/Xorg /usr/lib64/xorg-server/Xorg And as a path, I specified the full path in sesman.ini: path=/usr/lib64/xorg-server/Xorg [root@iaas-kunpeng-bljcs-02 bwda]# cat .xorgxrdp.10.log [ 3160.803] X.Org X Server 1.20.8 X Protocol Version 11, Revision 0 [ 3160.803] Build Operating System: localhost 4.19.90-23.0.v2101.ky10.aarch64 [ 3160.803] Current Operating System: Linux iaas-kunpeng-bljcs-02 4.19.90-24.4.v2101.ky10.aarch64 #1 SMP Mon May 24 14:45:37 CST 2021 aarch64 [ 3160.803] Kernel command line: BOOT_IMAGE=/vmlinuz-4.19.90-24.4.v2101.ky10.aarch64 root=/dev/mapper/system-lv_root ro biosdevname=0 net.ifnames=0 console=ttyS0,115200n8 crashkernel=1024M,high smmu.bypassdev=0x1000:0x17 smmu.bypassdev=0x1000:0x15 video=efifb:off video=VGA-1:640x480-32@60me [ 3160.803] Build Date: 27 February 2021 05:07:25PM [ 3160.803] Build ID: xorg-x11-server 1.20.8-3.p01.ky10 [ 3160.803] Current version of pixman: 0.40.0 [ 3160.803] Before reporting problems, check http://wiki.x.org to make sure that you have the latest version. [ 3160.803] Markers: (--) probed, () from config file, (==) default setting, (++) from command line, (!!) notice, (II) informational, (WW) warning, (EE) error, (NI) not implemented, (??) unknown. [ 3160.803] (++) Log file: ".xorgxrdp.10.log", Time: Wed May 24 10:46:06 2023 [ 3160.803] (EE) Unable to locate/open config file: "xrdp/xorg.conf" [ 3160.803] (==) Using config directory: "/etc/X11/xorg.conf.d" [ 3160.803] (==) Using system config directory "/usr/share/X11/xorg.conf.d" [ 3160.804] (==) No Layout section. Using the first Screen section. [ 3160.804] (==) No screen section available. Using defaults. [ 3160.804] () |-->Screen "Default Screen Section" (0) [ 3160.804] (**) | |-->Monitor "" [ 3160.804] (==) No monitor specified for screen "Default Screen Section". Using a default monitor configuration. [ 3160.804] (==) Automatically adding devices [ 3160.804] (==) Automatically enabling devices [ 3160.804] (==) Automatically adding GPU devices [ 3160.804] (==) Automatically binding GPU devices [ 3160.804] (==) Max clients allowed: 256, resource mask: 0x1fffff [ 3160.804] (==) FontPath set to: catalogue:/etc/X11/fontpath.d, built-ins [ 3160.804] (==) ModulePath set to "/usr/lib64/xorg/modules" [ 3160.804] (II) The server relies on udev to provide the list of input devices. If no devices become available, reconfigure udev or disable AutoAddDevices. [ 3160.804] (II) Loader magic: 0xaaab76d30dd8 [ 3160.804] (II) Module ABI versions: [ 3160.804] X.Org ANSI C Emulation: 0.4 [ 3160.804] X.Org Video Driver: 24.1 [ 3160.804] X.Org XInput driver : 24.1 [ 3160.804] X.Org Server Extension : 10.0 [ 3160.805] (II) systemd-logind: took control of session /org/freedesktop/login1/session/c8 [ 3160.806] (II) xfree86: Adding drm device (/dev/dri/card0) [ 3160.806] (EE) systemd-logind: failed to take device /dev/dri/card0: Operation not permitted [ 3160.806] (EE) /dev/dri/card0: failed to set DRM interface version 1.4: Permission denied [ 3160.811] (--) PCI:*(6@0:0:0) 19e5:1711:19e5:1711 rev 1, Mem @ 0xe4000000/33554432, 0xe7200000/2097152, BIOS @ 0x????????/65536 [ 3160.811] (II) LoadModule: "glx" [ 3160.811] (II) Loading /usr/lib64/xorg/modules/extensions/libglx.so [ 3160.813] (II) Module glx: vendor="X.Org Foundation" [ 3160.813] compiled for 1.20.8, module version = 1.0.0 [ 3160.813] ABI class: X.Org Server Extension, version 10.0 [ 3160.813] (==) Matched modesetting as autoconfigured driver 0 [ 3160.813] (==) Matched fbdev as autoconfigured driver 1 [ 3160.813] (==) Assigned the driver to the xf86ConfigLayout [ 3160.813] (II) LoadModule: "modesetting" [ 3160.813] (II) Loading /usr/lib64/xorg/modules/drivers/modesetting_drv.so [ 3160.813] (II) Module modesetting: vendor="X.Org Foundation" [ 3160.813] compiled for 1.20.8, module version = 1.20.8 [ 3160.813] Module class: X.Org Video Driver [ 3160.813] ABI class: X.Org Video Driver, version 24.1 [ 3160.813] (II) LoadModule: "fbdev" [ 3160.813] (II) Loading /usr/lib64/xorg/modules/drivers/fbdev_drv.so [ 3160.813] (II) Module fbdev: vendor="X.Org Foundation" [ 3160.813] compiled for 1.20.8, module version = 0.5.0 [ 3160.813] Module class: X.Org Video Driver [ 3160.813] ABI class: X.Org Video Driver, version 24.1 [ 3160.813] (II) modesetting: Driver for Modesetting Kernel Drivers: kms [ 3160.813] (II) FBDEV: driver for framebuffer: fbdev [ 3160.813] (EE) Fatal server error: [ 3160.813] (EE) parse_vt_settings: Cannot open /dev/tty0 (Permission denied) [ 3160.813] (EE) [ 3160.813] (EE) Please consult the openEuler Project support at http://wiki.x.org for help. [ 3160.813] (EE) Please also check the log file at ".xorgxrdp.10.log" for additional information. [ 3160.813] (EE) [ 3160.813] (WW) xf86CloseConsole: KDSETMODE failed: Bad file descriptor [ 3160.813] (WW) xf86CloseConsole: VT_GETMODE failed: Bad file descriptor [ 3160.813] (EE) Server terminated with error (1). Closing log file.
gharchive/issue
2023-05-01T21:47:01
2025-04-01T06:39:44.634922
{ "authors": [ "AlexanderZhirov", "FirstLicz", "matt335672" ], "repo": "neutrinolabs/xrdp", "url": "https://github.com/neutrinolabs/xrdp/issues/2652", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1976716590
Project license Hi - what is this code licensed under? I'd like to fork it for CS2 reverse engineering purposes. Many thanks Hey and sorry, we completely forgot about the license. 1063e42a9e25e4b290e52d0222bf33b3c6b4fb10 Many thanks for the quick update. One question - the underlying library that you forked from is MIT licensed. I believe it requires that license acknowledged in your copy (this is how I did it: https://github.com/saul/CS2SchemaGen/blob/main/ACKNOWLEDGEMENTS) Many thanks for the quick update. One question - the underlying library that you forked from is MIT licensed. I believe it requires that license acknowledged in your copy (this is how I did it: https://github.com/saul/CS2SchemaGen/blob/main/ACKNOWLEDGEMENTS) Thank you very much! We did what you said, it took a bit more time as I was doing some reverse for schema at the time.
gharchive/issue
2023-11-03T18:13:29
2025-04-01T06:39:44.643966
{ "authors": [ "cpz", "es3n1n", "saul" ], "repo": "neverlosecc/source2gen", "url": "https://github.com/neverlosecc/source2gen/issues/19", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
3173922
use within other scripts hey there, I'm writing a grunt task that generates docs and I was thinking of using groc. I was wondering if it's possible to use groc within another script as opposed to on the CLI. I have a very rough API that you could use, though it's definitely going to be a moving target for a while. (I'm totally happy to work with you to solidify it over time though :)) Take a look at where the CLI code builds a Project object for the low level API. Alternatively, you can fake a command line call w/o the shell as an intermediary by including the CLI module directly, and calling it similar to the bin file - this interface is less likely to change, but is definitely more awkward. Thinking further on this, a better and probably relatively stable API would be to decouple the argument parsing, and have a call that takes the resultant options hash - easy to configure, and also plays nicely with the .groc.json file
gharchive/issue
2012-02-10T16:07:04
2025-04-01T06:39:44.654283
{ "authors": [ "nevir", "svnlto" ], "repo": "nevir/groc", "url": "https://github.com/nevir/groc/issues/22", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1223617008
Minor balance changes also cleaned duplicates in learnsets.ts I was removing duplicate moves. Round was a mistake though, so I added it back
gharchive/pull-request
2022-05-03T03:20:30
2025-04-01T06:39:44.657116
{ "authors": [ "jmf61" ], "repo": "new-meta-incorporated/pokemon-showdown", "url": "https://github.com/new-meta-incorporated/pokemon-showdown/pull/35", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
62487578
Firefox Add-On? Is there any idea/motivation for a Firefox AddOn? Would be my first, if nobody wants to take it … Someone has posted the following on reddit.com/r/GNUTerryPratchett but I don't use Firefox so I haven't tested it myself: https://addons.mozilla.org/en-US/firefox/addon/gnu_terry_pratchett/ I don't know anything about Firefox development, but then I didn't really know about Chrome dev before I tried this. My original motivation for this was really just for my own use (I use Chrome) and I released it after seeing how much interest there was!
gharchive/issue
2015-03-17T19:04:58
2025-04-01T06:39:44.660653
{ "authors": [ "Ryuno-Ki", "newfolder0" ], "repo": "newfolder0/chrome-clacks", "url": "https://github.com/newfolder0/chrome-clacks/issues/5", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2463632448
FR: Toggle Functionality for Drop Rule A functionality that allows for both individual and bulk enabling or disabling of drop rules. This flexibility is crucial for efficient rule management and operational agility. The nerdgraph API does not offer an option to disable a drop rule and therefore can not be added as a feature to this app. This would require a change to the product via a feature request. You could possibly get the effect of a disbaled rule by adding a filter to the where clause that would never match like where true=false - Im not sure this is ideal. Internal FR created FRB-00007240
gharchive/issue
2024-08-13T15:26:44
2025-04-01T06:39:44.665272
{ "authors": [ "jsbnr" ], "repo": "newrelic-experimental/nr1-drop-rule-explorer", "url": "https://github.com/newrelic-experimental/nr1-drop-rule-explorer/issues/9", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
955569426
set aws region to be alertable Relevant information Describe what you have done and any details that you think are relevant or that you might want to discuss with us. Checklist [ ] I've read the guidelines and understand the acceptance criteria. [ ] The value of the attribute marked as identifier will be unique and valid. [ ] I've confirmed that my entity type wasn't already defined. If it is I'm providing an explanation above. Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.Cian Buckley seems not to be a GitHub user. You need a GitHub account to be able to sign the CLA. If you have already a GitHub account, please add the email address used for this commit to your account.You have signed the CLA already but the status is still pending? Let us recheck it.
gharchive/pull-request
2021-07-29T07:59:50
2025-04-01T06:39:44.669381
{ "authors": [ "CLAassistant", "cianBuckley" ], "repo": "newrelic/entity-definitions", "url": "https://github.com/newrelic/entity-definitions/pull/285", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
726036127
fix: namespace's prefix deleted(manager.yaml) fix: namespace's prefix deleted(manager.yaml) so when I ran this with my original version of kustomize (3.8.2) I was seeing a different $ diff master.yaml kaojiri-master.yaml 6c6 < name: newrelic-kubernetes-operator-system --- > name: system $ kustomize version {Version:3.8.2 GitCommit:e2973f6ecc9be6187cfd5ecf5e180f842249b3c6 BuildDate:2020-08-29T19:15:49+01:00 GoOs:darwin GoArch:amd64} However, I updated to the latest version of kustomize (3.8.5), no difference found $ diff master.yaml kaojiri-master.yaml $ kustomize version {Version:kustomize/v3.8.5 GitCommit:4052cd4fd8c76a17b5f64e32509f3fba9713fe75 BuildDate:2020-10-08T05:35:40+01:00 GoOs:darwin GoArch:amd64} I'm concerned that with merging this any users with older versions of kustomize may encounter unexpected issues. Let me check with some others and get a second opinion on this. after further internal discussion among users of this operator and in light of the fact that this is restoring previous behavior before it changed by kustomize, I'll go ahead and merge this.
gharchive/pull-request
2020-10-21T00:29:45
2025-04-01T06:39:44.679554
{ "authors": [ "kaojiri", "thande" ], "repo": "newrelic/newrelic-kubernetes-operator", "url": "https://github.com/newrelic/newrelic-kubernetes-operator/pull/123", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2234184037
feat: New singlestore quickstart Summary Added a new quickstart that utilizes an example OTeL tool collector to get metrics from SingleStore's API and ingest it into NR. Pre merge checklist [x] Did you check you NRQL syntax? - Does it work? [x] Did you include a Data source and Documentation reference? [x] Are all documentation links publicly accessible? [x] Did you check your descriptive content for voice and tone? [x] Did you check your descriptive content for spelling and grammar errors? [x] Did you review your content with a subject matter expert? (e.g. a Browser agent quickstart is reviewed with a member of the Browser Agent team) Dashboards [ ] Does the PR contain a screenshot for each of your dashboards? [ ] Do your screenshots show data? [ ] Has the sanitization script been run on each dashboard? Alerts [ ] Did you check that your alerts actually work? Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you all sign our Contributor License Agreement before we can accept your contribution.0 out of 2 committers have signed the CLA.:x: jcountsNR:x: MichelLosierYou have signed the CLA already but the status is still pending? Let us recheck it.
gharchive/pull-request
2024-04-09T19:43:40
2025-04-01T06:39:44.685937
{ "authors": [ "CLAassistant", "jcountsNR" ], "repo": "newrelic/newrelic-quickstarts", "url": "https://github.com/newrelic/newrelic-quickstarts/pull/2362", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2582386023
Docs say that Grape, Padrino, and Sinatra aren't supported for Ruby 3.0+ but they certainly are Some old documentation that presumably predates Ruby v3.0 being commonplace and well supported by the agent still notes: Please note that Grape, Padrino, and Sinatra aren't supported for Ruby 3.0+. https://docs.newrelic.com/docs/apm/agents/ruby-agent/getting-started/ruby-agent-requirements-supported-frameworks/ Resolved by https://github.com/newrelic/docs-website/pull/19465
gharchive/issue
2024-10-12T01:06:43
2025-04-01T06:39:44.687903
{ "authors": [ "fallwith", "kaylareopelle" ], "repo": "newrelic/newrelic-ruby-agent", "url": "https://github.com/newrelic/newrelic-ruby-agent/issues/2909", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
614994335
A bit of CI upkeep This PR mixes a number of small changes, so I think it's fine to bundle them like this. First: back when we were only using Travis, we ran all our jobs on the same Ubuntu version. Official repos didn't contain all the C++ compilers we need, so we also added r/test PPA and apt.llvm.org. This made the .travis.yml slightly less complicated than it would've been otherwise. Now that we're using Cirrus CI and Docker, it's far simpler to run a couple different Ubuntu versions such that they cover the whole range of compilers we support. That's what I did here: Ubuntu 18.04 keeps shouldering most of the load, while 16.04 is used for GCC 4.9, and 20.04 is used for GCC 9, GCC 10, and Clang 10. Second: .travis.yml had some leftovers from my failed attempt to run builds with more warnings enabled, and run tests with more checks (like stack protector). This PR adds these jobs to Cirrus. I'm not yet sure how useful they will be; I made a note to re-evaluate in two months. Fixes #838 Fixes #922 Reviews welcome. Will merge in 24 hours. Coverage remained the same at 56.199% when pulling 275f8cba7acbd4bf49e67c863d6f7e8de5ddb2c7 on feature/dockerfiles-upkeep into b1884ff27f341c8d400432d50567ec4857374f0b on master.
gharchive/pull-request
2020-05-08T21:22:59
2025-04-01T06:39:44.716600
{ "authors": [ "Minoru", "coveralls" ], "repo": "newsboat/newsboat", "url": "https://github.com/newsboat/newsboat/pull/925", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1551263381
adding chatGPT files FINALLY! I didn't click something when making my access token. I was able to do this on Friday last week, I just didn't update it. My bad.
gharchive/pull-request
2023-01-20T18:12:42
2025-04-01T06:39:44.730298
{ "authors": [ "spookybeetle" ], "repo": "newtfire/textAnalysis-Hub", "url": "https://github.com/newtfire/textAnalysis-Hub/pull/102", "license": "CC0-1.0", "license_type": "permissive", "license_source": "github-api" }
364840792
请问JPVideoPlayerManager.m中 ,这个断言的作用是防止什么?我总是崩溃在这里 我是崩溃在这个断言, 作者为什么要加这么多断言? pod install 下来,支持三个url,超过就会崩溃,JPVideoPlayerManager.m总是崩溃。 我把断言那句话,用if容错掉,然后就没有崩溃了 Me too 我把断言那句话,用if容错掉,然后就没有崩溃了 你知道, 作者为什么要加这么多assertion吗? 断言去掉了,为了排除 bug 加的断言.
gharchive/issue
2018-09-28T10:56:55
2025-04-01T06:39:44.734359
{ "authors": [ "CranberryYam", "invokeHao", "mjf1986", "newyjp" ], "repo": "newyjp/JPVideoPlayer", "url": "https://github.com/newyjp/JPVideoPlayer/issues/205", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1318879766
Package Resources are not associated to packages in application Packages scan step update In updating the application Packages scanning step in the scan_codebase pipeline for #447, I noticed that the CodebaseResources of a Package are not associated with the Packages that was scanned. Looking a little deeper, I see that the Package Resources in scancode.io cannot be properly associated to the Package they are from by using the .assemble() methods from packagedcode's Package handlers. This is because of how for_packages is implemented on the CodebaseResource model in scancode.io and the Resource model in commoncode. In scancode.io, for_packages on the CodebaseResource model is a property that collects the purls for Packages that have been related to that CodebaseResource. In scancode-toolkit/commoncode, for_packages on the Resource model is a list that contains package_uid strings for the Package the Resource is from. When you run the .assemble() methods from packagedcode Package handlers on a CodebaseResource object, it attempts to append package_uid strings to the for_packages field. This does not properly work since CodebaseResource.for_packages is a property, not an attribute that can be used the same way. This issue has been fixed, following an update to scancode-toolkit where the code has been modified to be able to accept different functions that adds a package to a resource. https://github.com/nexB/scancode-toolkit/pull/3035
gharchive/issue
2022-07-27T00:21:18
2025-04-01T06:39:44.762427
{ "authors": [ "JonoYang" ], "repo": "nexB/scancode.io", "url": "https://github.com/nexB/scancode.io/issues/476", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
745535611
Script hangs in the middle of execution for up to an hour. time sh -c 'sh -c "$(curl -sL https://nextdns.io/diag)"' Welcome to NextDNS network diagnostic tool. This tool will download a small binary to capture latency and routing information regarding the connectivity of your network with NextDNS. In order to perform a traceroute, root permission is required. You may therefore be asked to provide your password for sudo. The source code of this tool is available at https://github.com/nextdns/diag Do you want to continue? (press enter to accept) Testing IPv6 connectivity available: true Fetching https://test.nextdns.io status: ok client: 193.110.******* protocol: DOH dest IP: 45.90.28.0 server: rix-hkg-1 Traceroute for primary IPv4 (45.90.28.0) 1 193.110.****** 0ms 0ms 0ms ^C real 72m1.162s user 0m0.419s sys 0m0.174s What is your platform/OS/version? What is your platform/OS/version? uname -a Linux ******** 5.8.0-29-generic #31~20.04.1-Ubuntu SMP Fri Nov 6 16:10:42 UTC 2020 x86_64 x86_64 x86_64 GNU/Linux Traceroute is hanging when trying to run diag against 1.4.3 install. See this output where it's stuck: Testing IPv6 connectivity available: false Fetching https://test.nextdns.io status: ok client: 24.229.200.248 protocol: DOH dest IP: 162.220.223.23 server: anexia-ewr-1 Fetching PoP name for ultra low latency primary IPv4 (ipv4.dns1.nextdns.io) vultr-ewr: 19.638ms Fetching PoP name for ultra low latency secondary IPv4 (ipv4.dns2.nextdns.io) anexia-ewr: 13.054ms Fetching PoP name for anycast primary IPv4 (45.90.28.0) zepto-xrs: 26.782ms Fetching PoP name for anycast secondary IPv4 (45.90.30.0) anexia-ewr: 16.078ms Pinging PoPs anexia-ewr: 11.887ms hetzner-iad: 20.496ms smarthost-bos: 20.537ms vultr-ewr: 20.505ms anexia-mnz: 20.451ms zepto-xrs: 20.544ms zepto-iad: 20.465ms axcelx-bos: 20.506ms teraswitch-pit: 24.632ms cloudzy-pit: 52.182ms Traceroute for ultra low latency primary IPv4 (66.135.25.157) 1 10.229.192.1 13ms 5ms 7ms 2 207.44.112.149 9ms 8ms 7ms 3 172.16.10.101 13ms 14ms 13ms 4 172.16.10.170 13ms 14ms 13ms 5 213.200.112.170 14ms 15ms 12ms 6 173.205.45.234 74ms 44ms 48ms
gharchive/issue
2020-11-18T10:30:58
2025-04-01T06:39:45.359124
{ "authors": [ "IceCodeNew", "dmfrey", "rs" ], "repo": "nextdns/diag", "url": "https://github.com/nextdns/diag/issues/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1176591207
small typo: java.imageio java.imageio mentioned in the changelog should be javax.imageio: https://github.com/nextjournal/clerk/blob/fc297cb8b1cd879c888f573d1c7913dafd1d66dd/CHANGELOG.md#05346-2022-01-27 @daslu thanks, fixed!
gharchive/issue
2022-03-22T10:40:29
2025-04-01T06:39:45.396471
{ "authors": [ "daslu", "mk" ], "repo": "nextjournal/clerk", "url": "https://github.com/nextjournal/clerk/issues/119", "license": "ISC", "license_type": "permissive", "license_source": "github-api" }
1865197638
🛑 /getMusicFeed - Get music feed by URL or ID is down In e33b4f6, /getMusicFeed - Get music feed by URL or ID (https://nextpost.tech/tiktok-api/getMusicFeed?query=7129970471625362182&cache_timeout=0&license_key=$API_KEY) was down: HTTP code: 403 Response time: 13 ms Resolved: /getMusicFeed - Get music feed by URL or ID is back up in 3095fcc after 298 days, 4 hours, 19 minutes.
gharchive/issue
2023-08-24T13:45:46
2025-04-01T06:39:45.399263
{ "authors": [ "sergeykomlev" ], "repo": "nextpost-tech/tiktok-status", "url": "https://github.com/nextpost-tech/tiktok-status/issues/1280", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1878696221
[BUG] - Modal scrollBehavior="outside" breaks placement prop NextUI Version latest Describe the bug By using the scrollBehavior="outside" prop, placement is always at the top (with a my-16 margin) Cause: https://github.com/nextui-org/nextui/blob/main/packages/core/theme/src/components/modal.ts#L171 For now the only way to fix it on the user side is to not use scrollBehavior="outside" but enter <Modal classNames={{ wrapper: "items-end overflow-y-auto", base: "my-1", }} /> for example if one prefers placement="bottom" and scrollBehavior="outside" Your Example Website or App No response Steps to Reproduce the Bug or Issue Create a Add scrollBehaviour="bottom" Modal will be placed at top Expected behavior Modal should always respect the placement prop Screenshots or Videos No response Operating System Version any Browser Other (add additonal context) It might be due to incorrect usage of parameters. bottom is not a valid parameter for scrollBehavior. The valid parameters for scrollBehavior are normal | inside | outside https://nextui.org/docs/components/modal#api It might be due to incorrect usage of parameters. bottom is not a valid parameter for scrollBehavior. The valid parameters for scrollBehavior are normal | inside | outside https://nextui.org/docs/components/modal#api No. The bug is emerging when scrollBehavior=outside and placement=bottom is used together. Look at your screenshot of the docs page, the prop just below scrollBehavior scrollBehavior=“outside” Thank you for correcting the reproduction steps. https://codesandbox.io/p/sandbox/sleepy-hugle-d6tt88?file=/App.jsx:37,42&utm_medium=sandpack There are two scenarios: Scenario 1: When the text does not exceed the container. In this case, scrollBehavior='outside' should not be used. placement='top' placement='bottom' In this case, the placement property is effective. Scenario 2: When the text exceeds the container. In this case, the scrollBehavior property must be added. When scrollBehavior='outside' and placement='bottom' are used together, the behavior is such that the placement property does not take effect. Regardless of the placement value set, the behavior remains the same, with a distance from the top, and a scroll-down state, as shown in the image below. scrollBehavior='outside' and placement='top' In this case, the placement property does not take effect. Is it your expectation that when setting scrollBehavior='outside' and placement='bottom', the initial pop-up state is at the bottom of the scrollbar, as shown in the image below? @wingkwong Can I work on this issue and add the 3 cases suggested by @robodove ? @wingkwong This behavior is taking place because of the items-start class in scrollBehavior:outside variant and disabling it should provide the solution for OP. Should we allow the user to change the placement property when scrollBehavior = outside, or we should keep the things unchanged?
gharchive/issue
2023-09-02T13:46:39
2025-04-01T06:39:45.461107
{ "authors": [ "CHENFANGC", "abhinav700", "robodove" ], "repo": "nextui-org/nextui", "url": "https://github.com/nextui-org/nextui/issues/1548", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2419937462
Test2 test2 branch test Deployment started for this pull request. Deployment successful. Deployed application.
gharchive/pull-request
2024-07-19T21:34:31
2025-04-01T06:39:45.495600
{ "authors": [ "Hamed-Ayodeji", "neyo55" ], "repo": "neyo55/hng-stage4-pr-with-github-bot", "url": "https://github.com/neyo55/hng-stage4-pr-with-github-bot/pull/15", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2740392522
[-𝗙𝗨𝗟𝗟-𝐯𝐢𝐫𝐚𝐥!𝐜𝐥𝐢𝐩𝐬-]— Mona Alam Video Original Video Link Mona Alam Video Viral On Social Media X Trending Now 54 seconds ago L𝚎aked Video Mona Alam Original Video Viral Video L𝚎aked on X Twitter Telegram ►✅ 𝘾𝙇𝙄𝘾𝙆 𝙃𝙀𝙍𝙀 ==►► 𝙁𝙪𝙡𝙡 𝙑𝙞𝙙𝙚𝙤❤️❤️⬇️⬇️​ ►✅ 𝘾𝙇𝙄𝘾𝙆 𝙃𝙀𝙍𝙀 ==►► 𝙁𝙪𝙡𝙡 𝙑𝙞𝙙𝙚𝙤❤️❤️⬇️⬇️​ WATCH NOW [-𝐅𝐔𝐋𝐋-𝐕𝐈𝐑𝐀𝐋-]— Mona Alam Video Original Video Link Mona Alam Video Viral On Social Media X Now [1TNPO] [-wATCH-]— Mona Alam Video Original Video Link Mona Alam Video Viral On Social Media X Trending Now [-wATCH-]— Mona Alam Video Original Video Link Mona Alam Video Viral On Social Media X Trending Now [-wATCH-]— Mona Alam ʟᴇᴀᴋᴇᴅ Video ᴠɪʀᴀʟ On Social Media ˣ ᵀʷⁱᵗᵗᵉʳ [-wATCH-]— Mona Alam ʟᴇᴀᴋᴇᴅ Video ᴠɪʀᴀʟ On Social Media ˣ ᵀʷⁱᵗᵗᵉʳ [-wATCH-]— Mona Alam Video Original Video Link Mona Alam Video Viral On Social Media X Trending Now Mona Alam Original Video video took the internet by storm and amazed viewers on various social media platforms. Mona Alam, a young and talented digital creator, recently became famous thanks to this interesting video. L𝚎aked Video Mona Alam Original Video Viral Video L𝚎aked on X Twitter Mona Alam Original Video video oficial twitter L𝚎aked Video Mona Alam Original Video Viral Video L𝚎aked on X Twitter.. . . . .. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 🟢 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑾𝑨𝑻𝑪𝑯 𝑵𝑶𝑾 🔴 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑫𝒐𝒘𝒏𝒍𝒐𝒂𝒅 𝑵𝒐𝒘 ... ... .... 🟢 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑾𝑨𝑻𝑪𝑯 𝑵𝑶𝑾 🔴 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑫𝒐𝒘𝒏𝒍𝒐𝒂𝒅 𝑵𝒐𝒘 ... ... .... 🟢 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑾𝑨𝑻𝑪𝑯 𝑵𝑶𝑾 🔴 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑫𝒐𝒘𝒏𝒍𝒐𝒂𝒅 𝑵𝒐𝒘 ... ... .... 🟢 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑾𝑨𝑻𝑪𝑯 𝑵𝑶𝑾 🔴 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑫𝒐𝒘𝒏𝒍𝒐𝒂𝒅 𝑵𝒐𝒘 ... ... .... 🟢 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑾𝑨𝑻𝑪𝑯 𝑵𝑶𝑾 🔴 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑫𝒐𝒘𝒏𝒍𝒐𝒂𝒅 𝑵𝒐𝒘 ... ... .... 🟢 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑾𝑨𝑻𝑪𝑯 𝑵𝑶𝑾 🔴 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑫𝒐𝒘𝒏𝒍𝒐𝒂𝒅 𝑵𝒐𝒘 ... ... .... 🟢 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑾𝑨𝑻𝑪𝑯 𝑵𝑶𝑾 🔴 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑫𝒐𝒘𝒏𝒍𝒐𝒂𝒅 𝑵𝒐𝒘 ... ... .... 🟢 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑾𝑨𝑻𝑪𝑯 𝑵𝑶𝑾 🔴 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑫𝒐𝒘𝒏𝒍𝒐𝒂𝒅 𝑵𝒐𝒘 ... ... .... 🟢 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑾𝑨𝑻𝑪𝑯 𝑵𝑶𝑾 🔴 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑫𝒐𝒘𝒏𝒍𝒐𝒂𝒅 𝑵𝒐𝒘 ... ... .... 🟢 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑾𝑨𝑻𝑪𝑯 𝑵𝑶𝑾 🔴 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑫𝒐𝒘𝒏𝒍𝒐𝒂𝒅 𝑵𝒐𝒘 ... ... .... 🟢 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑾𝑨𝑻𝑪𝑯 𝑵𝑶𝑾 🔴 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑫𝒐𝒘𝒏𝒍𝒐𝒂𝒅 𝑵𝒐𝒘 ... ... .... 🟢 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑾𝑨𝑻𝑪𝑯 𝑵𝑶𝑾 🔴 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑫𝒐𝒘𝒏𝒍𝒐𝒂𝒅 𝑵𝒐𝒘 ... ... .... 🟢 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑾𝑨𝑻𝑪𝑯 𝑵𝑶𝑾 🔴 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑫𝒐𝒘𝒏𝒍𝒐𝒂𝒅 𝑵𝒐𝒘 ... ... .... 🟢 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑾𝑨𝑻𝑪𝑯 𝑵𝑶𝑾 🔴 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑫𝒐𝒘𝒏𝒍𝒐𝒂𝒅 𝑵𝒐𝒘 ... ... .... 🟢 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑾𝑨𝑻𝑪𝑯 𝑵𝑶𝑾 🔴 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑫𝒐𝒘𝒏𝒍𝒐𝒂𝒅 𝑵𝒐𝒘 ... ... .... 🟢 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑾𝑨𝑻𝑪𝑯 𝑵𝑶𝑾 🔴 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑫𝒐𝒘𝒏𝒍𝒐𝒂𝒅 𝑵𝒐𝒘 ... ... .... 🟢 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑾𝑨𝑻𝑪𝑯 𝑵𝑶𝑾 🔴 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑫𝒐𝒘𝒏𝒍𝒐𝒂𝒅 𝑵𝒐𝒘 ... ... .... 🟢 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑾𝑨𝑻𝑪𝑯 𝑵𝑶𝑾 🔴 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑫𝒐𝒘𝒏𝒍𝒐𝒂𝒅 𝑵𝒐𝒘 ... ... .... 🟢 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑾𝑨𝑻𝑪𝑯 𝑵𝑶𝑾 🔴 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑫𝒐𝒘𝒏𝒍𝒐𝒂𝒅 𝑵𝒐𝒘 ... ... .... 🟢 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑾𝑨𝑻𝑪𝑯 𝑵𝑶𝑾 🔴 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑫𝒐𝒘𝒏𝒍𝒐𝒂𝒅 𝑵𝒐𝒘 ... ... .... 🟢 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑾𝑨𝑻𝑪𝑯 𝑵𝑶𝑾 🔴 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑫𝒐𝒘𝒏𝒍𝒐𝒂𝒅 𝑵𝒐𝒘 ... ... .... 🟢 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑾𝑨𝑻𝑪𝑯 𝑵𝑶𝑾 🔴 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑫𝒐𝒘𝒏𝒍𝒐𝒂𝒅 𝑵𝒐𝒘 ... ... .... 🟢 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑾𝑨𝑻𝑪𝑯 𝑵𝑶𝑾 🔴 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑫𝒐𝒘𝒏𝒍𝒐𝒂𝒅 𝑵𝒐𝒘 ... ... .... 🟢 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑾𝑨𝑻𝑪𝑯 𝑵𝑶𝑾 🔴 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑫𝒐𝒘𝒏𝒍𝒐𝒂𝒅 𝑵𝒐𝒘 ... ... .... 🟢 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑾𝑨𝑻𝑪𝑯 𝑵𝑶𝑾 🔴 𝑪𝑳𝑰𝑪𝑲 𝑯𝑬𝑹𝑬➤🌐👉👉 𝑫𝒐𝒘𝒏𝒍𝒐𝒂𝒅 𝑵𝒐𝒘 ... .... .... ... ... ... .... .... ... ... .... ... ... .... .... ... ... ... .... .... ... ... .... ... ... .... .... ... ... .... ... ... .... .... ... ... .... ... ... .... .... ... ... .... ... ... .... .... ... ... .... ... .... ... ... .... .... ... ... ... .... .... ... ... .... ... ... .... .... ... ... .... ... ... .... .... ... ... .... ... ... .... .... ... ... .... ... ... .... .... ... ... .... ... .... ... ... .... .... ... ... ... .... .... ... ... .... ... ... .... .... ... ... .... ... ... .... .... ... ... .... ... ... .... .... ... ... .... ... ... .... .... ... ... .... ... .... ... ... .... .... ... ... .... ... ... .... .... ... ... .... ... ... .... .... ... ... .... ... ... .... .... ... ... .... ... .... ... ... .... .... ... ... ... .... .... ... ... .... ... ... .... .... ... ... .... ... ... .... .... ... ... .... ... ... .... .... ... ... .... ... ... .... .... ... ... .... ... .... ... ... .... .... ... ... ... .... .... ... ... .... ... ... .... .... ... ... .... ... ... .... .... ... ... .... ... ... .... .... ... ... .... ... ... .... .... ... ... .... ... .... ... ..... . . ### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄) . . . ### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄) . . . ### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄) . . . . . . . . . . . . . . . . . . . . . . . . ### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄) . . . ### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄) . . . ### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄) . . . . . . . . . . . . . . . . . . . . . . . . . . . . ### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄) . . . ### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄) . . . ### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄) . . . ### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄) . . . . . .
gharchive/issue
2024-12-15T08:06:38
2025-04-01T06:39:45.604230
{ "authors": [ "SudarshanDettman", "ghost", "lliansm" ], "repo": "nezhahq/nezha", "url": "https://github.com/nezhahq/nezha/issues/794", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1608228652
Check supported models and handle warnings Description of feature Check if requested alleles and lengths combinations are supported by epitope prediction tools and handle warnings/errors properly! Given in MERGE_PREDICTIONS_BUFFER.out.ch_prediction_warnings, use and output directly! Another check for the module lengths is introduced in PR #75 The peptide length will be checked for an available model and if multiple alleles are given only the matching models are used and the user is informed that the prediction will happen at reduced peptide length (e.g. for mouse alleles H2-Db and H2-Ld: H2-Db is available at length 9 and H2-Ld is available at length 8 and 9. Only peptides of length 9 will be predicted) This will be implemented as extra process. @skrakau Should I check in this process if the alleles are available and remove it from check_samplesheet_create_tables? related to https://github.com/nf-core/metapep/issues/74
gharchive/issue
2023-03-03T09:07:27
2025-04-01T06:39:45.613455
{ "authors": [ "skrakau", "tillenglert" ], "repo": "nf-core/metapep", "url": "https://github.com/nf-core/metapep/issues/48", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
433241086
Usage docs page missing Hello nf-core team, https://nf-co.re/usage_docs answers with a 404, so I'm letting you know! Well, a few pages are not there yet Nextflow tutorial Helper tools Guidelines etc. so I guess it's a work in progress. Good luck! Where did you find that url ? If I go to the main homepage, the links on top work :-) Would be interesting so we can fix the urls ! At the bottom menus: Getting Started Using nf-core Available pipelines Helper tools Nextflow tutorial Amazing that you found these - I'll update them to point to the correct subpages :-) Fixed, thanks a bunch @hmenager
gharchive/issue
2019-04-15T11:55:25
2025-04-01T06:39:45.626465
{ "authors": [ "apeltzer", "hmenager" ], "repo": "nf-core/nf-co.re", "url": "https://github.com/nf-core/nf-co.re/issues/107", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2729793583
Unify SummarizedExperiment output Description of feature Currently, the pipeline generates three SummarizedExperiment .rds files while it would be more convenient and less confusing to generate a single one with different assay layers. Current situation There are three rds files: All of which have a counts and abundance layer The abundance layer is the same in all three objects While the counts are different representing counts, length scaled counts and scaled counts, respectively. Proposed solution Create a single SummarizedExperiment with the following assays: tpm (what is currently in "abundance", but the name is more explicit) counts counts_scaled counts_length_scaled gene_lengths (what is currently in salmon.merged.gene_lengths.tsv) possibly, it could be worth adding a vst layer generated by DESeq2::vst(). Create another one for the transcripts with the same layers. @drpatelh would you accept a PR for this?
gharchive/issue
2024-12-10T11:03:27
2025-04-01T06:39:45.632699
{ "authors": [ "grst" ], "repo": "nf-core/rnaseq", "url": "https://github.com/nf-core/rnaseq/issues/1465", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
293316817
iOS 11.3 Hey, do you now if there is an issue caused by iOS 11.3 ? It seems that my homebridge doesn't work anymore since I update my iPad and iPhone to 11.3, but I don't know if it's the cause works for me I've also been having a strange issue since I updated to iOS 11.3 and tvOS 11.3 . No problem at all with any of my homekit accessories when I'm on my home network where my pi is attached, but the minute I leave the network all homebridge accessories are showing "no response". No error or anything strange looking in the homebridge log, any ideas? I removed my bridge (so all my accessories) then reinstall it, and now it works perfectly
gharchive/issue
2018-01-31T21:24:52
2025-04-01T06:39:45.644362
{ "authors": [ "Hackashaq666", "Marsgames", "zackaweed" ], "repo": "nfarina/homebridge", "url": "https://github.com/nfarina/homebridge/issues/1728", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
183938427
Callback? Hi there. How do my device "tells" homebridge some status changing? For instance, lights have been turned off directly. Could you be provide some example? Thanks. If you turn a lamp off manually, I don't see how homebridge would ever get the message, unless you had somesort of 'smartsocket' that could self monitor electrical usage. Then that would have to 'report back' via a Homebridge plugin https://www.npmjs.com/search?q=homebridge-plugin
gharchive/issue
2016-10-19T11:42:12
2025-04-01T06:39:45.645892
{ "authors": [ "FrankC2000", "lagunacomputer" ], "repo": "nfarina/homebridge", "url": "https://github.com/nfarina/homebridge/issues/858", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
325647944
tf.contrib.learn.preprocessing will be deprecated since tf.contrib.learn.preprocessing will be deprecated in future version of TensorFlow, the code in chapter 7 about the bag-of-word should be updated by using tf.data or something else. Thanks @EtoDemerzel0427 for bringing this up. I'll see what I can do. who can give me a example?
gharchive/issue
2018-05-23T10:39:13
2025-04-01T06:39:45.652779
{ "authors": [ "Cumberbatch08", "EtoDemerzel0427", "nfmcclure" ], "repo": "nfmcclure/tensorflow_cookbook", "url": "https://github.com/nfmcclure/tensorflow_cookbook/issues/130", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
229476176
Modularizar los comportamientos del bot, para que otras comunidades puedan utilizarlo Estoy enviando un ... (marque con una "x") [] Error reportado [*] Solicitud de característica, mejora. Comportamiento Actual No existe Comportamiento Esperado Se espera que el comportamiento de este bot, se pueda encapsular en módulos y/o clases, de tal manera que se puedan implementar sólo los features necesarios para cada bot y personalizarlo para cada comunidad. Ejemplo: Wengy puede (A la fecha del 17 Abril del 2017): Dar los buenos días (en ciertas condiciones). Publicar un mensaje una vez al día (en caso de que hayan nuevas entradas en el blog que se configuró en config.js. Responder mensajes en los que se les menciona o se les responde directamente. Saludar a los nuevos miembros del grupo, despedir a los que salen del grupo. Crea un gist automáticamente cuando el usuario coloca un bloque de código (bajo ciertas condiciones). Se requiere que cada una de estas características puedan ser activadas/desactivadas según la funcionalidad que se quiere con el bot, esto puede implicar mover gran parte del código a otro repo y que los bots que quieran utilizar ciertas características del core lo puedan hacer (incluso wengy). Reproducción del Problema No aplica Cual es el motivo / Caso util para cambiar el comportamiento? Un esqueleto de comportamientos para futuros bots de otras comunidades y que puedan personalizarlo. Por favor cuentamos sobre tu ambiente de desarrollo: No aplica ¿Se entiende @Villanuevand @edwinvrgs? @leocabeza, y si en vez de separar el código en distintos repositorios, más bien se utiliza un enfoque estilo Angular? Me explico: En su repo tienen esta estructura: packages/ animations/ common/ compiler-cli/ core/ ... Cada carpeta de esas es un módulo totalmente independiente, cada uno con su respectivo package.json, y pudiéndose descargar de npm. Me encanta la idea de este issue. 😃 Aunque leí otro repo, y por mi mente pasó un repo por cada feature... Pero por lo que entiendo, un repo para toda los módulos, y otro donde sea Wengy como tal y los use. ¿Es así como lo comentas? 🙄 Pues igual no hay nada definido. Y si, así como comentas es como se me ocurre @osnoser1 , aunque pensandolo bien no sé si sea la mejor manera, porque cada módulo estaría con sus propias dependencias... Ejemplo: Nosotros utilizamos node-fetch para hacer peticiones http. Entonces... ¡Qué pasa si alguien quiere utilizar axios u otra libreria para hacer peticiones http a sus proyetos? Otro caso es que pasa si alguien en vez de api.ai quieren utilizar el AI de Facebook, o el de Microsoft... Entonces no sé que tan mantenible sea, a la larga, crear una interfaz para esas funcionalidades. Quizás en el proyecto de Angular sea un poco distinto, porqué ya tienen todas las dependencias predefinidas. Es importante que se deben comentar todas las funciones con jsdoc junto con este issue, y también a mi parecer, agregar pruebas unitarias a todo Quizás esto sería una buena opción: https://github.com/lerna/lerna Tenemos que hacer un roadmap y hacer los cambios poco a poco @dianjuar Tell me more! Qué les parece si en vez de implementar una arquitectura desde 0 utilizamos una ya estructurada, conocida y documentada? https://nestjs.com/ A partir de ahí podemos implementar nuestro propias abstracciones, patrones o diseños. Yo creo que es mucha complejidad añadida @dianjuar, ¿Y si dejamos todo configurable para que el que quiera utilizar wengy-ven solo le haga fork y ya?
gharchive/issue
2017-05-17T20:24:29
2025-04-01T06:39:45.829574
{ "authors": [ "Villanuevand", "dianjuar", "leocabeza", "osnoser1" ], "repo": "ngVenezuela/wengy-ven", "url": "https://github.com/ngVenezuela/wengy-ven/issues/74", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
67641760
WSGI.py (or other WSGI Confguration) I noticed that this repo had a wsgi.py which would be useful for deploying with wsgi on Apache https://github.com/jaycrossler/geoq-django Understand that the project may favor nginx, but would it be possible to supply a default wsgi configuration for multithreaded wsgi deployment on Apache? We actually have used Apache / wsgi for other implementations of the system. I'll see if we have a default configuration we can make available
gharchive/issue
2015-04-10T17:18:25
2025-04-01T06:39:45.832126
{ "authors": [ "Prindle19", "stephenrjones" ], "repo": "ngageoint/geoq", "url": "https://github.com/ngageoint/geoq/issues/308", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
158760001
GEOWAVE-346: landsat8 utilities from commandline tool apologies for the fairly large diff - it includes general raster improvements working through landsat 8 as the use case this commit history got messed up...closing this and opening a new clean PR
gharchive/pull-request
2016-06-06T19:52:55
2025-04-01T06:39:45.833326
{ "authors": [ "rfecher" ], "repo": "ngageoint/geowave", "url": "https://github.com/ngageoint/geowave/pull/820", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
720253224
postgresql secure upstream I am trying to secure a postgresql behind nginx. I've followed the documentation from here (https://docs.nginx.com/nginx/admin-guide/security-controls/securing-tcp-traffic-upstream/) the config file looks like this: stream { upstream postgres { server localhost:5432; } server { listen 443; server_name db.myservername.com; proxy_pass postgres; proxy_ssl on; ssl_certificate SSL_CERT; ssl_certificate_key SSL_KEY; ssl_trusted_certificate SSL_CHAIN_CERT; } } Nginx complains that conf.d is not the correct path for this kind of file. Could you tell me if that can even work with nginx-le ? nginx-le doesn't do any magic. All it does is LE certificate update and replacement of SSL_CERT, SSL_CERT and SSL_CHAIN_CERT with those cert paths. And the file you mapped to /etc/nginx/service.conf will be copied to /etc/nginx.d/conf/ and this about it. You can see what it made by yourself inside the container, i.e. docker exec -it nginx cat /etc/nginx/conf.d/service.conf. As long as your source service.conf valid the result should be valid as well, so whatever works with nginx will be working fine with nginx-le because nginx-le is the nginx. Thanks for your reply. With a bit of hacking I managed to make it work (but I finally realized that I could not do SSL termination for that TCP stream as it requires NGINX Plus (see https://docs.nginx.com/nginx/admin-guide/security-controls/terminating-ssl-tcp/). Can you please post what you did at least briefly? Otherwise somebody will find this thread and will not be able to learn about how to do what you did. Sure, in the docker-compose.yml, I add a volume like so: - ./path_to/service_db.conf:/etc/nginx/stream.d/service_db.conf and the content of service_db.conf (see the hardcoded value for ssl_certificate as nginx-le script does not take care of this) upstream postgres { server localhost:5432; } server { listen 5433; proxy_pass postgres; proxy_ssl on; ssl_certificate /etc/nginx/ssl/le-crt.pem; ssl_certificate_key /etc/nginx/ssl/le-key.pem; ssl_trusted_certificate /etc/nginx/ssl/le-chain-crt.pem; ssl_protocols TLSv1.2 TLSv1.3; ssl_ciphers 'ECDHE-ECDSA-CHACHA20-POLY1305:ECDHE-RSA-CHACHA20-POLY1305:ECDHE-ECDSA-AES128-GCM-SHA256:ECDHE-RSA-AES128-GCM-SHA256:ECDHE-ECDSA-AES256-GCM-SHA384:ECDHE-RSA-AES256-GCM-SHA384:DHE-RSA-AES128-GCM-SHA256:DHE-RSA-AES256-GCM-SHA384:ECDHE-ECDSA-AES128-SHA256:ECDHE-RSA-AES128-SHA256:ECDHE-ECDSA-AES128-SHA:ECDHE-RSA-AES256-SHA384:ECDHE-RSA-AES128-SHA:ECDHE-ECDSA-AES256-SHA384:ECDHE-ECDSA-AES256-SHA:ECDHE-RSA-AES256-SHA:DHE-RSA-AES128-SHA256:DHE-RSA-AES128-SHA:DHE-RSA-AES256-SHA256:DHE-RSA-AES256-SHA:ECDHE-ECDSA-DES-CBC3-SHA:ECDHE-RSA-DES-CBC3-SHA:EDH-RSA-DES-CBC3-SHA:AES128-GCM-SHA256:AES256-GCM-SHA384:AES128-SHA256:AES256-SHA256:AES128-SHA:AES256-SHA:DES-CBC3-SHA:!DSS'; ssl_prefer_server_ciphers on; ssl_dhparam /etc/nginx/ssl/dhparams.pem; ssl_ecdh_curve secp384r1; ssl_session_timeout 10m; } Maybe one enhancement of neginx-le script could be to retrieve all files that look like stream_*.conf and replace certificate path the same way as service*.conf and move the files to stream.dfolder. @vtoupet thanks for the idea, will be done in #48. I believe it was proposed some time ago in another issue but we didn't pick it up that time.
gharchive/issue
2020-10-13T13:29:12
2025-04-01T06:39:45.849010
{ "authors": [ "paskal", "umputun", "vtoupet" ], "repo": "nginx-le/nginx-le", "url": "https://github.com/nginx-le/nginx-le/issues/45", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
365322254
Model developers to tune architecture With "Efficient Neural Architecture Search via Parameter Sharing" Planned major changes To better support architecture tuning with ENAS, I'm planning changes to Rafiki's current model training framework: Replacing budget option MODEL_TRIAL_COUNT with TIME_HOURS Context Currently, when application developers create model training jobs, they pass a budget like { 'GPU_COUNT': 1, 'MODEL_TRIAL_COUNT': 20 }, with MODEL_TRIAL_COUNT deciding the no. of trials to conduct for each model template. Change Replace MODEL_TRIAL_COUNT option with TIME_HOURS option, which specifies how long the train job should run for. It is a soft time target. At the same time, I'll be reworking the Advisor component (which proposes trials' knobs) such that it is additionally in charge of deciding how many trials to run, when to stop each worker, when to stop the train job, given the budget e.g. GPU_COUNT and TIME_HOURS. Reasons for change May not be intuitive to the application developer to specify no. of trials while creating a train job ("how many trials should I put as budget? how long do I need to wait?"), especially if they're not supposed to be familiar with details like how model are trained and tuned. In contrast, TIME_HOURS is more straightforward. Currently, different models & model tuning strategies would require different no. of trials to be effective. For example, the original ENAS tuning strategy requires maybe (301x150+10+1) trials for sufficient train-eval cycles. In the future, it gives more flexibility for model tuning strategies at the Advisor component - for example, I'll be adding a new type of tuning strategy that takes all the models with no hyperparameters (e.g. model's knob config only consists of fixed values) and just conducts a single trial (since there's nothing to tune). It's also possible that a new tuning strategy can situationally conduct more/fewer trials based on feedback from workers. Introducing PolicyKnob Motivation I have been integrating ENAS as a new model tuning strategy on Rafiki (e.g. at the Advisor component). If model templates want to do architecture tuning with ENAS, the model's training code needs to switch between different "modes": During the ENAS architecture search phase, the model needs to alternate between "train my parameters for 1 epoch" and "don't train my parameters; just evaluate on the validation dataset" At the end of the architecture search, the model needs to switch to training its parameters from scratch with a full-sized architecture stacked with more cells, and train for 310 epochs Similarly, when you think about a standard hyperparameter tuning procedure, you might want the model to do early-stopping for the first e.g. 100 trials, then conduct a final trial for a full e.g. 300 epochs. In both architecture tuning & hyperparameter tuning, the model needs to be configured by Rafiki somehow to switch between these "modes" on a trial-basis. Change We can model the configuration of a model template for different training "modes" with different model policies. For example, if a model is to engage in policy QUICK_TRAIN, it is to prematurely speed up its training step e.g. by either doing early-stopping or reducing the no. of epochs. The model communicates to Rafiki which policies it supports by adding PolicyKnob(policy_name) to its knob_config. On the other hand, Rafiki configures the activation of the model's policies on a trial-basis by realising the values of PolicyKnobs to either True (activated) or False (not activated). For example, here is a example knob config of a model which supports the policy QUICK_TRAIN: Whenever the model is to do early-stopping, Rafiki will pass quick_train=True as part of the model's knobs. Otherwise, the model defaults to full-length training. Here is my current documentation for PolicyKnob: ''' Knob type representing whether a certain policy should be activated, as a boolean. E.g. the `QUICK_TRAIN` policy knob decides whether the model should stop model training early, or not. Offering the ability to activate different policies can optimize hyperparameter search for your model. Activation of all policies default to false. ===================== ===================== **Policy** Description --------------------- --------------------- ``SHARE_PARAMS`` Whether model supports parameter sharing ``QUICK_TRAIN`` Whether model should stop training early in `train()`, e.g. with use of early stopping or reduced no. of epochs ``SKIP_TRAIN`` Whether model should skip training its parameters ``QUICK_EVAL`` Whether model should stop evaluation early in `evaluate()`, e.g. by evaluating on only a subset of the validation dataset ``DOWNSCALE`` Whether a smaller version of the model should be constructed e.g. with fewer layers ===================== ===================== ''' @nudles I have added some details & reasoning on the major changes I'm going to make for architecture tuning. Let me know if you have any comments & advice on them! In terms budget, can we let users to configure either hours or trials? rename QUICK_TRAIN to EARLY_STOP? when will DOWNSCALE be used? Okay, I will keep MODEL_TRIAL_COUNT for backward compatibility as well. Noted on the suggestion. DOWNSCALE is used in architecture search for both NAS and ENAS. During the architecture search phase, the model constructed has fewer layers (e.g. 6 layers) to speed up the search, and its performance is somewhat a proxy of actual performance. On the other hand, at the final train phase, the final model trained from scratched is full-sized (e.g. 15 layers), which will be the one that is going to give the best performance.
gharchive/issue
2018-10-01T06:18:48
2025-04-01T06:39:45.860757
{ "authors": [ "nginyc", "nudles" ], "repo": "nginyc/rafiki", "url": "https://github.com/nginyc/rafiki/issues/50", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
503162231
performance optimisation when stage has been disposed of I realised that there were some functions still being called recursively at every frame even after removing the scene. In hope of getting rid of as much as I could, I was removing everything by doing: stage.removeAllComponents(); stage.dispose(); stage.viewer.container.innerHTML = ""; stage = null; This might not be too much of a problem in some pages, but if you happen to add and remove stages a lot (popup/modal with ngl viewer, or single-page applications) then this can build up. After adding and removing a stage 20 times (using a modified version of /examples/embedded.html) This is what I would get on the performance tab of Chrome. Every single frame I would have 20 calls to mouseObserver._listen() and 20 to viewer.animate() that would amount to about 5ms of useless use of the main thread, and memory usage growing (even if it does get caught by the garbage collector eventually). Compare to this screenshot using the code in this PR Here with have absolutely nothing happening after doing the same steps, which is way better regarding resource usage. OK, so I had installed this LGTM analysis thing on my forked repo and force-pushed to retrigger a check but I don't think it's something that needs to be done from my side. Let me know if I have to do something in order to pass the check This looks basically OK, though I'm not expert on this code, but perhaps you should protect against accidentally calling window.cancelAnimationFrame with an undefined value, and set the request back to undefined once the request is canceled. (And type frameRequest to include undefined.) but perhaps you should protect against accidentally calling window.cancelAnimationFrame This is not strictly necessary. cancelAnimationFrame does nothing when called with undefined or any invalid value. Reviewed and tested after a rebase. This works as advertised. Thanks @aurel-l for your contribution! Reviewed and tested after a rebase. This works as advertised. Thanks @aurel-l for your contribution!
gharchive/pull-request
2019-10-06T21:19:49
2025-04-01T06:39:45.867406
{ "authors": [ "aurel-l", "garyo", "ppillot", "sto3psl" ], "repo": "nglviewer/ngl", "url": "https://github.com/nglviewer/ngl/pull/698", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2059184376
Scythe off by one tick Here are my settings. It is always ready to attack when the bar is still one tick from full. Kicking and other attack cooldowns seem to work fine. Idk if it's just the red scythe or also the default scythe. I use the red scythe. This seems to be a new issue, I wasn't experiencing this when running tob last week but all this week it's been happening. It is still happening today. Scythe and Shadow are off by one tick. the bar doesn't quite fill up to the highest before you can attack again. It functions perfectly with Tbow though. If you need a video I can make one, but probably not necessary? Try mimicking the same check marks. LMK Doesn't work how I like it. I use it as an action bar that fills up and when the bar is full I am ready to attack. Here is a video showing what is wrong. Tbow attacks when bar is filled. (correct) Shadow and Scythe attack when bar is one tick away from filled. (incorrect) https://www.youtube.com/watch?v=birD-b951qA Well the magic animation starts for tumeken's at 3 seconds (and it matches the video timestamps too). Keep in mind these are projectiles. The magic animation for shadow and range animation (on rapid) for Twisted bow have different speeds, but if you don't move and stay in the same spot will end up hitting the same if the monster is right in front of you. Try this setting? I was using regular scythe and it worked for me where it would be fill to the max then empty. I will try that. Projectiles dont have anything to do with the attack animation itself. I went back and looked and I could just be wrong here. I'm sick right now so maybe my timing is just off. I'll report back later, thanks. I tried it. It's still wrong by one tick for both shadow and scythe. I was butterflying Akkha and everytime I go to attack it's one tick away from full bar. I've been using this plugin for over a thousand raids and it's never had this issue until this last week. I'm not sure if this is wrong - can you get a video that includes xp drops from shadow and tbow? And just to confirm - you didn't change "Empties Before Attack", "Fills Before Attack", or "Attack Bar Fills or Drains" between the previous behavior or now? Two more things: Can you also enable the "Show Attack Cooldown Ticks" option? Can you confirm the attack animation IDs for the scythe and shadow are 8056 and 9493, respectively?
gharchive/issue
2023-12-29T08:13:57
2025-04-01T06:39:45.881694
{ "authors": [ "Puzzle-Drops", "cdog2323", "ngraves95" ], "repo": "ngraves95/attacktimer", "url": "https://github.com/ngraves95/attacktimer/issues/31", "license": "BSD-2-Clause", "license_type": "permissive", "license_source": "github-api" }
1988616446
requirements.txt is missing The README suggests to pip install -r requirements.txt but the file seems not to have been included in this repository. The requirements.txt file has been added to the repository. Thank you, however I am getting the following error: ERROR: Cannot install -r requirements.txt (line 13), -r requirements.txt (line 4) and torch==2.0.0 because these package versions have conflicting dependencies. The conflict is caused by: The user requested torch==2.0.0 lightning 2.0.1.post0 depends on torch<4.0 and >=1.11.0 torchvision 0.15.2 depends on torch==2.0.1 On a side note (that might be an issue on my side), I am experiencing issue with some packages (pillow, scikit_learn, ...) that needs to be built from source (on Ubuntu 20.04.6 LTS) in the required versions as the wheel seems not to be available. How was the requirements.txt file generated? Maybe it would be worth updating to latest, if it can be verified that the results of the benchmark stays consistent. The requirements.txt was generated using pipreqs (https://pypi.org/project/pipreqs/). I fixed the dependency errors in the new requirements.txt and verified that all these requirements can be downloaded and installed using "pip install -r requirementx.txt". As I mentioned in the requirements.txt, "lightning" was only used for the lightning version of the ENOT baseline, which we ended up not using for the evaluation (we used the pure pytorch version). For clarity and to resolve dependency issues, I have removed this version of ENOT. In general, the benchmark code uses only basic math operations from torch and numpy. All benchmark parameters are downloaded from Google Drive. Therefore, it is not critical to have exactly the same version as long as numpy and torch perform math operations in the same way in all versions. There is plotting and metric code for some benchmark pairs in the mixtures_benchmark_visualization_eot.ipynb notebook. You can compare it with the output you get in your Python environment. The results should be the same if your torch/numpy version has the same random generators, and may be slightly different if not.
gharchive/issue
2023-11-11T00:50:43
2025-04-01T06:39:45.900515
{ "authors": [ "ngushchin", "stepelu" ], "repo": "ngushchin/EntropicOTBenchmark", "url": "https://github.com/ngushchin/EntropicOTBenchmark/issues/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1028122877
🛑 CDN is down In c688e23, CDN (https://vncdn.vn/) was down: HTTP code: 0 Response time: 0 ms Resolved: CDN is back up in fda422c.
gharchive/issue
2021-10-16T18:56:00
2025-04-01T06:39:45.903144
{ "authors": [ "nguyen-tv1" ], "repo": "nguyen-tv1/vnetwork-uptime", "url": "https://github.com/nguyen-tv1/vnetwork-uptime/issues/50", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
405668282
Documentation errors In Scenes section ‘started’ : The Scene is dormant, waiting to be started or to be destroyed. ‘stopped’ : The Scene is started. In Navigators section: During the lifetime of a Navigator it can go from ‘inactive’ to ‘inactive’ and vice versa multiple times, until it reaches the ‘destroyed’ state. In Usage section: Acorn is tactically divided in several modules to be able to separate different concerns from eachother. I'm sorry I cannot create a pull request at the moment. Thanks! No problem, I will update them :+1: Fixed!
gharchive/issue
2019-02-01T11:47:35
2025-04-01T06:39:45.906873
{ "authors": [ "jamolkhon", "nhaarman" ], "repo": "nhaarman/Acorn", "url": "https://github.com/nhaarman/Acorn/issues/115", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1926111087
support for netstandard2.0 This should enable using the library with .NET framework targets as well. Went over the unit tests (with api key and developer mode) and built the doc again to verify things are still OK. remark: setting as work in progress as I'm facing some issue with transitive FSharp.Data dependency when running samples under fsianycpu.exe (the .NET Framework version of FSI). It is actually working, it is just necessary to reference the same version of FSharp.Data (4.2.4 as of now) in order for things to bind correctly. I think this can be merged, it works with .NET Framework FSI now. I'm facing some issue with transitive FSharp.Data dependency when running samples under fsianycpu.exe (the .NET Framework version of FSI). I don't recall if there is an issue using netstandard or not. There's weirdness using FSharp.Data as a Library for parsing Json as we are here. I thought targetting net5.0/6.0 was because of a prior FAKE dependency, but maybe there are issues. FYI I pushed a commit fixing the github action workflow workflow. I don't recall if there is an issue using netstandard or not. There's weirdness using FSharp.Data as a Library for parsing Json as we are here. I thought targetting net5.0/6.0 was because of a prior FAKE dependency, but maybe there are issues. I was thinking to make a separate PR to remove FAKE dependencies from paket.dependencies (and running paket install again), do you want me to adjust it in this PR? The error I'm facing is, I think, because FSharp.Data split the types in separate assembly, and as of 4.2.4 it still expects everything under a single assembly. We should focus on upgrading FSharp.Data to the earliest version that introduced the split, but this would require a bit of research. FYI I pushed a commit fixing the github action workflow workflow. No problem! it is just necessary to reference the same version of FSharp.Data (4.2.4 as of now) I think #r nuget doesn't pick up dependency restrictions properly (I've had this issue elsewhere). You have to manually specify versions. Anyway, thanks for putting this together! Confirming it works with 4.2.10, so it doesn't have to be the exact same version, just the types have to be in same assembly as the library is compiled for. The earlier version of FSharp.Data that bring the split seems to be 5.0.1: https://www.nuget.org/packages/FSharp.Data.Json.Core/#versions-body-tab So my suggestion, on a next major update is to update to 5.0.1, and people should be able to use any later release, it would simplify the `#r "nuget: FSharp.Data" to not have to bake the version. I was thinking to make a separate PR to remove FAKE dependencies from paket.dependencies (and running paket install again), do you want me to adjust it in this PR? It would be excellent if you removed them here. The error I'm facing is, I think, because FSharp.Data split the types in separate assembly, and as of 4.2.4 it still expects everything under a single assembly. We should focus on upgrading FSharp.Data to the earliest version that introduced the split, but this would require a bit of research. About that time they merged a breaking change to HTTP/Json parsing which is the problem. Phillip knew it was breaking when he merged it but our usage here is a bit of an edge case. I think 5.0.2 is the last compatible FSharp.Data before the break. Happy for you to update it in this PR if you'd like, but I'm happy to merge without this change. It seems non trivial to update it without breaking documentation generation for now, I'll try to pay a look at another time. 5.0.2 seems to be same as the 4.2.10 based on the release note, best for now is to keep it as it is. Thank you @smoothdeveloper for providing this improvement! It should show up on nuget soon as v0.2.0. Great, will be able to use it out of box off nuget then :) Thanks for the great package!
gharchive/pull-request
2023-10-04T12:30:23
2025-04-01T06:39:45.923440
{ "authors": [ "nhirschey", "smoothdeveloper" ], "repo": "nhirschey/FSharp.Data.Fred", "url": "https://github.com/nhirschey/FSharp.Data.Fred/pull/17", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }