id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
973440272
Update LICENSE 2021 fixes # Checklist: [ ] Add tests that demonstrate the correct behavior of the change. Tests should fail without the change. [ ] Add or update relevant docs, in the docs folder and in code. [ ] Add an entry in CHANGES.rst summarizing the change and linking to the issue. [ ] Add .. versionchanged:: entries in any relevant code docs. [ ] Run pre-commit hooks and fix any issues. [ ] Run pytest and tox, no tests failed. Thanks for trying to contribute, but I think we don't currently change the copyright year that appears in these license files.
gharchive/pull-request
2021-08-18T09:02:20
2025-04-01T04:35:27.849875
{ "authors": [ "jab", "masguit42" ], "repo": "pallets/jinja", "url": "https://github.com/pallets/jinja/pull/1483", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
172428764
疑似内存泄露的现象 测试用例: SpritePrefab引用了一张贴图A,在Lua端执行如下代码 function main() local prefab = Resources.Load('SpritePrefab',GameObject) Resources.UnloadUnusedAssets() end 利用Unity的Profiler内存分析 可以看到,即使在调用UnloadUnusedAssets后, 贴图A仍有静态引用。 但是在c#端运行同样的代码,则贴图A会得到正常释放。。 知道问题了, 首先要调用lua的 collectgarbage() 释放lua的引用, 再调用 Resources.UnloadUnusedAssets()才可生效.
gharchive/issue
2016-08-22T11:23:15
2025-04-01T04:35:27.943401
{ "authors": [ "wlgys8" ], "repo": "pangweiwei/slua", "url": "https://github.com/pangweiwei/slua/issues/154", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2406486694
Correct the target and actor in Slack Audit log UserPrivilegeEscalation plus clean up Background Updating the Slack Audit log UserPrivilegeEscalation to correctly user target instead of actor. We got alerts for our IT team promoted to Admin or Owner. Realized it was an error in the alert and corrected it. Changes Updated the Slack UserPrivilegeEscalation alert to correctly differentiate between the actor and target. Actor being the user making the change and target being the user, the chances are occurring on. Used the already definded dict USER_PRIV_ESC_ACTIONS for alert Titles Updated alert Titles to correctly reference the target instead of the actor. Add local variable action to improve readability. Updated tests to test the new behavior Updated tests to be YAML Formatted YAML file with Prettier Testing pipenv run panther_analysis_tool test --skip-disabled-tests --sort-test-results @bmbeverst thanks for the PR! Since you've provided logic for identifying the actor and the entity, we figured we could update the titles to be even more descriptive of the event. However, I'm concerned that some of our unit test events are missing information - in particular, I'm curious if a permission_assigned event indicates what permissions were granted. Do you have access to one of the events, and could you share the structure so we can add update the unit test and title? @ben-githubs That would be great! I am back from vacation. Unluckily, I do not have any permission_assigned events, search for the last 3 months. And I don't have a test slack workspace to create a test event with. There appears to be a few sets of permissions as well, roles and account types. Understood Brooks! I'll make a note to followup on those tests at a later date, but otherwise we'll merge this to avoid holding the logic changes up!
gharchive/pull-request
2024-07-12T23:32:37
2025-04-01T04:35:27.987131
{ "authors": [ "ben-githubs", "bmbeverst" ], "repo": "panther-labs/panther-analysis", "url": "https://github.com/panther-labs/panther-analysis/pull/1288", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
228091357
Question: League Standings I hope this is the proper place to ask a question. If not please forgive me. I have been playing around with mlbgame and am looking for a way to generate the current standings by league and division. Is this at all possible? Example: American League East 1 Yankees 21-10 2 Orioles 21-11 3 Red Sox 17-16 4 Rays 17-19 5 Blue Jays 13-21 ...... It is possible to get the information as a json payload, but not currently a part of the mlbgame code base. It could be added as a new feature, I'm sure. If I get some spare time, I'll look making a pull request if @panzarino doesn't get to it first. @digitalSaint I would definitely accept a PR with this functionality. Should this issue be closed? @bradleyhurley I'm going to close the issue once standings are implemented into an official release (which should be coming soon). Closing even though the release has not happened yet (it is coming soon).
gharchive/issue
2017-05-11T19:08:52
2025-04-01T04:35:28.029471
{ "authors": [ "bradleyhurley", "chrissmithto", "digitalSaint", "panzarino" ], "repo": "panzarino/mlbgame", "url": "https://github.com/panzarino/mlbgame/issues/31", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
860652860
Fix KeyError in calculate_splits.py key is "CID" not "cid" Thanks for the catch! Thanks for the catch! You're welcome! But really it was Python that caught it and gave the error message, and since it's correct elsewhere in that line it was a pretty easy fix. Remind me to ramble about the bugs I fixed in NetHack sometime
gharchive/pull-request
2021-04-18T12:29:23
2025-04-01T04:35:28.031050
{ "authors": [ "paoloruiz", "psymar" ], "repo": "paoloruiz/ootp-pt-analysis", "url": "https://github.com/paoloruiz/ootp-pt-analysis/pull/4", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
842025875
Enhancement: Filter by name I have about 100 queues/topics in a single ServiceBus and it takes a while to navigate. A text box above the queues/topics tree for filtering items would be great. If I were to submit a PR would this be something you'd consider including? PRs are almost always most welcome, but have you seen the existing filter capability? No, I was unable to find the option. Thanks! I see, it filters on the server. That's a shame, I wanted more than StartsWith because all of my queues/topics end with different suffixes that I want to filter by :( There was an issue about the same functionality that you are wishing for, #452. @mrpmorris, I think it would be great if you created a PR for that. I am sure @SeanFeldman and @paolosalvatori agrees. I like your idea of having a textbox above where the filter would be visible. Also, as far as I know, the Startswith functionality does not exist in the latest SDK for Service Bus so if you build something that does not rely on StartsWith it would make it easier to migrate it to the latest SDK in the future. I don't think the service bus api supports any kind of "contains" @mrpmorris when I created the Service Bus Explorer years ago I provided the ability to specify filters in the Connect form and context menus. I used and implemented the filters that were available at that time. It looks that nothing has changed since then at least in the old client SDK Microsoft.ServiceBus.dll. For example, the NamespaceManager.GetQueues(String) method retrieves an enumerable collection of all queues in the service namespace with the specified filter. You can filter by a combination of entity name (including starts with), entity length (Gt or Lt), created, updated, and accessed time (Gt or Lt). Hence, no support for the contains predicate. The ManagementClient.GetQueuesAsync method in the new Microsoft.Azure.ServiceBus library doesn't even support the possibility to pass a string which contains a filter.- Yeah, I saw that :( Seeing as the whole list is retrieved when connecting we don't actually save anything by filtering on the server. So maybe the client could do the filtering? I think it would be great if the client did the filtering. Since it is a bit confusing with three SDKs for Service Bus here is a link to what I consider the preferred method for getting queues in the latest SDK. It might be quite a task to replace the existing filtering with something based on that method, but it would be highly appreciated. @ErikMogensen yes we can certainly implement filtering on the client-side, but there are customers that have namespaces with hundreds not say thousands of queues and topics and in that case, it would be useful to have the filtering capability on the server-side to avoid downloading a ton of entities from the cloud 😢 @paolosalvatori, I think most of the time is spent retrieving the runtime information for the queues, so client-side would be almost as effective. Also, if we don't accept client-side filtering we won't be able to migrate to the latest Service Bus SDK unless the Service Bus PG implements server-side filtering. @paolosalvatori As soon as you connect, the client downloads all of the queues + topics anyway so that time is already spent. Filtering on that list in memory would be quicker than asking the server to filter the list in memory and then return it over the network. Or am I missing something obvious? @mrpmorris well, the client does not download queues and topics data unless you explicitly call the GetQueues and GetTopics methods 😃 Anyway, if you guys have the bandwidth to implement client-side filtering, go with that, I'm fully supportive 😄 I had a crack at it, but the presentation and operational logic is mixed together and I found it too much work to understand & separate them. Thanks for trying @mrpmorris! How would you feel about application code being migrated to something like Fluxor so it can be completely separated from the code needed to present the data in the UI? The top of the search in https://github.com/mrpmorris/Fluxor. Are you the author of it? Could you describe the pros and cons of migrating to it. Consider that Service Bus Explorer contains about 50.000 lines of code (depends on how you count it) in Windows Forms and there are no dedicated developers. Everything is driven by voluntary effort, sometimes financed by organizations wanting a certain feature. It's a clean way to separate the operational logic out of the code you need to control UI (disabling buttons, handling clicks etc). Implementing the client-side filter would have been easy. You'd fire a FilterChangedAction(value of the input) every time the input changed - and then a single function would only include nodes that matched the filter. It wouldn't be a small amount of work. But that's not because the Flux pattern is a lot of work, but because you have so much code and it's mixed together. @mrpmorris I started the project as a sample 10 years ago when the Service Bus was actually launched. Then I migrated the project to GitHub in September 2014. Since then I've been the primary contributor, but now most of the changes come from the community. There's no direct involvement of the product group in the project, despite I'm a Microsoft employee, and I'm too busy to embark on a full rewrite of the tool. So even I agree with your idea, I doubt that we have the bandwidth for a migration.
gharchive/issue
2021-03-26T14:48:39
2025-04-01T04:35:28.042288
{ "authors": [ "ErikMogensen", "mrpmorris", "paolosalvatori" ], "repo": "paolosalvatori/ServiceBusExplorer", "url": "https://github.com/paolosalvatori/ServiceBusExplorer/issues/526", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
678812591
Add chat widget component support for Expo/React Native A lot of people want to be able to use our chat widget in Expo/React Native -- this should be possible with a few tweaks to our web widget. TODOs: Probably need to create a separate repo from https://github.com/papercups-io/chat-widget and remove all browser-specific code (e.g. anything that uses localStorage, window, or document) Remove theme UI dependency, try theme-ui-native or something else (see this thread https://github.com/system-ui/theme-ui/issues/170) for better mobile support Honestly, it may not be necessary to be using these dependencies... need to investigate a bit but if we can just swap it out that might be easier in the short term Figure out how to position widget in React Native (assuming position: fixed won't work) Render the ChatWindow component (which typically gets rendered in an iframe) in a WebView May need to add a mobile option to the ChatWindow component so that we can make design tweaks when rendering in mobile (probably need to enlarge some things?) done in https://github.com/papercups-io/chat-widget-native
gharchive/issue
2020-08-14T00:13:39
2025-04-01T04:35:28.057292
{ "authors": [ "reichert621" ], "repo": "papercups-io/papercups", "url": "https://github.com/papercups-io/papercups/issues/139", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1794481870
Total peer additions metric Describe the feature Geth has the following metric: p2p/dials/success - the total number of dials that result in a peer being added to the peerset It would be great to add this to reth as well! cc @maxwolff Additional context No response will do
gharchive/issue
2023-07-07T22:28:12
2025-04-01T04:35:28.091734
{ "authors": [ "Rjected", "maxwolff" ], "repo": "paradigmxyz/reth", "url": "https://github.com/paradigmxyz/reth/issues/3667", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1818918094
Use a trait for size methods in primitive types Describe the feature Many primitive types have methods which calculate a heuristic for the heap size of the type, for example: https://github.com/paradigmxyz/reth/blob/b69a18dc47255304b3943ca859dc6eb9da579b19/crates/primitives/src/block.rs#L192-L200 This should be captured in a trait, and it should be possible to derive this trait for all of these types. One crate that is interesting for this is DeepsizeOf. An unanswered question is whether or not we should use this crate, because its implementation of DeepSizeOf for Vec seems to contradict the rust docs. The implementation: impl<T> DeepSizeOf for std::vec::Vec<T> where T: DeepSizeOf { fn deep_size_of_children(&self, context: &mut Context) -> usize { // Size of heap allocations for each child self.iter().map(|child| child.deep_size_of_children(context)).sum() + self.capacity() * std::mem::size_of::<T>() // Size of Vec's heap allocation } } The Guarantees section of the rust docs state: However, the pointer might not actually point to allocated memory. In particular, if you construct a Vec with capacity 0 via Vec::new, vec![], Vec::with_capacity(0), or by calling shrink_to_fit on an empty Vec, it will not allocate memory. Similarly, if you store zero-sized types inside a Vec, it will not allocate space for them. Note that in this case the Vec might not report a capacity of 0. Vec will allocate if and only if mem::size_of::<T>()*capacity() > 0. In general, Vec’s allocation details are very subtle — if you intend to allocate memory using a Vec and use it for something else (either to pass to unsafe code, or to build your own memory-backed collection), be sure to deallocate this memory by using from_raw_parts to recover the Vec and then dropping it. If a Vec has allocated memory, then the memory it points to is on the heap (as defined by the allocator Rust is configured to use by default), and its pointer points to len initialized, contiguous elements in order (what you would see if you coerced it to a slice), followed by capacity-len logically uninitialized, contiguous elements. This leads me to believe that the correct implementation should subtract len from capacity before multiplying by std::mem::size_of::<T>(). If this is true, then the crate may slightly overestimate the memory used by a Vec<T>. Ultimately, this is not a huge problem, and using DeepSizeOf would be the easiest and most straightforward way to derive this size heuristic, and we could upstream a fix if required. Our implementation also does not subtract the len from capacity, so using the DeepSizeOf crate should also match our existing logic. Additional context No response Can I work on this? I checked out the DeepsizeOf crate and it also notifies that it does not count the allocator's overhead, which I assume is something hard to estimate correctly. I guess since it also overestimates sometimes, it might output a balanced result.
gharchive/issue
2023-07-24T18:26:33
2025-04-01T04:35:28.101895
{ "authors": [ "Rjected", "altugbakan" ], "repo": "paradigmxyz/reth", "url": "https://github.com/paradigmxyz/reth/issues/3893", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2395979809
chore: bump revm 11 new revm release with new call schemes closing in favor of #156
gharchive/pull-request
2024-07-08T15:44:01
2025-04-01T04:35:28.102983
{ "authors": [ "mattsse" ], "repo": "paradigmxyz/revm-inspectors", "url": "https://github.com/paradigmxyz/revm-inspectors/pull/159", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
700942561
version and repo tag Hello, The package on pip is versioned 0.2.0 whereas the lastet tag / __version__ in pyls_isort/__about__.py in master is 0.1.1 Nice catch, I forgot to push... Should be okay now https://github.com/paradoxxxzero/pyls-isort/releases/tag/0.2.0 perfect, thanks!
gharchive/issue
2020-09-14T09:42:29
2025-04-01T04:35:28.105018
{ "authors": [ "fesaille", "paradoxxxzero" ], "repo": "paradoxxxzero/pyls-isort", "url": "https://github.com/paradoxxxzero/pyls-isort/issues/6", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
962300317
v0.2.3 has a broken Cargo.toml I'm trying to follow https://substrate.dev/docs/en/tutorials/create-your-first-substrate-chain/setup but my build fails with: $ cargo build --release error: failed to download `parity-db v0.2.3` Caused by: unable to get packages from source Caused by: failed to parse manifest at `/home/bear/.cargo/registry/src/github.com-1ecc6299db9ec823/parity-db-0.2.3/Cargo.toml` Caused by: failed to parse the version requirement `0.11 ` for dependency `parking_lot` Caused by: expected comma after minor version number, found '\t' I downloaded parity-db==0.2.3 via cargo-download and found that Cargo.toml indeed has an extra \t on the parking_lot declaration: # THIS FILE IS AUTOMATICALLY GENERATED BY CARGO # # When uploading crates to the registry Cargo will automatically # "normalize" Cargo.toml files for maximal compatibility # with all versions of Cargo and also rewrite `path` dependencies # to registry (e.g., crates.io) dependencies # # If you believe there's an error in this file please file an # issue against the rust-lang/cargo repository. If you're # editing this file be aware that the upstream Cargo.toml # will likely look very different (and much more reasonable) [package] edition = "2018" name = "parity-db" version = "0.2.3" authors = ["Parity Technologies <admin@parity.io>"] description = "Key-value database for the blockchain" homepage = "https://substrate.dev" license = "MIT OR Apache-2.0" repository = "https://github.com/paritytech/parity-db/" [profile.release] lto = "fat" codegen-units = 1 debug = true panic = "abort" [dependencies.blake2-rfc] version = "0.2.18" [dependencies.crc32fast] version = "1.2.0" [dependencies.fs2] version = "0.4.3" [dependencies.hex] version = "0.4.2" [dependencies.libc] version = "0.2" [dependencies.log] version = "0.4.8" [dependencies.memmap2] version = "0.2" [dependencies.parking_lot] version = "0.11\t" [dependencies.rand] version = "0.8.2" [dev-dependencies.env_logger] version = "0.8.2" 0.2.2 has it too ```error: failed to parse manifest at /cargo-home/registry/src/github.com-1ecc6299db9ec823/parity-db-0.2.2/Cargo.toml Caused by: failed to parse the version requirement 0.11 for dependency parking_lot Caused by: expected comma after minor version number, found '\t'``` This was fixed in 0.2.4. Try using the v3.0.0+monthly-2021-07 tag when following the tutorial. @arkpar Can you please do it for every package that has this one as dependency? I don't have it directly, it's somewhere in 3rd party module. @Atomosk The only package that directly depends on this is sc-client-db. It is going to be published with other substrate crates, but I'm not sure when. In any case, you can update your local dependency with cargo update -p parity-db even if it is indirect. Or use patch section in Cargo.toml @Atomosk just FYI Substrate Recipes is broken with the same problem. Reproduced it on both master and v3.0.0 tag. cargo update -p parity-db solved the problem.
gharchive/issue
2021-08-06T00:48:36
2025-04-01T04:35:28.194845
{ "authors": [ "Atomosk", "arkpar", "bernardoaraujor" ], "repo": "paritytech/parity-db", "url": "https://github.com/paritytech/parity-db/issues/39", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
928172089
polkadot-launch cannot register parachain to rococo-local of polkadot-v0.9.5 it works fine in polkadot-v0.9.4, but the registration doesn't work anymore in polkadot-v0.9.5 close as the master + #117 worked for me
gharchive/issue
2021-06-23T11:50:54
2025-04-01T04:35:28.198765
{ "authors": [ "GopherJ" ], "repo": "paritytech/polkadot-launch", "url": "https://github.com/paritytech/polkadot-launch/issues/116", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1563373150
ensure_none naming is misleading FRAME System's ensure_none ensures that the origin o represents an unsigned extrinsic. This naming is consistent with the inner member of the RawOrigin enum: pub enum RawOrigin<AccountId> { Root, Signed(AccountId), None, } During a Milestone Review for the Substrate Builders Program I had to clarify a team that was misled by the ensure_none naming. The team had interpreted that ensure_none meant no one could call this extrinsic, which is literally the opposite of the meaning we're aiming to convey. Even though the documentation already provides an accurate description, we could further avoid confusions by renaming ensure_none to ensure_unsigned. Hey @bernardoaraujor , I would like to take this up.
gharchive/issue
2023-01-30T23:32:22
2025-04-01T04:35:28.274532
{ "authors": [ "bernardoaraujor", "gitofdeepanshu" ], "repo": "paritytech/substrate", "url": "https://github.com/paritytech/substrate/issues/13279", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
836775466
frame contracts: expected associated type, found enum pallet::Event i just tried building from latest 'master' branch and got the following error: error[E0308]: mismatched types --> /Users/me/code/paritytech/substrate/frame/contracts/src/exec.rs:914:30 | 914 | <T as Config>::Event::from(event).into(), | ^^^^^ expected associated type, found enum `pallet::Event` | = note: expected associated type `<T as frame_system::Config>::Event` found enum `pallet::Event<T>` = help: consider constraining the associated type `<T as frame_system::Config>::Event` to `pallet::Event<T>` or calling a method that returns `<T as frame_system::Config>::Event` = note: for more information, visit https://doc.rust-lang.org/book/ch19-03-advanced-traits.html Do you use latest stable? Do you use latest stable? i actually tried using the latest 'master', maybe i should go back to using the latest stable :-) i just tried building the latest stable release https://github.com/paritytech/substrate/releases/tag/v3.0.0 at commit 49a4103f4bfef55be20a5c6d26e18ff3003c3353 git fetch upstream v3.0.0:v3.0.0 git checkout v3.0.0 cargo build --release but it gives the following error. maybe it's because i'm using an old version of rustc nightly-2020-10-06-x86_64-apple-darwin Compiling pallet-offences-benchmarking v3.0.0 (/Users/ls2/code/paritytech/substrate/frame/offences/benchmarking) error[E0308]: mismatched types --> frame/offences/benchmarking/src/lib.rs:288:6 | 288 | pallet_balances::Event::<T>::Endowed(reporter, (reward_amount / r).into()) | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected associated type, found enum `pallet_balances::Event` | = note: expected associated type `<T as frame_system::Config>::Event` found enum `pallet_balances::Event<T>` = help: consider constraining the associated type `<T as frame_system::Config>::Event` to `pallet_balances::Event<T>` = note: for more information, visit https://doc.rust-lang.org/book/ch19-03-advanced-traits.html error: aborting due to previous error For more information about this error, try `rustc --explain E0308`. error: could not compile `pallet-offences-benchmarking` i was using a clone of an old fork i made of the substrate repo, and i'd updated it locally to the latest substrate master, then switched back to v3.0.0. so i've now deleted that clone, and recloning directly from substrate's repo, then i'll switch to v3.0.0 branch, and try building that hmm, we'll it just finished building after a fresh clone , and i got the same error.
gharchive/issue
2021-03-20T10:32:57
2025-04-01T04:35:28.279168
{ "authors": [ "bkchr", "ltfschoen" ], "repo": "paritytech/substrate", "url": "https://github.com/paritytech/substrate/issues/8408", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
524028459
Bump dependencies, respecting semver We really need a bot to make these PRs, but in the meantime, I can do so with almost no effort. ✄ ----------------------------------------------------------------------------- closing since it's now stale. this really needs to be done and merged immediately upon CI pass.
gharchive/pull-request
2019-11-17T19:04:23
2025-04-01T04:35:28.280865
{ "authors": [ "DemiMarie-parity", "gavofyork" ], "repo": "paritytech/substrate", "url": "https://github.com/paritytech/substrate/pull/4117", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
363910527
JDK 11 compatibility Nassau does not currently compile on JDK 11, most likely due to the Error Prone usage. This was fixed in 805ccc3184061964c94dbf1580dd15f9a9c9a0a9.
gharchive/issue
2018-09-26T08:22:34
2025-04-01T04:35:28.291053
{ "authors": [ "jvirtanen" ], "repo": "paritytrading/nassau", "url": "https://github.com/paritytrading/nassau/issues/52", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
297718992
XLC : Unable to execute in single thread (spawns multiple threads by default) XLC runs workload on many CPUs despite of no parallel tasks in code. GCC executes in single CPU only but still gives better performance than XLC. CPU pinning gives poorest performance to XLC here. https://github.com/parthsl/TSP_XeonPhi/blob/79c0642083fcb871fb7b2062d8ef5408b35c08f1/two_opt/Parallel/hill_climb/src/hill_climb.c#L18-L23 Calling OpenMP threads before the call to two_opt_max_swap_single() function results in XLC to spawn multiple threads. Removing every OpenMP calls from the code results in XLC to call for one thread only and gives better performance compared to gcc.
gharchive/issue
2018-02-16T09:09:04
2025-04-01T04:35:28.469701
{ "authors": [ "parthsl" ], "repo": "parthsl/TSP_XeonPhi", "url": "https://github.com/parthsl/TSP_XeonPhi/issues/11", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2748227357
[profile]: đổi lại ảnh avatar ảnh avatar đang dùng của quizlet, cần đổi lại Đã xong
gharchive/issue
2024-12-17T15:30:57
2025-04-01T04:35:28.473616
{ "authors": [ "nqmgaming" ], "repo": "pass-with-high-score/quick_mem", "url": "https://github.com/pass-with-high-score/quick_mem/issues/103", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1911405924
Make ApiSecret a required property If a property is not nullable and does not have a default value, then it must be marked as required. That way, if the user doesn't set it, it becomes a compile-time error and not a run-time error. @Tyrrrz I don't have any feedback of value here - I know way to little about polyfilling. I trust you that this works. @justindbaur do you have any thoughts? 🙂 We are validating that ApiSecret is non-null through: https://github.com/passwordless/passwordless-dotnet/blob/3d841dbe021e2d30382b0c0253017618b623b6f0/src/Passwordless/ServiceCollectionExtensions.cs#L16 but this doesn't happen until first construction of IOptions<PasswordlessOptions>. I'm guessing Activator.CreateInstance<PasswordlessOptions>() doesn't respect the required? Yeah, it's validating at run-time, this adds compile-time validation. It won't help in the builder overload but will help in those where PasswordlessOptions are provided directly, including in internal code.
gharchive/pull-request
2023-09-25T12:29:36
2025-04-01T04:35:28.476710
{ "authors": [ "Tyrrrz", "abergs", "justindbaur" ], "repo": "passwordless/passwordless-dotnet", "url": "https://github.com/passwordless/passwordless-dotnet/pull/40", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
679266251
How do I style this? I'd like to add a bit of css to style this component. What's the best way of achieving this? This is just a wrapper for the leaflet minimap plugin. The minimap plugin has several CSS classes added you can use to style it any way you wish. The top level CSS class is leaflet-control-minimap. You can see the other CSS classes available by inspecting the DOM in your browser dev tools. If you want to include the default minimap CSS styling you can add it by loading the CSS file in a SFC style tag, or loading it separately. <style> @import '~leaflet-minimap/dist/Control.MiniMap.min.css'; </style>
gharchive/issue
2020-08-14T16:24:31
2025-04-01T04:35:28.496199
{ "authors": [ "CodyBontecou", "patrickcate" ], "repo": "patrickcate/vue-leaflet-minimap", "url": "https://github.com/patrickcate/vue-leaflet-minimap/issues/5", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
514772775
Pfe-accordion animations on desktop and iOS Safari Pfe-accordion animations do not complete in desktop and iOS safari. Expected: the arrow should rotate 90º on state change Actual: (iOS) opening the panel works correctly, but on close the arrow is rotated incorrectly until changing the focus. Screenshot: Actual (desktop): The close animation does not complete if the mouse is hovering over the accordion's open button. Screenshot: Can you confirm if this is still an issue? Yeah, looks like it's resolved! Not seeing the issue anymore in IOS and Safari 13. Closing.
gharchive/issue
2019-10-30T15:39:13
2025-04-01T04:35:28.517899
{ "authors": [ "Djfaucette", "castastrophe" ], "repo": "patternfly/patternfly-elements", "url": "https://github.com/patternfly/patternfly-elements/issues/597", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
360350778
Type-Ahead Filter We require a Type-Ahead Filter component that supports a variety of filtering use case across products. Some requirements for this can be found here: https://docs.google.com/document/d/15w2gpUb7__4oINMtnU7hwSqtlD6cwDIvkDeHM1TMfiQ/edit?usp=sharing and also a preliminary design here: Unified Filter Design Concept: https://docs.google.com/presentation/d/1yK7HUtUQZLBZGv-NjHGuMpDTs5Iq1mTqQCggGfDb1AQ/edit?usp=sharing This should be designed and implemented to use a consistent type-ahead approach with the Type-Ahead Select (Ref #722 ). Closing this because this example will appear in Filter Demo
gharchive/issue
2018-09-14T15:25:09
2025-04-01T04:35:28.520022
{ "authors": [ "mcarrano", "rachael-phillips" ], "repo": "patternfly/patternfly-next", "url": "https://github.com/patternfly/patternfly-next/issues/731", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
355622826
feature(stylelint): add rule to match pattern for custom properties fix https://github.com/patternfly/patternfly-next/issues/668 yeah!! let's make the linter evil! 🕶 Deploy preview for pf-next ready! Built with commit a22030081000192c9ff94d8da262380039ff3eee https://deploy-preview-669--pf-next.netlify.com
gharchive/pull-request
2018-08-30T14:53:50
2025-04-01T04:35:28.521962
{ "authors": [ "andresgalante", "michael-coker", "patternfly-build" ], "repo": "patternfly/patternfly-next", "url": "https://github.com/patternfly/patternfly-next/pull/669", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
165343968
question about error not mentioned in documentation Hi, when using ProxyMockFactory to mock a trait, I get a compilation error: <from inside IntelliJ 2016.2: using scala version 2.11.7, jvm version 1.8.0_91,> I'm using Scalatest FlatSpec class in the test class not the FlatSpecLike trait. Inside the IDE, no compilation error shows up, but when the maven compile runs, it emits the following: Error: could not find implicit value for parameter mockContext: org.scalamock.context.MockContext val m = mock[TransactionProcessor] I could not see anything in the docs on the web site about implicit MockContext. Can anyone tell me how to fix the error I'm seeing? Many thanks. I think that you should read this http://scalamock.org/user-guide/sharing-scalatest/ Thanks for the reply, but I don't think it addresses the question. Where in the document link you refer to, is the implicit mock context to be found. I have, in fact, tried the style of solution referred to in to web site document, but the compilation error remains present and therefore the test cannot run. You should construct your suite the other way. Your test should look like that: class MyTest extends FlatSpec with MockFactory. Maybe my way is not correct somehow but I have never used ProxyMockFactory directly. Nevertheless, try with what I proposed you and let me know if it works. Hi Bartek Thanks again for taking the time to reply. I am trying to test functions in a trait, and mocks from MockFactory will not compile with traits. The scalamock documents indicate that when testing traits, the proxyMockFactory mock is the correct one to use. So, when I try the "ordinary" mock, there is a compiler error immediately visible when providing a trait name in the mock parameter. Thus, I have the problem described, the mock from ProxyMockFactory depends on an implicit mockContext that is not visible / reachable. Hi, In addition, I've tried explicitly creating a MockContext and that too, results in compilation errors; e.g. Error: trait MockContext in package context cannot be accessed in package org.scalamock.context implicit val mockCtx = new MockContext { override protected def newExpectationException(message: String, methodName: Option[Symbol]): this.type = ??? override type ExpectationException = this.type } Error: not found: type MockContext implicit val mockCtx = new MockContext { It looks like I'll have to give up on Scalamock, and try to find another (working) solution. Thanks for your replies. Not sure what to do with this, need a code snippet to reproduce or comment. Please try with the latest version and raise a new issue if you still have a problem. I'm running into the same problem. This is a mostly whittled-down code sample that demonstrates the error: trait SomeTrait { val hasThings = Vector("a", "b", "c") def doesStuff = println("hello") } class SomeClass { this: SomeTrait => } class TestSpec extends ProxyMockFactory with FunSpecLike { val fakeTrait = mock[SomeTrait] val sut = new SomeClass with fakeTrait } Errors are: Error:(34, 23) could not find implicit value for parameter mockContext: org.scalamock.context.MockContext val fakeTrait = mock[SomeTrait] Error:(34, 23) not enough arguments for method mock: (implicit evidence$1: scala.reflect.ClassTag[com.shiftwise.rtisystem.test.actors.SomeTrait], implicit mockContext: org.scalamock.context.MockContext)com.shiftwise.rtisystem.test.actors.SomeTrait with org.scalamock.proxy.Mock. Unspecified value parameter mockContext. val fakeTrait = mock[SomeTrait] If I had to guess, I'd assume it has to do with the self-type annotation, but I'm new to Scala so I could be doing a number of other things wrong, too. I'm using version 3.3.0 with Scalatest 3.0.0 with Scala version 2.11.8. Thanks! can you try again, with 3.4.2 please? Also, I think you should be using org.scalamock.scalatest.proxy.MockFactory instead of ProxyMockFactory I've updated to 3.4.2 and my code now looks like: trait SomeTrait { val hasThings = Vector("a", "b", "c") def doesStuff = println("hello") } class SomeClass { this: SomeTrait => } class TestSpec extends org.scalamock.scalatest.proxy.MockFactory with FunSpecLike { val fakeTrait = mock[SomeTrait] val sut = new SomeClass with fakeTrait } and the errors have changes, but it's still not compiling: Error:(34, 27) not found: type mock trait fakeTrait extends mock[SomeTrait] Error:(35, 17) illegal inheritance; self-type SomeClass with TestSpec.this.fakeTrait does not conform to SomeClass's selftype SomeClass with .SomeTrait val sut = new SomeClass with fakeTrait I realized this morning that the latter change was suggested above; sorry about that. I appreciate your patience and thank you for maintaining the project! Hi, see my example below. Not sure if you can SomeClass to SomeClass2 in your code, but I don't see a good way of mocking out the SomeClass version with the selftype, sorry. import org.scalamock.scalatest.proxy.MockFactory import org.scalatest.{FunSpec, FunSpecLike} trait SomeTrait { val hasThings = Vector("a", "b", "c") def doesStuff() = println("hello") } class SomeClass { this: SomeTrait => } trait SomeClass2 { val someTrait: SomeTrait } class Issue146Test extends FunSpec with MockFactory { describe("foo") { it("should work") { val fakeTrait = mock[SomeTrait] // does not compile //val sut = new SomeClass with fakeTrait val sut2 = new SomeClass2 { val someTrait = fakeTrait } } } } Thank you for the example; while it's a little more complicated, it isn't tooo bad. I think I've figured out how to rejigger my non-example code so that I can use the trait without selftyping and still test the functionality that uses the trait behavior, but if it doesn't work out, I'll try implementing the test based on your sample. And that looks like a fantastic blog post. I appreciate your time and hope you have a good day! :)
gharchive/issue
2016-07-13T15:02:45
2025-04-01T04:35:28.577458
{ "authors": [ "barkhorn", "bkowalik", "lkemmerer", "london-coder" ], "repo": "paulbutcher/ScalaMock", "url": "https://github.com/paulbutcher/ScalaMock/issues/146", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
463826216
Feature Request: Long-text or HTML editor column type support We have spreadsheets that often have very long text values such as titles and descriptions of images. I haven't found an obvious way to have a column wrap text or adjust the row height to display the full field value. Also, since the existing columns are based on jSuite input types, it would be great to support the mini text editor. As a work-around I'm currently considering using a separate stand-alone field to act like Excel's formula bar for long-text entry that will copy in the value of the currently active field and save the value back on a cell change. Do you know of any examples where others have done similar? Thanks! Long texts, please use wordWrap:true in the column directives. Instead of a input will be render a textarea. https://bossanova.uk/jexcel/v2/examples/text-wrapping This works in version 3 as well. In regards to the HTML editor, I will consider for future releases. Thanks for the suggestion. This has been implemented in both distributions with type: html https://bossanova.uk/jexcel/v4/examples/richtext-html-editor
gharchive/issue
2019-07-03T15:44:04
2025-04-01T04:35:28.596418
{ "authors": [ "paulhodel", "seth-shaw-unlv" ], "repo": "paulhodel/jexcel", "url": "https://github.com/paulhodel/jexcel/issues/419", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2152980130
Validate Forms We need to validate forms to make sure users input the correct type of data. We also got a list from the project partner telling us which fields are required, so we'll need to change the form to reflect that. Here is the list of fields we need to make required: Entity Name Status Industrial/Chemical Headquarters’ Country Stakeholder Description with required checking done, the only thing left to do for this is to do type validation, meaning changing fields to their proper types. currently everything is varchar, but int fields should be ints, date fields dates, etc. no longer needed
gharchive/issue
2024-02-25T22:57:19
2025-04-01T04:35:28.602910
{ "authors": [ "paullpp" ], "repo": "paullpp/hemp-db", "url": "https://github.com/paullpp/hemp-db/issues/57", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1835468976
Resolve unintentional MIT License omission Hi Paul, It appears that the subs in this repo are either: copies of the stubs in https://github.com/Josverl/micropython-stubs generated by the use of the micropython-stubber tool : https://github.com/Josverl/micropython-stubbber That in itself is not a problem at all, happy that you have been able to use them for the exact purpose that I have built these for. However both are released under MIT licence which requires that that license is respected. The relevant parts: MIT License Copyright (c) 2020-2023 Jos Verlinde The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software and stubs. So it is no problem that you place this under a different license, however you cannot remove the earlier license and notice. in the development and publication of the micropython-stubs that have used as the base, I have taken specific effort to respect , and where relevant copy over relevant licenses with the stubs. I kindly ask you to do the same. Regards, Jos Verlinde @Josverl I'm sorry that I missed that one and of course I'll add the license header to the stubber which seems to be missing in the minified copy I got from your repo some months ago. The repo does not contain copies of your stubs repo but it contains the auto-generated ones from the stubber and were than each one hand modified. I also thought about in the past to maybe close this repo and contact you if it would be possible to incorporate your stubs into the MicroPico extension. In return I would to my contributions back to your stubs repo instead of this one. My only problem was that your repo is quite big and contains many different stubs for the pico-w. Also most of them don't contain the parameters for functions and doc-strings for classes and methods or maybe I haven't found them. Paul, thanks for resolving this. really appreciated. Regards, Jos Verlinde Jos, sounds great. I thought about making the extension more suitable for different boards so giving the user the option to chose between pico, pico-w and other stubs would be a great step towards the right direction. If you agree, then I'll let him know and see how we can work together on this. Sure OK I'll ping Jim on discord. Do you have a discord handle? OK I'll ping Jim on discord. Do you have a discord handle? Yes: paulober I've let him know, but he has not responded so far
gharchive/issue
2023-08-03T17:36:42
2025-04-01T04:35:28.609813
{ "authors": [ "Josverl", "paulober" ], "repo": "paulober/Pico-W-Stub", "url": "https://github.com/paulober/Pico-W-Stub/issues/11", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
617628502
Syntax error with deps dependencies Hello! I was super excited to find this extension and installed it, but it appears there's an error with one of my dependencies: bump: desc: Bump the version deps: - task: build:generated-dir Note that this is proper syntax, and is outlined in the section on task dependencies: https://taskfile.dev/#/usage?id=task-dependencies version: '2' tasks: default: deps: - task: echo_sth vars: {TEXT: "before 1"} - task: echo_sth vars: {TEXT: "before 2"} cmds: - echo "after" echo_sth: cmds: - echo {{.TEXT}} @klauern Good catch, the schema was missing the alternative deps format. Just added it and the fix should be in the next release This should now have made it to the extension This should now have made it to the extension
gharchive/issue
2020-05-13T17:31:38
2025-04-01T04:35:28.621594
{ "authors": [ "klauern", "paulvarache" ], "repo": "paulvarache/vscode-taskfile", "url": "https://github.com/paulvarache/vscode-taskfile/issues/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2283797225
Game Scene - Layers should be stored within a singular vector/list versus the normal elements list Layers are a crucial part of the scene object and to avoid eternal typecasting up from the branch class they should be stored in a vector or list within the scene class. This way the scene class can interact with the layers(currently engine allows up to 32 per scene) so that typecasting is avoided as much as possible, and code can be easier readable. Merged via https://github.com/pawbyte/Game-Pencil-Engine-Editor/pull/149
gharchive/issue
2024-05-07T16:39:38
2025-04-01T04:35:28.641285
{ "authors": [ "nhurde" ], "repo": "pawbyte/Game-Pencil-Engine-Editor", "url": "https://github.com/pawbyte/Game-Pencil-Engine-Editor/issues/141", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2462263281
Some arguments type is missing in afterOperation hook Link to reproduction No response Payload Version 3.0.0-beta.77 Node Version 20 Next.js Version 15.0.0-canary.53 Describe the Bug Some arguments type is missing in afterOperation hook Reproduction Steps export const afterOperation: CollectionAfterOperationHook = async ({ args, operation, result, req, }) => { if (operation === 'create') { const { url, download } = args.data } } (parameter) result: unknown Property 'url' does not exist on type 'RequiredDataFromCollectionSlug' Adapters and Plugins No response Hey @smoothdvd can you pass your collection slug into the type? eg: CollectionAfterOperationHook<'posts'> Then your data there will be inferred correctly from the generated types. These should generate automatically but if not then pnpm generate:types or pnpm payload generate:types should do that for you. Also please note that on the beta versions 77+ you should be on Next canary.104 as a minimum. I tested the above in a fresh project, so let me know if this fixes your issues, otherwise please provide a reproduction so I can take a closer look! @paulpopus Thank you, it worked after I follow your instruct (I also upgraded to beta.79 and next 15.0.0-canary.104).
gharchive/issue
2024-08-13T02:44:33
2025-04-01T04:35:28.676631
{ "authors": [ "paulpopus", "smoothdvd" ], "repo": "payloadcms/payload", "url": "https://github.com/payloadcms/payload/issues/7642", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2477687786
Extend formBuilderPlugin fields: not persist data Link to reproduction Payload Version 3.0.0-beta.86 Node Version v20 Next.js Version 15.0.0-canary.121 Describe the Bug If I add a custom field in the plugin configuration, it displays correctly in the admin UI but the data is not persisted. If I add the "slug" property the data is persisted but there is a typescript error because is missing in the interface FieldConfig. Reproduction Steps Declare the custom field import { fields, formBuilderPlugin } from "@payloadcms/plugin-form-builder"; import { Block } from "payload"; export const formBuilderPluginConfig = formBuilderPlugin({ redirectRelationships: ["pages"], fields: { hidden: { // slug: "hidden", // <- With this prop works labels: { singular: 'Hidden', plural: 'Hidden' }, fields: [ ...(fields.text as Block).fields // Other fields ] } }, }) Open the Admin UI to add the custom field Save Data is not persisted Adapters and Plugins "@payloadcms/db-mongodb": "beta", "@payloadcms/email-nodemailer": "beta", "@payloadcms/next": "beta", "@payloadcms/plugin-cloud-storage": "beta", "@payloadcms/plugin-form-builder": "beta", "@payloadcms/plugin-nested-docs": "beta", "@payloadcms/plugin-seo": "beta", "@payloadcms/richtext-lexical": "beta", "@payloadcms/storage-s3": "beta", "@payloadcms/ui": "beta", @matteo-naif did you find a work around for this? running into this same issue currently. Also trying to add a custom "hidden" field to the formBuilder plugin. Hi @siobahnschuck, I paste my working example with minimal setup import { formBuilderPlugin } from "@payloadcms/plugin-form-builder"; import { fields } from "@payloadcms/plugin-form-builder"; import { FieldConfig } from "@payloadcms/plugin-form-builder/types"; import { Block } from "payload"; export const textFields = (fields.text as Block).fields export const InputHidden: FieldConfig = { // @ts-ignore slug: "hidden", labels: { singular: 'Hidden', plural: 'Hidden' }, fields: [ ...textFields // Other fields ] }; export const formBuilderPluginConfig = formBuilderPlugin({ redirectRelationships: ["pages"], fields: { hidden: InputHidden } }) The result is the following: @matteo-naif thank you so much! This works well 🌟
gharchive/issue
2024-08-21T10:35:08
2025-04-01T04:35:28.679955
{ "authors": [ "matteo-naif", "siobahnschuck" ], "repo": "payloadcms/payload", "url": "https://github.com/payloadcms/payload/issues/7787", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2592102768
CSS z-hight for lexical toolbar is off Describe the Bug The toolbar for lexical editor appears over a dropdown menu: I use PayloadCMS v3 beta-114 Link to the code that reproduces this issue npx create-payload-app@beta -t website Reproduction Steps npx create-payload-app@beta Visit http://localhost:4444/admin/collections/pages/670fcbf993abb42cdb3e4bd7 Which area(s) are affected? (Select all that apply) Not sure Environment Info Node.js v22.6.0 Binaries: Node: 22.6.0 npm: 10.9.0 Yarn: N/A pnpm: 9.7.1 Relevant Packages: payload: 3.0.0-beta.114 next: 15.0.0-canary.173 @payloadcms/db-mongodb: 3.0.0-beta.114 @payloadcms/email-nodemailer: 3.0.0-beta.114 @payloadcms/graphql: 3.0.0-beta.114 @payloadcms/live-preview: 3.0.0-beta.114 @payloadcms/live-preview-react: 3.0.0-beta.114 @payloadcms/next/utilities: 3.0.0-beta.114 @payloadcms/plugin-cloud: 3.0.0-beta.114 @payloadcms/plugin-cloud-storage: 3.0.0-beta.114 @payloadcms/plugin-form-builder: 3.0.0-beta.114 @payloadcms/plugin-nested-docs: 3.0.0-beta.114 @payloadcms/plugin-redirects: 3.0.0-beta.114 @payloadcms/plugin-search: 3.0.0-beta.114 @payloadcms/plugin-seo: 3.0.0-beta.114 @payloadcms/richtext-lexical: 3.0.0-beta.114 @payloadcms/storage-s3: 3.0.0-beta.114 @payloadcms/translations: 3.0.0-beta.114 @payloadcms/ui/shared: 3.0.0-beta.114 react: 19.0.0-rc-3edc000d-20240926 react-dom: 19.0.0-rc-3edc000d-20240926 Operating System: Platform: darwin Arch: x64 Version: Darwin Kernel Version 23.6.0: Mon Jul 29 21:13:00 PDT 2024; root:xnu-10063.141.2~1/RELEASE_X86_64 Available memory (MB): 32768 Available CPU cores: 16 Hey @exoer, I believe this to be a duplicate of #8728 - which already has a PR up for a fix. @akhrarovsaid is right, closing
gharchive/issue
2024-10-16T14:39:26
2025-04-01T04:35:28.685233
{ "authors": [ "akhrarovsaid", "exoer", "r1tsuu" ], "repo": "payloadcms/payload", "url": "https://github.com/payloadcms/payload/issues/8740", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2707013419
The custom view of AfterNavLinks has a position mismatch in Safari on initial state Describe the Bug. macOS: 15.1.1, Safari 18.1.1 Chrome: 131.0.6778.86 My component code: 'use client' import Link from 'next/link' import { usePathname } from 'next/navigation' import { NavGroup } from '@payloadcms/ui' import React from 'react' const AfterNavLinks: React.FC = () => { const pathname = usePathname() const active = pathname === '/bull-board' return ( <NavGroup key="bull-board" label="Bull Board"> <Link href="/bull-board" target="_blank"> Bull Board </Link> </NavGroup> ) } export default AfterNavLinks The view displays correctly after I expand and collapse all navigation groups in Safari. Reproduction Steps Add custom component on AfterNavLinks hook. Open the navigation Environment Info Payload: Binaries: Node: 22.11.0 npm: 10.9.0 Yarn: N/A pnpm: 9.14.4 Relevant Packages: payload: 3.2.2 Operating System: Platform: darwin Arch: arm64 Version: Darwin Kernel Version 24.1.0: Thu Oct 10 21:05:14 PDT 2024; root:xnu-11215.41.3~2/RELEASE_ARM64_T8103 Available memory (MB): 8192 Available CPU cores: 8 Node: 22 Next.js: 15 It looks like this issue was fixed in version 3.4.0.
gharchive/issue
2024-11-30T08:51:12
2025-04-01T04:35:28.687196
{ "authors": [ "smoothdvd" ], "repo": "payloadcms/payload", "url": "https://github.com/payloadcms/payload/issues/9626", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1533569620
Region is Missing Error Getting the current error on s3 upload. I can't find anything in the docs about "Region" in configuration. [00:11:06] ERROR (payload): Region is missing Error: Region is missing at default (/path/node_modules/@aws-sdk/config-resolver/dist-cjs/regionConfig/config.js:10:15) at /path/node_modules/@aws-sdk/node-config-provider/dist-cjs/fromStatic.js:6:83 at /path/node_modules/@aws-sdk/property-provider/dist-cjs/chain.js:11:28 at processTicksAndRejections (internal/process/task_queues.js:95:5) at async coalesceProvider (/path/node_modules/@aws-sdk/property-provider/dist-cjs/memoize.js:14:24) at async /path/node_modules/@aws-sdk/property-provider/dist-cjs/memoize.js:26:28 at async useFipsEndpoint (/path/node_modules/@aws-sdk/config-resolver/dist-cjs/regionConfig/resolveRegionConfig.js:21:74) at async resolveParams (/path/node_modules/@aws-sdk/middleware-endpoint/dist-cjs/adaptors/getEndpointFromInstructions.js:29:40) at async getEndpointFromInstructions (/path/node_modules/@aws-sdk/middleware-endpoint/dist-cjs/adaptors/getEndpointFromInstructions.js:7:28) at async /path/node_modules/@aws-sdk/middleware-endpoint/dist-cjs/endpointMiddleware.js:8:26 [00:11:14] ERROR (payload): There was an error while uploading files corresponding to the collection media with filename Screen Shot 2023-01-14 at 6.15.24 PM-1.png: [00:11:14] ERROR (payload): Region is missing Error: Region is missing at default (/path/node_modules/@aws-sdk/config-resolver/dist-cjs/regionConfig/config.js:10:15) at /path/node_modules/@aws-sdk/node-config-provider/dist-cjs/fromStatic.js:6:83 at /path/node_modules/@aws-sdk/property-provider/dist-cjs/chain.js:11:28 [00:11:14] ERROR (payload): Region is missing Error: Region is missing at default (/path/node_modules/@aws-sdk/config-resolver/dist-cjs/regionConfig/config.js:10:15) at /path/node_modules/@aws-sdk/node-config-provider/dist-cjs/fromStatic.js:6:83 at /path/node_modules/@aws-sdk/property-provider/dist-cjs/chain.js:11:28 at processTicksAndRejections (internal/process/task_queues.js:95:5) at async coalesceProvider (/path/node_modules/@aws-sdk/property-provider/dist-cjs/memoize.js:14:24) at async /path/node_modules/@aws-sdk/property-provider/dist-cjs/memoize.js:26:28 at async useFipsEndpoint (/path/node_modules/@aws-sdk/config-resolver/dist-cjs/regionConfig/resolveRegionConfig.js:21:74) at async resolveParams (/path/node_modules/@aws-sdk/middleware-endpoint/dist-cjs/adaptors/getEndpointFromInstructions.js:29:40) at async getEndpointFromInstructions (/path/node_modules/@aws-sdk/middleware-endpoint/dist-cjs/adaptors/getEndpointFromInstructions.js:7:28) at async /path/node_modules/@aws-sdk/middleware-endpoint/dist-cjs/endpointMiddleware.js:8:26 hey @dundas, you could add the region to the configuration adapter: s3Adapter({ config: { credentials: { accessKeyId: process.env.S3_ACCESS_KEY_ID, secretAccessKey: process.env.S3_SECRET_ACCESS_KEY, }, region: process.env.S3_REGION, }, bucket: process.env.S3_BUCKET, }), }, Yes, @dundas - the above is correct. You just need to specify your region. I will close this one out. If that didn't solve your problem, please let us know!
gharchive/issue
2023-01-15T00:15:52
2025-04-01T04:35:28.689065
{ "authors": [ "dundas", "jmikrut", "tomsvogel" ], "repo": "payloadcms/plugin-cloud-storage", "url": "https://github.com/payloadcms/plugin-cloud-storage/issues/33", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
210699237
Allow any intent to be executed on the client side Right now only sale is supported. Fixed and pushed.
gharchive/issue
2017-02-28T06:32:53
2025-04-01T04:35:28.697864
{ "authors": [ "bluepnume" ], "repo": "paypal/paypal-checkout", "url": "https://github.com/paypal/paypal-checkout/issues/205", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2035589759
🛑 Departamento is down In 5389320, Departamento (http://cca.unb.br/) was down: HTTP code: 0 Response time: 0 ms Resolved: Departamento is back up in e216a20 after 21 minutes.
gharchive/issue
2023-12-11T12:37:55
2025-04-01T04:35:28.699366
{ "authors": [ "pazkero" ], "repo": "pazkero/status.cacic.bsb.br", "url": "https://github.com/pazkero/status.cacic.bsb.br/issues/1573", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2201844591
🛑 Departamento is down In 20bfb8d, Departamento (http://cca.unb.br/) was down: HTTP code: 0 Response time: 0 ms Resolved: Departamento is back up in 3907ee4 after 12 minutes.
gharchive/issue
2024-03-22T07:10:18
2025-04-01T04:35:28.700844
{ "authors": [ "pazkero" ], "repo": "pazkero/status.cacic.bsb.br", "url": "https://github.com/pazkero/status.cacic.bsb.br/issues/2869", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
379100070
show solution: fails when replaying commands with o/xxx in them Problem I use git checkout -b fixForMR o/feature in a solution. When replaying this command via show solution, there are some transcoding errors as it tries to run 2 commands: $ git checkout -b fixForMR o&#x2F The ref o&#x2F does not exist or is unknown $ feature The command "feature" isn't supported, sorry! ... How to reproduce Gist level id: 13b665274699425ce772ad5fe6705cbd Direct access: https://learngitbranching.js.org/?NODEMO&gist_level_id=13b665274699425ce772ad5fe6705cbd&locale=en_US type in: show solution and click Confirm sorry for the delay here, busy week. neat that you found / are using the gist level IDs! yeah theres some weirddd escaping. let me try to debug the solution command correct is correct in the gist: "solutionCommand": "git checkout -b fixForMR o/feature;git commit;git push origin fixForMR:feature;git rebase fixForMR feature",
gharchive/issue
2018-11-09T10:11:49
2025-04-01T04:35:28.741435
{ "authors": [ "arnaudjolly", "pcottle" ], "repo": "pcottle/learnGitBranching", "url": "https://github.com/pcottle/learnGitBranching/issues/497", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
361744384
Split reserve into 2 functions Currently reserve() has an optional timeout parameter. This leads to an optional return type; since if a timeout happens no exception is thrown yet we don't get a Job as return value. I propose: Change reserve() to: /** @throws someException **/ public function reserve(): Job Add reserveWithTimeout: /** @throws someException @throws TimeoutException **/ public function reserveWithTimeout(int $timeout): Job Another option could be with a Null Object pattern - a subclass of Job that has no contents. Its a little cleaner than a nullable job (function reserve(): ?Job), which is possible from PHP 7.1. It's not an exceptional event if the queue is empty and equally, a reserve could end up not returning a job either with socket timeouts - or other fairly ordinary issues. So, not having a job to work on is an event that should be handled - not normally as a particular problem though. I've written code that would pause for a little longer than usual on an empty queue, especially if it's not likely to be urgent, or there will often be large gaps between things to do. either with socket timeouts - or other fairly ordinary issues. These issues are already exceptions, and they should be...
gharchive/issue
2018-09-19T13:16:15
2025-04-01T04:35:28.752179
{ "authors": [ "SamMousa", "alister" ], "repo": "pda/pheanstalk", "url": "https://github.com/pda/pheanstalk/issues/178", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
860131936
Added information to readme on supported models and setup requirements Included some useful information (mainly from closed issues) to the readme Users continue to raise issues (#11, #14) about compatible models which might be preempted with updated documentation
gharchive/pull-request
2021-04-16T20:12:07
2025-04-01T04:35:28.760307
{ "authors": [ "lukestein" ], "repo": "pdemarco925/homebridge-icomfort", "url": "https://github.com/pdemarco925/homebridge-icomfort/pull/13", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
803085768
Memory leak? Bug report A description of the bug Somewhere in pdfminer.six there appears to be a memory leak; even after a Python process is done handling a PDF, the memory it required is not released. Steps to reproduce the bug pip install memory_profiler Save this script as memory.py: from pdfminer.high_level import extract_text from memory_profiler import profile import gc path = "samples/nonfree/nlp2004slides.pdf" @profile def main(): text = extract_text(path) gc.collect() main() Run python memory.py Output: Line # Mem usage Increment Occurences Line Contents ============================================================ 7 44.3 MiB 44.3 MiB 1 @profile 8 def main(): 9 51.5 MiB 7.2 MiB 1 text = extract_text(path) 10 51.5 MiB 0.0 MiB 1 gc.collect() Note: The problem does not seem specific to extract_text; other, lower-level usages of pdfminer.six seem to create similar effects. Also: Passing caching=False to extract_text does not appear to reduce the memory leak. I have spent some time trying to track down the source of apparent leak, but have not had much luck. I'll continue looking, but am raising this issue in the meantime, in case the cause is more readily apparent to others. I just faced two consecutive outages of the system I maintain - I tracked down the reason to one PDF a user wanted to upload - my system extracts the text to index the PDF document. I can provide the document to a maintainer ( it should not go public). Basically, the PDF consists of slides - when I open the PDF with evince I can view it, but the pages get rendered very, very slowly. So there is obviously something bogus. Is anybody aware of a tool which "fixes" bogus PDF files? If not, I think I'll just screenshot the pages and create a new one. I can reproduce this issue. Will have a look at the PR's. Would love to merge to PR's related to this issue but they don't seem to solve the issue. I'm experiencing a regression of the recursion error in issue https://github.com/pdfminer/pdfminer.six/issues/404 with version 20220524. Could this be related? Hi, I'm still running into this issue on version 20220524 of pdfminer.six. I can't provide the pdf that causes the issue since it needs to be anonymized, but these are the LAParams we're using for pdfminer: LAParams( line_overlap= 0.5, char_margin= 1.1, word_margin= 0.2, line_margin= 0.5, boxes_flow= 0.5, all_texts=True, ) Traceback: RecursionError: maximum recursion depth exceeded (245 additional frame(s) were not displayed) ... File "pdfminer/layout.py", line 705, in analyze super().analyze(laparams) File "pdfminer/layout.py", line 439, in analyze obj.analyze(laparams) File "pdfminer/layout.py", line 705, in analyze super().analyze(laparams) File "pdfminer/layout.py", line 439, in analyze obj.analyze(laparams) File "pdfminer/layout.py", line 439, in analyze obj.analyze(laparams) maximum recursion depth exceeded Is there any update on this issue? No update here yet. The root cause needs to be found first. Hey, is there any update on this issue?
gharchive/issue
2021-02-07T23:14:32
2025-04-01T04:35:28.767642
{ "authors": [ "Rumpelcita", "dominikhait", "jsvine", "jugmac00", "pietermarsman", "yoavkedem1" ], "repo": "pdfminer/pdfminer.six", "url": "https://github.com/pdfminer/pdfminer.six/issues/580", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1373366151
[ldd-msn] Change cardinality of mission_phase_identifier and mission_phase_name Describe the issue identified (if applicable) In some cases, particularly in bundle and collection labels, the mission_phase_identifier and mission_phase_name attributes may be multivalued. However, the cardinality for these two attributes is 0..1. Describe the solution you'd like Change the cardinality for both mission_phase_identifier and mission_phase_name to 0..Unbounded. Describe alternatives you've considered An alternative approach would be to have multiple instances of the Mission_Information object, each with a single instance of mission_phase_name and/or mission_phase_identifier. This approach is valid under the current version of the MSN LDD, but is unnecessarily verbose. LDD Dictionary Version 1300 PDS4 IM Version 1I00 Need-by Date Needed by: 2022-09-30. This change is needed for labels currently under designed. Resolved by pds-data-dictionaries/ldd-msn#32
gharchive/issue
2022-09-14T17:47:00
2025-04-01T04:35:28.786612
{ "authors": [ "cgobat", "jmafi" ], "repo": "pds-data-dictionaries/PDS4-LDD-Issue-Repo", "url": "https://github.com/pds-data-dictionaries/PDS4-LDD-Issue-Repo/issues/238", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
67762273
Add new layout attributes. Added foregroundColor, backgroundColor, headings, paragraphs and lastUpdated. Ping @matthewtole Coverage remained the same at 100.0% when pulling 4396365e760db08b3c693bf070355959d1ca770e on feature/headings-paragraphs-colors into faca07ea269664dc68c0bf71a8b5f1f07e5395ae on master. Coverage remained the same at 100.0% when pulling 4396365e760db08b3c693bf070355959d1ca770e on feature/headings-paragraphs-colors into faca07ea269664dc68c0bf71a8b5f1f07e5395ae on master. Aside from minor nits about error messages, :+1: Coverage remained the same at 100.0% when pulling e9776ce4f6ea0df8a73749ddc3f4eb4165be6130 on feature/headings-paragraphs-colors into faca07ea269664dc68c0bf71a8b5f1f07e5395ae on master.
gharchive/pull-request
2015-04-11T09:40:17
2025-04-01T04:35:28.800973
{ "authors": [ "Neal", "coveralls", "matthewtole" ], "repo": "pebble/pebble-api-node", "url": "https://github.com/pebble/pebble-api-node/pull/10", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1237008823
Create fr.json add locales FR Hey @scorpio686, could you update the json with this? Ty! "headOverlays": { "title": "Appearance", "hair": { "title": "Hair", "style": "Style", "color": "Color", "highlight": "Highlight" "highlight": "Highlight", "fade": "Fade" // This need to be added }, "opacity": "Opacity", "style": "Style", "color": "Color", "blemishes": "Blemishes", "beard": "Beard", "eyebrows": "Eyebrows", "ageing": "Ageing", "makeUp": "Make up", "blush": "Bblush", "complexion": "Complexion", "sunDamage": "Sun damage", "lipstick": "Lipstick", "moleAndFreckles": "Mole and Freckles", "chestHair": "Chest hair", "bodyBlemishes": "Body blemishes", "eyeColor": "Eye color" }, added on #80
gharchive/pull-request
2022-05-16T11:24:14
2025-04-01T04:35:28.803133
{ "authors": [ "franfdezmorales", "scorpio686" ], "repo": "pedr0fontoura/fivem-appearance", "url": "https://github.com/pedr0fontoura/fivem-appearance/pull/44", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
258531237
npm install --save node-pdfbox When i try to install on centos 6 typing npm install --save node-pdfbox i got this error: gyp ERR! build error gyp ERR! stack Error: make failed with exit code: 2 gyp ERR! stack at ChildProcess.onExit (/usr/local/lib/node_modules/npm/node_modules/node-gyp/lib/build.js:270:23) gyp ERR! stack at emitTwo (events.js:87:13) gyp ERR! stack at ChildProcess.emit (events.js:172:7) gyp ERR! stack at Process.ChildProcess._handle.onexit (internal/child_process.js:200:12) gyp ERR! System Linux 2.6.32-696.1.1.el6.x86_64 gyp ERR! command "/usr/local/bin/node" "/usr/local/lib/node_modules/npm/node_modules/node-gyp/bin/node-gyp.js" "rebuild" gyp ERR! cwd /var/nodejs/imagick/node_modules/node-pdfbox/node_modules/java gyp ERR! node -v v4.2.2 gyp ERR! node-gyp -v v3.0.3 gyp ERR! not ok npm ERR! Linux 2.6.32-696.1.1.el6.x86_64 npm ERR! argv "/usr/local/bin/node" "/usr/local/bin/npm" "install" npm ERR! node v4.2.2 npm ERR! npm v2.14.7 npm ERR! code ELIFECYCLE npm ERR! java@0.8.0 install: node-gyp rebuild npm ERR! Exit status 1 npm ERR! npm ERR! Failed at the java@0.8.0 install script 'node-gyp rebuild'. npm ERR! This is most likely a problem with the java package, npm ERR! not with npm itself. npm ERR! Tell the author that this fails on your system: npm ERR! node-gyp rebuild npm ERR! You can get their info via: npm ERR! npm owner ls java npm ERR! There is likely additional logging output above. npm ERR! Please include the following file with any support request: npm ERR! /var/nodejs/imagick/npm-debug.log but javac is installed Can you help me? After changing java version to the latest, I still got this error. After reading this article, compared with java and node version, I changed the node version to the older v8.11.1 and it installed correctly !
gharchive/issue
2017-09-18T15:55:13
2025-04-01T04:35:28.809798
{ "authors": [ "Giancarlo1974", "dashtom3" ], "repo": "pedro-lucas/node-pdfbox", "url": "https://github.com/pedro-lucas/node-pdfbox/issues/7", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
357337829
Dimensões maiores que 70cm aumenta o valor do frete em R$20,00 (SOLUÇÃO) Segue arquivos em anexo corrigidos com a opção no admin para atualizar o valor dos correios sempre que necessário (atualmente 79 reais). (apagar o .txt do final do arquivo e substituir). ect/system.xml model/carrier/correiosMethod.php CorreiosMethod.php.txt system.xml.txt Valeu @mateussantin! Se puder enviar na forma de uma PR, talvez o @pedro-teixeira possa mesclar a extensão. De qualquer forma, obrigado 👍 Boa noite, fiz um edit na minha issue detalhando melhor, sou meio novo no git. Vou mandar aqui o arquivo editado também, é um prazer contribuir com a comunidade espero que o Pedro possa incluir ou melhorar o código. ** Arquivo: CorreiosMethod.php ** Localizado: app/code/community/PedroTeixeira/Correios/Model/Carrier/CorreiosMethod.php Criado variável na linha: 58; Inlcuido função _generateAument na linha: 293; Chamada da função e soma do valor na linha: 347 348 e 349; Em qua, 5 de set de 2018 às 16:03, Rafael Patro notifications@github.com escreveu: Valeu @mateussantin https://github.com/mateussantin! Se puder enviar na forma de uma PR, talvez o @pedro-teixeira https://github.com/pedro-teixeira possa mesclar a extensão. De qualquer forma, obrigado 👍 — You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub https://github.com/pedro-teixeira/correios/issues/393#issuecomment-418845195, or mute the thread https://github.com/notifications/unsubscribe-auth/Am-kiAs0z_ku2YvWwi1pZcQBv5oPuCJVks5uYCAegaJpZM4WbZgQ . Alguém poderia enviar a alteração completa do CorreiosMethod.php, não tenho conhecimento em PHP, tentei fazer as modificações explicadas acima mas só está dando erro. Agradeço desde já!! Eu também não consegui fazer, segui o passo a passo mas deu o erro: Parse error: syntax error, unexpected 'protected' (T_PROTECTED) in /home/public_html/app/code/community/PedroTeixeira/Correios/Model/Carrier/CorreiosMethod.php on line 196 Obrigado @mateussantin Deu certo! Vou colocar o código inteiro de como ficou o arquivo: ` * @copyright 2015 Pedro Teixeira (http://pedroteixeira.io) * @license http://opensource.org/licenses/MIT MIT * @link https://github.com/pedro-teixeira/correios */ class PedroTeixeira_Correios_Model_Carrier_CorreiosMethod extends Mage_Shipping_Model_Carrier_Abstract implements Mage_Shipping_Model_Carrier_Interface { /** * _code property * * @var string */ protected $_code = 'pedroteixeira_correios'; protected $_isFixed = true; /** * _result property * * @var Mage_Shipping_Model_Rate_Result|Mage_Shipping_Model_Tracking_Result */ protected $_result = null; /** * ZIP code vars */ protected $_fromZip = null; protected $_toZip = null; /** * Value and Weight */ protected $_packageValue = null; protected $_packageWeight = null; protected $_volumeWeight = null; protected $_freeMethodWeight = null; protected $_midSize = null; protected $_splitUp = 0; protected $_postingDays = 0; /** * Post methods */ protected $_postMethods = null; protected $_postMethodsFixed = null; protected $_postMethodsExplode = null; /** * Free method request */ protected $_freeMethodRequest = false; protected $_freeMethodRequestResult = null; /** * Collect Rates * * @param Mage_Shipping_Model_Rate_Request $request Mage request * * @return bool|Mage_Shipping_Model_Rate_Result|Mage_Shipping_Model_Tracking_Result */ public function collectRates(Mage_Shipping_Model_Rate_Request $request) { // Do initial check if ($this->_inicialCheck($request) === false) { return false; } // Check package value if ($this->_packageValue < $this->getConfigData('min_order_value') || $this->_packageValue > $this->getConfigData('max_order_value') ) { $this->_throwError('valueerror', 'Value limits', __LINE__); return $this->_result; } // Check ZIP Code if (!preg_match('/^([0-9]{8})$/', $this->_toZip)) { $this->_throwError('zipcodeerror', 'Invalid Zip Code', __LINE__); return $this->_result; } if ($this->_packageWeight == 0) { $this->_packageWeight = $this->_getNominalWeight($request); } if ($this->getConfigData('weight_type') == PedroTeixeira_Correios_Model_Source_WeightType::WEIGHT_GR) { $this->_packageWeight = number_format($this->_packageWeight / 1000, 2, '.', ''); } // Check weight zero if ($this->_packageWeight <= 0) { $this->_throwError('weightzeroerror', 'Weight zero', __LINE__); return $this->_result; } $this->_postMethods = $this->getConfigData('postmethods'); $this->_postMethodsFixed = $this->_postMethods; $this->_postMethodsExplode = explode(',', $this->getConfigData('postmethods')); // Generate Volume Weight if ($this->_generateVolumeWeight($request) === false || $this->_removeInvalidServices() === false) { $this->_throwError('dimensionerror', 'Dimension error', __LINE__); return $this->_result; } $this->_filterMethodByItemRestriction($request); if (empty($this->_postMethods)) { return false; } //Show Quotes $this->_getQuotes(); // Use descont codes $this->_updateFreeMethodQuote($request); return $this->_result; } /** * Retrieve all visible items from request * * @param Mage_Shipping_Model_Rate_Request $request Mage request * * @return array */ protected function _getRequestItems($request) { $allItems = $request->getAllItems(); $items = array(); foreach ($allItems as $item) { if (!$item->getParentItemId()) { $items[] = $item; } } $items = $this->_loadBundleChildren($items); return $items; } /** * Gets Nominal Weight * * @param Mage_Shipping_Model_Rate_Request $request Mage request * * @return number */ protected function _getNominalWeight($request) { $weight = 0; $items = $this->_getRequestItems($request); foreach ($items as $item) { $product = Mage::getModel('catalog/product')->load($item->getProductId()); $weight += $product->getWeight(); } return $weight; } /** * Get shipping quote * * @return Mage_Shipping_Model_Rate_Result|Mage_Shipping_Model_Tracking_Result */ protected function _getQuotes() { $softErrors = explode(',', $this->getConfigData('soft_errors')); $correiosReturn = $this->_getCorreiosReturn(); if ($correiosReturn !== false) { $errorList = array(); $correiosReturn = $this->_addPostMethods($correiosReturn); foreach ($correiosReturn as $servicos) { $errorId = (string) $servicos->Erro; $errorList[$errorId] = $servicos->MsgErro; if ($errorId != '0' && !in_array($errorId, $softErrors)) { continue; } $servicos->Valor = $this->_getFormatPrice((string) $servicos->Valor); $this->_appendShippingReturn($servicos); } $this->_appendShippingErrors($errorList); } else { return $this->_result; } if ($this->_freeMethodRequest === true) { return $this->_freeMethodRequestResult; } else { return $this->_result; } } /** * Make initial checks and iniciate module variables * * @param Mage_Shipping_Model_Rate_Request $request Mage request * * @return bool */ protected function _inicialCheck(Mage_Shipping_Model_Rate_Request $request) { if (!$this->getConfigFlag('active')) { // Disabled Mage::log('pedroteixeira_correios: Disabled'); return false; } $origCountry = Mage::getStoreConfig('shipping/origin/country_id', $this->getStore()); $destCountry = $request->getDestCountryId(); if ($origCountry != 'BR' || $destCountry != 'BR') { // Out of delivery area Mage::log('pedroteixeira_correios: Out of delivery area'); return false; } $this->_fromZip = Mage::getStoreConfig('shipping/origin/postcode', $this->getStore()); $this->_toZip = $request->getDestPostcode(); // Fix ZIP code $this->_fromZip = str_replace(array('-', '.'), '', trim($this->_fromZip)); $this->_toZip = str_replace(array('-', '.'), '', trim($this->_toZip)); if (!preg_match('/^([0-9]{8})$/', $this->_fromZip)) { Mage::log('pedroteixeira_correios: From ZIP Code Error'); return false; } if (!trim($this->_toZip)) { return false; } $this->_result = Mage::getModel('shipping/rate_result'); $this->_packageValue = $request->getBaseCurrency()->convert( $request->getPackageValue(), $request->getPackageCurrency() ); $this->_packageWeight = number_format($request->getPackageWeight(), 2, '.', ''); $this->_freeMethodWeight = number_format($request->getFreeMethodWeight(), 2, '.', ''); } /** * Get Correios return * * @return bool|SimpleXMLElement[] * * @throws Exception */ protected function _getCorreiosReturn() { $filename = $this->getConfigData('url_ws_correios'); try { $client = new Zend_Http_Client($filename); $client->setConfig( array( 'timeout' => $this->getConfigData('ws_timeout'), 'adapter' => Mage::getModel('pedroteixeira_correios/http_client_adapter_socket') ) ); $client->setParameterGet('StrRetorno', 'xml'); $client->setParameterGet('nCdServico', $this->_postMethods); $client->setParameterGet('nVlPeso', $this->_packageWeight); $client->setParameterGet('sCepOrigem', $this->_fromZip); $client->setParameterGet('sCepDestino', $this->_toZip); $client->setParameterGet('nCdFormato', 1); $client->setParameterGet('nVlComprimento', $this->_midSize); $client->setParameterGet('nVlAltura', $this->_midSize); $client->setParameterGet('nVlLargura', $this->_midSize); if ($this->getConfigData('mao_propria')) { $client->setParameterGet('sCdMaoPropria', 'S'); } else { $client->setParameterGet('sCdMaoPropria', 'N'); } if ($this->getConfigData('aviso_recebimento')) { $client->setParameterGet('sCdAvisoRecebimento', 'S'); } else { $client->setParameterGet('sCdAvisoRecebimento', 'N'); } if ($this->getConfigData('valor_declarado') || in_array($this->getConfigData('acobrar_code'), $this->_postMethodsExplode) ) { $client->setParameterGet('nVlValorDeclarado', number_format($this->_packageValue, 2, ',', '')); } else { $client->setParameterGet('nVlValorDeclarado', 0); } $nCdEmpresa = $this->getConfigData('cod_admin'); $sDsSenha = $this->getConfigData('senha_admin'); if (!empty($nCdEmpresa) && !empty($sDsSenha)) { $client->setParameterGet('nCdEmpresa', $nCdEmpresa); $client->setParameterGet('sDsSenha', $sDsSenha); } $content = $client->request()->getBody(); if ($content == '') { throw new Exception('No XML returned [' . __LINE__ . ']'); } libxml_use_internal_errors(true); $sxe = simplexml_load_string($content); if (!$sxe) { throw new Exception('Bad XML [' . __LINE__ . ']'); } $xml = new SimpleXMLElement($content); if (count($xml->cServico) <= 0) { throw new Exception('No tag cServico in Correios XML [' . __LINE__ . ']'); } return $xml->cServico; } catch (Exception $e) { $this->_throwError('urlerror', 'URL Error - ' . $e->getMessage(), __LINE__); return false; } } /** * Apend shipping value to return * * @param SimpleXMLElement $servico Service Data * * @return void */ /* Inicio - Aumenta taxa dos correios */ protected function _generateAument(){ $items = Mage::getModel('checkout/cart')->getQuote()->getAllVisibleItems(); foreach($items as $item){ $_product = $this->_getSimpleProduct($item->getProduct()); if ($_product->getData('volume_altura') == '' || (int) $_product->getData('volume_altura') == 0) { $itemAltura = $this->getConfigData('altura_padrao'); } else { $itemAltura = $_product->getData('volume_altura'); } if ($_product->getData('volume_largura') == '' || (int) $_product->getData('volume_largura') == 0) { $itemLargura = $this->getConfigData('largura_padrao'); } else { $itemLargura = $_product->getData('volume_largura'); } if ($_product->getData('volume_comprimento') == '' || (int) $_product->getData('volume_comprimento') == 0) { $itemComprimento = $this->getConfigData('comprimento_padrao'); } else { $itemComprimento = $_product->getData('volume_comprimento'); } } if ($itemAltura > 70 || $itemLargura > 70 || $itemComprimento > 70) { $this->add_aumento = 79; } else { $this->add_aumento; } return true; } /* Fim - Aumenta taxa dos correios */ protected function _appendShippingReturn(SimpleXMLElement $servico) { $correiosDelivery = (int) $servico->PrazoEntrega; $shippingMethod = (string) $servico->Codigo; $shippingPrice = (float) $servico->Valor; if ($shippingPrice <= 0) { return; } /* Inicio - Aumenta taxa dos correios */ $this->_generateAument(); $vl_aumento = $this->add_aumento; $shippingPrice = $shippingPrice + $vl_aumento; /* Fim - Aumenta taxa dos correios */ $errorMsg = $this->_getSoftErrorMsg((string) $servico->Erro); $method = Mage::getModel('shipping/rate_result_method'); $method->setCarrier($this->_code); $method->setCarrierTitle($this->getConfigData('title') . $this->_getSplitUpMsg() . $errorMsg); $method->setMethod($shippingMethod); $shippingCost = $shippingPrice; $shippingPrice = $shippingPrice + $this->getConfigData('handling_fee'); $shippingData = Mage::helper('pedroteixeira_correios')->getShippingLabel($shippingMethod); $shippingData = Mage::helper('pedroteixeira_correios')->__($shippingData); if ($shippingMethod == $this->getConfigData('acobrar_code')) { $shippingData = $shippingData . ' ( R$' . number_format($shippingPrice, 2, ',', '.') . ' )'; $shippingPrice = 0; } if ($this->getConfigFlag('prazo_entrega')) { if ($correiosDelivery > 0) { $method->setMethodTitle( sprintf( $this->getConfigData('msgprazo'), $shippingData, (int) ($correiosDelivery + $this->getConfigData('add_prazo') + $this->_postingDays) ) ); } } else { $method->setMethodTitle($shippingData); } $method->setPrice($shippingPrice); $method->setCost($shippingCost); if ($this->_freeMethodRequest === true) { $this->_freeMethodRequestResult->append($method); } else { $this->_result->append($method); } } /** * Throw error * * @param string $message Message placeholder * @param string $log Message * @param string|int $line Line of log * @param string $custom Custom variables for placeholder * * @return void */ protected function _throwError($message, $log = null, $line = 'NO LINE', $custom = null) { $this->_result = null; $this->_result = Mage::getModel('shipping/rate_result'); $error = Mage::getModel('shipping/rate_result_error'); $error->setCarrier($this->_code); $error->setCarrierTitle($this->getConfigData('title')); if (is_null($custom) || $this->getConfigData($message) == '') { Mage::log($this->_code . ' [' . $line . ']: ' . $log); $error->setErrorMessage($this->getConfigData($message)); } else { Mage::log($this->_code . ' [' . $line . ']: ' . $log); $error->setErrorMessage(sprintf($this->getConfigData($message), $custom)); } $this->_result->append($error); } /** * Retrieves a simple product * * @param Mage_Catalog_Model_Product $product Catalog Product * * @return Mage_Catalog_Model_Product */ protected function _getSimpleProduct($product) { $type = $product->getTypeInstance(true); if ($type->getProduct($product)->hasCustomOptions() && ($simpleProductOption = $type->getProduct($product)->getCustomOption('simple_product')) ) { $simpleProduct = $simpleProductOption->getProduct($product); if ($simpleProduct) { return $this->_getSimpleProduct($simpleProduct); } } return $type->getProduct($product); } /** * Generate Volume weight * * @param Mage_Shipping_Model_Rate_Request $request Mage request * * @return bool */ protected function _generateVolumeWeight($request) { $pesoCubicoTotal = 0; $items = $this->_getRequestItems($request); foreach ($items as $item) { $_product = $this->_getSimpleProduct($item->getProduct()); if ($_product->getData('volume_altura') == '' || (int) $_product->getData('volume_altura') == 0) { $itemAltura = $this->getConfigData('altura_padrao'); } else { $itemAltura = $_product->getData('volume_altura'); } if ($_product->getData('volume_largura') == '' || (int) $_product->getData('volume_largura') == 0) { $itemLargura = $this->getConfigData('largura_padrao'); } else { $itemLargura = $_product->getData('volume_largura'); } if ($_product->getData('volume_comprimento') == '' || (int) $_product->getData('volume_comprimento') == 0) { $itemComprimento = $this->getConfigData('comprimento_padrao'); } else { $itemComprimento = $_product->getData('volume_comprimento'); } if ($this->getConfigFlag('check_dimensions')) { foreach ($this->_postMethodsExplode as $key => $method) { $sizeMax = max($itemAltura, $itemLargura, $itemComprimento); $sumMax = ($itemAltura + $itemLargura + $itemComprimento); $isValid = ($sizeMax <= $this->getConfigData("validate/serv_{$method}/max/size")); $isValid &= ($sumMax <= $this->getConfigData("validate/serv_{$method}/max/sum")); if (!$isValid) { unset($this->_postMethodsExplode[$key]); } } if (count($this->_postMethodsExplode) == 0) { return false; } $this->_postMethods = implode(',', $this->_postMethodsExplode); $this->_postMethodsFixed = $this->_postMethods; } $itemAltura = $this->_getFitHeight($item); $pesoCubicoTotal += (($itemAltura * $itemLargura * $itemComprimento) * $item->getTotalQty()) / $this->getConfigData('coeficiente_volume'); $this->_postingDays = max($this->_postingDays, (int) $_product->getData('posting_days')); } $this->_volumeWeight = number_format($pesoCubicoTotal, 2, '.', ''); return true; } /** * Generate free shipping for a product * * @param string $freeMethod Free method * * @return void */ protected function _setFreeMethodRequest($freeMethod) { $this->_freeMethodRequest = true; $this->_freeMethodRequestResult = Mage::getModel('shipping/rate_result'); $this->_postMethods = $freeMethod; $this->_postMethodsExplode = array($freeMethod); if ($this->getConfigData('weight_type') == PedroTeixeira_Correios_Model_Source_WeightType::WEIGHT_GR) { $this->_freeMethodWeight = number_format($this->_freeMethodWeight / 1000, 2, '.', ''); } $this->_packageWeight = $this->_freeMethodWeight; $this->_pacWeight = $this->_freeMethodWeight; } /** * Check if current carrier offer support to tracking * * @return bool true */ public function isTrackingAvailable() { return true; } /** * Get Tracking Info * * @param mixed $tracking Tracking * * @return mixed */ public function getTrackingInfo($tracking) { $result = $this->getTracking($tracking); if ($result instanceof Mage_Shipping_Model_Tracking_Result) { if ($trackings = $result->getAllTrackings()) { return $trackings[0]; } } elseif (is_string($result) && !empty($result)) { return $result; } return false; } /** * Get Tracking * * @param array $trackings Trackings * * @return Mage_Shipping_Model_Tracking_Result */ public function getTracking($trackings) { $this->_result = Mage::getModel('shipping/tracking_result'); foreach ((array) $trackings as $code) { $this->_getTracking($code); } return $this->_result; } /** * Loads the parameters and calls the webservice using SOAP * * @param string $code Code * * @return bool|array * * @throws Exception */ protected function _getTrackingRequest($code) { $response = false; $params = array( 'usuario' => $this->getConfigData('sro_username'), 'senha' => $this->getConfigData('sro_password'), 'tipo' => $this->getConfigData('sro_type'), 'resultado' => 'T', 'lingua' => $this->getConfigData('sro_language'), 'objetos' => $code, ); try { $client = new SoapClient( $this->getConfigData('url_sro_correios'), Mage::helper('pedroteixeira_correios')->getStreamContext() ); $response = $client->buscaEventos($params); if (empty($response)) { throw new Exception("Empty response"); } } catch (Exception $e) { Mage::log("Soap Error: {$e->getMessage()}"); } return $response; } /** * Loads tracking progress details * * @param SimpleXMLElement $evento XML Element Node * @param bool $isDelivered Delivery Flag * * @return array */ protected function _getTrackingProgressDetails($evento, $isDelivered = false) { $date = new Zend_Date($evento->data, 'dd/MM/YYYY', new Zend_Locale('pt_BR')); $track = array( 'deliverydate' => $date->toString('YYYY-MM-dd'), 'deliverytime' => $evento->hora . ':00', 'status' => $evento->descricao, ); if (!$isDelivered) { $msg = array($evento->descricao); if (isset($evento->destino) && isset($evento->destino->local)) { $msg = array("{$evento->descricao} para {$evento->destino->local}"); } $track['activity'] = implode(' | ', $msg); $track['deliverylocation'] = "{$evento->local} - {$evento->cidade}/{$evento->uf}"; } return $track; } /** * Loads progress data using the WSDL response * * @param string $request Request response * * @return array */ protected function _getTrackingProgress($request) { $track = array(); $progress = array(); $eventTypes = explode(',', $this->getConfigData("sro_event_type_last")); if (count($request->return->objeto->evento) == 1) { $progress[] = $this->_getTrackingProgressDetails($request->return->objeto->evento); } else { foreach ($request->return->objeto->evento as $evento) { $progress[] = $this->_getTrackingProgressDetails($evento); $isDelivered = ((int) $evento->status < 2 && in_array($evento->tipo, $eventTypes)); if ($isDelivered) { $track = $this->_getTrackingProgressDetails($evento, $isDelivered); } } } $progress[] = $track; return $progress; } /** * Protected Get Tracking, opens the request to Correios * * @param string $code Code * * @return bool */ protected function _getTracking($code) { $error = Mage::getModel('shipping/tracking_result_error'); $error->setTracking($code); $error->setCarrier($this->_code); $error->setCarrierTitle($this->getConfigData('title')); $error->setErrorMessage($this->getConfigData('urlerror')); $request = $this->_getTrackingRequest($code); if (!isset($request->return)) { $this->_result->append($error); return false; } $progress = $this->_getTrackingProgress($request); if (!empty($progress)) { $track = array_pop($progress); $track['progressdetail'] = $progress; $tracking = Mage::getModel('shipping/tracking_result_status'); $tracking->setTracking($code); $tracking->setCarrier($this->_code); $tracking->setCarrierTitle($this->getConfigData('title')); $tracking->addData($track); $this->_result->append($tracking); return true; } else { $this->_result->append($error); return false; } } /** * Returns the allowed carrier methods * * @return array */ public function getAllowedMethods() { $output = array($this->_code => $this->getConfigData('title')); $serviceObject = Mage::getSingleton('pedroteixeira_correios/postmethod'); foreach ($serviceObject->getCollection() as $service) { $output[ $service->getMethodCode() ] = "{$service->getMethodCode()} - {$service->getMethodTitle()}"; } return $output; } /** * Define ZIP Code as required * * @param string $countryId Country ID * * @return bool */ public function isZipCodeRequired($countryId = null) { return true; } /** * Retrieve an average size. * For optimization purposes all tree box sizes are converted in one medium dimension. * Result cant exceed the minimum transportation limits. * * @return PedroTeixeira_Correios_Model_Carrier_CorreiosMethod */ protected function _loadMidSize() { $volumeFactor = $this->getConfigData('coeficiente_volume'); $volumeTotal = $this->_volumeWeight * $volumeFactor; $pow = round(pow((int) $volumeTotal, (1 / 3))); $min = $this->getConfigData('midsize_min'); $this->_midSize = max($pow, $min); return $this; } /** * Validate post methods removing invalid services from quotation. * * @return boolean|PedroTeixeira_Correios_Model_Carrier_CorreiosMethod */ protected function _removeInvalidServices() { $tmpMethods = $this->_postMethodsExplode; $tmpMethods = $this->_filterMethodByConfigRestriction($tmpMethods); $isDivisible = (count($tmpMethods) == 0); if ($isDivisible) { return $this->_splitPack(); } $this->_postMethodsExplode = $tmpMethods; $this->_postMethods = implode(',', $this->_postMethodsExplode); $this->_postMethodsFixed = $this->_postMethods; return $this; } /** * Include an additional method to quote content before showing. * When requested the new method is added in xml content as specified in config.xml like below: * * * 10065 * 2.45 * 5 * * 00000000 * 0.0 * 0 * * * 99999999 * 0.1 * 150 * * * * @param SimpleXMLElement $cServico XML Node * * @see http://www.correios.com.br/para-voce/consultas-e-solicitacoes/precos-e-prazos/servicos-nacionais_pasta/carta * * @return SimpleXMLElement */ protected function _addPostMethods($cServico) { $addMethods = $this->getConfigData("add_postmethods"); if (empty($addMethods) || !is_array($addMethods)) { return $cServico; } foreach ($addMethods as $configData) { $isValid = true; $isValid &= $this->_packageWeight >= $configData['from']['weight']; $isValid &= $this->_packageWeight <= $configData['to']['weight']; $isValid &= $this->_midSize >= $configData['from']['size']; $isValid &= $this->_midSize <= $configData['to']['size']; $isValid &= $this->_toZip >= $configData['from']['zip']; $isValid &= $this->_toZip <= $configData['to']['zip']; if ($isValid) { $price = $configData['price']; $days = $configData['days']; $method = $configData['code']; foreach ($cServico as $servico) { if ($servico->Codigo == $method) { if (!empty($price)) { $servico->Valor = number_format($price, 2, ',', ''); } if (!empty($days)) { $servico->PrazoEntrega = $days; } $servico->EntregaDomiciliar = 'S'; $servico->EntregaSabado = 'S'; $servico->Erro = '0'; $servico->MsgErro = ' '; } } } } return $cServico; } /** * This keeps only postmethods available for all items in cart. * In other words you can set post methods by products. * Methods not available for all items in cart are removed. * Require attribute creation called postmethods. * Example: * code: postmethods * type: multiselect * label: [free] * value 1: 41068 * value 2: 40096 * ... * value 99: 81019 * * @param Mage_Shipping_Model_Rate_Request $request Mage request * * @return PedroTeixeira_Correios_Model_Carrier_CorreiosMethod */ protected function _filterMethodByItemRestriction($request) { if ($this->getConfigFlag('filter_by_item')) { $items = $this->_getRequestItems($request); $intersection = $this->_postMethodsExplode; foreach ($items as $item) { $product = $this->_getSimpleProduct($item->getProduct()); $prodPostMethods = explode( ',', $product->getResource()->getAttributeRawValue( $product->getId(), 'postmethods', $request->getStoreId() ) ); $intersection = array_intersect($prodPostMethods, $intersection); } $this->_postMethodsExplode = $intersection; $this->_postMethods = implode(',', $intersection); $this->_postMethodsFixed = $this->_postMethods; } return $this; } /** * Added a fit size for items in large quantities. * Means you can join items like two or more glasses, pots and vases. * The calc is applied only for height side. * Required attribute fit_size. Example: * * code: fit_size * type: varchar * * After you can set a fit size for all products and improve your sells * * @param Mage_Eav_Model_Entity_Abstract $item Order Item * * @return number */ protected function _getFitHeight($item) { $product = $this->_getSimpleProduct($item->getProduct()); $height = $product->getData('volume_altura'); $height = ($height > 0) ? $height : (int) $this->getConfigData('altura_padrao'); $fitSize = (float) $product->getData('fit_size'); if ($item->getQty() > 1 && is_numeric($fitSize) && $fitSize > 0) { $totalSize = $height + ($fitSize * ($item->getQty() - 1)); $height = $totalSize / $item->getQty(); } return $height; } /** * Splits the package in two parts. * If the package is already splited, each piece will be splited in two equal parts. * * @return boolean|PedroTeixeira_Correios_Model_Carrier_CorreiosMethod */ protected function _splitPack() { $isSplitEnabled = $this->getConfigFlag('split_pack'); $isMethodAvailable = (count($this->_postMethodsExplode) > 0); if ($isSplitEnabled && $isMethodAvailable) { $this->_splitUp++; $this->_volumeWeight /= 2; $this->_packageWeight /= 2; $this->_packageValue /= 2; return $this->_removeInvalidServices(); } return false; } /** * Receive a list of methods, and validate one-by-one using the config settings. * Returns a list of valid methods or empty. * * @param array $postmethods Services List * * @return array */ protected function _filterMethodByConfigRestriction($postmethods) { $validMethods = array(); $this->_loadMidSize(); foreach ($postmethods as $key => $method) { $isOverSize = ($this->_midSize > $this->getConfigData("validate/serv_{$method}/max/size")); $isOverSize |= ($this->_midSize * 3 > $this->getConfigData("validate/serv_{$method}/max/sum")); $isOverWeight = ($this->_packageWeight > $this->getConfigData("validate/serv_{$method}/max/weight")); $isOverCubic = ($this->_volumeWeight > $this->getConfigData("validate/serv_{$method}/max/volume_weight")); $isZipAllowed = $this->_validateZipRestriction($method); if (!$isOverSize && !$isOverWeight && !$isOverCubic && $isZipAllowed) { $validMethods[] = $method; } } return $validMethods; } /** * Loads the zip range list. * Returns TRUE only if zip target is included in the range. * * @param array $method Current Post Method * * @return boolean */ protected function _validateZipRestriction($method) { $zipConfig = $this->getConfigData("validate/serv_{$method}/zips"); foreach ($zipConfig as $data) { $zipRange = explode(',', $data); $isBetweenRange = true; $isBetweenRange &= ($this->_toZip >= $zipRange[0]); $isBetweenRange &= ($this->_toZip <= $zipRange[1]); if ($isBetweenRange) { return true; } } return false; } /** * Some special errors must be sent to users. * If not applicable, the default error will be sent. * * @param array $errorList Error List * * @return boolean */ protected function _appendShippingErrors($errorList) { $output = false; $successCode = '0'; $hasValidQuote = array_key_exists($successCode, $errorList); if (!$hasValidQuote) { $displayErrorList = explode(',', $this->getConfigData('hard_errors')); if ($this->getConfigFlag('show_soft_errors')) { $softErrorList = explode(',', $this->getConfigData('soft_errors')); $displayErrorList = array_merge($displayErrorList, $softErrorList); } foreach ($errorList as $errorCode => $errorMsg) { $isDisplayError = in_array($errorCode, $displayErrorList); if ($isDisplayError) { $error = Mage::getModel('shipping/rate_result_error'); $error->setCarrier($this->_code); $error->setErrorMessage($errorMsg); $this->_result->append($error); $output = true; } } if (!$output) { $logMsg = implode(',', $errorList); Mage::log("{$this->_code}: Warning! There is no valid quotes, and no one error was throwed: {$logMsg}"); } } return $output; } /** * Returns a short message showing the number of the packs that will be needed. * * @return string */ protected function _getSplitUpMsg() { $msg = ""; if ($this->_splitUp > 0) { $qty = pow(2, $this->_splitUp); $msg.= " / {$qty} volumes"; } return $msg; } /** * Returns a short warning message. * * @param string $error Error Id * * @return string */ protected function _getSoftErrorMsg($error) { $msg = ""; if ($this->getConfigFlag('show_soft_errors')) { $softErrorList = explode(',', $this->getConfigData('soft_errors')); $isSoftError = in_array($error, $softErrorList); if ($isSoftError) { $msg.= " / Área de Risco"; } } return $msg; } /** * Returns the price as float, and fixed by pack division. * * @param string $price Price String * * @return float */ protected function _getFormatPrice($price) { $stringPrice = str_replace('.', '', $price); $stringPrice = str_replace(',', '.', $stringPrice); $shippingPrice = floatval($stringPrice); $shippingPrice *= pow(2, $this->_splitUp); return $shippingPrice; } /** * Filter visible and bundle children products. * * @param array $items Product Items * * @return array */ protected function _loadBundleChildren($items) { $visibleAndBundleChildren = array(); /* @var $item Mage_Sales_Model_Quote_Item */ foreach ($items as $item) { $product = $item->getProduct(); $isBundle = ($product->getTypeId() == Mage_Catalog_Model_Product_Type::TYPE_BUNDLE); if ($isBundle) { /* @var $child Mage_Sales_Model_Quote_Item */ foreach ($item->getChildren() as $child) { $visibleAndBundleChildren[] = $child; } } else { $visibleAndBundleChildren[] = $item; } } return $visibleAndBundleChildren; } } ` Obrigado @fernandoplanejar91 Consegui fazer funcionar com a modificação inteira que você enviou. E muito obrigado @mateussantin pela solução!!! Olá @fernandoplanejar91 e @mateussantin, tentei colar o código acima e não deu certo. Vocês tem como enviar o arquivo por gentileza? Obrigado pela ajuda. Segue arquivos em anexo corrigidos com a opção no admin para atualizar o valor dos correios sempre que necessário (atualmente 79 reais). (apagar o .txt do final do arquivo e substituir). ect/system.xml model/carrier/correiosMethod.php CorreiosMethod.php.txt system.xml.txt Segue arquivos em anexo corrigidos com a opção no admin para atualizar o valor dos correios sempre que necessário (atualmente 79 reais). (apagar o .txt do final do arquivo e substituir). ect/system.xml model/carrier/correiosMethod.php CorreiosMethod.php.txt system.xml.txt @mateussantin, muito obrigado pela disposição em ajudar. Funcionou corretamente. Abs, Olá pessoal, Encontrei um erro, quando você compra somente 1 produto na medida por exemplo 105x3x3 é calculado o aumento de R$ 79,00 e o frete fica R$ 124,50, OK, até ai funcionando. Mas se eu adicionar também no carrinho um produto menor que 70cm, por exemplo 16x2x11 o frete fica R$ 47,80, ou seja, quando tem mais de um produto no carrinho ele está ignorando que um deles é maior que 70cm. Por favor alguém sabe uma solução para esse caso? Abraços! Também estou com esse problema, alguém conseguiu resolver??? Olá pessoal, Encontrei um erro, quando você compra somente 1 produto na medida por exemplo 105x3x3 é calculado o aumento de R$ 79,00 e o frete fica R$ 124,50, OK, até ai funcionando. Mas se eu adicionar também no carrinho um produto menor que 70cm, por exemplo 16x2x11 o frete fica R$ 47,80, ou seja, quando tem mais de um produto no carrinho ele está ignorando que um deles é maior que 70cm. Por favor alguém sabe uma solução para esse caso? Abraços! Bom dia @ricardopi70 Essa alteração reajusta o valor da taxa de manipulação, caso algum dos itens do carrinho exceda 70 cm. Localize o trecho a seguir: https://github.com/pedro-teixeira/correios/blob/2c828ac48627de551b7faed83f22f814b5c8bdef/app/code/community/PedroTeixeira/Correios/Model/Carrier/CorreiosMethod.php#L486 ...e cole esse código, exatamente após o trecho localizado. if (empty($this->getIsCorreiosHandlingFixed()) && $sizeMax > 70) { $correiosFee = 20; $handlingFee = $this->getConfigData('handling_fee'); $path = "carriers/{$this->getCarrierCode()}/handling_fee"; Mage::app()->getStore($this->getStore())->setConfig($path, $correiosFee + $handlingFee); $this->setIsCorreiosHandlingFixed(true); } P.S. Habilitar opção de Validar Dimensões dos Produtos Não compatível com as modificações previamente sugeridas neste tópico. Aplicar este ajuste nos arquivos originais do módulo. Obrigado @rafaelpatro, está funcionando corretamente agora. Valeu! Para mim não atualizou nada, mesmo baixando e instalando esses arquivos Segue arquivos em anexo corrigidos com a opção no admin para atualizar o valor dos correios sempre que necessário (atualmente 79 reais). (apagar o .txt do final do arquivo e substituir). ect/system.xml model/carrier/correiosMethod.php CorreiosMethod.php.txt system.xml.txt Olá boa noite Gostaria saber como o Pedro Teixeira chama os atributos de altura largura e profundidade? Segue a imagem de exemplo como a Jadlog chama Obrigada pela atenção Malka Gruenberg Bom dia @ricardopi70 Essa alteração reajusta o valor da taxa de manipulação, caso algum dos itens do carrinho exceda 70 cm. Localize o trecho a seguir: https://github.com/pedro-teixeira/correios/blob/2c828ac48627de551b7faed83f22f814b5c8bdef/app/code/community/PedroTeixeira/Correios/Model/Carrier/CorreiosMethod.php#L486 ...e cole esse código, exatamente após o trecho localizado. if (empty($this->getIsCorreiosHandlingFixed()) && $sizeMax > 70) { $correiosFee = 20; $handlingFee = $this->getConfigData('handling_fee'); $path = "carriers/{$this->getCarrierCode()}/handling_fee"; Mage::app()->getStore($this->getStore())->setConfig($path, $correiosFee + $handlingFee); $this->setIsCorreiosHandlingFixed(true); } P.S. Habilitar opção de Validar Dimensões dos Produtos Não compatível com as modificações previamente sugeridas neste tópico. Aplicar este ajuste nos arquivos originais do módulo. Olá a todos, agradeço a colaboração do pessoal. Ficou muito boa a solução. Mesmo com a correção do @ricardopi70, tive problemas ainda quando eu coloco apenas 1 produto no carrinho com tamanho acima de 70 ele estava adicionando 2 vezes o valor do manuseio. Fiz de uma outra forma e estou colocando aqui para vocês, caso mais alguém esteja com o mesmo problema. public $check_size = 0; protected function _generateAument() { $items = Mage::getModel('checkout/cart')->getQuote()->getAllVisibleItems(); foreach($items as $item){ $_product = $this->_getSimpleProduct($item->getProduct()); if ($_product->getData('volume_altura') == '' || (int) $_product->getData('volume_altura') == 0) { $itemAltura = $this->getConfigData('altura_padrao'); } else { $itemAltura = $_product->getData('volume_altura'); } if ($_product->getData('volume_largura') == '' || (int) $_product->getData('volume_largura') == 0) { $itemLargura = $this->getConfigData('largura_padrao'); } else { $itemLargura = $_product->getData('volume_largura'); } if ($_product->getData('volume_comprimento') == '' || (int) $_product->getData('volume_comprimento') == 0) { $itemComprimento = $this->getConfigData('comprimento_padrao'); } else { $itemComprimento = $_product->getData('volume_comprimento'); } if ($itemAltura > 70 || $itemLargura > 70 || $itemComprimento > 70) { $this->check_size = 1; } } if ($this->check_size == 1) { $this->add_aumento = $this->getConfigData('aument'); } else { $this->add_aumento; } return true; }
gharchive/issue
2018-09-05T17:53:40
2025-04-01T04:35:28.845361
{ "authors": [ "cesaroito", "fernandoplanejar91", "gilliattbastos", "malkahakuk", "mateussantin", "rafaelpatro", "ricardopi70" ], "repo": "pedro-teixeira/correios", "url": "https://github.com/pedro-teixeira/correios/issues/393", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
867053084
RSTP Server? anyone have any recommendation on using online rtsp server? thanks I'm currently using rtmp with Youtube and it works great. You have a list of servers in the wiki page: https://github.com/pedroSG94/rtmp-rtsp-stream-client-java/wiki/Media-servers-tested
gharchive/issue
2021-04-25T16:28:35
2025-04-01T04:35:28.852909
{ "authors": [ "ToysoftInc", "pedroSG94" ], "repo": "pedroSG94/rtmp-rtsp-stream-client-java", "url": "https://github.com/pedroSG94/rtmp-rtsp-stream-client-java/issues/846", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
115136395
Optional custom top-level domains Hello! I've added opt-in support for custom (non-standard) top-level domains, primarily for supporting use of parse-domains on local environments, where the TLD could feasibly be set to anything. The aim was to implement this in a way that hopefully eases factoring between development and production builds, via the parseDomain.applyCustomTlds() method – it returns a function that matches the original signature of the parse-domains module, so it can feasibly be exported in a custom app-specific module and used in development builds in place of the "vanilla" version of parse-domains. 😄 Thx @mattlubner. Sorry for not responding, I get too much notifications and it's hard to filter the important ones (like PRs). I really like your suggestion and I see the necessity to define custom TLDs. :+1: However, I am not quite happy with the proposed API. I usually try to keep the API surface as small as possible. Your proposed applyCustomTlds() introduces a whole new way to configure the module, which I believe is not necessary for most use-cases. I also don't like to add an optional argument customTlds because I always stumble upon these arguments when reading the source code weeks/years later again ("what was this array for? ... ahh these are custom tlds"). What about passing an options object as second argument? Like parseDomain("some.domain.local", { customTlds: ["local"] }) then you could still make your own version of the function function parseCustomTlds(url) { return parseDomain(url, { customTlds: ["local"] }); } Additionally I would not accept either a string or an array. Imho it's best to avoid ambiguity with options (unless it provides a huge usability benefit). Since it's a list of things, arrays are the best type for this option. What do you think? Apart from that, how does your implementation handle custom tlds like "dev.local"? :grinning: Could you add tests for these cases? Thx @mattlubner. Sorry for not responding, I get too much notifications and it's hard to filter the important ones (like PRs). I really like your suggestion and I see the necessity to define custom TLDs. :+1: However, I am not quite happy with the proposed API. I usually try to keep the API surface as small as possible. Your proposed applyCustomTlds() introduces a whole new way to configure the module, which I believe is not necessary for most use-cases. I also don't like to add an optional argument customTlds because I always stumble upon these arguments when reading the source code weeks/years later again ("what was this array for? ... ahh these are custom tlds"). What about passing an options object as second argument? Like parseDomain("some.domain.local", { customTlds: ["local"] }) then you could still make your own version of the function function parseCustomTlds(url) { return parseDomain(url, { customTlds: ["local"] }); } Additionally I would not accept either a string or an array. Imho it's best to avoid ambiguity with options (unless it provides a huge usability benefit). Since it's a list of things, arrays are the best type for this option. What do you think? Apart from that, how does your implementation handle custom tlds like "dev.local"? :grinning: Could you add tests for these cases? Wonderful suggestions! I also added a custom RegExp pass-thru as well, to cut down on repeated hydration of new RegExp objects (eg, for use within a server). Am cleaning up the README.md a bit right now… 😄 @jhnns Alrighty, I've tidied up this PR; hopefully it's ready to merge in! No biggie if you have some more feedback though, I'm happy to oblige. I don't at all mind removing the partial application helper – it's trivial to move that code into consuming modules, and you know, it actually looks a bit more efficient like this! 😉 Btw, I've never come across README.md syntax like this before. I tidied it up a bit, but apologies if it's auto-generated and I subsequently made things difficult for ya… 😄 @matthaias thanks for your effort. That looks good now. There's just one thing that bothers me: I think it would be better to allow the config option customTlds to be either a regexp or an array. The current implementation allows to specify both at the same time which may confuse users (which option takes precedence?). Thanks for correcting grammar mistakes in the README. Regarding the order, I prefer examples before any installation instructions :grin:. But you can just leave the README, I'll adjust it then. @jhnns Glad to help out! 😎 I'm actually not too keen on options.customTldsRegExp either. It seemed like the best way to reduce RegExp constructions, although if you aren't against the library overwriting options.customTlds with the constructed RegExp object, that definitely simplifies the API! The current implementation is good, thank you :+1: Shipped with 0.2.1
gharchive/pull-request
2015-11-04T20:07:56
2025-04-01T04:35:28.889183
{ "authors": [ "jhnns", "mattlubner" ], "repo": "peerigon/parse-domain", "url": "https://github.com/peerigon/parse-domain/pull/3", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2080338954
Simplifying Setup As of now, the repo is not very well organized to get started with running the experiments. Makes it difficult for a new user to come to the repo and start scaling experiments. The aim here is to make it easy for any user to come and start running experiments as easily as possible. [ ] run-factorizer-enc script, has hard-coded values that need to be changed when running the script. [ ] Less information on how to get the data and factorizer models and where to place them. [ ] Setting up the python environment for running the experiment. Seems like conda works a bit differently than venv and it creates confusion Modifiable parts for any experiment : Config files : base.prep.yml and base.conf.yml. Whenever, we will have to change the language pair, this needs to be changed. Bash Script Files : run-factorizer-enc.sh. Here, we have hardcoded parameter values which change from experiment to experiment. They change more frequently. Getting Datasets Although, it is easy to get the files and tokenize them using mosestokenizer, but for different language pair tokenization can vary. Can \ Shall we automate this process ? As of now, here are the things I am working on to make this process easier : [ ] Add a script to get the factorizer models and setting up the datasets directory ( if feasible ) [ ] Removing hardcoded parameters from bash run scripts [ ] Minimize the number of parameters, using repo_setup.env file ( stores the variables like, dataset setup dir, factorizer model dir, exp collection dir, config store dir ) [ ] (Maybe a script to generate base conf files for different language pairs )
gharchive/issue
2024-01-13T15:12:25
2025-04-01T04:35:28.925116
{ "authors": [ "pegasus-lynx" ], "repo": "pegasus-lynx/bytetok-nmt", "url": "https://github.com/pegasus-lynx/bytetok-nmt/issues/1", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
603143123
When workable version of code will be uploaded? When workable version of code will be uploaded? Hi, I will have some time working on it in the next few days and hopefully will release a workable version sometime in the next week. @peiyunh Any updates on workable version?
gharchive/issue
2020-04-20T10:53:50
2025-04-01T04:35:28.973511
{ "authors": [ "chowkamlee81", "peiyunh" ], "repo": "peiyunh/opcseg", "url": "https://github.com/peiyunh/opcseg/issues/1", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2032919558
ImageCLEFmedCaption_2022 Please tell me how to obtain the ImageCLEFmedCaption_2022 data set. It seems that the clf official website cannot download this data set now. Thank you! Please tell me how to obtain the ImageCLEFmedCaption_2022 data set. It seems that the clf official website cannot download this data set now. Thank you! Do you have a download method for this dataset?Thank you
gharchive/issue
2023-12-08T15:56:31
2025-04-01T04:35:29.059675
{ "authors": [ "deng1604", "tntdc585" ], "repo": "pengfeiliHEU/M2I2", "url": "https://github.com/pengfeiliHEU/M2I2/issues/11", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
740676055
Property 'yaml' does not exist on type 'typeof languages' i'm using ngx-monaco-editor (https://github.com/atularen/ngx-monaco-editor#readme) library in my Angular project. also i would like to use monaco-yaml: monaco.languages.yaml.yamlDefaults.setDiagnosticsOptions({ ... }) for some reason i get following error: error TS2339: Property 'yaml' does not exist on type 'typeof languages'. i cannot understand, what's the problem: ngx-monaco-editor already contains monaco-editor lib, and monaco.languages is defined, but there is no yaml property. that's kinda strange for me, since in monaco-yaml is declared same monaco namespace with property yaml, so it's global scope, but my code see only monaco namespase properties defined in ngx-monaco-editor i'll be happy if anyone has any ideas about what could that be I have the same issue. I am not sure why typescript can't find the Monaco.d.ts. Can you for now ignore that by // @ts-ignore until we find a solution for this.
gharchive/issue
2020-11-11T11:01:37
2025-04-01T04:35:29.064329
{ "authors": [ "mraimon", "yazaabed" ], "repo": "pengx17/monaco-yaml", "url": "https://github.com/pengx17/monaco-yaml/issues/37", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
55009681
[PMD-994] - Add Redshift dialect to Pentaho Metadata add Redshift to kettle-database-types.xml @mbatchelor @lucboudreau this fixes failing projects that use metadata(like PIR) Do we need this as a native plugin,r would it work as a PDI Database plugin? As a rule I'd like to kethe list of native types small if possible, as the default (built-in) list is quite long already. If this needs to be a native plugin then that's fine, just wondering. Matt, I've merged this one because it is a followup of something previously checked in.
gharchive/pull-request
2015-01-21T11:58:50
2025-04-01T04:35:29.099204
{ "authors": [ "aliakseihaidukou", "lucboudreau", "mattyb149" ], "repo": "pentaho/pentaho-kettle", "url": "https://github.com/pentaho/pentaho-kettle/pull/965", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2047416728
Endpoints and client work for Search #10 PR fills out the skeleton for #10. In progress, but tasks tackled thus far: Added basic client route for search results, /search/[query] SearchBar now redirects to /search/[query] updating /api/search to return result oriented data updating validators for this data + checking query results ~remaining TODO is basic search result table.~ basic search result table implemented now, too.
gharchive/pull-request
2023-12-18T20:40:51
2025-04-01T04:35:29.103754
{ "authors": [ "ejmg" ], "repo": "penumbra-zone/cuiloa", "url": "https://github.com/penumbra-zone/cuiloa/pull/44", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2195963207
cnidarium: implement incremental jmt migration Close #3506, WIP rustc segfault? 😈 Yeah for sure, waiting for CI to pass
gharchive/pull-request
2024-03-19T20:39:08
2025-04-01T04:35:29.104815
{ "authors": [ "erwanor" ], "repo": "penumbra-zone/penumbra", "url": "https://github.com/penumbra-zone/penumbra/pull/4053", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
162867779
Wrong warning for public imports Just like #381 this produces a warning one cannot get rid of. package T1 public class A package T2 import public T1 package T3 import T2 <-- The import T2 is never used directly class B extends A When removed this of course causes a compilation error. Well, you can get rid of it by importing T1 instead of T2. Why do you want to import T2, if you are not using anything from T2 directly? because I have many public imports in T1 and many in T2 that are used in many other package where I want to keep the import list small by moving them into public imports. fixed and now a warning shows you excessive imports - nice.
gharchive/issue
2016-06-29T08:25:32
2025-04-01T04:35:29.140653
{ "authors": [ "Frotty", "peq" ], "repo": "peq/WurstScript", "url": "https://github.com/peq/WurstScript/issues/449", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1581407566
[Question] Why PerAct couldn't solve "Push Buttons" (compares to HiveFormer)? Dear author, Thank you for sharing your awesome work! In the readme, you mention that " Impossible tasks: Some tasks like push_buttons are not solvable by PerAct since it doesn't have any memory." Is it because the PerAct's policy is P(a_t | O_t, Instruction) ? Since history observation is not provided, there's no way for PerAct to keep track of its progress for Instruction following ? Best, @berniewang8177, yes. If the task is "press the red button, then the blue button, and then the green button", the agent won't know which button was pressed previously, and will have to guess which part of the sequence it's at.
gharchive/issue
2023-02-12T20:58:56
2025-04-01T04:35:29.142890
{ "authors": [ "MohitShridhar", "berniewang8177" ], "repo": "peract/peract", "url": "https://github.com/peract/peract/issues/14", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2568211603
K8SPG-648 add PG17 support CHANGE DESCRIPTION Problem: Short explanation of the problem. Cause: Short explanation of the root cause of the issue if applicable. Solution: Short explanation of the solution we are providing with this PR. CHECKLIST Jira [ ] Is the Jira ticket created and referenced properly? [ ] Does the Jira ticket have the proper statuses for documentation (Needs Doc) and QA (Needs QA)? [ ] Does the Jira ticket link to the proper milestone (Fix Version field)? Tests [ ] Is an E2E test/test case added for the new feature/change? [ ] Are unit tests added where appropriate? Config/Logging/Testability [ ] Are all needed new/changed options added to default YAML files? [ ] Did we add proper logging messages for operator actions? [ ] Did we ensure compatibility with the previous version or cluster upgrade process? [ ] Does the change support oldest and newest supported PG version? [ ] Does the change support oldest and newest supported Kubernetes version? Test name Status custom-extensions passed custom-tls passed demand-backup passed finalizers passed init-deploy passed major-upgrade failure monitoring passed one-pod passed operator-self-healing passed pitr passed scaling passed scheduled-backup passed self-healing passed start-from-backup passed tablespaces passed telemetry-transfer passed upgrade-consistency passed upgrade-minor passed users passed We run 19 out of 19 commit: https://github.com/percona/percona-postgresql-operator/pull/921/commits/f651334605a27f83266361e92758941f5320c5a5 image: perconalab/percona-postgresql-operator:PR-921-f65133460 Test name Status custom-extensions passed custom-tls passed demand-backup passed finalizers passed init-deploy passed major-upgrade failure monitoring passed one-pod passed operator-self-healing passed pitr passed scaling passed scheduled-backup passed self-healing passed start-from-backup passed tablespaces passed telemetry-transfer passed upgrade-consistency passed upgrade-minor passed users passed We run 19 out of 19 commit: https://github.com/percona/percona-postgresql-operator/pull/921/commits/551a6e8091adaa99e84ee841d178494f2a5a9717 image: perconalab/percona-postgresql-operator:PR-921-551a6e809 Test name Status custom-extensions passed custom-tls passed demand-backup passed finalizers passed init-deploy failure major-upgrade failure monitoring passed one-pod passed operator-self-healing passed pitr passed scaling passed scheduled-backup passed self-healing failure start-from-backup passed tablespaces passed telemetry-transfer passed upgrade-consistency passed upgrade-minor passed users passed We run 19 out of 19 commit: https://github.com/percona/percona-postgresql-operator/pull/921/commits/4019bb7cb78ef3d118035f4bf63e8ed7233e8c78 image: perconalab/percona-postgresql-operator:PR-921-4019bb7cb Test name Status custom-extensions passed custom-tls passed demand-backup passed finalizers passed init-deploy passed major-upgrade failure monitoring passed one-pod passed operator-self-healing passed pitr passed scaling passed scheduled-backup passed self-healing passed start-from-backup passed tablespaces passed telemetry-transfer passed upgrade-consistency passed upgrade-minor passed users passed We run 19 out of 19 commit: https://github.com/percona/percona-postgresql-operator/pull/921/commits/4019bb7cb78ef3d118035f4bf63e8ed7233e8c78 image: perconalab/percona-postgresql-operator:PR-921-4019bb7cb Test name Status custom-extensions passed custom-tls passed demand-backup passed finalizers passed init-deploy passed monitoring passed one-pod passed operator-self-healing passed pitr passed scaling passed scheduled-backup passed self-healing passed start-from-backup passed tablespaces passed telemetry-transfer passed upgrade-consistency passed upgrade-minor passed users passed We run 18 out of 18 commit: https://github.com/percona/percona-postgresql-operator/pull/921/commits/1b55139985d6f3e1b7aac3465be2c8662db75d7e image: perconalab/percona-postgresql-operator:PR-921-1b5513998 Test name Status custom-extensions failure custom-tls failure demand-backup failure finalizers failure init-deploy failure monitoring failure one-pod failure operator-self-healing failure pitr failure scaling failure scheduled-backup failure self-healing failure start-from-backup failure tablespaces failure telemetry-transfer failure upgrade-consistency failure upgrade-minor passed users failure We run 18 out of 18 commit: https://github.com/percona/percona-postgresql-operator/pull/921/commits/3f4916ea622d037bce8aa17f1790d7c2364d6b02 image: perconalab/percona-postgresql-operator:PR-921-3f4916ea6 Test name Status custom-extensions failure custom-tls passed demand-backup failure finalizers passed init-deploy passed monitoring failure one-pod failure operator-self-healing passed pitr passed scaling passed scheduled-backup passed self-healing failure start-from-backup failure tablespaces failure telemetry-transfer passed upgrade-consistency failure upgrade-minor failure users passed We run 18 out of 18 commit: https://github.com/percona/percona-postgresql-operator/pull/921/commits/51fc6b5b59511d5d85138de834e005df6735c0a0 image: perconalab/percona-postgresql-operator:PR-921-51fc6b5b5 Test name Status custom-extensions passed custom-tls passed demand-backup failure finalizers passed init-deploy passed monitoring failure one-pod failure operator-self-healing passed pitr passed scaling passed scheduled-backup passed self-healing failure start-from-backup failure tablespaces failure telemetry-transfer passed upgrade-consistency passed upgrade-minor passed users passed We run 18 out of 18 commit: https://github.com/percona/percona-postgresql-operator/pull/921/commits/5469b2b45001a8a4465a6a98e18564b87ddd9e29 image: perconalab/percona-postgresql-operator:PR-921-5469b2b45 Test name Status custom-extensions passed custom-tls passed demand-backup passed finalizers passed init-deploy passed monitoring passed one-pod passed operator-self-healing passed pitr passed scaling passed scheduled-backup passed self-healing passed start-from-backup passed tablespaces passed telemetry-transfer passed upgrade-consistency passed upgrade-minor passed users passed We run 18 out of 18 commit: https://github.com/percona/percona-postgresql-operator/pull/921/commits/83fd40ee01f1950d3b8e658e3406e905e2dd76c1 image: perconalab/percona-postgresql-operator:PR-921-83fd40ee0 Test name Status custom-extensions passed custom-tls passed demand-backup failure finalizers passed init-deploy passed monitoring passed one-pod passed operator-self-healing passed pitr passed scaling passed scheduled-backup passed self-healing passed sidecars passed start-from-backup passed tablespaces passed telemetry-transfer passed upgrade-consistency passed upgrade-minor failure users passed We run 19 out of 19 commit: https://github.com/percona/percona-postgresql-operator/pull/921/commits/86af60d8ef9a59eb95e49c6adb605b05e2e5bec2 image: perconalab/percona-postgresql-operator:PR-921-86af60d8e Test name Status custom-extensions passed custom-tls passed demand-backup passed finalizers passed init-deploy passed monitoring passed one-pod passed operator-self-healing passed pitr passed scaling passed scheduled-backup passed self-healing passed sidecars passed start-from-backup passed tablespaces passed telemetry-transfer passed upgrade-consistency passed upgrade-minor passed users passed We run 19 out of 19 commit: https://github.com/percona/percona-postgresql-operator/pull/921/commits/4eae78c7ff6c0e867380c2826240ae3e1ef75841 image: perconalab/percona-postgresql-operator:PR-921-4eae78c7f Test name Status custom-extensions failure custom-tls passed demand-backup passed finalizers passed init-deploy passed monitoring failure one-pod passed operator-self-healing passed pitr passed scaling passed scheduled-backup passed self-healing passed sidecars passed start-from-backup passed tablespaces passed telemetry-transfer passed upgrade-consistency passed upgrade-minor passed users passed We run 19 out of 19 commit: https://github.com/percona/percona-postgresql-operator/pull/921/commits/a2497125c5120491bd3255ef0bfd8bc4bcb326e0 image: perconalab/percona-postgresql-operator:PR-921-a2497125c Test name Status custom-extensions passed custom-tls passed demand-backup passed finalizers passed init-deploy passed monitoring passed one-pod passed operator-self-healing passed pitr passed scaling passed scheduled-backup passed self-healing passed sidecars passed start-from-backup passed tablespaces passed telemetry-transfer passed upgrade-consistency passed upgrade-minor passed users passed We run 19 out of 19 commit: https://github.com/percona/percona-postgresql-operator/pull/921/commits/c3fe5387299f0f6ec20ef79ba2b327db8275770e image: perconalab/percona-postgresql-operator:PR-921-c3fe53872
gharchive/pull-request
2024-10-05T16:40:26
2025-04-01T04:35:29.284466
{ "authors": [ "JNKPercona", "hors" ], "repo": "percona/percona-postgresql-operator", "url": "https://github.com/percona/percona-postgresql-operator/pull/921", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
164608658
/routine/Str contains doc for method Str.val sub val is a sub among methods what may cause the confusion. Although how we generate /routine/*s can be improved, the very issue has disappeared(?) Please check it. Methods … … sub val No, that's not resolved. Most of we list all routines available below heading "Methods", while someone substitutes it with "Subroutine". It's an issue of doc conventions. What is the issue? Moving the documentation elsewhere? I have been checking Int and it does list routines under Methods, although it uses the word routine, instead of sub. chr would be in the same situation as val right? Would changing sub to routine help here? Besides, Str itself has several routines And I see there was an attempt to convert to routine which failed for some reason https://github.com/perl6/doc/commit/e66006021d55b8e72b13989cddebebd770ed37c9
gharchive/issue
2016-07-08T20:19:58
2025-04-01T04:35:29.343599
{ "authors": [ "AlexDaniel", "JJ", "W4anD0eR96", "gfldex" ], "repo": "perl6/doc", "url": "https://github.com/perl6/doc/issues/683", "license": "Artistic-2.0", "license_type": "permissive", "license_source": "github-api" }
213518633
Add Perl 6 from haskell / python Add Perl 6 from haskell / python to "Resources" Looks good. Thanks :)
gharchive/pull-request
2017-03-11T09:42:05
2025-04-01T04:35:29.353484
{ "authors": [ "hankache", "samcv" ], "repo": "perl6/perl6.org", "url": "https://github.com/perl6/perl6.org/pull/79", "license": "Artistic-2.0", "license_type": "permissive", "license_source": "github-api" }
952134158
Automatic Release | Validate opa/opal version added workflow to validate opa/opal version added release workflow to GitHub & pypi it is also possible to add release notes to the automation (ref: https://github.com/softprops/action-gh-release) fixed the trigger for on release workflow putting it on hold for now, @RazcoDev you might want to give it a go once you return. I am opening a ticket on you (PER-1052).
gharchive/pull-request
2021-07-24T19:24:47
2025-04-01T04:35:29.382100
{ "authors": [ "asafc", "nir1212" ], "repo": "permitio/opal", "url": "https://github.com/permitio/opal/pull/114", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2215572662
Error when importing lib in Next 14 I'm submitting a ... [x] bug report [ ] feature request [ ] question about the decisions made in the repository [ ] question about how to use this project Summary I am getting below error, while integrating this with Nextjs14: Any idea why? This is the snippet: > 4 | const permit = new Permit({ token: process.env.PERMIT_IO_API_KEY }); As a standalone js file that syncs resources works fine. Other information (e.g. detailed explanation, stack traces, related issues, suggestions how to fix, links for us to have context, eg. StackOverflow, personal fork, etc.) Nextjs14 Fixed in version 2.4.1. Thanks for reporting.
gharchive/issue
2024-03-29T15:26:38
2025-04-01T04:35:29.385637
{ "authors": [ "jeremyrx7", "obsd" ], "repo": "permitio/permit-node", "url": "https://github.com/permitio/permit-node/issues/76", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1834789641
Choose a validation lib and use it to make basic validation on each form From the discussion we had with @sjcobb @Gladorme, best candidates for now seems to be a conjunction of https://react-hook-form.com/ + https://zod.dev/ From the discussion we had with @sjcobb @Gladorme, best candidates for now seems to be a conjunction of https://react-hook-form.com/ + https://zod.dev/ Yes, let’s start experimenting with this asap and agreed that the libs in description seem like top candidates All big forms are validated by Zod and React-Hook-Form (#2044), except for plugin spec inputs. It still need to be done. However, plugin should be able to inject their validation schema in the ValidationProvider in order to have full validation.
gharchive/issue
2023-08-03T10:33:16
2025-04-01T04:35:29.394122
{ "authors": [ "Gladorme", "celian-garcia", "sjcobb" ], "repo": "perses/perses", "url": "https://github.com/perses/perses/issues/1348", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1618284813
[ENHANCEMENT] Fetch updates to show full error message Update fetch to throw UserFriendlyError, tested with invalid panel properties Screenshots Checklist [ ] Pull request has a descriptive title and context useful to a reviewer. [ ] Pull request title follows the [<catalog_entry>] <commit message> naming convention using one of the following catalog_entry values: FEATURE, ENHANCEMENT, BUGFIX, BREAKINGCHANGE, IGNORE. [ ] All commits have DCO signoffs. [ ] Changes that impact the UI include screenshots and/or screencasts of the relevant changes. yeah I'm totally fine with this change ! For sure we can try to improve the error message but sometimes it's quite harder unfortunately :(. When it's related to the cue, hopefully the next version of cue will help around that.
gharchive/pull-request
2023-03-10T03:03:44
2025-04-01T04:35:29.398389
{ "authors": [ "Nexucis", "sjcobb" ], "repo": "perses/perses", "url": "https://github.com/perses/perses/pull/1015", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
786584413
🛑 Procal-tech is down In c6d0796, Procal-tech (http://iot.procal-tech.com) was down: HTTP code: 0 Response time: 0 ms Resolved: Procal-tech is back up in b4e4d17. Resolved: Procal-tech is back up in b4e4d17.
gharchive/issue
2021-01-15T06:12:21
2025-04-01T04:35:29.401498
{ "authors": [ "personlin" ], "repo": "personlin/website_uptime", "url": "https://github.com/personlin/website_uptime/issues/76", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
186411159
Updated Jetty dependency and a test broken by Jetty changes. See issue: #694 9.4.0.v20161208 is out now @ph-hs I noticed, started updating the pull request last night, but there were complications. It is interesting that it is only the test that's broken with the Jetty upgrade. In real life, you can do the maven thing and deploy Spark with the 9.4.x of Jetty quite happily. The bits that you're stuck on could be, solved by: - WebSocketUpgradeFilter webSocketUpgradeFilter = - (WebSocketUpgradeFilter) servletContextHandler.getAttribute("org.eclipse.jetty.websocket.server.WebSocketUpgradeFilter"); + WebSocketUpgradeFilter webSocketUpgradeFilter = (WebSocketUpgradeFilter) servletContextHandler.getServletHandler().getFilter("Jetty_WebSocketUpgradeFilter").getFilter(); and - WebSocketServerFactory webSocketServerFactory = webSocketUpgradeFilter.getFactory(); + WebSocketServerFactory webSocketServerFactory = webSocketUpgradeFilter.getConfiguration().getFactory(); and MappedResource<WebSocketCreator> mappedResource = webSocketUpgradeFilter.getMappings().getMatch("/websocket"); + MappedResource<WebSocketCreator> mappedResource = webSocketUpgradeFilter.getConfiguration().getMatch("/websocket"); Thanks, I'll take a look tonight. I got burned by not getting the upgrade filter from the attributes and didn't have the time/energy to research into it last night. One of the unrelated tests is now failing: testDirectoryTraversalProtectionLocal(spark.staticfiles.StaticFilesTest) Time elapsed: 0.13 sec <<< FAILURE! java.lang.AssertionError: expected:<404> but was:<400> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:834) at org.junit.Assert.assertEquals(Assert.java:645) at org.junit.Assert.assertEquals(Assert.java:631) at spark.staticfiles.StaticFilesTest.testDirectoryTraversalProtectionLocal(StaticFilesTest.java:145) Needs to be investigated, I think. Seems the unexpected response comes all the way from Jetty: WARN org.eclipse.jetty.http.HttpParser - bad HTTP parsed: 400 Bad URI for HttpChannelOverHttp@4db7b425{r=1,c=false,a=IDLE,uri=//localhost:4567/..%5Cspark%5CSpark.class} @jakaarl would be great if you could have a look at this again.
gharchive/pull-request
2016-10-31T22:03:16
2025-04-01T04:35:29.405949
{ "authors": [ "jakaarl", "paul-hammant", "ph-hs", "tipsy" ], "repo": "perwendel/spark", "url": "https://github.com/perwendel/spark/pull/698", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
316119345
ie: fix errors found by vet This pull request was made with an automated tool. The suggested change fixes one or more problems discovered by "go vet". See https://github.com/functionary/functionary for more details on the @functionary GitHub user. Duplicate. I hope you're not planning on opening the same PR every day.
gharchive/pull-request
2018-04-20T03:56:43
2025-04-01T04:35:29.408824
{ "authors": [ "functionary", "pesco" ], "repo": "pesco/go", "url": "https://github.com/pesco/go/pull/2", "license": "ISC", "license_type": "permissive", "license_source": "github-api" }
380935999
Petablox loses context on Queue.peek() Consider the following program: import java.util.*; class Main { public static void main(String args[]) throws InterruptedException { Queue<Main> mains = new ArrayDeque(); mains.add(new Main()); mains.peek().notfound(); } public void notfound () { System.out.println("notfound"); } } When running the reachable-methods analysis, Petablox does not see that notfound is reachable. I use version b95fd27 of Petablox. Thanks for this report. We are looking at the issue and will fix it soon.
gharchive/issue
2018-11-14T23:41:11
2025-04-01T04:35:29.415991
{ "authors": [ "KihongHeo", "kalhauge" ], "repo": "petablox-project/petablox", "url": "https://github.com/petablox-project/petablox/issues/28", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
111687626
Minor memory efficiency improvement for Ruby 2.1+ Using the memory_profiler gem, I noticed a significant number of "VV" allocations during a job that checks the language of many relatively short strings. Thanks! Will see if we can get this on a 1.0.7 soon.
gharchive/pull-request
2015-10-15T19:15:03
2025-04-01T04:35:29.437783
{ "authors": [ "djudd", "peterc" ], "repo": "peterc/whatlanguage", "url": "https://github.com/peterc/whatlanguage/pull/35", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
421912693
Animated theme support It'd be fantastic to have support for animated themes. Animated GIFs or even a series of .png would do! Working on it :) Done.
gharchive/issue
2019-03-17T11:47:28
2025-04-01T04:35:29.443250
{ "authors": [ "mileyvirus", "peterprd" ], "repo": "peterprd/Xeon", "url": "https://github.com/peterprd/Xeon/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1721337512
Safari compatibility/Performance Dear Peter, after all this time I finally have a case where I can make really good use of your venerable workspacer extension! Thanks for this :) I’m not sure if you are interested in further development of it, but still I wanted to share an issue I found regarding usage in Safari 16.4. When trying to type anything into the editor the text is apparently added after the code-area element and so it turns up next to the Sava as button: An easy workaround of course is not to use Safari, since the editor works fine in Chrome and Firefox. I’m still curious why this happens though. Another issue is that it takes quite some time to load the workspacer environment if the number of files is rather large. In my specific case there is a folder containing roughly 9.500 files and it takes about a minute to load. I tried implementing the possibility to exclude certain folders, but my programming still were not sufficient … I had a brief look at the code and did not manage to see what was going wrong with the editor. It would help if I could run Safari on my computer, but my computer isn't a Mac. I am now thinking about how to redesign Workspacer so that large numbers of files will not slow down the page loading. Thanks Peter! It’s no big deal, only a minor annoyance, since it’s obviously easy to switch browsers. Better performance would be nice though, but that is also just a matter of comfortability, so please don’t put too much effort in it. Update regarding performance: I now have the same installation running on a proper server and it’s loading the same ammount of files almost instantly. ;) So it might just be a performance problem with local development setups or certain types of servers. I have added the option to exclude directories in the preferences. The editor still has bugs to fix. Brilliant, thanks!
gharchive/issue
2023-05-23T06:47:42
2025-04-01T04:35:29.457701
{ "authors": [ "animaux", "petertron" ], "repo": "petertron/workspacer", "url": "https://github.com/petertron/workspacer/issues/7", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
243247577
Process text from stdin As a potential solution to #2, I was thinking that titlecase could inspect the number of arguments it's being passed, and in case if there are no arguments, would try to read from stdin. I'm not sure if you agree with this approach or would prefer to use - (echo 'hello world' | titlecase -), similar to other *nix programs, or some command-line flag, but I'd be happy to try my hand at implementing those approaches if you'd like. Closes #2 Looks good to me. Thank you for the submission!
gharchive/pull-request
2017-07-16T15:40:56
2025-04-01T04:35:29.459421
{ "authors": [ "cust0dian", "peti" ], "repo": "peti/titlecase", "url": "https://github.com/peti/titlecase/pull/3", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
51523208
Relationship to tj's co? Promises newb here, so sorry if this is a dumb question, but could someone better informed than I comment on tj's co library as it relates to this project? Do the two projects "compete" in any way or are they complimentary? If they compete, how do they compare (and should benchmarks be added)? If they're complimentary, how so exactly? Thanks so much for any help clearing this up! why there is no benchmark for co anymore? I'm going to assume because co relies on promises nowdays - in co what you yield are promises now anyway so at best you'd yield bluebird promises (which'd still be slower than promise.coroutine) On Wed, Feb 4, 2015 at 3:51 PM, Anton Wilhelm notifications@github.com wrote: why there is no benchmark for co anymore? — Reply to this email directly or view it on GitHub https://github.com/petkaantonov/bluebird/issues/400#issuecomment-72856945 . Yes promise.coroutine can avoid allocating a new promise if you return bluebird promises to it. Other than that it's implementation efficiency differences.
gharchive/issue
2014-12-10T06:25:02
2025-04-01T04:35:29.463317
{ "authors": [ "benjamingr", "petkaantonov", "taoeffect", "timaschew" ], "repo": "petkaantonov/bluebird", "url": "https://github.com/petkaantonov/bluebird/issues/400", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
108493044
Promise.map and co throws TypeError: fn should be a function Hey, I'm refactoring my code base from a full promises api and trying to mix with generators, But I don't know how to execute a yield operator inside Bluebird.map, exports.saveAll = function* (items, options) { options || (options = {}) options.upsert = true return Promise.map(items, item => { const gen = function* () { const query = createQuery(item) const args = [query, item, options] return yield collection.findAndModify.apply(this, args) } return co(gen) }) } And somewhere else I call yield Collection.saveAll(items) ``` I'm having this error `TypeError: fn should be a function` How do I iterate a array of object with bluebird collection api and execute a generator in there? Please ask support questions on StackOverflow, this issue tracker is for feature requests and bug reports. Make sure to include the complete stacktrace of your error when you post it there, also a bit more context about the "somewhere". PS: My guess is on the this inside your gen, or that you forgot to co.wrap (or Promise.coroutine) your saveAll method so that it returns a promise. Thank's @bergus I'll ask there next time :) So I'll try what your suggestions too, But after wrapping findAndModify(co-monk) on a try catch block I see the error comes from there and I dont know why, Thx
gharchive/issue
2015-09-26T20:12:27
2025-04-01T04:35:29.466577
{ "authors": [ "bergus", "rafaeljesus" ], "repo": "petkaantonov/bluebird", "url": "https://github.com/petkaantonov/bluebird/issues/783", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2394136454
Website Loading on Desktop Browsers on Safari Description: Verify the website loads correctly on Safari (latest version) on macOS. Preconditions: The website https://oportunitatisicariere.ro/ is live and accessible. Step 1 : Open Safari on a macOS desktop. Expected Result: The website should load without any visual or functional issues. Step 2 : Navigate to https://oportunitatisicariere.ro/ Expected Result: The website should load without any visual or functional issues on Safari (latest version) on macOS. Step 3 : Observe the page load time and any potential loading issues. Expected Result: The website should load within 3 seconds. Step 4 : Verify that the layout, images, text, and other elements are displayed correctly. Expected Result: The page should fully render all elements correctly without visual issues (e.g., broken images, misaligned text). Step 5 : Interact with the website (e.g., click links, fill out forms) to ensure functionality. Expected Result: There should be no loading errors (e.g., 404 errors, server errors).
gharchive/issue
2024-07-07T17:05:44
2025-04-01T04:35:29.485202
{ "authors": [ "Georgianaredenstein" ], "repo": "peviitor-ro/oportunitatisicariere", "url": "https://github.com/peviitor-ro/oportunitatisicariere/issues/243", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2290852699
"Fântânele", "Bozieni", villages are listed in the drop-down menu. The villages are listed in the drop-down menu alongside its corresponding county and commune, according to the law Preconditions : the website (https://peviitor-ro.github.io/ui.orase/) is up and runing. After every step delete the text you typed from the search bar. Step 1 Type "Fântânele" in the search bar. Expected results The location is listed in the drop-down menu as "Sat Fântânele, PRAHOVA (Fântânele)". Step 2 Type "Bozieni" in the search bar. Expected results The location is listed in the drop-down menu as "Sat Bozieni, PRAHOVA (Fântânele)".
gharchive/issue
2024-05-11T11:35:52
2025-04-01T04:35:29.488844
{ "authors": [ "Elena1303996" ], "repo": "peviitor-ro/ui.orase", "url": "https://github.com/peviitor-ro/ui.orase/issues/5466", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2309740513
"Bârsești", "Topești", are listed in the drop-down menu "Bârsești", "Topești" are listed in the drop-down menu alongside its corresponding county and commune, according to the law Preconditions : the website (https://peviitor-ro.github.io/ui.orase/) is up and running. After every step delete the text you typed from the search bar. Step 1 Write in the search bar "Bârsești " Expected results The location appears in the drop-down menu as "Sat Bârsești , VRANCEA (Bârsești)" Step 2 Press "x" button. Expected results The location was deleted from the search bar. Step 3 Write in the search bar "Topești" Expected results The location appears in the drop-down menu as "Sat Topești , VRANCEA (Bârsești)".
gharchive/issue
2024-05-22T07:03:01
2025-04-01T04:35:29.492792
{ "authors": [ "yonutzu" ], "repo": "peviitor-ro/ui.orase", "url": "https://github.com/peviitor-ro/ui.orase/issues/6550", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2323802657
"Șopârlița" is listed in the drop-down menu Description: When user writes ”Șopârlița” the location appear in the drop-down menu next to the comune and the county which is a part according to the law. Precondition: The website is up an running. After every step delete the text you typed from the search bar. Step 1 Type "Șopârlița" in the search bar. Expected results The location is listed in the drop-down menu as "Comuna Șopârlița, OLT". Step 2 Write in the search bar "Șopârlița". Expected results The location is listed in the drop-down menu as "Sat Șopârlița OLT (Șopârlița)".
gharchive/issue
2024-05-29T16:59:23
2025-04-01T04:35:29.496524
{ "authors": [ "Elena1303996" ], "repo": "peviitor-ro/ui.orase", "url": "https://github.com/peviitor-ro/ui.orase/issues/7178", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
786147173
Constructing model with Matrix with missing values it fails using Mill x = ArrayNode([1f0 2f0; missing missing]) reflectinmodel(x, d -> Chain(Dense(d,10, selu), Dense(10, 10))) crashes but using Mill x = ArrayNode([1f0 2f0; missing missing]) reflectinmodel(x, d ->Dense(d, 10)) works, which suggest the problem in make_imputing I have fixed it by changing _make_imputing(x, t::Chain) = Chain(t[1:end-1], _make_imputing(x, t[end])) to _make_imputing(x, t::Chain) = Chain(_make_imputing(x, t[1]), t[2:end]...) I have fixed it by changing _make_imputing(x, t::Chain) = Chain(t[1:end-1], _make_imputing(x, t[end])) to _make_imputing(x, t::Chain) = Chain(_make_imputing(x, t[1]), t[2:end]...) Some tests for this added in b05e0ef4757ba187861a7306042b67fa6f39faf6 Some tests for this added in b05e0ef4757ba187861a7306042b67fa6f39faf6
gharchive/issue
2021-01-14T16:48:24
2025-04-01T04:35:29.499381
{ "authors": [ "Sheemon7", "pevnak" ], "repo": "pevnak/Mill.jl", "url": "https://github.com/pevnak/Mill.jl/issues/56", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1350279632
How to apply 100% completed Chinese translation in iOS version? Hi dev, since the Chinese translation is 100% complete, how can I apply it to the game? Thanks! Simplified Chinese (100% complete! 🎉) Hi! Check your app store for updates, the latest update should contain Chinese in the languages settings. Hi! Check your app store for updates, the latest update should contain Chinese in the languages settings. Hi @SKPG-Tech, the screenshots below show my PPL version, and Language Settings, which doesn't seem to have a Chinese language option. It seems that no new version of PPL has been released recently (the App Store page doesn't show What's New, but PP does). PPL: https://apps.apple.com/us/app/pewpew-live/id1556977519 PP: https://apps.apple.com/us/app/pewpew/id314964252 Thanks. The iOS version of PPL has not yet been updated. I'm working on it, but it takes time. The iOS version of PPL has not yet been updated. I'm working on it, but it takes time. Oh, I see, thanks. Closing this issue for now, we'll comment again once the update is out.
gharchive/issue
2022-08-25T03:50:21
2025-04-01T04:35:29.504682
{ "authors": [ "SKPG-Tech", "haolloyin", "jyaif" ], "repo": "pewpewlive/ppl-i18n", "url": "https://github.com/pewpewlive/ppl-i18n/issues/271", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
993475167
Update scalafmt-core to 3.0.3 Updates org.scalameta:scalafmt-core from 3.0.0 to 3.0.3. GitHub Release Notes - Version Diff I'll automatically update this PR to resolve conflicts as long as you don't change it yourself. If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below. Configure Scala Steward for your repository with a .scala-steward.conf file. Have a fantastic day writing Scala! Ignore future updates Add this to your .scala-steward.conf file to ignore future updates of this dependency: updates.ignore = [ { groupId = "org.scalameta", artifactId = "scalafmt-core" } ] labels: library-update, semver-patch Superseded by #51.
gharchive/pull-request
2021-09-10T18:14:54
2025-04-01T04:35:29.508128
{ "authors": [ "scala-steward" ], "repo": "pfcoperez/thebutlerdidit", "url": "https://github.com/pfcoperez/thebutlerdidit/pull/49", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1460073064
Support Helm Chart depends on #8 fixes #6 This PR introduces helm chart support. chart itself in charts/gcp-workload-identity-federation-webhook dir chart is linted in GitHub Actions (ci.yaml) chart validation by chart-testing generated manifests by default values.yaml are validated by kubeconform release workflow is also paved by helm-gh-pages. I tested the workflow in my forked repo. index.yaml and chart tgz will be pushed in gh-pages branch. (sample in my fork) So, you can try:helm repo add gcp-workload-identity-federation-webhook https://everpeace.github.io/gcp-workload-identity-federation-webhook helm repo update helm show chart gcp-workload-identity-federation-webhook/gcp-workload-identity-federation-webhook TODO [x] prepare empty gh-pages branch [x] configure Github Pages on gh-pages branch sure, tomorrow I'll review it. Thanks a lot! 🙇 Just one tip working with helm charts, you can use Kind in the GH Runner to deploy your charts to a local cluster, that could give you more feedback about invalid APIs or sections in the manifests. For instance, in KEDA we use it to validate the chart with several k8s versions Cool! I will try to do this in another PR. I'm merging this PR for now and will release v0.1.1 to make helm chart available.
gharchive/pull-request
2022-11-22T15:19:37
2025-04-01T04:35:29.516681
{ "authors": [ "JorTurFer", "everpeace" ], "repo": "pfnet-research/gcp-workload-identity-federation-webhook", "url": "https://github.com/pfnet-research/gcp-workload-identity-federation-webhook/pull/9", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
97682410
Negative padding I fixed pooling functions to support negative value for padding. fix #138 I notice that when pad < 0, cudnn cannot pass gradient check. It makes a fatal conflict. I'll close this pr because I cannot merge the current master.
gharchive/pull-request
2015-07-28T11:32:32
2025-04-01T04:35:29.518231
{ "authors": [ "unnonouno" ], "repo": "pfnet/chainer", "url": "https://github.com/pfnet/chainer/pull/274", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
722084123
Add metrics to evaluator Add some hooks to calculate metrics easily in the evaluation, this avoids having to override evaluators with all the boilerplate code /test LGTM!
gharchive/pull-request
2020-10-15T07:40:34
2025-04-01T04:35:29.519347
{ "authors": [ "asi1024", "emcastillo" ], "repo": "pfnet/pytorch-pfn-extras", "url": "https://github.com/pfnet/pytorch-pfn-extras/pull/79", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
842283286
Language and province selection in under hamburger menu We would like to add the ability for users to switch provinces and language once they are in the app. The languages vary based on province. Zeplin screens here See the video of the flow: https://user-images.githubusercontent.com/21990736/112690528-2631a580-8e39-11eb-9521-2c761c1df967.mov Figma prototype here Notes: If a user is using AA a language that is not available in MB (eg. Chinese), once they click Manitoba in the province selection, a dialog shows up that tells them "[Language] not available - Select language". Clicking the Select Language button takes them to the language selection drawer page. Clicking Cancel takes them back to the province selection drawer page, but BC will be checked (not Manitoba). If they click Manitoba again, they will get the dialog again that they need to Select Language. The "[Language] not available - Select Language" dialog is localized to their most recently selected language (in the previous example - Chinese). Once they click "Select language" and hit the language selection drawer page, it will switch to English. If they clicked "Cancel" they go back to the province selection page, and that page is localized to their previous language (in this example - Chinese). Acceptance Criteria [ ] Steps for QA: Clicking "continue" takes user to Manitoba in English! Cancel keeps user in BC in original language as expected. https://user-images.githubusercontent.com/85197659/128097518-0168e932-64cd-4ed0-bc8e-b22539f48551.mp4
gharchive/issue
2021-03-26T20:43:00
2025-04-01T04:35:29.527041
{ "authors": [ "alysha-PG", "cherrielam" ], "repo": "pg-irc/pathways-frontend", "url": "https://github.com/pg-irc/pathways-frontend/issues/1451", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
2624441461
Qualcomm Atheros QCNFA765: ath11k_pci crashing Description Appliance seems to detect the hardware but firmware crashes on load. mhi mhi0: Direct firmware load for ath11k/WCN6855/hw2.1/amss.bin failed with error -2 Host operating system FreeBSD 13.3-RELEASE-p7 GENERIC amd64 Wireless NIC Qualcomm Technologies, Inc QCNFA765 Wireless Network Adapter Wifibox version 0.14.0 Disk image type and version wifibox-alpine 20240911 Changes to the default configuration files No response Logs Unable to post now (no network) Additional context Add any other context about the problem here that might help the investigation. Have you tried to turn it on and off? [X] Yes, I have read all the manual pages first! It seems /lib/firmware/ath11k/WCN6855/hw2.1/ directory is missing from the Alpine appliance image. Thanks for reporting the issue! Based on this, I believe I was able to identify the root cause. Some of the files that are shipped with linux-firmware are symbolic links and they are not stored in the git repository (from where the contents of the respective tarball is extracted) but need to be added through calling a package builder script. I created a fix for that in the fix/net/wifibox-alpine/linux-firmware-symlinks branch of the pgj/freebsd-wifibox-port repository. Please try it by reinstalling the net/wifibox-alpine port (version 20241101) from there: https://github.com/pgj/freebsd-wifibox-port/tree/fix/net/wifibox-alpine/linux-firmware-symlinks Thank you. Now it does not crash but it seems the driver is waiting for something that never happens: [ 0.603295] ath11k_pci 0000:00:06.0: MSI vectors: 1 [ 0.603356] ath11k_pci 0000:00:06.0: wcn6855 hw2.1 [ 0.605637] NET: Registered PF_QIPCRTR protocol family [ 0.761336] mhi mhi0: Requested to power ON [ 0.761345] mhi mhi0: Power on setup success [ 0.847684] mhi mhi0: Wait for device to enter SBL or Mission mode``` I have done some investigation, and it seems this a known bug of the ath11k driver. Essentially, the driver does not tolerate well when it is run in a virtualized environment because it assumes that the location of the MSI table matches with that of the host. There is a patch that may address this boot issue, we can take a chance with that if you are available for testing it. Thanks a lot for the research! Of course I would be glad to help with the testing. Can you apply the patch to the branch? Unfortunately, I have just noticed that this patch has been made part of both Linux 6.6.50 and Linux 6.10.9 that are integrated into wifibox-alpine 20240911. This means that the problem must be with something else in this case. Well that's sad. Please let me know if you think of anything else or if there is some more data I could give to help solve this. For the record this is the wifi chip for Lenovo T14s Gen4 AMD laptops.
gharchive/issue
2024-10-30T15:15:12
2025-04-01T04:35:29.557936
{ "authors": [ "Defenso-QTH", "pgj" ], "repo": "pgj/freebsd-wifibox", "url": "https://github.com/pgj/freebsd-wifibox/issues/137", "license": "BSD-2-Clause", "license_type": "permissive", "license_source": "github-api" }
2008764452
Update to Npsql 8 - Issue with TypeHandlerResolverFactory Describe the bug When using Microsoft.SemanticKernel.Connectors.Memory.Postgres 1.0.0-beta8 with Npsql 8.0 I'm not able to run the application. It fails with this exception: System.TypeLoadException: 'Could not load type 'Npgsql.Internal.TypeHandling.TypeHandlerResolverFactory' from assembly 'Npgsql, Version=8.0.0.0, Culture=neutral, PublicKeyToken=5d8b90d52f46fda7'.' According to this issue the TypeHandlerResolverFactory was removed but VectorExtensions.UseVector seems to require this. Any chance that you can look into that? Without that I'm not able to use the semantic-kernel with Npsql 8 / .NET 8. Hi @Cotspheer, just pushed version 0.2.0 of the Pgvector package, which supports Npgsql 8. Tahts awesome news! Thank you! Can you add a Pgvector.Dapper 2.0 nuget package which references Pgvector 2.0?
gharchive/issue
2023-11-23T20:34:52
2025-04-01T04:35:29.569516
{ "authors": [ "Cotspheer", "ankane", "redx177" ], "repo": "pgvector/pgvector-dotnet", "url": "https://github.com/pgvector/pgvector-dotnet/issues/26", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
601949722
default cut-offs for mob_typer Dear devs, thanks a lot for a very useful piece of software. I've been browsing the documentation and paper but it seems I can't find what is the default threshold values for mob_typer similarity cut_offs (min seq. ident.; min. coverage & min evalue) for each database elements (ie. rep; mob; mpf; orit; etc) Is this listed somewhere ? Best regards, Joseph. Hello Joseph, Thank you for the report, I will have the defaults of the parameters listed in the readme in a future update. I hadn't realized that the defaults weren't visible when you use --help. Here are the defaults from the arguments in the code. I am currently updating the MOB-suite with a pretty significant overhaul so I will add that to the documentation as part of that update ('--min_rep_evalue', type=str, required=False, help='Minimum evalue threshold for replicon blastn', default=0.00001) ('--min_mob_evalue', type=str, required=False, help='Minimum evalue threshold for relaxase tblastn', default=0.00001) ('--min_con_evalue', type=str, required=False, help='Minimum evalue threshold for contig blastn', default=0.00001) ('--min_ori_evalue', type=str, required=False, help='Minimum evalue threshold for oriT elements blastn', default=0.00001) ('--min_mpf_evalue', type=str, required=False, help='Minimum evalue threshold for mpf elements blastn', default=0.00001) ('--min_rep_ident', type=int, required=False, help='Minimum sequence identity for replicons', default=80) ('--min_mob_ident', type=int, required=False, help='Minimum sequence identity for relaxases', default=80) ('--min_ori_ident', type=int, required=False, help='Minimum sequence identity for oriT elements', default=90) ('--min_mpf_ident', type=int, required=False, help='Minimum sequence identity for mpf elements', default=80) ('--min_rep_cov', type=int, required=False, help='Minimum percentage coverage of replicon query by input assembly', default=80) ('--min_mob_cov', type=int, required=False, help='Minimum percentage coverage of relaxase query by input assembly', default=80) ('--min_ori_cov', type=int, required=False, help='Minimum percentage coverage of oriT', default=90) ('--min_mpf_cov', type=int, required=False, help='Minimum percentage coverage of mpf', default=80) ('--min_overlap', type=int, required=False, help='Minimum overlap of fragments', default=10) Thanks a lot for the swift response ! Le ven. 17 avr. 2020 à 18:36, jrober84 notifications@github.com a écrit : Hello Joseph, Thank you for the report, I will have the defaults of the parameters listed in the readme in a future update. I hadn't realized that the defaults weren't visible when you use --help. Here are the defaults from the arguments in the code. I am currently updating the MOB-suite with a pretty significant overhaul so I will add that to the documentation as part of that update ('--min_rep_evalue', type=str, required=False, help='Minimum evalue threshold for replicon blastn', default=0.00001) ('--min_mob_evalue', type=str, required=False, help='Minimum evalue threshold for relaxase tblastn', default=0.00001) ('--min_con_evalue', type=str, required=False, help='Minimum evalue threshold for contig blastn', default=0.00001) ('--min_ori_evalue', type=str, required=False, help='Minimum evalue threshold for oriT elements blastn', default=0.00001) ('--min_mpf_evalue', type=str, required=False, help='Minimum evalue threshold for mpf elements blastn', default=0.00001) ('--min_rep_ident', type=int, required=False, help='Minimum sequence identity for replicons', default=80) ('--min_mob_ident', type=int, required=False, help='Minimum sequence identity for relaxases', default=80) ('--min_ori_ident', type=int, required=False, help='Minimum sequence identity for oriT elements', default=90) ('--min_mpf_ident', type=int, required=False, help='Minimum sequence identity for mpf elements', default=80) ('--min_rep_cov', type=int, required=False, help='Minimum percentage coverage of replicon query by input assembly', default=80) ('--min_mob_cov', type=int, required=False, help='Minimum percentage coverage of relaxase query by input assembly', default=80) ('--min_ori_cov', type=int, required=False, help='Minimum percentage coverage of oriT', default=90) ('--min_mpf_cov', type=int, required=False, help='Minimum percentage coverage of mpf', default=80) ('--min_overlap', type=int, required=False, help='Minimum overlap of fragments', default=10) — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/phac-nml/mob-suite/issues/48#issuecomment-615345733, or unsubscribe https://github.com/notifications/unsubscribe-auth/ABEFZXQRRARHBIRBHM3VINTRNCAPVANCNFSM4MKW4DSA . -- Joseph Nesme Roskildevej 191, st th 2500 Valby Denmark Tel. (fr): +33 6 30 72 33 53 Tel. (dk): +45 50 65 41 50
gharchive/issue
2020-04-17T12:44:36
2025-04-01T04:35:29.603869
{ "authors": [ "jnesme", "jrober84" ], "repo": "phac-nml/mob-suite", "url": "https://github.com/phac-nml/mob-suite/issues/48", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
514995170
Release tarball doesn't include submodules Apparently this is a problem across many projects. Doesn't apply to cloned repo. Include source package (tar.gz to directly install in R) in the releases
gharchive/issue
2019-10-30T21:31:33
2025-04-01T04:35:29.605005
{ "authors": [ "TheZetner" ], "repo": "phac-nml/wade", "url": "https://github.com/phac-nml/wade/issues/36", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
124772514
How to update an existing install ? Hi, I am happy with my fresh install of Koel but I am wondering : how to update it ? After git pull, do I need to run npm install and/or composer install ? The instruction is right in the README. Thank you, I have read carefully the wiki,not the README...
gharchive/issue
2016-01-04T14:52:15
2025-04-01T04:35:29.627163
{ "authors": [ "Roudaille77", "phanan" ], "repo": "phanan/koel", "url": "https://github.com/phanan/koel/issues/179", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
913388278
create a matrix to display admirals AdAM IG coverage (1) goal: come up with a way to measure how much of the ADaM IG standards are covered by admiral show overall coverage show Roche coverage show GSK coverage question for testing: how much is covered by the testing companies? I see this was moved to Sprint Done. Can we close this issue?
gharchive/issue
2021-06-07T10:23:16
2025-04-01T04:35:29.629649
{ "authors": [ "bms63", "koegerr" ], "repo": "pharmaverse/admiral", "url": "https://github.com/pharmaverse/admiral/issues/205", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
323615191
Missing browse scoped option In Nautilus, a really really cool option is "browse scoped" allowing to browse only a subset of the system (a package, a class...). This allows browsing senders/implementors/references... only in a part of the system, which is useful when we look for common methods. This should probably be implemented in Calypso before Pharo 7 release. It is already here for a long time. But approach is different. Scopes are just available in any method browser. No need to press any button to have them. Notice "Current image" item in method browser. It is combo box to switch these scopes. Try to expand list. Scopes are inherited from parent browser which spawn them. And it also includes items from selected methods. Same widget is available for refactoring previews. So after discussion with Stef and Guille I realized what scenario is actually missing. And now it is available in 0.11.4 (already integrated): I added menu item «Set up scope» where you select multiple packages. This scope is propogated to other places. Indication is done with special label «Scoped pckg» for «Packages» radio button. In addition this radio have extra logic when you click on it in checked state. In scoped mode it resets scope. In normal mode it sets up the scope of selected packages.
gharchive/issue
2018-05-16T13:13:52
2025-04-01T04:35:29.645509
{ "authors": [ "dionisiydk", "jecisc" ], "repo": "pharo-ide/Calypso", "url": "https://github.com/pharo-ide/Calypso/issues/271", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
140034318
YamlReader is not working as expected Is inheritance always followed? The id field does not appear to be populated This is partly related to https://github.com/phenopackets/phenopacket-format/issues/32 The underlying issue is that Jackson doesn't know what it's supposed to be creating as YAML requires a type e.g. !<person> in order to know what it needs to deserialise as. entities: - !<person> strain: null types: null negated_types: null description: null id: "person#1" label: "Joe Bloggs" taxon: null sex: null date_of_birth: null - !<variant> descriptionHGVS: "c.1234A>G" id: "variant#1" label: "c.1234A>G" phenotype_profile: - phenotype: !<phenotype> measurements: null types: - id: "HP:0200055" label: "Small hands" negated_types: null description: null has_location: null onset: null offset: null severity: null environment: null entity: !<person> strain: null types: null negated_types: null description: null id: "person#1" label: "Joe Bloggs" taxon: null sex: null date_of_birth: null evidence: - types: - id: "ECO:0000033" label: "TAS" negated_types: null description: null supporting_entities: null source: - "PMID:23455423" JSON uses a "type" key to acheive the same thing. You want to be able to do use YAML tags of sorts, but only ever refer to the id of an entity once defined: entities: - id: "person#1" type: person strain: null types: null negated_types: null description: null label: "Joe Bloggs" taxon: null sex: null date_of_birth: null - id: "variant#1" descriptionHGVS: "c.1234A>G" type: variant label: "c.1234A>G" phenotype_profile: - entity: "person#1" phenotype: measurements: null types: - id: "HP:0200055" label: "Small hands" negated_types: null description: null has_location: null onset: null offset: null severity: null environment: null evidence: - types: - id: "ECO:0000033" label: "TAS" source: - "PMID:23455423" Plus I guess you don't want the keys with null values written out either. This is now read as expected: --- id: "phenopkt#1" title: "Patient X phenotypes and potentially causative variant" persons: - id: "person#1" label: "Joe Bloggs" sex: "M" variants: - id: "variant#1" descriptionHGVS: "c.1234A>G" label: "c.1234A>G" phenotype_profile: - entity: "person#1" phenotype: types: - id: "HP:0200055" label: "Small hands" evidence: - types: - id: "ECO:0000033" label: "TAS" source: - id: "PMID:23455423"
gharchive/issue
2016-03-10T23:01:34
2025-04-01T04:35:29.742171
{ "authors": [ "cmungall", "julesjacobsen" ], "repo": "phenopackets/phenopacket-reference-implementation", "url": "https://github.com/phenopackets/phenopacket-reference-implementation/issues/16", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
1117621679
Add options to change order of custom commands This PR adds one improvement and implements a feature request #4. Command buttons are now removed from ribbon immediately, no restart necessary I've added up/down arrow buttons in command settings to reorder commands (closes #4) That looks really good! Thank you
gharchive/pull-request
2022-01-28T16:46:53
2025-04-01T04:35:29.819703
{ "authors": [ "miluoshi", "phibr0" ], "repo": "phibr0/obsidian-customizable-sidebar", "url": "https://github.com/phibr0/obsidian-customizable-sidebar/pull/20", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
472584873
Fatal Exception with A_HelloWorld As a recap, I couldn't get any of the examples to compile. I was using a git version of ESP8266/Arduino. I didn't want to change that version back to 2.5.2 so I went to my ESP32 machine, installed ESP8266/Arduino 2.5.2 release code. That solved the compile issue I raised so I closed it. While I had compiled it I didn't have an ESP8266 board to try. I didn't want to mix ESP32 Arduino stuff with ESP8266 Arduino stuff. That was the reason I got the second computer. I still didn't want to change my git version on my ESP8266 machine. I have wanted to try the current Esparto and today I decided to proceed on my ESP32 machine. All seemed good until I loaded the flash with A_HelloWorld. I also loaded the Data directory into SPIFFS. I then did a reset. And I kept getting immediate fatal exceptions. ets Jan 8 2013,rst cause:1, boot mode:(3,6) load 0x4010f000, len 1384, room 16 tail 8 chksum 0x2d csum 0x2d v8b899c12 ~ld SDK:2.2.1(cfd48f3)/Core:2.5.2=20502000/lwIP:STABLE-2_1_2_RELEASE/glue:1.1-7-g82abda3/BearSSL:a143020 SPIFFSImpl: allocating 512+240+1400=2152 bytes SPIFFSImpl: mounting fs @300000, size=fb000, block=2000, page=100 SPIFFSImpl: mount rc=0 Fatal exception 28(LoadProhibitedCause): epc1=0x40234ed0, epc2=0x00000000, epc3=0x00000000, excvaddr=0x00000000, depc=0x00000000 Exception (28): epc1=0x40234ed0 epc2=0x00000000 epc3=0x00000000 excvaddr=0x00000000 depc=0x00000000 I tried a bunch of things and got nowhere. In the end I took an SSD drive and installed Arduino 1.8.9 on it. (a portable - self contained copy). I installed the 2.5.2 release code of the ESP8266/Arduino framework. I installed the current SPIFFS upload tool and the exception decoder. I installed the Esparto and listed required libraries. I am using a 4MB ESP-12E module with the following settings. Decoding stack results 0x4020277c: std::operator , std::allocator >(std::basic_string , std::allocator > const&, std::basic_string , std::allocator > const&) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2\bits/basic_string.h line 2573 0x40203250: std::map , std::allocator > >::operator[](std::string const&) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2\bits/stl_tree.h line 1141 0x40226600: malloc(size_t) at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\umm_malloc\umm_malloc.cpp line 1685 0x402036d9: ESPArto::ESPArto(std::map , std::allocator > >) at E:\sketchbook\libraries\esparto-master\src\ESPArto.cpp line 464 0x402069bc: ESPArto::_mqTat(std::vector >) at E:\sketchbook\libraries\esparto-master\src\mqtt.cpp line 230 0x4022b6b4: std::_Function_base::_Base_manager >)>::_M_manager(std::_Any_data&, std::_Any_data const&, std::_Manager_operation) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2/functional line 1934 0x40202534: std::_Function_handler >), void (*)(std::vector >)>::_M_invoke(std::_Any_data const&, std::vector >) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2/functional line 2069 0x402069e0: ESPArto::_mqTdaily(std::vector >) at E:\sketchbook\libraries\esparto-master\src\mqtt.cpp line 232 0x4022b6b4: std::_Function_base::_Base_manager >)>::_M_manager(std::_Any_data&, std::_Any_data const&, std::_Manager_operation) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2/functional line 1934 0x40202534: std::_Function_handler >), void (*)(std::vector >)>::_M_invoke(std::_Any_data const&, std::vector >) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2/functional line 2069 0x40100504: millis() at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\core_esp8266_wiring.cpp line 186 0x4022b6b4: std::_Function_base::_Base_manager >)>::_M_manager(std::_Any_data&, std::_Any_data const&, std::_Manager_operation) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2/functional line 1934 0x40202534: std::_Function_handler >), void (*)(std::vector >)>::_M_invoke(std::_Any_data const&, std::vector >) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2/functional line 2069 0x4020115f: _GLOBAL__sub_I_Esparto() at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2\bits/stl_tree.h line 671 0x40222260: String::String(char const*) at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\WString.cpp line 36 0x402054ac: _GLOBAL__sub_I__ZN7ESPArto4_udpE() at E:\sketchbook\libraries\esparto-master\src\ESPArto.cpp line 497 0x40223435: init_done() at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\core_esp8266_main.cpp line 149 0x4023975a: tcp_input at core/tcp_in.c line 321 0x4023ad00: udp_sendto_if_src at core/udp.c line 855 0x4022e952: _printf_float at /home/earle/src/esp-quick-toolchain/repo/newlib/newlib/libc/stdio/nano-vfprintf_float.c line 241 0x402018fb: std::__introsort_loop__gnu_cxx::__normal_iterator >, int, ESPArto::_matchTasks(std::function , std::function )::__lambda22>(__gnu_cxx::__normal_iterator > >, __gnu_cxx::__normal_iterator > >, int) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2\bits/stl_algo.h line 2328 0x40203412: std::map , std::allocator > >::operator[](std::string&&) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2\bits/stl_tree.h line 140 0x40238642: tcp_parseopt at core/tcp_in.c line 1946 0x4023811f: dns_gethostbyname_addrtype at core/dns.c line 1566 0x4010044c: __wrap_spi_flash_read(uint32_t, uint32_t*, size_t) at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\core_esp8266_phy.cpp line 308 0x4023813a: dns_gethostbyname_addrtype at core/dns.c line 1580 0x40238642: tcp_parseopt at core/tcp_in.c line 1946 0x40100854: uart_isr(void*) at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\uart.cpp line 364 0x4023811f: dns_gethostbyname_addrtype at core/dns.c line 1566 0x4010044c: __wrap_spi_flash_read(uint32_t, uint32_t*, size_t) at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\core_esp8266_phy.cpp line 308 0x4023813a: dns_gethostbyname_addrtype at core/dns.c line 1580 0x401004e4: millis() at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\core_esp8266_wiring.cpp line 183 0x401004e4: millis() at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\core_esp8266_wiring.cpp line 183 0x402394c6: tcp_input at core/tcp_in.c line 189 0x4023975a: tcp_input at core/tcp_in.c line 321 0x401008c8: uart_isr(void*) at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\uart.cpp line 388 0x4023923f: tcp_receive at core/tcp_in.c line 1814 0x401004c0: millis() at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\core_esp8266_wiring.cpp line 180 0x4023925a: tcp_receive at core/tcp_in.c line 1814 I used two different ESP-12 module boards that I have been using for a couple of years. I have never had any issues with them. Can you think of what the issue might be? One other note. I did a complete flash contents write to remove any previous information before I programmed the code and SPIFFS. Just pushed a quick bugfix (3.3.2) to solve a problem some folk were having with the early examples e.g. "Hello World" If that's you, update your copy and try again - many apologies of course On Thu, Jul 25, 2019 at 1:08 AM Rudy notifications@github.com wrote: One other note. I did a complete flash contents write to remove any previous information before I programmed the code and SPIFFS. — You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHub https://github.com/philbowles/esparto/issues/22?email_source=notifications&email_token=AGT3V7S7ENCRRL7N776DIPDQBDOHHA5CNFSM4IGVL7SKYY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGOD2X3LVI#issuecomment-514831829, or mute the thread https://github.com/notifications/unsubscribe-auth/AGT3V7UYGPAB5BPJRSMCFRLQBDOHHANCNFSM4IGVL7SA . I tried the new code and it didn't solve the problem. It still crashes immediately on startup. How do I go about troubleshooting this? Decoding stack results 0x4020277c: std::operator , std::allocator >(std::basic_string , std::allocator > const&, std::basic_string , std::allocator > const&) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2\bits/basic_string.h line 2573 0x40203250: std::map , std::allocator > >::operator[](std::string const&) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2\bits/stl_tree.h line 1141 0x4022695c: malloc(size_t) at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\umm_malloc\umm_malloc.cpp line 1685 0x402036d9: ESPArto::ESPArto(std::map , std::allocator > >) at E:\sketchbook\libraries\esparto-master\src\ESPArto.cpp line 464 0x402069e4: ESPArto::_mqTat(std::vector >) at E:\sketchbook\libraries\esparto-master\src\mqtt.cpp line 230 0x4022ba10: std::_Function_base::_Base_manager >)>::_M_manager(std::_Any_data&, std::_Any_data const&, std::_Manager_operation) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2/functional line 1934 0x40202534: std::_Function_handler >), void (*)(std::vector >)>::_M_invoke(std::_Any_data const&, std::vector >) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2/functional line 2069 0x40206a08: ESPArto::_mqTdaily(std::vector >) at E:\sketchbook\libraries\esparto-master\src\mqtt.cpp line 232 0x4022ba10: std::_Function_base::_Base_manager >)>::_M_manager(std::_Any_data&, std::_Any_data const&, std::_Manager_operation) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2/functional line 1934 0x40202534: std::_Function_handler >), void (*)(std::vector >)>::_M_invoke(std::_Any_data const&, std::vector >) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2/functional line 2069 0x40100504: millis() at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\core_esp8266_wiring.cpp line 186 0x4022ba10: std::_Function_base::_Base_manager >)>::_M_manager(std::_Any_data&, std::_Any_data const&, std::_Manager_operation) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2/functional line 1934 0x40202534: std::_Function_handler >), void (*)(std::vector >)>::_M_invoke(std::_Any_data const&, std::vector >) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2/functional line 2069 0x4020115f: _GLOBAL__sub_I_Esparto() at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2\bits/stl_tree.h line 671 0x402225bc: String::String(char const*) at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\WString.cpp line 36 0x402054ac: _GLOBAL__sub_I__ZN7ESPArto4_udpE() at E:\sketchbook\libraries\esparto-master\src\ESPArto.cpp line 497 0x40223791: init_done() at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\core_esp8266_main.cpp line 149 0x4023ad00: udp_input at core/udp.c line 195 0x4022e952: isspace at /home/earle/src/esp-quick-toolchain/repo/newlib/newlib/libc/ctype/isspace.c line 43 0x402018fb: std::__introsort_loop__gnu_cxx::__normal_iterator >, int, ESPArto::_matchTasks(std::function , std::function )::__lambda22>(__gnu_cxx::__normal_iterator > >, __gnu_cxx::__normal_iterator > >, int) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2\bits/stl_algo.h line 2328 0x40203412: std::map , std::allocator > >::operator[](std::string&&) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2\bits/stl_tree.h line 140 0x4023811f: dns_recv at core/dns.c line 1206 0x4010044c: __wrap_spi_flash_read(uint32_t, uint32_t*, size_t) at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\core_esp8266_phy.cpp line 308 0x4023813a: dns_recv at core/dns.c line 1221 0x40100854: uart_isr(void*) at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\uart.cpp line 364 0x4023811f: dns_recv at core/dns.c line 1206 0x4010044c: __wrap_spi_flash_read(uint32_t, uint32_t*, size_t) at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\core_esp8266_phy.cpp line 308 0x4023813a: dns_recv at core/dns.c line 1221 0x401004e4: millis() at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\core_esp8266_wiring.cpp line 183 0x401004e4: millis() at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\core_esp8266_wiring.cpp line 183 0x402394c6: tcp_receive at core/tcp_in.c line 1781 0x401008c8: uart_isr(void*) at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\uart.cpp line 388 0x4023923f: tcp_receive at core/tcp_in.c line 2117 0x401004c0: millis() at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\core_esp8266_wiring.cpp line 180 0x4023925a: tcp_receive at core/tcp_in.c line 1618 Hmm. Can you check that you have indeed updated to v3.3.2 it absolutely should have fixed this. ALso take any of the later examples that use WiFi and try one of them (the problem was with empty/ non-existent config block - WiFI examples - by definition - must have config data) then on the serial monitor type cmd/info and show me the output?
gharchive/issue
2019-07-24T22:48:08
2025-04-01T04:35:29.831405
{ "authors": [ "RudyFiero", "philbowles" ], "repo": "philbowles/esparto", "url": "https://github.com/philbowles/esparto/issues/22", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1524970898
请求帮助,使用task_create无法建任务 tasks.data 返回了这些 {'title': 'xxxx-建任务', 'customFields': {'story': '相关需求', 'estStarted': '预计开始', 'deadline': '截止日期', 'mailto': '抄送给', 'pri': '优先级', 'estimate': '最初预计'}..... 官方文档上的这些参数改如何填写呢? $params = array( 'project' => 1, 'type' => 'ui', 'module' => 0, 'assignedTo[]' => 'lisi', 'testAssignedTo[]' => 'lisi', 'color' => '', 'name' => '测试添加任务2', 'pri' => 2, 'estimate' => 1, 'desc' => '测试添加任务描述测试添加任务描述', 'estStarted' => '2019-11-11', 'deadline' => '2019-11-12', 'mailto[1]' => 'lisi' ); // 请求参数 更新到0.3.0版本,post参数可使用data传入,例如 response = zentao.task_create( executionID=2, storyID=0, moduleID=0, ... data={ "execution": 2, "type": "design", "name": "锦囊喵叽", "assignedTo[]": "老六", "pri": 3, "desc": "暴打小柯基" ... }, ) 成啦,感谢!
gharchive/issue
2023-01-09T05:48:40
2025-04-01T04:35:29.847339
{ "authors": [ "GbingSong", "philip1134" ], "repo": "philip1134/pyzentao", "url": "https://github.com/philip1134/pyzentao/issues/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
642516225
请教关于生成订阅的问题。 我有4台VPS,想生成一个订阅便于管理,不需要流量控制这些功能,请问有什么好软件推荐吗? 可以用我写的另一个工具 Shawdowrockets 訂閱鏈接編輯器 如果不想用我的API,把这个工具生成的 BASE64 放到一个纯文本档案里再放到你的服务器或 Github Gist 上也可以达成订阅的效果。 好的,谢谢。
gharchive/issue
2020-06-21T07:56:37
2025-04-01T04:35:29.885339
{ "authors": [ "LeiD215", "phlinhng" ], "repo": "phlinhng/v2ray-tcp-tls-web", "url": "https://github.com/phlinhng/v2ray-tcp-tls-web/issues/18", "license": "WTFPL", "license_type": "permissive", "license_source": "github-api" }
1989700784
Merge Branch Merging Heinrich's and cleaning again bibs-Merge1 ended up merging all Omar/Princess/Heinrich/Sam
gharchive/issue
2023-11-13T01:07:13
2025-04-01T04:35:29.891067
{ "authors": [ "bibiygarcia" ], "repo": "phobs-co/phobs-co.github.io", "url": "https://github.com/phobs-co/phobs-co.github.io/issues/2", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
36567594
WP8 push notification I have easily applied push notification service for my android app, but confused how to do in windows phone. if(device.platform == "Win32NT"){ pushNotification.register( channelHandler, errorHandler, { "channelName": channelName, "ecb": "onNotificationWP8", "uccb": "channelHandler", "errcb": "jsonErrorHandler" }); } I doesn't understand what the above parameters. eg: channelName. where can i get the values for these parameters?? Also, does anyone give me the link to full tutorial applying PNS in wp8 including server side too? Regards, Lomas Joshi Hello postixo, My uri is always empty. I don't get fix the problem with a your example using angularjs.Your application is still working? Are you could send the html code complete? fabiofapeli, you have to ask for the uri several times. yes, so strange, but Microsoft is Microsoft :) @joshilomas your code is clean , can you write a blog about it so that everyone can see when they google it instead of coming here :) Over a year later and Cordova & WP8 still has garbage information on doing this (or really anything) for the Windows Phone. I am now in the same boat as you all. Anyone please help me out for Integrating Push Notification in Windows Phone Cordova. I tried hard but unsucessfull. For me its working fine. just keep loop RegisterMPNS Function until unless you are not getting uri. function registerMPNS(){ var pushNotification; pushNotification = window.plugins.pushNotification; pushNotification.register( channelHandler, errorHandler, { "channelName": "xyz", //channelName is whateveryou like "ecb": "onNotificationWP8", "uccb": "channelHandler", "errcb": "jsonErrorHandler" }); } $window.onNotificationWP8 = function(e){ } function channelHandler(event){ var uri = event.uri; if(uri==''){ registerMPNS(); } }
gharchive/issue
2014-06-26T11:34:16
2025-04-01T04:35:29.907386
{ "authors": [ "fabiofapeli", "gustter", "joshilomas", "kundan1988", "mohammadmanzoor8972", "revaret", "tetreault" ], "repo": "phonegap-build/PushPlugin", "url": "https://github.com/phonegap-build/PushPlugin/issues/255", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }