id
stringlengths 4
10
| text
stringlengths 4
2.14M
| source
stringclasses 2
values | created
timestamp[s]date 2001-05-16 21:05:09
2025-01-01 03:38:30
| added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| metadata
dict |
---|---|---|---|---|---|
973440272
|
Update LICENSE 2021
fixes #
Checklist:
[ ] Add tests that demonstrate the correct behavior of the change. Tests should fail without the change.
[ ] Add or update relevant docs, in the docs folder and in code.
[ ] Add an entry in CHANGES.rst summarizing the change and linking to the issue.
[ ] Add .. versionchanged:: entries in any relevant code docs.
[ ] Run pre-commit hooks and fix any issues.
[ ] Run pytest and tox, no tests failed.
Thanks for trying to contribute, but I think we don't currently change the copyright year that appears in these license files.
|
gharchive/pull-request
| 2021-08-18T09:02:20 |
2025-04-01T04:35:27.849875
|
{
"authors": [
"jab",
"masguit42"
],
"repo": "pallets/jinja",
"url": "https://github.com/pallets/jinja/pull/1483",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
}
|
172428764
|
疑似内存泄露的现象
测试用例:
SpritePrefab引用了一张贴图A,在Lua端执行如下代码
function main()
local prefab = Resources.Load('SpritePrefab',GameObject)
Resources.UnloadUnusedAssets()
end
利用Unity的Profiler内存分析 可以看到,即使在调用UnloadUnusedAssets后, 贴图A仍有静态引用。
但是在c#端运行同样的代码,则贴图A会得到正常释放。。
知道问题了,
首先要调用lua的 collectgarbage() 释放lua的引用, 再调用 Resources.UnloadUnusedAssets()才可生效.
|
gharchive/issue
| 2016-08-22T11:23:15 |
2025-04-01T04:35:27.943401
|
{
"authors": [
"wlgys8"
],
"repo": "pangweiwei/slua",
"url": "https://github.com/pangweiwei/slua/issues/154",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2406486694
|
Correct the target and actor in Slack Audit log UserPrivilegeEscalation plus clean up
Background
Updating the Slack Audit log UserPrivilegeEscalation to correctly user target instead of actor. We got alerts for our IT team promoted to Admin or Owner. Realized it was an error in the alert and corrected it.
Changes
Updated the Slack UserPrivilegeEscalation alert to correctly differentiate between the actor and target. Actor being the user making the change and target being the user, the chances are occurring on.
Used the already definded dict USER_PRIV_ESC_ACTIONS for alert Titles
Updated alert Titles to correctly reference the target instead of the actor.
Add local variable action to improve readability.
Updated tests to test the new behavior
Updated tests to be YAML
Formatted YAML file with Prettier
Testing
pipenv run panther_analysis_tool test --skip-disabled-tests --sort-test-results
@bmbeverst thanks for the PR! Since you've provided logic for identifying the actor and the entity, we figured we could update the titles to be even more descriptive of the event. However, I'm concerned that some of our unit test events are missing information - in particular, I'm curious if a permission_assigned event indicates what permissions were granted. Do you have access to one of the events, and could you share the structure so we can add update the unit test and title?
@ben-githubs That would be great! I am back from vacation.
Unluckily, I do not have any permission_assigned events, search for the last 3 months. And I don't have a test slack workspace to create a test event with. There appears to be a few sets of permissions as well, roles and account types.
Understood Brooks! I'll make a note to followup on those tests at a later date, but otherwise we'll merge this to avoid holding the logic changes up!
|
gharchive/pull-request
| 2024-07-12T23:32:37 |
2025-04-01T04:35:27.987131
|
{
"authors": [
"ben-githubs",
"bmbeverst"
],
"repo": "panther-labs/panther-analysis",
"url": "https://github.com/panther-labs/panther-analysis/pull/1288",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
228091357
|
Question: League Standings
I hope this is the proper place to ask a question. If not please forgive me. I have been playing around with mlbgame and am looking for a way to generate the current standings by league and division. Is this at all possible?
Example:
American League East
1 Yankees 21-10
2 Orioles 21-11
3 Red Sox 17-16
4 Rays 17-19
5 Blue Jays 13-21
......
It is possible to get the information as a json payload, but not currently a part of the mlbgame code base. It could be added as a new feature, I'm sure. If I get some spare time, I'll look making a pull request if @panzarino doesn't get to it first.
@digitalSaint I would definitely accept a PR with this functionality.
Should this issue be closed?
@bradleyhurley I'm going to close the issue once standings are implemented into an official release (which should be coming soon).
Closing even though the release has not happened yet (it is coming soon).
|
gharchive/issue
| 2017-05-11T19:08:52 |
2025-04-01T04:35:28.029471
|
{
"authors": [
"bradleyhurley",
"chrissmithto",
"digitalSaint",
"panzarino"
],
"repo": "panzarino/mlbgame",
"url": "https://github.com/panzarino/mlbgame/issues/31",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
860652860
|
Fix KeyError in calculate_splits.py
key is "CID" not "cid"
Thanks for the catch!
Thanks for the catch!
You're welcome! But really it was Python that caught it and gave the error message, and since it's correct elsewhere in that line it was a pretty easy fix.
Remind me to ramble about the bugs I fixed in NetHack sometime
|
gharchive/pull-request
| 2021-04-18T12:29:23 |
2025-04-01T04:35:28.031050
|
{
"authors": [
"paoloruiz",
"psymar"
],
"repo": "paoloruiz/ootp-pt-analysis",
"url": "https://github.com/paoloruiz/ootp-pt-analysis/pull/4",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
842025875
|
Enhancement: Filter by name
I have about 100 queues/topics in a single ServiceBus and it takes a while to navigate.
A text box above the queues/topics tree for filtering items would be great.
If I were to submit a PR would this be something you'd consider including?
PRs are almost always most welcome, but have you seen the existing filter capability?
No, I was unable to find the option.
Thanks!
I see, it filters on the server.
That's a shame, I wanted more than StartsWith because all of my queues/topics end with different suffixes that I want to filter by :(
There was an issue about the same functionality that you are wishing for, #452.
@mrpmorris, I think it would be great if you created a PR for that. I am sure @SeanFeldman and @paolosalvatori agrees. I like your idea of having a textbox above where the filter would be visible.
Also, as far as I know, the Startswith functionality does not exist in the latest SDK for Service Bus so if you build something that does not rely on StartsWith it would make it easier to migrate it to the latest SDK in the future.
I don't think the service bus api supports any kind of "contains"
@mrpmorris when I created the Service Bus Explorer years ago I provided the ability to specify filters in the Connect form and context menus. I used and implemented the filters that were available at that time. It looks that nothing has changed since then at least in the old client SDK Microsoft.ServiceBus.dll. For example, the NamespaceManager.GetQueues(String) method
retrieves an enumerable collection of all queues in the service namespace with the specified filter. You can filter by a combination of entity name (including starts with), entity length (Gt or Lt), created, updated, and accessed time (Gt or Lt). Hence, no support for the contains predicate.
The ManagementClient.GetQueuesAsync method in the new Microsoft.Azure.ServiceBus library doesn't even support the possibility to pass a string which contains a filter.-
Yeah, I saw that :(
Seeing as the whole list is retrieved when connecting we don't actually save anything by filtering on the server. So maybe the client could do the filtering?
I think it would be great if the client did the filtering. Since it is a bit confusing with three SDKs for Service Bus here is a link to what I consider the preferred method for getting queues in the latest SDK.
It might be quite a task to replace the existing filtering with something based on that method, but it would be highly appreciated.
@ErikMogensen yes we can certainly implement filtering on the client-side, but there are customers that have namespaces with hundreds not say thousands of queues and topics and in that case, it would be useful to have the filtering capability on the server-side to avoid downloading a ton of entities from the cloud 😢
@paolosalvatori, I think most of the time is spent retrieving the runtime information for the queues, so client-side would be almost as effective. Also, if we don't accept client-side filtering we won't be able to migrate to the latest Service Bus SDK unless the Service Bus PG implements server-side filtering.
@paolosalvatori As soon as you connect, the client downloads all of the queues + topics anyway so that time is already spent.
Filtering on that list in memory would be quicker than asking the server to filter the list in memory and then return it over the network.
Or am I missing something obvious?
@mrpmorris well, the client does not download queues and topics data unless you explicitly call the GetQueues and GetTopics methods 😃 Anyway, if you guys have the bandwidth to implement client-side filtering, go with that, I'm fully supportive 😄
I had a crack at it, but the presentation and operational logic is mixed together and I found it too much work to understand & separate them.
Thanks for trying @mrpmorris!
How would you feel about application code being migrated to something like Fluxor so it can be completely separated from the code needed to present the data in the UI?
The top of the search in https://github.com/mrpmorris/Fluxor. Are you the author of it?
Could you describe the pros and cons of migrating to it. Consider that Service Bus Explorer contains about 50.000 lines of code (depends on how you count it) in Windows Forms and there are no dedicated developers. Everything is driven by voluntary effort, sometimes financed by organizations wanting a certain feature.
It's a clean way to separate the operational logic out of the code you need to control UI (disabling buttons, handling clicks etc).
Implementing the client-side filter would have been easy. You'd fire a FilterChangedAction(value of the input) every time the input changed - and then a single function would only include nodes that matched the filter.
It wouldn't be a small amount of work. But that's not because the Flux pattern is a lot of work, but because you have so much code and it's mixed together.
@mrpmorris I started the project as a sample 10 years ago when the Service Bus was actually launched. Then I migrated the project to GitHub in September 2014. Since then I've been the primary contributor, but now most of the changes come from the community. There's no direct involvement of the product group in the project, despite I'm a Microsoft employee, and I'm too busy to embark on a full rewrite of the tool. So even I agree with your idea, I doubt that we have the bandwidth for a migration.
|
gharchive/issue
| 2021-03-26T14:48:39 |
2025-04-01T04:35:28.042288
|
{
"authors": [
"ErikMogensen",
"mrpmorris",
"paolosalvatori"
],
"repo": "paolosalvatori/ServiceBusExplorer",
"url": "https://github.com/paolosalvatori/ServiceBusExplorer/issues/526",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
678812591
|
Add chat widget component support for Expo/React Native
A lot of people want to be able to use our chat widget in Expo/React Native -- this should be possible with a few tweaks to our web widget.
TODOs:
Probably need to create a separate repo from https://github.com/papercups-io/chat-widget and remove all browser-specific code (e.g. anything that uses localStorage, window, or document)
Remove theme UI dependency, try theme-ui-native or something else (see this thread https://github.com/system-ui/theme-ui/issues/170) for better mobile support
Honestly, it may not be necessary to be using these dependencies... need to investigate a bit but if we can just swap it out that might be easier in the short term
Figure out how to position widget in React Native (assuming position: fixed won't work)
Render the ChatWindow component (which typically gets rendered in an iframe) in a WebView
May need to add a mobile option to the ChatWindow component so that we can make design tweaks when rendering in mobile (probably need to enlarge some things?)
done in https://github.com/papercups-io/chat-widget-native
|
gharchive/issue
| 2020-08-14T00:13:39 |
2025-04-01T04:35:28.057292
|
{
"authors": [
"reichert621"
],
"repo": "papercups-io/papercups",
"url": "https://github.com/papercups-io/papercups/issues/139",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1794481870
|
Total peer additions metric
Describe the feature
Geth has the following metric:
p2p/dials/success - the total number of dials that result in a peer being added to the peerset
It would be great to add this to reth as well!
cc @maxwolff
Additional context
No response
will do
|
gharchive/issue
| 2023-07-07T22:28:12 |
2025-04-01T04:35:28.091734
|
{
"authors": [
"Rjected",
"maxwolff"
],
"repo": "paradigmxyz/reth",
"url": "https://github.com/paradigmxyz/reth/issues/3667",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1818918094
|
Use a trait for size methods in primitive types
Describe the feature
Many primitive types have methods which calculate a heuristic for the heap size of the type, for example:
https://github.com/paradigmxyz/reth/blob/b69a18dc47255304b3943ca859dc6eb9da579b19/crates/primitives/src/block.rs#L192-L200
This should be captured in a trait, and it should be possible to derive this trait for all of these types.
One crate that is interesting for this is DeepsizeOf. An unanswered question is whether or not we should use this crate, because its implementation of DeepSizeOf for Vec seems to contradict the rust docs.
The implementation:
impl<T> DeepSizeOf for std::vec::Vec<T> where T: DeepSizeOf {
fn deep_size_of_children(&self, context: &mut Context) -> usize {
// Size of heap allocations for each child
self.iter().map(|child| child.deep_size_of_children(context)).sum()
+ self.capacity() * std::mem::size_of::<T>() // Size of Vec's heap allocation
}
}
The Guarantees section of the rust docs state:
However, the pointer might not actually point to allocated memory. In particular, if you construct a Vec with capacity 0 via Vec::new, vec![], Vec::with_capacity(0), or by calling shrink_to_fit on an empty Vec, it will not allocate memory. Similarly, if you store zero-sized types inside a Vec, it will not allocate space for them. Note that in this case the Vec might not report a capacity of 0. Vec will allocate if and only if mem::size_of::<T>()*capacity() > 0. In general, Vec’s allocation details are very subtle — if you intend to allocate memory using a Vec and use it for something else (either to pass to unsafe code, or to build your own memory-backed collection), be sure to deallocate this memory by using from_raw_parts to recover the Vec and then dropping it.
If a Vec has allocated memory, then the memory it points to is on the heap (as defined by the allocator Rust is configured to use by default), and its pointer points to len initialized, contiguous elements in order (what you would see if you coerced it to a slice), followed by capacity-len logically uninitialized, contiguous elements.
This leads me to believe that the correct implementation should subtract len from capacity before multiplying by std::mem::size_of::<T>(). If this is true, then the crate may slightly overestimate the memory used by a Vec<T>.
Ultimately, this is not a huge problem, and using DeepSizeOf would be the easiest and most straightforward way to derive this size heuristic, and we could upstream a fix if required. Our implementation also does not subtract the len from capacity, so using the DeepSizeOf crate should also match our existing logic.
Additional context
No response
Can I work on this? I checked out the DeepsizeOf crate and it also notifies that it does not count the allocator's overhead, which I assume is something hard to estimate correctly. I guess since it also overestimates sometimes, it might output a balanced result.
|
gharchive/issue
| 2023-07-24T18:26:33 |
2025-04-01T04:35:28.101895
|
{
"authors": [
"Rjected",
"altugbakan"
],
"repo": "paradigmxyz/reth",
"url": "https://github.com/paradigmxyz/reth/issues/3893",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2395979809
|
chore: bump revm 11
new revm release with new call schemes
closing in favor of #156
|
gharchive/pull-request
| 2024-07-08T15:44:01 |
2025-04-01T04:35:28.102983
|
{
"authors": [
"mattsse"
],
"repo": "paradigmxyz/revm-inspectors",
"url": "https://github.com/paradigmxyz/revm-inspectors/pull/159",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
700942561
|
version and repo tag
Hello,
The package on pip is versioned 0.2.0 whereas the lastet tag / __version__ in pyls_isort/__about__.py in master is 0.1.1
Nice catch, I forgot to push...
Should be okay now https://github.com/paradoxxxzero/pyls-isort/releases/tag/0.2.0
perfect, thanks!
|
gharchive/issue
| 2020-09-14T09:42:29 |
2025-04-01T04:35:28.105018
|
{
"authors": [
"fesaille",
"paradoxxxzero"
],
"repo": "paradoxxxzero/pyls-isort",
"url": "https://github.com/paradoxxxzero/pyls-isort/issues/6",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
962300317
|
v0.2.3 has a broken Cargo.toml
I'm trying to follow https://substrate.dev/docs/en/tutorials/create-your-first-substrate-chain/setup
but my build fails with:
$ cargo build --release
error: failed to download `parity-db v0.2.3`
Caused by:
unable to get packages from source
Caused by:
failed to parse manifest at `/home/bear/.cargo/registry/src/github.com-1ecc6299db9ec823/parity-db-0.2.3/Cargo.toml`
Caused by:
failed to parse the version requirement `0.11 ` for dependency `parking_lot`
Caused by:
expected comma after minor version number, found '\t'
I downloaded parity-db==0.2.3 via cargo-download and found that Cargo.toml indeed has an extra \t on the parking_lot declaration:
# THIS FILE IS AUTOMATICALLY GENERATED BY CARGO
#
# When uploading crates to the registry Cargo will automatically
# "normalize" Cargo.toml files for maximal compatibility
# with all versions of Cargo and also rewrite `path` dependencies
# to registry (e.g., crates.io) dependencies
#
# If you believe there's an error in this file please file an
# issue against the rust-lang/cargo repository. If you're
# editing this file be aware that the upstream Cargo.toml
# will likely look very different (and much more reasonable)
[package]
edition = "2018"
name = "parity-db"
version = "0.2.3"
authors = ["Parity Technologies <admin@parity.io>"]
description = "Key-value database for the blockchain"
homepage = "https://substrate.dev"
license = "MIT OR Apache-2.0"
repository = "https://github.com/paritytech/parity-db/"
[profile.release]
lto = "fat"
codegen-units = 1
debug = true
panic = "abort"
[dependencies.blake2-rfc]
version = "0.2.18"
[dependencies.crc32fast]
version = "1.2.0"
[dependencies.fs2]
version = "0.4.3"
[dependencies.hex]
version = "0.4.2"
[dependencies.libc]
version = "0.2"
[dependencies.log]
version = "0.4.8"
[dependencies.memmap2]
version = "0.2"
[dependencies.parking_lot]
version = "0.11\t"
[dependencies.rand]
version = "0.8.2"
[dev-dependencies.env_logger]
version = "0.8.2"
0.2.2 has it too
```error: failed to parse manifest at /cargo-home/registry/src/github.com-1ecc6299db9ec823/parity-db-0.2.2/Cargo.toml
Caused by:
failed to parse the version requirement 0.11 for dependency parking_lot
Caused by:
expected comma after minor version number, found '\t'```
This was fixed in 0.2.4. Try using the v3.0.0+monthly-2021-07 tag when following the tutorial.
@arkpar Can you please do it for every package that has this one as dependency? I don't have it directly, it's somewhere in 3rd party module.
@Atomosk The only package that directly depends on this is sc-client-db. It is going to be published with other substrate crates, but I'm not sure when.
In any case, you can update your local dependency with cargo update -p parity-db even if it is indirect. Or use patch section in Cargo.toml
@Atomosk just FYI Substrate Recipes is broken with the same problem.
Reproduced it on both master and v3.0.0 tag.
cargo update -p parity-db solved the problem.
|
gharchive/issue
| 2021-08-06T00:48:36 |
2025-04-01T04:35:28.194845
|
{
"authors": [
"Atomosk",
"arkpar",
"bernardoaraujor"
],
"repo": "paritytech/parity-db",
"url": "https://github.com/paritytech/parity-db/issues/39",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
928172089
|
polkadot-launch cannot register parachain to rococo-local of polkadot-v0.9.5
it works fine in polkadot-v0.9.4, but the registration doesn't work anymore in polkadot-v0.9.5
close as the master + #117 worked for me
|
gharchive/issue
| 2021-06-23T11:50:54 |
2025-04-01T04:35:28.198765
|
{
"authors": [
"GopherJ"
],
"repo": "paritytech/polkadot-launch",
"url": "https://github.com/paritytech/polkadot-launch/issues/116",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1563373150
|
ensure_none naming is misleading
FRAME System's ensure_none ensures that the origin o represents an unsigned extrinsic.
This naming is consistent with the inner member of the RawOrigin enum:
pub enum RawOrigin<AccountId> {
Root,
Signed(AccountId),
None,
}
During a Milestone Review for the Substrate Builders Program I had to clarify a team that was misled by the ensure_none naming. The team had interpreted that ensure_none meant no one could call this extrinsic, which is literally the opposite of the meaning we're aiming to convey.
Even though the documentation already provides an accurate description, we could further avoid confusions by renaming ensure_none to ensure_unsigned.
Hey @bernardoaraujor , I would like to take this up.
|
gharchive/issue
| 2023-01-30T23:32:22 |
2025-04-01T04:35:28.274532
|
{
"authors": [
"bernardoaraujor",
"gitofdeepanshu"
],
"repo": "paritytech/substrate",
"url": "https://github.com/paritytech/substrate/issues/13279",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
836775466
|
frame contracts: expected associated type, found enum pallet::Event
i just tried building from latest 'master' branch and got the following error:
error[E0308]: mismatched types
--> /Users/me/code/paritytech/substrate/frame/contracts/src/exec.rs:914:30
|
914 | <T as Config>::Event::from(event).into(),
| ^^^^^ expected associated type, found enum `pallet::Event`
|
= note: expected associated type `<T as frame_system::Config>::Event`
found enum `pallet::Event<T>`
= help: consider constraining the associated type `<T as frame_system::Config>::Event` to `pallet::Event<T>` or calling a method that returns `<T as frame_system::Config>::Event`
= note: for more information, visit https://doc.rust-lang.org/book/ch19-03-advanced-traits.html
Do you use latest stable?
Do you use latest stable?
i actually tried using the latest 'master', maybe i should go back to using the latest stable :-)
i just tried building the latest stable release https://github.com/paritytech/substrate/releases/tag/v3.0.0 at commit 49a4103f4bfef55be20a5c6d26e18ff3003c3353
git fetch upstream v3.0.0:v3.0.0
git checkout v3.0.0
cargo build --release
but it gives the following error. maybe it's because i'm using an old version of rustc nightly-2020-10-06-x86_64-apple-darwin
Compiling pallet-offences-benchmarking v3.0.0 (/Users/ls2/code/paritytech/substrate/frame/offences/benchmarking)
error[E0308]: mismatched types
--> frame/offences/benchmarking/src/lib.rs:288:6
|
288 | pallet_balances::Event::<T>::Endowed(reporter, (reward_amount / r).into())
| ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected associated type, found enum `pallet_balances::Event`
|
= note: expected associated type `<T as frame_system::Config>::Event`
found enum `pallet_balances::Event<T>`
= help: consider constraining the associated type `<T as frame_system::Config>::Event` to `pallet_balances::Event<T>`
= note: for more information, visit https://doc.rust-lang.org/book/ch19-03-advanced-traits.html
error: aborting due to previous error
For more information about this error, try `rustc --explain E0308`.
error: could not compile `pallet-offences-benchmarking`
i was using a clone of an old fork i made of the substrate repo, and i'd updated it locally to the latest substrate master, then switched back to v3.0.0.
so i've now deleted that clone, and recloning directly from substrate's repo, then i'll switch to v3.0.0 branch, and try building that
hmm, we'll it just finished building after a fresh clone , and i got the same error.
|
gharchive/issue
| 2021-03-20T10:32:57 |
2025-04-01T04:35:28.279168
|
{
"authors": [
"bkchr",
"ltfschoen"
],
"repo": "paritytech/substrate",
"url": "https://github.com/paritytech/substrate/issues/8408",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
524028459
|
Bump dependencies, respecting semver
We really need a bot to make these PRs, but in the meantime, I can do so with almost no effort.
✄ -----------------------------------------------------------------------------
closing since it's now stale. this really needs to be done and merged immediately upon CI pass.
|
gharchive/pull-request
| 2019-11-17T19:04:23 |
2025-04-01T04:35:28.280865
|
{
"authors": [
"DemiMarie-parity",
"gavofyork"
],
"repo": "paritytech/substrate",
"url": "https://github.com/paritytech/substrate/pull/4117",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
363910527
|
JDK 11 compatibility
Nassau does not currently compile on JDK 11, most likely due to the Error Prone usage.
This was fixed in 805ccc3184061964c94dbf1580dd15f9a9c9a0a9.
|
gharchive/issue
| 2018-09-26T08:22:34 |
2025-04-01T04:35:28.291053
|
{
"authors": [
"jvirtanen"
],
"repo": "paritytrading/nassau",
"url": "https://github.com/paritytrading/nassau/issues/52",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
297718992
|
XLC : Unable to execute in single thread (spawns multiple threads by default)
XLC runs workload on many CPUs despite of no parallel tasks in code. GCC executes in single CPU only but still gives better performance than XLC. CPU pinning gives poorest performance to XLC here.
https://github.com/parthsl/TSP_XeonPhi/blob/79c0642083fcb871fb7b2062d8ef5408b35c08f1/two_opt/Parallel/hill_climb/src/hill_climb.c#L18-L23
Calling OpenMP threads before the call to two_opt_max_swap_single() function results in XLC to spawn multiple threads.
Removing every OpenMP calls from the code results in XLC to call for one thread only and gives better performance compared to gcc.
|
gharchive/issue
| 2018-02-16T09:09:04 |
2025-04-01T04:35:28.469701
|
{
"authors": [
"parthsl"
],
"repo": "parthsl/TSP_XeonPhi",
"url": "https://github.com/parthsl/TSP_XeonPhi/issues/11",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2748227357
|
[profile]: đổi lại ảnh avatar
ảnh avatar đang dùng của quizlet, cần đổi lại
Đã xong
|
gharchive/issue
| 2024-12-17T15:30:57 |
2025-04-01T04:35:28.473616
|
{
"authors": [
"nqmgaming"
],
"repo": "pass-with-high-score/quick_mem",
"url": "https://github.com/pass-with-high-score/quick_mem/issues/103",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1911405924
|
Make ApiSecret a required property
If a property is not nullable and does not have a default value, then it must be marked as required. That way, if the user doesn't set it, it becomes a compile-time error and not a run-time error.
@Tyrrrz I don't have any feedback of value here - I know way to little about polyfilling. I trust you that this works.
@justindbaur do you have any thoughts? 🙂
We are validating that ApiSecret is non-null through: https://github.com/passwordless/passwordless-dotnet/blob/3d841dbe021e2d30382b0c0253017618b623b6f0/src/Passwordless/ServiceCollectionExtensions.cs#L16 but this doesn't happen until first construction of IOptions<PasswordlessOptions>.
I'm guessing Activator.CreateInstance<PasswordlessOptions>() doesn't respect the required?
Yeah, it's validating at run-time, this adds compile-time validation. It won't help in the builder overload but will help in those where PasswordlessOptions are provided directly, including in internal code.
|
gharchive/pull-request
| 2023-09-25T12:29:36 |
2025-04-01T04:35:28.476710
|
{
"authors": [
"Tyrrrz",
"abergs",
"justindbaur"
],
"repo": "passwordless/passwordless-dotnet",
"url": "https://github.com/passwordless/passwordless-dotnet/pull/40",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
679266251
|
How do I style this?
I'd like to add a bit of css to style this component. What's the best way of achieving this?
This is just a wrapper for the leaflet minimap plugin. The minimap plugin has several CSS classes added you can use to style it any way you wish.
The top level CSS class is leaflet-control-minimap. You can see the other CSS classes available by inspecting the DOM in your browser dev tools.
If you want to include the default minimap CSS styling you can add it by loading the CSS file in a SFC style tag, or loading it separately.
<style>
@import '~leaflet-minimap/dist/Control.MiniMap.min.css';
</style>
|
gharchive/issue
| 2020-08-14T16:24:31 |
2025-04-01T04:35:28.496199
|
{
"authors": [
"CodyBontecou",
"patrickcate"
],
"repo": "patrickcate/vue-leaflet-minimap",
"url": "https://github.com/patrickcate/vue-leaflet-minimap/issues/5",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
514772775
|
Pfe-accordion animations on desktop and iOS Safari
Pfe-accordion animations do not complete in desktop and iOS safari.
Expected: the arrow should rotate 90º on state change
Actual: (iOS) opening the panel works correctly, but on close the arrow is rotated incorrectly until changing the focus. Screenshot:
Actual (desktop): The close animation does not complete if the mouse is hovering over the accordion's open button. Screenshot:
Can you confirm if this is still an issue?
Yeah, looks like it's resolved! Not seeing the issue anymore in IOS and Safari 13. Closing.
|
gharchive/issue
| 2019-10-30T15:39:13 |
2025-04-01T04:35:28.517899
|
{
"authors": [
"Djfaucette",
"castastrophe"
],
"repo": "patternfly/patternfly-elements",
"url": "https://github.com/patternfly/patternfly-elements/issues/597",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
360350778
|
Type-Ahead Filter
We require a Type-Ahead Filter component that supports a variety of filtering use case across products. Some requirements for this can be found here: https://docs.google.com/document/d/15w2gpUb7__4oINMtnU7hwSqtlD6cwDIvkDeHM1TMfiQ/edit?usp=sharing and also a preliminary design here:
Unified Filter Design Concept: https://docs.google.com/presentation/d/1yK7HUtUQZLBZGv-NjHGuMpDTs5Iq1mTqQCggGfDb1AQ/edit?usp=sharing
This should be designed and implemented to use a consistent type-ahead approach with the Type-Ahead Select (Ref #722 ).
Closing this because this example will appear in Filter Demo
|
gharchive/issue
| 2018-09-14T15:25:09 |
2025-04-01T04:35:28.520022
|
{
"authors": [
"mcarrano",
"rachael-phillips"
],
"repo": "patternfly/patternfly-next",
"url": "https://github.com/patternfly/patternfly-next/issues/731",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
355622826
|
feature(stylelint): add rule to match pattern for custom properties
fix https://github.com/patternfly/patternfly-next/issues/668
yeah!! let's make the linter evil! 🕶
Deploy preview for pf-next ready!
Built with commit a22030081000192c9ff94d8da262380039ff3eee
https://deploy-preview-669--pf-next.netlify.com
|
gharchive/pull-request
| 2018-08-30T14:53:50 |
2025-04-01T04:35:28.521962
|
{
"authors": [
"andresgalante",
"michael-coker",
"patternfly-build"
],
"repo": "patternfly/patternfly-next",
"url": "https://github.com/patternfly/patternfly-next/pull/669",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
165343968
|
question about error not mentioned in documentation
Hi, when using ProxyMockFactory to mock a trait, I get a compilation error:
<from inside IntelliJ 2016.2: using scala version 2.11.7, jvm version 1.8.0_91,>
I'm using Scalatest FlatSpec class in the test class not the FlatSpecLike trait.
Inside the IDE, no compilation error shows up, but when the maven compile runs, it emits the following:
Error: could not find implicit value for parameter mockContext: org.scalamock.context.MockContext
val m = mock[TransactionProcessor]
I could not see anything in the docs on the web site about implicit MockContext. Can anyone tell me how to fix the error I'm seeing?
Many thanks.
I think that you should read this http://scalamock.org/user-guide/sharing-scalatest/
Thanks for the reply, but I don't think it addresses the question. Where in the document link you refer to, is the implicit mock context to be found. I have, in fact, tried the style of solution referred to in to web site document, but the compilation error remains present and therefore the test cannot run.
You should construct your suite the other way. Your test should look like that:
class MyTest extends FlatSpec with MockFactory. Maybe my way is not correct somehow but I have never used ProxyMockFactory directly. Nevertheless, try with what I proposed you and let me know if it works.
Hi Bartek
Thanks again for taking the time to reply.
I am trying to test functions in a trait, and mocks from MockFactory will not compile with traits. The scalamock documents indicate that when testing traits, the proxyMockFactory mock is the correct one to use. So, when I try the "ordinary" mock, there is a compiler error immediately visible when providing a trait name in the mock parameter. Thus, I have the problem described, the mock from ProxyMockFactory depends on an implicit mockContext that is not visible / reachable.
Hi,
In addition, I've tried explicitly creating a MockContext and that too, results in compilation errors; e.g.
Error: trait MockContext in package context cannot be accessed in package org.scalamock.context
implicit val mockCtx = new MockContext {
override protected def newExpectationException(message: String, methodName: Option[Symbol]): this.type = ???
override type ExpectationException = this.type
}
Error: not found: type MockContext
implicit val mockCtx = new MockContext {
It looks like I'll have to give up on Scalamock, and try to find another (working) solution.
Thanks for your replies.
Not sure what to do with this, need a code snippet to reproduce or comment.
Please try with the latest version and raise a new issue if you still have a problem.
I'm running into the same problem. This is a mostly whittled-down code sample that demonstrates the error:
trait SomeTrait {
val hasThings = Vector("a", "b", "c")
def doesStuff = println("hello")
}
class SomeClass {
this: SomeTrait =>
}
class TestSpec extends ProxyMockFactory
with FunSpecLike {
val fakeTrait = mock[SomeTrait]
val sut = new SomeClass with fakeTrait
}
Errors are:
Error:(34, 23) could not find implicit value for parameter mockContext: org.scalamock.context.MockContext
val fakeTrait = mock[SomeTrait]
Error:(34, 23) not enough arguments for method mock: (implicit evidence$1: scala.reflect.ClassTag[com.shiftwise.rtisystem.test.actors.SomeTrait], implicit mockContext: org.scalamock.context.MockContext)com.shiftwise.rtisystem.test.actors.SomeTrait with org.scalamock.proxy.Mock.
Unspecified value parameter mockContext.
val fakeTrait = mock[SomeTrait]
If I had to guess, I'd assume it has to do with the self-type annotation, but I'm new to Scala so I could be doing a number of other things wrong, too.
I'm using version 3.3.0 with Scalatest 3.0.0 with Scala version 2.11.8.
Thanks!
can you try again, with 3.4.2 please?
Also, I think you should be using org.scalamock.scalatest.proxy.MockFactory instead of ProxyMockFactory
I've updated to 3.4.2 and my code now looks like:
trait SomeTrait {
val hasThings = Vector("a", "b", "c")
def doesStuff = println("hello")
}
class SomeClass {
this: SomeTrait =>
}
class TestSpec extends org.scalamock.scalatest.proxy.MockFactory
with FunSpecLike {
val fakeTrait = mock[SomeTrait]
val sut = new SomeClass with fakeTrait
}
and the errors have changes, but it's still not compiling:
Error:(34, 27) not found: type mock
trait fakeTrait extends mock[SomeTrait]
Error:(35, 17) illegal inheritance;
self-type SomeClass with TestSpec.this.fakeTrait does not conform to SomeClass's selftype SomeClass with .SomeTrait
val sut = new SomeClass with fakeTrait
I realized this morning that the latter change was suggested above; sorry about that. I appreciate your patience and thank you for maintaining the project!
Hi, see my example below.
Not sure if you can SomeClass to SomeClass2 in your code, but I don't see a good way of mocking out the SomeClass version with the selftype, sorry.
import org.scalamock.scalatest.proxy.MockFactory
import org.scalatest.{FunSpec, FunSpecLike}
trait SomeTrait {
val hasThings = Vector("a", "b", "c")
def doesStuff() = println("hello")
}
class SomeClass {
this: SomeTrait =>
}
trait SomeClass2 {
val someTrait: SomeTrait
}
class Issue146Test extends FunSpec with MockFactory {
describe("foo") {
it("should work") {
val fakeTrait = mock[SomeTrait]
// does not compile
//val sut = new SomeClass with fakeTrait
val sut2 = new SomeClass2 {
val someTrait = fakeTrait
}
}
}
}
Thank you for the example; while it's a little more complicated, it isn't tooo bad. I think I've figured out how to rejigger my non-example code so that I can use the trait without selftyping and still test the functionality that uses the trait behavior, but if it doesn't work out, I'll try implementing the test based on your sample.
And that looks like a fantastic blog post. I appreciate your time and hope you have a good day! :)
|
gharchive/issue
| 2016-07-13T15:02:45 |
2025-04-01T04:35:28.577458
|
{
"authors": [
"barkhorn",
"bkowalik",
"lkemmerer",
"london-coder"
],
"repo": "paulbutcher/ScalaMock",
"url": "https://github.com/paulbutcher/ScalaMock/issues/146",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
463826216
|
Feature Request: Long-text or HTML editor column type support
We have spreadsheets that often have very long text values such as titles and descriptions of images. I haven't found an obvious way to have a column wrap text or adjust the row height to display the full field value.
Also, since the existing columns are based on jSuite input types, it would be great to support the mini text editor.
As a work-around I'm currently considering using a separate stand-alone field to act like Excel's formula bar for long-text entry that will copy in the value of the currently active field and save the value back on a cell change. Do you know of any examples where others have done similar?
Thanks!
Long texts, please use wordWrap:true in the column directives. Instead of a input will be render a textarea.
https://bossanova.uk/jexcel/v2/examples/text-wrapping
This works in version 3 as well.
In regards to the HTML editor, I will consider for future releases. Thanks for the suggestion.
This has been implemented in both distributions with type: html
https://bossanova.uk/jexcel/v4/examples/richtext-html-editor
|
gharchive/issue
| 2019-07-03T15:44:04 |
2025-04-01T04:35:28.596418
|
{
"authors": [
"paulhodel",
"seth-shaw-unlv"
],
"repo": "paulhodel/jexcel",
"url": "https://github.com/paulhodel/jexcel/issues/419",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2152980130
|
Validate Forms
We need to validate forms to make sure users input the correct type of data. We also got a list from the project partner telling us which fields are required, so we'll need to change the form to reflect that.
Here is the list of fields we need to make required:
Entity Name
Status
Industrial/Chemical
Headquarters’ Country
Stakeholder Description
with required checking done, the only thing left to do for this is to do type validation, meaning changing fields to their proper types. currently everything is varchar, but int fields should be ints, date fields dates, etc.
no longer needed
|
gharchive/issue
| 2024-02-25T22:57:19 |
2025-04-01T04:35:28.602910
|
{
"authors": [
"paullpp"
],
"repo": "paullpp/hemp-db",
"url": "https://github.com/paullpp/hemp-db/issues/57",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1835468976
|
Resolve unintentional MIT License omission
Hi Paul,
It appears that the subs in this repo are either:
copies of the stubs in https://github.com/Josverl/micropython-stubs
generated by the use of the micropython-stubber tool : https://github.com/Josverl/micropython-stubbber
That in itself is not a problem at all, happy that you have been able to use them for the exact purpose that I have built these for.
However both are released under MIT licence which requires that that license is respected.
The relevant parts:
MIT License
Copyright (c) 2020-2023 Jos Verlinde
The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software and stubs.
So it is no problem that you place this under a different license, however you cannot remove the earlier license and notice.
in the development and publication of the micropython-stubs that have used as the base, I have taken specific effort to respect , and where relevant copy over relevant licenses with the stubs.
I kindly ask you to do the same.
Regards,
Jos Verlinde
@Josverl I'm sorry that I missed that one and of course I'll add the license header to the stubber which seems to be missing in the minified copy I got from your repo some months ago.
The repo does not contain copies of your stubs repo but it contains the auto-generated ones from the stubber and were than each one hand modified.
I also thought about in the past to maybe close this repo and contact you if it would be possible to incorporate your stubs into the MicroPico extension. In return I would to my contributions back to your stubs repo instead of this one.
My only problem was that your repo is quite big and contains many different stubs for the pico-w. Also most of them don't contain the parameters for functions and doc-strings for classes and methods or maybe I haven't found them.
Paul,
thanks for resolving this.
really appreciated.
Regards,
Jos Verlinde
Jos,
sounds great. I thought about making the extension more suitable for different boards so giving the user the option to chose between pico, pico-w and other stubs would be a great step towards the right direction.
If you agree, then I'll let him know and see how we can work together on this.
Sure
OK I'll ping Jim on discord. Do you have a discord handle?
OK I'll ping Jim on discord. Do you have a discord handle?
Yes: paulober
I've let him know, but he has not responded so far
|
gharchive/issue
| 2023-08-03T17:36:42 |
2025-04-01T04:35:28.609813
|
{
"authors": [
"Josverl",
"paulober"
],
"repo": "paulober/Pico-W-Stub",
"url": "https://github.com/paulober/Pico-W-Stub/issues/11",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
617628502
|
Syntax error with deps dependencies
Hello! I was super excited to find this extension and installed it, but it appears there's an error with one of my dependencies:
bump:
desc: Bump the version
deps:
- task: build:generated-dir
Note that this is proper syntax, and is outlined in the section on task dependencies: https://taskfile.dev/#/usage?id=task-dependencies
version: '2'
tasks:
default:
deps:
- task: echo_sth
vars: {TEXT: "before 1"}
- task: echo_sth
vars: {TEXT: "before 2"}
cmds:
- echo "after"
echo_sth:
cmds:
- echo {{.TEXT}}
@klauern Good catch, the schema was missing the alternative deps format. Just added it and the fix should be in the next release
This should now have made it to the extension
This should now have made it to the extension
|
gharchive/issue
| 2020-05-13T17:31:38 |
2025-04-01T04:35:28.621594
|
{
"authors": [
"klauern",
"paulvarache"
],
"repo": "paulvarache/vscode-taskfile",
"url": "https://github.com/paulvarache/vscode-taskfile/issues/3",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2283797225
|
Game Scene - Layers should be stored within a singular vector/list versus the normal elements list
Layers are a crucial part of the scene object and to avoid eternal typecasting up from the branch class they should be stored in a vector or list within the scene class.
This way the scene class can interact with the layers(currently engine allows up to 32 per scene) so that typecasting is avoided as much as possible, and code can be easier readable.
Merged via https://github.com/pawbyte/Game-Pencil-Engine-Editor/pull/149
|
gharchive/issue
| 2024-05-07T16:39:38 |
2025-04-01T04:35:28.641285
|
{
"authors": [
"nhurde"
],
"repo": "pawbyte/Game-Pencil-Engine-Editor",
"url": "https://github.com/pawbyte/Game-Pencil-Engine-Editor/issues/141",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2462263281
|
Some arguments type is missing in afterOperation hook
Link to reproduction
No response
Payload Version
3.0.0-beta.77
Node Version
20
Next.js Version
15.0.0-canary.53
Describe the Bug
Some arguments type is missing in afterOperation hook
Reproduction Steps
export const afterOperation: CollectionAfterOperationHook = async ({
args,
operation,
result,
req,
}) => {
if (operation === 'create') {
const { url, download } = args.data
}
}
(parameter) result: unknown
Property 'url' does not exist on type 'RequiredDataFromCollectionSlug'
Adapters and Plugins
No response
Hey @smoothdvd can you pass your collection slug into the type?
eg:
CollectionAfterOperationHook<'posts'>
Then your data there will be inferred correctly from the generated types. These should generate automatically but if not then pnpm generate:types or pnpm payload generate:types should do that for you.
Also please note that on the beta versions 77+ you should be on Next canary.104 as a minimum.
I tested the above in a fresh project, so let me know if this fixes your issues, otherwise please provide a reproduction so I can take a closer look!
@paulpopus Thank you, it worked after I follow your instruct (I also upgraded to beta.79 and next 15.0.0-canary.104).
|
gharchive/issue
| 2024-08-13T02:44:33 |
2025-04-01T04:35:28.676631
|
{
"authors": [
"paulpopus",
"smoothdvd"
],
"repo": "payloadcms/payload",
"url": "https://github.com/payloadcms/payload/issues/7642",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2477687786
|
Extend formBuilderPlugin fields: not persist data
Link to reproduction
Payload Version
3.0.0-beta.86
Node Version
v20
Next.js Version
15.0.0-canary.121
Describe the Bug
If I add a custom field in the plugin configuration, it displays correctly in the admin UI but the data is not persisted.
If I add the "slug" property the data is persisted but there is a typescript error because is missing in the interface FieldConfig.
Reproduction Steps
Declare the custom field
import { fields, formBuilderPlugin } from "@payloadcms/plugin-form-builder";
import { Block } from "payload";
export const formBuilderPluginConfig = formBuilderPlugin({
redirectRelationships: ["pages"],
fields: {
hidden: {
// slug: "hidden", // <- With this prop works
labels: {
singular: 'Hidden',
plural: 'Hidden'
},
fields: [
...(fields.text as Block).fields
// Other fields
]
}
},
})
Open the Admin UI to add the custom field
Save
Data is not persisted
Adapters and Plugins
"@payloadcms/db-mongodb": "beta", "@payloadcms/email-nodemailer": "beta", "@payloadcms/next": "beta", "@payloadcms/plugin-cloud-storage": "beta", "@payloadcms/plugin-form-builder": "beta", "@payloadcms/plugin-nested-docs": "beta", "@payloadcms/plugin-seo": "beta", "@payloadcms/richtext-lexical": "beta", "@payloadcms/storage-s3": "beta", "@payloadcms/ui": "beta",
@matteo-naif did you find a work around for this? running into this same issue currently. Also trying to add a custom "hidden" field to the formBuilder plugin.
Hi @siobahnschuck, I paste my working example with minimal setup
import { formBuilderPlugin } from "@payloadcms/plugin-form-builder";
import { fields } from "@payloadcms/plugin-form-builder";
import { FieldConfig } from "@payloadcms/plugin-form-builder/types";
import { Block } from "payload";
export const textFields = (fields.text as Block).fields
export const InputHidden: FieldConfig = {
// @ts-ignore
slug: "hidden",
labels: {
singular: 'Hidden',
plural: 'Hidden'
},
fields: [
...textFields
// Other fields
]
};
export const formBuilderPluginConfig = formBuilderPlugin({
redirectRelationships: ["pages"],
fields: {
hidden: InputHidden
}
})
The result is the following:
@matteo-naif thank you so much! This works well 🌟
|
gharchive/issue
| 2024-08-21T10:35:08 |
2025-04-01T04:35:28.679955
|
{
"authors": [
"matteo-naif",
"siobahnschuck"
],
"repo": "payloadcms/payload",
"url": "https://github.com/payloadcms/payload/issues/7787",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2592102768
|
CSS z-hight for lexical toolbar is off
Describe the Bug
The toolbar for lexical editor appears over a dropdown menu:
I use PayloadCMS v3 beta-114
Link to the code that reproduces this issue
npx create-payload-app@beta -t website
Reproduction Steps
npx create-payload-app@beta
Visit http://localhost:4444/admin/collections/pages/670fcbf993abb42cdb3e4bd7
Which area(s) are affected? (Select all that apply)
Not sure
Environment Info
Node.js v22.6.0
Binaries:
Node: 22.6.0
npm: 10.9.0
Yarn: N/A
pnpm: 9.7.1
Relevant Packages:
payload: 3.0.0-beta.114
next: 15.0.0-canary.173
@payloadcms/db-mongodb: 3.0.0-beta.114
@payloadcms/email-nodemailer: 3.0.0-beta.114
@payloadcms/graphql: 3.0.0-beta.114
@payloadcms/live-preview: 3.0.0-beta.114
@payloadcms/live-preview-react: 3.0.0-beta.114
@payloadcms/next/utilities: 3.0.0-beta.114
@payloadcms/plugin-cloud: 3.0.0-beta.114
@payloadcms/plugin-cloud-storage: 3.0.0-beta.114
@payloadcms/plugin-form-builder: 3.0.0-beta.114
@payloadcms/plugin-nested-docs: 3.0.0-beta.114
@payloadcms/plugin-redirects: 3.0.0-beta.114
@payloadcms/plugin-search: 3.0.0-beta.114
@payloadcms/plugin-seo: 3.0.0-beta.114
@payloadcms/richtext-lexical: 3.0.0-beta.114
@payloadcms/storage-s3: 3.0.0-beta.114
@payloadcms/translations: 3.0.0-beta.114
@payloadcms/ui/shared: 3.0.0-beta.114
react: 19.0.0-rc-3edc000d-20240926
react-dom: 19.0.0-rc-3edc000d-20240926
Operating System:
Platform: darwin
Arch: x64
Version: Darwin Kernel Version 23.6.0: Mon Jul 29 21:13:00 PDT 2024; root:xnu-10063.141.2~1/RELEASE_X86_64
Available memory (MB): 32768
Available CPU cores: 16
Hey @exoer,
I believe this to be a duplicate of #8728 - which already has a PR up for a fix.
@akhrarovsaid is right, closing
|
gharchive/issue
| 2024-10-16T14:39:26 |
2025-04-01T04:35:28.685233
|
{
"authors": [
"akhrarovsaid",
"exoer",
"r1tsuu"
],
"repo": "payloadcms/payload",
"url": "https://github.com/payloadcms/payload/issues/8740",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2707013419
|
The custom view of AfterNavLinks has a position mismatch in Safari on initial state
Describe the Bug.
macOS: 15.1.1, Safari 18.1.1
Chrome: 131.0.6778.86
My component code:
'use client'
import Link from 'next/link'
import { usePathname } from 'next/navigation'
import { NavGroup } from '@payloadcms/ui'
import React from 'react'
const AfterNavLinks: React.FC = () => {
const pathname = usePathname()
const active = pathname === '/bull-board'
return (
<NavGroup key="bull-board" label="Bull Board">
<Link href="/bull-board" target="_blank">
Bull Board
</Link>
</NavGroup>
)
}
export default AfterNavLinks
The view displays correctly after I expand and collapse all navigation groups in Safari.
Reproduction Steps
Add custom component on AfterNavLinks hook.
Open the navigation
Environment Info
Payload:
Binaries:
Node: 22.11.0
npm: 10.9.0
Yarn: N/A
pnpm: 9.14.4
Relevant Packages:
payload: 3.2.2
Operating System:
Platform: darwin
Arch: arm64
Version: Darwin Kernel Version 24.1.0: Thu Oct 10 21:05:14 PDT 2024; root:xnu-11215.41.3~2/RELEASE_ARM64_T8103
Available memory (MB): 8192
Available CPU cores: 8
Node: 22
Next.js: 15
It looks like this issue was fixed in version 3.4.0.
|
gharchive/issue
| 2024-11-30T08:51:12 |
2025-04-01T04:35:28.687196
|
{
"authors": [
"smoothdvd"
],
"repo": "payloadcms/payload",
"url": "https://github.com/payloadcms/payload/issues/9626",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1533569620
|
Region is Missing Error
Getting the current error on s3 upload. I can't find anything in the docs about "Region" in configuration.
[00:11:06] ERROR (payload): Region is missing
Error: Region is missing
at default (/path/node_modules/@aws-sdk/config-resolver/dist-cjs/regionConfig/config.js:10:15)
at /path/node_modules/@aws-sdk/node-config-provider/dist-cjs/fromStatic.js:6:83
at /path/node_modules/@aws-sdk/property-provider/dist-cjs/chain.js:11:28
at processTicksAndRejections (internal/process/task_queues.js:95:5)
at async coalesceProvider (/path/node_modules/@aws-sdk/property-provider/dist-cjs/memoize.js:14:24)
at async /path/node_modules/@aws-sdk/property-provider/dist-cjs/memoize.js:26:28
at async useFipsEndpoint (/path/node_modules/@aws-sdk/config-resolver/dist-cjs/regionConfig/resolveRegionConfig.js:21:74)
at async resolveParams (/path/node_modules/@aws-sdk/middleware-endpoint/dist-cjs/adaptors/getEndpointFromInstructions.js:29:40)
at async getEndpointFromInstructions (/path/node_modules/@aws-sdk/middleware-endpoint/dist-cjs/adaptors/getEndpointFromInstructions.js:7:28)
at async /path/node_modules/@aws-sdk/middleware-endpoint/dist-cjs/endpointMiddleware.js:8:26
[00:11:14] ERROR (payload): There was an error while uploading files corresponding to the collection media with filename Screen Shot 2023-01-14 at 6.15.24 PM-1.png:
[00:11:14] ERROR (payload): Region is missing
Error: Region is missing
at default (/path/node_modules/@aws-sdk/config-resolver/dist-cjs/regionConfig/config.js:10:15)
at /path/node_modules/@aws-sdk/node-config-provider/dist-cjs/fromStatic.js:6:83
at /path/node_modules/@aws-sdk/property-provider/dist-cjs/chain.js:11:28
[00:11:14] ERROR (payload): Region is missing
Error: Region is missing
at default (/path/node_modules/@aws-sdk/config-resolver/dist-cjs/regionConfig/config.js:10:15)
at /path/node_modules/@aws-sdk/node-config-provider/dist-cjs/fromStatic.js:6:83
at /path/node_modules/@aws-sdk/property-provider/dist-cjs/chain.js:11:28
at processTicksAndRejections (internal/process/task_queues.js:95:5)
at async coalesceProvider (/path/node_modules/@aws-sdk/property-provider/dist-cjs/memoize.js:14:24)
at async /path/node_modules/@aws-sdk/property-provider/dist-cjs/memoize.js:26:28
at async useFipsEndpoint (/path/node_modules/@aws-sdk/config-resolver/dist-cjs/regionConfig/resolveRegionConfig.js:21:74)
at async resolveParams (/path/node_modules/@aws-sdk/middleware-endpoint/dist-cjs/adaptors/getEndpointFromInstructions.js:29:40)
at async getEndpointFromInstructions (/path/node_modules/@aws-sdk/middleware-endpoint/dist-cjs/adaptors/getEndpointFromInstructions.js:7:28)
at async /path/node_modules/@aws-sdk/middleware-endpoint/dist-cjs/endpointMiddleware.js:8:26
hey @dundas, you could add the region to the configuration
adapter: s3Adapter({ config: { credentials: { accessKeyId: process.env.S3_ACCESS_KEY_ID, secretAccessKey: process.env.S3_SECRET_ACCESS_KEY, }, region: process.env.S3_REGION, }, bucket: process.env.S3_BUCKET, }), },
Yes, @dundas - the above is correct. You just need to specify your region.
I will close this one out. If that didn't solve your problem, please let us know!
|
gharchive/issue
| 2023-01-15T00:15:52 |
2025-04-01T04:35:28.689065
|
{
"authors": [
"dundas",
"jmikrut",
"tomsvogel"
],
"repo": "payloadcms/plugin-cloud-storage",
"url": "https://github.com/payloadcms/plugin-cloud-storage/issues/33",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
210699237
|
Allow any intent to be executed on the client side
Right now only sale is supported.
Fixed and pushed.
|
gharchive/issue
| 2017-02-28T06:32:53 |
2025-04-01T04:35:28.697864
|
{
"authors": [
"bluepnume"
],
"repo": "paypal/paypal-checkout",
"url": "https://github.com/paypal/paypal-checkout/issues/205",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2035589759
|
🛑 Departamento is down
In 5389320, Departamento (http://cca.unb.br/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Departamento is back up in e216a20 after 21 minutes.
|
gharchive/issue
| 2023-12-11T12:37:55 |
2025-04-01T04:35:28.699366
|
{
"authors": [
"pazkero"
],
"repo": "pazkero/status.cacic.bsb.br",
"url": "https://github.com/pazkero/status.cacic.bsb.br/issues/1573",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2201844591
|
🛑 Departamento is down
In 20bfb8d, Departamento (http://cca.unb.br/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Departamento is back up in 3907ee4 after 12 minutes.
|
gharchive/issue
| 2024-03-22T07:10:18 |
2025-04-01T04:35:28.700844
|
{
"authors": [
"pazkero"
],
"repo": "pazkero/status.cacic.bsb.br",
"url": "https://github.com/pazkero/status.cacic.bsb.br/issues/2869",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
379100070
|
show solution: fails when replaying commands with o/xxx in them
Problem
I use git checkout -b fixForMR o/feature in a solution.
When replaying this command via show solution, there are some transcoding errors as it tries to run 2 commands:
$ git checkout -b fixForMR o/
The ref o/ does not
exist or is unknown
$ feature
The command "feature" isn't
supported, sorry!
...
How to reproduce
Gist level id: 13b665274699425ce772ad5fe6705cbd
Direct access: https://learngitbranching.js.org/?NODEMO&gist_level_id=13b665274699425ce772ad5fe6705cbd&locale=en_US
type in: show solution and click Confirm
sorry for the delay here, busy week. neat that you found / are using the gist level IDs!
yeah theres some weirddd escaping. let me try to debug
the solution command correct is correct in the gist:
"solutionCommand": "git checkout -b fixForMR o/feature;git commit;git push origin fixForMR:feature;git rebase fixForMR feature",
|
gharchive/issue
| 2018-11-09T10:11:49 |
2025-04-01T04:35:28.741435
|
{
"authors": [
"arnaudjolly",
"pcottle"
],
"repo": "pcottle/learnGitBranching",
"url": "https://github.com/pcottle/learnGitBranching/issues/497",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
361744384
|
Split reserve into 2 functions
Currently reserve() has an optional timeout parameter.
This leads to an optional return type; since if a timeout happens no exception is thrown yet we don't get a Job as return value.
I propose:
Change reserve() to:
/** @throws someException **/
public function reserve(): Job
Add reserveWithTimeout:
/**
@throws someException
@throws TimeoutException
**/
public function reserveWithTimeout(int $timeout): Job
Another option could be with a Null Object pattern - a subclass of Job that has no contents. Its a little cleaner than a nullable job (function reserve(): ?Job), which is possible from PHP 7.1.
It's not an exceptional event if the queue is empty and equally, a reserve could end up not returning a job either with socket timeouts - or other fairly ordinary issues. So, not having a job to work on is an event that should be handled - not normally as a particular problem though.
I've written code that would pause for a little longer than usual on an empty queue, especially if it's not likely to be urgent, or there will often be large gaps between things to do.
either with socket timeouts - or other fairly ordinary issues.
These issues are already exceptions, and they should be...
|
gharchive/issue
| 2018-09-19T13:16:15 |
2025-04-01T04:35:28.752179
|
{
"authors": [
"SamMousa",
"alister"
],
"repo": "pda/pheanstalk",
"url": "https://github.com/pda/pheanstalk/issues/178",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
860131936
|
Added information to readme on supported models and setup requirements
Included some useful information (mainly from closed issues) to the readme
Users continue to raise issues (#11, #14) about compatible models which might be preempted with updated documentation
|
gharchive/pull-request
| 2021-04-16T20:12:07 |
2025-04-01T04:35:28.760307
|
{
"authors": [
"lukestein"
],
"repo": "pdemarco925/homebridge-icomfort",
"url": "https://github.com/pdemarco925/homebridge-icomfort/pull/13",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
803085768
|
Memory leak?
Bug report
A description of the bug
Somewhere in pdfminer.six there appears to be a memory leak; even after a Python process is done handling a PDF, the memory it required is not released.
Steps to reproduce the bug
pip install memory_profiler
Save this script as memory.py:
from pdfminer.high_level import extract_text
from memory_profiler import profile
import gc
path = "samples/nonfree/nlp2004slides.pdf"
@profile
def main():
text = extract_text(path)
gc.collect()
main()
Run python memory.py
Output:
Line # Mem usage Increment Occurences Line Contents
============================================================
7 44.3 MiB 44.3 MiB 1 @profile
8 def main():
9 51.5 MiB 7.2 MiB 1 text = extract_text(path)
10 51.5 MiB 0.0 MiB 1 gc.collect()
Note: The problem does not seem specific to extract_text; other, lower-level usages of pdfminer.six seem to create similar effects. Also: Passing caching=False to extract_text does not appear to reduce the memory leak.
I have spent some time trying to track down the source of apparent leak, but have not had much luck. I'll continue looking, but am raising this issue in the meantime, in case the cause is more readily apparent to others.
I just faced two consecutive outages of the system I maintain - I tracked down the reason to one PDF a user wanted to upload - my system extracts the text to index the PDF document. I can provide the document to a maintainer ( it should not go public). Basically, the PDF consists of slides - when I open the PDF with evince I can view it, but the pages get rendered very, very slowly.
So there is obviously something bogus.
Is anybody aware of a tool which "fixes" bogus PDF files? If not, I think I'll just screenshot the pages and create a new one.
I can reproduce this issue. Will have a look at the PR's.
Would love to merge to PR's related to this issue but they don't seem to solve the issue.
I'm experiencing a regression of the recursion error in issue https://github.com/pdfminer/pdfminer.six/issues/404 with version 20220524. Could this be related?
Hi, I'm still running into this issue on version 20220524 of pdfminer.six. I can't provide the pdf that causes the issue since it needs to be anonymized, but these are the LAParams we're using for pdfminer:
LAParams(
line_overlap= 0.5,
char_margin= 1.1,
word_margin= 0.2,
line_margin= 0.5,
boxes_flow= 0.5,
all_texts=True,
)
Traceback:
RecursionError: maximum recursion depth exceeded
(245 additional frame(s) were not displayed)
...
File "pdfminer/layout.py", line 705, in analyze
super().analyze(laparams)
File "pdfminer/layout.py", line 439, in analyze
obj.analyze(laparams)
File "pdfminer/layout.py", line 705, in analyze
super().analyze(laparams)
File "pdfminer/layout.py", line 439, in analyze
obj.analyze(laparams)
File "pdfminer/layout.py", line 439, in analyze
obj.analyze(laparams)
maximum recursion depth exceeded
Is there any update on this issue?
No update here yet. The root cause needs to be found first.
Hey, is there any update on this issue?
|
gharchive/issue
| 2021-02-07T23:14:32 |
2025-04-01T04:35:28.767642
|
{
"authors": [
"Rumpelcita",
"dominikhait",
"jsvine",
"jugmac00",
"pietermarsman",
"yoavkedem1"
],
"repo": "pdfminer/pdfminer.six",
"url": "https://github.com/pdfminer/pdfminer.six/issues/580",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1373366151
|
[ldd-msn] Change cardinality of mission_phase_identifier and mission_phase_name
Describe the issue identified (if applicable)
In some cases, particularly in bundle and collection labels, the mission_phase_identifier and mission_phase_name attributes may be multivalued. However, the cardinality for these two attributes is 0..1.
Describe the solution you'd like
Change the cardinality for both mission_phase_identifier and mission_phase_name to 0..Unbounded.
Describe alternatives you've considered
An alternative approach would be to have multiple instances of the Mission_Information object, each with a single instance of mission_phase_name and/or mission_phase_identifier. This approach is valid under the current version of the MSN LDD, but is unnecessarily verbose.
LDD Dictionary Version
1300
PDS4 IM Version
1I00
Need-by Date
Needed by: 2022-09-30. This change is needed for labels currently under designed.
Resolved by pds-data-dictionaries/ldd-msn#32
|
gharchive/issue
| 2022-09-14T17:47:00 |
2025-04-01T04:35:28.786612
|
{
"authors": [
"cgobat",
"jmafi"
],
"repo": "pds-data-dictionaries/PDS4-LDD-Issue-Repo",
"url": "https://github.com/pds-data-dictionaries/PDS4-LDD-Issue-Repo/issues/238",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
67762273
|
Add new layout attributes.
Added foregroundColor, backgroundColor, headings, paragraphs and
lastUpdated.
Ping @matthewtole
Coverage remained the same at 100.0% when pulling 4396365e760db08b3c693bf070355959d1ca770e on feature/headings-paragraphs-colors into faca07ea269664dc68c0bf71a8b5f1f07e5395ae on master.
Coverage remained the same at 100.0% when pulling 4396365e760db08b3c693bf070355959d1ca770e on feature/headings-paragraphs-colors into faca07ea269664dc68c0bf71a8b5f1f07e5395ae on master.
Aside from minor nits about error messages, :+1:
Coverage remained the same at 100.0% when pulling e9776ce4f6ea0df8a73749ddc3f4eb4165be6130 on feature/headings-paragraphs-colors into faca07ea269664dc68c0bf71a8b5f1f07e5395ae on master.
|
gharchive/pull-request
| 2015-04-11T09:40:17 |
2025-04-01T04:35:28.800973
|
{
"authors": [
"Neal",
"coveralls",
"matthewtole"
],
"repo": "pebble/pebble-api-node",
"url": "https://github.com/pebble/pebble-api-node/pull/10",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1237008823
|
Create fr.json
add locales FR
Hey @scorpio686, could you update the json with this? Ty!
"headOverlays": {
"title": "Appearance",
"hair": {
"title": "Hair",
"style": "Style",
"color": "Color",
"highlight": "Highlight"
"highlight": "Highlight",
"fade": "Fade" // This need to be added
},
"opacity": "Opacity",
"style": "Style",
"color": "Color",
"blemishes": "Blemishes",
"beard": "Beard",
"eyebrows": "Eyebrows",
"ageing": "Ageing",
"makeUp": "Make up",
"blush": "Bblush",
"complexion": "Complexion",
"sunDamage": "Sun damage",
"lipstick": "Lipstick",
"moleAndFreckles": "Mole and Freckles",
"chestHair": "Chest hair",
"bodyBlemishes": "Body blemishes",
"eyeColor": "Eye color"
},
added on #80
|
gharchive/pull-request
| 2022-05-16T11:24:14 |
2025-04-01T04:35:28.803133
|
{
"authors": [
"franfdezmorales",
"scorpio686"
],
"repo": "pedr0fontoura/fivem-appearance",
"url": "https://github.com/pedr0fontoura/fivem-appearance/pull/44",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
258531237
|
npm install --save node-pdfbox
When i try to install on centos 6 typing
npm install --save node-pdfbox
i got this error:
gyp ERR! build error
gyp ERR! stack Error: make failed with exit code: 2
gyp ERR! stack at ChildProcess.onExit (/usr/local/lib/node_modules/npm/node_modules/node-gyp/lib/build.js:270:23)
gyp ERR! stack at emitTwo (events.js:87:13)
gyp ERR! stack at ChildProcess.emit (events.js:172:7)
gyp ERR! stack at Process.ChildProcess._handle.onexit (internal/child_process.js:200:12)
gyp ERR! System Linux 2.6.32-696.1.1.el6.x86_64
gyp ERR! command "/usr/local/bin/node" "/usr/local/lib/node_modules/npm/node_modules/node-gyp/bin/node-gyp.js" "rebuild"
gyp ERR! cwd /var/nodejs/imagick/node_modules/node-pdfbox/node_modules/java
gyp ERR! node -v v4.2.2
gyp ERR! node-gyp -v v3.0.3
gyp ERR! not ok
npm ERR! Linux 2.6.32-696.1.1.el6.x86_64
npm ERR! argv "/usr/local/bin/node" "/usr/local/bin/npm" "install"
npm ERR! node v4.2.2
npm ERR! npm v2.14.7
npm ERR! code ELIFECYCLE
npm ERR! java@0.8.0 install: node-gyp rebuild
npm ERR! Exit status 1
npm ERR!
npm ERR! Failed at the java@0.8.0 install script 'node-gyp rebuild'.
npm ERR! This is most likely a problem with the java package,
npm ERR! not with npm itself.
npm ERR! Tell the author that this fails on your system:
npm ERR! node-gyp rebuild
npm ERR! You can get their info via:
npm ERR! npm owner ls java
npm ERR! There is likely additional logging output above.
npm ERR! Please include the following file with any support request:
npm ERR! /var/nodejs/imagick/npm-debug.log
but javac is installed
Can you help me?
After changing java version to the latest, I still got this error. After reading this article, compared with java and node version, I changed the node version to the older v8.11.1 and it installed correctly !
|
gharchive/issue
| 2017-09-18T15:55:13 |
2025-04-01T04:35:28.809798
|
{
"authors": [
"Giancarlo1974",
"dashtom3"
],
"repo": "pedro-lucas/node-pdfbox",
"url": "https://github.com/pedro-lucas/node-pdfbox/issues/7",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
357337829
|
Dimensões maiores que 70cm aumenta o valor do frete em R$20,00 (SOLUÇÃO)
Segue arquivos em anexo corrigidos com a opção no admin para atualizar o valor dos correios sempre que necessário (atualmente 79 reais).
(apagar o .txt do final do arquivo e substituir).
ect/system.xml
model/carrier/correiosMethod.php
CorreiosMethod.php.txt
system.xml.txt
Valeu @mateussantin!
Se puder enviar na forma de uma PR, talvez o @pedro-teixeira possa mesclar a extensão.
De qualquer forma, obrigado 👍
Boa noite, fiz um edit na minha issue detalhando melhor, sou meio novo no
git.
Vou mandar aqui o arquivo editado também, é um prazer contribuir com a
comunidade espero que o Pedro possa incluir ou melhorar o código.
** Arquivo: CorreiosMethod.php
** Localizado:
app/code/community/PedroTeixeira/Correios/Model/Carrier/CorreiosMethod.php
Criado variável na linha: 58;
Inlcuido função _generateAument na linha: 293;
Chamada da função e soma do valor na linha: 347 348 e 349;
Em qua, 5 de set de 2018 às 16:03, Rafael Patro notifications@github.com
escreveu:
Valeu @mateussantin https://github.com/mateussantin!
Se puder enviar na forma de uma PR, talvez o @pedro-teixeira
https://github.com/pedro-teixeira possa mesclar a extensão.
De qualquer forma, obrigado 👍
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
https://github.com/pedro-teixeira/correios/issues/393#issuecomment-418845195,
or mute the thread
https://github.com/notifications/unsubscribe-auth/Am-kiAs0z_ku2YvWwi1pZcQBv5oPuCJVks5uYCAegaJpZM4WbZgQ
.
Alguém poderia enviar a alteração completa do CorreiosMethod.php, não tenho conhecimento em PHP, tentei fazer as modificações explicadas acima mas só está dando erro. Agradeço desde já!!
Eu também não consegui fazer, segui o passo a passo mas deu o erro:
Parse error: syntax error, unexpected 'protected' (T_PROTECTED) in /home/public_html/app/code/community/PedroTeixeira/Correios/Model/Carrier/CorreiosMethod.php on line 196
Obrigado @mateussantin
Deu certo! Vou colocar o código inteiro de como ficou o arquivo:
`
* @copyright 2015 Pedro Teixeira (http://pedroteixeira.io)
* @license http://opensource.org/licenses/MIT MIT
* @link https://github.com/pedro-teixeira/correios
*/
class PedroTeixeira_Correios_Model_Carrier_CorreiosMethod
extends Mage_Shipping_Model_Carrier_Abstract
implements Mage_Shipping_Model_Carrier_Interface
{
/**
* _code property
*
* @var string
*/
protected $_code = 'pedroteixeira_correios';
protected $_isFixed = true;
/**
* _result property
*
* @var Mage_Shipping_Model_Rate_Result|Mage_Shipping_Model_Tracking_Result
*/
protected $_result = null;
/**
* ZIP code vars
*/
protected $_fromZip = null;
protected $_toZip = null;
/**
* Value and Weight
*/
protected $_packageValue = null;
protected $_packageWeight = null;
protected $_volumeWeight = null;
protected $_freeMethodWeight = null;
protected $_midSize = null;
protected $_splitUp = 0;
protected $_postingDays = 0;
/**
* Post methods
*/
protected $_postMethods = null;
protected $_postMethodsFixed = null;
protected $_postMethodsExplode = null;
/**
* Free method request
*/
protected $_freeMethodRequest = false;
protected $_freeMethodRequestResult = null;
/**
* Collect Rates
*
* @param Mage_Shipping_Model_Rate_Request $request Mage request
*
* @return bool|Mage_Shipping_Model_Rate_Result|Mage_Shipping_Model_Tracking_Result
*/
public function collectRates(Mage_Shipping_Model_Rate_Request $request)
{
// Do initial check
if ($this->_inicialCheck($request) === false) {
return false;
}
// Check package value
if ($this->_packageValue < $this->getConfigData('min_order_value')
|| $this->_packageValue > $this->getConfigData('max_order_value')
) {
$this->_throwError('valueerror', 'Value limits', __LINE__);
return $this->_result;
}
// Check ZIP Code
if (!preg_match('/^([0-9]{8})$/', $this->_toZip)) {
$this->_throwError('zipcodeerror', 'Invalid Zip Code', __LINE__);
return $this->_result;
}
if ($this->_packageWeight == 0) {
$this->_packageWeight = $this->_getNominalWeight($request);
}
if ($this->getConfigData('weight_type') == PedroTeixeira_Correios_Model_Source_WeightType::WEIGHT_GR) {
$this->_packageWeight = number_format($this->_packageWeight / 1000, 2, '.', '');
}
// Check weight zero
if ($this->_packageWeight <= 0) {
$this->_throwError('weightzeroerror', 'Weight zero', __LINE__);
return $this->_result;
}
$this->_postMethods = $this->getConfigData('postmethods');
$this->_postMethodsFixed = $this->_postMethods;
$this->_postMethodsExplode = explode(',', $this->getConfigData('postmethods'));
// Generate Volume Weight
if ($this->_generateVolumeWeight($request) === false || $this->_removeInvalidServices() === false) {
$this->_throwError('dimensionerror', 'Dimension error', __LINE__);
return $this->_result;
}
$this->_filterMethodByItemRestriction($request);
if (empty($this->_postMethods)) {
return false;
}
//Show Quotes
$this->_getQuotes();
// Use descont codes
$this->_updateFreeMethodQuote($request);
return $this->_result;
}
/**
* Retrieve all visible items from request
*
* @param Mage_Shipping_Model_Rate_Request $request Mage request
*
* @return array
*/
protected function _getRequestItems($request)
{
$allItems = $request->getAllItems();
$items = array();
foreach ($allItems as $item) {
if (!$item->getParentItemId()) {
$items[] = $item;
}
}
$items = $this->_loadBundleChildren($items);
return $items;
}
/**
* Gets Nominal Weight
*
* @param Mage_Shipping_Model_Rate_Request $request Mage request
*
* @return number
*/
protected function _getNominalWeight($request)
{
$weight = 0;
$items = $this->_getRequestItems($request);
foreach ($items as $item) {
$product = Mage::getModel('catalog/product')->load($item->getProductId());
$weight += $product->getWeight();
}
return $weight;
}
/**
* Get shipping quote
*
* @return Mage_Shipping_Model_Rate_Result|Mage_Shipping_Model_Tracking_Result
*/
protected function _getQuotes()
{
$softErrors = explode(',', $this->getConfigData('soft_errors'));
$correiosReturn = $this->_getCorreiosReturn();
if ($correiosReturn !== false) {
$errorList = array();
$correiosReturn = $this->_addPostMethods($correiosReturn);
foreach ($correiosReturn as $servicos) {
$errorId = (string) $servicos->Erro;
$errorList[$errorId] = $servicos->MsgErro;
if ($errorId != '0' && !in_array($errorId, $softErrors)) {
continue;
}
$servicos->Valor = $this->_getFormatPrice((string) $servicos->Valor);
$this->_appendShippingReturn($servicos);
}
$this->_appendShippingErrors($errorList);
} else {
return $this->_result;
}
if ($this->_freeMethodRequest === true) {
return $this->_freeMethodRequestResult;
} else {
return $this->_result;
}
}
/**
* Make initial checks and iniciate module variables
*
* @param Mage_Shipping_Model_Rate_Request $request Mage request
*
* @return bool
*/
protected function _inicialCheck(Mage_Shipping_Model_Rate_Request $request)
{
if (!$this->getConfigFlag('active')) {
// Disabled
Mage::log('pedroteixeira_correios: Disabled');
return false;
}
$origCountry = Mage::getStoreConfig('shipping/origin/country_id', $this->getStore());
$destCountry = $request->getDestCountryId();
if ($origCountry != 'BR' || $destCountry != 'BR') {
// Out of delivery area
Mage::log('pedroteixeira_correios: Out of delivery area');
return false;
}
$this->_fromZip = Mage::getStoreConfig('shipping/origin/postcode', $this->getStore());
$this->_toZip = $request->getDestPostcode();
// Fix ZIP code
$this->_fromZip = str_replace(array('-', '.'), '', trim($this->_fromZip));
$this->_toZip = str_replace(array('-', '.'), '', trim($this->_toZip));
if (!preg_match('/^([0-9]{8})$/', $this->_fromZip)) {
Mage::log('pedroteixeira_correios: From ZIP Code Error');
return false;
}
if (!trim($this->_toZip)) {
return false;
}
$this->_result = Mage::getModel('shipping/rate_result');
$this->_packageValue = $request->getBaseCurrency()->convert(
$request->getPackageValue(),
$request->getPackageCurrency()
);
$this->_packageWeight = number_format($request->getPackageWeight(), 2, '.', '');
$this->_freeMethodWeight = number_format($request->getFreeMethodWeight(), 2, '.', '');
}
/**
* Get Correios return
*
* @return bool|SimpleXMLElement[]
*
* @throws Exception
*/
protected function _getCorreiosReturn()
{
$filename = $this->getConfigData('url_ws_correios');
try {
$client = new Zend_Http_Client($filename);
$client->setConfig(
array(
'timeout' => $this->getConfigData('ws_timeout'),
'adapter' => Mage::getModel('pedroteixeira_correios/http_client_adapter_socket')
)
);
$client->setParameterGet('StrRetorno', 'xml');
$client->setParameterGet('nCdServico', $this->_postMethods);
$client->setParameterGet('nVlPeso', $this->_packageWeight);
$client->setParameterGet('sCepOrigem', $this->_fromZip);
$client->setParameterGet('sCepDestino', $this->_toZip);
$client->setParameterGet('nCdFormato', 1);
$client->setParameterGet('nVlComprimento', $this->_midSize);
$client->setParameterGet('nVlAltura', $this->_midSize);
$client->setParameterGet('nVlLargura', $this->_midSize);
if ($this->getConfigData('mao_propria')) {
$client->setParameterGet('sCdMaoPropria', 'S');
} else {
$client->setParameterGet('sCdMaoPropria', 'N');
}
if ($this->getConfigData('aviso_recebimento')) {
$client->setParameterGet('sCdAvisoRecebimento', 'S');
} else {
$client->setParameterGet('sCdAvisoRecebimento', 'N');
}
if ($this->getConfigData('valor_declarado')
|| in_array($this->getConfigData('acobrar_code'), $this->_postMethodsExplode)
) {
$client->setParameterGet('nVlValorDeclarado', number_format($this->_packageValue, 2, ',', ''));
} else {
$client->setParameterGet('nVlValorDeclarado', 0);
}
$nCdEmpresa = $this->getConfigData('cod_admin');
$sDsSenha = $this->getConfigData('senha_admin');
if (!empty($nCdEmpresa) && !empty($sDsSenha)) {
$client->setParameterGet('nCdEmpresa', $nCdEmpresa);
$client->setParameterGet('sDsSenha', $sDsSenha);
}
$content = $client->request()->getBody();
if ($content == '') {
throw new Exception('No XML returned [' . __LINE__ . ']');
}
libxml_use_internal_errors(true);
$sxe = simplexml_load_string($content);
if (!$sxe) {
throw new Exception('Bad XML [' . __LINE__ . ']');
}
$xml = new SimpleXMLElement($content);
if (count($xml->cServico) <= 0) {
throw new Exception('No tag cServico in Correios XML [' . __LINE__ . ']');
}
return $xml->cServico;
} catch (Exception $e) {
$this->_throwError('urlerror', 'URL Error - ' . $e->getMessage(), __LINE__);
return false;
}
}
/**
* Apend shipping value to return
*
* @param SimpleXMLElement $servico Service Data
*
* @return void
*/
/* Inicio - Aumenta taxa dos correios */
protected function _generateAument(){
$items = Mage::getModel('checkout/cart')->getQuote()->getAllVisibleItems();
foreach($items as $item){
$_product = $this->_getSimpleProduct($item->getProduct());
if ($_product->getData('volume_altura') == '' || (int) $_product->getData('volume_altura') == 0) {
$itemAltura = $this->getConfigData('altura_padrao');
} else {
$itemAltura = $_product->getData('volume_altura');
}
if ($_product->getData('volume_largura') == '' || (int) $_product->getData('volume_largura') == 0) {
$itemLargura = $this->getConfigData('largura_padrao');
} else {
$itemLargura = $_product->getData('volume_largura');
}
if ($_product->getData('volume_comprimento') == '' || (int) $_product->getData('volume_comprimento') == 0) {
$itemComprimento = $this->getConfigData('comprimento_padrao');
} else {
$itemComprimento = $_product->getData('volume_comprimento');
}
}
if ($itemAltura > 70 || $itemLargura > 70 || $itemComprimento > 70) {
$this->add_aumento = 79;
} else {
$this->add_aumento;
}
return true;
}
/* Fim - Aumenta taxa dos correios */
protected function _appendShippingReturn(SimpleXMLElement $servico)
{
$correiosDelivery = (int) $servico->PrazoEntrega;
$shippingMethod = (string) $servico->Codigo;
$shippingPrice = (float) $servico->Valor;
if ($shippingPrice <= 0) {
return;
}
/* Inicio - Aumenta taxa dos correios */
$this->_generateAument();
$vl_aumento = $this->add_aumento;
$shippingPrice = $shippingPrice + $vl_aumento;
/* Fim - Aumenta taxa dos correios */
$errorMsg = $this->_getSoftErrorMsg((string) $servico->Erro);
$method = Mage::getModel('shipping/rate_result_method');
$method->setCarrier($this->_code);
$method->setCarrierTitle($this->getConfigData('title') . $this->_getSplitUpMsg() . $errorMsg);
$method->setMethod($shippingMethod);
$shippingCost = $shippingPrice;
$shippingPrice = $shippingPrice + $this->getConfigData('handling_fee');
$shippingData = Mage::helper('pedroteixeira_correios')->getShippingLabel($shippingMethod);
$shippingData = Mage::helper('pedroteixeira_correios')->__($shippingData);
if ($shippingMethod == $this->getConfigData('acobrar_code')) {
$shippingData = $shippingData . ' ( R$' . number_format($shippingPrice, 2, ',', '.') . ' )';
$shippingPrice = 0;
}
if ($this->getConfigFlag('prazo_entrega')) {
if ($correiosDelivery > 0) {
$method->setMethodTitle(
sprintf(
$this->getConfigData('msgprazo'),
$shippingData,
(int) ($correiosDelivery + $this->getConfigData('add_prazo') + $this->_postingDays)
)
);
}
} else {
$method->setMethodTitle($shippingData);
}
$method->setPrice($shippingPrice);
$method->setCost($shippingCost);
if ($this->_freeMethodRequest === true) {
$this->_freeMethodRequestResult->append($method);
} else {
$this->_result->append($method);
}
}
/**
* Throw error
*
* @param string $message Message placeholder
* @param string $log Message
* @param string|int $line Line of log
* @param string $custom Custom variables for placeholder
*
* @return void
*/
protected function _throwError($message, $log = null, $line = 'NO LINE', $custom = null)
{
$this->_result = null;
$this->_result = Mage::getModel('shipping/rate_result');
$error = Mage::getModel('shipping/rate_result_error');
$error->setCarrier($this->_code);
$error->setCarrierTitle($this->getConfigData('title'));
if (is_null($custom) || $this->getConfigData($message) == '') {
Mage::log($this->_code . ' [' . $line . ']: ' . $log);
$error->setErrorMessage($this->getConfigData($message));
} else {
Mage::log($this->_code . ' [' . $line . ']: ' . $log);
$error->setErrorMessage(sprintf($this->getConfigData($message), $custom));
}
$this->_result->append($error);
}
/**
* Retrieves a simple product
*
* @param Mage_Catalog_Model_Product $product Catalog Product
*
* @return Mage_Catalog_Model_Product
*/
protected function _getSimpleProduct($product)
{
$type = $product->getTypeInstance(true);
if ($type->getProduct($product)->hasCustomOptions()
&& ($simpleProductOption = $type->getProduct($product)->getCustomOption('simple_product'))
) {
$simpleProduct = $simpleProductOption->getProduct($product);
if ($simpleProduct) {
return $this->_getSimpleProduct($simpleProduct);
}
}
return $type->getProduct($product);
}
/**
* Generate Volume weight
*
* @param Mage_Shipping_Model_Rate_Request $request Mage request
*
* @return bool
*/
protected function _generateVolumeWeight($request)
{
$pesoCubicoTotal = 0;
$items = $this->_getRequestItems($request);
foreach ($items as $item) {
$_product = $this->_getSimpleProduct($item->getProduct());
if ($_product->getData('volume_altura') == '' || (int) $_product->getData('volume_altura') == 0) {
$itemAltura = $this->getConfigData('altura_padrao');
} else {
$itemAltura = $_product->getData('volume_altura');
}
if ($_product->getData('volume_largura') == '' || (int) $_product->getData('volume_largura') == 0) {
$itemLargura = $this->getConfigData('largura_padrao');
} else {
$itemLargura = $_product->getData('volume_largura');
}
if ($_product->getData('volume_comprimento') == '' || (int) $_product->getData('volume_comprimento') == 0) {
$itemComprimento = $this->getConfigData('comprimento_padrao');
} else {
$itemComprimento = $_product->getData('volume_comprimento');
}
if ($this->getConfigFlag('check_dimensions')) {
foreach ($this->_postMethodsExplode as $key => $method) {
$sizeMax = max($itemAltura, $itemLargura, $itemComprimento);
$sumMax = ($itemAltura + $itemLargura + $itemComprimento);
$isValid = ($sizeMax <= $this->getConfigData("validate/serv_{$method}/max/size"));
$isValid &= ($sumMax <= $this->getConfigData("validate/serv_{$method}/max/sum"));
if (!$isValid) {
unset($this->_postMethodsExplode[$key]);
}
}
if (count($this->_postMethodsExplode) == 0) {
return false;
}
$this->_postMethods = implode(',', $this->_postMethodsExplode);
$this->_postMethodsFixed = $this->_postMethods;
}
$itemAltura = $this->_getFitHeight($item);
$pesoCubicoTotal += (($itemAltura * $itemLargura * $itemComprimento) *
$item->getTotalQty()) / $this->getConfigData('coeficiente_volume');
$this->_postingDays = max($this->_postingDays, (int) $_product->getData('posting_days'));
}
$this->_volumeWeight = number_format($pesoCubicoTotal, 2, '.', '');
return true;
}
/**
* Generate free shipping for a product
*
* @param string $freeMethod Free method
*
* @return void
*/
protected function _setFreeMethodRequest($freeMethod)
{
$this->_freeMethodRequest = true;
$this->_freeMethodRequestResult = Mage::getModel('shipping/rate_result');
$this->_postMethods = $freeMethod;
$this->_postMethodsExplode = array($freeMethod);
if ($this->getConfigData('weight_type') == PedroTeixeira_Correios_Model_Source_WeightType::WEIGHT_GR) {
$this->_freeMethodWeight = number_format($this->_freeMethodWeight / 1000, 2, '.', '');
}
$this->_packageWeight = $this->_freeMethodWeight;
$this->_pacWeight = $this->_freeMethodWeight;
}
/**
* Check if current carrier offer support to tracking
*
* @return bool true
*/
public function isTrackingAvailable()
{
return true;
}
/**
* Get Tracking Info
*
* @param mixed $tracking Tracking
*
* @return mixed
*/
public function getTrackingInfo($tracking)
{
$result = $this->getTracking($tracking);
if ($result instanceof Mage_Shipping_Model_Tracking_Result) {
if ($trackings = $result->getAllTrackings()) {
return $trackings[0];
}
} elseif (is_string($result) && !empty($result)) {
return $result;
}
return false;
}
/**
* Get Tracking
*
* @param array $trackings Trackings
*
* @return Mage_Shipping_Model_Tracking_Result
*/
public function getTracking($trackings)
{
$this->_result = Mage::getModel('shipping/tracking_result');
foreach ((array) $trackings as $code) {
$this->_getTracking($code);
}
return $this->_result;
}
/**
* Loads the parameters and calls the webservice using SOAP
*
* @param string $code Code
*
* @return bool|array
*
* @throws Exception
*/
protected function _getTrackingRequest($code)
{
$response = false;
$params = array(
'usuario' => $this->getConfigData('sro_username'),
'senha' => $this->getConfigData('sro_password'),
'tipo' => $this->getConfigData('sro_type'),
'resultado' => 'T',
'lingua' => $this->getConfigData('sro_language'),
'objetos' => $code,
);
try {
$client = new SoapClient(
$this->getConfigData('url_sro_correios'), Mage::helper('pedroteixeira_correios')->getStreamContext()
);
$response = $client->buscaEventos($params);
if (empty($response)) {
throw new Exception("Empty response");
}
} catch (Exception $e) {
Mage::log("Soap Error: {$e->getMessage()}");
}
return $response;
}
/**
* Loads tracking progress details
*
* @param SimpleXMLElement $evento XML Element Node
* @param bool $isDelivered Delivery Flag
*
* @return array
*/
protected function _getTrackingProgressDetails($evento, $isDelivered = false)
{
$date = new Zend_Date($evento->data, 'dd/MM/YYYY', new Zend_Locale('pt_BR'));
$track = array(
'deliverydate' => $date->toString('YYYY-MM-dd'),
'deliverytime' => $evento->hora . ':00',
'status' => $evento->descricao,
);
if (!$isDelivered) {
$msg = array($evento->descricao);
if (isset($evento->destino) && isset($evento->destino->local)) {
$msg = array("{$evento->descricao} para {$evento->destino->local}");
}
$track['activity'] = implode(' | ', $msg);
$track['deliverylocation'] = "{$evento->local} - {$evento->cidade}/{$evento->uf}";
}
return $track;
}
/**
* Loads progress data using the WSDL response
*
* @param string $request Request response
*
* @return array
*/
protected function _getTrackingProgress($request)
{
$track = array();
$progress = array();
$eventTypes = explode(',', $this->getConfigData("sro_event_type_last"));
if (count($request->return->objeto->evento) == 1) {
$progress[] = $this->_getTrackingProgressDetails($request->return->objeto->evento);
} else {
foreach ($request->return->objeto->evento as $evento) {
$progress[] = $this->_getTrackingProgressDetails($evento);
$isDelivered = ((int) $evento->status < 2 && in_array($evento->tipo, $eventTypes));
if ($isDelivered) {
$track = $this->_getTrackingProgressDetails($evento, $isDelivered);
}
}
}
$progress[] = $track;
return $progress;
}
/**
* Protected Get Tracking, opens the request to Correios
*
* @param string $code Code
*
* @return bool
*/
protected function _getTracking($code)
{
$error = Mage::getModel('shipping/tracking_result_error');
$error->setTracking($code);
$error->setCarrier($this->_code);
$error->setCarrierTitle($this->getConfigData('title'));
$error->setErrorMessage($this->getConfigData('urlerror'));
$request = $this->_getTrackingRequest($code);
if (!isset($request->return)) {
$this->_result->append($error);
return false;
}
$progress = $this->_getTrackingProgress($request);
if (!empty($progress)) {
$track = array_pop($progress);
$track['progressdetail'] = $progress;
$tracking = Mage::getModel('shipping/tracking_result_status');
$tracking->setTracking($code);
$tracking->setCarrier($this->_code);
$tracking->setCarrierTitle($this->getConfigData('title'));
$tracking->addData($track);
$this->_result->append($tracking);
return true;
} else {
$this->_result->append($error);
return false;
}
}
/**
* Returns the allowed carrier methods
*
* @return array
*/
public function getAllowedMethods()
{
$output = array($this->_code => $this->getConfigData('title'));
$serviceObject = Mage::getSingleton('pedroteixeira_correios/postmethod');
foreach ($serviceObject->getCollection() as $service) {
$output[ $service->getMethodCode() ] = "{$service->getMethodCode()} - {$service->getMethodTitle()}";
}
return $output;
}
/**
* Define ZIP Code as required
*
* @param string $countryId Country ID
*
* @return bool
*/
public function isZipCodeRequired($countryId = null)
{
return true;
}
/**
* Retrieve an average size.
* For optimization purposes all tree box sizes are converted in one medium dimension.
* Result cant exceed the minimum transportation limits.
*
* @return PedroTeixeira_Correios_Model_Carrier_CorreiosMethod
*/
protected function _loadMidSize()
{
$volumeFactor = $this->getConfigData('coeficiente_volume');
$volumeTotal = $this->_volumeWeight * $volumeFactor;
$pow = round(pow((int) $volumeTotal, (1 / 3)));
$min = $this->getConfigData('midsize_min');
$this->_midSize = max($pow, $min);
return $this;
}
/**
* Validate post methods removing invalid services from quotation.
*
* @return boolean|PedroTeixeira_Correios_Model_Carrier_CorreiosMethod
*/
protected function _removeInvalidServices()
{
$tmpMethods = $this->_postMethodsExplode;
$tmpMethods = $this->_filterMethodByConfigRestriction($tmpMethods);
$isDivisible = (count($tmpMethods) == 0);
if ($isDivisible) {
return $this->_splitPack();
}
$this->_postMethodsExplode = $tmpMethods;
$this->_postMethods = implode(',', $this->_postMethodsExplode);
$this->_postMethodsFixed = $this->_postMethods;
return $this;
}
/**
* Include an additional method to quote content before showing.
* When requested the new method is added in xml content as specified in config.xml like below:
*
*
* 10065
* 2.45
* 5
*
* 00000000
* 0.0
* 0
*
*
* 99999999
* 0.1
* 150
*
*
*
* @param SimpleXMLElement $cServico XML Node
*
* @see http://www.correios.com.br/para-voce/consultas-e-solicitacoes/precos-e-prazos/servicos-nacionais_pasta/carta
*
* @return SimpleXMLElement
*/
protected function _addPostMethods($cServico)
{
$addMethods = $this->getConfigData("add_postmethods");
if (empty($addMethods) || !is_array($addMethods)) {
return $cServico;
}
foreach ($addMethods as $configData) {
$isValid = true;
$isValid &= $this->_packageWeight >= $configData['from']['weight'];
$isValid &= $this->_packageWeight <= $configData['to']['weight'];
$isValid &= $this->_midSize >= $configData['from']['size'];
$isValid &= $this->_midSize <= $configData['to']['size'];
$isValid &= $this->_toZip >= $configData['from']['zip'];
$isValid &= $this->_toZip <= $configData['to']['zip'];
if ($isValid) {
$price = $configData['price'];
$days = $configData['days'];
$method = $configData['code'];
foreach ($cServico as $servico) {
if ($servico->Codigo == $method) {
if (!empty($price)) {
$servico->Valor = number_format($price, 2, ',', '');
}
if (!empty($days)) {
$servico->PrazoEntrega = $days;
}
$servico->EntregaDomiciliar = 'S';
$servico->EntregaSabado = 'S';
$servico->Erro = '0';
$servico->MsgErro = ' ';
}
}
}
}
return $cServico;
}
/**
* This keeps only postmethods available for all items in cart.
* In other words you can set post methods by products.
* Methods not available for all items in cart are removed.
* Require attribute creation called postmethods.
* Example:
* code: postmethods
* type: multiselect
* label: [free]
* value 1: 41068
* value 2: 40096
* ...
* value 99: 81019
*
* @param Mage_Shipping_Model_Rate_Request $request Mage request
*
* @return PedroTeixeira_Correios_Model_Carrier_CorreiosMethod
*/
protected function _filterMethodByItemRestriction($request)
{
if ($this->getConfigFlag('filter_by_item')) {
$items = $this->_getRequestItems($request);
$intersection = $this->_postMethodsExplode;
foreach ($items as $item) {
$product = $this->_getSimpleProduct($item->getProduct());
$prodPostMethods = explode(
',', $product->getResource()->getAttributeRawValue(
$product->getId(), 'postmethods', $request->getStoreId()
)
);
$intersection = array_intersect($prodPostMethods, $intersection);
}
$this->_postMethodsExplode = $intersection;
$this->_postMethods = implode(',', $intersection);
$this->_postMethodsFixed = $this->_postMethods;
}
return $this;
}
/**
* Added a fit size for items in large quantities.
* Means you can join items like two or more glasses, pots and vases.
* The calc is applied only for height side.
* Required attribute fit_size. Example:
*
* code: fit_size
* type: varchar
*
* After you can set a fit size for all products and improve your sells
*
* @param Mage_Eav_Model_Entity_Abstract $item Order Item
*
* @return number
*/
protected function _getFitHeight($item)
{
$product = $this->_getSimpleProduct($item->getProduct());
$height = $product->getData('volume_altura');
$height = ($height > 0) ? $height : (int) $this->getConfigData('altura_padrao');
$fitSize = (float) $product->getData('fit_size');
if ($item->getQty() > 1 && is_numeric($fitSize) && $fitSize > 0) {
$totalSize = $height + ($fitSize * ($item->getQty() - 1));
$height = $totalSize / $item->getQty();
}
return $height;
}
/**
* Splits the package in two parts.
* If the package is already splited, each piece will be splited in two equal parts.
*
* @return boolean|PedroTeixeira_Correios_Model_Carrier_CorreiosMethod
*/
protected function _splitPack()
{
$isSplitEnabled = $this->getConfigFlag('split_pack');
$isMethodAvailable = (count($this->_postMethodsExplode) > 0);
if ($isSplitEnabled && $isMethodAvailable) {
$this->_splitUp++;
$this->_volumeWeight /= 2;
$this->_packageWeight /= 2;
$this->_packageValue /= 2;
return $this->_removeInvalidServices();
}
return false;
}
/**
* Receive a list of methods, and validate one-by-one using the config settings.
* Returns a list of valid methods or empty.
*
* @param array $postmethods Services List
*
* @return array
*/
protected function _filterMethodByConfigRestriction($postmethods)
{
$validMethods = array();
$this->_loadMidSize();
foreach ($postmethods as $key => $method) {
$isOverSize = ($this->_midSize > $this->getConfigData("validate/serv_{$method}/max/size"));
$isOverSize |= ($this->_midSize * 3 > $this->getConfigData("validate/serv_{$method}/max/sum"));
$isOverWeight = ($this->_packageWeight > $this->getConfigData("validate/serv_{$method}/max/weight"));
$isOverCubic = ($this->_volumeWeight > $this->getConfigData("validate/serv_{$method}/max/volume_weight"));
$isZipAllowed = $this->_validateZipRestriction($method);
if (!$isOverSize && !$isOverWeight && !$isOverCubic && $isZipAllowed) {
$validMethods[] = $method;
}
}
return $validMethods;
}
/**
* Loads the zip range list.
* Returns TRUE only if zip target is included in the range.
*
* @param array $method Current Post Method
*
* @return boolean
*/
protected function _validateZipRestriction($method)
{
$zipConfig = $this->getConfigData("validate/serv_{$method}/zips");
foreach ($zipConfig as $data) {
$zipRange = explode(',', $data);
$isBetweenRange = true;
$isBetweenRange &= ($this->_toZip >= $zipRange[0]);
$isBetweenRange &= ($this->_toZip <= $zipRange[1]);
if ($isBetweenRange) {
return true;
}
}
return false;
}
/**
* Some special errors must be sent to users.
* If not applicable, the default error will be sent.
*
* @param array $errorList Error List
*
* @return boolean
*/
protected function _appendShippingErrors($errorList)
{
$output = false;
$successCode = '0';
$hasValidQuote = array_key_exists($successCode, $errorList);
if (!$hasValidQuote) {
$displayErrorList = explode(',', $this->getConfigData('hard_errors'));
if ($this->getConfigFlag('show_soft_errors')) {
$softErrorList = explode(',', $this->getConfigData('soft_errors'));
$displayErrorList = array_merge($displayErrorList, $softErrorList);
}
foreach ($errorList as $errorCode => $errorMsg) {
$isDisplayError = in_array($errorCode, $displayErrorList);
if ($isDisplayError) {
$error = Mage::getModel('shipping/rate_result_error');
$error->setCarrier($this->_code);
$error->setErrorMessage($errorMsg);
$this->_result->append($error);
$output = true;
}
}
if (!$output) {
$logMsg = implode(',', $errorList);
Mage::log("{$this->_code}: Warning! There is no valid quotes, and no one error was throwed: {$logMsg}");
}
}
return $output;
}
/**
* Returns a short message showing the number of the packs that will be needed.
*
* @return string
*/
protected function _getSplitUpMsg()
{
$msg = "";
if ($this->_splitUp > 0) {
$qty = pow(2, $this->_splitUp);
$msg.= " / {$qty} volumes";
}
return $msg;
}
/**
* Returns a short warning message.
*
* @param string $error Error Id
*
* @return string
*/
protected function _getSoftErrorMsg($error)
{
$msg = "";
if ($this->getConfigFlag('show_soft_errors')) {
$softErrorList = explode(',', $this->getConfigData('soft_errors'));
$isSoftError = in_array($error, $softErrorList);
if ($isSoftError) {
$msg.= " / Área de Risco";
}
}
return $msg;
}
/**
* Returns the price as float, and fixed by pack division.
*
* @param string $price Price String
*
* @return float
*/
protected function _getFormatPrice($price)
{
$stringPrice = str_replace('.', '', $price);
$stringPrice = str_replace(',', '.', $stringPrice);
$shippingPrice = floatval($stringPrice);
$shippingPrice *= pow(2, $this->_splitUp);
return $shippingPrice;
}
/**
* Filter visible and bundle children products.
*
* @param array $items Product Items
*
* @return array
*/
protected function _loadBundleChildren($items)
{
$visibleAndBundleChildren = array();
/* @var $item Mage_Sales_Model_Quote_Item */
foreach ($items as $item) {
$product = $item->getProduct();
$isBundle = ($product->getTypeId() == Mage_Catalog_Model_Product_Type::TYPE_BUNDLE);
if ($isBundle) {
/* @var $child Mage_Sales_Model_Quote_Item */
foreach ($item->getChildren() as $child) {
$visibleAndBundleChildren[] = $child;
}
} else {
$visibleAndBundleChildren[] = $item;
}
}
return $visibleAndBundleChildren;
}
}
`
Obrigado @fernandoplanejar91
Consegui fazer funcionar com a modificação inteira que você enviou.
E muito obrigado @mateussantin pela solução!!!
Olá @fernandoplanejar91 e @mateussantin, tentei colar o código acima e não deu certo. Vocês tem como enviar o arquivo por gentileza?
Obrigado pela ajuda.
Segue arquivos em anexo corrigidos com a opção no admin para atualizar o valor dos correios sempre que necessário (atualmente 79 reais).
(apagar o .txt do final do arquivo e substituir).
ect/system.xml
model/carrier/correiosMethod.php
CorreiosMethod.php.txt
system.xml.txt
Segue arquivos em anexo corrigidos com a opção no admin para atualizar o valor dos correios sempre que necessário (atualmente 79 reais).
(apagar o .txt do final do arquivo e substituir).
ect/system.xml
model/carrier/correiosMethod.php
CorreiosMethod.php.txt
system.xml.txt
@mateussantin, muito obrigado pela disposição em ajudar. Funcionou corretamente.
Abs,
Olá pessoal,
Encontrei um erro, quando você compra somente 1 produto na medida por exemplo 105x3x3 é calculado o aumento de R$ 79,00 e o frete fica R$ 124,50, OK, até ai funcionando.
Mas se eu adicionar também no carrinho um produto menor que 70cm, por exemplo 16x2x11 o frete fica R$ 47,80, ou seja, quando tem mais de um produto no carrinho ele está ignorando que um deles é maior que 70cm.
Por favor alguém sabe uma solução para esse caso?
Abraços!
Também estou com esse problema, alguém conseguiu resolver???
Olá pessoal,
Encontrei um erro, quando você compra somente 1 produto na medida por exemplo 105x3x3 é calculado o aumento de R$ 79,00 e o frete fica R$ 124,50, OK, até ai funcionando.
Mas se eu adicionar também no carrinho um produto menor que 70cm, por exemplo 16x2x11 o frete fica R$ 47,80, ou seja, quando tem mais de um produto no carrinho ele está ignorando que um deles é maior que 70cm.
Por favor alguém sabe uma solução para esse caso?
Abraços!
Bom dia @ricardopi70
Essa alteração reajusta o valor da taxa de manipulação, caso algum dos itens do carrinho exceda 70 cm.
Localize o trecho a seguir:
https://github.com/pedro-teixeira/correios/blob/2c828ac48627de551b7faed83f22f814b5c8bdef/app/code/community/PedroTeixeira/Correios/Model/Carrier/CorreiosMethod.php#L486
...e cole esse código, exatamente após o trecho localizado.
if (empty($this->getIsCorreiosHandlingFixed()) && $sizeMax > 70) {
$correiosFee = 20;
$handlingFee = $this->getConfigData('handling_fee');
$path = "carriers/{$this->getCarrierCode()}/handling_fee";
Mage::app()->getStore($this->getStore())->setConfig($path, $correiosFee + $handlingFee);
$this->setIsCorreiosHandlingFixed(true);
}
P.S.
Habilitar opção de Validar Dimensões dos Produtos
Não compatível com as modificações previamente sugeridas neste tópico. Aplicar este ajuste nos arquivos originais do módulo.
Obrigado @rafaelpatro, está funcionando corretamente agora. Valeu!
Para mim não atualizou nada, mesmo baixando e instalando esses arquivos
Segue arquivos em anexo corrigidos com a opção no admin para atualizar o valor dos correios sempre que necessário (atualmente 79 reais).
(apagar o .txt do final do arquivo e substituir).
ect/system.xml
model/carrier/correiosMethod.php
CorreiosMethod.php.txt
system.xml.txt
Olá boa noite
Gostaria saber como o Pedro Teixeira chama os atributos de altura largura e profundidade?
Segue a imagem de exemplo como a Jadlog chama
Obrigada pela atenção
Malka Gruenberg
Bom dia @ricardopi70
Essa alteração reajusta o valor da taxa de manipulação, caso algum dos itens do carrinho exceda 70 cm.
Localize o trecho a seguir:
https://github.com/pedro-teixeira/correios/blob/2c828ac48627de551b7faed83f22f814b5c8bdef/app/code/community/PedroTeixeira/Correios/Model/Carrier/CorreiosMethod.php#L486
...e cole esse código, exatamente após o trecho localizado.
if (empty($this->getIsCorreiosHandlingFixed()) && $sizeMax > 70) {
$correiosFee = 20;
$handlingFee = $this->getConfigData('handling_fee');
$path = "carriers/{$this->getCarrierCode()}/handling_fee";
Mage::app()->getStore($this->getStore())->setConfig($path, $correiosFee + $handlingFee);
$this->setIsCorreiosHandlingFixed(true);
}
P.S.
Habilitar opção de Validar Dimensões dos Produtos
Não compatível com as modificações previamente sugeridas neste tópico. Aplicar este ajuste nos arquivos originais do módulo.
Olá a todos, agradeço a colaboração do pessoal. Ficou muito boa a solução. Mesmo com a correção do @ricardopi70, tive problemas ainda quando eu coloco apenas 1 produto no carrinho com tamanho acima de 70 ele estava adicionando 2 vezes o valor do manuseio. Fiz de uma outra forma e estou colocando aqui para vocês, caso mais alguém esteja com o mesmo problema.
public $check_size = 0;
protected function _generateAument()
{
$items = Mage::getModel('checkout/cart')->getQuote()->getAllVisibleItems();
foreach($items as $item){
$_product = $this->_getSimpleProduct($item->getProduct());
if ($_product->getData('volume_altura') == '' || (int) $_product->getData('volume_altura') == 0) {
$itemAltura = $this->getConfigData('altura_padrao');
} else {
$itemAltura = $_product->getData('volume_altura');
}
if ($_product->getData('volume_largura') == '' || (int) $_product->getData('volume_largura') == 0) {
$itemLargura = $this->getConfigData('largura_padrao');
} else {
$itemLargura = $_product->getData('volume_largura');
}
if ($_product->getData('volume_comprimento') == '' || (int) $_product->getData('volume_comprimento') == 0) {
$itemComprimento = $this->getConfigData('comprimento_padrao');
} else {
$itemComprimento = $_product->getData('volume_comprimento');
}
if ($itemAltura > 70 || $itemLargura > 70 || $itemComprimento > 70) {
$this->check_size = 1;
}
}
if ($this->check_size == 1) {
$this->add_aumento = $this->getConfigData('aument');
} else {
$this->add_aumento;
}
return true;
}
|
gharchive/issue
| 2018-09-05T17:53:40 |
2025-04-01T04:35:28.845361
|
{
"authors": [
"cesaroito",
"fernandoplanejar91",
"gilliattbastos",
"malkahakuk",
"mateussantin",
"rafaelpatro",
"ricardopi70"
],
"repo": "pedro-teixeira/correios",
"url": "https://github.com/pedro-teixeira/correios/issues/393",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
867053084
|
RSTP Server?
anyone have any recommendation on using online rtsp server? thanks
I'm currently using rtmp with Youtube and it works great.
You have a list of servers in the wiki page:
https://github.com/pedroSG94/rtmp-rtsp-stream-client-java/wiki/Media-servers-tested
|
gharchive/issue
| 2021-04-25T16:28:35 |
2025-04-01T04:35:28.852909
|
{
"authors": [
"ToysoftInc",
"pedroSG94"
],
"repo": "pedroSG94/rtmp-rtsp-stream-client-java",
"url": "https://github.com/pedroSG94/rtmp-rtsp-stream-client-java/issues/846",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
115136395
|
Optional custom top-level domains
Hello!
I've added opt-in support for custom (non-standard) top-level domains, primarily for supporting use of parse-domains on local environments, where the TLD could feasibly be set to anything.
The aim was to implement this in a way that hopefully eases factoring between development and production builds, via the parseDomain.applyCustomTlds() method – it returns a function that matches the original signature of the parse-domains module, so it can feasibly be exported in a custom app-specific module and used in development builds in place of the "vanilla" version of parse-domains. 😄
Thx @mattlubner. Sorry for not responding, I get too much notifications and it's hard to filter the important ones (like PRs).
I really like your suggestion and I see the necessity to define custom TLDs. :+1:
However, I am not quite happy with the proposed API. I usually try to keep the API surface as small as possible. Your proposed applyCustomTlds() introduces a whole new way to configure the module, which I believe is not necessary for most use-cases. I also don't like to add an optional argument customTlds because I always stumble upon these arguments when reading the source code weeks/years later again ("what was this array for? ... ahh these are custom tlds").
What about passing an options object as second argument? Like
parseDomain("some.domain.local", {
customTlds: ["local"]
})
then you could still make your own version of the function
function parseCustomTlds(url) {
return parseDomain(url, {
customTlds: ["local"]
});
}
Additionally I would not accept either a string or an array. Imho it's best to avoid ambiguity with options (unless it provides a huge usability benefit). Since it's a list of things, arrays are the best type for this option. What do you think?
Apart from that, how does your implementation handle custom tlds like "dev.local"? :grinning: Could you add tests for these cases?
Thx @mattlubner. Sorry for not responding, I get too much notifications and it's hard to filter the important ones (like PRs).
I really like your suggestion and I see the necessity to define custom TLDs. :+1:
However, I am not quite happy with the proposed API. I usually try to keep the API surface as small as possible. Your proposed applyCustomTlds() introduces a whole new way to configure the module, which I believe is not necessary for most use-cases. I also don't like to add an optional argument customTlds because I always stumble upon these arguments when reading the source code weeks/years later again ("what was this array for? ... ahh these are custom tlds").
What about passing an options object as second argument? Like
parseDomain("some.domain.local", {
customTlds: ["local"]
})
then you could still make your own version of the function
function parseCustomTlds(url) {
return parseDomain(url, {
customTlds: ["local"]
});
}
Additionally I would not accept either a string or an array. Imho it's best to avoid ambiguity with options (unless it provides a huge usability benefit). Since it's a list of things, arrays are the best type for this option. What do you think?
Apart from that, how does your implementation handle custom tlds like "dev.local"? :grinning: Could you add tests for these cases?
Wonderful suggestions! I also added a custom RegExp pass-thru as well, to cut down on repeated hydration of new RegExp objects (eg, for use within a server). Am cleaning up the README.md a bit right now… 😄
@jhnns Alrighty, I've tidied up this PR; hopefully it's ready to merge in! No biggie if you have some more feedback though, I'm happy to oblige.
I don't at all mind removing the partial application helper – it's trivial to move that code into consuming modules, and you know, it actually looks a bit more efficient like this! 😉
Btw, I've never come across README.md syntax like this before. I tidied it up a bit, but apologies if it's auto-generated and I subsequently made things difficult for ya… 😄
@matthaias thanks for your effort. That looks good now.
There's just one thing that bothers me: I think it would be better to allow the config option customTlds to be either a regexp or an array. The current implementation allows to specify both at the same time which may confuse users (which option takes precedence?).
Thanks for correcting grammar mistakes in the README. Regarding the order, I prefer examples before any installation instructions :grin:. But you can just leave the README, I'll adjust it then.
@jhnns Glad to help out! 😎
I'm actually not too keen on options.customTldsRegExp either. It seemed like the best way to reduce RegExp constructions, although if you aren't against the library overwriting options.customTlds with the constructed RegExp object, that definitely simplifies the API!
The current implementation is good, thank you :+1:
Shipped with 0.2.1
|
gharchive/pull-request
| 2015-11-04T20:07:56 |
2025-04-01T04:35:28.889183
|
{
"authors": [
"jhnns",
"mattlubner"
],
"repo": "peerigon/parse-domain",
"url": "https://github.com/peerigon/parse-domain/pull/3",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2080338954
|
Simplifying Setup
As of now, the repo is not very well organized to get started with running the experiments. Makes it difficult for a new user to come to the repo and start scaling experiments. The aim here is to make it easy for any user to come and start running experiments as easily as possible.
[ ] run-factorizer-enc script, has hard-coded values that need to be changed when running the script.
[ ] Less information on how to get the data and factorizer models and where to place them.
[ ] Setting up the python environment for running the experiment. Seems like conda works a bit differently than venv and it creates confusion
Modifiable parts for any experiment :
Config files : base.prep.yml and base.conf.yml. Whenever, we will have to change the language pair, this needs to be changed.
Bash Script Files : run-factorizer-enc.sh. Here, we have hardcoded parameter values which change from experiment to experiment. They change more frequently.
Getting Datasets
Although, it is easy to get the files and tokenize them using mosestokenizer, but for different language pair tokenization can vary. Can \ Shall we automate this process ?
As of now, here are the things I am working on to make this process easier :
[ ] Add a script to get the factorizer models and setting up the datasets directory ( if feasible )
[ ] Removing hardcoded parameters from bash run scripts
[ ] Minimize the number of parameters, using repo_setup.env file ( stores the variables like, dataset setup dir, factorizer model dir, exp collection dir, config store dir )
[ ] (Maybe a script to generate base conf files for different language pairs )
|
gharchive/issue
| 2024-01-13T15:12:25 |
2025-04-01T04:35:28.925116
|
{
"authors": [
"pegasus-lynx"
],
"repo": "pegasus-lynx/bytetok-nmt",
"url": "https://github.com/pegasus-lynx/bytetok-nmt/issues/1",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
603143123
|
When workable version of code will be uploaded?
When workable version of code will be uploaded?
Hi, I will have some time working on it in the next few days and hopefully will release a workable version sometime in the next week.
@peiyunh Any updates on workable version?
|
gharchive/issue
| 2020-04-20T10:53:50 |
2025-04-01T04:35:28.973511
|
{
"authors": [
"chowkamlee81",
"peiyunh"
],
"repo": "peiyunh/opcseg",
"url": "https://github.com/peiyunh/opcseg/issues/1",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
2032919558
|
ImageCLEFmedCaption_2022
Please tell me how to obtain the ImageCLEFmedCaption_2022 data set. It seems that the clf official website cannot download this data set now. Thank you!
Please tell me how to obtain the ImageCLEFmedCaption_2022 data set. It seems that the clf official website cannot download this data set now. Thank you!
Do you have a download method for this dataset?Thank you
|
gharchive/issue
| 2023-12-08T15:56:31 |
2025-04-01T04:35:29.059675
|
{
"authors": [
"deng1604",
"tntdc585"
],
"repo": "pengfeiliHEU/M2I2",
"url": "https://github.com/pengfeiliHEU/M2I2/issues/11",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
740676055
|
Property 'yaml' does not exist on type 'typeof languages'
i'm using ngx-monaco-editor (https://github.com/atularen/ngx-monaco-editor#readme) library in my Angular project. also i would like to use monaco-yaml:
monaco.languages.yaml.yamlDefaults.setDiagnosticsOptions({ ... })
for some reason i get following error:
error TS2339: Property 'yaml' does not exist on type 'typeof languages'.
i cannot understand, what's the problem: ngx-monaco-editor already contains monaco-editor lib, and monaco.languages is defined, but there is no yaml property. that's kinda strange for me, since in monaco-yaml is declared same monaco namespace with property yaml, so it's global scope, but my code see only monaco namespase properties defined in ngx-monaco-editor
i'll be happy if anyone has any ideas about what could that be
I have the same issue. I am not sure why typescript can't find the Monaco.d.ts. Can you for now ignore that by // @ts-ignore until we find a solution for this.
|
gharchive/issue
| 2020-11-11T11:01:37 |
2025-04-01T04:35:29.064329
|
{
"authors": [
"mraimon",
"yazaabed"
],
"repo": "pengx17/monaco-yaml",
"url": "https://github.com/pengx17/monaco-yaml/issues/37",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
55009681
|
[PMD-994] - Add Redshift dialect to Pentaho Metadata
add Redshift to kettle-database-types.xml
@mbatchelor @lucboudreau this fixes failing projects that use metadata(like PIR)
Do we need this as a native plugin,r would it work as a PDI Database plugin? As a rule I'd like to kethe list of native types small if possible, as the default (built-in) list is quite long already.
If this needs to be a native plugin then that's fine, just wondering.
Matt, I've merged this one because it is a followup of something previously checked in.
|
gharchive/pull-request
| 2015-01-21T11:58:50 |
2025-04-01T04:35:29.099204
|
{
"authors": [
"aliakseihaidukou",
"lucboudreau",
"mattyb149"
],
"repo": "pentaho/pentaho-kettle",
"url": "https://github.com/pentaho/pentaho-kettle/pull/965",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2047416728
|
Endpoints and client work for Search #10
PR fills out the skeleton for #10.
In progress, but tasks tackled thus far:
Added basic client route for search results, /search/[query]
SearchBar now redirects to /search/[query]
updating /api/search to return result oriented data
updating validators for this data + checking query results
~remaining TODO is basic search result table.~
basic search result table implemented now, too.
|
gharchive/pull-request
| 2023-12-18T20:40:51 |
2025-04-01T04:35:29.103754
|
{
"authors": [
"ejmg"
],
"repo": "penumbra-zone/cuiloa",
"url": "https://github.com/penumbra-zone/cuiloa/pull/44",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2195963207
|
cnidarium: implement incremental jmt migration
Close #3506, WIP
rustc segfault? 😈
Yeah for sure, waiting for CI to pass
|
gharchive/pull-request
| 2024-03-19T20:39:08 |
2025-04-01T04:35:29.104815
|
{
"authors": [
"erwanor"
],
"repo": "penumbra-zone/penumbra",
"url": "https://github.com/penumbra-zone/penumbra/pull/4053",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
162867779
|
Wrong warning for public imports
Just like #381 this produces a warning one cannot get rid of.
package T1
public class A
package T2
import public T1
package T3
import T2 <-- The import T2 is never used directly
class B extends A
When removed this of course causes a compilation error.
Well, you can get rid of it by importing T1 instead of T2. Why do you want to import T2, if you are not using anything from T2 directly?
because I have many public imports in T1 and many in T2 that are used in many other package where I want to keep the import list small by moving them into public imports.
fixed and now a warning shows you excessive imports - nice.
|
gharchive/issue
| 2016-06-29T08:25:32 |
2025-04-01T04:35:29.140653
|
{
"authors": [
"Frotty",
"peq"
],
"repo": "peq/WurstScript",
"url": "https://github.com/peq/WurstScript/issues/449",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1581407566
|
[Question] Why PerAct couldn't solve "Push Buttons" (compares to HiveFormer)?
Dear author,
Thank you for sharing your awesome work!
In the readme, you mention that
" Impossible tasks: Some tasks like push_buttons are not solvable by PerAct since it doesn't have any memory."
Is it because the PerAct's policy is P(a_t | O_t, Instruction) ? Since history observation is not provided, there's no way for PerAct to keep track of its progress for Instruction following ?
Best,
@berniewang8177, yes. If the task is "press the red button, then the blue button, and then the green button", the agent won't know which button was pressed previously, and will have to guess which part of the sequence it's at.
|
gharchive/issue
| 2023-02-12T20:58:56 |
2025-04-01T04:35:29.142890
|
{
"authors": [
"MohitShridhar",
"berniewang8177"
],
"repo": "peract/peract",
"url": "https://github.com/peract/peract/issues/14",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2568211603
|
K8SPG-648 add PG17 support
CHANGE DESCRIPTION
Problem:
Short explanation of the problem.
Cause:
Short explanation of the root cause of the issue if applicable.
Solution:
Short explanation of the solution we are providing with this PR.
CHECKLIST
Jira
[ ] Is the Jira ticket created and referenced properly?
[ ] Does the Jira ticket have the proper statuses for documentation (Needs Doc) and QA (Needs QA)?
[ ] Does the Jira ticket link to the proper milestone (Fix Version field)?
Tests
[ ] Is an E2E test/test case added for the new feature/change?
[ ] Are unit tests added where appropriate?
Config/Logging/Testability
[ ] Are all needed new/changed options added to default YAML files?
[ ] Did we add proper logging messages for operator actions?
[ ] Did we ensure compatibility with the previous version or cluster upgrade process?
[ ] Does the change support oldest and newest supported PG version?
[ ] Does the change support oldest and newest supported Kubernetes version?
Test name
Status
custom-extensions
passed
custom-tls
passed
demand-backup
passed
finalizers
passed
init-deploy
passed
major-upgrade
failure
monitoring
passed
one-pod
passed
operator-self-healing
passed
pitr
passed
scaling
passed
scheduled-backup
passed
self-healing
passed
start-from-backup
passed
tablespaces
passed
telemetry-transfer
passed
upgrade-consistency
passed
upgrade-minor
passed
users
passed
We run 19 out of 19
commit: https://github.com/percona/percona-postgresql-operator/pull/921/commits/f651334605a27f83266361e92758941f5320c5a5
image: perconalab/percona-postgresql-operator:PR-921-f65133460
Test name
Status
custom-extensions
passed
custom-tls
passed
demand-backup
passed
finalizers
passed
init-deploy
passed
major-upgrade
failure
monitoring
passed
one-pod
passed
operator-self-healing
passed
pitr
passed
scaling
passed
scheduled-backup
passed
self-healing
passed
start-from-backup
passed
tablespaces
passed
telemetry-transfer
passed
upgrade-consistency
passed
upgrade-minor
passed
users
passed
We run 19 out of 19
commit: https://github.com/percona/percona-postgresql-operator/pull/921/commits/551a6e8091adaa99e84ee841d178494f2a5a9717
image: perconalab/percona-postgresql-operator:PR-921-551a6e809
Test name
Status
custom-extensions
passed
custom-tls
passed
demand-backup
passed
finalizers
passed
init-deploy
failure
major-upgrade
failure
monitoring
passed
one-pod
passed
operator-self-healing
passed
pitr
passed
scaling
passed
scheduled-backup
passed
self-healing
failure
start-from-backup
passed
tablespaces
passed
telemetry-transfer
passed
upgrade-consistency
passed
upgrade-minor
passed
users
passed
We run 19 out of 19
commit: https://github.com/percona/percona-postgresql-operator/pull/921/commits/4019bb7cb78ef3d118035f4bf63e8ed7233e8c78
image: perconalab/percona-postgresql-operator:PR-921-4019bb7cb
Test name
Status
custom-extensions
passed
custom-tls
passed
demand-backup
passed
finalizers
passed
init-deploy
passed
major-upgrade
failure
monitoring
passed
one-pod
passed
operator-self-healing
passed
pitr
passed
scaling
passed
scheduled-backup
passed
self-healing
passed
start-from-backup
passed
tablespaces
passed
telemetry-transfer
passed
upgrade-consistency
passed
upgrade-minor
passed
users
passed
We run 19 out of 19
commit: https://github.com/percona/percona-postgresql-operator/pull/921/commits/4019bb7cb78ef3d118035f4bf63e8ed7233e8c78
image: perconalab/percona-postgresql-operator:PR-921-4019bb7cb
Test name
Status
custom-extensions
passed
custom-tls
passed
demand-backup
passed
finalizers
passed
init-deploy
passed
monitoring
passed
one-pod
passed
operator-self-healing
passed
pitr
passed
scaling
passed
scheduled-backup
passed
self-healing
passed
start-from-backup
passed
tablespaces
passed
telemetry-transfer
passed
upgrade-consistency
passed
upgrade-minor
passed
users
passed
We run 18 out of 18
commit: https://github.com/percona/percona-postgresql-operator/pull/921/commits/1b55139985d6f3e1b7aac3465be2c8662db75d7e
image: perconalab/percona-postgresql-operator:PR-921-1b5513998
Test name
Status
custom-extensions
failure
custom-tls
failure
demand-backup
failure
finalizers
failure
init-deploy
failure
monitoring
failure
one-pod
failure
operator-self-healing
failure
pitr
failure
scaling
failure
scheduled-backup
failure
self-healing
failure
start-from-backup
failure
tablespaces
failure
telemetry-transfer
failure
upgrade-consistency
failure
upgrade-minor
passed
users
failure
We run 18 out of 18
commit: https://github.com/percona/percona-postgresql-operator/pull/921/commits/3f4916ea622d037bce8aa17f1790d7c2364d6b02
image: perconalab/percona-postgresql-operator:PR-921-3f4916ea6
Test name
Status
custom-extensions
failure
custom-tls
passed
demand-backup
failure
finalizers
passed
init-deploy
passed
monitoring
failure
one-pod
failure
operator-self-healing
passed
pitr
passed
scaling
passed
scheduled-backup
passed
self-healing
failure
start-from-backup
failure
tablespaces
failure
telemetry-transfer
passed
upgrade-consistency
failure
upgrade-minor
failure
users
passed
We run 18 out of 18
commit: https://github.com/percona/percona-postgresql-operator/pull/921/commits/51fc6b5b59511d5d85138de834e005df6735c0a0
image: perconalab/percona-postgresql-operator:PR-921-51fc6b5b5
Test name
Status
custom-extensions
passed
custom-tls
passed
demand-backup
failure
finalizers
passed
init-deploy
passed
monitoring
failure
one-pod
failure
operator-self-healing
passed
pitr
passed
scaling
passed
scheduled-backup
passed
self-healing
failure
start-from-backup
failure
tablespaces
failure
telemetry-transfer
passed
upgrade-consistency
passed
upgrade-minor
passed
users
passed
We run 18 out of 18
commit: https://github.com/percona/percona-postgresql-operator/pull/921/commits/5469b2b45001a8a4465a6a98e18564b87ddd9e29
image: perconalab/percona-postgresql-operator:PR-921-5469b2b45
Test name
Status
custom-extensions
passed
custom-tls
passed
demand-backup
passed
finalizers
passed
init-deploy
passed
monitoring
passed
one-pod
passed
operator-self-healing
passed
pitr
passed
scaling
passed
scheduled-backup
passed
self-healing
passed
start-from-backup
passed
tablespaces
passed
telemetry-transfer
passed
upgrade-consistency
passed
upgrade-minor
passed
users
passed
We run 18 out of 18
commit: https://github.com/percona/percona-postgresql-operator/pull/921/commits/83fd40ee01f1950d3b8e658e3406e905e2dd76c1
image: perconalab/percona-postgresql-operator:PR-921-83fd40ee0
Test name
Status
custom-extensions
passed
custom-tls
passed
demand-backup
failure
finalizers
passed
init-deploy
passed
monitoring
passed
one-pod
passed
operator-self-healing
passed
pitr
passed
scaling
passed
scheduled-backup
passed
self-healing
passed
sidecars
passed
start-from-backup
passed
tablespaces
passed
telemetry-transfer
passed
upgrade-consistency
passed
upgrade-minor
failure
users
passed
We run 19 out of 19
commit: https://github.com/percona/percona-postgresql-operator/pull/921/commits/86af60d8ef9a59eb95e49c6adb605b05e2e5bec2
image: perconalab/percona-postgresql-operator:PR-921-86af60d8e
Test name
Status
custom-extensions
passed
custom-tls
passed
demand-backup
passed
finalizers
passed
init-deploy
passed
monitoring
passed
one-pod
passed
operator-self-healing
passed
pitr
passed
scaling
passed
scheduled-backup
passed
self-healing
passed
sidecars
passed
start-from-backup
passed
tablespaces
passed
telemetry-transfer
passed
upgrade-consistency
passed
upgrade-minor
passed
users
passed
We run 19 out of 19
commit: https://github.com/percona/percona-postgresql-operator/pull/921/commits/4eae78c7ff6c0e867380c2826240ae3e1ef75841
image: perconalab/percona-postgresql-operator:PR-921-4eae78c7f
Test name
Status
custom-extensions
failure
custom-tls
passed
demand-backup
passed
finalizers
passed
init-deploy
passed
monitoring
failure
one-pod
passed
operator-self-healing
passed
pitr
passed
scaling
passed
scheduled-backup
passed
self-healing
passed
sidecars
passed
start-from-backup
passed
tablespaces
passed
telemetry-transfer
passed
upgrade-consistency
passed
upgrade-minor
passed
users
passed
We run 19 out of 19
commit: https://github.com/percona/percona-postgresql-operator/pull/921/commits/a2497125c5120491bd3255ef0bfd8bc4bcb326e0
image: perconalab/percona-postgresql-operator:PR-921-a2497125c
Test name
Status
custom-extensions
passed
custom-tls
passed
demand-backup
passed
finalizers
passed
init-deploy
passed
monitoring
passed
one-pod
passed
operator-self-healing
passed
pitr
passed
scaling
passed
scheduled-backup
passed
self-healing
passed
sidecars
passed
start-from-backup
passed
tablespaces
passed
telemetry-transfer
passed
upgrade-consistency
passed
upgrade-minor
passed
users
passed
We run 19 out of 19
commit: https://github.com/percona/percona-postgresql-operator/pull/921/commits/c3fe5387299f0f6ec20ef79ba2b327db8275770e
image: perconalab/percona-postgresql-operator:PR-921-c3fe53872
|
gharchive/pull-request
| 2024-10-05T16:40:26 |
2025-04-01T04:35:29.284466
|
{
"authors": [
"JNKPercona",
"hors"
],
"repo": "percona/percona-postgresql-operator",
"url": "https://github.com/percona/percona-postgresql-operator/pull/921",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
164608658
|
/routine/Str contains doc for method Str.val
sub val is a sub among methods what may cause the confusion.
Although how we generate /routine/*s can be improved, the very issue has disappeared(?)
Please check it.
Methods
…
…
sub val
No, that's not resolved.
Most of we list all routines available below heading "Methods", while someone substitutes it with "Subroutine". It's an issue of doc conventions.
What is the issue? Moving the documentation elsewhere?
I have been checking Int and it does list routines under Methods, although it uses the word routine, instead of sub. chr would be in the same situation as val right? Would changing sub to routine help here?
Besides, Str itself has several routines
And I see there was an attempt to convert to routine which failed for some reason https://github.com/perl6/doc/commit/e66006021d55b8e72b13989cddebebd770ed37c9
|
gharchive/issue
| 2016-07-08T20:19:58 |
2025-04-01T04:35:29.343599
|
{
"authors": [
"AlexDaniel",
"JJ",
"W4anD0eR96",
"gfldex"
],
"repo": "perl6/doc",
"url": "https://github.com/perl6/doc/issues/683",
"license": "Artistic-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
213518633
|
Add Perl 6 from haskell / python
Add Perl 6 from haskell / python to "Resources"
Looks good. Thanks :)
|
gharchive/pull-request
| 2017-03-11T09:42:05 |
2025-04-01T04:35:29.353484
|
{
"authors": [
"hankache",
"samcv"
],
"repo": "perl6/perl6.org",
"url": "https://github.com/perl6/perl6.org/pull/79",
"license": "Artistic-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
952134158
|
Automatic Release | Validate opa/opal version
added workflow to validate opa/opal version
added release workflow to GitHub & pypi
it is also possible to add release notes to the automation (ref: https://github.com/softprops/action-gh-release)
fixed the trigger for on release workflow
putting it on hold for now, @RazcoDev you might want to give it a go once you return.
I am opening a ticket on you (PER-1052).
|
gharchive/pull-request
| 2021-07-24T19:24:47 |
2025-04-01T04:35:29.382100
|
{
"authors": [
"asafc",
"nir1212"
],
"repo": "permitio/opal",
"url": "https://github.com/permitio/opal/pull/114",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2215572662
|
Error when importing lib in Next 14
I'm submitting a ...
[x] bug report
[ ] feature request
[ ] question about the decisions made in the repository
[ ] question about how to use this project
Summary
I am getting below error, while integrating this with Nextjs14:
Any idea why? This is the snippet:
> 4 | const permit = new Permit({ token: process.env.PERMIT_IO_API_KEY });
As a standalone js file that syncs resources works fine.
Other information (e.g. detailed explanation, stack traces, related issues, suggestions how to fix, links for us to have context, eg. StackOverflow, personal fork, etc.)
Nextjs14
Fixed in version 2.4.1.
Thanks for reporting.
|
gharchive/issue
| 2024-03-29T15:26:38 |
2025-04-01T04:35:29.385637
|
{
"authors": [
"jeremyrx7",
"obsd"
],
"repo": "permitio/permit-node",
"url": "https://github.com/permitio/permit-node/issues/76",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1834789641
|
Choose a validation lib and use it to make basic validation on each form
From the discussion we had with @sjcobb @Gladorme, best candidates for now seems to be a conjunction of https://react-hook-form.com/ + https://zod.dev/
From the discussion we had with @sjcobb @Gladorme, best candidates for now seems to be a conjunction of https://react-hook-form.com/ + https://zod.dev/
Yes, let’s start experimenting with this asap and agreed that the libs in description seem like top candidates
All big forms are validated by Zod and React-Hook-Form (#2044), except for plugin spec inputs. It still need to be done. However, plugin should be able to inject their validation schema in the ValidationProvider in order to have full validation.
|
gharchive/issue
| 2023-08-03T10:33:16 |
2025-04-01T04:35:29.394122
|
{
"authors": [
"Gladorme",
"celian-garcia",
"sjcobb"
],
"repo": "perses/perses",
"url": "https://github.com/perses/perses/issues/1348",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1618284813
|
[ENHANCEMENT] Fetch updates to show full error message
Update fetch to throw UserFriendlyError, tested with invalid panel properties
Screenshots
Checklist
[ ] Pull request has a descriptive title and context useful to a reviewer.
[ ] Pull request title follows the [<catalog_entry>] <commit message> naming convention using one of the following catalog_entry values: FEATURE, ENHANCEMENT, BUGFIX, BREAKINGCHANGE, IGNORE.
[ ] All commits have DCO signoffs.
[ ] Changes that impact the UI include screenshots and/or screencasts of the relevant changes.
yeah I'm totally fine with this change !
For sure we can try to improve the error message but sometimes it's quite harder unfortunately :(.
When it's related to the cue, hopefully the next version of cue will help around that.
|
gharchive/pull-request
| 2023-03-10T03:03:44 |
2025-04-01T04:35:29.398389
|
{
"authors": [
"Nexucis",
"sjcobb"
],
"repo": "perses/perses",
"url": "https://github.com/perses/perses/pull/1015",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
786584413
|
🛑 Procal-tech is down
In c6d0796, Procal-tech (http://iot.procal-tech.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Procal-tech is back up in b4e4d17.
Resolved: Procal-tech is back up in b4e4d17.
|
gharchive/issue
| 2021-01-15T06:12:21 |
2025-04-01T04:35:29.401498
|
{
"authors": [
"personlin"
],
"repo": "personlin/website_uptime",
"url": "https://github.com/personlin/website_uptime/issues/76",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
186411159
|
Updated Jetty dependency and a test broken by Jetty changes.
See issue: #694
9.4.0.v20161208 is out now
@ph-hs I noticed, started updating the pull request last night, but there were complications.
It is interesting that it is only the test that's broken with the Jetty upgrade. In real life, you can do the maven thing and deploy Spark with the 9.4.x of Jetty quite happily.
The bits that you're stuck on could be, solved by:
- WebSocketUpgradeFilter webSocketUpgradeFilter =
- (WebSocketUpgradeFilter) servletContextHandler.getAttribute("org.eclipse.jetty.websocket.server.WebSocketUpgradeFilter");
+ WebSocketUpgradeFilter webSocketUpgradeFilter = (WebSocketUpgradeFilter) servletContextHandler.getServletHandler().getFilter("Jetty_WebSocketUpgradeFilter").getFilter();
and
- WebSocketServerFactory webSocketServerFactory = webSocketUpgradeFilter.getFactory();
+ WebSocketServerFactory webSocketServerFactory = webSocketUpgradeFilter.getConfiguration().getFactory();
and
MappedResource<WebSocketCreator> mappedResource = webSocketUpgradeFilter.getMappings().getMatch("/websocket");
+ MappedResource<WebSocketCreator> mappedResource = webSocketUpgradeFilter.getConfiguration().getMatch("/websocket");
Thanks, I'll take a look tonight. I got burned by not getting the upgrade filter from the attributes and didn't have the time/energy to research into it last night.
One of the unrelated tests is now failing:
testDirectoryTraversalProtectionLocal(spark.staticfiles.StaticFilesTest) Time elapsed: 0.13 sec <<< FAILURE!
java.lang.AssertionError: expected:<404> but was:<400>
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.failNotEquals(Assert.java:834)
at org.junit.Assert.assertEquals(Assert.java:645)
at org.junit.Assert.assertEquals(Assert.java:631)
at spark.staticfiles.StaticFilesTest.testDirectoryTraversalProtectionLocal(StaticFilesTest.java:145)
Needs to be investigated, I think.
Seems the unexpected response comes all the way from Jetty:
WARN org.eclipse.jetty.http.HttpParser - bad HTTP parsed: 400 Bad URI for HttpChannelOverHttp@4db7b425{r=1,c=false,a=IDLE,uri=//localhost:4567/..%5Cspark%5CSpark.class}
@jakaarl would be great if you could have a look at this again.
|
gharchive/pull-request
| 2016-10-31T22:03:16 |
2025-04-01T04:35:29.405949
|
{
"authors": [
"jakaarl",
"paul-hammant",
"ph-hs",
"tipsy"
],
"repo": "perwendel/spark",
"url": "https://github.com/perwendel/spark/pull/698",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
316119345
|
ie: fix errors found by vet
This pull request was made with an automated tool.
The suggested change fixes one or more problems discovered by "go vet".
See https://github.com/functionary/functionary for more details on
the @functionary GitHub user.
Duplicate. I hope you're not planning on opening the same PR every day.
|
gharchive/pull-request
| 2018-04-20T03:56:43 |
2025-04-01T04:35:29.408824
|
{
"authors": [
"functionary",
"pesco"
],
"repo": "pesco/go",
"url": "https://github.com/pesco/go/pull/2",
"license": "ISC",
"license_type": "permissive",
"license_source": "github-api"
}
|
380935999
|
Petablox loses context on Queue.peek()
Consider the following program:
import java.util.*;
class Main {
public static void main(String args[]) throws InterruptedException {
Queue<Main> mains = new ArrayDeque();
mains.add(new Main());
mains.peek().notfound();
}
public void notfound () {
System.out.println("notfound");
}
}
When running the reachable-methods analysis, Petablox does not see that
notfound is reachable.
I use version b95fd27 of Petablox.
Thanks for this report. We are looking at the issue and will fix it soon.
|
gharchive/issue
| 2018-11-14T23:41:11 |
2025-04-01T04:35:29.415991
|
{
"authors": [
"KihongHeo",
"kalhauge"
],
"repo": "petablox-project/petablox",
"url": "https://github.com/petablox-project/petablox/issues/28",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
}
|
111687626
|
Minor memory efficiency improvement for Ruby 2.1+
Using the memory_profiler gem, I noticed a significant number of "VV" allocations during a job that checks the language of many relatively short strings.
Thanks! Will see if we can get this on a 1.0.7 soon.
|
gharchive/pull-request
| 2015-10-15T19:15:03 |
2025-04-01T04:35:29.437783
|
{
"authors": [
"djudd",
"peterc"
],
"repo": "peterc/whatlanguage",
"url": "https://github.com/peterc/whatlanguage/pull/35",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
421912693
|
Animated theme support
It'd be fantastic to have support for animated themes. Animated GIFs or even a series of .png would do!
Working on it :)
Done.
|
gharchive/issue
| 2019-03-17T11:47:28 |
2025-04-01T04:35:29.443250
|
{
"authors": [
"mileyvirus",
"peterprd"
],
"repo": "peterprd/Xeon",
"url": "https://github.com/peterprd/Xeon/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1721337512
|
Safari compatibility/Performance
Dear Peter,
after all this time I finally have a case where I can make really good use of your venerable workspacer extension! Thanks for this :)
I’m not sure if you are interested in further development of it, but still I wanted to share an issue I found regarding usage in Safari 16.4. When trying to type anything into the editor the text is apparently added after the code-area element and so it turns up next to the Sava as button:
An easy workaround of course is not to use Safari, since the editor works fine in Chrome and Firefox. I’m still curious why this happens though.
Another issue is that it takes quite some time to load the workspacer environment if the number of files is rather large. In my specific case there is a folder containing roughly 9.500 files and it takes about a minute to load. I tried implementing the possibility to exclude certain folders, but my programming still were not sufficient …
I had a brief look at the code and did not manage to see what was going wrong with the editor. It would help if I could run Safari on my computer, but my computer isn't a Mac.
I am now thinking about how to redesign Workspacer so that large numbers of files will not slow down the page loading.
Thanks Peter! It’s no big deal, only a minor annoyance, since it’s obviously easy to switch browsers.
Better performance would be nice though, but that is also just a matter of comfortability, so please don’t put too much effort in it.
Update regarding performance: I now have the same installation running on a proper server and it’s loading the same ammount of files almost instantly. ;) So it might just be a performance problem with local development setups or certain types of servers.
I have added the option to exclude directories in the preferences. The editor still has bugs to fix.
Brilliant, thanks!
|
gharchive/issue
| 2023-05-23T06:47:42 |
2025-04-01T04:35:29.457701
|
{
"authors": [
"animaux",
"petertron"
],
"repo": "petertron/workspacer",
"url": "https://github.com/petertron/workspacer/issues/7",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
243247577
|
Process text from stdin
As a potential solution to #2, I was thinking that titlecase could inspect the number of arguments it's being passed, and in case if there are no arguments, would try to read from stdin.
I'm not sure if you agree with this approach or would prefer to use - (echo 'hello world' | titlecase -), similar to other *nix programs, or some command-line flag, but I'd be happy to try my hand at implementing those approaches if you'd like.
Closes #2
Looks good to me. Thank you for the submission!
|
gharchive/pull-request
| 2017-07-16T15:40:56 |
2025-04-01T04:35:29.459421
|
{
"authors": [
"cust0dian",
"peti"
],
"repo": "peti/titlecase",
"url": "https://github.com/peti/titlecase/pull/3",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
51523208
|
Relationship to tj's co?
Promises newb here, so sorry if this is a dumb question, but could someone better informed than I comment on tj's co library as it relates to this project?
Do the two projects "compete" in any way or are they complimentary? If they compete, how do they compare (and should benchmarks be added)? If they're complimentary, how so exactly?
Thanks so much for any help clearing this up!
why there is no benchmark for co anymore?
I'm going to assume because co relies on promises nowdays - in co what
you yield are promises now anyway so at best you'd yield bluebird promises
(which'd still be slower than promise.coroutine)
On Wed, Feb 4, 2015 at 3:51 PM, Anton Wilhelm notifications@github.com
wrote:
why there is no benchmark for co anymore?
—
Reply to this email directly or view it on GitHub
https://github.com/petkaantonov/bluebird/issues/400#issuecomment-72856945
.
Yes promise.coroutine can avoid allocating a new promise if you return bluebird promises to it. Other than that it's implementation efficiency differences.
|
gharchive/issue
| 2014-12-10T06:25:02 |
2025-04-01T04:35:29.463317
|
{
"authors": [
"benjamingr",
"petkaantonov",
"taoeffect",
"timaschew"
],
"repo": "petkaantonov/bluebird",
"url": "https://github.com/petkaantonov/bluebird/issues/400",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
108493044
|
Promise.map and co throws TypeError: fn should be a function
Hey,
I'm refactoring my code base from a full promises api and trying to mix with generators,
But I don't know how to execute a yield operator inside Bluebird.map,
exports.saveAll = function* (items, options) {
options || (options = {})
options.upsert = true
return Promise.map(items, item => {
const gen = function* () {
const query = createQuery(item)
const args = [query, item, options]
return yield collection.findAndModify.apply(this, args)
}
return co(gen)
})
}
And somewhere else I call
yield Collection.saveAll(items)
```
I'm having this error `TypeError: fn should be a function`
How do I iterate a array of object with bluebird collection api and execute a generator in there?
Please ask support questions on StackOverflow, this issue tracker is for feature requests and bug reports. Make sure to include the complete stacktrace of your error when you post it there, also a bit more context about the "somewhere".
PS: My guess is on the this inside your gen, or that you forgot to co.wrap (or Promise.coroutine) your saveAll method so that it returns a promise.
Thank's @bergus I'll ask there next time :)
So I'll try what your suggestions too,
But after wrapping findAndModify(co-monk) on a try catch block I see the error comes from there and I dont know why,
Thx
|
gharchive/issue
| 2015-09-26T20:12:27 |
2025-04-01T04:35:29.466577
|
{
"authors": [
"bergus",
"rafaeljesus"
],
"repo": "petkaantonov/bluebird",
"url": "https://github.com/petkaantonov/bluebird/issues/783",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2394136454
|
Website Loading on Desktop Browsers on Safari
Description: Verify the website loads correctly on Safari (latest version) on macOS.
Preconditions: The website https://oportunitatisicariere.ro/ is live and accessible.
Step 1 : Open Safari on a macOS desktop.
Expected Result: The website should load without any visual or functional issues.
Step 2 : Navigate to https://oportunitatisicariere.ro/
Expected Result: The website should load without any visual or functional issues on Safari (latest version) on macOS.
Step 3 : Observe the page load time and any potential loading issues.
Expected Result: The website should load within 3 seconds.
Step 4 : Verify that the layout, images, text, and other elements are displayed correctly.
Expected Result: The page should fully render all elements correctly without visual issues (e.g., broken images, misaligned text).
Step 5 : Interact with the website (e.g., click links, fill out forms) to ensure functionality.
Expected Result: There should be no loading errors (e.g., 404 errors, server errors).
|
gharchive/issue
| 2024-07-07T17:05:44 |
2025-04-01T04:35:29.485202
|
{
"authors": [
"Georgianaredenstein"
],
"repo": "peviitor-ro/oportunitatisicariere",
"url": "https://github.com/peviitor-ro/oportunitatisicariere/issues/243",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2290852699
|
"Fântânele", "Bozieni", villages are listed in the drop-down menu.
The villages are listed in the drop-down menu alongside its corresponding county and commune, according to the law
Preconditions : the website (https://peviitor-ro.github.io/ui.orase/) is up and runing. After every step delete the text you typed from the search bar.
Step 1
Type "Fântânele" in the search bar.
Expected results
The location is listed in the drop-down menu as "Sat Fântânele, PRAHOVA (Fântânele)".
Step 2
Type "Bozieni" in the search bar.
Expected results
The location is listed in the drop-down menu as "Sat Bozieni, PRAHOVA (Fântânele)".
|
gharchive/issue
| 2024-05-11T11:35:52 |
2025-04-01T04:35:29.488844
|
{
"authors": [
"Elena1303996"
],
"repo": "peviitor-ro/ui.orase",
"url": "https://github.com/peviitor-ro/ui.orase/issues/5466",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2309740513
|
"Bârsești", "Topești", are listed in the drop-down menu
"Bârsești", "Topești" are listed in the drop-down menu alongside its corresponding county and commune, according to the law
Preconditions : the website (https://peviitor-ro.github.io/ui.orase/) is up and running. After every step delete the text you typed from the search bar.
Step 1
Write in the search bar "Bârsești "
Expected results
The location appears in the drop-down menu as "Sat Bârsești , VRANCEA (Bârsești)"
Step 2
Press "x" button.
Expected results
The location was deleted from the search bar.
Step 3
Write in the search bar "Topești"
Expected results
The location appears in the drop-down menu as "Sat Topești , VRANCEA (Bârsești)".
|
gharchive/issue
| 2024-05-22T07:03:01 |
2025-04-01T04:35:29.492792
|
{
"authors": [
"yonutzu"
],
"repo": "peviitor-ro/ui.orase",
"url": "https://github.com/peviitor-ro/ui.orase/issues/6550",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2323802657
|
"Șopârlița" is listed in the drop-down menu
Description:
When user writes ”Șopârlița” the location appear in the drop-down menu next to the comune and the county which is a part according to the law.
Precondition:
The website is up an running. After every step delete the text you typed from the search bar.
Step 1
Type "Șopârlița" in the search bar.
Expected results
The location is listed in the drop-down menu as "Comuna Șopârlița, OLT".
Step 2
Write in the search bar "Șopârlița".
Expected results
The location is listed in the drop-down menu as "Sat Șopârlița OLT (Șopârlița)".
|
gharchive/issue
| 2024-05-29T16:59:23 |
2025-04-01T04:35:29.496524
|
{
"authors": [
"Elena1303996"
],
"repo": "peviitor-ro/ui.orase",
"url": "https://github.com/peviitor-ro/ui.orase/issues/7178",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
786147173
|
Constructing model with Matrix with missing values it fails
using Mill
x = ArrayNode([1f0 2f0; missing missing])
reflectinmodel(x, d -> Chain(Dense(d,10, selu), Dense(10, 10)))
crashes
but
using Mill
x = ArrayNode([1f0 2f0; missing missing])
reflectinmodel(x, d ->Dense(d, 10))
works, which suggest the problem in make_imputing
I have fixed it by changing
_make_imputing(x, t::Chain) = Chain(t[1:end-1], _make_imputing(x, t[end]))
to
_make_imputing(x, t::Chain) = Chain(_make_imputing(x, t[1]), t[2:end]...)
I have fixed it by changing
_make_imputing(x, t::Chain) = Chain(t[1:end-1], _make_imputing(x, t[end]))
to
_make_imputing(x, t::Chain) = Chain(_make_imputing(x, t[1]), t[2:end]...)
Some tests for this added in b05e0ef4757ba187861a7306042b67fa6f39faf6
Some tests for this added in b05e0ef4757ba187861a7306042b67fa6f39faf6
|
gharchive/issue
| 2021-01-14T16:48:24 |
2025-04-01T04:35:29.499381
|
{
"authors": [
"Sheemon7",
"pevnak"
],
"repo": "pevnak/Mill.jl",
"url": "https://github.com/pevnak/Mill.jl/issues/56",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1350279632
|
How to apply 100% completed Chinese translation in iOS version?
Hi dev, since the Chinese translation is 100% complete, how can I apply it to the game? Thanks!
Simplified Chinese (100% complete! 🎉)
Hi! Check your app store for updates, the latest update should contain Chinese in the languages settings.
Hi! Check your app store for updates, the latest update should contain Chinese in the languages settings.
Hi @SKPG-Tech, the screenshots below show my PPL version, and Language Settings, which doesn't seem to have a Chinese language option.
It seems that no new version of PPL has been released recently (the App Store page doesn't show What's New, but PP does).
PPL: https://apps.apple.com/us/app/pewpew-live/id1556977519
PP: https://apps.apple.com/us/app/pewpew/id314964252
Thanks.
The iOS version of PPL has not yet been updated. I'm working on it, but it takes time.
The iOS version of PPL has not yet been updated. I'm working on it, but it takes time.
Oh, I see, thanks.
Closing this issue for now, we'll comment again once the update is out.
|
gharchive/issue
| 2022-08-25T03:50:21 |
2025-04-01T04:35:29.504682
|
{
"authors": [
"SKPG-Tech",
"haolloyin",
"jyaif"
],
"repo": "pewpewlive/ppl-i18n",
"url": "https://github.com/pewpewlive/ppl-i18n/issues/271",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
993475167
|
Update scalafmt-core to 3.0.3
Updates org.scalameta:scalafmt-core from 3.0.0 to 3.0.3.
GitHub Release Notes - Version Diff
I'll automatically update this PR to resolve conflicts as long as you don't change it yourself.
If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below.
Configure Scala Steward for your repository with a .scala-steward.conf file.
Have a fantastic day writing Scala!
Ignore future updates
Add this to your .scala-steward.conf file to ignore future updates of this dependency:
updates.ignore = [ { groupId = "org.scalameta", artifactId = "scalafmt-core" } ]
labels: library-update, semver-patch
Superseded by #51.
|
gharchive/pull-request
| 2021-09-10T18:14:54 |
2025-04-01T04:35:29.508128
|
{
"authors": [
"scala-steward"
],
"repo": "pfcoperez/thebutlerdidit",
"url": "https://github.com/pfcoperez/thebutlerdidit/pull/49",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1460073064
|
Support Helm Chart
depends on #8
fixes #6
This PR introduces helm chart support.
chart itself in charts/gcp-workload-identity-federation-webhook dir
chart is linted in GitHub Actions (ci.yaml)
chart validation by chart-testing
generated manifests by default values.yaml are validated by kubeconform
release workflow is also paved by helm-gh-pages.
I tested the workflow in my forked repo.
index.yaml and chart tgz will be pushed in gh-pages branch. (sample in my fork)
So, you can try:helm repo add gcp-workload-identity-federation-webhook https://everpeace.github.io/gcp-workload-identity-federation-webhook
helm repo update
helm show chart gcp-workload-identity-federation-webhook/gcp-workload-identity-federation-webhook
TODO
[x] prepare empty gh-pages branch
[x] configure Github Pages on gh-pages branch
sure, tomorrow I'll review it. Thanks a lot! 🙇
Just one tip working with helm charts, you can use Kind in the GH Runner to deploy your charts to a local cluster, that could give you more feedback about invalid APIs or sections in the manifests. For instance, in KEDA we use it to validate the chart with several k8s versions
Cool! I will try to do this in another PR.
I'm merging this PR for now and will release v0.1.1 to make helm chart available.
|
gharchive/pull-request
| 2022-11-22T15:19:37 |
2025-04-01T04:35:29.516681
|
{
"authors": [
"JorTurFer",
"everpeace"
],
"repo": "pfnet-research/gcp-workload-identity-federation-webhook",
"url": "https://github.com/pfnet-research/gcp-workload-identity-federation-webhook/pull/9",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
97682410
|
Negative padding
I fixed pooling functions to support negative value for padding.
fix #138
I notice that when pad < 0, cudnn cannot pass gradient check.
It makes a fatal conflict.
I'll close this pr because I cannot merge the current master.
|
gharchive/pull-request
| 2015-07-28T11:32:32 |
2025-04-01T04:35:29.518231
|
{
"authors": [
"unnonouno"
],
"repo": "pfnet/chainer",
"url": "https://github.com/pfnet/chainer/pull/274",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
722084123
|
Add metrics to evaluator
Add some hooks to calculate metrics easily in the evaluation, this avoids having to override evaluators with all the boilerplate code
/test
LGTM!
|
gharchive/pull-request
| 2020-10-15T07:40:34 |
2025-04-01T04:35:29.519347
|
{
"authors": [
"asi1024",
"emcastillo"
],
"repo": "pfnet/pytorch-pfn-extras",
"url": "https://github.com/pfnet/pytorch-pfn-extras/pull/79",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
842283286
|
Language and province selection in under hamburger menu
We would like to add the ability for users to switch provinces and language once they are in the app. The languages vary based on province.
Zeplin screens here
See the video of the flow:
https://user-images.githubusercontent.com/21990736/112690528-2631a580-8e39-11eb-9521-2c761c1df967.mov
Figma prototype here
Notes:
If a user is using AA a language that is not available in MB (eg. Chinese), once they click Manitoba in the province selection, a dialog shows up that tells them "[Language] not available - Select language". Clicking the Select Language button takes them to the language selection drawer page. Clicking Cancel takes them back to the province selection drawer page, but BC will be checked (not Manitoba). If they click Manitoba again, they will get the dialog again that they need to Select Language.
The "[Language] not available - Select Language" dialog is localized to their most recently selected language (in the previous example - Chinese). Once they click "Select language" and hit the language selection drawer page, it will switch to English. If they clicked "Cancel" they go back to the province selection page, and that page is localized to their previous language (in this example - Chinese).
Acceptance Criteria
[ ]
Steps for QA:
Clicking "continue" takes user to Manitoba in English! Cancel keeps user in BC in original language as expected.
https://user-images.githubusercontent.com/85197659/128097518-0168e932-64cd-4ed0-bc8e-b22539f48551.mp4
|
gharchive/issue
| 2021-03-26T20:43:00 |
2025-04-01T04:35:29.527041
|
{
"authors": [
"alysha-PG",
"cherrielam"
],
"repo": "pg-irc/pathways-frontend",
"url": "https://github.com/pg-irc/pathways-frontend/issues/1451",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2624441461
|
Qualcomm Atheros QCNFA765: ath11k_pci crashing
Description
Appliance seems to detect the hardware but firmware crashes on load.
mhi mhi0: Direct firmware load for ath11k/WCN6855/hw2.1/amss.bin failed with error -2
Host operating system
FreeBSD 13.3-RELEASE-p7 GENERIC amd64
Wireless NIC
Qualcomm Technologies, Inc
QCNFA765 Wireless Network Adapter
Wifibox version
0.14.0
Disk image type and version
wifibox-alpine 20240911
Changes to the default configuration files
No response
Logs
Unable to post now (no network)
Additional context
Add any other context about the problem here that might help the investigation.
Have you tried to turn it on and off?
[X] Yes, I have read all the manual pages first!
It seems /lib/firmware/ath11k/WCN6855/hw2.1/ directory is missing from the Alpine appliance image.
Thanks for reporting the issue! Based on this, I believe I was able to identify the root cause. Some of the files that are shipped with linux-firmware are symbolic links and they are not stored in the git repository (from where the contents of the respective tarball is extracted) but need to be added through calling a package builder script.
I created a fix for that in the fix/net/wifibox-alpine/linux-firmware-symlinks branch of the pgj/freebsd-wifibox-port repository. Please try it by reinstalling the net/wifibox-alpine port (version 20241101) from there:
https://github.com/pgj/freebsd-wifibox-port/tree/fix/net/wifibox-alpine/linux-firmware-symlinks
Thank you. Now it does not crash but it seems the driver is waiting for something that never happens:
[ 0.603295] ath11k_pci 0000:00:06.0: MSI vectors: 1
[ 0.603356] ath11k_pci 0000:00:06.0: wcn6855 hw2.1
[ 0.605637] NET: Registered PF_QIPCRTR protocol family
[ 0.761336] mhi mhi0: Requested to power ON
[ 0.761345] mhi mhi0: Power on setup success
[ 0.847684] mhi mhi0: Wait for device to enter SBL or Mission mode```
I have done some investigation, and it seems this a known bug of the ath11k driver. Essentially, the driver does not tolerate well when it is run in a virtualized environment because it assumes that the location of the MSI table matches with that of the host. There is a patch that may address this boot issue, we can take a chance with that if you are available for testing it.
Thanks a lot for the research! Of course I would be glad to help with the testing. Can you apply the patch to the branch?
Unfortunately, I have just noticed that this patch has been made part of both Linux 6.6.50 and Linux 6.10.9 that are integrated into wifibox-alpine 20240911. This means that the problem must be with something else in this case.
Well that's sad. Please let me know if you think of anything else or if there is some more data I could give to help solve this.
For the record this is the wifi chip for Lenovo T14s Gen4 AMD laptops.
|
gharchive/issue
| 2024-10-30T15:15:12 |
2025-04-01T04:35:29.557936
|
{
"authors": [
"Defenso-QTH",
"pgj"
],
"repo": "pgj/freebsd-wifibox",
"url": "https://github.com/pgj/freebsd-wifibox/issues/137",
"license": "BSD-2-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
2008764452
|
Update to Npsql 8 - Issue with TypeHandlerResolverFactory
Describe the bug
When using Microsoft.SemanticKernel.Connectors.Memory.Postgres 1.0.0-beta8 with Npsql 8.0 I'm not able to run the application. It fails with this exception:
System.TypeLoadException: 'Could not load type 'Npgsql.Internal.TypeHandling.TypeHandlerResolverFactory' from assembly 'Npgsql, Version=8.0.0.0, Culture=neutral, PublicKeyToken=5d8b90d52f46fda7'.'
According to this issue the TypeHandlerResolverFactory was removed but VectorExtensions.UseVector seems to require this.
Any chance that you can look into that? Without that I'm not able to use the semantic-kernel with Npsql 8 / .NET 8.
Hi @Cotspheer, just pushed version 0.2.0 of the Pgvector package, which supports Npgsql 8.
Tahts awesome news! Thank you!
Can you add a Pgvector.Dapper 2.0 nuget package which references Pgvector 2.0?
|
gharchive/issue
| 2023-11-23T20:34:52 |
2025-04-01T04:35:29.569516
|
{
"authors": [
"Cotspheer",
"ankane",
"redx177"
],
"repo": "pgvector/pgvector-dotnet",
"url": "https://github.com/pgvector/pgvector-dotnet/issues/26",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
601949722
|
default cut-offs for mob_typer
Dear devs, thanks a lot for a very useful piece of software.
I've been browsing the documentation and paper but it seems I can't find what is the default threshold values for mob_typer similarity cut_offs (min seq. ident.; min. coverage & min evalue) for each database elements (ie. rep; mob; mpf; orit; etc)
Is this listed somewhere ?
Best regards,
Joseph.
Hello Joseph,
Thank you for the report, I will have the defaults of the parameters listed in the readme in a future update. I hadn't realized that the defaults weren't visible when you use --help. Here are the defaults from the arguments in the code. I am currently updating the MOB-suite with a pretty significant overhaul so I will add that to the documentation as part of that update
('--min_rep_evalue', type=str, required=False,
help='Minimum evalue threshold for replicon blastn',
default=0.00001)
('--min_mob_evalue', type=str, required=False,
help='Minimum evalue threshold for relaxase tblastn',
default=0.00001)
('--min_con_evalue', type=str, required=False, help='Minimum evalue threshold for contig blastn',
default=0.00001)
('--min_ori_evalue', type=str, required=False,
help='Minimum evalue threshold for oriT elements blastn',
default=0.00001)
('--min_mpf_evalue', type=str, required=False,
help='Minimum evalue threshold for mpf elements blastn',
default=0.00001)
('--min_rep_ident', type=int, required=False, help='Minimum sequence identity for replicons',
default=80)
('--min_mob_ident', type=int, required=False, help='Minimum sequence identity for relaxases',
default=80)
('--min_ori_ident', type=int, required=False,
help='Minimum sequence identity for oriT elements', default=90)
('--min_mpf_ident', type=int, required=False,
help='Minimum sequence identity for mpf elements', default=80)
('--min_rep_cov', type=int, required=False,
help='Minimum percentage coverage of replicon query by input assembly',
default=80)
('--min_mob_cov', type=int, required=False,
help='Minimum percentage coverage of relaxase query by input assembly',
default=80)
('--min_ori_cov', type=int, required=False,
help='Minimum percentage coverage of oriT',
default=90)
('--min_mpf_cov', type=int, required=False,
help='Minimum percentage coverage of mpf',
default=80)
('--min_overlap', type=int, required=False,
help='Minimum overlap of fragments',
default=10)
Thanks a lot for the swift response !
Le ven. 17 avr. 2020 à 18:36, jrober84 notifications@github.com a écrit :
Hello Joseph,
Thank you for the report, I will have the defaults of the parameters
listed in the readme in a future update. I hadn't realized that the
defaults weren't visible when you use --help. Here are the defaults from
the arguments in the code. I am currently updating the MOB-suite with a
pretty significant overhaul so I will add that to the documentation as part
of that update
('--min_rep_evalue', type=str, required=False,
help='Minimum evalue threshold for replicon blastn',
default=0.00001)
('--min_mob_evalue', type=str, required=False,
help='Minimum evalue threshold for relaxase tblastn',
default=0.00001)
('--min_con_evalue', type=str, required=False, help='Minimum evalue
threshold for contig blastn',
default=0.00001)
('--min_ori_evalue', type=str, required=False,
help='Minimum evalue threshold for oriT elements blastn',
default=0.00001)
('--min_mpf_evalue', type=str, required=False,
help='Minimum evalue threshold for mpf elements blastn',
default=0.00001)
('--min_rep_ident', type=int, required=False, help='Minimum sequence
identity for replicons',
default=80)
('--min_mob_ident', type=int, required=False, help='Minimum sequence
identity for relaxases',
default=80)
('--min_ori_ident', type=int, required=False,
help='Minimum sequence identity for oriT elements', default=90)
('--min_mpf_ident', type=int, required=False,
help='Minimum sequence identity for mpf elements', default=80)
('--min_rep_cov', type=int, required=False,
help='Minimum percentage coverage of replicon query by input assembly',
default=80)
('--min_mob_cov', type=int, required=False,
help='Minimum percentage coverage of relaxase query by input assembly',
default=80)
('--min_ori_cov', type=int, required=False,
help='Minimum percentage coverage of oriT',
default=90)
('--min_mpf_cov', type=int, required=False,
help='Minimum percentage coverage of mpf',
default=80)
('--min_overlap', type=int, required=False,
help='Minimum overlap of fragments',
default=10)
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub
https://github.com/phac-nml/mob-suite/issues/48#issuecomment-615345733,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/ABEFZXQRRARHBIRBHM3VINTRNCAPVANCNFSM4MKW4DSA
.
--
Joseph Nesme
Roskildevej 191, st th
2500 Valby
Denmark
Tel. (fr): +33 6 30 72 33 53
Tel. (dk): +45 50 65 41 50
|
gharchive/issue
| 2020-04-17T12:44:36 |
2025-04-01T04:35:29.603869
|
{
"authors": [
"jnesme",
"jrober84"
],
"repo": "phac-nml/mob-suite",
"url": "https://github.com/phac-nml/mob-suite/issues/48",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
514995170
|
Release tarball doesn't include submodules
Apparently this is a problem across many projects. Doesn't apply to cloned repo.
Include source package (tar.gz to directly install in R) in the releases
|
gharchive/issue
| 2019-10-30T21:31:33 |
2025-04-01T04:35:29.605005
|
{
"authors": [
"TheZetner"
],
"repo": "phac-nml/wade",
"url": "https://github.com/phac-nml/wade/issues/36",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
124772514
|
How to update an existing install ?
Hi,
I am happy with my fresh install of Koel but I am wondering : how to update it ?
After git pull, do I need to run npm install and/or composer install ?
The instruction is right in the README.
Thank you, I have read carefully the wiki,not the README...
|
gharchive/issue
| 2016-01-04T14:52:15 |
2025-04-01T04:35:29.627163
|
{
"authors": [
"Roudaille77",
"phanan"
],
"repo": "phanan/koel",
"url": "https://github.com/phanan/koel/issues/179",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
913388278
|
create a matrix to display admirals AdAM IG coverage (1)
goal: come up with a way to measure how much of the ADaM IG standards are covered by admiral
show overall coverage
show Roche coverage
show GSK coverage
question for testing:
how much is covered by the testing companies?
I see this was moved to Sprint Done. Can we close this issue?
|
gharchive/issue
| 2021-06-07T10:23:16 |
2025-04-01T04:35:29.629649
|
{
"authors": [
"bms63",
"koegerr"
],
"repo": "pharmaverse/admiral",
"url": "https://github.com/pharmaverse/admiral/issues/205",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
323615191
|
Missing browse scoped option
In Nautilus, a really really cool option is "browse scoped" allowing to browse only a subset of the system (a package, a class...). This allows browsing senders/implementors/references... only in a part of the system, which is useful when we look for common methods.
This should probably be implemented in Calypso before Pharo 7 release.
It is already here for a long time. But approach is different.
Scopes are just available in any method browser. No need to press any button to have them.
Notice "Current image" item in method browser. It is combo box to switch these scopes. Try to expand list.
Scopes are inherited from parent browser which spawn them. And it also includes items from selected methods.
Same widget is available for refactoring previews.
So after discussion with Stef and Guille I realized what scenario is actually missing.
And now it is available in 0.11.4 (already integrated):
I added menu item «Set up scope» where you select multiple packages. This scope is propogated to other places.
Indication is done with special label «Scoped pckg» for «Packages» radio button.
In addition this radio have extra logic when you click on it in checked state. In scoped mode it resets scope. In normal mode it sets up the scope of selected packages.
|
gharchive/issue
| 2018-05-16T13:13:52 |
2025-04-01T04:35:29.645509
|
{
"authors": [
"dionisiydk",
"jecisc"
],
"repo": "pharo-ide/Calypso",
"url": "https://github.com/pharo-ide/Calypso/issues/271",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
140034318
|
YamlReader is not working as expected
Is inheritance always followed? The id field does not appear to be populated
This is partly related to https://github.com/phenopackets/phenopacket-format/issues/32
The underlying issue is that Jackson doesn't know what it's supposed to be creating as YAML requires a type e.g. !<person> in order to know what it needs to deserialise as.
entities:
- !<person>
strain: null
types: null
negated_types: null
description: null
id: "person#1"
label: "Joe Bloggs"
taxon: null
sex: null
date_of_birth: null
- !<variant>
descriptionHGVS: "c.1234A>G"
id: "variant#1"
label: "c.1234A>G"
phenotype_profile:
- phenotype: !<phenotype>
measurements: null
types:
- id: "HP:0200055"
label: "Small hands"
negated_types: null
description: null
has_location: null
onset: null
offset: null
severity: null
environment: null
entity: !<person>
strain: null
types: null
negated_types: null
description: null
id: "person#1"
label: "Joe Bloggs"
taxon: null
sex: null
date_of_birth: null
evidence:
- types:
- id: "ECO:0000033"
label: "TAS"
negated_types: null
description: null
supporting_entities: null
source:
- "PMID:23455423"
JSON uses a "type" key to acheive the same thing.
You want to be able to do use YAML tags of sorts, but only ever refer to the id of an entity once defined:
entities:
- id: "person#1"
type: person
strain: null
types: null
negated_types: null
description: null
label: "Joe Bloggs"
taxon: null
sex: null
date_of_birth: null
- id: "variant#1"
descriptionHGVS: "c.1234A>G"
type: variant
label: "c.1234A>G"
phenotype_profile:
- entity: "person#1"
phenotype:
measurements: null
types:
- id: "HP:0200055"
label: "Small hands"
negated_types: null
description: null
has_location: null
onset: null
offset: null
severity: null
environment: null
evidence:
- types:
- id: "ECO:0000033"
label: "TAS"
source:
- "PMID:23455423"
Plus I guess you don't want the keys with null values written out either.
This is now read as expected:
---
id: "phenopkt#1"
title: "Patient X phenotypes and potentially causative variant"
persons:
- id: "person#1"
label: "Joe Bloggs"
sex: "M"
variants:
- id: "variant#1"
descriptionHGVS: "c.1234A>G"
label: "c.1234A>G"
phenotype_profile:
- entity: "person#1"
phenotype:
types:
- id: "HP:0200055"
label: "Small hands"
evidence:
- types:
- id: "ECO:0000033"
label: "TAS"
source:
- id: "PMID:23455423"
|
gharchive/issue
| 2016-03-10T23:01:34 |
2025-04-01T04:35:29.742171
|
{
"authors": [
"cmungall",
"julesjacobsen"
],
"repo": "phenopackets/phenopacket-reference-implementation",
"url": "https://github.com/phenopackets/phenopacket-reference-implementation/issues/16",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1117621679
|
Add options to change order of custom commands
This PR adds one improvement and implements a feature request #4.
Command buttons are now removed from ribbon immediately, no restart necessary
I've added up/down arrow buttons in command settings to reorder commands (closes #4)
That looks really good! Thank you
|
gharchive/pull-request
| 2022-01-28T16:46:53 |
2025-04-01T04:35:29.819703
|
{
"authors": [
"miluoshi",
"phibr0"
],
"repo": "phibr0/obsidian-customizable-sidebar",
"url": "https://github.com/phibr0/obsidian-customizable-sidebar/pull/20",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
472584873
|
Fatal Exception with A_HelloWorld
As a recap, I couldn't get any of the examples to compile. I was using a git version of ESP8266/Arduino. I didn't want to change that version back to 2.5.2 so I went to my ESP32 machine, installed ESP8266/Arduino 2.5.2 release code. That solved the compile issue I raised so I closed it.
While I had compiled it I didn't have an ESP8266 board to try. I didn't want to mix ESP32 Arduino stuff with ESP8266 Arduino stuff. That was the reason I got the second computer. I still didn't want to change my git version on my ESP8266 machine.
I have wanted to try the current Esparto and today I decided to proceed on my ESP32 machine. All seemed good until I loaded the flash with A_HelloWorld. I also loaded the Data directory into SPIFFS. I then did a reset. And I kept getting immediate fatal exceptions.
ets Jan 8 2013,rst cause:1, boot mode:(3,6)
load 0x4010f000, len 1384, room 16
tail 8
chksum 0x2d
csum 0x2d
v8b899c12
~ld
SDK:2.2.1(cfd48f3)/Core:2.5.2=20502000/lwIP:STABLE-2_1_2_RELEASE/glue:1.1-7-g82abda3/BearSSL:a143020
SPIFFSImpl: allocating 512+240+1400=2152 bytes
SPIFFSImpl: mounting fs @300000, size=fb000, block=2000, page=100
SPIFFSImpl: mount rc=0
Fatal exception 28(LoadProhibitedCause):
epc1=0x40234ed0, epc2=0x00000000, epc3=0x00000000, excvaddr=0x00000000, depc=0x00000000
Exception (28):
epc1=0x40234ed0 epc2=0x00000000 epc3=0x00000000 excvaddr=0x00000000 depc=0x00000000
I tried a bunch of things and got nowhere. In the end I took an SSD drive and installed Arduino 1.8.9 on it. (a portable - self contained copy). I installed the 2.5.2 release code of the ESP8266/Arduino framework. I installed the current SPIFFS upload tool and the exception decoder. I installed the Esparto and listed required libraries.
I am using a 4MB ESP-12E module with the following settings.
Decoding stack results
0x4020277c: std::operator , std::allocator >(std::basic_string , std::allocator > const&, std::basic_string , std::allocator > const&) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2\bits/basic_string.h line 2573
0x40203250: std::map , std::allocator > >::operator[](std::string const&) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2\bits/stl_tree.h line 1141
0x40226600: malloc(size_t) at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\umm_malloc\umm_malloc.cpp line 1685
0x402036d9: ESPArto::ESPArto(std::map , std::allocator > >) at E:\sketchbook\libraries\esparto-master\src\ESPArto.cpp line 464
0x402069bc: ESPArto::_mqTat(std::vector >) at E:\sketchbook\libraries\esparto-master\src\mqtt.cpp line 230
0x4022b6b4: std::_Function_base::_Base_manager >)>::_M_manager(std::_Any_data&, std::_Any_data const&, std::_Manager_operation) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2/functional line 1934
0x40202534: std::_Function_handler >), void (*)(std::vector >)>::_M_invoke(std::_Any_data const&, std::vector >) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2/functional line 2069
0x402069e0: ESPArto::_mqTdaily(std::vector >) at E:\sketchbook\libraries\esparto-master\src\mqtt.cpp line 232
0x4022b6b4: std::_Function_base::_Base_manager >)>::_M_manager(std::_Any_data&, std::_Any_data const&, std::_Manager_operation) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2/functional line 1934
0x40202534: std::_Function_handler >), void (*)(std::vector >)>::_M_invoke(std::_Any_data const&, std::vector >) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2/functional line 2069
0x40100504: millis() at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\core_esp8266_wiring.cpp line 186
0x4022b6b4: std::_Function_base::_Base_manager >)>::_M_manager(std::_Any_data&, std::_Any_data const&, std::_Manager_operation) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2/functional line 1934
0x40202534: std::_Function_handler >), void (*)(std::vector >)>::_M_invoke(std::_Any_data const&, std::vector >) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2/functional line 2069
0x4020115f: _GLOBAL__sub_I_Esparto() at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2\bits/stl_tree.h line 671
0x40222260: String::String(char const*) at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\WString.cpp line 36
0x402054ac: _GLOBAL__sub_I__ZN7ESPArto4_udpE() at E:\sketchbook\libraries\esparto-master\src\ESPArto.cpp line 497
0x40223435: init_done() at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\core_esp8266_main.cpp line 149
0x4023975a: tcp_input at core/tcp_in.c line 321
0x4023ad00: udp_sendto_if_src at core/udp.c line 855
0x4022e952: _printf_float at /home/earle/src/esp-quick-toolchain/repo/newlib/newlib/libc/stdio/nano-vfprintf_float.c line 241
0x402018fb: std::__introsort_loop__gnu_cxx::__normal_iterator >, int, ESPArto::_matchTasks(std::function , std::function )::__lambda22>(__gnu_cxx::__normal_iterator > >, __gnu_cxx::__normal_iterator > >, int) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2\bits/stl_algo.h line 2328
0x40203412: std::map , std::allocator > >::operator[](std::string&&) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2\bits/stl_tree.h line 140
0x40238642: tcp_parseopt at core/tcp_in.c line 1946
0x4023811f: dns_gethostbyname_addrtype at core/dns.c line 1566
0x4010044c: __wrap_spi_flash_read(uint32_t, uint32_t*, size_t) at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\core_esp8266_phy.cpp line 308
0x4023813a: dns_gethostbyname_addrtype at core/dns.c line 1580
0x40238642: tcp_parseopt at core/tcp_in.c line 1946
0x40100854: uart_isr(void*) at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\uart.cpp line 364
0x4023811f: dns_gethostbyname_addrtype at core/dns.c line 1566
0x4010044c: __wrap_spi_flash_read(uint32_t, uint32_t*, size_t) at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\core_esp8266_phy.cpp line 308
0x4023813a: dns_gethostbyname_addrtype at core/dns.c line 1580
0x401004e4: millis() at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\core_esp8266_wiring.cpp line 183
0x401004e4: millis() at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\core_esp8266_wiring.cpp line 183
0x402394c6: tcp_input at core/tcp_in.c line 189
0x4023975a: tcp_input at core/tcp_in.c line 321
0x401008c8: uart_isr(void*) at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\uart.cpp line 388
0x4023923f: tcp_receive at core/tcp_in.c line 1814
0x401004c0: millis() at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\core_esp8266_wiring.cpp line 180
0x4023925a: tcp_receive at core/tcp_in.c line 1814
I used two different ESP-12 module boards that I have been using for a couple of years. I have never had any issues with them.
Can you think of what the issue might be?
One other note. I did a complete flash contents write to remove any previous information before I programmed the code and SPIFFS.
Just pushed a quick bugfix (3.3.2) to solve a problem some folk were having
with the early examples e.g. "Hello World" If that's you, update your copy
and try again - many apologies of course
On Thu, Jul 25, 2019 at 1:08 AM Rudy notifications@github.com wrote:
One other note. I did a complete flash contents write to remove any
previous information before I programmed the code and SPIFFS.
—
You are receiving this because you are subscribed to this thread.
Reply to this email directly, view it on GitHub
https://github.com/philbowles/esparto/issues/22?email_source=notifications&email_token=AGT3V7S7ENCRRL7N776DIPDQBDOHHA5CNFSM4IGVL7SKYY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGOD2X3LVI#issuecomment-514831829,
or mute the thread
https://github.com/notifications/unsubscribe-auth/AGT3V7UYGPAB5BPJRSMCFRLQBDOHHANCNFSM4IGVL7SA
.
I tried the new code and it didn't solve the problem. It still crashes immediately on startup.
How do I go about troubleshooting this?
Decoding stack results
0x4020277c: std::operator , std::allocator >(std::basic_string , std::allocator > const&, std::basic_string , std::allocator > const&) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2\bits/basic_string.h line 2573
0x40203250: std::map , std::allocator > >::operator[](std::string const&) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2\bits/stl_tree.h line 1141
0x4022695c: malloc(size_t) at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\umm_malloc\umm_malloc.cpp line 1685
0x402036d9: ESPArto::ESPArto(std::map , std::allocator > >) at E:\sketchbook\libraries\esparto-master\src\ESPArto.cpp line 464
0x402069e4: ESPArto::_mqTat(std::vector >) at E:\sketchbook\libraries\esparto-master\src\mqtt.cpp line 230
0x4022ba10: std::_Function_base::_Base_manager >)>::_M_manager(std::_Any_data&, std::_Any_data const&, std::_Manager_operation) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2/functional line 1934
0x40202534: std::_Function_handler >), void (*)(std::vector >)>::_M_invoke(std::_Any_data const&, std::vector >) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2/functional line 2069
0x40206a08: ESPArto::_mqTdaily(std::vector >) at E:\sketchbook\libraries\esparto-master\src\mqtt.cpp line 232
0x4022ba10: std::_Function_base::_Base_manager >)>::_M_manager(std::_Any_data&, std::_Any_data const&, std::_Manager_operation) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2/functional line 1934
0x40202534: std::_Function_handler >), void (*)(std::vector >)>::_M_invoke(std::_Any_data const&, std::vector >) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2/functional line 2069
0x40100504: millis() at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\core_esp8266_wiring.cpp line 186
0x4022ba10: std::_Function_base::_Base_manager >)>::_M_manager(std::_Any_data&, std::_Any_data const&, std::_Manager_operation) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2/functional line 1934
0x40202534: std::_Function_handler >), void (*)(std::vector >)>::_M_invoke(std::_Any_data const&, std::vector >) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2/functional line 2069
0x4020115f: _GLOBAL__sub_I_Esparto() at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2\bits/stl_tree.h line 671
0x402225bc: String::String(char const*) at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\WString.cpp line 36
0x402054ac: _GLOBAL__sub_I__ZN7ESPArto4_udpE() at E:\sketchbook\libraries\esparto-master\src\ESPArto.cpp line 497
0x40223791: init_done() at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\core_esp8266_main.cpp line 149
0x4023ad00: udp_input at core/udp.c line 195
0x4022e952: isspace at /home/earle/src/esp-quick-toolchain/repo/newlib/newlib/libc/ctype/isspace.c line 43
0x402018fb: std::__introsort_loop__gnu_cxx::__normal_iterator >, int, ESPArto::_matchTasks(std::function , std::function )::__lambda22>(__gnu_cxx::__normal_iterator > >, __gnu_cxx::__normal_iterator > >, int) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2\bits/stl_algo.h line 2328
0x40203412: std::map , std::allocator > >::operator[](std::string&&) at e:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\tools\xtensa-lx106-elf-gcc\2.5.0-3-20ed2b9\xtensa-lx106-elf\include\c++\4.8.2\bits/stl_tree.h line 140
0x4023811f: dns_recv at core/dns.c line 1206
0x4010044c: __wrap_spi_flash_read(uint32_t, uint32_t*, size_t) at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\core_esp8266_phy.cpp line 308
0x4023813a: dns_recv at core/dns.c line 1221
0x40100854: uart_isr(void*) at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\uart.cpp line 364
0x4023811f: dns_recv at core/dns.c line 1206
0x4010044c: __wrap_spi_flash_read(uint32_t, uint32_t*, size_t) at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\core_esp8266_phy.cpp line 308
0x4023813a: dns_recv at core/dns.c line 1221
0x401004e4: millis() at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\core_esp8266_wiring.cpp line 183
0x401004e4: millis() at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\core_esp8266_wiring.cpp line 183
0x402394c6: tcp_receive at core/tcp_in.c line 1781
0x401008c8: uart_isr(void*) at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\uart.cpp line 388
0x4023923f: tcp_receive at core/tcp_in.c line 2117
0x401004c0: millis() at E:\arduino-1.8.9-windows\arduino-1.8.9\portable\packages\esp8266\hardware\esp8266\2.5.2\cores\esp8266\core_esp8266_wiring.cpp line 180
0x4023925a: tcp_receive at core/tcp_in.c line 1618
Hmm. Can you check that you have indeed updated to v3.3.2 it absolutely should have fixed this. ALso take any of the later examples that use WiFi and try one of them (the problem was with empty/ non-existent config block - WiFI examples - by definition - must have config data) then on the serial monitor type cmd/info and show me the output?
|
gharchive/issue
| 2019-07-24T22:48:08 |
2025-04-01T04:35:29.831405
|
{
"authors": [
"RudyFiero",
"philbowles"
],
"repo": "philbowles/esparto",
"url": "https://github.com/philbowles/esparto/issues/22",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1524970898
|
请求帮助,使用task_create无法建任务
tasks.data 返回了这些
{'title': 'xxxx-建任务', 'customFields': {'story': '相关需求', 'estStarted': '预计开始', 'deadline': '截止日期', 'mailto': '抄送给', 'pri': '优先级', 'estimate': '最初预计'}.....
官方文档上的这些参数改如何填写呢?
$params = array(
'project' => 1,
'type' => 'ui',
'module' => 0,
'assignedTo[]' => 'lisi',
'testAssignedTo[]' => 'lisi',
'color' => '',
'name' => '测试添加任务2',
'pri' => 2,
'estimate' => 1,
'desc' => '测试添加任务描述测试添加任务描述',
'estStarted' => '2019-11-11',
'deadline' => '2019-11-12',
'mailto[1]' => 'lisi'
); // 请求参数
更新到0.3.0版本,post参数可使用data传入,例如
response = zentao.task_create(
executionID=2,
storyID=0,
moduleID=0,
...
data={
"execution": 2,
"type": "design",
"name": "锦囊喵叽",
"assignedTo[]": "老六",
"pri": 3,
"desc": "暴打小柯基"
...
},
)
成啦,感谢!
|
gharchive/issue
| 2023-01-09T05:48:40 |
2025-04-01T04:35:29.847339
|
{
"authors": [
"GbingSong",
"philip1134"
],
"repo": "philip1134/pyzentao",
"url": "https://github.com/philip1134/pyzentao/issues/3",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
642516225
|
请教关于生成订阅的问题。
我有4台VPS,想生成一个订阅便于管理,不需要流量控制这些功能,请问有什么好软件推荐吗?
可以用我写的另一个工具 Shawdowrockets 訂閱鏈接編輯器
如果不想用我的API,把这个工具生成的 BASE64 放到一个纯文本档案里再放到你的服务器或 Github Gist 上也可以达成订阅的效果。
好的,谢谢。
|
gharchive/issue
| 2020-06-21T07:56:37 |
2025-04-01T04:35:29.885339
|
{
"authors": [
"LeiD215",
"phlinhng"
],
"repo": "phlinhng/v2ray-tcp-tls-web",
"url": "https://github.com/phlinhng/v2ray-tcp-tls-web/issues/18",
"license": "WTFPL",
"license_type": "permissive",
"license_source": "github-api"
}
|
1989700784
|
Merge Branch
Merging Heinrich's and cleaning again
bibs-Merge1
ended up merging all Omar/Princess/Heinrich/Sam
|
gharchive/issue
| 2023-11-13T01:07:13 |
2025-04-01T04:35:29.891067
|
{
"authors": [
"bibiygarcia"
],
"repo": "phobs-co/phobs-co.github.io",
"url": "https://github.com/phobs-co/phobs-co.github.io/issues/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
36567594
|
WP8 push notification
I have easily applied push notification service for my android app, but confused how to do in windows phone.
if(device.platform == "Win32NT"){
pushNotification.register(
channelHandler,
errorHandler,
{
"channelName": channelName,
"ecb": "onNotificationWP8",
"uccb": "channelHandler",
"errcb": "jsonErrorHandler"
});
}
I doesn't understand what the above parameters. eg: channelName. where can i get the values for these parameters??
Also, does anyone give me the link to full tutorial applying PNS in wp8 including server side too?
Regards,
Lomas Joshi
Hello postixo,
My uri is always empty.
I don't get fix the problem with a your example using angularjs.Your application is still working? Are you could send the html code complete?
fabiofapeli, you have to ask for the uri several times. yes, so strange, but Microsoft is Microsoft :)
@joshilomas your code is clean , can you write a blog about it so that everyone can see when they google it instead of coming here :)
Over a year later and Cordova & WP8 still has garbage information on doing this (or really anything) for the Windows Phone. I am now in the same boat as you all.
Anyone please help me out for Integrating Push Notification in Windows Phone Cordova.
I tried hard but unsucessfull.
For me its working fine.
just keep loop RegisterMPNS Function until unless you are not getting uri.
function registerMPNS(){
var pushNotification;
pushNotification = window.plugins.pushNotification;
pushNotification.register(
channelHandler,
errorHandler,
{
"channelName": "xyz", //channelName is whateveryou like
"ecb": "onNotificationWP8",
"uccb": "channelHandler",
"errcb": "jsonErrorHandler"
});
}
$window.onNotificationWP8 = function(e){
}
function channelHandler(event){
var uri = event.uri;
if(uri==''){
registerMPNS();
}
}
|
gharchive/issue
| 2014-06-26T11:34:16 |
2025-04-01T04:35:29.907386
|
{
"authors": [
"fabiofapeli",
"gustter",
"joshilomas",
"kundan1988",
"mohammadmanzoor8972",
"revaret",
"tetreault"
],
"repo": "phonegap-build/PushPlugin",
"url": "https://github.com/phonegap-build/PushPlugin/issues/255",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.