id
stringlengths 4
10
| text
stringlengths 4
2.14M
| source
stringclasses 2
values | created
timestamp[s]date 2001-05-16 21:05:09
2025-01-01 03:38:30
| added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| metadata
dict |
---|---|---|---|---|---|
795324005
|
I'd love to help!
This is one of the best mods I've ever used! I'd love to help you troubleshoot it on 1.16.4. Are you still working on this?
I have to admit that I keep meaning to learn how to use github so bear with me...
Sorry I noticed I wasn't good enough in Java and didn't have enough knowledge in modding
That's ok 😁. I don't have much either. Thanks for getting back to me. If
you ever change your mind, let me know 👍
On Sun., Nov. 14, 2021, 10:30 Daniel Mendes, @.***>
wrote:
Sorry I noticed I wasn't good enough in Java and didn't have enough
knowledge in modding
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub
https://github.com/Daniel-Mendes/Simply-Caterpillar/issues/1#issuecomment-968322903,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/AFQUZFBUW2JC536GMNGP6F3UL7PZFANCNFSM4WVWDVYQ
.
Triage notifications on the go with GitHub Mobile for iOS
https://apps.apple.com/app/apple-store/id1477376905?ct=notification-email&mt=8&pt=524675
or Android
https://play.google.com/store/apps/details?id=com.github.android&referrer=utm_campaign%3Dnotification-email%26utm_medium%3Demail%26utm_source%3Dgithub.
|
gharchive/issue
| 2021-01-27T18:03:41 |
2025-04-01T04:54:52.689670
|
{
"authors": [
"Daniel-Mendes",
"phrayz"
],
"repo": "Daniel-Mendes/Simply-Caterpillar",
"url": "https://github.com/Daniel-Mendes/Simply-Caterpillar/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
593423454
|
BaseReferenceDrawer constant uses wrong height for a custom struct
Issue
After generating a Reference ScriptableObject script with the SO code generator using a custom serializable struct with its own propertyDrawer, the representation of the struct has the wrong height when folded out in the editor. As far as I know, this only happens for SO References, not any of the other SO classes. It requires the "Use Constant" flag to be true. The height for the resulting representation of the property is always wrong, as the foldout icon takes up an entire line's height.
Reproduction flow
Take a simple struct, in this case called CameraView. It contains two Vector3 values, to determine the desired rotation (in Euler angles) and position of a camera. This is the struct definition:
[Serializable]
public struct CameraView {
public Vector3 cameraPosition;
public Vector3 cameraRotation;
}
Since the struct is serializable by the UnityEditor, we have a very basic PropertyDrawer, that just draws the fields as intended:
OnGUI:
EditorGUI.PropertyField(position, property, label, true);
And GetPropertyHeight:
return EditorGUI.GetPropertyHeight(property);
After this, we generate SO Architecture code for the CameraView type, including a CameraViewReference. This allows us to create a few CameraViewVariable assets, which look fine in the editor:
(Of course, we're ignoring the developer description as that was already mentioned in a different issue.)
When we finally include this reference in a MonoBehaviour, like so:
public CameraViewReference cameraView;
It looks fine as a variable:
But when we select "Use Constant", we are immediately presented with the foldout icon, positioned a bit strangely:
This isn't a major issue though, as it's still possible to click the rightmost edge of the foldout icon, which results in the images shown at the top of this issue.
Another issue would be the indentation of this struct, as it would be better suited using the full width underneath. However, that is not the main focus.
Potential solution
After digging around in the BaseReferenceDrawer, I found a potential solution. In GetPropertyHeight, the SupportsMultiLine bool is false for custom structs. When manually set to true for testing purposes, the height changes when the field is folded out.
Closed:
Folded out:
But as you can see, the height is still a bit strange, as it now results in a bit of unused space. Regardless, I noticed the use of the [MultiLine] attribute for classes would allow one to make use of this behaviour. However, as it stands, this is only reserved for classes, and not for structs.
Thanks for reporting this issue in so much detail!
I should hopefully have some time in the coming week to catch up with the increasing list of issues :)
I've encountered the same problem in the collection editor.
I did things a bit different this time along. I generated these scriptable objects for a class instead of a struct, and putting the MultiLine attribute above the class declaration doesn't seem to have any effect.
I figure that this is probably the result of the same problem. To reproduce this one, just follow the original reproduction flow, but generate a collection as well. Using a class VS a struct might make a difference but I'm not completely sure about that.
I'm working on a fix by implementing #110 - this is, as you might imagine, quite a big feature. Once it's done everything should render properly, since I'm also re-writing all the editors.
Once it's done you won't have to implement PropertyDrawers either, unless you want some custom behaviour, of course.
This will be fixed in the new update. Just added the finishing touches.
This update includes drawing custom objects without property drawers. The above is an example of that, the type simply including the following:
[System.Serializable]
public class TestType
{
[SerializeField]
private float floatValue;
[SerializeField]
private Quaternion quaternionValue;
[SerializeField]
private SubType subType;
[System.Serializable]
public class SubType
{
public double publicValue;
[SerializeField]
private string privateValue;
}
}
|
gharchive/issue
| 2020-04-03T14:29:44 |
2025-04-01T04:54:52.700489
|
{
"authors": [
"DanielEverland",
"Robert-Oost"
],
"repo": "DanielEverland/ScriptableObject-Architecture",
"url": "https://github.com/DanielEverland/ScriptableObject-Architecture/issues/119",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
802835565
|
Update BaseVariableEditor.cs
fixes #140
if(IsClampable) -> if(_isClamped.boolValue)
fixes #141
now invokes Raise event when value changed in Inspector
Adds a "Raise" button to manually invoke event.
whoops, I accidentally included some "developer description" stuff.... You don't need to include that.
Please submit separate PRs pr. fix/feature
|
gharchive/pull-request
| 2021-02-07T02:08:13 |
2025-04-01T04:54:52.702691
|
{
"authors": [
"DanielEverland",
"Monsoonexe"
],
"repo": "DanielEverland/ScriptableObject-Architecture",
"url": "https://github.com/DanielEverland/ScriptableObject-Architecture/pull/142",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
977666705
|
could not be resolved to an NgModule class
ERROR in node_modules/ng-pick-datetime/date-time/date-time.module.d.ts:1:22 - error NG6002: Appears in the NgModule.imports of AppModule, but could not be resolved to an NgModule class.
This likely means that the library (ng-pick-datetime) which declares OwlDateTimeModule has not been processed correctly by ngcc, or is not compatible with Angular Ivy. Check if a newer
version of the library is available, and update if so. Also consider checking with the library's authors to see if the library is expected to be compatible with Ivy.
Also getting this error.
Ok so to resolve this you need to do the following
run this npm i ng-pick-datetime-ex
run this npm install dayjs
put this in the css or scss style sheet @import "../node_modules/ng-pick-datetime-ex/assets/style/picker.min.css";
4.put these in the module where the picker is needed OwlDateTimeModule,
OwlNativeDateTimeModule,
and you should be fine
if there is any error do npm i
It work for me just now.
@iamade install those two packages seemed to fix my issue. Im using pnpm and was getting a similar problem. Did those packages needed to be added because they are missing peer dependencies?
It still doesn't work on angular cli version 13
@iamade
I understand from your answer that we should no more use
"ng-pick-datetime": "^7.0.0",
but instead we should use
"ng-pick-datetime-ex": "^15.0.0",
am I correct?
|
gharchive/issue
| 2021-08-24T03:31:19 |
2025-04-01T04:54:52.716038
|
{
"authors": [
"JosehGaks",
"crh225",
"harvanchik",
"iamade",
"malek-itani",
"sheng-snd"
],
"repo": "DanielYKPan/date-time-picker",
"url": "https://github.com/DanielYKPan/date-time-picker/issues/740",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2341469084
|
FAILED SYNC IMAGE docker.io/library/busybox:latest
SYNC docker.io/library/busybox:latest
镜像不存在, 请根据模板提示, 填写正确的镜像
|
gharchive/issue
| 2024-06-08T04:40:16 |
2025-04-01T04:54:52.722509
|
{
"authors": [
"ddgit123",
"wzshiming"
],
"repo": "DaoCloud/public-image-mirror",
"url": "https://github.com/DaoCloud/public-image-mirror/issues/3346",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
530403484
|
Features and enhancements request
There are 4 in total,sorry for my poor English first,hope these pics can help.
Gallery info alignment
2.Add option to allow user to choose column quantity in both portrait and landscape,personally i request a 3 columns option in portrait
3.Add option to turn on/off gallery name
4.Shorten the distance between two page rows and add page number displaying to each page.
该意见对提升app的体验还是有点帮助的,特别是3和4项
@Dar9586
Thanks,version 1.9.8 solved 1 3 4 and only "3 column in portrait" remains to be done.
Added 3 column while in portrait in 1.9.9
Perfect,thank you.
|
gharchive/issue
| 2019-11-29T15:38:09 |
2025-04-01T04:54:52.726424
|
{
"authors": [
"Dar9586",
"ZerOri",
"hohommx2015"
],
"repo": "Dar9586/NClientV2",
"url": "https://github.com/Dar9586/NClientV2/issues/39",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
491251195
|
Support getting version for macOS Betas
Modify the version parsing code to support macOS Beta versions which don't include a patch version (e.g. 10.15).
Signed-off-by: Zac Brown zacbrown@users.noreply.github.com
The lint failures do not appear to be related to my change.
Codecov Report
Merging #103 into master will not change coverage.
The diff coverage is n/a.
@@ Coverage Diff @@
## master #103 +/- ##
=======================================
Coverage 82.67% 82.67%
=======================================
Files 3 3
Lines 127 127
=======================================
Hits 105 105
Misses 22 22
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 5672990...3ae2ce9. Read the comment docs.
Thank you for the pull request. As far as I can see, the Travis build is failing because of TryFrom used in cargo-audit which isn't related to your changes. I'm going to merge and release a new version soon.
|
gharchive/pull-request
| 2019-09-09T18:21:37 |
2025-04-01T04:54:52.733325
|
{
"authors": [
"DarkEld3r",
"codecov-io",
"zacbrown"
],
"repo": "DarkEld3r/os_info",
"url": "https://github.com/DarkEld3r/os_info/pull/103",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
418079856
|
NetworkServerConnection and NetworkClientConnection missing overrides.
In the documents here the example for NetworkServerConnection and NetworkClientConnection have methods that need overrides but does not have it in the examples.
Fixed in next version!
|
gharchive/issue
| 2019-03-07T01:02:45 |
2025-04-01T04:54:52.749913
|
{
"authors": [
"Derek-R-S",
"JamJar00"
],
"repo": "DarkRiftNetworking/DarkRift.Documentation",
"url": "https://github.com/DarkRiftNetworking/DarkRift.Documentation/issues/3",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
553068747
|
Don't add trailing / in api calls
Always appending / to the resource can cause some POST requests for new resources to fail when using _callAPI / _buildRequestURI.
PR Summary
Fixes #230
PR Checklist
No new help or unit tests to write.
[N/A ] Write Help
[N/A ] Write Unit Test
[x] Update CHANGELOG.md
Thanks for the update and support. I will look into merging this PR this week.
|
gharchive/pull-request
| 2020-01-21T19:03:50 |
2025-04-01T04:54:52.801962
|
{
"authors": [
"DarqueWarrior",
"daveneeley"
],
"repo": "DarqueWarrior/vsteam",
"url": "https://github.com/DarqueWarrior/vsteam/pull/231",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
344077564
|
BTRX-277 - Implement "Download Full Application" button
[ ] Submitter: Include the JIRA issue number in the PR description
[ ] Submitter: Make sure Swagger is updated if API changes
[ ] Submitter: Check documentation and code comments. Add explanatory PR comments if helpful.
[ ] Submitter: JIRA ticket checks:
Acceptance criteria exists and is met
Note any changes to implementation from the description
Add notes on what you've tested
[ ] Submitter: Update RC_XXX release ticket with any config or environment changes necessary
[ ] Submitter: Update FISMA documentation if changes to:
Authentication
Authorization
Encryption
Audit trails
[ ] Submitter: If you're adding new libraries, sign us up to security updates for them
[ ] Tell the tech lead (TL) that the PR exists if they wants to look at it
[ ] Anoint a lead reviewer (LR). Assign PR to LR
Review cycle:
LR reviews
Rest of team may comment on PR at will
LR assigns to submitter for feedback fixes
Submitter rebases to develop again if necessary
Submitter makes further commits. DO NOT SQUASH
Submitter updates documentation as needed
Submitter reassigns to LR for further feedback
[ ] TL sign off
[ ] LR sign off
[ ] Product Owner sign off
[ ] Assign to submitter to finalize
[ ] Submitter: Verify all tests go green, including CI tests
[ ] Submitter: Squash commits and merge to develop
[ ] Submitter: Delete branch after merge
[ ] Submitter: Test this change works on dev environment after deployment. YOU own getting it fixed if dev isn't working for ANY reason!
[ ] Submitter: Verify swagger UI on dev environment still works after deployment
[ ] Submitter: Inform other teams of any API changes via hipchat and/or email
[ ] Submitter: Mark JIRA issue as resolved once this checklist is completed
Coverage increased (+0.05%) to 61.684% when pulling 2c4e071532fae34ca6d6888cf5f23424c3771e31 on BTRX-277-DownloadFullApplication into 9d378ecce2ad5f59a5b01adb160f068513793b19 on develop.
|
gharchive/pull-request
| 2018-07-24T15:01:17 |
2025-04-01T04:54:52.844246
|
{
"authors": [
"coveralls",
"vvicario"
],
"repo": "DataBiosphere/consent",
"url": "https://github.com/DataBiosphere/consent/pull/289",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
2570129523
|
(fleet) add network healthchecks to the script
This PR adds a couple network healthcheck to URLs we require to work for some features (APM instrumentation today). This should help diagnose if those urls are accessible.
/merge
|
gharchive/pull-request
| 2024-10-07T11:17:32 |
2025-04-01T04:54:52.866800
|
{
"authors": [
"arbll"
],
"repo": "DataDog/agent-linux-install-script",
"url": "https://github.com/DataDog/agent-linux-install-script/pull/282",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1427873178
|
[BUG] There's no way to know when the trace agent is ready
Agent Environment
Agent is started locally via docker:
docker run --rm \
--name dd-agent \
-p8125:8125 \
-p8126:8126 \
-v /var/run/docker.sock:/var/run/docker.sock:ro \
-v /proc/:/host/proc/:ro \
-v /sys/fs/cgroup/:/host/sys/fs/cgroup:ro \
-e DD_API_KEY="OMIT" \
-e DD_APP_KEY="OMIT" \
-e DD_AGENT_MAJOR_VERSION="7" \
-e DD_SERVICE="myapp" \
-e DD_HOSTNAME="localhost" \
-e DD_ENV="dev" \
gcr.io/datadoghq/agent:7
Describe what happened:
The start script for my application:
while ! nc localhost 8126 </dev/null; do
echo "waiting for statsd server to start..."
sleep 1
done
ddtrace-run myapp
If I run my start script it will sit and wait for statsd to start. If I then go and run the above docker command to start the dd agent, my start script eventually connects and then proceeds to run my app via ddtrace-run.
As my app is starting up though, I get this error:
[2022-10-28 14:42:48][ddtrace.internal.writer][ERROR] - failed to send traces to Datadog Agent at http://localhost:8126/v0.4/traces
Traceback (most recent call last):
File "myapp/env/lib/python3.10/site-packages/tenacity/__init__.py", line 407, in __call__
result = fn(*args, **kwargs)
File "myapp/env/lib/python3.10/site-packages/ddtrace/internal/writer.py", line 446, in _send_payload
response = self._put(payload, headers)
File "myapp/env/lib/python3.10/site-packages/ddtrace/internal/writer.py", line 399, in _put
resp = compat.get_connection_response(self._conn)
File "myapp/env/lib/python3.10/site-packages/ddtrace/internal/compat.py", line 235, in get_connection_response
return conn.getresponse()
File "/opt/homebrew/Cellar/python@3.10/3.10.5/Frameworks/Python.framework/Versions/3.10/lib/python3.10/http/client.py", line 1374, in getresponse
response.begin()
File "/opt/homebrew/Cellar/python@3.10/3.10.5/Frameworks/Python.framework/Versions/3.10/lib/python3.10/http/client.py", line 318, in begin
version, status, reason = self._read_status()
File "/opt/homebrew/Cellar/python@3.10/3.10.5/Frameworks/Python.framework/Versions/3.10/lib/python3.10/http/client.py", line 287, in _read_status
raise RemoteDisconnected("Remote end closed connection without"
http.client.RemoteDisconnected: Remote end closed connection without response
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "myapp/env/lib/python3.10/site-packages/ddtrace/internal/writer.py", line 559, in flush_queue
self._retry_upload(self._send_payload, encoded, n_traces)
File "myapp/env/lib/python3.10/site-packages/tenacity/__init__.py", line 404, in __call__
do = self.iter(retry_state=retry_state)
File "myapp/env/lib/python3.10/site-packages/tenacity/__init__.py", line 361, in iter
raise retry_exc from fut.exception()
tenacity.RetryError: RetryError[<Future at 0x106c2e9b0 state=finished raised RemoteDisconnected>]
This error comes 1 second before I finally see:
2022-10-28 21:42:48 UTC | TRACE | INFO | (run.go:254 in Infof) | Listening for traces at http://0.0.0.0:8126
In the logs for the docker dd-agent. So what's happening is my app is trying to send traces before the dd-agent is ready to send them. There needs to be some way for me to ask dd-agent if it's ready to go so that my app can wait until that time to proceed with traces.
Describe what you expected:
I expect there to be some way in either bash or python for me to ask the dd-agent, "is it ok to send traces?"
Steps to reproduce the issue:
This will happen for any application that uses ddtrace-run and can not rely on dd-agent to be running already at start up time.
Additional environment details (Operating System, Cloud provider, etc):
I see this error on linux and mac systems, aws and heroku.
Hi @lexicalunit, thanks for using Datadog.
Since in this case the port on the host is opened by docker rather than the agent itself, it indeed happens before the agent is ready to handle requests. It may be better to wait until a full http request succeeds, for example by running curl http://localhost:8126/info instead of nc. If you need further assistance, feel free to reach out to our support.
Thanks.
|
gharchive/issue
| 2022-10-28T21:46:26 |
2025-04-01T04:54:52.886403
|
{
"authors": [
"lexicalunit",
"vickenty"
],
"repo": "DataDog/datadog-agent",
"url": "https://github.com/DataDog/datadog-agent/issues/14103",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1321396639
|
Adds 'print-default-build-tags' invoke task
What does this PR do?
Adds an extra invoke task to print out go build tags for a given build configuration.
eg:
$ inv print-default-build-tags
apm,consul,containerd,cri,docker,ec2,etcd,gce,jetson,jmx,kubeapiserver,kubelet,netcgo,orchestrator,otlp,podman,process,python,secrets,systemd,zk,zlib
$ inv print-default-build-tags -a arm64 -b agent -f iotoo
'iotoo' does not correspond to an agent flavor. Options: ['base', 'iot', 'heroku', 'dogstatsd']
$ inv print-default-build-tags --help
Usage: inv[oke] [--core-opts] print-default-build-tags [--options] [other tasks here ...]
Docstring:
Build the default list of tags based on the build type and current platform.
Prints as comma separated list suitable for go tooling (eg, gopls, govulncheck)
The container integrations are currently only supported on Linux, disabling on
the Windows and Darwin builds.
Options:
-a STRING, --arch=STRING
-b STRING, --build=STRING
-f STRING, --flavor=STRING
Motivation
Various go tooling takes in a build-tags parameter to run checks/analysis against a codebase.
I implemented this as a convenience for developers to configure their tooling correctly when working on the agent codebase.
Additional Notes
Possible Drawbacks / Trade-offs
Describe how to test/QA your changes
N/A
Reviewer's Checklist
[ ] If known, an appropriate milestone has been selected; otherwise the Triage milestone is set.
[ ] Use the major_change label if your change either has a major impact on the code base, is impacting multiple teams or is changing important well-established internals of the Agent. This label will be use during QA to make sure each team pay extra attention to the changed behavior. For any customer facing change use a releasenote.
[ ] A release note has been added or the changelog/no-changelog label has been applied.
[ ] Changed code has automated tests for its functionality.
[ ] Adequate QA/testing plan information is provided if the qa/skip-qa label is not applied.
[ ] At least one team/.. label has been applied, indicating the team(s) that should QA this change.
[ ] If applicable, docs team has been notified or an issue has been opened on the documentation repo.
[ ] If applicable, the need-change/operator and need-change/helm labels have been applied.
[ ] If applicable, the k8s/<min-version> label, indicating the lowest Kubernetes version compatible with this feature.
[ ] If applicable, the config template has been updated.
Maybe you can use something like
@task
def print_default_build_tags(ctx, build="agent", arch="x64", flavor=AgentFlavor.base.name):
"""
Build the default list of tags based on the build type and current platform.
Prints as comma separated list suitable for go tooling (eg, gopls, govulncheck)
The container integrations are currently only supported on Linux, disabling on
the Windows and Darwin builds.
"""
try:
flavor = AgentFlavor[flavor]
except KeyError:
raise Exit(message=f"`{flavor}` is not an agent flavor", code=1)
print(",".join(sorted(get_default_build_tags(build, arch, flavor))))
for the flavor issue
Thanks for the suggestion, works great
No particular rush for this to get in, moving it to 7.40
|
gharchive/pull-request
| 2022-07-28T19:43:02 |
2025-04-01T04:54:52.894616
|
{
"authors": [
"paulcacheux",
"scottopell"
],
"repo": "DataDog/datadog-agent",
"url": "https://github.com/DataDog/datadog-agent/pull/12907",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1957563100
|
Attach disk with docker images to micro-vms on boot
What does this PR do?
This PR optimizes micro-vm setup time by attaching a second disk to each VM on boot, and mounting it at a specified location.
This disk contains the docker images required for system-probe tests. This approach eliminates the need for sharing the docker images with each micro-vm and performing docker load, thus significantly reducing the tests jobs time.
Motivation
Additional Notes
Corresponding test-infra-definitions PR:
https://github.com/DataDog/test-infra-definitions/pull/387
https://datadoghq.atlassian.net/browse/EBPF-243
Possible Drawbacks / Trade-offs
Describe how to test/QA your changes
Reviewer's Checklist
[ ] If known, an appropriate milestone has been selected; otherwise the Triage milestone is set.
[ ] Use the major_change label if your change either has a major impact on the code base, is impacting multiple teams or is changing important well-established internals of the Agent. This label will be use during QA to make sure each team pay extra attention to the changed behavior. For any customer facing change use a releasenote.
[ ] A release note has been added or the changelog/no-changelog label has been applied.
[ ] Changed code has automated tests for its functionality.
[ ] Adequate QA/testing plan information is provided if the qa/skip-qa label is not applied.
[ ] At least one team/.. label has been applied, indicating the team(s) that should QA this change.
[ ] If applicable, docs team has been notified or an issue has been opened on the documentation repo.
[ ] If applicable, the need-change/operator and need-change/helm labels have been applied.
[ ] If applicable, the k8s/<min-version> label, indicating the lowest Kubernetes version compatible with this feature.
[ ] If applicable, the config template has been updated.
should we consider using memory optimized machine types?
We are not memory-constrained at the moment.
Looks like this failed?
error: Process exited with status 1: running " sudo virsh vol-create ci-22271412-4670-kernel-matrix-testing-22271412-global-pool /tmp/volume-ci-22271412-4670-kernel-matrix-testing-22271412-global-pool-docker-arm64.qcow2.xml":
error: Failed to create vol from /tmp/volume-ci-22271412-4670-kernel-matrix-testing-22271412-global-pool-docker-arm64.qcow2.xml
error: inaccessible backing store volume /home/kernel-version-testing/docker-arm64.qcow2: No such file or directory
Looks like this failed?
error: Process exited with status 1: running " sudo virsh vol-create ci-22271412-4670-kernel-matrix-testing-22271412-global-pool /tmp/volume-ci-22271412-4670-kernel-matrix-testing-22271412-global-pool-docker-arm64.qcow2.xml":
error: Failed to create vol from /tmp/volume-ci-22271412-4670-kernel-matrix-testing-22271412-global-pool-docker-arm64.qcow2.xml
error: inaccessible backing store volume /home/kernel-version-testing/docker-arm64.qcow2: No such file or directory
Yeah the job for arm64 docker disk in ami-builder failed due to docker rate limiting. Will try again today.
|
gharchive/pull-request
| 2023-10-23T16:33:54 |
2025-04-01T04:54:52.903450
|
{
"authors": [
"brycekahle",
"usamasaqib"
],
"repo": "DataDog/datadog-agent",
"url": "https://github.com/DataDog/datadog-agent/pull/20340",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2078454450
|
Handle Triage milestone in lint-milestone
What does this PR do?
Handle Triage milestone in lint-milestone.
Motivation
It is not a real milestone label, the PR would be ignored when generating QA cards.
Additional Notes
Possible Drawbacks / Trade-offs
Describe how to test/QA your changes
Reviewer's Checklist
[ ] If known, an appropriate milestone has been selected; otherwise the Triage milestone is set.
[ ] Use the major_change label if your change either has a major impact on the code base, is impacting multiple teams or is changing important well-established internals of the Agent. This label will be use during QA to make sure each team pay extra attention to the changed behavior. For any customer facing change use a releasenote.
[ ] A release note has been added or the changelog/no-changelog label has been applied.
[ ] Changed code has automated tests for its functionality.
[ ] Adequate QA/testing plan information is provided. Except if the qa/skip-qa label, with required either qa/done or qa/no-code-change labels, are applied.
[ ] At least one team/.. label has been applied, indicating the team(s) that should QA this change.
[ ] If applicable, docs team has been notified or an issue has been opened on the documentation repo.
[ ] If applicable, the need-change/operator and need-change/helm labels have been applied.
[ ] If applicable, the k8s/<min-version> label, indicating the lowest Kubernetes version compatible with this feature.
[ ] If applicable, the config template has been updated.
/merge
|
gharchive/pull-request
| 2024-01-12T09:46:00 |
2025-04-01T04:54:52.909978
|
{
"authors": [
"pgimalac"
],
"repo": "DataDog/datadog-agent",
"url": "https://github.com/DataDog/datadog-agent/pull/22015",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
800359260
|
pkg/trace/api: add discovery endpoint
TODO:
[x] Add a test to ensure JSON keys never change regardless of field changes in the code.
Example response from http://localhost:8126/hello:
{
"Version": "0.99.0",
"GitCommit": "fab047e10",
"GitBranch": "master",
"BuildDate": "2020-12-04 15:57:06.74187 +0200 EET m=+0.029001792",
"GoVersion": "1.15.6",
"Endpoints": [
"/v0.3/traces",
"/v0.3/services",
"/v0.4/traces",
"/v0.4/services",
"/v0.5/traces",
"/v0.5/stats",
"/profiling/v1/input"
],
"Features": [
"nostats"
],
"Config": {
"Enabled": true,
"Hostname": "datadog-gbbr.local",
"DefaultEnv": "gabriel.aszalos",
"ConfigPath": "/opt/datadog-agent/etc/datadog.yaml",
"Endpoints": [
{
"Host": "http://localhost:8888",
"NoProxy": false
}
],
"BucketInterval": 10000000000,
"ExtraAggregators": null,
"ExtraSampleRate": 1,
"TargetTPS": 10,
"MaxEPS": 200,
"ReceiverHost": "localhost",
"ReceiverPort": 8126,
"ReceiverSocket": "",
"ConnectionLimit": 0,
"ReceiverTimeout": 20,
"MaxRequestBytes": 52428800,
"StatsWriter": {
"ConnectionLimit": 0,
"QueueSize": 0,
"FlushPeriodSeconds": 0
},
"TraceWriter": {
"ConnectionLimit": 0,
"QueueSize": 0,
"FlushPeriodSeconds": 0
},
"ConnectionResetInterval": 0,
"StatsdHost": "localhost",
"StatsdPort": 8125,
"LogLevel": "debug",
"LogFilePath": "/var/log/datadog/trace-agent.log",
"LogThrottling": false,
"MaxMemory": 0,
"MaxCPU": 0,
"WatchdogInterval": 10000000000,
"ProxyURL": null,
"SkipSSLValidation": false,
"Ignore": {},
"ReplaceTags": null,
"AnalyzedRateByServiceLegacy": {},
"AnalyzedSpansByService": {},
"DDAgentBin": "/opt/datadog-agent/bin/agent/agent",
"Obfuscation": null
}
}
Hi @gbbr
What do values like those below mean? It looks to me like a feature flag hasn't been set. The client wouldn't expect these to be null but booleans. These would parse as false in some languages but I believe they are actually set by default in the trace agent.
"Obfuscation": null,
"ReplaceTags": null
This one looks like an actually nullable value, rather than the consequence of an unset feature flag, and I'd argue complete absence is a better policy in this case.
"ProxyURL": null
@richardstartin first off, nothing that is in the Config is a feature flag. A "feature flag" is simply an experimental option that may be enabled or not (and will eventually be removed in subsequent versions) currently showing up in FeatureFlags section.
Obfuscation and ReplaceTags are both objects, not booleans. I have made a new commit to leave this out from the response when they are null. They specify the configuration for the obfuscator, basically the contents of that section in the datadog.yaml file (see this link. As for the tag replacer, see this link. Most of what's in the Config key is from that file...
If it helps, here's a fully populated object from the tests: https://github.com/DataDog/datadog-agent/blob/c16442c8acbd858bd47bc40c3aef451e6aa6866b/pkg/trace/api/api_test.go#L687-L843
Removed ProxyURL, probably not interesting at all to a user...
@dougqh @richardstartin I took all of your feedback, and made the Go code uglier in order to get prettier JSON. I also manually selected which config options should show up, and removed any which seemed like they might contain sensitive information or be irrelevant to the user. Furthermore, I made the obfuscation configurations as booleans as requested.
Please check the updated PR description and let me know your thoughts.
|
gharchive/pull-request
| 2021-02-03T14:03:23 |
2025-04-01T04:54:52.917534
|
{
"authors": [
"gbbr",
"richardstartin"
],
"repo": "DataDog/datadog-agent",
"url": "https://github.com/DataDog/datadog-agent/pull/7344",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1242255139
|
Fix json.Marshal for aggregate timeseries response
What does this PR do?
json.Marshal for aggregate log timeseries responses will always return null values. This change fixes the json marshal to correctly return the values for the timeseries.
What inspired you to submit this pull request?
I need to query the API for aggregated logs and the values for the timeseries are sort of important.
Link to the issue describing the bug that you're fixing.
Fixes https://github.com/DataDog/datadog-api-client-go/issues/1505
Additional Notes
Review checklist
Please check relevant items below:
[ ] This PR includes all newly recorded cassettes for any modified tests.
[x] This PR does not rely on API client schema changes.
[ ] The CI should be fully passing.
[ ] Or, this PR relies on API schema changes and this is a Draft PR to include tests for that new functionality.
Note: CI shouldn't be run on this Draft PR, as its expected to fail without the corresponding schema changes.
Thanks, we need a fix in the template, this should be done in #1508.
|
gharchive/pull-request
| 2022-05-19T19:56:21 |
2025-04-01T04:54:52.922140
|
{
"authors": [
"oblogic7",
"therve"
],
"repo": "DataDog/datadog-api-client-go",
"url": "https://github.com/DataDog/datadog-api-client-go/pull/1506",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
807070489
|
tests: remove unstable warnings, fix test script
This removes the error in go 1.13, plus a few warnings
/azp run
|
gharchive/pull-request
| 2021-02-12T09:12:41 |
2025-04-01T04:54:52.923033
|
{
"authors": [
"therve"
],
"repo": "DataDog/datadog-api-client-go",
"url": "https://github.com/DataDog/datadog-api-client-go/pull/717",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1330380432
|
Registration types in registering components should be fx.Out and have group:"true" defined
From @ogaca-dd in #1:
I think having provides struct requires a lot of typing. What do you think about:
// define in ipcserver/component.go
type RouteProvider struct {
fx.Out
IPCRoute *ipcserver.Route `group:"true"`
}
// ipcserver.NewRoute will returns RouteProvider instead
func newFlare(deps dependencies) (Component, RouteProvider) {
....
return f, ipcserver.NewRoute("/agent/flare", f.ipcHandler),
}
Pros: There is no need to define provides struct and it hides the logic of the group.
Note: Same technique could also applied to fx.in, ie, newFlare could take 2 args, dependencies and a new struct containing Registrations []*Registration group:"true". I don't know if it worth it.
Originally posted by @ogaca-dd in https://github.com/DataDog/dd-agent-comp-experiments/pull/1#discussion_r938975316
WIP - https://github.com/DataDog/dd-agent-comp-experiments/pull/new/issue13
|
gharchive/issue
| 2022-08-05T20:45:46 |
2025-04-01T04:54:52.970581
|
{
"authors": [
"djmitche"
],
"repo": "DataDog/dd-agent-comp-experiments",
"url": "https://github.com/DataDog/dd-agent-comp-experiments/issues/13",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2603442560
|
chore: Update releaser scripts to generate CHANGELOG.
What and why?
Generate the CHANGELOG for a package from conventional commit messages instead of relying on manually updating the changelog.
This does not satisfy all the requirements of RUM-5667 but eliminates a particularly manual portion of releasing.
refs: RUM-5667
Review checklist
[ ] This pull request has appropriate unit and / or integration tests
[ ] This pull request references a Github or JIRA issue
Merging now that 2.8.0 is out.
|
gharchive/pull-request
| 2024-10-21T19:15:53 |
2025-04-01T04:54:52.973128
|
{
"authors": [
"fuzzybinary"
],
"repo": "DataDog/dd-sdk-flutter",
"url": "https://github.com/DataDog/dd-sdk-flutter/pull/658",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2236200142
|
Table names surrounded by back ticks are considered BIND_PARAMETER
I was missing the WithDBMS(mysql) lexer option
|
gharchive/issue
| 2024-04-10T18:39:06 |
2025-04-01T04:54:54.314023
|
{
"authors": [
"tanner-bruce"
],
"repo": "DataDog/go-sqllexer",
"url": "https://github.com/DataDog/go-sqllexer/issues/34",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1127340712
|
BREAKING CHANGE Remove pyhdb
pyhdb library has some stability issue so we cannnot, in good faith, leave users exposed to it. From 7.35 onwards users will need to perform a manual install of hdbcli to use SAP Hana integration.
This PR:
Removes pyhdb from the agent
Removes pyhdb compatibility from the integration
Updates documentation to inform the user that they need to manually install hdbcli to use the integration, while before it was an optional step
This PR is best viewed hiding whitespace changes
/azp run PR Changes
|
gharchive/pull-request
| 2022-02-08T14:36:21 |
2025-04-01T04:54:54.316214
|
{
"authors": [
"hithwen",
"sarah-witt"
],
"repo": "DataDog/integrations-core",
"url": "https://github.com/DataDog/integrations-core/pull/11469",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
696226322
|
Use database config template in existing specs
What does this PR do?
Apply db template to integrations with the spec.yaml using the custom query options
Motivation
Additional Notes
Review checklist (to be filled by reviewers)
[ ] Feature or bugfix MUST have appropriate tests (unit, integration, e2e)
[ ] PR title must be written as a CHANGELOG entry (see why)
[ ] Files changes must correspond to the primary purpose of the PR as described in the title (small unrelated changes should have their own PR)
[ ] PR must have changelog/ and integration/ labels attached
And postgres?
|
gharchive/pull-request
| 2020-09-08T22:34:42 |
2025-04-01T04:54:54.319058
|
{
"authors": [
"ChristineTChen",
"hithwen"
],
"repo": "DataDog/integrations-core",
"url": "https://github.com/DataDog/integrations-core/pull/7548",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
1691352861
|
npm installable javascript DJ client
Summary
This adds an ES6 javascript DJ client that's npm installable. This PR only includes the get endpoints to pull entities but in a follow-up PR I'll add the ability to create entities as well.
Test Plan
N/A (will follow up with tests)
[ ] PR has an associated issue: #
[ ] make check passes
[ ] make test shows 100% unit test coverage
Deployment Plan
N/A
Thanks @shangyian, updated the python builds to ignore the client/javascript directory then added an eslint scripts and some workflow steps to build and lint the js client.
@agorajek or @betodealmeida not urgent but we should add the javascript jobs to be required checks for merging PRs. Also we have to setup a release job using an npm access token just like we did with pypi. I went ahead and published the 0.0.1 version for now: https://www.npmjs.com/package/datajunction
|
gharchive/pull-request
| 2023-05-01T20:41:36 |
2025-04-01T04:54:54.323156
|
{
"authors": [
"samredai"
],
"repo": "DataJunction/dj",
"url": "https://github.com/DataJunction/dj/pull/484",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1517222593
|
[Bug] 无主键的表, 怎么同步?
Search before asking
[X] I had searched in the issues and found no similar issues.
What happened
在测试cdc单表及cdcsource整库同步时, 有遇到有些表是没有设置主键的, 这时候同步就会报错。
这个问题实际上看起来有点不合理, 因为没有主键的表,要实现cdc方式的同步。但是有些库,特别是bi库,会大量存在无主键,
但多字段可确定唯一记录的场景(不会在建表的时候就把复合主键建上)
大家是怎么解决的?因为有时候是不方便给这些表去加主键的。
这个后续会有解决方案吗?
What you expected to happen
希望无主键的表, 也能实现同步
How to reproduce
无
Anything else
No response
Version
0.6.7
Are you willing to submit PR?
[ ] Yes I am willing to submit a PR!
Code of Conduct
[X] I agree to follow this project's Code of Conduct
可以在Sink环节去订制无主键的处理,社区源码里所提供的是有主键的处理方式,从而实现思路上来讲,这些都很容易扩展,关键看如何去定义参数来确定哪些字段是主键效果。
感谢回复.
你指的是去改造源码, 还是指在建sink表时有处理方法?
可以在Sink环节去订制无主键的处理,社区源码里所提供的是有主键的处理方式,从而实现思路上来讲,这些都很容易扩展,关键看如何去定义参数来确定哪些字段是主键效果。
感谢回复.
你指的是去改造源码, 还是指在建sink表时有处理方法?
Sink时处理
可以在Sink环节去订制无主键的处理,社区源码里所提供的是有主键的处理方式,从而实现思路上来讲,这些都很容易扩展,关键看如何去定义参数来确定哪些字段是主键效果。
感谢回复.
你指的是去改造源码, 还是指在建sink表时有处理方法?
Sink时处理
感谢回复.
我昨天测试出问题所在了, 源物料表可以无主键, 但是过程中的sink, 且目标表必须有主键, 否则同步会失败.
且mysql的处理方式还比较特殊.
mysql 整库集成的参数中可否添加设置black.list 过滤下有问题的表(无主键的表,或者不想同步的表),虽然有table-name参数能控制要同步的表,有时表太多操作起来不方便,当定位到1000张表中有1张导致整库集成失败时,通过black.list 参数过滤掉就不会影响到整个作业999张表的执行。
mysql 整库集成的参数中可否添加设置black.list 过滤下有问题的表(无主键的表,或者不想同步的表),虽然有table-name参数能控制要同步的表,有时表太多操作起来不方便,当定位到1000张表中有1张导致整库集成失败时,通过black.list 参数过滤掉就不会影响到整个作业999张表的执行。
这个建议我支持, 有时候排掉几张表要方便的多, 不过table-name 好像是支持正则的, 只不过写正则其实真的不容易
|
gharchive/issue
| 2023-01-03T10:52:15 |
2025-04-01T04:54:54.329951
|
{
"authors": [
"aiwenmo",
"cooltnt",
"deepthinkin"
],
"repo": "DataLinkDC/dlink",
"url": "https://github.com/DataLinkDC/dlink/issues/1496",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2407582238
|
chore(main): release 0.6.0
:robot: I have created a release beep boop
0.6.0 (2024-07-14)
Features
Move the Python bindings into a standalone repository (c789ee4)
This PR was generated with Release Please. See documentation.
:robot: Created releases:
0.6.0
:sunflower:
|
gharchive/pull-request
| 2024-07-14T20:15:47 |
2025-04-01T04:54:54.345894
|
{
"authors": [
"DataTriny"
],
"repo": "DataTriny/accesskit-python",
"url": "https://github.com/DataTriny/accesskit-python/pull/1",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
2400707772
|
(feat) - Ollama
Didn't add token consumption re-implementation in this PR. will add another one.
Make input and output same as what openai endpoints expect for ollama.
For instance, when output is same formatting, this then becomes directly compatible with our other components which extract output from choices.
If you want to keep both options, then by default you can keep openai formatting & add a varibale named let's say 'api_format' = 'original' to use ollama endpoints.
|
gharchive/pull-request
| 2024-07-10T13:03:43 |
2025-04-01T04:54:54.347244
|
{
"authors": [
"omkar-334",
"satpalsr"
],
"repo": "DataformerAI/dataformer",
"url": "https://github.com/DataformerAI/dataformer/pull/13",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
219551319
|
Memory leak on log.setup
Version 8.3 (8E162) . Swift 3.1
XCGLogger at "4.1.0"
class AppDelegate: UIResponder, UIApplicationDelegate, UNUserNotificationCenterDelegate {
...
let log = XCGLogger.default
...
func application(_ application: UIApplication, didFinishLaunchingWithOptions launchOptions: [UIApplicationLaunchOptionsKey: Any]?) -> Bool {
...
log.setup(level: .debug, showFunctionName: false, showThreadName: false, showLevel: true, showFileNames: true, showLineNumbers: true, writeToFile: logPath, fileLevel: .debug)
On last line app crash reporting
Message from debugger: Terminated due to memory issue
This is the same issue as #193
Thanks. Tracking this under #193. But your XCGLogger at "4.1.0" comment has me concerned. Since there is no 4.1.0 yet. #195 seems related to that as well.
@xandros6 issue is fixed now. Can you please close this?
Think this was an issue with the bad tags, which should now be fixed with the 5.0.1 release. But please re-open this if you continue to see issues.
|
gharchive/issue
| 2017-04-05T11:26:09 |
2025-04-01T04:54:54.358157
|
{
"authors": [
"DaveWoodCom",
"sohayb",
"xandros6"
],
"repo": "DaveWoodCom/XCGLogger",
"url": "https://github.com/DaveWoodCom/XCGLogger/issues/194",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2260987503
|
upgrade to new pandas api, replace append with concat
upgrade to pandas 2.0
Hey, I thought I should say that this change is approved but isn't merged yet!
For those who have the error after running the example in the README.md, you can install @ertosns 's working fork of this package from the command line.
pip3 uninstall Historic_Crypto
pip3 install git+https://github.com/ertosns/Historic_Crypto.git
its about one year that your project notifications coming for me. why this is happening?
|
gharchive/pull-request
| 2024-04-24T10:44:02 |
2025-04-01T04:54:54.361342
|
{
"authors": [
"BamdadTabari",
"augustus-thomas",
"ertosns"
],
"repo": "David-Woroniuk/Historic_Crypto",
"url": "https://github.com/David-Woroniuk/Historic_Crypto/pull/25",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1759865988
|
Problems Syncing with Blueretro Module
Trying to sync with Nintendo Switch and PC work fine, but trying to sync with bluetooth modules that require holding the sync button while turning on the controller do not work. For example, when trying to connect to a blueretro module such as this one (https://www.amazon.com/gp/product/B0BTLVLV4W/ref=ppx_yo_dt_b_asin_title_o02_s00?ie=UTF8&psc=1) I cannot get it to connect. The instructions on the module say to hold the sync button (which works on my other switch controllers) or to reprogram the bluetooth address using a computer for other controllers (such as the PS3).
What would be the best way to connect this controller via bluetooth to the blueretro module if it asks to hold the sync button? Or is there a way to manually type in a bluetooth mac address?
Hey, just got one built. The Pico seems to attempt to auto sync with a BlueRetro without a button when powered on, causing the BlueRetro device to stop blinking it's blue LED, however after a few seconds the Pico seemingly automatically disconnects, and the BlueRetro device resumes blinking. The Pico works with my switch and the BlueRetro works with other controllers, but not with each other unfortunately. Hopefully there will be an update to have it work.
Hi! Could you solve the compatibility between BlueRetro and this project?
Hey, just got one built. The Pico seems to attempt to auto sync with a BlueRetro without a button when powered on, causing the BlueRetro device to stop blinking it's blue LED, however after a few seconds the Pico seemingly automatically disconnects, and the BlueRetro device resumes blinking. The Pico works with my switch and the BlueRetro works with other controllers, but not with each other unfortunately.
what firmwares are you guys using? My picos just brick for any FW after the initial one.
|
gharchive/issue
| 2023-06-16T03:38:22 |
2025-04-01T04:54:54.374017
|
{
"authors": [
"Roet-Ivar",
"Sbrivas",
"TheIcedTeaAddict",
"pavementtar"
],
"repo": "DavidPagels/retro-pico-switch",
"url": "https://github.com/DavidPagels/retro-pico-switch/issues/8",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2217421426
|
Build c51f988 Failing on Pterodactyl
Built using mvn
container@pterodactyl~ java -Xms128M -XX:MaxRAMPercentage=95.0 -jar Barrel.jar
Starting Barrel Proxy software
SLF4J: Failed to load class "org.slf4j.impl.StaticLoggerBinder".
SLF4J: Defaulting to no-operation (NOP) logger implementation
SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for further details.
Config file not found! Terminating...
Assuming theres a missing dep in pom.xml but thats a guess, hopefully its an easy fix
Okay so that was from a missing config (after digging through the main Barrel issues >.>) , now facing a connection issue to my bedrock server (running the right release afaik)
WARNING: An exceptionCaught() event was fired, and it reached at the tail of the pipeline. It usually means the last handler in the pipeline did not handle the exception.
io.netty.handler.codec.DecoderException: org.cloudburstmc.protocol.bedrock.codec.PacketSerializeException: Error whilst deserializing UpdateBlockPacket(flags=[], blockPosition=(13, 67, 67), definition=null, dataLayer=0)
at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:98)
at io.netty.handler.codec.MessageToMessageCodec.channelRead(MessageToMessageCodec.java:111)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:442)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412)
at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:444)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412)
at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103)
at io.netty.handler.codec.MessageToMessageCodec.channelRead(MessageToMessageCodec.java:111)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:442)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412)
at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:444)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412)
at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103)
at io.netty.handler.codec.MessageToMessageCodec.channelRead(MessageToMessageCodec.java:111)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:442)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412)
at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:440)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919)
at org.cloudburstmc.netty.handler.codec.raknet.ProxyInboundRouter.channelRead(ProxyInboundRouter.java:66)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:444)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412)
at org.cloudburstmc.netty.handler.codec.raknet.common.EncapsulatedToMessageHandler.channelRead0(EncapsulatedToMessageHandler.java:15)
at org.cloudburstmc.netty.handler.codec.raknet.common.EncapsulatedToMessageHandler.channelRead0(EncapsulatedToMessageHandler.java:8)
at io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:99)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:444)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412)
at org.cloudburstmc.netty.handler.codec.raknet.AdvancedChannelInboundHandler.channelRead(AdvancedChannelInboundHandler.java:48)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:444)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412)
at org.cloudburstmc.netty.handler.codec.raknet.AdvancedChannelInboundHandler.channelRead(AdvancedChannelInboundHandler.java:48)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:444)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412)
at org.cloudburstmc.netty.handler.codec.raknet.AdvancedChannelInboundHandler.channelRead(AdvancedChannelInboundHandler.java:48)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:444)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412)
at org.cloudburstmc.netty.handler.codec.raknet.common.RakSessionCodec.onOrderedReceived(RakSessionCodec.java:352)
at org.cloudburstmc.netty.handler.codec.raknet.common.RakSessionCodec.checkForOrdered(RakSessionCodec.java:333)
at org.cloudburstmc.netty.handler.codec.raknet.common.RakSessionCodec.handleDatagram(RakSessionCodec.java:326)
at org.cloudburstmc.netty.handler.codec.raknet.common.RakSessionCodec.channelRead(RakSessionCodec.java:217)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:442)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412)
at io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:102)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:444)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412)
at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103)
at io.netty.handler.codec.MessageToMessageCodec.channelRead(MessageToMessageCodec.java:111)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:442)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412)
at org.cloudburstmc.netty.handler.codec.raknet.AdvancedChannelInboundHandler.channelRead(AdvancedChannelInboundHandler.java:48)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:444)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412)
at org.cloudburstmc.netty.handler.codec.raknet.client.RakClientProxyRouteHandler.channelRead(RakClientProxyRouteHandler.java:50)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:442)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412)
at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:440)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919)
at io.netty.channel.nio.AbstractNioMessageChannel$NioMessageUnsafe.read(AbstractNioMessageChannel.java:97)
at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:788)
at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724)
at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650)
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562)
at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
at java.base/java.lang.Thread.run(Thread.java:829)
Caused by: org.cloudburstmc.protocol.bedrock.codec.PacketSerializeException: Error whilst deserializing UpdateBlockPacket(flags=[], blockPosition=(13, 67, 67), definition=null, dataLayer=0)
at org.cloudburstmc.protocol.bedrock.codec.BedrockCodec.tryDecode(BedrockCodec.java:59)
at org.cloudburstmc.protocol.bedrock.netty.codec.packet.BedrockPacketCodec.decode(BedrockPacketCodec.java:59)
at org.cloudburstmc.protocol.bedrock.netty.codec.packet.BedrockPacketCodec.decode(BedrockPacketCodec.java:19)
at io.netty.handler.codec.MessageToMessageCodec$2.decode(MessageToMessageCodec.java:81)
at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:88)
... 85 more
Caused by: java.lang.NullPointerException
at org.cloudburstmc.protocol.bedrock.codec.v291.serializer.UpdateBlockSerializer_v291.deserialize(UpdateBlockSerializer_v291.java:35)
at org.cloudburstmc.protocol.bedrock.codec.v291.serializer.UpdateBlockSerializer_v291.deserialize(UpdateBlockSerializer_v291.java:15)
at org.cloudburstmc.protocol.bedrock.codec.BedrockCodec.tryDecode(BedrockCodec.java:57)
... 89 more
Do you use custom blocks on your PocketMine server?
No the Bedrock server was running Vanilla Bedrock, tried each release of 1.20.70 I could find - does it require PocketMine specifically?
@DavyCraft648 this is the problem i face you need to fix this for create a egg for pterodactyl.
Ignore my grammar
@DavyCraft648 this is the problem i face you need to fix this for create a egg for pterodactyl. Ignore my grammar
Using LiteLoader I was able to connect, thats not something I can do egg side - Eggs just run files they cant code them
I've never tried it with Bedrock Dedicated Server yes, I did make it for PocketMine
Bedrock Dedicated Server doesn't have public code so I think it would be more difficult to debug it However, try turning off the settings in BDS regarding hashed block IDs
Connecting via LiteLoader though I drowned, had the same message as #3 and was unable to break blocks or swim (as I was underwater) - Assuming known problem? xD
|
gharchive/issue
| 2024-04-01T03:08:58 |
2025-04-01T04:54:54.410559
|
{
"authors": [
"DavyCraft648",
"GabBiswajit",
"ImKringle"
],
"repo": "DavyCraft648/Barrel",
"url": "https://github.com/DavyCraft648/Barrel/issues/12",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
814317702
|
issue needs/triage comment updated to include community channels
What kind of PR is this?:
/kind chore
/request @izzycsy to review
What this PR does / why we need it:
Added community channels to be part of needs/triage comment message for users to get help faster.
/request @monstrobishi to review again because it got stale
|
gharchive/pull-request
| 2021-02-23T10:15:14 |
2025-04-01T04:54:54.495936
|
{
"authors": [
"fuxingloh"
],
"repo": "DeFiCh/app",
"url": "https://github.com/DeFiCh/app/pull/540",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
221452137
|
Make all cogs be plugins.
So, I have been thinking of a few particular reason to make all cogs plugins.
Plugins have an Json Text file reader provided by the Core. This means that any changes to that file can be noticed when reloading the plugin.
When changing the current json text file for the cogs in here Any changes would not be noticed without restarting the bot. This can drastically be improved as well as improve some issues in the core itself.
Each cog can then have their own json file that contains all of their message responses. This would imporve a lot of things as well as stated from #2 on this list.
As such converting all cogs that are not plugins to plugins would be an excelent way to not only get what I want for #2 & #3 on the list but also decrease the number of restarts on the bot to restarts when dependencies update that I always pip install when I know they updated. A Prime example would be the aiohttp v2.0.7 release happening soon.
Done in https://github.com/DecoraterBot-devs/DecoraterBot-cogs/commit/f238e15121208e038f51e536ab670222ef9b6d00
|
gharchive/issue
| 2017-04-13T03:58:25 |
2025-04-01T04:54:54.934480
|
{
"authors": [
"AraHaan"
],
"repo": "DecoraterBot-devs/DecoraterBot-cogs",
"url": "https://github.com/DecoraterBot-devs/DecoraterBot-cogs/issues/2",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
735151056
|
Improve funcionality for gate and garage door
Hello,
This is a nice work, but i'd love if you also do it for a garage door, and for a gate?
Since the principal is the same its just a visual change and everyone whould love.
Thanks
Garage door:
Gate:
I was wondering if anyone has adapted this card to use for the garage door?
|
gharchive/issue
| 2020-11-03T09:15:48 |
2025-04-01T04:54:54.962182
|
{
"authors": [
"DrSpaldo",
"joaopequeno-git"
],
"repo": "Deejayfool/hass-shutter-card",
"url": "https://github.com/Deejayfool/hass-shutter-card/issues/29",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
673214300
|
each of session id will use the all padding data, if it will cause information to travel through
hi, each of session id will use the all padding data, if it will cause information to travel through?
Could you please further explain your question? I didn't understand you.
no problem, thank you for response ~
|
gharchive/issue
| 2020-08-05T02:58:09 |
2025-04-01T04:54:54.964424
|
{
"authors": [
"Songweiping",
"seven-xu"
],
"repo": "DeepGraphLearning/RecommenderSystems",
"url": "https://github.com/DeepGraphLearning/RecommenderSystems/issues/13",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
640216038
|
Why only send the position of root joint of each frame into the model as input?
Hi,
I was kind of curious that why only send the position of root joint of each frame into the model as input?
I mean after parsing the .bvh file, we could get the positions of every joint at every frame. But we just use the root joint's position as input, is it because other joints' position info is redundant?
Mant thanks!
As we only generate the rotation and global position, we find it's sufficient to use rotation as the network's input. However, we do use the position information in rec_loss and ee_loss to get more accurate result.
As we only generate the rotation and global position, we find it's sufficient to use rotation as the network's input. However, we do use the position information in rec_loss and ee_loss to get more accurate result.
Oh, it makes sense! To reconstruct the bvh file, we just need the position info of the root joint. I should have known that.
|
gharchive/issue
| 2020-06-17T07:41:36 |
2025-04-01T04:54:54.978707
|
{
"authors": [
"PeizhuoLi",
"crissallan"
],
"repo": "DeepMotionEditing/deep-motion-editing",
"url": "https://github.com/DeepMotionEditing/deep-motion-editing/issues/36",
"license": "BSD-2-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
410753386
|
Managing number of steps in a batch
iterate_rnn in Traj class makes iterator of batches. A tail of the batches are zero padded for arranging length of episodes. For this reason we couldn't control number of steps in a batch.
This problem can be solved by incorporating random_batch_rnn to iterate_rnn.
|
gharchive/issue
| 2019-02-15T12:43:58 |
2025-04-01T04:54:55.012275
|
{
"authors": [
"rarilurelo"
],
"repo": "DeepX-inc/machina",
"url": "https://github.com/DeepX-inc/machina/issues/119",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
419150742
|
fix #140
fix #140 and done run_diayn,py test in my computer.
Please check!
Could you add test for DIAYN?
Done adding a test of DIAYN.
Please check 👍
|
gharchive/pull-request
| 2019-03-10T06:03:14 |
2025-04-01T04:54:55.013505
|
{
"authors": [
"rarilurelo",
"ven-kyoshiro"
],
"repo": "DeepX-inc/machina",
"url": "https://github.com/DeepX-inc/machina/pull/170",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2076968974
|
Update network with RWS-structures
We'll update our OSM-based network to follow RWS-structures as presented 12 dec 2023 (see below).
Tasks:
[ ] generate a hydamo:sluis populated via a notebook
[ ] force the network to follow hydamo:sluis
@ngoorden; we miss in the South-West Delta:
Hartelkering
Vokeraksluizen
Bergsediepsluis
Katse Heule: https://downloads.rijkswaterstaatdata.nl/rws_legger/dwarsprofielen/48E-114-01.html
Sluiscomplex Terneuzen
Case (Borgharen):
network has node within distance defined by crest_width (doorstroombreedte) of weir-node.
Nodes and edges are moved towards weir-node
original:
moved network:
Case (linne stoney):
[ ] network has no node on edge crest_width / 2
[ ] project point on edge and split edge; creating a new node
[ ] move new node to point
original
Merwedekanaal-Utrecht-Nieuwegein (uit NIS, ter verificatie met RWS).
[ ] Koninginnensluis
[ ] Zuidersluis
[ ] Noordersluis
[ ] sluis Oog en Al
[ ] sluis bij Den Hommel
[ ] sluisje Vleutense wetering 31H-357
[ ] Gemaal, sluis en brug Vleutense Wetering 31H-356
[ ] Muntsluis
In DM zit:
Merwedekanaal: Muntsluis -Oog en Al - Noordergemaal
Merwedekanaal: Zuidersluis- syfon Noordersluis onder ARK- (Doorslagsluis) - Koninginnensluis
RWS kunstwerken in Prinses Margrietkanaal-Eemskanaal (in NIS)
[ ] Prinses Margrietsluis (Lemmer)
[ ] Terhornse sluis (staat iha open, bij Sneek)
[ ] Gaarkeuken
[ ] Dorkwerdersluis
[ ] Oostersluis
[ ] Driewegsluis
[ ] Slochtersluis
[ ] Groevesluis-Zuid
[ ] Groevesluis-Noord
[ ] Zeesluizen Farsum
In DM kennen we:
Prinses-Margrietkanaal: Lemmer -Friese boezem-Gaarkeuken
Gaarkeuken/Electraboezem-Driewegsluis-Fivolingo boezem
Gaarkeuken/Electraboezem-Oostersluis-Eemskanaal
Eelectraboezem - Reitdiep/Dorkwerd-verbindingskanaal-Eemskanaal<->Winschoterdiep
@gijsber; terechte opmerkingen. Ons lijstje bevat vrijwel uitsluitend alle kunstwerken die op de randen liggen van KRW-lichamen:
Brabantse kanalen: 1 KRW-lichaam
ARK: 1 KRW-lichaam
Lemmer-Delftzijl; geen KRW-lichaam. Ook zitten deze waarschijnlijk al in de regionale modellen
Merwedekanaal (alles niet ARK rond Utrecht) zit waarschijnlijk in het HDSR-model
Voorstel:
1 en 2 pakken we op in een verfijning van het Ribasim-HWS model; nw issue na inbouw van deze 61 complexen, sturing met hoofd- en nevenkranen.
3 en 4 pakken we op bij de integratie van de regionale modellen van Wetterskip, NZV en HDSR
Hoi @d2hydro lijkt me uitstekend. deze aanvulling was vooral om te zorgen dat we de aanvullende RWS-kunstwerken (en kanaalverbindingen) scherp hebben.
Nav gesprek met Geert:
NBMLK systeem
In het NBMLK systeem komt water vanuit de ZuidWillemsvaart (ZWV) binnen vanuit Belgie (Loozen). Onder normale omstandigheden stroomt op de kruising een klein deel rechtsaf naar Panheel, gaat het grootste deel linksaf en een beetje rechtdoor de Noordervaart op.
De oorsprong van dit water is de Maas Monsin) waar via het Maasakkoord een deel toebedeeld wordt aan de Zuid-Willemsvaart.
Geert raad aan om uiteindelijk wel het model door te trekken tot Monsin en het waterakkoord in het model op te nemen.
Als de Maas weinig water heeft komt er weinig water in de ZWV en is er te weinig voor Brabant beschikbaar. Het Kanaal Wessem-Nederweert (KWN) zakt uit vanwege de schutsluis en dit water wordt (op basis van peilbeheer) opgetopt door het gemaal Panheel. Hiervoor zit in DM een verdeelsleutel die alleen bij lage Maasdebieten water aftakt naar het kanaal Wessem-Nederweert in de wetenschap dat de ZWV op dat moment te weinig water heeft gekregen in Maastricht.
Wat we voorlopig kunnen doen is twee keer toepassen van een fractional flow, namelijk (1) bij kruising ZWV-KWN, sluis(bij lage debieten niets rechtsaf naar Panheel, bij hogere debieten een deel rechtsaf naar Panheel), (2) bij Panheel: bij lage debieten instroom naar KWN, bij hoge debieten geen instroom.
Op termijn kan dit via peilbeheer.
Sluis I vs Markkanaal. Dit zou ik nu dichtzetten en op termijn via peilbeheer Markkanaal regelen.
ARK
KWA (nog niet opgenomen) kan gekoppeld aan Lobith debiet.
Irenesluizen uiteindelijk op peilbeheer ARK-Noord. Voor nu is 25m3/s doorvoer voldoende.
Stuw Hagestein uiteindelijk op peilbeheer totdat het debiet bij de Waal (Tiel) een ondergrens bereikt. Dan gaat de keersluis van de Bernhardsluizen open en bepaalt de Waal het peil op de Nederrijn. (keersluis kan alleen hoge Waal keren, niet hoge waterstand in ARK-betuwepand). Marijkesluis staat open.
Bernhardsluizen voorlopig kunnen een vast debiet (schutverlies Waal-->ARK) hebben.
All done (long time ago)
|
gharchive/issue
| 2024-01-11T15:30:45 |
2025-04-01T04:54:55.157780
|
{
"authors": [
"DanielTollenaar",
"d2hydro",
"gijsber"
],
"repo": "Deltares/Ribasim-NL",
"url": "https://github.com/Deltares/Ribasim-NL/issues/50",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1987993149
|
Fix getindex of local truth_values with global condition_idx
This fixes a
BoundsError: attempt to access 2-element Vector{String} at index [4]
A test model that doesn't work without this is here (I also have code to generate this):
2Basins_2Pumps_1_Boezem_TargetLevelBasin1.zip
Though perhaps it's better to see if we can easily update one of our own test models to trigger this path.
[x] add test
Reported by @harm-nomden-sweco.
Added to #716.
|
gharchive/pull-request
| 2023-11-10T16:55:18 |
2025-04-01T04:54:55.160940
|
{
"authors": [
"visr"
],
"repo": "Deltares/Ribasim",
"url": "https://github.com/Deltares/Ribasim/pull/772",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2035772397
|
Foutje in voorbeeldscript van evaluate_gxg
In GitLab by @evaschoonderwoerd on May 30, 2022, 12:02
Dit staat er nu:
import imod
heads = imod.idf.open("head*.idf").sel(time=heads.time.dt.year >= 2000)
gxg = imod.evaluate.calculate_gxg(heads)
Dit moet het worden:
import imod
heads = imod.idf.open("head*.idf")
heads = heads.sel(time=heads.time.dt.year >= 2000)
gxg = imod.evaluate.calculate_gxg(heads)
In GitLab by @JoerivanEngelen on Feb 17, 2023, 10:19
Fixed in https://gitlab.com/deltares/imod/imod-python/-/merge_requests/176
|
gharchive/issue
| 2022-05-30T10:02:49 |
2025-04-01T04:54:55.163020
|
{
"authors": [
"Manangka"
],
"repo": "Deltares/imod-python",
"url": "https://github.com/Deltares/imod-python/issues/244",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2035800987
|
disv_converter: Broadcasting bug with well package
In GitLab by @JoerivanEngelen on Jul 14, 2023, 14:18
See example here
Due to a bug with broadcasting n_times too many wells are generated:
94 times & 94 indices.
The disv_converter is likely going to be removed in v1.0.
|
gharchive/issue
| 2023-12-11T14:25:02 |
2025-04-01T04:54:55.164729
|
{
"authors": [
"JoerivanEngelen",
"Manangka"
],
"repo": "Deltares/imod-python",
"url": "https://github.com/Deltares/imod-python/issues/475",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2501393787
|
Shapely STRtree versus scipy KDTree for snapping
I've currently used the scipy KDTree with a sparse distance matrix to get snapping candidates, but this obviously requires scipy.
It might be worthwhile to use the shapely STRtree instead; we just have to build the CSR form ourselves.
With some 1.3 million points:
def locate_scipy(geometry, max_distance):
xy = shapely.get_coordinates(geometry)
tree = KDTree(xy)
return tree.sparse_distance_matrix(tree, max_distance=max_distance).tocsr()
%timeit matrix = locate_scipy(out.geometry, 2.0)
1.67 s ± 55.4 ms per loop (mean ± std. dev. of 7 runs, 1 loop each)
|
gharchive/issue
| 2024-09-02T17:48:03 |
2025-04-01T04:54:55.166468
|
{
"authors": [
"Huite"
],
"repo": "Deltares/pandamesh",
"url": "https://github.com/Deltares/pandamesh/issues/40",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
506830264
|
feat(directive): add whitelist for allowed attributes on element
fix #11
:tada: This PR is included in version 2.3.0 :tada:
The release is available on:
npm package (@latest dist-tag)
GitHub release
Your semantic-release bot :package::rocket:
|
gharchive/pull-request
| 2019-10-14T19:20:22 |
2025-04-01T04:54:55.168950
|
{
"authors": [
"Demivan"
],
"repo": "Demivan/fluent-vue",
"url": "https://github.com/Demivan/fluent-vue/pull/16",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2664062862
|
Training details on data augmentation
Hi!
This work is very impressive. I have a question regarding the training pipeline for the student model with pseudo labels:
Are you following the exact same loss for the student training compared with V1?
That means: applying the data augmentation (color distortion, gaussian blurring, CutMix) during training.
Yes. But in V2, we find that when training smaller models (e.g., ViT-S and ViT-B-based models) with the pseudo label from the largest ViT-G-based model, the augmentations are not necessary.
|
gharchive/issue
| 2024-11-16T09:13:51 |
2025-04-01T04:54:55.199944
|
{
"authors": [
"LiheYoung",
"zaiweizhang"
],
"repo": "DepthAnything/Depth-Anything-V2",
"url": "https://github.com/DepthAnything/Depth-Anything-V2/issues/207",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1166654532
|
🛑 Larimarca is down
In 0cef03f, Larimarca (http://larimarca.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Larimarca is back up in 3fa2b01.
|
gharchive/issue
| 2022-03-11T17:04:11 |
2025-04-01T04:54:55.263604
|
{
"authors": [
"Deuris117"
],
"repo": "Deuris117/upptime",
"url": "https://github.com/Deuris117/upptime/issues/8",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
170622813
|
DxDataGrid CellTemplate Not Working Properly
Hi,
When a dxTemplate is defined inside a dxDataGrid and the datasource is not ready before first load it fails drawing the grid. I hope that you can reproduce the issue introducing the dxDataGrid dataSource from a service or setting it inside ngAfterViewInit().
The error: n.owner is not a function at DxTemplate. (dx.all.js:10) from _cleanPreviousContent -> o.cleanData -> _cleanTemplates
It seems that you are cleaning content of nothing.
If the dataSource is ready then it works as expected. If you have a dataSource before first load and then you change it, it works too.
Thank You
¿#85?
Hello,
Yes, looks like #85 should fix it. We will research it as soon as possible.
It seems that #85 Fixed this.
|
gharchive/issue
| 2016-08-11T11:20:37 |
2025-04-01T04:54:55.282896
|
{
"authors": [
"Aden-git",
"Angelminster"
],
"repo": "DevExpress/devextreme-angular2",
"url": "https://github.com/DevExpress/devextreme-angular2/issues/74",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
802960316
|
typeText replace option doesn't work when the input already has a value that fit maxlength
What is your Test Scenario?
Replacing the old value in the text input with a new value. The old value has a length that fit maxlength.
What is the Current behavior?
The new value doesn't replace the old value.
What is the Expected behavior?
The new value replace the old value.
What is your web application and your TestCafe test code?
Your website URL (or attach your complete example):
The example app is just a simple html form.
<form>
<input type="text" id="name" maxlength="5">
</form>
Your complete test code (or attach your test files):
import { Selector } from 'testcafe';
fixture `Example`
.page `file:///path/to/form.html`;
test('Example test', async t => {
await t
.typeText('#name', '12345') // Input the old value.
.typeText('#name', 'abcde', { replace: true }) // Replace the new value.
.expect(Selector('#name').value).eql('abcde'); // Assert the input value.
});
Your complete configuration file (if any):
Not using configuration file.
Your complete test report:
Actually it's not getting any error.
Running tests in:
- Chrome 88.0.4324.146 / macOS 10.15.7
Example
✖ Example test
1) AssertionError: expected '12345' to deeply equal 'abcde'
+ expected - actual
-'12345'
+'abcde'
Screenshots:
Steps to Reproduce:
Create an example html file.
<form>
<input type="text" id="name" maxlength="5">
</form>
Create an example test file.
import { Selector } from 'testcafe';
fixture `Example`
.page `file:///path/to/form.html`;
test('Example test', async t => {
await t
.typeText('#name', '12345') // Input the old value.
.typeText('#name', 'abcde', { replace: true }) // Replace the new value.
.expect(Selector('#name').value).eql('abcde'); // Assert the input value.
});
Execute this command
testcafe chrome example.js
Check the behavior.
Your Environment details:
testcafe version: 1.10.1
node.js version: v14.6.0
command-line arguments: testcafe chrome example.js
browser name and version: Chrome 88.0.4324.146
platform and version: macOS 10.15.7
Thank you for the example. I reproduced the issue.
@AlexKamaev
Hi, thanks for reproducing the issue.
If it's not a critical bug, I'm wondering that can I try to fix it by myself?
You need to take a look at the following method:
https://github.com/DevExpress/testcafe/blob/46460d16e044abff52bff497aaec29b20022ed21/src/client/automation/playback/type/type-text.js#L242
I modified it as follows:
function _typeTextToTextEditable (element, text) {
const elementValue = domUtils.getElementValue(element);
const textLength = text.length;
let startSelection = textSelection.getSelectionStart(element);
let endSelection = textSelection.getSelectionEnd(element);
const isInputTypeNumber = domUtils.isInputElement(element) && element.type === 'number';
if (!simulateBeforeInput(element, text, browserUtils.isChrome))
return;
let needProcessInput = simulateTextInput(element, text);
if (needProcessInput)
needProcessInput = simulateBeforeInput(element, text, browserUtils.isSafari);
if (!needProcessInput)
return;
// NOTE: the 'maxlength' attribute doesn't work in all browsers. IE still doesn't support input with the 'number' type
let elementMaxLength = !browserUtils.isIE && isInputTypeNumber ? null : parseInt(element.maxLength, 10);
if (elementMaxLength < 0)
elementMaxLength = browserUtils.isIE && browserUtils.version < 17 ? 0 : null;
const newElementValue = elementValue.substring(0, startSelection) + text + elementValue.substring(endSelection, elementValue.length);
if (elementMaxLength === null || isNaN(elementMaxLength) || elementMaxLength >= newElementValue.length) {
// NOTE: B254013
if (isInputTypeNumber && browserUtils.isIOS && elementValue[elementValue.length - 1] === '.') {
startSelection += 1;
endSelection += 1;
}
domUtils.setElementValue(element, newElementValue);
textSelection.select(element, startSelection + textLength, startSelection + textLength);
}
// NOTE: We should simulate the 'input' event after typing a char (B253410, T138385)
eventSimulator.input(element, text);
}
And now, the test is successfully passed.
If you want to create a PR, you can take my method and test it with more complex scenarios.
Also, you need to add your test to the following folder:
https://github.com/DevExpress/testcafe/tree/master/test/functional/fixtures/regression
Please take a look at any other test in this folder and make a similar one.
Thank you in advance.
|
gharchive/issue
| 2021-02-07T14:06:11 |
2025-04-01T04:54:55.294126
|
{
"authors": [
"AlexKamaev",
"rueyaa332266"
],
"repo": "DevExpress/testcafe",
"url": "https://github.com/DevExpress/testcafe/issues/5921",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2472116289
|
Blank on top window, hides everything else
Current behavior
Running the application, first displays a blank window and then the app window. The window that is displayed first, remains on top of all others, cannot be moved or closed and only goes away when the application is closed. However, it renders the application (and rest of the computer) unusable.
At times, it works as if it's click-through (even though the mouse if over it, it can affect UI elements that are behind it).
This also appeared for me when I first installed preview 4 (the first v2 I installed), but stopped after a restart. Now, after several restarts, it doesn't go away.
Also, it appears even on multiple desktops. I may open DevToys on desktop 1, switch to desktop 2 and the blank window will re-appear
How to reproduce it (as minimally and precisely as possible)
In my case, it is enough to start the app.
Expected behavior
Even if this window must open before everything else, it should be hidden/go away.
Screenshots
Workaround
None...
Affected platforms
Windows
Affected DevToys kind
DevToys (app with GUI)
DevToys Version
Version 2.0-preview.5
Relevant Assets/Logs
2024-08-18T21:38:09.9511471+02:00 Information [DevToys.Core.FileHelper] [0] Cleared temp files in 0.2746ms
2024-08-18T21:38:09.9601615+02:00 Information [DevToys.Core.Mef.MefComposer] [LogDiscoveringPlugin] Discovering plugin in 'C:\Users\there\AppData\Local\Programs\DevToys Preview\Plugins\DevToys.Tools'...
2024-08-18T21:38:10.0650859+02:00 Information [DevToys.Core.Mef.MefComposer] [LogMefComposition] MEF composed 114 parts from 6 assemblies in 107.3382ms
2024-08-18T21:38:10.0826349+02:00 Information [DevToys.Windows.MainWindow] [2] MEF, services and logging initialized in 190.7751 ms
2024-08-18T21:38:10.0827085+02:00 Information [DevToys.Windows.MainWindow] [1] App is starting...
2024-08-18T21:38:10.8626166+02:00 Information [DevToys.Windows.MainWindow] [4] Running Windows with WebView 127.0.2651.105
2024-08-18T21:38:10.8872613+02:00 Information [DevToys.Windows.MainWindow] [3] App main window's UI loaded in 804.2242 ms
2024-08-18T21:38:10.9254514+02:00 Information [DevToys.Core.Tools.GuiToolInstance] [LogInitialized] Initialized 'Base64ImageEncoderDecoder' tool instance manager.
2024-08-18T21:38:10.9268356+02:00 Information [DevToys.Core.Tools.GuiToolInstance] [LogInitialized] Initialized 'Base64TextEncoderDecoder' tool instance manager.
2024-08-18T21:38:10.9278538+02:00 Information [DevToys.Core.Tools.GuiToolInstance] [LogInitialized] Initialized 'CertificateDecoder' tool instance manager.
2024-08-18T21:38:10.9287863+02:00 Information [DevToys.Core.Tools.GuiToolInstance] [LogInitialized] Initialized 'ColorBlindnessSimulator' tool instance manager.
2024-08-18T21:38:10.9297010+02:00 Information [DevToys.Core.Tools.GuiToolInstance] [LogInitialized] Initialized 'CronParser' tool instance manager.
2024-08-18T21:38:10.9305977+02:00 Information [DevToys.Core.Tools.GuiToolInstance] [LogInitialized] Initialized 'DateConverter' tool instance manager.
2024-08-18T21:38:10.9314671+02:00 Information [DevToys.Core.Tools.GuiToolInstance] [LogInitialized] Initialized 'EscapeUnescape' tool instance manager.
2024-08-18T21:38:10.9323364+02:00 Information [DevToys.Core.Tools.GuiToolInstance] [LogInitialized] Initialized 'GZipEncoderDecoder' tool instance manager.
2024-08-18T21:38:10.9332096+02:00 Information [DevToys.Core.Tools.GuiToolInstance] [LogInitialized] Initialized 'HashAndChecksumGenerator' tool instance manager.
2024-08-18T21:38:10.9340840+02:00 Information [DevToys.Core.Tools.GuiToolInstance] [LogInitialized] Initialized 'HtmlEncoderDecoder' tool instance manager.
2024-08-18T21:38:10.9351007+02:00 Information [DevToys.Core.Tools.GuiToolInstance] [LogInitialized] Initialized 'ImageConverter' tool instance manager.
2024-08-18T21:38:10.9359755+02:00 Information [DevToys.Core.Tools.GuiToolInstance] [LogInitialized] Initialized 'JsonFormatter' tool instance manager.
2024-08-18T21:38:10.9368498+02:00 Information [DevToys.Core.Tools.GuiToolInstance] [LogInitialized] Initialized 'JSONPathTester' tool instance manager.
2024-08-18T21:38:10.9377312+02:00 Information [DevToys.Core.Tools.GuiToolInstance] [LogInitialized] Initialized 'JsonTableConverter' tool instance manager.
2024-08-18T21:38:10.9386534+02:00 Information [DevToys.Core.Tools.GuiToolInstance] [LogInitialized] Initialized 'JsonWebTokenEncoderDecoder' tool instance manager.
2024-08-18T21:38:10.9396346+02:00 Information [DevToys.Core.Tools.GuiToolInstance] [LogInitialized] Initialized 'JsonYamlConverter' tool instance manager.
2024-08-18T21:38:10.9404875+02:00 Information [DevToys.Core.Tools.GuiToolInstance] [LogInitialized] Initialized 'ListCompare' tool instance manager.
2024-08-18T21:38:10.9413381+02:00 Information [DevToys.Core.Tools.GuiToolInstance] [LogInitialized] Initialized 'LoremIpsumGenerator' tool instance manager.
2024-08-18T21:38:10.9422219+02:00 Information [DevToys.Core.Tools.GuiToolInstance] [LogInitialized] Initialized 'MarkdownPreview' tool instance manager.
2024-08-18T21:38:10.9430943+02:00 Information [DevToys.Core.Tools.GuiToolInstance] [LogInitialized] Initialized 'NumberBaseConverter' tool instance manager.
2024-08-18T21:38:10.9439496+02:00 Information [DevToys.Core.Tools.GuiToolInstance] [LogInitialized] Initialized 'PasswordGenerator' tool instance manager.
2024-08-18T21:38:10.9447856+02:00 Information [DevToys.Core.Tools.GuiToolInstance] [LogInitialized] Initialized 'QRCodeEncoderDecoder' tool instance manager.
2024-08-18T21:38:10.9456238+02:00 Information [DevToys.Core.Tools.GuiToolInstance] [LogInitialized] Initialized 'RegExTester' tool instance manager.
2024-08-18T21:38:10.9464658+02:00 Information [DevToys.Core.Tools.GuiToolInstance] [LogInitialized] Initialized 'SqlFormatter' tool instance manager.
2024-08-18T21:38:10.9473022+02:00 Information [DevToys.Core.Tools.GuiToolInstance] [LogInitialized] Initialized 'TextAnalyzerAndUtilities' tool instance manager.
2024-08-18T21:38:10.9481627+02:00 Information [DevToys.Core.Tools.GuiToolInstance] [LogInitialized] Initialized 'TextCompare' tool instance manager.
2024-08-18T21:38:10.9490527+02:00 Information [DevToys.Core.Tools.GuiToolInstance] [LogInitialized] Initialized 'UrlEncoderDecoder' tool instance manager.
2024-08-18T21:38:10.9499290+02:00 Information [DevToys.Core.Tools.GuiToolInstance] [LogInitialized] Initialized 'UUIDGenerator' tool instance manager.
2024-08-18T21:38:10.9508424+02:00 Information [DevToys.Core.Tools.GuiToolInstance] [LogInitialized] Initialized 'XmlFormatter' tool instance manager.
2024-08-18T21:38:10.9517674+02:00 Information [DevToys.Core.Tools.GuiToolInstance] [LogInitialized] Initialized 'XMLTester' tool instance manager.
2024-08-18T21:38:10.9537155+02:00 Information [DevToys.Core.Tools.GuiToolInstance] [LogInitialized] Initialized 'Extensions Manager' tool instance manager.
2024-08-18T21:38:10.9543269+02:00 Information [DevToys.Core.Tools.GuiToolInstance] [LogInitialized] Initialized 'Settings' tool instance manager.
2024-08-18T21:38:10.9550508+02:00 Information [DevToys.Core.Tools.GuiToolInstance] [LogInitialized] Initialized 'NoSearchResults' tool instance manager.
2024-08-18T21:38:10.9582272+02:00 Information [DevToys.Core.Tools.GuiToolProvider] [LogToolInstancesCreated] Instantiated 32 tools in 38.1213ms
2024-08-18T21:38:12.2389846+02:00 Information [DevToys.Blazor.Core.Services.FontService] [LogImportThirdPartyFontsAsync] Loaded 2 font(s) in 32.2464 ms and injected them in 485.4802 ms
2024-08-18T21:38:12.4395053+02:00 Information [DevToys.Blazor.BuiltInTools.ExtensionsManager.ExtensionInstallationManager] [0] Extension DevToys.Tools is up to date
2024-08-18T21:38:18.9919159+02:00 Information [DevToys.Core.Tools.GuiToolProvider] [LogSetMostRecentUsedTool] Set 'Settings' as the most recently used tool.
2024-08-18T21:38:18.9936640+02:00 Information [DevToys.Core.Tools.GuiToolInstance] [LogInstanceCreated] Instance of 'Settings' tool created.
2024-08-18T21:38:42.6608387+02:00 Information [DevToys.Core.Settings.SettingsProvider] [LogSetSetting] Setting 'DevToys.Core.MainWindowBounds' changed to '51,222,1653,1018'
2024-08-18T21:38:42.6623959+02:00 Information [DevToys.Core.Settings.SettingsProvider] [LogSetSetting] Setting 'DevToys.Core.MainWindowMaximized' changed to 'False'
2024-08-18T21:38:42.6631975+02:00 Information [DevToys.Core.FileHelper] [0] Cleared temp files in 0.0888ms
2024-08-18T21:38:42.6639622+02:00 Information [DevToys.Core.Settings.SettingsProvider] [LogSetSetting] Setting 'DevToys.Blazor.UserPreferredNavBarState' changed to 'Expanded'
Hi,
This is the first time I see such an issue. I'd love your help to understand it better as I currently do not reproduce the issue.
A few questions:
What GPU do you have?
Do you have any sort of accessibility feature turned on in Windows?
Are you able to click through that gray rectangle in the middle of the screen? Does clicking through allows to, for example, navigate to a tool?
Are able to click on components outside of that rectangle?
Is there any menu showing up when doing a right click on this rectangle?
Are you able to move the window through the title bar at the top of the window? Or do you need to click and drag that rectangle in order to move the window?
You said it happened the first time you ran 2.0.4.0, then it did not reprpduce for some time, and now it happens constantly, is that correct?
Did you install DevToys through the Microsoft Store, installer or a Zip? Does the issue reproduce if you install with one of these options different than the one you used already?
Does the issue reproduce if you switch Windows or DevToys to Dark theme?
Thanks 😃
Hey there! Thanks for the immediate response! Replies inline:
What GPU do you have?
That's actually what I'm concentrating on myself. I'm on an ASUS Zenbook with a (totally inadequate) built-in Intel IrisXE GPU. The display is a high density OLED though, which pushes the GPU. The rest of the machine specs are: 16Gb RAM/1TB SSD/Win11 fully patched. GPU details:
I'm also using the latest drivers from Intel.
Do you have any sort of accessibility feature turned on in Windows?
No, nothing that I knowingly enabled.
Are you able to click through that gray rectangle in the middle of the screen? Does clicking through allows to, for example, navigate to a tool?
Yes, when the DevToys window is active, I can click through the rectangle (which btw is not always in the center of the screen - it changes position everytime I start DevToys - it behaves as a simple window with its default settings, ie. no startup position settings in the code). Mouseover events work as well. However, if I change focus to another window (eg. Explorer), the rectangle blocks any interaction.
Are able to click on components outside of that rectangle?
Yes, I can click normally
Is there any menu showing up when doing a right click on this rectangle?
No, nothing
Are you able to move the window through the title bar at the top of the window? Or do you need to click and drag that rectangle in order to move the window?
The rectangle cannot be moved. If I click on the DevToys titlebar, I can move the window, but still the rectangle takes so much space, that I cannot do anything meaningful.
You said it happened the first time you ran 2.0.4.0, then it did not reprpduce for some time, and now it happens constantly, is that correct?
Yes and... no. After your reply I ran some more tests. I was actually able to get it working sometimes. It is happening MOST of the time. I randomly got the normal behaviour (just a few times tho). And it did show SOME consistency running ok when I set the compatibility down to Windows 7, in which case it was 90% times ok, 10% not.
Did you install DevToys through the Microsoft Store, installer or a Zip? Does the issue reproduce if you install with one of these options different than the one you used already?
Unfortunately I don't remember how I installed 2.0.4.0, even though I think it was through GitHub-downloaded installer. Then I upgraded to 2.0.5.0 However, after your message, I uninstalled it, tried the portable 64bit version, the 64bit installer, the portable 32bit version - all behaved the same.
Does the issue reproduce if you switch Windows or DevToys to Dark theme?
Yes, it happens with both themes. However, there is one interesting thing here... If I change the theme to dark, it immediately goes to black bg, white text etc as it should. If I close the application and restart it, I get the rectangle and then white bg with white text and the app is even more unusable :)
I used SysInternal's Process Explorer which has a good tool to select processes by targeting windows, but it didn't actually help. Both the normal DevToys window and the problematic rectangle actually showed the same thing:
One last thing - don't think it will help that much, but you never know. I accessed my laptop remotely (meaning I got degraded colors etc) and got this:
I know it's not much, but there IS a pattern in there (which I can't see when I'm on the laptop). Maybe that pattern is familiar?
As I said in the beginning, it feels much more a problem of my GPU rather than the app. On my second Win11 machine and Mac it works fine :)
Still, I thought it'd be good to report it...
Let me know if this helped or if you need anything more...
Hey, after getting to the laptop and checking this post, I realised that the last image did not make sense. I tweaked it enough to make the design pattern stand out... Colours are messed up. Just note the design in case it rings a bell - not the colours.
Hi,
Thank you for all these details! It's very helpful.
It really sounds like a GPU-related issue, indeed. It seems like the background / Mica effect of the window is bleeding through the web view that displays DevToys UI.
Here is a potential workaround: Can you try disabling transparency effect in Windows? You can find it in Windows Settings app > Personalization > Colors > Transparency effects.
I'm also curious whether you encounter these "odd" colors in different part of Windows 11, such as Start Menu, Widgets, Settings app, Calculator app, Emoji (Win+;), Clipboard History (Win+V), Sounds setting flyout (Ctrl+Win+V).
Thank you :)
Hi again
ok, we got a winner!
As expected, it is GPU-related. it was actually not the Windows transparency effects, but your mention of the Mica effect, jolted my memory. Few months ago, for the heck of it I installed an app called Mica or Everyone (https://github.com/MicaForEveryone/MicaForEveryone)
THIS one is what was causing the issue. I don't actually know why, but closing the app or just disabling customizations for the DevToys process, is enough to fix the issue. I even tried upgrading MfE to the latest version, but it didn't help.
Any customization I try to apply using it, causes DevToys to show that blank window. So I just set it to ignore DevToys and it's all good!
Thanks so much for your time! Keep up the good work... Hopefully I'll be able to contribute with a couple of plugins in the near future...
Closing this issue
Glad you got it fixed 🙂
Just if you're curious, I've been chatting with the author of Mica For Everyone on the matter:
Wow, thanks for investigating this even further! You're the best!
|
gharchive/issue
| 2024-08-18T19:41:10 |
2025-04-01T04:54:55.336677
|
{
"authors": [
"theremon",
"veler"
],
"repo": "DevToys-app/DevToys",
"url": "https://github.com/DevToys-app/DevToys/issues/1385",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1402351216
|
Duplicate Section of FAQ, Tracks
Scrolling through the Devtrack website, I noticed that there were duplicate sections (What do we do? Who are we? cards, FAQ, and Tracks) present. I have attached the screenshot. I would like to work on this issue.
@harshmehta14 you are assigned.
Thanks for assigning it to me. Can you tell me which screen would you like to keep?
2nd screenshot you keep.
|
gharchive/issue
| 2022-10-09T18:03:25 |
2025-04-01T04:54:55.339951
|
{
"authors": [
"PGautam27",
"harshmehta14"
],
"repo": "DevTrackClub/DevTrackWebsite",
"url": "https://github.com/DevTrackClub/DevTrackWebsite/issues/51",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
428354012
|
Layout de detalhes da Venda Quebrado
Ao instalar o módulo o layout fica assim , mas quando desativo o layout volta ao normal!
[2019-04-04 19:25:46] main.CRITICAL: Unsuccessful. Please, contact Rede. {"exception":"[object] (Rede\Exception\RedeException(code: 370): Unsuccessful. Please, contact Rede. at /var/www/html/XXXX/siteroot/vendor/developersrede/erede-php/src/Rede/Service/AbstractTransactionsService.php:86)"}
A forma de pagamento não é rede , minha loja usa dois métodos . acredito que esse erro é referente a pedidos que não estão na base da rede . Podemos pensar em uma solução juntos .
|
gharchive/issue
| 2019-04-02T17:33:19 |
2025-04-01T04:54:55.352908
|
{
"authors": [
"viniqueiroz"
],
"repo": "DevelopersRede/magento2",
"url": "https://github.com/DevelopersRede/magento2/issues/4",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2342034260
|
Requirement for PyGObject <3.47 causes conflicts on Ubuntu 24.04
Commit DevilXD/TwitchDropsMiner@27dcc9c pinned the PyGObject library to version <3.47 to maintain compatibility with Ubuntu 20.04.
The default version of the gir1.2-gtk-3.0 package on 24.04, which to my knowledge is a dependency of the app, installs a version of PyGObject >3.47.
Pip will try to remove the existing newer package when installing the older one. However, since the package wasn't installed using pip, it cannot be uninstalled using pip.
I am not sure of a solution that would work for every case, but one possible solution for users may be to install an older version of the gir1.2-gtk-3.0 package if possible. I am yet to test this workaround.
In my situation this issue occurred in a docker container with a 24.04 base image. For my use case, it was sufficient to just change the base image to 22.04 and tweak a couple things (SimonAkers/docker-tdm@03c4bfa). I wanted to document this as an issue in case it shows up for someone else.
This problem can be entirely avoided by using a virtual environment to install the dependencies, so they don't conflict with system packages. You were installing the dependencies with --break-system-packages, which is usually a really bad idea.
Besides, unless you have some specific reason to require the latest version of Ubuntu, I'd recommend staying on the oldest LTS version anyway.
@guihkx thanks for the reply, here is my thought process.
This problem can be entirely avoided by using a virtual environment to install the dependencies, so they don't conflict with system packages.
To build some of the dependencies you need access to libraries which cannot easily be installed inside a venv (to my knowledge) such as for cairo. There exists vext, but I do not believe it supports all required packages. I may be wrong though, or there may be another solution.
I would appreciate any advice on how to run TDM strictly inside a venv if possible, as I'm aware that's an ideal solution.
You were installing the dependencies with --break-system-packages, which is usually a really bad idea.
For my use case in a rough docker container, it's acceptable usage. If it were on my host machine, then I agree it would not be very safe. However, if I can get everything inside a venv, I would not need to worry about it. Any advice?
I think you're mixing things up.
PyGObject only depends on pycairo, and so to build it, you only need the headers package (libcairo2-dev, which you do), and a compiler toolkit (usually just installing the build-essential package group is enough).
When those two main things are ready, running pip install will build and install those dependencies from source automatically for you.
I would appreciate any advice on how to run TDM strictly inside a venv if possible, as I'm aware that's an ideal solution.
I'm honestly not too experienced with Docker, but you can probably come up with something adapted from this article I found.
@SimonAkers Docker usage is against the Project Goals, and is thus unsupported. If you run into issues, you're on your own.
Project Goals can be found here: https://github.com/DevilXD/TwitchDropsMiner?tab=readme-ov-file#project-goals
@guihkx While I really appreciate you helping other people with their Linux-related issues, please remember that most of the existence of this project depends solely on my availability, and I'm already struggling with keeping up to Twitch changing things around (see #462). TDM was never meant to be deployed anywhere, and making that easier leads us closer to people deploying farms with thousands of instances, and then Twitch putting in efforts to stop that and prevent TDM from working. This has already happened multiple times in the past (read more under the project goals), and while we can't prevent it entirely, we can at least try to discourage doing so.
@DevilXD Appreciate your work, and I understand your reasons for not supporting docker.
My motivations for opening this issue were not to promote docker usage, but to make this potential issue known. It's not necessarily limited to docker, this issue can happen in standard Ubuntu 24.04 desktop. So I wanted to put any potential solutions out there in case a typical user has the problem.
I agree with closing the issue though, it's not something you technically are responsible for anyways. Again, thanks for your hard work!
I see. Thank you for understanding =)
|
gharchive/issue
| 2024-06-09T05:38:51 |
2025-04-01T04:54:55.367323
|
{
"authors": [
"DevilXD",
"SimonAkers",
"guihkx"
],
"repo": "DevilXD/TwitchDropsMiner",
"url": "https://github.com/DevilXD/TwitchDropsMiner/issues/496",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1669400021
|
Add Command Line Interface mode
Add support for pure Command Line Interface, by reimplementing GUIManager as CLIManager, for the case running on server without GUI Desktop as a 24*7-hour bot.
In the CLI mode, I only implemented the display of a few core information, because many functions are probably not required when it is simply running as a bot on a server (and it requires less work). Configuration through the user interface is not supported, so you need to edit settings.json manually (should not be a problem for someone using command line).
Logging in is supported. The user just needs to open the link in their browser (not required to be on the same machine as the server) and enter the code.
Usage
Add --cli argument to launch in CLI mode.
Screenshot
It works fine on my Ubuntu server:
Help wanted: In main.py, to make it works with CLI mode, I have to move the parser to before creating the window temporarily, so that works in the following flow:
parse arguments
skip creating window if --cli flag is set (or else an error will occur if it tries to create a window in a non-GUI environment)
In the current implementation, an error will occur if the parser output something to the message box. A better method is needed, maybe direct the parser output to the console? I am not sure whether it will bring any additional problems, so I did not fix it.
Just from my personal perspective: I'm a user in China (UTC+8 Time zone), where some Twitch live streaming in place like Europe starts at an unfriendly local time (like at mid-night) here. I don't want to keep my laptop on and idling for a whole night, but sometimes if I don't it's hard to get the full drop rewards. (In particular, the game I'm playing now, "Tom Clancy's Rainbow Six Siege", requires 4 hours' watching to get a single drop. And sometimes drops are even only enable for official live rooms, so I have to follow the official campaign schedule to get the full rewards.)
In this case, the best solution is to run this program on my VPS (and it's display-less). That's the initial reason why I wrote this fork.
It might be more convenient of allowing it to run on server non-stop, even for personal use. Also this PR doesn't cancel the limit of linked accounts, maybe still be able to prevent abuse to some degree (though adding CLI mode still facilitates for multi-instance mining and profiting, of course).
Anyway, I can understand your considerations if you decide to refuse adding this feature eventually. And thanks for your review comments <3
Honestly this is 1 of 2 things I really hope gets implemented.. the second hopefully being a lot easier once this gets implemented and that would be getting this to run in docker.
@JourneyOver You should check out the project goals: https://github.com/DevilXD/TwitchDropsMiner#project-goals
This PR isn't getting merged, because it leads to abuse. It goes as far as me considering pulling down the entire project, but it hasn't been bad enough yet to do so. At least as far as I'm aware.
I honestly have no idea, why do you consider using it through cli to be "abuse".
It's not like without it it's impossible to set it up on some private server to run 24/7. You literally designed it to launch in a tray on startup and mine as long as pc is running.
I'm in a situation, where there's a surprise campaign running, but out of like 50 enabled channels there's only one live and I'm not sure if I'd be able to get all drops before campaign ends, especially if that channel goes offline. I'd like to leave this program running overnight, but my gaming pc's fans are quite loud even when idling, so I'd prefer not to have to sleep in a room with pc on. I could run it on a raspberry pi that I own, which has passive cooling and is completely silent, but it's not plugged to any monitors, so it boots in headless mode and it would be a lot of hassle to boot it in a way where I can connect via vitual desktop.
So I'm failing to understand your thought process. I could run it 24/7 in a background on my pc. I could run it 24/7 on any remote server with remote desktop protocol. But the convenience of being able to run it on my own machine that happens to boot without gui is just too much, and you would consider this "abuse"?
@ignis05 The reasoning behind it is already explained in the Project Goals section, a link to which you seem to have intentionally ignored. There are links to issues within said section, that exactly explain why I really don't want this project to be possible to run in headless mode. I wanted to do it at the beginning, but I don't want to now. It only leads to abuse and nothing else. I don't want thousands of instances running on some remote server, for profit purposes.
In case you still can't understand it - this is a hobby project. I'm a one-man development team, currently taken away by a 6/10 work cycle since a month ago, and for the next month or so. I don't have the time to resolve the existing issues, yet alone design a foolproof system that won't end up crashing and sending the miner into an endless restart loop. Even if I wanted to, this project is far from being ready to become "headless" in a sense where you can "set and forget" it.
In other words, this project isn't ready for headless, and I'm not doing anything to help it, because I don't have the time for it, and it only leads to abuse, so why bother. Due to abusers, we can't have nice things - this is the sad truth of it.
For more abuse examples, see #84, also linked in the project goals section: https://github.com/DevilXD/TwitchDropsMiner#project-goals
I did not ignore anything, it's just that the project goals don't really mention anything other than it being possible to "abuse" and being at the bottom of the priority list. While on this pr you simply said This PR isn't getting merged, because it leads to abuse..
Like I said, it's not like running it in its current state as a 24/7 farm is impossible, it's just a bit more tedious to set up. I understand your reasons for not wanting your personal project to be used against your will, even if I disagree with the sentiment that automatically farming drops with like 60% uptime on a personal pc is great, but doing the same thing on a headless server with better uptime or including farming channel points and drops for unlinked accounts is taking it too far and straight up abuse. But I guess when it comes to morality, everyone draws the line one step ahead of where they stand.
You asked earlier in the thread in regard to the headless mode, Why exactly is this needed for a personal, single-account utility program like this one?, I just tried to give my perspective on a scenario that I was in that I believe might have met those criteria.
But while I might disagree on what counts as abuse, I understand that in the end, this is your project, and you are free to do anything you want with it. I appreciate the unpaid work you've already put into this and that you chose to share it with everyone, and I wish you the best, regardless of what you choose to do with this pr.
Personally, I use it on a headless Linux server, as I described in #17. Once you figure it out, it really requires little to no effort already. I even have a script that can updated to latest Master (.bat), so I could have it be 0 maintenance if I combined those.
The reason for doing this is partly because of #80 and partly so that I don't have to have my PC running 24/7 (for drops that are only one stream from one channel during a specific event), which spares additional wear on my PC and saves power.
Given it's really easy to implement headless, whether by finding this pull request, the docker pull request, or using Xvfb as I did, any server farm described in https://github.com/DevilXD/TwitchDropsMiner/issues/84#issuecomment-1331709361 could easily be deployed anyways. I think it's reasonable to assume, that anyone seeking profit would be willing to put the extra bit of effort in and thus it's only really inconveniencing ethical users, who might not have the knowledge, time or incentive to implement the bot in this way.
This project started as a small, personal script, that I eventually had an idea of sharing with others. From the very beginning of this project, the idea was to have a console application, capable of being deployed on a remote server and mostly forgotten about. As a personal bot of sorts, it'd take care of only a single Twitch account, and that's it. A simple, helpful tool, reliving you of having to keep track of which channels are online, changing the channel when the watching one goes offline, and having to be there on time to claim the drop. All of this info is still present in the README.md file, with only a couple of corrections here and there, ever since the first very few versions have been published.
However, some people didn't understand what this project is for. They wanted to "go big" - and have massive farms of hundreds, thousands of accounts, all mining drops for personal profit. I've been literally involuntarily dragged into a "deal" between two people, where one was selling a "server with 10 instances on it" to the other, where the other person was contacting me and demanding to know when I'd fix a critical-at-the-time issue (#68), preventing the miner from claiming the drops.
It changed me. I no longer want to make it possible to run this on a remote server, and this is the least I can do to try and limit the misuse of this project. Yes, it makes it "inconvenient" for people who would want to do that, but as I cannot control how many instances you're going to run on that servers of yours, I cannot allow for it. I do not support misuse and abuse - and if it's going to continue, this project will simply be discontinued entirely. I'm not going to waste my time for the profit of others, since again - this isn't the purpose of this project.
anyone seeking profit would be willing to put the extra bit of effort in
There are other projects like this out there, that make it much easier to deploy. I don't get why someone cannot just use that one instead. As far as my own personal experience goes, "putting in effort" is enough of a deterrent to limit the abuse on it's own. There's no way to stop it, unfortunately, as any open-source project can be repurposed to bypass it's limitations, with enough effort put in. It's been doing a good enough job, ever since "the change" happened, and I didn't head from anyone trying to make a giant farm since.
thus it's only really inconveniencing ethical users, who might not have the knowledge, time or incentive to implement the bot in this way
Good. Again, being able to run this on a remote server isn't the purpose of this project - not anymore. Nobody forces anyone to use this thing either. If one doesn't find it suitable for their purpose - they shouldn't use it. Simple. If anyone would like to have a miner they can deploy on a server, as I said before, there are other, more suitable projects out there, that can do exactly that, Please use that one instead.
While I disagree, it's a matter of opinion at this point. At the end of the day, I'm just grateful you publish your work for free, making it available to all of us. Thank you. <3
|
gharchive/pull-request
| 2023-04-15T13:36:35 |
2025-04-01T04:54:55.385280
|
{
"authors": [
"DevilXD",
"Fidelxyz",
"JourneyOver",
"Windows200000",
"ignis05"
],
"repo": "DevilXD/TwitchDropsMiner",
"url": "https://github.com/DevilXD/TwitchDropsMiner/pull/173",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2513104523
|
Last changes similar to specimen-procesor
Changes after functional testing, similar changes as we made in the processing service
Fixed the test coverage
|
gharchive/pull-request
| 2024-09-09T07:11:07 |
2025-04-01T04:54:55.406215
|
{
"authors": [
"samleeflang"
],
"repo": "DiSSCo/dissco-core-digital-media-object-processor",
"url": "https://github.com/DiSSCo/dissco-core-digital-media-object-processor/pull/39",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
489124946
|
[pact-jvm-provider-junit5] GC overhead limit exceeded
Hi, I'm using the latest pact-jvm-provider-junit5 beta, (but the issue seems to occur with previous versions too) and trying to run a simple test like this
@Provider("valkyrie")
@PactBroker(tags = ["master"], host = "host.url", scheme = "https")
@Tag("pact")
class ValkyrieProviderTest {
@TestTemplate
@ExtendWith(PactVerificationInvocationContextProvider::class)
fun test(context: PactVerificationContext) {
context.verifyInteraction()
}
@BeforeEach
fun setup(context: PactVerificationContext) {
context.target = AmpqTestTarget()
}
@PactVerifyProvider("a hermes verdict")
fun `an hermes verdict is generated`(): String {
val stream = mock<MessageChannel>()
val outboundStreamingChannel = mock<OutboundStreamingChannel> {
on { stream() } doReturn stream
}
val producer = VerdictStreamProducer(...)
producer.transmit(...)
val messageCaptor = ArgumentCaptor.forClass(GenericMessage::class.java)
verify(stream).send(messageCaptor.capture(), any())
return messageCaptor.value.payload as String
}
}
When running it from the IDE, it works perfectly fine, however, when running it from gradle using a custom test task, see below, it constantly crashes.
tasks {
create<Test>("pact") {
useJUnitPlatform { includeTags("pact") }
// set system properties for pact
}
}
Verification Failed - Uncaught exception during scan
Failures:
0) a hermes verdict
Uncaught exception during scan
0 - Uncaught exception during scan
java.lang.OutOfMemoryError: GC overhead limit exceeded
java.lang.AssertionError:
0 - Uncaught exception during scan
java.lang.OutOfMemoryError: GC overhead limit exceeded
at au.com.dius.pact.provider.junit5.PactVerificationContext.verifyInteraction(PactJUnit5VerificationProvider.kt:78)
at com.olx.karma.valkyrie.pact.ValkyrieProviderTest.test(ValkyrieProviderTest.kt:38)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.junit.platform.commons.util.ReflectionUtils.invokeMethod(ReflectionUtils.java:532)
at org.junit.jupiter.engine.execution.ExecutableInvoker.invoke(ExecutableInvoker.java:115)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.lambda$invokeTestMethod$6(TestMethodTestDescriptor.java:171)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:72)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.invokeTestMethod(TestMethodTestDescriptor.java:167)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.execute(TestMethodTestDescriptor.java:114)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.execute(TestMethodTestDescriptor.java:59)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$4(NodeTestTask.java:108)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:72)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:98)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:74)
at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.submit(SameThreadHierarchicalTestExecutorService.java:32)
at org.junit.platform.engine.support.hierarchical.NodeTestTask$DefaultDynamicTestExecutor.execute(NodeTestTask.java:157)
at org.junit.jupiter.engine.descriptor.TestTemplateTestDescriptor.execute(TestTemplateTestDescriptor.java:133)
at org.junit.jupiter.engine.descriptor.TestTemplateTestDescriptor.lambda$execute$2(TestTemplateTestDescriptor.java:101)
at java.util.stream.ForEachOps$ForEachOp$OfRef.accept(ForEachOps.java:184)
at java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
at java.util.stream.ReferencePipeline$2$1.accept(ReferencePipeline.java:175)
at java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
at java.util.ArrayList$ArrayListSpliterator.forEachRemaining(ArrayList.java:1382)
at java.util.stream.ReferencePipeline$Head.forEach(ReferencePipeline.java:580)
at java.util.stream.ReferencePipeline$7$1.accept(ReferencePipeline.java:270)
at java.util.ArrayList$ArrayListSpliterator.forEachRemaining(ArrayList.java:1382)
at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:482)
at java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:472)
at java.util.stream.ForEachOps$ForEachOp.evaluateSequential(ForEachOps.java:151)
at java.util.stream.ForEachOps$ForEachOp$OfRef.evaluateSequential(ForEachOps.java:174)
at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
at java.util.stream.ReferencePipeline.forEach(ReferencePipeline.java:418)
at org.junit.jupiter.engine.descriptor.TestTemplateTestDescriptor.execute(TestTemplateTestDescriptor.java:101)
at org.junit.jupiter.engine.descriptor.TestTemplateTestDescriptor.execute(TestTemplateTestDescriptor.java:38)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$4(NodeTestTask.java:108)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:72)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:98)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:74)
at java.util.ArrayList.forEach(ArrayList.java:1257)
at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.invokeAll(SameThreadHierarchicalTestExecutorService.java:38)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$4(NodeTestTask.java:112)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:72)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:98)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:74)
at java.util.ArrayList.forEach(ArrayList.java:1257)
at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.invokeAll(SameThreadHierarchicalTestExecutorService.java:38)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$4(NodeTestTask.java:112)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:72)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:98)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:74)
at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.submit(SameThreadHierarchicalTestExecutorService.java:32)
at org.junit.platform.engine.support.hierarchical.HierarchicalTestExecutor.execute(HierarchicalTestExecutor.java:57)
at org.junit.platform.engine.support.hierarchical.HierarchicalTestEngine.execute(HierarchicalTestEngine.java:51)
at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:220)
at org.junit.platform.launcher.core.DefaultLauncher.lambda$execute$6(DefaultLauncher.java:188)
at org.junit.platform.launcher.core.DefaultLauncher.withInterceptedStreams(DefaultLauncher.java:202)
at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:181)
at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:128)
at org.gradle.api.internal.tasks.testing.junitplatform.JUnitPlatformTestClassProcessor$CollectAllTestClassesExecutor.processAllTestClasses(JUnitPlatformTestClassProcessor.java:102)
at org.gradle.api.internal.tasks.testing.junitplatform.JUnitPlatformTestClassProcessor$CollectAllTestClassesExecutor.access$000(JUnitPlatformTestClassProcessor.java:82)
at org.gradle.api.internal.tasks.testing.junitplatform.JUnitPlatformTestClassProcessor.stop(JUnitPlatformTestClassProcessor.java:78)
at org.gradle.api.internal.tasks.testing.SuiteTestClassProcessor.stop(SuiteTestClassProcessor.java:61)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:36)
at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:24)
at org.gradle.internal.dispatch.ContextClassLoaderDispatch.dispatch(ContextClassLoaderDispatch.java:33)
at org.gradle.internal.dispatch.ProxyDispatchAdapter$DispatchingInvocationHandler.invoke(ProxyDispatchAdapter.java:94)
at com.sun.proxy.$Proxy2.stop(Unknown Source)
at org.gradle.api.internal.tasks.testing.worker.TestWorker.stop(TestWorker.java:132)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:36)
at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:24)
at org.gradle.internal.remote.internal.hub.MessageHubBackedObjectConnection$DispatchWrapper.dispatch(MessageHubBackedObjectConnection.java:182)
at org.gradle.internal.remote.internal.hub.MessageHubBackedObjectConnection$DispatchWrapper.dispatch(MessageHubBackedObjectConnection.java:164)
at org.gradle.internal.remote.internal.hub.MessageHub$Handler.run(MessageHub.java:412)
at org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:64)
at org.gradle.internal.concurrent.ManagedExecutorImpl$1.run(ManagedExecutorImpl.java:48)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at org.gradle.internal.concurrent.ThreadFactoryImpl$ManagedThreadRunnable.run(ThreadFactoryImpl.java:56)
at java.lang.Thread.run(Thread.java:748)
Am I doing something wrong? Moreover, this failure occurs only with an AmpqTestTarget and not with a HttpTarget
As explained in https://stackoverflow.com/questions/1393486/error-java-lang-outofmemoryerror-gc-overhead-limit-exceeded, your test must be generating a lot of temporary objects. And the error Uncaught exception during scan is probably because you have a large classpath that is being scanned to find the annotated method to call.
Can you provide the Java version you are using as well as the memory options used (like -Xmx..). Also, if you enable the system property pact.verifier.classpathscan.verbose in the tests it might show what it is doing.
You can also try restricting the scan to just your packages by providing the list of packages to the AmpqTestTarget() constructor.
Thanks @uglyog, with the package passed as parameter the issue disappeared. I discovered now that I have a couple of jvm settings in the gradle.properties file, so, that might explain the difference between an IDE run and a Gradle one.
|
gharchive/issue
| 2019-09-04T12:18:46 |
2025-04-01T04:54:55.412695
|
{
"authors": [
"gmariotti",
"uglyog"
],
"repo": "DiUS/pact-jvm",
"url": "https://github.com/DiUS/pact-jvm/issues/933",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1394874517
|
Race condition on stale params
We currently have a race condition where we:
Set parameters on the detector
Stale params has not yet gone high
We wait for stale params to go low but it's already low
We set the bit depth on the filewriter to that on the detector but the detector one is "stale"
e.g.:
Params set up here....
[2022-10-03 15:48:13,449] Artemis eiger INFO: before wait: stale_params 0 bit_depth 32
Wait for 0 here (but still 0)
[2022-10-03 15:48:13,451] Artemis eiger INFO: after wait: stale_params 0 bit_depth 32
Start waiting for 0 again:
[2022-10-03 15:48:13,552] Artemis eiger INFO: after wait 0: stale_params 1 bit_depth 32
[2022-10-03 15:48:13,653] Artemis eiger INFO: after wait 1: stale_params 1 bit_depth 32
[2022-10-03 15:48:13,754] Artemis eiger INFO: after wait 2: stale_params 1 bit_depth 32
[2022-10-03 15:48:13,855] Artemis eiger INFO: after wait 3: stale_params 1 bit_depth 32
[2022-10-03 15:48:13,956] Artemis eiger INFO: after wait 4: stale_params 1 bit_depth 16
[2022-10-03 15:48:14,057] Artemis eiger INFO: after wait 5: stale_params 1 bit_depth 16
[2022-10-03 15:48:14,159] Artemis eiger INFO: after wait 6: stale_params 1 bit_depth 16
[2022-10-03 15:48:14,260] Artemis eiger INFO: after wait 7: stale_params 1 bit_depth 16
[2022-10-03 15:48:14,361] Artemis eiger INFO: after wait 8: stale_params 1 bit_depth 16
[2022-10-03 15:48:14,462] Artemis eiger INFO: after wait 9: stale_params 1 bit_depth 16
[2022-10-03 15:48:14,462] Artemis eiger INFO: Setting to data type UInt16, currently at 16
Potential solution is immediately after setting the detector params monitor stale params and confirm it goes high -> low
Any ideas on a good way to do this @GDYendell?
Are you doing put with callback on the parameters? That should block until stale is set high.
Not currently, will there be any timing differences between:
Put and wait for callback for each parameter
Put to all params, gathering all the callback as we go then wait for all callbacks at the end
Do you have a preference? I think 2 is cleaner
Either should be fine.
|
gharchive/issue
| 2022-10-03T14:52:07 |
2025-04-01T04:54:55.416917
|
{
"authors": [
"DominicOram",
"GDYendell"
],
"repo": "DiamondLightSource/python-artemis",
"url": "https://github.com/DiamondLightSource/python-artemis/issues/241",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
1323269409
|
⚠️ Halloluise has degraded performance
In cab5be8, Halloluise (https://halloluise.de) experienced degraded performance:
HTTP code: 200
Response time: 3334 ms
Resolved: Halloluise performance has improved in b841ea8.
|
gharchive/issue
| 2022-07-30T19:43:12 |
2025-04-01T04:54:55.419583
|
{
"authors": [
"DianaHaidarbaigi"
],
"repo": "DianaHaidarbaigi/LuisenGym",
"url": "https://github.com/DianaHaidarbaigi/LuisenGym/issues/93",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2680994685
|
Command Migrations, Async Server removal
What's included in this PR:
Integration tests removal/move from async to resp (This makes it ready to remove async server)
Command seggregation at /eval/commands.go file fo better readability
Command migration: HELLO, SLEEP, FLUSHDB, COMMAND
Document removal for BIGWRITEAOF, SELECT
Removed QWatch and QUnwatch (commands, tests)
Removed AsyncServer
Multithreading is enabled by default, you cannot stop it, you need to use numshards = 1 if you want a single shared system
I have removed support for bitop @apoorvyadav1111
|
gharchive/pull-request
| 2024-11-21T21:24:12 |
2025-04-01T04:54:55.422020
|
{
"authors": [
"AshwinKul28"
],
"repo": "DiceDB/dice",
"url": "https://github.com/DiceDB/dice/pull/1318",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
2562706724
|
[#835] docs: updated LPUSH documentation
Updates documentation for LPUSH command as per #835
Thanks for the contribution @Aditya-Chowdhary. Thanks for the reviews @apoorvyadav1111. LGTM.
|
gharchive/pull-request
| 2024-10-02T22:01:33 |
2025-04-01T04:54:55.422987
|
{
"authors": [
"Aditya-Chowdhary",
"JyotinderSingh"
],
"repo": "DiceDB/dice",
"url": "https://github.com/DiceDB/dice/pull/932",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
1428294974
|
Version 3.0.0 release tracker
The 3.0.0 release is here! New major changes are here, the deps have been highly updated, and more further plans will come.
See resume of the changes
Community stuff
Notify if the Windows executable build is skipped (aaec68e4c16e1a272be386147da993bb1a1fbdc6, a1d2ca898f346c4731093711ada289528dcf70e0)
Update the guide (a8fa0b4c2343bea0571d48ce414a9593492682cd, 06bd147415a3a8cd44d6f06941d4c3a3999eca0b)
New behavior, or gaming-related stuff
The title menu! A new "welcome interface" for starting the game! (#60, #64)
Background stars! The background is now more colorful with a lot of stars (#73)
Dependencies
Updated dependencies:
Pyxel: 1.7.1 -> 1.8.22 (#54, #61, #63, #65, #66, #67, #68, #69, #70, #71, #72, #78, #80, #81, #82)
Black: 22.3.0 -> 22.10.0 (#50, #62, #74)
flake8: 4.0.1 -> 5.0.4 (#55, #57, #58)
setup-python (CI): 3 -> 4.3.0 (#49, #53, #56, #75)
Coming soon...
Since Pyxel now supports web support, we're planning to add an in-browser option to play!
I'll follow the same process I've used in previous releases, just give me some time :)
https://github.com/DiddiLeija/diddi-and-the-bugs/releases/tag/3.0.0
I'm currently working on the itch.io thing :P
Done! https://diddileija.itch.io/diddi-and-the-bugs/devlog/445179/version-300
Closing for now, since we finished the release process :)
|
gharchive/issue
| 2022-10-29T14:41:41 |
2025-04-01T04:54:55.429638
|
{
"authors": [
"DiddiLeija"
],
"repo": "DiddiLeija/diddi-and-the-bugs",
"url": "https://github.com/DiddiLeija/diddi-and-the-bugs/issues/83",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
355452681
|
3rd party drivers don't update
I've heard reports that 3rd party drivers don't update (--ext-co isn't fetching updated packages?), but so far I've been unable to replicate, as ebuild.sh seems to actually just nuke any 3rd party package directories, then just clones them again.
Can't replicate and the ebuild.sh seems solid enough.
|
gharchive/issue
| 2018-08-30T07:09:44 |
2025-04-01T04:54:55.430921
|
{
"authors": [
"Dids"
],
"repo": "Dids/clobber",
"url": "https://github.com/Dids/clobber/issues/7",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2360189016
|
🛑 Email Service - Datascan is down
In 622b9d7, Email Service - Datascan (https://marge.datascan-site.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Email Service - Datascan is back up in d5445e1 after 20 minutes.
|
gharchive/issue
| 2024-06-18T15:51:07 |
2025-04-01T04:54:55.455626
|
{
"authors": [
"gsantovena"
],
"repo": "Digifact-FEL/uptime_monitoring",
"url": "https://github.com/Digifact-FEL/uptime_monitoring/issues/116",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
140420613
|
Abstract complex logic into a new Model
This should allow all important operations to be called internally from Python, as well as via the HTTP API.
@manthey Can you take a look at this? I'm having trouble setting up the tests locally, so I'm not sure exactly why they're failing.
@manthey PTAL.
@brianhelba I changed what exceptions are thrown, since we don't want Rest exceptions in the general model. I'll merge this in; if you want changes, we'll do it as another PR.
This is perfect, thanks @manthey.
|
gharchive/pull-request
| 2016-03-12T22:01:02 |
2025-04-01T04:54:55.634532
|
{
"authors": [
"brianhelba",
"manthey"
],
"repo": "DigitalSlideArchive/large_image",
"url": "https://github.com/DigitalSlideArchive/large_image/pull/32",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
813144209
|
🛑 ZL1OTD Repeater is down
In e8ab060, ZL1OTD Repeater (https://zl1otd.dvnz.nz) was down:
HTTP code: 502
Response time: 19448 ms
Resolved: ZL1OTD Repeater is back up in e57e2b8.
|
gharchive/issue
| 2021-02-22T05:45:08 |
2025-04-01T04:54:55.637187
|
{
"authors": [
"ZL2RO"
],
"repo": "DigitalVoiceNZ/upptime",
"url": "https://github.com/DigitalVoiceNZ/upptime/issues/113",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1693398314
|
🛑 API - Debank Gas Price Dict is down
In fa50fad, API - Debank Gas Price Dict ($DEBANK_GAS_PRICE_DICT_API_URL) was down:
HTTP code: 429
Response time: 597 ms
Resolved: API - Debank Gas Price Dict is back up in 88a7a58.
|
gharchive/issue
| 2023-05-03T04:29:44 |
2025-04-01T04:54:55.643121
|
{
"authors": [
"AutomationDimension"
],
"repo": "DimensionDev/status",
"url": "https://github.com/DimensionDev/status/issues/1254",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1787675376
|
🛑 API - Debank Gas Price Dict is down
In 431e1ae, API - Debank Gas Price Dict ($DEBANK_GAS_PRICE_DICT_API_URL) was down:
HTTP code: 429
Response time: 254 ms
Resolved: API - Debank Gas Price Dict is back up in c031bfd.
|
gharchive/issue
| 2023-07-04T10:42:47 |
2025-04-01T04:54:55.645200
|
{
"authors": [
"AutomationDimension"
],
"repo": "DimensionDev/status",
"url": "https://github.com/DimensionDev/status/issues/1801",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
619768863
|
Turnip prices should update to show user input
This is work for issue #110. It doesn't include localizations, but conditionally shows turnip prices. I just took a crack at it and try to match the style of the app already.
I'm really not sure about that. Maybe the text should be more explicit? Like "Your in game price this morning is X and the predictions for this afternoon is Y". Right now it feels weird.
@TheVaan @renaudjenny on this.
I agree that it’s not that easy to find the right text. When we touch the text, we should discuss if there is more to do.
E.g. it doesn’t make sense to show the average price in the morning when it is later then 12pm and you inserted prices for the afternoon.
Maybe we need steps like:
no values for today:
Your in game price is predicted to be around X in the morning and around Y in the afternoon
value morning inserted, value afternoon missing
Your in game price was X in the morning and is predicted to be around Y in the afternoon
value in the afternoon inserted (doesn’t matter if morning value inserted)
Your in game price tomorrow is predicted to be around X in the morning and around Y in the afternoon
Just an idea.
I'm sorry we'll redo this one internally for now. Not yet clear on the features and I'm working on it to experiment.
|
gharchive/pull-request
| 2020-05-17T18:32:18 |
2025-04-01T04:54:55.648386
|
{
"authors": [
"Dimillian",
"TheVaan",
"maeganjwilson"
],
"repo": "Dimillian/ACHNBrowserUI",
"url": "https://github.com/Dimillian/ACHNBrowserUI/pull/156",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
748907641
|
Is it still active?
The current version crashes on Big Sur for me
I'm also interested in knowing if this is still active.
It's not abandoned, but SwiftUI have so many issues on macOS that it's hard to justify working it. I might continue this year, but I guess I'll be much more motivated when a new version of macOS will be released with a better SwiftUI.
There was a release 8 days ago. So, it seems like there's work going on.
@Dimillian Is this still maintained ?
~7 months since the last push and the release...
Well I mean, it’s working :)
Well I mean, it’s working :)
However, It is buggy as hell :(
It crashes here and there, disabled sidebar items from the settings is ignored and shown everywhere, It can't load u/usernames.... etc
I remember the original developer mentioning that this was due to Swift 4. I have since assumed he gave up, and so i removed it and gave up waiting as well ;-)
|
gharchive/issue
| 2020-11-23T15:41:08 |
2025-04-01T04:54:55.655987
|
{
"authors": [
"Dimillian",
"Magniquick",
"armaandh",
"jhjacobs81",
"samyak-jain"
],
"repo": "Dimillian/RedditOS",
"url": "https://github.com/Dimillian/RedditOS/issues/20",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
660309150
|
Support for 15254
Currently the app requires 17763+. Would it be possible to support 15254 or even 14393?
Sorry, it's impossible. This project references the WindowsCommunityToolkit, which requires 17763.
|
gharchive/issue
| 2020-07-18T18:43:06 |
2025-04-01T04:54:55.658799
|
{
"authors": [
"DinoChan",
"hihain"
],
"repo": "DinoChan/OnePomodoro",
"url": "https://github.com/DinoChan/OnePomodoro/issues/42",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
109668277
|
corrected grammar mistakes in CONTRIBUTING.md.
There're several grammar and spelling mistakes in this file. I just corrected it 2 minutes ago. I am still understanding the relationship inside the your Java (Minecraft Rewrite) programming.
A problem behind is: My Eclipse IDE can't push the project into the same folder in GitHub. It will push to a new folder in GitHub. I am finding ways to fix it.
Eclipse have problems to import multi-module maven project with nested modules.
NetBeans/Intellij seems to work with this.
Do you want add more here? Or I can merge it now?
And thanks for contributing ;)
You may merge now. Probably, there's nothing to change now. I just corrected spelling and grammar. May be in the future, add some Tips / TODO.
|
gharchive/pull-request
| 2015-10-04T07:47:28 |
2025-04-01T04:54:55.690635
|
{
"authors": [
"GotoFinal",
"mob41"
],
"repo": "Diorite/Diorite",
"url": "https://github.com/Diorite/Diorite/pull/13",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1185906466
|
🛑 Master Bot is down
In c1f3b06, Master Bot (https://MasterBot.DLCDevelopment.repl.co) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Master Bot is back up in b587dbe.
|
gharchive/issue
| 2022-03-30T05:31:26 |
2025-04-01T04:54:55.725856
|
{
"authors": [
"samosaman73"
],
"repo": "Discord-Development-Centre/status",
"url": "https://github.com/Discord-Development-Centre/status/issues/1132",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1001233815
|
[feature request] loading custom font support
deno-canvas supports writing text to a canvas:
import { createCanvas } from 'https://deno.land/x/canvas/mod.ts'
const canvas = createCanvas(500,600)
const ctx = canvas.getContext('2d')
ctx.fillStyle='red'
ctx.fillText(50,50,"Hello World")
await Deno.writeFile("image.png", canvas.toBuffer());
this is currently only limited to fonts your system knows about (it might even be more limited than that). Canvas kit has an api for loading custom fonts https://skia.org/docs/user/modules/quickstart/#text-shaping. Itd be great if deno-canvas supported loading custom fonts. I think pulling in the whole paragraph builder api might be substantial, but all I would personally be interested in is mirroring the browser's canvas apis with the addition of being able to load custom fonts. E.g.
import { createCanvas, registerFont } from 'https://deno.land/x/canvas/mod.ts'
const canvas = createCanvas(500,600)
const ctx = canvas.getContext('2d')
await registerFont({
font_family: 'Comic Sans',
// extra fancy would be supporting a url here (e.g. file:///my-fonts/comic-sans.ttf or any web url)
src: './my-fonts/comic-sans.ttf'
})
ctx.font = 'Comic Sans'
ctx.fillStyle='red'
ctx.fillText(50,50,"Hello World")
await Deno.writeFile("image.png", canvas.toBuffer());
the registerFont method is very similar to the css interface for loading fonts:
@font-face {
font-family: 'KulminoituvaRegular';
src: url('http://www.miketaylr.com/f/kulminoituva.ttf');
}
[edit]
it seems like ctx.font = is not a supported api at all in the current version. Setting font to anything, including fonts available to the system looks like:
> ctx.font = 'Roboto'
Uncaught TypeError: Cannot read property 'family' of null
at q.set [as font] (https://deno.land/x/canvas@v1.3.0/src/lib.js:2264:37)
at <anonymous>:2:10
For loading custom fonts, there is canvas.loadFont.
For setting font you'll need to specify size too.
oh if this is already supported thats fantastic. I am looking at the code now https://github.com/DjDeveloperr/deno-canvas/blob/master/src/types.ts#L1065, what should the descriptors field be?
ah, I figured it out:
canvas.loadFont(fontBuffer, {
family: 'Comic Sans',
style: 'normal',
weight: 'normal',
variant: 'normal'
})
it probably wouldnt hurt to make the type signatures more accurate than Record<string, string>. Perhaps:
type FontDescriptors = {
/* identifying name of font */
family: string
style: 'normal' | 'italic'
variant: 'normal' | ...
weight: 'normal' | 'bold' | ...
}
loadFont(
bytes: ArrayBuffer | Uint8Array,
descriptors: FontDescriptors,
): void;
seems like the context.measureText results are very inaccurate.
const text = "Hello There"
// load a font
const font = await Deno.readFile(family)
const font_identifier = new Date().toString()
canvas.loadFont(font, {
family: font_identifier
})
context.font = `${size}px ${font_identifier}`
// get the font measurements
const metrics = context.measureText(text)
// draw a rect around it
context.fillStyle = 'white'
context.fillRect(0, 0, metrics.width, metrics.fontBoundingBoxAscent + metrics.actualBoundingBoxDescent)
// draw the text
context.fillStyle = "black"
context.fillText(text_chunk, 0, 0)
if this looks like a real issue I can create a separate issue or change the title on this one
Here are some repros. First a jsfiddle which correctly measures the width of text:
https://jsfiddle.net/8dk71toq/2/
and second, deno-canvas incorrectly measuring the same text:
import { createCanvas } from 'https://deno.land/x/canvas@v1.3.0/mod.ts'
const canvas = createCanvas(500, 200)
const context = canvas.getContext('2d')
function draw_text(text: string, x: number, y: number) {
const metrics = context.measureText(text)
console.log(metrics)
context.fillStyle = 'red'
context.fillRect(
x,
y,
metrics.actualBoundingBoxLeft + metrics.actualBoundingBoxRight,
metrics.fontBoundingBoxAscent,
)
context.fillStyle = 'black'
context.fillText(
text,
x + metrics.actualBoundingBoxLeft,
y + metrics.fontBoundingBoxAscent,
)
}
context.fillStyle = 'white'
context.fillRect(0, 0, canvas.width, canvas.height)
const font_buffer = await Deno.readFile('./fonts/tangerine/Tangerine-Regular.ttf')
canvas.loadFont(font_buffer, {family: 'Tangerine'})
context.font = '50px Tangerine'
draw_text("Hello World", 50, 50)
await Deno.writeFile('canvas.png', canvas.toBuffer())
both examples use this font https://fonts.google.com/specimen/Tangerine?query=tangerine
some further debugging shows that the builtin font (monospace in the case of my mac) correctly measures the text. 50px monospace is shown below.
loading a different font and specifying it via context.font does produce different measurements, so its clear that canvaskit is reading the new font. It just doesnt quite interpret the sizes correctly. Is this an issue I should move upstream to canvaskit?
All values in TextMetrics except width were added with a hacky workaround, so they might not be right. And no, actual canvaskit does not even implement measureText as they just say to use Paragraph API instead. So this is a issue in deno-canvas only.
Got it. Well it is unfortunate that measuring text can't be done with deno canvas. I wanted to build out text captions with borders. I suppose there's an extremely hacky workaround for me where I dump the image data and find the max & min x & y
Right.. text rendering as a whole is not-so-good with canvaskit-wasm. Btw, deno-canvas does expose Skia related APIs, such as Paragraph API if that's any helpful for you. I plan on porting Node.js skia-canvas using Deno FFI, if that works out we'd have a more performant and compatible canvas API. Though can't say for sure it'll be a thing, or anytime soon as FFI is limited a lot right now.
It would definitely be cool to see an ffi bridge to skia. For what I need though, if the paragraph api is exposed that would probably get me everything I need. Would you mind showing an example of how to use it in deno canvas (I have seen canvaskits docs but I'm not sure it's the same api here)
Skia related APIs are all exposed in a namespace that is default exported from mod.ts.
import Skia from "https://deno.land/x/canvas@v1.3.0/mod.ts";
// Use Skia.Paragraph, Skia.ParagraphBuilder, etc.
// API should be same as canvaskit.
@DjDeveloperr could you give me a primer on contributing to this library? Would it be this file https://github.com/DjDeveloperr/deno-canvas/blob/master/src/lib.js#L2750? Its hard to tell if this is a source file because some of this code looks minified. Could you show where another skia specific method is used as an example?
Yeah, in src/lib.js. Right.. it was minified at first. a variable contains everything exported as default in mod.ts, so all Skia APIs. You can use ctrl+f to find implementation of measureText.
@andykais your first attempt at the implementation looks great. I think it would be good enough for first pass (better than currently inaccurate implementation)
And yes, we do have access to Font in context (it's a minified property, which we can access through context.we), and even FontMgr (in canvas.Cf).
So I think you can PR this function (but slightly modified to accept FontMgr and Font instead of), and I can add it in lib.js 🤔
whats the current CanvasKit version?
It is 0.32.0
I think I've figure out how to measure text correctly.
First, ignore measureText() method. It is not accurate.
// Make sure to set the font size first!
context2d.font = `16px sans-serif`;
// The text to display
const text: string = `Hello World`;
// Measure the width of a single character
const charWidth: number = Math.floor(context2d.measureText("X").width);
// Get the length of the text string
const textLength: number = text.length;
// textWidth at selected fontSize
const textWidth: number = charWidth * textLength;
@webdev3301 I believe there is still an issue with measuring custom loaded fonts. See this message above: https://github.com/DjDeveloperr/deno-canvas/issues/17#issuecomment-926166920
|
gharchive/issue
| 2021-09-20T17:48:02 |
2025-04-01T04:54:55.839706
|
{
"authors": [
"DjDeveloperr",
"andykais",
"webdev3301"
],
"repo": "DjDeveloperr/deno-canvas",
"url": "https://github.com/DjDeveloperr/deno-canvas/issues/17",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1408133821
|
socket_bot hangs - no bot response
Hi, thanks for sharing this tool! I've been through the steps to get this set up with Invoke AI today. I've managed to run through your installation instructions as expected, but seem to fail without an error. At your last instruction, I run socket_bot.py and the last four messages I receive are:
DEBUG:slack_sdk.socket_mode.builtin.client:A new message enqueued (current queue size: 1)
DEBUG:slack_sdk.socket_mode.builtin.client:A message dequeued (current queue size: 0)
DEBUG:slack_sdk.socket_mode.builtin.client:Message processing started (type: hello, envelope_id: None)
DEBUG:slack_sdk.socket_mode.builtin.client:Message processing completed (type: hello, envelope_id: None)
Then the tool hangs for at least 20 minutes (I haven't waited longer)
I've tried restarting the env, I've also tried having invoke AI running in advance of starting socket_bot, as well as not having it running at all.
Can you please help me? Is it possible I've missed something in the slack bot set up?
Further info: In the slack bot set up I was forced to choose a specific event subscription in order to save event subscriptions being 'on' - I did so with the only one available from the list.
Thanks!
"display_information": {
"name": "Dreamboat"
},
"features": {
"bot_user": {
"display_name": "Dreamboat",
"always_online": false
}
},
"oauth_config": {
"scopes": {
"bot": [
"app_mentions:read",
"chat:write",
"files:write",
"groups:write",
"im:write",
"mpim:write",
"reactions:write",
"chat:write.public"
]
}
},
"settings": {
"event_subscriptions": {
"bot_events": [
"app_mention"
]
},
"interactivity": {
"is_enabled": true
},
"org_deploy_enabled": false,
"socket_mode_enabled": true,
"token_rotation_enabled": false
}
}```
Issue was potentially caused by typo in env variables.
|
gharchive/issue
| 2022-10-13T16:52:32 |
2025-04-01T04:54:55.859874
|
{
"authors": [
"EddieVec"
],
"repo": "Djings/slacksd",
"url": "https://github.com/Djings/slacksd/issues/5",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
221837233
|
DjvuPage.BuildImage dimension mismatches for foreground, background and mask are not compensated for
In function which constructs final image by blending data from background, foreground and mask in the case width and height of images is not exactly same AccessViolotaionException is thrown.
Issue resolved with commit f9cd2f2
|
gharchive/issue
| 2017-04-14T16:00:50 |
2025-04-01T04:54:55.860980
|
{
"authors": [
"4creators"
],
"repo": "DjvuNet/DjvuNet",
"url": "https://github.com/DjvuNet/DjvuNet/issues/41",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
178469379
|
Fix button showing up on secret page.
What's this PR do?
We never hooked up the variables to the "Connect with Northstar" button on the secret login page ✨ 🔒 ✨ so you can actually secretly log in on Thor/Production!
How should this be reviewed?
The variable is being passed now, yeah?
Any background context you want to provide?
No.
Relevant tickets
Fixes 🚯.
Checklist
[ ] Documentation added for new features/changed endpoints.
[ ] Tested on staging.
[ ] Pinged a PM if this is a larger PR that would benefit from some additional testing love.
👍
|
gharchive/pull-request
| 2016-09-21T21:39:45 |
2025-04-01T04:54:55.944457
|
{
"authors": [
"DFurnes",
"weerd"
],
"repo": "DoSomething/phoenix",
"url": "https://github.com/DoSomething/phoenix/pull/7071",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1620655757
|
File System Todo: fix doctests
The doctests currently fail due to file persistence. Unlike GenServers, you can't can't simply start a new link/pid and begin afresh!
The current doctests will still fail if the code is run repeatedly, so it's not a silver bullet, though I'm not currently sure of a better approach except for simply removing the doctests entirely.
I think we should simply avoid running these doctests by removing iex> from.
Alternatively, we could remove the file before each doctest to avoid persistence issues, but I don't like this option as much.
I made the changes. Thank you for the PR!
|
gharchive/pull-request
| 2023-03-13T02:43:12 |
2025-04-01T04:54:55.960125
|
{
"authors": [
"BrooklinJazz",
"matt-humphrey"
],
"repo": "DockYard-Academy/curriculum",
"url": "https://github.com/DockYard-Academy/curriculum/pull/886",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
227212608
|
Create Account does not work after two.
"I want to create many accounts in short time. But after two success account create can't create more not working. I guess steam block you after two account create.
Also i add third minutes for each create still not work after two."
Above is quoted from someone, and I am meeting this same problem now, any suggestions to solve this? Thanks!
Slow down more.
Just curious but, isn't it supposed to be after 5?
|
gharchive/issue
| 2017-05-09T00:24:11 |
2025-04-01T04:54:55.978780
|
{
"authors": [
"DoctorMcKay",
"Raikiree",
"ylms"
],
"repo": "DoctorMcKay/node-steam-user",
"url": "https://github.com/DoctorMcKay/node-steam-user/issues/106",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1563865947
|
🛑 Ceresana is down
In 42de2e9, Ceresana (https://www.ceresana.com) was down:
HTTP code: 500
Response time: 761 ms
Resolved: Ceresana is back up in 7cfbf47.
|
gharchive/issue
| 2023-01-31T08:24:38 |
2025-04-01T04:54:55.981364
|
{
"authors": [
"Dodger77"
],
"repo": "Dodger77/upptime",
"url": "https://github.com/Dodger77/upptime/issues/1694",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1284985718
|
🛑 Apodeko is down
In 72c7ac4, Apodeko (https://www.apodeko.de) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Apodeko is back up in 8d0a84a.
|
gharchive/issue
| 2022-06-26T17:38:26 |
2025-04-01T04:54:55.983597
|
{
"authors": [
"Dodger77"
],
"repo": "Dodger77/upptime",
"url": "https://github.com/Dodger77/upptime/issues/843",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
266323169
|
Add dist folder containing current 'dist' files.
This is just an issue for me to upload the dist files.
I will do this myself, thank you!
I am creating a branch to start the pull now.
I have merged pull request #8 but will keep the issue open as there is more to do.
|
gharchive/issue
| 2017-10-18T00:33:37 |
2025-04-01T04:54:55.984771
|
{
"authors": [
"willtheorangeguy"
],
"repo": "Dog-Face-Development/Bars",
"url": "https://github.com/Dog-Face-Development/Bars/issues/7",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
82642758
|
Allow optional "x" to dismiss toast messages.
Continuing the discussion from #1389.
The Problem: Many users aren't aware that they can close the toast messages by swiping on it. This is apparent especially on Desktop devices where the user is not accustomed to swiping.
Here is a more recent complaint on the Gitter board: https://gitter.im/Dogfalo/materialize?at=5568b4cfa891bec12ef4d221
The solution: Provide the user with some hint on how to close the modal.
I am in favor of two options, although I do prefer the latter.
Make the click event close the message (This would work, but it may cause some unexpected closes; this may not be the ideal option since the problem of the user not knowing that it's closable still exists until the user clicks/taps the message).
Add an "x" button to the message (This is ideal since the user will immediately know that they have the option to manually close the modal).
Refer to the docs for the current method of dismissing notifications: http://materializecss.com/dialogs.html
I'm not saying that the swiping method be replaced with one of these options; it's the opposite, actually. By adding the "x" to the message, it triggers an instinct in users that the toast message is indeed closable, something that is not apparent in the current setting.
cc: @jcapogna @acburst @mdcsfk @Dogfalo
I vote for adding an an X close. The mechanism I mentioned in #1389 has been very well received by people I've showed it to.
Modals should also have an "X" to close, with respective options and callbacks
:+1:
+1
:+1:
If anyone's interested in a quick easy workaround:
$(document).on('click', '#toast-container .toast', function() {
$(this).fadeOut(function(){
$(this).remove();
});
});
+1
@davidbielik in addition, you can add this css:
#toast-container .toast::after {
font-family: 'Material Icons';
content: "close";
-webkit-font-feature-settings: 'liga';
color: #dd2c00;
font-size: 1.5rem;
font-weight: 300;
float: right;
padding-left: 3rem;
}
Complete code
HTML File
<script>
$(document).on('click', '#toast-container .toast', function() {
$(this).fadeOut(function(){
$(this).remove();
});
});
</script>
CSS File
#toast-container .toast {
cursor: pointer;
}
#toast-container .toast::after {
font-family: 'Material Icons';
content: "close";
-webkit-font-feature-settings: 'liga';
color: #dd2c00;
font-size: 1.5rem;
font-weight: 300;
float: right;
padding-left: 3rem;
}
@flexbrane awesome snippet. I used it. It works.
So here is what I did:
function closetoast(){ $('#toast-container').fadeOut(); };
Materialize.toast('Toasttext <a onclick="closetoast()" href="#!">X</a>', 5000);
0-1 actions, not dismiss or cancel
If an action is present, comply with dialog spacing and affordance rules. For two or more actions, use a dialog, not a snackbar, even when one of the actions is a dismiss action. If the action described in the snackbar is important enough to block the use of the screen, it should be a dialog.
Source: https://material.io/guidelines/components/snackbars-toasts.html#snackbars-toasts-usage
Interestingly, inbox.google.com uses the following when you press "done".
I got it working using the toast's own dismiss function. This should make the usual callbacks.
<a onclick='M.Toast.getInstance(this.parentElement).dismiss();'>close or icon</a>
|
gharchive/issue
| 2015-05-30T00:27:48 |
2025-04-01T04:54:55.995214
|
{
"authors": [
"GloverDonovan",
"artus9033",
"cwg999",
"davidbielik",
"eutervogel",
"flexbrane",
"gizotti",
"hsali",
"jcapogna",
"koustuvsinha",
"nrozic",
"pumpknhd",
"rapito",
"tomscholz"
],
"repo": "Dogfalo/materialize",
"url": "https://github.com/Dogfalo/materialize/issues/1453",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
104297859
|
Does materialize work with angular?
I have a project that uses angular and I want to use materialize on it. Is that possible?
Yep, though not all features have been implemented yet.
https://github.com/krescruz/angular-materialize
It wasn't quite working for me, but did you ever give it a try @Toujimon ?
I'm currently working on a project with this and it's working just fine. My major complains are about the datepicker (it works really funny and I just used the one referenced in the angular ui selection).
Can you be more specific about your problem using it?
Well, in an ongoing project, I tried to use the select option as set on https://github.com/krescruz/angular-materialize.
That did not work at all, because I had to set the select functionality by hand on my JS like this:
$('select').material_select();
So we ended up throwing away jquery and all materialize.js and used angular material. =/
I understand. I must admit that for dropdowns, I prefer to use the "browser-default" class option instead.
Thanks @Toujimon for the responses.
The materialize DatePicker didn't work for me with AngularJS.
|
gharchive/issue
| 2015-09-01T15:59:17 |
2025-04-01T04:54:55.999262
|
{
"authors": [
"Dogfalo",
"Toujimon",
"dcoferraz",
"ferchoman09"
],
"repo": "Dogfalo/materialize",
"url": "https://github.com/Dogfalo/materialize/issues/1991",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
246602205
|
Bug with the class on NavBar "navbar-fixed"
Good evening,
I have a bug on the NavBar Mobile Collapse, which appears when you add the class "navbar-fixed", the left pane becomes all grayed out, and you can not select any link Below, and I've linked you live
You can see it in the screenshot below, and in the link, which leads directly to the bug
And here is a screenshot without the class "navbar-fixed" there is no problem, so I guess the bug vient of the class "navbar-fixed", I put a screenshot below.
You can see the bug, live on this link, with the code below.
<div class="navbar-fixed">
<nav>
<div class="nav-wrapper">
<a href="#!" class="brand-logo">Logo</a>
<a href="#" data-activates="mobile-demo" class="button-collapse"><i class="material-icons">menu</i></a>
<ul class="right hide-on-med-and-down">
<li><a href="sass.html">Sass</a></li>
<li><a href="badges.html">Components</a></li>
<li><a href="collapsible.html">Javascript</a></li>
<li><a href="mobile.html">Mobile</a></li>
</ul>
<ul class="side-nav" id="mobile-demo">
<li><a href="sass.html">Sass</a></li>
<li><a href="badges.html">Components</a></li>
<li><a href="collapsible.html">Javascript</a></li>
<li><a href="mobile.html">Mobile</a></li>
</ul>
</div>
</nav>
</div>
See you soon.
Try to move the sidenav ul outside of its wrappers.
Duplicate of https://github.com/Dogfalo/materialize/issues/3844
tomscholz marked this as a duplicate of #3844 18 hours ago
For what not to correct it since version 0.9.7 it has no way?
@tomscholz, myself and some others joined as contributors this year. The best solution has to be found but there are more things to be done than this. PR's are very welcome.
|
gharchive/issue
| 2017-07-30T19:00:59 |
2025-04-01T04:54:56.005122
|
{
"authors": [
"CodingWeb",
"DanielRuf",
"tomscholz"
],
"repo": "Dogfalo/materialize",
"url": "https://github.com/Dogfalo/materialize/issues/5047",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
249197212
|
Autocomplete doesn't complete the last character
Hello guys,
I was searching for issues related to my problem and I thought this one could be duplicate:
https://github.com/Dogfalo/materialize/issues/4230
It was said that the related autocomplete problem was solved. I've downloaded the last materialize-sass gem version but it seems it wasn't fixed.
The problem I'm facing with is: I have the tag "notebook". the autocomplete works until "noteboo", when I enter the "k", the tag simply doesn't appear.
That's how I'm declaring my autocomplete component:
$('.chips-autocomplete').material_chip({
autocompleteOptions: {
data: tagoptions.autoCompleteData,
limit: Infinity,
minLength: 1
},
secondaryPlaceholder: 'Adicione a tag',
placeholder: '+Tag',
data: tagoptions.data
});
The tagoptions.autoCompleteData object is defined as:
{
eraser:null,
notebook:null,
pencil: null
}
The tagoptions.data is an array of chips:
[{
id,
tag
}]
Is it a real problem or am I missing something?
Thank you,
Guilherme
same problem in node and meteor
What version? I cannot reproduce this on: http://materializecss.com/forms.html#autocomplete
materialize:materialize@=0.99.0
Thank you, guys.
I'm using 0.97.8.
@neutron92 , did you solve it with this snippet?
i have this errors
materialize_materialize.js?hash=38aa0de00bf1da3e83504deced815b530abaefa4:7826 Uncaught Error: Scss compiler error: Undefined variable: "$medium-and-down".
at materialize_materialize.js?hash=38aa0de00bf1da3e83504deced815b530abaefa4:7826
at materialize_materialize.js?hash=38aa0de00bf1da3e83504deced815b530abaefa4:7830
at materialize_materialize.js?hash=38aa0de00bf1da3e83504deced815b530abaefa4:7842
(anonymous) @ materialize_materialize.js?hash=38aa0de00bf1da3e83504deced815b530abaefa4:7826
(anonymous) @ materialize_materialize.js?hash=38aa0de00bf1da3e83504deced815b530abaefa4:7830
(anonymous) @ materialize_materialize.js?hash=38aa0de00bf1da3e83504deced815b530abaefa4:7842
global-imports.js?hash=5e0f0a71522c063bb881071904e7eeca01779ebd:16 Uncaught TypeError: Cannot read property 'Materialize' of undefined
at global-imports.js?hash=5e0f0a71522c063bb881071904e7eeca01779ebd:16
(anonymous) @ global-imports.js?hash=5e0f0a71522c063bb881071904e7eeca01779ebd:16
app.js?hash=8d04954afed1b707ef74ab0f7fa58a663df0fea6:1 Uncaught ReferenceError: meteorInstall is not defined
at app.js?hash=8d04954afed1b707ef74ab0f7fa58a663df0fea6:1
same prob here
https://codepen.io/AlphaDx/pen/xgmXgw
This should be fixed on the latest version
|
gharchive/issue
| 2017-08-10T00:10:57 |
2025-04-01T04:54:56.011522
|
{
"authors": [
"Dogfalo",
"acburst",
"guilhermemaranhao",
"neutron92"
],
"repo": "Dogfalo/materialize",
"url": "https://github.com/Dogfalo/materialize/issues/5094",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2038074898
|
For limit order get status EXPIRED_IN_MATCH on Binance testnet
ver: 2.0.3 + 2.0.4.post3 + 2.0.4.post2
main.account_name: Demo - Binance
main.exchange: binance
main.srv_version: 1.4.1
12/12 09:03:16 Exception in fetch_order: <AioRpcError of RPC that terminated with:
status = StatusCode.UNKNOWN
details = "Unexpected <class 'google.protobuf.json_format.ParseError'>: Message type "martin.FetchOrderResponse" has no field named "preventedMatchId" at "FetchOrderResponse".
Available Fields(except extensions): "['symbol', 'orderId', 'orderListId', 'clientOrderId', 'price', 'origQty', 'executedQty', 'cummulativeQuoteQty', 'status', 'timeInForce', 'type', 'side', 'stopPrice', 'icebergQty', 'time', 'updateTime', 'isWorking', 'workingTime', 'origQuoteOrderQty', 'selfTradePreventionMode']""
debug_error_string = "UNKNOWN:Error received from peer ipv4:127.0.0.1:50051 {created_time:"2023-12-12T09:03:16.12978971+00:00", grpc_status:2, grpc_message:"Unexpected <class \'google.protobuf.json_format.ParseError\'>: Message type \"martin.FetchOrderResponse\" has no field named \"preventedMatchId\" at \"FetchOrderResponse\".\n Available Fields(except extensions): \"[\'symbol\', \'orderId\', \'orderListId\', \'clientOrderId\', \'price\', \'origQty\', \'executedQty\', \'cummulativeQuoteQty\', \'status\', \'timeInForce\', \'type\', \'side\', \'stopPrice\', \'icebergQty\', \'time\', \'updateTime\', \'isWorking\', \'workingTime\', \'origQuoteOrderQty\', \'selfTradePreventionMode\']\""}"
>
For some pair can't off STP mode:
'defaultSelfTradePreventionMode': 'EXPIRE_MAKER', 'allowedSelfTrade
PreventionModes': ['EXPIRE_TAKER', 'EXPIRE_MAKER', 'EXPIRE_BOTH']
|
gharchive/issue
| 2023-12-12T09:31:25 |
2025-04-01T04:54:56.014057
|
{
"authors": [
"DogsTailFarmer"
],
"repo": "DogsTailFarmer/exchanges-wrapper",
"url": "https://github.com/DogsTailFarmer/exchanges-wrapper/issues/42",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
459881442
|
[Suggestions] Feature Requests
Thanks for putting this addon together! Had a few suggestions in using it (currently in the Beta):
1.) Is it possible to close the window on Esc key? This is done by adding the frame's name to the UISpecialFields table:
table.insert(UISpecialFrames, windowFrame:GetName());
2.) Is there a faster way to clear the selected dropdown's multi-select? If I have 4-5 selected and want to clear it there is no easy way to do this other than to unclick all of them.
3.) Perhaps a "fix" to #2, is there an "All" category? That just shows me all groups? This would have to be removed from the dropdown for the "Create Group" tab.
4.) Is there a checkbox or some feature that can be set to "Check Dungeons Out of my Level Range". Would be helpful if im leveling an alt while looking for a dungeon group for my main / higher level alt. You would still of course need to either change to your main to whisper the group or say "Hey I can bring my level XX YY and I can log over in Z minutes"
5.) Dropdowns match the input fields in look. from this window it looks like the input field for invite message "invite please" is part of the dropdown for broadcast channel:
That's all I have for now having just installed it. I'll be sure to report any Lua errors etc too!
Hey thanks for your feedback, I really appreciate it!
#1 Yeah sure, thats not a problem, will be in for the next build!
#2 The UI you see right now is actually a complete rework that I just published yesterday, I will add the possibility to select and deselect all items at the same time!
#3 See #2
#4 I did not want to overwhelm people with a huge dungeon list, also some people might not know which dungeons to run on a given level. Experiences players don't need this hand-holding so I will make it an option in the settings menu!
#5 I realized this too, I am still trying to figure out how to properly make the dropdown items stand out a bit more without breaking the visual style, but a fix should come soon!
Out of curiosity, how did you get to know of this addon?
Update:
#1, #2, #3 and #4 are in, still trying to figure out what the best way to handle #5 is.
The new version should be available on curseforge/twitch by now
Awesome :) Thanks! I know you get a bad rep for being called "LFG" but hope there is a decent size of people to propagate information and make finding a group easier without having to stand in a city for an hour much easier :)
Even if you are the only user on the realm (which I hope is not going to be the case) you still have all the groups looking for people in a nicely organized window :)
Dropdown menus now have a slightly different color and should more recognizable now, closing this issue now!
|
gharchive/issue
| 2019-06-24T13:06:49 |
2025-04-01T04:54:56.048715
|
{
"authors": [
"DomenikIrrgang",
"sinkda"
],
"repo": "DomenikIrrgang/ClassicLFG",
"url": "https://github.com/DomenikIrrgang/ClassicLFG/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1368650973
|
🛑 USPS - Registration is down
In 183f772, USPS - Registration (https://reg.usps.com/) was down:
HTTP code: 400
Response time: 128 ms
Resolved: USPS - Registration is back up in 1af280b.
|
gharchive/issue
| 2022-09-10T15:18:19 |
2025-04-01T04:54:56.088244
|
{
"authors": [
"dotgovdev-admin"
],
"repo": "DotGov-Dev/govuptime",
"url": "https://github.com/DotGov-Dev/govuptime/issues/7",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
274510295
|
Autoformat should keep indentation level
When I make a text selection like this:
And then I hit cmd-shift-f to invoke auto-format, I get this:
From the "n:explorer" element downwards the indentation level is wrong. It seems everythign is aligned on the last closing tag which is on most left column.
Actually It should always use the indentation level based on the whole document, even if just the selection is autoformatted.
I will not have time to work on this in the near future. I will accept PRs for this after v2.0.0 is released.
|
gharchive/issue
| 2017-11-16T13:10:09 |
2025-04-01T04:54:56.090847
|
{
"authors": [
"DotJoshJohnson",
"mojo2012"
],
"repo": "DotJoshJohnson/vscode-xml",
"url": "https://github.com/DotJoshJohnson/vscode-xml/issues/131",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
966292788
|
Invalid json using collect over a relationship in bolt client
This PR addresses https://github.com/DotNet4Neo4j/Neo4jClient/issues/417 issue
Fix typo Deserilizes => Deserializes
StatementResultHelper.cs is fixed to build valid JSON in case of IRelationship
New test is provided that emulates the initial problem
Thank you @andrey-kastryukhin !
Thanks! Greetings from Ibiza
No problem at all! Greetings to Ibiza
PS - Generally PRs like this which are tested are super easy to go through - so if there are things like this, please feel free to PR them!!
|
gharchive/pull-request
| 2021-08-11T09:08:09 |
2025-04-01T04:54:56.093303
|
{
"authors": [
"Clooney24",
"andrey-kastryukhin",
"cskardon"
],
"repo": "DotNet4Neo4j/Neo4jClient",
"url": "https://github.com/DotNet4Neo4j/Neo4jClient/pull/418",
"license": "MS-PL",
"license_type": "permissive",
"license_source": "github-api"
}
|
195665101
|
Could a setting be considered for SA1501/SA1503 to relax on some kind of statements?
Somewhat related to #1175 our team uses something in the lines of the following rule to account for common guards:
When the body of an if statement consists of a single jump statement (e.g. return, throw, break, continue) we allow that statement without braces and on the same line.
Which would mean that the initial code on that linked issue will be valid for us:
if (condition) return;
But not so other kinds of statements like:
if (condition) someVariable = "someValue";
if (condition) someMethod();
or even
if (condition)
someMethod();
if (condition)
return;
We feel this is terse and removes clutter without causing readability concerns since the situation is pretty clear with jump statements which break the flow of execution (basically it allows us to have multiple one-liner guards, taking less space).
// Just some example, maybe not the best.
if (obj == null) throw new ArgumentNullException("obj");
if (obj.Whatever == null) return;
foreach (var item in obj.Whatever.List)
{
if (!item.Condition) continue;
Process(item);
}
Now, I know that it's impossible to account for every team particular requirement or nuance and that maybe this is not even in the spirit of StyleCop at all (cannot say, while we feel we follow common sense standards, we have never adopted the tool until we discovered this great version using analyzers) but having the option to allow these kind of statement while also enforcing the already existing related restrictions is something that will be very good for us.
I imagined that before going and doing our own thing to enforce it or change the way we work to accomodate, it wouldn't hurt asking the project and the community if they would be interested in considering this behavior as optional or maybe even allow it as a contribution.
Thanks for the well written proposal. I see benefit in this (narrow) case so a 👍 from me. This will have to be optional, with the default behavior to remain the same as it is now.
The statement on the same line is paramount here, as it reduces the chance of somebody typing a next statement and assuming that it will be executed as part of the if statement. This could be further strengthened by demanding an empty line, but that should probably be a new rule.
I believe the selective omission of braces where the child element is on the next line is covered by leaving SA1519 and SA1520 enabled, but disabling SA1503.
I'm :-1: on an additional option allowing the child statement to appear on the same line in these cases, as that option would be equivalent to the current ability to disable SA1501 in essentially all real-world cases.
@sharwell I do not agree that disabling SA1501 would cover all cases. This proposal talks about allowing this only for if statements. Disabling SA1501 would also open up other statements (like lock, for, while, etc.) to become single line statements, something that was never the intention of the proposal.
An alternative for the proposed solution could be to only allow configuration of SA1501 so that it will allow single line if statements. The braces will then be mandatory, as SA1503 is not modified.
@vweijsters Thanks, I misread a section of the proposal and was under the impression that the following was allowed:
if (something)
someMethod();
It seems like there are two options at play here:
May braces be omitted for an if statement with a single child that branches?
If so, is that single child placed on the current line or the following line?
I have no real preference for question 1.
I'd say that the answer to the 2nd question should be yes.
💭 There may be benefit in adding it for people that cannot / may not upgrade to Visual Studio 2017 in the near future.
Thanks for the thoughtful discussion guys. As mentioned, we only allow this for if statements with jumps as their single child, and we do like the current validations for both SA1501 and SA1503 enabled to happen on other cases than these.
I also agree that those seem to be the options at play here and of course our answers would be "yes" and "current line", but were you to decide to go forward with this I'm not imagining any other way than having these as settings affecting the two rules, which I realize seems rather specific.
I'm going to give a mild 👎 . If implemented, it should be disabled by default. I think returns or throws should be the first thing on a line and be by themselves on a line to make them easier to find. I say this in spite of recognizing that it is a pretty frequent coding style to do validation and return at the start of a method, that there may be several of these, and that it extends quite a bit of vertical space when that is done.
I ask another question...what is the blank line spacing around the "if" if there are no braces. Is it legit to do:
if (condition1) return;
if (condition2) return;
or do there need to be blank lines between them. This may just open a whole new can of worms that shouldn't be opened.
In principle I like this proposal. I always allow this in my guidelines, but with StyleCop it currently all or nothing in this case (SA1503). I want this fine grain ment of this rule, so a big YES ( 👍 ) to question 1.
For question 2 (same or following line) I see that most comments are answering this with 'current line'. I understand this answer, because it saves a line and looks clear on simple examples.
But after using both current line and following line in the past, I ended up with the following line as preference, because of two reasons:
Whenever you want to place breakpoint for debugging, you can't place this only on the jump statement (return, throw, continue). When you are only interested in the 1 out of 1000 case that it throws, your will be glad you put a breakpoint only on the throw statement.
With conditions that are longer and in variable sizes it becomes more unclear what will happen when the condition is meet. Your eyes need to scan the line for the jump statement. If the jump statement is on the following line the logic is easier to spot. Also with 'current line' the line length more often exceeds my preferred max line length (130 characters). See the following more real life example:
Current line
if (resource != "products" && resource != "people" && resource != "pricegroups") throw new NotImplementedException("Resource is not implemented yet!");
if (-99999 <= anotherlimit && anotherlimit <= 99999) throw new ArgumentOutOfRangeException("limit", "Must be between -99999 and 99999!");
if (limit > 100) throw new ArgumentOutOfRangeException("limit", "Must be less or equal to 100!");
if (obj == null) throw new ArgumentNullException("obj");
if (obj.Whatever == null) return;
foreach (var item in obj.Whatever.List)
{
if (!item.Condition && someValue == otherValue && value1 == 2) continue;
Process(item);
}
Following line
if (resource != "products" && resource != "people" && resource != "pricegroups")
throw new NotImplementedException("Resource is not implemented yet!");
if (-99999 <= anotherlimit && anotherlimit <= 99999)
throw new ArgumentOutOfRangeException("limit", "Must be between -99999 and 99999!");
if (limit > 100)
throw new ArgumentOutOfRangeException("limit", "Must be less or equal to 100!");
if (obj == null)
throw new ArgumentNullException("obj");
if (obj.Whatever == null)
return;
foreach (var item in obj.Whatever.List)
{
if (!item.Condition && someValue == otherValue && value1 == 2)
continue;
Process(item);
}
:memo: I didn't have a question 1 and question 2. I was saying if implemented, there would be two new options in stylecop.json to allow each project to "answer the questions" with respect to just that project.
@sharwell Aha, didn't catch that :-). Then yes, those would be the two new options to make SA1503 not too strict.
I'm personally in favor to add those options, because I always end op disable the stylecop rule and have a bad feeling about it, because SA1503 in itself is a good rule to have.
My 2 cents as this is dear to me as well
// Allowed scenario 1: terse and **risk free** code
if (false) Console.WriteLine("statement1");
// Disallowed scenario 2: this is a trap for a future developer
if (false)
Console.WriteLine("statement2");
// Disallowed scenario 3: the trap is stepped on as the developer thinks statement4 is part of the if
if (false)
Console.WriteLine("statement3");
Console.WriteLine("statement4");
In the code snippets above, only the string "statement4" is printed out
I love to use scenario 1, "the 1-liner", but can't because there is no way I will disable SA1503BracesMustNotBeOmitted
As soon as it is not a 1-liner anymore you one should introduce braces.
Otherwise a developer that sees scenario 2 can add a statement at the same indentation thinking it is part of the if (which it is not).
The following avoids the trap by adding braces to prevent the trap
// Allowed scenario 4
if (false){
Console.WriteLine("statement3");
}
@buckleyGI Your disallowed scenario 3 is impossible when SA1137 is enabled. It should also be impossible when the built-in FormattingAnalyzer is enabled (IDE0055).
@sharwell Thanks, I didn't know about IDE0055 and it seems a solution if we can't enforce it.
But The version my company uses, VS 2022 Version 17.2.1, still allows me to setup the developer trap as shown in the screenshot below.
All the scenario's are allowed by this VS (no squiggles) and our Azure DevOps build server also doesn't complain, it just compiles.
Do we have to explicitly turn on IDE0055?
It is not 100% clear how these IDExxx rules work. Are they rules only for the IDE (suggestion/quick fixes) or can you also enforce them at compilation time?
Is it because we still use ruleset instead of the editorconfig and should we migrate to editorconfig to enforce IDE055 by DevOps? I read in the documentation that you can influence the IDExxx rules only with editorconfig.
@sharwell Thanks, I didn't know about SA1137 and the following configuration will allow the one liner without introducing the trap.
<Rule Id="SA1503" Action="None" />
<Rule Id="SA1137" Action="Error" />
|
gharchive/issue
| 2016-12-14T22:33:07 |
2025-04-01T04:54:56.113675
|
{
"authors": [
"GregReddick",
"buckleyGI",
"liquidsnk",
"remyvd",
"sharwell",
"vweijsters"
],
"repo": "DotNetAnalyzers/StyleCopAnalyzers",
"url": "https://github.com/DotNetAnalyzers/StyleCopAnalyzers/issues/2252",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
67815674
|
Code fix: SA1405 (DebugAssertMustProvideMessageText)
Implement a code fix for #94.
A candidate for the generated string is rendering the assertion expression to a string literal.
Do we really want a code fix here? The goal is that the developer provides a meaningful message, something which will be impossible to generate. I would personally prefer to have no code fix over having a code fix that generates meaningless boilerplate text (something that will not improve code quality)
(Same holds for #658)
My initial thought was a reasonable message for the following:
Debug.Assert(x != null || Foo());
Would be: Assertion failed: x != null || Foo()
I think that message has no value, other than eliminating the useful warning. I would prefer forcing developers to enter a message that is more meaningful than that. Of course it is impossible to really enforce that from a tool, but I would opt for not providing an easy way out.
Remember that Debug.Assert is typically only even evaluated in a debug build. Personally if I can see the assertion that failed, I already have a very good idea what the problem was.
I agree, I normally disable SA1405 because it has limited value.
I would also think there should be no CodeFix for this one. The developer should add their own message if this rule is enabled. Maybe this issue should be closed?
|
gharchive/issue
| 2015-04-11T18:19:13 |
2025-04-01T04:54:56.118336
|
{
"authors": [
"dlemstra",
"sharwell",
"vweijsters"
],
"repo": "DotNetAnalyzers/StyleCopAnalyzers",
"url": "https://github.com/DotNetAnalyzers/StyleCopAnalyzers/issues/657",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
221609655
|
POMs have lots duplication and different versions for dependancies
The current POMs have a lot of duplication (i.e.: same plugins copy+pasted) and use the same dependancies (i.e.: version specified in 2 different POMs) or use different versions (i.e.: OSGi framework).
Would make it easier going forward to have a parent POM to control versions/common plugins across the components.
do you think that upgrade to gradle would make more sense?
@buzdin ; i don't have any experience of gradle / not seen any customers use it in my day-to-day job, but happy to create an issue to add support and investigate.
|
gharchive/issue
| 2017-04-13T16:07:05 |
2025-04-01T04:54:56.128908
|
{
"authors": [
"buzdin",
"garethahealy"
],
"repo": "DozerMapper/dozer",
"url": "https://github.com/DozerMapper/dozer/issues/326",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1146255516
|
🛑 Moegirl Commons is down
In 33096d2, Moegirl Commons (https://commons.moegirl.org.cn/Mainpage) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Moegirl Commons is back up in 8c5b344.
|
gharchive/issue
| 2022-02-21T21:54:30 |
2025-04-01T04:54:56.136208
|
{
"authors": [
"Dragon-Fish"
],
"repo": "Dragon-Fish/moegirl-uptime",
"url": "https://github.com/Dragon-Fish/moegirl-uptime/issues/566",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2562934641
|
Docker WebUI want start
OS: Fedora Linux 40 (Workstation Edition) x86_64
Kernel: 6.10.7-200.fc40.x86_64
CPU: AMD Ryzen 9 5900X (24) @ 3.700GHz
GPU: NVIDIA GeForce RTX 3090
Memory: 96457MiB
Docker WebUI want start
If I execute docker run -it --rm --gpus all -p 7860:7860 athomasson2/ebook2audiobookxtts:latest the Webui won't start. I simply drop to a root shell on the container, diplaying the following:
==========
== CUDA ==
==========
CUDA Version 11.8.0
Container image Copyright (c) 2016-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
This container image and its contents are governed by the NVIDIA Deep Learning Container License.
By pulling and using the container, you accept the terms and conditions of this license:
https://developer.nvidia.com/ngc/nvidia-deep-learning-container-license
A copy of this license is made available in this container at /NGC-DL-CONTAINER-LICENSE for your convenience.
If I then try to manually start the webui, the installation seems broken
root@42368a76209a:/ebook2audiobookXTTS# python custom_model_ebook2audiobookXTTS_gradio.py
starting...
Traceback (most recent call last):
File "/ebook2audiobookXTTS/custom_model_ebook2audiobookXTTS_gradio.py", line 7, in <module>
from pydub import AudioSegment
ModuleNotFoundError: No module named 'pydub'
Can't build container from Dockerfile
If I try to build the container myself using the Dockerfile it looks like scapy can't compile.
Could be related to this (already closed) issue https://github.com/explosion/spaCy/issues/13449.
597.0 Downloading spacy-3.0.6.tar.gz (7.1 MB)
599.8 ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 7.1/7.1 MB 2.5 MB/s eta 0:00:00
600.2 Installing build dependencies: started
607.1 Installing build dependencies: finished with status 'done'
607.1 Getting requirements to build wheel: started
608.2 Getting requirements to build wheel: finished with status 'error'
608.2 error: subprocess-exited-with-error
608.2
608.2 × Getting requirements to build wheel did not run successfully.
608.2 │ exit code: 1
608.2 ╰─> [164 lines of output]
608.2
608.2 Error compiling Cython file:
608.2 ------------------------------------------------------------
608.2 ...
608.2 int length
608.2
608.2
608.2 cdef class Vocab:
608.2 cdef Pool mem
608.2 cpdef readonly StringStore strings
608.2 ^
608.2 ------------------------------------------------------------
608.2
608.2 spacy/vocab.pxd:28:10: Variables cannot be declared with 'cpdef'. Use 'cdef' instead.
608.2
608.2 Error compiling Cython file:
608.2 ------------------------------------------------------------
608.2 ...
608.2
608.2
608.2 cdef class Vocab:
608.2 cdef Pool mem
608.2 cpdef readonly StringStore strings
608.2 cpdef public Morphology morphology
608.2 ^
608.2 ------------------------------------------------------------
608.2
608.2 spacy/vocab.pxd:29:10: Variables cannot be declared with 'cpdef'. Use 'cdef' instead.
608.2
608.2 Error compiling Cython file:
608.2 ------------------------------------------------------------
608.2 ...
608.2
608.2 cdef class Vocab:
608.2 cdef Pool mem
608.2 cpdef readonly StringStore strings
608.2 cpdef public Morphology morphology
608.2 cpdef public object vectors
608.2 ^
608.2 ------------------------------------------------------------
608.2
608.2 spacy/vocab.pxd:30:10: Variables cannot be declared with 'cpdef'. Use 'cdef' instead.
608.2
608.2 Error compiling Cython file:
608.2 ------------------------------------------------------------
608.2 ...
608.2 cdef class Vocab:
608.2 cdef Pool mem
608.2 cpdef readonly StringStore strings
608.2 cpdef public Morphology morphology
608.2 cpdef public object vectors
608.2 cpdef public object _lookups
608.2 ^
608.2 ------------------------------------------------------------
608.2
608.2 spacy/vocab.pxd:31:10: Variables cannot be declared with 'cpdef'. Use 'cdef' instead.
608.2
608.2 Error compiling Cython file:
608.2 ------------------------------------------------------------
608.2 ...
608.2 cdef Pool mem
608.2 cpdef readonly StringStore strings
608.2 cpdef public Morphology morphology
608.2 cpdef public object vectors
608.2 cpdef public object _lookups
608.2 cpdef public object writing_system
608.2 ^
608.2 ------------------------------------------------------------
608.2
608.2 spacy/vocab.pxd:32:10: Variables cannot be declared with 'cpdef'. Use 'cdef' instead.
608.2
608.2 Error compiling Cython file:
608.2 ------------------------------------------------------------
608.2 ...
608.2 cpdef readonly StringStore strings
608.2 cpdef public Morphology morphology
608.2 cpdef public object vectors
608.2 cpdef public object _lookups
608.2 cpdef public object writing_system
608.2 cpdef public object get_noun_chunks
608.2 ^
608.2 ------------------------------------------------------------
608.2
608.2 spacy/vocab.pxd:33:10: Variables cannot be declared with 'cpdef'. Use 'cdef' instead.
608.2
608.2 Error compiling Cython file:
608.2 ------------------------------------------------------------
608.2 ...
608.2 cdef float prior_prob
608.2
608.2
608.2 cdef class KnowledgeBase:
608.2 cdef Pool mem
608.2 cpdef readonly Vocab vocab
608.2 ^
608.2 ------------------------------------------------------------
608.2
608.2 spacy/kb.pxd:31:10: Variables cannot be declared with 'cpdef'. Use 'cdef' instead.
608.2 Copied /tmp/pip-install-s1yoqgkp/spacy_a3a04683ec794ce8805d0f2ae47f9987/setup.cfg -> /tmp/pip-install-s1yoqgkp/spacy_a3a04683ec794ce8805d0f2ae47f9987/spacy/tests/package
608.2 Copied /tmp/pip-install-s1yoqgkp/spacy_a3a04683ec794ce8805d0f2ae47f9987/pyproject.toml -> /tmp/pip-install-s1yoqgkp/spacy_a3a04683ec794ce8805d0f2ae47f9987/spacy/tests/package
608.2 Cythonizing sources
608.2 Compiling spacy/training/example.pyx because it changed.
608.2 Compiling spacy/parts_of_speech.pyx because it changed.
608.2 Compiling spacy/strings.pyx because it changed.
608.2 Compiling spacy/lexeme.pyx because it changed.
608.2 Compiling spacy/vocab.pyx because it changed.
608.2 Compiling spacy/attrs.pyx because it changed.
608.2 Compiling spacy/kb.pyx because it changed.
608.2 Compiling spacy/ml/parser_model.pyx because it changed.
608.2 Compiling spacy/morphology.pyx because it changed.
608.2 Compiling spacy/pipeline/dep_parser.pyx because it changed.
608.2 Compiling spacy/pipeline/morphologizer.pyx because it changed.
608.2 Compiling spacy/pipeline/multitask.pyx because it changed.
608.2 Compiling spacy/pipeline/ner.pyx because it changed.
608.2 Compiling spacy/pipeline/pipe.pyx because it changed.
608.2 Compiling spacy/pipeline/trainable_pipe.pyx because it changed.
608.2 Compiling spacy/pipeline/sentencizer.pyx because it changed.
608.2 Compiling spacy/pipeline/senter.pyx because it changed.
608.2 Compiling spacy/pipeline/tagger.pyx because it changed.
608.2 Compiling spacy/pipeline/transition_parser.pyx because it changed.
608.2 Compiling spacy/pipeline/_parser_internals/arc_eager.pyx because it changed.
608.2 Compiling spacy/pipeline/_parser_internals/ner.pyx because it changed.
608.2 Compiling spacy/pipeline/_parser_internals/nonproj.pyx because it changed.
608.2 Compiling spacy/pipeline/_parser_internals/_state.pyx because it changed.
608.2 Compiling spacy/pipeline/_parser_internals/stateclass.pyx because it changed.
608.2 Compiling spacy/pipeline/_parser_internals/transition_system.pyx because it changed.
608.2 Compiling spacy/pipeline/_parser_internals/_beam_utils.pyx because it changed.
608.2 Compiling spacy/tokenizer.pyx because it changed.
608.2 Compiling spacy/training/align.pyx because it changed.
608.2 Compiling spacy/training/gold_io.pyx because it changed.
608.2 Compiling spacy/tokens/doc.pyx because it changed.
608.2 Compiling spacy/tokens/span.pyx because it changed.
608.2 Compiling spacy/tokens/token.pyx because it changed.
608.2 Compiling spacy/tokens/span_group.pyx because it changed.
608.2 Compiling spacy/tokens/graph.pyx because it changed.
608.2 Compiling spacy/tokens/morphanalysis.pyx because it changed.
608.2 Compiling spacy/tokens/_retokenize.pyx because it changed.
608.2 Compiling spacy/matcher/matcher.pyx because it changed.
608.2 Compiling spacy/matcher/phrasematcher.pyx because it changed.
608.2 Compiling spacy/matcher/dependencymatcher.pyx because it changed.
608.2 Compiling spacy/symbols.pyx because it changed.
608.2 Compiling spacy/vectors.pyx because it changed.
608.2 [ 1/41] Cythonizing spacy/attrs.pyx
608.2 [ 2/41] Cythonizing spacy/kb.pyx
608.2 Traceback (most recent call last):
608.2 File "/usr/local/lib/python3.10/site-packages/pip/_vendor/pyproject_hooks/_in_process/_in_process.py", line 353, in <module>
608.2 main()
608.2 File "/usr/local/lib/python3.10/site-packages/pip/_vendor/pyproject_hooks/_in_process/_in_process.py", line 335, in main
608.2 json_out['return_val'] = hook(**hook_input['kwargs'])
608.2 File "/usr/local/lib/python3.10/site-packages/pip/_vendor/pyproject_hooks/_in_process/_in_process.py", line 118, in get_requires_for_build_wheel
608.2 return hook(config_settings)
608.2 File "/tmp/pip-build-env-sbvqvvd7/overlay/lib/python3.10/site-packages/setuptools/build_meta.py", line 332, in get_requires_for_build_wheel
608.2 return self._get_build_requires(config_settings, requirements=[])
608.2 File "/tmp/pip-build-env-sbvqvvd7/overlay/lib/python3.10/site-packages/setuptools/build_meta.py", line 302, in _get_build_requires
608.2 self.run_setup()
608.2 File "/tmp/pip-build-env-sbvqvvd7/overlay/lib/python3.10/site-packages/setuptools/build_meta.py", line 318, in run_setup
608.2 exec(code, locals())
608.2 File "<string>", line 224, in <module>
608.2 File "<string>", line 211, in setup_package
608.2 File "/tmp/pip-build-env-sbvqvvd7/overlay/lib/python3.10/site-packages/Cython/Build/Dependencies.py", line 1154, in cythonize
608.2 cythonize_one(*args)
608.2 File "/tmp/pip-build-env-sbvqvvd7/overlay/lib/python3.10/site-packages/Cython/Build/Dependencies.py", line 1321, in cythonize_one
608.2 raise CompileError(None, pyx_file)
608.2 Cython.Compiler.Errors.CompileError: spacy/kb.pyx
608.2 [end of output]
608.2
608.2 note: This error originates from a subprocess, and is likely not a problem with pip.
608.2 error: subprocess-exited-with-error
608.2
608.2 × Getting requirements to build wheel did not run successfully.
608.2 │ exit code: 1
608.2 ╰─> See above for output.
608.2
608.2 note: This error originates from a subprocess, and is likely not a problem with pip.
------
Dockerfile:30
--------------------
28 | # Install Python dependencies
29 | RUN pip install --upgrade pip
30 | >>> RUN pip install bs4 pydub nltk beautifulsoup4 ebooklib tqdm mecab-python3 tts==0.21.3
31 |
32 | # Download unidic
--------------------
ERROR: failed to solve: process "/bin/sh -c pip install bs4 pydub nltk beautifulsoup4 ebooklib tqdm mecab-python3 tts==0.21.3" did not complete successfully: exit code: 1
Interesting....
Try running this docker?
It's the docker of this space tbh, but it's up and running so we know it works at least
HUGGINGFACE SPACE
'''bash
docker run -it -p 7860:7860 --platform=linux/amd64
registry.hf.space/drewthomasson-ebook2audiobookxtts:latest python app.py
'''
I'll see about updating the docker file in the readme,
To make it more inline with this one
Docker hub-page
just modified the dockerfile with the GPU one instead
|
gharchive/issue
| 2024-10-03T01:30:56 |
2025-04-01T04:54:56.196021
|
{
"authors": [
"DrewThomasson",
"matthiss"
],
"repo": "DrewThomasson/ebook2audiobookXTTS",
"url": "https://github.com/DrewThomasson/ebook2audiobookXTTS/issues/13",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2229768250
|
Flipping tiles on an axis
Great library. I'm render a Tiled map and am unable to find any way to flip a tile on an axis. Does this a feature exist, or would implementing this feature be possible?
That should be possible with custom shader code. Lemme merge some WIP code for that and then i'll add an example on how to do it. Prolly next weekend-ish.
@cedtwo Can you check out the custom_shader_code example in 0.7.3? That should help with flipping tiles :)
|
gharchive/issue
| 2024-04-07T13:16:29 |
2025-04-01T04:54:56.204805
|
{
"authors": [
"Droggelbecher",
"cedtwo"
],
"repo": "Droggelbecher/bevy-fast-tilemap",
"url": "https://github.com/Droggelbecher/bevy-fast-tilemap/issues/24",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1898525418
|
Use explicit API for core/model module
Idea Description
You can see the 2022 pr
Reference images and links
https://github.com/DroidKaigi/conference-app-2022/pull/554/files
🤚🏻
Thanks! Assigned 👍
|
gharchive/issue
| 2023-09-15T14:19:14 |
2025-04-01T04:54:56.206550
|
{
"authors": [
"mikanIchinose",
"takahirom"
],
"repo": "DroidKaigi/conference-app-2023",
"url": "https://github.com/DroidKaigi/conference-app-2023/issues/1220",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
369346589
|
Change namespace of converts to sub namespace "currencies"
Fix phpcs issues
Remove todo rule from phpcs.xml
Coverage remained the same at 100.0% when pulling 0549c922e87b2d8514db103025fed5fa186cf816 on some_changes into 61fbf0a83546621457cbfce1748be3058c44206a on master.
|
gharchive/pull-request
| 2018-10-11T23:41:50 |
2025-04-01T04:54:56.221898
|
{
"authors": [
"Dropelikeit",
"coveralls"
],
"repo": "Dropelikeit/PriceCalculator",
"url": "https://github.com/Dropelikeit/PriceCalculator/pull/12",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
907126949
|
Reward System?
I love RPG games even if they are text-based, Good to know it's in development. I was wondering you could add reward system like if player founds a key in the dungeon then he shall be rewarded with a clue or something so he can pass the next stage easily.
Maybe that could be a nice addition once I get around to adding dungeons.
|
gharchive/issue
| 2021-05-31T06:39:47 |
2025-04-01T04:54:56.246280
|
{
"authors": [
"DuhonTheGuy",
"itsOwen"
],
"repo": "DuhonTheGuy/RPG-Generator",
"url": "https://github.com/DuhonTheGuy/RPG-Generator/issues/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
718904075
|
[Feature] Do not allow players to re-enter played games in 2020
So that they cannot retry to get a higher score.
Done.
|
gharchive/issue
| 2020-10-11T19:00:03 |
2025-04-01T04:54:56.262595
|
{
"authors": [
"yo1995"
],
"repo": "DukeMobileDevCenter/Jigsaw",
"url": "https://github.com/DukeMobileDevCenter/Jigsaw/issues/99",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
1327218290
|
Add Subscription Modal
Changes
create Modal Wrapper Component
create Subscription Modal Body
fix the home view dependent on the user's role
fix the add room form such that student role is dependent on the teacher role.
Screenshots
Great code, just 2 comments.
Links should open in new tab
why so many changes to package-lock?
|
gharchive/pull-request
| 2022-08-03T13:27:14 |
2025-04-01T04:54:56.265083
|
{
"authors": [
"ShenyiCui",
"astraxq"
],
"repo": "Dulwich-Bookings/Dulwich-Bookings-Frontend",
"url": "https://github.com/Dulwich-Bookings/Dulwich-Bookings-Frontend/pull/45",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.