id
stringlengths 4
10
| text
stringlengths 4
2.14M
| source
stringclasses 2
values | created
timestamp[s]date 2001-05-16 21:05:09
2025-01-01 03:38:30
| added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| metadata
dict |
---|---|---|---|---|---|
2541249730 | Task_excercise code review request
Modified code to match google standard, added more comments, modified classes and changed main.
Remove files
Then add a git ignore files for the build directory. Exclude something that is not [ cmake, cpp , h or .hpp. ]
| gharchive/pull-request | 2024-09-22T20:42:18 | 2025-04-01T06:37:05.489529 | {
"authors": [
"DJOA-UP",
"Tonix22"
],
"repo": "JERL88/FMDAF",
"url": "https://github.com/JERL88/FMDAF/pull/6",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1938702412 | TROJ_GEN.R002V01IT23
This is probably a false positive, but since the penultimate version this has been detected in Virustotal
Obviously it's a false positive. Don't know why now it's reporting the app as a malware but it's completely false. The project is open source and everyone can read the code and search malicious behaviors.
| gharchive/issue | 2023-10-11T20:47:47 | 2025-04-01T06:37:05.498422 | {
"authors": [
"JGeek00",
"zekabra"
],
"repo": "JGeek00/adguard-home-manager",
"url": "https://github.com/JGeek00/adguard-home-manager/issues/62",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2313720122 | Microzed freezes after some hours of operation
This is an unpredictable bug: the server on Microzed looses connection with BLACS worker after some operation time.
I suggest using sipyco, zprocess or Cap&Proto
Also protobuf could work since it is well maintained.
| gharchive/issue | 2024-05-23T19:57:05 | 2025-04-01T06:37:05.568826 | {
"authors": [
"restelli"
],
"repo": "JQIamo/jane",
"url": "https://github.com/JQIamo/jane/issues/6",
"license": "BSD-2-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
577099016 | Possible improvements for library
This PR adds:
Add types for Observable Notification and Network
Add documentation to operators
Change network.connection API to use fromEventPattern
Add errors where approriate
These probably need some test coverage running in a real browser such as Karma tests.
This is amazing! thanks a lot for all the work you put into that! Are you still working on something or can I merge this?
@JWO719 if the tests pass green it's ready to merge :)
At the moment they are highly mocked. I tried investigating Jest in the browser, I can get puppeteer to work but not webdriver. Also JSDom causes similar issues.
Without figuring out a way to test them natively (and not sure geolocation is possible) this was the best solution I could come up with.
@JWO719 See #2 for more improvements on this branch. That change has a bigger change in the repo itself (using @nrwl/nx) but now also has the outline of an app that can be used to run in browser, also run E2E tests (as Jest will only ever be mockable)
I'll close this in favor of #2
| gharchive/pull-request | 2020-03-06T18:22:46 | 2025-04-01T06:37:05.577437 | {
"authors": [
"JWO719",
"tanepiper"
],
"repo": "JWO719/rxjs-web",
"url": "https://github.com/JWO719/rxjs-web/pull/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
188235538 | Operators in Constraints
I would like to know if it was possible to include the option to add up properties of variables in the constraints section. Let's see an example: I have a scenario where a farmer wants to optimise his profits by evaluating the most profitable growing option:
{ "name": "Crop Rotation Problem", "optimize": "Gross Margin", "opType": "max", "constraints": { "field0": { "max": 1 }, "field1": { "max": 1 }, "Wheat": { "max": 10 } }, "variables": { "field00": { "field0": 1, "Wheat": 3.6, "Gross Margin": 3052.44 }, "field01": { "field0": 1, "Summer Wheat": 3.6, "Gross Margin": 1787.58 }, "field02": { "field0": 1, "Summer Barley": 3.6, "Gross Margin": 1789.65 }, "field10": { "field1": 1, "Corn": 11.9, "Gross Margin": 7243.830000000001 }, "field11": { "field1": 1, "Soybeans": 11.9, "Gross Margin": 4710.235 }, "field12": { "field1": 1, "Onions": 11.9, "Gross Margin": 46602.325000000004 }, }, "ints": { "field00": 1, "field01": 1, "field02": 1, "field10": 1, "field11": 1, "field12": 1, } }
Now for example imagine there was a policy restriction only allowing the farmer to grow a maximum of 75% of his entire acerage with the sum of Wheat and Summer Barley. To solve this, it would be nice if the Solver would be able to handle the following:
{ "name": "Crop Rotation Problem", "optimize": "Gross Margin", "opType": "max", "constraints": { "field0": { "max": 1 }, "field1": { "max": 1 }, "Wheat" + "Summer Barley": { "max": 10 } },...
Maybe there is a nicer and mathematically more appealing way to achieve this, but for now that's the only way I could think of.
Thanks already for the awesome work!
Hi Toffi,
I think that the simplest way would be to add the contribution to a constraint that corresponds to the maximum total area (maxFields here) for all the variables that have an impact on the total area (all the variables in this case):
"constraints": {
"maxFields": { "max": 10 },
"field0": { "max": 1 },
"field1": { "max": 1 },
"Wheat": { "max": 10 } }
},
"variables": {
"field00": { "maxFields": 1, "field0": 1, "Wheat": 3.6, "Gross Margin": 3052.44 }, }
"field01": { "maxFields": 1, "field0": 1, "Summer Wheat": 3.6, "Gross Margin": 1787.58 },
"field02": { "maxFields": 1, "field0": 1, "Summer Barley": 3.6, "Gross Margin": 1789.65 },
"field10": { "maxFields": 1, "field1": 1, "Corn": 11.9, "Gross Margin": 7243.83 },
"field11": { "maxFields": 1, "field1": 1, "Soybeans": 11.9, "Gross Margin": 4710.235 },
"field12": { "maxFields": 1, "field1": 1, "Onions": 11.9, "Gross Margin": 46602.325 },
}
What do you think?
Nice, that makes a lot of sense! Though I'm still struggling with another concept related to this issue: the policy may force our example farmer to grow a minimum of 2 crops on his farm. In GAMS one could create a variable CropsUsed and check whether the sum of CropsUsed was greater than 2. Now I don't see a way to get this to match the setup above, I was thinking of assigning different values for each crop, but realised that it would not get me any further. Do you probably have an idea? Thanks a lot already!
I am not sure I understand. If you want to force a minimum number of fields, you can achieve it the following way:
"constraints": {
"cropsUsed": { "min": 2, "max": 10 },
"field0": { "max": 1 },
"field1": { "max": 1 },
"Wheat": { "max": 10 } }
},
"variables": {
"field00": { "cropsUsed": 1, "field0": 1, "Wheat": 3.6, "Gross Margin": 3052.44 }, }
"field01": { "cropsUsed": 1, "field0": 1, "Summer Wheat": 3.6, "Gross Margin": 1787.58 },
"field02": { "cropsUsed": 1, "field0": 1, "Summer Barley": 3.6, "Gross Margin": 1789.65 },
"field10": { "cropsUsed": 1, "field1": 1, "Corn": 11.9, "Gross Margin": 7243.83 },
"field11": { "cropsUsed": 1, "field1": 1, "Soybeans": 11.9, "Gross Margin": 4710.235 },
"field12": { "cropsUsed": 1, "field1": 1, "Onions": 11.9, "Gross Margin": 46602.325 },
}
Notice that I just renamed maxFields into cropsUsed and added a lower bound to the constraint.
Sorry for being unclear, I think the first example is too minified. The farmer can only grow one kind of plant on each of his fields. He has multiple options though, for field0 example he has 9 options (field00 - field 08), for field1 he has only 5 options. He would like to chose the combination of options leading to the highest gross margin, which in this case would be to grow Onions on both field0 and field1. But the policy requires him to grow at least 2 crops on his entire farm (or even 3 if he had more than 2 fields), so any other combination would be acceptable, like field0 Onions and field1 Winter Wheat.
{
"name": "Crop Rotation",
"optimize": "Gross Margin",
"opType": "max",
"constraints": {
"field0": {
"max": 1
},
"feld1": {
"max": 1
},
"Potato": {
"max": 10
}
},
"variables": {
"field00": {
"field": 1,
"Sugar Beets": 11.9,
"Gross Margin": 9915.792
},
"field01": {
"field": 1,
"Greening": 11.9,
"Gross Margin": 11.9
},
"field02": {
"field": 1,
"Winter Wheat": 11.9,
"Gross Margin": 10121.07
},
"field03": {
"field": 1,
"Winter Barley": 11.9,
"Gross Margin": 8123.695000000001
},
"field04": {
"field": 1,
"Winter Rye": 11.9,
"Gross Margin": 6995.218
},
"field05": {
"field": 1,
"Summer Barley": 11.9,
"Gross Margin": 5935.2
},
"field06": {
"field": 1,
"Oats": 11.9,
"Gross Margin": 7517.308
},
"field07": {
"field": 1,
"Triticale": 11.9,
"Gross Margin": 6527.24
},
"field08": {
"field": 1,
"Onions": 11.9,
"Gross Margin": 46494.196
},
"feld10": {
"feld1": 1,
"Summer Wheat": 2.8,
"Gross Margin": 1389.98
},
"feld11": {
"feld1": 1,
"Summer Barley": 2.8,
"Gross Margin": 1391.65
},
"feld12": {
"feld1": 1,
"Winter Rye": 2.8,
"Gross Margin": 1639.4109999999998
},
"feld13": {
"feld1": 1,
"Triticale": 2.8,
"Gross Margin": 1529.98
},
"feld14": {
"feld1": 1,
"Onions": 11.9,
"Gross Margin": 46494.196
}
},
"ints": {
"field0": 1,
"field1": 1,
"field2": 1,
"field3": 1,
"field4": 1,
"field5": 1,
"field6": 1,
"field7": 1,
"field8": 1,
"feld10": 1,
"feld11": 1,
"feld12": 1,
"feld13": 1,
"feld14": 1
}
}
I hope that helps clarifying the issue, I've been thinking about it all day and it's driving me crazy. Thought about assigning prime numbers to each crop and then dividing the sum of them through each individual prime number (if the results was even, then just one crop would be grown), but a) this only works for 2 crops (not 3 or even more) and b) I still can't use operators in the constraint section.
Thanks for helping me out!
Basically you want to get at least n different types of crops. The solution would be to use the "big M" method. In your example, the idea is to force diversity of crops to 2. Let's say there are only 3 types of crops b, r and o (for barley, rye and onion), we can create 3 binary variables y_b, y_r and y_o corresponding to whether each crop is grown or not (binary variable => value either 0 or 1).
Then we can create a constraint y_b + y_r + y_o >= 2 (diversity constraint), that will force our diversity to be at least 2.
Now, we need to force those binary variables to 0 or 1 when the correspond crops are grown. 2 constraints per type of crops are needed: one to restrain the binary variable to 0 and one to force it to 1:
The restrain constraint is quite straightforward: y_o <= nb_onion_fields (can be rewritten 0 <= nb_onion_fields - y_o). This constraint ensures that if no onion field is grown, the associated binary variable will be 0.
The force constraint is a tiny bit more tricky: nb_onion_fields <= y_o * M (can be rewritten nb_onion_fields - y_o * M <= 0). This constraint ensures that if onion is grown, the associated binary variable will be 1. Notice that the value for M should be big enough so that it does not constraint the number of onion fields you can grow.
Here is the json formulation for what I stated above (by the way, the simpler the example the better).
{
"name": "Crop Rotation",
"optimize": "Gross Margin",
"opType": "max",
"constraints": {
"field0": {
"max": 6
},
"feld1": {
"max": 6
},
"cropsUsed": {
"max": 10
},
"diversity": {
"min": 2
},
"restrainOnion": {
"min": 0
},
"restrainSummerBarley": {
"min": 0
},
"restrainWinterRye": {
"min": 0
},
"forceOnion": {
"max": 0
},
"forceSummerBarley": {
"max": 0
},
"forceWinterRye": {
"max": 0
}
},
"variables": {
"growOnion": {
"diversity": 1,
"restrainOnion": -1,
"forceOnion": -999
},
"growSummerBarley": {
"diversity": 1,
"restrainSummerBarley": -1,
"forceSummerBarley": -999
},
"growWinterRye": {
"diversity": 1,
"restrainWinterRye": -1,
"forceWinterRye": -999
},
"field00": {
"field0": 1,
"cropsUsed": 1,
"Winter Rye": 11.9,
"Gross Margin": 6995.218,
"restrainWinterRye": 1,
"forceWinterRye": 1
},
"field01": {
"field0": 1,
"cropsUsed": 1,
"Summer Barley": 11.9,
"Gross Margin": 5935.2,
"restrainSummerBarley": 1,
"forceSummerBarley": 1
},
"field02": {
"field0": 1,
"cropsUsed": 1,
"Onions": 11.9,
"Gross Margin": 46494.196,
"restrainOnion": 1,
"forceOnion": 1
},
"feld10": {
"feld1": 1,
"cropsUsed": 1,
"Summer Barley": 2.8,
"Gross Margin": 1391.65,
"restrainSummerBarley": 1,
"forceSummerBarley": 1
},
"feld11": {
"feld1": 1,
"cropsUsed": 1,
"Winter Rye": 2.8,
"Gross Margin": 1639.411,
"restrainWinterRye": 1,
"forceWinterRye": 1
},
"feld12": {
"feld1": 1,
"cropsUsed": 1,
"Onions": 11.9,
"Gross Margin": 46494.196,
"restrainOnion": 1,
"forceOnion": 1
}
},
"ints": {
"field00": 1,
"field01": 1,
"field02": 1,
"feld10": 1,
"feld11": 1,
"feld12": 1,
},
"binaries": {
"growOnion": 1,
"growSummerBarley": 1,
"growWinterRye": 1
}
}
You can see that I decided to set M = 999 but M=10 would have been sufficient since no more than 10 crops can be grown.
Wow that perfectly solves my issue! Thanks so much for your advice, hope farmers will benefit from it someday 👍
no problem, let us know if you need further help
Sorry for coming back at you, but as I was implementing and checking the code I realised it wasn't solving as expected. I attached a MWE for recreating the issue, sorry for that it's so long.
When solving the attached problem, the diversity constraint (diversity > 3) is fulfilled by growPotatoes having a value of 2, and growOnions having a value of 1, despite them being declared as binary variables. It looks like my declaration of these variables as binaries is being ignored somehow. Another constraint that restricts the amount of each pair of crops grown to 95% of the entire acerage also seems to be ignored (as this constraint should also enforce the diversity of crops to a minimum of 3).
Thanks for your help once again in advance!
crop_rotation.txt
Sorry for the delay,
One possible reason why it won't solve for particular problem configurations is because the solver can be numerically unstable (and loop forever) but it's rare. It is a work in progress to make it stable but it takes a consequent amount of time. If that is the reason, you might just have to wait for the update (a reasonable ETA may be not before 3 months).
Is it important for that configuration to work? Does your work relying on the solver need to be ready soon?
Well thanks for having a look once again!
The solver is used (as you already figured I guess) to optimize crop rotations for farmers considering different aspects. It's a university project, so it will be free of charge when available. Right now, I'm hoping to get it online some time next year, also releasing it on github then. If you're interested in the project or the way the tableau is created just drop me an email, as I don't feel the code is ready to be published yet.
Without digging too deep into it, what about something like this:
You have a farm that has 4 fields on it. You can only grow 1 crop per field. Your crop choices are:
Barley (not available in field 4 because ...)
Wheat
Summer Wheat
Soy (not available for field 3 because of ...)
Onions
You can only have 1 crop / field.
{
name: "problem_2",
opType: "max",
optimize: "profit",
constraints: {
acres: {max: 100},
a_limiter: {max: 75},
f1: {max: 1},
f2: {max: 1},
f3: {max: 1},
f4: {max: 1}
},
variables: {
// Field 1 can grow these crops
f1_barley: {acres: 1, a_limiter: 1, f1: 1, profit: ?},
f1_wheat: {acres: 1, a_limiter: 1, f1: 1, profit: ?},
f1_summer_wheat: {acres: 1, f1: 1, profit: ?},
f1_soy: {acres: 1, profit: f1: 1, profit: ?},
f1_onions: {acres: 1, f1: 1, profit: ?},
// Field 2 can grow these crops
f2_barley: {acres: 1, a_limiter: 1, f2: 1, profit: ?},
f2_wheat: {acres: 1, a_limiter: 1, f2: 1, profit: ?},
f2_summer_wheat: {acres: 1, f2: 1, profit: ?},
f2_soy: {acres: 1, f1: 1, f2: 1, profit: ?},
f2_onions: {acres: 1, f2: 1, profit: ?},
// Field 3 can grow these crops...Notice Soy is Gone
f3_barley: {acres: 1, a_limiter: 1, f3: 1,profit: ?},
f3_wheat: {acres: 1, a_limiter: 1, f3: 1,profit: ?},
f3_summer_wheat: {acres: 1, f3: 1, profit: ?},
f3_onions: {acres: 1, f3: 1, profit: ?},
// Field 4 can grow these crops...Notice Barley is Gone
f4_wheat: {acres: 1, a_limiter: 1, f4: 1,profit: ?},
f4_summer_wheat: {acres: 1, f4: 1, profit: ?},
f4_soy: {acres: 1, f4: 1, profit: ?},
f4_onions: {acres: 1, f4: 1, profit: ?},
},
ints: {
f1_barley: 1,
f2_barley: 1,
...
..
.
f4_onions: 1
}
}
Thanks for the contribution! Being new to GitHub I just realized that I should close this issue, as the original question was resolved by bchevalier's answer (using the big M-Method).
The setup currently used is working similar to your approach, with some differences in the constraints to fulfill the requirements of the Common Agricultural Policy of the EU. The first issue is that each field is different in size, so restricting each crop production area to 75% of the total has to be done through the actual crop area grown. Assuming the total area would be 100, then
Wheat: 75 Soy: 75 ...
and in the variable
field1_option1: {field1: 1, Wheat: 5.6, Profit: ?} field1_option2: {field1: 1, Soy: 5.6, Profit: ?}
Then, each combination of crops cannot exceed 95% of the cropping area, therefore the following is added to the constraints
WheatBarley: 95 WheatSoy: 95 ...
and in the variables
field1_option1: {field1: 1, Wheat: 5.6, WheatBarley: 5.6, WheatSoy: 5.6, ..., Profit: ?} field1_option2: {field1: 1, Soy: 5.6, WheatSoy: 5.6, SoyOnions: 5.6, ..., Profit: ?}
As you can already see, the more growing options there are, the longer the constraints are going to be for each field.
Right now, as described above, this seems to be the source of the issues whith really long solve times for problems where this 95% constraints become binding.
To get a little closer to reality, even more constraints will be added, like available work hours during different times of the year, available machine hours and others. It would be perfect to achieve all this within JavaScript, as the available options for including GAMS are not as promising.
| gharchive/issue | 2016-11-09T12:25:52 | 2025-04-01T06:37:05.609074 | {
"authors": [
"JWally",
"Toffi-123",
"bchevalier"
],
"repo": "JWally/jsLPSolver",
"url": "https://github.com/JWally/jsLPSolver/issues/47",
"license": "unlicense",
"license_type": "permissive",
"license_source": "bigquery"
} |
2003566736 | 🛑 Glitch is down
In 64c1de5, Glitch ($GLITCH) was down:
HTTP code: 503
Response time: 244 ms
Resolved: Glitch is back up in 3adf780 after 26 minutes.
| gharchive/issue | 2023-11-21T06:40:29 | 2025-04-01T06:37:05.612783 | {
"authors": [
"JYFUX"
],
"repo": "JYFUX/upptime",
"url": "https://github.com/JYFUX/upptime/issues/2718",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2616666463 | ci: Adjust size label thresholds for pull requests
Pull Request
Description
Please include a summary of the change and which issue is fixed. Please also include relevant motivation and context. List any dependencies that are required for this change.
#211 👈
main
This stack of pull requests is managed by Graphite. Learn more about stacking.
Join @JackPlowman and the rest of your teammates on Graphite
Merge activity
Oct 27, 11:19 AM EDT: A user merged this pull request with Graphite.
| gharchive/pull-request | 2024-10-27T15:15:36 | 2025-04-01T06:37:05.656279 | {
"authors": [
"JackPlowman"
],
"repo": "JackPlowman/github-stats",
"url": "https://github.com/JackPlowman/github-stats/pull/211",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
241468114 | custom_commands.json, command groups, and aliases
still need to work on:
custom_commands system,
command groups (secondary/main),
command aliases,
parameters
Command Group: Ready
Custom Commands: Ready
Command aliases/parameters: Not Ready
Still need to move some of my custom commnds, doing it later
| gharchive/issue | 2017-07-08T17:33:38 | 2025-04-01T06:37:05.706181 | {
"authors": [
"Jackzmc"
],
"repo": "Jackzmc/JackzCo-Bot",
"url": "https://github.com/Jackzmc/JackzCo-Bot/issues/8",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
1103133309 | 计算机学院:请求完成通信工程2019级部分
[ ] Update more experimental reports
[ ] Update more notes
[ ] Much more professional evaluations
Add "TODO-List heading" and thanks for your future contributions!
| gharchive/issue | 2022-01-14T07:01:25 | 2025-04-01T06:37:05.707694 | {
"authors": [
"Jacob953",
"LegendZi"
],
"repo": "Jacob953/evalcsu",
"url": "https://github.com/Jacob953/evalcsu/issues/10",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
} |
156766326 | Bullets within a bullet
Currently unable to list bullets within a bullet.
Jingwei
This is solved in my newest version. I'll close the issue once I merge my branch.
I've merged branches, this is now implemented.
| gharchive/issue | 2016-05-25T14:33:14 | 2025-04-01T06:37:05.713661 | {
"authors": [
"Sodaaaa",
"szymczdm"
],
"repo": "JacquesCarette/literate-scientific-software",
"url": "https://github.com/JacquesCarette/literate-scientific-software/issues/7",
"license": "bsd-2-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
59929313 | Add main.py to make package executable
This is convenient for running repository version with python -m psdash
Thanks!
| gharchive/pull-request | 2015-03-05T09:59:27 | 2025-04-01T06:37:05.750362 | {
"authors": [
"Jahaja",
"techtonik"
],
"repo": "Jahaja/psdash",
"url": "https://github.com/Jahaja/psdash/pull/46",
"license": "cc0-1.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
183342082 | RxBinding need a new action for every click
I found I cannot use it to manager all my click code together. I alway setOnClickListenr for the same one ,and than I can create less OnClickListener object.
This is not a problem since these objects are so small. You would need to be allocating thousands of click listeners per second before it started to matter. No plans to change this.
thx. #close
thx.
| gharchive/issue | 2016-10-17T07:15:00 | 2025-04-01T06:37:05.762751 | {
"authors": [
"JakeWharton",
"zouzhenglu"
],
"repo": "JakeWharton/RxBinding",
"url": "https://github.com/JakeWharton/RxBinding/issues/296",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
1827421069 | Type '{ participants: true; }' is not assignable to type 'never'.ts(2322) (property) include: never
Type '{ participants: true; }' is not assignable to type 'never'.ts(2322)
(property) include: never
Thank you for reporting this issue, and for providing the screenshot. To help me understand and resolve the problem, could you please provide more details?
What were you doing when this issue occurred?
Can you provide any error messages or logs?
What browser and operating system are you using?
Are there specific steps I can follow to reproduce the issue?
Any additional information you can provide will be very helpful.
This happened from a fresh installation, just after I setup the database then I tried to run it, it happened.
Commands I used on setting up prisma;
yarn
yarn prisma init
yarn prisma migrate
yarn prisma migrate dev
yarn prisma db pull
npm run build && npm start
The error is likely happening because the Prisma Client that's being used to run the code doesn't match the Prisma schema. This can happen if the Prisma Client hasn't been regenerated after making changes to the schema.
Since you did a fresh clone and ran the Prisma setup commands, it's possible that the Prisma Client was not regenerated after running the migrations.
Here's what you can do to try and fix the issue:
You need to regenerate the Prisma Client to ensure that it matches the schema. You can do this by running the following command:
npx prisma generate
After regenerating the Prisma Client, you should rebuild the project to make sure that the newly generated client is being used. You can do this by running:
yarn build
Restart your development server to make sure that the changes are picked up.:
yarn dev
Also, based on the commands you listed, it seems you are using both npm and yarn at the same time, which I would recommend you not doing.
| gharchive/issue | 2023-07-29T09:37:30 | 2025-04-01T06:37:05.785284 | {
"authors": [
"JaleelB",
"rodrigoricky"
],
"repo": "JaleelB/callsquare",
"url": "https://github.com/JaleelB/callsquare/issues/4",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
196539445 | Add support for Google Secure Scaffold
Its being every time more common the need of work on top of that scaffold.
https://github.com/google/gae-secure-scaffold-python
It would be great to have it as an option in the generator.
Is it doable?
Definitely is doable, but I think we should wait until we have a viable angular project for the generator as there isn't really a point to setting up the google scaffold for a react project.
| gharchive/issue | 2016-12-19T22:41:40 | 2025-04-01T06:37:05.794294 | {
"authors": [
"miguelmoraleda",
"njam3"
],
"repo": "Jam3/generator-jam3",
"url": "https://github.com/Jam3/generator-jam3/issues/255",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
438330424 | Add status command to bot
Might require some further formatting and code quality fixes but making the initial PR for now if anyone wants to have a look at it.
Fixes #2
Happy enough that the feature works and no compromising information is revealed from the status command, only things revealed are:
Instance name:
Instance State:
Public IP Address: (Should not be compromising assuming server network traffic rules are correctly configured, plus bot hoster can remove this line if they wish)
Last startup time:
May be worth moving the status lines messaged out to the config.json file but this will be under another issue if I decide it's worthwhile
| gharchive/pull-request | 2019-04-29T13:59:13 | 2025-04-01T06:37:05.802909 | {
"authors": [
"JamesMatchett"
],
"repo": "JamesMatchett/AWS-Discord-Bot",
"url": "https://github.com/JamesMatchett/AWS-Discord-Bot/pull/3",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1917829529 | Add more tests
For example to make sure that positional and keyword only identifiers are parsed correctly. *args and **kwargs.
Plucking some docstring_parser test holes here: https://github.com/JanEricNitschke/pymend/pull/69
| gharchive/issue | 2023-09-28T15:53:19 | 2025-04-01T06:37:05.824185 | {
"authors": [
"JanEricNitschke"
],
"repo": "JanEricNitschke/pymend",
"url": "https://github.com/JanEricNitschke/pymend/issues/8",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
803607847 | Statistics pages sorting
https://eoa.ee/?kool=true and https://eoa.ee/?hof=true are not sorted properly.
https://eoa.ee/?kool=true is not sorted at all.
https://eoa.ee/?hof=true seems to be sorted by "1. KOHTI" but secont sorting parameter is random.
What would be good criteria for sorting these tables? First places, then second places, then third places? Total of the count of the first three places?
I think overall participation might be the most relevant. For example, a student that has participated in 7 competitions but does not have any high places should be higher on the list than a person who has participated in one olympiad and achieved the third place there.
In that case, should all people who have participated in at least one contest be shown in the hall of fame? Currently, only students with at least one placement 1st-3rd are shown (source).
I prefer sorting based on participation. But I think that we have to decide how many participations are required to get there. Some statistics (participation-how many students):
25-1 24-1 23-2 22-1 21-4
20-3 19-1 18-3 17-6 16-3
15-5 14-11 13-10 12-13 11-23
10-16 9-24 8-26 7-55 6-95
5-110 4-238 3-362 2-599 1-1729
So I think we should include everyone who has got TOP3 or at least 8 participations (This is open for discussion).
Note that there are currently 450 people with at least one top 3 placement in the database (502 in total when also counting others with >= 8 participations). Should the hall of fame display be based on how many people would be shown?
| gharchive/issue | 2021-02-08T14:19:15 | 2025-04-01T06:37:05.859447 | {
"authors": [
"JarlPatrick",
"kaarelkivisalu",
"marko213"
],
"repo": "JarlPatrick/eoa",
"url": "https://github.com/JarlPatrick/eoa/issues/23",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1321224327 | Adding Shaders
Adding the Shaders.hx file
This is part 1 of it. I'll do the PlayState next.
| gharchive/pull-request | 2022-07-28T16:57:14 | 2025-04-01T06:37:05.878605 | {
"authors": [
"Gabriel2019r"
],
"repo": "JasmineIsSwagger/FridayNightFunkinZR-RE-Ality",
"url": "https://github.com/JasmineIsSwagger/FridayNightFunkinZR-RE-Ality/pull/1",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1482773665 | Refactoring for Binary Number Literals, Constants, etc.
If this exists in code:
private const int MaximumSize = 80000000;
var maximumSize = 80000000;
Offer a refactoring to reformat it like this:
private const int MaximumSize = 80_000_000;
var maximumSize = 80_000_000;
I personally find that much easier to read.
Note that VS has a "Separate thousands" refactoring:
However, I'd like this to be solution-wide, so I'll still do this as an interested reader looking for exercise :)
| gharchive/issue | 2022-12-07T20:30:09 | 2025-04-01T06:37:05.885992 | {
"authors": [
"JasonBock"
],
"repo": "JasonBock/Transpire",
"url": "https://github.com/JasonBock/Transpire/issues/22",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
494521403 | Throwing Exception in validation flow
Hello, I inspect the repository and I got an issue about the validation way in the application layer.
According to following link, there is a RequestValidationBehavior to validate input data by using FluentValidation that is similar to ValidatorBehavior in eShopOnContainers. One of the issue is throwing exception in the code below.
https://github.com/JasonGT/NorthwindTraders/blob/c942193b02762561968c3981d9172e6e6fc6d274/Northwind.Application/Infrastructure/RequestValidationBehavior.cs#L33
I like ValidatorBehavior that MediatR have provided for us, but if throwing exception is used in handling validation there could be a big problem by using it. The problem is controlling the flow of business by exceptions that is wrong. I have thought about modifying this line of code to avoid throwing exception, but I couldn't come up with an idea.
Hi @A-Soltani - thanks for your feedback. I think about this approach sometimes too. Perhaps a better approach would be to use a RequestResult. This could include properties such as Succeeded (was the request successful) and Errors[] (if not successful, what were the errors).
In this way, we have a consistent approach to return results, errors or otherwise.
What do you think - do you like this approach better?
Feedback from everyone else welcome.
As an example:
public class RequestResult<T>
{
public bool Succeeded { get; set; }
public IEnumerable<RequestError> Errors { get; set; }
public T Result { get; set; }
}
public class RequestError
{
public string Code { get; set; }
public string Description { get; set; }
}
public class GetCustomerDetailQuery : IRequest<RequestResult<CustomerDetailVm>>
{
public string Id { get; set; }
}
The problem with this approach, is that the RequestError is quite generic in nature. An error type might be required to differentiate between authorisation, validation, not found and so on. Different types will be handled differently by an API client, 401, 400, 404, etc.
Thoughts?
Just reviewing eShopOnContainers, and they take the same ValidationException approach:
if (failures.Any())
{
_logger.LogWarning("Validation errors - {CommandType} - Command: {@Command} - Errors: {@ValidationErrors}", typeName, request, failures);
throw new OrderingDomainException($"Command Validation Errors for type {typeof(TRequest).Name}",
new ValidationException("Validation exception", failures));
}
The goal should be to find the simplest approach. In this case, we could state that the initial responsibility is with the client to validate the request. If the client fails to do so, then this is an exceptional circumstance - hence the ValidationException is thrown. This is therefore the simplest approach.
Thoughts?
No feedback so I am going to close this issue.
If you're interested in an alternative approach take a look at the new implementation for the UserManagerService which returns a Result type; https://github.com/JasonGT/NorthwindTraders/blob/master/Src/Infrastructure/Identity/UserManagerService.cs
| gharchive/issue | 2019-09-17T09:54:14 | 2025-04-01T06:37:05.897334 | {
"authors": [
"A-Soltani",
"JasonGT"
],
"repo": "JasonGT/NorthwindTraders",
"url": "https://github.com/JasonGT/NorthwindTraders/issues/166",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
137968902 | [BUG] Sensitivity
When setting the sensitivity of L2, R2, LS, RS to 0.8 there is always this error on editing the profile and editing is not possible:
System.ArgumentOutOfRangeException: Der Wert 0 ist für Value ungültig. Value sollte zwischen 'Minimum' und 'Maximum' liegen.
Parametername: Value
bei System.Windows.Forms.NumericUpDown.set_Value(Decimal value)
bei DS4Windows.Options.Reload(Int32 deviceNum, String name)
bei DS4Windows.DS4Form.ShowOptions(Int32 devID, String profile)
bei DS4Windows.DS4Form.tsBNEditProfile_Click(Object sender, EventArgs e)
bei System.Windows.Forms.ToolStripItem.RaiseEvent(Object key, EventArgs e)
bei System.Windows.Forms.ToolStripMenuItem.OnClick(EventArgs e)
bei System.Windows.Forms.ToolStripItem.HandleClick(EventArgs e)
bei System.Windows.Forms.ToolStripItem.HandleMouseUp(MouseEventArgs e)
bei System.Windows.Forms.ToolStrip.OnMouseUp(MouseEventArgs mea)
bei System.Windows.Forms.ToolStripDropDown.OnMouseUp(MouseEventArgs mea)
bei System.Windows.Forms.Control.WmMouseUp(Message& m, MouseButtons button, Int32 clicks)
bei System.Windows.Forms.Control.WndProc(Message& m)
bei System.Windows.Forms.ToolStrip.WndProc(Message& m)
bei System.Windows.Forms.ToolStripDropDown.WndProc(Message& m)
bei System.Windows.Forms.NativeWindow.Callback(IntPtr hWnd, Int32 msg, IntPtr wparam, IntPtr lparam)
#102 fixes this.
| gharchive/issue | 2016-03-02T19:31:26 | 2025-04-01T06:37:06.071901 | {
"authors": [
"kiliansch",
"oehlrich9"
],
"repo": "Jays2Kings/DS4Windows",
"url": "https://github.com/Jays2Kings/DS4Windows/issues/100",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
728274906 | A lot of "Something went very wrong errors"
Hello Jean :D. Good to see this repository is doing well. I'm here because I decided to mess around with my MacOS Rice a little. I decided to reinstall your version of simple-bar rather than a friend's take on it. I see that on the usage of the command
git clone https://github.com/Jean-Tinland/simple-bar $HOME/Library/Application\ Support/Übersicht/widgets/simple-bar
This freaks simple-bar out and I don't know how to solve it
Thanks in advance for any help you can provide!
Also, I can't interact with the bar like I did the first time. It should be also be worth noting that yabai isn't installed. Is it a requirement?
Actually, installing Yabai fixed the issue. I'm just a idiot xd.
Hi, I was indeed thinking about that since the 2 widgets which weren't working were depending on yabai to get their data.
No problem, I think I need to cleanup the readme because the "Compatibility & requirements" is a bit lost in the middle of everything.
Glad it is working now!
This happened to me, but I had yabai installed. I managed to fix it by entering the right yabai path in the cmd + , settings, and upgrading/reinstalling yabai. Works beautifully now. Just wanted to say it for anyone that is having the same problem
This happened to me, but I had yabai installed. I managed to fix it by entering the right yabai path in the cmd + , settings, and upgrading/reinstalling yabai. Works beautifully now. Just wanted to say it for anyone that is having the same problem
| gharchive/issue | 2020-10-23T14:40:55 | 2025-04-01T06:37:06.092887 | {
"authors": [
"Jean-Tinland",
"Obl1que",
"WarpWing"
],
"repo": "Jean-Tinland/simple-bar",
"url": "https://github.com/Jean-Tinland/simple-bar/issues/45",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2326648784 | Is it possible to only retrieve information about a specific flight by searching for the flight number?
Por exemplo, apenas retornar as informações do voo LA3467 da LATAM
Unfortunately, no. You can only retrive by its airline or its registration.
| gharchive/issue | 2024-05-30T22:49:12 | 2025-04-01T06:37:06.095171 | {
"authors": [
"JeanExtreme002",
"yunathan51"
],
"repo": "JeanExtreme002/FlightRadarAPI",
"url": "https://github.com/JeanExtreme002/FlightRadarAPI/issues/75",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
739847571 | Syntax errors in bloblang expressions not checked by linter
The linter assumes that every bloblang expression is valid, this can cause invalid bloblang expressions anywhere in your configuration to prevent benthos from starting, despite the lint passing.
Expected behaviour: Linter flags all syntax errors in all bloblang expressions (including interpolated expressions).
This is done and will be in the next release: https://github.com/Jeffail/benthos/commit/cd97a28ab8d87c95fd07029309d92136d28bf5a0
| gharchive/issue | 2020-11-10T11:57:31 | 2025-04-01T06:37:06.111814 | {
"authors": [
"Jeffail",
"nicktelford"
],
"repo": "Jeffail/benthos",
"url": "https://github.com/Jeffail/benthos/issues/547",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
335698428 | Want to differentiate between retweets and "normal" tweets
Is there a way to differentiate between retweets, and normal tweets? I am scraping tweets from a set of usernames, eg. tweetCriteria = got.manager.TweetCriteria().setUsername('barackobama').
Using this criteria I get both normal tweets and retweets, however they are not differentiated.
Any advice on making adjustments to the script is welcome.
This would be very useful, and I too would like this feature.
There are several ways, try following in Python if you want to modify code and do it
Search for div with class "QuoteTweet u-block js-tweet-details-fixer". If this div exist that means it's a retweet.
Check if div with class "js-tweet-text-container" has or plain text. If it has Plain text, its not Retweet, if it has then it is
You can get the details of Retweeted data by looking into <'div class="QuoteTweet u-block js-tweet-details-fixer"><'div class="QuoteTweet-container">...
you can also search for text " Retweeted " in entire json response. If it exists, it is a Retweet.
| gharchive/issue | 2018-06-26T07:49:08 | 2025-04-01T06:37:06.115179 | {
"authors": [
"bellwolf",
"rahulha",
"rax87"
],
"repo": "Jefferson-Henrique/GetOldTweets-python",
"url": "https://github.com/Jefferson-Henrique/GetOldTweets-python/issues/197",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
201158599 | Super Large Sourcemap
Related to #108, but thought this would warrant it's own possible discussion thread.
As you can see in the output below the app.js.map file is 12.4mb. What would cause a sourcemap file to grow so large? Anything I'm possibly doing wrong?
> cross-env NODE_ENV=production webpack --progress --hide-modules
DONE Compiled successfully in 34886ms
Asset Size Chunks Chunk Names
/app.js 2.53 MB 1 [emitted] [big] app
fonts/icons.eot?25a32416abee198dd821b0b17a198a8f 76.5 kB [emitted]
fonts/icons.ttf?1dc35d25e61d819a9c357074014867ab 153 kB [emitted]
fonts/icons.woff?c8ddf1e5e5bf3682bc7bebf30f394148 90.4 kB [emitted]
fonts/icons.woff2?e6cf7c6ec7c2d6f670ae9d762604cb0b 71.9 kB [emitted]
/0.js 65.3 kB 0 [emitted]
fonts/icons.svg?d7c639084f684d66a1bc66855d193ed8 392 kB [emitted] [big]
/styles.css 496 kB 1 [emitted] [big] app
/0.js.map 470 kB 0 [emitted]
/app.js.map 12.4 MB 1 [emitted] app
/styles.css.map 88 bytes 1 [emitted] app
manifest.json 539 bytes [emitted]
Here are my config files
Any ideas?
Thanks again!
Your app.js is very large too. As far as I know, sourcemaps are often 5 to 10 times larger that the source file itself. But the browser doesn't download the sourcemaps, if they are not needed.
But I think 2.53 MB after minification is your bigger problem.
I've got it down to 1.1mb now after a few other optimizations, but still getting 10mb+ sourcemaps. You're saying that's to be expected?
I think you're talking about these?
new webpack.optimize.UglifyJsPlugin({
sourceMap: true,
compress: {
warnings: false,
screw_ie8: true,
conditionals: true,
unused: true,
comparisons: true,
sequences: true,
dead_code: true,
evaluate: true,
if_return: true,
join_vars: true,
},
output: {
comments: false
},
})
I was curious and tried that too. My output was exactly the same size. Only the the output.comments = false saved 1 KB on one file (from 280KB to 279KB).
I have no idea why that saved so much in your situation.
But I also don't see the mix-manifest.json in your output. Do you use the current laravel-mix version?
Here my output to compare:
> cross-env NODE_ENV=production webpack --progress --hide-modules
DONE Compiled successfully in 40789ms
Asset Size Chunks Chunk Names
/js/0.e97fe5c0195b81c24655.js.map 5.48 MB 0, 6 [emitted]
/js/0.e97fe5c0195b81c24655.js 600 kB 0, 6 [emitted] [big]
/js/2.e1d5ef9a8751c57c2226.js 5.17 kB 2, 6 [emitted]
/js/vendor.75c7220f46ad969b6a63.js 280 kB 3, 6 [emitted] [big] vendor
/js/frontend.c5d9b369644b972943d0.js 50.7 kB 4, 6 [emitted] frontend
/js/backend.2aef4bf88b748fd757aa.js 6.05 kB 5, 6 [emitted] backend
/js/manifest.d41d8cd98f00b204e980.js 1.59 kB 6 [emitted] manifest
/css/app.1842a9963041feec44c9.css 159 kB 4, 6 [emitted] frontend
/js/1.f1f62619ad496dab1015.js 13.9 kB 1, 6 [emitted]
/js/1.f1f62619ad496dab1015.js.map 73.8 kB 1, 6 [emitted]
/js/2.e1d5ef9a8751c57c2226.js.map 37.5 kB 2, 6 [emitted]
/js/vendor.75c7220f46ad969b6a63.js.map 2.05 MB 3, 6 [emitted] vendor
/js/frontend.c5d9b369644b972943d0.js.map 337 kB 4, 6 [emitted] frontend
/css/app.1842a9963041feec44c9.css.map 110 bytes 4, 6 [emitted] frontend
/js/backend.2aef4bf88b748fd757aa.js.map 41.7 kB 5, 6 [emitted] backend
/js/manifest.d41d8cd98f00b204e980.js.map 14.4 kB 6 [emitted] manifest
mix-manifest.json 628 bytes [emitted]
As you can see my souremaps are also ~10 times as large as my source files are. I think thats normal. The same is true if I look out the output from vue.js project which was generated with the vue-cli.
Did you look at your app.js without your UglifyJsPlugin modifications? Was is minified etc?
Does your app.js contain the sourcemap?
Some configurations inline the sourcemaps to your .js file - which creates a massive .js file
Yeah the sourcemaps will be large for an app that big. It shouldn't affect your production deploy.
| gharchive/issue | 2017-01-17T02:11:57 | 2025-04-01T06:37:06.129814 | {
"authors": [
"JeffreyWay",
"OwenMelbz",
"jkudish",
"strebl"
],
"repo": "JeffreyWay/laravel-mix",
"url": "https://github.com/JeffreyWay/laravel-mix/issues/109",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
349234590 | Just a question. What is the reasoning behind this vue loader config?
Why does laravel-mix set esModule to false?
vue: {
preLoaders: {},
postLoaders: {},
esModule: false
},
I'm asking because setups provided by vue-cli have this set to true. I tend to use require().default in specific situations, and using it becomes inconsistent for laravel-mix because this config makes me drop .default.
esModule: false option is no longer available in vue-loader v14+, laravel mix is using v13.
It has been set to false to keep mix behaviour non breaking
Of course
| gharchive/issue | 2018-08-09T18:12:12 | 2025-04-01T06:37:06.133006 | {
"authors": [
"ankurk91",
"raniesantos"
],
"repo": "JeffreyWay/laravel-mix",
"url": "https://github.com/JeffreyWay/laravel-mix/issues/1735",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
204256969 | Doesn't compile both less and sass files
I have following configuration:
mix.js('resources/assets/js/app.js', 'public/js').extract(['vue'])
.less('resources/assets/less/theme.less', 'public/css')
.sass('resources/assets/sass/app.scss', 'public/css');
However less files are not compiled, it works fine if I comment out .sass line.
Support for this was added just recently. Update your laravel-mix dependency to 0.6.0
| gharchive/issue | 2017-01-31T10:33:41 | 2025-04-01T06:37:06.134846 | {
"authors": [
"adriaanzon",
"vedmant"
],
"repo": "JeffreyWay/laravel-mix",
"url": "https://github.com/JeffreyWay/laravel-mix/issues/247",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
421651727 | When Alita walks near enemies her rotation goes crazy
Description:
When Alita walks near enemies her rotation goes crazy
Build:
v0.2.9.0
Type:
Alita
Steps to reproduce:
Walk near an enemy
Frequency:
Sometimes
This was the enemies pushing the enemy, but it should be alright now.
| gharchive/issue | 2019-03-15T18:29:58 | 2025-04-01T06:37:06.146651 | {
"authors": [
"OscarHernandezG",
"ValdiviaDev"
],
"repo": "JellyBitStudios/JellyBitEngine",
"url": "https://github.com/JellyBitStudios/JellyBitEngine/issues/50",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
72705815 | Will there be a DNX Core 5 version of this library created?
It would be great if this was useable w/ DNX core projects.
Yes, there will be.
Any rough time frame on when you are planning on this?
Thanks!
The initial work is done, but it was done against the VS2015 beta before the aspnetcore -> dnxcore rename, so that all still needs doing.
All the automatic integration with ASP.NET MVC needs writing too, but this needs the new validation API in MVC6 to be finished first.
I'd say this is still several months away from being finished, and I'm away for all of July, August and half of September, so it's unlikely to be until later in the year - sorry.
Cleanup: Closing this issue since it hasn't had activity in about a year.
| gharchive/issue | 2015-05-02T22:10:30 | 2025-04-01T06:37:06.161969 | {
"authors": [
"JeremySkinner",
"SeanKilleen",
"csaloio",
"jrharmon"
],
"repo": "JeremySkinner/FluentValidation",
"url": "https://github.com/JeremySkinner/FluentValidation/issues/60",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
444298478 | Spelling corrections and readability changes
Thanks for a great lib!
Came across a couple of small changes needed to the ASP.NET Core docs as I was reading through - hence this PR
Merged, thanks!
| gharchive/pull-request | 2019-05-15T08:16:50 | 2025-04-01T06:37:06.163391 | {
"authors": [
"JeremySkinner",
"ry8806"
],
"repo": "JeremySkinner/FluentValidation",
"url": "https://github.com/JeremySkinner/FluentValidation/pull/1122",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
118312930 | error in Android
not finding a certain file on landing page of Android Build
patched in latest PR and merged with master. was just a typo in a reference line.
| gharchive/issue | 2015-11-23T03:20:16 | 2025-04-01T06:37:06.180477 | {
"authors": [
"EMCP"
],
"repo": "JesusGuerrero/MAHRIO",
"url": "https://github.com/JesusGuerrero/MAHRIO/issues/41",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
954073296 | Bugfix/161 correct temporary dataset path
Now the data.zip is downloaded to the correct path.
This fixes the issue: Incorrect temporary folder for the cats-vs-dogs dataset LINK
fixes #161
Hi @zaleslaw! Did you have a chance to take a look at it?
Fixed in #235
| gharchive/pull-request | 2021-07-27T16:49:54 | 2025-04-01T06:37:06.185639 | {
"authors": [
"funkstermonster",
"zaleslaw"
],
"repo": "JetBrains/KotlinDL",
"url": "https://github.com/JetBrains/KotlinDL/pull/163",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2382128022 | Exception in thread "main" java.lang.NoSuchMethodError: 'void androidx.compose.ui.graphics.SkiaBackedCanvas_skikoKt.setAlphaMultiplier(androidx.compose.ui.graphics.Canvas, float)'
Describe the bug
When I try to build and run Compose Desktop App in my Ubuntu 23.10, I have an error
But when I run in Android Target, it runs normally.
Affected platforms
Desktop (Windows, Linux, macOS)
Versions
Libraries:
Compose Multiplatform version: 1.6.10
Navigation Multiplatform version: 2.7.0-alpha07
...
Kotlin version: 2.0.0
OS version(s) : Ubuntu 23.10
OS architecture (x86 or arm64): x86_64
JDK (for desktop issues): JetBrains Runtime 17.0.10
To Reproduce
Steps to reproduce the behavior:
Run this code snippet: @Composable
fun MainScreen() {
val navController = rememberNavController()
Scaffold(
bottomBar = {
BottomNavigationBar(navController)
}
) {
BottomNavGraph(
navController = navController
)
}
}
//In App.kt
@Composable
@Preview
fun App() {
MaterialTheme {
MainScreen()
}
}
Run ./gradlew run
Expected behavior
Should not crash and show navigation bar like Android Target
Screenshots
Additional context
Crash log:
Task :composeApp:run
(java:33049): Gtk-WARNING **: 14:15:46.931: Theme directory 16x16/panel of theme Mkos-Big-Sur-Night has no size field
(java:33049): Gtk-WARNING **: 14:15:46.936: Theme directory 16x16@2x/places of theme Mkos-Big-Sur-Night has no size field
(java:33049): Gtk-WARNING **: 14:15:46.942: Theme directory 24x24/apps of theme Mkos-Big-Sur-Night has no size field
(java:33049): Gtk-WARNING **: 14:15:46.945: Theme directory 24x24@2x/panel of theme Mkos-Big-Sur-Night has no size field
(java:33049): Gtk-WARNING **: 14:15:46.946: Theme directory 24x24@2x/panel of theme Mkos-Big-Sur-Night has no size field
Exception in thread "main" java.lang.NoSuchMethodError: 'void androidx.compose.ui.graphics.SkiaBackedCanvas_skikoKt.setAlphaMultiplier(androidx.compose.ui.graphics.Canvas, float)'
at androidx.compose.ui.platform.RenderNodeLayer.performDrawLayer(RenderNodeLayer.skiko.kt:295)
at androidx.compose.ui.platform.RenderNodeLayer.drawLayer(RenderNodeLayer.skiko.kt:244)
at androidx.compose.ui.node.NodeCoordinator.draw(NodeCoordinator.kt:348)
at androidx.compose.ui.node.LayoutNode.draw$ui(LayoutNode.kt:926)
at androidx.compose.ui.node.InnerNodeCoordinator.performDraw(InnerNodeCoordinator.kt:174)
at androidx.compose.ui.node.NodeCoordinator.drawContainedDrawModifiers(NodeCoordinator.kt:361)
at androidx.compose.ui.node.NodeCoordinator.draw(NodeCoordinator.kt:353)
at androidx.compose.ui.node.LayoutNode.draw$ui(LayoutNode.kt:926)
at androidx.compose.ui.node.RootNodeOwner.draw(RootNodeOwner.skiko.kt:197)
at androidx.compose.ui.scene.MultiLayerComposeSceneImpl.draw(MultiLayerComposeScene.skiko.kt:257)
at androidx.compose.ui.scene.BaseComposeScene.render(BaseComposeScene.skiko.kt:171)
at androidx.compose.ui.scene.ComposeSceneMediator.onRender(ComposeSceneMediator.desktop.kt:537)
at org.jetbrains.skiko.SkiaLayer.update$skiko(SkiaLayer.awt.kt:485)
at org.jetbrains.skiko.redrawer.AWTRedrawer.update(AWTRedrawer.kt:54)
at org.jetbrains.skiko.redrawer.LinuxOpenGLRedrawer.redrawImmediately(LinuxOpenGLRedrawer.kt:83)
at org.jetbrains.skiko.SkiaLayer.paint(SkiaLayer.awt.kt:325)
at androidx.compose.ui.scene.skia.WindowSkiaLayerComponent$contentComponent$1.paint(WindowSkiaLayerComponent.desktop.kt:54)
at java.desktop/javax.swing.JComponent.paintChildren(JComponent.java:955)
at java.desktop/javax.swing.JComponent.paint(JComponent.java:1124)
at java.desktop/javax.swing.JLayeredPane.paint(JLayeredPane.java:586)
at java.desktop/javax.swing.JComponent.paintChildren(JComponent.java:955)
at java.desktop/javax.swing.JComponent.paint(JComponent.java:1124)
at androidx.compose.ui.window.Window_desktopKt$Window$12$1.invoke(Window.desktop.kt:434)
at androidx.compose.ui.window.Window_desktopKt$Window$12$1.invoke(Window.desktop.kt:419)
at androidx.compose.ui.window.AwtWindow_desktopKt$AwtWindow$3.invoke(AwtWindow.desktop.kt:78)
at androidx.compose.ui.window.AwtWindow_desktopKt$AwtWindow$3.invoke(AwtWindow.desktop.kt:76)
at androidx.compose.ui.util.UpdateEffect_desktopKt$UpdateEffect$2$performUpdate$1.invoke(UpdateEffect.desktop.kt:59)
at androidx.compose.ui.util.UpdateEffect_desktopKt$UpdateEffect$2$performUpdate$1.invoke(UpdateEffect.desktop.kt:55)
at androidx.compose.runtime.snapshots.Snapshot$Companion.observe(Snapshot.kt:2304)
at androidx.compose.runtime.snapshots.SnapshotStateObserver$ObservedScopeMap.observe(SnapshotStateObserver.kt:504)
at androidx.compose.runtime.snapshots.SnapshotStateObserver.observeReads(SnapshotStateObserver.kt:260)
at androidx.compose.ui.util.UpdateEffect_desktopKt$UpdateEffect$2.invoke$performUpdate(UpdateEffect.desktop.kt:55)
at androidx.compose.ui.util.UpdateEffect_desktopKt$UpdateEffect$2.invoke(UpdateEffect.desktop.kt:64)
at androidx.compose.ui.util.UpdateEffect_desktopKt$UpdateEffect$2.invoke(UpdateEffect.desktop.kt:47)
at androidx.compose.runtime.DisposableEffectImpl.onRemembered(Effects.kt:82)
at androidx.compose.runtime.CompositionImpl$RememberEventDispatcher.dispatchRememberObservers(Composition.kt:1295)
at androidx.compose.runtime.CompositionImpl.applyChangesInLocked(Composition.kt:984)
at androidx.compose.runtime.CompositionImpl.applyChanges(Composition.kt:1005)
at androidx.compose.runtime.Recomposer.composeInitial$runtime(Recomposer.kt:1099)
at androidx.compose.runtime.CompositionImpl.composeInitial(Composition.kt:633)
at androidx.compose.runtime.CompositionImpl.setContent(Composition.kt:619)
at androidx.compose.ui.window.Application_desktopKt$awaitApplication$2$1$2.invokeSuspend(Application.desktop.kt:221)
at kotlin.coroutines.jvm.internal.BaseContinuationImpl.resumeWith(ContinuationImpl.kt:33)
at kotlinx.coroutines.DispatchedTask.run(DispatchedTask.kt:102)
at java.desktop/java.awt.event.InvocationEvent.dispatch(InvocationEvent.java:318)
at java.desktop/java.awt.EventQueue.dispatchEventImpl(EventQueue.java:792)
at java.desktop/java.awt.EventQueue$3.run(EventQueue.java:739)
at java.desktop/java.awt.EventQueue$3.run(EventQueue.java:733)
at java.base/java.security.AccessController.doPrivileged(AccessController.java:399)
at java.base/java.security.ProtectionDomain$JavaSecurityAccessImpl.doIntersectionPrivilege(ProtectionDomain.java:86)
at java.desktop/java.awt.EventQueue.dispatchEvent(EventQueue.java:761)
at java.desktop/java.awt.EventDispatchThread.pumpOneEventForFilters(EventDispatchThread.java:207)
at java.desktop/java.awt.EventDispatchThread.pumpEventsForFilter(EventDispatchThread.java:128)
at java.desktop/java.awt.EventDispatchThread.pumpEventsForHierarchy(EventDispatchThread.java:117)
at java.desktop/java.awt.EventDispatchThread.pumpEvents(EventDispatchThread.java:113)
at java.desktop/java.awt.EventDispatchThread.pumpEvents(EventDispatchThread.java:105)
at java.desktop/java.awt.EventDispatchThread.run(EventDispatchThread.java:92)
It works on a simple project with:
Compose Multiplatform version: 1.6.10
Navigation Multiplatform version: 2.7.0-alpha07
Kotlin version: 2.0.0
Code:
import androidx.compose.foundation.layout.Row
import androidx.compose.material3.Button
import androidx.compose.material3.MaterialTheme
import androidx.compose.material3.Scaffold
import androidx.compose.runtime.Composable
import androidx.compose.runtime.DisposableEffect
import androidx.navigation.compose.NavHost
import androidx.navigation.compose.composable
import androidx.navigation.compose.rememberNavController
@Composable
fun App() {
MaterialTheme {
val navController = rememberNavController()
Scaffold(
bottomBar = {
Row {
Button({
navController.navigate("Home")
}) {}
Button({
navController.navigate("Settings")
}) {}
}
}
) {
NavHost(
navController = navController,
startDestination = BottomBarScreen.Home.route
) {
composable(route = BottomBarScreen.Home.route) {
DisposableEffect(Unit) {
println("Home")
onDispose { }
}
}
composable(route = BottomBarScreen.Settings.route) {
DisposableEffect(Unit) {
println("Settings")
onDispose { }
}
}
}
}
}
}
sealed class BottomBarScreen(
val route: String,
) {
object Home: BottomBarScreen(
route = "Home",
)
object Settings: BottomBarScreen(
route = "Settings",
)
}
NoSuchMethodError usually indicates there is a binary incompatibility, could you provide your project or just build.gradle with the dependencies and gradle/lib.versions.toml?
Please check the following ticket on YouTrack for follow-ups to this issue. GitHub issues will be closed in the coming weeks.
| gharchive/issue | 2024-06-30T07:30:32 | 2025-04-01T06:37:06.196085 | {
"authors": [
"igordmn",
"maxrave-dev",
"okushnikov"
],
"repo": "JetBrains/compose-multiplatform",
"url": "https://github.com/JetBrains/compose-multiplatform/issues/5050",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
946367262 | Deprecate Color.RGB, Color.HSL etc. functions in favor of top-level rgb, hsl an so on
We've decided that this:
borderColor(rgb(10, 20, 30))
is better than this
borderColor(Color.rgb(200, 20, 10))
Please check the following ticket on YouTrack for follow-ups to this issue. GitHub issues will be closed in the coming weeks.
| gharchive/issue | 2021-07-16T14:55:03 | 2025-04-01T06:37:06.198887 | {
"authors": [
"Schahen",
"okushnikov"
],
"repo": "JetBrains/compose-multiplatform",
"url": "https://github.com/JetBrains/compose-multiplatform/issues/902",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
770863596 | size_unit for geom_text should take into account user specified label_format
Otherwise labels can overlap.
IMO, better would be to have a parameter specifying number of symbols that constitute label taken as unit.
| gharchive/issue | 2020-12-18T12:51:50 | 2025-04-01T06:37:06.216452 | {
"authors": [
"IKrukov-HORIS",
"alshan"
],
"repo": "JetBrains/lets-plot",
"url": "https://github.com/JetBrains/lets-plot/issues/270",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
86851982 | Compatibility fix for Chocolatey v0.9.9.6
The chocolatey executable files were moved a couple of versions back breaking this meta-runner.
This change updates the path to the new location and appends the '-y' option to suppress the user confirmation to install the package.
:+1:
| gharchive/pull-request | 2015-06-10T05:57:43 | 2025-04-01T06:37:06.217807 | {
"authors": [
"gep13",
"striglone"
],
"repo": "JetBrains/meta-runner-power-pack",
"url": "https://github.com/JetBrains/meta-runner-power-pack/pull/51",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
1012589533 | Add high school
High school where one of the faculties is computer science.
School website: https://brzozowa5.edu.pl/
@dmprusak Pull request merged. Thank you!
| gharchive/pull-request | 2021-09-30T20:24:11 | 2025-04-01T06:37:06.220743 | {
"authors": [
"dmprusak",
"philipto"
],
"repo": "JetBrains/swot",
"url": "https://github.com/JetBrains/swot/pull/12611",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2156650615 | Create ITTS.txt
please add my collage
@masghi The first verification stage is complete: the official domain for **Institut Teknologi Tangerang Selatan ** is itts.ac.id, to the best of my knowledge. This is just an informational note. The pull request is still under review. The review may take some more time. I greatly appreciate your patience. Here is the proof of domain ownership: https://opencourse.itts.ac.id/
@masghi Pull request merged. Thank you.
| gharchive/pull-request | 2024-02-27T13:52:32 | 2025-04-01T06:37:06.222843 | {
"authors": [
"masghi",
"philipto"
],
"repo": "JetBrains/swot",
"url": "https://github.com/JetBrains/swot/pull/20215",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2731572531 | Add files via upload
Blount County
@bobderf Please provide us with:
the school official website URL
the school street address, including city and country
a proof, which shows that the school recognizes the domain you are submitting as an official email domain for the students.
@bobderf You might miss my previous comment. Please react.
SorryOn Sat, Dec 14, 2024 at 6:02 AM Philip Torchinsky @.***> wrote:
@bobderf You might miss my previous comment. Please react.
—Reply to this email directly, view it on GitHub, or unsubscribe.You are receiving this because you were mentioned.Message ID: @.***>
On Thu, Dec 12, 2024 at 5:37 AM Philip Torchinsky @.***> wrote:
@bobderf Please provide us with:
the school official website URL = https://heritagehigh.blountk12.org/the school street address, including city and country = 3741 E Lamar Alexander Pkwy, Maryville, TN 37804a proof, which shows that the school recognizes the domain you are submitting as an official email domain for the students. I’m not sure what you mean I am a student ?
—Reply to this email directly, view it on GitHub, or unsubscribe.You are receiving this because you were mentioned.Message ID: @.***>
@bobderf Thank you for the clarifications. Pull request merged. Please start requesting the licenses in about an hour, to let the changes to propagate through our system.
| gharchive/pull-request | 2024-12-11T01:07:55 | 2025-04-01T06:37:06.228061 | {
"authors": [
"bobderf",
"philipto"
],
"repo": "JetBrains/swot",
"url": "https://github.com/JetBrains/swot/pull/25775",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
546287756 | Adds Harrow International School Beijing
北京哈罗国际学校
Harrow International School Beijing
www.harrowbeijing.cn
@SacretFlyer Pull request merged. Thank you!
| gharchive/pull-request | 2020-01-07T13:37:39 | 2025-04-01T06:37:06.229386 | {
"authors": [
"SacretFlyer",
"philipto"
],
"repo": "JetBrains/swot",
"url": "https://github.com/JetBrains/swot/pull/7427",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
604294472 | Uploaded: Ron Dearing UTC - University Technical College
Website: https://www.rondearingutc.com/
@haydhook Pull request merged. Thank you!
| gharchive/pull-request | 2020-04-21T21:21:26 | 2025-04-01T06:37:06.230644 | {
"authors": [
"haydhook",
"philipto"
],
"repo": "JetBrains/swot",
"url": "https://github.com/JetBrains/swot/pull/8177",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
725781687 | Adding hwbcymru domain
https://www.cardiffhigh.cardiff.sch.uk/
https://i.imgur.com/XQofrpK.png
@uhCHS Please provide us with a proof, which shows that the school recognizes the domain you are submitting as an official email domain for the students. I am sorry, the screenshot of a mailbox is not enough. The screenshot shows that you probably has an access to the mailbox there, but it says nothing about recognition of the domain by the school.
@philipto Thanks for getting back to me, You can see this link which is from the government that the domain hwbcymru has been setup for 85% of Wales schools. This has been something that was enforced with the government and not with the the school its self. Thanks
forgot to provide the link https://gov.wales/wales-leads-way-microsoft-schools
@uhCHS Pull request merged. Thank you!
FTR: additional proof of the domain validity is at https://hwb.gov.wales/news/articles/dc715915-79e3-4582-8789-5ea9f664c908
@philipto Does it take time to activate?
https://i.imgur.com/XOsnS0q.png
working now thanks :+1:
| gharchive/pull-request | 2020-10-20T17:02:27 | 2025-04-01T06:37:06.234834 | {
"authors": [
"philipto",
"uhCHS"
],
"repo": "JetBrains/swot",
"url": "https://github.com/JetBrains/swot/pull/9653",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
319842827 | Does not compile with Boost Test 1.67
Compiler MSVS 2015 VC++ 14.0
Boost Test 1.67
teamcity_boost.cpp(78): error C2259: 'JetBrains::TeamcityBoostLogFormatter': cannot instantiate abstract class
teamcity_boost.cpp(78): note: due to following members:
teamcity_boost.cpp(78): note: 'void boost::unit_test::unit_test_log_formatter::log_entry_context(std::ostream &,boost::unit_test::log_level,boost::unit_test::const_string)': is abstract
boost/test/unit_test_log_formatter.hpp(267): note: see declaration of 'boost::unit_test::unit_test_log_formatter::log_entry_context'
teamcity_boost.cpp(78): note: 'void boost::unit_test::unit_test_log_formatter::entry_context_finish(std::ostream &,boost::unit_test::log_level)': is abstract
unit_test_log_formatter.hpp(274): note: see declaration of 'boost::unit_test::unit_test_log_formatter::entry_context_finish'
@dmitry-treskunov Fixed by https://github.com/JetBrains/teamcity-cpp/commit/6620272df69790f0e7fcce35d2f4c7bdef57e58e. You can close it.
| gharchive/issue | 2018-05-03T09:15:10 | 2025-04-01T06:37:06.238431 | {
"authors": [
"k15tfu",
"yuchdev"
],
"repo": "JetBrains/teamcity-cpp",
"url": "https://github.com/JetBrains/teamcity-cpp/issues/25",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
589015268 | Facing css import issue in create react app in typescript
I am facing css import issue in create react app in typescript
Can u please let me know how I can resolve this issue as I have imported the css into my typescript file as below
import 'suneditor/dist/css/suneditor.min.css';
Which version are you using?
suneditor@2.25.0
I have added the path of the suneditor.min.css in CopyWebpackPlugin in Webpack
as node_modules/suneditor/dist/css/suneditor.min.css
And in
new HtmlWebpackIncludeAssetsPlugin({
assets: [
'css/suneditor.min.css'] })
And use in my typescript file as
import 'suneditor/dist/css/suneditor.min.css';
It works in my dev env but fails in Prod after build as shown in figure above
It seems that the web font is not loaded.
You must have a file loader.
{
test: /\.(eot|svg|ttf|woff|woff2)(\?v=[0-9]\.[0-9]\.[0-9])?$/,
use: [{
loader: "file-loader",
options: {
publicPath: '../',
name: 'fonts/[hash].[ext]',
limit: 5000,
mimetype: 'application/font-woff'
}
}]
}
Or
The latest version uses "inline svg" rather than a web font.
Please update the editor to the latest version.
ok how do I use font import in CopyWebpackPlugin
Sorry, I don't know..😭
I recommend updating the version.
I set it like this in my webpack environment.
{
test: /\.(eot|svg|ttf|woff|woff2)(\?v=[0-9]\.[0-9]\.[0-9])?$/,
use: [{
loader: "file-loader",
options: {
publicPath: '../',
name: 'fonts/[hash].[ext]',
limit: 5000,
mimetype: 'application/font-woff'
}
}]
}
Even if the web font is loaded with this setting, it is recommended to use the latest version.
Can u let me know what is the improvement in switching to newer version
Please refer to the release history.
https://github.com/JiHong88/SunEditor/releases
I was experiencing similar issues, as I have
import 'suneditor/dist/css/suneditor.min.css';
...in my code as well.
Here's what I've learned: in development mode, webpack is just copying the contents of that file and putting it into a <style> element, without interacting with the source. In production mode, webpack will open the source, copy the contents into memory and attempt to minify it with Terser, however there must be a syntax error in 'suneditor/dist/css/suneditor.min.css', because it fails to parse and therefore fails to include it in the final production bundle. There is also evidence that 'suneditor/dist/css/suneditor.min.css' has a syntax error, because my text editor cannot apply css syntax highlighting to it. I will report back if I can find more information.
According to csslint.com: Expected RBRACE at line 1, col 30372.
I think both my CSS syntax highlighter, Terser and CSS Lint don't understand the 1turn value here:
@keyframes spinner{
to {
transform: rotate(1turn);
}
}
When I change it to:
@keyframes spinner{
to {
transform: rotate(360deg);
}
}
CSS Lint is happy.
@rwaldron Thank for your feedback!
I have checked this issue, and it seems that "cssnano" converts "360deg" to "1turn".
https://github.com/cssnano/cssnano/issues/823
This issue does not seem to be resolved. :(
I will fix this issue by changing "360deg" to "361deg".
The 2.30.0 version has been updated.
If this issue has not been resolved, please reopen this issue.
Thank you.
| gharchive/issue | 2020-03-27T09:55:07 | 2025-04-01T06:37:06.251302 | {
"authors": [
"FineStrokes",
"JiHong88",
"rwaldron"
],
"repo": "JiHong88/SunEditor",
"url": "https://github.com/JiHong88/SunEditor/issues/300",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1292810935 | 关于运行app,无法运行
No such module 'CoreBitcoin.libscrypt'
报错
下载后直接运行就好了,不需要pod
运行BlockChainWallet这个文件夹里的项目文件
| gharchive/issue | 2022-07-04T08:24:14 | 2025-04-01T06:37:06.253014 | {
"authors": [
"JianBinWu",
"bolianghui"
],
"repo": "JianBinWu/Wallet-iOS",
"url": "https://github.com/JianBinWu/Wallet-iOS/issues/4",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
675784038 | collectionView controller as contentviewcontroller ?
it looks like all examples are using tableview controller as content view controller. Is collection viewcontroller as content supported ? any example?
nvm, it also works 👍
There seems no limitation to what type of VC content view controller it can accept. Pretty general
| gharchive/issue | 2020-08-09T22:32:31 | 2025-04-01T06:37:06.258375 | {
"authors": [
"10000TB"
],
"repo": "Jiar/SegementSlide",
"url": "https://github.com/Jiar/SegementSlide/issues/60",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2172074950 | Can’t open tour from getHeaderActions but does work from widget button.
What happened?
Launching a tour from a page header action does not work, I have the following:
protected function getHeaderActions(): array
{
return [
Action::make('Tour')->dispatch('filament-tour::open-tour', ['tour_dashboard']),
];
}
When clicking the button I get this error Tour with id 'tour_dashboard' not found.
However, if I add this to a widget’s view on the same page, it will launch the tour as expected.
<button wire:click="$dispatch('filament-tour::open-tour', 'tour_dashboard')">Tour</button>
How to reproduce the bug
Create a tour on a custom dashboard then include a header action to dispatch the open-tour event.
Package Version
3.1.0.3
PHP Version
8.2.15
Laravel Version
10.46.0
Which operating systems does with happen with?
Windows
Notes
No response
I have the same issue : console displays "Tour with id 'blabla' not found
Same for me.
Solution here #8 Adding tour_ before the id
Action::make('demo')->link()->action(fn($livewire)=>$livewire->dispatch('filament-tour::open-tour', 'tour_grow-page-list')),
// or
Action::make('demo')->link()->dispatch('filament-tour::open-tour', ['tour_grow-page-list']),
Hi ! This bug will be fixed in the next release !
| gharchive/issue | 2024-03-06T17:35:46 | 2025-04-01T06:37:06.262665 | {
"authors": [
"FDT2k",
"JibayMcs",
"RicLeP",
"invaders-xx"
],
"repo": "JibayMcs/filament-tour",
"url": "https://github.com/JibayMcs/filament-tour/issues/12",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
363784797 | Enable TCP_NODELAY
It might make sense to enable TCP_NODELAY in our SOCKS server to avoid unnecessary Nagling.
It might make sense to enable TCP_NODELAY in our SOCKS server to avoid unnecessary Nagling.
Kode
I think this is fixed with the migration. to Go?
Yeah, Go is nodelay by default.
| gharchive/issue | 2018-09-25T22:07:03 | 2025-04-01T06:37:06.265837 | {
"authors": [
"bemasc",
"fortuna",
"shatlyktma"
],
"repo": "Jigsaw-Code/Intra",
"url": "https://github.com/Jigsaw-Code/Intra/issues/80",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
707614250 | bugged "Click To Update!"
as the title says, the click to update at the top is bugged and can't be clicked on.
image
im having the same issue :/
| gharchive/issue | 2020-09-23T19:12:11 | 2025-04-01T06:37:06.266982 | {
"authors": [
"Sebas3525",
"honk7777"
],
"repo": "Jiiks/BetterDiscordApp",
"url": "https://github.com/Jiiks/BetterDiscordApp/issues/955",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
497129848 | Add a license
Can you please add a license to the repository?
The absence of a license makes usage and contribution tricky.
Thanks for the suggestion! I'll add MIT license.
| gharchive/issue | 2019-09-23T14:22:58 | 2025-04-01T06:37:06.276701 | {
"authors": [
"Jinmo",
"tmr232"
],
"repo": "Jinmo/idapkg",
"url": "https://github.com/Jinmo/idapkg/issues/13",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2161125710 | Versao04
Criação do pull request da Versao04
Show de bola
| gharchive/pull-request | 2024-02-29T12:18:55 | 2025-04-01T06:37:06.284201 | {
"authors": [
"JoaoVitor2807",
"YgorGoncalves754"
],
"repo": "JoaoVitor2807/AtividadeGit",
"url": "https://github.com/JoaoVitor2807/AtividadeGit/pull/4",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2407637123 | Add decred to macos/ios scripts and rescan page bug fix
Builds on #12.
Tested on simnet: https://github.com/JoeGruffins/cake_wallet/issues/11#issuecomment-2227516005
Modified commits:
https://github.com/JoeGruffins/cake_wallet/commit/ceef7e0f23689659fd96fc608c22e6e423b969c1
https://github.com/JoeGruffins/cake_wallet/commit/97ad5c308c89d6aae5b0b555de93aef2e6d46f67
https://github.com/JoeGruffins/cake_wallet/commit/f5305670a2a775a54534820e32283efd54abdf66
Is 7531942 for testing or do we need those changes?
Is 7531942 for testing or do we need those changes?
Errm, most of it was from @itswisdomagain commit on the stale branch. But we need the script and readme changes.
We would clean that commit of test changes once we are ready to push upstream.
@JoeGruffins you might need to replace the upstream commit with this one: https://github.com/JoeGruffins/cake_wallet/pull/13/commits/55d837ccc2a1a42083b95d7d9794e4e184a170ea
Ok I updated it.
Thanks.
getting an error when building:
lib/entities/provider_types.dart:50:29: Error: A non-null value must be returned since the return type 'List<ProviderType>' doesn't allow null.
- 'List' is from 'dart:core'.
- 'ProviderType' is from 'package:cake_wallet/entities/provider_types.dart' ('lib/entities/provider_types.dart').
static List<ProviderType> getAvailableBuyProviderTypes(WalletType walletType) {
^
lib/entities/provider_types.dart:87:29: Error: A non-null value must be returned since the return type 'List<ProviderType>' doesn't allow null.
- 'List' is from 'dart:core'.
- 'ProviderType' is from 'package:cake_wallet/entities/provider_types.dart' ('lib/entities/provider_types.dart').
static List<ProviderType> getAvailableSellProviderTypes(WalletType walletType) {
I think this comment was intended for #14, it has been resolved.
I think this comment was intended for https://github.com/JoeGruffins/cake_wallet/pull/14, it has been resolved.
Yeah sorry about that.
Changes have been added to the upstream pr so closing.
| gharchive/pull-request | 2024-07-14T23:20:47 | 2025-04-01T06:37:06.293422 | {
"authors": [
"JoeGruffins",
"ukane-philemon"
],
"repo": "JoeGruffins/cake_wallet",
"url": "https://github.com/JoeGruffins/cake_wallet/pull/13",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1191137051 | BUG
Al traducir texto a Braille desde un archivo se genera lo adjunto. Revisar!
En realidad no es un bug porque no se tiene registro de algunos caracteres especiales que se agregó como nota en el mapping general.
Se buscó la representación de estos caracteres para representarlos en braille pero no se encontraron los mismos.
TODO: tal vez investigar localmente si existe un mappeo especial para estos caracteres al braille y agregarlos manualmente.
| gharchive/issue | 2022-04-04T01:36:25 | 2025-04-01T06:37:06.318501 | {
"authors": [
"JoelVG"
],
"repo": "JoelVG/text-to-braille",
"url": "https://github.com/JoelVG/text-to-braille/issues/18",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1628121859 | Pick Color Scheme & name
For fun easy issue
On This now as well as figma
https://www.figma.com/file/ExU4vVOZXjSc01fPM60rU5/Untitled?node-id=10%3A39&t=AV6VEJNCelnS5bnE-1
| gharchive/issue | 2023-03-16T19:18:15 | 2025-04-01T06:37:06.339324 | {
"authors": [
"John4064"
],
"repo": "John4064/WoW-Ranking",
"url": "https://github.com/John4064/WoW-Ranking/issues/7",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
527831334 | Show Battery
It would be nice if there was a setting to show the current battery level or to hide it if the battery was fully charged
Adding this as a maybe
This is now in beta3 : https://github.com/JohnCoates/Aerial/releases/tag/v1.6.5beta3
Beware after installing, you will need to set again your video format (1080p, 4K, HDR) as I changed some more internal stuff.
| gharchive/issue | 2019-11-25T04:05:26 | 2025-04-01T06:37:06.340845 | {
"authors": [
"admiral-ackbar",
"glouel"
],
"repo": "JohnCoates/Aerial",
"url": "https://github.com/JohnCoates/Aerial/issues/887",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1092024936 | Static typing and sizing
Currently Forscape does all type and dimension checking at runtime. We would like to transfer this responsibility to compile time, not only for performance, but also for semantic formatting. We can highlight errors prior to running the interpreter, and even bold matrix identifiers.
There are challenges to consider:
Generic functions
Symbolic dimensions, e.g. an m×n matrix
Aggregate types, e.g. lists of X, sets of X, etc.
Nonhomogeneous aggregate types
Subtypes, e.g. int is a subtype of rational is a subtype of real
Inference
Trying something too difficult and failing
Probably use "call-site template function instantiation", or whatever it's really called. I'm having trouble finding educational material.
With call-site instantiation of generic functions:
The function declaration doesn't give you much, just the number of parameters, number of default args, and their types
A function plus a list of arg types lets you deduce the return type
The function to call is the result of an expression, so you need to be sure it is instantiated, and you need to deduce the return type without being able to statically determine the function to call
Can you return a generic function, then deduce its type at call sites? You need the body to deduce the type. When the called function could evaluate to different options, you need all the options.
I halfway think it can work, but it's doing my head in. Maybe I should sleep.
Actually, I think I'm biting off much more than I can chew with generic functions. I don't need to get bogged down on tangents, just make a restriction that the function declaration includes a fully specified signature, and make a note to relax that restriction later.
Well declare-time deduction is incredibly limited. It has challenges with recursion and functions passed as arguments. The closure examples are a little contrived, but the isolated function definitions have ambiguous signatures. The root finding example fails because for f(x), f could be a function ℝ → ℝ or just a member of ℝ resulting in implicit multiplication. To have any chance of success, you'll need to consider the call site. But then you're statically tying identifiers to functions, which is not cool.
But static typing is the gateway to most of the interesting features, and it should be done non-intrusively. Get it together!
Well this just won't happen quickly. First things first, I need another Very Sophisticated Vector™ to represent types. I expect it will be similar to the "tree" in KiCAS. Then with that support in place, a lot of experimentation to find what works. Probably not a weekend project...
Perhaps a call graph would be helpful?
The problem with any type deduction involving the call site is that we cannot generally determine which function will be called. Take for example:
alg add(x, y) return x + y
alg mult(x, y) return x * y
f = addfalsemulttrue
print(f(3, 3))
However, I believe we can determine all the functions which could possibly be called, and instantiate them with the correct types.
Static type checking basically works. It still lacks some tedious details around references captures, function prototypes, and recursion. It needs to clone functions rather than just use the original definition (vital for recursion) while keeping valid references (can think about frame offsets in the symbol table). Meaningful error messages will require some book-keeping to report at the declare and/or call sites. The whole thing could use a refactor, especially before pressing on to static dimensioning and autodiff. But it works 😭.
I almost have test coverage, but I worry that "cursed_factorial" will expose my recursion strategy as too naive. I had thought that if you hit recursion, you resolve as much as possible in your pass, then take an additional pass to either finish or conclude you have a cycle without an exit. It may not be the simple, but I still think it's feasible. This is a nice puzzle.
It seems to work now. Hopefully the recursion handling doesn't admit invalid Forscape programs. The IDE crashes constantly now, and the typing error messages are terrible, but that's cleanup work.
This is set to close with https://github.com/JohnDTill/Forscape/pull/38. There is still more to explore with instantiation as opposed to just checking, and static dimensioning, but the scope of this issue has been large enough!
| gharchive/issue | 2022-01-02T13:52:41 | 2025-04-01T06:37:06.350422 | {
"authors": [
"JohnDTill"
],
"repo": "JohnDTill/Forscape",
"url": "https://github.com/JohnDTill/Forscape/issues/19",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
165219842 | mark iOS and tvOS frameworks save for app extensions
Just ticking the boxes in the iOS and tvOS schemes to mark them safe for app extensions
Thanks! 👍 This seems ok, but I'll need to raise GCDKit's version as well or Carthage won't see this change if it cached the previous version's repository. I'll merge your fork and push it as 1.2.6 later when I get free time :) (Same with CoreStore)
| gharchive/pull-request | 2016-07-13T01:55:10 | 2025-04-01T06:37:06.352388 | {
"authors": [
"JohnEstropia",
"jannon"
],
"repo": "JohnEstropia/GCDKit",
"url": "https://github.com/JohnEstropia/GCDKit/pull/13",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
209272719 | Add handling for disambiguation pages in wikipedia
Fixes #97
Hey @scarescrow -- Thanks for this. 👍 I'll probably lint it against flake8 locally unless you'd like to do so yourself.
However, I ran into an issue: When I hit a disambiguation page, I seemed to get an error. If I remember correctly I tried .wiki Test and .wiki A (disambiguation).
I'm willing to look into it and fix it up, but I'd have to find some free time. 🙂 It'd be nice if you could verify that you are seeing the same thing, or if you could provide a working disambiguation page to see how it works! Thanks again! 😄
Hey @JohnMaguire , did you get errors for both pages, Test and A (disambiguation)?
I checked for .wiki hello and got the link to the disambiguation page correctly.
I'll check it once more with the pages you mentioned to see if I'm getting the errors as well.
Will reopen if a fix is posted.
| gharchive/pull-request | 2017-02-21T21:06:59 | 2025-04-01T06:37:06.357775 | {
"authors": [
"JohnMaguire",
"scarescrow"
],
"repo": "JohnMaguire/Cardinal",
"url": "https://github.com/JohnMaguire/Cardinal/pull/126",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2599837236 | RemoteEvent / BindableEvent / RemoteFunction / BindableFunction should be "unknown?" instead of "any"
As the firing of these Roblox instances have no way to enforce type safety, I feel it's cleanest to make developers validate the types that are received.
In this below example, no errors are detected - everything is fine. That player variable could be passed into another function and cause a really weird error somewhere far away from the event.
local bindableEvent: BindableEvent
bindableEvent.Event:Connect(function(player: Player)
-- the player is actually a string, but the dev is expecting a player
end)
bindableEvent:Fire(Players.LocalPlayer.Name)
What I propose is that it's better for the code to accept this uncertainty, and refine it in the following code. For example:
local bindableEvent: BindableEvent
bindableEvent.Event:Connect(function(player: unknown?)
assert(typeof(player) == "Instance", `player is not instance, got type "{typeof(player)}"`)
assert(player:IsA("Player"), `player is not player, got class "{player.ClassName}"`)
-- the type is now certainly a player, or has given the developer a clear error at the earliest point possible
end)
bindableEvent:Fire(Players.LocalPlayer.Name)
I feel this handling of it is safe and more accurate to what's actually happening in the code, similar to the improvement made to attributes earlier.
Thank you for your time!
The main reason I am hesistant about this is from a user experience PoV, as switching this from any to unknown will lead to loads of type errors throughout existing code bases I imagine.
I can understand that, I guess I just feel it's a purer version of type safety that some developers will find desirable. If having it as the default behavior is too extreme, would considering it as a setting / flag be viable? Similar to strict datamodels.
| gharchive/issue | 2024-10-20T02:02:52 | 2025-04-01T06:37:06.365119 | {
"authors": [
"JohnnyMorganz",
"nightcycle"
],
"repo": "JohnnyMorganz/luau-lsp",
"url": "https://github.com/JohnnyMorganz/luau-lsp/issues/799",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
271424386 | Calling malloc/free should be done by scheduler
void * mymalloc(size_t size, gthread_task_t* owner);
void myfree(void * data, gthread_task_t* owner);
Both of these need to be invoked by scheduler to indicate the calling thread. Currently, I am not sure as to what the best way to do this is (what files need to be changed) so if one of you guys with more knowledge of the scheduler code could do it, that would be cool.
I'll take a look at this. There's one or two places where the scheduler uses libc malloc() and free() and I want to get rid of those as well. I know the assignment description says there should be an indication for mymalloc()s done by the scheduler and "kernel" but I think we can get rid of these and only have mymalloc() used by user code.
| gharchive/issue | 2017-11-06T10:23:12 | 2025-04-01T06:37:06.420740 | {
"authors": [
"JonNRb",
"khalkash"
],
"repo": "JonNRb/gthread",
"url": "https://github.com/JonNRb/gthread/issues/15",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2255243754 | added roc support
Added support of Roc language.
Thanks for the contribution!
| gharchive/pull-request | 2024-04-21T20:11:32 | 2025-04-01T06:37:06.451628 | {
"authors": [
"JoosepAlviste",
"jluzny"
],
"repo": "JoosepAlviste/nvim-ts-context-commentstring",
"url": "https://github.com/JoosepAlviste/nvim-ts-context-commentstring/pull/105",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2608374859 | Code Fix for CS8603 Possible Null Reference Return
We have been using the Nulless.bang analyzer to ensure null safety in our .NET C# projects, particularly in helping us migrate to strict nullable reference types mode while maintaining legacy code.
However, we would like to address the CS8603 warning for possible null reference returns. Currently, the following code results in a CS8603 warning:
public string demo()
{
return null;
}
To resolve this, we would like to introduce a code fix to modify the code to handle nullable returns explicitly. Here's how the code should look after the fix:
public string? demo()
{
return null;
}
This change should ensure that nullable reference types are correctly handled, and CS8603 warnings are resolved.
Steps to Implement:
Modify the code fix to identify CS8603 warnings.
Automatically update return types to nullable (string? instead of string) when the method can return null.
Ensure this change is compatible with existing code and maintains functionality.
Expected Outcome:
The analyzer should modify methods that return null to use nullable return types, resolving the CS8603 warning.
Null safety should be improved while keeping flexibility for legacy code.
Nice !
| gharchive/issue | 2024-10-23T12:13:01 | 2025-04-01T06:37:06.461337 | {
"authors": [
"JoostVanVelthoven",
"TLA020"
],
"repo": "JoostVanVelthoven/Nullness.Bang",
"url": "https://github.com/JoostVanVelthoven/Nullness.Bang/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
344834727 | Add setting to limit the amount of logic updates per render update
To prevent enemies from jumping too much.
Downside is that the game speed will be lowered
Done
| gharchive/issue | 2018-07-26T13:02:50 | 2025-04-01T06:37:06.467096 | {
"authors": [
"JordyMoos"
],
"repo": "JordyMoos/elm-pixel-boulder-game",
"url": "https://github.com/JordyMoos/elm-pixel-boulder-game/issues/61",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2213992048 | KREAd frontend should remove legacy queries
Description
Creating this issue in Kread repo to match https://github.com/Agoric/agoric-sdk/issues/9145
In the upcoming agoric-upgrade-16, we plan to upgrade our cosmos-sdk major version to 0.47. This version of cosmos-sdk removes support for "legacy queries", which had been deprecated since version 0.45. Our codebase originally demonstrated use of legacy queries as the mechanism for reading "vstorage", and this pattern had been copied by some versions of the KREAd frontend code.
Proposed Solution
Proposed Solution
The Kryha/KREAd frontend appear to have eliminated legacy queries on at least recent version of the development branch. However, we are uncertain if that is the version currently running in production.
This ticket is to ensure that the public-facing KREAd frontend does not use legacy queries, whether that involves development, deployment, or if it's already done this ticket can be closed.
See https://github.com/Agoric/agoric-sdk/issues/9096 for more information on legacy queries and how to switch to alternative query methods.
Acceptance criteria
Confirmation legacy queries not used
Additional info
No response
Conclusion
The current state of KREAd running in production, commit 1f281a83b1bdf374a5bf11358c537829da614508 , the agoric/rpc pkg being imported has the version "^0.6.0", package.json.
This confirms that the KREAd version running in production still uses legacy queries.
Note: the same is true for the development branch, package.json.
Context
At the PR #40 , were the rpc pkg is upgraded to version 0.6.0 , the ChainStorageWatcher executes a method used to query the vstorage called batchVstorageQuery, where we can see that the structure passed as option to the fetch method is the same as described as legacy queries on the issue #9096
const options = {
method: 'POST',
body: JSON.stringify(
paths.map((path, index) => ({
jsonrpc: '2.0',
id: index,
method: 'abci_query',
params: { path: `/custom/vstorage/${path[0]}/${path[1]}` },
})),
),
};
Source
On the PR #55 the batchVstorageQuery is update to use the JSON API instead of the deprecated RPC method.
On the PR #65, the batchQuery file is replaced with vstorageQuery , which still use the expected JSON API to execute the queries.
This feature is present from the @agoric/rpc version: 0.7.2 and forward.
The latest version is 0.9.0
| gharchive/issue | 2024-03-28T19:39:30 | 2025-04-01T06:37:06.475648 | {
"authors": [
"Jorge-Lopes",
"otoole-brendan"
],
"repo": "Jorge-Lopes/KREAd",
"url": "https://github.com/Jorge-Lopes/KREAd/issues/4",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
89881169 | Decreasing Average Fitness
Of the last four simulations I've ran, the average fitness consistently decreases each generation. This is obviously an issue. There is virtually no improvement occurring.
With my next simulation, I'm going to try and give more weight to chromosomes with a higher fitness when selecting parents during reproduction.
And of course, there is always the possibility that my fundamental implementation of a genetic algorithm is flawed, as this is only my second project utilizing such.
| gharchive/issue | 2015-06-21T08:16:38 | 2025-04-01T06:37:06.507003 | {
"authors": [
"JoshuaBrockschmidt"
],
"repo": "JoshuaBrockschmidt/ideal_ANN",
"url": "https://github.com/JoshuaBrockschmidt/ideal_ANN/issues/4",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
459164926 | Can not install better-sqlite3
I keep getting these errors when I try to install it
gyp ERR! configure error
gyp ERR! stack Error: Command failed: C:\Users\Ahmad Alfalasi\AppData\Local\Programs\Python\Python37-32\python.EXE -c import sys; print "%s.%s.%s" % sys.version_info[:3];
gyp ERR! stack File "", line 1
gyp ERR! stack import sys; print "%s.%s.%s" % sys.version_info[:3];
gyp ERR! stack ^
gyp ERR! stack SyntaxError: invalid syntax
gyp ERR! stack
gyp ERR! stack at ChildProcess.exithandler (child_process.js:294:12)
gyp ERR! stack at ChildProcess.emit (events.js:182:13)
gyp ERR! stack at maybeClose (internal/child_process.js:962:16)
gyp ERR! stack at Process.ChildProcess._handle.onexit (internal/child_process.js:251:5)
gyp ERR! System Windows_NT 10.0.17134
gyp ERR! command "C:\Program Files\nodejs\node.exe" "C:\Users\Ahmad Alfalasi\AppData\Roaming\npm\node_modules\npm\node_modules\node-gyp\bin\node-gyp.js" "rebuild"
gyp ERR! cwd C:\Users\Ahmad Alfalasi\Desktop\Current bots\bot\node_modules\integer
gyp ERR! node -v v10.15.0
gyp ERR! node-gyp -v v3.8.0
gyp ERR! not ok
npm WARN enoent ENOENT: no such file or directory, open 'C:\Users\Ahmad Alfalasi\Desktop\Current bots\bot\package.json'
npm WARN discord.js@11.5.1 requires a peer of bufferutil@^4.0.0 but none is installed. You must install peer dependencies yourself.
npm WARN discord.js@11.5.1 requires a peer of erlpack@discordapp/erlpack but none is installed. You must install peer dependencies yourself.
npm WARN discord.js@11.5.1 requires a peer of libsodium-wrappers@^0.7.3 but none is installed. You must install peer dependencies yourself.
npm WARN discord.js@11.5.1 requires a peer of node-opus@^0.2.7 but none is installed. You must install peer dependencies yourself.
npm WARN discord.js@11.5.1 requires a peer of opusscript@^0.0.6 but none is installed. You must install peer dependencies yourself.
npm WARN discord.js@11.5.1 requires a peer of sodium@^2.0.3 but none is installed. You must install peer dependencies yourself.
npm WARN discord.js@11.5.1 requires a peer of @discordjs/uws@^10.149.0 but none is installed. You must install peer dependencies yourself.
npm WARN bot No description
npm WARN bot No repository field.
npm WARN bot No README data
npm WARN bot No license field.
npm ERR! code ELIFECYCLE
npm ERR! errno 1
npm ERR! integer@2.1.0 install: node-gyp rebuild
npm ERR! Exit status 1
npm ERR!
npm ERR! Failed at the integer@2.1.0 install script.
npm ERR! This is probably not a problem with npm. There is likely additional logging output above.
npm ERR! A complete log of this run can be found in:
npm ERR! C:\Users\Ahmad Alfalasi\AppData\Roaming\npm-cache_logs\2019-06-21T12_06_10_612Z-debug.log
#279 is a duplicate of this, but I'll favor that one since it already got a response.
| gharchive/issue | 2019-06-21T12:06:56 | 2025-04-01T06:37:06.519358 | {
"authors": [
"Dipsip6969",
"JoshuaWise"
],
"repo": "JoshuaWise/better-sqlite3",
"url": "https://github.com/JoshuaWise/better-sqlite3/issues/278",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1122842980 | Possible to use CURRENT_TIMESTAMP in update function?
better-sqlite3 provides a helper function db.update(table, data, where, whiteList). For the data parameter, we can pass in a key-value pairs to specify the column name and column value.
However, it treats the value as literal value, e.g. db.update('user', {updated_at:'CURRENT_TIMESTAMP'}, {id:1}) will set the updated_at column to be string literal 'CURRENT_TIMESTAMP', instead of setting the column to the current database timestamp.
Work around is to use prepared statement, however that will be much more verbose. In the current design, is there a way to express sql expression on the column value part?
better-sqlite3 provides a helper function db.update(table, data, where, whiteList)
Does it? Can't find it in the docs and that sounds oddly specific and high level, not like something this package would offer. https://github.com/JoshuaWise/better-sqlite3/blob/master/docs/api.md#class-database
const Database = require('better-sqlite3');
const db = new Database(':memory:');
console.log(db.update);
Logs undefined.
I don't think this question is related to better-sqlite3. If it is, can you provide a minimal example?
Oh I was referring to better-sqlite3-helper, submitted to the wrong repo
| gharchive/issue | 2022-02-03T09:25:59 | 2025-04-01T06:37:06.523185 | {
"authors": [
"Prinzhorn",
"beenotung"
],
"repo": "JoshuaWise/better-sqlite3",
"url": "https://github.com/JoshuaWise/better-sqlite3/issues/760",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1087272162 | removing the new script
Upper
I am updating the file, with some delating files
comment here 👍
| gharchive/pull-request | 2021-12-23T00:16:15 | 2025-04-01T06:37:06.524332 | {
"authors": [
"Josue-PonceCaro"
],
"repo": "Josue-PonceCaro/Testing",
"url": "https://github.com/Josue-PonceCaro/Testing/pull/1",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
971986072 | Can not drag the blocks to body?
How can i use this plugin? I tried the demo but the blocks is not draggle.
Not sure why it's not possible to drag the blocks onto the canvas anymore.
| gharchive/issue | 2021-08-16T18:16:41 | 2025-04-01T06:37:06.583788 | {
"authors": [
"Ju99ernaut",
"shkhalid"
],
"repo": "Ju99ernaut/grapesjs-ga",
"url": "https://github.com/Ju99ernaut/grapesjs-ga/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1125215544 | [INIT] Create data model from UML diagram
Create the data model via classes and their associated tables in an SQLite database
https://github.com/Juansero29/Loop/blob/main/docs/uml.svg
Folders per modules
modules/models.ts
modules/services.ts
const create = ...
const update = ...
export default { create, update }
For exporting useful methods from services
https://ionicframework.com/docs/native/sqlite
https://github.com/storesafe/cordova-sqlite-storage
https://github.com/storesafe/cordova-sqlite-storage#android-database-provider
https://github.com/alex-steinberg/ionic-react-sqlite-example/blob/master/src/pages/Home.tsx
https://github.com/typeorm/typeorm
Dependency Injection of repositories: https://thomasburlesonia.medium.com/https-medium-com-thomasburlesonia-universal-dependency-injection-86a8c0881cbc
That dependency injection system looks quite nice.
npm install @mindspace-io/utils --save
It allows decoupling of use and construction, and easy testing via mocks. I didn't fully understand however how to tell when to use the mock and when to use the "real" implementation. Do I need to comment out the real one to use the mock? Seems weird => need to investigate further this library
Also:
Don’t forget to build a custom hook to make DI lookups super easy!
Looks nice
Updated typeorm site: https://typeorm.io/
Also, updated dependency injector is this one: https://www.npmjs.com/package/@mindspace-io/react - oddly enough it requires use of --legacy-deps in order to be installed
| gharchive/issue | 2022-02-06T14:19:08 | 2025-04-01T06:37:06.630494 | {
"authors": [
"Juansero29"
],
"repo": "Juansero29/Loop",
"url": "https://github.com/Juansero29/Loop/issues/11",
"license": "Unlicense",
"license_type": "permissive",
"license_source": "github-api"
} |
730100108 | Apple pay window does not show for the first time apple pay is initiated
Version used:
2.1.0
On the first launch of the app and when the first time the apple pay is initiated, the apple pay does not
show and the user canceled message is triggered.
On further attempts, this works without any problem.
Please check out 3.1.0 and see if it persists: https://www.npmjs.com/package/judokit-react-native
Tried this latest version and got an error with apple pay. More details here https://github.com/Judopay/JudoKit-ReactNative/issues/82
Issue was handled at #82
It occurs in the 3.3.5 version of judokit-react-native
Issue is happening now in version 3.3.7 of judokit-react-native also.
| gharchive/issue | 2020-10-27T04:40:52 | 2025-04-01T06:37:06.710523 | {
"authors": [
"PavithraPurushothaman-T2S",
"mpetrenco",
"radhakrishnant2s",
"selvamariappant2s"
],
"repo": "Judopay/JudoKit-ReactNative",
"url": "https://github.com/Judopay/JudoKit-ReactNative/issues/72",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
963568821 | Safe keyboard-direct checks
Blocks keyboard on iOS/tvOS native (I lack the hardware atm to test), allows conditional Android and Switch keyboard support (default off).
Tasty. Good work!
| gharchive/pull-request | 2021-08-09T01:20:51 | 2025-04-01T06:37:06.712609 | {
"authors": [
"JujuAdams",
"offalynne"
],
"repo": "JujuAdams/Input",
"url": "https://github.com/JujuAdams/Input/pull/139",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
804915552 | Add underline and strikethrough
I don't know if you can do this with normal game maker studio text functions let alone with scribble but this feels like a nice thing to have...
[slant] can be used to emulate italics (but tends to be frowned on by typographers). Bold doesn't have an equivalent; I don't know what Juju's stance on adding one would be. With that said, you can set up a font with actual bold and italics attributes in GameMaker and change the font to it on the fly with [font_name]`.
Strikethrough and underline are unaccounted for since those are just lines being drawn and not actually font attributes. I imagine those cold be added.
it would be easier if this was built into scribble, it eliminates a few extra steps...
Actually, I was looking at an old version of the documentation: you can now combine font families together, and switch between them using formatting tags: https://github.com/JujuAdams/Scribble/wiki/Functions-(Font-Modification)
As Dragonite kindly points out, swapping between bold and italic (and bold+italic) fonts is supported using the [b] and [i] and [bi] command tags. You need to set these up manually using scribble_font_set_style_family as Scribble does not procedurally associate fonts (it used to, but it was unreliable).
Adding underline and strikethrough is... something I'll think about. I'm not 100% sure if GM reveals enough information to handle these features, though I will check.
I'm sure you can put together something that's "good enough" just based on the dimensions of each glyph and or the final positions of each vertex. No idea how much doing something simple like that would offend typographers though.
I just implemented underline, I'm adding my (badly written) functions to get individual lines width and draw an underline in case it's useful as a starting point. It's of course only for horizontal, left to right languages.
The textargument is the scribble element.
{
var _line_count = text.get_line_count();
var _glyph_count = text.get_glyph_count();
var _previous_x = 99999;
var _last_leftmost_glyph = -1;
text[$ "__lines_width"] = array_create(_line_count);
if(_line_count == 1) text.__lines_width[0] = text.get_bbox().width;
else
{
var j = 0;
for(var i = 0; i < _glyph_count; i++)
{
var _current_x = text.get_glyph_data(i).left;
if(_current_x < _previous_x || i == _glyph_count - 1)
{
if(_last_leftmost_glyph != -1)
{
text.__lines_width[j] = text.get_glyph_data(i - 1).right - text.get_glyph_data(_last_leftmost_glyph).left;
j++;
}
_last_leftmost_glyph = i;
}
_previous_x = _current_x;
}
}
text[$ "__total_width"] = 0;
for(var i = 0; i < _line_count; i++) text.__total_width+= text.__lines_width[i];
}
function scribble_draw_underline(text,xx,yy,underline_pos,typist = undefined,color = c_white,alpha = 1,thickness = 3)
{
if(underline_pos < 0) exit;
if(text[$ "__lines_width"] == undefined) show_debug_message("Error trying to draw underline before having called 'scribble_get_lines_width'");
draw_set_color(color);
draw_set_alpha(alpha)
if(typist != undefined) var _underline_pos = min(typist.__revealed_total_width,underline_pos);
else _underline_pos = underline_pos;
var _current_left = 0;
var _bbox = text.get_bbox();
for(var i = 0; i < text.get_line_count(); i++)
{
if(_underline_pos < _current_left) break;
if(_underline_pos >= _current_left + text.__lines_width[i]) var line_progress = text.__lines_width[i];
else line_progress = _underline_pos - _current_left;
var dist = line_height/2 + (i * line_height);
var x1 = xx + lengthdir_x(dist,-90) + (_bbox.width - text.__lines_width[i])/2;
var y1 = yy + lengthdir_y(dist,-90);
draw_rectangle(x1,y1,x1 + line_progress,y1 + thickness,false);
_current_left+= text.__lines_width[i];
}
}```
| gharchive/issue | 2021-02-09T20:52:27 | 2025-04-01T06:37:06.718195 | {
"authors": [
"DragoniteSpam",
"JujuAdams",
"OmegaX1000",
"patchuby"
],
"repo": "JujuAdams/Scribble",
"url": "https://github.com/JujuAdams/Scribble/issues/184",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1920373396 | TagBot trigger issue
This issue is used to trigger TagBot; feel free to unsubscribe.
If you haven't already, you should update your TagBot.yml to include issue comment triggers.
Please see this post on Discourse for instructions and more details.
If you'd like for me to do this for you, comment TagBot fix on this issue.
I'll open a PR within a few hours, please be patient!
Triggering TagBot for merged registry pull request: https://github.com/JuliaRegistries/General/pull/92349
Triggering TagBot for merged registry pull request: https://github.com/JuliaRegistries/General/pull/92881
Triggering TagBot for merged registry pull request: https://github.com/JuliaRegistries/General/pull/93181
Triggering TagBot for merged registry pull request: https://github.com/JuliaRegistries/General/pull/93821
Triggering TagBot for merged registry pull request: https://github.com/JuliaRegistries/General/pull/98497
Triggering TagBot for merged registry pull request: https://github.com/JuliaRegistries/General/pull/108275
| gharchive/issue | 2023-09-30T19:23:59 | 2025-04-01T06:37:06.742072 | {
"authors": [
"JuliaTagBot"
],
"repo": "JuliaAI/MLJBalancing.jl",
"url": "https://github.com/JuliaAI/MLJBalancing.jl/issues/6",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1092862129 | For a 0.19.2 release
#706
Codecov Report
Merging #707 (483263e) into master (5d8c78c) will decrease coverage by 2.68%.
The diff coverage is 100.00%.
:exclamation: Current head 483263e differs from pull request most recent head a852ddc. Consider uploading reports for the commit a852ddc to get more accurate results
@@ Coverage Diff @@
## master #707 +/- ##
==========================================
- Coverage 86.53% 83.84% -2.69%
==========================================
Files 36 36
Lines 3401 2904 -497
==========================================
- Hits 2943 2435 -508
- Misses 458 469 +11
Impacted Files
Coverage Δ
src/MLJBase.jl
100.00% <100.00%> (+7.14%)
:arrow_up:
src/interface/data_utils.jl
91.30% <100.00%> (-2.03%)
:arrow_down:
src/sources.jl
70.00% <0.00%> (-18.00%)
:arrow_down:
src/composition/models/transformed_target_model.jl
85.45% <0.00%> (-14.55%)
:arrow_down:
src/data/datasets.jl
86.84% <0.00%> (-13.16%)
:arrow_down:
src/measures/continuous.jl
89.13% <0.00%> (-7.17%)
:arrow_down:
src/show.jl
29.92% <0.00%> (-7.12%)
:arrow_down:
src/measures/measures.jl
64.63% <0.00%> (-6.24%)
:arrow_down:
src/measures/probabilistic.jl
58.46% <0.00%> (-4.70%)
:arrow_down:
... and 26 more
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 5d8c78c...a852ddc. Read the comment docs.
| gharchive/pull-request | 2022-01-03T22:23:00 | 2025-04-01T06:37:06.758299 | {
"authors": [
"ablaom",
"codecov-commenter"
],
"repo": "JuliaAI/MLJBase.jl",
"url": "https://github.com/JuliaAI/MLJBase.jl/pull/707",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2610471795 | is it possible to embed javascript in Config()
I have been playing around with Apache ECharts and the JSON object it uses for plotting can take Javascript functions as callbacks. For example
{
type: "scatter",
symbol: "circle",
symbolSize: function (val) {return val[2] * 2; }
}
Config() object can take in string, but how do I render the callback functions in JSON without the quotes for string. Performing JSON3.write() gets me "symbolSize: function (val) {return val[2] * 2; }" but I would like to render symbolSize: function (val) {return val[2] * 2; }
I think you'd need to find/create a type that JSON3 will write as you want.
| gharchive/issue | 2024-10-24T05:03:31 | 2025-04-01T06:37:06.801556 | {
"authors": [
"asbisen",
"joshday"
],
"repo": "JuliaComputing/EasyConfig.jl",
"url": "https://github.com/JuliaComputing/EasyConfig.jl/issues/9",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2022762812 | ComputationalHeatTransfer with Immersed Layers
Added functionality of Dirichlet and Neumann problems with method dispatching for unbounded problems
I've merged the pull request, but there are still a few small issues with the documentation. There needs to be some documentation for the APIs.
| gharchive/pull-request | 2023-12-03T23:22:35 | 2025-04-01T06:37:07.000022 | {
"authors": [
"jdeldre",
"masteral456"
],
"repo": "JuliaIBPM/ComputationalHeatTransfer.jl",
"url": "https://github.com/JuliaIBPM/ComputationalHeatTransfer.jl/pull/9",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
422345670 | Failed to start kernel
Trying to open up a notebook after starting Jupyter up with IJulia.notebook() on Julia 1.1 nets me the following error message:
Traceback (most recent call last):
File "/Users/soderhos/.julia/conda/3/lib/python3.7/site-packages/tornado/web.py", line 1592, in _execute
result = yield result
File "/Users/soderhos/.julia/conda/3/lib/python3.7/site-packages/tornado/gen.py", line 1133, in run
value = future.result()
File "/Users/soderhos/.julia/conda/3/lib/python3.7/site-packages/tornado/gen.py", line 1141, in run
yielded = self.gen.throw(*exc_info)
File "/Users/soderhos/.julia/conda/3/lib/python3.7/site-packages/notebook/services/sessions/handlers.py", line 73, in post
type=mtype))
File "/Users/soderhos/.julia/conda/3/lib/python3.7/site-packages/tornado/gen.py", line 1133, in run
value = future.result()
File "/Users/soderhos/.julia/conda/3/lib/python3.7/site-packages/tornado/gen.py", line 1141, in run
yielded = self.gen.throw(*exc_info)
File "/Users/soderhos/.julia/conda/3/lib/python3.7/site-packages/notebook/services/sessions/sessionmanager.py", line 79, in create_session
kernel_id = yield self.start_kernel_for_session(session_id, path, name, type, kernel_name)
File "/Users/soderhos/.julia/conda/3/lib/python3.7/site-packages/tornado/gen.py", line 1133, in run
value = future.result()
File "/Users/soderhos/.julia/conda/3/lib/python3.7/site-packages/tornado/gen.py", line 1141, in run
yielded = self.gen.throw(*exc_info)
File "/Users/soderhos/.julia/conda/3/lib/python3.7/site-packages/notebook/services/sessions/sessionmanager.py", line 92, in start_kernel_for_session
self.kernel_manager.start_kernel(path=kernel_path, kernel_name=kernel_name)
File "/Users/soderhos/.julia/conda/3/lib/python3.7/site-packages/tornado/gen.py", line 1133, in run
value = future.result()
File "/Users/soderhos/.julia/conda/3/lib/python3.7/site-packages/tornado/gen.py", line 326, in wrapper
yielded = next(result)
File "/Users/soderhos/.julia/conda/3/lib/python3.7/site-packages/notebook/services/kernels/kernelmanager.py", line 160, in start_kernel
super(MappingKernelManager, self).start_kernel(**kwargs)
File "/Users/soderhos/.julia/conda/3/lib/python3.7/site-packages/jupyter_client/multikernelmanager.py", line 110, in start_kernel
km.start_kernel(**kwargs)
File "/Users/soderhos/.julia/conda/3/lib/python3.7/site-packages/jupyter_client/manager.py", line 259, in start_kernel
**kw)
File "/Users/soderhos/.julia/conda/3/lib/python3.7/site-packages/jupyter_client/manager.py", line 204, in _launch_kernel
return launch_kernel(kernel_cmd, **kw)
File "/Users/soderhos/.julia/conda/3/lib/python3.7/site-packages/jupyter_client/launcher.py", line 128, in launch_kernel
proc = Popen(cmd, **kwargs)
File "/Users/soderhos/.julia/conda/3/lib/python3.7/subprocess.py", line 769, in __init__
restore_signals, start_new_session)
File "/Users/soderhos/.julia/conda/3/lib/python3.7/subprocess.py", line 1516, in _execute_child
raise child_exception_type(errno_num, err_msg, err_filename)
FileNotFoundError: [Errno 2] No such file or directory: '/Applications/Julia-1.0.app/Contents/Resources/julia/bin/julia': '/Applications/Julia-1.0.app/Contents/Resources/julia/bin/julia'
It seems like Jupyter is trying to locate Julia 1.0, even though the notebook session was started with Julia 1.1. Why would this happen?
When you installed the new version of Julia and deleted the old one, you have to launch the Julia command line and run build IJulia at the package prompt (to tell Jupyter where to find Julia).
See https://github.com/JuliaLang/IJulia.jl#updating-julia-and-ijulia
| gharchive/issue | 2019-03-18T17:31:54 | 2025-04-01T06:37:07.014989 | {
"authors": [
"TheSodesa",
"stevengj"
],
"repo": "JuliaLang/IJulia.jl",
"url": "https://github.com/JuliaLang/IJulia.jl/issues/824",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1062575020 | Add alternative installation instructions
Linux & Mac: I couldn't find specific instructions, but I'm guessing they exist, as it sounds like Juliaup should work on these platforms.
Windows: For some, installing from the Store may not be an option. How should these people install Juliaup?
It would be nice if the README listed these installation instructions.
The only option that is ready at this point is the Windows Store option. We are getting closer with Linux and MacOS support, but at this point it is not ready for general consumption or feedback, so for now we should wait with install instructions in the README until things are ready :) But then, yes, agreed, we need to add them! I'll keep this issue open to track that.
Related to this: I understand that juliaup should update itself. Does this, too, rely on a connection with the Windows Store? Getting the Windows Store to work is very hard, so after the first install, it would be good to not rely on juliaup communicating with the store.
Yes, the version from the Windows Store gets updated by the Windows Store. It solves a lot of problems as doing background updates is not trivial...
My current plan is to next figure out automatic self-update for Linux and Mac, and then come back to Windows and see whether we can improve the situation for folks where the Windows Store is in some form blocked.
[…] doing background updates is not trivial...
What exactly do you mean by 'background updates'?
Off the top of my head, Powershell (pwsh) warns that an upgrade is available on start, and pip likewise warns at least when used to install a package (and maybe in more cases). Both require the user to initiate the upgrade – typically through manual download in the case of pwsh, but a simple one-liner in the case of pip (and rustup, and stack, and…). I'm fine with the latter strategy.
Does the new-ish Windows package manager, winget, change the picture at all, or is that just a CLI for the Store app?
[…] or is that just a CLI for the Store app?
It is not, but it can function as one:
PS ~ > winget search julia
Name Id Version Source
-----------------------------------------------------
Julia 9NJNWW8PVKMN Unknown msstore
Julian Date Selector 9NSGP4VDNW0R Unknown msstore
Julia Julialang.Julia 1.6.2 winget
I tried to circumvent the Store by using winget, but it doesn't work: installing 9NJNWW8PVKMN still requires logging in.
I tried to circumvent the Store by using winget, but it doesn't work: installing 9NJNWW8PVKMN still requires logging in.
Blocked by https://github.com/microsoft/winget-cli/issues/1585#issuecomment-974509235.
| gharchive/issue | 2021-11-24T15:36:44 | 2025-04-01T06:37:07.137222 | {
"authors": [
"DNF2",
"ItzLevvie",
"MatthijsBlom",
"davidanthoff"
],
"repo": "JuliaLang/juliaup",
"url": "https://github.com/JuliaLang/juliaup/issues/175",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1044719950 | Ambiguity in ldiv! since ArrayLayouts@0.7.7
The following worked on 0.7.6 but resulted in an ambiguity error on 0.7.7:
using BlockArrays
Q = PseudoBlockMatrix(rand(5,5),[1,4],[3,2])
q = PseudoBlockVector(rand(5),[1,4])
u = Q \ q
Output:
ERROR: LoadError: MethodError: ldiv!(::LinearAlgebra.LU{Float64, PseudoBlockMatrix{Float64, Matrix{Float64}, Tuple{BlockedUnitRange{Vector{Int64}}, BlockedUnitRange{Vector{Int64}}}}}, ::PseudoBlockVector{Float64, Vector{Float64}, Tuple{BlockedUnitRange{Vector{Int64}}}}) is ambiguous. Candidates:
ldiv!(A::LinearAlgebra.Factorization, x::ArrayLayouts.LayoutVector{T} where T) in ArrayLayouts at C:\Users\krcools\.julia\packages\ArrayLayouts\O51Gr\src\ldiv.jl:135
ldiv!(L::LinearAlgebra.LU{var"#s798", var"#s799"} where {var"#s798", var"#s799"<:(ArrayLayouts.LayoutMatrix{T} where T)}, B) in ArrayLayouts at C:\Users\krcools\.julia\packages\ArrayLayouts\O51Gr\src\factorizations.jl:403
Possible fix, define
ldiv!(::LinearAlgebra.LU{var"#s798", var"#s799"} where {var"#s798", var"#s799"<:(ArrayLayouts.LayoutMatrix{T} where T)}, ::ArrayLayouts.LayoutVector{T} where T)
Stacktrace:
[1] ldiv!(Y::PseudoBlockVector{Float64, Vector{Float64}, Tuple{BlockedUnitRange{Vector{Int64}}}}, A::LinearAlgebra.LU{Float64, PseudoBlockMatrix{Float64, Matrix{Float64}, Tuple{BlockedUnitRange{Vector{Int64}}, BlockedUnitRange{Vector{Int64}}}}}, B::PseudoBlockVector{Float64, Vector{Float64}, Tuple{BlockedUnitRange{Vector{Int64}}}})
@ LinearAlgebra C:\buildbot\worker\package_win64\build\usr\share\julia\stdlib\v1.6\LinearAlgebra\src\factorization.jl:142
[2] _ldiv!
@ C:\Users\krcools\.julia\packages\ArrayLayouts\O51Gr\src\ldiv.jl:82 [inlined]
[3] copyto!
@ C:\Users\krcools\.julia\packages\ArrayLayouts\O51Gr\src\ldiv.jl:98 [inlined]
[4] ldiv!
@ C:\Users\krcools\.julia\packages\ArrayLayouts\O51Gr\src\ldiv.jl:92 [inlined]
[5] _ldiv!
@ C:\Users\krcools\.julia\packages\ArrayLayouts\O51Gr\src\ldiv.jl:81 [inlined]
[6] copyto!
@ C:\Users\krcools\.julia\packages\ArrayLayouts\O51Gr\src\ldiv.jl:98 [inlined]
[7] copy
@ C:\Users\krcools\.julia\packages\ArrayLayouts\O51Gr\src\ldiv.jl:21 [inlined]
[8] materialize
@ C:\Users\krcools\.julia\packages\ArrayLayouts\O51Gr\src\ldiv.jl:22 [inlined]
[9] ldiv
@ C:\Users\krcools\.julia\packages\ArrayLayouts\O51Gr\src\ldiv.jl:86 [inlined]
[10] \(A::PseudoBlockMatrix{Float64, Matrix{Float64}, Tuple{BlockedUnitRange{Vector{Int64}}, BlockedUnitRange{Vector{Int64}}}}, x::PseudoBlockVector{Float64, Vector{Float64}, Tuple{BlockedUnitRange{Vector{Int64}}}})
@ ArrayLayouts C:\Users\krcools\.julia\packages\ArrayLayouts\O51Gr\src\ldiv.jl:168
Environment:
Project SL_2021_ThinSheetMultiTraceCalderon v0.1.0
Status `C:\Users\krcools\.julia\dev\SL_2021_ThinSheetMultiTraceCalderon\Project.toml`
[4c555306] ArrayLayouts v0.7.7
[8e7c35d0] BlockArrays v0.16.9
This is related to my request to have PBMs factorize over LU isn't it? Sorry not to catch this when I tried the feature branch. It did not give me this error message at the time.
Are you able to make a PR? Just add in factorizations.jl an overload ldiv!(::LU{<:Any,<:LayoutMatrix}, ::LayoutVector).
I'm quite busy right now so don't have tim
Sure, I noticed that you started a PR #82 yourself with the exact same modifications though.
oh, nevermind, I just need to fix that PR....
I just checked, it works fine now.
| gharchive/issue | 2021-11-04T12:38:38 | 2025-04-01T06:37:07.141456 | {
"authors": [
"dkarrasch",
"dlfivefifty",
"krcools"
],
"repo": "JuliaLinearAlgebra/ArrayLayouts.jl",
"url": "https://github.com/JuliaLinearAlgebra/ArrayLayouts.jl/issues/83",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
306023557 | Improve documentation
I had a misunderstanding of plan_* functions.
Main reason of this was from the term "inverse" on documentation
It says
You can compute the inverse-transform plan by inv(P) and apply the inverse plan with P \ Â (the inverse plan is cached and reused for subsequent calls to inv or ), and apply the inverse plan to a pre-allocated output array A with ldiv!(A, P, Â).
The purpose of this explanation was
you have FORWARD transform plan
using mul! : FORWARD transform
using ldiv! : BACKWARD transform
plan_fft is for FORWARD transform so it doesn't matter.
However, other plan_* documentation says basically "Same as plan_fft", so if you apply this structure to BACKWARD transform, documentation have two meaning
you have BACKWARD transform plan
using mul! : BACKWARD transform
using ldiv! : FORWARD transform
or
you have BACKWARD transform plan
using ldiv : BACKWARD transform
because the term "INVERSE" and "BACKWARD" are often used interchangeably. It was hard to distinguish without investigating source code. I know there is a "-" between inverse and transform, but it is really confusing.
Would we have more clear explanation of this?
"forward" and "backward" just refer to the sign of the exponent in the Fourier transform. To get the inverse of a given transform, you have to flip the sign of the exponent (going from forward to backward or vice versa) and also scale by 1/n where n is the length of the transform.
The inverse of a plan (as computed by inv or applied by ldiv!) is actually this inverse: e.g. the inverse of a forwards transform is a backwards transform scaled by 1/n. The inverse of an inverse plan is the original plan.
| gharchive/issue | 2018-03-16T17:57:28 | 2025-04-01T06:37:07.148008 | {
"authors": [
"appleparan",
"stevengj"
],
"repo": "JuliaMath/AbstractFFTs.jl",
"url": "https://github.com/JuliaMath/AbstractFFTs.jl/issues/13",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2379029207 | GitHub: get API url and server URL from env
I encountered difficulties trying to use CompatHelper.jl on an Enterprise server, but it turns out I just needed to set the API url and the hostname correctly.
This pull request updates auto_detect_ci_service to automatically handle non-GitHub.com GitHub servers.
I am not sure how to add tests for this.
@DilumAluthge I see you are probably the most active here. Can you take a look at this and advise on tests?
So, if we add this functionality into CompatHelper itself, it'll be a little difficult for us to test it.
Instead, could you modify your CompatHelper.yml file?
For example, if you look at the recommended CompatHelper.yml file, part of it looks like this:
https://github.com/JuliaRegistries/CompatHelper.jl/blob/97e9dcdde383ea5bea89d51de221698958b47630/.github/workflows/CompatHelper.yml#L37-L41
You could instead modify your CompatHelper.yml file to look like this:
- name: "Run CompatHelper"
run: |
import CompatHelper
my_ci_cfg = CompatHelper.GitHubActions(;
username = "...",
email = "...",
api_hostname = "...",
clone_hostname = "...",
)
CompatHelper.main(ENV, my_ci_cfg)
shell: julia --color=yes {0}
Where you'd fill in the relevant values in the GitHubActions(; ...) constructor.
That way, you can specify the exact values you need, without us needed to modify the source code of CompatHelper.jl (and thus needing to figure out a way to test it).
| gharchive/pull-request | 2024-06-27T20:21:37 | 2025-04-01T06:37:07.234008 | {
"authors": [
"DilumAluthge",
"gbruer15"
],
"repo": "JuliaRegistries/CompatHelper.jl",
"url": "https://github.com/JuliaRegistries/CompatHelper.jl/pull/498",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
665809612 | ERROR: LoadError: UndefVarError: include not defined
hi
include is part of the Julia language. Thus, I do not understand the reason for this error. thanks for your time
2020-07-26T16:03:27.4974582Z VegaLite
2020-07-26T16:03:27.9252965Z ERROR: LoadError: UndefVarError: include not defined
2020-07-26T16:03:28.5092161Z Stacktrace:
2020-07-26T16:03:28.5093738Z [1] top-level scope at /tmp/jl_dglOg0/packages/InvariantCausalPrediction/dnTFi/src/InvariantCausalPrediction.jl:39
2020-07-26T16:03:28.6607536Z [2] include(::Module, ::String) at ./Base.jl:377
2020-07-26T16:03:28.6608348Z [3] top-level scope at none:2
2020-07-26T16:03:28.6637636Z [4] eval at ./boot.jl:331 [inlined]
2020-07-26T16:03:28.6638488Z [5] eval(::Expr) at ./client.jl:449
2020-07-26T16:03:28.6639540Z [6] top-level scope at ./none:3
2020-07-26T16:03:28.6640050Z in expression starting at /tmp/jl_dglOg0/packages/InvariantCausalPrediction/dnTFi/src/InvariantCausalPrediction.jl:39
2020-07-26T16:03:29.9026532Z ERROR: Failed to precompile InvariantCausalPrediction [5fe40f08-422b-4ec7-90aa-ba60e31ac74e] to /tmp/jl_dglOg0/compiled/v1.4/InvariantCausalPrediction/3Q6yc_c0C9F.ji.
2020-07-26T16:03:30.0841146Z Stacktrace:
It sounds like you have having some issues getting your package to work.
It doesn't sound like these issues are specific to AutoMerge or the registration process.
Please ask questions on the Julia Discourse forum.
hi,
I have asked the question in the forum.
https://discourse.julialang.org/t/automerge-decision-new-package-error-loaderror-undefvarerror-include-not-defined/43730
thanks
| gharchive/issue | 2020-07-26T16:06:01 | 2025-04-01T06:37:07.239171 | {
"authors": [
"DilumAluthge",
"drcxcruz"
],
"repo": "JuliaRegistries/General",
"url": "https://github.com/JuliaRegistries/General/issues/18468",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
917119614 | New version: SPIRV_LLVM_Translator_jll v9.0.0+4
Autogenerated JLL package registration
Registering JLL package SPIRV_LLVM_Translator_jll.jl
Repository: https://github.com/JuliaBinaryWrappers/SPIRV_LLVM_Translator_jll.jl
Version: v9.0.0+4
Commit: 72ec82700b502b6b770a9eb04fefad091a1d82b0
Revision on Yggdrasil: https://github.com/JuliaPackaging/Yggdrasil/commit/55777a3d0eecb57e7f1362cef33da805c20c9aa2
Created by: @maleadt
Same as https://github.com/JuliaRegistries/General/pull/38546.
| gharchive/pull-request | 2021-06-10T09:23:04 | 2025-04-01T06:37:07.242345 | {
"authors": [
"jlbuild",
"maleadt"
],
"repo": "JuliaRegistries/General",
"url": "https://github.com/JuliaRegistries/General/pull/38547",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1062713220 | New version: NDTensors v0.1.33
Registering package: NDTensors
Repository: https://github.com/ITensor/ITensors.jl
Created by: @mtfishman
Version: v0.1.33
Commit: f029107358903335a7e0242fc8fe74f12f35b8d2
Reviewed by: @mtfishman
Reference: https://github.com/ITensor/ITensors.jl/commit/f029107358903335a7e0242fc8fe74f12f35b8d2#commitcomment-60824698
Description: A Julia library for efficient tensor computations and tensor network calculations
@mtfishman please see the message above. Once you fix the automerge issues you can register again the new revision without changing the version number and this pull request will be automatically updated.
[noblock]
| gharchive/pull-request | 2021-11-24T17:51:27 | 2025-04-01T06:37:07.245974 | {
"authors": [
"JuliaRegistrator",
"giordano"
],
"repo": "JuliaRegistries/General",
"url": "https://github.com/JuliaRegistries/General/pull/49332",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1385185519 | New version: Torch_jll v1.10.2+0
Autogenerated JLL package registration
Registering JLL package Torch_jll.jl
Repository: https://github.com/JuliaBinaryWrappers/Torch_jll.jl
Version: v1.10.2+0
Commit: 2bafb001677b1eb77ac26303f5d958319f302c45
Revision on Yggdrasil: https://github.com/JuliaPackaging/Yggdrasil/commit/dc06b9903eba8408087120ed962ed7e5d3716a31
Created by: @Wimmerer
@stemann Now the problem is MKL:
ERROR: InitError: could not load library "/tmp/jl_mho9f5/artifacts/929e07419d06d190327bd30982e5b0b510a49664/lib/libtorch.so"
libmkl_intel_lp64.so.2: cannot open shared object file: No such file or directory
Can we avoid MKL entirely? It's a mess. We can't link to the libmkl_intel libraries in a sane way, enjoy reading https://github.com/JuliaPackaging/Yggdrasil/pull/1075 if you want to learn more.
[noblock]
@stemann Now the problem is MKL:
ERROR: InitError: could not load library "/tmp/jl_mho9f5/artifacts/929e07419d06d190327bd30982e5b0b510a49664/lib/libtorch.so"
libmkl_intel_lp64.so.2: cannot open shared object file: No such file or directory
Can we avoid MKL entirely? It's a mess. We can't link to the libmkl_intel libraries in a sane way, enjoy reading JuliaPackaging/Yggdrasil#1075 if you want to learn more.
[noblock]
OK, I see.
I will try to exclude building with MKL: https://github.com/JuliaPackaging/Yggdrasil/pull/5583
[noblock]
| gharchive/pull-request | 2022-09-25T22:32:38 | 2025-04-01T06:37:07.251500 | {
"authors": [
"giordano",
"jlbuild",
"stemann"
],
"repo": "JuliaRegistries/General",
"url": "https://github.com/JuliaRegistries/General/pull/68946",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2099013920 | New package: GeoStatsFunctions v0.1.0
Registering package: GeoStatsFunctions
Repository: https://github.com/JuliaEarth/GeoStatsFunctions.jl
Created by: @juliohm
Version: v0.1.0
Commit: a6a2c8dad93770da666d93fb8f20ae0671bd4027
Reviewed by: @juliohm
Reference: https://github.com/JuliaEarth/GeoStatsFunctions.jl/commit/a6a2c8dad93770da666d93fb8f20ae0671bd4027#commitcomment-137741950
Description: Geostatistical functions for the GeoStats.jl framework
Even though this seems to be part of a larger framework, I'd still ask that you add some documentation before registering. This could be a README that's a little more expansive, with a list of functions and a paragraph or two on how these functions fit into GeoStats, and/or a minimal Documenter-based page with all the docstrings. Or, a link to the specific part of the GeoStats documentation where these functions are documented.
Even though this seems to be part of a larger framework, I'd still ask that you add some documentation before registering. This could be a README that's a little more expansive, with a list of functions and a paragraph or two on how these functions fit into GeoStats, and/or a minimal Documenter-based page with all the docstrings. Or, a link to the specific part of the GeoStats documentation where these functions are documented.
Sorry @goerz, but I disagree. We are documenting the framework in a single place to avoid outdated README's in submodules of the project. That is why the README explicitly mentions the official GeoStats.jl docs and community channel. These modules are not intended for end-users, they are intended for developers of the framework only.
[noblock]
[noblock] They gotta be documented somewhere. Just picking something at random, ballsearch.jl has a docstring for BallSearchAccum, but if I search https://juliaearth.github.io/GeoStatsDocs/stable/search.html?q=BallSearchAccum I'm not seeing that docstring.
I have a similar situation with my own packages, where there's a *Base package that's not for public consumption, but even that can have a minimal listing of docstrings: https://juliaquantumcontrol.github.io/QuantumControlBase.jl/dev/
You definitely don't have to put anything in the README that could become outdated, but a few words about the functions contained in this package and where to find their documentation would be helpful.
Thank you for the suggestions, we are constantly improving our documentation, it is just not the right place in our viewpoint. Many docstrings are not intended for end-users either, and we are constantly evolving internals to accommodate documented functionality in the main documentation website.
Appreciate if you can add a [noblock] to your first comment to avoid blocking the auto-merge by the bot.
it is just not the right place
Then what is the right place?
It doesn't seem like a high bar to require that every registered package, even an auxiliary package, has some form – any form – of documentation.
Appreciate if you can add a [noblock] to your first comment
Sorry, no, the requirement that every package needs some minimal form of documentation is a line I'm willing to hold. So if you insist on not having any documentation, you would have to get a registry maintainer to override my veto. Or, preferably, just add a documentation stub like the example I gave before. I don't think it's something that'll take more than 20 minutes or so to set up, and I'd be happy to unblock then.
If this package is indeed 100% internal functions, it probably shouldn't be a package, but a submodule of GeoStats.
Then what is the right place?
I already explained that the documentation of this module lives inside the main documentation of the project, and that internal functions not intended for end-users are not present on purpose.
Sorry, no, the requirement that every package needs some minimal form of documentation is a line I'm willing to hold. So if you insist on not having any documentation, you would have to get a registry maintainer to override my veto.
Do you really feel that this requirement is helping with the quality of the general registry? We have so many other modules already registered, all pointing to the main docs of the project:
https://github.com/JuliaEarth/GeoStatsBase.jl
https://github.com/JuliaEarth/GeoTables.jl
https://github.com/JuliaEarth/GeoStatsTransforms.jl
The requirement of documentation is important, but you are forcing us to write the documentation in a specific place of your preference. We prefer to point end-users to a central well-maintained documentation, and yes, we believe that it still makes sense to develop this as a separate package in a separate repository where people can contribute specific PRs.
Besides, it is useless to enforce contributors of packages to add a README at registration time. They can always undo the README later on with additional commits. You have to give contributors the freedom to write good documentation where they feel is best for the community.
[noblock]
[noblock] Yeah, looking at the project organization, I can see the pattern. So I think it's okay. The overall documentation of the organization is far better than many other new registrations, so it seems unfair to block it.
I'd still say if I was a user of GeoStats, and I'd ever have to explicitly import GeoStatsFunctions (or Meshes, or any of the other listed package), I'd find it useful to have a complete reference API documentation for that sub-package. But yeah, that's up to you. It's certainly not a prerequisite for registration that a package has perfect documentation according to my tastes ;-)
[noblock] Thank you, it is a lot of work to keep these packages synced in
both source and docs. That is the most productive way we found. GeoStats.jl
is an umbrella package that loads the full stack of packages using
Reexport.jl it hosts the docs and is the goto place for end-users.
Em qui., 25 de jan. de 2024 11:09, Michael Goerz @.***>
escreveu:
[noblock] Yeah, looking at the project organization
https://juliaearth.github.io/GeoStatsDocs/dev/index.html#Project-organization,
I can see the pattern. So I think it's okay. The overall documentation of
the organization is far better than many other new registrations, so it
seems unfair to block it.
I'd still say if I was a user of GeoStats, and I'd ever have to explicitly
import GeoStatsFunctions (or Meshes, or any of the other listed package),
I'd find it useful to have a complete reference API documentation for that
sub-package. But yeah, that's up to you. It's certainly not a prerequisite
for registration that a package has perfect documentation according to
my tastes ;-)
—
Reply to this email directly, view it on GitHub
https://github.com/JuliaRegistries/General/pull/99463#issuecomment-1910290229,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/AAZQW3MNU65EC4VPHFLYQD3YQJRQ7AVCNFSM6AAAAABCJLTJFCVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMYTSMJQGI4TAMRSHE
.
You are receiving this because you were mentioned.Message ID:
@.***>
| gharchive/pull-request | 2024-01-24T20:15:49 | 2025-04-01T06:37:07.271008 | {
"authors": [
"JuliaRegistrator",
"goerz",
"juliohm"
],
"repo": "JuliaRegistries/General",
"url": "https://github.com/JuliaRegistries/General/pull/99463",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1884273180 | Use Enzyme or DifferentiationInterface for autodiff
Current Status
The AD part of the package has not undergone any major overhaul since I first implemented it around the start of the project. Back then I relied on Flux/Zygote, because the package was tailored to Flux models anyway (at the time), I was entirely new to AD and Julia; and I could use Zygote to differentiate through structs, like so:
"""
∂ℓ(
generator::AbstractGradientBasedGenerator,
ce::AbstractCounterfactualExplanation,
)
The default method to compute the gradient of the loss function at the current counterfactual state for gradient-based generators.
It assumes that `Zygote.jl` has gradient access.
"""
function ∂ℓ(
generator::AbstractGradientBasedGenerator, ce::AbstractCounterfactualExplanation
)
return Flux.gradient(ce -> ℓ(generator, ce), ce)[1][:counterfactual_state]
end
Pain Points
The current implementation is less than ideal for various reasons:
Zygote cannot handle nested AD, which is necessary for some counterfactual generators (see #376).
Gradients are still taken implicitly, which is not in line with where the broader ecosystem is headed, I believe.
The previous point also makes it difficult to implement forward-over-reverse to solve the nested AD issue.
The AD implementation has never been optimized for performance, so I guess there's a lot of room for improvement here.
To Do
[ ] Double-check https://gdalle.github.io/JuliaCon2024-AutoDiff/#/title-slide
[ ] Try out DifferentiationInterface.jl
This seems like a good idea (he said, completely unbiased). Want a hand with that @pat-alt ?
@gdalle fancy seeing you here 😄
That would be amazing, of course, if you could help out, but only if it's not too much trouble for you. I want to look at this soon, but I might look at #495 first, because I think I may need this for a research project I'm currently working on.
I've updated the description a little bit. If you have any pointers, I'd much appreciate if you could share them here.
Thanks!
| gharchive/issue | 2023-09-06T15:32:53 | 2025-04-01T06:37:07.365891 | {
"authors": [
"gdalle",
"pat-alt"
],
"repo": "JuliaTrustworthyAI/CounterfactualExplanations.jl",
"url": "https://github.com/JuliaTrustworthyAI/CounterfactualExplanations.jl/issues/300",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
365832651 | Fix py2 urlopen
urlopen() was used as a context manager on python 2 too.
This happened when resolving remote refs and not using requests, for instance when resolving file:/// refs.
This solves the same problem as #439, but better ;).
Hi! Thanks a lot for this, it'll close #439.
Left a comment on how to shorten up the test -- it also looks like the test might be failing on Windows.
I've applied the comments.
In order to fix the added test on windows I need to find a windows host so it will take some days.
| gharchive/pull-request | 2018-10-02T10:21:38 | 2025-04-01T06:37:07.368205 | {
"authors": [
"Julian",
"gaetano-guerriero"
],
"repo": "Julian/jsonschema",
"url": "https://github.com/Julian/jsonschema/pull/472",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
904391180 | Reset the cursor to the top of the infoview when updating.
Closes: #27
Bit hairy getting some testing in place, but @gebner care to review this does what you expect?
Hmm, unfortunately this does not work as well as I'd hoped. The cursor jumps back to the top as soon as you switch back to the window for the lean file. Maybe it just takes getting used to, but it is very counter-intuitive that switching windows moves the cursor. The infoview moves to the top while I'm still reading it.
-- long trace message for testing
#eval let y := (List.range 100).map fun x => dbg_trace x; x + 1 dbg_trace "this is important" y.length
Maybe a better heuristic is to move the cursor to the top only when you move to a new line.
The behavior of scrolloff=10 would also be okay, but that doesn't seem to work for automatic updates.
Aha, ok, I think I follow, you want it to move just if the new contents are shorter than the window length? Or do you want it to move even if they're long but have changed? (Whereas now it moves always even if nothing whatsoever changes in the infoview contents, just CursorHold fires again and it repopulates)
I think I can imagine both of those other behaviors being the expected one in different scenarios, right?
Aha, ok, I think I follow, you want it to move just if there are some new infoview contents are shorter than the window length?
My main issue with the current (= main branch) behavior is that it is easy to miss errors: you scroll down a long trace message.
Then you move a line up, the infoview is empty, and now you think that everything is fine or that lean is still processing the file.
So for me it's only important to scroll up when the content changes and becomes shorter.
Or do you want it to move even if they're long but have changed?
Ideally not. When I look at the middle of a long trace message and it's updated to similarly long message, then chances are I still want to see the same part, and see if anything's changed.
I think I can imagine both of those other behaviors being the expected one in different scenarios, right?
Good question, and I can't really tell. That's why I'd like to keep the automatic movements to a minimum. Scrolling the infoview has never bothered me in vscode, but it's also never empty.
OK, I think I convinced myself this is actually likely a neovim bug, which I filed as neovim/neovim#14663.
I found a way to hack around it for now, which I pushed to this PR (it actually I think has nothing to do with the cursor position but I'm not 100% and didn't go off and read the neovim C source yet).
Feel free to have another look, hopefully this is closer to what you expect, though perhaps we could be even more sticky on restoring line position. But in trying your long trace example and adding some other lines around it, I suspect this is an improvement.
This is exactly what I wanted! Thanks!
| gharchive/pull-request | 2021-05-27T23:43:28 | 2025-04-01T06:37:07.374051 | {
"authors": [
"Julian",
"gebner"
],
"repo": "Julian/lean.nvim",
"url": "https://github.com/Julian/lean.nvim/pull/29",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
198018091 | scality service doesn't have the server data
when try to run the s3 scality service to install s3server using this blueprint
g8client__main:
url: 'du-conv-2.demo.greenitglobe.com'
login: 's3user'
password: '123456789'
account: 's3_acc'
vdc__scality22:
g8client: 'main'
location: 'du-conv-2'
disk.ovc__disk1:
size: 1000
s3__s3vm:
vdc: 'scality22'
disk:
- 'disk1'
hostprefix: 's3app22'
scenario:
1- create new repo from cockpit portal
2- execute the bp from cockpit portal
3- the run status will be OK and all the steps are OK too
4- however if I go to the scality service to get the sercet/access keys, it will be emtpy json
root@vm-14:/optvar/cockpit_repos/s3server4/services/vdcfarm!auto_82/vdc!scality24/node!s3vm/os!s3vm/node!app/os!app/scality!app# cat data.json
{
"domain":"",
"keyAccess":"",
"keySecret":"",
"os":"app",
"storageData":"\/data\/data",
"storageMeta":"\/data\/meta"
5- now from the cockpit machine, do ays destroy & ays blueprint & ays install
now the data.json will be OK and include all the needed info
root@vm-14:/optvar/cockpit_repos/s3server4/services/vdcfarm!auto_82/vdc!scality24/node!s3vm/os!s3vm/node!app/os!app/scality!app# cat data.json
{
"domain":"s3app21-3232242740.gigapps.io",
"keyAccess":"W7X2BxGxWXxO",
"keySecret":"YBbEjNdeQrqp",
"os":"app",
"storageData":"\/data\/data",
"storageMeta":"\/data\/meta"
Doing the execute action from the portal only units the services. These values are set during the install.
To have these values set, you should've also run the install action from the portal.
This is designed behavior.
| gharchive/issue | 2016-12-29T14:46:04 | 2025-04-01T06:37:07.405795 | {
"authors": [
"ramezsaeed",
"rkhamis"
],
"repo": "Jumpscale/jscockpit",
"url": "https://github.com/Jumpscale/jscockpit/issues/326",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
192003006 | cuisine executor? does not show output while going
When I start, I see no output till command is done,
this is not ok
need to see while going.
Checked on Ubuntu: works
https://gist.github.com/xmonader/ac7398c0edeae39fd16364c410e660da
Needs to be validated on osx
verfied on osx
| gharchive/issue | 2016-11-28T13:13:55 | 2025-04-01T06:37:07.408221 | {
"authors": [
"abdulrahmantkhalifa",
"despiegk",
"xmonader"
],
"repo": "Jumpscale/jumpscale_core8",
"url": "https://github.com/Jumpscale/jumpscale_core8/issues/588",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
656931289 | Download YouTube playlist videos in a different way
Please tick all applicable boxes.
[X] I am using Python 3.5.3 or higher (run python --version on the command line)
[X] I have followed the official guides to install the bot for my system
[X] I have updated my dependencies to the latest version using the appropriate update script
Which version are you using?
[ ] The latest master version (release-260419)
[X] The latest review version
What type of issue are you creating?
[X] Bug
[ ] Feature request
[ ] Question
Description of issue
When the bot is commanded to play a YouTube playlist, it downloads all the videos of the playlist all at once... which in a large playlist makes YouTube IP block the VPS the bot is running on.
This behavior doesn’t make sense!
Why doesn’t the bot instead only download the first song, plays it, when it’s done it downloads the second song, plays it, etc...
Much better than downloading all the songs of a playlist at once! And will for sure prevent YouTube’s annoying IP block.
Steps to reproduce
Make it play any YouTube playlist with many videos (example: https://www.youtube.com/playlist?list=PL4o29bINVT4EG_y-k5jGoOu3-Am8Nvi10)
It will try to download all of the videos all at once and will get IP blocked by Youtube in no time.
Log file
Please attach your MusicBot log file (located at logs/musicbot.log) to this issue. You can do so by dragging and dropping the file here. If you do not include your log file, you WILL be asked to provide one.
[30.515432358] 2020-07-14 21:31:48,849 - INFO - launcher: Checking for Python 3.5+
[30.718088150] 2020-07-14 21:31:48,849 - INFO - launcher: Checking console encoding
[30.900955200] 2020-07-14 21:31:48,849 - INFO - launcher: Ensuring we're in the right environment
[448.570251465] 2020-07-14 21:31:49,267 - INFO - launcher: Required checks passed.
[449.181795120] 2020-07-14 21:31:49,268 - INFO - launcher: Optional checks passed.
[449.469566345] 2020-07-14 21:31:49,268 - INFO - launcher: Moving old musicbot log
######################### PRE-RUN SANITY CHECKS PASSED #########################
[1.0954689979553223] 2020-07-14 21:31:49,914 - WARNING - musicbot.config | In config.py::MainThread(140250957956928), line 123 in run_checks: i18n file does not exist. Trying to fallback to config/i18n/en.json.
[1.0958342552185059] 2020-07-14 21:31:49,914 - INFO - musicbot.config | In config.py::MainThread(140250957956928), line 134 in run_checks: Using i18n: config/i18n/en.json
[1.0980954170227051] 2020-07-14 21:31:49,917 - DEBUG - musicbot.bot | In bot.py::MainThread(140250957956928), line 239 in _setup_logging: Set logging level to INFO
[1.1010162830352783] 2020-07-14 21:31:49,920 - DEBUG - musicbot.json | In json.py::MainThread(140250957956928), line 8 in __init__: Init JSON obj with config/i18n/en.json
[1.1118435859680176] 2020-07-14 21:31:49,930 - INFO - musicbot.bot | In bot.py::MainThread(140250957956928), line 90 in __init__: Starting MusicBot release-260819-72-g2350384
[1.1124629974365234] 2020-07-14 21:31:49,931 - INFO - musicbot.bot | In bot.py::MainThread(140250957956928), line 96 in __init__: Loaded autoplaylist with 2531 entries
[1.1938691139221191] 2020-07-14 21:31:50,012 - DEBUG - musicbot.spotify | In spotify.py::MainThread(140250957956928), line 81 in get_token: Created a new access token: {'access_token': 'BQBYcGaTU42K96Gnt82HevLsnJEwJOVWAES6nNenj54CuwRTc6jCZH0ua598Lsd790dqgTGSY1IefZahBno', 'token_type': 'Bearer', 'expires_in': 3600, 'scope': '', 'expires_at': 1594765910}
[1.1942601203918457] 2020-07-14 21:31:50,013 - INFO - musicbot.bot | In bot.py::MainThread(140250957956928), line 121 in __init__: Authenticated with Spotify successfully using client ID and secret.
[3.6479089260101318] 2020-07-14 21:31:52,466 - DEBUG - musicbot.bot | In bot.py::MainThread(140250957956928), line 965 in on_ready: Connection established, ready to go.
[3.6483943462371826] 2020-07-14 21:31:52,467 - DEBUG - musicbot.bot | In bot.py::MainThread(140250957956928), line 356 in _cache_app_info: Caching app info
[3.7539911270141602] 2020-07-14 21:31:52,573 - DEBUG - musicbot.bot | In bot.py::MainThread(140250957956928), line 787 in _scheck_ensure_env: Ensuring data folders exist
[3.7552502155303955] 2020-07-14 21:31:52,574 - DEBUG - musicbot.bot | In bot.py::MainThread(140250957956928), line 803 in _scheck_server_permissions: Checking server permissions
[3.7554564476013184] 2020-07-14 21:31:52,574 - DEBUG - musicbot.bot | In bot.py::MainThread(140250957956928), line 807 in _scheck_autoplaylist: Auditing autoplaylist
[3.7556393146514893] 2020-07-14 21:31:52,574 - DEBUG - musicbot.bot | In bot.py::MainThread(140250957956928), line 811 in _scheck_configs: Validating config
[3.7558276653289795] 2020-07-14 21:31:52,574 - DEBUG - musicbot.config | In config.py::MainThread(140250957956928), line 237 in async_validate: Validating options...
[3.7559976577758789] 2020-07-14 21:31:52,575 - DEBUG - musicbot.config | In config.py::MainThread(140250957956928), line 251 in async_validate: Acquired owner id via API
[3.7561569213867188] 2020-07-14 21:31:52,575 - DEBUG - musicbot.bot | In bot.py::MainThread(140250957956928), line 814 in _scheck_configs: Validating permissions config
[3.7563233375549316] 2020-07-14 21:31:52,575 - DEBUG - musicbot.permissions | In permissions.py::MainThread(140250957956928), line 94 in async_validate: Validating permissions...
[3.7565703392028809] 2020-07-14 21:31:52,575 - DEBUG - musicbot.permissions | In permissions.py::MainThread(140250957956928), line 98 in async_validate: Fixing automatic owner group
[3.7567837238311768] 2020-07-14 21:31:52,575 - INFO - musicbot.bot | In bot.py::MainThread(140250957956928), line 979 in on_ready: Connected: 732710014216830987/Merde#6205
[3.7572257518768311] 2020-07-14 21:31:52,576 - INFO - musicbot.bot | In bot.py::MainThread(140250957956928), line 987 in on_ready: Owner: 193120890899267585/ٴٴ#9103
[3.7575387954711914] 2020-07-14 21:31:52,576 - INFO - musicbot.bot | In bot.py::MainThread(140250957956928), line 993 in on_ready: Guild List:
[3.7578756809234619] 2020-07-14 21:31:52,576 - INFO - musicbot.bot | In bot.py::MainThread(140250957956928), line 997 in on_ready: - Harder
[3.7582430839538574] 2020-07-14 21:31:52,577 - INFO - musicbot.bot | In bot.py::MainThread(140250957956928), line 1053 in on_ready: Not bound to any text channels
[3.7584948539733887] 2020-07-14 21:31:52,577 - INFO - musicbot.bot | In bot.py::MainThread(140250957956928), line 1079 in on_ready: Not autojoining any voice channels
[3.7594964504241943] 2020-07-14 21:31:52,578 - INFO - musicbot.bot | In bot.py::MainThread(140250957956928), line 286 in _join_startup_channels: Found owner in "General"
[3.7597861289978027] 2020-07-14 21:31:52,578 - INFO - musicbot.bot | In bot.py::MainThread(140250957956928), line 295 in _join_startup_channels: Attempting to join Harder/General
[6.4041831493377686] 2020-07-14 21:31:55,223 - DEBUG - musicbot.bot | In bot.py::MainThread(140250957956928), line 747 in deserialize_queue: Deserializing queue for 732295950714011738
[6.6349246501922607] 2020-07-14 21:31:55,453 - DEBUG - musicbot.bot | In bot.py::MainThread(140250957956928), line 439 in get_player: Created player via deserialization for guild 732295950714011738 with 1 entries
[6.6354134082794189] 2020-07-14 21:31:55,454 - INFO - musicbot.bot | In bot.py::MainThread(140250957956928), line 311 in _join_startup_channels: Joined Harder/General
[6.6366314888000488] 2020-07-14 21:31:55,455 - DEBUG - musicbot.entry | In entry.py::MainThread(140250957956928), line 62 in get_ready_future: Created future for None
[6.6373357772827148] 2020-07-14 21:31:55,456 - INFO - musicbot.entry | In entry.py::MainThread(140250957956928), line 366 in _really_download: Download started: https://soundcloud.com/izanagibang/best-song-ever-created
[8.2095448970794678] 2020-07-14 21:31:57,028 - INFO - musicbot.entry | In entry.py::MainThread(140250957956928), line 376 in _really_download: Download complete: https://soundcloud.com/izanagibang/best-song-ever-created
[8.2132229804992676] 2020-07-14 21:31:57,032 - FFMPEG - musicbot.player | In player.py::MainThread(140250957956928), line 293 in _play: Creating player with options: -nostdin -vn audio_cache/soundcloud-90213207-Best_Song_Ever_Created_Maximbady_-_Hey_baby.mp3
[8.2230744361877441] 2020-07-14 21:31:57,042 - DEBUG - musicbot.player | In player.py::MainThread(140250957956928), line 306 in _play: Playing <musicbot.player.SourcePlaybackCounter object at 0x7f8eb49a2400> using <discord.voice_client.VoiceClient object at 0x7f8eb521fb50>
[8.2273652553558350] 2020-07-14 21:31:57,046 - DEBUG - musicbot.bot | In bot.py::MainThread(140250957956928), line 474 in on_player_play: Running on_player_play
[8.2278752326965332] 2020-07-14 21:31:57,046 - DEBUG - musicbot.bot | In bot.py::MainThread(140250957956928), line 723 in serialize_queue: Serializing queue for 732295950714011738
[27.0019505023956299] 2020-07-14 21:32:15,820 - INFO - musicbot.bot | In bot.py::MainThread(140250957956928), line 2716 in on_message: 193120890899267585/ٴٴ#9103: !skip
[27.0131967067718506] 2020-07-14 21:32:15,832 - DEBUG - musicbot.player | In player.py::Thread-3(140250671998720), line 220 in _playback_finished: Deleting file: audio_cache/soundcloud-90213207-Best_Song_Ever_Created_Maximbady_-_Hey_baby.mp3
[27.0143411159515381] 2020-07-14 21:32:15,833 - DEBUG - musicbot.player | In player.py::Thread-3(140250671998720), line 225 in _playback_finished: File deleted: audio_cache/soundcloud-90213207-Best_Song_Ever_Created_Maximbady_-_Hey_baby.mp3
[27.0178744792938232] 2020-07-14 21:32:15,836 - DEBUG - musicbot.bot | In bot.py::MainThread(140250957956928), line 550 in on_player_finished_playing: Running on_player_finished_playing
[27.2232422828674316] 2020-07-14 21:32:16,042 - DEBUG - musicbot.bot | In bot.py::MainThread(140250957956928), line 723 in serialize_queue: Serializing queue for 732295950714011738
[27.2246158123016357] 2020-07-14 21:32:16,043 - DEBUG - musicbot.bot | In bot.py::MainThread(140250957956928), line 546 in on_player_stop: Running on_player_stop
[104.0940260887145996] 2020-07-14 21:33:32,913 - INFO - musicbot.bot | In bot.py::MainThread(140250957956928), line 2716 in on_message: 193120890899267585/ٴٴ#9103: !stream https://youtu.be/-aLYvZ5sX28
[104.9392480850219727] 2020-07-14 21:33:33,758 - ERROR - musicbot.bot | In bot.py::MainThread(140250957956928), line 2845 in on_message: Error in stream: ExtractionError: Unknown error: [0;31mERROR:[0m Unable to download webpage: HTTP Error 429: Too Many Requests (caused by <HTTPError 429: 'Too Many Requests'>)
Traceback (most recent call last):
File "/usr/local/lib/python3.8/dist-packages/youtube_dl/extractor/common.py", line 627, in _request_webpage
return self._downloader.urlopen(url_or_request)
File "/usr/local/lib/python3.8/dist-packages/youtube_dl/YoutubeDL.py", line 2238, in urlopen
return self._opener.open(req, timeout=self._socket_timeout)
File "/usr/lib/python3.8/urllib/request.py", line 531, in open
response = meth(req, response)
File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
response = self.parent.error(
File "/usr/lib/python3.8/urllib/request.py", line 563, in error
result = self._call_chain(*args)
File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
result = func(*args)
File "/usr/lib/python3.8/urllib/request.py", line 755, in http_error_302
return self.parent.open(new, timeout=req.timeout)
File "/usr/lib/python3.8/urllib/request.py", line 531, in open
response = meth(req, response)
File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
response = self.parent.error(
File "/usr/lib/python3.8/urllib/request.py", line 569, in error
return self._call_chain(*args)
File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
result = func(*args)
File "/usr/lib/python3.8/urllib/request.py", line 649, in http_error_default
raise HTTPError(req.full_url, code, msg, hdrs, fp)
urllib.error.HTTPError: HTTP Error 429: Too Many Requests
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/lib/python3.8/dist-packages/youtube_dl/YoutubeDL.py", line 797, in extract_info
ie_result = ie.extract(url)
File "/usr/local/lib/python3.8/dist-packages/youtube_dl/extractor/common.py", line 530, in extract
ie_result = self._real_extract(url)
File "/usr/local/lib/python3.8/dist-packages/youtube_dl/extractor/youtube.py", line 1782, in _real_extract
video_webpage, urlh = self._download_webpage_handle(url, video_id)
File "/usr/local/lib/python3.8/dist-packages/youtube_dl/extractor/youtube.py", line 276, in _download_webpage_handle
return super(YoutubeBaseInfoExtractor, self)._download_webpage_handle(
File "/usr/local/lib/python3.8/dist-packages/youtube_dl/extractor/common.py", line 660, in _download_webpage_handle
urlh = self._request_webpage(url_or_request, video_id, note, errnote, fatal, data=data, headers=headers, query=query, expected_status=expected_status)
File "/usr/local/lib/python3.8/dist-packages/youtube_dl/extractor/common.py", line 645, in _request_webpage
raise ExtractorError(errmsg, sys.exc_info()[2], cause=err)
youtube_dl.utils.ExtractorError: Unable to download webpage: HTTP Error 429: Too Many Requests (caused by <HTTPError 429: 'Too Many Requests'>)
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/ikk157/MusicBot/musicbot/playlist.py", line 123, in add_stream_entry
info = await self.downloader.extract_info(self.loop, song_url, download=False)
File "/home/ikk157/MusicBot/musicbot/downloader.py", line 84, in extract_info
return await loop.run_in_executor(self.thread_pool, functools.partial(self.unsafe_ytdl.extract_info, *args, **kwargs))
File "/usr/lib/python3.8/concurrent/futures/thread.py", line 57, in run
result = self.fn(*self.args, **self.kwargs)
File "/usr/local/lib/python3.8/dist-packages/youtube_dl/YoutubeDL.py", line 820, in extract_info
self.report_error(compat_str(e), e.format_traceback())
File "/usr/local/lib/python3.8/dist-packages/youtube_dl/YoutubeDL.py", line 625, in report_error
self.trouble(error_message, tb)
File "/usr/local/lib/python3.8/dist-packages/youtube_dl/YoutubeDL.py", line 595, in trouble
raise DownloadError(message, exc_info)
youtube_dl.utils.DownloadError: [0;31mERROR:[0m Unable to download webpage: HTTP Error 429: Too Many Requests (caused by <HTTPError 429: 'Too Many Requests'>)
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/ikk157/MusicBot/musicbot/bot.py", line 2823, in on_message
response = await handler(**handler_kwargs)
File "/home/ikk157/MusicBot/musicbot/bot.py", line 1709, in cmd_stream
await player.playlist.add_stream_entry(song_url, channel=channel, author=author)
File "/home/ikk157/MusicBot/musicbot/playlist.py", line 138, in add_stream_entry
raise ExtractionError("Unknown error: {}".format(e))
musicbot.exceptions.ExtractionError: Unknown error: [0;31mERROR:[0m Unable to download webpage: HTTP Error 429: Too Many Requests (caused by <HTTPError 429: 'Too Many Requests'>)
[133.1778171062469482] 2020-07-14 21:34:01,996 - INFO - musicbot.bot | In bot.py::MainThread(140250957956928), line 2716 in on_message: 193120890899267585/ٴٴ#9103: !shutdown
[133.707516432] launcher-INFO: All done. ```
This won't fix anything about the IP block, because you'll still be downloading from the same IP. We do want to add a built in proxy in the feature to help prevent the IP block however.
This won't fix anything about the IP block, because you'll still be downloading from the same IP. We do want to add a built in proxy in the feature to help prevent the IP block however.
This won't fix anything about the IP block, because you'll still be downloading from the same IP. We do want to add a built in proxy in the feature to help prevent the IP block however.
Ah i see... I appreciate your efforts!
This won't fix anything about the IP block, because you'll still be downloading from the same IP. We do want to add a built in proxy in the feature to help prevent the IP block however.
Ah i see... I appreciate your efforts!
| gharchive/issue | 2020-07-14T22:15:55 | 2025-04-01T06:37:07.455680 | {
"authors": [
"AutumnClove",
"ibrahimk157"
],
"repo": "Just-Some-Bots/MusicBot",
"url": "https://github.com/Just-Some-Bots/MusicBot/issues/2094",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
359307385 | Autoplaylist Functionalities Phase 1: Autostream Implementation
After creating your pull request, tick these boxes if they are applicable to you.
[x] I have tested my changes against the review branch (the latest developmental version), and this pull request is targeting that branch as a base
[x] I have tested my changes on Python 3.5/3.6
Description
I planned to do several pull requests to rewrite and add functionalities to autoplaylist which cover a range of feature requests. This will happen in different phases. Phases are arranged in the order that it won't interfere with each other very much.
phase 1: add autostream and related functionalities
phase 2: make options to allow totally different autoplaylist.txt in different servers (or guild)
phase 3: add an option to show autoplaylist "now playing" notifications
This pull request introduces autostream. It's similar to autoplaylist, but it's for streams.
You can add streams for autoplaying in config/autostream.txt.
If the configuration for randomization and toggling is not set, autostream will play after autoplaylist.
Update 1: autostream and autoplaylist could be set to skip automatically when someone adds stuff to the queue.
Update 2: autostream and autoplaylist can play in 2 modes:
merge: merge autoplaylist and autostream when playback
toggle: use the toggleplaylist command to toggle between them
Update 3:
modes are saved
different modes for different servers now possible
I am currently satisfied with these changes. I don't think I will push more commits except commits for clarity and bugs fixing from b06c8b3 onwards. (This mark an end to more feature implementation to this little PR)
As of now, if #1740 got merged before this. This PR will require some modifications to work properly. If there's plan to merge #1740, merge it first. I will make the required modifications on this PR after #1740 got merged. If #1740 got merged and this message is still here, DO NOT MERGE THIS PR as it will definitely break. I will merge changes required from TheerapakG/MusicBot:autostream_permissions into this.
Related issues (if applicable)
(#1462, #1587)
not a direct implementation of #1587 but introduce "roughly equivalent" features, here is the breakdown:
autostream.txt now exist
you can add or remove a stream using autostream command
toggling can be set using the config file
state of toggling does save so next startup it'll be on whatever mode you've set
you can use the config file to skip the autos when added the new entry (the self-closing stream do exist so I added the config to let you choose to force the autostream to skip or not)
Waiting AutoStream so much! Thx for work man!
Any ETA ?
Any ETA for autostream function ?
You can try it out now by pulling from the branch in my repository. There might be some bugs (I would expect only two or three bugs at most) which I haven't found. If you try it out and found them please let me know! For the date when this will be merged into the main branch, I don't know yet. It depends on the collaborator of this repository.
| gharchive/pull-request | 2018-09-12T04:03:37 | 2025-04-01T06:37:07.464882 | {
"authors": [
"ReddeR1337",
"TheerapakG"
],
"repo": "Just-Some-Bots/MusicBot",
"url": "https://github.com/Just-Some-Bots/MusicBot/pull/1727",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1903308752 | Allow to create .plugin file from any plugin
Might make sense in accordance with https://github.com/JustAnyones/Plugin-creator-website/issues/2 to allow to create a .plugin from any plugin or previous plugin made in PCA in case of an update.
It has been decided to not implement it for any kind of plugin. The API for encrypting plugins can be made public, but PCA will not support this feature out of the box, only for its own projects.
| gharchive/issue | 2023-09-19T16:06:06 | 2025-04-01T06:37:07.468698 | {
"authors": [
"JustAnyones"
],
"repo": "JustAnyones/Plugin-creator-website",
"url": "https://github.com/JustAnyones/Plugin-creator-website/issues/3",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
195473333 | 请教目录缓存
请问大神 ,BookReader的目录缓存是怎么做的啊,每次 进来的时候都请求目录然后缓存吗?
目前进入阅读界面是先加载缓存,再进行请求。具体参照com.justwayward.reader.ui.presenter.BookReadPresenter 目前这个也不是很完善,暂时是这么做的。
大神 这个好像懂了 ,看了您的这个类。
我还有一个疑问 就是 ,我们每次读取一章的时候,是不是第一次就会把章节的所有内容都拿出来,而不是每一次翻页都会去内存里面读,然后显示出来啊 。
是每次翻页的时候 都会计算分页吗
翻页的时候 根据position 去读取一页的内容,实时的 不是一次性读出来
好勒 谢谢大神
客气客气~~
| gharchive/issue | 2016-12-14T08:54:04 | 2025-04-01T06:37:07.499049 | {
"authors": [
"cocoas",
"smuyyh"
],
"repo": "JustWayward/BookReader",
"url": "https://github.com/JustWayward/BookReader/issues/80",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
842646266 | Can you share the word crop code
In the paper : "We crop from the proposed multilingual dataset. We discard images with widths shorter than 32 pixels as they are too blurry, and obtain 4.1M word images in total."
But I ended up with more than 7 million text line images.
How did you crop the text regions? Did you use axis-aligned boxes or quadrilaterals?
@Jyouhou I use axis-aligned boxes,and only the rectangle with width and height greater than 32 is reserved
Thanks for the reply.
Most text are highly oriented in the dataset. I filtered by the shortest edge of the quadrilaterals (not the axis-aligned boxes).
@Jyouhou Can you share your wechat? It's more convenient to communicate
Sure. You can send your wechat account to my cmu email: shangbal@cs.cmu.edu
| gharchive/issue | 2021-03-28T02:41:00 | 2025-04-01T06:37:07.512761 | {
"authors": [
"Jyouhou",
"wushilian"
],
"repo": "Jyouhou/UnrealText",
"url": "https://github.com/Jyouhou/UnrealText/issues/23",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2616837114 | Changes by create-pull-request action
Automated changes by create-pull-request GitHub action
Your image
keinos/sqlite3:latest
Current base image
alpine:latest
Overview
Image reference
keinos/sqlite3:3.46.1
keinos/sqlite3:latest
- digest
ae78ae013f46
cdc9a2c3c976
- tag
3.46.1
latest
- stream
latest
- provenance
https://github.com/KEINOS/Dockerfile_of_SQLite3/commit/9e8d27bfac0f790f9a6de7babbb20270ab9611b7
- vulnerabilities
- platform
linux/amd64
linux/amd64
- size
7.4 MB
9.7 MB (+2.3 MB)
- packages
17
14 (-3)
Base Image
alpine:3also known as:• 3.20• 3.20.3• latest
alpine:latestalso known as:• 3• 3.20• 3.20.3
- vulnerabilities
Policies (0 improved, 0 worsened, 7 missing data)
Policy Name
keinos/sqlite3:3.46.1
keinos/sqlite3:latest
Change
Standing
Default non-root user
:white_check_mark:
:question: No data
No AGPL v3 licenses
:white_check_mark:
:question: No data
No fixable critical or high vulnerabilities
:white_check_mark:
:question: No data
No high-profile vulnerabilities
:white_check_mark:
:question: No data
No outdated base images
:white_check_mark:
:question: No data
No unapproved base images
:white_check_mark:
:question: No data
Supply chain attestations
:white_check_mark:
:question: No data
Packages and Vulnerabilities (5 package changes and 0 vulnerability changes)
:heavy_minus_sign: 3 packages removed
:infinity: 2 packages changed
12 packages unchanged
Changes for packages of type apk (5 changes)
Package
Versionkeinos/sqlite3:3.46.1
Versionkeinos/sqlite3:latest
:heavy_minus_sign:
ca-certificates
20240705-r0
:infinity:
libcrypto3
3.3.2-r0
3.3.2-r1
:infinity:
libssl3
3.3.2-r0
3.3.2-r1
:heavy_minus_sign:
openssl
3.3.2-r0
:heavy_minus_sign:
pax-utils
1.3.7-r2
| gharchive/pull-request | 2024-10-27T20:04:21 | 2025-04-01T06:37:07.576566 | {
"authors": [
"KEINOS"
],
"repo": "KEINOS/Dockerfile_of_SQLite3",
"url": "https://github.com/KEINOS/Dockerfile_of_SQLite3/pull/61",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
233352048 | How could I disable Google search?
I'm using Cerebro as apps launcher and never want to use it for web search. However, sometimes I end up executing search queries which I don't want to do, especially considering it is a Google search (my search engine of choice is Duckduckgo).
I can't find corresponding plugin and do not see an option to disable this. I remember you've being planning to decouple everything into plugins, will it be possible to disable/remove search entirely then?
It would be nice to have google search as a plugin so we could have more flexibility
It is already extracted from the main repository as a separate plugin, so we are all eagerly awaiting the 0.2.9 release to finally be able to disable it 🙂
I see.
It seems that the same applies to Yandex translate:
https://github.com/KELiON/cerebro-yandex-translate
Yes, I've finished extracting plugins, but there are few minor bugs that prevents from releasing it:) But it will be soon, I promise! :D
In version 0.3.0 you can uninstall google plugin as any other
| gharchive/issue | 2017-06-03T07:09:02 | 2025-04-01T06:37:07.582426 | {
"authors": [
"KELiON",
"danielmelogpi",
"maximbaz",
"nazar-pc"
],
"repo": "KELiON/cerebro",
"url": "https://github.com/KELiON/cerebro/issues/333",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.