id
stringlengths 4
10
| text
stringlengths 4
2.14M
| source
stringclasses 2
values | created
timestamp[s]date 2001-05-16 21:05:09
2025-01-01 03:38:30
| added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| metadata
dict |
---|---|---|---|---|---|
125509210
|
Apply project structure to develop branch
Please review and merge the structure to the develop branch.
That structure doesn't make much sense for me. Why we have 03 api folder (android-api-client, ios-api-client, rest-api)?
My structure would be some thing like:
api
mobile
-- android
-- ios
web
@samuelcastro, there aren't 3 api folders. It's 1 api (REST services) folder and 2 client folders one for android (pure java code) and one for ios (can be in c++ or objective-c).
My proposal intent is to maintain separation of responsibilities in the code itself.
The clients apis will have the responsibility to handle the communication with the server and provide a higher level of abstraction to the mobiles applications per se. As an example the android team (and android code) can focus on implementing the interface and call methods without bothering how to handle the connection and so on.
For sure the code could be merged into de mobile application itself. But I think the separation is handy to others to help with the code if the separation of projects is evident. If someone have java skills, it can modify the android-api-client code, test it and not bother installing any kind of IDE, framework or whatever dependency related to android.
Other interesting point is that we suppose to use http://swagger.io to describe and document the REST API. If we're going to use it why not use it's tools to generate the clients ([whatever]-api-client), it do have tools to do that.
It's important to have some comments of the iOS and Android teams, as they're the most impacted ones. So @matheusruschel, @jhowcs and @felipefcm can you tell what you prefer? In really, it's not a good this we segregate the code in libs if all of you don't feel comfortable with this approach.
Just a last comment: api-client is not an api, it's an implementation to access the api (basic a library to access the server).
Either way is fine by me!
—
Sent from Mailbox
On Fri, Jan 8, 2016 at 9:58 AM, Leonardo Cabral notifications@github.com
wrote:
@samuelcastro, there aren't 3 api folders. It's 1 api (REST services) folder and 2 client folders one for android (pure java code) and one for ios (can be in c++ or objective-c).
My proposal intent is to maintain separation of responsibilities in the code itself.
The clients apis will have the responsibility to handle the communication with the server and provide a higher level of abstraction to the mobiles applications per se. As an example the android team (and android code) can focus on implementing the interface and call methods without bothering how to handle the connection and so on.
For sure the code could be merged into de mobile application itself. But I think the separation is handy to others to help with the code if the separation of projects is evident. If someone have java skills, it can modify the android-api-client code, test it and not bother installing any kind of IDE, framework or whatever dependency related to android.
Other interesting point is that we suppose to use http://swagger.io to describe and document the REST API. If we're going to use it why not use it's tools to generate the clients ([whatever]-api-client), it do have tools to do that.
It's important to have some comments of the iOS and Android teams, as they're the most impacted ones. So @matheusruschel, @jhowcs and @felipefcm can you tell what you prefer? In really, it's not a good this we segregate the code in libs if all of you don't feel comfortable with this approach.
Just a last comment: api-client is not an api, it's an implementation to access the api (basic a library to access the server).
Reply to this email directly or view it on GitHub:
https://github.com/VanHack/binners-project/pull/7#issuecomment-169980064
Initially I don't see any problems with the api/app separate folders, but personally prefer less sibling folders in the root. Maybe we can merge @leomcabral and @samuelcastro ideas:
api
--android
--ios
mobile
--android
--ios
web
Does that make sense?
I'm scheduling a meeting with @leomcabral to discuss about that and maybe you guys should joying us.
To be honest, for the nature of that project, I prefer to use a simple approach instead of use "client api", for me this approach will create messy structure and if if understood, we'll create a new request layer, like so:
mobile app > [request] > client-api > [request] > api
And, talking about mobile platform itself, we have big limitation with memory/network/process and so on, so I'm not sure if the @leomcabral's approach is the best one. But I'm open to discuss it, maybe there is misunderstanding in my side and we are a team, so whatever the team decide I'm gonna use for sure.
Alright, When should we meet?
On Fri, Jan 8, 2016 at 1:55 PM, Samuel Castro notifications@github.com
wrote:
I'm scheduling a meeting with @leomcabral https://github.com/leomcabral
to discuss about that and maybe you guys should joying us.
To be honest, for the nature of that project, I prefer to use a simple
approach instead of use "client api", for me this approach will create
messy structure and if if understood, we'll create a new request layer,
like so:
mobile app > [request] > client-api > [request] > api
And, talking about mobile platform itself, we have big limitation with
memory/network/process and so on, so I'm not sure if the @leomcabral
https://github.com/leomcabral's approach is the best one. But I'm open
to discuss it, maybe there is misunderstanding in my side and we are a
team, so whatever the team decide I'm gonna use for sure.
—
Reply to this email directly or view it on GitHub
https://github.com/VanHack/binners-project/pull/7#issuecomment-170038485
.
Just let me know your best time guys, lets see a good one.
Everyday from 12 pm to 00 am works for me
On Fri, Jan 8, 2016 at 2:26 PM, Samuel Castro notifications@github.com
wrote:
Just let me know your best time guys, lets see a good one.
—
Reply to this email directly or view it on GitHub
https://github.com/VanHack/binners-project/pull/7#issuecomment-170047183
.
@samuelcastro I think you get it complete wrong. There is no request between the mobile app to the client-api. The project of a client api is not a service that have to be managed. A client api project will result in a jar file (I don't know if you are aware of Java terms, but it a jar is like a DLL).
The mobile apps will only use objects that are implemented in a DLL (jar in java case).
mobile_app + client-api-lib -> [request] -> api
I've already worked with Java (JEE/SE) and you're correct I just got some misunderstanding, is more clear now.
You suggestion make total sense when working in a big project with a big team, I still don't think it will make much sense in our project, because we're small team and so on.
And, we are not just talking about Java and jars, there is the IOS side, that will be created using swift, but anyway, this is up to you guys, I'll probably have time just to help in the API side, so, feel free to choose whatever you think is better.
So we can end this thread now. I'll close it and we'll keep it simple for the start. Let's start with @samuelcastro structure:
api
mobile
-- android
-- ios
web
We can expand it later as needed.
|
gharchive/pull-request
| 2016-01-07T23:32:28 |
2025-04-01T04:55:45.240510
|
{
"authors": [
"felipefcm",
"leomcabral",
"matheusruschel",
"samuelcastro"
],
"repo": "VanHack/binners-project",
"url": "https://github.com/VanHack/binners-project/pull/7",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
730141777
|
hot reload cogs if owner
add owner only command to reload cogs, saves stopping bot and restarting it when event is happening
https://discordpy.readthedocs.io/en/latest/ext/commands/extensions.html
|
gharchive/issue
| 2020-10-27T06:08:01 |
2025-04-01T04:55:45.248687
|
{
"authors": [
"aadibajpai"
],
"repo": "VandyHacks/the-squirrel-from-VandyHacks",
"url": "https://github.com/VandyHacks/the-squirrel-from-VandyHacks/issues/8",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1393808001
|
updating readme
I would love to update readme with more precise steps for contributing to the repo
sure
|
gharchive/issue
| 2022-10-02T15:43:39 |
2025-04-01T04:55:45.251121
|
{
"authors": [
"Vanshika2063",
"deep-poharkar"
],
"repo": "Vanshika2063/Leetcode-Questions",
"url": "https://github.com/Vanshika2063/Leetcode-Questions/issues/113",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2311697647
|
Improve Button CSS for Better Visual Appeal
Enhance the button design to be more visually appealing and interactive. The current design is too complex and needs to be simplified for better user experience and maintainability.
Ask for feedback from friends
|
gharchive/issue
| 2024-05-23T00:20:22 |
2025-04-01T04:55:45.311853
|
{
"authors": [
"Vaten0x"
],
"repo": "Vaten0x/LectureSurfers",
"url": "https://github.com/Vaten0x/LectureSurfers/issues/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
448752040
|
proxy issues, getting captcha challenge
Hi, open this new ticket.
When i putting proxies on cloudscraper, i always get captcha challenge.
Ps: Can i ini session on local ip and then navigate in the sites with proxies ?
Ps: Can i ini session on local ip and then navigate in the sites with proxies ?
yes, just {} the proxies in your scrapper session.
The only "major" thing is see is that requests uses urllib3 which inturn uses httplib to proxiy https requests... and this changes the request methodology from a GET to a CONNECT...
I still need to investigate more
I tried this : `def IniSession(proxies,k):
while 1 :
print('Monitor {} - Try to ini session '.format(k))
L = []
s = cfscrape.create_scraper()
proxie=None
try:
a = s.get('https://www.nakedcph.com/adidas-originals-yeezy-boost-350-v2-eg5293/p/7677', proxies=proxie)
if a.status_code == 200 :
print('Monitor {} - Ini success !'.format(k))
proxie = random.choice(proxies)
a = s.get('https://www.nakedcph.com/adidas-originals-yeezy-boost-350-v2-eg5293/p/7677', proxies=proxie)
L.append(s)
L.append(proxie)
return L
except Exception as e:
print('Monitor {} - ERROR TO INI !{}'.format(k,e))
pass`
output :
Monitor 0 - Try to ini session
Monitor 0 - Ini success !
Monitor -0 : Error in Monitorlinks ... Captcha
So... after much debugging... i just decided to install my own socks5 proxy locally to test with,... i don't get any reCaptcha's ... @JeanRon have you tested this out with a local proxy instead of a random public one?
That said... I am in the process of implementing some 3rd party reCaptcha solvers... so if you do want to use public proxies... you can solve the reCaptcha that way... but of course you will need to get an account with said 3rd party... but tbh they cheap as...
That not free public proxies, but they could be flag. Where did you make your proxies ?
sounds good
I used danted
@JeanRon re those proxies?
Hi, proxies are form different servers, Linode, vultr etc ..
@JeanRon tbh, i think its because the exit nodes are known to CF... anyway i have started to introduce 3rd party reCaptcha resolvers into the code, currently anti-captcha.com is in there, and im working on deathbycaptcha atm, but there service seems to keep going down and giving me 500 errors... so it makes it hard to test / write for...
I'm going to close this off as i do not believe this is a code based issue, as when i push it through private proxies i dont get captcha's but i do when i push it out open or well known proxy providers, including tor exit nodes... which leads me to believe CF is aware of the end points
|
gharchive/issue
| 2019-05-27T09:18:34 |
2025-04-01T04:55:45.335227
|
{
"authors": [
"JeanRon",
"VeNoMouS"
],
"repo": "VeNoMouS/cloudscraper",
"url": "https://github.com/VeNoMouS/cloudscraper/issues/43",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
674074469
|
Saving and reloading a bndb corrupts segment mappings when a user segment precedes an auto segment
Binary Ninja Version: 2.1.2322-dev Personal, 38f36d9a
Platform: Arch Linux
At some point between 1.1.1338-dev and 2.1.2263, Binary Ninja started corrupting databases with a specific segment layout on save, causing them to produce an incorrect map when reloaded. I know the issue happens on save because I can open a database saved with 1.1.1338-dev in 2.1.2263 and not see any corruption until I resave it with 2.1.2263. I have confirmed that this bug is still present in the latest development build 2.1.2322-dev.
Please see my repository here, which contains a minimal repro of the bug as well as a README with a full description of when/how it manifests.
Thanks for reporting. This issue is fixed in 2.1.2394-dev.
Note that rebasing is performed in relation to the lowest addressed auto segment.
|
gharchive/issue
| 2020-08-06T07:05:00 |
2025-04-01T04:55:45.338426
|
{
"authors": [
"bpotchik",
"tchebb"
],
"repo": "Vector35/binaryninja-api",
"url": "https://github.com/Vector35/binaryninja-api/issues/1872",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
928983342
|
Races between multiple worker threads - random API failures
Binary Ninja Version: 2.4.2875-dev, adb7c635
Platform: macOS 10.15
When running a standalone script (commercial license) using binaryninja API and multiple worker threads, the API calls start to behave non-deterministically. Sometimes, certain objects like func.low_level_il lose reference to its medium_level_il or vice versa. Other times, the func.medium_level_il.ssa_form[x] failes for index x, while it should not.
The problematic situation is transient and does not occur every time, but the script fails when it does.
Initially I though there is a race condition between multiple worker threads, where one thread performs analysis update and the other starts using the objects. So, I added bv.set_analysis_hold(True) to stop that. It did not help.
Then, I set the number of worker threads to 1. The issue is not reproducible since then.
Are you holding a reference to the initial IL function object once you start working with it, or are you grabbing it from the top level Function object every time?
Essentially, the IL function objects are handled as a snapshot. Once you grab and hold onto a reference to any of them, the related IL functions will persist and you can interact with them/query them safely. So going from say, low_level_il_func.ssa_form[x] or low_level_il_func[x].ssa_form should be safe, it relies on you having a persistent reference to one of these IL function objects. Similarly, jumping between levels from IL function objects should be safe: low_level_il_func.medium_level_il for example.
Native function objects are different, and when you query for the IL function object from one of them you'll be given the IL function object that was generated most recently. So if you have some long-running analysis and you are accessing native_func.low_level_il[x] repeatedly, all of those are opportunities for you to be getting results from a newer collection of IL function objects. Something that might result in you seeing results from multiple analysis results at once might be low_level_il_func.source_function.medium_level_il, because the original low_level_il might be much older than the new MLIL function object you get.
If you were to restructure your code slightly to hold onto one of the IL function objects and only get access to the other IL forms from that first IL function object I think your issues would go away.
In summary: try to only grab one of the IL functions from the native Function object once. Then get to the other IL forms by jumping directly between the IL functions, not by hopping through the native Function object (which will always give you the latest IL function). Multiple hops between IL functions are also safe: low_level_il.medium_level_il.high_level_il.
I am actually working with the IL function object. I am not repeatedly grabbing it via native function.
But, I do use func_mlil.source_parameters reference for the following situations:
I need to find cross references (call sites) for the func_mlil
I need to inspect list of input parameters to func_mlil
I need to get destination MLIL function object for a call inside func_mlil
It seems, this is where things become flaky?
Is there a way to do 1., 2. and 3. without going back to the native function reference?
Those sound like they should be safe. Going back to the native function object is fine, so long as you don’t use that native function object to go back to the IL functions (since they may not be the ones you came from) in a way that could result in mixing results from IL functions from different analysis updates of the same function.
The rest of this is written assuming that func_mlil.source_parameters there should be func_mlil.source_function.
By 3. do you mean get the MLIL function object for the call target of a call in some function? If so, that should be fine because the concern about mixing results between IL function objects is only relevant within a single function. Also another way to do that via the mlil call instruction directly might be bv.get_function_at(mlil_call_inst.dest.constant) (provided the call target is in fact constant).
So to go back to your original post, this bit stands out: func.medium_level_il.ssa_form[x] resulting in spurious failures. Because the IL function objects don’t change once they’re generated, I’m curious where x is coming from. Normally I’d expect that failing to be a result of getting an instruction index from one IL function object and using it in another accidentally (the most common form of this is accidentally mixing an index from an older IL function object, the underlying function being reanalyzed, then using the index with the most recent version). An index is only valid for the specific IL function object it came from since they’re not intended to be stable across function updates or forms. So if that’s not the problem we might be giving you an invalid instruction index from somewhere else and that’s what we need to chase down.
I tracked the issue down to disappearing MLIL objects:
The corresponding disassembly is there, the low level IL is there, but all of a sudden the MLIL is gone.
That's generally the result of the UI noticing that analysis is in progress and the object isn't ready yet, not that it's gone missing. We generate the ILs lazily on demand because they can be memory-intensive, so generally speaking when navigating to a function in the UI there's simply no IL to display at first.
Once you've got a IL function object that you're using nothing should happen to change it out from under you.
Do you by any chance have a script that can reproduce the behavior you're seeing when run on a sufficiently large binary?
We've been unable to reproduce this and it doesn't sound like there are any other examples that can repro it at this point.
For now we're going to close as "can't reproduce" but definitely feel free to re-open if you see it pop up again or get a good repro.
|
gharchive/issue
| 2021-06-24T08:33:09 |
2025-04-01T04:55:45.349295
|
{
"authors": [
"psifertex",
"rssor",
"wipawel"
],
"repo": "Vector35/binaryninja-api",
"url": "https://github.com/Vector35/binaryninja-api/issues/2526",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2041744315
|
Value set analysis loses track of stack offset value
Version and Platform (required):
Binary Ninja Version: latest dev (will update this with exact number when I get home)
OS: Linux Ubuntu
OS Version: 22.04 LTS
CPU Architecture: x64
Bug Description:
I have a snippet of code like this:
0x1446f95f1: mov qword [rsp], rbp
0x1446f95f5: sub rsp, 0x40
When the first instruction is executing, rbp contains a certain offset in the current frame:
(f is the function containing this)
>>> f.get_reg_value_at(0x1446f95f1, 'rbp')
<stack frame offset -0x210> # <- this value is important
>>> f.get_reg_value_at(0x1446f95f1, 'rsp')
<stack frame offset -0x250> # <- current stack offset
>>> f.get_stack_contents_at(0x1446f95f1, -0x250, 8)
<undetermined> # <- makes sense, we haven't assigned the value yet
So far, so good, however immediately after the instruction has executed, Binja does not know what is on top of the stack:
>>> f.get_stack_contents_at(0x1446f95f5, -0x250, 8)
<undetermined> # <- this should be "<stack frame offset -0x210>", not "<undetermined>"
This is an issue because later in the function the code restores the rbp and uses it as a base, something like:
mov rbp, [rsp + xxx] # pointing to the same location as above
mov rax, [rbp + 0x30] # Can not resolve where on the stack this points to
jmp rax # destination of this jump unknown even though it's a constant value in reality
Expected Behavior:
I expect the value set analysis to keep track of the register value when pushed on the stack.
Any chance you can share the binary or what the MLILSSA looks like?
MLILSSA looks like this:
I can not share the binary publicly but @fuzyll already have it and I can send it to you on Slack as well.
Apparently I forgot to put it in this issue, but the magic words for finding the binary are "playful brisk dairy sauna".
So I tried this but I can't reproduce the issue. Though I'm getting different values for the stack at location:
>>> f.get_reg_value_at(0x1446f95f1, 'rbp')
<entry rbp>
>>> f.get_reg_value_at(0x1446f95f1, 'rsp')
<stack frame offset -0x10>
>>> f.get_stack_contents_at(0x1446f95f1, -0x10, 8)
<undetermined>
>>> f.get_stack_contents_after(0x1446f95f1, -0x10, 8)
<entry rbp>
I will take a look with the latest binja version and see what results I get. I will get back on this.
Closing this issue as it seems to be resolved
|
gharchive/issue
| 2023-12-14T13:50:13 |
2025-04-01T04:55:45.356705
|
{
"authors": [
"ZetaTwo",
"fuzyll",
"plafosse"
],
"repo": "Vector35/binaryninja-api",
"url": "https://github.com/Vector35/binaryninja-api/issues/4832",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
56266388
|
Scroll performance on OSX
After installing the current version 1.4.2 on OSX 10.10.2 the scroll performance, even on small files with a few 100 lines of code, slows down significantly. Scrolling is no longer smooth, instead it skips every other line.
PHPStorm 8.0.2 on a Macbook Pro Retina Mid 2014 with GT 750M
Update: this is still present on the lastest PHPStorm 10.x version
I have a bad feeling that this is probably going to be related to natural scrolling or retina, both are really difficult to debug without running OSX.
I am going to be doing a general performance pass in the next release so it might go away, but if its every other line then maybe not :-/
might be related to https://github.com/Vektah/CodeGlance/issues/93
Huge overhaul to give pixel perfect scrolling in 1.5.0. Should be fixed
|
gharchive/issue
| 2015-02-02T17:29:51 |
2025-04-01T04:55:45.363340
|
{
"authors": [
"Vektah",
"ingokasch"
],
"repo": "Vektah/CodeGlance",
"url": "https://github.com/Vektah/CodeGlance/issues/67",
"license": "bsd-2-clause",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1944187028
|
chore(dev-deps): update commitlint monorepo to v17.8.0
This PR contains the following updates:
Package
Type
Update
Change
@commitlint/cli (source)
devDependencies
minor
17.7.2 -> 17.8.0
@commitlint/config-conventional (source)
devDependencies
minor
17.7.0 -> 17.8.0
@commitlint/prompt (source)
devDependencies
minor
17.7.2 -> 17.8.0
Release Notes
conventional-changelog/commitlint (@commitlint/cli)
v17.8.0
Compare Source
Note: Version bump only for package @commitlint/cli
17.7.2 (2023-09-28)
Note: Version bump only for package @commitlint/cli
17.7.1 (2023-08-10)
Note: Version bump only for package @commitlint/cli
conventional-changelog/commitlint (@commitlint/config-conventional)
v17.8.0
Compare Source
Note: Version bump only for package @commitlint/config-conventional
conventional-changelog/commitlint (@commitlint/prompt)
v17.8.0
Compare Source
Note: Version bump only for package @commitlint/prompt
17.7.2 (2023-09-28)
Note: Version bump only for package @commitlint/prompt
17.7.1 (2023-08-10)
Note: Version bump only for package @commitlint/prompt
Configuration
📅 Schedule: Branch creation - At any time (no schedule defined), Automerge - At any time (no schedule defined).
🚦 Automerge: Enabled.
♻ Rebasing: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox.
🔕 Ignore: Close this PR and you won't be reminded about these updates again.
[ ] If you want to rebase/retry this PR, check this box
This PR has been generated by Renovate Bot.
:tada: This PR is included in version 1.2.3 :tada:
The release is available on GitHub release
Your semantic-release bot :package::rocket:
|
gharchive/pull-request
| 2023-10-16T01:45:18 |
2025-04-01T04:55:45.405940
|
{
"authors": [
"Venatum"
],
"repo": "Venatum/bull-board-docker",
"url": "https://github.com/Venatum/bull-board-docker/pull/75",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1667022549
|
More support for labyrinth_goal
Currently, the keywords are exactly those KoLmafia understands in the "shadowLabyrinthGoal" property.
KoLmafia has no default for that. This script has to set that property with a default, since otherwise KoLmafia will stop automation and tell the user to continue in the browser.
Available options:
"muscle", "mysticality", "moxie" - gain 90-100 substats.
"maxHP - 30 turns of Maximum HP +300%
"maxMP - 30 turns of Maximum MP +300%
"resistance" - Superhuman spooky, hot, and sleaze resistance
"effects" - +3 turns to 3 random effects.
We could add a pseudo-option - "mainstat" - which we transform at runtime into "muscle", "mysticality", or "moxie".
We arbitrarily chose "maxHP" as the default for VSR.LabyrinthGoal - the value that will be used if not overriden on the command line.
Perhaps "mainstat" would be better?
Perhaps "effects" would be better? It might even extend Shadow Affinity.
I just made "effects" the default.
I still want a "mainstat" keyword.
Revision https://github.com/Veracity0/shadow-rift/pull/18 adds the "mainstat" keyword.
|
gharchive/issue
| 2023-04-13T19:35:43 |
2025-04-01T04:55:45.433400
|
{
"authors": [
"Veracity0"
],
"repo": "Veracity0/shadow-rift",
"url": "https://github.com/Veracity0/shadow-rift/issues/9",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1324720884
|
feat: being able to provide custom main() method for initialization on startup
Description
At the moment Dart Frog provides an own main() method. This causes problems since as a developer I can't run my own initialization on startup. Examples are:
Database migration
One-time-initialization of other packages, such as Reflectable
Requirements
[ ] Being able to provide own main() method
[ ] Or alternative getting some other initialization call on startup
Additional Context
If there are already other solutions for the above two cases, then please let me know
Could this be a function in an _initialize.dart file, kinda like _middleware.dart? That way you dart_frog doesn't have to support a completely custom main(), but you could still inject those types of functions on startup.
Hi @mtwichel,
thanks for your prompt response.
TLDR: _initialize() would definitely help! Having no main.dart file might still cause problems with other packages though.
Having _initialize.dart would definitely already help. There is one drawback though. It seems that some packages, such as Reflectable, only look in main.dart for certain things. At least this is the situation I currently have. I added a fake main.dart file just so that the Reflectable code generator "finds" the file and generates the reflection code for the dart files included from main. When renaming main.dart to e.g. main2.dart, Reflectable will not generate code for it.
This is not really a Frog issue. However, the approach of hiding the main() method might break compatibility with other packages which depend on main being available.
Especially Reflectable is useful in combination with Frog, since it allows setting up a simple ORM system.
I also agree that having an entry point is important -- especially for initializing processes/components that are vital for the server to run properly. It also makes it easier to handle errors at startup.
Currently, the only workaround is to create a top level variable as suggested in #142. The main issue with such approach is that it makes handling errors difficult/impossible since the error won't appear until that variable is called.
Hi everyone! This is on our roadmap to support very soon. An initial proposal I have is to allow developers to optionally create a main.dart file which contains a run method with the following signature:
import 'dart:async';
import 'dart:io';
import 'package:dart_frog/dart_frog.dart';
FutureOr<HttpServer> run(
Handler handler,
InternetAddress ip,
int port,
) {
// This will be called on startup and the HttpServer can be configured here.
return serve(handler, ip, port);
}
Let us know what you think, thanks! 🙏
After being asked on Twitter about this proposal, I would like to add some further insight.
Another use case for a place for initialization code is error monitoring. This is often done by wrapping the entry point with runZonedGuarded and using its error callback to send error reports. This approach is used by Sentry, Datadog, et al.
Also the initialization of those error tracking SDKs needs to be done once at the startup of the server.
As far as I can tell, the approach, as outlined above, works for this use case.
|
gharchive/issue
| 2022-08-01T17:08:20 |
2025-04-01T04:55:45.452686
|
{
"authors": [
"felangel",
"mtwichel",
"osaxma",
"ueman",
"zottel4242"
],
"repo": "VeryGoodOpenSource/dart_frog",
"url": "https://github.com/VeryGoodOpenSource/dart_frog/issues/226",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1462417349
|
feat: add publishable flag
Status
READY
Description
Adds a publisable flag to the brick which will be used to decide if we will add the pana workflow and the publish_to field to the pubspec.
Fixes #22
Type of Change
[x] ✨ New feature (non-breaking change which adds functionality)
[ ] 🛠️ Bug fix (non-breaking change which fixes an issue)
[ ] ❌ Breaking change (fix or feature that would cause existing functionality to change)
[ ] 🧹 Code refactor
[ ] ✅ Build configuration change
[ ] 📝 Documentation
[ ] 🗑️ Chore
Thoughts on also removing CHANGELOG and LICENSE when it is not publishable?
|
gharchive/pull-request
| 2022-11-23T21:17:08 |
2025-04-01T04:55:45.456200
|
{
"authors": [
"erickzanardo",
"wolfenrain"
],
"repo": "VeryGoodOpenSource/very_good_dart_package",
"url": "https://github.com/VeryGoodOpenSource/very_good_dart_package/pull/24",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
216797916
|
fix Circle CI build
The ember-try config is actually broken currently which breaks the CircleCLI build…
This also updates the node version used on Circle to 4 from 0.10.
The build is running now but failing. It seems that these failures should be fixed by #341 though. Not sure how you want to handle that (merge this first although the build is failing and then rebase #341?)
@marcoow thanks! I'll merge this now and look at rebasing #341 as you suggest. Thanks for taking the time to update this.
|
gharchive/pull-request
| 2017-03-24T14:13:57 |
2025-04-01T04:55:45.462649
|
{
"authors": [
"bantic",
"marcoow"
],
"repo": "Vestorly/torii",
"url": "https://github.com/Vestorly/torii/pull/342",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1360916994
|
record layout test suite
Adds a test suite for record layout.
Runs permutations of targets and tests
Has expected failures for some tests due to various zig/arocc issues, including parser failures.
will error if a test passes when it currently expects to fail
Can run a specific platform/test for local debugging
readme for more info
@TwoClocks I made some adjustments, could you double check that everything looks OK?
Also could we encode the expected compile errors in the test cases somehow or are they expected to change?
@TwoClocks I made some adjustments, could you double check that everything looks OK?
Looks fine. But I'm not sure why you want to ignore warnings? The output isn't/shouldn't be displayed unless the results are unexpected. Are you seeing warnings spitting out?
Also, both pack(32) and pack(0) are valid, but compiler support may differ. pack(0) should behave the same as pack(). The spec just says "small powers of two", but never defines what "small" is.
Also could we encode the expected compile errors in the test cases somehow or are they expected to change?
I've researched most of the errors in the test cases. I believe are all because of issues in arocc / zig #387. As fixes are merged, I'd expect more tests to pass. I'd expect all tests to pass for all platforms error free at some point.
As issues are addressed, it seemed easier to alter the runner than to alter the individual tests, as a single fix might change many tests. But I can put the expected failures in the tests if you feel that's better.
Looks fine. But I'm not sure why you want to ignore warnings? The output isn't/shouldn't be displayed unless the results are unexpected. Are you seeing warnings spitting out?
Are you talking about setting some diagnostics off? Those were there before but weren't specific to the test cases causing them.
Also, both pack(32) and pack(0) are valid, but compiler support may differ. pack(0) should behave the same as pack(). The spec just says "small powers of two", but never defines what "small" is.
What compiler supports pack(32)? Neither clang, GCC or MSVC seem to accept it. Also what about pack(5), pack(3), etc. in test 64?
I've researched most of the errors in the test cases. I believe are all because of issues in arocc / zig #387. As fixes are merged, I'd expect more tests to pass. All tests for all targets should be error-free at some point. As issues are addressed, it seemed easier to alter the runner than to alter the individual tests, as a single fix might change many tests. For example, #381 is causing all targets to fail on 4 tests. But I can put the expected failures in the tests if you feel that's better.
All 980? Must be a lot of duplicates then.
Are you talking about setting some diagnostics off? Those were there before but weren't specific to the test cases causing them.
I was referring to the changes to the test files. I was concerned you were seeing warnings.
What compiler supports pack(32)? Neither clang, GCC or MSVC seem to accept it. Also what about pack(5), pack(3), etc. in test 64?
Clang accepts it.
Clang compiles 0064 w/o error (lots of warnings). Clang is doing record layout w/ those pack values. So while the values might be "invalid" record layout is happening. 0064 is testing that arocc's layout matches Clang's.
The tests do not opine on what correct is. They are just testing that the layout matches the target:compiler.
While working on this I've run many tests through clang/gcc/msvc because I thought the test was invalid. While I have not done this will all the tests/targets, I have done it with many. I have not found a test that clang/gcc/msvc fail.
All 980? Must be a lot of duplicates then.
Yeah. lots of duplicates. I test every combination of test file and target that's valid for a given test. Some tests are specific to MSVC, or SysV-like targets.
You can see from groups of asserts that a large number of targets are expected to be the same for any individual test.
I did not reduce the tests to "indicative" targets for each test because the grouping of results differs greatly across tests. There are targets that are always together across all tests, but not enough to speed up the running time much.
So I test every perrmutation. I can try to reduce the number of targets if you like.
#381 causes all targets to fail when running test 0064. When it's fixed, I'd expect some subset of targets to pass. Then we'll need to explore the rermaining failures to see if they are a new issue, or an existing one.
This is what I have been doing for months now. Opening bugs, and when you fix them, altering the test results then going to see if the "unmasked" failures are new bugs or dupes. Rinse and repeat.
I'm open to any other process you think would be better.
I was referring to the changes to the test files. I was concerned you were seeing warnings.
There should not be any changes to the test files, I originally modified few but later reverted them.
Clang accepts it. Clang compiles 0064 w/o error (lots of warnings). Clang is doing record layout w/ those pack values. So while the values might be "invalid" record layout is happening. 0064 is testing that arocc's layout matches Clang's.
The tests do not opine on what correct is. They are just testing that the layout matches the target:compiler. While working on this I've run many tests through clang/gcc/msvc because I thought the test was invalid. While I have not done this will all the tests/targets, I have done it with many. I have not found a test that clang/gcc/msvc fail.
What do you mean by accept? For me clang gives a warning about invalid pack parameter and then does not modify the pack value. Either the file should be generated with #pragma GCC diagnostic ignored "-Wignored-pragmas" or those invalid #pragma packs should not be added.
What do you mean by accept? For me clang gives a warning about invalid pack parameter and then does not modify the pack value. Either the file should be generated with #pragma GCC diagnostic ignored "-Wignored-pragmas" or those invalid #pragma packs should not be added.
clang-14 -DONE ./test/records/0064_test.c
This compiles, creates an object file, and no _Static_assert fire. It does not error. I'd call this acceptance. The test checks arocc produces whatever clang produces.
If you want to make arocc error, while clang just warns, then we can remove that part of the test.
|
gharchive/pull-request
| 2022-09-03T17:07:49 |
2025-04-01T04:55:45.476938
|
{
"authors": [
"TwoClocks",
"Vexu"
],
"repo": "Vexu/arocc",
"url": "https://github.com/Vexu/arocc/pull/394",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1050856575
|
1.5.x introduces API breaking changes
Describe the bug
Version 1.5.x is mislabeled. Its version number suggests that there are new features with no breaking API changes. But in fact old code is not compatible anymore.
To Reproduce
Have a flutter project that relies on version 1.3.x or 1.4.x of this plugin.
Run flutter clean, remove the pubspec.lock
Run flutter pub upgrade
See error messages
Expected behavior
The plugin should continue to work with the old code, as it's a minor series upgrade. What is now called 1.5.x should have been a 2.x
Screenshots
None, but changes like https://github.com/VictorKachalov/health_kit_reporter/commit/f08e62901555af0d859ec829bbc682a73ee9d38c#diff-bb8f6fdce8b91b43f4156f21e5fed48f710f08ebaad88347e4b3cc896fdccd6dR167 are likely to have caused the problem.
Desktop (please complete the following information):
All of them
Smartphone (please complete the following information):
All of them
Additional context
Please see https://dart.dev/tools/pub/versioning
Hi @maltemedocs
thank you, makes sense, I will soon publish version 2.0.0 with some other breaking changes.
|
gharchive/issue
| 2021-11-11T11:02:48 |
2025-04-01T04:55:45.521555
|
{
"authors": [
"VictorKachalov",
"maltemedocs"
],
"repo": "VictorKachalov/health_kit_reporter",
"url": "https://github.com/VictorKachalov/health_kit_reporter/issues/33",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
341250808
|
Pathfinding problems with multiple lane roads
I'm finding a problem when using any multiple lane road, almost all cars are refusing to change lanes once they enter the multiple lane section of road. A very small minority are able to change lanes and generally act normally but the majority of vehicles remain stuck on their lanes unable to change lanes to go in a different direction that would get to their destination quicker. This means that I'm ending up with traffic taking extremely circuitous routes or going onto roads not equipped to deal with that level of traffic. I've got a few examples of whats going on in some screenshots below with traffic entering the highways at the highlighted sections (I'm using left hand traffic in my city);
All of these occur when lane connections have been used at one or both of the junctions. As can be seen traffic entering the highways is forced to follow the lane they have entered on and are unable to change lanes and take a faster route to their destination. This is also happening on many of the multi lane city streets as well (I haven't screenshot those as it's harder to see what's going on). Maximising advanced lane selection does not solve the issue, instead cars will spread out and use all of the lanes before all attempting to get back into the same lane as previously. This has all meant that my traffic flow has dropped to about 65-70% rather than the usual 80-85%.
Could you please provide your savegame?
Probably related to #191
So O.k. I had a look at my game and some of the path finds are really weird. Like going from one district to to a neighboring one, but the trucks take a route which is going through another district far away from the ideal route. I would provide my savegame but you would need about 40 GB disk space for assets and at least 90+ GB ram to load it in its current state.
I fixed an error in the cost calculation of the Vehicle AI (see issue #191). If this issue is caused by the same problem then this should also be fixed now.
If this is in the 1.10.10 version its still present on my game I'm using this version actually. Or did I miss an update?
As far I can see in my savegame, the pathfinding works as it should.
Yup the error was fixed alongside #191, it all works as it should, thank you very much.
|
gharchive/issue
| 2018-07-14T17:05:10 |
2025-04-01T04:55:45.527769
|
{
"authors": [
"LordSmeghead",
"SquareBeard",
"VictorPhilipp",
"Videonauth"
],
"repo": "VictorPhilipp/Cities-Skylines-Traffic-Manager-President-Edition",
"url": "https://github.com/VictorPhilipp/Cities-Skylines-Traffic-Manager-President-Edition/issues/192",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2069434882
|
Entrega de tarea lab 1 y 3 A01748692 Diego Sánchez Valle
Programa de Python
La entrega final de la tarea se encuentra en la carpeta 03 de labs
No fue posible incorporar la entrada de texto por .txt, todo es por la entrada standard de la línea de comandos.
el commit "pass or fail messages" es la entrega final
|
gharchive/pull-request
| 2024-01-08T01:44:18 |
2025-04-01T04:55:45.529706
|
{
"authors": [
"SopitaNipona"
],
"repo": "VictorRodriguez/Computational-Methods",
"url": "https://github.com/VictorRodriguez/Computational-Methods/pull/111",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
599875717
|
how many bytes overhead for storing one data item?
how many bytes overhead for storing one data item?
for small byte size and for large 64kb byte sizes.
how many additional bytes are needed per entry? Can you please put this as FAQ on front page?
thx
A bit of info https://github.com/VictoriaMetrics/fastcache/issues/13#issuecomment-498747925
fastcache.New accepts maxBytes arg. This is the maximum cache size in bytes. The number of elements is limited by the same maxBytes arg - each element occupies at least len(key)+len(value)+len(header) bytes, where header is per-entry header with non-zero size.
non-zero size is how big? i would like to allocate sufficient space / know how many elements is possible inside the cache.
fastcache.New accepts maxBytes arg. This is the maximum cache size in bytes. The number of elements is limited by the same maxBytes arg - each element occupies at least len(key)+len(value)+len(header) bytes, where header is per-entry header with non-zero size.
non-zero size is how big? i would like to allocate sufficient space / know how many elements is possible inside the cache.
Does bytes below is the header?
var kvLenBuf [4]byte
|
gharchive/issue
| 2020-04-14T21:38:52 |
2025-04-01T04:55:45.586181
|
{
"authors": [
"Cylkal",
"cristaloleg",
"gitmko0"
],
"repo": "VictoriaMetrics/fastcache",
"url": "https://github.com/VictoriaMetrics/fastcache/issues/32",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2625693730
|
求增加画中画旋转功能
求增加画中画旋转功能 油管有些视频画中画的时候是横向的 如果画中画有可以旋转的按钮就好了
求增加画中画旋转功能 油管有些视频画中画的时候是横向的 如果画中画有可以旋转的按钮就好了
方便加个微信反馈群吗?后续插件会新增部分功能并进行买断制,但反馈的用户都会获得免费使用。微信群
|
gharchive/issue
| 2024-10-31T01:03:33 |
2025-04-01T04:55:45.596167
|
{
"authors": [
"gxy5202",
"nightwav"
],
"repo": "VideoRoll/VideoRoll",
"url": "https://github.com/VideoRoll/VideoRoll/issues/235",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
141373867
|
Topic/build system
Update to the latest version of Autorevision and clean up the project file.
What is the benefit to moving the stuff out of the project file in to xcconfig files? I like being able to see the current build settings in the xcode UI...
All the setting are still visible in the xcode UI, mostly it is a cleaner way to keep track of changes; among other things changes are not written out in triplicate.
|
gharchive/pull-request
| 2016-03-16T19:15:30 |
2025-04-01T04:55:45.598315
|
{
"authors": [
"dak180",
"josh64x2"
],
"repo": "ViennaRSS/vienna-rss",
"url": "https://github.com/ViennaRSS/vienna-rss/pull/586",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1244136251
|
🛑 Rue de la coloc is down
In ee9ea69, Rue de la coloc ($COLOC) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Rue de la coloc is back up in 7988ab0.
|
gharchive/issue
| 2022-05-22T03:27:22 |
2025-04-01T04:55:45.606431
|
{
"authors": [
"Vikingfr"
],
"repo": "Vikingfr/upptime",
"url": "https://github.com/Vikingfr/upptime/issues/633",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1259564036
|
🛑 Rue de la coloc is down
In 2aef448, Rue de la coloc ($COLOC) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Rue de la coloc is back up in 9d815ff.
|
gharchive/issue
| 2022-06-03T07:40:45 |
2025-04-01T04:55:45.608673
|
{
"authors": [
"Vikingfr"
],
"repo": "Vikingfr/upptime",
"url": "https://github.com/Vikingfr/upptime/issues/695",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1260982574
|
🛑 SCOP276 is down
In f0a1a34, SCOP276 ($SCOP276) was down:
HTTP code: 0
Response time: 0 ms
Resolved: SCOP276 is back up in f0f637e.
|
gharchive/issue
| 2022-06-05T07:29:14 |
2025-04-01T04:55:45.610680
|
{
"authors": [
"Vikingfr"
],
"repo": "Vikingfr/upptime",
"url": "https://github.com/Vikingfr/upptime/issues/716",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
562855223
|
Shape from particle.js
Hello,
it's more like a qestion than an issue.
Is there possibility to form a shape from particles using particle.js
Well it will be great if particles could form a shape - like for example - this brain here: http://tomb.pl/projects/thread/images/brain.jpg
You can use CSS with the clip-path property.
|
gharchive/issue
| 2020-02-10T22:01:57 |
2025-04-01T04:55:45.618851
|
{
"authors": [
"SoulDesignerTom",
"bego-mx"
],
"repo": "VincentGarreau/particles.js",
"url": "https://github.com/VincentGarreau/particles.js/issues/416",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
661166428
|
Add Project to the List of OS Projects
I have always had a list of random resources that I kept on GitHub as a sort of non-browser specific bookmark list.
Thought it would make a great OS project that anyone could contribute to with their own resources.
https://github.com/saramccombs/dev-resources
While I don't have the time at the moment to add it, I thought I'd create an issue to get it added in case someone has time before I do.
Please tag me as the reviewer should you take on this issue.
Cheers!
This is a great idea! I have a few things in my bookmarks that I could add to this.
Hi @saramccombs I'm happy to take this, it will be my very first issue
|
gharchive/issue
| 2020-07-19T23:38:54 |
2025-04-01T04:55:45.633115
|
{
"authors": [
"saramccombs",
"tkshill",
"vanessacor"
],
"repo": "Virtual-Coffee/open-source",
"url": "https://github.com/Virtual-Coffee/open-source/issues/10",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1006200136
|
Add scalafix, remove unused imports
The final solution should be to run ./mill all __. fix. There is a problem that several modules are compiling to scala 3 and in this version there is no flag -Ywarn-unused yet.
If we run ./mill all __. fix,it tries to run scalafix on all modules in the defined scala version, but there are modules that compile to every scala version, so it throws an error that -Ywarn-unused flag not exists in scala 3.
I haven't found a solution that can easily exclude scalafix from running for modules compiling to scala 3, therefore I created a scalafix () method in which I explicitly run fix() on each module.
I opened https://github.com/VirtusLab/scala-cli/pull/158 with just https://github.com/VirtusLab/scala-cli/pull/148/commits/8cc457c0f688fe7ebf3153c14eeeae0432a285c0. Once it's merged, we can just squash-and-merge the PR here.
|
gharchive/pull-request
| 2021-09-24T08:10:18 |
2025-04-01T04:55:45.732575
|
{
"authors": [
"alexarchambault",
"lwronski"
],
"repo": "VirtusLab/scala-cli",
"url": "https://github.com/VirtusLab/scala-cli/pull/148",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2216768103
|
Update scalafmt-cli_2.13, scalafmt-core to 3.8.1
About this PR
📦 Updates
org.scalameta:scalafmt-cli_2.13
org.scalameta:scalafmt-core
from 3.7.17 to 3.8.1
📜 GitHub Release Notes - Version Diff
Usage
✅ Please merge!
I'll automatically update this PR to resolve conflicts as long as you don't change it yourself.
If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below.
Configure Scala Steward for your repository with a .scala-steward.conf file.
Have a fantastic day writing Scala!
🔍 Files still referring to the old version number
The following files still refer to the old version number (3.7.17).
You might want to review and update them manually.
website/docs/reference/cli-options.md
website/docs/reference/scala-command/cli-options.md
website/docs/reference/scala-command/runner-specification.md
website/docs/release_notes.md
⚙ Adjust future updates
Add this to your .scala-steward.conf file to ignore future updates of this dependency:
updates.ignore = [ { groupId = "org.scalameta" } ]
Or, add this to slow down future updates of this dependency:
dependencyOverrides = [{
pullRequests = { frequency = "30 days" },
dependency = { groupId = "org.scalameta" }
}]
labels: library-update, early-semver-minor, semver-spec-minor, old-version-remains, commit-count:1
Blocked by https://github.com/scalameta/scalafmt/issues/3891
|
gharchive/pull-request
| 2024-03-31T04:08:28 |
2025-04-01T04:55:45.738891
|
{
"authors": [
"Gedochao",
"scala-steward"
],
"repo": "VirtusLab/scala-cli",
"url": "https://github.com/VirtusLab/scala-cli/pull/2830",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2109631812
|
[Bug] icon name is missing while using customRender
Version
0.18.3
Link to Minimal Reproduction
Link
Steps to Reproduce
open console in the reproduction page
click the third column and they click the forth one.
there first console shows name as undefined and the second one shows and expected.
Current Behavior
icon name is missing in the icon_click event while icon attribute works just fine.
Expected Behavior
icon name is provided just like using icon attribute.
Environment
- OS: Windows 10
- Browser: Microsoft Edge 121
- Framework: React@18
Any additional comments?
No response
codesanbox打不开你的demo 是不是有什么权限
codesanbox 打不开你的 demo 是不是有什么权限
似乎是有点问题,我刚刚修改了一下,麻烦再看看
|
gharchive/issue
| 2024-01-31T09:15:02 |
2025-04-01T04:55:45.745736
|
{
"authors": [
"fangsmile",
"martesi"
],
"repo": "VisActor/VTable",
"url": "https://github.com/VisActor/VTable/issues/1041",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2692874812
|
[Bug] 明细表配置radio column 报错
Version
1.11.1
Link to Minimal Reproduction
https://codesandbox.io/p/sandbox/vtable-export-umd-forked-55p5y4?workspaceId=ca8fcece-a151-491b-9e81-01f767ea3eab
Steps to Reproduce
按照文档配置
Current Behavior
编译报错
Expected Behavior
Environment
- OS:
- Browser:
- Framework:
Any additional comments?
No response
复现链接没有开通权限
复现链接没有开通权限
已更新
headerType 不支持radio @youhan26
|
gharchive/issue
| 2024-11-26T02:44:29 |
2025-04-01T04:55:45.750037
|
{
"authors": [
"fangsmile",
"youhan26"
],
"repo": "VisActor/VTable",
"url": "https://github.com/VisActor/VTable/issues/2909",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
322144385
|
add the LICENSE
no message
submit
|
gharchive/pull-request
| 2018-05-11T02:14:10 |
2025-04-01T04:55:45.847568
|
{
"authors": [
"VizXu"
],
"repo": "VizXu/GobangGame",
"url": "https://github.com/VizXu/GobangGame/pull/1",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1155088208
|
added next auth secret env variable
NextAuth.js is a library in the front end we use to manage the user session.
According to the docs, a new env variable NEXTAUTH_SECRET is required.
@alexeh I was not sure if we also have to edit the file kubernetes/modules/client/main.tf
@alexeh I was not sure if we also have to edit the file kubernetes/modules/client/main.tf
Whatever you need during image build, needs to be injected via GH. I understand what are you pointing out, actually we think those Kubernetes env vars are not required so once we have time to test it, we are gonna delete them and see if anything explodes
|
gharchive/pull-request
| 2022-03-01T09:24:10 |
2025-04-01T04:55:45.857837
|
{
"authors": [
"alexeh",
"davidsingal"
],
"repo": "Vizzuality/landgriffon",
"url": "https://github.com/Vizzuality/landgriffon/pull/279",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1681377969
|
Sync fork with upstream
This PR rebases off of strangelove-ventures/lens, and merges changes from main.
This gives us the cosmos-sdk upgrades that we need for https://github.com/VolumeFi/paloma/issues/115
This is ready for review. There is one remaining bug in that goreleaser fails for darwin_arm64. This bug is actually present in upstream strangelove-ventures/lens as well. It can be addressed in a follow-up PR.
Got it. Looks like it wasn't that bad and didn't require extensive refactoring of Matijia's updates. Good news
@ToasterTheBrave do we want to remove the offending goreleaser for darwin here too?
|
gharchive/pull-request
| 2023-04-24T14:15:18 |
2025-04-01T04:55:45.892036
|
{
"authors": [
"ToasterTheBrave",
"taariq"
],
"repo": "VolumeFi/lens-1",
"url": "https://github.com/VolumeFi/lens-1/pull/7",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1353307518
|
chore(components): implement new note design
Closes #581
there is an initiative to conduct a convention for how rectangles are designed across the components system (badge, button, nav-item, text field, etc').
I call it appearance model and it features the following predefined categories -
filled
subtle
outlined
duotone
ghost
each of them may require additional subset of states like - idle, disabled, hover, active, selected etc’.
This was never thoroughly progressed by design but is implemented properly in development.
Aligning to that convention by setting the note to a subtle appearance value produces the following
with its current design, the note doesn't fall under any of the available categories in the set. It isn’t really close to any of the above which should raise a flag or at least a discussion.
an natural action should either be creating a new value for a rectangle model (for this new design to fit in), or we align the note to an existing value and change it's new appearance to comply with what is already part of the convention.
Big question is - does its new appearance have any common ground with the rest? was it design arbitrarily? was it designed with the above described convention in mind?
@AyalaBu @rachelbt
|
gharchive/pull-request
| 2022-08-28T10:11:26 |
2025-04-01T04:55:45.897537
|
{
"authors": [
"yinonov"
],
"repo": "Vonage/vivid-3",
"url": "https://github.com/Vonage/vivid-3/pull/634",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2454595346
|
Adding reprojection step in calculate_stats if geotiffs' and stats borders' crs don't match; COUNTRY=mozambique
Description
This fixes #1298 by reprojecting MODIS rasters to match the coordinates used by our stats borders.
How to test the feature:
[ ] Open Mozambique and navigate to analysis tab
[ ] Select a MODIS derived hazard layer
[ ] Run analysis
[ ] Verify the stats match what you'd expected based on the raster
Checklist - did you ...
Test your changes with
[ ] REACT_APP_COUNTRY=rbd yarn start
[ ] REACT_APP_COUNTRY=cambodia yarn start
[x] REACT_APP_COUNTRY=mozambique yarn start
[ ] Add / update necessary tests?
[ ] Add / update outdated documentation?
Screenshot/video of feature:
Nice work @gislawill. I think it's probably worth adding a few comments throughout the code to explain what's happening.
I would say though that I am tempted to do the reprojection in the STAC processing functions of our backend. It feels a bit more homogeneous that the .tif users download from PRISM are always using the same projection, i.e. ESPG 4326
And let's add one or two tests around this if possible :)
I would say though that I am also tempted to do the reprojection in the STAC processing functions of our backend. It feels a bit more homogeneous that the .tif users download from PRISM are always using the same projection, i.e. ESPG 4326
Good call, I shifted this functionality over to the generate_geotiff_from_stac_api method and simplified the functionality. Now, if the raster get from STAC isn't in EPSG:4326, we project it to EPSG:4326 using rioxarray as recommended by @lowriech. rioxarray readily handles the sinusoidal projection with the custom transformation needed for rasterio's reprojection.
And let's add one or two tests around this if possible :)
Sure thing, I'll add a couple before raising for review
This is ready for full review @ericboucher and @lowriech
Looks great so far! Let's talk more about mocking the data, so that we have something static to test on
Looks great so far! Let's talk more about mocking the data, so that we have something static to test on
Thanks @lowriech. I'm curious to hear more of your thoughts on this. I think that we have a pretty static set up currently by using a stored image and mocking all of our s3 calls (here). Curious if you've seen it done differently/better?
|
gharchive/pull-request
| 2024-08-08T00:40:40 |
2025-04-01T04:55:45.940003
|
{
"authors": [
"ericboucher",
"gislawill",
"lowriech"
],
"repo": "WFP-VAM/prism-app",
"url": "https://github.com/WFP-VAM/prism-app/pull/1327",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
234031852
|
Adapt to #41
Remove references to settable HttpOnly as this (no longer?) works in modern browsers https://github.com/WICG/async-cookies-api/issues/41
@craigfrancis does this resolve #41 ?
I can't check on my computer for a couple of days, but that looks about right... thanks for the quick update :-)
Just had a quick look over the repo, and that looks good to me... thanks @bsittler
|
gharchive/pull-request
| 2017-06-06T21:34:47 |
2025-04-01T04:55:45.942183
|
{
"authors": [
"bsittler",
"craigfrancis"
],
"repo": "WICG/async-cookies-api",
"url": "https://github.com/WICG/async-cookies-api/pull/43",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2341821186
|
EOPNOTSUPP using raspberry pi I2C interface
I am playing around with an PN532 NFC HAT on a RaspberryPi 3B using I2C. I got everything to compile, but always received EOPNOTSUPP when trying to do anything with the device. The linux example code worked fine.
After some debugging I narrowed the problem down to this line: https://github.com/WMT-GmbH/pn532/blob/25a0e4dc7f21dd59786a96436b5becdc95568223/src/i2c.rs#L64
For some reason, the raspberry i2c (or the implementation from linux-embedded-hal?) does not like the two Read Commands. I changed the code to this:
fn read(&mut self, buf: &mut [u8]) -> Result<(), Self::Error> {
let mut kbuf = vec![0; buf.len() + 1];
let res = self.i2c.exec(
I2C_ADDRESS,
&mut [Operation::Read(&mut kbuf)],
);
buf.copy_from_slice(&kbuf[1..]);
res
}
And with this, everything works wonderfully. The implementation is somewhat ugly because of the temporary buffer and the copying over, but I did not find another way. That said, I am not very experienced with I2C, so maybe I am doing it wrong?
I am happy to provide a merge request if the above solution is acceptable to you.
Hi, thanks for writing and taking the time to analyze this.
From looking at the linux embedded hal crate this could be caused by https://github.com/rust-embedded/linux-embedded-hal/issues/82
Unfortunately we can't use your fix, since vec! won't be available on no_std targets.
Maybe we could look into resolving the linux driver issue instead
While I in principal agree with you that the fix in the linked repo would be right place to fix this. However, this is above my technical abilities and I still need a working solution for my project. So my suggestion is to do a workaround for a moment in this repo. I see two paths:
Extend the static buffer by one byte and read into this, then drop the first byte when doing anything with the data. This is a bit cumbersome in the code right now, but I think could be made to work.
Use conditional compilation (i.e. if no-std is on, use the current code, otherwise use my suggested implementation). I think this is the easiest workaround for now and should not cause further issues, because linux-embedded kinda implies that std is available.
What is your preferred solution? I can try to provide a patch.
Looking further into it there seems to be a fix ready: https://github.com/rust-embedded/rust-i2cdev/pull/85
Maybe you could ask the devs of rust-i2cdev and linux-embedded-hal to publish a new release.
That being said, if you'd like I would accept a PR with your fix gated behind a linux-i2c-compat feature. (that feature would also enable the std feature)
Note that we'd probably remove that fix in a future release, when the upstream crates published new releases
Using linux-embedded-hal = "0.4.0" will pull in i2cdev = "0.6.1" which contains the fix you need.
I'll close this issue but if you encounter further problems, please reach out
|
gharchive/issue
| 2024-06-08T20:21:18 |
2025-04-01T04:55:46.031150
|
{
"authors": [
"SirVer",
"dimpolo"
],
"repo": "WMT-GmbH/pn532",
"url": "https://github.com/WMT-GmbH/pn532/issues/19",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
231010260
|
Spelling mistake memeory
memeory should be memory.
https://github.com/codersantosh/ns-theme-check/blob/master/inc/checks.php#L99
@codersantosh This can be closed, we've made some changes to the way we're sniffing stuff so this has been taken care of :)
|
gharchive/issue
| 2017-05-24T11:38:25 |
2025-04-01T04:55:46.053365
|
{
"authors": [
"codersantosh",
"dingo-d"
],
"repo": "WPTRT/ns-theme-check",
"url": "https://github.com/WPTRT/ns-theme-check/issues/49",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1654822495
|
Non-Functional Tests ALL
https://github.com/WSUCptSCapstone-Fall2022Spring2023/msft-imagedataprocessing/issues/129
https://github.com/WSUCptSCapstone-Fall2022Spring2023/msft-imagedataprocessing/issues/128
https://github.com/WSUCptSCapstone-Fall2022Spring2023/msft-imagedataprocessing/issues/127
https://github.com/WSUCptSCapstone-Fall2022Spring2023/msft-imagedataprocessing/issues/126
https://gyazo.com/bfc62cd40069fb9f3eedc8bc66f325dd
https://gyazo.com/986672cd91ab955ec1b5a678f99ce109
@zach6215 View links and close this issue. Then I will close the child issues.
Good job
|
gharchive/issue
| 2023-04-05T02:23:44 |
2025-04-01T04:55:46.065532
|
{
"authors": [
"jordanzyx",
"zach6215"
],
"repo": "WSUCptSCapstone-Fall2022Spring2023/msft-imagedataprocessing",
"url": "https://github.com/WSUCptSCapstone-Fall2022Spring2023/msft-imagedataprocessing/issues/135",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1697920104
|
Merge manual prod deployment maptiler update
Problem
Maptiler in the Prod environment has been using Sagging.
Solution
We added Maptiler variable in the manual-deploy-prod.yml file that points to the Prod Maptiler URL
@Rhipeus, we're unsure how you want to handle your last merge into Master and how to keep it away from Prod.
Let's chat when you have the time.
|
gharchive/pull-request
| 2023-05-05T17:01:14 |
2025-04-01T04:55:46.067069
|
{
"authors": [
"amabdallah"
],
"repo": "WSWCWaterDataExchange/WestDAAT",
"url": "https://github.com/WSWCWaterDataExchange/WestDAAT/pull/288",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
487755698
|
fix(npm-script): npm run build で でsassのコンパイルもする
READMEの通りにすると、sassが生成されなかったので、rootのpackage.jsonの npm run build でsassまで生成するようにしました。
本当はts-nodeでTypeScriptをJavaScriptにする作業を(見た目上)省略できるのが理想だと思いますが、それは私の方で進めようと思います。
@matsuyuki-a
別に compile コマンドは、開発時に sass のコンパイルといった、アセットの生成だけを行う処理が必要で設けていました。
compile を削除して build にまとめた場合、build だとビルドの実行ファイルの生成もあるので時間が掛かる傾向だと思うのですが、 sass のコンパイルだけを行いたい場合はどのように実行しようと考えていますか?
確かに説明不足でした、
コマンドラインで下位パッケージのディレクトリに入り、そこでnpm run compile を想定しています
@matsuyuki-a
なるほど。
npm run start 実行中に sass の変更を反映したい場合は少し手間になりますね。 🤔
cd packages/styles
npm run compile
cd ../engine
npm run deploy
そう考えると、リポジトリの切り方を考えた方がよさそうですね
sassもengineの中に入れちゃうのがいいと思います
理想論で言うと、engineからビューを切り離して、そこにスタイル突っ込むのが理想ですが、そこまでは今は厳しそう
@matsuyuki-a
なるほど。
リポジトリの切り方になってくると別に時間を設けたほうが良さそうなので、ひとまず今の状態で進みましょう。
了解ですー
|
gharchive/pull-request
| 2019-08-31T10:25:37 |
2025-04-01T04:55:46.071680
|
{
"authors": [
"aokashi",
"matsuyuki-a"
],
"repo": "WWAWing/WWAWing",
"url": "https://github.com/WWAWing/WWAWing/pull/150",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1402265380
|
Fix Eslint github action
Describe the bug
The Github actions are failing... maybe the plugin is old?
To Reproduce
Steps to reproduce the behavior:
Go to the PRs
Click on a PR
Scroll down to the checks and see the failed check
See an error on hallee/eslint https://github.com/WWCodeTokyo/speak-her-db/actions/runs/3214168172/jobs/5254367000
Expected behavior
Eslint check should run
Screenshots
hallee/eslint is deprecated in favor of https://github.com/actions/toolkit/blob/master/docs/problem-matchers.md
I removed hallee/eslint but still see a failed eslint check. Is this some netlify thing? I can't figure out how to turn it off
@ann-kilzer should we close this issue or do you still want to use eslint at some point?
|
gharchive/issue
| 2022-10-09T12:51:51 |
2025-04-01T04:55:46.087912
|
{
"authors": [
"RageZBla",
"ann-kilzer"
],
"repo": "WWCodeTokyo/speak-her-db",
"url": "https://github.com/WWCodeTokyo/speak-her-db/issues/280",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
852062150
|
いいところ診断の結果に優しさを追加したい
以下の結果を追加したい
'{userName}のいいところは優しさです。あなたの優しい雰囲気や立ち振る舞いに多くの人が癒されています。'
こちらから対応します。
|
gharchive/issue
| 2021-04-07T06:35:28 |
2025-04-01T04:55:46.099316
|
{
"authors": [
"Wallacedudu"
],
"repo": "Wallacedudu/assessment",
"url": "https://github.com/Wallacedudu/assessment/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2113721469
|
Pass chainId to registerRequestHandler
Is your feature request related to a problem? Please describe.
When having multiple EVM chains connected, it is hard to determine which chain to interact with
Describe the solution you'd like
Pass chainId parameter to registerRequestHandler handler function. Since we already have it in _onSessionRequest, it won't be hard to implement.
The following method passes the session topic and request parameters to the handler. It would be nice to also get the chainId
@override
void registerRequestHandler({
required String chainId,
required String method,
dynamic Function(String, dynamic)? handler,
}) {
try {
return signEngine.registerRequestHandler(
chainId: chainId,
method: method,
handler: handler,
);
} catch (e) {
rethrow;
}
}
Ok but what do you mean by "having multiple EVM connected"? You can not have "multiple EVM chains connected". Your paired session between dapp/wallet always works on the current active chain.
Yes, you can only have one active chain at a time, but you can pass multiple chains as optional namespaces.
Created a pull request that solves this issue
Hello @Rimantovas! Here you can see the PR that would fix your issue! TL;DR: After this PR is merged you will have to handle your methods with onSessionRequest subscription (as it was supposed to be originally but was broken). method handlers registration will still work perfectly but if you want to access the current chainId then onSessionRequest subscription is your way to go https://github.com/WalletConnect/WalletConnectFlutterV2/pull/269
|
gharchive/issue
| 2024-02-01T23:36:45 |
2025-04-01T04:55:46.103081
|
{
"authors": [
"Rimantovas",
"quetool"
],
"repo": "WalletConnect/WalletConnectFlutterV2",
"url": "https://github.com/WalletConnect/WalletConnectFlutterV2/issues/258",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1891944834
|
Connection is broken after ~1 min of inactivity
The connection is broken after ~1 min of inactivity, and I get some TaskCanceledExceptions in console.
After this, the request is sent to my phone, but the response is not received back.
I am using nuget package version 2.1.0
Can you provide an example that reproduces the issue? Can you provide the full log you get in the console relating to the TaskCanceledExceptions
After this, the request is sent to my phone, but the response is not received back.
Do you mean after you get the TaskCanceledExceptions in the console, you get the request sent to your phone?
Can you provide an example that reproduces the issue? Can you provide the full log you get in the console relating to the TaskCanceledExceptions
I get something like this in the console.
Steps to reproduce the issue:
Scan QR and approve the connection
Wait >1 min (don't make any requests)
Make a request after 1-2 mins
Request is received on my phone
Response is not received back
Do you mean after you get the TaskCanceledExceptions in the console, you get the request sent to your phone?
Yes. I suppose that there are some background tasks canceled or something because of the exceptions and that's why the response is not received.
Same issue here, only see the thread has exited logs. Nothing more.
Closing this as it is a duplicate of #79, let's continue our discussion on that thread.
|
gharchive/issue
| 2023-09-12T08:20:45 |
2025-04-01T04:55:46.108514
|
{
"authors": [
"VR-Innovations",
"axenteoctavian",
"gigajuwels"
],
"repo": "WalletConnect/WalletConnectSharp",
"url": "https://github.com/WalletConnect/WalletConnectSharp/issues/117",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2351058715
|
:feat exclude wallet RDNS
Changes
feat: Added a new excludeWalletRDNS param. This array contains RDNS entries for wallet domain names. Previously, developers could not exclude wallets injected from EIP6963. Now, they can add the RDNS names of wallets they wish to filter out.
fixes https://linear.app/walletconnect/issue/APKT-473/customers-cannot-override-6963-to-exclude-non-compatible-wallets
Closing as this is not the ideal solution
|
gharchive/pull-request
| 2024-06-13T12:46:57 |
2025-04-01T04:55:46.113538
|
{
"authors": [
"svenvoskamp"
],
"repo": "WalletConnect/web3modal",
"url": "https://github.com/WalletConnect/web3modal/pull/2414",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1747545221
|
🛑 Administrative Portal is down
In f31c860, Administrative Portal ($PROD_ADMIN_URL) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Administrative Portal is back up in 247e2ba.
|
gharchive/issue
| 2023-06-08T10:05:42 |
2025-04-01T04:55:46.115799
|
{
"authors": [
"apratheek"
],
"repo": "WalletInc/status",
"url": "https://github.com/WalletInc/status/issues/202",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2282162192
|
Where can I find the /biobert-pytorch/embedding/getbiovec.sh file?
In the get_embedding_sh function within the pre-processing.py file, the getbiovec.sh file cannot be found. The file is also not located within the biobert-pytorch GitHub repository.
I have the same problem.
Hi,
getbiovec.sh file is the file with disclosed code written in the biobert-pytorch GitHub repository and has been linked in the readme: https://github.com/dmis-lab/biobert-pytorch/tree/master/embedding.
The details are as follows:
If you are willing to cooperate, please feel free to contact me
Warm regards,
Yuchen
|
gharchive/issue
| 2024-05-07T02:50:25 |
2025-04-01T04:55:46.121180
|
{
"authors": [
"WangyuchenCS",
"yefeng241302",
"zhangxudongbio"
],
"repo": "WangyuchenCS/scGREAT",
"url": "https://github.com/WangyuchenCS/scGREAT/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2704455297
|
Comment commands like gcc and gc{motion} do not work within kickstart.nvim
Problem
Comment commands within kickstart.nvim partially do not work (empty checkboxes indicate issues):
[ ] gcc
[ ] gc{motion}
[x] {Visual}gc
Steps to Reproduce
Install kickstart.nvim.
Add the following to your configuration:
-- ...
require('lazy').setup {
{
'Wansmer/langmapper.nvim',
lazy = false,
priority = 1, -- High priority is needed if you will use `autoremap()`
config = function()
local function escape(str)
-- You need to escape these characters to work correctly
local escape_chars = [[;,."|\]]
return vim.fn.escape(str, escape_chars)
end
-- Recommended to use lua template string
local en = [[`qwertyuiop[]asdfghjkl;'zxcvbnm]]
local ru = [[ёйцукенгшщзхъфывапролджэячсмить]]
local en_shift = [[~QWERTYUIOP{}ASDFGHJKL:"ZXCVBNM<>]]
local ru_shift = [[ËЙЦУКЕНГШЩЗХЪФЫВАПРОЛДЖЭЯЧСМИТЬБЮ]]
vim.opt.langmap = vim.fn.join({
-- | `to` should be first | `from` should be second
escape(ru_shift) .. ';' .. escape(en_shift),
escape(ru) .. ';' .. escape(en),
}, ',')
require('langmapper').setup({--[[ your config ]]})
end,
},
-- ...
require('langmapper').automapping({ buffer = false })
-- The line beneath this is called `modeline`. See `:help modeline`
--- vim: ts=2 sts=2 sw=2 et
Solution
You can resolve this issue in one of two ways:
Use global automapping
require('langmapper').automapping({ global = true, buffer = false })
-- The line beneath this is called `modeline`. See `:help modeline`
--- vim: ts=2 sts=2 sw=2 et
Use automapping in the configuration of the echasnovski/mini.nvim plugin
Click to expand
-- ...
{ -- Collection of various small independent plugins/modules
'echasnovski/mini.nvim',
config = function()
-- Better Around/Inside textobjects
--
-- Examples:
-- - va) - [V]isually select [A]round [)]paren
-- - yinq - [Y]ank [I]nside [N]ext [Q]uote
-- - ci' - [C]hange [I]nside [']quote
require('mini.ai').setup { n_lines = 500 }
-- Add/delete/replace surroundings (brackets, quotes, etc.)
--
-- - saiw) - [S]urround [A]dd [I]nner [W]ord [)]Paren
-- - sd' - [S]urround [D]elete [']quotes
-- - sr)' - [S]urround [R]eplace [)] [']
require('mini.surround').setup()
-- Simple and easy statusline.
-- You could remove this setup call if you don't like it,
-- and try some other statusline plugin
local statusline = require 'mini.statusline'
-- set use_icons to true if you have a Nerd Font
statusline.setup { use_icons = vim.g.have_nerd_font }
-- You can configure sections in the statusline by overriding their
-- default behavior. For example, here we set the section for
-- cursor location to LINE:COLUMN
---@diagnostic disable-next-line: duplicate-set-field
statusline.section_location = function()
return '%2l:%-2v'
end
-- ... and there is more!
-- Check out: https://github.com/echasnovski/mini.nvim
-- ADD THIS LINE
-- ⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄
require('langmapper').automapping({ buffer = false })
end,
},
-- ...
Note to Author
You can close this issue as solved.
Solution
You can resolve this issue in one of two ways:
Use global automapping
require('langmapper').automapping({ global = true, buffer = false })
-- The line beneath this is called `modeline`. See `:help modeline`
--- vim: ts=2 sts=2 sw=2 et
Use automapping in the configuration of the echasnovski/mini.nvim plugin
Click to expand
-- ...
{ -- Collection of various small independent plugins/modules
'echasnovski/mini.nvim',
config = function()
-- Better Around/Inside textobjects
--
-- Examples:
-- - va) - [V]isually select [A]round [)]paren
-- - yinq - [Y]ank [I]nside [N]ext [Q]uote
-- - ci' - [C]hange [I]nside [']quote
require('mini.ai').setup { n_lines = 500 }
-- Add/delete/replace surroundings (brackets, quotes, etc.)
--
-- - saiw) - [S]urround [A]dd [I]nner [W]ord [)]Paren
-- - sd' - [S]urround [D]elete [']quotes
-- - sr)' - [S]urround [R]eplace [)] [']
require('mini.surround').setup()
-- Simple and easy statusline.
-- You could remove this setup call if you don't like it,
-- and try some other statusline plugin
local statusline = require 'mini.statusline'
-- set use_icons to true if you have a Nerd Font
statusline.setup { use_icons = vim.g.have_nerd_font }
-- You can configure sections in the statusline by overriding their
-- default behavior. For example, here we set the section for
-- cursor location to LINE:COLUMN
---@diagnostic disable-next-line: duplicate-set-field
statusline.section_location = function()
return '%2l:%-2v'
end
-- ... and there is more!
-- Check out: https://github.com/echasnovski/mini.nvim
-- ADD THIS LINE
-- ⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄⌄
require('langmapper').automapping({ buffer = false })
end,
},
-- ...
There are many plugins in kickstart.nvim that can affect the setting of key bindings. The default gcc and gc works together with langmapeer on an empty config.
To continue working on this issue, please provide the minimal config where the problem is reproduced (not the whole kickstart.nvim).
local lazypath = vim.fn.stdpath 'data' .. '/lazy/lazy.nvim'
if not (vim.uv or vim.loop).fs_stat(lazypath) then
local lazyrepo = 'https://github.com/folke/lazy.nvim.git'
local out = vim.fn.system { 'git', 'clone', '--filter=blob:none', '--branch=stable', lazyrepo, lazypath }
if vim.v.shell_error ~= 0 then
error('Error cloning lazy.nvim:\n' .. out)
end
end ---@diagnostic disable-next-line: undefined-field
vim.opt.rtp:prepend(lazypath)
vim.opt.number = true
require('lazy').setup({
{
'Wansmer/langmapper.nvim',
lazy = false,
priority = 1, -- High priority is needed if you will use `autoremap()`
init = function()
local function escape(str)
-- You need to escape these characters to work correctly
local escape_chars = [[;,."|\]]
return vim.fn.escape(str, escape_chars)
end
-- Recommended to use lua template string
local en = [[`qwertyuiop[]asdfghjkl;'zxcvbnm]]
local ru = [[ёйцукенгшщзхъфывапролджэячсмить]]
local en_shift = [[~QWERTYUIOP{}ASDFGHJKL:"ZXCVBNM<>]]
local ru_shift = [[ËЙЦУКЕНГШЩЗХЪФЫВАПРОЛДЖЭЯЧСМИТЬБЮ]]
vim.opt.langmap = vim.fn.join({
-- | `to` should be first | `from` should be second
escape(ru_shift) .. ';' .. escape(en_shift),
escape(ru) .. ';' .. escape(en),
}, ',')
end,
config = function()
require('langmapper').setup {}
require('langmapper').hack_get_keymap()
end,
},
{
'folke/which-key.nvim',
-- event = 'VimEnter',
},
})
require('langmapper').automapping { buffer = false }
-- The line beneath this is called `modeline`. See `:help modeline`
-- vim: ts=2 sts=2 sw=2 et
So, this is because you using which-key. Trying this solution: https://github.com/Wansmer/langmapper.nvim/discussions/11#discussioncomment-11210683
I close the issue because it is not related to langmapper, but to an external dependency
|
gharchive/issue
| 2024-11-29T09:11:50 |
2025-04-01T04:55:46.131299
|
{
"authors": [
"Wansmer",
"drowning-cat"
],
"repo": "Wansmer/langmapper.nvim",
"url": "https://github.com/Wansmer/langmapper.nvim/issues/42",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2057178641
|
env.ini
Hello, thank you for sharing your amazing work!
I've an issue, I'm unable to load the page, it automatically redirect to ./public/{local}/notsupported page.
I addressed the issue to env.ini file location, probably. As you can see it's unable to "convert" $ENV into "assets/" position.
I tried to place env.ini everywhere without success.
Any hint ?
Thanks
The value of the path setting should be:
The env.ini file should be in the root directory where the composer.json file is located.
The value of the path setting should be:
The env.ini file should be in the root directory where the composer.json file is located.
I've done that, but nothing changed.
I tried to restart the server, re-compile it, but nothing.
I modified the file config.ini located in app/config
deleted the browser cache, and now it works :D
Thank you again for your work!
|
gharchive/issue
| 2023-12-27T10:10:20 |
2025-04-01T04:55:46.169758
|
{
"authors": [
"Warsaalk",
"zeck93"
],
"repo": "Warsaalk/trashsim-public",
"url": "https://github.com/Warsaalk/trashsim-public/issues/3",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1992558602
|
[DOCS] missing crun instructions in https://wasmedge.org/docs/develop/deploy/podman/#prerequisites
Description
Summary
Step 3 of https://wasmedge.org/docs/develop/deploy/podman/#prerequisites: It seems like some instructions are missing.
Build and configure crun with WasmEdge support
Next, configure and build a crun binary with WasmEdge support.
Maybe there should be a git clone, similar to https://wasmedge.org/docs/develop/deploy/oci-runtime/crun/#prerequisites?
Issue refers to:
https://github.com/WasmEdge/WasmEdge/issues/3028
Screenshots
No response
Additional information
No response
I'll work on this
|
gharchive/issue
| 2023-11-14T11:31:06 |
2025-04-01T04:55:46.200633
|
{
"authors": [
"dhruvmehtaaa"
],
"repo": "WasmEdge/docs",
"url": "https://github.com/WasmEdge/docs/issues/195",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1307173752
|
Consider putting bucket paths outside the script
Relates to #1
Maybe as env variable. Maybe target path (workspace) also deserves that.
Done due to change in approach for handling workflow
|
gharchive/issue
| 2022-07-17T18:17:53 |
2025-04-01T04:55:46.205701
|
{
"authors": [
"Observer46"
],
"repo": "WaterlinePL/project-download-job",
"url": "https://github.com/WaterlinePL/project-download-job/issues/3",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
839310748
|
工作流中数据开发组件关联脚本选项,选择脚本后提示关联脚本失败 NullPointerException
Describe the bug
工作流中数据开发组件关联脚本选项,选择脚本后提示关联脚本失败 NullPointerException
../dss_linkis/linkis/linkis-publicservice/logs/linkis.log 有如下异常提示
linkis版本0.11.0 dss版本0.9.0
同样的问题,DSS0.9.1 \ linkis0.9.4
Hello, this problem has been solved in DSS DSS-230
|
gharchive/issue
| 2021-03-24T03:35:33 |
2025-04-01T04:55:46.255859
|
{
"authors": [
"lordk911",
"peacewong",
"zoneyeah"
],
"repo": "WeBankFinTech/Linkis",
"url": "https://github.com/WeBankFinTech/Linkis/issues/691",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1385824154
|
Add one key flow example
Adds the one key flow example for Web3Auth. This works using firebase.
Things to note while documenting: Need to understand how the users are being classified - accordingly when to call torus libraries and when to call web3auth. webpack config changes. this works only on custom verifiers, with custom jwts - why.
|
gharchive/pull-request
| 2022-09-26T10:29:16 |
2025-04-01T04:55:46.271095
|
{
"authors": [
"yashovardhan"
],
"repo": "Web3Auth/examples",
"url": "https://github.com/Web3Auth/examples/pull/9",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
266545870
|
Create pritamnegi.txt
This Pull Request resolves Issue #1
Thanks for contributing!
|
gharchive/pull-request
| 2017-10-18T15:52:57 |
2025-04-01T04:55:46.308323
|
{
"authors": [
"MJ10",
"pritamnegi"
],
"repo": "WebClub-NITK/Hacktoberfest-2k17",
"url": "https://github.com/WebClub-NITK/Hacktoberfest-2k17/pull/507",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1194258814
|
🛑 SSO is down
In b6365a5, SSO (https://sso.zeltik.es/auth) was down:
HTTP code: 0
Response time: 0 ms
Resolved: SSO is back up in 883f317.
|
gharchive/issue
| 2022-04-06T08:29:50 |
2025-04-01T04:55:46.833554
|
{
"authors": [
"aqp696"
],
"repo": "Webdeporte/statuspage",
"url": "https://github.com/Webdeporte/statuspage/issues/82",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
65987760
|
w-include - dynamic path
User variables as include path.
The child (included) template should be parsed by passing local-scope and global-scope variables to it. The result of that should be put in place of the w-include tag.
|
gharchive/issue
| 2015-04-02T18:44:35 |
2025-04-01T04:55:46.834538
|
{
"authors": [
"SvenAlHamad"
],
"repo": "Webiny/Htpl",
"url": "https://github.com/Webiny/Htpl/issues/24",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2251280092
|
How can I get email from attribute object?
Hello,
I need to get the email from Webklex\PHPIMAP\Attribute, when I die my object dd($this->email->from) I got this:
Webklex\PHPIMAP\Attribute^ {#844
#name: ? string
#values: []
+"0": Webklex\PHPIMAP\Address^ {#849
+personal: "Kouceyla Hadji"
+mailbox: "kouceyla"
+host: "domain.com"
+mail: "kouceyla@domain.com"
+full: "Kouceyla Hadji <kouceyla@domain.com>"
}
}
I need to get "mail"
Would be $this->email->getFrom()->first()->mail in your case. Cheers :)
Thanks :)
|
gharchive/issue
| 2024-04-18T18:14:10 |
2025-04-01T04:55:46.837097
|
{
"authors": [
"Julian1998",
"kossa"
],
"repo": "Webklex/php-imap",
"url": "https://github.com/Webklex/php-imap/issues/488",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
202395473
|
Add external/custom fonts
How to add custom font files?
https://gist.github.com/Turbo87/e8e941e68308d3b40ef6
Following the original starter project, it's updated and they resolve issues
https://github.com/AngularClass/angular2-webpack-starter
|
gharchive/issue
| 2017-01-22T17:08:56 |
2025-04-01T04:55:46.857727
|
{
"authors": [
"klihelp"
],
"repo": "Websitescity/ng2-f6-starter",
"url": "https://github.com/Websitescity/ng2-f6-starter/issues/3",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1410536188
|
Could not load file or assembly Col.Mod.Better.Shared
Hello, I'm not able to load this mod in the latest version of the game. Folder name checks, all vanilla, integrity checked. Log attached.
22-10-16_18-06-36_9583.log
You has not copy "Col.Mod.Better.Shared.dll" in mod root directory. Add the file from last realease and is working.
This mod hasnt assets.
Log please.
i get the same error mentioned in the beginning, the .shared.dll file does exist in the coi.mod.better folder as downloaded in the release
22-10-27_06-04-00_3357.log
I am getting the same error as well with new save.
|
gharchive/issue
| 2022-10-16T16:11:37 |
2025-04-01T04:55:46.870524
|
{
"authors": [
"Cengro",
"Wehmeyer100",
"adiamaq",
"askewdread"
],
"repo": "Wehmeyer100/CoI.Mod.Better",
"url": "https://github.com/Wehmeyer100/CoI.Mod.Better/issues/34",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1555032989
|
trying eqwalizer on elixir projects
We wanted to try eqwalizer on our elixir projects. I know it would be a lot better to implement a parser for Elixir projects, but as a first step we decompiled the beam files to erlang sources, and
we were running eqwalizer on those sources.
We have found some problems:
in the decompiled source there is a -file attribute:
-file("lib/asdfqqqq.ex", 1).
-module('Elixir.Asdfqqqq').
When this attribute is there eqwalizer simply ignores all problems. If I manually delete this line, it works.
there is also an info/1 function in the source:
'__info__'(module) -> 'Elixir.Asdfqqqq';
'__info__'(functions) -> [{double, 1}];
'__info__'(macros) -> [];
'__info__'(exports_md5) -> <<"\035S8}I7\030\232\236g-\023">>;
this last function clause will cause this problem:
thread 'main' panicked at 'called `Result::unwrap()` on an `Err` value: Utf8Error { valid_up_to: 429, error_len: Some(1) }', crates/elp/src/bin/[reporting.rs:101](http://reporting.rs:101/):78
note: run with `RUST_BACKTRACE=1` environment variable to display a backtrace
actually, when I manually fix these problems eqwalizer works quite well, and it is able to spot the problems in the code (of course the line numbers are wrong). One remaining problem is that I could not figure out how to tell to eqwalizer where the declarations of Elixir types are. So for instance:
-spec double({error, 1}) -> 'Elixir.String':t().
double(_x@1) -> 2 * _x@1.
Elixir.String type is declared in $ELIXIR_DIR//lib/elixir/lib/string.ex. Similarly, I want to add all files where elixir types are declared. How can I add those using project.json?
I have one more question...
Let's say I have my_app application, which uses my_lib as a dependency. In my_app.erl I have the following:
-include_lib("my_lib/include/my_lib.hrl").
I've tried it with this project.json:
{
"apps": [
{
"name": "my_app",
"dir": "",
"ebin": "_build/default/lib/my_app/ebin",
"extra_src_dirs": [""],
"include_dirs": ["include"],
"macros": ["TEST"],
"src_dirs": ["src"]
}
],
"deps": [
{
"name": "my_lib",
"dir": "",
"ebin": "_build/default/lib/my_lib/ebin",
"extra_src_dirs": [""],
"include_dirs": ["_build/default/lib/my_lib/include"],
"macros": ["TEST"],
"src_dirs": ["src"]
}
],
"root": ""
}
elp is still complaining about being not able to find the include file:
eqWAlizing ████████████████████ 3/3 error: parse_error
┌─ src/my_app.erl:3:14
│
103 │ -include_lib("my_lib/include/my_lib.hrl").
│ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ can't find include lib "my_lib/include/my_lib.hrl"
Do you have an idea what should I put into my project.json ?
We were trying to make it work using the following project.json :
{
"apps": [
{
"name": "asdfqqqq",
"dir": "/home/joe/Projects/asdfqqqq",
"ebin": "/home/joe/Projects/asdfqqqq/_build/dev/lib/asdfqqqq/ebin",
"extra_src_dirs": ["test"],
"include_dirs": ["include"],
"macros": ["TEST"],
"src_dirs": ["erlangized_src/"]
}
],
"deps": [
{
"name": "elixir",
"dir": "/home/joe/.manager/elixir",
"ebin": "/home/joe/.managerl/erlang-OTP-23.2.3/elixirs/elixir-v1.12.2/lib/elixir/lib/elixir/ebin",
"extra_src_dirs": [],
"include_dirs": [],
"macros": [],
"src_dirs": ["lib"]
}
],
"source_root": "/home/joe/Projects/asdfqqqq",
"otp_lib_dir": "/home/joe/.managerl/erlang-OTP-23.2.3/lib/"
}
Elixir.String.t() type is declared in the above-mentioned folder. With these settings, it still was not able to find the declaration of this type. Is it possible that to find the declaration eqwalizer also needs the erlang source code present in the src directory, or do you have any other ideas on what's going on wrong?
We don't provide any support for type-checking Elixir projects out of the box and we don't plan to do it on our own in the near future
Nevertheless, I think that it's feasible to to use it for Elixir, - but someone needs to drive it. And it's hard to estimate the amount of tech work.
Contributions are welcome.
|
gharchive/issue
| 2023-01-24T13:56:52 |
2025-04-01T04:55:46.896221
|
{
"authors": [
"ilya-klyuchnikov",
"tothlac"
],
"repo": "WhatsApp/eqwalizer",
"url": "https://github.com/WhatsApp/eqwalizer/issues/17",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2648456959
|
Trademark infringement complain from enf.whatsapp.2548904@enf-meta.com
A few month ago I developed any app Status Saver to save WhatsApp Status but I got
Trademark Owner:
WhatsApp LLC
Complainant Email:
enf.whatsapp.2548904@enf-meta.com
Complainant Details:
The app uses the trademarks, banners, badges and logos of WhatsApp without authorization, including, but not limited to, the use of ley elements of WhatsApp's Famous Trademarked Logo in the app's icon. The developer’s use of WhatsApp's trademarks in app’s title and/or icon implies an association or other relationship that is neither authorized nor endorsed by WhatsApp.
I tried to contact them throw email but it has been 5 days but no reply from WhatsApp LLC. I am ready to do all required changes in app. I have done lots of hard work to develop this app. Please help me out.
Hi!
This repo is specifically for proxy related issues and I am not familiar with the infringement complain process unfortunately. I'll close this task.
|
gharchive/issue
| 2024-11-11T07:37:47 |
2025-04-01T04:55:46.899690
|
{
"authors": [
"LokikSoni-Dev",
"eozturk1"
],
"repo": "WhatsApp/proxy",
"url": "https://github.com/WhatsApp/proxy/issues/329",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2503042203
|
🛑 Munin is down
In e706e0c, Munin (https://munin.wicloz.rocks/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Munin is back up in 3e7b96c after 10 minutes.
|
gharchive/issue
| 2024-09-03T14:30:43 |
2025-04-01T04:55:46.930957
|
{
"authors": [
"Wicloz"
],
"repo": "Wicloz/upptime",
"url": "https://github.com/Wicloz/upptime/issues/709",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
136818373
|
Add color functions, improve variables and imports
imports now work in any place in file (i. e. Numix used to use this, see v2.0)
variables logic heavily rewritten (based on postcss-simple-vars)
color functions, powered by postcss-functions
migrate to ES6, add babel
switch ESLint config to better suit your style :)
Thank you for your contribution! If you want to continue contributing to this project, keep in mind that I am only aiming for GTK 3.20 compatibility right now. There have been made some big GTK CSS changes in this new version, so backward compatibility with older theme files is too complicated to be top priority.
|
gharchive/pull-request
| 2016-02-26T21:24:36 |
2025-04-01T04:55:46.936352
|
{
"authors": [
"Wietsedv",
"iamale"
],
"repo": "Wietsedv/postcss-web-gtk",
"url": "https://github.com/Wietsedv/postcss-web-gtk/pull/1",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
69728392
|
added support for custom React ID Attribute
This commit adds support for cases when React is changed to use a custom React ID attribute. In browserify, the react ID can be changed as follows:
require('react/lib/DOMProperty').ID_ATTRIBUTE_NAME = 'data-reactid-...';
So this commit makes sure that ReactShadow gets the right ID Attribute in case its running as a commonjs module
Thanks @ziahamza! :+1:
|
gharchive/pull-request
| 2015-04-21T02:46:35 |
2025-04-01T04:55:46.977501
|
{
"authors": [
"Wildhoney",
"ziahamza"
],
"repo": "Wildhoney/ReactShadow",
"url": "https://github.com/Wildhoney/ReactShadow/pull/13",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
367394332
|
Black and White Output
Hello William! Congratulations, this is a wonderfull script. Well I found something interesting and related with the issue #11 . Every time the program runs, it outputs a mask with only one chanel. That's why it generates black and white images. This is intentional? Thank you very much for this library!
Yes, it's generating a mask. You need to apply a bitwise operator to produce a coloured image. You can find more on stackoverflow.
StackOverflow - opencv-apply-mask-to-a-color-image
|
gharchive/issue
| 2018-10-05T23:03:30 |
2025-04-01T04:55:46.979404
|
{
"authors": [
"Valdecy",
"WillBrennan"
],
"repo": "WillBrennan/SkinDetector",
"url": "https://github.com/WillBrennan/SkinDetector/issues/14",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
173031408
|
Exception: Unclosed quoted field on line 4.
So here's a random edge case... winrm-fs fails when there's an "All Users, All Hosts" type of powershell profile that writes output on the target node. winrm-fs cannot parse the response as CSV successfully because the values written out in the profile will most likely not be properly comma separated.
The fix would probably be to run the script without loading a profile, using the -noprofile switch when calling PowerShell.. but that call appears to be in the winrm definition. So maybe this issue should be logged there?
Anyway... The fix for me is to just remove the write-output from my profile.
To recreate the issue:
get a "destination", for me I just did kitchen create any windows host will work
create a powershell profile script in the remote node's $PSHome directory..
run kitchen converge or some other command that uses the function parse_response
`C:\Windows\System32\WindowsPowerShell\v1.0\Microsoft.PowerShell_profile.ps`` >
Write-Output 'Hello!!'
Run kitchen converge
Here's the trace:
D [WinRM] opening remote shell on http://10.12.235.80:5985/wsman
D [WinRM] remote shell 70A3F068-8133-44F6-9C38-4F8FAA875DDE is open on http://10.12.235.80:5985/wsman
D creating hash for file $env:TEMP/default-Windows-long_script.ps1
D Running check_files.ps1
D @{
D "67907d49d035159f3f525a8fbf5c5368" = @{
D "target" = "$env:TEMP/default-Windows-long_script.ps1";
D "src_basename" = "default-Windows-long_script.ps1";
D "dst" = "$env:TEMP/default-Windows-long_script.ps1"
D }
D }
D Parsing CSV Response
D Resetting $HOMEDRIVE & $HOMEPATH for you...
Initializing the chef shell...
All set!
"chk_exists","src_md5","dst_md5","chk_dirty","verifies","target_is_folder"
"True","67907d49d035159f3f525a8fbf5c5368","67907d49d035159f3f525a8fbf5c5368","False","True","False"
D Cleaning up local sandbox in C:/Users/nhudacin/AppData/Local/Temp/default-Windows-sandbox-20160824-9900-1eijbbu
>>>>>> ------Exception-------
>>>>>> Class: Kitchen::ActionFailed
>>>>>> Message: 1 actions failed.
>>>>>> Failed to complete #converge action: [Unclosed quoted field on line 4.] on default-Windows
>>>>>> ----------------------
>>>>>> Please see .kitchen/logs/kitchen.log for more details
>>>>>> Also try running `kitchen diagnose --all` for configuration
D ------Exception-------
D Class: Kitchen::ActionFailed
D Message: 1 actions failed.
>>>>>> Failed to complete #converge action: [Unclosed quoted field on line 4.] on default-Windows
D ----------------------
D ------Backtrace-------
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/command.rb:187:in `report_errors'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/command.rb:178:in `run_action'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/command/action.rb:39:in `block in call'
D C:/opscode/chefdk/embedded/lib/ruby/2.1.0/benchmark.rb:279:in `measure'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/command/action.rb:37:in `call'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/cli.rb:56:in `perform'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/cli.rb:185:in `block (2 levels) in <class:CLI>'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/thor-0.19.1/lib/thor/command.rb:27:in `run'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/thor-0.19.1/lib/thor/invocation.rb:126:in `invoke_command'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/cli.rb:326:in `invoke_task'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/thor-0.19.1/lib/thor.rb:359:in `dispatch'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/thor-0.19.1/lib/thor/base.rb:440:in `start'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/bin/kitchen:13:in `block in <top (required)>'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/errors.rb:174:in `with_friendly_errors'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/bin/kitchen:13:in `<top (required)>'
D C:/opscode/chefdk/bin/kitchen:21:in `load'
D C:/opscode/chefdk/bin/kitchen:21:in `<main>'
D ----End Backtrace-----
D -Composite Exception--
D Class: Kitchen::ActionFailed
D Message: Failed to complete #converge action: [Unclosed quoted field on line 4.] on default-Windows
D ----------------------
D ------Backtrace-------
D C:/opscode/chefdk/embedded/lib/ruby/2.1.0/csv.rb:1872:in `block in shift'
D C:/opscode/chefdk/embedded/lib/ruby/2.1.0/csv.rb:1779:in `loop'
D C:/opscode/chefdk/embedded/lib/ruby/2.1.0/csv.rb:1779:in `shift'
D C:/opscode/chefdk/embedded/lib/ruby/2.1.0/csv.rb:1721:in `each'
D C:/opscode/chefdk/embedded/lib/ruby/2.1.0/csv.rb:1735:in `to_a'
D C:/opscode/chefdk/embedded/lib/ruby/2.1.0/csv.rb:1735:in `read'
D C:/opscode/chefdk/embedded/lib/ruby/2.1.0/csv.rb:1293:in `parse'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/winrm-fs-0.4.3/lib/winrm-fs/core/file_transporter.rb:404:in `parse_response'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/winrm-fs-0.4.3/lib/winrm-fs/core/file_transporter.rb:203:in `check_files'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/winrm-fs-0.4.3/lib/winrm-fs/core/file_transporter.rb:80:in `block in upload'
D C:/opscode/chefdk/embedded/lib/ruby/2.1.0/benchmark.rb:279:in `measure'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/winrm-fs-0.4.3/lib/winrm-fs/core/file_transporter.rb:78:in `upload'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/transport/winrm.rb:134:in `upload'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/transport/winrm.rb:394:in `run_from_file_command'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/transport/winrm.rb:102:in `execute'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/provisioner/base.rb:72:in `block in call'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/transport/base.rb:94:in `initialize'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/transport/winrm.rb:470:in `new'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/transport/winrm.rb:470:in `create_new_connection'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/transport/winrm.rb:74:in `connection'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/provisioner/base.rb:71:in `call'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/instance.rb:373:in `block in converge_action'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/instance.rb:513:in `call'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/instance.rb:513:in `synchronize_or_call'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/instance.rb:478:in `block in action'
D C:/opscode/chefdk/embedded/lib/ruby/2.1.0/benchmark.rb:279:in `measure'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/instance.rb:477:in `action'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/instance.rb:369:in `converge_action'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/instance.rb:348:in `block in transition_to'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/instance.rb:347:in `each'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/instance.rb:347:in `transition_to'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/instance.rb:138:in `converge'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/command.rb:201:in `public_send'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/command.rb:201:in `run_action_in_thread'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/command.rb:173:in `block (2 levels) in run_action'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/logging-2.1.0/lib/logging/diagnostic_context.rb:450:in `call'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/logging-2.1.0/lib/logging/diagnostic_context.rb:450:in `block in create_with_logging_context'
D ----End Backtrace-----
D ---Nested Exception---
D Class: Kitchen::ActionFailed
D Message: Failed to complete #converge action: [Unclosed quoted field on line 4.]
D ----------------------
D ------Backtrace-------
D C:/opscode/chefdk/embedded/lib/ruby/2.1.0/csv.rb:1872:in `block in shift'
D C:/opscode/chefdk/embedded/lib/ruby/2.1.0/csv.rb:1779:in `loop'
D C:/opscode/chefdk/embedded/lib/ruby/2.1.0/csv.rb:1779:in `shift'
D C:/opscode/chefdk/embedded/lib/ruby/2.1.0/csv.rb:1721:in `each'
D C:/opscode/chefdk/embedded/lib/ruby/2.1.0/csv.rb:1735:in `to_a'
D C:/opscode/chefdk/embedded/lib/ruby/2.1.0/csv.rb:1735:in `read'
D C:/opscode/chefdk/embedded/lib/ruby/2.1.0/csv.rb:1293:in `parse'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/winrm-fs-0.4.3/lib/winrm-fs/core/file_transporter.rb:404:in `parse_response'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/winrm-fs-0.4.3/lib/winrm-fs/core/file_transporter.rb:203:in `check_files'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/winrm-fs-0.4.3/lib/winrm-fs/core/file_transporter.rb:80:in `block in upload'
D C:/opscode/chefdk/embedded/lib/ruby/2.1.0/benchmark.rb:279:in `measure'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/winrm-fs-0.4.3/lib/winrm-fs/core/file_transporter.rb:78:in `upload'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/transport/winrm.rb:134:in `upload'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/transport/winrm.rb:394:in `run_from_file_command'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/transport/winrm.rb:102:in `execute'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/provisioner/base.rb:72:in `block in call'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/transport/base.rb:94:in `initialize'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/transport/winrm.rb:470:in `new'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/transport/winrm.rb:470:in `create_new_connection'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/transport/winrm.rb:74:in `connection'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/provisioner/base.rb:71:in `call'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/instance.rb:373:in `block in converge_action'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/instance.rb:513:in `call'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/instance.rb:513:in `synchronize_or_call'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/instance.rb:478:in `block in action'
D C:/opscode/chefdk/embedded/lib/ruby/2.1.0/benchmark.rb:279:in `measure'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/instance.rb:477:in `action'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/instance.rb:369:in `converge_action'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/instance.rb:348:in `block in transition_to'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/instance.rb:347:in `each'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/instance.rb:347:in `transition_to'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/instance.rb:138:in `converge'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/command.rb:201:in `public_send'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/command.rb:201:in `run_action_in_thread'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/test-kitchen-1.11.1/lib/kitchen/command.rb:173:in `block (2 levels) in run_action'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/logging-2.1.0/lib/logging/diagnostic_context.rb:450:in `call'
D C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/logging-2.1.0/lib/logging/diagnostic_context.rb:450:in `block in create_with_logging_context'
D ----End Backtrace-----
D [WinRM] closing remote shell 70A3F068-8133-44F6-9C38-4F8FAA875DDE on http://10.12.235.80:5985/wsman
D [WinRM] remote shell 70A3F068-8133-44F6-9C38-4F8FAA875DDE closed
PS C:\_source\git\cookbook-consilio-chefdk>
Interesting. Yes we really should be specifying -noprofile in winrm. However we are on the verge of releasing winrm v2 which uses true powershell tempting and does not load user profiles.
|
gharchive/issue
| 2016-08-24T18:59:09 |
2025-04-01T04:55:47.014917
|
{
"authors": [
"mwrock",
"nhudacin"
],
"repo": "WinRb/winrm-fs",
"url": "https://github.com/WinRb/winrm-fs/issues/48",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
53940299
|
Fixed issue 2 - upload a dir multiple times
Fixed issue 2 and redesigned how file upload works and reduced a lot of duplication.
Added PowerShell scripts and templating
Fixed multiple directory uploads and directory dirty checking
Collapsed file decoding to a single script
Use deterministic paths for temp files.
Externalizing as .ps1.erb is an improvement, but getting it to be pure powershell (by using powershell parameters to avoid the need for ERB variables) would be even better.
@maxlinc Now that you mention using PS args, yes that would be a nice improvement.
|
gharchive/pull-request
| 2015-01-10T02:18:44 |
2025-04-01T04:55:47.017747
|
{
"authors": [
"maxlinc",
"sneal"
],
"repo": "WinRb/winrm-fs",
"url": "https://github.com/WinRb/winrm-fs/pull/4",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
128618795
|
Why isn't there a designer WYSIWYG for Hamburger menu items?
Other than the image for the hamburger itself, there doesn't seem to be a designer view of Hamburger menu items and button icons. I was hoping it would render similar to Windows Phone ApplicationBar which has a design-time view.
Is the only option at the moment to handcode XAML and view the results at runtime?
Thanks,
Rick
There is a designer view (in Minimal template at least). Look in Shell.xaml. That's where the hamburger menu is coded, so the design-time view is in there.
Unfortunately, due to the way T10 does things, I don't think there is a way to see the hamburger menu over any of your pages at design-time.
Not sure what could be done different at this point.
|
gharchive/issue
| 2016-01-25T19:36:39 |
2025-04-01T04:55:47.021827
|
{
"authors": [
"JerryNixon",
"callummoffat",
"rickengle"
],
"repo": "Windows-XAML/Template10",
"url": "https://github.com/Windows-XAML/Template10/issues/574",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
320433088
|
Help me use SQLite!
General Help Thread
This thread is for anyone who needs help using SQLite or Node.js with regards to better-sqlite3.
It's common for issues to appear with titles like "Help me do ABC" or "How do I build an app like XYZ?". These types of issues aren't real issues with better-sqlite3, so they will be closed and considered a duplicated of this thread. Therefore, post your questions here instead. If you're lucky, friendly members of the community may chime in here to help you.
By default, SQLite is very bad at performing concurrent reads and writes. However, performance can be improved tremendously by activating WAL mode.
db.pragma('journal_mode = WAL');
For more information, read here.
Is it possible to insert multiple rows from an array of objects? Say I have this:
var data = [
{ foo: 1, bar: 2 },
{ foo: 3, bar: 4 }
];
I'd like to be able to do this:
db.prepare('INSERT INTO baz VALUES ($foo, $bar)').run(data);
Not sure if running INSERT statements in a loop is efficient.
As a side note, sqlite does does support insertion of multiple values.
Eg, INSERT INTO baz VALUES (1, 2), (3, 4) will insert 2 rows.
Using a loop is quite efficient, @dimitry-ishenko.
There are two things to be aware of:
you should prepare the statement only once, and run it multiple times
you can further increase performance by wrapping the loop in a transaction (BEGIN...COMMIT)
to learn how to use manual transactions, see here.
Here's an example of what (I believe) you're trying to do:
const insert = db.prepare('INSERT INTO baz VALUES ($foo, $bar)');
const insertMany = asTransaction((data) => {
for (const obj of data) insert.run(obj);
});
insertMany([
{ foo: 1, bar: 2 },
{ foo: 3, bar: 4 }
]);
Hi, I am curious about the performance of the method you just mentioned @JoshuaWise. How does this compare to actually running a batch write? The tradeoff being if the insertMany method below cannot use a prepared statement.
const insertMany = data => {
const valuesStr = Array(data.length).fill('(?, ?)').join(',')
const sql = `INSERT INTO baz VALUES ${valuesStr}`
const flattened = data.reduce((acc, d) => {
acc.push(Object.values(d))
return acc
}, [])
db.run(sql, flattened)
}
when comparing the two executing directly like so cat insert-many.sql | sqlite3 I believe my method would win out. Wrapped in better-sqlite though, is it faster to run many inserts as opposed to batching to avoid preparing the statement each time?
Hi, can i just use the prepare statement + run statement for creating tables or creating trigger?
const statement = db.prepare('CREATE TABLE IF NOT EXISTS ? (?)');
{some foo}
statement.run(tableName, tableColumns);
@Nikumarekko the only thing wrong with your code is that SQLite3 does not allow you to bind strings for identifier names (such as table or column names). You can only use bound values for actual run-time values.
// No bound parameters for table/column names
const createTable = db.prepare(`CREATE TABLE IF NOT EXISTS ${tableName} (${tableColumns})`);
createTable.run();
// Use bound parameters for actual strings
const select = db.prepare(`SELECT * FROM ${tableName} WHERE foo = ?`);
const results = select.all('someString');
Keep in mind that because you have to use string interpolation for table and column names, they should only come from trusted sources.
@Nikumarekko the only thing wrong with your code is that SQLite3 does not allow you to bind strings for identifier names (such as table or column names). You can only use bound values for actual run-time values.
// No bound parameters for table/column names
const createTable = db.prepare(`CREATE TABLE IF NOT EXISTS ${tableName} (${tableColumns})`);
createTable.run();
// Use bound parameters for actual strings
const select = db.prepare(`SELECT * FROM ${tableName} WHERE foo = ?`);
const results = select.all('someString');
Keep in mind that because you have to use string interpolation for table and column names, they should only come from trusted sources.
Thanks for your quick reply, helps me a lot
I have a file called json.sqlite that stores my data and im not sure how I will VACUUM it.
Also what is the name of my db file if my using quick.db? (quick.db is a wrapper for better-sqlite3)
https://stackoverflow.com/questions/18126997/how-to-vacuum-sqlite-database this is how you vacuum a sqlite file how will i vacuum a better-sqlite3 file?
hi @JoshuaWise
I'm trying to use this module with Express.js but for some reason when I call run on the statement, the http connection gets closed without any error, just upstream prematurely closed connection while reading response header from upstream in nginx log
code is pretty simple
router.post("/path1", function(req, res, next) {
db.prepare(sqlQuery).run(params);
res.redirect("/path2");
})
any suggestion how to find the cause of the problem?
thanks
@bombard-kb better-sqlite3 and express don't interact in any way whatsoever. It could be related to a timeout occurring in express, but it likely is completely unrelated to better-sqlite3.
@JoshuaWise it looks like I was trying to use "on conflict statement" with sqlite3 version, which doesn't support it
How do I use better-sqlite3 with a precompile version of SQLite?
@JefStat better-sqlite3 has to be compiled with SQLite as a static library, not a precompiled dynamic library. You can, however, use a custom amalgamation with your own custom compile-time options. Learn how here
Hi
Can I store an image as a Blob in sqlite/better-sqlite3 to then use URL.createObjectURL(blob) for displaying the image in the render window?
As I am a newbie with sqlite, it might as well be a stupid question :-)
I am developing an electron/desktop app where I am storing a lot of images. Before I have used the browsers internal indexeddb where the images were stored as a Blob. I then created an objectURL which I used to display the images in the render window.
As indexeddb is limited in some ways, I came across better-sqlite3. Storing and reading the images as base64 worked like a charm, but I would have hoped that I do not have to shuffle around so much data and can make use of objectURLs.
Can I have table name as a query parameter? For example, in mysql, I can do CREATE TABLE ?? (col1 varchar, ...) with a parameter array of [tableName]. I saw in the API doc you can insert parameters with the standard ? in a query, but how about parameters that are table names? Does that have it's own symbol, like how mysql has ???
Can I store an image as a Blob in sqlite/better-sqlite3
@fakob, you can store any arbitrary data in a Blob, so the answer is yes.
Can I have table name as a query parameter?
@andrewkolos, SQLite3 provides no way of using parameters for table/column names. However, if you properly quote/escape the table name, you can use JavaScript's string interpolation:
function quote(name) {
return '\x60' + String(name).replace(/\x60/g, '\x60\x60') + '\x60';
}
const stmt = db.prepare(`SELECT * FROM ${quote(tableName)}`);
@JoshuaWise
Thanks for the answer. Though there might be a difference between Blobs stored in sqlite3 compared to the indexedDB database. From Blobs stored in sqlite3 I was not able to create objectURLs with URL.createObjectURL(blob). But maybe I have misunderstood something.
Hi,
I get some 'out of memory errors', I guess because of my use of better-sqlite. I don't know why : https://stackoverflow.com/questions/55657313/out-of-memory-puppeteer-or-better-sqlite3
Any hint ?
so how you guys generally use this in your app?
Is there any tools like query builder so I don't have to maintain sql strings by hand which is error prone imo. I found knex.js only supports node-sqlite3.
cc @JoshuaWise
Hi there,
How would I use a trigger function using better-sqlite to perform a file delete once the row is deleted.
Here's the scenario:
Table: Student
Columns:
id integer (primary key not null)
Name Varchar(30) not null
Table : StudentFiles
Columns :
id integer (primary key not null)
studentFilePath: varchar(200) // this is stored as a json string.
studentId: foreign key references Student.id on delete cascade on update cascade
When a row is deleted from the Student table, the corresponding row will also be removed from the StudentFiles table.
How would I create a trigger to call my javascript function to delete the file as pointed by studentFilePath before the row inside StudentFiles disappear?
Thanks.
@mygithubid1, Here's the basic idea:
const fs = require('fs');
const db = require('better-sqlite3')('data.db');
db.function('delete_file', (filename) => {
fs.unlinkSync(String(filename));
});
db.exec(`
CREATE TEMPORARY TRIGGER delete_student_files
AFTER DELETE ON StudentFiles
BEGIN
SELECT delete_file(old.studentFilePath);
END;
`);
@JoshuaWise Thank you. Will give this a shot.
@JoshuaWise Worked. Thank you.
Hello,
I encountered a problem while executing ATTACH statement.
const attach = this.db.prepare("ATTACH DATABASE '" + databasePath + "' AS temp;")
this.db.exec(attach)
Error 'expected first argument to be a string' is reported. Does anybody know what is the issue?
databasePath is relative
Best regards,
Igor
@vurdeljica, the db.prepare() function returns a Statement object. To execute a statement, you should use statement.run(). The db.exec() function you're using expects an SQL string, not a statement object.
This is the correct code:
const attach = this.db.prepare('ATTACH DATABASE ? AS temp;')
attach.run(databasePath)
Please read the very detailed documentation to learn how to use better-sqlite3.
Hi, How would I use transaction and call model layer function in my controller layer file?
My project use MVC architecture.
I have two model file: preset.model and task.model, and one controller file.
The preset schema only save name field, task schema has name and presetUid fields
I need to create preset first step and get presetUid, so that I can add the task.
preset.model.js
function add(preset) {
const sql = `INSERT INTO preset (name) VALUES ($name)`
const result = db.prepare(sql).run(preset.name)
return result
}
task.model.js
function add(task) {
const sql = `INSERT INTO preset (name, presetUid) VALUES ($name, $presetUid)`
const param = {
name: task.name,
presetUid: task.presetUid
}
const result = db.prepare(sql).run(param)
return result
}
In my controller.js file, I want to use preset.add and task.add function in model file
like this:
controller.js
import preset from './preset.model'
import task from './task.model'
import db from './db'
async function combo() {
const data = { name: 'preset01'}
try {
await db.beginTransaction() // I want to start for transaction
const preset = await preset.add(data)
const task = {
name: 'task01',
presetUid: preset.lastInsertRowid
}
await task.add(task)
await db.commit() // I want to do commit
} catch(error) {
await db.rollback() // I want to rollback
}
}
I know the documentation of better-sqlite does not have db.beginTransaction(), db.commit() and db.rollback() api. They are example to show what I want to do.
My problem is how can I do then I will get the same feature in my controller file.
Thank you.
@s7130457
function combo() {
const data = { name: 'preset01'}
try {
db.exec("BEGIN TRANSACTION;");
const preset = preset.add(data)
const task = {
name: 'task01',
presetUid: preset.lastInsertRowid
}
task.add(task)
db.exec("END TRANSACTION;");
} catch(error) {
db.exec("ROLLBACK TRANSACTION;");
}
}
https://www.sqlite.org/lang_transaction.html
One more thing. There's no need for await in your functions. The library is synchronous.
@s7130457 @mygithubid1 so there is actually a transaction higher order function as part of the api https://github.com/JoshuaWise/better-sqlite3/blob/master/docs/api.md#transactionfunction---function that was added in v5.
usage:
import preset from './preset.model'
import task from './task.model'
import db from './db'
function comboInserts() {
const data = { name: 'preset01' }
const presetReturn = preset.add(data)
const taskData = { name: 'task01', presetUid: presetReturn.lastInsertRowId }
task.add(taskData)
}
const comboTransaction = db.transaction(comboInserts)
// usage
comboTransaction()
I’m new to this so I hope you can help me..
I am trying to display all the datas I have under the table list in the names column but can’t make it work., this is my code:
row = db.prepare(SELECT names FROM list).all()
message.channel.send(row);
It displays:
[object Object]
the [object Object] repeats depending on how many entry is in the db, I just wanted it to display the all the names in the list table, I’m still a noob so hopefully you can help me.
Thank you.
@nierchi
Channel#send's first argument is content, which should be a string, however,
you provided an array of objects.
The default serialization/conversion for a string to an object.
Function Object:
stringify(function (){}) -> [object Function]
Array Object:
stringify([]) -> [object Array]
RegExp Object:
stringify(/x/) -> [object RegExp]
Date Object
stringify(new Date) -> [object Date]
Object Object
stringify({}) -> [object Object]
Several More!
In conclusion, you have to map the property or use for loop.
@chroventer ohh.. alright thank you for giving me that idea, I’ll tweak my code.
Thanks for the help 😊
Hi! Are there any performance benefits in preparing a statement once and then using it multiple times? E.g. i now have this:
function select(params){
return db.prepare('...').run(params)
}
would this be more performant (or are there any drawbacks):
const statement = db.prepare('...');
function select(params){
return statement.run(params);
}
Thanks in advance!
db.prepare returns a statement.
Part of this operation involves reaching out to the sqlite3 engine with a call to sqlite_prepare_v3 to compile the sql.
https://github.com/JoshuaWise/better-sqlite3/blob/master/docs/api.md#class-statement
Hi! Are there any performance benefits in preparing a statement once and then using it multiple times? E.g. i now have this:
function select(params){
return db.prepare('...').run(params)
}
would the following have better performance (or are there any drawbacks)?
const statement = db.prepare('...');
function select(params){
return statement.run(params);
}
Thanks in advance!
Yes, if your query is relatively fast, reusing the prepared statement will result in a huge performance gain.
Hi,
I am new to this, so have few questions. How i can get the response back from the transaction which happen. For e,g.
function doMultipleTransaction(sqlQuery, sqlData) { try { log.info("SQl Query : ", sqlQuery); log.info("SQl Data : ", sqlData); log.info("Transaction Started"); const insert = db.prepare(sqlQuery); const insertMany = db.transaction((sqlDataRows) => { for (const sqlDataRow of sqlDataRows) { insert.run(sqlDataRow); } }); return insertMany(sqlData); } catch (err) { // (transaction was forcefully rolled back) if any error while doing transaction if (!db.inTransaction) throw err; } }
From the above function, how can i get the response back of all the inserted records.
Or is there any better way of doing this.
you can return values from the transaction function. This will work:
function doMultipleTransaction(sqlQuery, sqlData) {
try {
log.info("SQl Query : ", sqlQuery);
log.info("SQl Data : ", sqlData);
log.info("Transaction Started");
const insert = db.prepare(sqlQuery);
const insertMany = db.transaction((sqlDataRows) => {
return sqlDataRows.map(sqlDataRow => insert.run(sqlDataRow))
});
return insertMany(sqlData);
} catch (err) {
// (transaction was forcefully rolled back) if any error while doing transaction
if (!db.inTransaction) throw err;
}
}
I believe this is an undocumented feature
@manishrana87 I was about to reply exactly as @andykais did, but he beat me to it: Either collect the results into an array or into an object and return. .transaction() returns what it received.
Hi, I am having another question.
I have a array of item ["001", "002", "003"] .
These are primary key from a table "image". If I want to fetch the details from the "image" table.
In normal SQL editor, the query is,
Select * from image where image_id in ('001', '002', '003');`
How I can implement the same in terms of better-sqlite3.
@manishrana87, If your array always has three items, you can do this:
// First, open the database
const db = require('better-sqlite3')('my-database.db');
// Then, create a prepared statement
const statement = db.prepare("Select * from image where image_id in (?, ?, ?)");
// Lastly, run the query to get the results
const results = statement.all(["001", "002", "003"]);
However, if your array can have any number of items, you should do this:
// Same as before
const db = require('better-sqlite3')('my-database.db');
// This time, the statement only fetches a single item
const statement = db.prepare("Select * from image where image_id = ?");
// Run the query once for each item in the array
const results = array.map(item => statement.get(item));
@JoshuaWise, If I know I am working with a single nodejs process accessing my sqlite database, do I need to put a transaction around something like this:
const db = require('better-sqlite3')('my-database.db')
const selectStmt = db.prepare('SELECT * FROM cache WHERE cacheKey = ?')
const insertStmt = db.prepare('INSERT INTO cache (data, cacheKey) VALUES (?, ?)')
const ee = new EventEmitter()
ee.on('event', data => {
// I want the select then insert to happen transactionally
const cacheKey = JSON.stringify(data)
const row = selectStmt.get(cacheKey)
if (row === undefined) insertStmt.run(data, cacheKey)
})
// imagine these events are coming in asynchronously at any time
ee.emit('event', {some:'data'})
ee.emit('event', {some:'moredata'})
side note: I specifically cannot use a unique index here because sometimes I will insert into the database and not want the inserts to be unique, partial indexes will not work here because when I want to use the cache, I may access older rows which were created when the cache was not active.
@andykais Hi there. Why should your cache should support duplicate keys?
@andykais Since your SQL statements are executed synchronously (within the same event-loop tick), you don't need to worry about other events happening at the same time (assuming no other thread or process is accessing the same database, as you mentioned).
Why should your cache should support duplicate keys?
@mygithubid1 its a bit of a weird use case. This table servers both as a cache and storage for request/responses. Sometimes I want to use the cache (e.g. if a request has already been made with those parameters, grab the response from the database), and sometimes I do not want to use the cache (e.g. I know hitting the same endpoint will return different data each time, so I just want to store the responses in the database).
I dont want to derail this thread with more detailed explanations though, so lets leave it at that.
Hi,
My INTEGERS passed into db.prepare().run() are being cast/translated to REALs when executed.
I have a messages table (id INTEGER, json TEXT) that I want to update to show when a user has an unread or read message.
The json text/object is v. simple
id
json
1
'{"id": 1, "isRead": 0, ......}'
2
'{"id": 2, "isRead": 1, ......}'
where isRead is a 0 or 1 flag.
To update the isRead property within the json field, I pass a simple object containing only integers
msgObj = {
isRead: 1,
id: 1
}
But when I run this
db.prepare(`
UPDATE messages
SET json = json_set(json, '$.isRead', :isRead)
WHERE id = :id
`).run({
isRead: msgObj.isRead,
id: msgObj.id
})
isRead becomes {..."isRead": 1.0, ...} within the json text field.
I turned on {verbose: console.log} in my db require and the console shows me that the generated SQL is
UPDATE messages
SET json = json_set(json, '$.isRead', 1.0)
WHERE id = 1.0
The id still identifies the correct row, as the UPDATE works, so the id field in the WHERE clause must be equating 1.0 with the INTEGER id of 1 in the table.
Is this normal/correct behaviour, or am I missing something?
Apologies if this seems straightforward but how do I create a table with a dynamic name?
Originally I thought I could do
const stmt= db.prepare(`CREATE TABLE ${MY_CUSTOM_NAME} (
id INTEGER PRIMARY KEY AUTOINCREMENT,
some_property TEXT NOT NULL
)`);
const i = stmt.run();
But that gives me the following error, assuming MY_CUSTOM_NAME was passed in as '2'
SqliteError {message: "near "2": syntax error", stack: "SqliteError: near "2": syntax error↵ at VueComp…ents/NacPage.vue?vue&type=script&lang=js&:165:10)", code: "SQLITE_ERROR", __ob__: Observer}
code: "SQLITE_ERROR"
message: "near "2": syntax error"
stack: "SqliteError: near "2": syntax error↵ at VueComponent._callee$ (webpack-internal:///./node_modules/babel-loader/lib/index.js!./node_modules/vue-loader/lib/index.js?!./src/renderer/components/NacPage.vue?vue&type=script&lang=js&:130:35)↵ at tryCatch (webpack-internal:///./node_modules/regenerator-runtime/runtime.js:62:40)↵ at Generator.invoke [as _invoke] (webpack-internal:///./node_modules/regenerator-runtime/runtime.js:296:22)↵ at Generator.prototype.<computed> [as next] (webpack-internal:///./node_modules/regenerator-runtime/runtime.js:114:21)↵ at step (webpack-internal:///./node_modules/babel-runtime/helpers/asyncToGenerator.js:17:30)↵ at eval (webpack-internal:///./node_modules/babel-runtime/helpers/asyncToGenerator.js:35:14)↵ at new Promise (<anonymous>)↵ at new F (webpack-internal:///./node_modules/core-js/library/modules/_export.js:36:28)↵ at eval (webpack-internal:///./node_modules/babel-runtime/helpers/asyncToGenerator.js:14:12)↵ at VueComponent.addDrug (webpack-internal:///./node_modules/babel-loader/lib/index.js!./node_modules/vue-loader/lib/index.js?!./src/renderer/components/NacPage.vue?vue&type=script&lang=js&:165:10)"
__ob__: Observer {value: SqliteError, dep: Dep, vmCount: 0}
__proto__: Error
@simeon9696
According to the docs at https://www.sqlite.org/lang_createtable.html, , the only names forbidden are those that begin with sqlite_
However, if a table name contains a space or starts with a number, it needs to be in double quotes https://www.sqlite.org/lang_keywords.html, so
const stmt= db.prepare(`CREATE TABLE "${MY_CUSTOM_NAME}" (
id INTEGER PRIMARY KEY AUTOINCREMENT,
some_property TEXT NOT NULL
)`);
should work.
@DefCheese works like a charm! Thanks for pointing out the documentation for me!
So I'm having a bit of difficulty understanding how the INSERT statement works with parameter binding. From the docs:
// The following are equivalent.
const stmt = db.prepare('INSERT INTO people VALUES (@firstName, @lastName, @age)');
stmt.run({
firstName: 'John',
lastName: 'Smith',
age: 45
});
This means that @firstName = John etc but does firstName correspond to a column in the table? If the first column was named cities and the second towns and the above statement was run. Would the entry in cities be John and towns be Smith because that's the order of the data in the INSERT statement?
TL:DR: Are bindings in INSERT statements bound to column names in a table or the order in which the columns appear in the table. I ask because rn I'm inserting data with bound parameters and the object has the keys and values mapped correctly but some of the column data is swapped around when it comes out.
@simeon9696
INSERT can be used on specific column names in your table, or all of them
Assuming your table is
CREATE TABLE locations (
street TEXT,
town TEXT,
zip TEXT
);
then INSERT can be used on all columns like this:
const stmt = db.prepare('INSERT INTO locations VALUES (@a, @b, @c)');
stmt.run({
a: 'High Street',
b: 'BigTown',
c: '12345'
});
...and the values would be inserted into the columns in the order they were defined in the CREATE TABLE above (a goes into street, b to town, etc.)
However, you can also specify the order yourself and even omit columns) using
const stmt = db.prepare('INSERT INTO locations (zip, town) VALUES (@a, @b)');
stmt.run({
a: '12345',
b: 'BigTown'
});
street would then be NULL for this inserted row.
The bind names you use (a, b, c) don't relate directly to the column names, just the order you use them in the SQL statement.
So
const stmt = db.prepare('INSERT INTO locations (zip, town) VALUES (@b, @a)');
stmt.run({
a: 'BigTown',
b: '12345'
});
will also work fine. Obviously, using meaningful bind variable names is preferred over a, b, c, etc. :)
Is it possible to pass in a table name as a parameter in a transaction?
//From the docs
const insert = db.prepare('INSERT INTO cats (name, age) VALUES (@name, @age)');
const insertMany = db.transaction((cats) => {
for (const cat of cats) insert.run(cat);
});
insertMany([
{ name: 'Joey', age: 2 },
{ name: 'Sally', age: 4 },
{ name: 'Junior', age: 1 },
]);
But lets say each cat had its own table so instead of saying
const insert = db.prepare('INSERT INTO cats (name, age) VALUES (@name, @age)');
You'd write
const insert = db.prepare('INSERT INTO "${cat_name}"(name, age) VALUES (@name, @age)');
I was thinking of putting it inside the function itself but that just seems inefficient. I did see in the docs Any arguments passed to the transaction function will be forwarded to the wrapped function but I don't know how to pass the arguments passed into the function into the prepared statement.
I have one query.
My create table statement is like this.
CREATE TABLE IF NOT EXISTS cron_job_header (id INTEGER PRIMARY KEY AUTOINCREMENT, cronExpression TEXT, scheduleUploadStatus TEXT, scheduleCreatedTime TEXT)
Here id is a integer and should be autoincremented.
What will be the insert statement for this table having below data.
{ cronExpression: 'At 06:30 PM, on day 15 of the month, only in September', scheduleUploadStatus: 'Pending', scheduleCreatedTime: 2020-10-21T10:54:51.001Z }
Because when I am trying to insert this record in a table, it is giving me below error.
Error while creating a record TypeError: SQLite3 can only bind numbers, strings, Buffers, and null.
Can someone help me with this.
[ ] [_**a**_](
[ ] 1. url
)
[ ] [_**a**_](
[ ] 1. url
)
Using a loop is quite efficient, @dimitry-ishenko.
This is not correct, @JoshuaWise . The larger the bulk insert, the more times you avoid updating the various indices. If you bulk insert 100,000 rows on a table with 3 indices, you have avoided 299,997 index updates (even with the WAL, this takes a noticable amount of time.)
@dimitry-ishenko - the surrounding library does not make this choice. What you're asking is part of the query language.
The answer is "yes, SQLite gained bulk inserts in 3.7.11."
Hey i was using the ? character
db.prepare('CREATE TABLE ? (?)')
But when i was running the code it saysme this error
I cant use the ? character? And if i cant, why?
Im writing a good amount of code that requires re-running inserts on unique rows. I have essentially two options at my disposal. Either run an insert, catch the error and run a select, or I add an ON CONFLICT clause, and perform an update so that I can use the RETURNING clause. Examples below:
Option 1
db.exec(`CREATE TABLE tag (
id INTEGER PRIMARY KEY NOT NULL,
name TEXT UNIQUE NOT NULL,
created_at DATETIME NOT NULL);`)
const insert_stmt = db.prepare('INSERT INTO tag (name, created_at) VALUES (?, ?)')
const select_stmt = db.prepare('SELECT id FROM tag WHERE name = ?')
function create_tag(name) {
try {
return insert_stmt.run(name, new Date().toString()).lastInsertRowid
} catch (e) {
// also, no good way to distinguish a 'unique constraint' sqlite error from others
return select_stmt.get(name).id
}
}
create_tag('red') // 1
create_tag('red') // 1
Option 2
db.exec(`CREATE TABLE tag (
id INTEGER PRIMARY KEY NOT NULL,
name TEXT UNIQUE NOT NULL,
created_at DATETIME NOT NULL);`)
const insert_stmt = db.prepare(`INSERT INTO tag (name, created_at) VALUES (?, ?)
ON CONFLICT DO UPDATE SET name = name
RETURNING id`)
function create_tag(name: string) {
const row = insert_stmt_2.get(name, new Date().toString())
return row.id
}
console.log(create_tag('green')) // 2
console.log(create_tag('green')) // 2
The prior is error prone since at best I can capture all SqliteErrors, but the latter does two write operations (I think) which feels bad, and uses RETURNING + .get on inserts, which I am unsure of, but am afraid will have performance implications.
Is this the best way to get the id of conflicting rows?
Is this the best way to get the id of conflicting rows?
the best way is to prevent them, rather than to make them then clean them up
you should either insert where not exists or insert or replace into
insert where not exists wouldnt give me back the id of the row that already exists. The point of doing it this way is to return the id for some data thats unique with a single query.
How is insert or replace into different than insert .. on conflict do update?
one's sql, the other is a nonstandard extension
one has significantly more overhead than the other
one performs the write and then undoes it, the other prevents the write in
the first place in many cases
one has handler overhead, the other doesn't
one uses common things, the other uses obscure things that most people
don't know the rules about (including you)
the reason i didn't return the id is that there are several good ways to do
that. you could just put select on the query in the obvious way, but
generally, it's better to do the inserts and then do the select afterwards,
because it's faster, it gets you a point answer without a transaction, it
accommodates other online clients, and because it lets you gather just the
relevant columns and omit ones you added for thoroughness
On Tue, Aug 24, 2021 at 7:35 AM Andrew Kaiser @.***>
wrote:
insert where not exists wouldnt give me back the id of the row that
already exists. The point of doing it this way is to return the id for some
data thats unique with a single query.
How is insert or replace into different than insert .. on conflict do
update?
—
You are receiving this because you commented.
Reply to this email directly, view it on GitHub
https://github.com/JoshuaWise/better-sqlite3/issues/125#issuecomment-904699018,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/AAAS5KQOC3CR4ZUTDA7NZ6TT6OU4PANCNFSM4E6NEIMA
.
one has significantly more overhead than the other
what is the overhead? I just ran a benchmark and insert or replace vs on conflict do update are nearly identical:
insert_or_select avg speed: 0.065ms over 1000 runs
upsert avg speed: 0.583ms over 1000 runs
insert_or_replace avg speed: 0.560ms over 1000 runs
benchmark code
import { performance } from 'perf_hooks'
import Sqlite from 'better-sqlite3'
const db = new Sqlite('sqlite.db')
const BENCHMARK_RUNS = 1000
function benchmark(name:string,func:() => any) {
let time_sum = 0
for (let i = 0; i < BENCHMARK_RUNS; i++) {
const start = performance.now()
func()
const end = performance.now()
time_sum += (end - start)
}
const average = time_sum/BENCHMARK_RUNS
console.log(`${name} avg speed: ${average.toFixed(3)}ms over ${BENCHMARK_RUNS} runs`)
}
db.exec(`DROP TABLE IF EXISTS tag;`)
db.exec(`CREATE TABLE tag (
id INTEGER PRIMARY KEY NOT NULL,
name TEXT UNIQUE NOT NULL,
created_at DATETIME NOT NULL);`)
const insert_stmt = db.prepare('INSERT INTO tag (name, created_at) VALUES (?, ?)')
const select_stmt = db.prepare('SELECT id FROM tag WHERE name = ?')
function insert_or_select(name: string) {
try {
return insert_stmt.run(name, new Date().toString()).lastInsertRowid
} catch (e) {
if (e instanceof Sqlite.SqliteError) return select_stmt.get(name).id
else throw e
}
}
const upsert_stmt = db.prepare(`INSERT INTO tag (name, created_at) VALUES (?, ?)
ON CONFLICT DO UPDATE SET name = name
RETURNING id`)
function upsert(name: string) {
const row = upsert_stmt.get(name, new Date().toString())
return row.id
}
const insert_or_replace_stmt = db.prepare(`INSERT OR REPLACE INTO tag (name, created_at) VALUES (?, ?)`)
function insert_or_replace(name: string) {
const row = insert_or_replace_stmt.run(name, new Date().toString())
return row.lastInsertRowid
}
benchmark('insert_or_select', () => insert_or_select('red'))
benchmark('upsert', () => upsert('green'))
benchmark('insert_or_replace', () => insert_or_replace('green'))
generally, it's better to do the inserts and then do the select afterwards,
because it's faster
this does appear to be the case. It complicates my code slightly, but this does appear to be the faster solution by far
Its also worth pointing out that insert or replace is not going to be equivilant to an upsert even outside of the internals of sqlite sql parsing. on conflict do update will update a row, but insert or replace will delete the old row and create a new row. This means if you try doing an insert or replace on a row that is referenced by another table, youre going to break your foreign keys.
ok good luck
Is there a recommended way or only one way (like this) to prepare one-off statements?
I don't know enough C or C++ to find out how prepare() is implemented in better-sqlite3, but my guess is that those are only finalized() before closing database connection. What if I don't want to reopen the database just to finalize() them?
I'm asking this because exec() is highly discouraged due to injection risk, but I don't see another simple way to execute dynamically generated SQLs (they are complicated in my case because of relational division) other than the above. What's the rough performance difference between exec() and prepare() since "All SQL must be converted into a prepared statement before it can be run" from SQLite doc?
Is there a way for better-sqlite3 to evaluate json when using json_object?
In this example:
SELECT m.*, json_group_array(json_object('name', p.name, 'rating', p.rating, 'team', t.team)) teams FROM matches m LEFT JOIN teams t ON m.id = t.match LEFT JOIN players p ON p.id = t.player ORDER BY date ASC
When I run .all() on this statement I get an array of match objects, but each teams is a JSON string that I need to parse.
Hi,
I want to filter records using in(). How can I do that with named parameters?
db.prepare('select * from test_table where string_column in ($values)').all({ values: [ 'a', 'b' ] })
should end up as:
select * from test_table where string_column in ('a', 'b')
I know that I can kind of do that by dynamically generating the SQL string before passing it to .prepare() but that's probably a little unreadable especially when in combination with other, 'normal' parameters, also because I can't use named parameters.
const asValues = [ 'a', 'b' ]
const sPlaceholders = asValues.map(sV => '?').join(', ');
const sSql = `select * from test_table where first_col = ? and other_col in (${sPlaceholders})`;
db.prepare(sSql).all('value_first_col', asValues);
Thanks.
PS.: As far as I understand, filling in the parameter values in the SQL statement is a thing SQLite itself does and not better-sqlite3, right? So maybe I'm wrong here.
I apologize if I missed this somewhere, but unfortunately the type of statement I'm after contains a field name in between its key words which makes it hard to search for. Anyway, here's the background:
I have a database that contains two tables. The first table, library, contains the main data entries. The second table, tags, contains text tags that can be used to describe and group the contents of the library table. The primary key of the library table is the column, hash. The tags table relates its entries on the same field, hash, as a foreign key into the library table.
Here's the question:
I want to run this query:
select * from library where hash in (select hash from tags where text in ('new', 'test'));
In my Javascript it looks like this:
const tags = ['new', 'test']
const sql = 'select * from library where hash in (select hash from tags where text in (?));'
const results = db.prepare(sql).all(tags)
The code generates the error: RangeError: Too many parameter values were provided
I have also tried:
const tags = ['new', 'test']
const sql = 'select * from library where hash in (select hash from tags where text in (?));'
const results = db.prepare(sql).all(...tags)
This generates the same error: RangeError: Too many parameter values were provided
The stumper for me is that if my array only contains one value, then it works:
const tags = ['new']
const sql = 'select * from library where hash in (select hash from tags where text in (?));'
const results = db.prepare(sql).all(...tags)
How do I use better-sqlite3 for this kind of statement? Thanks!
You have two tags, you need two placeholders.
const tags = ['new', 'test']
const sql = 'select * from library where hash in (select hash from tags where text in (?, ?))'
const results = db.prepare(sql).all(tags)
If your tags are dynamic so you don't know in advance how many tags there will be you can dynamically generate the placeholders too.
const tags = ['new', 'test']
const placeholders = Array.from(tags).fill('?').join(', ')
const sql = `select * from library where hash in (select hash from tags where text in (${placeholders}))`
const results = db.prepare(sql).all(tags)
@elyobo - dynamically generating queries is a severe security risk and needs to be left in 1990 with php. please don't give that advice. that's the most common cause of injections..
@driverjb - generally speaking sqlite does not support the parameter binding of arrays; there's nothing better-sqlite can do about that. the typical approach is to create a temporary table, insert your arguments into that, and join against it instead. it's a pain in the butt. on the bright side, it's also typically faster, and unlike in, is fairly re-usable on those occasions that that matters.
Dynamically generating queries with use input yes; dynamically generating queries with safe placeholders to use with prepared statements is fine.
I have a general query. Is there any overhead in having many SQLite databases open at the same time?
I'm trying to understand if we should have a single DB or several smaller DBs (e.g 100s individual dbs).
We are currently having a single database and prefixing the tables with customer_id. But it'd be a lot easier if we could have a db per customer.
I have a general query. Is there any overhead in having many SQLite databases open at the same time?
A small, probably unimportant amount. SQLite needs several hundred K of ram.
I'm trying to understand if we should have a single DB or several smaller DBs (e.g 100s individual dbs).
You almost certainly should have a single database. This is not related to needs of the tool. You can treat the single database like it's distinct; the converse is not true. One will eventually need to be retooled; the other will not.
prefixing the tables with customer_id.
Use a single table with a defining column instead.
But it'd be a lot easier if we could have a db per customer.
You absolutely can. You just shouldn't. You will lose out on the ability to make any decisions across the set.
@danielepolencic The correct way of handling this would be to have a single database that has a table that contains your customers with the primary key of say, customer_id. Then all your other tables have a foreign key matching up to that customer_id so that a given customers data can be selected from the table. This allows a single table for each set of data instead of juggling what sounds like a nightmare scenario of tables all with different names that represent the same data sets for different customers.
Hi,
I wrote this code:
let statement = db.prepare(`SELECT COUNT(*) FROM main WHERE userId = ?`); let row = statement.get(user.id); console.log(row);
which returns { 'COUNT(*)': 1 }
I made it so I can check easier if a row with that userId already exist or not. But how would I check for the count since row.count(*) won't work?
I would change it to this:
function userExists(userId) {
const statement = db.prepare(`SELECT 1) FROM main WHERE userId = ?`);
const row = statement.get(user.id);
return row !== undefined
}
const doTheyExist = userExists(user.id)
@Dezaku I would change it to this:
function userExists(userId) {
const statement = db.prepare(`SELECT 1 FROM main WHERE userId = ?`);
const row = statement.get(user.id);
return row !== undefined
}
const doTheyExist = userExists(user.id)
Thats a way easier way, thank you!
That was the long version that makes it clear what's happening. You could shorten it to this:
function userExists(userId) {
return db.prepare('SELECT 1 FROM main WHERE userId = ?').get(userId) !== undefined
}
const doTheyExist = userExists(user.id)
That was the long version that makes it clear what's happening. You could shorten it to this:
function userExists(userId) {
return db.prepare('SELECT 1 FROM main WHERE userId = ?').get(userId) !== undefined
}
const doTheyExist = userExists(user.id)
Alright, thank you!
Hello guys, I'm currently trying to get the top 10 values of my "total" column. Somehow two errors occur and I can't seem to fix them. This is my code
const statement =
db.prepare(`SELECT total FROM main ORDER BY total DESC LIMIT 10
`);
const rows = statement.get(); // I also tried statement.all() but the same error occurs
console.log(rows);
Since I also tried statement.all() I assume something is wrong with the statement itself but I didn't find any solution yet.
This is the error:
Received one or more errors
1 ExpectedValidationError > s.literal(V)
| Expected values to be equals
|
| Expected:
| | null
|
| Received:
| | ''
2 ExpectedConstraintError > s.string.lengthGreaterThanOrEqual
| Invalid string length
|
| Expected: expected.length >= 1
|
| Received:
| | ''
Hello guys, I get two errors which I can't seem to fix. This is the code:
const userArray = [];
const totalArray = [];
function getUserIDs() {
const stmt = db.prepare(
`SELECT userid FROM Users ORDER BY total DESC LIMIT 10`
);
for (const row of stmt.iterate()) {
userArray.push(row.userid);
}
}
function getTotals() {
const stmt = db.prepare(
`SELECT total FROM Users ORDER BY total DESC LIMIT 10`
);
for (const row of stmt.iterate()) {
totalArray.push(row.total ? row.total : "0");
}
}
try {
getUserIDs();
getTotals();
} catch (error) {
console.log(error);
}
console.log(userArray, totalArray);
And these are the errors:
CombinedError (2)
Received one or more errors
1 ExpectedValidationError > s.literal(V)
| Expected values to be equals
|
| Expected:
| | null
|
| Received:
| | ''
2 ExpectedConstraintError > s.string.lengthGreaterThanOrEqual
| Invalid string length
|
| Expected: expected.length >= 1
|
| Received:
| | ''
at UnionValidator.handle (C:\Development\Bots\v14\CakeBot\node_modules\@sapphire\shapeshift\dist\index.js:1085:23)
at UnionValidator.parse (C:\Development\Bots\v14\CakeBot\node_modules\@sapphire\shapeshift\dist\index.js:142:88)
at EmbedBuilder.setDescription (C:\Development\Bots\v14\CakeBot\node_modules\@discordjs\builders\dist\index.js:231:26)
at Object.execute (C:\Development\Bots\v14\CakeBot\src\commands\currencySystem\leaderboard.js:44:8)
at process.processTicksAndRejections (node:internal/process/task_queues:95:5)
at async Object.execute (C:\Development\Bots\v14\CakeBot\src\events\client\interactionCreate.js:11:9)
node:events:505
throw er; // Unhandled 'error' event
^
userArray and totalArray both still get logged into the console normally and the output it shows is what it should be
repeatedly preparing the query is wasteful
const exists_q = db.prepare('select 1 from main where userid = ?'),
userExists = userId => exists_q.get(userId) !== undefined;
Is there a way to wait for a database entry to appear?
Say my code looks like this:
events.on('someEvent', event => {
db.prepare('INSERT INTO events (id, someValue) VALUES (?, ?)')
.run(event.id, event.someValue);
});
events.on('otherEvent', event => {
const previousEvent = db.prepare('SELECT * FROM events WHERE id = ?')
.get(event.referenceId);
});
Now say that there's some code in someEvent that takes a bit of time to finish before inserting into the database. Is there some way to wait for event.referenceId to appear in the database, potentially with a timeout if it never does?
I'm uploading images to my node.js server and saving those in a table in my sqlite db. I'd like to stream that file upload straight to the table rather than reading the whole image in memory before doing the insert.
I saw that SQLite should support this via blob_write and better-sqlite3 appears to support blob_write, however, I'm not certain how to make use of this.
What I have is a data type of AsyncIterable<Uint8Array>. I'm not sure how to convert this to whatever better-sqlite3 needs for my insert:
imageDb
.prepare(
/* sql */ `
INSERT INTO "Images"
(updatedAt, contentType, blob)
VALUES (@updatedAt, @contentType, @blob)
`,
)
.run({updatedAt: new Date().toISOString(), contentType, blob: data}) // <-- data is the Uint8Array in an AsyncIterable
Tips are appreciated. Thanks!
Hello, I am inserting some values into a table:
const sqlite3 = require('better-sqlite3');
const db = sqlite3(":memory:");
db.pragma('journal_mode = WAL');
db.exec("CREATE TABLE mytable (id INTEGER PRIMARY KEY AUTOINCREMENT, value TEXT NOT NULL)")
const sql =`
INSERT INTO mytable (value) VALUES (:value)
`;
const stmt = db.prepare(sql);
stmt.run({'value'=>'some text'});
// Get the generated mytable.id here
How I can retrieve the inserted id?
@pc-magas - sqlite3 .run returns an object with two properties. one is the count of changed rows; the other is the last insert id.
const stmt = db.prepare('INSERT INTO cats (name, age) VALUES (?, ?)');
const info = stmt.run('Joey', 2);
console.log(info.changes); // => 1
console.log(info.lastInsertId); // => if first cat, 1; otherwise higher
SELECT COUNT(progress_id) total_done, SUM(session_length) FROM progress_table WHERE strftime("%Y-%m-%d",time_stamp,"localtime") > (SELECT DATE("now", "-7 day","localtime"))
What I am doing wrong here, In my "DB Browser for SQLITE", this query does not throw any error, How ever in better-sqlite, node is throwing
SqliteError: no such column: now
Is there a different way of writing this query for better-sqlite to work?.
Use single quotes for strings and you'll be fine.
SQL strings are generally single quoted, double quotes generally used for quoting identifiers e.g. if you have a column name with unsafe chars like a space in them. I've noticed that SQLite is inconsistently forgiving in this regard (there's presumably a pattern to it, but I haven't looked - I just use single quotes by habit anyway, but a colleague ran into this in a subtle and confusing bug).
Use single quotes for strings and you'll be fine.
SQL strings are generally single quoted, double quotes generally used for quoting identifiers e.g. if you have a column name with unsafe chars like a space in them. I've noticed that SQLite is inconsistently forgiving in this regard (there's presumably a pattern to it, but I haven't looked - I just use single quotes by habit anyway, but a colleague ran into this in a subtle and confusing bug).
Thank you so much for your quick, it worked.
Hi there, I wonder how I can do operations across databases. Like,
-- Analysis DB
CREATE TABLE Reports ( date: bigint, ruleID: int)
-- Basic DB
CREATE TABLE Rules (ruleID: int, desc: text)
How I can join these two tables on ruleID across databases? I don't want to fetch data from these dbs separately and do 'join like' operations in js (forEach & Map).
Conventional usage with Express?
Most other DBs rely on connection pools that handle reconnects for us (allowing concurrent reads as well).
Hovewer, for SQLite it's not recommended to use pooling:
I personally wouldn't use better-sqlite-pool. It's possible there's something I'm missing, but it seems like an anti-pattern. The best way to use better-sqlite3 in a single-threaded environment such as Node.js is to only use a single connection for the entire application, and to only run transactions synchronously (within a single event-loop tick).
https://github.com/WiseLibs/better-sqlite3/issues/234
Then the question arises: how to handle DB disconnects?
According to the docs new Database() creates a new connection. If I just reconnect per HTTP request, I can't benefit from prepare statements, not to say it's likely a waste of CPU to reconnect so often...
But, if I connect once, without a connection pool, the connection will eventually be lost and the app will crash.
Surprisingly, I couldn't find more information about reconnects and proper connection management within a web server setup. So asking it here 🤔
@ivan-kleshnin , sqlite is not a server client model. it's a flat file parser local library. there aren't many realistic situations under which it would "disconnect."
@JoshuaWise If you could help with the above that would be awesome. This is a bit out of the depths of mostly everyone I know that uses sqlite ;)
What I feel I'm looking for would look like :
filter2(predicate) {
this.#db.aggregate('filter', {
start: [],
step: (accumulator, [value, key]) => {
const parsed = JSON.parse(currentValue);
if (predicate(parsed)) {
accumulator.push(parsed);
}
return accumulator;
},
result: (accumulator) => JSON.stringify(accumulator),
});
const results = this.#db
.prepare(`SELECT filter(value, key) FROM mytable`)
.pluck()
.get();
return JSON.parse(results);
}
this gives me an error about the wrong number of arguments. Arrays and objects don't work either. I Really hope there's a way :(
@JoshuaWise If you could help with the above that would be awesome. This is a bit out of the depths of mostly everyone I know that uses sqlite ;)
What I feel I'm looking for would look like :
filter2(predicate) {
this.#db.aggregate('filter', {
start: [],
step: (accumulator, [value, key]) => {
const parsed = JSON.parse(value);
if (predicate(parsed, key)) {
accumulator.push(parsed);
}
return accumulator;
},
result: (accumulator) => JSON.stringify(accumulator),
});
const results = this.#db
.prepare(`SELECT filter(value, key) FROM mytable`)
.pluck()
.get();
return JSON.parse(results);
}
this gives me an error about the wrong number of arguments. Arrays and objects don't work either. I Really hope there's a way :(
You almost got it, just get rid of the array brackets:
db.aggregate('filter', {
start: [],
step: (accumulator, value, key) => {
const parsed = JSON.parse(value);
if (predicate(parsed, key)) {
accumulator.push(parsed);
}
return accumulator;
},
result: (accumulator) => JSON.stringify(accumulator),
});
const results = db
.prepare(`SELECT filter(value, key) FROM mytable`)
.pluck()
.get();
return JSON.parse(results);
@JoshuaWise This is awesome! I guess it really was a question of matching parameter counts between the functions. Might I suggest this as an example in the docs somewhere? I even went and looked at your tests which all look at a single row, so I was wondering about the possibility, even. But now I know, and Enmap will be better for it, so thank you!
How can I find where SQLite stores data when using an in-memory database (:memory:)?
The memory consumption of a Node.js server process using an SQLite in-memory database decreased from 1.5 GB to 50 MB in Task Manager, while the SQLite database still contains 1 GB of data. The database size was found using this query: SELECT (page_count - freelist_count) * page_size AS size FROM pragma_page_count(), pragma_freelist_count(), pragma_page_size().
|
gharchive/issue
| 2018-05-04T21:17:35 |
2025-04-01T04:55:47.137003
|
{
"authors": [
"5t4rdu5t",
"DeepJoyPo",
"DefCheese",
"Dezaku",
"JSmental",
"JYone3A",
"JefStat",
"JoshuaWise",
"JustIzumi",
"KaustubhMogha11",
"Nikumarekko",
"StoneCypher",
"TheColorman",
"VyrekXD",
"andrewkolos",
"andykais",
"arne",
"bog-sb",
"bombard-kb",
"chroventer",
"danielepolencic",
"dimitry-ishenko",
"driverjb",
"elyobo",
"eslachance",
"fakob",
"ivan-kleshnin",
"janith-dev",
"kentcdodds",
"manishrana87",
"mceachen",
"memoz",
"mygithubid1",
"nierchi",
"pc-magas",
"s7130457",
"shadeofgod",
"simeon9696",
"vurdeljica"
],
"repo": "WiseLibs/better-sqlite3",
"url": "https://github.com/WiseLibs/better-sqlite3/issues/125",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
200664309
|
NPE on auto-test done by Google
Hello,
When I publish my app, Google starts auto-test on various devices. Since two releases Google give a NPE with this stack trace:
java.lang.NullPointerException: Attempt to invoke interface method 'android.os.Bundle com.android.vending.billing.IInAppBillingService.getBuyIntent(int, java.lang.String, java.lang.String, java.lang.String, java.lang.String)' on a null object reference
at com.smartmobilesoftware.util.IabHelper.launchPurchaseFlow(IabHelper.java:386)
at com.smartmobilesoftware.util.IabHelper.launchPurchaseFlow(IabHelper.java:338)
at jp.wizcorp.wizpurchase.WizPurchase.buy(WizPurchase.java:416)
at jp.wizcorp.wizpurchase.WizPurchase.access$200(WizPurchase.java:40)
at jp.wizcorp.wizpurchase.WizPurchase$4.run(WizPurchase.java:296)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1112)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:587)
at java.lang.Thread.run(Thread.java:818)
In a physical use, no exception...
Google told me that the tested device is Galaxy S6 under Android 5.1. Others 5.1 devices pass all the tests.
Thanks for pointing out this issue. 👍
I'll try to have a look at this this week.
@CyrilleGuimezanes By the way, do you have any JS code to share that could help reproduce/understand this issue?
Actually it's a really simple use of your plugin and in a normal use (when I use my phone under android 6.0), it not crash
Here is the logcat attached to Google bug report if that can help:
logcat.txt
Hello,
Thanks for the logcat file! I tried to trace back your error but I couldn't find why you have this exception as everything seems to be properly initialized.
As I have no way to reproduce it currently, would you be able to make a new build with debug logs activated in IabHelper and attach the resulting logs?
This can be done either by:
directly replacing IabHelper.java l.74 by: boolean mDebugLog = true;
or adding in WizPurchase.java l.359 this line: mHelper.enableDebugLogging(true);
|
gharchive/issue
| 2017-01-13T15:54:58 |
2025-04-01T04:55:47.162645
|
{
"authors": [
"CyrilleGuimezanes",
"jrouault"
],
"repo": "Wizcorp/cordova-plugin-wizpurchase",
"url": "https://github.com/Wizcorp/cordova-plugin-wizpurchase/issues/48",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2080246559
|
BUG: Status page does not load.
its not loading
Flxed Uploaded Files
|
gharchive/issue
| 2024-01-13T11:16:06 |
2025-04-01T04:55:47.163579
|
{
"authors": [
"Bocon778",
"Wizqdev"
],
"repo": "Wizqdev/Status-Page-Api",
"url": "https://github.com/Wizqdev/Status-Page-Api/issues/3",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1516299086
|
🛑 万界规划局共享资源库 (Wjghj Project Static Database) is down
In ed8dcc5, 万界规划局共享资源库 (Wjghj Project Static Database) (https://common.wjghj.cn) was down:
HTTP code: 0
Response time: 0 ms
Resolved: 万界规划局共享资源库 (Wjghj Project Static Database) is back up in 4c8bbbb.
|
gharchive/issue
| 2023-01-02T11:31:57 |
2025-04-01T04:55:47.167247
|
{
"authors": [
"dragon-fish"
],
"repo": "Wjghj-Project/status",
"url": "https://github.com/Wjghj-Project/status/issues/434",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1284446867
|
Package test
Hi @LogSkele , I wanted to do a test package before doing it on authenticlick. I have already created a package but can't seem to import it after downloading it.
You have to do it in a IDE with a terminal like VS Code?
Then in the same directory as dist,
python3 -m twine upload dist/*
If you are on windows.
Bro, I uploaded the package already. Just couldn't download it
:>
|
gharchive/issue
| 2022-06-25T04:42:26 |
2025-04-01T04:55:47.169095
|
{
"authors": [
"LogSkele",
"Wjplatformer"
],
"repo": "Wjplatformer/package-py-test",
"url": "https://github.com/Wjplatformer/package-py-test/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
628166607
|
Request for export with Flat Shading
I tried to export with surface shader in maya as it is the only flat shader available but the textures didnt get exported with that material. Is it possible to convert the lambert shader automatically into Flat shader when exporting as gltf?
i need it to get exported as this material meshbasicmaterial. instead it gets exported as meshstandardmaterial as i checked by importing into threejs. Please let me know how to export as meshbasic material from maya as it a flat shader
I'm not sure I follow you. Maya doesn't really have a "flat shader" as far as I know.
You just set the vertex normals to be the face normals
Then you assign a Lambert or our PBR shader, apply textures, and export to GLTF (I have several flat shaded simple models in this repository).
Regarding ThreeJS, I don't know what rules it applies to convert the GLTF material to its internal shaders, maybe others could help here?
@Ziriax I mean when exported as gltf from maya instead of using meshstandardshader is it possible to force meshbasicshader as this is the flat shader in threejs.
I'm not sure I follow you. Maya doesn't really have a "flat shader" as far as I know.
You just set the vertex normals to be the face normals
Then you assign a Lambert or our PBR shader, apply textures, and export to GLTF (I have several flat shaded simple models in this repository).
Regarding ThreeJS, I don't know what rules it applies to convert the GLTF material to its internal shaders, maybe others could help here?
Thank you, how to convert vertex normal to face normal?
This tutorial might help, but Google "Maya set to face normals" will help too.
https://www.youtube.com/watch?v=0xB7nb0E448&t=11m
|
gharchive/issue
| 2020-06-01T05:37:12 |
2025-04-01T04:55:47.231672
|
{
"authors": [
"Ziriax",
"thickaris"
],
"repo": "WonderMediaProductions/Maya2glTF",
"url": "https://github.com/WonderMediaProductions/Maya2glTF/issues/124",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
200453489
|
Recognize sanitize_hex_color as a sanitization function
WPCS should recognize sanitize_hex_color() as a sanitization function.
This was fixed in #781
|
gharchive/issue
| 2017-01-12T19:10:32 |
2025-04-01T04:55:47.238718
|
{
"authors": [
"jrfnl",
"paulschreiber"
],
"repo": "WordPress-Coding-Standards/WordPress-Coding-Standards",
"url": "https://github.com/WordPress-Coding-Standards/WordPress-Coding-Standards/issues/780",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2328438242
|
Implementation Plan: Augment the catalog database with suitable Rekognition tags
Due date:
2024-07-05
Assigned reviewers
[ ] @sarayourfriend - for your thoughts and context around sensitive data and machine-generated labels
[ ] @stacimc - for your deeper knowledge of the batched update process and the catalog in general
Description
Fixes #4040
This is the last and final IP for the Rekognition data project! It might look a little daunting, but I've tried to add as much rational and explanation to the document as I can. Based on the timezones of the reviewers, if a synchronous session for discussing this IP would be helpful in accelerating any review, I'm sure we could make it happen!
Current round
This discussion is following the Openverse decision-making process. Information
about this process can be found
on the Openverse documentation site.
Requested reviewers or participants will be following this process. If you are
being asked to give input on a specific detail, you do not need to familiarise
yourself with the process and follow it.
This discussion is currently in the Decision round.
The deadline for review of this round is 2024-07-05
Thanks @sarayourfriend for your input! I'm interested in hearing @stacimc's thoughts on the raised points before I make any changes to the IP.
Taking a look at this today, @AetherUnbound! 👀
@AetherUnbound, I finally found the tag analysis I did in April of 2023! In this gist there is a list of all 2,595 unique tags in the Openverse Rekognition data, and counts for the number of times each tag occurs. The script I used is included as well.
https://gist.github.com/zackkrida/ecba17d8fc89d8e4d8aa78e5e19199c0
There's a ton of excellent discussion as part of this IP, thank you to everyone who added input to it 💖 I'm going to put this in draft while we clarify some steps at other pieces of the pipeline, namely #4455 and #4456. That will give us an indication of how we want to handle the Rekognition loading (and subsequently, the Clarifai filtering) going forward. I'll focus on those, then bring the outcomes of them back to this IP for revision!
@AetherUnbound @stacimc after Madison is back from WCEU, regardless of whether revisions are finished, I think it we could finish this discussion in the synchronous chat. Even if Madison is still working on revisions, we can chat and make sure we'll be ready to approve once Madison is finished with revisions. I wanted to propose that last week, but kept forgetting and thinking it would be okay to do this week, and remembered only too late that Madison is AFK this week for the conference.
Anyway, I think we can get this totally squared away pretty quickly with a synchronous chat, and avoid this dragging beyond the end of next week.
@stacimc @sarayourfriend I've gone over this IP and rewritten pieces that assumed we'd be removing data from the catalog to reflect our recently refined "data warehouse" approach based on #4465 and other discussions. I've also made some changes to the filtering approach for Clarifai and explicitly noted gendered terms that might come up in the "acceptable" demographics. I believe this should be ready for final decision - I'm going to leave it open for the next two weeks while I'm AFK so there's plenty of time to gather more feedback if necessary! Thanks so much for your input on this!
Thank you both!
|
gharchive/pull-request
| 2024-05-31T19:07:11 |
2025-04-01T04:55:47.574640
|
{
"authors": [
"AetherUnbound",
"sarayourfriend",
"stacimc",
"zackkrida"
],
"repo": "WordPress/openverse",
"url": "https://github.com/WordPress/openverse/pull/4417",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
116962941
|
Delete Absolute Orientation fields and Movement Manner field
AbsOriPalm, AbsOriFing, and MovMan can be deleted from the models & database, and also from the detail view, the search panel Phonology, the CSV export, field choices in Django, and .... any other place?
Ha, finally my refactoring pays off :D. Because of all kinds of changes I made months ago, this was just a matter of removing three strings from a number of lists. I removed it from:
The detail view
The search view
The CSV export (for which I discovered the headers were broken by all internationalization, so I fixed that as well in the process)
The field choices
But not yet from the underlying database and model, because it's much handier to do that while cleaning other fields as well, as described in #114
|
gharchive/issue
| 2015-11-15T00:19:06 |
2025-04-01T04:55:47.662058
|
{
"authors": [
"Woseseltops",
"ocrasborn"
],
"repo": "Woseseltops/signbank",
"url": "https://github.com/Woseseltops/signbank/issues/109",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1804011447
|
[Feature] Add Firebase Crashlytics to the project
WIP
Please don't – not only does this feed data to Google (privacy issues), it's also proprietary code. If you need analytics in a FOSS app, why not use a FOSS analytics library? I've heard good things about Plausible Analytics and Swetrix, for example (but then, I'm no Android dev myself so I cannot really compare). And if it's just about crash reporting, ACRA should be fine (F-Droid uses that and, as far as I heard, the devs there are quite happy with it). All three (and others from the list) can be used in a privacy-friendly way – which cannot be said about Crashlytics.
That said: I was looking for the APK, but it seems you provide your app only via Google's walled garden (which us privacy folks rather avoid: e.g. my devices all run without the Google stuff and I get my apps from F-Droid). Would you consider providing the APK (e.g. attached to its corresponding release), or bring your app to F-Droid (Crashlytics would be a show-stopper for that, as F-Droid only accepts fully FOSS code)?
Hey Izzy! Tks for comment! I really appreciate you care about privaty issues of my application, I will look those other options of analytics libraries!
In my mind is very interesting provide my application to everyone can use but sorry this project is just for studies, and I don't have plans to provide the app into other app store.
How this project is 100% open source you can duplicate the code or fork in your github and generate your apk/abb version! Please feel free to do so!
I really appreciate you care about privaty issues of my application, I will look those other options of analytics libraries!
Wonderful!
I don't have plans to provide the app into other app store.
Just provide the APK, and it can be in my repo within 24h if there are no show-stoppers :smiley:
you can duplicate the code or fork in your github and generate your apk/abb version!
I'm no Android developer (just a "power user" maybe). So yes, I probably could run gradle to produce the APK, but I'd be lost adjusting the code :see_no_evil:
|
gharchive/issue
| 2023-07-14T01:50:55 |
2025-04-01T04:55:47.667136
|
{
"authors": [
"IzzySoft",
"Wottrich"
],
"repo": "Wottrich/android-smart-checklist",
"url": "https://github.com/Wottrich/android-smart-checklist/issues/70",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2537977589
|
how to run
ModuleNotFoundError: No module named 'configuration_mplugowl3'
Can simply use transformers to load config and model, the package is downloaded within the model.
Does this code span work for you?
import torch
from transformers import AutoConfig, AutoModel
model_path = 'mPLUG/mPLUG-Owl3-7B-240728'
config = AutoConfig.from_pretrained(model_path, trust_remote_code=True)
print(config)
# model = mPLUGOwl3Model(config).cuda().half()
model = AutoModel.from_pretrained(model_path, attn_implementation='sdpa', torch_dtype=torch.half, trust_remote_code=True)
model.eval().cuda()
|
gharchive/issue
| 2024-09-20T06:21:16 |
2025-04-01T04:55:47.888189
|
{
"authors": [
"AntyRia",
"LukeForeverYoung",
"ecoli-hit"
],
"repo": "X-PLUG/mPLUG-Owl",
"url": "https://github.com/X-PLUG/mPLUG-Owl/issues/245",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1651135694
|
🛑 XAONIX.com is down
In de83190, XAONIX.com (https://xaonix.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: XAONIX.com is back up in 4803a3f.
|
gharchive/issue
| 2023-04-02T23:38:04 |
2025-04-01T04:55:47.895639
|
{
"authors": [
"ricardomaia"
],
"repo": "XAONIX/status",
"url": "https://github.com/XAONIX/status/issues/391",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2244308079
|
Add a simply function to plot the dependency tree
Before you submit this PR: make sure to put all operations-related information in a wiki-note, a PR should be about code and is publicly accessible
What does the code in this PR do / what does it improve?
The function is different from the one in docs/source/build_datastructure_doc.py:build_datastructure_doc. The new function is a much more simplified version.
Can you briefly describe how it works?
Can you give a minimal working example (or illustrate with a figure)?
Please include the following if applicable:
[ ] Update the docstring(s)
[ ] Update the documentation
[ ] Tests to check the (new) code is working as desired.
[ ] Does it solve one of the open issues on github?
Notes on testing
Until the automated tests pass, please mark the PR as a draft.
On the XENONnT fork we test with database access, on private forks there is no database access for security considerations.
All italic comments can be removed from this template.
coverage: 91.202% (-0.1%) from 91.349%
when pulling e7e8d0ec7ef4b95f24a7c6231ae37b4539414b0a on dependency_tree
into e7d95878b091ebe0a6d80732b256ba677c7a6ba1 on master.
|
gharchive/pull-request
| 2024-04-15T18:07:22 |
2025-04-01T04:55:47.917780
|
{
"authors": [
"coveralls",
"dachengx"
],
"repo": "XENONnT/straxen",
"url": "https://github.com/XENONnT/straxen/pull/1363",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
2667881552
|
Can non-square images be trained on ControlNet?
I have a few questions:
Is it mandatory to use 1024x1024 resolution when fine-tuning ControlNet? For example, is it possible to train using 1280x768 images?
Related to question 1, does ControlNet support training with various bucket resolutions?
Has the pre-trained ControlNet model been trained on images of diverse resolutions?
If the answer to question 3 is no, is the model still capable of performing inference on images with different resolutions?
anyone please answer me ToT
Yes, it does support different resolutions, but both width and height should be divisible by 8.
oh, thank you. Can a mixed dataset with varying image sizes still be trained in the same way?
SD3's paper, section5.3 described this in detail
|
gharchive/issue
| 2024-11-18T09:45:23 |
2025-04-01T04:55:47.932949
|
{
"authors": [
"Artemis1111",
"chenlijn",
"oukohou"
],
"repo": "XLabs-AI/x-flux",
"url": "https://github.com/XLabs-AI/x-flux/issues/136",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1109076953
|
config.json reciding and backup after process alteration
Hi,
it kinda stings me that changes made to the configuration by the executeable are not shown in the advanced configuration area.
I assume you are working with two different config files here, one which the user is editing and one that recides in /data/user/0/<project>/files/config.json (as per the log). On save the editor content is copied to /data/user/0/<project>/files/config.json.
That is a step I am not fully understanding (would be nice if you could enlighten me).
As far as my experience with xmrig goes, the config is not locked on execution start and is monitored for the exact purpose of alteration while running. The autosave value makes sure the executeable can save to the config as well.
Using a unified file as intended would save advanced users a lot of pain once a proper config has been figured out.
Right now the changes made by the executeable are not reflected within the apps editor, forcing the app to go through benchmark and config check whenever a mining process ist started.
The edit file is saved as string in local Data Base, each start the content sent in base64 to the native module (Kotlin), little modified (adds xmrig httpd - to pull stats like hashrate), output colors and intervals.
In last versions each update in backend (xmrig) important parts are sent back to the UI and saved in the data base.
Maybe its better to save a separate json file and give the user the ability to edit the file directly - but its can cause different set of bugs - if the user disable the xmrig httpd - 50% of the app will stop working.
We need to thing about it, mabye some hybrid version between the solutions... I don't know....
Maybe its better to save a separate json file and give the user the ability to edit the file directly - but its can cause different set of bugs - if the user disable the xmrig httpd - 50% of the app will stop working.
Usually one would hide such a thing behind a setting in the settings menu marked as something cryptic like developer options and after that a data path to where the file rescides. Very unlikely that anyone who shouldn't touch this, would.
|
gharchive/issue
| 2022-01-20T09:42:55 |
2025-04-01T04:55:47.945727
|
{
"authors": [
"PaxJaromeMalues",
"garrylachman"
],
"repo": "XMRig-for-Android/xmrig-for-android",
"url": "https://github.com/XMRig-for-Android/xmrig-for-android/issues/89",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2161944133
|
fixReducedOffersV1 still not covering all cases and book blocking (Version: 2.1.0)
Issue Description
Amendment fixReducedOffersV1 https://xrpl.org/resources/known-amendments#fixreducedoffersv1 is still not fixing all case where the order book is blocked
two cases have now been identified on mainnet.
Supporting Files
@lathanbritz, I'm disappointed but not surprised that I didn't fix all the ways that order books can be blocked. I have something else I'm working on right now, but this issue is next on my list after that's done.
It would help a lot if you can identify a specific historic transaction that failed because of this blockage. Of course you can only point at transactions that fail with a tec code. So there may not be any such transactions on the network. But being able to replay such a transaction might help me see the reason behind the blockage. If you know of a transaction like that please mention it here. Thanks.
@scottschurr rLtdxMMXPs21PYnnSiwm85EabdHDH4oVhi has blocked the USD book (its the one with the partially filled order), it does not have too many tx
@lathanbritz that helps, thanks. That's current? Like that's the state of the book for, say, ledger 86325875? Is there a way you can get the ledger index of the blocking offer (from rLtdxMMXPs21PYnnSiwm85EabdHDH4oVhi). That would be a big help. Then I can look back at that offer's history.
@scottschurr we have another instance today
ledger index 86,858,489 its present
account https://xrpscan.com/account/rJTyFN9PYy9uWJ3ixsSBfvhZGgvvjNt8MV
@lathanbritz, thanks for the information on that recent blockage.
I have a unit test that is able to reproduce the last blockage that you reported (by rLtdxMMXPs21PYnnSiwm85EabdHDH4oVhi). I'm currently looking into how to prevent the source of that particular blockage.
When I get an opening I'll research this more recent blockage, since it may not have the same source as the older one that I'm currently looking at.
@shortthefomo, first the good news: I found the source of the problem for the USD/Bitstamp:XRP blockage that you reported earlier. I have a proposed fix (not yet a pull request), but that will of course need to go through code review.
Regarding the reported order book blockage in ledger 86,858,489 I have less good news. I analyzed the top-most 150 offers in the...
"TakerGets": "CSC"/"rCSCManTZ8ME9EoLrSHHYKW8PPwWMgkwr",
"TakerPays": "XRP"
order book in that ledger. That included the one you identified with TakerPays: "4" (that was offer 49). I don't see how any of those 150 offers could be blockers. I'm not saying you didn't see something funny with that order book. But I don't understand the nature of the symptom you were seeing.
All-in-all, however, it is forward progress to identify an additional source of order book blockage and find a fix for it. Thanks for the help.
Great news!
The symptoms that I see that makes it simple to spot is simply taking the first offers off ASK/BID books and then checking for a negative spread. If it's negative those books (ASK and BID) had an order that is "stuck".
That's how ive been identifying them, not sure if that information helps here.
Thank you for all the effort on this @scottschurr
this is expected to be fixed by fixReducedOffersV2, introduced in 2.3.0-b1
feel free to re-open this issue if more needs to be done
|
gharchive/issue
| 2024-02-29T19:20:33 |
2025-04-01T04:55:47.956219
|
{
"authors": [
"intelliot",
"lathanbritz",
"scottschurr",
"shortthefomo"
],
"repo": "XRPLF/rippled",
"url": "https://github.com/XRPLF/rippled/issues/4937",
"license": "ISC",
"license_type": "permissive",
"license_source": "github-api"
}
|
985939422
|
OVR Plugins conflict
XRTK - Mixed Reality Toolkit Bug Report
Describe the bug
Multiple plugins with the same name 'ovrplugin' (found at 'Packages/com.oculus.unity-integration/Runtime/VR/Plugins/1.61.0/Win64/OVRPlugin.dll' and 'Packages/com.xrtk.oculus/Runtime/Plugins/1.61.0/Win64/OVRPlugin.dll'). That means one or more plugins are set to be compatible with Editor. Only one plugin at the time can be used by Editor.
To Reproduce
Go to Open development branch of XRTK.Core https://github.com/XRTK/XRTK-Core/pull/862/commits/3a5469734e22e324ecb7636aaaad9458e6ed2936
Expected behavior
No conflicting plugins
Actual behavior
Conflicting Plugins
Screenshots
Your Setup
Operation System:
Windows
Unity Version: 2019.4.29f1
XRTK Packages
[x] com.xrtk.core 0.3.0-preview.15 https://github.com/XRTK/XRTK-Core/pull/862/commits/3a5469734e22e324ecb7636aaaad9458e6ed2936
This is a known issue and will be documented as such
|
gharchive/issue
| 2021-09-02T01:24:27 |
2025-04-01T04:55:47.962496
|
{
"authors": [
"StephenHodgson"
],
"repo": "XRTK/Oculus",
"url": "https://github.com/XRTK/Oculus/issues/129",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
465224026
|
Controller Visualisation not behaving as expected
XRTK - Mixed Reality Toolkit Bug Report
Describe the bug
The original intention with the Controller Visualisation Configuration, was that it provided ways to OVERRIDE the default configuration of the Visualiser.
In such, with nothing configured, it would visualise the Gizmo models (the model defaults) for al platforms, regardless of whether any overrides were configured. However, today it only renders a model if an override is applied.
To Reproduce
Create new XRTK project and configure
View Visualizer configuration and remove any configured overrides
Run the XRTK project with a headset / controllers connected
Expected behavior
With no override configuration provided, controllers should display with their global defaults
Actual behavior
Unless an override is configured, no controllers show.
Your Setup (please complete the following information)
Unity Version 2019.1.9
XRTK Version 0.1.11
Target Platform (please complete the following information)
ALL platforms with controllers
Resolved by #234
|
gharchive/issue
| 2019-07-08T12:14:44 |
2025-04-01T04:55:47.966248
|
{
"authors": [
"SimonDarksideJ"
],
"repo": "XRTK/XRTK-Core",
"url": "https://github.com/XRTK/XRTK-Core/issues/233",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2707660485
|
Add Opt-Out
Respectfully, this is cool, but I just don't like automated error reporting.
It's implemented now: https://github.com/Xalalau/GMod-Lua-Error-API/commit/885fe685feb8c442b710fc0edd1694d741483636
|
gharchive/issue
| 2024-11-30T16:59:09 |
2025-04-01T04:55:48.089042
|
{
"authors": [
"Denneisk",
"Xalalau"
],
"repo": "Xalalau/GMod-Lua-Error-API",
"url": "https://github.com/Xalalau/GMod-Lua-Error-API/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
858419646
|
Flutter SDK 打了patch 后对没有aspectd 项目有影响
19:29:39 > Task :flutter:compileFlutterBuildRelease
19:29:39 Unhandled exception:
19:29:39 Invalid argument(s): Missing canonical name for Reference to Widget.
19:29:39 #0 BinaryPrinter.writeNonNullReference (package:kernel/binary/ast_to_binary.dart:868:9)
19:29:39 #1 BinaryPrinter.visitSuperInitializer (package:kernel/binary/ast_to_binary.dart:1312:5)
19:29:39 #2 SuperInitializer.accept (package:kernel/ast.dart:2580:45)
19:29:39 #3 BinaryPrinter.writeNode (package:kernel/binary/ast_to_binary.dart:382:10)
19:29:39 #4 BinaryPrinter.writeNodeList (package:kernel/binary/ast_to_binary.dart:292:7)
19:29:39 #5 BinaryPrinter.visitConstructor (package:kernel/binary/ast_to_binary.dart:1183:5)
19:29:39 #6 Constructor.accept (package:kernel/ast.dart:1976:40)
19:29:39 #7 BinaryPrinter.writeConstructorNode (package:kernel/binary/ast_to_binary.dart:438:10)
19:29:39 #8 BinaryPrinter.writeConstructorNodeList (package:kernel/binary/ast_to_binary.dart:337:7)
19:29:39 #9 BinaryPrinter.visitClass (package:kernel/binary/ast_to_binary.dart:1141:5)
19:29:39 #10 Class.accept (package:kernel/ast.dart:1379:38)
19:29:39 #11 BinaryPrinter.writeClassNode (package:kernel/binary/ast_to_binary.dart:424:10)
19:29:39 #12 BinaryPrinter.writeClassNodeList (package:kernel/binary/ast_to_binary.dart:319:7)
19:29:39 #13 BinaryPrinter.visitLibrary (package:kernel/binary/ast_to_binary.dart:973:5)
19:29:39 #14 Library.accept (package:kernel/ast.dart:607:38)
19:29:39 #15 BinaryPrinter.writeLibraryNode (package:kernel/binary/ast_to_binary.dart:403:10)
19:29:39 #16 BinaryPrinter.writeLibraries (package:kernel/binary/ast_to_binary.dart:710:9)
19:29:39 #17 BinaryPrinter.writeComponentFile. (package:kernel/binary/ast_to_binary.dart:553:7)
19:29:39 #18 Timeline.timeSync (dart:developer/timeline.dart:163:22)
19:29:39 #19 BinaryPrinter.writeComponentFile (package:kernel/binary/ast_to_binary.dart:536:14)
19:29:39 #20 FrontendCompiler.writeDillFile (package:frontend_server/frontend_server.dart:751:15)
19:29:39 #21 FrontendCompiler.compile (package:frontend_server/frontend_server.dart:562:13)
19:29:39
19:29:39 #22 _FlutterFrontendCompiler.compile (file:///home/flutter/flutter_linux_1.17.3-stable/.pub-cache/git/event_track-afa105d576be0588df429ecfc5a0d1bff3309fd9/lib/src/flutter_frontend_server/server.dart:58:22)
19:29:39 #23 starter (file:///home/flutter/flutter_linux_1.17.3-stable/.pub-cache/git/event_track-afa105d576be0588df429ecfc5a0d1bff3309fd9/lib/src/flutter_frontend_server/server.dart:197:27)
19:29:39 #24 main (file:///home/flutter/flutter_linux_1.17.3-stable/.pub-cache/git/event_track-afa105d576be0588df429ecfc5a0d1bff3309fd9/lib/src/flutter_frontend_server/starter.dart:13:30)
19:29:39 #25 _startIsolate. (dart:isolate-patch/isolate_patch.dart:299:32)
19:29:39 #26 _RawReceivePortImpl._handleMessage (dart:isolate-patch/isolate_patch.dart:168:12)
打patch 过后的:
@kangwang1988
自身问题,已解决
|
gharchive/issue
| 2021-04-15T02:37:13 |
2025-04-01T04:55:48.138888
|
{
"authors": [
"JsonYuu"
],
"repo": "XianyuTech/aspectd",
"url": "https://github.com/XianyuTech/aspectd/issues/113",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1525130777
|
[DUCK]支持“三月五号的九天后”
支持“三月五号的九天后”,输出日期:3月14号
done
|
gharchive/issue
| 2023-01-09T08:18:39 |
2025-04-01T04:55:48.141697
|
{
"authors": [
"zhangsonglei"
],
"repo": "XiaoMi/MiNLP",
"url": "https://github.com/XiaoMi/MiNLP/issues/153",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
363280780
|
Cannot connect to AWS instance using remmina(VNC error)
Hello,
Today I was trying to connect to the AWS instance using remmina for which i followed the set-up guide. But once I hit connect I get error saying that its a error :VNC problem connecting, and says some problem.
Regards
Adarsh Sawant
I looks like Remmina is trying to use the VNC protocol to connect. You need to make sure to use RDP (Remote Desktop Protocol) instead.
Hi Thomas,
I opened Remmina using RDP.
In the next step I have to select between whether to use XORG or XVNC .
If I choose XORG the screen crashes and the Remote monitor just vanishes aways.
If I choose XVNC the then I get the same error as mentioned above.
Regards
Hi Thomas,
Is there any way of installing Desktop environment on EC2 instance ? If you yes can you please refer me to it ?
Regards
Adarsh Sawant
On the main README page for these labs, you find a link to setup instructions. Follow these and you will find how to install GNOME on your EC2 instance and enable RDP services:
https://github.com/Xilinx/SDAccel-AWS-F1-Developer-Labs/blob/master/setup/instructions.md#installing-a-gui-desktop
|
gharchive/issue
| 2018-09-24T19:07:21 |
2025-04-01T04:55:48.145254
|
{
"authors": [
"ThomasXilinx",
"adu81020799"
],
"repo": "Xilinx/SDAccel-AWS-F1-Developer-Labs",
"url": "https://github.com/Xilinx/SDAccel-AWS-F1-Developer-Labs/issues/7",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
844318586
|
vai_c_tensorflow2 compile error, if layer name contains '/'
Hi
I'm trying Vitis-AI 1.3.2 (34794f6) and docker xilinx/vitis-ai-gpu, encounting error in vai_c_tensorflow2
I'm using very simple model, to test docker environment.
The network contains only one layer keras.layers.Conv2D.
If layer name = 'conv_conv', vai_c_tensorflow goes on.
But layer name = 'conv/conv', vai_c_tensorflow falls error.
this is reported error message.
(vitis-ai-tensorflow2) Vitis-AI /workspace > vai_c_tensorflow2 -m ./quantized_vsm.h5 -a /opt/vitis_ai/compiler/arch/DPUCAHX8H/U50/arch.json -o temp -n vsm --options '{"input_shape": "1,224,224,3"}'
**************************************************
* VITIS_AI Compilation - Xilinx Inc.
**************************************************
[INFO] Namespace(inputs_shape=['1,224,224,3'], layout='NHWC', model_files=['./quantized_vsm.h5'], model_type='tensorflow2', out_filename='temp/vsm_org.xmodel', proto=None)
in_shapes: [[1, 224, 224, 3]]
[INFO] tensorflow2 model: quantized_vsm.h5
Traceback (most recent call last):
File "/opt/vitis_ai/conda/envs/vitis-ai-tensorflow2/bin/xnnc-run", line 33, in <module>
sys.exit(load_entry_point('xnnc==1.3.0', 'console_scripts', 'xnnc-run')())
File "/opt/vitis_ai/conda/envs/vitis-ai-tensorflow2/lib/python3.7/site-packages/xnnc/__main__.py", line 194, in main
normal_run(args)
File "/opt/vitis_ai/conda/envs/vitis-ai-tensorflow2/lib/python3.7/site-packages/xnnc/__main__.py", line 178, in normal_run
in_shapes=in_shapes if len(in_shapes) > 0 else None,
File "/opt/vitis_ai/conda/envs/vitis-ai-tensorflow2/lib/python3.7/site-packages/xnnc/xconverter.py", line 131, in run
xmodel = CORE.make_xmodel(model_files, model_type, _layout, in_shapes)
File "/opt/vitis_ai/conda/envs/vitis-ai-tensorflow2/lib/python3.7/site-packages/xnnc/core.py", line 104, in make_xmodel
model_files, layout, in_shapes=in_shapes, model_type=model_t
File "/opt/vitis_ai/conda/envs/vitis-ai-tensorflow2/lib/python3.7/site-packages/xnnc/translator/tensorflow_translator.py", line 89, in to_xmodel
model_fmt, (model_name, raw_nodes) = cls.load_raw_model(model_files)
File "/opt/vitis_ai/conda/envs/vitis-ai-tensorflow2/lib/python3.7/site-packages/xnnc/translator/tensorflow_translator.py", line 127, in load_raw_model
return "h5", cls.__load_from_hdf5(model_files)
File "/opt/vitis_ai/conda/envs/vitis-ai-tensorflow2/lib/python3.7/site-packages/xnnc/translator/tensorflow_translator.py", line 1802, in __load_from_hdf5
for x in param.attrs.get("weight_names").tolist()
AttributeError: 'NoneType' object has no attribute 'tolist'
Hi @three-cups
So far as I know, there is no limitation for Xcompiler about the layer name. Xcompiler will find the weight according to the layer's name. And the error shows that Xcompiler can't find the weight according to the layer's name. It seems that you change only the layer name without synchronously changing the mapping link between layer and weights.
A quick test, you can change the layer name from conv_conv to conv_conv_test to see if the compiler works.
Hi, @qianglin-xlnx
Thank you for your replying, but I can't get result you mentioned.
My test sequence is...
1st : set layer name "conv_conv" ... success
2nd : set layer name "conv/conv" ... fail
3rd : set layer name "conv_conv_test" ... success
It may the matter of vitis_quantizer, not of Xcompiler.
Please check attached my model file (not quantized) and quantize script.
vsm_test.zip
Hi @three-cups
Thank you for the files you provided. They are helpful and we have confirmed that it's a bug. We will fix this bug in the next release.
Before we fix it, please avoid change the layer name with "/" . Thank you very much.
|
gharchive/issue
| 2021-03-30T10:05:39 |
2025-04-01T04:55:48.151260
|
{
"authors": [
"qianglin-xlnx",
"three-cups"
],
"repo": "Xilinx/Vitis-AI",
"url": "https://github.com/Xilinx/Vitis-AI/issues/355",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1704048657
|
AssertionError: Im2Col_0: ConvolutionInputGenerator1D works only for 1D convs
Discussed in https://github.com/Xilinx/finn/discussions/602
Originally posted by sakshigoel155 May 27, 2022
Hi,
I have a model with 4 2D convolution layers followed by 3 dense layers. I generate an ONNX model for it with input size (1,1,4,512).
While trying out DataflowBuildConfig on it to generate a FINN representation, I get the following error.
Final outputs will be generated in output_estimates_only
Build log is at output_estimates_only/build_dataflow.log
Running step: step_qonnx_to_finn [1/8]
Running step: step_tidy_up [2/8]
Running step: step_streamline [3/8]
Traceback (most recent call last):
File "/workspace/finn/src/finn/builder/build_dataflow.py", line 166, in build_dataflow_cfg
model = transform_step(model, cfg)
File "/workspace/finn/src/finn/builder/build_dataflow_steps.py", line 295, in step_convert_to_hls
model = model.transform(to_hls.InferConvInpGen())
File "/workspace/finn-base/src/finn/core/modelwrapper.py", line 141, in transform
(transformed_model, model_was_changed) = transformation.apply(
File "/workspace/finn/src/finn/transformation/fpgadataflow/convert_to_hls_layers.py", line 201, in apply
assert is_1d_convolution, (
AssertionError: Im2Col_0: ConvolutionInputGenerator1D works only for 1D convs
=======================================================================
I also tried the flow using bnn-pynq tutorial and after the streamlining step, while converting to HLS with createdataflowpartition, I get the same error,
I looked at the error further and it seems, it can only work for square input images. My input is not square input image but a 4X512 input, which uses EEG time series data from 4 channels, 512 samples at a time.
The graph after streamlining, just before
parent_model = model.transform(CreateDataflowPartition()), without fpgadataflow.convert_to_hls_layers.InferConvInpGen, looks like in the attached image
Can someone please help in finding out a solution on how can i go ahead?. Will be really thankful if someone can help me, as my project deadline is near.
Thanks
To add some context, running into almost the exact same issue though did not want to repost as the discussion existed, so just turned it from a discussion to an issue.
Hi @joshua99911 ,
Thanks for providing all this information, could you please convert the request by using the provided template: https://github.com/Xilinx/finn/blob/main/.github/ISSUE_TEMPLATE/bug_report.md
|
gharchive/issue
| 2023-05-10T14:24:55 |
2025-04-01T04:55:48.172584
|
{
"authors": [
"auphelia",
"joshua99911"
],
"repo": "Xilinx/finn",
"url": "https://github.com/Xilinx/finn/issues/814",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
2401104182
|
Bump to upstream 'e869c5fe'
This bumps enough of upstream so that the merge-base is compatible with our LLVM version (with https://github.com/Xilinx/llvm-project/pull/214)
e869c5fe bumps the LLVM submodule of ONNX-MLIR. Shouldn't we also incorporate this bump?
e869c5fe bumps the LLVM submodule of ONNX-MLIR. Shouldn't we also incorporate this bump?
Yes, good observation. I'm waiting for https://github.com/Xilinx/llvm-project/pull/214 to land to bump to that merge here. This is also why the tests on this PR still fail.
Automation gave me a4f8191f09077832bdc8a28b7dc4d805e9685176 as a last good commit to merge to, which is just one commit after the merge base 2cc3f15b0d8a3d5157c5721d618f44a629cd6d5a. We can take this but I see little value in making a bump for one commit... so I guess we're pretty much blocked on ONNX-MLIR until there is an LLVM bump. What do you think?
Automation gave me a4f8191f09077832bdc8a28b7dc4d805e9685176 as a last good commit to merge to, which is just one commit after the merge base 2cc3f15b0d8a3d5157c5721d618f44a629cd6d5a. We can take this but I see little value in making a bump for one commit... so I guess we're pretty much blocked on ONNX-MLIR until there is an LLVM bump. What do you think?
Yes, we can let onnx-mlir be until we have a newer LLVM version; if we also cannot bump LLVM, we could also go to a newer onnx-mlir while manually keeping it compatible with an older LLVM version (usually those are only a few changes).
|
gharchive/pull-request
| 2024-07-10T15:43:50 |
2025-04-01T04:55:48.178727
|
{
"authors": [
"cferry-AMD",
"mgehre-amd"
],
"repo": "Xilinx/onnx-mlir",
"url": "https://github.com/Xilinx/onnx-mlir/pull/139",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2526984725
|
🛑 Haty Server is down
In 626debb, Haty Server (https://hashi.sbs/_matrix/consent) was down:
HTTP code: 502
Response time: 684 ms
Resolved: Haty Server is back up in da9569c after 5 minutes.
|
gharchive/issue
| 2024-09-15T15:54:51 |
2025-04-01T04:55:48.202339
|
{
"authors": [
"MomentQYC"
],
"repo": "YAT-Publish/status",
"url": "https://github.com/YAT-Publish/status/issues/776",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1975456416
|
Referral tracking parameter
Code of Conduct
[X] I agree to follow this project's Code of Conduct.
Submission validity
[X] This is not a personal support request, that should be posted on discussions community.
[X] I checked current issues and this request isn't a duplicate of an existing issue, opened or closed.
Summary
Track a parameter on the end of short urls to allow tracking of where those link were clicked from.
Examples: myourls.com/hello?ref=qrcode or myourls.com/hello?ref=discord
When these links are accessed admins can see how many time a link has been clicked based on the ref parameter. This is helpful to know where a shortened link was accessed.
Motivation
Currently to do something similar you need to create duplicate short urls that all need to be unique. They also aren't dynamic.
This is useful to see where traffic is coming from and where urls are possibly gaining traction.
Lastly, not every site or source will pass something of "referral" header so this is a great way to guarantee some method of tracking a link and its placement.
Site note: this is a feature on questionpro.com that i recently used and really liked being able to see where people were accessing a survey it put out giving me another piece of information into demographics.
Describe the solution
Add a referral join-table
example:
keyword
ref
count
hello
discord
4
hello
qrcode
3
And update UI to show break down based on ref param
Thanks for opening this issue @jaredkotoff.
Isn't this something manageable with a plugin?
It could be, but it's also somewhat common feature of other url shortens.
And unless the database scheme has a place to track this sort of thing, I didn't see anything in the docs about plug-ins making changes to the database. Wasn't sure if that was frowned upon.
In my opinion a referring parameter would be a value-add to core. Personally it would provide me more meaningful data than having location stats.
Sorry I'm a bit late in the discussion. What is @YOURLS/core thoughts about this ?
I'm definitely gettting "plugin territory" vibes on this one, I don't think this is something 99% of users need. It kinda defeats the "short URL" concept to add stuff to your short URL. It's been years and hundreds of thousands of YOURLS installs, and this is the first time this feature is requested.
Likely not the first time someone's wanted it, but I agree it sounds like plugin territory.
this is the first time this feature is requested.
Im surprised no one has wanted UTM params before.
Is there documentation for modifing the db for plug-ins?
There have been a couple Google Analytics plugins, see https://github.com/YOURLS/awesome
There is no documentation for modifying the DB in plugins. A couple of them do it. It's just a matter of performing a SQL request. I would rely on the "plugin_$(plugin-name)" hook to create or alter tables, to make sure it's done once only. See for instance https://github.com/YOURLS/cache-stats-pages/blob/master/plugin.php#L48-L55
Hi @jaredkotoff
if you set the parameter 'YOURLS_UNIQUE_URLS' to false in the file user/config.php you are able to create duplicate urls but with different short code urls: I am using this to track the same URL but with different short urls. Then I generate QR Codes to better understand from which print product (poster, catalog etc.) users are coming or if for example poster qr codes are scanned i.e. are qr codes something people do really scan ;)
cheers
@sbarbieri i mention how this is not a viable solution. It's cumbersome and doesn't solve the need to create them on the fly.
It also makes it hard to compare the traffic sources.
Anyways, Im already working on the plug-in
|
gharchive/issue
| 2023-11-03T04:52:17 |
2025-04-01T04:55:48.217600
|
{
"authors": [
"LeoColomb",
"dgw",
"jarebear6expepjozn6rakjq5iczi3irqwphcvb",
"jaredkotoff",
"ozh",
"sbarbieri"
],
"repo": "YOURLS/YOURLS",
"url": "https://github.com/YOURLS/YOURLS/issues/3659",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2634500028
|
added gsap
🛠️ Fixes Issue
Fixes: #429
👨💻 Description
What does this PR do?
Briefly describe the feature or bug fix you've implemented.
Highlight any major changes or added functionality.
Mention how it enhances the project.
📄 Type of Change
[ ] Bug fix (non-breaking change which fixes an issue)
[x] New feature (non-breaking change which adds functionality)
[ ] Breaking change (fix or feature that would cause existing functionality to not work as expected)
[ ] Documentation update (adds or updates related documentation)
📷 Screenshots/GIFs (if any)
Include screenshots or GIFs to demonstrate your changes
✅ Checklist
[x] I am a participant of GSSoC-ext.
[x] I have followed the contribution guidelines of this project.
[x] I have made this change from my own.
[x] I have taken help from some online resources.
[x] My code follows the style guidelines of this project.
[x] I have performed a self-review of my own code.
[x] I have added documentation to explain my changes.
Mandatory Tasks
[x] Make sure you have self-reviewed the code. A decent size PR without self-review might be rejected.
🤝 GSSoC Participation
[x] This PR is submitted under the GSSoC program.
[x] I have taken prior approval for this feature/fix.
https://github.com/user-attachments/assets/b45ee9fd-7791-42e0-80bc-ea0caa25716b
|
gharchive/pull-request
| 2024-11-05T05:57:51 |
2025-04-01T04:55:48.239707
|
{
"authors": [
"priyashuu"
],
"repo": "YadavAkhileshh/Alien-Invasion-Defense",
"url": "https://github.com/YadavAkhileshh/Alien-Invasion-Defense/pull/674",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
718625436
|
No error and not work
I run two commands.
python run_basics.py
python demo.py -i images -o result --isDlib True
There seems to be no error. But it didn't seem to be work.
Anyone can help me, please
|
gharchive/issue
| 2020-10-10T13:57:34 |
2025-04-01T04:55:48.242170
|
{
"authors": [
"fashionguy"
],
"repo": "YadiraF/PRNet",
"url": "https://github.com/YadiraF/PRNet/issues/196",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
121151754
|
[1363] Fix overflowing availability number [master]
Resolves #1363
Sure
|
gharchive/pull-request
| 2015-12-09T03:25:27 |
2025-04-01T04:55:48.244544
|
{
"authors": [
"orenyk",
"squidgetx"
],
"repo": "YaleSTC/reservations",
"url": "https://github.com/YaleSTC/reservations/pull/1366",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
940573788
|
Create JSON file with stats for all the YANGPageMain.json files
XXXYANGPageMain.json, where XXX is prefix contains statistics on total number of parsed modules, passed & failed ones.
Create one JSON file which will contain summary statistics on all the XXXYANGPageMain.json files
Done
|
gharchive/issue
| 2021-07-09T09:08:03 |
2025-04-01T04:55:48.252410
|
{
"authors": [
"SlavomirMazurPantheon"
],
"repo": "YangCatalog/sdo_analysis",
"url": "https://github.com/YangCatalog/sdo_analysis/issues/88",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.