id
stringlengths 4
10
| text
stringlengths 4
2.14M
| source
stringclasses 2
values | created
timestamp[s]date 2001-05-16 21:05:09
2025-01-01 03:38:30
| added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| metadata
dict |
---|---|---|---|---|---|
1845019449
|
Typo in docs
Describe the bug
The variants docs has a wrong property. It should be called compoundVariants according to the source code
Your Example Website or App
No response
Steps to Reproduce the Bug or Issue
Custom variants
Expected behavior
It should be called compoundVariants
Screenshots or Videos
No response
Operating System Version
All
Browser
Chrome
Good catch thank you!
Fix is ont its way
https://github.com/nextui-org/nextui/commit/cef7235c4bb3c8c3938bee2496b19beb6df83d22
|
gharchive/issue
| 2023-08-10T11:48:39 |
2025-04-01T04:35:12.605900
|
{
"authors": [
"ccorreia64",
"jrgarciadev"
],
"repo": "nextui-org/nextui",
"url": "https://github.com/nextui-org/nextui/issues/1343",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1980927590
|
[BUG] - Accordion error Error: Unknown element <[object Object]> in collection.
NextUI Version
2.0.24
Describe the bug
I use @nextui-org/accordion @ 2.0.24 version can't make it work!
Here's the full code:
import {Accordion, AccordionItem} from "@nextui-org/accordion"
export default function Home() {
const defaultContent =
"Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat.";
return( <Accordion>
<AccordionItem key="1" aria-label="Accordion 1" title="Accordion 1">
{defaultContent}
</AccordionItem>
<AccordionItem key="2" aria-label="Accordion 2" title="Accordion 2">
{defaultContent}
</AccordionItem>
<AccordionItem key="3" aria-label="Accordion 3" title="Accordion 3">
{defaultContent}
</AccordionItem>
</Accordion>);
}
The code is copied from https://nextui.org/docs/components/accordion
this is the full console error:
- error node_modules/.pnpm/@react-stately+collections@3.10.2_react@18.2.0/node_modules/@react-stately/collections/dist/import.mjs (205:0) @ $eb2240fc39a57fa5$export$bf788dd355e3a401.getFullNode
- error Error: Unknown element <[object Object]> in collection.
at getFullNode.next (<anonymous>)
at iterateCollection.next (<anonymous>)
at Generator.next (<anonymous>)
Your Example Website or App
No response
Steps to Reproduce the Bug or Issue
pnpm i @nextui-org/accordion
copy example code from "https://nextui.org/docs/components/accordion"
pnpm dev
you can get the error
Expected behavior
works as normal
Screenshots or Videos
No response
Operating System Version
macOS 13.4.1 (c)
Browser
Chrome
Duplicated https://github.com/nextui-org/nextui/issues/1619
|
gharchive/issue
| 2023-11-07T09:36:44 |
2025-04-01T04:35:12.611276
|
{
"authors": [
"jrgarciadev",
"universeroc"
],
"repo": "nextui-org/nextui",
"url": "https://github.com/nextui-org/nextui/issues/1912",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1279138512
|
Prettier?
🤔
Yeah we could have a config but I think that may be too opinionated? I personally just use the default that comes with the vscode extension. If someone wants prettier they can install/configure it themselves imo. What do you think?
Make it optional ? When generating a project, it prompts whether to add a prettier config file and packages or not.
Thought a bit about this and I don't think it meaningfully differs from my stance on "add Jotai/Zustand"
You can just NPM install if you want prettier. No need for us to include. I think I let my desire for prettier get ahead of me 😂
I think that most people have prettier installed in their editor and if they have a preferred config then can do that on their own. Closing this now after discussing on discord.
|
gharchive/issue
| 2022-06-21T21:42:26 |
2025-04-01T04:35:12.633642
|
{
"authors": [
"TheoBr",
"booleans-oss",
"nexxeln"
],
"repo": "nexxeln/create-t3-app",
"url": "https://github.com/nexxeln/create-t3-app/issues/25",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2739789935
|
[★𝐕𝐈𝐑𝐀𝐋★] 𝐒𝐮𝐛𝐡𝐚𝐬𝐡𝐫𝐞𝐞 𝐒𝐚𝐡𝐮 Video 𝙻𝚎𝚊𝚔𝚎𝚍 𝙽𝚘𝚠 𝘖𝘳𝘪𝘨𝘪𝘯𝘢𝘭 𝖳𝗋𝖾𝗇𝖽𝗂𝗇𝗀 Clips LINK On Social Media X Twitter
20 sec ago [★𝐕𝐈𝐑𝐀𝐋★] 𝐒𝐮𝐛𝐡𝐚𝐬𝐡𝐫𝐞𝐞 𝐒𝐚𝐡𝐮 Video 𝙻𝚎𝚊𝚔𝚎𝚍 𝙽𝚘𝚠 𝘖𝘳𝘪𝘨𝘪𝘯𝘢𝘭 𝖳𝗋𝖾𝗇𝖽𝗂𝗇𝗀 Clips LINK On Social Media X Twitter
..
..
..
..
..
..
𝐒𝐮𝐛𝐡𝐚𝐬𝐡𝐫𝐞𝐞 𝐒𝐚𝐡𝐮, a young and talented digital creator, recently gained widespread attention on social media platforms with her viral video. The video quickly became a trending topic across various platforms, sparking a significant amount of discussion among viewers. As a rising star in the digital world, 𝐒𝐮𝐛𝐡𝐚𝐬𝐡𝐫𝐞𝐞 𝐒𝐚𝐡𝐮's creativity and content have captivated audiences, contributing to her growing popularity on platforms like X and Twitter.
This viral moment has sparked conversations about the impact of digital content on social media trends. 𝐒𝐮𝐛𝐡𝐚𝐬𝐡𝐫𝐞𝐞 𝐒𝐚𝐡𝐮's video, while generating attention, highlights the growing influence of young content creators who are reshaping the landscape of online media. With her engaging presence, 𝐒𝐮𝐛𝐡𝐚𝐬𝐡𝐫𝐞𝐞 𝐒𝐚𝐡𝐮 continues to inspire others to explore the possibilities of digital creation and interaction on social platforms.
.
.
.
..
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
..
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
..
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
..
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
..
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
..
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
..
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
..
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
..
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
To Reproduce
Additional Context
Validation
[X] I confirm this is a bug about nezha (Nezha Dashboard).
[X] I have searched Issues and confirm this bug has been reported before.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
|
gharchive/issue
| 2024-12-14T11:27:01 |
2025-04-01T04:35:12.755927
|
{
"authors": [
"ahidiqba",
"lliansm",
"wahidsarker",
"ylmenendez"
],
"repo": "nezhahq/nezha",
"url": "https://github.com/nezhahq/nezha/issues/739",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1743786754
|
Improve multiqc report
Just a couple of small improvements for the multiqc report.
PR checklist
[x] This comment contains a description of changes (with reason).
[ ] If you've fixed a bug or added code that should be tested, add tests!
[ ] If you've added a new tool - have you followed the pipeline conventions in the contribution docs
[ ] If necessary, also make a PR on the nf-core/nanostring branch on the nf-core/test-datasets repository.
[x] Make sure your code lints (nf-core lint).
[x] Ensure the test suite passes (nextflow run . -profile test,docker --outdir <OUTDIR>).
[ ] Usage Documentation in docs/usage.md is updated.
[ ] Output Documentation in docs/output.md is updated.
[ ] CHANGELOG.md is updated.
[ ] README.md is updated (including new tool citations and authors/contributors).
Maybe a CHANGELOG update to draw people';s attention the output might look different might nice, but MultiQC wise looks good 👍 (I'm assuming teh write_out_prepared_gex is related to that)
Thank you!
|
gharchive/pull-request
| 2023-06-06T12:39:10 |
2025-04-01T04:35:12.786436
|
{
"authors": [
"christopher-mohr"
],
"repo": "nf-core/nanostring",
"url": "https://github.com/nf-core/nanostring/pull/23",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
209791084
|
Issue with notifications when user is outside wifi and was an invite
Hi
Should notifications for an invited user work outside of wifi or not? I only receive them when in wifi never over cellular. The 'owner' always receives then
I am using iPad as the home hub
Also based on a motion sensor
That's how motion works
|
gharchive/issue
| 2017-02-23T15:17:42 |
2025-04-01T04:35:12.796820
|
{
"authors": [
"brownad"
],
"repo": "nfarina/homebridge",
"url": "https://github.com/nfarina/homebridge/issues/1161",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
113545661
|
Stock Count - The loading sign keeps running when validate stock count
For some stock counts containing many items, when trying to validate, system cannot complete the process since the loading sign on validate button keeps running.
stock count speed improved here: https://github.com/nfco/netforce/commit/e51e7cbe434ff7cf0b0c4ce58667d8a015046ffa
|
gharchive/issue
| 2015-10-27T09:51:45 |
2025-04-01T04:35:12.798082
|
{
"authors": [
"bankwirat",
"nf-dj"
],
"repo": "nfco/netforce",
"url": "https://github.com/nfco/netforce/issues/23",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1583810073
|
[BUG] Term search uses collapse to hide extended info
This creates issues with safari rendering, children will not get collapsed correctly, so parent box size stays the same.
@Brilator Can you check if this issue still happens in Swate-standalone?
I can only assume that Excel uses deprecated browser versions for add-ins even on mac, as this should be fully supported:
https://developer.mozilla.org/en-US/docs/Web/CSS/visibility?retiredLocale=de#browser_compatibility
still occurs in Safari, not in Firefox on Mac
|
gharchive/issue
| 2023-02-14T09:30:52 |
2025-04-01T04:35:12.800328
|
{
"authors": [
"Brilator",
"Freymaurer"
],
"repo": "nfdi4plants/Swate",
"url": "https://github.com/nfdi4plants/Swate/issues/270",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1858257057
|
[BUG] <Roster file has bad ID for Jay Cutler in 2010 and no ID in 2011>
Is there an existing issue for this?
[X] I have searched the existing issues
Have you installed the latest development version of the package(s) in question?
[X] I have installed the latest development version of the package.
What version of the package do you have?
na direct pull
Describe the bug
In the roster file, Jay Cutler's 2010 record uses Rashied Davies ID's. Jay Cutler does not have a record for 2011. Both Jay Cutler and Rashid Davies have inconsistent draft data (different records say they were drafted by different teams in different rounds)
Reprex
import pandas as pd
import numpy
## load roster files ##\
roster_url = 'https://github.com/nflverse/nflverse-data/releases/download/rosters'
rosters = []
for season in range(2006,2018):
## pull roster for that season ##
temp = pd.read_csv(
'{0}/roster_{1}.csv?raw=true'.format(
roster_url,
season
),
low_memory=False
)
rosters.append(temp)
## combine rosters ##
r = pd.concat(rosters)
## can see that jay cutler has wrong ID 2010, is missing 2011, and strange draft data ##
r[
r['full_name'] == 'Jay Cutler'
][[
'season','team','full_name','gsis_id',
'espn_id','pff_id','pfr_id','esb_id',
'entry_year','draft_club','draft_number'
]]
## if you inspect the wrong ID, you see it's Rashied Davies ##
r[
r['gsis_id'] == '00-0023429'
][[
'season','team','full_name','gsis_id',
'espn_id','pff_id','pfr_id','esb_id',
'entry_year','draft_club','draft_number'
]]
## to confirm it's not an issue with concatination of DFs, you can see issue is at the file ##
## level ##
r2010 = pd.read_csv(
'{0}/roster_{1}.csv?raw=true'.format(
roster_url,
2010
)
)
r2010[
r2010['full_name'] == 'Jay Cutler'
][[
'season','team','full_name','gsis_id',
'espn_id','pff_id','pfr_id','esb_id',
'entry_year','draft_club','draft_number'
]]
Expected Behavior
Expected behavior is that the roster file would have consistent information for both players and not have missing seasons
nflverse_sitrep
na did in python
Screenshots
No response
Additional context
No response
This has been resolved:
nflreadr::load_rosters(2010:2011) |>
dplyr::filter(full_name == "Jay Cutler") |>
dplyr::select(season, week, full_name, gsis_id)
#> ── nflverse roster data ────────────────────────────────────────────────────────
#> ℹ Data updated: 2023-09-06 10:40:56 PDT
#> # A tibble: 2 × 4
#> season week full_name gsis_id
#> <int> <int> <chr> <chr>
#> 1 2010 20 Jay Cutler 00-0024226
#> 2 2011 16 Jay Cutler 00-0024226
|
gharchive/issue
| 2023-08-20T20:18:53 |
2025-04-01T04:35:12.806261
|
{
"authors": [
"greerreNFL",
"john-b-edwards"
],
"repo": "nflverse/nflverse-data",
"url": "https://github.com/nflverse/nflverse-data/issues/31",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
607109161
|
Hyper-V cluster prepare isn't working
It's been a while since I've created a Hyper-V cluster and while working on the latest node templates #811, I noticed that deploying Hyper-V cluster is failing in the prepare step.
The problem is that neon-cli is trying to run the Hyper-V/Get-NetAdapter cmdlet but this no longer seems to be present in the Hyper-V namespace. The command exists without the namespace.
This used to work, so MSFT must have relocated the cmdlet or something.
We need to backport this to a new 1.x release.
This is fixed for v1.2.3 and v2.1.0
|
gharchive/issue
| 2020-04-26T20:44:16 |
2025-04-01T04:35:12.811087
|
{
"authors": [
"jefflill"
],
"repo": "nforgeio/neonKUBE",
"url": "https://github.com/nforgeio/neonKUBE/issues/812",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1126892391
|
chore(deps): update dependency @angular/platform-browser to v13
This PR contains the following updates:
Package
Type
Update
Change
@angular/platform-browser
dependencies
major
~10.1.3 -> ~13.2.0
Release Notes
angular/angular
v13.2.1
Compare Source
animations
Commit
Type
Description
4644886aaf
perf
remove no longer needed CssKeyframes classes (#44903) (#44919)
common
Commit
Type
Description
b4e4617807
fix
include query parameters for open HTTP requests in verify (#44917)
compiler-cli
Commit
Type
Description
0778e6f7d7
fix
accept nullish coalescing operator for any and unknown types (#44862)
07185f4ed1
fix
enable nullish coalescing check only with strictNullChecks (#44862)
4a5ad1793f
fix
ensure casing of logical paths is preserved (#44798)
core
Commit
Type
Description
7ec482d9c2
fix
Add back support for namespace URIs in createElement of dom renderer (#44914)
250dc40a46
fix
flush delayed scoping queue while setting up TestBed (#44814)
forms
Commit
Type
Description
1aebbf8714
fix
ensure OnPush ancestors are marked dirty when the promise resolves (#44886)
6b7fffcbeb
fix
Update the typed forms migration schematic to find all files. (#44881)
Special Thanks
Alan, Andrew Kushnir, Andrew Scott, Aristeidis Bampakos, Arjen, Daniel Díaz, David Shevitz, Doug Parker, Dylan Hunn, Esteban Gehring, George Kalpakas, Jessica Janiuk, JoostK, Juri Strumpflohner, Lee Robinson, Maarten Tibau, Paul Gschwendtner, Theodore Brown, arturovt, dario-piotrowicz, fru2, markostanimirovic and mgechev
v13.2.0
Compare Source
Deprecations
The CachedResourceLoader and RESOURCE_CACHE_PROVIDER symbols were previously necessary in some cases to test AOT-compiled components with View Engine, but they are no longer needed since Ivy.
The ComponentFactory and ComponentFactoryResolver classes are deprecated. Since Ivy, there is no need to resolve Component factories. Please use other APIs where you Component classes can be used directly (without resolving their factories).
Since Ivy, the CompilerOptions.useJit and CompilerOptions.missingTranslation config options are unused, passing them has no effect.
Commit
Type
Description
9c11183e74
docs
deprecate CachedResourceLoader and RESOURCE_CACHE_PROVIDER symbols (#44749)
9f12e7fea4
docs
deprecate ComponentFactory and ComponentFactoryResolver symbols (#44749)
4e95a316ce
docs
deprecate unused config options from the CompilerOptions interface (#44749)
compiler
Commit
Type
Description
a4ab6d6b72
feat
add support for safe calls in templates (#44580)
abd1bc8039
fix
correct spans when parsing bindings with comments (#44785)
ed67a074ce
fix
properly compile DI factories when coverage reporting is enabled (#44732)
compiler-cli
Commit
Type
Description
fa835b5a29
feat
enable extended diagnostics by default (#44712)
73424def13
feat
provide the animations for DirectiveMeta (#44630)
fe3e4d6865
fix
Handle ng-template with structural directive in indexer (#44788)
7316e72ec5
fix
properly index elements when on a template (#44785)
100091ebf0
fix
remove leftover _extendedTemplateDiagnostics requirements (#44777)
d2ae96f742
fix
skip ExtendedTemplateCheckerImpl construction if there were configuration errors (#44778)
core
Commit
Type
Description
5626b34264
fix
consistently use namespace short name rather than URI (#44766)
94bfcdd9de
fix
error if NgZone.isInAngularZone is called with a noop zone (#44800)
forms
Commit
Type
Description
72092ebd26
feat
Allow a FormControl to use initial value as default. (#44434)
f7aa937cac
fix
Make some minor fixups for forward-compatibility with typed forms. (#44540)
router
Commit
Type
Description
5a4ddfd4f5
feat
Allow symbol keys for Route data and resolve properties (#44519)
Special Thanks
Alex Rickabaugh, Andrew Kushnir, Andrew Scott, Dario Piotrowicz, Derek Cormier, Doug Parker, Douglas Parker, Dylan Hunn, George Kalpakas, Jessica Janiuk, JoostK, Kristiyan Kostadinov, Martin Probst, Oleg Postoev, Stephanie Tuerk, Tim Bowersox, Wiley Marques, Yousaf Nawaz, dario-piotrowicz, iRealNirmal, ivanwonder and shejialuo
v13.1.3
Compare Source
animations
Commit
Type
Description
af0a152a2c
fix
apply setStyles to only rootTimelines (#44515)
compiler-cli
Commit
Type
Description
626f3f230b
perf
reduce analysis work during incremental rebuilds (#44731)
ngcc
Commit
Type
Description
f9ca4d8499
fix
support element accesses for export declarations (#44669)
Special Thanks
Alan Agius, Andrew Kushnir, AnkitSharma-007, Daniel Díaz, Dmytro Mezhenskyi, Jessica Janiuk, Joey Perrott, JoostK, Ramesh Thiruchelvam, dario-piotrowicz, iRealNirmal and Łukasz Holeczek
v13.1.2
Compare Source
animations
Commit
Type
Description
abc217b28e
fix
retain triggers values for moved tracked list items (#44578)
compiler
Commit
Type
Description
59eef29a6c
fix
correct spans when parsing bindings with comments (#44678)
compiler-cli
Commit
Type
Description
08049fa23f
fix
enable narrowing of using type guard methods (#44447)
a26afce68c
fix
fix crash during type-checking of library builds (#44587)
1e918b6f31
fix
handle property reads of ThisReceiver in the indexer (#44678)
63c8e56a3a
fix
incorrectly interpreting $any calls with a property read (#44657)
60fb27f12d
fix
properly index elements (#44678)
language-service
Commit
Type
Description
f5addee488
fix
revert the test files for Ivy (#44528)
Special Thanks
Abdurrahman Abu-Hijleh, Adam Plumer, Alex Rickabaugh, AlirezaEbrahimkhani, Andrew Kushnir, Andrew Scott, Borja Paz Rodríguez, Chihab Otmani, Chris Mancini, Dario Piotrowicz, Doug Parker, George Kalpakas, Joey Perrott, JoostK, Kristiyan Kostadinov, Kyoz, Patrick Prakash, Paul Gschwendtner, Serhey Dolgushev, Yousaf Nawaz, Yuchao Wu, alkavats1, dario-piotrowicz, huangqing, ivanwonder, shejialuo, twerske, wszgrcy and zuckjet
v13.1.1
Compare Source
animations
Commit
Type
Description
bb1d4ff315
fix
don't consume instructions for animateChild (#44357)
d8b6adb7bc
fix
should not invoke disabled child animations (#37724)
forms
Commit
Type
Description
bce108ab49
fix
_reduceValue arrow function now has correct types. (#44483)
998c1e63fe
fix
I indroduced a minor error in a previous PR: pendingValue is a value not a boolean flag. (#44450)
Special Thanks
Aristeidis Bampakos, Dylan Hunn, George Kalpakas, JoostK, Kristiyan Kostadinov, Paul Gschwendtner, Spej, Yousaf Nawaz, dario-piotrowicz, faso-dev, jaybell and zuckjet
v13.1.0
Compare Source
Deprecations
The downgradeModule function calls with NgModule factories are deprecated. Please use NgModule class based downgradeModule calls instead.
common
TestRequest from @angular/common/http/testing no longer
accepts ErrorEvent when simulating XHR errors. Instead instances of
ProgressEvent should be passed, matching with the native browser behavior.
Commit
Type
Description
dbc46d68b9
docs
deprecate factory-based signature of the downgradeModule function (#44090)
common
Commit
Type
Description
489cf42cd0
fix
incorrect error type for XHR errors in TestRequest (#36082)
13362972bb
perf
code size reduction of ngFor directive (#44315)
compiler
Commit
Type
Description
c85bcb0c63
feat
reference ICU message IDs from their placeholders (#43534)
core
Commit
Type
Description
5dff077d50
feat
add migration to remove entryComponents (#44308)
e65a245a0b
feat
add migration to remove entryComponents (#44322)
d56e3f43a1
feat
support TypeScript 4.5 (#44164)
http
Commit
Type
Description
d452b388bd
feat
add has() method to HttpContext class (#43887)
localize
Commit
Type
Description
d3cf222a81
feat
support "associated message ids" for placeholders (#43534)
ngcc
Commit
Type
Description
41265919aa
fix
correctly resolve UMD dependencies (#44381)
upgrade
Commit
Type
Description
34f990986c
feat
support NgModule class as an argument of the downgradeModule function (#43973)
Special Thanks
Alex Rickabaugh, Andrew Kushnir, Andrew Scott, Doug Parker, Dustin M. Eastway, Dylan Hunn, George Kalpakas, HyperLife1119, Jelle Bruisten, Jessica Janiuk, Joey Perrott, JoostK, Kristiyan Kostadinov, Markus Doggweiler, Paul Gschwendtner, Pei Wang, Pete Bacon Darwin and dario-piotrowicz
v13.0.3
Compare Source
compiler-cli
Commit
Type
Description
6cdbfdbe6e
fix
downlevel transform incorrectly extracting constructor parameters for nested classes (#44281)
305b76b45f
fix
interpret string concat calls (#44167)
core
Commit
Type
Description
0ca5c5bd09
fix
add missing info about a component in the "pipe could not be found" error message (#44081)
907da3977a
fix
destroy hooks not set up for useClass provider using forwardRef (#44281)
bcd3b4959b
fix
support cyclic metadata in TestBed overrides (#44215)
forms
Commit
Type
Description
96fedd249e
fix
make the FormControlStatus available as a public API (#44183)
language-service
Commit
Type
Description
cabc1786de
fix
Correctly parse inputs and selectors with dollar signs (#44268)
ngcc
Commit
Type
Description
b68994d20a
fix
correctly report error when collecting dependencies of UMD module (#44245)
6f5c0c1515
fix
ensure that ngcc does not write a lock-file into node_modules package directories (#44228)
bf5f734e9c
fix
support the UMD wrapper function format emitted by Webpack (#44245)
router
Commit
Type
Description
d265d0d241
fix
prevent componentless routes from being detached (#44240)
Special Thanks
Alan Agius, Alex Rickabaugh, Andrew Kushnir, Andrew Scott, Artur, Christian-E, David Shevitz, Doug Parker, Douglas Parker, Dylan Hunn, George Kalpakas, Jessica Janiuk, Joey Perrott, JoostK, Kristiyan Kostadinov, Marc Redemske, Paul Gschwendtner, Pei Wang, Pete Bacon Darwin, Ramesh Thiruchelvam, Ravi Chandra, Rohan Pednekar, Ruslan Usmanov, dario-piotrowicz, profanis and unknown
v13.0.2
Compare Source
This release contains various API docs improvements.
Special Thanks
Andrew Kushnir, Armen Vardanyan, Dylan Hunn, Joey Perrott, Martin von Gagern, Paul Gschwendtner, Pete Bacon Darwin, Ramesh Thiruchelvam, dario-piotrowicz and fusho-takahashi
v13.0.1
Compare Source
compiler
Commit
Type
Description
ee2031d9f4
fix
ensure that partially compiled queries can handle forward references (#44113)
e5a960b159
fix
generate correct code for safe method calls (#44088)
compiler-cli
Commit
Type
Description
dede29e4f3
fix
ensure literal types are retained when strictNullInputTypes is disabled (#38305)
04df3a0b92
fix
handle pre-release versions when checking version (#44109)
core
Commit
Type
Description
4c700b6244
fix
do not use Function constructors in development mode to avoid CSP violations (#43587)
platform-browser
Commit
Type
Description
30a27adf9a
fix
use correct parent in animation removeChild callback (#44033)
Special Thanks
A. Singh, Alan Agius, Alex Rickabaugh, Andrew Kushnir, Andrew Scott, Aristeidis Bampakos, George Kalpakas, Joe Martin (Crowdstaffing), Joel Lefkowitz, Joey Perrott, JoostK, Kristiyan Kostadinov, Michael Urban, Paul Gschwendtner, Pavan Kumar Jadda, Pei Wang, Pete Bacon Darwin, Roman Frołow, dario-piotrowicz, iRealNirmal, ileil, kreuzerk, mgechev, profanis and raman
v13.0.0
Compare Source
Blog post "Angular v13 is now available".
Breaking Changes
common
The behavior of the SpyLocation used by the RouterTestingModule has changed
to match the behavior of browsers. It no longer emits a 'popstate' event
when Location.go is called. In addition, simulateHashChange now
triggers both a hashchange and a popstate event.
Tests which use location.go and expect the changes to be picked up by
the Router should likely change to simulateHashChange instead.
Each test is different in what it attempts to assert so there is no
single change that works for all tests. Each test using the SpyLocation to
simulate browser URL changes should be evaluated on a case-by-case basis.
core
TypeScript versions older than 4.4.2 are no longer supported.
NodeJS versions older than v12.20.0 are no longer
supported due to the Angular packages using the NodeJS package exports
feature with subpath patterns.
The WrappedValue class can no longer be imported from @angular/core,
which may result in compile errors or failures at runtime if outdated
libraries are used that are still using WrappedValue. The usage of
WrappedValue should be removed as no replacement is available.
forms
A new type called FormControlStatus has been introduced, which is a union of all possible status strings for form controls. AbstractControl.status has been narrowed from string to FormControlStatus, and statusChanges has been narrowed from Observable<any> to Observable<FormControlStatus>. Most applications should consume the new types seamlessly. Any breakage caused by this change is likely due to one of the following two problems: (1) the app is comparing AbstractControl.status against a string which is not a valid status; or, (2) the app is using statusChanges events as if they were something other than strings.
router
The default url serializer would previously drop
everything after and including a question mark in query parameters. That
is, for a navigation to /path?q=hello?&other=123, the query
params would be parsed to just {q: 'hello'}. This is
incorrect because the URI spec allows for question mark characers in
query data. This change will now correctly parse the params for the
above example to be {v: 'hello?', other: '123'}.
Previously null and undefined inputs for routerLink were
equivalent to empty string and there was no way to disable the link's
navigation.
In addition, the href is changed from a property HostBinding() to an
attribute binding (HostBinding('attr.href')). The effect of this
change is that DebugElement.properties['href'] will now return the
href value returned by the native element which will be the full URL
rather than the internal value of the RouterLink href property.
When storing and retrieving a DetachedRouteHandle, the Router traverses
the Route children in order to correctly allow storing a parent route when
there are several possible child Route configs that can be stored. This allows
a RouteReuseStrategy to store a parent Route and a child, while preserving
the ability to change the child route while restoring the parent. Some
implementations of RouteReuseStrategy will need to be updated to correctly
store and retrieve the DetachedRouteHandle of descendants as well as the stored
parent ActivatedRouteSnapshot. Previously, the Router would only store
the parent, making it impossible to change descendant paths when a stored parent
was retrieved. See #20114.
The router will no longer replace the browser URL when a new navigation
cancels an ongoing navigation. This often causes URL flicker and was
only in place to support some AngularJS hybrid applications. Hybrid
applications which rely on the navigationId being present on initial
navigations that were handled by the Angular router should instead
subscribe to NavigationCancel events and perform the
location.replaceState themselves to add navigationId to the Router
state.
In addition, tests which assert urlChanges on the SpyLocation may
need to be adjusted to account for the replaceState which is no longer
triggered.
It is no longer possible to use Route.loadChildren using a string
value. The following supporting classes were removed from
@angular/core:
NgModuleFactoryLoader
SystemJsNgModuleFactoryLoader
The @angular/router package no longer exports these symbols:
SpyNgModuleFactoryLoader
DeprecatedLoadChildren
The signature of the setupTestingRouter function from
@angular/core/testing has been changed to drop its NgModuleFactoryLoader
parameter, as an argument for that parameter can no longer be created.
service-worker
The return type of SwUpdate#activateUpdate and SwUpdate#checkForUpdate changed to Promise<boolean>.
Although unlikely, it is possible that this change will cause TypeScript type-checking to fail in
some cases. If necessary, update your types to account for the new
return type.
Deprecations
core
Angular no longer requires component factories to dynamically create components. The factory-based signature of the ViewContainerRef.createComponent function is deprecated in favor of a different signature that allows passing component classes instead.
The getModuleFactory function is deprecated in favor of the getNgModuleById one. With Ivy it's possible to work with NgModule classes directly, without retrieving corresponding factories, so the getNgModuleById should be used instead.
Ivy made it possible to avoid the need to resolve Component and NgModule factories. Framework APIs allow to use Component and NgModule Types directly. As a result, the PlatformRef.bootstrapModuleFactory and a factory-based signature of the ApplicationRef.bootstrap method are now obsolete and are now deprecated. The PlatformRef.bootstrapModuleFactory calls can be replaced with PlatformRef.bootstrapModule ones. The ApplicationRef.bootstrap method allows to provide Component Type, so this can be used a replacement for the factory-based calls.
In ViewEngine, JIT compilation required special providers (like Compiler, CompilerFactory, etc) to be injected in the app and corresponding methods to be invoked. With Ivy, JIT compilation takes place implicitly if the Component, NgModule, etc have not already been AOT compiled. Those special providers were made available in Ivy for backwards-compatibility with ViewEngine to make the transition to Ivy smoother. Since ViewEngine is deprecated and will soon be removed, those symbols are now deprecated as well:
ModuleWithComponentFactories
Compiler
CompilerFactory
JitCompilerFactory
NgModuleFactory
Important note: this deprecation doesn't affect JIT mode in Ivy (JIT remains available with Ivy).
In Ivy, AOT summary files are unused in TestBed. Passing AOT summary files in TestBed has no effect, so the aotSummaries usage in TestBed is deprecated and will be removed in a future version of Angular.
platform-server
The renderModuleFactory symbol in @angular/platform-server is no longer necessary as of Angular v13.
The renderModuleFactory calls can be replaced with renderModule.
service-worker
The SwUpdate#activated observable is deprecated.
The SwUpdate#activated observable only emits values as a direct response to calling
SwUpdate#activateUpdate() and was only useful for determining whether the call resulted in an
update or not. Now, the return value of SwUpdate#activateUpdate() can be used to determine the
outcome of the operation and therefore using SwUpdate#activated does not offer any benefit.
The SwUpdate#availalbe observable is deprecated.
The new SwUpdate#versionUpdates observable provides the same information and more. Therefore, it
is possible to rebuild the same behavior as SwUpdate#availalbe using the events emitted by
SwUpdate#versionUpdates and filtering for VersionReadyEvent events.
As a result, the SwUpdate#availalbe observable is now redundant.
Commit
Type
Description
747553dd68
docs
deprecate ViewEngine-based renderModuleFactory (#43757)
bazel
Commit
Type
Description
62d7005a52
feat
add strict_templates and experimental_extended_template_diagnostics to ng_module() rule (#43582)
d977701a43
feat
allow for custom conditions to be set in ng_package targets (#43764)
4886585875
feat
create transition for enabling partial compilation (#43431)
cd1b52483e
feat
expose esm2020 and es2020 conditions in APF package exports (#43740)
49b82ae561
feat
implement partial compilation APF v13 for ng_package rule (#43431)
274cb38e0b
feat
switch prodmode output to ES2020 (#43431)
73ac50c447
feat
wire up partial compilation build setting in ng_module (#43431)
e0a72857cc
fix
construct a manifest file even when warnings are emitted (#43582)
dbe656d1e0
fix
ngc-wrapped should not rely on linker for external workspaces (#43690)
common
Commit
Type
Description
adf4481211
feat
add injection token for default date pipe timezone (#43611)
c6a93001eb
fix
synchronise location mock behavior with the navigators (#41730)
compiler
Commit
Type
Description
14b492df26
fix
do not error if $any is used inside a listener (#43866)
compiler-cli
Commit
Type
Description
bed121c34f
feat
inline resources when generating class metadata calls (#43178)
263feba5c2
fix
handle nullable expressions correctly in the nullish coalescing extended template diagnostic (#43572)
8f7fdc59af
fix
not evaluating new signature for __spreadArray (#43618)
426a3ecae7
fix
updates ngc to pass the build when only warnings are emitted (#43673)
core
Commit
Type
Description
a3960846da
feat
add createNgModuleRef function to create NgModuleRef based on NgModule class (#43580)
fe1f6421d2
feat
add getNgModuleById function to retrieve loaded NgModules by id (#43580)
81c7eb813c
feat
add migration to opt out existing apps from new test module teardown behavior (#43353)
e57691c9c5
feat
Add migration to update empty routerLinks in templates (#43176)
7dccbdd27b
feat
add support for Types in ViewContainerRef.createComponent (#43022)
c14085e434
feat
drop support for TypeScript 4.2 and 4.3 (#43642)
94ba59bc9d
feat
enable test module teardown by default (#43353)
ea61ec2562
feat
support TypeScript 4.4 (#43281)
e0a0d05d45
feat
update node version support range to support v16 (#43740)
7396021e4b
fix
avoid duplicating comments in TestBed teardown migration (#43776)
7fd0428aae
fix
don't rethrow errors if test teardown has been disabled (#43635)
66fb311d20
fix
incorrect signature for initTestEnvironment (#43615)
8ae99821d6
fix
support InjectFlags argument in NodeInjector.get() (#41592)
8878183521
perf
remove support for the deprecated WrappedValue (#43507)
elements
Commit
Type
Description
a468213f34
fix
remove ng-add schematic (#43975)
f544a53f5f
fix
remove incorrect @angular/platform-browser peer dependency (#43975)
forms
Commit
Type
Description
d9d8f950e9
feat
allow disabling min/max validators dynamically (by setting the value to null) (#42978)
e49fc96ed3
feat
Make Form Statuses use stricter types. (#42952)
language-service
Commit
Type
Description
b10d90bef6
feat
Add method for retrieving the component template at the cursor location (#43208)
d5f9890c92
feat
auto-apply optional chaining on nullable symbol (#42995)
69957f72e2
feat
provide snippets for attribute (#43590)
fc3b50e427
fix
exclude the SafePropertyRead when applying the optional chaining (#43321)
migrations
Commit
Type
Description
95a68c5dc3
fix
account for CRLF characters in template migrations (#44013)
77bd2538cb
fix
apply individual expression edits to preserve newline characters (#43519)
d849350c7b
fix
Ensure routerLink migration doesn't update unrelated files (#43519)
2efc18e675
fix
migration failed finding tsconfig file (#43343)
b6f2a55147
fix
prevent migrations from updating external templates multiple times (#44013)
router
Commit
Type
Description
4f3beffdbf
feat
emit activate/deactivate events when an outlet gets attached/detached (#43333)
faf9f5a3bc
feat
new output that would notify when link is activated (#43280)
3c6b653089
feat
Option to correctly restore history on failed navigation (#43289)
784671597e
fix
Allow question marks in query param values (#31187)
796da641f0
fix
Do not modify parts of URL excluded from with 'eager' updates (#43421)
772e08d14e
fix
fix Router's public API for canceledNavigationResolution (#43842)
ccb09b4558
fix
null/undefined routerLink should disable navigation (#43087)
9e039ca68b
fix
Only trigger router navigation on popstate events from Location subscription (#43328)
c5d0bd4966
fix
Prevent URL flicker when new navigations cancel ongoing ones (#43496)
adc68b100b
fix
reuse route strategy fix (#43791)
361273fad5
refactor
remove support for loadChildren string syntax (#43591)
service-worker
Commit
Type
Description
59225f5586
feat
SwUpdate#activeUpdate and SwUpdate#checkForUpdate should have a meaningful outcome (#43668)
0dc45446fe
feat
expose more version update events (#43668)
Special Thanks
Ahmed Ayed, Alan Agius, Alex Rickabaugh, Andrew Kushnir, Andrew Scott, Bjarki, Charles Lyding, Dmitrij Kuba, Doug Parker, Dylan Hunn, George Kalpakas, Jessica Janiuk, Jochen Kraushaar, Joe Martin (Crowdstaffing), Joey Perrott, Jon Rimmer, JoostK, Kristiyan Kostadinov, Maximilian Köller, Paul Gschwendtner, Pei Wang, Pete Bacon Darwin, Tomasz Domański, Willy Schott, anandtiwary, dario-piotrowicz, iRealNirmal, ivanwonder, krzysztof-grzybek, mgechev and vthinkxie
v12.2.16
Compare Source
12.2.16 (2022-01-27)
ngcc
Commit
Type
Description
460befd693
fix
support element accesses for export declarations (#44824)
Special Thanks
Andrew Kushnir and JoostK
v12.2.15
Compare Source
ngcc
Commit
Type
Description
b6554d75cd
fix
correctly resolve UMD dependencies (#44382)
Special Thanks
George Kalpakas
v12.2.14
Compare Source
compiler
Commit
Type
Description
e3db0385b6
fix
ensure that partially compiled queries can handle forward references (#44124)
ngcc
Commit
Type
Description
a8be244113
fix
correctly report error when collecting dependencies of UMD module (#44245)
fc072935ee
fix
support the UMD wrapper function format emitted by Webpack (#44245)
Special Thanks
George Kalpakas, Pete Bacon Darwin and iRealNirmal
v12.2.13
Compare Source
compiler-cli
Commit
Type
Description
9a89db790f
fix
avoid broken references in .d.ts files due to @internal markers (#43965)
core
Commit
Type
Description
8f402c9d06
fix
support InjectFlags argument in NodeInjector.get() (#41592)
Special Thanks
Alan Agius, George Kalpakas, Jochen Kraushaar, Joe Martin (Crowdstaffing), JoostK and vthinkxie
v12.2.12
Compare Source
compiler-cli
Commit
Type
Description
112557497c
fix
avoid broken references in .d.ts files due to @internal markers (#43527)
ngcc
Commit
Type
Description
067ae54d46
fix
support alternate UMD layout when adding new imports (#43931)
Special Thanks
Alan Agius, Andrew Kushnir, George Kalpakas, Jessica Janiuk, Joey Perrott, JoostK, Mladen Jakovljević, Virginia Dooley, amayer42, dirk diebel and ericcheng2005
v12.2.11
Compare Source
ngcc
Commit
Type
Description
cab21cea7a
fix
support alternate wrapper function layout for UMD (#43879)
router
Commit
Type
Description
58c11865ac
fix
Do not clear currentNavigation if already set to next one (#43852)
Special Thanks
Alan Agius, Andrew Kushnir, Andrew Scott, David Shevitz, George Kalpakas, Joe Martin (Crowdstaffing), Natalia Venditto, Pete Bacon Darwin, Younes Jaaidi and dario-piotrowicz
v12.2.10
Compare Source
Special Thanks
Alan Agius, Daniel Díaz, David Shevitz, Doug Parker, George Kalpakas, Joe Martin (Crowdstaffing), Tanguy Nodet, Thomas Turrell-Croft, dario-piotrowicz, hchiam, markostanimirovic and mgechev
v12.2.9
Compare Source
core
Commit
Type
Description
b4b441077a
fix
handle invalid constructor parameters in partial factory declarations (#43619)
router
Commit
Type
Description
7f6050587d
fix
unset attachRef when router-outlet is destroyed to avoid mounting a destroyed component (#43697)
service-worker
Commit
Type
Description
c4ecc07838
fix
make ngsw.json generation deterministic and correct (#43679)
Special Thanks
Alan Agius, Daniel Díaz, George Kalpakas, JoostK, Kristiyan Kostadinov, Pete Bacon Darwin, Wey-Han Liaw, dario-piotrowicz, iRealNirmal, little-pinecone, mgechev, ultrasonicsoft and xiaohanxu-nick
v12.2.8
Compare Source
compiler-cli
Commit
Type
Description
c1338bf837
fix
correctly interpret token arrays in @Injectable deps (#43226)
language-service
Commit
Type
Description
c8f8d7d3b1
fix
provide dom event completions (#43299)
ngcc
Commit
Type
Description
69299f7d4d
fix
do not fail for packages which correspond with Object members (#43589)
service-worker
Commit
Type
Description
3cf41354ae
fix
do not unassign clients from a broken version (#43518)
Special Thanks
Adrien Crivelli, Alex Rickabaugh, Andrew Scott, Bobby Galli, Chris, Daniel Díaz, Dario Piotrowicz, George Kalpakas, Joe Martin (Crowdstaffing), JoostK, Pete Bacon Darwin, Rafael Santana, Raj Sekhar, Ricardo Chavarria, Teri Glover, Virginia Dooley, dario-piotrowicz, enisfr and wszgrcy
v12.2.7
Compare Source
common
Commit
Type
Description
2bb4bf1468
fix
titlecase pipe incorrectly handling numbers (#43476)
compiler
Commit
Type
Description
9c8a1f8a71
fix
include leading whitespace in source-spans of i18n messages (#43132)
compiler-cli
Commit
Type
Description
defb02f11e
fix
handle directives that refer to a namespaced class in a type parameter bound (#43511)
platform-browser
Commit
Type
Description
adc7c56ede
fix
improve error message for missing animation trigger (#41356)
Special Thanks
Andrew Scott, Daniel Díaz, George Kalpakas, JoostK, Kristiyan Kostadinov, Mwiku, Pete Bacon Darwin, Teri Glover, Virginia Dooley, Xiaohanxu1996, dario-piotrowicz and kirjs
v12.2.6
Compare Source
animations
Commit
Type
Description
141fde1632
fix
emit pure annotations to static property initializers (#43344)
core
Commit
Type
Description
ca510c87c5
fix
emit pure annotations to static property initializers (#43344)
router
Commit
Type
Description
[4034f252c9](https://togithub.com/angular/angular/commit/4034f252c9707dabd0
Configuration
📅 Schedule: At any time (no schedule defined).
🚦 Automerge: Disabled by config. Please merge this manually once you are satisfied.
♻ Rebasing: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox.
🔕 Ignore: Close this PR and you won't be reminded about this update again.
[ ] If you want to rebase/retry this PR, click this checkbox.
This PR has been generated by Renovate Bot.
⚠ Artifact update problem
Renovate failed to update an artifact related to this branch. You probably do not want to merge this PR as-is.
♻ Renovate will retry this branch, including artifacts, only when one of the following happens:
any of the package files in this branch needs updating, or
the branch becomes conflicted, or
you click the rebase/retry checkbox if found above, or
you rename this PR's title to start with "rebase!" to trigger it manually
The artifact failure details are included below:
File name: bprr-ui/package-lock.json
npm WARN old lockfile
npm WARN old lockfile The package-lock.json file was created with an old version of npm,
npm WARN old lockfile so supplemental metadata must be fetched from the registry.
npm WARN old lockfile
npm WARN old lockfile This is a one-time fix-up, please be patient...
npm WARN old lockfile
npm ERR! code ERESOLVE
npm ERR! ERESOLVE unable to resolve dependency tree
npm ERR!
npm ERR! While resolving: bprr-ui@0.0.0
npm ERR! Found: @angular/common@10.1.6
npm ERR! node_modules/@angular/common
npm ERR! @angular/common@"~10.1.3" from the root project
npm ERR!
npm ERR! Could not resolve dependency:
npm ERR! peer @angular/common@"13.2.1" from @angular/platform-browser@13.2.1
npm ERR! node_modules/@angular/platform-browser
npm ERR! @angular/platform-browser@"~13.2.0" from the root project
npm ERR!
npm ERR! Fix the upstream dependency conflict, or retry
npm ERR! this command with --force, or --legacy-peer-deps
npm ERR! to accept an incorrect (and potentially broken) dependency resolution.
npm ERR!
npm ERR! See /tmp/renovate/cache/others/npm/eresolve-report.txt for a full report.
npm ERR! A complete log of this run can be found in:
npm ERR! /tmp/renovate/cache/others/npm/_logs/2022-02-08T07_34_00_941Z-debug.log
|
gharchive/pull-request
| 2022-02-08T07:34:10 |
2025-04-01T04:35:13.247335
|
{
"authors": [
"ngeor"
],
"repo": "ngeor/bitbucket-pr-report",
"url": "https://github.com/ngeor/bitbucket-pr-report/pull/47",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
636180189
|
Add option to store UUID
in the agent.conf, you need to set store_UUID = True for persistence to work on a container or demo environment (where the underlying instance might be swapped but you want it to report the same on the controller). We should add a variable to set this to true. This solves that issue.
Current my workaround is this
sudo sed -i.bak 's/store_uuid = False/store_uuid = True/' /etc/controller-agent/agent.conf
just added this as default behavior whenever nginx_controller_hostname or nginx_controller_instance_name is defined.
With the expectation that if those are not defined and thus get written to the agent.conf via the installer, the hostname might change.
Instance_name will always remain fixed and is set at installation.
However, nginx_controller_hostname could safely change.
|
gharchive/issue
| 2020-06-10T11:46:57 |
2025-04-01T04:35:13.261095
|
{
"authors": [
"brianehlert",
"magicalyak"
],
"repo": "nginxinc/ansible-role-nginx-controller-agent",
"url": "https://github.com/nginxinc/ansible-role-nginx-controller-agent/issues/6",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1145032124
|
feat(cli): add ability to create a store inside a class
PR Checklist
Please check if your PR fulfills the following requirements:
[x] The commit message follows our guidelines: https://github.com/ngneat/elf/blob/master/CONTRIBUTING.md#commit
[x] Tests for the changes have been added (for bug fixes / features)
[x] Docs have been added / updated (for bug fixes / features)
PR Type
What kind of change does this PR introduce?
[ ] Bugfix
[x] Feature
[ ] Code style update (formatting, local variables)
[ ] Refactoring (no functional changes, no api changes)
[ ] Build related changes
[ ] CI related changes
[x] Documentation content changes
[ ] Other... Please describe:
What is the current behavior?
Issue Number: #162
What is the new behavior?
Users can choose whether a store should be created inside a class or not.
Does this PR introduce a breaking change?
[ ] Yes
[x] No
Other information
@EricPoul what do you say about https://github.com/ngneat/elf/discussions/165
It can be useful for repoTemplate: "function" or for creating a store outside of a class. Inside the class, we create this function anyway. It still takes a lot of space so I'd move it to the private method.
Anyway, I don't see this as a bother, especially when it generates not by me. One main con of creating a state separately from the store is the identical names of state and arguments of reducer functions in the class.
const { `state`, config } = createState();
const store = new Store({ name: 'todos', state, config });
update(): void {
store.update((`state`) => { ...`state`, some: 'new' })
}
|
gharchive/pull-request
| 2022-02-20T14:16:12 |
2025-04-01T04:35:13.285144
|
{
"authors": [
"EricPoul",
"NetanelBasal"
],
"repo": "ngneat/elf",
"url": "https://github.com/ngneat/elf/pull/167",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2760339144
|
Lỗi liên quan đến việc chuyển một trường nhập liệu giữa hai trạng thái uncontrolled và controlled trong React
Coi thử xem nó là gì và fix được không
@ngodat0103 ok
da fix xem pull resquest
Có lỗi thì fix đi đã, ko được hãy tạo issua @nguyenthiyenly0407
|
gharchive/issue
| 2024-12-27T03:22:49 |
2025-04-01T04:35:13.289956
|
{
"authors": [
"VietDucc",
"ngodat0103",
"nguyenthiyenly0407"
],
"repo": "ngodat0103/se347",
"url": "https://github.com/ngodat0103/se347/issues/102",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
802434859
|
what's the callback url for OIDC auth?
Hi.
I'm wanting to set up powerdns admin to work with my orgs oidc provider but i need a callback url and i can't find it anywhere. Could some tell me what it should be?
Thanks.
It should be address of powerdnsadmin plus /oidc/authorized
E.g. if you host pdnsadmin at https://dns.domain.com callback url should be https://dns.domain.com/oidc/authorized
|
gharchive/issue
| 2021-02-05T19:56:28 |
2025-04-01T04:35:13.291568
|
{
"authors": [
"Prototik",
"nktech1135"
],
"repo": "ngoduykhanh/PowerDNS-Admin",
"url": "https://github.com/ngoduykhanh/PowerDNS-Admin/issues/881",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
285023211
|
Chuyển tiền và tiền thối
Em chưa hiểu về giao dịch và tiền thối lắm, cho em hỏi vd A có 10KCoin, muốn chuyển B 3KCoin thì phải tạo 1 transaction trong đó output gồm 2 phần tử, một cái chứa địa chỉ của B 3KCoin, một cái chứa địa chỉ A 7KCoin phải ko ạ? Và referencedOutputHash là hash của giao dịch gần nhất mà A đã chuyển?
A có 10K có nghĩa là A có tổng giá trị transaction ouput chưa sử dụng là 10K
Khi A muốn chuyển B 3K thì A phải tìm một số transaction ouput để làm input (dĩ nhiên là chưa dùng) với tổng >= 3K nghĩa là nhiều transaction output (hash + index) chứ không nhất thiết phải là một.
Sau đó để output ở transaction bạn tạo:
B: 3K
A: 7K
A ở đây có thể là địa chỉ nào đó của A
@nguyenkha cho em hỏi về trường hợp này:
Ban đầu addrA và addrB, mỗi địa chỉ được hệ thống KCoin của thầy chuyển vào 10K ở transaction có hash là hash1
Bây giờ muốn chuyển 3K từ addrA vào addrB thì input là 2 phần tử A,B hay chỉ 1 phần tử A, output là [A: 7, B:3] hay [A:7, B:13] ?
Input chỉ bao gồm A, Output gồm B (3K) và địa chỉ nhận tiền thối (7K) - không nhất thiết phải là A, mà là của người sở hữu địa chỉ A/địa chỉ của hệ thống của bạn => tùy cách bạn cài đặt hệ thống.
|
gharchive/issue
| 2017-12-29T02:29:22 |
2025-04-01T04:35:13.304266
|
{
"authors": [
"nguyenkha",
"vukhoa30"
],
"repo": "nguyenkha/kcoin-blockchain",
"url": "https://github.com/nguyenkha/kcoin-blockchain/issues/14",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
521431435
|
docs(readme): add section on hot reloading of translations
I added a section in the FAQ to explain how to apply hot reloading of translations in an application. Thanks to the explanation of @keradus in #874
Until then...
|
gharchive/pull-request
| 2019-11-12T09:32:30 |
2025-04-01T04:35:13.315376
|
{
"authors": [
"Odonno"
],
"repo": "ngx-translate/core",
"url": "https://github.com/ngx-translate/core/pull/1139",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
638406393
|
Add truncate and ftruncate
This change is
here we go
|
gharchive/pull-request
| 2020-06-14T18:37:51 |
2025-04-01T04:35:13.316597
|
{
"authors": [
"blackheaven"
],
"repo": "nh2/hatrace",
"url": "https://github.com/nh2/hatrace/pull/81",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
463356969
|
isAllDay always appears in the popup creation form
Version
1.12.1
Development Environment
Windows 10.0 build 17134
XAMPP for Windows 5.6.34(Apache 2.4.33 + PHP 7.2.5 + MySQL 5.0.12)
No any NodeJS environment.
Current Behavior
I found it when I trying to modify the example files to fit my requirements. If the "isAllDay" option set to false, the checkbox of "All Day" inside popup creation form seemed not to hide at all. So when someone checked it, everything would stop.
Expected Behavior
Since the popup creation form layout seemed to be declared inside "calendar.js", so it should be a bug when the "isAllDay" of the calendar was set to false but not removing the checkbox.
@LZong-tw
The isAllDay option is not an option to control the UI within the creation popup.
It determines whether the schedule is an 'All day' check or not.
We do not yet support the features you want.
If you need it right now, modify the following files and use them.
src/js/view/popup/scheduleCreationPopup.js
src/js/view/template/popup/scheduleCreationPopup.hbs
|
gharchive/issue
| 2019-07-02T17:46:13 |
2025-04-01T04:35:13.319910
|
{
"authors": [
"LZong-tw",
"jungeun-cho"
],
"repo": "nhn/tui.calendar",
"url": "https://github.com/nhn/tui.calendar/issues/342",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
307894501
|
feat: New Design
Version
v0.8.0
Development Environment
Any
Current Behavior
Expected Behavior
Apply new design.
Finished
|
gharchive/issue
| 2018-03-23T03:56:16 |
2025-04-01T04:35:13.321897
|
{
"authors": [
"dongsik-yoo"
],
"repo": "nhnent/tui.calendar",
"url": "https://github.com/nhnent/tui.calendar/issues/31",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1717366333
|
examples: Fix nidaqmx mypy errors (except aio-related errors)
What does this Pull Request accomplish?
Fix non-aio-related mypy errors for nidaqmx examples.
The currently released version of mypy-protobuf (3.4.0) doesn't support aio yet, but it looks like support will be available in a future release.
Why should this Pull Request be merged?
Demonstrate type-checking best practices (I hope).
Prerequisite for updating the CI workflow to validate nidaqmx examples in https://github.com/ni/grpc-device/pull/931
What testing has been done?
Ran validate_examples.py -p nidaqmx before and after.
Before:
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:116: error: "None" has no attribute "GetErrorString" [attr-defined]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:126: error: Module has no attribute "aio" [attr-defined]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:131: error: Incompatible types in "await" (actual type "Union[GetTaskAttributeUInt32Response, Any]", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:131: error: Item "None" of "Optional[NiDAQmxStub]" has no attribute "GetTaskAttributeUInt32" [union-attr]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:140: error: Incompatible types in "await" (actual type "Union[GetNthTaskDeviceResponse, Any]", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:140: error: Item "None" of "Optional[NiDAQmxStub]" has no attribute "GetNthTaskDevice" [union-attr]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:145: error: Incompatible types in "await" (actual type "Union[GetDeviceAttributeInt32Response, Any]", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:145: error: Item "None" of "Optional[NiDAQmxStub]" has no attribute "GetDeviceAttributeInt32" [union-attr]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:170: error: Incompatible types in "await" (actual type "CreateTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:174: error: Incompatible types in "await" (actual type "CreateAIVoltageChanResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:184: error: Incompatible types in "await" (actual type "CfgSampClkTimingResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:194: error: Incompatible types in "await" (actual type "CreateTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:198: error: Incompatible types in "await" (actual type "CreateAOVoltageChanResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:207: error: Incompatible types in "await" (actual type "CfgSampClkTimingResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:217: error: Incompatible types in "await" (actual type "CreateTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:221: error: Incompatible types in "await" (actual type "CreateAIVoltageChanResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:231: error: Incompatible types in "await" (actual type "CfgSampClkTimingResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:241: error: Incompatible types in "await" (actual type "CreateTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:245: error: Incompatible types in "await" (actual type "CreateAOVoltageChanResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:254: error: Incompatible types in "await" (actual type "CfgSampClkTimingResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:269: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:276: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:283: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:290: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:303: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:310: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:317: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:328: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:335: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:342: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:354: error: Incompatible types in "await" (actual type "CfgDigEdgeStartTrigResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:361: error: Incompatible types in "await" (actual type "CfgDigEdgeStartTrigResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:368: error: Incompatible types in "await" (actual type "CfgDigEdgeStartTrigResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:380: error: Incompatible types in "await" (actual type "WriteAnalogF64Response", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:392: error: Incompatible types in "await" (actual type "WriteAnalogF64Response", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:415: error: Incompatible types in "await" (actual type "Tuple[Tuple[str, Union[str, bytes]], ...]", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:439: error: Incompatible types in "await" (actual type "Tuple[Tuple[str, Union[str, bytes]], ...]", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:442: error: Incompatible types in "await" (actual type "StartTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:446: error: Incompatible types in "await" (actual type "StartTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:450: error: Incompatible types in "await" (actual type "StartTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:455: error: Incompatible types in "await" (actual type "StartTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:462: error: "CallIterator[RegisterEveryNSamplesEventResponse]" has no attribute "__aiter__" (not async iterable) [attr-defined]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:464: error: Incompatible types in "await" (actual type "Union[ReadAnalogF64Response, Any]", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:464: error: Item "None" of "Optional[NiDAQmxStub]" has no attribute "ReadAnalogF64" [union-attr]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:477: error: Incompatible types in "await" (actual type "Union[ReadAnalogF64Response, Any]", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:477: error: Item "None" of "Optional[NiDAQmxStub]" has no attribute "ReadAnalogF64" [union-attr]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:511: error: Incompatible types in "await" (actual type "StopTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:515: error: Incompatible types in "await" (actual type "StopTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:519: error: Incompatible types in "await" (actual type "StopTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:523: error: Incompatible types in "await" (actual type "StopTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:531: error: Cannot determine type of "key" [has-type]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:532: error: Cannot determine type of "value" [has-type]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:550: error: Incompatible types in "await" (actual type "ClearTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:553: error: Incompatible types in "await" (actual type "ClearTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:556: error: Incompatible types in "await" (actual type "ClearTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:559: error: Incompatible types in "await" (actual type "ClearTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\digital-output.py:75: error: Incompatible types in assignment (expression has type "WriteDigitalU32Response", variable has type "CreateTaskResponse") [assignment]
examples\nidaqmx\digital-output.py:93: error: Cannot determine type of "key" [has-type]
examples\nidaqmx\digital-output.py:94: error: Cannot determine type of "value" [has-type]
examples\nidaqmx\digital-input.py:75: error: Incompatible types in assignment (expression has type "ReadDigitalU32Response", variable has type "CreateTaskResponse") [assignment]
examples\nidaqmx\digital-input.py:88: error: "CreateTaskResponse" has no attribute "read_array" [attr-defined]
examples\nidaqmx\digital-input.py:92: error: Cannot determine type of "key" [has-type]
examples\nidaqmx\digital-input.py:93: error: Cannot determine type of "value" [has-type]
examples\nidaqmx\counter-output.py:92: error: Cannot determine type of "key" [has-type]
examples\nidaqmx\counter-output.py:93: error: Cannot determine type of "value" [has-type]
examples\nidaqmx\counter-input.py:83: error: Incompatible types in assignment (expression has type "ReadCounterScalarF64Response", variable has type "CreateTaskResponse") [assignment]
examples\nidaqmx\counter-input.py:90: error: "CreateTaskResponse" has no attribute "value" [attr-defined]
examples\nidaqmx\counter-input.py:94: error: Cannot determine type of "key" [has-type]
examples\nidaqmx\counter-input.py:95: error: Cannot determine type of "value" [has-type]
examples\nidaqmx\analog-output.py:96: error: Cannot determine type of "key" [has-type]
examples\nidaqmx\analog-output.py:97: error: Cannot determine type of "value" [has-type]
examples\nidaqmx\analog-input-every-n-samples-aio.py:54: error: Module has no attribute "aio" [attr-defined]
examples\nidaqmx\analog-input-every-n-samples-aio.py:61: error: Item "None" of "Optional[NiDAQmxStub]" has no attribute "GetErrorString" [union-attr]
examples\nidaqmx\analog-input-every-n-samples-aio.py:68: error: Incompatible types in "await" (actual type "CreateTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\analog-input-every-n-samples-aio.py:73: error: Incompatible types in "await" (actual type "CreateAIVoltageChanResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\analog-input-every-n-samples-aio.py:86: error: Incompatible types in "await" (actual type "CfgSampClkTimingResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\analog-input-every-n-samples-aio.py:106: error: Incompatible types in "await" (actual type "Tuple[Tuple[str, Union[str, bytes]], ...]", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\analog-input-every-n-samples-aio.py:112: error: Incompatible types in "await" (actual type "Tuple[Tuple[str, Union[str, bytes]], ...]", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\analog-input-every-n-samples-aio.py:114: error: Incompatible types in "await" (actual type "StartTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\analog-input-every-n-samples-aio.py:117: error: Incompatible types in "await" (actual type "GetTaskAttributeUInt32Response", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\analog-input-every-n-samples-aio.py:129: error: "CallIterator[RegisterEveryNSamplesEventResponse]" has no attribute "__aiter__" (not async iterable) [attr-defined]
examples\nidaqmx\analog-input-every-n-samples-aio.py:131: error: Incompatible types in "await" (actual type "Union[ReadAnalogF64Response, Any]", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\analog-input-every-n-samples-aio.py:131: error: Item "None" of "Optional[NiDAQmxStub]" has no attribute "ReadAnalogF64" [union-attr]
examples\nidaqmx\analog-input-every-n-samples-aio.py:158: error: "CallIterator[RegisterDoneEventResponse]" has no attribute "__aiter__" (not async iterable) [attr-defined]
examples\nidaqmx\analog-input-every-n-samples-aio.py:168: error: Incompatible types in "await" (actual type "StopTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\analog-input-every-n-samples-aio.py:174: error: Cannot determine type of "key" [has-type]
examples\nidaqmx\analog-input-every-n-samples-aio.py:175: error: Cannot determine type of "value" [has-type]
examples\nidaqmx\analog-input-every-n-samples-aio.py:186: error: Incompatible types in "await" (actual type "ClearTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\analog-input-every-n-samples.py:63: error: Item "None" of "Optional[NiDAQmxStub]" has no attribute "GetErrorString" [union-attr]
examples\nidaqmx\analog-input-every-n-samples.py:132: error: Item "None" of "Optional[NiDAQmxStub]" has no attribute "ReadAnalogF64" [union-attr]
examples\nidaqmx\analog-input-every-n-samples.py:183: error: Cannot determine type of "key" [has-type]
examples\nidaqmx\analog-input-every-n-samples.py:184: error: Cannot determine type of "value" [has-type]
examples\nidaqmx\analog-input.py:93: error: Incompatible types in assignment (expression has type "GetTaskAttributeUInt32Response", variable has type "CreateTaskResponse") [assignment]
examples\nidaqmx\analog-input.py:98: error: "CreateTaskResponse" has no attribute "value" [attr-defined]
examples\nidaqmx\analog-input.py:100: error: Incompatible types in assignment (expression has type "ReadAnalogF64Response", variable has type "CreateTaskResponse") [assignment]
examples\nidaqmx\analog-input.py:113: error: "CreateTaskResponse" has no attribute "read_array" [attr-defined]
examples\nidaqmx\analog-input.py:114: error: "CreateTaskResponse" has no attribute "samps_per_chan_read" [attr-defined]
examples\nidaqmx\analog-input.py:119: error: Cannot determine type of "key" [has-type]
examples\nidaqmx\analog-input.py:120: error: Cannot determine type of "value" [has-type]
Found 99 errors in 9 files (checked 9 source files)
After:
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:127: error: Module has no attribute "aio" [attr-defined]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:132: error: Incompatible types in "await" (actual type "GetTaskAttributeUInt32Response", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:141: error: Incompatible types in "await" (actual type "GetNthTaskDeviceResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:146: error: Incompatible types in "await" (actual type "GetDeviceAttributeInt32Response", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:170: error: Incompatible types in "await" (actual type "CreateTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:174: error: Incompatible types in "await" (actual type "CreateAIVoltageChanResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:184: error: Incompatible types in "await" (actual type "CfgSampClkTimingResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:194: error: Incompatible types in "await" (actual type "CreateTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:198: error: Incompatible types in "await" (actual type "CreateAOVoltageChanResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:207: error: Incompatible types in "await" (actual type "CfgSampClkTimingResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:217: error: Incompatible types in "await" (actual type "CreateTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:221: error: Incompatible types in "await" (actual type "CreateAIVoltageChanResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:231: error: Incompatible types in "await" (actual type "CfgSampClkTimingResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:241: error: Incompatible types in "await" (actual type "CreateTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:245: error: Incompatible types in "await" (actual type "CreateAOVoltageChanResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:254: error: Incompatible types in "await" (actual type "CfgSampClkTimingResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:269: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:276: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:283: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:290: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:303: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:310: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:317: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:328: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:335: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:342: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:354: error: Incompatible types in "await" (actual type "CfgDigEdgeStartTrigResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:361: error: Incompatible types in "await" (actual type "CfgDigEdgeStartTrigResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:368: error: Incompatible types in "await" (actual type "CfgDigEdgeStartTrigResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:380: error: Incompatible types in "await" (actual type "WriteAnalogF64Response", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:392: error: Incompatible types in "await" (actual type "WriteAnalogF64Response", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:415: error: Incompatible types in "await" (actual type "Tuple[Tuple[str, Union[str, bytes]], ...]", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:439: error: Incompatible types in "await" (actual type "Tuple[Tuple[str, Union[str, bytes]], ...]", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:442: error: Incompatible types in "await" (actual type "StartTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:446: error: Incompatible types in "await" (actual type "StartTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:450: error: Incompatible types in "await" (actual type "StartTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:455: error: Incompatible types in "await" (actual type "StartTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:463: error: "CallIterator[RegisterEveryNSamplesEventResponse]" has no attribute "__aiter__" (not async iterable) [attr-defined]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:464: error: Incompatible types in "await" (actual type "ReadAnalogF64Response", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:475: error: Incompatible types in "await" (actual type "ReadAnalogF64Response", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:508: error: Incompatible types in "await" (actual type "StopTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:512: error: Incompatible types in "await" (actual type "StopTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:516: error: Incompatible types in "await" (actual type "StopTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:520: error: Incompatible types in "await" (actual type "StopTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:527: error: "_Metadatum" object is not iterable [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:528: error: Cannot determine type of "key" [has-type]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:529: error: Cannot determine type of "value" [has-type]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:547: error: Incompatible types in "await" (actual type "ClearTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:550: error: Incompatible types in "await" (actual type "ClearTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:553: error: Incompatible types in "await" (actual type "ClearTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:556: error: Incompatible types in "await" (actual type "ClearTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\analog-input-every-n-samples-aio.py:51: error: Module has no attribute "aio" [attr-defined]
examples\nidaqmx\analog-input-every-n-samples-aio.py:66: error: Incompatible types in "await" (actual type "CreateTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\analog-input-every-n-samples-aio.py:69: error: Incompatible types in "await" (actual type "CreateAIVoltageChanResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\analog-input-every-n-samples-aio.py:82: error: Incompatible types in "await" (actual type "CfgSampClkTimingResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\analog-input-every-n-samples-aio.py:102: error: Incompatible types in "await" (actual type "Tuple[Tuple[str, Union[str, bytes]], ...]", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\analog-input-every-n-samples-aio.py:108: error: Incompatible types in "await" (actual type "Tuple[Tuple[str, Union[str, bytes]], ...]", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\analog-input-every-n-samples-aio.py:110: error: Incompatible types in "await" (actual type "StartTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\analog-input-every-n-samples-aio.py:113: error: Incompatible types in "await" (actual type "GetTaskAttributeUInt32Response", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\analog-input-every-n-samples-aio.py:125: error: "CallIterator[RegisterEveryNSamplesEventResponse]" has no attribute "__aiter__" (not async iterable) [attr-defined]
examples\nidaqmx\analog-input-every-n-samples-aio.py:126: error: Incompatible types in "await" (actual type "ReadAnalogF64Response", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\analog-input-every-n-samples-aio.py:152: error: "CallIterator[RegisterDoneEventResponse]" has no attribute "__aiter__" (not async iterable) [attr-defined]
examples\nidaqmx\analog-input-every-n-samples-aio.py:162: error: Incompatible types in "await" (actual type "StopTaskResponse", expected type "Awaitable[Any]") [misc]
examples\nidaqmx\analog-input-every-n-samples-aio.py:167: error: "_Metadatum" object is not iterable [misc]
examples\nidaqmx\analog-input-every-n-samples-aio.py:168: error: Cannot determine type of "key" [has-type]
examples\nidaqmx\analog-input-every-n-samples-aio.py:169: error: Cannot determine type of "value" [has-type]
examples\nidaqmx\analog-input-every-n-samples-aio.py:180: error: Incompatible types in "await" (actual type "ClearTaskResponse", expected type "Awaitable[Any]") [misc]
Found 67 errors in 2 files (checked 9 source files)
|
gharchive/pull-request
| 2023-05-19T14:40:12 |
2025-04-01T04:35:13.352332
|
{
"authors": [
"bkeryan"
],
"repo": "ni/grpc-device",
"url": "https://github.com/ni/grpc-device/pull/930",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1058006854
|
"Scripting Tool" source to be relinked to "grpc-labview-support" source for dev purpose
Currently, "grpc-labview-support" source is built and deployed under the vi.lib as "grpc-labview-support-release" , and the "scripting tool" is dependent on the released version of "grpc-labview-support-release". So, when the "scripting tool" is deployed into vi.lib as "scripting tool release", it will maintain the dependency from the vi.lib for "grpc-labview-support-release".
This might make the release process simple. But the development process needs clear documentation so the developers know to treat the source as two different projects and make sure the "grpc-labview-support" changes are installed before making changes to the "scripting tool" source.
During the last discussion with @ccifra , the decision was to make the "scripting tool" source dependent on the "grpc-labview-support" source, so that the development process is easier, and cifra will take careful action during the release in his end.
@ccifra I assume there are a few folks who are making changes to the repo and this change might be another merge issue headache. Since, people are working through this, we could make this change before a release or when there is downtime in development? let me know when to plan this.
|
gharchive/issue
| 2021-11-19T00:42:30 |
2025-04-01T04:35:13.355628
|
{
"authors": [
"navinsubramani"
],
"repo": "ni/grpc-labview",
"url": "https://github.com/ni/grpc-labview/issues/40",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2730180481
|
Site tweaks to improve strict CSP hosting
Pull Request
🤨 Rationale
The reason for this PR was trying to evaluate Nimble-based apps hosting in Web App hosting under our strict CSP policy. This was done by exercising the nimble site build which covers the following: Angular, Blazor, Vite, Storybook.
In doing so identified some changes that are either useful or minimally invasive as workarounds for WebApp hosting issues or improving the experience of Nimble site if hosted in Web App hosting.
👩💻 Implementation
Vite:
A minimal change to provide a relative base configuration for vite was needed. This causes vite to change from paths relative to root (/script.js) to relative paths in the current directory (./script.js). Reason is WebApps are not hosted in root but from a subdirectory path.
Angular:
A minimal change to disable inlineCritical styles is needed so that stylesheets included in angular.json load correctly (which seems to be something we figured out and inherited in all apps but doesn't seem to be documented anywhere 🤷 it's like a herd immunity inherited in copy-paste generations 💉). Reason is described in blog / we don't allow unsafe-inline.
Blazor:
During investigation found that Blazor does not like to serve from index.html urls and relies on path urls. You can actually see it on the current published site if you use a blazor index.html style url instead of a directory url. The page fails to load and has many console errors as Blazor does poor URL parsing / manipulation to load JS resources. It's probably worth creating an issue but I did not yet.
I thought of a workaround for nimble site by specifying a base url manually of <base href="./" />. This is not compatible with our strict CSP setting and is ignored due to base-uri: 'none';'' (the OWASP strict policy example) causing the errors to continue. We potentially could switch our CSP to base-uri: 'self' but it's not clear to me what the security implications are and I'm not recommending it yet (see following).
I also found that Blazor seems generally against the idea of serving off index.html vs the directory path as index.html resolution is not supported out of the box in the Router either.
I thought of a Router workaround for index.html resolution by serving the same component from both the path url and the index.html url. It seems to work but I'm not aware of any other concerns relying on that workaround.
Even with the above a Blazor 8 app using Nimble will still not run as some Blazor 8 template binding features require using eval and are only addressed in Blazor 9.
Based on the above, while I'll propose the changes for Nimble's Blazor site page to support index.html based urls working, I'm not going to propose changes to Web App hosting strict CSP base-uri configuration to support the workaround described above. Instead, I think we should say Nimble Blazor WebApp hosting is contingent on Blazor 9 support in Nimble and fixing path serving in Web App hosting. I don't think we should recommend the workarounds I figured out above to Blazor WebApp devs as from the linked issues Blazor does not seem interested in supporting that pattern.
Storybook:
Depends on unsafe-inline (which we do not allow) and they seem resistant to the change. Need to reply on the issue with a convincing discussion, not sure if MDN and OWASP recommendations are sufficient for them. Did not provide a comment / create an issue.
All:
Updated links to point to index.html paths as it's minimally invasive (makes the URL bar uglier) but works around AzDo 2941644. This could be reverted in Nimble once the WebApp directory hosting issue is address.
Updated each page to have a link to the parent page to make them easier to navigate between when hosted in the WebApp hosting iframe.
🧪 Testing
Manual and via built storybook.
Also example hosted on dev:
Landing (vite)
Wafer Performance (vite)
Angular
Blazor broken, see above
Storybook expected to be broken, see above (can't explicitly test as deleted from dev package due to size upload bug)
✅ Checklist
[x] I have updated the project documentation to reflect my changes or determined no changes are needed.
@jattasNI @msmithNI @atmgrifter00 these are minor non-source impacting changes just to the site so bypassing but can address concerns in a follow-up
|
gharchive/pull-request
| 2024-12-10T13:49:46 |
2025-04-01T04:35:13.370020
|
{
"authors": [
"rajsite"
],
"repo": "ni/nimble",
"url": "https://github.com/ni/nimble/pull/2492",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1971571849
|
drag and drop functionality in tags to form a sentence
Please add drag and drop functionality in tags to form a sentence and change the words.
words as tags and dragged and drop to form a sentence.
Thanks
Sai
I am not very clear about your needs, could you please provide more detailed instructions, such as GIFs, web pages and so on.
|
gharchive/issue
| 2023-11-01T02:16:52 |
2025-04-01T04:35:13.379112
|
{
"authors": [
"nicedouble",
"saisaigraph"
],
"repo": "nicedouble/StreamlitAntdComponents",
"url": "https://github.com/nicedouble/StreamlitAntdComponents/issues/27",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2442086752
|
ytmapi_rs: Remove raw_query and processed_query functions from public API.
Replace with more ergonomic alternatives without lifetime requirements. e.g deserialize_query that will return T = DeserializeOwned and string_query that will return pretty printed.
Since RawResult and ProcessedResult need to be a part of the public API for AuthToken and ParseFrom traits, may not be worth doing this.
Closing due to above comment
|
gharchive/issue
| 2024-08-01T10:26:22 |
2025-04-01T04:35:13.413998
|
{
"authors": [
"nick42d"
],
"repo": "nick42d/youtui",
"url": "https://github.com/nick42d/youtui/issues/112",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
203915694
|
Added Boss, Skill Points, and fixed some minor text
Added some things and fixed some things. Planning on adding other types of monsters later.
I'll check it out tomorrow and some things might need to change but that won't be until the story gets added. I'll also need to reorganize the files for the boss and other enemies. The "Gaben boss" might become an Easter egg/rare boss but I would like to thank you for your work. I'm currently stuck doing exams so it is hard for me to work on my projects.
Just an FYI, you can also update the README.md with your new commands or I can do it for you later.
Also just a tip, try to comment your code. We are all guilty of this and I'll comment it more once exams are over but it will be better for people who just saw this project to understand it.
Sorry, pressed wrong button but if you can, try to change them.
Alright
|
gharchive/pull-request
| 2017-01-30T02:12:04 |
2025-04-01T04:35:13.433689
|
{
"authors": [
"KingofKFCJamal",
"nickthegamer5"
],
"repo": "nickthegamer5/SteamBattleBot",
"url": "https://github.com/nickthegamer5/SteamBattleBot/pull/5",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2049824271
|
Add prettier editor config for code consistency
My aim initially was to simply investigate. When I saved a file I noticed my editor removed all semicolons as it is my personal preference. I decided it would be useful to prevent editors like mine from transforming files on save and keep code as consistent as possible for future contributors on this very useful project.
Great shout - thanks!
|
gharchive/pull-request
| 2023-12-20T04:47:18 |
2025-04-01T04:35:13.436186
|
{
"authors": [
"josephdburdick",
"nicoalbanese"
],
"repo": "nicoalbanese/kirimase",
"url": "https://github.com/nicoalbanese/kirimase/pull/110",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2186153265
|
Initialize db if it doesn't exist instead of throwing an error
running docker-compose -f ./docker/docker-compose.yml run app gives me the following output:
Traceback (most recent call last):
File "/commandjobs/src/menu.py", line 429, in <module>
curses.wrapper(main)
File "/usr/local/lib/python3.12/curses/__init__.py", line 94, in wrapper
return func(stdscr, *args, **kwds)
^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/commandjobs/src/menu.py", line 425, in main
app = MenuApp(stdscr, logger)
^^^^^^^^^^^^^^^^^^^^^^^
File "/commandjobs/src/menu.py", line 54, in __init__
self.db_manager = DatabaseManager(self.db_path) # Specify the path
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/commandjobs/src/database_manager.py", line 6, in __init__
self.conn = sqlite3.connect(db_path)
^^^^^^^^^^^^^^^^^^^^^^^^
sqlite3.OperationalError: unable to open database file
I can see the path to database file in the .env file, but that path does not exist by default. I created repo/job_listings.db in the base directory of the repo, but the error remains.
here is what I did:
mkdir repo
touch repo/job_listings.db
docker-compose -f docker/docker-compose.yml build
docker-compose -f docker/docker-compose.yml run app
Hello @hammadfauz, thank you for reporting this. Sorry about that. You are right, the app should just use a default db name and initialize it
Recently did some changes on the paths and forgot to update the config/sample.env file with the proper ones
Just pushed this PR that fixes it
https://github.com/nicobrenner/commandjobs/pull/42
You could git pull on your local repo and copy the sample.env file again, or if you prefer, you can just manually remove the /repo/ part of the path of the BASE_RESUME and DB_PATH variables
|
gharchive/issue
| 2024-03-14T11:51:56 |
2025-04-01T04:35:13.439741
|
{
"authors": [
"hammadfauz",
"nicobrenner"
],
"repo": "nicobrenner/commandjobs",
"url": "https://github.com/nicobrenner/commandjobs/issues/41",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2195914228
|
Packaging/windows
About
This PR adds makefiles to build Window MSI installers. The versioning is tied to the latest git-tag. Notably, this does not include the necessary updates to create patch files; Windows has a pretty serious disconnect between Installation packages and Patch packages. This means that if someone wants to upgrade, they would need to explicitly uninstall the existing lfs-s3 and then reinstall the new one.
https://github.com/nicolas-graves/lfs-s3/issues/60
@Bekreth I was able to left a review with a single question, thanks for explaining everything.
Thanks again @Bekreth! How about also adding a word in the README for Windows users about how to install/use it on Windows? It could be just one line in the Download & Install section.
|
gharchive/pull-request
| 2024-03-19T20:15:28 |
2025-04-01T04:35:13.445834
|
{
"authors": [
"Bekreth",
"nicolas-graves"
],
"repo": "nicolas-graves/lfs-s3",
"url": "https://github.com/nicolas-graves/lfs-s3/pull/64",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
345372528
|
openApi Applying security
Hello,
Today I learned about defining securitySchemes
(step 1 in https://swagger.io/docs/specification/authentication/):
openApi = OpenApi().apply {
...
components.securitySchemes["BearerAuth"] = mapOf(
"type" to "http",
"scheme" to "bearer"
)
// paths["/corporations"]["security"] = Operation.create(...)
...
}
How can I apply security (step 2) to a path (e.g. "/corporations")?
Thanks
Hi in0rdr,
currently you can't, but feel free to contribute with a pull request.
regards Niels
|
gharchive/issue
| 2018-07-27T21:32:20 |
2025-04-01T04:35:13.457798
|
{
"authors": [
"in0rdr",
"nielsfalk"
],
"repo": "nielsfalk/ktor-swagger",
"url": "https://github.com/nielsfalk/ktor-swagger/issues/13",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2075879415
|
Update models
Summary
update models
LGTM
|
gharchive/pull-request
| 2024-01-11T06:40:48 |
2025-04-01T04:35:13.459202
|
{
"authors": [
"aokumasan",
"fuku2014"
],
"repo": "nifcloud/nifcloud-cli",
"url": "https://github.com/nifcloud/nifcloud-cli/pull/32",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1823085869
|
NgHttpCachingBrowserStorage based storage does not work correctly
When I try to configure ng-http-caching to use Local or Session storage, the cache is being populated, but the results are never served from the cache. A new HTTP request is always triggered.
I believe its because the 'version' field is not being set when writing to the cache.
https://github.com/nigrosimone/ng-http-caching/blob/72b896306d36abee0b28b2ec90ac51d225b7cd23/projects/ng-http-caching/src/lib/storage/ng-http-caching-browser-storage.ts#L131
Hi, thanks! I have fixed in 16.0.5 just published on NPM. Now, it works as expected?
Thanks, it works now.
|
gharchive/issue
| 2023-07-26T20:07:51 |
2025-04-01T04:35:13.500630
|
{
"authors": [
"bryhemm",
"nigrosimone"
],
"repo": "nigrosimone/ng-http-caching",
"url": "https://github.com/nigrosimone/ng-http-caching/issues/9",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
936217971
|
🛑 RFH Status is down
In d1e8195, RFH Status (https://rfh-koeln.sciebo.de/) was down:
HTTP code: 503
Response time: 801 ms
Resolved: RFH Status is back up in 0abcc90.
|
gharchive/issue
| 2021-07-03T09:45:07 |
2025-04-01T04:35:13.528713
|
{
"authors": [
"niklasmtj"
],
"repo": "niklasmtj/sciebo-check",
"url": "https://github.com/niklasmtj/sciebo-check/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
185705442
|
fix(thread_local): Removes unnecessary semicolon
Not sure if this is explicitly a rust 1.9.0 thing or not
Oh, interesting, the semicolon was only added to the thread_local! pattern in 1.11.0. It's completely optional though, so your fix is fine.
Maybe add 1.9.0 to .travis.yml? That way we always know the required minimum, and make sure that it works. This is the true minimum right now too, for std::panic support. I don't know if @nikomatsakis has a policy on rustc versions being a breaking change -- opinions vary and I'm in the camp that it is breaking. Either way, it's good to have that in CI so we know what works right now.
@cuviper I'm not sure if I've formed a strong opinion, but I do expect users to basically keep up with stable --- that is to say, if there is a new stable feature, I wouldn't hesitate to add a dependency on it. But it seems reasonable to consider that a breaking change, at the same time.
In any case, I always forget that ; isn't allowed in thread_local!.
|
gharchive/pull-request
| 2016-10-27T15:40:32 |
2025-04-01T04:35:13.531575
|
{
"authors": [
"cuviper",
"durango",
"nikomatsakis"
],
"repo": "nikomatsakis/rayon",
"url": "https://github.com/nikomatsakis/rayon/pull/126",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
182448167
|
readme update for npm
I was reading https://github.com/nilbus/Backbone.dualStorage/pull/141 and realized it might not be clear how to add this via npm if it isnt registered with https://npmjs.com.
This will add "Backbone.dualStorage": "github:nilbus/backbone.dualstorage" to your package.json's dependencies and install the node module.
Thanks!
|
gharchive/pull-request
| 2016-10-12T06:47:59 |
2025-04-01T04:35:13.534902
|
{
"authors": [
"digitalvapor",
"nilbus"
],
"repo": "nilbus/Backbone.dualStorage",
"url": "https://github.com/nilbus/Backbone.dualStorage/pull/153",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1311989313
|
Enable downstream dependency analysis with updated javadoc
This PR enables Annotator to process downstream dependencies while making decisions for public methods with non-primitive return types. This PR is a follow up for #624
Before starting the main process, annotator will do the followings:
Collects public method with non-primitive-return-type
Collects regions in downstream dependencies that will potentially introduce a new error if a method in target module is annotated.
Constructs the conflict graph and computes the effect in the collected regions for each method.
Aggregates the result of analysis on each sub-module and makes a uniformed report.
Stores the effect on submodule and will provide the information to main process while making decisions.
To activate this feature, flags below must be passed to Annotator:
-ddbc or --downstream-dependencies-build-command list of commands to run downstream dependencies separated by comma,
-nlmlp or --nullaway-library-model-loader-path path to library model loader for NullAway.
This PR also adds javadoc on huge segment of the code.
@lazaroclapp @ketkarameya I removed the request for review on this PR since I haven't added the proper javadoc on this PR methods. Will request a review as soon as I add the javadocs on this :)
Quick high-level question. Based on the description, this is focused on checking when making a return type @Nullable will introduce errors to a downstream target. What about the case of errors in a downstream target due to a parameter not being @Nullable? Isn't it possible we would want to introduce some @Nullable parameter annotations based on downstream targets?
Quick high-level question. Based on the description, this is focused on checking when making a return type @Nullable will introduce errors to a downstream target. What about the case of errors in a downstream target due to a parameter not being @Nullable? Isn't it possible we would want to introduce some @Nullable parameter annotations based on downstream targets?
@msridhar Hi Manu, sure that’s also a useful information from downstream dependencies. This PR is mostly focused on preparing the infrastructure and pipelines to enable this process. Once this lands, we can add features as such incrementally in followup PRs :)
Sorry I'm on holiday until August 1. I can try to look after that
For the future: for ease of review, a PR should not make a lot of unrelated changes. The ideal is one PR per feature and separate PRs for specific kinds of refactorings. This PR was both way too large and attempting to do too many different things for me to truly review it properly.
This is my mistake as well, since I told you to go ahead with javadocs, but I thought you meant on the new feature for this PR, not on unrelated files. Ideally, it should have been one PR to implement downstream dependency analysis and then one or more separate PRs doing refactorings/documentation, I think.
That said, no need to break it down at this point. I did what I could to review the full PR. In the end, I skimmed some of the refactorings+javadoc changes, particularly around index/ and trackers/ in the core. I did go over submodules/ and the changes to injector/ a bit more carefully, though.
Other than that, one thing that definitely stands out to me, though, is that all test changes seem to be due to an unrelated refactoring. I see no tests for the new functionality. How feasible is to add a multi-target test as a sanity check here? (If not very easy, maybe on Wednesday we can discuss the state of any internal manual testing? And leave automated testing for a follow up PR?)
Edit: I see some testing infra added, but it happened in the middle of my review pass, will check that on the next pass. At this point, please avoid changes within this PR that aren't needed to address the comments or to test/fix the downstream target analysis feature, @nimakarimipour (happy to see further refactors/documentation improvements, but on new PRs!)
Hi @lazaroclapp @ketkarameya, sorry for the inconveniences this PR caused in the review process. I will definitely keep that in mind for future PRs. While I was adding javadoc to the code, in many cases, it was required to also add documents to the files that are not necessarily changed but used in the current class. This lead to documenting large segment of the code. Also during adding documents I understood some parts could have been written better and did the refactoring to speed up the review process which now I know I should have created multiple PRs with better order of processing. Will definitely do that for all future PRs. For now as you said, I will only address comments and keep all future changes (such as completing test infra) in future PRs. Please let me know if you think I should break this PR in future.
@ketkarameya @lazaroclapp This PR is ready for another review.
@lazaroclapp @ketkarameya Thank you for your comments, I will add follow up PRs to complete all required needs.
|
gharchive/pull-request
| 2022-07-20T21:53:46 |
2025-04-01T04:35:13.695334
|
{
"authors": [
"msridhar",
"nimakarimipour"
],
"repo": "nimakarimipour/NullAwayAnnotator",
"url": "https://github.com/nimakarimipour/NullAwayAnnotator/pull/26",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
197451637
|
Dark theme with textfield textcolor = white
Hi - just stumbled upon this and running it through some tests and so far love it! Very good job.
I'm not a good dev can get by - but not good enough to pull, figure out and fix. So if you don't mind I'll just make some suggestion here as I go along and see if they make sense to you.
I like your dark theme - good thinking there. I'm trying a textfield in it and the text is black so it is hard to see in a dark text field.
@rjpalermo1
Good catch! I will fix this ASAP and let you know once it's fixed.
Hi @rjpalermo1,
Thanks again for bringing this to my attention. I've added a fix for this in V1.2.9. Cheers 👍
|
gharchive/issue
| 2016-12-24T06:48:48 |
2025-04-01T04:35:13.698101
|
{
"authors": [
"nimati",
"rjpalermo1"
],
"repo": "nimati/FCAlertView",
"url": "https://github.com/nimati/FCAlertView/issues/35",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
408310396
|
Make all requests reloadable
Requests which render their DOM themselves (instead of having it predefined in HTML) cannot be reloaded properly. The hash-targeted element is created after the browser tries to focus it.
Requests with that problem: export, ...?
Solved by #174
|
gharchive/issue
| 2019-02-08T20:24:23 |
2025-04-01T04:35:13.707049
|
{
"authors": [
"Bettelstab"
],
"repo": "nimiq/keyguard-next",
"url": "https://github.com/nimiq/keyguard-next/issues/172",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
267215645
|
cowboy_clock crashes during lookup of rfc1123
I get below error log intermittently, possibly during high load in the system.
2017-10-20 04:56:50.535 [error] <0.720.0> Lager event handler error_logger_lager_h exited with reason {'EXIT',{{case_clause,[http,<0.4193.0>,1,<0.4194.0>,badarg,[{ets,lookup_element,[cowboy_clock,rfc1123,2],[]},{cowboy_clock,rfc1123,0,[{file,"/drone/src/code.xaptum.xyz/xaptum-development/xmx/_build/default/lib/cowboy/src/cowboy_clock.erl"},{line,54}]},{cowboy_req,response_headers,2,[{file,"/drone/src/code.xaptum.xyz/xaptum-development/xmx/_build/default/lib/cowboy/src/cowboy_req.erl"},{line,778}]},{cowboy_req,do_reply,4,[{file,"/drone/src/code.xaptum.xyz/xaptum-development/xmx/_build/defau..."},...]},...]]},...}}
2017-10-20 04:56:50.588 [error] <0.4197.0>@cowboy_clock:rfc1123:54 CRASH REPORT Process <0.4197.0> with 0 neighbours crashed with reason: bad argument in call to ets:lookup_element(cowboy_clock, rfc1123, 2) in cowboy_clock:rfc1123/0 line 54
2017-10-20 04:56:50.588 [error] <0.720.0> Lager event handler error_logger_lager_h exited with reason {'EXIT',{{case_clause,[http,<0.4196.0>,1,<0.4197.0>,badarg,[{ets,lookup_element,[cowboy_clock,rfc1123,2],[]},{cowboy_clock,rfc1123,0,[{file,"/drone/src/code.xaptum.xyz/xaptum-development/xmx/_build/default/lib/cowboy/src/cowboy_clock.erl"},{line,54}]},{cowboy_req,response_headers,2,[{file,"/drone/src/code.xaptum.xyz/xaptum-development/xmx/_build/default/lib/cowboy/src/cowboy_req.erl"},{line,778}]},{cowboy_req,do_reply,4,[{file,"/drone/src/code.xaptum.xyz/xaptum-development/xmx/_build/defau..."},...]},...]]},...}}```
Is there anything else? Because I think this could happen for one of two cases:
the cowboy_clock process has crashed and is being restarted; the ets table is temporarily gone
there is a bug in ets and the writes are not atomic
Would be good if you can confirm whether cowboy_clock restarted (I guess the pid number can give a good hint) and if it has, see if you can find the crash reason.
Hi Irina,
I have the same issue when I did not start cowboy as application. The cowboy_clock process is started as worker by cowboy_sup that is started by cowboy_app.
I hope it will help in your debugging.
Regards,
Alexei.
Hi Alexei,
Thanks so much!
I have cowboy app running for sure as it's working for me all along.
However, I saw this error again and this time I was mindful of what you
said above and realized that this happened while my server was being
restarted (obviously along with all the other apps restarting with it -- so
looks like something is still trying to use cowboy when it's already
stopped/stopping and that's why I get this error). I can't be sure now
about the other times I was seeing it, maybe it was always this situation.
I'll keep an eye on it to see if I can reproduce when the server isn't
restarting, but I suspect now that this was probably always the case.
Cheers,
Irina.
On Sun, Oct 22, 2017 at 11:08 PM, alekras notifications@github.com wrote:
Hi Irina,
I have the same issue when I did not start cowboy as application. The
cowboy_clock process is started as worker by cowboy_sup that is started by
cowboy_app.
I hope it will help in your debugging.
Regards,
Alexei.
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub
https://github.com/ninenines/cowboy/issues/1228#issuecomment-338542884,
or mute the thread
https://github.com/notifications/unsubscribe-auth/ALcepvOrNw0OFyi69vsi0D-kk2XtcWABks5svBFFgaJpZM4QA15J
.
I can see that happening if Ranch supervises the connections because those will still run while Cowboy is being restarted, and if Cowboy is not there then cowboy_clock isn't either, so the ets table doesn't exist (anymore).
I suppose a graceful way to deal with that is to catch the error and then compute the value directly if the ets table is gone.
Hi Loïc,
I've been watching our logs and seems like it does happen upon server exit.
So I guess it isn't really an issue.
Thanks so much for your help,
Irina.
On Mon, Oct 23, 2017 at 11:01 AM, Loïc Hoguin notifications@github.com
wrote:
I can see that happening if Ranch supervises the connections because those
will still run while Cowboy is being restarted, and if Cowboy is not there
then cowboy_clock isn't either, so the ets table doesn't exist (anymore).
I suppose a graceful way to deal with that is to catch the error and then
compute the value directly if the ets table is gone.
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub
https://github.com/ninenines/cowboy/issues/1228#issuecomment-338708338,
or mute the thread
https://github.com/notifications/unsubscribe-auth/ALcepieu9wPpm-5c-0AXFCSEfrsYKDcmks5svLhSgaJpZM4QA15J
.
Cheers, good to know! I will still do the proposed change in my previous comment so that these logs don't appear anymore and to not have connections get killed because of this.
Fixed locally, this will be part of 2.1. Thanks!
Awesome! Thanks so much Loïc!
|
gharchive/issue
| 2017-10-20T15:35:00 |
2025-04-01T04:35:13.720101
|
{
"authors": [
"alekras",
"essen",
"iguberman"
],
"repo": "ninenines/cowboy",
"url": "https://github.com/ninenines/cowboy/issues/1228",
"license": "isc",
"license_type": "permissive",
"license_source": "bigquery"
}
|
526573969
|
Required has_body missing for response with code = 400
Here is an error message from log:
Log access failed for: [400, #{<<"connection">> => <<"close">>,<<"content-length">> => <<"0">>}, #{headers => #{<<"accept">> => <<"image/gif, i"...>>,<<"accept-chars"...>> => <<"iso-8859-1,*"...>>,<<"accept-langu"...>> => <<"en">>,<<"cache-co"...>> => <<"no-cache">>,...},method => <<"GET">>,path => <<"/acute-cp/">>,peer => {{10756,18944,...},48951},...}]
with: error:function_clause
stacktrace: in call to cowboy_req:has_body(#{headers => #{<<"accept">> => <<"image/gif, image/x-xbitm...">...) at /home/jenkins/jenkins-agent/jenkins-agent/workspace/ey_private_fistful-server_master/_build/default/lib/cowboy/src/cowboy_req.erl:466
called from cowboy_access_log_h:get_request_body_length/1 at /home/jenkins/jenkins-agent/jenkins-agent/workspace/ey_private_fistful-server_master/_build/default/lib/cowboy_access_log/src/cowboy_access_log_h.erl:122
called from cowboy_access_log_h:prepare_meta/3 at /home/jenkins/jenkins-agent/jenkins-agent/workspace/ey_private_fistful-server_master/_build/default/lib/cowboy_access_log...
It look like there is no has_body attribute in request, possibly because where is a problem with request and it goes thru cowboy_http:error_terminate/4.
What do you mean by required? If it goes through error_terminate then it calls early_error which gives you a PartialReq :: map() and not a cowboy_req:req(). In other words you get the values Cowboy figured out up until that point.
I mean it required by type specification here and by code here.
So if I get PartialReq :: map() how can I detect it?
You are under the impression that early_error receives the Req, it does not, see https://ninenines.eu/docs/en/cowboy/2.7/manual/cowboy_stream/#_callbacks
PartialReq - cowboy_req:req(), except all fields are optional
It doesn't really matter if there's a body or not as far as Cowboy is concerned because it will just close the connection without attempting to read more. Sometimes it has not read enough to be able to know there is or is not a request body.
I see, thanks!
|
gharchive/issue
| 2019-11-21T12:48:44 |
2025-04-01T04:35:13.725318
|
{
"authors": [
"essen",
"kpy3"
],
"repo": "ninenines/cowboy",
"url": "https://github.com/ninenines/cowboy/issues/1417",
"license": "isc",
"license_type": "permissive",
"license_source": "bigquery"
}
|
295945129
|
Is this project dead?
@hgschmie @electrum pinging you since Github shows you as the only contributors.
I see no PRs were merged (nor commented on) since 2016
Build is trying to use java versions no longer supported on travis
All in all it looks like there are no active maintainers.
I'd like to help here if possible (e.g. by becoming a contributor) instead of making a hard fork
Are you interested?
wow, haven't checked that repo in a long time.
Yes, it is. See https://github.com/basepom/dependency-versions-check-maven-plugin
I know about basepom one but for some reason, that I no longer remember, I found something in this project useful that wasn't there.
(I was wondering if I should also wait the customary 4 years before replying but meh :D)
sorry. I don't really check the old ning repos anymore. basepom is more up-to-date. :-)
I would close and archive those if I could but I lost my admin rights for the ning repos and it seems no one is left that can do this.
|
gharchive/issue
| 2018-02-09T17:16:24 |
2025-04-01T04:35:13.730884
|
{
"authors": [
"hgschmie",
"jakub-bochenski"
],
"repo": "ning/maven-dependency-versions-check-plugin",
"url": "https://github.com/ning/maven-dependency-versions-check-plugin/issues/17",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
243847179
|
Router
When i run the server on windows with a router and connect to myselfI get a bunch of "couldn't reroute" errors regarding the Ping packet.
I'm a bit late sorry, but the good news is that I have a repro!
I will push a fix soon
|
gharchive/issue
| 2017-07-18T20:50:41 |
2025-04-01T04:35:13.731896
|
{
"authors": [
"Nickito12",
"ningirsu"
],
"repo": "ningirsu/stepmania-server",
"url": "https://github.com/ningirsu/stepmania-server/issues/40",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
206229038
|
ENH: Speed up rendering by sharing data
We have a lot of data that could be shared (vertex locations), and we should do it:
http://docs.enthought.com/mayavi/mayavi/tips.html#sharing-the-same-data-between-scenes
I'm working on this at #191
|
gharchive/issue
| 2017-02-08T15:23:15 |
2025-04-01T04:35:13.747308
|
{
"authors": [
"Eric89GXL",
"christianbrodbeck"
],
"repo": "nipy/PySurfer",
"url": "https://github.com/nipy/PySurfer/issues/177",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2435274556
|
AVIF detection is not working when a file is represented as an ArrayBuffer
Hey, thanks for the great library!
I found an issue when detecting avif files. When work with an ArrayBuffer, isAvifStringIncluded utility returns false for an avif file, hence isAVIF function also returns false.
getFileChunk preliminary converts ArrayBuffer into Uint8Array, in which numbers are stored in decimal format, therefore there is no need to parse them as hexadecimal here https://github.com/nir11/file-type-checker/blob/main/src/utils/index.ts#L158.
This implementation of the function would be sufficient:
export function isAvifStringIncluded(fileChunk: Array<number>): boolean {
const signature = fileChunk
.slice(4, 12)
.map(hex => String.fromCharCode(hex))
.join('')
return signature === 'ftypavif'
}
Hey, great catch!
I’ve fixed it now.
Thanks!
|
gharchive/issue
| 2024-07-29T12:22:48 |
2025-04-01T04:35:13.756087
|
{
"authors": [
"nir11",
"prplx"
],
"repo": "nir11/file-type-checker",
"url": "https://github.com/nir11/file-type-checker/issues/13",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2228966090
|
[Question] What are the different Connection Strategies?
Hey!
Would it be possible to get a brief summary of the different Connection Strategies and when to use which.
Found it in etc/aawgd.env
|
gharchive/issue
| 2024-04-06T00:05:54 |
2025-04-01T04:35:13.760140
|
{
"authors": [
"NabeelUppel"
],
"repo": "nisargjhaveri/WirelessAndroidAutoDongle",
"url": "https://github.com/nisargjhaveri/WirelessAndroidAutoDongle/issues/100",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2248444420
|
如何设置代理服务器
是要挂vpn一类的吗
手机上直接开全局科学上网
|
gharchive/issue
| 2024-04-17T14:29:16 |
2025-04-01T04:35:13.772763
|
{
"authors": [
"YANYANG1234",
"niuhuan"
],
"repo": "niuhuan/pikapika",
"url": "https://github.com/niuhuan/pikapika/issues/274",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
666978385
|
Switch CDN from "unpkg.com" to "cdn.jsdelivr.net"
In Febrary 2020 (#1022), docsify has switch its CDN from unpkg.com to cdn.jsdelivr.net.
As a main plugin, it would be nice if docsify-edit-on-github can be hosted on cdn.jsdelivr.net ; more coherent with other docsify plugins.
@axel3rd enabled,thx for suggestion.
https://cdn.jsdelivr.net/npm/docsify-edit-on-github
|
gharchive/issue
| 2020-07-28T10:53:59 |
2025-04-01T04:35:13.805512
|
{
"authors": [
"axel3rd",
"njleonzhang"
],
"repo": "njleonzhang/docsify-edit-on-github",
"url": "https://github.com/njleonzhang/docsify-edit-on-github/issues/12",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1802290456
|
資料共有フォーマットを作成
資料共有フォーマットを作成し完成したのでプルリクエストを送らせてもらいました。
レビューよろしくお願いします。
レイアウト完璧です!
ありがとうございます
マージします!
|
gharchive/pull-request
| 2023-07-13T06:31:36 |
2025-04-01T04:35:13.816515
|
{
"authors": [
"GodHexagon",
"yu-za66"
],
"repo": "nkc-ug/PSI",
"url": "https://github.com/nkc-ug/PSI/pull/14",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
126656666
|
nothing happens when running bootprint swagger
When I run bootprint swagger http://petstore.swagger.io/v2/swagger.json doc, I get the following output:
Loading bootprint-swagger 0.13.1
Loading bootprint-json-schema 0.8.4
Loading bootprint-base 0.6.3
But nothing happens after that. No files were put in the doc directory.
Hi there,
have you made a fresh install or have you updated your version? Have a look at bootprint#12. This issue has appeared 2 days ago and should be resolved by now... You may have to clear the npm-cache, remove bootprint and reinstall to get the current version of deep-aplus (@1.0.2)
That was it, thanks!
|
gharchive/issue
| 2016-01-14T13:40:08 |
2025-04-01T04:35:13.821810
|
{
"authors": [
"bendavis78",
"nknapp"
],
"repo": "nknapp/bootprint-swagger",
"url": "https://github.com/nknapp/bootprint-swagger/issues/49",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
366930555
|
semaphore wait is causing stall in app
There is few possible configurations that will cause app stall because of semaphore.wait. Most often is read/write characteristic from within notificationCallback. This is caused because notification/indication procedure is waiting for confirmation, but in mean time we want to request read/write from peer device and 2 semaphores block each other.
Workaround to this situation is to create new task from within notificationCallback that will request read/write. This task will be executed right after notificationCallback ends. We have to remember that we cant in this case to call vTaskDelay/delay after creating such task, because this will cause semaphores blocking.
I am working now on few upgrades and bugfixes but this is also on my list TODO.
Yeah... There is a take and a wait for the notify. I commented this out
but having stack overflow issues so an not sure if that helped.
On Thu, Oct 4, 2018, 11:26 PM chegewara notifications@github.com wrote:
Its hard to say what kind of confirmation it is, it can be confirmation
from bt stack that notification has been send. The fact is that event CONF
is triggered.
—
You are receiving this because you commented.
Reply to this email directly, view it on GitHub
https://github.com/nkolban/esp32-snippets/issues/673#issuecomment-427235270,
or mute the thread
https://github.com/notifications/unsubscribe-auth/AAmH_fbqaXPJvEiEN1VSbMiCJg-0ys5Iks5uhtFYgaJpZM4XIu-x
.
Ok, with some more tests i can for sure say that this semaphore. wait is not suppose to be in here:
https://github.com/nkolban/esp32-snippets/blob/master/cpp_utils/BLECharacteristic.cpp#L572
It not only causes issues with stal, it also causing other issues. I am testing now enhancement with one esp32 server and 3 esp32 clients connected. Each client writes to server characteristic and when characteristic is write then server sends notification to every connected client. I have 2xesp32 that are writing every 250ms and one esp32 writing every 500ms and without that semaphore.wait i have no issue, but with that line some notifications are dropped.
I (738524) SampleClient: 8--> 7460
E (738524) SampleClient: Notify callback for characteristic beb5483e-36e1-4688-b7f5-ea07361b26a8 of data Time since boot: 430.216617 length 27
E (738614) SampleClient: Notify callback for characteristic beb5483e-36e1-4688-b7f5-ea07361b26a8 of data Time since boot com3: 738.438042 length 32
E (738614) SampleClient: Notify callback for characteristic beb5483e-36e1-4688-b7f5-ea07361b26a8 of data Time since boot com9: 513.238381 length 32
I (738774) SampleClient: 8--> 7460
E (738854) SampleClient: Notify callback for characteristic beb5483e-36e1-4688-b7f5-ea07361b26a8 of data Time since boot com3: 738.688038 length 32
E (738854) SampleClient: Notify callback for characteristic beb5483e-36e1-4688-b7f5-ea07361b26a8 of data Time since boot com9: 513.488417 length 32
E (739004) SampleClient: Notify callback for characteristic beb5483e-36e1-4688-b7f5-ea07361b26a8 of data Time since boot: 430.71851788417 length 27
I`
Implementing a keyboard with a notification going out on every key press, I've noticed the .notify() takes quite a long time. In fact so long that it introduces a noticeable delay after a key has been pressed. Rapidly hitting a key 3 times in a row only registers 2 keypresses with bluetooth host. Getting rid of that semaphore wait seems to have fixed the problem (discovered that on my own by trial and error, then found this thread – means I am not crazy?). This is happening on a single-core device with lots of other tasks running.
@toxuin I have 2 questions, do you have input->notify() in separate task? and what is your Tick rate in menuconfig->FreeRTOS?
@chegewara yes, it's a task that does very little apart from
inputCharacteristic->setValue(report, sizeof(report));
inputCharacteristic->notify();```
Tick rate is default(?) 100hz.
@toxuin Now with tick rate 100, do you have vTaskDelay() in it and if yes do you have port_TICK_PERIOD_MS to convert ticks to ms? Also try to create notify task with higher priority than other tasks if its possible. I can send message from my example code without any delays between characters.
Bluetooth task in my code waits on a queue receive, so no task delays.
Thanks, Che! I will set the priority and see if that helps. Just to confirm: is it ok to remove that semaphore wait on notify?..
You can try, i didnt encounter any issues when i commented it out.
discussion moved here
https://github.com/nkolban/esp32-snippets/issues/681
|
gharchive/issue
| 2018-10-04T19:11:53 |
2025-04-01T04:35:13.833089
|
{
"authors": [
"chegewara",
"toxuin",
"wegunterjr"
],
"repo": "nkolban/esp32-snippets",
"url": "https://github.com/nkolban/esp32-snippets/issues/673",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2137919755
|
[機能リクエスト]クリップボードからの画像添付
よくスクリーンショットをSNSに画像として添付することが多いので、画像をクリップボードからのペーストで添付できると助かります。
@bee7813993 さん、機能リクエストありがとうございます!こちらのIssueは Issue formの整備により移行させていただきました!
状況の確認を行いたい場合は、#2 へ Subscribeしていただければ幸いです。
上記対応により、本Issueはクローズさせていただきます。
|
gharchive/issue
| 2024-02-16T05:49:43 |
2025-04-01T04:35:13.834550
|
{
"authors": [
"bee7813993",
"nkte8"
],
"repo": "nkte8/skyshare",
"url": "https://github.com/nkte8/skyshare/issues/1",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
366460056
|
When key is not found print the key name into error too
e.g. when I do
std::string result = jsonObject["hello"];
and "hello" is not found it just does an assert and I don't know what key is being looked up. It would be nice to also print out the key name.
Duplicate #932.
|
gharchive/issue
| 2018-10-03T18:04:21 |
2025-04-01T04:35:13.841039
|
{
"authors": [
"graph",
"nlohmann"
],
"repo": "nlohmann/json",
"url": "https://github.com/nlohmann/json/issues/1273",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
776585649
|
no matching function for call to ‘nlohmann::basic_json<>::basic_json()’
Hi Team,
I am trying to convert json to managed object and in this regard, I am using the to_json and from_json functions provided by nlohmann/json.
When I write a sample program, I am able to see no issues with this APIs but when I integrate with the actual project, observing the below error when i call get_to with int as datatype.
no matching function for call to ‘nlohmann::basic_json<>::basic_json()’
CPP file (MTop.cpp)
void MTop::to_json(json& j, const MTop& mo)
{
j = json{{"OClass", mOClass}};
}
void MTop::from_json(const json& j, MTop& mo)
{
mo.mOClass = j.at("OClass").get_to< int >(); ---> This line is showing compilation error.
}
Headerfile (MTop.hpp)
#include <MObjectCls.h>
#include <nlohmann/json.hpp>
using json = nlohmann::json;
class MTop : public MObject
{
public:
virtual void to_json(json& j, const MTop& mo);
virtual void from_json(const json& j, MTop& mo);
private:
MObjectCls mOClass;
};
Headerfile (MObjectCls.h)
class MObjectCls
{
public:
MObjectCls(int value);
};
Since there to_json and from_json are used by the object of this class only, these two functions are not maintained under any namespace.
if the from_json( ) logic is modified as below -
j.at("OClass").get_to(mo.mOClass);
MTop.cpp -->
error: no matching function for call to ‘nlohmann::basic_json<>::basic_json(<brace-enclosed initializer list>)’
error: no matching function for call to ‘nlohmann::basic_json<>::get_to(MObjectCls&) const’
../include/nlohmann/json.hpp:3017:28: error: no type named ‘type’ in ‘struct std::enable_if<false, int>’
../include/nlohmann/json.hpp:3030:25: error: no type named ‘type’ in ‘struct std::enable_if<false, int>’
Please guide me what could be the issue.
Note: Since the header and cpp files are big, I have shared only the contents which are of focus here.
Deepak
The functions from_json and to_json must be defined in the same namespace as the type. They must not be member functions. Please try moving them out of MTop.
About "error: no matching function for call to ‘nlohmann::basic_json<>::get_to(MObjectCls&) const’"
You misused get_to. It should be either
int value = j.get<int>();
or
int value;
j.get_to( value );
First one in your case.
@nlohmann
Thanks for your response.
I have modified the files as below -
MTop.cpp
#include <nlohmann/json.hpp>
using json = nlohmann::json;
namespace MTop_ns
{
typedef enum
{
STATE_LOCKED = 0,
STATE_UNLOCKED = 1,
STATE_SHUTTINGDOWN = 2
} MAdminStateEnum;
NLOHMANN_JSON_SERIALIZE_ENUM ( MAdminStateEnum, {
{STATE_LOCKED, "locked"},
{STATE_UNLOCKED, "unlocked"},
{STATE_SHUTTINGDOWN, "shutdown"},
});
struct MTop000
{
int mOInstId;
std::string mOInstName;
MAdminStateEnum mAdminState;
};
void to_json(json& j, const MTop000& mo)
{
j = json{{"mOInstId", mo.mOInstId},
{"mOInstName", mo.mOInstName},
};
}
void from_json(const json& j, MTop000& mo)
{
j.at("mOInstId").get_to(mo.mOInstId);
j.at("mOInstName").get_to(mo.mOInstName);
}
}
MTop.hpp
#include <MObjectCls.h>
class MTop : public MObject
{
public:
private:
MObjectCls mOClass;
};
Now, I don't see any compilation errors as the from_json and to_json are not part of the member functions.
I would be accessing this as below in other file -
MTop_ns::MTop000 mT = {"10, "mTopClass", STATE_LOCKED"};
json j = mT;
Since I have multiple managed objects, there will be many class files. If I had to write to_json/from_json for all those, then I need to do the same for all right ?
Is there any other way to do it ?
Deepak
Yes, you have to define it for every type. Maybe this article helps: https://json.nlohmann.me/features/arbitrary_types/#simplify-your-life-with-macros
@nlohmann
Thanks for the reference.
Looking into that, I made changes to my code and here is the change -
Headerfile (MTop.hpp)
#include <MObjectCls.h>
#include <nlohmann/json.hpp>
using json = nlohmann::json;
namespace MTop_ns
{
class MTop : public MObject
{
public:
NLOHMANN_DEFINE_TYPE_INTRUSIVE(MTop, moClass)
private:
MObjectCls mOClass;
};
}
Headerfile (MObjectCls.h)
class MObjectCls
{
public:
MObjectCls( );
MObjectCls(int value);
virtual ~MObjectCls( );
};
WIth this, I am observing below errors -
MTop.h->
error: no match for ‘operator=’ (operand types are ‘nlohmann::basic_json<>::value_type’ {aka ‘nlohmann::basic_json<>’} and ‘const MibSccpInactivityTestTimer’)
../include/nlohmann/detail/macro_scope.hpp:268:81: error: no matching function for call to ‘nlohmann::basic_json<>::get_to(MibAdministrativeState&) const’
../include/nlohmann/json.hpp:3017:28: error: no type named ‘type’ in ‘struct std::enable_if<false, int>’
If I modify as below, then I don't see any error.
Headerfile (MTop.hpp)
#include <MObjectCls.h>
#include <nlohmann/json.hpp>
using json = nlohmann::json;
namespace MTop_ns
{
class MTop : public MObject
{
public:
NLOHMANN_DEFINE_TYPE_INTRUSIVE(MTop, moClass)
private:
int mOClass;
};
}
This is a nested class where we need to access parameters from base class..
Should this macro be called inside all the inherited classes with the namespace ?
Deepak
Yes.
Yes.
@nlohmann
Thanks for your continuous support.
According to the document, we see that if the macro NLOHMANN_DEFINE_TYPE_INTRUSIVE is used, there is no need to define the macro within the namespace. Currently, I am following the same. Is this correct ?
NLOHMANN_DEFINE_TYPE_NON_INTRUSIVE(name, member1, member2, ...) is to be defined inside of the namespace of the class/struct to create code for.
NLOHMANN_DEFINE_TYPE_INTRUSIVE(name, member1, member2, ...) is to be defined inside of the class/struct to create code for. This macro can also access private members.
#include <MObjectCls.h>
#include <nlohmann/json.hpp>
using json = nlohmann::json;
class MTop : public MObject
{
public:
NLOHMANN_DEFINE_TYPE_INTRUSIVE(MTop, moClass)
private:
int mOClass;
};
MTop.h: error: invalid use of non-static data member ‘MTop::mOClass’
../include/nlohmann/detail/macro_scope.hpp:276:130: error: ‘NLOHMANN_JSON_TO’ was not declared in this scope
Are these errors related with not using namespace ?
Deepak
@nlohmann
Thanks for your continuous support.
According to the document, we see that if the macro NLOHMANN_DEFINE_TYPE_INTRUSIVE is used, there is no need to define the macro within the namespace. Currently, I am following the same. Is this correct ?
NLOHMANN_DEFINE_TYPE_NON_INTRUSIVE(name, member1, member2, ...) is to be defined inside of the namespace of the class/struct to create code for.
NLOHMANN_DEFINE_TYPE_INTRUSIVE(name, member1, member2, ...) is to be defined inside of the class/struct to create code for. This macro can also access private members.
#include <MObjectCls.h>
#include <nlohmann/json.hpp>
using json = nlohmann::json;
class MTop : public MObject
{
public:
NLOHMANN_DEFINE_TYPE_INTRUSIVE(MTop, moClass)
private:
int mOClass;
};
MTop.h: error: invalid use of non-static data member ‘MTop::mOClass’
../include/nlohmann/detail/macro_scope.hpp:276:130: error: ‘NLOHMANN_JSON_TO’ was not declared in this scope
Are these errors related with not using namespace ?
Deepak
Looks like a typo. In NLOHMANN_DEFINE_TYPE_INTRUSIVE, you write moClass (lower-case o), but the member is called mOClass (upper-case O).
Looks like a typo. In NLOHMANN_DEFINE_TYPE_INTRUSIVE, you write moClass (lower-case o), but the member is called mOClass (upper-case O).
@nlohmann
My apology. The code which was pasted had this typo error.
In the code what is being used, the variable passed is mOClass only.
NLOHMANN_DEFINE_TYPE_INTRUSIVE(MTop, mOClass)
Need your inputs on my below query -
According to the document, we see that if the macro NLOHMANN_DEFINE_TYPE_INTRUSIVE is used, there is no need to define the macro within the namespace. Currently, I am following the same. Is this correct ?
Deepak
@nlohmann
My apology. The code which was pasted had this typo error.
In the code what is being used, the variable passed is mOClass only.
NLOHMANN_DEFINE_TYPE_INTRUSIVE(MTop, mOClass)
Need your inputs on my below query -
According to the document, we see that if the macro NLOHMANN_DEFINE_TYPE_INTRUSIVE is used, there is no need to define the macro within the namespace. Currently, I am following the same. Is this correct ?
Deepak
What compiler are you using? Looks like there's a problem with the macro expansion.
What compiler are you using? Looks like there's a problem with the macro expansion.
Can you try replacing that macro with this?
friend void to_json(nlohmann::json& nlohmann_json_j, const Type& nlohmann_json_t) {
NLOHMANN_JSON_TO(mOClass)
}
friend void from_json(const nlohmann::json& nlohmann_json_j, Type& nlohmann_json_t) {
NLOHMANN_JSON_FROM(mOClass)
}
If that doesn't work, then try this:
friend void to_json(nlohmann::json& nlohmann_json_j, const Type& nlohmann_json_t) {
nlohmann_json_j["mOClass"] = nlohmann_json_t.mOClass;
}
friend void from_json(const nlohmann::json& nlohmann_json_j, Type& nlohmann_json_t) {
nlohmann_json_j.at("mOClass").get_to(nlohmann_json_t.mOClass);
}
This will at least narrow down the failure.
Can you try replacing that macro with this?
friend void to_json(nlohmann::json& nlohmann_json_j, const Type& nlohmann_json_t) {
NLOHMANN_JSON_TO(mOClass)
}
friend void from_json(const nlohmann::json& nlohmann_json_j, Type& nlohmann_json_t) {
NLOHMANN_JSON_FROM(mOClass)
}
If that doesn't work, then try this:
friend void to_json(nlohmann::json& nlohmann_json_j, const Type& nlohmann_json_t) {
nlohmann_json_j["mOClass"] = nlohmann_json_t.mOClass;
}
friend void from_json(const nlohmann::json& nlohmann_json_j, Type& nlohmann_json_t) {
nlohmann_json_j.at("mOClass").get_to(nlohmann_json_t.mOClass);
}
This will at least narrow down the failure.
@gregmarr
Thanks for your inputs.
Will modify as suggested and check.
Deepak
@gregmarr
Thanks for your inputs.
Will modify as suggested and check.
Deepak
@gregmarr
Forgot to share the compiler details.
I am using WSL2 (running Ubuntu 20.4)
$ gcc --version
gcc (Ubuntu 9.3.0-17ubuntu1~20.04) 9.3.0
Copyright (C) 2019 Free Software Foundation, Inc.
This is free software; see the source for copying conditions. There is NO
warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.
$ uname -a
Linux IN-9KHT253 4.19.128-microsoft-standard #1 SMP Tue Jun 23 12:58:10 UTC 2020 x86_64 x86_64 x86_64 GNU/Linux
@gregmarr
Forgot to share the compiler details.
I am using WSL2 (running Ubuntu 20.4)
$ gcc --version
gcc (Ubuntu 9.3.0-17ubuntu1~20.04) 9.3.0
Copyright (C) 2019 Free Software Foundation, Inc.
This is free software; see the source for copying conditions. There is NO
warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.
$ uname -a
Linux IN-9KHT253 4.19.128-microsoft-standard #1 SMP Tue Jun 23 12:58:10 UTC 2020 x86_64 x86_64 x86_64 GNU/Linux
@gregmarr
I tried your suggestion but end up getting different error
#include <MObjectCls.h>
#include <nlohmann/json.hpp>
using json = nlohmann::json;
class MTop : public MObject
{
public:
friend void to_json(nlohmann::json& nlohmann_json_j, const Type& nlohmann_json_t) {
nlohmann_json_j["mOClass"] = nlohmann_json_t.mOClass;
}
friend void from_json(const nlohmann::json& nlohmann_json_j, Type& nlohmann_json_t) {
nlohmann_json_j.at("mOClass").get_to(nlohmann_json_t.mOClass);
}
private:
int mOClass;
};
Below error is observed -
error: ‘Type’ does not name a type
error: ‘Type’ has not been declared
mOClass is integer. Its Unsigned integer. It should be supported by JSON API right ?
Deepak
@gregmarr
I tried your suggestion but end up getting different error
#include <MObjectCls.h>
#include <nlohmann/json.hpp>
using json = nlohmann::json;
class MTop : public MObject
{
public:
friend void to_json(nlohmann::json& nlohmann_json_j, const Type& nlohmann_json_t) {
nlohmann_json_j["mOClass"] = nlohmann_json_t.mOClass;
}
friend void from_json(const nlohmann::json& nlohmann_json_j, Type& nlohmann_json_t) {
nlohmann_json_j.at("mOClass").get_to(nlohmann_json_t.mOClass);
}
private:
int mOClass;
};
Below error is observed -
error: ‘Type’ does not name a type
error: ‘Type’ has not been declared
mOClass is integer. Its Unsigned integer. It should be supported by JSON API right ?
Deepak
Right, I forgot to mention to change Type to the type of mOClass.
Right, I forgot to mention to change Type to the type of mOClass.
@gregmarr
Thanks for the clarification.
This resolved the compilation error.. Hopefully, I should be able to integrate successfully.
Thanks for your support.
@gregmarr
Thanks for the clarification.
This resolved the compilation error.. Hopefully, I should be able to integrate successfully.
Thanks for your support.
|
gharchive/issue
| 2020-12-30T18:38:10 |
2025-04-01T04:35:13.882137
|
{
"authors": [
"YarikTH",
"deepakd82",
"gregmarr",
"nlohmann"
],
"repo": "nlohmann/json",
"url": "https://github.com/nlohmann/json/issues/2559",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
138709671
|
Initial commit
Initial work on MPIM
closing =)
|
gharchive/pull-request
| 2016-03-05T19:10:59 |
2025-04-01T04:35:13.883551
|
{
"authors": [
"james-lawrence",
"nlopes"
],
"repo": "nlopes/slack",
"url": "https://github.com/nlopes/slack/pull/62",
"license": "bsd-2-clause",
"license_type": "permissive",
"license_source": "bigquery"
}
|
801219685
|
Add 1 second wait before storing another annotation
Up to an hour and still running. not sure why - Edit: Completed, running, but this did work.
@thomasyu888 I would like to understand better the cause of the issue so we can find a reliable solution.
@thomasyu888 I would like to understand better the cause of the issue so we can find a reliable solution.
@tschaffter My guess of why this worked is that I am waiting a second before doing another rest call. All of the code that actually returns the response in the server is abstracted away and is part of connexion (unmodifiable). My experiment of adding a sleep in the controller itself caused the same RemoteDisconnect issue, which led me to think that the client could be making too many calls too fast.
Ive been trying to search more about this issue, but I can't seem to find more than what I've already found.
@thomasyu888 A better approach would be to block/loop over the response until the status code 201 is received. This way we always wait enough time instead of assuming that the request will be completed within 1 second (response time depends on hardware and connection bandwidth).
The ideal solution would be to send the requests asynchronously. In JavaScript/TypeScript, I would use mergeMap to send a bunch of requests at once, which is one instruction that would then return once all the requests have completed. The uWSGI server of the data node is configured to process one request at a time but this could be changed in a later revision to decrease processing time (e.g. pulling clinical notes concurrently, sending requests to tools concurrently, etc.).
Could you give a try to the first solution proposed (wait for 201 response)?
@tschaffter Thanks for the resources! I took a look at where I would even look for a response == 201, the issue is the api_instance.create_annotation(... function returns the json object already, so I would have to look deeper in the code. Upon investigate, they seemed to have generated request functions within the classes which is great but also, there are more response codes than just 201.
I'm trying something different. Apparently I can add a async_req=True command to use asynchronous HTTP calls. lets see if that works.
|
gharchive/pull-request
| 2021-02-04T12:01:24 |
2025-04-01T04:35:13.888531
|
{
"authors": [
"thomasyu888",
"tschaffter"
],
"repo": "nlpsandbox/nlpsandbox-client",
"url": "https://github.com/nlpsandbox/nlpsandbox-client/pull/93",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
906390100
|
Review Rules one more time before launch
https://www.synapse.org/#!Synapse:syn22277124/wiki/604831
Doc link can be found here. Main feedback:
**Organization and Structure **
Lots of good content here. Tried to group relevant bullet points together.
(1) Registering for NLP Sandbox: more tactical rules that affect developer workflow
(2) Publication Considerations: concerning publications
(3) Service Agreement: less tactical rules, more the obligatory legal jargon for scenarios of conflict or malicious intent
Consolidating some sections
Wondering if we can consolidate some of the below sections, comments in line
Terms of Use
Terms of Use currently says, not sure if this is a placeholder for a template we need from governance? “BE SURE TO CONTACT GOVERNANCE ABOUT THIS PROJECT”
A point of consideration for the future: per Meg Doerr's readability guidelines, I used Microsoft word to get some readability statistics.
The NLP Sandbox rules are currently readable for a college student or graduate.
If we want to make them readable to a wider audience of Americans, we may want to do a deeper dive on breaking down some of the language. However, given that there is a lot of obligatory legal language here, it may be challenging.
No action required, just a consideration for the future.
|
gharchive/issue
| 2021-05-29T06:24:33 |
2025-04-01T04:35:13.892933
|
{
"authors": [
"jiaxinmachine88",
"tschaffter"
],
"repo": "nlpsandbox/nlpsandbox.io",
"url": "https://github.com/nlpsandbox/nlpsandbox.io/issues/92",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
213584967
|
Stanford segmenter nltk Could not find SLF4J in your classpath
I set up a nltk and stanford environment, nltk and stanford jars has downloaded, the program with nltk was ok, but I had a trouble with stanford segmenter. just make a simple program via stanford segmenter, I got a error is Could not find SLF4J in your classpath, although I had export all jars include slf4j-api.jar. Detail as follows
Python3.5 NLTK, 3.2.2, Standford jars 3.7
OS: Centos
Environment variable:
export JAVA_HOME=/usr/java/jdk1.8.0_60
export NLTK_DATA=/opt/nltk_data
export STANFORD_SEGMENTER_PATH=/opt/stanford/stanford-segmenter-3.7
export CLASSPATH=$CLASSPATH:$STANFORD_SEGMENTER_PATH/stanford-segmenter.jar
export STANFORD_POSTAGGER_PATH=/opt/stanford/stanford-postagger-full-2016-10-31
export CLASSPATH=$CLASSPATH:$STANFORD_POSTAGGER_PATH/stanford-postagger.jar
export STANFORD_NER_PATH=/opt/stanford/stanford-ner-2016-10-31
export CLASSPATH=$CLASSPATH:$STANFORD_NER_PATH/stanford-ner.jar
export STANFORD_MODELS=$STANFORD_NER_PATH/classifiers:$STANFORD_POSTAGGER_PATH/models
export STANFORD_PARSER_PATH=/opt/stanford/stanford-parser-full-2016-10-31
export CLASSPATH=$CLASSPATH:$STANFORD_PARSER_PATH/stanford-parser.jar:$STANFORD_PARSER_PATH/stanford-parser-3.6.0-models.jar:$STANFORD_PARSER_PATH/slf4j-api.jar:$STANFORD_PARSER_PATH/ejml-0.23.jar
export STANFORD_CORENLP_PATH=/opt/stanford/stanford-corenlp-full-2016-10-31
export CLASSPATH=$CLASSPATH:$STANFORD_CORENLP_PATH/stanford-corenlp-3.7.0.jar:$STANFORD_CORENLP_PATH/stanford-corenlp-3.7.0-models.jar:$STANFORD_CORENLP_PATH/javax.json.jar:$STANFORD_CORENLP_PATH/joda-time.jar:$STANFORD_CORENLP_PATH/jollyday.jar:$STANFORD_CORENLP_PATH/protobuf.jar:$STANFORD_CORENLP_PATH/slf4j-simple.jar:$STANFORD_CORENLP_PATH/xom.jar
export STANFORD_CORENLP=$STANFORD_CORENLP_PATH
The program as follows:
from nltk.tokenize import StanfordSegmenter
>>> segmenter = StanfordSegmenter(
path_to_sihan_corpora_dict="/opt/stanford/stanford-segmenter-3.7/data/",
path_to_model="/opt/stanford/stanford-segmenter-3.7/data/pku.gz",
path_to_dict="/opt/stanford/stanford-segmenter-3.7/data/dict-chris6.ser.gz"
)
>>> res = segmenter.segment(u"这是斯坦福中文分词器测试")
The error as follows:
Exception in thread "main" java.lang.ExceptionInInitializerError
at edu.stanford.nlp.ie.AbstractSequenceClassifier.<clinit>(AbstractSequenceClassifier.java:88)
Caused by: java.lang.IllegalStateException: Could not find SLF4J in your classpath
at edu.stanford.nlp.util.logging.RedwoodConfiguration$Handlers.lambda$static$530(RedwoodConfiguration.java:190)
at edu.stanford.nlp.util.logging.RedwoodConfiguration$Handlers$7.buildChain(RedwoodConfiguration.java:309)
at edu.stanford.nlp.util.logging.RedwoodConfiguration$Handlers$7.apply(RedwoodConfiguration.java:318)
at edu.stanford.nlp.util.logging.RedwoodConfiguration.lambda$handlers$535(RedwoodConfiguration.java:363)
at edu.stanford.nlp.util.logging.RedwoodConfiguration.apply(RedwoodConfiguration.java:41)
at edu.stanford.nlp.util.logging.Redwood.<clinit>(Redwood.java:609)
... 1 more
Caused by: edu.stanford.nlp.util.MetaClass$ClassCreationException: java.lang.ClassNotFoundException: edu.stanford.nlp.util.logging.SLF4JHandler
at edu.stanford.nlp.util.MetaClass.createFactory(MetaClass.java:364)
at edu.stanford.nlp.util.MetaClass.createInstance(MetaClass.java:381)
at edu.stanford.nlp.util.logging.RedwoodConfiguration$Handlers.lambda$static$530(RedwoodConfiguration.java:186)
... 6 more
Caused by: java.lang.ClassNotFoundException: edu.stanford.nlp.util.logging.SLF4JHandler
at java.net.URLClassLoader.findClass(URLClassLoader.java:381)
at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:331)
at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
at java.lang.Class.forName0(Native Method)
at java.lang.Class.forName(Class.java:264)
at edu.stanford.nlp.util.MetaClass$ClassFactory.construct(MetaClass.java:135)
at edu.stanford.nlp.util.MetaClass$ClassFactory.<init>(MetaClass.java:202)
at edu.stanford.nlp.util.MetaClass$ClassFactory.<init>(MetaClass.java:69)
at edu.stanford.nlp.util.MetaClass.createFactory(MetaClass.java:360)
... 8 more
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/usr/local/python3/lib/python3.5/site-packages/nltk/tokenize/stanford_segmenter.py", line 96, in segment
return self.segment_sents([tokens])
File "/usr/local/python3/lib/python3.5/site-packages/nltk/tokenize/stanford_segmenter.py", line 123, in segment_sents
stdout = self._execute(cmd)
File "/usr/local/python3/lib/python3.5/site-packages/nltk/tokenize/stanford_segmenter.py", line 143, in _execute
cmd,classpath=self._stanford_jar, stdout=PIPE, stderr=PIPE)
File "/usr/local/python3/lib/python3.5/site-packages/nltk/internals.py", line 134, in java
raise OSError('Java command failed : ' + str(cmd))
OSError: Java command failed : ['/usr/java/jdk1.8.0_60/bin/java', '-mx2g', '-cp', '/opt/stanford/stanford-segmenter-3.7/stanford-segmenter.jar:/opt/stanford/stanford-parser-full-2016-10-31/slf4j-api.jar', 'edu.stanford.nlp.ie.crf.CRFClassifier', '-sighanCorporaDict', '/opt/stanford/stanford-segmenter-3.7/data/', '-textFile', '/tmp/tmpkttpldl6', '-sighanPostProcessing', 'true', '-keepAllWhitespaces', 'false', '-loadClassifier', '/opt/stanford/stanford-segmenter-3.7/data/pku.gz', '-serDictionary', '/opt/stanford/stanford-segmenter-3.7/data/dict-chris6.ser.gz', '-inputEncoding', 'UTF-8']
FYI - An workaround I used is to set path_to_slf4j to sth that does not include slf4j, e.g. path_to_slf4j = "/home/chen/Downloads/stanford-segmenter-2016-10-31/stanford-segmenter-3.7.0.jar"
This should be resolved in our new interface to the Stanford Tools. If not, please reopen.
|
gharchive/issue
| 2017-03-12T06:11:13 |
2025-04-01T04:35:13.897567
|
{
"authors": [
"benbenguo",
"cchunsheng",
"stevenbird"
],
"repo": "nltk/nltk",
"url": "https://github.com/nltk/nltk/issues/1652",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
120471371
|
Update the default download URL to use https://
The index file that the downloader uses is currently served from GitHub
Pages (via http://www.nltk.org), but while GitHub Pages doesn't directly
offer https://, they provide an alternate way to easily download files
securely over https:// via their raw.githubusercontent.com FQDN.
This helps protect the index and the file URLs listed in it from being
MITM'd and modified to something malicious and just helps with ensuring
file integrity.
The other half of this is https://github.com/nltk/nltk_data/pull/40.
Great idea, thanks @reedloden.
|
gharchive/pull-request
| 2015-12-04T19:37:43 |
2025-04-01T04:35:13.900501
|
{
"authors": [
"reedloden",
"stevenbird"
],
"repo": "nltk/nltk",
"url": "https://github.com/nltk/nltk/pull/1220",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
193268218
|
Fixed missing parenthesis on MASI distance found by @drevicko
Fixed missing parenthesis found by @drevicko
There was a missing parenthesis on my MASI code, see
https://github.com/DrDub/nltk/commit/e8bbbd4cd4cc7cb6414cb78959b0641ef22ba9f9#commitcomment-18577279
for details. Thanks to @drevicko for the analysis and for
pointing this out.
This patch also fixed the doctests.
Thanks @drdub
|
gharchive/pull-request
| 2016-12-03T05:48:15 |
2025-04-01T04:35:13.902512
|
{
"authors": [
"DrDub",
"stevenbird"
],
"repo": "nltk/nltk",
"url": "https://github.com/nltk/nltk/pull/1535",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
273232027
|
RAN integration
It would show the power of staart to integrate it with RAN
would be interested in this!
|
gharchive/issue
| 2017-11-12T13:52:27 |
2025-04-01T04:35:13.907048
|
{
"authors": [
"nmaro",
"sw-yx"
],
"repo": "nmaro/staart",
"url": "https://github.com/nmaro/staart/issues/15",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1622785609
|
meteva 为啥要设置绘图字体?
引入 meteva 包后发现 plt.rcParams['font.sans-serif'] 被设置成 simhei ,但是我是 ubuntu 服务器,已经设置好 matplotlib 的字体参数,为什么要重写matplolib字体,是有什么逻辑上的要求么?
绘图函数中使用了plt.rcParams['font.sans-serif']=['SimHei'] 语句来设置中文字体。这不是强制性的要求,如果需要更换字体,在调用 meteva中的绘图函数功能前,可以先运行 plt.rcParams['font.sans-serif'] = [其它字体参数]。
|
gharchive/issue
| 2023-03-14T05:58:02 |
2025-04-01T04:35:13.908182
|
{
"authors": [
"liucouhua",
"qinglonger"
],
"repo": "nmcdev/meteva",
"url": "https://github.com/nmcdev/meteva/issues/150",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1374898779
|
Installation challenges; add node.js to required packages?
Hi all,
I installed ipygoldenlayout today and didn't already have node.js installed. Maybe we should add to the README that you may need to
conda install -c conda-forge nodejs
before you install the labextension?
Also, I'm getting the following error when I run the labextension install command:
Node v18.9.0
Yarn configuration loaded.
> /opt/miniconda3/bin/npm pack ipygoldenlayout
npm notice
npm notice 📦 ipygoldenlayout@0.1.0
npm notice === Tarball Contents ===
npm notice 1.5kB LICENSE.txt
npm notice 752B README.md
npm notice 700.9kB dist/index.js
npm notice 754B lib/extension.js
npm notice 361B lib/index.js
npm notice 1.1kB lib/plugin.js
npm notice 533B lib/version.js
npm notice 5.1kB lib/widget.js
npm notice 2.8kB package.json
npm notice === Tarball Details ===
npm notice name: ipygoldenlayout
npm notice version: 0.1.0
npm notice filename: ipygoldenlayout-0.1.0.tgz
npm notice package size: 205.6 kB
npm notice unpacked size: 713.8 kB
npm notice shasum: 2b0b5b6c782524848fdd46d25a056a611b7d6f5e
npm notice integrity: sha512-OwsrKUV8mTpm9[...]KwKZff+tdos5g==
npm notice total files: 9
npm notice
ipygoldenlayout-0.1.0.tgz
Node v18.9.0
Yarn configuration loaded.
Building jupyterlab assets (production, minimized)
> node /opt/miniconda3/lib/python3.9/site-packages/jupyterlab/staging/yarn.js install --non-interactive
yarn install v1.21.1
[1/5] Validating package.json...
[2/5] Resolving packages...
success Already up-to-date.
Done in 0.16s.
> node /opt/miniconda3/lib/python3.9/site-packages/jupyterlab/staging/yarn.js yarn-deduplicate -s fewer --fail
yarn run v1.21.1
$ /opt/miniconda3/share/jupyter/lab/staging/node_modules/.bin/yarn-deduplicate -s fewer --fail
Done in 0.22s.
> node /opt/miniconda3/lib/python3.9/site-packages/jupyterlab/staging/yarn.js run build:prod:minimize
yarn run v1.21.1
$ webpack --config webpack.prod.minimize.config.js
[webpack-cli] ModuleNotFoundError: Module not found: Error: Can't resolve './custom.css' in '/opt/miniconda3/share/jupyter/lab/staging/node_modules/ipygoldenlayout/lib'
at /opt/miniconda3/share/jupyter/lab/staging/node_modules/webpack/lib/Compilation.js:1768:28
at /opt/miniconda3/share/jupyter/lab/staging/node_modules/webpack/lib/NormalModuleFactory.js:742:13
at eval (eval at create (/opt/miniconda3/share/jupyter/lab/staging/node_modules/tapable/lib/HookCodeFactory.js:33:10), <anonymous>:10:1)
at /opt/miniconda3/share/jupyter/lab/staging/node_modules/webpack/lib/NormalModuleFactory.js:276:22
at eval (eval at create (/opt/miniconda3/share/jupyter/lab/staging/node_modules/tapable/lib/HookCodeFactory.js:33:10), <anonymous>:9:1)
at /opt/miniconda3/share/jupyter/lab/staging/node_modules/webpack/lib/NormalModuleFactory.js:405:22
at /opt/miniconda3/share/jupyter/lab/staging/node_modules/webpack/lib/NormalModuleFactory.js:119:11
at /opt/miniconda3/share/jupyter/lab/staging/node_modules/webpack/lib/NormalModuleFactory.js:650:24
at /opt/miniconda3/share/jupyter/lab/staging/node_modules/webpack/lib/NormalModuleFactory.js:813:8
at /opt/miniconda3/share/jupyter/lab/staging/node_modules/webpack/lib/NormalModuleFactory.js:933:5
resolve './custom.css' in '/opt/miniconda3/share/jupyter/lab/staging/node_modules/ipygoldenlayout/lib'
using description file: /opt/miniconda3/share/jupyter/lab/staging/node_modules/ipygoldenlayout/package.json (relative path: ./lib)
Field 'browser' doesn't contain a valid alias configuration
using description file: /opt/miniconda3/share/jupyter/lab/staging/node_modules/ipygoldenlayout/package.json (relative path: ./lib/custom.css)
no extension
Field 'browser' doesn't contain a valid alias configuration
/opt/miniconda3/share/jupyter/lab/staging/node_modules/ipygoldenlayout/lib/custom.css doesn't exist
.js
Field 'browser' doesn't contain a valid alias configuration
/opt/miniconda3/share/jupyter/lab/staging/node_modules/ipygoldenlayout/lib/custom.css.js doesn't exist
.json
Field 'browser' doesn't contain a valid alias configuration
/opt/miniconda3/share/jupyter/lab/staging/node_modules/ipygoldenlayout/lib/custom.css.json doesn't exist
.wasm
Field 'browser' doesn't contain a valid alias configuration
/opt/miniconda3/share/jupyter/lab/staging/node_modules/ipygoldenlayout/lib/custom.css.wasm doesn't exist
as directory
/opt/miniconda3/share/jupyter/lab/staging/node_modules/ipygoldenlayout/lib/custom.css doesn't exist
error Command failed with exit code 2.
info Visit https://yarnpkg.com/en/docs/cli/run for documentation about this command.
JupyterLab failed to build
Are you trying to install ipygoldenlayout to JupyterLab>=4?
Probably ipygoldenlayout does not support a version >=4 of JupyterLab, yet.
|
gharchive/issue
| 2022-09-15T18:05:42 |
2025-04-01T04:35:13.912576
|
{
"authors": [
"bmorris3",
"jhgoebbert"
],
"repo": "nmearl/ipygoldenlayout",
"url": "https://github.com/nmearl/ipygoldenlayout/issues/35",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1161910763
|
document Instability in initial F near Fmsy
There is a non-monotonic relationship that can interfere with initial F estimation if the option to allow spawner-recruitment to operate inside initial year is used and if the input initial catch is near MSY. In the attached file there is a profile on initial F bracketing Fmsy. It shows, quite logically, that values of F greater or less than Fmsy give initial catch < MSY and that only init F = Fmsy produces an equilibrium catch equal to MSY. So there are two pathological problems:
if initial catch is near MSY (of course, SS has not yet calculated Fmsy yet), then the search for init F will encounter the non-monotonic situation
if initial catch is > MSY, then the mismatch between observed and estimated initial catch cannot be reduced by increasing init F further. However, it might be possible for SS to increase R0 or other parameters in order to increase MSY indirectly.
original issue https://github.com/nmfs-stock-synthesis/stock-synthesis/issues/18
@Rick-Methot-NOAA it seems like this is detailed on page 33 of the user manual (also section 6.3.0.2 in the html file under Benchmark Calculations) when the Fishing Mortality info was updated. Is my understanding of that info correct?
We need to add the text from Kathryn's comment (which I wrote) to section 8.7.1 regarding initial fishing mortality. The topic is already addressed in 8.6.0.1, so the 8.7.1 text can say something like: "If the initial equilibrium catch is near MSY, then a logical inconsistency may occur as documented in section 8.6.0.1."
|
gharchive/issue
| 2022-03-07T20:52:10 |
2025-04-01T04:35:13.922929
|
{
"authors": [
"Rick-Methot-NOAA",
"e-perl-NOAA",
"k-doering-NOAA"
],
"repo": "nmfs-ost/ss3-doc",
"url": "https://github.com/nmfs-ost/ss3-doc/issues/98",
"license": "CC0-1.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1750079595
|
Number of sent messages quota: Quota.Application
Readiness checklist
[x] I added/updated unit tests.
[ ] I added/updated integration tests.
[x] I ensured that the PR title is good enough for the changelog.
[x] I labeled the PR.
With the latest fixes to MetricStatus, we are back to the Transport tests tripping on that one test that fails most times, i.e., the other tests are now passing.
@tnotheis I have merged the target branch with main as well.
|
gharchive/pull-request
| 2023-06-09T15:12:10 |
2025-04-01T04:35:13.926104
|
{
"authors": [
"Dannyps"
],
"repo": "nmshd/backbone",
"url": "https://github.com/nmshd/backbone/pull/160",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1977438004
|
Compability for jam and profile pages.
The jam pages and profile pages don't have some of the color present in project pages #1
Description container for:
Project page
.formatted_description
Profile page
.user_profile
Jam page
.jam_content
Devlog
.post_body
Impossible
|
gharchive/issue
| 2023-11-04T17:26:45 |
2025-04-01T04:35:13.940386
|
{
"authors": [
"nndda"
],
"repo": "nndda/pitch-css",
"url": "https://github.com/nndda/pitch-css/issues/3",
"license": "CC0-1.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
115390138
|
Multiple Charts on a single canvas
I think a useful feature would be the ability to overlay charts on a single canvas. I know this would be difficult though this could provide a very useful level of functionality.
+1
What's wrong with using multiple canvases?
Well I would like to try and overlay certain graphs (i.e a doughnut chart
over a pie graph, to allow for 2 sets of the data to be correlated, or a
column graph with a line chart) and allow to show correlation or save room
on the page
On 10 November 2015 at 01:58, Jake notifications@github.com wrote:
What's wrong with using multiple canvases?
—
Reply to this email directly or view it on GitHub
https://github.com/nnnick/Chart.js/issues/1624#issuecomment-155086251.
@Maraket combo charts between bars and libes are supported in v2. The relevant sample file shows you how to accomplish this.
Closing since v2 supports multiple datasets in pie/doughnut charts. We are not going to support multiple charts on a single canvas at this time.
|
gharchive/issue
| 2015-11-05T22:30:40 |
2025-04-01T04:35:13.947221
|
{
"authors": [
"Maraket",
"etimberg",
"fercreek",
"jakesyl"
],
"repo": "nnnick/Chart.js",
"url": "https://github.com/nnnick/Chart.js/issues/1624",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1391766563
|
[Version] update version to 1.8.3
Release 1.8.2 for Tizen 7.0 M2. Start implementation for next release (v1.8.3).
Signed-off-by: Jaeyun jy1210.jung@samsung.com
:memo: TAOS-CI Version: 1.5.20200925. Thank you for submitting PR #217. Please a submit 1commit/1PR (one commit per one PR) policy to get comments quickly from reviewers. Your PR must pass all verificiation processes of cibot before starting a review process from reviewers. If you are new member to join this project, please read manuals in documentation folder and wiki page. In order to monitor a progress status of your PR in more detail, visit http://ci.nnstreamer.ai/.
|
gharchive/pull-request
| 2022-09-30T03:26:22 |
2025-04-01T04:35:13.958199
|
{
"authors": [
"jaeyun-jung",
"taos-ci"
],
"repo": "nnstreamer/api",
"url": "https://github.com/nnstreamer/api/pull/217",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
998947553
|
[resnet/unittest] Add unittest for resnet
This patch adds resnet18 unittest model generation with genModelTests.
Further, the input data range is changed from 0 to x from -x to x as
relu based models work better with 0 to x data range to prevent loss of
information.
This patch adds resnet models unittest for resnet18.
The verification has been done offline for 2 iterations for the output
of all layers with precision of 1.1e-4.
Derivaitves and gradients have higher error because of relu: when some
value is close to 0, it can be positive or negative with some error (of
the order of e-7). Although this error is way within the error limit.
however, this exacerbates the error in backwarding where derivatives (which
are significant in values) can flow if the relu value was over 0, and
not flow if under zero. This is manageable in smaller models but
difficult to avoid in unittests for larger models.
Other bug fixes in this patch:
max error reported by unittest_nntrainer_models has been fixed
error reporting now includes layer type as well
ModelTestOption MINIMUM has been renamed to NO_THROW_RUN
Resnet unittest is disabled as the golden data exceeds 70MB. This will be enabled when the golden data can be generated at runtime.
Signed-off-by: Parichay Kapoor pk.kapoor@samsung.com
:memo: TAOS-CI Version: 1.5.20200925. Thank you for submitting PR #1578. Please a submit 1commit/1PR (one commit per one PR) policy to get comments quickly from reviewers. Your PR must pass all verificiation processes of cibot before starting a review process from reviewers. If you are new member to join this project, please read manuals in documentation folder and wiki page. In order to monitor a progress status of your PR in more detail, visit http://nnstreamer.mooo.com/.
:octocat: cibot: @kparichay, A builder checker could not be completed because one of the checkers is not completed. In order to find out a reason, please go to http://nnstreamer.mooo.com/nntrainer/ci/repo-workers/pr-checker/1578-202109231315220.69390988349915-c29de3dcac3d6dd3d4f8b936bac7a2b365d3b31d/.
|
gharchive/pull-request
| 2021-09-17T05:08:46 |
2025-04-01T04:35:13.964326
|
{
"authors": [
"kparichay",
"taos-ci"
],
"repo": "nnstreamer/nntrainer",
"url": "https://github.com/nnstreamer/nntrainer/pull/1578",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2305265518
|
[WIP] [Wait for #2584] [ TEST ] add torch input and output test data for mixed precision
In this PR
This PR adds torch mixed precision golden data generation and input and
output for the test.
. some fixes to test.
Resolves:
Self evaluation:
Build test: [X]Passed [ ]Failed [ ]Skipped
Run test: [X]Passed [ ]Failed [ ]Skipped
Signed-off-by: jijoong.moon jijoong.moon@samsung.com
:memo: TAOS-CI Version: 1.5.20200925. Thank you for submitting PR #2591. Please a submit 1commit/1PR (one commit per one PR) policy to get comments quickly from reviewers. Your PR must pass all verificiation processes of cibot before starting a review process from reviewers. If you are new member to join this project, please read manuals in documentation folder and wiki page. In order to monitor a progress status of your PR in more detail, visit http://ci.nnstreamer.ai/.
:octocat: cibot: @jijoongmoon, test/include/nntrainer_test_util.h does not include Doxygen tags such as @file @brief @author @bug. You must include the Doxygen tags in the source code. Please refer to a Doxygen manual at http://github.com/nnstreamer/TAOS-CI/blob/main/ci/doc/doxygen-documentation.md
:octocat: cibot: @jijoongmoon, A builder checker could not be completed because one of the checkers is not completed. In order to find out a reason, please go to http://ci.nnstreamer.ai/nntrainer/ci/repo-workers/pr-checker/2591-202405201604200.90595698356628-66d13bcda6fb4d33dc335e99842756eb04ec4ec6/.
:octocat: cibot: @jijoongmoon, test/include/nntrainer_test_util.h does not include Doxygen tags such as @file @brief @author @bug. You must include the Doxygen tags in the source code. Please refer to a Doxygen manual at http://github.com/nnstreamer/TAOS-CI/blob/main/ci/doc/doxygen-documentation.md
:octocat: cibot: @jijoongmoon, A builder checker could not be completed because one of the checkers is not completed. In order to find out a reason, please go to http://ci.nnstreamer.ai/nntrainer/ci/repo-workers/pr-checker/2591-202405220946440.50683903694153-be6c1eede9268cb8bf348eee51c0b0c0046e6ac3/.
:octocat: cibot: @jijoongmoon, test/include/nntrainer_test_util.h does not include Doxygen tags such as @file @brief @author @bug. You must include the Doxygen tags in the source code. Please refer to a Doxygen manual at http://github.com/nnstreamer/TAOS-CI/blob/main/ci/doc/doxygen-documentation.md
:octocat: cibot: @jijoongmoon, A builder checker could not be completed because one of the checkers is not completed. In order to find out a reason, please go to http://ci.nnstreamer.ai/nntrainer/ci/repo-workers/pr-checker/2591-202405221527370.5227530002594-728fb074979c591cd4b8d6328fb18db5b020c5ed/.
:octocat: cibot: @jijoongmoon, test/include/nntrainer_test_util.h does not include Doxygen tags such as @file @brief @author @bug. You must include the Doxygen tags in the source code. Please refer to a Doxygen manual at http://github.com/nnstreamer/TAOS-CI/blob/main/ci/doc/doxygen-documentation.md
:octocat: cibot: @jijoongmoon, A builder checker could not be completed because one of the checkers is not completed. In order to find out a reason, please go to http://ci.nnstreamer.ai/nntrainer/ci/repo-workers/pr-checker/2591-202405221623260.88832211494446-4f547971ce01761e435f2c6cd9d3a804f3199c92/.
:octocat: cibot: @jijoongmoon, A builder checker could not be completed because one of the checkers is not completed. In order to find out a reason, please go to http://ci.nnstreamer.ai/nntrainer/ci/repo-workers/pr-checker/2591-202405221719140.62206411361694-2d603edd465885797ab17968529b538f96f3c0a2/.
:octocat: cibot: @jijoongmoon, A builder checker could not be completed because one of the checkers is not completed. In order to find out a reason, please go to http://ci.nnstreamer.ai/nntrainer/ci/repo-workers/pr-checker/2591-202405230753120.3925359249115-6c293942e0d1aadb7c3a2368c2f2e9136c3a228e/.
closed by #2663
|
gharchive/pull-request
| 2024-05-20T07:04:17 |
2025-04-01T04:35:13.978882
|
{
"authors": [
"jijoongmoon",
"taos-ci"
],
"repo": "nnstreamer/nntrainer",
"url": "https://github.com/nnstreamer/nntrainer/pull/2591",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
506932503
|
Tests are failing on older python versions
For python version 2.6, 3.2, 3.3 travis is failing with this error message:
Downloading archive: https://storage.googleapis.com/travis-ci-language-archives/python/binaries/ubuntu/16.04/x86_64/python-2.6.tar.bz2
0.14s$ curl -sSf -o python-2.6.tar.bz2 ${archive_url}
curl: (22) The requested URL returned error: 404 Not Found
Unable to download 2.6 archive. The archive may not exist. Please consider a different version.
Your fix was merged (I managed to confuse github at the same time though, so it marked the PR as closed). Thank you for helping out!
Your fix was merged (I managed to confuse github at the same time though, so it marked the PR as closed). Thank you for helping out!
|
gharchive/issue
| 2019-10-15T00:11:51 |
2025-04-01T04:35:13.982226
|
{
"authors": [
"masell",
"noahmorrison"
],
"repo": "noahmorrison/chevron",
"url": "https://github.com/noahmorrison/chevron/issues/61",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
54027583
|
khashier displays this transaction incorrectly
See http://khashier.com/tx/24acea41498b2f387ff751997f8dc90a4551f7b638e7cfc00750a9966f3bc018
It says the fee is -3.9961 when it was really 0.0039.
It says it has 839 outputs when it really has 838.
It lists one of the outputs twice, with two different 'spent' links:
http://khashier.com/tx/5d46a5d2596e41b3f42c161a362532ce82a54c599b2fb5e75e153aa941acb553
http://khashier.com/tx/53a63ff503e39486cafc279d4ae3feabbbaf539c396404c1fff4ad70512244dd
The 'output' link in both of the above points at http://khashier.com/tx/24acea41498b2f387ff751997f8dc90a4551f7b638e7cfc00750a9966f3bc018#o465
The output in question staked twice, but one of them was orphaned.
The bug is more obvious here, where the fee is off by around 2000 CLAMs:
http://khashier.com/tx/3423091ad3e4dbb17822baeeae773dc9741995a3db9b9ce77732167f8e9a1913
I just noticed it again:
http://khashier.com/tx/f1cad229a6fbe5765e854ac194f892049c9af8528d93d7c64145326a9d27b22d
|
gharchive/issue
| 2015-01-12T07:45:39 |
2025-04-01T04:35:13.990972
|
{
"authors": [
"dooglus"
],
"repo": "nochowderforyou/clams",
"url": "https://github.com/nochowderforyou/clams/issues/130",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1838270578
|
[SDXL] Add SDXL pipeline to SHARK
-- This commit adds SDXL pipeline to SHARK.
Signed-off-by: Abhishek Varma abhishek@nod-labs.com
Unrelated CI failure of arith.maxf issue. CC: @monorimet
Merging for now, CI issue is just stale artifacts on one of the runners.
|
gharchive/pull-request
| 2023-08-06T16:37:06 |
2025-04-01T04:35:14.005804
|
{
"authors": [
"Abhishek-Varma",
"monorimet"
],
"repo": "nod-ai/SHARK",
"url": "https://github.com/nod-ai/SHARK/pull/1731",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
427708324
|
Payload from HttpNode changed in 0.20 when posting a FormData
What are the steps to reproduce?
I have a Angular app that POST a manually build FormData to a HttpNode in NodeRed with a file as Base64 encoded. Something like this.
uploadImaged(dadosImaged: UploadImaged): Observable<HttpEvent<Imaged>> {
const url = `${environment.NODE_RED_URL}/uploadImaged`;
let body = new FormData();
body.set('codTipoDoc', `${dadosImaged.codTipoDoc}`);
body.set('descricao', dadosImaged.descricao);
body.set('file', dadosImaged.file);
body.set('filename', dadosImaged.filename);
body.set('indices', JSON.stringify(dadosImaged.indices));
const req = new HttpRequest('POST', url, body, { reportProgress: true });
return this.http.request<any>(req).pipe(catchError(this.errorHandler));
}
What happens?
Until version 0.19.X the PAYLOAD of HttpNode was show as a buffer or a raw string in debug node.
In version 0.20.X when POSTing to HttpNode an exception is raised. TypeError: invalid media type.
Activating the flag Accept file uploads? the payload is parsed but I need the buffer version to make an HttpRequest to another backend.
Example:
Evidence Image/imgur
What do you expect to happen?
Payload should be a buffer formdata like in version 0.19
Please tell us about your environment:
[x] Node-RED version:
0.20.X
[x] node.js version:
8.15.1
[x] npm version:
6.4.1
[x] Platform/OS:
Linux Mint
[x] Browser:
Any
Fixed in master - will be in 0.20.4 this week
|
gharchive/issue
| 2019-04-01T13:33:50 |
2025-04-01T04:35:14.024478
|
{
"authors": [
"knolleary",
"marcioluis"
],
"repo": "node-red/node-red",
"url": "https://github.com/node-red/node-red/issues/2122",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
359616940
|
Error code: 48001, message: api unauthorized
请问以下是什么问题?我该如何解决?
“Error code: 48001, message: api unauthorized”
若还需要什么详情资料,请通知我。谢谢!
The WeChat API Call encountered the following exception:
Traceback (most recent call last):
File "C:\STB\STAN_MG_API\WeChat\wc_follower_total.py", line 30, in extract
data = client.user.get_followers()
File "C:\STB\STAN_MG_API\WeChat\wechatpy\client\api\user.py", line 60, in get_followers
params=params
File "C:\STB\STAN_MG_API\WeChat\wechatpy\client\api\base.py", line 13, in _get
return self._client.get(url, **kwargs)
File "C:\STB\STAN_MG_API\WeChat\wechatpy\client\base.py", line 182, in get
**kwargs
File "C:\STB\STAN_MG_API\WeChat\wechatpy\client\base.py", line 115, in _request
res, method, url, result_processor, **kwargs
File "C:\STB\STAN_MG_API\WeChat\wechatpy\client\base.py", line 173, in _handle_result
response=res
wechatpy.exceptions.WeChatClientException: Error code: 48001, message: api unauthorized hint: [FNYrYa0206k455!]
The repo just used for Node.js users.
|
gharchive/issue
| 2018-09-12T19:19:19 |
2025-04-01T04:35:14.031871
|
{
"authors": [
"JacksonTian",
"STANAPI"
],
"repo": "node-webot/wechat-api",
"url": "https://github.com/node-webot/wechat-api/issues/253",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
146541005
|
Demo doesn't work
No icons, no fonts, no anything:
Try later, or another demo instance, it will refresh after 1 hour.
|
gharchive/issue
| 2016-04-07T08:07:33 |
2025-04-01T04:35:14.046929
|
{
"authors": [
"matusszabo",
"wzup"
],
"repo": "nodee-apps/cms",
"url": "https://github.com/nodee-apps/cms/issues/7",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
109114356
|
Add wrapped object factory test
Similar to the example at: https://nodejs.org/api/addons.html#addons_factory_of_wrapped_objects
Thanks, looks good apart from future-proofing.
I pushed some changes but please do not merge it yet! It fails when I try to run it on node 0.10.40, will investigate. Closing for now, maybe will raise a new PR.
The problems should now be fixed, I tried it on Node 0.10.40 and 4.1.1 (was OK on Node 0.12.7 and 4.0.0 in the past). I added another change where the wrapped object can return another wrapped object (I need this functionality for my own project), and rebased on the latest master changes.
Looks good. Test failure on travis seems to be a network error downloading something.
Travis aargh @#$%
Thanks for looking at this.
I pushed a couple more changes:
Small fix to test/js/wrappedobjectfactory-test.js:
--- a/test/js/wrappedobjectfactory-test.js
+++ b/test/js/wrappedobjectfactory-test.js
@@ -10,7 +10,7 @@ const test = require('tap').test
, testRoot = require('path').resolve(__dirname, '..')
, bindings = require('bindings')({ module_root: testRoot, bindings: 'wrappedobjectfactory' });
-test('objectwraphandle', function (t) {
+test('wrappedobjectfactory', function (t) {
t.plan(4);
var obj = bindings.newFactoryObjectInstance(123);
Also fix test/cpp/objectwraphandle.cpp to move constructor object into a private static inline function
Merged as 4392aa9df745960d553a47e61ae58d8bb811753c bff133ab771eef56315bb2f498a0d95777246f1f 3a4a0ae15bcf35e177d35087063146151c56449f
Thanks for merging my changes!
|
gharchive/pull-request
| 2015-09-30T15:18:11 |
2025-04-01T04:35:14.202984
|
{
"authors": [
"brodybits",
"kkoopa"
],
"repo": "nodejs/nan",
"url": "https://github.com/nodejs/nan/pull/478",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
624685008
|
gyp: fix XCode CLT version detection on Catalina
Since Catalina XCode Command Line Tools don't show up among installed packages,
which breaks version detection in gyp/pylib/gyp/xcode_emulation.py.
The workaround is to remove the CLT installation directory, and run
xcode-select --install which reinstalls an older version of CLT. This older
version will eventually be upgraded by Software Update, which breaks gyp again.
Checklist
[x] commit message follows commit guidelines
Description of change
This patch adds a fallback mechanism to detect newer CLT versions based on
software update history.
After every CLT upgrade on 10.15, I have to reinstall Command Line Tools.
Please merge and release this, soon.
Since we vendor in gyp from https://github.com/nodejs/gyp-next this should really land there.
Can we delete the code from this repository that we vendor in from some other repository?
we could use git submodules but they're such a developer-hostile solution that it's usually better to just copy the code
I think we just need to improve our documentation and get more pro-active with pointing people over there when they open a PR here. The code's the same so moving the PR shouldn't be too difficult since you can just copy your edited files.
In times of Github Actions, a bot would help here. If changes are in specific paths, just post a comment about the situation.
Unfortunately GitHub Actions running on pull requests from forks do not have permissions to post comments.
Unless you are using a own personal access token.
|
gharchive/pull-request
| 2020-05-26T08:09:12 |
2025-04-01T04:35:14.212190
|
{
"authors": [
"cclauss",
"jkroepke",
"richardlau",
"rvagg",
"szekelyisz"
],
"repo": "nodejs/node-gyp",
"url": "https://github.com/nodejs/node-gyp/pull/2141",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1167149020
|
Test Issue created at Sat, 12 Mar 2022 02:23:47 GMT
Test issue body Sat, 12 Mar 2022 02:23:47 GMT
Comment on issue at Sat, 12 Mar 2022 02:23:49 GMT
|
gharchive/issue
| 2022-03-12T02:23:48 |
2025-04-01T04:35:14.487029
|
{
"authors": [
"nodemationqa"
],
"repo": "nodemationqa/nodeQA",
"url": "https://github.com/nodemationqa/nodeQA/issues/570",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
810518559
|
feat(enricher): Adds the rh-metering flag for adding Red Hat Runtimes metering to your application.
The new flag is --rh-metering. can also be used as --rh-metering.component-version=NODE_VERSION
This is currently an undocumented feature since it is more for those using Red Hat Runtimes build of Node.js
Coverage decreased (-0.3%) to 98.299% when pulling 30d26e79c049dd1aa4e54f0824f3d4fc70972c5f on lholmquist:NODE-880-meterting-metadata into 089809e0fe7c8b95be1c6334175dd87b937075b6 on nodeshift:master.
|
gharchive/pull-request
| 2021-02-17T20:35:29 |
2025-04-01T04:35:14.495465
|
{
"authors": [
"coveralls",
"lholmquist"
],
"repo": "nodeshift/nodeshift",
"url": "https://github.com/nodeshift/nodeshift/pull/543",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
857834213
|
1.2.10: Startup Crash @[java.lang.VerifyError: class mythicbotany.data.BlockTagProvider overrides final method]
Hi.
After upgrading to 1.2.10 I get the following crash at startup:
java.lang.VerifyError: class mythicbotany.data.BlockTagProvider overrides final method io.github.noeppi_noeppi.libx.data.provider.BlockTagProviderBase.func_200432_c()V
The full crash report can be found here: https://pastebin.com/pSZiay8g
I have not changed anything else. If there is any further information you need to help track this down, please do let me know.
Cheers.
You should not use LibX alpha builds to play the game.
Will be fixed soon. For now just downgrade LibX to the latest release version.
You should not use LibX alpha builds to play the game.
I do so mainly for testing purposes as it often helps to pick up bugs. I do not use them in my master pack, just in the testing branch.
|
gharchive/issue
| 2021-04-14T12:09:54 |
2025-04-01T04:35:14.498700
|
{
"authors": [
"noeppi-noeppi",
"sciguyryan"
],
"repo": "noeppi-noeppi/MythicBotany",
"url": "https://github.com/noeppi-noeppi/MythicBotany/issues/31",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2450932013
|
refactor views to auto-generate permission_required
ref app.test.abstract.views.AllViews
Reason: Reduce what is required to create a view. !43
@pytest.mark.skip(reason='write test')
def test_view_index_attribute_missing_permission_required(self):
""" Attribute missing Test
Ensure that `permission_required` attribute is not defined within the view.
this can be done by mocking the inherited class with the `permission_required` attribute
set to a value that if it changed would be considered defined in the created view.
## Why?
This attribute can be dynamically added based of of the view name along with attributes
`model._meta.model_name` and `str(__class__.__name__).lower()`.
Additional test:
- ensure that the attribute does get automagically created.
- ensure that the classes name is one of add, change, delete, display or index.
"""
Requirements
[ ] Remove permission_required attribute from all views
[ ] ensure class names for views are correctly set.
added 5m of time spent
marked this issue as related to #74
marked this issue as related to #15
|
gharchive/issue
| 2024-07-20T13:28:29 |
2025-04-01T04:35:14.503726
|
{
"authors": [
"jon-nfc"
],
"repo": "nofusscomputing/centurion_erp",
"url": "https://github.com/nofusscomputing/centurion_erp/issues/144",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1682061457
|
chore: Use pollster as async executor
This drops the direct tokio dependency (it still exists inside reqwest) and uses pollster for block_on. Their package readme has a good "why" for itself.
I was trying to understand if we can block_on when compiling to JS with wasm-bindgen and found this comment: https://github.com/async-rs/async-std/issues/913#issuecomment-1296282060 which recommended pollster. I tried it in a demo wasm-pack project I have an it seems to block on an async function, so I think we can layer JS promises and it should work 🤞
Converting to draft because it seems that Condvar::wait throws when interop'ing with JS promises.
Learned a lot here but closing this. We might use pollster to block_on async functions in something like nargo_cli.
|
gharchive/pull-request
| 2023-04-24T21:20:00 |
2025-04-01T04:35:14.506340
|
{
"authors": [
"phated"
],
"repo": "noir-lang/aztec_backend",
"url": "https://github.com/noir-lang/aztec_backend/pull/135",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2706128704
|
chore: test against noir 1.0.0-beta.0
Description
Problem*
Resolves
Summary*
We're looking to flip 1.0.0-beta.0 to an official release so we should test compatibility with it.
Additional Context
PR Checklist*
[x] I have tested the changes locally.
[x] I have formatted the changes with Prettier and/or cargo fmt on default settings.
i hadn't seen this pr, however, since we're doing a few changes in parallel to this (documentation, fixing one of my mistakes, etcetc), this change will be handled in another pr (#74), so i'm closing this for simplicity.
|
gharchive/pull-request
| 2024-11-29T20:42:56 |
2025-04-01T04:35:14.509003
|
{
"authors": [
"TomAFrench",
"jtriley-eth"
],
"repo": "noir-lang/noir-bignum",
"url": "https://github.com/noir-lang/noir-bignum/pull/63",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1687471085
|
chore: Add AI codereviewer
Related issue(s)
Resolves #
Description
This adds codeball as a PR reviewer
Summary of changes
Dependency additions / changes
Test additions / changes
Checklist
[ ] I have tested the changes locally.
[ ] I have formatted the changes with Prettier and/or cargo fmt with default settings.
[ ] I have linked this PR to the issue(s) that it resolves.
[ ] I have reviewed the changes on GitHub, line by line.
[ ] I have ensured all changes are covered in the description.
Documentation needs
[ ] This PR requires documentation updates when merged.
Additional context
Okay thought this was gonna be a nice version of codereview.gpt -- it only applies a label :(
|
gharchive/pull-request
| 2023-04-27T19:59:55 |
2025-04-01T04:35:14.513200
|
{
"authors": [
"kevaundray"
],
"repo": "noir-lang/noir",
"url": "https://github.com/noir-lang/noir/pull/1242",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1701671909
|
chore: fix installation link in readme
Related issue(s)
Fixes https://github.com/noir-lang/noir/issues/1089#issuecomment-1539010915 (again)
Description
Summary of changes
Dependency additions / changes
Test additions / changes
Checklist
[x] I have tested the changes locally.
[x] I have formatted the changes with Prettier and/or cargo fmt with default settings.
[x] I have linked this PR to the issue(s) that it resolves.
[x] I have reviewed the changes on GitHub, line by line.
[x] I have ensured all changes are covered in the description.
Documentation needs
[ ] This PR requires documentation updates when merged.
Additional context
(I think you'll need approval from someone on the Noir team :) )
(I think you'll need approval from someone on the Noir team :) )
Ah wasn't aware you guys couldn't as we can approve in Aztec org.
|
gharchive/pull-request
| 2023-05-09T09:09:40 |
2025-04-01T04:35:14.517859
|
{
"authors": [
"TomAFrench",
"iAmMichaelConnor"
],
"repo": "noir-lang/noir",
"url": "https://github.com/noir-lang/noir/pull/1326",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1308959606
|
スマホや iPad でシーンの下部が画面外にはみ出る
予期しない動作の内容 | Describe the unexpected behavior
スマホや iPad でシーンの下部 2〜5 % ほどが画面外にはみ出る
再現方法 | To Reproduce
発生環境(モバイルの場合) | Smartphone (please complete the following information):
デバイス | Device: Android, iPad
OS: Android, iOS
バージョン | Version: -
ブラウザ (Web 版) | Browser: -
備考 | Additional context
Reported by @Reputeless
Commented by @su8ru on Siv3D Slack
#app の height: 100vh 指定が原因かもしれません。
モバイルでの 100vh 指定はいろいろ問題を抱えており、js で window.innerHeight を参照する回避が一般的だと思います。
Commented by @nokotan on Siv3D Slack
該当箇所確認できました
Siv3D プロジェクトテンプレートに付属している html テンプレートに以下の処理を追加してみています
appContainer.style.height = `${window.innerHeight}px`;
|
gharchive/issue
| 2022-07-19T04:25:23 |
2025-04-01T04:35:14.530052
|
{
"authors": [
"nokotan"
],
"repo": "nokotan/OpenSiv3D",
"url": "https://github.com/nokotan/OpenSiv3D/issues/21",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
328708253
|
Modernize codebase
Drop Bower support
Add ES module support
Use StandardJS
Update dependencies
Also add tests to get us to 100% code coverage
|
gharchive/pull-request
| 2018-06-02T03:15:23 |
2025-04-01T04:35:14.533011
|
{
"authors": [
"nolanlawson"
],
"repo": "nolanlawson/chord-magic",
"url": "https://github.com/nolanlawson/chord-magic/pull/5",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1148548517
|
Popping behavior
Check that this is really a bug
[X] I confirm
Reproduction link
https://atroposjs.com/
Bug description
When a multi-component Atropos collection with a common eventsEl is moving, some of the elements will "pop" instead of smoothly animating. This manifests differently in different browsers. On Chrome the popping is with the scale of the element. On Safari the popping seems to be with the glare.
https://user-images.githubusercontent.com/157106/155403478-689bf63c-f65c-4967-ab44-2102c40dd683.mov
https://user-images.githubusercontent.com/157106/155403501-544484ec-9116-43c8-9834-6dd77706e8d9.mov
Expected Behavior
No response
Actual Behavior
No response
Atropos version
1.0.2
Platform/Target and Browser Versions
macOS: Chrome, Safari
Validations
[X] Follow our Code of Conduct
[X] Read the docs.
[X] Check that there isn't already an issue that request the same feature to avoid creating a duplicate.
[X] Make sure this is a Atropos issue and not a framework-specific issue
Would you like to open a PR for this bug?
[ ] I'm willing to open a PR
Is there maybe already a solution or possible quickfix for this ?
|
gharchive/issue
| 2022-02-23T20:33:14 |
2025-04-01T04:35:14.538300
|
{
"authors": [
"kylemcdonald",
"triiickstaa"
],
"repo": "nolimits4web/atropos",
"url": "https://github.com/nolimits4web/atropos/issues/19",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
755151109
|
Navigation Arrows not showing on IE
This is a (multiple allowed):
[x] bug
[ ] enhancement
[ ] feature-discussion (RFC)
Swiper Version: EXACT RELEASE VERSION OR COMMIT HASH, HERE.
Platform/Target and Browser Versions: PLATFORM CLIENT YOU ARE TARGETING SUCH AS macOS, Windows, CORDOVA, IOS, ANDROID, CHROME, ETC.
Live Link or JSFiddle/Codepen or website with isssue: PREFERABLY (IF YOU WANT YOUR ISSUE TO BE RESOLVED ASAP).
What you did
I couldn't see the arrows.
Expected Behavior
See Arrows
Actual Behavior
Arrows only visible on Firefox and Chrome
P.S. Remember, an issue is not the place to ask questions. You can use Stack Overflow
for that.
Before you open an issue, please check if a similar issue already exists or has been closed before.
IE is not supported by Swiper
|
gharchive/issue
| 2020-12-02T10:35:32 |
2025-04-01T04:35:14.542874
|
{
"authors": [
"EchoChen012",
"nolimits4web"
],
"repo": "nolimits4web/swiper",
"url": "https://github.com/nolimits4web/swiper/issues/3996",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2302203413
|
Closing a modal changes active slide and breaks loop mode
Check that this is really a bug
[X] I confirm
Reproduction link
https://codesandbox.io/p/sandbox/swiper-element-boostrap-modal-x5535x
Bug description
I have a swiper that shows multiple slides per view, and when clicking the anchor inside the slide, a boostrap modal is triggered. However, when clicking a slide that is not the central one and closing the modal, the active slide is changed to the one that triggered the modal. Also, if it is the last one, loop mode completely breaks and no next slide is shown.
The boostrap modal is not inside the swiper-container.
Here's a demo video from the codesandbox I provided:
screen-capture.webm
Expected Behavior
Closing a modal should not change the current active slide.
Actual Behavior
No response
Swiper version
11.1.3
Platform/Target and Browser Versions
Windows 10, Chrome 124
Validations
[X] Follow our Code of Conduct
[X] Read the docs.
[X] Check that there isn't already an issue that request the same feature to avoid creating a duplicate.
[X] Make sure this is a Swiper issue and not a framework-specific issue
Would you like to open a PR for this bug?
[ ] I'm willing to open a PR
Thanks for the quick fix!! Any estimated date for the release?
|
gharchive/issue
| 2024-05-17T09:09:51 |
2025-04-01T04:35:14.548329
|
{
"authors": [
"mdmontesinos"
],
"repo": "nolimits4web/swiper",
"url": "https://github.com/nolimits4web/swiper/issues/7540",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2506913541
|
TypeError when Swiper instance is destroyed right after zooming
Check that this is really a bug
[X] I confirm
Reproduction link
https://codesandbox.io/p/devbox/swiper-react-zoom-pqy2gq
Bug description
On a project we have an image gallery build on top of Swiper. We have the zoom functionality enabled. On mobile we have a functionality in place that checks whether or not you have swiped the image in the vertical direction (gallery set to horizontal mode) and we close the gallery when this happens. However, both things - zooming + swiping (onTouchStart - onTouchEnd) - can happen at the same time. This means there are cases in which the swiper gallery is closed right after the zoom happens.
During zooming, there is a setTimeout call with callback that invokes the allowTouchMove
https://github.com/nolimits4web/swiper/blob/d0b6abd74805398ac3cbf41aeebf141a805ec64b/src/modules/zoom/zoom.mjs#L249
However, the asynchronous nature of setTimeout in our use case makes it happen after the instance is destroyed. Which leads to TypeErrors:
../../node_modules/swiper/modules/zoom.mjs in allowTouchMove at line 216:1
TypeError
Cannot set properties of undefined (setting 'preventTouchMoveFromPointerMove')
Reproduction link note: Our setup to reproduce this issue is quite complex and unreliable, but I think the issue is hopefully described/obvious enough. If not, I can try to setup a proper repro sandbox.
Expected Behavior
No errors are thrown after the Swiper instance is destroyed right after zooming.
Actual Behavior
Unreliably the error is thrown after the Swiper instance is destroyed right after zooming.
Swiper version
11.1.4
Platform/Target and Browser Versions
Android Chrome 128
Validations
[X] Follow our Code of Conduct
[X] Read the docs.
[X] Check that there isn't already an issue that request the same feature to avoid creating a duplicate.
[X] Make sure this is a Swiper issue and not a framework-specific issue
Would you like to open a PR for this bug?
[X] I'm willing to open a PR
Fixed by https://github.com/nolimits4web/swiper/pull/7723
|
gharchive/issue
| 2024-09-05T06:45:28 |
2025-04-01T04:35:14.554807
|
{
"authors": [
"Heptonius",
"nolimits4web"
],
"repo": "nolimits4web/swiper",
"url": "https://github.com/nolimits4web/swiper/issues/7722",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2038949947
|
🛑 f.cz is down
In ab5e208, f.cz (https://f.cz/nodeinfo/2.0) was down:
HTTP code: 0
Response time: 0 ms
Resolved: f.cz is back up in 371375e after 5 minutes.
|
gharchive/issue
| 2023-12-13T04:50:42 |
2025-04-01T04:35:14.558007
|
{
"authors": [
"matejdivecky"
],
"repo": "nolog-it/mastodon-uptime",
"url": "https://github.com/nolog-it/mastodon-uptime/issues/447",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
103137290
|
change julia REQUIRE to 0.3
since Colors has no 0.2-compatible tagged versions at this time
ref https://github.com/nolta/Winston.jl/pull/235/files#r37931603
Also noticed in #237. But it would be fine to merge this separately.
Ah, whoops. Saw that you had done the same thing for ProfileView so held off there, but missed this one. Not too critical I think if bigger changes are in the works.
|
gharchive/pull-request
| 2015-08-25T23:22:19 |
2025-04-01T04:35:14.559548
|
{
"authors": [
"timholy",
"tkelman"
],
"repo": "nolta/Winston.jl",
"url": "https://github.com/nolta/Winston.jl/pull/238",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
63718497
|
Install takes ages
Hi,
Is there any way to optimize this tool? We use it on Travis-CI and install takes about 160 seconds, which is a while for single tool.
154.00s
$ gem install cupertino --no-ri --no-rdoc
Fetching: highline-1.6.21.gem (100%)
Successfully installed highline-1.6.21
Fetching: commander-4.2.1.gem (100%)
Successfully installed commander-4.2.1
Fetching: terminal-table-1.4.5.gem (100%)
Successfully installed terminal-table-1.4.5
Fetching: term-ansicolor-1.0.7.gem (100%)
Successfully installed term-ansicolor-1.0.7
Fetching: net-http-digest_auth-1.4.gem (100%)
Successfully installed net-http-digest_auth-1.4
Fetching: net-http-persistent-2.9.4.gem (100%)
Successfully installed net-http-persistent-2.9.4
Fetching: mime-types-1.25.1.gem (100%)
Successfully installed mime-types-1.25.1
Fetching: mini_portile-0.6.2.gem (100%)
Successfully installed mini_portile-0.6.2
Fetching: nokogiri-1.6.6.2.gem (100%)
Building native extensions. This could take a while...
Successfully installed nokogiri-1.6.6.2
Fetching: ntlm-http-0.1.1.gem (100%)
Successfully installed ntlm-http-0.1.1
Fetching: webrobots-0.1.1.gem (100%)
Successfully installed webrobots-0.1.1
Fetching: unf_ext-0.0.6.gem (100%)
Building native extensions. This could take a while...
Successfully installed unf_ext-0.0.6
Fetching: unf-0.1.4.gem (100%)
Successfully installed unf-0.1.4
Fetching: domain_name-0.5.23.gem (100%)
Successfully installed domain_name-0.5.23
Fetching: mechanize-2.5.1.gem (100%)
Successfully installed mechanize-2.5.1
Fetching: security-0.1.3.gem (100%)
Successfully installed security-0.1.3
Fetching: certified-1.0.0.gem (100%)
IMPORTANT: Remember to use the included executable `certifed-update` regularly to keep your certificate bundle up to date.
Successfully installed certified-1.0.0
Fetching: cupertino-1.3.0.gem (100%)
Successfully installed cupertino-1.3.0
18 gems installed
Most of the time was taken by "Building native extensions. This could take a while...".
That's just the price of doing business in Ruby. I'd entertain any PRs looking to speed things up, but this is not a priority.
|
gharchive/issue
| 2015-03-23T13:02:37 |
2025-04-01T04:35:14.571068
|
{
"authors": [
"AndrewDryga",
"mattt"
],
"repo": "nomad/cupertino",
"url": "https://github.com/nomad/cupertino/issues/194",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
53592878
|
build error for Xcode6.1.1
I got a following issue..
xcodebuild MyApp.xcworkspace
2015-01-07 11:52:13.902 xcodebuild[3989:42121] [MT] DVTAssertions: Warning in /SourceCache/IDEFrameworks/IDEFrameworks-6611/IDEFoundation/SourceControl/Model/IDESourceControlManager.m:423
Details: Error Domain=com.apple.dt.IDESourceControlErrorDomain Code=-1 "Missing extension: public.vcs.subversion" UserInfo=0x7fd084558ad0 {NSLocalizedDescription=Missing extension: public.vcs.subversion}
Object: <IDESourceControlManager: 0x7fd0840d7e40>
Method: -loadRepositories
Thread: <NSThread: 0x7fd082c35910>{number = 1, name = main}
Please file a bug at http://bugreport.apple.com with this warning message and any useful information you can provide.
** BUILD FAILED **
The following build commands failed:
Check dependencies
(1 failure)
** ARCHIVE FAILED **
sorry.. duplicated #176
|
gharchive/issue
| 2015-01-07T03:15:41 |
2025-04-01T04:35:14.574473
|
{
"authors": [
"tikidunpon"
],
"repo": "nomad/shenzhen",
"url": "https://github.com/nomad/shenzhen/issues/177",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1079277805
|
请问正常使用中突然报错重开显示TUNNEL - music.163.com:443
如图
同问,今天上午正常用,中午也遇到这情况了
同问 现在无法正常播放
https://github.com/nondanee/UnblockNeteaseMusic/issues/85 按照这个可以解决
#85按照这个可以解决
你好,linux客户端生成证书替换后,还是报错
TUNNEL - music.163.com:443
TUNNEL - music.163.com:443
TUNNEL - music.163.com:443
MITM > clientlog.music.163.com
#85按照这个可以解决
你好,linux客户端生成证书替换后,还是报错
TUNNEL - music.163.com:443 TUNNEL - music.163.com:443 TUNNEL - music.163.com:443 MITM > clientlog.music.163.com
添加HTTPS代理端口了吗
#85 按照这个可以解决
你好,linux客户端生成证书替换后,还是报错
TUNNEL - music.163.com:443 TUNNEL - music.163.com:443 TUNNEL - music.163.com:443 MITM > clientlog.music.163.com
添加HTTPS代理端口号码
没有,之前我是源代码部署的 只开启了一个http端口
node app.js -p 33123
#85 按照这个可以解决
你好,linux客户端生成证书替换后,还是报错
TUNNEL - music.163.com:443 TUNNEL - music.163.com:443 TUNNEL - music.163.com:443 MITM > clientlog.music.163.com
添加HTTPS代理端口号码
没有,之前我是源代码部署的 只开启了一个http端口 node app.js -p 33123
添加HTTPS代理试试 node app.js -p 33123:33124 这样的
#85 按照这个可以解决
你好,linux客户端生成证书替换后,还是报错
TUNNEL - music.163.com:443 TUNNEL - music.163.com:443 TUNNEL - music.163.com:443 MITM > clientlog.music.163.com
添加HTTPS代理端口号码
没有,之前我是源代码部署的 只开启了一个http端口 node app.js -p 33123
添加HTTPS代理试试 node app.js -p 33123:33124 这样的
请问一下,两个端口的话 网易代理配置 怎么设置呢
网易客户端只有 服务器+端口
#85 按照这个可以解决
你好,linux客户端生成证书替换后,还是报错
TUNNEL - music.163.com:443 TUNNEL - music.163.com:443 TUNNEL - music.163.com:443 MITM > clientlog.music.163.com
添加HTTPS代理端口号码
没有,之前我是源代码部署的 只开启了一个http端口 node app.js -p 33123
添加HTTPS代理试试 node app.js -p 33123:33124 这样的
请问一下,两个端口的话 网易代理配置 怎么设置呢
网易客户端只有 服务器+端口
网易客户端还按原来的配置就行
网易客户端还按原来的配置就行
直接走http的端口33123吗 不用配置https 33124?
网易客户端还按原来的配置就行
直接走http的端口33123吗 不用配置https 33124?
对
#85 按照这个可以解决
你好,linux 客户端生成证书替换后,还是报错
TUNNEL - music.163.com:443 TUNNEL - music.163.com:443 TUNNEL - music.163.com:443 MITM > clientlog.music.163.com
添加 HTTPS 代理端口号码
没有,之前我是源代码部署的 只开启了一个 http 端口 node app.js -p 33123
添加 HTTPS 代理试试 node app.js -p 33123:33124 这样的
请问一下,node用这个代码可以使用,但pm2为什么就不可以(我需要让它在window后台运行)
代码:
pm2 start app.js -p 8080:8081
Log:
address: '::',
0|app | port: 8080
0|app | }
0|app | Error: listen EADDRINUSE: address already in use :::8080
0|app | at Server.setupListenHandle [as _listen2] (net.js:1318:16)
0|app | at listenInCluster (net.js:1366:12)
0|app | at Server.listen (net.js:1452:7)
0|app | at D:\Program\UnblockNeteaseMusic-0.25.3\src\app.js:83:27
0|app | at processTicksAndRejections (internal/process/task_queues.js:93:5) {
0|app | code: 'EADDRINUSE',
0|app | errno: -4091,
0|app | syscall: 'listen',
0|app | address: '::',
0|app | port: 8080
0|app | }
我的今天不能用了,都显示网络不给力哦,请检查你的网络设置,关键是我docker和windows里面运行的都是这个错误了。估计是接口改了还是怎么地
问题已解决,感谢大佬们
网易客户端还原来的配置就行
直接走http的端口33123吗不用配置https 33124?
对
3Q,按照您的方法,已解决!!!
#85 按照这个可以解决
你好,linux客户端生成证书替换后,还是报错
TUNNEL - music.163.com:443 TUNNEL - music.163.com:443 TUNNEL - music.163.com:443 MITM > clientlog.music.163.com
添加HTTPS代理端口号码
没有,之前我是源代码部署的 只开启了一个http端口 node app.js -p 33123
添加HTTPS代理试试 node app.js -p 33123:33124 这样的
现在不行,网络连接错误,还有什么办法吗
|
gharchive/issue
| 2021-12-14T03:50:39 |
2025-04-01T04:35:14.612395
|
{
"authors": [
"ClankySun10936",
"GenBUPT",
"TDSofTM",
"XianyuOne",
"cg-llc",
"jingouwangzi",
"kk748548942",
"silveryvoid"
],
"repo": "nondanee/UnblockNeteaseMusic",
"url": "https://github.com/nondanee/UnblockNeteaseMusic/issues/795",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1852622783
|
Encoding error when writing proposed_changes.txt file
The bug
When 'iambic plan' command is run on updated yaml file(s) the changes should be written to a proposed_changes.txt file. But 'UnicodeEncodeError' exception occurs. Error Traceback is added at the end for reference.
To Reproduce
Run 'iambic plan' command on updated yaml file(s)
Expected behavior
Should write proposed_changes.txt file without any error
Desktop
OS: Windows 10
Version 19045.3324
Python 3.10.11
Error Traceback
2023/08/16 11:54:54 [error ] Unhandled Exception
error=UnicodeEncodeError('charmap', 'IAMbic Summary\r\n\r\nChange Detection\r\n\r\n* 1 Update actions.\r\n* 1 templates with changes.\r\n* 1 accounts affected.\r\n\r\nIAMbic Change Details\r\nUpdate\r\n└── resources/aws/iam/ma
naged_policy/redacted redacted/ssm-parameter-store-access.yaml\r\n └── redacted redacted - (redacted)\r\n └── aws:policy_document // arn:aws:iam::redacted:policy/ssm-parameter-store-access\r\n └── policy
_document:\r\n -iterable_item_removed\r\n {"root['Statement'][0]['Action'][1]": 'ssm:DeleteParameter', "root['Statement'][0]['Action'][5]": 'ssm:DeleteParameters'}\r\n', 145, 148, 'chara
cter maps to ')
exception=UnicodeEncodeError
stacktrace= File "", line 198, in _run_module_as_main
File "", line 88, in run_code
File "redacted\venv\Scripts\iambic.exe_main.py", line 7, in
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "redacted\click\core.py", line 783, in invoke
return __callback(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "redacted\iambic\main.py", line 360, in plan
run_plan(templates, repo_dir=repo_dir)
File "redacted\iambic\main.py", line 400, in run_plan
output_proposed_changes(template_changes)
File "redacted\iambic\main.py", line 54, in output_proposed_changes
file_render_resource_changes(output_path, template_changes)
File "redacted\iambic\output\text.py", line 25, in file_render_resource_changes
f.write(rendered_data)
File "redacted\Python\Python311\Lib\encodings\cp1252.py", line 19, in encode
return codecs.charmap_encode(input,self.errors,encoding_table)[0]
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
Pull Request
#576
Fix is included in iambic-core 0.11.48, available in PyPi
|
gharchive/issue
| 2023-08-16T06:51:58 |
2025-04-01T04:35:14.643247
|
{
"authors": [
"0xAJX",
"smoy"
],
"repo": "noqdev/iambic",
"url": "https://github.com/noqdev/iambic/issues/575",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
93614414
|
Add null-check when updating or destroying an instance.
Hey Jun,
(not sure if the best practice was to do another pull request) but here it is. So I've found the issue is with the scroll bar being a direct child of an angular ui-router element. I have added in the console.error() as you suggested. But using console isn't supported by all browsers, (IE) especially, and not sure it's good to use here. I'd suggest removing it, imo.
Thanks,
Adam
As you suggested, I just remove the line throwing an error and make it return early. Pls refer to 6c642d8a478ce231aad98aee7f6cd541cac816d2.
Thanks!
|
gharchive/pull-request
| 2015-07-07T19:54:36 |
2025-04-01T04:35:14.650890
|
{
"authors": [
"adammiles10",
"noraesae"
],
"repo": "noraesae/perfect-scrollbar",
"url": "https://github.com/noraesae/perfect-scrollbar/pull/359",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1466604244
|
feat: configure GoRelease to publish to homebrew
Scope:
configured GoReleaser to publish to homebrew
Test:
I didn't want to push a new tag and trigger a regular release so I could only test my changes by doing a local-only release with the command:
goreleaser release --snapshot --rm-dist
this generated dist/fman.rb formula file that would be published to homebrew-tab repo
Thanks
|
gharchive/pull-request
| 2022-11-28T15:15:08 |
2025-04-01T04:35:14.654576
|
{
"authors": [
"nore-dev",
"tomekz"
],
"repo": "nore-dev/fman",
"url": "https://github.com/nore-dev/fman/pull/61",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2345930314
|
boards: norik: Add support for Norik Octopus SoM & IO-Board
Add support for Norik Octopus SoM & IO-Board based on nRF9160 SiP.
Changed Norik systems d.o.o. to Norik Systems as requested.
@FPlohl you still have one rename left in "boards/norik/octopus_som/board.cmake" to be consistent.
@pfiser fixed remaining inconsistencies. Clear to open PR to main branch?
yes, @FPlohl go ahead and open PR on upstream zephyr repo. Fingers crossed :)
|
gharchive/pull-request
| 2024-06-11T09:41:39 |
2025-04-01T04:35:14.659264
|
{
"authors": [
"FPlohl",
"pfiser"
],
"repo": "norikdoo/zephyr",
"url": "https://github.com/norikdoo/zephyr/pull/2",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.