id
stringlengths 4
10
| text
stringlengths 4
2.14M
| source
stringclasses 2
values | created
timestamp[s]date 2001-05-16 21:05:09
2025-01-01 03:38:30
| added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| metadata
dict |
---|---|---|---|---|---|
296246467
|
Unable to install due to dependency broadlinkjs-rm@^0.2.4
Hi,
If I run npm install -g homebridge-broadlink-rm / with or without sudo I get
npm ERR! code ETARGET
npm ERR! notarget No matching version found for broadlinkjs-rm@^0.2.4
npm ERR! notarget In most cases you or one of your dependencies are requesting
npm ERR! notarget a package version that doesn't exist.
npm ERR! notarget
npm ERR! notarget It was specified as a dependency of 'homebridge-broadlink-rm'
If I check https://www.npmjs.com/package/broadlinkjs-rm it only shows version 0.2.2 is there some delay, or some command I can run to update the repo so that it can download 0.2.4. Even if I install broadlinkjs-rm directly with npm it pulls 0.2.2
Thanks
Having same issue here
Sorry, published it.
|
gharchive/issue
| 2018-02-12T01:26:41 |
2025-04-01T06:39:27.552761
|
{
"authors": [
"Silentbob101",
"dessmith",
"lprhodes"
],
"repo": "lprhodes/homebridge-broadlink-rm",
"url": "https://github.com/lprhodes/homebridge-broadlink-rm/issues/246",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2176988706
|
Removing alerts
It would be nice to be able to display errors when popups are disabled. What tool are you using to minify your javascript? It would be good to see some documentation for that.
Your pull request seems good. I am busy currently and still need to do a bit of review, but this will hopefully be merged soon. Thank you for noticing this!
@david-dick I've made one change, which is to use var instead of let. The intent is to work on as many browser versions as possible. Regarding the minifier tool, I used https://www.toptal.com/developers/javascript-minifier . Nothing special about that specific tool, it's just the first one that showed up.
Let me know if this is good to commit. Thanks!
works for me!
Hey @david-dick, I appreciate the PR a lot, you're the first person to ever write any PR or issue on any of my repos... thank you!! As for the PR itself, it's been merged. 👍
|
gharchive/pull-request
| 2024-03-09T01:19:55 |
2025-04-01T06:39:27.558648
|
{
"authors": [
"david-dick",
"lraj22"
],
"repo": "lraj22/browserfeatcl",
"url": "https://github.com/lraj22/browserfeatcl/pull/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
368542552
|
Add sha256sum for release file
Please provide SHA256 checksums for your releases
Hey @jjmartres, I am trying to revive this project since the maintainer seems to be gone. I also tried contacting him separately, but no response. If he comes back, I plan to submit any changes I make on my back to this one.
For your particular request, I added checksums with this MR here and release of 0.2.0.
Close due to archiving repository.
Sorry for not presenting so long. I've been working on another project and don't have time to for helm-unittest.
Please consider other working forks like quintush/helm-unittest.
|
gharchive/issue
| 2018-10-10T08:11:18 |
2025-04-01T06:39:27.562769
|
{
"authors": [
"jjmartres",
"lrills",
"walkafwalka"
],
"repo": "lrills/helm-unittest",
"url": "https://github.com/lrills/helm-unittest/issues/60",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
936858008
|
Remove page scroll on save
Is your feature request related to a problem?
When working on modelling exercises, my progress gets periodically saved. The saving is indicated by a green bar on top of the modelling window. When the bar is shown, the page moves down by the width of the bar. This moves the modelling canvas and leads to frustration when connecting elements.
Describe the solution you'd like
The saving could either be indicated by an element that does not increase page space above the modelling window or the space for the save bar is already allocated and the bar is just made visible.
Describe alternatives you've considered
No response
Additional context
This issue has been addressed.
|
gharchive/issue
| 2021-07-05T08:56:47 |
2025-04-01T06:39:27.580186
|
{
"authors": [
"N0W0RK",
"ge65cer"
],
"repo": "ls1intum/Artemis",
"url": "https://github.com/ls1intum/Artemis/issues/3680",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
659631926
|
Fix Assessment of modeling exercises without any submission
Checklist
[x] I tested all changes and all related features with different users (student, tutor, instructor, admin) locally
[x] Client: I added multiple screenshots/screencasts of my UI changes
Motivation and Context
Assessment of empty UML-Models should work
Description
Checked the places in the modeling-assessment component where it is assumed that there must be a model from apollon. In those places I added checks to prevent errors
Steps for Testing
Create an exam with at least 2 modeling exercises
Generate Student Exams and participate
Exercise 1: do not submit any solution for the exercise
Exercise 2: submit a model, then delete every element and submit again (empty submission should be counted)
After the exam has ended, switch to tutor account and go to the assessment of the modeling exercise
do an assessment with open browser console -> no error should be displayed (The error you can see in the screenshot is happening, because there isn't any other submission for this exercise which can be assessed )
Screenshots
I was also able to assess without any errors.
When assessing the exercise for which I did not submit, I am actually informed that no model was found:
(but again, assessing this worked without any issues)
The assessment works without any errors. Each non-submitted exercise turns into a submission, though. Is this intended? (e.g. I did not submit anything for the first exercise, but was still able to assess it as if it were an empty submission)
Yes this is intended at the moment, because when Artemis prepares the exercises it automatically creates an empty submission
Please approve if everything else is working correctly
|
gharchive/pull-request
| 2020-07-17T20:53:46 |
2025-04-01T06:39:27.586570
|
{
"authors": [
"TobiasPr",
"alexmardale",
"krusche"
],
"repo": "ls1intum/Artemis",
"url": "https://github.com/ls1intum/Artemis/pull/1918",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
464946768
|
Feature/hibernate query cache
Checklist
[ ] I tested the changes and all related features on the test server https://artemistest.ase.in.tum.de.
[ ] I documented my source code using the JavaDoc / JSDoc style.
[ ] I added integration test cases for the server (Spring) related to the features
[ ] I added integration test cases for the client (Jest) related to the features
[ ] I added screenshots/screencast of my UI changes
[ ] I translated all the newly inserted strings
Motivation and Context
Description
Steps for Testing
Log in to ArTEMiS
Navigate to Course Administration
...
Screenshots
There is not much going here, so I'll close this to keep the open pull requests small.
Feel free to reopen it, when there is additional progress
|
gharchive/pull-request
| 2019-07-07T12:18:57 |
2025-04-01T06:39:27.590976
|
{
"authors": [
"krusche",
"sleiss"
],
"repo": "ls1intum/Artemis",
"url": "https://github.com/ls1intum/Artemis/pull/625",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1916204675
|
DM-40926: Make the left and right sidebars sticky
Make the content of both sidebar sticky when visible on page and also allow them to independently scroll if longer than the viewport. This is useful for accessing the document outline, for example.
https://github.com/lsst-sqre/technote/assets/349384/51fcf423-c0fd-4a65-bdac-e5d5f868c01d
|
gharchive/pull-request
| 2023-09-27T19:42:14 |
2025-04-01T06:39:27.615123
|
{
"authors": [
"jonathansick"
],
"repo": "lsst-sqre/technote",
"url": "https://github.com/lsst-sqre/technote/pull/18",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1329520657
|
🛑 lsy223622.com is down
In 3fa17c7, lsy223622.com (http://lsy223622.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: lsy223622.com is back up in ccbe5c9.
|
gharchive/issue
| 2022-08-05T06:44:56 |
2025-04-01T06:39:27.620442
|
{
"authors": [
"lsy223622"
],
"repo": "lsy223622/status",
"url": "https://github.com/lsy223622/status/issues/290",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1140363852
|
🛑 Mirror-木生 is down
In 73789d4, Mirror-木生 (https://t.lsy223622.com/ncov-xidian-times) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Mirror-爬 is back up in 7e256c7.
|
gharchive/issue
| 2022-02-16T17:48:06 |
2025-04-01T06:39:27.622917
|
{
"authors": [
"lsy223622"
],
"repo": "lsy223622/xdncov-mirror-status",
"url": "https://github.com/lsy223622/xdncov-mirror-status/issues/1082",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1465270227
|
🛑 Mirror-爬 is down
In 12ea61c, Mirror-爬 (https://ncov.zhouym.tech/) was down:
HTTP code: 404
Response time: 948 ms
Resolved: Mirror-鱼露 is back up in f056bec.
|
gharchive/issue
| 2022-11-26T22:29:38 |
2025-04-01T06:39:27.625432
|
{
"authors": [
"lsy223622"
],
"repo": "lsy223622/xdncov-mirror-status",
"url": "https://github.com/lsy223622/xdncov-mirror-status/issues/11842",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1515050150
|
🛑 Mirror-没有女朋友跨年的群傻逼 is down
In de0fa87, Mirror-没有女朋友跨年的群傻逼 (https://x.ksfu.top/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Mirror-木生 is back up in cfa927e.
|
gharchive/issue
| 2022-12-31T07:46:10 |
2025-04-01T06:39:27.628145
|
{
"authors": [
"lsy223622"
],
"repo": "lsy223622/xdncov-mirror-status",
"url": "https://github.com/lsy223622/xdncov-mirror-status/issues/15781",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1316035989
|
🛑 Mirror-爬 is down
In 2777a48, Mirror-爬 (https://ncov.zhouym.tech/) was down:
HTTP code: 404
Response time: 1290 ms
Resolved: Mirror-鱼露 is back up in ee13cfe.
|
gharchive/issue
| 2022-07-24T23:22:32 |
2025-04-01T06:39:27.630664
|
{
"authors": [
"lsy223622"
],
"repo": "lsy223622/xdncov-mirror-status",
"url": "https://github.com/lsy223622/xdncov-mirror-status/issues/2876",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1342750121
|
🛑 Mirror-爬 is down
In d06b2d9, Mirror-爬 (https://ncov.zhouym.tech/) was down:
HTTP code: 404
Response time: 2341 ms
Resolved: Mirror-鱼露 is back up in 05c0b7b.
|
gharchive/issue
| 2022-08-18T08:39:06 |
2025-04-01T06:39:27.633120
|
{
"authors": [
"lsy223622"
],
"repo": "lsy223622/xdncov-mirror-status",
"url": "https://github.com/lsy223622/xdncov-mirror-status/issues/4913",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1352481783
|
🛑 Mirror-爬 is down
In c8f51bb, Mirror-爬 (https://ncov.zhouym.tech/) was down:
HTTP code: 404
Response time: 7523 ms
Resolved: Mirror-鱼露 is back up in 8e7f1ff.
|
gharchive/issue
| 2022-08-26T16:29:14 |
2025-04-01T06:39:27.635530
|
{
"authors": [
"lsy223622"
],
"repo": "lsy223622/xdncov-mirror-status",
"url": "https://github.com/lsy223622/xdncov-mirror-status/issues/5576",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1401913085
|
🛑 Mirror-爬 is down
In b38ca2f, Mirror-爬 (https://ncov.zhouym.tech/) was down:
HTTP code: 404
Response time: 1148 ms
Resolved: Mirror-鱼露 is back up in 32c8887.
|
gharchive/issue
| 2022-10-08T11:13:37 |
2025-04-01T06:39:27.637974
|
{
"authors": [
"lsy223622"
],
"repo": "lsy223622/xdncov-mirror-status",
"url": "https://github.com/lsy223622/xdncov-mirror-status/issues/8269",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
597648413
|
Hide Notification on ios
How to hide Notification on ios call from flutter?
What do you mean hide Notification? no notifications will be shown unless you ask for.
Closing due inactivity, if there is more info that you can add to the issue, comment here and we can re open the issue.
@erickzanardo not OP but I was subbed to this issue as I was looking for a solution to this myself.
I think OP was referring to clearing the notification center media player notification programmatically, as it is retained there long after it is actually used within your app (eg. if you only use the audio player on a specific page).
@luanpotter Can this issue be reopened? Currently, none of audioplayers' notificationService methods dismiss the iOS headless service and setNotification notification.
The only way to dismiss the notification is to write native code.
There should be a way to dismiss the notification from the player itself.
@searchy2 this is issue don't have much info on it, I think it would be better if a new issue were open, would you mind opening a new one so we can track this more easily?
Sure, I'll open a new issue.
Create a new issue https://github.com/luanpotter/audioplayers/issues/897
|
gharchive/issue
| 2020-04-10T01:51:05 |
2025-04-01T06:39:27.682825
|
{
"authors": [
"erickzanardo",
"jamie1192",
"luanpotter",
"quangquy87",
"searchy2"
],
"repo": "luanpotter/audioplayers",
"url": "https://github.com/luanpotter/audioplayers/issues/492",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2141659301
|
🛑 Školský šport is down
In e377595, Školský šport (https://skolskysport.sk/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Školský šport is back up in a92d32f after 7 minutes.
|
gharchive/issue
| 2024-02-19T07:03:20 |
2025-04-01T06:39:27.685659
|
{
"authors": [
"lubosm"
],
"repo": "lubosm/minedusk-uptime",
"url": "https://github.com/lubosm/minedusk-uptime/issues/1400",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1804319706
|
🛑 Školský šport is down
In a466ae9, Školský šport (https://skolskysport.sk/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Školský šport is back up in f63e3b7.
|
gharchive/issue
| 2023-07-14T07:02:32 |
2025-04-01T06:39:27.688158
|
{
"authors": [
"lubosm"
],
"repo": "lubosm/minedusk-uptime",
"url": "https://github.com/lubosm/minedusk-uptime/issues/881",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2367005227
|
Beaming the bottom of a splitter crash
If you shine a laser under (and probably above) a Polarizing Beam Splitter, the game crashes.
latest.log
crash-2024-06-21_14.55.04-server.txt
Steps to reproduce:
https://github.com/luccaPossamai/optical/assets/94794129/92eb4e46-3cae-4aeb-a49a-49d3b941dcc2
Yup, I uploaded to curseforge the version before testing. In newer versions this bug would not exist.
|
gharchive/issue
| 2024-06-21T17:57:37 |
2025-04-01T06:39:27.726334
|
{
"authors": [
"AtobaAzul",
"luccaPossamai"
],
"repo": "luccaPossamai/optical",
"url": "https://github.com/luccaPossamai/optical/issues/7",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2047040646
|
Add prisma adapter installation instructions
In the Getting started in Next.js App Router documentation page, it might be useful to show how to install the Prisma adapater before using it in the code, so that the onboarded user doesn't wonder if they missed a step.
Might be better if we just mention in the paragraph that the adapters are provided as a separate package?
Might be better if we just mention in the paragraph that the adapters are provided as a separate package?
I think what's weird on this page is that there is a code snippet that is provided, and it doesn't work because of the dependency issue. But if you need to read the full adapter documentation to set it up, it's not clear why the code is provided in the first place.
If it's just for illustrative purposes, maybe we should say that the code should ”looke like this”, and that the exact code depends on the adapter used, and then provide the adapters list to chose from.
I am not very good at writing, so I'm not even sure how to properly fix it. But I'm quite sure that it's a bit misleading to have a piece of code that doesn't work, and if it's not supposed to work, it's not clear why it's presented in the first place.
Hey, I appreciate the PR but we're merging v3 in a few hours so I'll be closing this for now to clean up the repo. You can create a new PR against the v2 branch if you'd like since we're planning to support it for a few more months.
|
gharchive/pull-request
| 2023-12-18T16:38:19 |
2025-04-01T06:39:27.742505
|
{
"authors": [
"alexpirine",
"pilcrowOnPaper"
],
"repo": "lucia-auth/lucia",
"url": "https://github.com/lucia-auth/lucia/pull/1307",
"license": "0BSD",
"license_type": "permissive",
"license_source": "github-api"
}
|
2283679742
|
Module '"lucide-react"' has no exported member 'Notebook'.ts(2305)
Package
[ ] lucide
[ ] lucide-angular
[ ] lucide-flutter
[ ] lucide-preact
[X] lucide-react
[ ] lucide-react-native
[ ] lucide-solid
[ ] lucide-svelte
[ ] lucide-vue
[ ] lucide-vue-next
[ ] Figma plugin
[ ] source/main
[ ] other/not relevant
Version
0.378.0
Can you reproduce this in the latest version?
[X] Yes
[ ] No
Browser
[X] Chrome/Chromium
[ ] Firefox
[ ] Safari
[ ] Edge
[ ] iOS Safari
[ ] Opera
[ ] Other/not relevant
Operating system
[X] Windows
[ ] Linux
[ ] macOS
[ ] ChromeOS
[ ] iOS
[ ] Android
[ ] Other/not relevant
Description
import {
ActivitySquareIcon,
FileUpIcon,
LayoutDashboardIcon,
BadgeDollarSign,
Contact,
Users,
LayoutPanelLeft,
Notebook,
} from "lucide-react";
when I tried to import some icons from the lucide-react package
I got the next error: Module '"lucide-react"' has no exported member 'Notebook'.ts(2305). Needles to say I looked in forums for similar problems. specifically I tried to downgrade to 0.263.0 version but that didn't work.
Steps to reproduce
install latest version
1.import 'notebook' icon
2.get the error
Checklist
[X] I have searched if someone has submitted a similar issue before and there weren't any. (Please make sure to also search closed issues, as this issue might already have been resolved.)
I cannot reproduce this issue, can you make sure an earlier version isn't stuck in some kind of cache and you're actually using v0.378.0, which definitely has this export?
(.next cache, node_modules etc)
I have version 0.378 both in package.json and pnpm-lock.yaml . I tried looking into the cache but vs code wont read it
Can you check if node_modules/lucide-react/dist/esm/notebook.js exists?
As for any cache, you shouldn't be looking into it, but clearing it (since you haven't provided any extra information about your frameworks, I cannot help you with how that should be done, but if it's next.js for example, you may try deleting the entire .next folder, same goes for node_modules, especially if the export above isn't there).
|
gharchive/issue
| 2024-05-07T15:39:47 |
2025-04-01T06:39:27.756702
|
{
"authors": [
"CayllahuaPedro",
"karsa-mistmere"
],
"repo": "lucide-icons/lucide",
"url": "https://github.com/lucide-icons/lucide/issues/2138",
"license": "ISC",
"license_type": "permissive",
"license_source": "github-api"
}
|
2628126118
|
SolidStart: createSignal and onMount, etc not working
Package
[ ] lucide
[ ] lucide-angular
[ ] lucide-flutter
[ ] lucide-preact
[ ] lucide-react
[ ] lucide-react-native
[X] lucide-solid
[ ] lucide-svelte
[ ] lucide-vue
[ ] lucide-vue-next
[ ] Figma plugin
[ ] source/main
[ ] other/not relevant
Version
0.454.0
Can you reproduce this in the latest version?
[X] Yes
[ ] No
Browser
[] Chrome/Chromium
[ ] Firefox
[ ] Safari
[ ] Edge
[ ] iOS Safari
[ ] Opera
[ X ] Arc
[ ] Other/not relevant
Operating system
[ ] Windows
[ ] Linux
[X] macOS
[ ] ChromeOS
[ ] iOS
[ ] Android
[ ] Other/not relevant
Description
In SolidStart, importing and using an icon (e.g., AArrowDown from lucide-solid) causes onMount, createSignal, etc not to work.
Steps to reproduce
import { AArrowDown } from "lucide-solid";
const App = () => {
const [count, setCount] = createSignal(0); // not working
onMount(() => {
console.log("Hi"); // not working
});
return (
<div>
<AArrowDown />
<button onClick={() => {
setCount(count() + 1); // not working
}>increase</button>
</div>
);
}
Checklist
[X] I have searched if someone has submitted a similar issue before and there weren't any. (Please make sure to also search closed issues, as this issue might already have been resolved.)
what the… this issue happens only with Arc Browser. Arc Browser is an ad blocker
|
gharchive/issue
| 2024-11-01T01:36:37 |
2025-04-01T06:39:27.765363
|
{
"authors": [
"gronxb"
],
"repo": "lucide-icons/lucide",
"url": "https://github.com/lucide-icons/lucide/issues/2574",
"license": "ISC",
"license_type": "permissive",
"license_source": "github-api"
}
|
830916805
|
Why
Traceback (most recent call last):
File "c:\users\marshall\appdata\local\programs\python\python38\lib\runpy.py", line 194, in _run_module_as_main
return run_code(code, main_globals, None,
File "c:\users\marshall\appdata\local\programs\python\python38\lib\runpy.py", line 87, in run_code
exec(code, run_globals)
File "C:\Users\Marshall\AppData\Local\Programs\Python\Python38\Scripts\imagine.exe_main.py", line 4, in
File "c:\users\marshall\appdata\local\programs\python\python38\lib\site-packages\deep_daze_init.py", line 1, in
from deep_daze.deep_daze import DeepDaze, Imagine
File "c:\users\marshall\appdata\local\programs\python\python38\lib\site-packages\deep_daze\deep_daze.py", line 25, in
assert torch.cuda.is_available(), 'CUDA must be available in order to use Deep Daze'
AssertionError: CUDA must be available in order to use Deep Daze
Hi, try this:
pip uninstall torch
pip install torch==1.8.0+cu111 torchvision==0.9.0+cu111 torchaudio===0.8.0 -f https://download.pytorch.org/whl/torch_stable.html
Whenever i try to open the torch-1.8.0+cu111-cp38-cp38-win_amd64.whl file it instantly closes and i cant see what it says
|
gharchive/issue
| 2021-03-13T14:58:56 |
2025-04-01T06:39:27.771728
|
{
"authors": [
"JiriSuster",
"PrivateServersGANG"
],
"repo": "lucidrains/deep-daze",
"url": "https://github.com/lucidrains/deep-daze/issues/75",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2118960064
|
🛑 Loxone is down
In 547c26e, Loxone (http://loxone.kerl.io) was down:
HTTP code: 502
Response time: 225 ms
Resolved: Loxone is back up in cddba64 after 5 minutes.
|
gharchive/issue
| 2024-02-05T16:30:16 |
2025-04-01T06:39:27.774276
|
{
"authors": [
"lucienkerl"
],
"repo": "lucienkerl/status",
"url": "https://github.com/lucienkerl/status/issues/1602",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
204422110
|
I have a problem with activate Redis
Good evening everyone!
I currently have a problem that I activate Redis on LuckyPerms my API which also uses Redis this my in error (Error: http://pastebin.com/nn3r6BTj)
It's been 3 days since I search for my I do not find
Thanks you!
Make sure you're packaging the newest Jedis version in your plugin. Seems like there's an issue there?
Could you give me the redis version you used for LuckyPerms? :)
Thanks!
Ah yeah. I just realised, this is a LuckPerms issue.
I shade this version.
https://github.com/lucko/jedis/releases/tag/jedis-2.9.0-shaded
I'll fix that when a get a chance. (probably tomorrow)
I tried to update it in my API
I think the connection system needs to change ..
I get an error in my code! (http://prnt.sc/e2vk2z)
I would probably do with my developer to look at my API :)
Thanks :)
Should be fixed in this build.
https://ci.lucko.me/job/LuckPerms/41/
Let me know if you have any further issues.
|
gharchive/issue
| 2017-01-31T21:01:25 |
2025-04-01T06:39:27.795434
|
{
"authors": [
"jimmy57000",
"lucko"
],
"repo": "lucko/LuckPerms",
"url": "https://github.com/lucko/LuckPerms/issues/158",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1454055829
|
Check the viability and implement Iris Replication
https://docs.unrealengine.com/5.1/en-US/unreal-engine-5.1-release-notes/#networkingandmultiplayer
I haven't found a way to make it work in the released UE5.1 (binary via Epic Launcher), only when compiling the engine from source.
Already tried to force the usage by overriding the target editor options (bOverrideBuildEnvironment) and adding a bUseIris = true, in addition to adding SetupIrisSupport(Target) in the modules, but it didn't work, it gives linking errors. And there is no IrisCore in the Intermediate folder of the Engine. 🥲
|
gharchive/issue
| 2022-11-17T21:59:33 |
2025-04-01T06:39:27.799193
|
{
"authors": [
"lucoiso"
],
"repo": "lucoiso/UEProject_Elementus",
"url": "https://github.com/lucoiso/UEProject_Elementus/issues/41",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
210261934
|
New readme for http
@skibinska I think we need a separate readme for http so we can discuss http methods and status codes properly. Are you okay with that?
yeap, I actually mentioned it in my workshop, but it makes more to introduce it here.
Should we rethink the way that we will be presenting this workshop or won't ask the questions about this topic?
I just added some questions! What about JSON? Should we add something about that too?
I would just add to first readme one sentence about JSON - it is a way to store information in an organized, easy-to-access manner and maybe that it looks like a JS object but its keys and values are in quotation mark.
|
gharchive/issue
| 2017-02-25T21:05:10 |
2025-04-01T06:39:27.801290
|
{
"authors": [
"lucymonie",
"skibinska"
],
"repo": "lucymonie/api-workshop",
"url": "https://github.com/lucymonie/api-workshop/issues/5",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
411872466
|
What does CENTER(SIZE)_VARIANCE mean in defaults.py?
I don`t know these configure in default.py:
Hard negative mining
_C.MODEL.CENTER_VARIANCE = 0.1
_C.MODEL.SIZE_VARIANCE = 0.2
↑
Theyr used in /ssd/util/box_utils.py when boxes invert into locations or locations invert into boxes.But I dont know why.
change MAX_PER_CLASS to 400 as official caffe code will slightly increase mAP(0.8025=>0.8063, 0.7783=>0.7798)
_C.TEST.MAX_PER_CLASS = 200
_C.TEST.MAX_PER_IMAGE = -1
↑
I dont know these either and I cant find where they`r used in project.
Can anyone help?Thanks all the time.
Variance is used to encode/decode prior bboxes .
https://github.com/weiliu89/caffe/blob/4817bf8b4200b35ada8ed0dc378dceaf38c539e4/examples/ssd/ssd_pascal.py#L322
MAX_PER_CLASS and MAX_PER_IMAGE are used in post_processor.py
|
gharchive/issue
| 2019-02-19T11:17:02 |
2025-04-01T06:39:27.809133
|
{
"authors": [
"RichardHWD",
"lufficc"
],
"repo": "lufficc/SSD",
"url": "https://github.com/lufficc/SSD/issues/40",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
127926786
|
Cluster isn't ready and enableOfflineQueue options is false, with defaults
Hey,
We are using ioredis for quite some time now and its great.
We noticed that sometimes (Can't reproduce really) we have logs of Cluster isn't ready and enableOfflineQueue options is false, even though we are using the default settings of enableOfflineQueue which is true.
Any ideas?
I think it happens when the library didn't finish loading and we are already sending commands to it
Sorry for my late response. It's really strange since this error only emits when the enableOfflineQueue is false: https://github.com/luin/ioredis/blob/master/lib/cluster.js#L479-L486.
What do you mean by the library didn't finish loading? Are you sending commands before setting cluster.options.enableOfflineQueue = true?
I'm not setting cluster.options.enableOfflineQueue = true at all as it is the default, but maybe its possible that the library didn't yet set this property itself and then when its getting to the line you marked its seeing the value as undefined or something...
I don't think it's possible since the options is defined in the constructor of Cluster (https://github.com/luin/ioredis/blob/master/lib/cluster.js#L64).
I actually managed to replicate it. Trying to create a simple script now
@luin found it. PR ready
|
gharchive/issue
| 2016-01-21T13:42:00 |
2025-04-01T06:39:27.813764
|
{
"authors": [
"luin",
"shaharmor"
],
"repo": "luin/ioredis",
"url": "https://github.com/luin/ioredis/issues/233",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
770375980
|
Plugin is crashing entire Homebridge because of tedee Server problems
Hi @lukasroegner
there is a major issue on tedee's server service at the moment. Thursday 17.12.2020 21:45 ongoing
The App also is showing this Issue and I talked to Andrzej from tedee.
However, I only noticed because my entire homebridge crashed.
The Plugin should probably not cause a complete crash and reboot loop just because tedee service is not available?!
Maybe you can take a look at it at the next opportunity so that the plugin can handle it better.
Thanks in advance.
[17.12.2020, 22:14:47] [TedeePlatform] Error while getting locks from API: Error: Request failed with status code 401
Error: Request failed with status code 401
at createError (/usr/local/lib/node_modules/homebridge-tedee/node_modules/axios/lib/core/createError.js:16:15)
at settle (/usr/local/lib/node_modules/homebridge-tedee/node_modules/axios/lib/core/settle.js:17:12)
at IncomingMessage.handleStreamEnd (/usr/local/lib/node_modules/homebridge-tedee/node_modules/axios/lib/adapters/http.js:236:11)
at IncomingMessage.emit (events.js:327:22)
at endReadableNT (internal/streams/readable.js:1327:12)
at processTicksAndRejections (internal/process/task_queues.js:80:21)
{
config: {
url: 'https://api.tedee.com/api/v1.15/my/lock',
method: 'get',
headers: {
Accept: 'application/json, text/plain, */*',
Authorization: 'Bearer xxxxxxxxxxxxxxxxxxxxxxxiOiJSUzI1NiIsImtpZCI6Ilg1ZVhrNHh5b2pORnVtMWtsMll0djhkbE5QNC1jNTdkTzZRR1RWQndhTmsifQ.eyJpc3MiOiJodHRwczovL3RlZGVlLmIyY2xvZ2luLmNvbS90ZnAvZTMxOTYxN2EtZjMyMi00Y2JhLTgxZWYtNGI1Yjk1MGU5MTcyL2IyY18xX3NpZ25pbl9yb3BjL3YyLjAvIiwiZXhwIjoxNjA4MjUwNDc1LCJuYmYiOjE2MDgyMzk2NzUsImF1ZCI6IjAyMTA2YjgyLTA1MjQtNGZkMy1hYzU3LWFmNzc0ZjM0MDk3OSIsImlkcCI6IkxvY2FsQWNjb3VudCIsIm9pZCI6ImNiNjY1ZTFiLxxxxxxxxxxxxiNjY1ZTFiLTZmY2QtNGY2Ny05ODQzLTcxN2ZlZWIzMWEzZCIsIm5hbWUiOiJESmF5IiwibmV3VXNlciI6ZmFsc2UsImVtYWlscyI6WyJkamF5QG1lLmNvbSJdLCJ0ZnAiOiJCMkNfMV9TaWduSW5fUm9wYyIsImF6cCI6IjAyMTA2YjgyLTA1MjQtNGZkMy1hYzU3LWFmNzc0ZjM0MDk3OSIsInZlciI6IjEuMCIsImlhdCI6MTYwODIzOTY3NX0.D1FcynocxrgCekS10JP8lCOteZy-jqiZgwxUCgM7EhtNjk5V9R_Mawq9i0AvXcow0oK5NynKpSMmt2Wp1KY2O9j-vRf-GiF-1xMTo97US6PnthkIyF41AEAcwdJSewBm6LJswa-DiGDAA2Le2nt1yg9hDcYamj1cuYoFqtatQ1YGK8NfnZZfnQzvwKE3EsT_Vby_uUiZgSwPZOwMkMMmYQ2qA0EFP8x0inni6mljwThps6xBBG3RGVCvGhmhn51XjpvYhyUsO7TKyXUANHKSsfi1GvRgNqA_RzSfyZKBLFT-bI2fgiUpyhhwETP5q62TSHWiaYvvcNY19gzXj3wPcg',
'User-Agent': 'axios/0.19.2'
},
transformRequest: [ [Function: transformRequest] ],
transformResponse: [ [Function: transformResponse] ],
timeout: 0,
adapter: [Function: httpAdapter],
xsrfCookieName: 'XSRF-TOKEN',
xsrfHeaderName: 'X-XSRF-TOKEN',
maxContentLength: -1,
validateStatus: [Function: validateStatus],
data: undefined
},
request: <ref *1> ClientRequest {
_events: [Object: null prototype] {
socket: [Function (anonymous)],
abort: [Function (anonymous)],
aborted: [Function (anonymous)],
error: [Function (anonymous)],
timeout: [Function (anonymous)],
prefinish: [Function: requestOnPrefinish]
},
_eventsCount: 6,
_maxListeners: undefined,
outputData: [],
outputSize: 0,
writable: true,
destroyed: false,
_last: true,
chunkedEncoding: false,
shouldKeepAlive: false,
_defaultKeepAlive: true,
useChunkedEncodingByDefault: false,
sendDate: false,
_removedConnection: false,
_removedContLen: false,
_removedTE: false,
_contentLength: 0,
_hasBody: true,
_trailer: '',
finished: true,
_headerSent: true,
socket: TLSSocket {
_tlsOptions: [Object],
_secureEstablished: true,
_securePending: false,
_newSessionPending: false,
_controlReleased: true,
secureConnecting: false,
_SNICallback: null,
servername: 'api.tedee.com',
alpnProtocol: false,
authorized: true,
authorizationError: null,
encrypted: true,
_events: [Object: null prototype],
_eventsCount: 10,
connecting: false,
_hadError: false,
_parent: null,
_host: 'api.tedee.com',
_readableState: [ReadableState],
_maxListeners: undefined,
_writableState: [WritableState],
allowHalfOpen: false,
_sockname: null,
_pendingData: null,
_pendingEncoding: '',
server: undefined,
_server: null,
ssl: [TLSWrap],
_requestCert: true,
_rejectUnauthorized: true,
parser: null,
_httpMessage: [Circular *1],
[Symbol(res)]: [TLSWrap],
[Symbol(verified)]: true,
[Symbol(pendingSession)]: null,
[Symbol(async_id_symbol)]: 5351,
[Symbol(kHandle)]: [TLSWrap],
[Symbol(kSetNoDelay)]: false,
[Symbol(lastWriteQueueSize)]: 0,
[Symbol(timeout)]: null,
[Symbol(kBuffer)]: null,
[Symbol(kBufferCb)]: null,
[Symbol(kBufferGen)]: null,
[Symbol(kCapture)]: false,
[Symbol(kBytesRead)]: 0,
[Symbol(kBytesWritten)]: 0,
[Symbol(connect-options)]: [Object],
[Symbol(RequestTimeout)]: undefined
},
_header: 'GET /api/v1.15/my/lock HTTP/1.1\r\n' +
'Accept: application/json, text/plain, */*\r\n' +
'Authorization: Bearer xxxxxxxxxxxxxxxxxxxxxxxiOiJSUzI1NiIsImtpZCI6Ilg1ZVhrNHh5b2pORnVtMWtsMll0djhkbE5QNC1jNTdkTzZRR1RWQndhTmsifQ.eyJpc3MiOiJodHRwczovL3RlZGVlLmIyY2xvZ2luLmNvbS90ZnAvZTMxOTYxN2EtZjMyMi00Y2JhLTgxZWYtNGI1Yjk1MGU5MTcyL2IyY18xX3NpZ25pbl9yb3BjL3YyLjAvIiwiZXhwIjoxNjA4MjUwNDc1LCJuYmYiOjE2MDgyMzk2NzUsImF1ZCI6IjAyMTA2YjgyLTA1MjQtNGZkMy1hYzU3LWFmNzc0ZjM0MDk3OSIsImlkcCI6IkxvY2FsQWNjb3VudCIsIm9pZCI6ImNiNjY1ZTFiLxxxxxxxxxxxxiNjY1ZTFiLTZmY2QtNGY2Ny05ODQzLTcxN2ZlZWIzMWEzZCIsIm5hbWUiOiJESmF5IiwibmV3VXNlciI6ZmFsc2UsImVtYWlscyI6WyJkamF5QG1lLmNvbSJdLCJ0ZnAiOiJCMkNfMV9TaWduSW5fUm9wYyIsImF6cCI6IjAyMTA2YjgyLTA1MjQtNGZkMy1hYzU3LWFmNzc0ZjM0MDk3OSIsInZlciI6IjEuMCIsImlhdCI6MTYwODIzOTY3NX0.D1FcynocxrgCekS10JP8lCOteZy-jqiZgwxUCgM7EhtNjk5V9R_Mawq9i0AvXcow0oK5NynKpSMmt2Wp1KY2O9j-vRf-GiF-1xMTo97US6PnthkIyF41AEAcwdJSewBm6LJswa-DiGDAA2Le2nt1yg9hDcYamj1cuYoFqtatQ1YGK8NfnZZfnQzvwKE3EsT_Vby_uUiZgSwPZOwMkMMmYQ2qA0EFP8x0inni6mljwThps6xBBG3RGVCvGhmhn51XjpvYhyUsO7TKyXUANHKSsfi1GvRgNqA_RzSfyZKBLFT-bI2fgiUpyhhwETP5q62TSHWiaYvvcNY19gzXj3wPcg\r\n' +
'User-Agent: axios/0.19.2\r\n' +
'Host: api.tedee.com\r\n' +
'Connection: close\r\n' +
'\r\n',
_keepAliveTimeout: 0,
_onPendingData: [Function: noopPendingOutput],
agent: Agent {
_events: [Object: null prototype],
_eventsCount: 2,
_maxListeners: undefined,
defaultPort: 443,
protocol: 'https:',
options: [Object],
requests: {},
sockets: [Object],
freeSockets: {},
keepAliveMsecs: 1000,
keepAlive: false,
maxSockets: Infinity,
maxFreeSockets: 256,
scheduling: 'fifo',
maxTotalSockets: Infinity,
totalSocketCount: 4,
maxCachedSessions: 100,
_sessionCache: [Object],
[Symbol(kCapture)]: false
},
socketPath: undefined,
method: 'GET',
maxHeaderSize: undefined,
insecureHTTPParser: undefined,
path: '/api/v1.15/my/lock',
_ended: true,
res: IncomingMessage {
_readableState: [ReadableState],
_events: [Object: null prototype],
_eventsCount: 3,
_maxListeners: undefined,
socket: [TLSSocket],
httpVersionMajor: 1,
httpVersionMinor: 1,
httpVersion: '1.1',
complete: true,
rawHeaders: [Array],
rawTrailers: [],
aborted: false,
upgrade: false,
url: '',
method: null,
statusCode: 401,
statusMessage: 'Unauthorized',
client: [TLSSocket],
_consuming: true,
_dumped: false,
req: [Circular *1],
responseUrl: 'https://api.tedee.com/api/v1.15/my/lock',
redirects: [],
[Symbol(kCapture)]: false,
[Symbol(kHeaders)]: [Object],
[Symbol(kHeadersCount)]: 22,
[Symbol(kTrailers)]: null,
[Symbol(kTrailersCount)]: 0,
[Symbol(RequestTimeout)]: undefined
},
aborted: false,
timeoutCb: null,
upgradeOrConnect: false,
parser: null,
maxHeadersCount: null,
reusedSocket: false,
host: 'api.tedee.com',
protocol: 'https:',
_redirectable: Writable {
_writableState: [WritableState],
_events: [Object: null prototype],
_eventsCount: 2,
_maxListeners: undefined,
_options: [Object],
_redirectCount: 0,
_redirects: [],
_requestBodyLength: 0,
_requestBodyBuffers: [],
_onNativeResponse: [Function (anonymous)],
_currentRequest: [Circular *1],
_currentUrl: 'https://api.tedee.com/api/v1.15/my/lock',
[Symbol(kCapture)]: false
},
[Symbol(kCapture)]: false,
[Symbol(kNeedDrain)]: false,
[Symbol(corked)]: 0,
[Symbol(kOutHeaders)]: [Object: null prototype] {
accept: [Array],
authorization: [Array],
'user-agent': [Array],
host: [Array]
}
},
response: {
status: 401,
statusText: 'Unauthorized',
headers: {
'transfer-encoding': 'chunked',
'content-type': 'application/json',
server: 'Microsoft-IIS/10.0',
'www-authenticate': 'Bearer error="invalid_token", error_description="The signature key was not found", Bearer error="invalid_token", error_description="The signature key was not found"',
'request-context': 'appId=cid-v1:xxxxxxxxxxxx-bf23-4ab2-b999-bbd05e97fbc2',
'strict-transport-security': 'max-age=xxxxxxxxxxxx',
'x-correlation-id': 'xxxxxxxxxxxx-0001-f800-b63f-xxxxxxxxxxxx',
'x-powered-by': 'ASP.NET',
date: 'Thu, 17 Dec 2020 21:14:47 GMT',
connection: 'close'
},
config: {
url: 'https://api.tedee.com/api/v1.15/my/lock',
method: 'get',
headers: [Object],
transformRequest: [Array],
transformResponse: [Array],
timeout: 0,
adapter: [Function: httpAdapter],
xsrfCookieName: 'XSRF-TOKEN',
xsrfHeaderName: 'X-XSRF-TOKEN',
maxContentLength: -1,
validateStatus: [Function: validateStatus],
data: undefined
},
request: <ref *1> ClientRequest {
_events: [Object: null prototype],
_eventsCount: 6,
_maxListeners: undefined,
outputData: [],
outputSize: 0,
writable: true,
destroyed: false,
_last: true,
chunkedEncoding: false,
shouldKeepAlive: false,
_defaultKeepAlive: true,
useChunkedEncodingByDefault: false,
sendDate: false,
_removedConnection: false,
_removedContLen: false,
_removedTE: false,
_contentLength: 0,
_hasBody: true,
_trailer: '',
finished: true,
_headerSent: true,
socket: [TLSSocket],
_header: 'GET /api/v1.15/my/lock HTTP/1.1\r\n' +
'Accept: application/json, text/plain, */*\r\n' +
'Authorization: Bearer xxxxxxxxxxxxxxxxxxxxxxxiOiJSUzI1NiIsImtpZCI6Ilg1ZVhrNHh5b2pORnVtMWtsMll0djhkbE5QNC1jNTdkTzZRR1RWQndhTmsifQ.eyJpc3MiOiJodHRwczovL3RlZGVlLmIyY2xvZ2luLmNvbS90ZnAvZTMxOTYxN2EtZjMyMi00Y2JhLTgxZWYtNGI1Yjk1MGU5MTcyL2IyY18xX3NpZ25pbl9yb3BjL3YyLjAvIiwiZXhwIjoxNjA4MjUwNDc1LCJuYmYiOjE2MDgyMzk2NzUsImF1ZCI6IjAyMTA2YjgyLTA1MjQtNGZkMy1hYzU3LWFmNzc0ZjM0MDk3OSIsImlkcCI6IkxvY2FsQWNjb3VudCIsIm9pZCI6ImNiNjY1ZTFiLxxxxxxxxxxxxiNjY1ZTFiLTZmY2QtNGY2Ny05ODQzLTcxN2ZlZWIzMWEzZCIsIm5hbWUiOiJESmF5IiwibmV3VXNlciI6ZmFsc2UsImVtYWlscyI6WyJkamF5QG1lLmNvbSJdLCJ0ZnAiOiJCMkNfMV9TaWduSW5fUm9wYyIsImF6cCI6IjAyMTA2YjgyLTA1MjQtNGZkMy1hYzU3LWFmNzc0ZjM0MDk3OSIsInZlciI6IjEuMCIsImlhdCI6MTYwODIzOTY3NX0.D1FcynocxrgCekS10JP8lCOteZy-jqiZgwxUCgM7EhtNjk5V9R_Mawq9i0AvXcow0oK5NynKpSMmt2Wp1KY2O9j-vRf-GiF-1xMTo97US6PnthkIyF41AEAcwdJSewBm6LJswa-DiGDAA2Le2nt1yg9hDcYamj1cuYoFqtatQ1YGK8NfnZZfnQzvwKE3EsT_Vby_uUiZgSwPZOwMkMMmYQ2qA0EFP8x0inni6mljwThps6xBBG3RGVCvGhmhn51XjpvYhyUsO7TKyXUANHKSsfi1GvRgNqA_RzSfyZKBLFT-bI2fgiUpyhhwETP5q62TSHWiaYvvcNY19gzXj3wPcg\r\n' +
'User-Agent: axios/0.19.2\r\n' +
'Host: api.tedee.com\r\n' +
'Connection: close\r\n' +
'\r\n',
_keepAliveTimeout: 0,
_onPendingData: [Function: noopPendingOutput],
agent: [Agent],
socketPath: undefined,
method: 'GET',
maxHeaderSize: undefined,
insecureHTTPParser: undefined,
path: '/api/v1.15/my/lock',
_ended: true,
res: [IncomingMessage],
aborted: false,
timeoutCb: null,
upgradeOrConnect: false,
parser: null,
maxHeadersCount: null,
reusedSocket: false,
host: 'api.tedee.com',
protocol: 'https:',
_redirectable: [Writable],
[Symbol(kCapture)]: false,
[Symbol(kNeedDrain)]: false,
[Symbol(corked)]: 0,
[Symbol(kOutHeaders)]: [Object: null prototype]
},
data: { success: false, errorMessages: [Array], statusCode: 401 }
},
isAxiosError: true,
toJSON: [Function (anonymous)]
}
Error: Request failed with status code 401
at createError (/usr/local/lib/node_modules/homebridge-tedee/node_modules/axios/lib/core/createError.js:16:15)
at settle (/usr/local/lib/node_modules/homebridge-tedee/node_modules/axios/lib/core/settle.js:17:12)
at IncomingMessage.handleStreamEnd (/usr/local/lib/node_modules/homebridge-tedee/node_modules/axios/lib/adapters/http.js:236:11)
at IncomingMessage.emit (events.js:327:22)
at endReadableNT (internal/streams/readable.js:1327:12)
at processTicksAndRejections (internal/process/task_queues.js:80:21)
{
config: {
url: 'https://api.tedee.com/api/v1.15/my/lock',
method: 'get',
headers: {
Accept: 'application/json, text/plain, */*',
Authorization: 'Bearer xxxxxxxxxxxxxxxxxxxxxxxiOiJSUzI1NiIsImtpZCI6Ilg1ZVhrNHh5b2pORnVtMWtsMll0djhkbE5QNC1jNTdkTzZRR1RWQndhTmsifQ.eyJpc3MiOiJodHRwczovL3RlZGVlLmIyY2xvZ2luLmNvbS90ZnAvZTMxOTYxN2EtZjMyMi00Y2JhLTgxZWYtNGI1Yjk1MGU5MTcyL2IyY18xX3NpZ25pbl9yb3BjL3YyLjAvIiwiZXhwIjoxNjA4MjUwNDc1LCJuYmYiOjE2MDgyMzk2NzUsImF1ZCI6IjAyMTA2YjgyLTA1MjQtNGZkMy1hYzU3LWFmNzc0ZjM0MDk3OSIsImlkcCI6IkxvY2FsQWNjb3VudCIsIm9pZCI6ImNiNjY1ZTFiLxxxxxxxxxxxxiNjY1ZTFiLTZmY2QtNGY2Ny05ODQzLTcxN2ZlZWIzMWEzZCIsIm5hbWUiOiJESmF5IiwibmV3VXNlciI6ZmFsc2UsImVtYWlscyI6WyJkamF5QG1lLmNvbSJdLCJ0ZnAiOiJCMkNfMV9TaWduSW5fUm9wYyIsImF6cCI6IjAyMTA2YjgyLTA1MjQtNGZkMy1hYzU3LWFmNzc0ZjM0MDk3OSIsInZlciI6IjEuMCIsImlhdCI6MTYwODIzOTY3NX0.D1FcynocxrgCekS10JP8lCOteZy-jqiZgwxUCgM7EhtNjk5V9R_Mawq9i0AvXcow0oK5NynKpSMmt2Wp1KY2O9j-vRf-GiF-1xMTo97US6PnthkIyF41AEAcwdJSewBm6LJswa-DiGDAA2Le2nt1yg9hDcYamj1cuYoFqtatQ1YGK8NfnZZfnQzvwKE3EsT_Vby_uUiZgSwPZOwMkMMmYQ2qA0EFP8x0inni6mljwThps6xBBG3RGVCvGhmhn51XjpvYhyUsO7TKyXUANHKSsfi1GvRgNqA_RzSfyZKBLFT-bI2fgiUpyhhwETP5q62TSHWiaYvvcNY19gzXj3wPcg',
'User-Agent': 'axios/0.19.2'
},
transformRequest: [ [Function: transformRequest] ],
transformResponse: [ [Function: transformResponse] ],
timeout: 0,
adapter: [Function: httpAdapter],
xsrfCookieName: 'XSRF-TOKEN',
xsrfHeaderName: 'X-XSRF-TOKEN',
maxContentLength: -1,
validateStatus: [Function: validateStatus],
data: undefined
},
request: <ref *1> ClientRequest {
_events: [Object: null prototype] {
socket: [Function (anonymous)],
abort: [Function (anonymous)],
aborted: [Function (anonymous)],
error: [Function (anonymous)],
timeout: [Function (anonymous)],
prefinish: [Function: requestOnPrefinish]
},
_eventsCount: 6,
_maxListeners: undefined,
outputData: [],
outputSize: 0,
writable: true,
destroyed: false,
_last: true,
chunkedEncoding: false,
shouldKeepAlive: false,
_defaultKeepAlive: true,
useChunkedEncodingByDefault: false,
sendDate: false,
_removedConnection: false,
_removedContLen: false,
_removedTE: false,
_contentLength: 0,
_hasBody: true,
_trailer: '',
finished: true,
_headerSent: true,
socket: TLSSocket {
_tlsOptions: [Object],
_secureEstablished: true,
_securePending: false,
_newSessionPending: false,
_controlReleased: true,
secureConnecting: false,
_SNICallback: null,
servername: 'api.tedee.com',
alpnProtocol: false,
authorized: true,
authorizationError: null,
encrypted: true,
_events: [Object: null prototype],
_eventsCount: 10,
connecting: false,
_hadError: false,
_parent: null,
_host: 'api.tedee.com',
_readableState: [ReadableState],
_maxListeners: undefined,
_writableState: [WritableState],
allowHalfOpen: false,
_sockname: null,
_pendingData: null,
_pendingEncoding: '',
server: undefined,
_server: null,
ssl: [TLSWrap],
_requestCert: true,
_rejectUnauthorized: true,
parser: null,
_httpMessage: [Circular *1],
[Symbol(res)]: [TLSWrap],
[Symbol(verified)]: true,
[Symbol(pendingSession)]: null,
[Symbol(async_id_symbol)]: 5351,
[Symbol(kHandle)]: [TLSWrap],
[Symbol(kSetNoDelay)]: false,
[Symbol(lastWriteQueueSize)]: 0,
[Symbol(timeout)]: null,
[Symbol(kBuffer)]: null,
[Symbol(kBufferCb)]: null,
[Symbol(kBufferGen)]: null,
[Symbol(kCapture)]: false,
[Symbol(kBytesRead)]: 0,
[Symbol(kBytesWritten)]: 0,
[Symbol(connect-options)]: [Object],
[Symbol(RequestTimeout)]: undefined
},
_header: 'GET /api/v1.15/my/lock HTTP/1.1\r\n' +
'Accept: application/json, text/plain, */*\r\n' +
'Authorization: Bearer xxxxxxxxxxxxxxxxxxxxxxxiOiJSUzI1NiIsImtpZCI6Ilg1ZVhrNHh5b2pORnVtMWtsMll0djhkbE5QNC1jNTdkTzZRR1RWQndhTmsifQ.eyJpc3MiOiJodHRwczovL3RlZGVlLmIyY2xvZ2luLmNvbS90ZnAvZTMxOTYxN2EtZjMyMi00Y2JhLTgxZWYtNGI1Yjk1MGU5MTcyL2IyY18xX3NpZ25pbl9yb3BjL3YyLjAvIiwiZXhwIjoxNjA4MjUwNDc1LCJuYmYiOjE2MDgyMzk2NzUsImF1ZCI6IjAyMTA2YjgyLTA1MjQtNGZkMy1hYzU3LWFmNzc0ZjM0MDk3OSIsImlkcCI6IkxvY2FsQWNjb3VudCIsIm9pZCI6ImNiNjY1ZTFiLxxxxxxxxxxxxiNjY1ZTFiLTZmY2QtNGY2Ny05ODQzLTcxN2ZlZWIzMWEzZCIsIm5hbWUiOiJESmF5IiwibmV3VXNlciI6ZmFsc2UsImVtYWlscyI6WyJkamF5QG1lLmNvbSJdLCJ0ZnAiOiJCMkNfMV9TaWduSW5fUm9wYyIsImF6cCI6IjAyMTA2YjgyLTA1MjQtNGZkMy1hYzU3LWFmNzc0ZjM0MDk3OSIsInZlciI6IjEuMCIsImlhdCI6MTYwODIzOTY3NX0.D1FcynocxrgCekS10JP8lCOteZy-jqiZgwxUCgM7EhtNjk5V9R_Mawq9i0AvXcow0oK5NynKpSMmt2Wp1KY2O9j-vRf-GiF-1xMTo97US6PnthkIyF41AEAcwdJSewBm6LJswa-DiGDAA2Le2nt1yg9hDcYamj1cuYoFqtatQ1YGK8NfnZZfnQzvwKE3EsT_Vby_uUiZgSwPZOwMkMMmYQ2qA0EFP8x0inni6mljwThps6xBBG3RGVCvGhmhn51XjpvYhyUsO7TKyXUANHKSsfi1GvRgNqA_RzSfyZKBLFT-bI2fgiUpyhhwETP5q62TSHWiaYvvcNY19gzXj3wPcg\r\n' +
'User-Agent: axios/0.19.2\r\n' +
'Host: api.tedee.com\r\n' +
'Connection: close\r\n' +
'\r\n',
_keepAliveTimeout: 0,
_onPendingData: [Function: noopPendingOutput],
agent: Agent {
_events: [Object: null prototype],
_eventsCount: 2,
_maxListeners: undefined,
defaultPort: 443,
protocol: 'https:',
options: [Object],
requests: {},
sockets: [Object],
freeSockets: {},
keepAliveMsecs: 1000,
keepAlive: false,
maxSockets: Infinity,
maxFreeSockets: 256,
scheduling: 'fifo',
maxTotalSockets: Infinity,
totalSocketCount: 4,
maxCachedSessions: 100,
_sessionCache: [Object],
[Symbol(kCapture)]: false
},
socketPath: undefined,
method: 'GET',
maxHeaderSize: undefined,
insecureHTTPParser: undefined,
path: '/api/v1.15/my/lock',
_ended: true,
res: IncomingMessage {
_readableState: [ReadableState],
_events: [Object: null prototype],
_eventsCount: 3,
_maxListeners: undefined,
socket: [TLSSocket],
httpVersionMajor: 1,
httpVersionMinor: 1,
httpVersion: '1.1',
complete: true,
rawHeaders: [Array],
rawTrailers: [],
aborted: false,
upgrade: false,
url: '',
method: null,
statusCode: 401,
statusMessage: 'Unauthorized',
client: [TLSSocket],
_consuming: true,
_dumped: false,
req: [Circular *1],
responseUrl: 'https://api.tedee.com/api/v1.15/my/lock',
redirects: [],
[Symbol(kCapture)]: false,
[Symbol(kHeaders)]: [Object],
[Symbol(kHeadersCount)]: 22,
[Symbol(kTrailers)]: null,
[Symbol(kTrailersCount)]: 0,
[Symbol(RequestTimeout)]: undefined
},
aborted: false,
timeoutCb: null,
upgradeOrConnect: false,
parser: null,
maxHeadersCount: null,
reusedSocket: false,
host: 'api.tedee.com',
protocol: 'https:',
_redirectable: Writable {
_writableState: [WritableState],
_events: [Object: null prototype],
_eventsCount: 2,
_maxListeners: undefined,
_options: [Object],
_redirectCount: 0,
_redirects: [],
_requestBodyLength: 0,
_requestBodyBuffers: [],
_onNativeResponse: [Function (anonymous)],
_currentRequest: [Circular *1],
_currentUrl: 'https://api.tedee.com/api/v1.15/my/lock',
[Symbol(kCapture)]: false
},
[Symbol(kCapture)]: false,
[Symbol(kNeedDrain)]: false,
[Symbol(corked)]: 0,
[Symbol(kOutHeaders)]: [Object: null prototype] {
accept: [Array],
authorization: [Array],
'user-agent': [Array],
host: [Array]
}
},
response: {
status: 401,
statusText: 'Unauthorized',
headers: {
'transfer-encoding': 'chunked',
'content-type': 'application/json',
server: 'Microsoft-IIS/10.0',
'www-authenticate': 'Bearer error="invalid_token", error_description="The signature key was not found", Bearer error="invalid_token", error_description="The signature key was not found"',
'request-context': 'appId=cid-v1:xxxxxxxxxxxx-bf23-4ab2-b999-bbd05e97fbc2',
'strict-transport-security': 'max-age=xxxxxxxxxxxx',
'x-correlation-id': 'xxxxxxxxxxxx-0001-f800-b63f-xxxxxxxxxxxx',
'x-powered-by': 'ASP.NET',
date: 'Thu, 17 Dec 2020 21:14:47 GMT',
connection: 'close'
},
config: {
url: 'https://api.tedee.com/api/v1.15/my/lock',
method: 'get',
headers: [Object],
transformRequest: [Array],
transformResponse: [Array],
timeout: 0,
adapter: [Function: httpAdapter],
xsrfCookieName: 'XSRF-TOKEN',
xsrfHeaderName: 'X-XSRF-TOKEN',
maxContentLength: -1,
validateStatus: [Function: validateStatus],
data: undefined
},
request: <ref *1> ClientRequest {
_events: [Object: null prototype],
_eventsCount: 6,
_maxListeners: undefined,
outputData: [],
outputSize: 0,
writable: true,
destroyed: false,
_last: true,
chunkedEncoding: false,
shouldKeepAlive: false,
_defaultKeepAlive: true,
useChunkedEncodingByDefault: false,
sendDate: false,
_removedConnection: false,
_removedContLen: false,
_removedTE: false,
_contentLength: 0,
_hasBody: true,
_trailer: '',
finished: true,
_headerSent: true,
socket: [TLSSocket],
_header: 'GET /api/v1.15/my/lock HTTP/1.1\r\n' +
'Accept: application/json, text/plain, */*\r\n' +
'Authorization: Bearer xxxxxxxxxxxxxxxxxxxxxxxiOiJSUzI1NiIsImtpZCI6Ilg1ZVhrNHh5b2pORnVtMWtsMll0djhkbE5QNC1jNTdkTzZRR1RWQndhTmsifQ.eyJpc3MiOiJodHRwczovL3RlZGVlLmIyY2xvZ2luLmNvbS90ZnAvZTMxOTYxN2EtZjMyMi00Y2JhLTgxZWYtNGI1Yjk1MGU5MTcyL2IyY18xX3NpZ25pbl9yb3BjL3YyLjAvIiwiZXhwIjoxNjA4MjUwNDc1LCJuYmYiOjE2MDgyMzk2NzUsImF1ZCI6IjAyMTA2YjgyLTA1MjQtNGZkMy1hYzU3LWFmNzc0ZjM0MDk3OSIsImlkcCI6IkxvY2FsQWNjb3VudCIsIm9pZCI6ImNiNjY1ZTFiLxxxxxxxxxxxxiNjY1ZTFiLTZmY2QtNGY2Ny05ODQzLTcxN2ZlZWIzMWEzZCIsIm5hbWUiOiJESmF5IiwibmV3VXNlciI6ZmFsc2UsImVtYWlscyI6WyJkamF5QG1lLmNvbSJdLCJ0ZnAiOiJCMkNfMV9TaWduSW5fUm9wYyIsImF6cCI6IjAyMTA2YjgyLTA1MjQtNGZkMy1hYzU3LWFmNzc0ZjM0MDk3OSIsInZlciI6IjEuMCIsImlhdCI6MTYwODIzOTY3NX0.D1FcynocxrgCekS10JP8lCOteZy-jqiZgwxUCgM7EhtNjk5V9R_Mawq9i0AvXcow0oK5NynKpSMmt2Wp1KY2O9j-vRf-GiF-1xMTo97US6PnthkIyF41AEAcwdJSewBm6LJswa-DiGDAA2Le2nt1yg9hDcYamj1cuYoFqtatQ1YGK8NfnZZfnQzvwKE3EsT_Vby_uUiZgSwPZOwMkMMmYQ2qA0EFP8x0inni6mljwThps6xBBG3RGVCvGhmhn51XjpvYhyUsO7TKyXUANHKSsfi1GvRgNqA_RzSfyZKBLFT-bI2fgiUpyhhwETP5q62TSHWiaYvvcNY19gzXj3wPcg\r\n' +
'User-Agent: axios/0.19.2\r\n' +
'Host: api.tedee.com\r\n' +
'Connection: close\r\n' +
'\r\n',
_keepAliveTimeout: 0,
_onPendingData: [Function: noopPendingOutput],
agent: [Agent],
socketPath: undefined,
method: 'GET',
maxHeaderSize: undefined,
insecureHTTPParser: undefined,
path: '/api/v1.15/my/lock',
_ended: true,
res: [IncomingMessage],
aborted: false,
timeoutCb: null,
upgradeOrConnect: false,
parser: null,
maxHeadersCount: null,
reusedSocket: false,
host: 'api.tedee.com',
protocol: 'https:',
_redirectable: [Writable],
[Symbol(kCapture)]: false,
[Symbol(kNeedDrain)]: false,
[Symbol(corked)]: 0,
[Symbol(kOutHeaders)]: [Object: null prototype]
},
data: { success: false, errorMessages: [Array], statusCode: 401 }
},
isAxiosError: true,
toJSON: [Function (anonymous)]
}
[17.12.2020, 22:14:47] Got SIGTERM, shutting down Homebridge...
@lukasroegner I again had this yesterday 07.01.2021 around 18:00 and 23:00 the last time.
[7.1.2021, 23:05:43] [HB Supervisor] Homebridge Process Ended. Code: 143, Signal: null
[7.1.2021, 23:05:48] [HB Supervisor] Restarting Homebridge...
[7.1.2021, 23:05:48] [HB Supervisor] Starting Homebridge with extra flags: -I
[7.1.2021, 23:05:48] [HB Supervisor] Started Homebridge v1.2.5 with PID: 7853
[7.1.2021, 23:05:49] Loaded config.json with 3 accessories and 8 platforms.
[7.1.2021, 23:05:56] ---
[7.1.2021, 23:05:56] Loaded plugin: homebridge-tedee@0.2.1
[7.1.2021, 23:05:56] Registering platform 'homebridge-tedee.TedeePlatform'
[7.1.2021, 23:05:56] ---
[7.1.2021, 23:05:56] Loading 8 platforms...
[7.1.2021, 23:05:56] [TedeePlatform] Initializing TedeePlatform platform...
[7.1.2021, 23:05:56] [TedeePlatform] Initializing platform...
Setup Payload:
X-HM://0023IV4HADIAK
Enter this code with your HomeKit app on your iOS device to pair with Homebridge:
┌────────────┐
│ 032-45-678 │
└────────────┘
[7.1.2021, 23:05:57] Homebridge v1.2.5 is running on port 51842.
[7.1.2021, 23:06:17] [TedeePlatform] Error while retrieving access token: Error: getaddrinfo EAI_AGAIN tedee.b2clogin.com
[7.1.2021, 23:07:17] [TedeePlatform] Error while retrieving access token: Error: getaddrinfo EAI_AGAIN tedee.b2clogin.com
[7.1.2021, 23:07:57] [TedeePlatform] Error while retrieving access token: Error: getaddrinfo EAI_AGAIN tedee.b2clogin.com
Error: getaddrinfo EAI_AGAIN tedee.b2clogin.com
at GetAddrInfoReqWrap.onlookup [as oncomplete] (dns.js:67:26)
{
errno: -3001,
code: 'EAI_AGAIN',
syscall: 'getaddrinfo',
hostname: 'tedee.b2clogin.com',
config: {
url: 'https://tedee.b2clogin.com/tedee.onmicrosoft.com/oauth2/v2.0/token?p=B2C_1_SignIn_Ropc',
method: 'post',
data: 'grant_type=password&username=xxx%40xxx.com&password=xxxxx-xxxxx-xxxxx&scope=openid%xxx-0524-4fd3-xxx-xxx&client_id=xxx-xxx-xxx-xxx-xxx&response_type=token%20id_token',
headers: {
Accept: 'application/json, text/plain, */*',
'Content-Type': 'application/x-www-form-urlencoded',
'User-Agent': 'axios/0.19.2',
'Content-Length': 202
},
transformRequest: [ [Function: transformRequest] ],
transformResponse: [ [Function: transformResponse] ],
timeout: 0,
adapter: [Function: httpAdapter],
xsrfCookieName: 'XSRF-TOKEN',
xsrfHeaderName: 'X-XSRF-TOKEN',
maxContentLength: -1,
validateStatus: [Function: validateStatus]
},
request: <ref *1> Writable {
_writableState: WritableState {
objectMode: false,
highWaterMark: 16384,
finalCalled: false,
needDrain: false,
ending: false,
ended: false,
finished: false,
destroyed: false,
decodeStrings: true,
defaultEncoding: 'utf8',
length: 0,
writing: false,
corked: 0,
sync: true,
bufferProcessing: false,
onwrite: [Function: bound onwrite],
writecb: null,
writelen: 0,
afterWriteTickInfo: null,
buffered: [],
bufferedIndex: 0,
allBuffers: true,
allNoop: true,
pendingcb: 0,
prefinished: false,
errorEmitted: false,
emitClose: true,
autoDestroy: true,
errored: null,
closed: false
},
_events: [Object: null prototype] {
response: [Function: handleResponse],
error: [Function: handleRequestError]
},
_eventsCount: 2,
_maxListeners: undefined,
_options: {
protocol: 'https:',
maxRedirects: 21,
maxBodyLength: 10485760,
path: '/tedee.onmicrosoft.com/oauth2/v2.0/token?p=B2C_1_SignIn_Ropc',
method: 'POST',
headers: [Object],
agent: undefined,
agents: [Object],
auth: undefined,
hostname: 'tedee.b2clogin.com',
port: null,
nativeProtocols: [Object],
pathname: '/tedee.onmicrosoft.com/oauth2/v2.0/token',
search: '?p=B2C_1_SignIn_Ropc'
},
_redirectCount: 0,
_redirects: [],
_requestBodyLength: 202,
_requestBodyBuffers: [ [Object] ],
_onNativeResponse: [Function (anonymous)],
_currentRequest: ClientRequest {
_events: [Object: null prototype],
_eventsCount: 6,
_maxListeners: undefined,
outputData: [],
outputSize: 0,
writable: true,
destroyed: false,
_last: true,
chunkedEncoding: false,
shouldKeepAlive: false,
_defaultKeepAlive: true,
useChunkedEncodingByDefault: true,
sendDate: false,
_removedConnection: false,
_removedContLen: false,
_removedTE: false,
_contentLength: null,
_hasBody: true,
_trailer: '',
finished: false,
_headerSent: true,
socket: [TLSSocket],
_header: 'POST /tedee.onmicrosoft.com/oauth2/v2.0/token?p=B2C_1_SignIn_Ropc HTTP/1.1\r\n' +
'Accept: application/json, text/plain, */*\r\n' +
'Content-Type: application/x-www-form-urlencoded\r\n' +
'User-Agent: axios/0.19.2\r\n' +
'Content-Length: 202\r\n' +
'Host: tedee.b2clogin.com\r\n' +
'Connection: close\r\n' +
'\r\n',
_keepAliveTimeout: 0,
_onPendingData: [Function: noopPendingOutput],
agent: [Agent],
socketPath: undefined,
method: 'POST',
maxHeaderSize: undefined,
insecureHTTPParser: undefined,
path: '/tedee.onmicrosoft.com/oauth2/v2.0/token?p=B2C_1_SignIn_Ropc',
_ended: false,
res: null,
aborted: false,
timeoutCb: null,
upgradeOrConnect: false,
parser: null,
maxHeadersCount: null,
reusedSocket: false,
host: 'tedee.b2clogin.com',
protocol: 'https:',
_redirectable: [Circular *1],
[Symbol(kCapture)]: false,
[Symbol(kNeedDrain)]: false,
[Symbol(corked)]: 0,
[Symbol(kOutHeaders)]: [Object: null prototype]
},
_currentUrl: 'https://tedee.b2clogin.com/tedee.onmicrosoft.com/oauth2/v2.0/token?p=B2C_1_SignIn_Ropc',
[Symbol(kCapture)]: false
},
response: undefined,
isAxiosError: true,
toJSON: [Function (anonymous)]
}
Error: getaddrinfo EAI_AGAIN tedee.b2clogin.com
at GetAddrInfoReqWrap.onlookup [as oncomplete] (dns.js:67:26)
{
errno: -3001,
code: 'EAI_AGAIN',
syscall: 'getaddrinfo',
hostname: 'tedee.b2clogin.com',
config: {
url: 'https://tedee.b2clogin.com/tedee.onmicrosoft.com/oauth2/v2.0/token?p=B2C_1_SignIn_Ropc',
method: 'post',
data: 'grant_type=password&username=xxx%40xxx.com&password=xxxxx-xxxxx-xxxxx&scope=openid%xxx-xxx-xxx-xxx-xxx&client_id=xxx-xxx-xxx-xxx-xxx&response_type=token%20id_token',
headers: {
Accept: 'application/json, text/plain, */*',
'Content-Type': 'application/x-www-form-urlencoded',
'User-Agent': 'axios/0.19.2',
'Content-Length': 202
},
transformRequest: [ [Function: transformRequest] ],
transformResponse: [ [Function: transformResponse] ],
timeout: 0,
adapter: [Function: httpAdapter],
xsrfCookieName: 'XSRF-TOKEN',
xsrfHeaderName: 'X-XSRF-TOKEN',
maxContentLength: -1,
validateStatus: [Function: validateStatus]
},
request: <ref *1> Writable {
_writableState: WritableState {
objectMode: false,
highWaterMark: 16384,
finalCalled: false,
needDrain: false,
ending: false,
ended: false,
finished: false,
destroyed: false,
decodeStrings: true,
defaultEncoding: 'utf8',
length: 0,
writing: false,
corked: 0,
sync: true,
bufferProcessing: false,
onwrite: [Function: bound onwrite],
writecb: null,
writelen: 0,
afterWriteTickInfo: null,
buffered: [],
bufferedIndex: 0,
allBuffers: true,
allNoop: true,
pendingcb: 0,
prefinished: false,
errorEmitted: false,
emitClose: true,
autoDestroy: true,
errored: null,
closed: false
},
_events: [Object: null prototype] {
response: [Function: handleResponse],
error: [Function: handleRequestError]
},
_eventsCount: 2,
_maxListeners: undefined,
_options: {
protocol: 'https:',
maxRedirects: 21,
maxBodyLength: 10485760,
path: '/tedee.onmicrosoft.com/oauth2/v2.0/token?p=B2C_1_SignIn_Ropc',
method: 'POST',
headers: [Object],
agent: undefined,
agents: [Object],
auth: undefined,
hostname: 'tedee.b2clogin.com',
port: null,
nativeProtocols: [Object],
pathname: '/tedee.onmicrosoft.com/oauth2/v2.0/token',
search: '?p=B2C_1_SignIn_Ropc'
},
_redirectCount: 0,
_redirects: [],
_requestBodyLength: 202,
_requestBodyBuffers: [ [Object] ],
_onNativeResponse: [Function (anonymous)],
_currentRequest: ClientRequest {
_events: [Object: null prototype],
_eventsCount: 6,
_maxListeners: undefined,
outputData: [],
outputSize: 0,
writable: true,
destroyed: false,
_last: true,
chunkedEncoding: false,
shouldKeepAlive: false,
_defaultKeepAlive: true,
useChunkedEncodingByDefault: true,
sendDate: false,
_removedConnection: false,
_removedContLen: false,
_removedTE: false,
_contentLength: null,
_hasBody: true,
_trailer: '',
finished: false,
_headerSent: true,
socket: [TLSSocket],
_header: 'POST /tedee.onmicrosoft.com/oauth2/v2.0/token?p=B2C_1_SignIn_Ropc HTTP/1.1\r\n' +
'Accept: application/json, text/plain, */*\r\n' +
'Content-Type: application/x-www-form-urlencoded\r\n' +
'User-Agent: axios/0.19.2\r\n' +
'Content-Length: 202\r\n' +
'Host: tedee.b2clogin.com\r\n' +
'Connection: close\r\n' +
'\r\n',
_keepAliveTimeout: 0,
_onPendingData: [Function: noopPendingOutput],
agent: [Agent],
socketPath: undefined,
method: 'POST',
maxHeaderSize: undefined,
insecureHTTPParser: undefined,
path: '/tedee.onmicrosoft.com/oauth2/v2.0/token?p=B2C_1_SignIn_Ropc',
_ended: false,
res: null,
aborted: false,
timeoutCb: null,
upgradeOrConnect: false,
parser: null,
maxHeadersCount: null,
reusedSocket: false,
host: 'tedee.b2clogin.com',
protocol: 'https:',
_redirectable: [Circular *1],
[Symbol(kCapture)]: false,
[Symbol(kNeedDrain)]: false,
[Symbol(corked)]: 0,
[Symbol(kOutHeaders)]: [Object: null prototype]
},
_currentUrl: 'https://tedee.b2clogin.com/tedee.onmicrosoft.com/oauth2/v2.0/token?p=B2C_1_SignIn_Ropc',
[Symbol(kCapture)]: false
},
response: undefined,
isAxiosError: true,
toJSON: [Function (anonymous)]
}
After this 23:10 it seemed to work again.
[7.1.2021, 23:08:30] Homebridge v1.2.5 is running on port 51842.
[7.1.2021, 23:08:50] [TedeePlatform] Error while retrieving access token: Error: getaddrinfo EAI_AGAIN tedee.b2clogin.com
��������������������������������������������������������������������������������������������������������������������������������������������������[7.1.2021, 22:17:16] [HB Supervisor] OS: Linux 4.19.66-v7+ arm
[7.1.2021, 22:17:16] [HB Supervisor] Node.js v14.15.4 /usr/local/bin/node
[7.1.2021, 22:17:16] [HB Supervisor] Homebridge Path: /usr/local/lib/node_modules/homebridge/bin/homebridge
[7.1.2021, 22:17:16] [HB Supervisor] UI Path: /usr/local/lib/node_modules/homebridge-config-ui-x/dist/bin/standalone.js
[7.1.2021, 22:17:16] [HB Supervisor] Starting Homebridge with extra flags: -I
[7.1.2021, 22:17:16] [HB Supervisor] Started Homebridge v1.2.5 with PID: 725
[7.1.2021, 22:17:19] Loaded config.json with 3 accessories and 8 platforms.
[7.1.2021, 23:10:32] Loaded plugin: homebridge-tedee@0.2.1
[7.1.2021, 23:10:32] Registering platform 'homebridge-tedee.TedeePlatform'
[7.1.2021, 23:10:32] ---
[7.1.2021, 23:10:32] Loading 8 platforms...
[7.1.2021, 23:10:32] [TedeePlatform] Initializing TedeePlatform platform...
[7.1.2021, 23:10:32] Loading 3 accessories...
[7.1.2021, 23:10:33] [TedeePlatform] Initializing platform...
Setup Payload:
X-HM://0023IV4HADIAK
Enter this code with your HomeKit app on your iOS device to pair with Homebridge:
┌────────────┐
│ 032-45-678 │
└────────────┘
[7.1.2021, 23:10:33] Homebridge v1.2.5 is running on port 51842.
[7.1.2021, 23:10:35] [TedeePlatform] [Wohnungstür] Initializing...
[7.1.2021, 23:10:35] [TedeePlatform] [Wohnungstür] Adding lock service
[7.1.2021, 23:10:35] [TedeePlatform] [Wohnungstür] Adding latch service
[7.1.2021, 23:10:35] [TedeePlatform] [Wohnungstür] Adding battery service
@lukasroegner I again had this yesterday 07.01.2021 around 18:00 and 23:00 the last time.
[7.1.2021, 23:05:43] [HB Supervisor] Homebridge Process Ended. Code: 143, Signal: null
[7.1.2021, 23:05:48] [HB Supervisor] Restarting Homebridge...
[7.1.2021, 23:05:48] [HB Supervisor] Starting Homebridge with extra flags: -I
[7.1.2021, 23:05:48] [HB Supervisor] Started Homebridge v1.2.5 with PID: 7853
[7.1.2021, 23:05:49] Loaded config.json with 3 accessories and 8 platforms.
[7.1.2021, 23:05:56] ---
[7.1.2021, 23:05:56] Loaded plugin: homebridge-tedee@0.2.1
[7.1.2021, 23:05:56] Registering platform 'homebridge-tedee.TedeePlatform'
[7.1.2021, 23:05:56] ---
[7.1.2021, 23:05:56] Loading 8 platforms...
[7.1.2021, 23:05:56] [TedeePlatform] Initializing TedeePlatform platform...
[7.1.2021, 23:05:56] [TedeePlatform] Initializing platform...
Setup Payload:
X-HM://0023IV4HADIAK
Enter this code with your HomeKit app on your iOS device to pair with Homebridge:
┌────────────┐
│ 032-45-678 │
└────────────┘
[7.1.2021, 23:05:57] Homebridge v1.2.5 is running on port 51842.
[7.1.2021, 23:06:17] [TedeePlatform] Error while retrieving access token: Error: getaddrinfo EAI_AGAIN tedee.b2clogin.com
[7.1.2021, 23:07:17] [TedeePlatform] Error while retrieving access token: Error: getaddrinfo EAI_AGAIN tedee.b2clogin.com
[7.1.2021, 23:07:57] [TedeePlatform] Error while retrieving access token: Error: getaddrinfo EAI_AGAIN tedee.b2clogin.com
Error: getaddrinfo EAI_AGAIN tedee.b2clogin.com
at GetAddrInfoReqWrap.onlookup [as oncomplete] (dns.js:67:26)
{
errno: -3001,
code: 'EAI_AGAIN',
syscall: 'getaddrinfo',
hostname: 'tedee.b2clogin.com',
config: {
url: 'https://tedee.b2clogin.com/tedee.onmicrosoft.com/oauth2/v2.0/token?p=B2C_1_SignIn_Ropc',
method: 'post',
data: 'grant_type=password&username=xxx%40xxx.com&password=xxxxx-xxxxx-xxxxx&scope=openid%xxx-0524-4fd3-xxx-xxx&client_id=xxx-xxx-xxx-xxx-xxx&response_type=token%20id_token',
headers: {
Accept: 'application/json, text/plain, */*',
'Content-Type': 'application/x-www-form-urlencoded',
'User-Agent': 'axios/0.19.2',
'Content-Length': 202
},
transformRequest: [ [Function: transformRequest] ],
transformResponse: [ [Function: transformResponse] ],
timeout: 0,
adapter: [Function: httpAdapter],
xsrfCookieName: 'XSRF-TOKEN',
xsrfHeaderName: 'X-XSRF-TOKEN',
maxContentLength: -1,
validateStatus: [Function: validateStatus]
},
request: <ref *1> Writable {
_writableState: WritableState {
objectMode: false,
highWaterMark: 16384,
finalCalled: false,
needDrain: false,
ending: false,
ended: false,
finished: false,
destroyed: false,
decodeStrings: true,
defaultEncoding: 'utf8',
length: 0,
writing: false,
corked: 0,
sync: true,
bufferProcessing: false,
onwrite: [Function: bound onwrite],
writecb: null,
writelen: 0,
afterWriteTickInfo: null,
buffered: [],
bufferedIndex: 0,
allBuffers: true,
allNoop: true,
pendingcb: 0,
prefinished: false,
errorEmitted: false,
emitClose: true,
autoDestroy: true,
errored: null,
closed: false
},
_events: [Object: null prototype] {
response: [Function: handleResponse],
error: [Function: handleRequestError]
},
_eventsCount: 2,
_maxListeners: undefined,
_options: {
protocol: 'https:',
maxRedirects: 21,
maxBodyLength: 10485760,
path: '/tedee.onmicrosoft.com/oauth2/v2.0/token?p=B2C_1_SignIn_Ropc',
method: 'POST',
headers: [Object],
agent: undefined,
agents: [Object],
auth: undefined,
hostname: 'tedee.b2clogin.com',
port: null,
nativeProtocols: [Object],
pathname: '/tedee.onmicrosoft.com/oauth2/v2.0/token',
search: '?p=B2C_1_SignIn_Ropc'
},
_redirectCount: 0,
_redirects: [],
_requestBodyLength: 202,
_requestBodyBuffers: [ [Object] ],
_onNativeResponse: [Function (anonymous)],
_currentRequest: ClientRequest {
_events: [Object: null prototype],
_eventsCount: 6,
_maxListeners: undefined,
outputData: [],
outputSize: 0,
writable: true,
destroyed: false,
_last: true,
chunkedEncoding: false,
shouldKeepAlive: false,
_defaultKeepAlive: true,
useChunkedEncodingByDefault: true,
sendDate: false,
_removedConnection: false,
_removedContLen: false,
_removedTE: false,
_contentLength: null,
_hasBody: true,
_trailer: '',
finished: false,
_headerSent: true,
socket: [TLSSocket],
_header: 'POST /tedee.onmicrosoft.com/oauth2/v2.0/token?p=B2C_1_SignIn_Ropc HTTP/1.1\r\n' +
'Accept: application/json, text/plain, */*\r\n' +
'Content-Type: application/x-www-form-urlencoded\r\n' +
'User-Agent: axios/0.19.2\r\n' +
'Content-Length: 202\r\n' +
'Host: tedee.b2clogin.com\r\n' +
'Connection: close\r\n' +
'\r\n',
_keepAliveTimeout: 0,
_onPendingData: [Function: noopPendingOutput],
agent: [Agent],
socketPath: undefined,
method: 'POST',
maxHeaderSize: undefined,
insecureHTTPParser: undefined,
path: '/tedee.onmicrosoft.com/oauth2/v2.0/token?p=B2C_1_SignIn_Ropc',
_ended: false,
res: null,
aborted: false,
timeoutCb: null,
upgradeOrConnect: false,
parser: null,
maxHeadersCount: null,
reusedSocket: false,
host: 'tedee.b2clogin.com',
protocol: 'https:',
_redirectable: [Circular *1],
[Symbol(kCapture)]: false,
[Symbol(kNeedDrain)]: false,
[Symbol(corked)]: 0,
[Symbol(kOutHeaders)]: [Object: null prototype]
},
_currentUrl: 'https://tedee.b2clogin.com/tedee.onmicrosoft.com/oauth2/v2.0/token?p=B2C_1_SignIn_Ropc',
[Symbol(kCapture)]: false
},
response: undefined,
isAxiosError: true,
toJSON: [Function (anonymous)]
}
Error: getaddrinfo EAI_AGAIN tedee.b2clogin.com
at GetAddrInfoReqWrap.onlookup [as oncomplete] (dns.js:67:26)
{
errno: -3001,
code: 'EAI_AGAIN',
syscall: 'getaddrinfo',
hostname: 'tedee.b2clogin.com',
config: {
url: 'https://tedee.b2clogin.com/tedee.onmicrosoft.com/oauth2/v2.0/token?p=B2C_1_SignIn_Ropc',
method: 'post',
data: 'grant_type=password&username=xxx%40xxx.com&password=xxxxx-xxxxx-xxxxx&scope=openid%xxx-xxx-xxx-xxx-xxx&client_id=xxx-xxx-xxx-xxx-xxx&response_type=token%20id_token',
headers: {
Accept: 'application/json, text/plain, */*',
'Content-Type': 'application/x-www-form-urlencoded',
'User-Agent': 'axios/0.19.2',
'Content-Length': 202
},
transformRequest: [ [Function: transformRequest] ],
transformResponse: [ [Function: transformResponse] ],
timeout: 0,
adapter: [Function: httpAdapter],
xsrfCookieName: 'XSRF-TOKEN',
xsrfHeaderName: 'X-XSRF-TOKEN',
maxContentLength: -1,
validateStatus: [Function: validateStatus]
},
request: <ref *1> Writable {
_writableState: WritableState {
objectMode: false,
highWaterMark: 16384,
finalCalled: false,
needDrain: false,
ending: false,
ended: false,
finished: false,
destroyed: false,
decodeStrings: true,
defaultEncoding: 'utf8',
length: 0,
writing: false,
corked: 0,
sync: true,
bufferProcessing: false,
onwrite: [Function: bound onwrite],
writecb: null,
writelen: 0,
afterWriteTickInfo: null,
buffered: [],
bufferedIndex: 0,
allBuffers: true,
allNoop: true,
pendingcb: 0,
prefinished: false,
errorEmitted: false,
emitClose: true,
autoDestroy: true,
errored: null,
closed: false
},
_events: [Object: null prototype] {
response: [Function: handleResponse],
error: [Function: handleRequestError]
},
_eventsCount: 2,
_maxListeners: undefined,
_options: {
protocol: 'https:',
maxRedirects: 21,
maxBodyLength: 10485760,
path: '/tedee.onmicrosoft.com/oauth2/v2.0/token?p=B2C_1_SignIn_Ropc',
method: 'POST',
headers: [Object],
agent: undefined,
agents: [Object],
auth: undefined,
hostname: 'tedee.b2clogin.com',
port: null,
nativeProtocols: [Object],
pathname: '/tedee.onmicrosoft.com/oauth2/v2.0/token',
search: '?p=B2C_1_SignIn_Ropc'
},
_redirectCount: 0,
_redirects: [],
_requestBodyLength: 202,
_requestBodyBuffers: [ [Object] ],
_onNativeResponse: [Function (anonymous)],
_currentRequest: ClientRequest {
_events: [Object: null prototype],
_eventsCount: 6,
_maxListeners: undefined,
outputData: [],
outputSize: 0,
writable: true,
destroyed: false,
_last: true,
chunkedEncoding: false,
shouldKeepAlive: false,
_defaultKeepAlive: true,
useChunkedEncodingByDefault: true,
sendDate: false,
_removedConnection: false,
_removedContLen: false,
_removedTE: false,
_contentLength: null,
_hasBody: true,
_trailer: '',
finished: false,
_headerSent: true,
socket: [TLSSocket],
_header: 'POST /tedee.onmicrosoft.com/oauth2/v2.0/token?p=B2C_1_SignIn_Ropc HTTP/1.1\r\n' +
'Accept: application/json, text/plain, */*\r\n' +
'Content-Type: application/x-www-form-urlencoded\r\n' +
'User-Agent: axios/0.19.2\r\n' +
'Content-Length: 202\r\n' +
'Host: tedee.b2clogin.com\r\n' +
'Connection: close\r\n' +
'\r\n',
_keepAliveTimeout: 0,
_onPendingData: [Function: noopPendingOutput],
agent: [Agent],
socketPath: undefined,
method: 'POST',
maxHeaderSize: undefined,
insecureHTTPParser: undefined,
path: '/tedee.onmicrosoft.com/oauth2/v2.0/token?p=B2C_1_SignIn_Ropc',
_ended: false,
res: null,
aborted: false,
timeoutCb: null,
upgradeOrConnect: false,
parser: null,
maxHeadersCount: null,
reusedSocket: false,
host: 'tedee.b2clogin.com',
protocol: 'https:',
_redirectable: [Circular *1],
[Symbol(kCapture)]: false,
[Symbol(kNeedDrain)]: false,
[Symbol(corked)]: 0,
[Symbol(kOutHeaders)]: [Object: null prototype]
},
_currentUrl: 'https://tedee.b2clogin.com/tedee.onmicrosoft.com/oauth2/v2.0/token?p=B2C_1_SignIn_Ropc',
[Symbol(kCapture)]: false
},
response: undefined,
isAxiosError: true,
toJSON: [Function (anonymous)]
}
After this 23:10 it seemed to work again.
[7.1.2021, 23:08:30] Homebridge v1.2.5 is running on port 51842.
[7.1.2021, 23:08:50] [TedeePlatform] Error while retrieving access token: Error: getaddrinfo EAI_AGAIN tedee.b2clogin.com
��������������������������������������������������������������������������������������������������������������������������������������������������[7.1.2021, 22:17:16] [HB Supervisor] OS: Linux 4.19.66-v7+ arm
[7.1.2021, 22:17:16] [HB Supervisor] Node.js v14.15.4 /usr/local/bin/node
[7.1.2021, 22:17:16] [HB Supervisor] Homebridge Path: /usr/local/lib/node_modules/homebridge/bin/homebridge
[7.1.2021, 22:17:16] [HB Supervisor] UI Path: /usr/local/lib/node_modules/homebridge-config-ui-x/dist/bin/standalone.js
[7.1.2021, 22:17:16] [HB Supervisor] Starting Homebridge with extra flags: -I
[7.1.2021, 22:17:16] [HB Supervisor] Started Homebridge v1.2.5 with PID: 725
[7.1.2021, 22:17:19] Loaded config.json with 3 accessories and 8 platforms.
[7.1.2021, 23:10:32] Loaded plugin: homebridge-tedee@0.2.1
[7.1.2021, 23:10:32] Registering platform 'homebridge-tedee.TedeePlatform'
[7.1.2021, 23:10:32] ---
[7.1.2021, 23:10:32] Loading 8 platforms...
[7.1.2021, 23:10:32] [TedeePlatform] Initializing TedeePlatform platform...
[7.1.2021, 23:10:32] Loading 3 accessories...
[7.1.2021, 23:10:33] [TedeePlatform] Initializing platform...
Setup Payload:
X-HM://0023IV4HADIAK
Enter this code with your HomeKit app on your iOS device to pair with Homebridge:
┌────────────┐
│ 032-45-678 │
└────────────┘
[7.1.2021, 23:10:33] Homebridge v1.2.5 is running on port 51842.
[7.1.2021, 23:10:35] [TedeePlatform] [Wohnungstür] Initializing...
[7.1.2021, 23:10:35] [TedeePlatform] [Wohnungstür] Adding lock service
[7.1.2021, 23:10:35] [TedeePlatform] [Wohnungstür] Adding latch service
[7.1.2021, 23:10:35] [TedeePlatform] [Wohnungstür] Adding battery service
Do you have scheduled restarts enabled for Homebridge?
No. Not scheduled. But homebridge starts again when a error occurred.
However, since 19 Days there was no such incident again. 🤷🏼♂️
Do you have scheduled restarts enabled for Homebridge?
No. Not scheduled. But homebridge starts again when a error occurred.
However, since 19 Days there was no such incident again. 🤷🏼♂️
|
gharchive/issue
| 2020-12-17T21:20:27 |
2025-04-01T06:39:27.841619
|
{
"authors": [
"DJay-X"
],
"repo": "lukasroegner/homebridge-tedee",
"url": "https://github.com/lukasroegner/homebridge-tedee/issues/9",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1040072620
|
Documentation plans
Hi Luke, nice to meet you. I'm new to Workers (and still exploring) and found your library from watching an episode of Fauna Live where the guest was Obinna Ekwuno. They used this library to build their application.
In that video, they mentioned setting the type in the wrangler.toml file to webpack, but I cannot find that reference anywhere here so I'm a little confused. I know this library is semi-"just starting out" but was curious if you had plans to do some documentation for other dumb-dumbs like myself who are a little slow and need hand holding.
You have great examples, it's just a little hard for me to visualize going from wrangler init to one of those examples. Would also love to contribute if you're up for it.
Thanks!
People use miniflare with esbuild. I know there's a decent starter out there for that, if not already official one. They should have that straight up in getting started.
Hey, so worktop has no relation or dependence on wrangler actually. You can build a worktop project any number of ways, including the worktop.build package (#94). Wrangler includes a webpack integration which only kicks in if you have type = "webpack" at the top, but my strong personal recommendation is to avoid this as it means you really have no idea what's going into your final worker.
Separately, Wrangler will deploy file(s) to your CF account as the Workers deploy step. There are other tools out there that does this too, including cfw which is what the /examples use. You can read more about Wrangler configuration here though.
An example wranger.toml file may look like this:
name = "example-worker"
type = "javascript"
account_id = "..."
zone_id = "..."
[build]
command = "npm run build"
[build.upload]
format = "service-worker"
# or, format = "modules"
# but "modules" requires addl config
Closing in favor of #61 as there will be a worktop-specific project scaffolder in near future.
|
gharchive/issue
| 2021-10-30T07:03:35 |
2025-04-01T06:39:27.869897
|
{
"authors": [
"grantholle",
"janat08",
"lukeed"
],
"repo": "lukeed/worktop",
"url": "https://github.com/lukeed/worktop/issues/103",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
582078589
|
vscodium@1.43.0: hash check failed
Updating 'vscodium' (1.42.1 -> 1.43.0)
Downloading new version
VSCodium-win32-x64-1.43.0.zip (82.9 MB) [=========================================================================================] 100%
Checking hash of VSCodium-win32-x64-1.43.0.zip ... ERROR Hash check failed!
App: extras/vscodium
URL: https://github.com/VSCodium/vscodium/releases/download/1.43.0/VSCodium-win32-x64-1.43.0.zip
First bytes: 50 4B 03 04 14 00 00 00
Expected: f81b637f0a680ce2aa2616564b34ff0654609e7a28c1830147d11c685271319c
Actual: e6d0880abafa2ab94589e61ad880d88d6f6cec34e6e202c66e15f532c4f75b53
Still failed hash checking, local cache has been cleaned.
Updating 'vscodium' (1.42.1 -> 1.43.0)
Downloading new version
VSCodium-win32-x64-1.43.0.zip (82.9 MB) [=========================================================================================] 100%
Checking hash of VSCodium-win32-x64-1.43.0.zip ... ERROR Hash check failed!
App: extras/vscodium
URL: https://github.com/VSCodium/vscodium/releases/download/1.43.0/VSCodium-win32-x64-1.43.0.zip
First bytes: 50 4B 03 04 14 00 00 00
Expected: f81b637f0a680ce2aa2616564b34ff0654609e7a28c1830147d11c685271319c
Actual: e6d0880abafa2ab94589e61ad880d88d6f6cec34e6e202c66e15f532c4f75b53
Please try again or create a new issue by using the following link and paste your console output:
https://github.com/lukesampson/scoop-extras/issues/new?title=vscodium%401.43.0%3A+hash+check+failed
...
scoop update; scoop update vscodium -f
thanks, update successfully!
|
gharchive/issue
| 2020-03-16T07:55:34 |
2025-04-01T06:39:27.879416
|
{
"authors": [
"Ash258",
"hago"
],
"repo": "lukesampson/scoop-extras",
"url": "https://github.com/lukesampson/scoop-extras/issues/3707",
"license": "Unlicense",
"license_type": "permissive",
"license_source": "github-api"
}
|
389448867
|
Use Windows binary of openssl over mingw binaries
Currently a mingw version of openssl and by default it checks for configuration files in /usr/share/. When openssl is ran programatically (like in https://www.npmjs.com/package/pem) we might run into "path format" issues.
Instead use a windows binary (https://slproweb.com/products/Win32OpenSSL.html)
https://github.com/lukesampson/scoop/blob/master/bucket/openssl-slp.json
Sorry, thanks.
If openssl has issues, perhaps we should rename openssl-slp.json to openssl.json, and rename openssl.json to openssl-mingw.json?
mingw version of openssl is using UNIX pathnames /usr/bin/... and fails immediately (which is a bumer for starters and packages) but the slp version uses the env variable. IDK but the mingw version has any such env features, it's not even much useful in WSL too.
Renaming might break stuff for old users but the slp version is the one everyone should be using.
If no one disagrees, I think we should make the switch. I only use the SLP version.
Can I make a PR then?
Sure, thanks!
|
gharchive/issue
| 2018-12-10T19:21:49 |
2025-04-01T06:39:27.883747
|
{
"authors": [
"r0mflip",
"rasa"
],
"repo": "lukesampson/scoop",
"url": "https://github.com/lukesampson/scoop/issues/2877",
"license": "unlicense",
"license_type": "permissive",
"license_source": "bigquery"
}
|
54700717
|
Installer fails
I'm having trouble installing this server on one of our Windows Server 2008 R2 (64bit) servers. The server appears to only have .Net 4.0 which I'm assuming is the problem. Is there a way to get this app to work on .Net 4.0 instead of 4.5?
Running a transacted installation.
Beginning the Install phase of the installation.
See the contents of the log file for the E:\statsd.net-v1.5.0.0\statsdnet.exe assembly's progress.
The file is located at E:\statsd.net-v1.5.0.0\statsdnet.InstallLog.
An exception occurred during the Install phase.
System.InvalidOperationException: Unable to get installer types in the E:\statsd.net-v1.5.0.0\statsdnet.exe assembly.
The inner exception System.Reflection.ReflectionTypeLoadException was thrown with the following error message: Unable to load one or more of the requested types. Retrieve the LoaderExceptions property for more information..
The Rollback phase of the installation is beginning.
See the contents of the log file for the E:\statsd.net-v1.5.0.0\statsdnet.exe assembly's progress.
The file is located at E:\statsd.net-v1.5.0.0\statsdnet.InstallLog.
An exception occurred during the Rollback phase of the System.Configuration.Install.AssemblyInstaller installer.
System.InvalidOperationException: Unable to get installer types in the E:\statsd.net-v1.5.0.0\statsdnet.exe assembly.
The inner exception System.Reflection.ReflectionTypeLoadException was thrown with the following error message: Unable to load one or more of the requested types. Retrieve the LoaderExceptions property for more information..
An exception occurred during the Rollback phase of the installation. This exception will be ignored and the rollback will continue. However, the machine might not fully revert to its initial state after the rollback is complete.
The Rollback phase completed successfully.
The transacted install has completed.
Running a transacted installation.
Beginning the Install phase of the installation.
See the contents of the log file for the E:\statsd.net-v1.5.0.0\statsdnet.exe assembly's progress.
The file is located at E:\statsd.net-v1.5.0.0\statsdnet.InstallLog.
An exception occurred during the Install phase.
System.InvalidOperationException: Unable to get installer types in the E:\statsd.net-v1.5.0.0\statsdnet.exe assembly.
The inner exception System.Reflection.ReflectionTypeLoadException was thrown with the following error message: Unable to load one or more of the requested types. Retrieve the LoaderExceptions property for more information..
The Rollback phase of the installation is beginning.
See the contents of the log file for the E:\statsd.net-v1.5.0.0\statsdnet.exe assembly's progress.
The file is located at E:\statsd.net-v1.5.0.0\statsdnet.InstallLog.
An exception occurred during the Rollback phase of the System.Configuration.Install.AssemblyInstaller installer.
System.InvalidOperationException: Unable to get installer types in the E:\statsd.net-v1.5.0.0\statsdnet.exe assembly.
The inner exception System.Reflection.ReflectionTypeLoadException was thrown with the following error message: Unable to load one or more of the requested types. Retrieve the LoaderExceptions property for more information..
An exception occurred during the Rollback phase of the installation. This exception will be ignored and the rollback will continue. However, the machine might not fully revert to its initial state after the rollback is complete.
The Rollback phase completed successfully.
The transacted install has completed.
Hi Joe,
Sorry, but the service needs .net 4.5 and upwards to run.
Thanks,
Luke
Sent from my iPhone
On 18 Jan 2015, at 6:03 PM, Joe Phillips notifications@github.com wrote:
I'm having trouble installing this server on one of our Windows Server 2008 R2 (64bit) servers. The server appears to only have .Net 4.0 which I'm assuming is the problem. Is there a way to get this app to work on .Net 4.0 instead of 4.5?
Running a transacted installation.
Beginning the Install phase of the installation.
See the contents of the log file for the E:\statsd.net-v1.5.0.0\statsdnet.exe assembly's progress.
The file is located at E:\statsd.net-v1.5.0.0\statsdnet.InstallLog.
An exception occurred during the Install phase.
System.InvalidOperationException: Unable to get installer types in the E:\statsd.net-v1.5.0.0\statsdnet.exe assembly.
The inner exception System.Reflection.ReflectionTypeLoadException was thrown with the following error message: Unable to load one or more of the requested types. Retrieve the LoaderExceptions property for more information..
The Rollback phase of the installation is beginning.
See the contents of the log file for the E:\statsd.net-v1.5.0.0\statsdnet.exe assembly's progress.
The file is located at E:\statsd.net-v1.5.0.0\statsdnet.InstallLog.
An exception occurred during the Rollback phase of the System.Configuration.Install.AssemblyInstaller installer.
System.InvalidOperationException: Unable to get installer types in the E:\statsd.net-v1.5.0.0\statsdnet.exe assembly.
The inner exception System.Reflection.ReflectionTypeLoadException was thrown with the following error message: Unable to load one or more of the requested types. Retrieve the LoaderExceptions property for more information..
An exception occurred during the Rollback phase of the installation. This exception will be ignored and the rollback will continue. However, the machine might not fully revert to its initial state after the rollback is complete.
The Rollback phase completed successfully.
The transacted install has completed.
Running a transacted installation.
Beginning the Install phase of the installation.
See the contents of the log file for the E:\statsd.net-v1.5.0.0\statsdnet.exe assembly's progress.
The file is located at E:\statsd.net-v1.5.0.0\statsdnet.InstallLog.
An exception occurred during the Install phase.
System.InvalidOperationException: Unable to get installer types in the E:\statsd.net-v1.5.0.0\statsdnet.exe assembly.
The inner exception System.Reflection.ReflectionTypeLoadException was thrown with the following error message: Unable to load one or more of the requested types. Retrieve the LoaderExceptions property for more information..
The Rollback phase of the installation is beginning.
See the contents of the log file for the E:\statsd.net-v1.5.0.0\statsdnet.exe assembly's progress.
The file is located at E:\statsd.net-v1.5.0.0\statsdnet.InstallLog.
An exception occurred during the Rollback phase of the System.Configuration.Install.AssemblyInstaller installer.
System.InvalidOperationException: Unable to get installer types in the E:\statsd.net-v1.5.0.0\statsdnet.exe assembly.
The inner exception System.Reflection.ReflectionTypeLoadException was thrown with the following error message: Unable to load one or more of the requested types. Retrieve the LoaderExceptions property for more information..
An exception occurred during the Rollback phase of the installation. This exception will be ignored and the rollback will continue. However, the machine might not fully revert to its initial state after the rollback is complete.
The Rollback phase completed successfully.
The transacted install has completed.
—
Reply to this email directly or view it on GitHub.
|
gharchive/issue
| 2015-01-18T16:03:14 |
2025-04-01T06:39:27.900903
|
{
"authors": [
"lukevenediger",
"phillijw"
],
"repo": "lukevenediger/statsd.net",
"url": "https://github.com/lukevenediger/statsd.net/issues/38",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
771405461
|
Support for Fiscal Printer
Does this support fiscal printer ?
I’m not sure what a fiscal printer is, could you provide a reference to this please?
https://docs.microsoft.com/en-us/previous-versions/windows/embedded/ms884287(v=winembedded.4)
Hi @lukevp ,
A thermal receipt you pass a free text to the sdk. A thermal fiscal receipt there are certain commands that should be passed to the SDK.
It's still ESC/POS commands. How can I pass ESC/POS commands using the SDK ?
LIST OF FISCAL COMMANDS - IN ASCENDING ORDER
HEX DEC Function
21h (33) Clear the display
23h (35) Show text on lower line of display
26h (38) Open non-fiscal receipt
27h (39) Close non-fiscal receipt
29h (41) Setting the memory switches
2Ah (42) Printing non-fiscal free text
2Bh (43) Set FOOTER and printing options
2Ch (44) Advance paper
2Dh (45) Paper cut
2Eh (46) Set HEADER (Name and address)
2Fh (47) Showing text on upper line of display
30h (48) Open fiscal receipt
31h (49) Register sale
32h (50) Tax rates set during selected period
33h (51) Subtotal
34h (52) Register sale and show on display
35h (53) Calculate TOTAL
36h (54) Print free fiscal text
38h (56) Close fiscal receipt
3Ch (60) Cancel fiscal receipt
3Dh (61) Set date and hour
3Eh (62) Get current date and hour
3Fh (63) Show date and hour on display
40h (64) Info on last fiscal entry
41h (65) Info on daily totals
43h (67) Info on daily paid sums
44h (68) Number of free fields in fiscal memory
45h (69) Daily financial report with/without closure
46h (70) Internal debiting/crediting
47h (71) Print diagnostic info
48h (72) Fiscalization
49h (73) Detailed report of the fiscal memory selected by number of entry
4Ah (74) Read statuses
4Ch (76) Status of the fiscal transaction
4Fh (79) Short report of the fiscal memory selected by date of entry
50h (80) Sound signal
53h (83) Set multiplier, decimals, currency name and disabled taxes
54h (84) Print a bar code
55h (85) Set additional payment names
56h (86) Get last fiscal memory date
59h (89) Program production test area
5Ah (90) Return diagnostic info
5Bh (91) Program serial number, country number and Fiscal memory number
5Eh (94) Detailed of fiscal memory (selected by date of entry)
5Fh (95) Short report of fiscal memory (selected by entry number)
60h (96) Set tax office text
61h (97) Return tax rates
62h (98) Set tax registration number
63h (99) Return set tax registration number
64h (100) Show free text on display
65h (101) Set operator’s password
66h (102) Enter operator’s name
67h (103) Info on current receipt
69h (105) Operator report
6Ah (106) Drawer kick-out
6Bh (107) Define items and items info
6Ch (108) Detailed daily report
6Dh (109) Print duplicate receipt
6Eh (110) Additional daily info
6Fh (111) Report on groups of items
70h (112) Reading info on operator
71h (113) Read the number of the last fiscal entry or period
72h (114) Read info on fiscal entry or period
73h (115) Program graphic logo
74h (116) Read fiscal memory block
76h (118) Register technical intervention
78h (120) Electronic journal support
79H (121) Read code memory (firmware)
7Eh (126) Erase electronic journal
7Fh (127) RAM reset
Hey @dicksonkimeu ,
this is not currently supported, but would be implemented as a FiscalEmitter. We currently only have the Epson emitter. You would not need to use that Fiscal SDK as that’s a different interface (UPOS) vs directly interacting with the printer like this library.
are you open to implementing the Fiscal support? I will assist you as I can but I do not have any hardware that works with this, so we would have to collaborate on a PR.
I've worked with Fiscal Printers here in Brazil 10 years ago (or so), however, I did not use direct command bytes, but rather each manufacture DLL. You will need the whole command to implement it. @dicksonkimeu, the list that you provided is only a summary. You will probably need the parameters info as well, if you provide a handbook I can also assist to implement the commands in a new emitter.
Send me a test email to dicky.tech@gmail.com I will forward you the documentation from Datecs fp300
|
gharchive/issue
| 2020-12-19T15:57:19 |
2025-04-01T06:39:27.912864
|
{
"authors": [
"dicksonkimeu",
"igorocampos",
"lukevp"
],
"repo": "lukevp/ESC-POS-.NET",
"url": "https://github.com/lukevp/ESC-POS-.NET/issues/95",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1461747381
|
Add 'latest' option for version
Hi,
you provide two ways of setting up version:
# Using 'latest' branch, the most recent CMake and ninja are installed.
uses: lukka/get-cmake@latest # <--= THIS IS THE ONE LINER YOU NEED
and
- name: Get specific version CMake, v3.24.3, and Ninja v1.11.1
uses: lukka/get-cmake@latest
with:
cmakeVersion: 3.24.3 # <--= optional, overrides the _latest_ version of CMake
ninjaVersion: 1.11.1 # <--= optional, overrides the _latest_ version of Ninja
however it makes problem if I would like to use matrix build for different cmake versions, especially if one of them is latest.
I cannot make
jobs:
strategy:
matrix:
cmake: [ 3.9.2, latest ]
steps:
- name: Get specific version CMake version 1
uses: lukka/get-cmake@${{ matrix.cmake }} # cannot use matrix component in the action name
- name: Get specific version CMake version 2
uses: lukka/get-cmake@latest
with:
- cmakeVersion: ${{ matrix.cmake }} # 'latest' will not work here
Could you maybe add accepting latest as the valid version which fallbacks to the default?
@rlalik totally agreed, usage like you described should be supported.
@rlalik you may give a try to the version on PR: `get-cmake@dev/any-version
solved in https://github.com/lukka/get-cmake/releases/tag/v3.25.1
|
gharchive/issue
| 2022-11-23T13:16:26 |
2025-04-01T06:39:27.924246
|
{
"authors": [
"lukka",
"rlalik"
],
"repo": "lukka/get-cmake",
"url": "https://github.com/lukka/get-cmake/issues/60",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1793842591
|
feat: FAQ - Add Network and Nodes Chapter
Removes the old Network and Validator pages and sets up the new Network and Nodes folder, including the following pages:
Blockchain Architecture
Network Configuration
Peer Connections
Node Setup
Validators
Security
Staking
Thanks, Johann, for taking the time 🙏🏻😙 I applied your suggestions and clarified the answers. Also added redirects and modified the headings as in the other sections
sadly we now have a few conflicts bc we merged the other PRs :/
should be easy to fi anyway
|
gharchive/pull-request
| 2023-07-07T16:42:09 |
2025-04-01T06:39:27.928180
|
{
"authors": [
"Hugoo",
"fhildeb"
],
"repo": "lukso-network/docs",
"url": "https://github.com/lukso-network/docs/pull/574",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
468609478
|
Parser special constructs implementation
Summary
Based on the parser framework, it's important to implement special constructs, like type definitions. These constructs could be in the future implemented using the meta-enso layer, so here we need to use the same layer, but internally.
Value
Correct parsing of all Enso constructs.
Acceptance Criteria & Test Cases
All language constructs should be parsed correctly.
Is there a PR @wdanilo ? If so, please connect it to the issue.
@wdanilo how can I test it?
|
gharchive/issue
| 2019-07-16T11:50:52 |
2025-04-01T06:39:27.933800
|
{
"authors": [
"sylwiabr",
"wdanilo"
],
"repo": "luna/enso",
"url": "https://github.com/luna/enso/issues/37",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
143167751
|
trace viewer
almost certainly depends on #173
timeline view like sublime text minimap with memory/register activity highlighted in color
colorized memory blocks based on the basic block that "owns" them, or which blocks touched them
navigate memory based on which basic blocks touched them, kinda like xrefs in ida but for memory access
also select memory/syscalls and visually see the taint flow and highlight the blocks in a graph view
time decayed memory
memory diffing
text tracing should be a driver that parses the binary trace
run actual emulator from state - if syscalls are reduced to operational transforms, we can forward/rewind
|
gharchive/issue
| 2016-03-24T06:43:27 |
2025-04-01T06:39:27.944469
|
{
"authors": [
"lunixbochs"
],
"repo": "lunixbochs/usercorn",
"url": "https://github.com/lunixbochs/usercorn/issues/174",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1872193356
|
On duplication in LEM integration
I think there are two factors of duplication in #629:
Misuse of genericity
This is essentially explained in a PR comment. That is, the types that depend on a generic C: Coprocessor<F> are redefined in #629 because the LEM logic cannot for now accomodate any coprocessor other than DummyCoproc<F>. However, for any GenericType<Foo, Bar>, there is nothing in Rust that prevents you from defining methods for type GenericType<Foo, usize>.
I have a branch that starts working out the deduplication in https://github.com/huitseeker/lurk-rs/tree/lem-integration-experiment
The reason I think it's important to resolve this is that the methods we'd define today on Foo<F, DummyCoproc<F>> will eventually be able to be defined on Foo<F, C: Coproc<F>>, which is going to be much easier to adapt to if they're not distinct types.
Not abstracting the Multiframe, and failing to define provers in terms of that abstraction
The MultiFrame is the fundamental type that's defined starkly differently in LEM and Lurk:
in LEM
in Lurk
This struct appears in the APIs of the prover:
https://github.com/lurk-lab/lurk-rs/blob/9aa8b75247066b70972f2806a1537c359f47c2c6/src/proof/mod.rs#L20
https://github.com/lurk-lab/lurk-rs/blob/9aa8b75247066b70972f2806a1537c359f47c2c6/src/proof/mod.rs#L33-L34
https://github.com/lurk-lab/lurk-rs/blob/9aa8b75247066b70972f2806a1537c359f47c2c6/src/proof/mod.rs#L107-L110
https://github.com/lurk-lab/lurk-rs/blob/9aa8b75247066b70972f2806a1537c359f47c2c6/src/proof/nova.rs#L121
https://github.com/lurk-lab/lurk-rs/blob/9aa8b75247066b70972f2806a1537c359f47c2c6/src/proof/nova.rs#L131-L141
https://github.com/lurk-lab/lurk-rs/blob/9aa8b75247066b70972f2806a1537c359f47c2c6/src/proof/nova.rs#L172-L181
https://github.com/lurk-lab/lurk-rs/blob/9aa8b75247066b70972f2806a1537c359f47c2c6/src/proof/nova.rs#L421-L428
morally speaking, the only thing we need (that is, the only thing Nova requires) in the position of this MultiFrame in the prover is some instance of StepCircuit. I suspect we could implement:
a trait MultiFrame that abstracts over both types of Multiframe (LEM, non-Lurk). The main APIs to provide there are associated types fixing the local notion of Store and Ptr, blank, from_frames and synthesize_frames (the later being expressed in terms of those associated types:
https://github.com/lurk-lab/lurk-rs/blob/9aa8b75247066b70972f2806a1537c359f47c2c6/src/proof/nova.rs#L421-L428
https://github.com/lurk-lab/lurk-rs/blob/9aa8b75247066b70972f2806a1537c359f47c2c6/src/circuit/circuit_frame.rs#L183-L192
Note the apparent difference between the instances of this backend function are not key, since their only usage is in the
impl nova::StepCircuit for MultiFrame and impl bellpepper_core::Circuit for MultiFrame. Whatever abstraction that allows implementing those two callsites on top of what the MultiFrame trait offers should be enough.
an implementation of the Nova prover that works on top of the Multiframe trait, rather than introspecting into the details of the two instances of their implementations above
Summarizing the story:
#629 was a meaty PR,
we worked out how to prove through an interface in #642 and #633
#663 committed isolated changes from #629
the blocker was then a pattern of genericity + mutability of the store which could make interleaving evaluation and proving really hard,
we moved to an interior mutability store in #680,
this allowed for the sought after genericity in #709
this paved the way for #717 and #718 which offer apples-to-apples guarantees of feature parity for evaluation and proving,
The present should close as soon as #717 and #718 are merged, because while we are indeed taking some duplication with those PRs, this is better than the alternative (worked out in #729) due to a lower complexity.
The missing pieces are NIVC support from #677 and #725, which should also resolve soon.
Closed with the merge of #717, #718.
|
gharchive/issue
| 2023-08-29T18:33:40 |
2025-04-01T06:39:27.966867
|
{
"authors": [
"huitseeker"
],
"repo": "lurk-lab/lurk-rs",
"url": "https://github.com/lurk-lab/lurk-rs/issues/639",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
634260321
|
auto update
how to disable automatic update?
This is not supported.
how to disable automatic update?
You can Install netron with pip: pip install netron which won't update automatically
|
gharchive/issue
| 2020-06-08T07:00:37 |
2025-04-01T06:39:27.983009
|
{
"authors": [
"handw-github",
"lutzroeder",
"oxygen-dioxide"
],
"repo": "lutzroeder/netron",
"url": "https://github.com/lutzroeder/netron/issues/513",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2246009931
|
Problema com o acesso aos dados do OMIE spot do dia seguinte
A minha instalação HA recebeu esta mensagem. Há algum problema com o acesso aos dados do OMIE spot do dia seguinte?
This error originated from a custom integration.
Logger: custom_components.omie.coordinator
Source: helpers/update_coordinator.py:328
integration: OMIE - electricity market operator for the Iberian Peninsula (documentation, issues)
First occurred: 12:30:01 PM (1 occurrences)
Last logged: 12:30:01 PM
Error requesting omie.spot_next data: Cannot connect to host www.omie.es:443 ssl:True [SSLCertVerificationError: (1, '[SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed: unable to get local issuer certificate (_ssl.c:1000)')]
No meu caso, com o HA versão 2024.3.2 não há problema nenhum e o omie tem os valores de hoje e amanhã
Noutro HA com a versão 2024.4.3 deixou de apresentar o preço para amanhã, entretanto tentei reiniciar a integração e o HA e agora não apresenta nenhuma informação, nem atual, nem de hoje nem amanhã.
O erro apresentado é o mesmo:
aiohttp.client_exceptions.ClientConnectorCertificateError: Cannot connect to host www.omie.es:443 ssl:True [SSLCertVerificationError: (1, '[SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed: unable to get local issuer certificate (_ssl.c:1000)')]
Desde a semana passada, tenho o HA com a versão 2024.4.3, mas o problema só começou hoje. Infelizmente, pouco depois de publicar este problema, reiniciei a integração e agora também não recebo os preços à vista de hoje.
Desde a semana passada, tenho o HA com a versão 2024.4.3, mas o problema só começou hoje. Infelizmente, pouco depois de publicar este problema, reiniciei a integração e agora também não recebo os preços à vista de hoje.
Estranho, a minha instalação com 2024.3.2, agora mesmo, está a funcionar bem, e mesmo depois de reiniciar o HA não deixa de funcionar. Parece-me de algum modo relacionado com a nova versão do HA.
Tentei remover, reboot, adicionar e permanece sem dados.
erro nos logs
Logger: homeassistant.helpers.event
Source: helpers/template.py:588
First occurred: 16:22:26 (16 occurrences)
Last logged: 16:22:26
Error while processing template: Template<template=({% set CG = 0.20 -%} {% set PM = state_attr('sensor.omie_spot_price_pt', 'tomorrow_average') -%} {% set ns = namespace(hourly_data=[]) %} {% for h, _ in state_attr('sensor.omie_spot_price_pt', 'tomorrow_hours').items() -%} {% if PM == None or MAJ == None %} {% set price=PM %} {% else %} {% set price=(PM*(1-CG)) | round(2) %} {% endif %} {% set ns.hourly_data=ns.hourly_data + [(h.isoformat(), price)] %} {% endfor -%} {{ dict(ns.hourly_data) }}) renders=2>
Error while processing template: Template<template=({% set K1 = 10 -%} {% set POMIEh = states('sensor.omie_spot_price_pt')|float -%} {{ [POMIEh - K1, 0] | max | round(2) }}) renders=2>
Error while processing template: Template<template=({{ state('sensor.omie_spot_price_pt') }}) renders=2>
Error while processing template: Template<template=({% set K1 = 10 -%} {% set ns = namespace(hourly_data=[]) %} {% for h, POMIEh in state_attr('sensor.omie_spot_price_pt', 'today_hours').items() -%} {% if POMIEh == None %} {% set price=POMIEh %} {% else %} {% set price=[(POMIEh - K1), 0] | max | round(2) %} {% endif %} {% set ns.hourly_data=ns.hourly_data + [(h.isoformat(), price)] %} {% endfor -%} {{ dict(ns.hourly_data) }}) renders=2>
Error while processing template: Template<template=({% set K1 = 10 -%} {% set ns = namespace(hourly_data=[]) %} {% for h, POMIEh in state_attr('sensor.omie_spot_price_pt', 'tomorrow_hours').items() -%} {% if POMIEh == None %} {% set price=POMIEh %} {% else %} {% set price=[(POMIEh - K1), 0] | max | round(2) %} {% endif %} {% set ns.hourly_data=ns.hourly_data + [(h.isoformat(), price)] %} {% endfor -%} {{ dict(ns.hourly_data) }}) renders=2>
Traceback (most recent call last):
File "/usr/src/homeassistant/homeassistant/helpers/template.py", line 586, in async_render
render_result = _render_with_context(self.template, compiled, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/src/homeassistant/homeassistant/helpers/template.py", line 2545, in _render_with_context
return template.render(**kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/jinja2/environment.py", line 1301, in render
self.environment.handle_exception()
File "/usr/local/lib/python3.12/site-packages/jinja2/environment.py", line 936, in handle_exception
raise rewrite_traceback_stack(source=source)
File "<template>", line 1, in top-level template code
File "/usr/src/homeassistant/homeassistant/helpers/template.py", line 2237, in forgiving_int_filter
raise_no_default("int", value)
File "/usr/src/homeassistant/homeassistant/helpers/template.py", line 1849, in raise_no_default
raise ValueError(
ValueError: Template error: int got invalid input 'None' when rendering template '{{ (state_attr('sensor.omie_spot_price_pt','today_average') | int /1000) | round (2) }}' but no default was specified
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/usr/src/homeassistant/homeassistant/helpers/template.py", line 712, in async_render_to_info
render_info._result = self.async_render(
^^^^^^^^^^^^^^^^^^
File "/usr/src/homeassistant/homeassistant/helpers/template.py", line 588, in async_render
raise TemplateError(err) from err
homeassistant.exceptions.TemplateError: ValueError: Template error: int got invalid input 'None' when rendering template '{{ (state_attr('sensor.omie_spot_price_pt','today_average') | int /1000) | round (2) }}' but no default was specified
Também já tentei reboot, verificar ligação, etc... e aqui deixou mesmo de funcionar e dá o erro que apresentam nos primeiros posts:
Core 2024.4.3
Supervisor 2024.04.0
Operating System 12.2
Frontend 20240404.2
deve ter havido alguma alteração no que toca a ssl verification.
client_exceptions.ClientConnectorCertificateError: Cannot connect to host www.omie.es:443 ssl:True [SSLCertVerificationError: (1, '[SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed: unable to get local issuer certificate (_ssl.c:1000)')]
Pela minha ultima análise parece-me que estará relacionado com a HA OS, fiz restore do HA para a versão anterior e continua sem funcionar, mas o OS ficou na mesma o 12.2.
Numa outra instalação com o OS 12.1 está a funcionar normalmente.
o Luis acabou de fazer release de uma versão temporária que faz disable ao ssl verify, confirmo que já funciona.
Pela minha ultima análise parece-me que estará relacionado com a HA OS, fiz restore do HA para a versão anterior e continua sem funcionar, mas o OS ficou na mesma o 12.2.
Numa outra instalação com o OS 12.1 está a funcionar normalmente.
Fiz downgrade da 12.2 para 12.1 e continua a não funcionar.
Já agora, a integração scrape para o omie.es também deixou de funcionar, mas se desligar a verificação de ssl já funciona.
Estive a investigar e o problema deriva da forma como o omie.es está a reportar os certificados HTTPS, não tem a ver com versões de HA, HA OS ou desta integração. Pode acontecer mais ou menos esporadicamente dependendo da forma como o omie.es responde a cada pedido (pode haver certificados diferentes em uso, cada pedido é um caso).
Infelizmente tive de desactivar a validação HTTPS para desbloquear a integração até perceber se há alguma solução melhor.
https://github.com/luuuis/hass_omie/releases/tag/v1.0.9
Estive a investigar e o problema deriva da forma como o omie.es está a reportar os certificados HTTPS, não tem a ver com versões de HA, HA OS ou desta integração. Pode acontecer mais ou menos esporadicamente dependendo da forma como o omie.es responde a cada pedido (pode haver certificados diferentes em uso, cada pedido é um caso).
Infelizmente tive de desactivar a validação HTTPS para desbloquear a integração até perceber se há alguma solução melhor.
https://github.com/luuuis/hass_omie/releases/tag/v1.0.9
Coincidência ou não, o addon nginx proxy manager começou a falhar ao renovar os certificados let's encrypt que tinha. o HTTP-01 challenge não funciona, tive de configurar um novo certificado usando o DNS-01 challenge.
O OMIE.es não está a enviar a cadeia completa de certificados necessária para a verificação, daí a falha na verificação.
https://www.ssllabs.com/ssltest/analyze.html?d=www.omie.es
Abri uma ocorrência no suporte técnico do OMIE e eles dizem que vão analisar. Entretanto vou manter este ticket aberto.
Deparei-me com este problema hoje e após algum debug cheguei à conclusão que o pacote ca-certificates não inclui o certificado intermédio GlobalSign GCC R3 DV TLS CA 2020 e que o servidor do OMIE não o está a devolver.
No meu caso, não é só o HA que está a dar erro com pedidos para https://www.omie.es, o curl também se queixa:
$ curl -I https://www.omie.es
curl: (60) SSL certificate problem: unable to get local issuer certificate
More details here: https://curl.se/docs/sslcerts.html
curl failed to verify the legitimacy of the server and therefore could not
establish a secure connection to it. To learn more about this situation and
how to fix it, please visit the web page mentioned above.
A solução que pareceu mais óbvia foi fazer o download do certificado intermédio em falta GlobalSign GCC R3 DV TLS CA 2020 e instalá-lo na "TrustStore" do sistema e do Python.
O certificado intermédio pode ser obtido facilmente, abrindo a página OMIE no browser (Chrome por exemplo), e depois:
clicar no ícone à esquerda do URL
seleccionar "Connection is secure"
seleccionar "Certificate is valid"
seleccionar tab "Details" em cima
escolher o "Global GCC R3 DV TLS CA 2020" na caixa de cima, chamada "Certificate Hierarchy"
clicar no botão "Export" e guardar o ficheiro com extensão .crt
Para quem confiar e não quiser ter trabalho, o certificado é este:
-----BEGIN CERTIFICATE-----
MIIEsDCCA5igAwIBAgIQd70OB0LV2enQSdd00CpvmjANBgkqhkiG9w0BAQsFADBM
MSAwHgYDVQQLExdHbG9iYWxTaWduIFJvb3QgQ0EgLSBSMzETMBEGA1UEChMKR2xv
YmFsU2lnbjETMBEGA1UEAxMKR2xvYmFsU2lnbjAeFw0yMDA3MjgwMDAwMDBaFw0y
OTAzMTgwMDAwMDBaMFMxCzAJBgNVBAYTAkJFMRkwFwYDVQQKExBHbG9iYWxTaWdu
IG52LXNhMSkwJwYDVQQDEyBHbG9iYWxTaWduIEdDQyBSMyBEViBUTFMgQ0EgMjAy
MDCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAKxnlJV/de+OpwyvCXAJ
IcxPCqkFPh1lttW2oljS3oUqPKq8qX6m7K0OVKaKG3GXi4CJ4fHVUgZYE6HRdjqj
hhnuHY6EBCBegcUFgPG0scB12Wi8BHm9zKjWxo3Y2bwhO8Fvr8R42pW0eINc6OTb
QXC0VWFCMVzpcqgz6X49KMZowAMFV6XqtItcG0cMS//9dOJs4oBlpuqX9INxMTGp
6EASAF9cnlAGy/RXkVS9nOLCCa7pCYV+WgDKLTF+OK2Vxw3RUJ/p8009lQeUARv2
UCcNNPCifYX1xIspvarkdjzLwzOdLahDdQbJON58zN4V+lMj0msg+c0KnywPIRp3
BMkCAwEAAaOCAYUwggGBMA4GA1UdDwEB/wQEAwIBhjAdBgNVHSUEFjAUBggrBgEF
BQcDAQYIKwYBBQUHAwIwEgYDVR0TAQH/BAgwBgEB/wIBADAdBgNVHQ4EFgQUDZjA
c3+rvb3ZR0tJrQpKDKw+x3wwHwYDVR0jBBgwFoAUj/BLf6guRSSuTVD6Y5qL3uLd
G7wwewYIKwYBBQUHAQEEbzBtMC4GCCsGAQUFBzABhiJodHRwOi8vb2NzcDIuZ2xv
YmFsc2lnbi5jb20vcm9vdHIzMDsGCCsGAQUFBzAChi9odHRwOi8vc2VjdXJlLmds
b2JhbHNpZ24uY29tL2NhY2VydC9yb290LXIzLmNydDA2BgNVHR8ELzAtMCugKaAn
hiVodHRwOi8vY3JsLmdsb2JhbHNpZ24uY29tL3Jvb3QtcjMuY3JsMEcGA1UdIARA
MD4wPAYEVR0gADA0MDIGCCsGAQUFBwIBFiZodHRwczovL3d3dy5nbG9iYWxzaWdu
LmNvbS9yZXBvc2l0b3J5LzANBgkqhkiG9w0BAQsFAAOCAQEAy8j/c550ea86oCkf
r2W+ptTCYe6iVzvo7H0V1vUEADJOWelTv07Obf+YkEatdN1Jg09ctgSNv2h+LMTk
KRZdAXmsE3N5ve+z1Oa9kuiu7284LjeS09zHJQB4DJJJkvtIbjL/ylMK1fbMHhAW
i0O194TWvH3XWZGXZ6ByxTUIv1+kAIql/Mt29PmKraTT5jrzcVzQ5A9jw16yysuR
XRrLODlkS1hyBjsfyTNZrmL1h117IFgntBA5SQNVl9ckedq5r4RSAU85jV8XK5UL
REjRZt2I6M9Po9QL7guFLu4sPFJpwR1sPJvubS2THeo7SxYoNDtdyBHs7euaGcMa
D/fayQ==
-----END CERTIFICATE-----
Posteriormente é necessário colocar o ficheiro na máquina do HA, em /usr/local/share/ca-certificates/GlobalSign_GCC_R3_DV_TLS_CA_2020.crt e também copiar o certificado para o cacert.pem do Python. A localização deste ficheiro muda dependendo do tipo de instalação mas pode ser obtido perguntando ao Python qual é com python -m certifi. No meu caso o HA corre em venv e o ficheiro está em /home/hassuser/homeassistant-green-3-11-0/lib/python3.11/site-packages/certifi/cacert.pem
É ainda necessário correr o comando sudo update-ca-certificates para o sistema saber que tem um novo certificado disponível e reiniciar o HA.
$ sudo update-ca-certificates
Updating certificates in /etc/ssl/certs...
rehash: warning: skipping ca-certificates.crt,it does not contain exactly one certificate or CRL
1 added, 0 removed; done.
Running hooks in /etc/ca-certificates/update.d...
done.
Posteriomente o curl já não se queixa e os erros de SSL nos logs do HA desapareceram:
$ curl -I https://www.omie.es
HTTP/2 200
date: Wed, 17 Apr 2024 13:08:10 GMT
content-type: text/html; charset=UTF-8
set-cookie: apgwCORS=1fe9e05987a1818347b021e7a21c81fe; Path=/; SameSite=None; Secure
set-cookie: apgw=1fe9e05987a1818347b021e7a21c81fe; Path=/
server: Apache/2.4.41 (Ubuntu)
cache-control: must-revalidate, no-cache, private
x-drupal-dynamic-cache: MISS
link: <https://www.omie.es/es>; rel="canonical", <https://www.omie.es/es>; rel="shortlink"
x-ua-compatible: IE=edge
content-language: es
x-content-type-options: nosniff
x-frame-options: SAMEORIGIN
permissions-policy: interest-cohort=()
expires: Sun, 19 Nov 1978 05:00:00 GMT
x-generator: Drupal 9 (https://www.drupal.org)
balancer_worker_route: 02
strict-transport-security: max-age=31536000; includeSubDomains; preload
x-xss-protection: 1; mode=block
referrer-policy: same-origin
feature-policy: autoplay 'self'; camera 'none'; document-domain 'self'; encrypted-media 'self'; fullscreen 'self'; geolocation 'self'; microphone 'none'; midi 'none'; payment 'none'; xr-spatial-tracking 'none'
content-security-policy: default-src https: data: 'self' 'unsafe-inline' 'unsafe-eval' https://www.google-analytics.com https://www.google.com/recaptcha/ https://www.gstatic.com/recaptcha/
content-security-policy: frame-src https://www.google.com/recaptcha/ https://cdnjs.cloudflare.com/ https://www.youtube.com/
content-security-policy: object-src 'self'
A solução que pareceu mais óbvia foi fazer o download do certificado intermédio em falta GlobalSign GCC R3 DV TLS CA 2020 e instalá-lo na "TrustStore" do sistema e do Python.
Não recomendo, pois toda a "confiança" deve advir dos Root CAs do sistema. Estás de certa forma a adulterar o processo de verificação para o sistema inteiro. O certificado intermédio pode ser descarregado automaticamente ou manualmente para usar na verificação mas não deve ser considerado um Root CA.
Exemplo verificação com openssl
$ openssl x509 -noout -subject -in cert.pem
subject=CN=www.omie.es
$ openssl verify cert.pem
CN=www.omie.es
error 20 at 0 depth lookup: unable to get local issuer certificate
error cert.pem: verification failed
C0FA22F101000000:error:80000002:system library:file_open:No such file or directory:providers/implementations/storemgmt/file_store.c:263:calling stat(/opt/homebrew/etc/openssl@3/certs)
$ openssl x509 -noout -text -in cert.pem | grep -A1 "Authority Information Access"
Authority Information Access:
CA Issuers - URI:http://secure.globalsign.com/cacert/gsgccr3dvtlsca2020.crt
$ wget -q http://secure.globalsign.com/cacert/gsgccr3dvtlsca2020.crt
$ openssl verify -untrusted gsgccr3dvtlsca2020.crt cert.pem
cert.pem: OK
O mais correcto será o OMIE corrigir a configuração do servidor. Nota: quem actualizar para o https://github.com/luuuis/hass_omie/releases/tag/v1.0.9 não precisa de se preocupar com nada disto.
A configuração já foi corrigida no www.omie.es, já não é necessário recorrer à v1.0.9.
https://www.ssllabs.com/ssltest/analyze.html?d=www.omie.es
|
gharchive/issue
| 2024-04-16T13:06:22 |
2025-04-01T06:39:28.009299
|
{
"authors": [
"Gusimelo",
"JeffersM1",
"abacao",
"filipecaldeira",
"frankesousa",
"joaocbatista",
"jpedrofigueiredo",
"luuuis"
],
"repo": "luuuis/hass_omie",
"url": "https://github.com/luuuis/hass_omie/issues/61",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2515145204
|
Add DDRNet model for semantic segmentation
This PR integrates the DDRNet model to enhance semantic segmentation performance.
Comparison with Current Model (MicroNet + Segmentation Head)
Parameters: DDRNet has 5.7M, compared to 2.1M for MicroNet.
Training Time: Both models have similar single-epoch training times.
Metrics After 1 Epoch
MicroNet + Segmentation Head:
Jaccard Index: 0.00873
F1 Score: 0.69811
DDRNet-23-slim:
Jaccard Index: 0.00952
F1 Score: 0.70069
Possible improvements
Compute Jaccard Index every n epochs to increase training speed.
Improve handling of auxiliary heads for export.
Use pretrained weights.
Add Online Hard Example Mining (OHEM) with Cross Entropy Loss.
Add EMA (Exponential Moving Average).
Would it make sense to have auxiliary segmentaion head part of the backbone (if use_aux_head=True) so we could then easily remove it from graph if needed (e.g. durgin export)? Because now user has to change use_aux_head=False in the config if they want to export optimally. CC: @kozlov721 also for thoughts
Yeah I think this would be a better solution. We could later add some more general support for train-only heads, but I think there's not gonna be many use-cases and they will mostly be solvable by making the head part of the backbone anyways.
Also, let's first merge #69, then sync this one and add a test case for the new predefined model to tests.integration.test_simple.test_predefined_models
|
gharchive/pull-request
| 2024-09-10T00:13:41 |
2025-04-01T06:39:28.021978
|
{
"authors": [
"kozlov721",
"sokovninn"
],
"repo": "luxonis/luxonis-train",
"url": "https://github.com/luxonis/luxonis-train/pull/70",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1374976741
|
LVGL v8 status ? Animation with Square Line Studio
We use GitHub issues for development related discussions.
Please use the forum to ask questions.
Describe the issue
Hi All,
I'm starting a nexw project based on Lilygo T display ESP32S3 and LVGL.
I'm able to create a screen with the nice and usefull tool SquareLine Studio. I can display it on the devkit.
I tried to add an animation on my image. I export the ui from sqaure line studio and tried to build on vscode. But call to function "" refer to undifined function.
After comparing v8 to v7 , i saw tat this function were not present in v7. So i need to go on v8 to benefit of it.
I tried to create a blank project, and add lvgl as a submodule to have v8 ready to use. But it seems that KConfig of lv_port_esp32 and last v8 lvgl are different. I don't find all the same options in menuconfig.
So should I only modify the Kconfig and add the missing options present in the lv_port_esp32 version may be? Or is there also other changes to do and it is more complex than that?
I saw the first ticket on this topics that exist for one year now but still open.
Thank you !
Code to reproduce the issue
Expected Results
Actual Results
ESP32 Chip version
ESP32S3
ESP-IDF version
v4.4.2
Development kit used
Lilygo T display ESP32S3
Development machine OS
Visual studio code
Compilation warnings/errors (if available)
implicit declaration of function 'lv_obj_get_x_aligned'; did you mean 'lv_obj_set_align'? [-Werror=implicit-function-declaration]
implicit declaration of function 'lv_obj_get_y_aligned'; did you mean 'lv_obj_set_align'? [-Werror=implicit-function-declaration]
implicit declaration of function 'lv_anim_set_user_data'; did you mean 'lv_obj_set_user_data'? [-Werror=implicit-function-declaration]
If possible, copy the compilation log into a file and attach it here
+1 for official LVGL v8.x support
+1 from me as well.
Still trying to get LVGL running with lvgl and regular drivers library
I also got some customer request for ESP32-LVGLv8 support and I really like to have this and lvgl_esp32_drivers repos updated.
As v7 and v8 differ only in some minor API changes in the drivers I think it's not that difficult to update them.
Unfortunately, I don't have enough hardware for a deep enough testing, however you might have already seen our sponsorship program. From our donations I'd be happy to give 300 USD for updating these repos to v8. Does it sound like a fair offer?
Would you be interested it?
cc @C47D
With some help from @sukesh-ak I have changed to use LGFX Master with LVGL v8. Have a look at https://github.com/sukesh-ak/ESP32-TUX
So far that has worked very well with ESP-IDF v5.x but I have only tried it with ESP-WROVER-Kit v4.1 and the Makerfabs 16bit parallel+touch devkits.
The only changes I have had to make to get a clean compile are:
lv_demo_stress.c: lIne 77 change first %d to %lu
lv_example_table_2.c: line 95 change first %"LV_PRIu32" to %d
So the burning need from our side is gone and the $300 can be saved by using LGFX
Hi @kisvegabor, thanks for CCing me but I'm pretty busy with day job in this days. Happy to see yall got it working!
|
gharchive/issue
| 2022-09-15T19:19:36 |
2025-04-01T06:39:28.040366
|
{
"authors": [
"C47D",
"MrMarteng",
"Olfox59",
"ammaree",
"kisvegabor"
],
"repo": "lvgl/lv_port_esp32",
"url": "https://github.com/lvgl/lv_port_esp32/issues/320",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1183443629
|
lxc remote add cannot take certificate fingerprint as command line argument
Security problem / feature request
Required information
Distribution: Ubuntu
Distribution version: 20.04
The output of "lxc info" or if that fails:
Kernel version:
LXC version:
LXD version: 4.24-c92c0b2
Storage backend in use: zfs
Issue description
It is not possible (or not documented how) to add a remote source in an automated and secure way.
When doing a
lxc remote add someserver somelocation --public
lxc displays the certificate fingerprint and asks for confirmation. This is secure, but requires interaction.
From within a script or automated installation system this is not possible. There you could do a
lxc remote add someserver somelocation --public --accept-certificate
which works automated, but insecure, since it accepts any certificate.
There should be a command line argument for the fingerprint, to add a remote server in an automated, but still secure way.
Steps to reproduce
Invite some evil hacker in your network or the common internet to fake the image server and polute it with dirty images.
lxc remote add someserver somelocation --public --accept-certificate from within a script, e.g. automated installation with cloud-iniit or similar.
Be doomed.
stgraber@dakara:~$ lxc config trust add --name blah
Client blah certificate add token: eyJjbGllbnRfbmFtZSI6ImJsYWgiLCJmaW5nZXJwcmludCI6IjQwMDI1MTc4N2Q2NzA0ZmY4OTdkYmZkOGQ0Mzg3OTcwYTJkOTVkOWRjOTA1MzAzYTI4OTM3MzE0YWE0YjhhODEiLCJhZGRyZXNzZXMiOlsiMTcyLjE3LjAuMjMyOjg0NDMiLCJbMjYwMjpmYzYyOmI6MTAwMDo1NDM2OjViMjU6NjRlNDpkODFhXTo4NDQzIiwiMTcyLjE3LjI1MC4xOjg0NDMiLCJbMjYwMjpmYzYyOmI6MjUwOjoxXTo4NDQzIiwiWzIwMDE6NDcwOmIyYjU6MTAzNzo6MTAwMF06ODQ0MyJdLCJzZWNyZXQiOiIzOTllY2M1MWMzYjdjZTc5Yjg1MTFmNGZiYzAxZTJjMjJjMGQyNzlkODA2NzZjYjUzMDM5M2JjNDMxNWE2MzFlIn0=
stgraber@dakara:~$ lxc remote add foo eyJjbGllbnRfbmFtZSI6ImJsYWgiLCJmaW5nZXJwcmludCI6IjQwMDI1MTc4N2Q2NzA0ZmY4OTdkYmZkOGQ0Mzg3OTcwYTJkOTVkOWRjOTA1MzAzYTI4OTM3MzE0YWE0YjhhODEiLCJhZGRyZXNzZXMiOlsiMTcyLjE3LjAuMjMyOjg0NDMiLCJbMjYwMjpmYzYyOmI6MTAwMDo1NDM2OjViMjU6NjRlNDpkODFhXTo4NDQzIiwiMTcyLjE3LjI1MC4xOjg0NDMiLCJbMjYwMjpmYzYyOmI6MjUwOjoxXTo4NDQzIiwiWzIwMDE6NDcwOmIyYjU6MTAzNzo6MTAwMF06ODQ0MyJdLCJzZWNyZXQiOiIzOTllY2M1MWMzYjdjZTc5Yjg1MTFmNGZiYzAxZTJjMjJjMGQyNzlkODA2NzZjYjUzMDM5M2JjNDMxNWE2MzFlIn0=
|
gharchive/issue
| 2022-03-28T13:27:44 |
2025-04-01T06:39:28.129970
|
{
"authors": [
"hadmut",
"stgraber"
],
"repo": "lxc/lxd",
"url": "https://github.com/lxc/lxd/issues/10137",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
171812798
|
Support operation cancellation in the client (and make more operations support it)
As evidenced by #2293, users rightly expect ctrl-c to cancel things.
This is not the case with LXD, with most operations not being cancel-able over the API anyway.
We should investigate making the following cancel-able and having them canceled on user interrupt:
Image transfer
Container transfer
Container publication
The rest of the time we probably should catch ctrl-c, show a warning message that this will NOT interrupt what's going on in the background, and after repeated ctrl-c (lets say 3 times), actually exit.
Per #3059 when fixing the daemon side of this issue we should add a hook that will automatically cancel all operations during daemon shutdown, ensuring a cleaner daemon shutdown and returning a clear message to clients about the daemon going down.
So I think it's time to look into this one again. I expect this is going to be a bit of ongoing work until we get all the existing operations to support cancellation.
Looking at the current code in operations.go, it looks like all the bits are in place to support cancellation for all operation types. We "just" need to have them define a onCancel function.
My guess is that the onCancel function will typically share a channel with the onRun function and will be limited to writing a value to the channel and then block on it. The onRun function can then select on that channel and if it notices the cancellation, it can do whatever's needed to cancel and then close the channel to have LXD consider the operation cancelled.
Anyway, I'd recommend you pick something easy to cancel and just try to get onCancel to behave, then test this manually through the API. No need to do client side plumbing for this right now. The new client library has proper support for cancellation so we should only do the client side of this once the port to the new library is complete (I'm really really close now :)).
We've got the initial pass of the server side of this implemented. I'll take the issue over to get the initial pass of the client integration done. We should be able to get that done with a helper in lxc/utils.go that wraps around a lxd.Operation struct and effectively calls Wait() + support cancellation through SIGHUP/SIGKILL.
|
gharchive/issue
| 2016-08-18T03:55:41 |
2025-04-01T06:39:28.134380
|
{
"authors": [
"stgraber"
],
"repo": "lxc/lxd",
"url": "https://github.com/lxc/lxd/issues/2294",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
389951946
|
Config for avoiding rootfs shifts/mitigating failures with snapshots
I'd like to add a flag or config to avoid performing operations that would cause an id shift on the container rootfs. In other words, this would allow something like the following flow:
lxc config set mycontainer raw.idmap "both 1000 1000"
lxc start --disallow-id-shift mycontainer # fails
lxc snapshot mycontainer pre-shift-snapshot
lxc start mycontainer # succeeds after remapping
# some time later
lxc delete mycontainer/pre-shift-snapshot
# or, if the remapping was interrupted
lxc restore mycontainer pre-shift-snapshot
AFAICS the same would be needed for lxc publish since that may also perform shifts under the hood.
I'd love to hear any thoughts on whether this seems sane, or if there's a better way to go about this before an in-kernel remapping solution is ready :) I'm happy to post PRs, of course.
Alternatives
One other idea I had was a config to have LXD internally perform a snapshot before shifting ids, and then remove the snapshot afterward. That felt a bit too much like it's moving policy into LXD though, and it seems weird that a start or publish operation would also create a snapshot... On the other hand, I think this would be easier for other LXD users to utilize.
The other idea was for Chrome OS's LXD control daemon tremplin to check the id maps itself via the volatile.* configs to know when to perform a snapshot, but that seems very brittle. IMO the concrete id map at runtime, aside from any manually mapped ids like "both 1000 1000", shouldn't be the business of anything except LXD.
Background
On Chrome OS it's not improbable that our VM will be inadvertently shut down while shifting the container's rootfs ids. This could be due to a power failure, kernel panic in VM guest or host, or a Chrome crash (yes, Chrome crashing will cause our VM/LXD instances to require shutdown :). We saw a few users hitting this in crbug.com/894299.
Until there's a solution to leave these ids unshifted on disk, we would like to ensure that we take a snapshot before performing id shifts. Our downstream tracking bug for this is crbug.com/912360.
I think it should actually be pretty simple for tremplin to determine whether a shift is going to happen by comparing volatile.last_state.idmap to volatile.idmap.next, if they differ, a shift will occur on startup.
In theory we could introduce a security.protection.shift config key which if set to true would prevent any shifting operation, requiring the user to set it to false temporarily as needed, though the check above may be enough until we get to shiftfs.
@sforshee is currently working on porting the shiftfs patches to 4.19 and I'm expecting to have the initial LXD support for it done in January with it shipping to our users as part of Ubuntu 19.04 in April.
That will unfortunately not be mainline at that point and it's still not completely clear what the mainline solution will end up being, whether that's just merging shiftfs or some VFS rework to make it possible to do this without being a filesystem, but shiftfs will be a clean standalone virtual filesystem so backporting it and maintaining a kernel with it shouldn't need much effort.
If you're comfortable with programs other than LXD examining volatile.last_state.idmap and volatile.idmap.next then I think that addresses my concerns with letting tremplin deal with that state.
I could see the security.protection.shift config key being useful just to protect users who want to run lxc publish. But if this doesn't seem like a super useful addition for LXD, I can just update the reddit crostini wiki to encourage snapshotting and we can close this :)
security.protection.shift should be easy enough to implement that if you think it's worth it to prevent potential issues with lxc publish, it's probably fine to do so.
Snapshots when published, at least on btrfs and zfs, get copied read-write, then shifted, exported and discarded, so even if something bad happens mid publish, the snapshot itself should be fine and so the protection doesn't need to extend to them.
So it'd effectively only control startup time shifting of the container itself and be only relevant on systems where shifting is used (so will be ignored when shiftfs is available).
|
gharchive/issue
| 2018-12-11T21:01:35 |
2025-04-01T06:39:28.142648
|
{
"authors": [
"smibarber",
"stgraber"
],
"repo": "lxc/lxd",
"url": "https://github.com/lxc/lxd/issues/5349",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1569441257
|
lxd/backup : update the error msg when "backup/index.yaml" can't be found
Signed-off-by: Gabriel Mougard gabriel.mougard@gmail.com
This pull request didn't trigger Jenkins as its author isn't in the allow list.
An organization member must perform one of the following:
To have this branch tested by Jenkins, use the "ok to test" command.
To have a one time test done, use the "test this please" command.
Those commands are simple Github comments of the format: "jenkins: COMMAND"
jenkins: test this please
|
gharchive/pull-request
| 2023-02-03T08:40:06 |
2025-04-01T06:39:28.145234
|
{
"authors": [
"gabrielmougard",
"lxc-jenkins",
"monstermunchkin"
],
"repo": "lxc/lxd",
"url": "https://github.com/lxc/lxd/pull/11328",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
319831969
|
xattr: Support empty values
Signed-off-by: Stéphane Graber stgraber@ubuntu.com
LGTM! Thanks for fixing it that quickly :-)
@brauner jenkins looks pretty happy with this
|
gharchive/pull-request
| 2018-05-03T08:40:01 |
2025-04-01T06:39:28.146727
|
{
"authors": [
"morphis",
"stgraber"
],
"repo": "lxc/lxd",
"url": "https://github.com/lxc/lxd/pull/4529",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
797755890
|
Update Ubuntu packages
Hi !
The "python3-pylxd" packages for the Focal, Groovy and Hirsute versions of Ubuntu are still pointing to the 2.2.10
Can you push the latest version to these repositories ?
Thanks :)
We don't maintain the packages in Ubuntu/Debian - you're probably best contacting the Debian maintainers - https://tracker.debian.org/pkg/python-pylxd
|
gharchive/issue
| 2021-01-31T16:51:11 |
2025-04-01T06:39:28.148533
|
{
"authors": [
"mazerty",
"sparkiegeek"
],
"repo": "lxc/pylxd",
"url": "https://github.com/lxc/pylxd/issues/462",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
356163789
|
fix activity path of AndroidManifest
com.lyft.domic.samples.redux.rxredux.MainActivity is wrong activity path.
I fix to '(com.lyft.domic.samples.mvvm).MainActivity. it's a mvvm sample.
Ah, I guess refactoring touched that, thanks!
|
gharchive/pull-request
| 2018-09-01T02:25:34 |
2025-04-01T06:39:28.184451
|
{
"authors": [
"artem-zinnatullin",
"satoshun"
],
"repo": "lyft/domic",
"url": "https://github.com/lyft/domic/pull/35",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
787180864
|
fix: failed data loading/refreshing in TaskExecutionDetails
TL;DR
Fixes a failure to load data in the TaskExecutionDetails view when coming from a page where the TaskExecution has already been loaded an is in a final state.
Also fixes a refresh issue with child NodeExecutions on that page.
Type
[x] Bug Fix
[ ] Feature
[ ] Plugin
Complete description
The enabled flag on useQuery is meant to delay running a dependent query until data from the parent is available. We were abusing it a little bit by attempting to use it to return cached data from a query without running the query. A better implementation of this is to conditionally set the staleTime to Infinity and disable refetch in the case where we want to just use whatever cached data is available. This updates useCondtionalQuery to follow that logic.
Also fixed an issue where the NodeExecutions list on the TaskExecutionDetails page was not refreshing when the parent generator task succeeded but the spawned NodeExecutions were still in progress. This required adding a refetchInterval to the query and fixing the logic in the shouldEnableQuery function.
Tracking Issue
https://github.com/lyft/flyte/issues/672
Codecov Report
Merging #142 (fcab137) into master (d8daf6c) will decrease coverage by 0.10%.
The diff coverage is 68.42%.
@@ Coverage Diff @@
## master #142 +/- ##
==========================================
- Coverage 74.43% 74.32% -0.11%
==========================================
Files 415 414 -1
Lines 7310 7310
Branches 1154 1159 +5
==========================================
- Hits 5441 5433 -8
- Misses 1869 1877 +8
Impacted Files
Coverage Δ
src/components/App/App.tsx
85.71% <ø> (-0.50%)
:arrow_down:
src/components/data/QueryAuthorizationObserver.tsx
29.41% <ø> (ø)
src/components/hooks/useFetchableData.ts
91.07% <ø> (-1.79%)
:arrow_down:
src/components/data/apiContext.ts
61.90% <20.00%> (-32.22%)
:arrow_down:
...utions/TaskExecutionDetails/TaskExecutionNodes.tsx
57.57% <33.33%> (ø)
src/components/Navigation/ProjectSelector.tsx
100.00% <100.00%> (ø)
...rc/components/Navigation/SearchableProjectList.tsx
100.00% <100.00%> (ø)
src/components/hooks/useConditionalQuery.ts
100.00% <100.00%> (ø)
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update dd62120...fcab137. Read the comment docs.
Codecov Report
Merging #142 (fcab137) into master (d8daf6c) will decrease coverage by 0.10%.
The diff coverage is 68.42%.
@@ Coverage Diff @@
## master #142 +/- ##
==========================================
- Coverage 74.43% 74.32% -0.11%
==========================================
Files 415 414 -1
Lines 7310 7310
Branches 1154 1159 +5
==========================================
- Hits 5441 5433 -8
- Misses 1869 1877 +8
Impacted Files
Coverage Δ
src/components/App/App.tsx
85.71% <ø> (-0.50%)
:arrow_down:
src/components/data/QueryAuthorizationObserver.tsx
29.41% <ø> (ø)
src/components/hooks/useFetchableData.ts
91.07% <ø> (-1.79%)
:arrow_down:
src/components/data/apiContext.ts
61.90% <20.00%> (-32.22%)
:arrow_down:
...utions/TaskExecutionDetails/TaskExecutionNodes.tsx
57.57% <33.33%> (ø)
src/components/Navigation/ProjectSelector.tsx
100.00% <100.00%> (ø)
...rc/components/Navigation/SearchableProjectList.tsx
100.00% <100.00%> (ø)
src/components/hooks/useConditionalQuery.ts
100.00% <100.00%> (ø)
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update dd62120...fcab137. Read the comment docs.
:tada: This PR is included in version 0.19.3 :tada:
The release is available on GitHub release
Your semantic-release bot :package::rocket:
:tada: This PR is included in version 0.19.3 :tada:
The release is available on GitHub release
Your semantic-release bot :package::rocket:
|
gharchive/pull-request
| 2021-01-15T20:43:13 |
2025-04-01T06:39:28.214202
|
{
"authors": [
"codecov-io",
"schottra",
"service-github-lyft-semantic-release"
],
"repo": "lyft/flyteconsole",
"url": "https://github.com/lyft/flyteconsole/pull/142",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
775736118
|
Switching from using model Metadata -> TaskMetadata
TaskMetadata will be maintained as a shadow and allows decoupling of
protocol buffer types from contributor code and user code. This allows
more flexiblity
another prime motivation behind this change is that - it allows making the interface less verbose. This makes it trivial to support default values for metadata.
I think I'm not seeing the bigger picture. At least I don't understand how this is a decoupling? Ultimately, at serialization time, the Python class will have to be converted into the model class right? I feel like this is more delaying the coupling rather than decoupling.
Which is fine, but I don't understand what this enables. +1 if you want. I think it's okay to use the model sometimes though - like currently we use the dynamic job spec model, literal map and parameter map models, all the literal models, auth, labels, annotations, etc. I don't think we should create parallel classes for all those.
can you take a look at test failures?
|
gharchive/pull-request
| 2020-12-29T06:59:12 |
2025-04-01T06:39:28.217018
|
{
"authors": [
"katrogan",
"kumare3",
"wild-endeavor"
],
"repo": "lyft/flytekit",
"url": "https://github.com/lyft/flytekit/pull/298",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
415163533
|
Update bazel-gazelle version
The current bazel version complains about use of cfg="data" on an attribute in the old version of bazel-gazelle. The current release doesn't have this problem.
@rodaine can you approve this to fix CI?
|
gharchive/pull-request
| 2019-02-27T14:52:02 |
2025-04-01T06:39:28.218068
|
{
"authors": [
"akonradi"
],
"repo": "lyft/protoc-gen-validate",
"url": "https://github.com/lyft/protoc-gen-validate/pull/148",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1859025503
|
Can't truly uninstall
When I disable and uninstall this plugin, and after restarting on all devices, it continues to create the Daily Planner directory and new day planner file. This is driving me crazy; it's bad enough I'm considering nuking my vault and apps entirely to remove it.
Well, I nuked the whole Vault and made a new one. Now, the old vault, with the same name, keeps getting made so that “Daily Planner” can be made with the daily note file underneath. This is crazy. I guess I could setup a chronjob to delete the dir every day, but what do I have to do here? I fully reinstalled Obsidian and it still keeps happening.
Okay, I think I figured it out.
I had moved on from my old vault on only 3 out of 4 devices. So the last one, that I forgot about, was still making that daily note constantly.
Can probably mark as solved.
|
gharchive/issue
| 2023-08-21T09:57:34 |
2025-04-01T06:39:28.220645
|
{
"authors": [
"BNHeadrick"
],
"repo": "lynchjames/obsidian-day-planner",
"url": "https://github.com/lynchjames/obsidian-day-planner/issues/217",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
600095483
|
Adapt to new nICEr API
After https://github.com/lynckia/nrappkit/pull/4 and https://github.com/lynckia/nICEr/pull/2 are merged, need to update licode to work with changed nICEr api.
This PR is showing how I did it. Not sure it is 100% correct, but it works in our environment and now is on testing stage.
Note: need to change GIT_TAG for project_nicer in erizo/src/third_party/nicer.cmake to new value.
This looks promising! thanks a lot for the contribution, why do you want to update nICEr? are there new features?
Mostly because of stability issues: we often see disconnection with "Ica failed" error. For some clients it happens during almost each session.
On our side we still have ability to switch between libnice and nicer, and when we switch to the latest libnice it works more stable. But you decided to go with nICEr, this is why we tried the latest version of it. For now I can say - it works better.
During merging I saw such new features:
changed ice-restart logic
mDNS support
I merged these commits in another PR to fix a couple of cases, thanks for the contribution
|
gharchive/pull-request
| 2020-04-15T07:51:33 |
2025-04-01T06:39:28.224188
|
{
"authors": [
"jcague",
"vpoddubchak"
],
"repo": "lynckia/licode",
"url": "https://github.com/lynckia/licode/pull/1564",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1517196859
|
fix(sbb-tag): fix styles
Preflight Checklist
[x] I have read the Contributing Guidelines for this project.
[x] I agree to follow the Code of Conduct that this project adheres to.
[x] I have searched the pull request tracker for a Pull Request (PR) that matches the one I want to submit, without success.
Issue
This PR fixes an issue reported by Aleksandar from team Lynx, related to both storybook and his aem environment:
Tag elements are not correctly displayed
Pull request checklist
Please check if your PR fulfills the following requirements:
[ ] Tests for the changes have been added (for bug fixes / features)
[ ] Docs have been reviewed and added / updated if needed (for bug fixes / features)
See Review Guidelines for more information what is checked during review process.
Changes
Changes in this pull request:
fixed tag element display (background, border, animations)
Browsers
I tested the build on the following browsers:
[x] Firefox Desktop
[x] Chrome Desktop
[ ] Edge Desktop
[x] Safari Desktop
[x] Chrome Mobile
[x] Safari Mobile
Screen readers
I tested the build on the following browsers:
[ ] JAWS Firefox Desktop
[ ] JAWS Chrome Desktop
[ ] NVDA Firefox Desktop
[ ] NVDA Chrome Desktop
[ ] VoiceOver Safari Desktop
[ ] VoiceOver Chrome Desktop
[ ] VoiceOver Safari Mobile
[ ] Android Accessibility Suite Chrome Mobile
Pull request type
Please check the type of change your PR introduces:
[x] Bugfix
[ ] Feature
[ ] Code style update (formatting, renaming)
[x] Refactoring (no functional changes, no api changes)
[ ] Build related changes
[ ] Documentation content changes
[ ] Other (please describe):
Does this introduce a breaking change?
[ ] Yes
[x] No
Other information
The ellipsis seem to be broken
working on it
Codecov Report
Merging #1516 (47821d4) into master (ab917b0) will decrease coverage by 1.62%.
The diff coverage is 62.05%.
@@ Coverage Diff @@
## master #1516 +/- ##
==========================================
- Coverage 54.85% 53.22% -1.63%
==========================================
Files 49 82 +33
Lines 1659 3438 +1779
Branches 406 958 +552
==========================================
+ Hits 910 1830 +920
- Misses 671 1474 +803
- Partials 78 134 +56
Impacted Files
Coverage Δ
...mponents/sbb-accordion-item/sbb-accordion-item.tsx
0.00% <0.00%> (-37.21%)
:arrow_down:
src/components/sbb-clock/sbb-clock.tsx
0.00% <0.00%> (ø)
src/components/sbb-link/sbb-link.tsx
59.57% <ø> (-31.86%)
:arrow_down:
src/components/sbb-logo/sbb-logo.tsx
0.00% <ø> (ø)
src/components/sbb-menu-action/sbb-menu-action.tsx
61.90% <ø> (ø)
src/components/sbb-menu/sbb-menu.tsx
25.89% <ø> (ø)
...ts/sbb-navigation-action/sbb-navigation-action.tsx
71.42% <ø> (ø)
...onents/sbb-navigation-list/sbb-navigation-list.tsx
84.00% <ø> (ø)
...ts/sbb-navigation-marker/sbb-navigation-marker.tsx
53.19% <ø> (ø)
...mponents/sbb-checkbox-group/sbb-checkbox-group.tsx
30.61% <30.61%> (ø)
... and 65 more
:mega: We’re building smart automated test selection to slash your CI/CD build times. Learn more
pushed fixes for ellipsis + active state + focus outline + minor bug on firefox
|
gharchive/pull-request
| 2023-01-03T10:27:29 |
2025-04-01T06:39:28.248238
|
{
"authors": [
"codecov-commenter",
"federicoisepponfincons"
],
"repo": "lyne-design-system/lyne-components",
"url": "https://github.com/lyne-design-system/lyne-components/pull/1516",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
100395468
|
NameError (uninitialized constant DeviseTokenAuth::Concerns::User::BCrypt)
I get error when doing multiple API requests in my ROR application running this gem which is deployed on heroku using Puma (based on this article https://devcenter.heroku.com/articles/deploying-rails-applications-with-the-puma-web-server). Especially the error occurs when I have set Puma parameter -w (defining number of workers) to amount bigger than 1. Interestingly the error does not occur when I have set parameter -t (number of threads) to amount biger than 1. So it is probably solely connected to web concurrency defined by puma workes. Please see below details of the error:
app[web.1]: NameError (uninitialized constant DeviseTokenAuth::Concerns::User::BCrypt):
app[web.1]: vendor/bundle/ruby/2.0.0/gems/devise_token_auth-0.1.30/app/models/devise_token_auth/concerns/user.rb:111:in token_is_current?' app[web.1]: vendor/bundle/ruby/2.0.0/gems/devise_token_auth-0.1.30/app/models/devise_token_auth/concerns/user.rb:86:in valid_token?'
app[web.1]: vendor/bundle/ruby/2.0.0/gems/devise_token_auth-0.1.30/app/controllers/devise_token_auth/concerns/set_user_by_token.rb:40:in set_user_by_token' app[web.1]: vendor/bundle/ruby/2.0.0/gems/devise_token_auth-0.1.30/lib/devise_token_auth/controllers/helpers.rb:115:in current_user'
app[web.1]: vendor/bundle/ruby/2.0.0/gems/devise_token_auth-0.1.30/lib/devise_token_auth/controllers/helpers.rb:103:in authenticate_user!' app[web.1]: vendor/bundle/ruby/2.0.0/gems/activesupport-4.1.9/lib/active_support/callbacks.rb:424:in block in make_lambda'
app[web.1]: vendor/bundle/ruby/2.0.0/gems/activesupport-4.1.9/lib/active_support/callbacks.rb:143:in call' app[web.1]: vendor/bundle/ruby/2.0.0/gems/activesupport-4.1.9/lib/active_support/callbacks.rb:143:in block in halting_and_conditional'
app[web.1]: vendor/bundle/ruby/2.0.0/gems/activesupport-4.1.9/lib/active_support/callbacks.rb:229:in call' app[web.1]: vendor/bundle/ruby/2.0.0/gems/activesupport-4.1.9/lib/active_support/callbacks.rb:229:in block in halting'
app[web.1]: vendor/bundle/ruby/2.0.0/gems/activesupport-4.1.9/lib/active_support/callbacks.rb:166:in call' app[web.1]: vendor/bundle/ruby/2.0.0/gems/activesupport-4.1.9/lib/active_support/callbacks.rb:166:in block in halting'
app[web.1]: vendor/bundle/ruby/2.0.0/gems/activesupport-4.1.9/lib/active_support/callbacks.rb:229:in call' app[web.1]: vendor/bundle/ruby/2.0.0/gems/activesupport-4.1.9/lib/active_support/callbacks.rb:229:in block in halting'
app[web.1]: vendor/bundle/ruby/2.0.0/gems/activesupport-4.1.9/lib/active_support/callbacks.rb:86:in call' app[web.1]: vendor/bundle/ruby/2.0.0/gems/activesupport-4.1.9/lib/active_support/callbacks.rb:86:in run_callbacks'
I am having this issue aswell
thanks @andersonbrandon
@booleanbetrayal any plan to release new gem version?
Would like to get a new gem version out in the next couple of days, but am waiting for one changeset to make it in. In the meantime, you could always point to master:
gem 'devise_token_auth', :git => 'https://github.com/lynndylanhurley/devise_token_auth.git', :branch => 'master'
ok, thanks!!!
|
gharchive/issue
| 2015-08-11T19:55:58 |
2025-04-01T06:39:28.255751
|
{
"authors": [
"booleanbetrayal",
"danielBlowingNose",
"salmanasiddiqui",
"toshimaru"
],
"repo": "lynndylanhurley/devise_token_auth",
"url": "https://github.com/lynndylanhurley/devise_token_auth/issues/333",
"license": "WTFPL",
"license_type": "permissive",
"license_source": "github-api"
}
|
1291593749
|
⚠️ App has degraded performance
In aa42580, App (https://mesirve.app) experienced degraded performance:
HTTP code: 200
Response time: 395 ms
Resolved: App performance has improved in 45a4b08.
|
gharchive/issue
| 2022-07-01T16:14:07 |
2025-04-01T06:39:28.258367
|
{
"authors": [
"leynier"
],
"repo": "lynot/mesirve-status",
"url": "https://github.com/lynot/mesirve-status/issues/44",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1722803774
|
💡 Implement React Testing Library
Feature Request 👨🏼💻
Implement React Testing Library for project testing 🧪
Use Case ✍🏼
As developers, we want to ensure the stability and reliability of our project by implementing a robust testing framework. By introducing React Testing Library, we aim to facilitate the testing process and improve the overall quality of our codebase.
Possible Solution 💡
Install React Testing Library as a project dependency.
Create separate folders for the tests.
Integrate testing into the project's development workflow.
List all current components that need to be tested (make US 🏷️)
Components To Test :
/App
Header
Footer
NavBar
LayerByMode
CopyUrlButton
Code
ActiveLink
/utils
encodedObjectValues
formatUrlWithQuery
loadFont
/hooks
useInputChange
useInputDateChange
useSelectedFont
/forms
colorInput
FontPicker
input
inputDate
selectInput
/remotion
/molecules
AvatarWithCaption
IconWithCaption
TalkDetails
|
gharchive/issue
| 2023-05-23T21:21:51 |
2025-04-01T06:39:28.264807
|
{
"authors": [
"Boyadjie",
"CruuzAzul"
],
"repo": "lyonjs/shortvid.io",
"url": "https://github.com/lyonjs/shortvid.io/issues/330",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
545650340
|
去除搜索结果中无版权的音乐
发现现在的搜索结果中会出现无版权的音乐,然后播放会提示请求异常之类的提示。能否考虑直接将无版权的音乐在搜索结果中去除或者提示出来呢。
这个是接口问题,我也不知道哪些有版权哪些无版权,不打算处理这个问题
|
gharchive/issue
| 2020-01-06T10:20:33 |
2025-04-01T06:39:28.266674
|
{
"authors": [
"lyswhut",
"maoshibai"
],
"repo": "lyswhut/lx-music-desktop",
"url": "https://github.com/lyswhut/lx-music-desktop/issues/107",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
863711574
|
功能添加请求:1、歌手专辑页面;2、歌手所有单曲页。
1、专辑页可以更好定位歌手某专。比如:今天我想听《八度空间》,明天我想听《七里香》(手动狗头);
2、歌手所有单曲页。是为了更好(懒)的听当前歌手的所有歌曲或下载。比如:今天我想沉浸在某胖伦的音乐海洋,不许其他任何人 打扰。(再次狗头)
太对了,强烈请求添加,呜呜呜
|
gharchive/issue
| 2021-04-21T10:20:38 |
2025-04-01T06:39:28.267643
|
{
"authors": [
"UTAKO-TAKI",
"qaz5391450"
],
"repo": "lyswhut/lx-music-desktop",
"url": "https://github.com/lyswhut/lx-music-desktop/issues/485",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2097799663
|
Ray caster output tensors size
Hi @lzhnb,
Thanks for the library. I am trying to employ it to do some ray casting on meshes. The question I have is what size the depths, normals, and primitive_ids tensors should have? I tried to initiate the tensors with torch zeros of size [n, 1], [n, 3], and [n, 1] accordingly but get wrong outputs consisting of all 10 for depths, -1 for primitive ids, and 0 for normals.
Thanks in advance.
Nevermind, seems like my rays were just not pointing onto the surface. Upon further inspection I found that everything works fine.
|
gharchive/issue
| 2024-01-24T09:17:52 |
2025-04-01T06:39:28.275351
|
{
"authors": [
"ilia-chelak"
],
"repo": "lzhnb/Primitive3D",
"url": "https://github.com/lzhnb/Primitive3D/issues/5",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2204293917
|
Update package.json
Hi @m-avagyan,
There is missing react-i18next package in dependencies,
and @types/react-i18next (This is a stub types definition. react-i18next provides its own type definitions, so you do not need this installed.) here
Thanks
Hi @rkmsnc , thanks for contribution 🙌
|
gharchive/pull-request
| 2024-03-24T10:58:23 |
2025-04-01T06:39:28.309018
|
{
"authors": [
"m-avagyan",
"rkmsnc"
],
"repo": "m-avagyan/webpack-react-typescript-template",
"url": "https://github.com/m-avagyan/webpack-react-typescript-template/pull/3",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
271164565
|
[WIP] Add ICMPv4 sockets
Add support for ICMP sockets
Add tests for ICMP sockets
Rename proto-type features to socket-type
Update documentation
Resolves: #64
See src/socket/icmp.rs:137-174 for the updated API using bind.
@dlrobertson The updated API doesn't look quite right. Shouldn't you be binding to the identifier field in the IP header rather than IP endpoint? Imagine you are implementing a traceroute program. With the current API this is completely impossible.
To send and receive ICMP messages that are not associated with a specific TCP/UDP port number (e.g., Echo, Echo Reply, Timestamp, Timestamp Reply, Information Request, Information Reply), the socket has to be bound to a specific ICMP identifier. The ICMP identifier is a 16-bit field present in bytes 5/6 in the header of these messages. Only messages containing the right identifier can be sent or received through a safe raw ICMP socket of this type.
So it isn't the identifier in the IP header (I was wrong in my previous comment) it is the identifier in the Icmp message. I used the IpEndpoint port value for consistency with other implementations and for the case where you're binding to ICMP error responses to a UDP port. Then the socket creation takes an additional parameter that determines how the port value of the bound IpEndpoint works. As a result, the current API includes the following two cases.
Bind to ICMP error responses for UDP packets sent from port 53.
use smoltcp::socket::{Socket, IcmpSocket, IcmpSocketType};
// Created with type Udp. IpEndpoint::port is truely a port.
let mut icmp_socket = match IcmpSocket::new(rx_buffer, tx_buffer, IcmpSocketType::Udp) {
Socket::Icmp(socket) => socket,
_ => unreachable!()
};
icmp_socket.bind(53).unwrap();
Bind to ICMP messages with the identifier 0x1234
use smoltcp::socket::{Socket, IcmpSocket, IcmpSocketType};
// Created with type Icmp. IpEndpoint::prt is actually the 16 bit identifier
let mut icmp_socket = match IcmpSocket::new(rx_buffer, tx_buffer, IcmpSocketType::Icmp) {
Socket::Icmp(socket) => socket,
_ => unreachable!()
};
icmp_socket.bind(0x1234).unwrap();
I chose to implement this using a parameter added to socket creation, but this could also be accomplished using an enum passed to bind. I'm starting to realize the use IpEndpoint makes sense for C, but I think it would be more readable if I use some sort of enum.
Ah, I see. Yes, I think a dedicated enum IcmpEndpoint would be best.
How important is it to keep the standard bind API? E.g. How bad would it be if we ended up with something like.
let mut socket = ...
socket.bind(IcmpEndpoint::Identifier(0x1234))
When bind is only given a u16 it gets to be a bit ambiguous. In theory bind could keep the same signature. Then we could implement the setting of the sockets endpoint with something like the following.
match self.socket_type {
SocketType::Udp => self.endpoint = IcmpEndpoint::Udp(endpoint.into()),
SocketType::Icmp => {
let tmp_endpoint: IpEndpoint = endpoint.into();
self.endpoint = IcmpEndpoint::Icmp(tmp_endpoint.port),
}
}
There's no "standard bind API", we do not try to implement POSIX. There aren't ICMP sockets in POSIX anyway. So, do the most clear thing.
Rebased on master and updated the ping example. I got it to work on my system, but otherwise I did very little testing of it.
After the comments above are fixed this is ready to be merged.
Thanks for the reviews. The final product was much much better than the first few revisions
|
gharchive/pull-request
| 2017-11-04T04:02:17 |
2025-04-01T06:39:28.351868
|
{
"authors": [
"dlrobertson",
"whitequark"
],
"repo": "m-labs/smoltcp",
"url": "https://github.com/m-labs/smoltcp/pull/69",
"license": "0BSD",
"license_type": "permissive",
"license_source": "github-api"
}
|
577234715
|
ModuleNotFoundError when importing mpu6050
the smbus installation worked, I was able to test the i2c and that works. I also got no errors when installing through pip install. However as soon as I try to run a script in Python3, it says that:
Traceback (most recent call last):
File "/home/pi/Documents/gyrompu.py", line 1, in
import mpu6050 as mpu6050
ModuleNotFoundError: No module named 'mpu6050'
I've read some threads about similar issues and it looks like it's all coming from using Python 3
seems like the issue is within Python because it is working when I'm using a script that's located in the same directory as the mpu6050.py
same issue
@sutanu86 Did you install the package? If so, how?
@chancelier did you use the same python version for installing the package as for running it?
Hi, I just released a version that might help for different platform: Jetson nano, raspberry pi, Beaglebone, pyboard, ODROID and more.
It is based on mr Tijn package.
https://github.com/romybompart/py_imu_mpu6050
https://pypi.org/project/py-imu-mpu6050/
Now that is working in different platform I will add self_calibration functions and
Magnetometer support
git clone https://github.com/Tijndagamer/mpu6050.git
cd mpu6050
sudo python setup.py install
Hello, you can see what is the directory that save the files when you install the mpu6050, for example in my case add in ./.local/lib/python2.7/site-packages... if you run you can try to run with python2 <filename.py>
git clone https://github.com/Tijndagamer/mpu6050.git
cd mpu6050
sudo python setup.py install
This didn't work for me.
"ModuleNotFoundError: No module named 'smbus'"
|
gharchive/issue
| 2020-03-06T23:56:59 |
2025-04-01T06:39:28.361254
|
{
"authors": [
"EbGu3",
"JohnAllen",
"MarCixn",
"chancelier",
"gwonglapierre",
"m-rtijn",
"romybompart",
"sutanu86"
],
"repo": "m-rtijn/mpu6050",
"url": "https://github.com/m-rtijn/mpu6050/issues/29",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1282689345
|
fix: signature for pairwise mapping
Fixed signature for map_pairwise_serial! and map_pairwise_parallel! for CellListPair type.
Currently, the package fails for the following example:
using CellListMap
box = Box([100,100,100],20)
b1 = [[100,100,100] .* rand(3) for _ in 1:10]
b2 = [[100,100,100] .* rand(3) for _ in 1:10]
cl = CellList(b1, b2, box)
CellListMap.map_pairwise_serial!((x,y,i,j,d2,output)->output, nothing, box, cl)
because keyword show_progress is by default show_progress which is not defined elsewhere.
Please let me know if you experience any other issue. Thanks again.
|
gharchive/pull-request
| 2022-06-23T16:53:02 |
2025-04-01T06:39:28.392278
|
{
"authors": [
"drelatgithub",
"lmiq"
],
"repo": "m3g/CellListMap.jl",
"url": "https://github.com/m3g/CellListMap.jl/pull/61",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
341005739
|
awesome font url broken for ipv.save(offline=true)
I get the following error:
HTTPError: 404 Client Error: Not Found for url: http://fontawesome.io/assets/font-awesome-4.7.0.zip
Hi Timo,
that should be fixed in master, by https://github.com/maartenbreddels/ipyvolume/pull/104
cheers,
Maarten
|
gharchive/issue
| 2018-07-13T12:46:11 |
2025-04-01T06:39:28.418122
|
{
"authors": [
"TimoFriedri",
"maartenbreddels"
],
"repo": "maartenbreddels/ipyvolume",
"url": "https://github.com/maartenbreddels/ipyvolume/issues/153",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
812429926
|
Nudge app appears off center when it's opened on an external monitor, then the user disconnects from monitor
When an update is past due, Nudge cannot be dismissed by the "Later" or "I understand" buttons (expected). However, if the user is working off an external monitor, when they unplug and work on the main laptop screen, Nudge does not recenter itself. This prevents the user from fully see and interact with Nudge.
Possible fix: https://developer.apple.com/documentation/appkit/nswindow/1419090-center
I think I have fixed it: https://github.com/macadmins/nudge/commit/ce198c101ff2e2758eed6e1c38be1d6d5a8fe87e
Please try this version: https://github.com/macadmins/nudge/releases/tag/v.1.0.0.02202021001333
It took two commits, but confirmed with @macbm that this fixes it. It will take Nudge 60 seconds to fix this as it's tied to the nudgeRefreshCycle key that defaults to 60 seconds, but that's the best I can do without increasing CPU usage and polling this constantly (which I don't think is worth the effort).
|
gharchive/issue
| 2021-02-19T23:29:15 |
2025-04-01T06:39:28.439926
|
{
"authors": [
"erikng",
"macbm"
],
"repo": "macadmins/nudge",
"url": "https://github.com/macadmins/nudge/issues/121",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2695163525
|
🛑 MacManager (portal.macalegin.com) is down
In 03f3ebd, MacManager (portal.macalegin.com) (https://portal.macalegin.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: MacManager (portal.macalegin.com) is back up in f11c4c2 after 9 minutes.
|
gharchive/issue
| 2024-11-26T15:43:03 |
2025-04-01T06:39:28.443561
|
{
"authors": [
"macalegin-admin"
],
"repo": "macalegin-admin/status",
"url": "https://github.com/macalegin-admin/status/issues/178",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
374478349
|
Searchbox showing incorrect options for some region queries
Type 1-1 in the search box and the first option shown is 1--19-21.
It's suggesting the 20b region around position 1 in chromosome 1, so chromosome 1, start -19, end 21.
|
gharchive/issue
| 2018-10-26T17:31:22 |
2025-04-01T06:39:28.444601
|
{
"authors": [
"nawatts"
],
"repo": "macarthur-lab/gnomadjs",
"url": "https://github.com/macarthur-lab/gnomadjs/issues/320",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
119507203
|
npm release is out of date
https://www.npmjs.com/package/angular-selectize2 reports version 1.2.3.
Please release an updated version for npm.
It looks like you tried to update package.json to include the 3.0.1 version number but the v3.0.1 tag is below that commit. So when npm looks up the v3.0.1 tag, it gets the v3.0.1 code (I think) but the package.json file that gets downloaded still says v.1.2.3, which can lead to issues when using npm or other tools like npm-shrinkwrap cuz they'll get tripped up on the version number in that file. Perhaps just moving the git v3.0.1 tag a few commit forward would help. Or publishing a v3.0.2 with everything in sync?
Agreed, I've created #151 to add a main entry to package.json, but it won't help if npm isn't being updated.
I think you should be able to reference the GitHub repo directly in your package dependencies if npm isn't current.
I wonder if this is the reason that I can't get it to load within my app.js file while using webpack...
|
gharchive/issue
| 2015-11-30T14:49:08 |
2025-04-01T06:39:28.463211
|
{
"authors": [
"SteveShaffer",
"adiakritos",
"avaragado",
"chrisparton1991",
"cudasteve"
],
"repo": "machineboy2045/angular-selectize",
"url": "https://github.com/machineboy2045/angular-selectize/issues/117",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2547242741
|
Fix Hungarian default date format and add Hungarian translations
Hey,
I fixed the Hungarian default date format as it was incorrect, you can find more information here:
https://en.wikipedia.org/wiki/Date_and_time_notation_in_Hungary#Date
Also I added translations for the Hungarian language!
Thank you @tamasori, since I am going to release a more stable version soon, I will keep this PR on hold and apply your changes directly. I will add you as contributor for reference
Ok, thank you @macmotp
All your changes have been included into the new release, thanks!
|
gharchive/pull-request
| 2024-09-25T07:53:00 |
2025-04-01T06:39:28.483569
|
{
"authors": [
"macmotp",
"tamasori"
],
"repo": "macmotp/locale",
"url": "https://github.com/macmotp/locale/pull/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
73086124
|
Bug displaying unicode characters above BMP with "syntax on"
From @GoogleCodeExporter on March 16, 2015 9:26
What steps will reproduce the problem?
Set syntax on (latex for instance) and add a calligraphic A in math mode
(U+1D49C) between parenthesis, itself and between dollars $(A)$. I use
EversonMono font.
What is the expected output? What do you see instead?
I expect it to be displayed properly but sometimes, a space is displayed after
the A, shift the rest of the line up to the $
What version of MacVim and OS X are you using (see "MacVim->About MacVim"
and "Apple Menu->About This Mac" menu items, e.g. "Snapshot 40, 10.5.6
Intel")?
Snapshot 73
Please provide any additional information below.
Attached the screenshots of normal buggy display, normal display and syntax off
display.
Original issue reported on code.google.com by harold.c...@gmail.com on 9 Jan 2015 at 2:09
Attachments:
bug1.tiff
bug2.tiff
bug3.tiff
Copied from original issue: douglasdrumond/macvim#524
For what it's worth, I can still reproduce this as well as of snapshot 76:
Copy this character: 𝒜 (the Unicode character referenced in the original post; hopefully Github preserves it).
Launch MacVim, paste the character inside a set of parentheses and type some text after (so you have a line like "(𝒜) hello." Make sure there's a line below this text.
Exit insert mode, move the cursor to the start of the line. Move left and right between the first and second lines and note how the display of the line shifts, sometimes visually truncating the last character in the line.
I can't reproduce it.
@jpetrie could you write more specific way to reproduce the issue? for example
Open an empty MacVim window
Enter insert mode
Paste "(𝒜) hello." and enter return.
Input "aaaa"
Exit insert mode
...
I can still see it by doing the following:
Make sure the Core Text renderer is enabled.
Launch MacVim or otherwise open a new, empty buffer.
Enter insert mode.
Paste the text inside quotes: " (𝒜) test" (that's space, open-paren, 𝒜, close paren, space, 'test')
Exit insert mode
Use h and l to scrub the cursor along the line, particular to the start of the line, observe the rendering artifacts.
what is 'the rendering artifacts'?? snapshot-81, 10.11, I can't see any issue. could you record screencast?
@jpetrie are you using snapshot-80 or earlier version?
snapshot-81 has no problem.
Yeah, it's fixed in 81+.
|
gharchive/issue
| 2015-05-04T18:05:29 |
2025-04-01T06:39:28.575503
|
{
"authors": [
"douglasdrumond",
"jpetrie",
"splhack"
],
"repo": "macvim-dev/macvim",
"url": "https://github.com/macvim-dev/macvim/issues/30",
"license": "Vim",
"license_type": "permissive",
"license_source": "github-api"
}
|
779464551
|
Refresh causes unknown endpoint
Current use of React BrowserRouter can not reload correctly after refresh
Could be fixed with React HashRouter
Maybe redirect to "/" url.
Fixed with "/* endpoint that returns index.html from build.
Fixed with "/* endpoint that returns index.html from build.
|
gharchive/issue
| 2021-01-05T19:28:27 |
2025-04-01T06:39:28.577480
|
{
"authors": [
"macwille"
],
"repo": "macwille/club-webstore",
"url": "https://github.com/macwille/club-webstore/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
96767878
|
Handle failed image upload in image block
If a the image upload fails, Sir Trevor will still create an empty image block, which can cause rendering problems.
That's an interesting one, if I have a failed image upload I still end up seeing the image (now as a base64 file) in an image block but adds a warning to inform that the image failed.
Although thinking about it, I don't use the default image block I've been using Image-extended so that might have something to do with it.
Admittedly, there could do with a way of replacing the uploaded image block without first removing it but that's more of a usability issue than anything else and isn't much of an issue at that.
Going through old issues as part of a block creation / config idea i'm working on.
It'll address bad uploads and allow re-uploading without deleting the block. Also you'll be able to say whether the image element is required, which will allow us to force re-upload or the user will need to delete the block before being able to save and submit the data.
As a side note. This would make image blocks easier to build rather than replicating and editing each time someone wants to customise a field.
Coolbeans, it's not a critical end of the world as we know it kinda issue but It's nice to know it's being looked at.
@raffij Is this feature still in progress?
This can also happen with a video block.
I thought the validation regex tweaks prevented that from happening on the vids? I'll see if i've got time this weekend and see if I can see how to validate the video block
@SgtOddball it's something i need to fix anyway for one of our projects.
The decision is whether an invalid video block should be ignored, or fail validation before a url is posted.
Ahh ok, I'd think that invalid blocks should be ignored otherwise a video link that becomes invalid later on could cause issues as it's already been accepted previously as valid (unless it gets re-evaluated each time the block is rendered).
That's my thinking too.
Just stumbled into this too. It seems that the current image block can no longer detect if an image upload failed in any way. The Promise will always trigger the success method, no matter what the server responded.
So, it is completely possible that i read the code wrong, but the image block (and maybe other blocks using the uploader and/or cancellable-promise) is no longer able to react to any problems during upload.
|
gharchive/issue
| 2015-07-23T09:11:41 |
2025-04-01T06:39:28.590274
|
{
"authors": [
"SgtOddball",
"ToBe998",
"ilyaporopudas",
"raffij",
"rnjailamba"
],
"repo": "madebymany/sir-trevor-js",
"url": "https://github.com/madebymany/sir-trevor-js/issues/370",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
128953333
|
Wrong RFC state
http://why-cant-we-have-nice-things.mwl.be/events shows https://wiki.php.net/rfc/invalid_strings_in_arithmetic as implemented, but the vote was canceled and it was put under discussion again.
I'm guessing the code is giving the votes priority over the status at some point, cause from a votes standpoint it would have been implemented. Will look into it.
|
gharchive/issue
| 2016-01-26T21:48:05 |
2025-04-01T06:39:28.600398
|
{
"authors": [
"Anahkiasen",
"kelunik"
],
"repo": "madewithlove/why-cant-we-have-nice-things",
"url": "https://github.com/madewithlove/why-cant-we-have-nice-things/issues/79",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
119316454
|
Fixes render="publish" does not work when used with an expression
based url #354. This is handy for rendering the main content of a page to get better SEO love.
Perfect. Thanks!
|
gharchive/pull-request
| 2015-11-28T20:26:55 |
2025-04-01T06:39:28.617464
|
{
"authors": [
"chirino",
"madoublet"
],
"repo": "madoublet/respond",
"url": "https://github.com/madoublet/respond/pull/360",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
246977271
|
[IMPAC-603] Create dashboard from template + designer mode
"Create from template" feature
Fix currencies drop-down when dashboard changed
Fix alerts settings button
@xaun I think this is ready - can you please review?
@xaun : up please, the release is tomorrow... thanks.
|
gharchive/pull-request
| 2017-08-01T07:21:14 |
2025-04-01T06:39:28.638550
|
{
"authors": [
"cesar-tonnoir"
],
"repo": "maestrano/impac-angular",
"url": "https://github.com/maestrano/impac-angular/pull/374",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
850201821
|
Plague Inc.: The Cure
Game name: Plague Inc.: The Cure
Engine version: i dunno
Platform: PC
Error message: Элемент с тем же ключом уже был добавлен.
Stack trace: в System.ThrowHelper.ThrowArgumentException(ExceptionResource resource)
в System.Collections.Generic.Dictionary2.Insert(TKey key, TValue value, Boolean add) в uTinyRipper.PlatformGameStructure.AddFile(IDictionary2 files, String name, String path)
в uTinyRipper.PlatformGameStructure.CollectSerializedGameFiles(DirectoryInfo root, IDictionary2 files) в uTinyRipper.MixedGameStructure.CollectFromDirectory(DirectoryInfo root, IDictionary2 files, IDictionary2 assemblies, ISet1 dataPathes)
в uTinyRipper.MixedGameStructure..ctor(IEnumerable1 pathes) в uTinyRipper.GameStructure.CheckMixed(List1 pathes)
в uTinyRipper.GameStructure.Load(List1 pathes, LayoutInfo layinfo) в uTinyRipper.GameStructure.Load(IEnumerable1 pathes, LayoutInfo layinfo)
в uTinyRipperGUI.MainWindow.LoadFiles(Object data)
No promises, but you can try my AssetRipper fork.
No promises, but you can try my AssetRipper fork.
Yooooo, an absolute legend, it works
|
gharchive/issue
| 2021-04-05T08:40:54 |
2025-04-01T06:39:28.642244
|
{
"authors": [
"BurAndBY",
"ds5678"
],
"repo": "mafaca/UtinyRipper",
"url": "https://github.com/mafaca/UtinyRipper/issues/877",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
224316558
|
Chromium link in regions.html is broken
The link in "Some applications, like Chromium, actually still do this in their own window decorations." doesn't seem to work. The code is still there, but the code search interface doesn't seem to like the "rcl" parameter.
Maybe a link to the git repo is more stable: https://chromium.googlesource.com/chromium/src/+/d7d447a09a95eb7ad399eb8512f677a66e127d69/ui/views/window/window_shape.cc#12
the wonder of the modern web
Replaced with the URI you suggested.
|
gharchive/issue
| 2017-04-26T01:22:28 |
2025-04-01T06:39:28.647760
|
{
"authors": [
"magcius",
"zmodem"
],
"repo": "magcius/xplain",
"url": "https://github.com/magcius/xplain/issues/27",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1527176192
|
Clarification request about the algorithm used here vs the Unicode text segmentation guidelines
Hi!
First: Thanks for this handy crate!
I compared the results of this crate against what https://unicode.org/reports/tr29/#Word_Boundaries mandates and was surprised to find some differences. I then checked the algorithm used by your crate and realized that it is a more pragmatic approach.
I wonder whether this is done on purpose (in which case it might make sense to explain it in the README) or whether this is an oversight?
The https://crates.io/crates/unicode-segmentation crate seems to claim to do it the "official" way (Disclaimer: I do not claim that I fully checked every rule and confirmed that it is correct) and comes to the conclusion that the example from the documentation at https://crates.io/crates/words-count should be 18 words long.
What is your thought on this?
This crate doesn't follow the rules in Word Boundaries.
In Chinese composition, I believe Chinese punctuation marks need to be counted as words. That's why Rust是由 Mozilla 主導開發的通用、編譯型程式語言。 counts 20 words (including 、 and 。). You can get the same result in LibreOffice Writer.
Also, you can try "a-good-word".
use unicode_segmentation::UnicodeSegmentation;
fn main() {
let s = "a-good-word";
println!("{:?}", words_count::count(s));
println!("{:?}", s.unicode_words().collect::<Vec<&str>>());
println!("{:?}", s.split_word_bounds().collect::<Vec<&str>>());
}
This crate tells you a-good-word is a word, while the unicode_words function in unicode-segmentation returns ["a", "good", "word"].
In LibreOffice Writer, a-good-word is counted as one word.
Thanks for the clarification! I can totally understand that sometimes a unicode standard may not exactly be what one would want. I have sent a suggestion how to potentially clarify this a bit more in the README :). It might make sense to reference https://crates.io/crates/unicode-segmentation, but I left it out for now and will leave that up to you!
|
gharchive/issue
| 2023-01-10T10:56:17 |
2025-04-01T06:39:28.848134
|
{
"authors": [
"Ablu",
"magiclen"
],
"repo": "magiclen/words-count",
"url": "https://github.com/magiclen/words-count/issues/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
638290480
|
Is the only way to access metrics via serde?
Hi,
Maybe I'm not reading the documentation correctly, but I'm trying to access the metrics generatd by the library (to format and print to the console) and it seems the only way to do so is to roundtrip via some serde serializer? I feel like I'm missing something.
Hi,
Generated metric registries are regular Rust structs, so you can access the different metrics directly, and do whatever reading or transforms you need. Metrics implement Serialize to hook various back-ends that require different formats, notably Prometheus which is supported since 0.4.
Ok, it sounds like I'm using the library incorrectly? For example, using the library manually I have
#[derive(Default, Debug)]
pub struct Perf {
fps: Throughput,
}
fn main() {
let metrics = Perf::default();
let fps = &metrics.fps;
loop {
measure!(fps, expensive()))
// println!("{}", fps.mean())
}
}
However, the internal Histogram isn't visible, so I can't call fps.mean()?. Am I supposed to implement my own Throughput type?
Hmm, that's an oversight, I feel like we should expose the inner histogram. Would you like to open that PR?
|
gharchive/issue
| 2020-06-14T05:56:08 |
2025-04-01T06:39:28.877875
|
{
"authors": [
"magnet",
"nemosupremo"
],
"repo": "magnet/metered-rs",
"url": "https://github.com/magnet/metered-rs/issues/25",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
99669386
|
Vamp Router does not start
I try out vamp 0.7.9 and installed vamp like you have defined it in the guide (http://vamp.io/installation/ubuntu/). But the Vamp Router does not start, i got an endless loop in the log file (/var/log/vamp-router/vamp-router.log ) of the following message.
I could not find a log file of the haproxy to see what is wrong.
██╗ ██╗ █████╗ ███╗ ███╗██████╗
██║ ██║██╔══██╗████╗ ████║██╔══██╗
██║ ██║███████║██╔████╔██║██████╔╝
╚██╗ ██╔╝██╔══██║██║╚██╔╝██║██╔═══╝
╚████╔╝ ██║ ██║██║ ╚═╝ ██║██║
╚═══╝ ╚═╝ ╚═╝╚═╝ ╚═╝╚═╝
router
version 0.7.9
by magnetic.io
08:13:53.615 main NOTI ==> Attempting to load config at /usr/share/vamp-router/configuration
08:13:53.615 main NOTI ==> Did not find a config...initializing empty config
08:13:53.616 main CRIT ==> Could not render initial config, exiting...
08:13:53.649 main INFO ==>
██╗ ██╗ █████╗ ███╗ ███╗██████╗
██║ ██║██╔══██╗████╗ ████║██╔══██╗
██║ ██║███████║██╔████╔██║██████╔╝
╚██╗ ██╔╝██╔══██║██║╚██╔╝██║██╔═══╝
╚████╔╝ ██║ ██║██║ ╚═╝ ██║██║
╚═══╝ ╚═╝ ╚═╝╚═╝ ╚═╝╚═╝
router
version 0.7.9
by magnetic.io
08:13:53.650 main NOTI ==> Attempting to load config at /usr/share/vamp-router/configuration
08:13:53.650 main NOTI ==> Did not find a config...initializing empty config
08:13:53.651 main CRIT ==> Could not render initial config, exiting...
08:13:53.670 main INFO ==>
@zutherb Which version of Ubuntu are you using?
Did you notice the issue reported in the documentation regarding vamp-router?
http://vamp.io/installation/ubuntu/#install-router
@plamola
I'm using:
Linux master1 3.16.0-30-generic #40~14.04.1-Ubuntu SMP Thu Jan 15 17:43:14 UTC 2015 x86_64 x86_64 x86_64 GNU/Linux
Yes notice the issue.
I'm able to reproduce this issue, when I install an old version of haproxy.
The haproxy version in the default Ubuntu repo is 1.4.x; vamp-router required haproxy 1.5+
Could you check which version of haproxy you got installed?
haproxy -v
See the instructions for adding the haproxy repo in our docs:
http://vamp.io/installation/ubuntu/#haproxy-1-5-x
Hi @plamola,
i'm using HA-Proxy version 1.5.14
HA-Proxy version 1.5.14 2015/07/02
Copyright 2000-2015 Willy Tarreau <willy@haproxy.org>```
Can you check the privileges on the /usr/share/vamp-router/data & /usr/share/vamp-router/configuration directories?
The vamp-router user needs read/write access on both directories.
I've done a clean install on an Ubuntu 14.04 machine and this is what is been setup:
$ ls -al /usr/share/vamp-router
total 10168
drwxr-xr-x 6 root root 4096 Aug 10 09:54 .
drwxr-xr-x 105 root root 4096 Aug 10 09:54 ..
drwxrwxr-x 2 root root 4096 Aug 10 09:54 bin
drwxr-xr-x 4 vamp-router vamp-router 4096 Aug 10 09:54 configuration
drwxr-xr-x 2 vamp-router vamp-router 4096 Aug 10 11:02 data
drwxrwxr-x 2 root root 4096 Aug 10 09:54 examples
lrwxrwxrwx 1 root root 20 Aug 3 09:12 logs -> /var/log/vamp-router
-rwxr-xr-x 1 root root 10383392 Jul 31 14:14 vamp-router
Looks good
vagrant@master1:~$ ls -al /usr/share/vamp-router
total 10168
drwxr-xr-x 6 root root 4096 Aug 7 06:04 .
drwxr-xr-x 118 root root 4096 Aug 7 06:06 ..
drwxrwxr-x 2 root root 4096 Aug 7 06:04 bin
drwxr-xr-x 2 vamp-router vamp-router 4096 Aug 7 07:27 configuration
drwxr-xr-x 2 vamp-router vamp-router 4096 Aug 7 07:27 data
drwxrwxr-x 2 root root 4096 Aug 7 06:04 examples
lrwxrwxrwx 1 root root 20 Aug 3 02:12 logs -> /var/log/vamp-router
-rwxr-xr-x 1 root root 10383392 Jul 31 07:14 vamp-router
I will try a clean installation of ubuntu as well.
Hi @plamola,
the fresh installation doesn't work. Here is what i did:
#!/usr/bin/env bash
sudo apt-get install -y openjdk-8
sudo add-apt-repository -y ppa:vbernat/haproxy-1.5
sudo apt-get update
sudo apt-get install -y haproxy
echo "deb https://dl.bintray.com/magnetic-io/upstart trusty main" | sudo tee -a /etc/apt/sources.list
curl https://bintray.com/user/downloadSubjectPublicKey?username=bintray | sudo apt-key add -
sudo apt-get update
sudo apt-get install -y vamp-core
cp -vf /vagrant/scripts/vamp/core-application.conf /usr/share/vamp-core/conf/application.conf
sudo service vamp-core restart
sudo apt-get install -y vamp-pulse
cp -vf /vagrant/scripts/vamp/pulse-application.conf /usr/share/vamp-pulse/conf/application.conf
sudo service vamp-pulse restart
sudo apt-get install -y vamp-router
/usr/share/vamp-router/vamp-router --binary=/usr/sbin/haproxy
sudo apt-get install -y vamp-cli
@zutherb Hi Bernd, we're checking this right now in a fresh Ubuntu 14.04 instance. We noticed a couple of things:
Openjdk 1.8 is not available on this version. We used the Oracle JDK. This is beside this issue of course.
When installing vamp-router, it should automatically start correctly. A ps tells me the following:
root 32259 1 0 09:27 ? 00:00:00 sudo -u vamp-router bin/vamp-router
vamp-ro+ 32261 32259 0 09:27 ? 00:00:00 bash bin/vamp-router
vamp-ro+ 32263 32261 0 09:27 ? 00:00:00 ./vamp-router
vamp-ro+ 32268 1 0 09:27 ? 00:00:00 /usr/sbin/haproxy -f /usr/share/vamp-router/configuration/haproxy_new.cfg -p /usr/share/vamp-router/data/haproxy-private.pid -D
a => root does a sudo -u preserving the user and starting our little bash script to kick off router.
b => vamp-router user runs bin/vamp-router bash script.
c => vamp-router user runs the actual vamp-router binary.
d => vamp-router user runs haproxy with the correct initialisation parameters.
For now, all is fine...a simple curl confirms its running:
tim@ubuntu-14-4-real:~$ curl localhost:10001/v1/info
{"Message":"Hi, I'm Vamp Router! How are you?","Version":"0.7.9","Status":{"Name":"HAProxy","Version":"1.5.14","Release_date":"2015/07/02","Nbproc":"1","Process_num":"1","Pid":"32268","Uptime":"0d 0h04m29s","Uptime_sec":"269","Memmax_MB":"0","Ulimit-n":"8223","Maxsock":"8223","Maxconn":"4096","Hard_maxconn":"4096","CurrConns":"0","CumConns":"92","CumReq":"92","MaxSslConns":"0","CurrSslConns":"0","CumSslConns":"0","Maxpipes":"0","PipesUsed":"0","PipesFree":"0","ConnRate":"0","ConnRateLimit":"0","MaxConnRate":"0","SessRate":"0","SessRateLimit":"0","MaxSessRate":"0","SslRate":"0","SslRateLimit":"0","MaxSslRate":"0","SslFrontendKeyRate":"0","SslFrontendMaxKeyRate":"0","SslFrontendSessionReuse_pct":"0","SslBackendKeyRate":"0","SslBackendMaxKeyRate":"0","SslCacheLookups":"0","SslCacheMisses":"0","CompressBpsIn":"0","CompressBpsOut":"0","CompressBpsRateLim":"0","ZlibMemUsage":"0","MaxZlibMemUsage":"0","Tasks":"5","Run_queue":"1","Idle_pct":"100"}}
Things start going wrong once we start/stop the processes directly using root or sudo. This will work once...but then all pid files, logs, haproxy socket files etc. are written under root. From that moment on, everything should be started under root, otherwise vamp-router cannot communicate with haproxy.
Hi Tim,
thx for your quick help.
i am using the Oracle JDK.
vagrant@master1:~$ java -version
java version "1.8.0_51"
Java(TM) SE Runtime Environment (build 1.8.0_51-b16)
Java HotSpot(TM) 64-Bit Server VM (build 25.51-b03, mixed mode)
The ps tells me. And i can't see a running haproxy-
root 1035 0.0 0.4 62108 3172 ? Ss 09:51 0:00 sudo -u vamp-core bin/vamp-core
vamp-co+ 1040 20.0 22.7 2150740 172460 ? Sl 09:51 1:24 java -Dlogback.configurationFile=/usr/share/vamp-core/bin/../conf/logback.xml -Dconfig.file=/usr/share/vamp-core/bin/../conf/application.conf -cp /usr/share/vamp-core/lib/vamp-core.jar Main
root 1045 0.0 0.4 62108 3116 ? Ss 09:51 0:00 sudo -u vamp-pulse bin/vamp-pulse
vamp-pu+ 1052 5.8 23.7 2189576 180060 ? Sl 09:51 0:24 java -Dlogback.configurationFile=/usr/share/vamp-pulse/bin/../conf/logback.xml -Dconfig.file=/usr/share/vamp-pulse/bin/../conf/application.conf -cp /usr/share/vamp-pulse/lib/vamp-pulse.jar Boot
root 12647 0.0 0.4 62108 3512 ? Ss 09:58 0:00 sudo -u vamp-router bin/vamp-router
vamp-ro+ 12649 0.0 0.2 9520 2204 ? S 09:58 0:00 bash bin/vamp-router
vamp-ro+ 12650 0.0 0.8 11300 6668 ? Sl 09:58 0:00 ./vamp-router
vagrant 12653 0.0 0.2 11728 1944 pts/3 S+ 09:58 0:00 grep --color=auto vamp
The curl doesn't work as well.
vagrant@master1:~$ curl localhost:10001/v1/info
curl: (7) Failed to connect to localhost port 10001: Connection refused
The vamp-core logfile (/var/log/vamp-core/application.log) show that it can connect to the proxy.
09:55:30.467 | ERROR | io.vamp.common.notification.Notification | | Router response error.
09:55:30.467 | ERROR | io.vamp.common.notification.Notification | | java.net.ConnectException: Connection refused: localhost/127.0.0.1:10001 to http://localhost:10001/v1/routes
java.util.concurrent.ExecutionException: java.net.ConnectException: Connection refused: localhost/127.0.0.1:10001 to http://localhost:10001/v1/routes
at com.ning.http.client.providers.netty.NettyResponseFuture.abort(NettyResponseFuture.java:342) ~[vamp-core.jar:0.7.9.1]
at com.ning.http.client.providers.netty.NettyConnectListener.operationComplete(NettyConnectListener.java:107) ~[vamp-core.jar:0.7.9.1]
at org.jboss.netty.channel.DefaultChannelFuture.notifyListener(DefaultChannelFuture.java:431) ~[vamp-core.jar:0.7.9.1]
at org.jboss.netty.channel.DefaultChannelFuture.notifyListeners(DefaultChannelFuture.java:422) ~[vamp-core.jar:0.7.9.1]
at org.jboss.netty.channel.DefaultChannelFuture.setFailure(DefaultChannelFuture.java:384) ~[vamp-core.jar:0.7.9.1]
at org.jboss.netty.channel.socket.nio.NioClientBoss.processSelectedKeys(NioClientBoss.java:109) ~[vamp-core.jar:0.7.9.1]
at org.jboss.netty.channel.socket.nio.NioClientBoss.process(NioClientBoss.java:79) ~[vamp-core.jar:0.7.9.1]
at org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:337) ~[vamp-core.jar:0.7.9.1]
at org.jboss.netty.channel.socket.nio.NioClientBoss.run(NioClientBoss.java:42) ~[vamp-core.jar:0.7.9.1]
at org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) ~[vamp-core.jar:0.7.9.1]
at org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42) ~[vamp-core.jar:0.7.9.1]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) ~[na:1.8.0_51]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) ~[na:1.8.0_51]
at java.lang.Thread.run(Thread.java:745) ~[na:1.8.0_51]
Caused by: java.net.ConnectException: Connection refused: localhost/127.0.0.1:10001 to http://localhost:10001/v1/routes
at com.ning.http.client.providers.netty.NettyConnectListener.operationComplete(NettyConnectListener.java:103) ~[vamp-core.jar:0.7.9.1]
... 12 common frames omitted
Caused by: java.net.ConnectException: Connection refused: localhost/127.0.0.1:10001
at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) ~[na:1.8.0_51]
at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717) ~[na:1.8.0_51]
at org.jboss.netty.channel.socket.nio.NioClientBoss.connect(NioClientBoss.java:152) ~[vamp-core.jar:0.7.9.1]
at org.jboss.netty.channel.socket.nio.NioClientBoss.processSelectedKeys(NioClientBoss.java:105) ~[vamp-core.jar:0.7.9.1]
... 8 common frames omitted
09:55:30.470 | ERROR | akka.actor.OneForOneStrategy | e/user/deployment-synchronization-actor$ | io.vamp.common.notification.NotificationErrorException cannot be cast to io.vamp.core.router_driver.DeploymentRoutes
java.lang.ClassCastException: io.vamp.common.notification.NotificationErrorException cannot be cast to io.vamp.core.router_driver.DeploymentRoutes
at io.vamp.core.operation.deployment.DeploymentSynchronizationActor.io$vamp$core$operation$deployment$DeploymentSynchronizationActor$$synchronize(DeploymentSynchronizationActor.scala:84) ~[vamp-core.jar:0.7.9.1]
at io.vamp.core.operation.deployment.DeploymentSynchronizationActor$$anonfun$receive$1.applyOrElse(DeploymentSynchronizationActor.scala:72) ~[vamp-core.jar:0.7.9.1]
at akka.actor.Actor$class.aroundReceive(Actor.scala:467) ~[vamp-core.jar:0.7.9.1]
at io.vamp.core.operation.deployment.DeploymentSynchronizationActor.aroundReceive(DeploymentSynchronizationActor.scala:43) ~[vamp-core.jar:0.7.9.1]
at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516) [vamp-core.jar:0.7.9.1]
at akka.actor.ActorCell.invoke(ActorCell.scala:487) [vamp-core.jar:0.7.9.1]
at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:238) [vamp-core.jar:0.7.9.1]
at akka.dispatch.Mailbox.run(Mailbox.scala:220) [vamp-core.jar:0.7.9.1]
at akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:397) [vamp-core.jar:0.7.9.1]
at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) [vamp-core.jar:0.7.9.1]
at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) [vamp-core.jar:0.7.9.1]
at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) [vamp-core.jar:0.7.9.1]
at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) [vamp-core.jar:0.7.9.1]
09:55:30.562 | ERROR | io.vamp.common.notification.Notification | | Router response error.
09:55:30.563 | ERROR | io.vamp.common.notification.Notification | | java.net.ConnectException: Connection refused: localhost/127.0.0.1:10001 to http://localhost:10001/v1/info
java.util.concurrent.ExecutionException: java.net.ConnectException: Connection refused: localhost/127.0.0.1:10001 to http://localhost:10001/v1/info
at com.ning.http.client.providers.netty.NettyResponseFuture.abort(NettyResponseFuture.java:342) ~[vamp-core.jar:0.7.9.1]
at com.ning.http.client.providers.netty.NettyConnectListener.operationComplete(NettyConnectListener.java:107) ~[vamp-core.jar:0.7.9.1]
at org.jboss.netty.channel.DefaultChannelFuture.notifyListener(DefaultChannelFuture.java:431) ~[vamp-core.jar:0.7.9.1]
at org.jboss.netty.channel.DefaultChannelFuture.addListener(DefaultChannelFuture.java:145) ~[vamp-core.jar:0.7.9.1]
at com.ning.http.client.providers.netty.NettyAsyncHttpProvider.doConnect(NettyAsyncHttpProvider.java:1138) ~[vamp-core.jar:0.7.9.1]
at com.ning.http.client.providers.netty.NettyAsyncHttpProvider.execute(NettyAsyncHttpProvider.java:935) ~[vamp-core.jar:0.7.9.1]
at com.ning.http.client.AsyncHttpClient.executeRequest(AsyncHttpClient.java:499) ~[vamp-core.jar:0.7.9.1]
at dispatch.HttpExecutor$class.apply(execution.scala:47) ~[vamp-core.jar:0.7.9.1]
at dispatch.Http.apply(execution.scala:12) ~[vamp-core.jar:0.7.9.1]
at dispatch.HttpExecutor$class.apply(execution.scala:42) ~[vamp-core.jar:0.7.9.1]
at dispatch.Http.apply(execution.scala:12) ~[vamp-core.jar:0.7.9.1]
at io.vamp.common.http.RestClient$.http(RestClient.scala:69) ~[vamp-core.jar:0.7.9.1]
at io.vamp.common.http.RestClient$.get(RestClient.scala:27) ~[vamp-core.jar:0.7.9.1]
at io.vamp.core.router_driver.DefaultRouterDriver.info(DefaultRouterDriver.scala:20) ~[vamp-core.jar:0.7.9.1]
at io.vamp.core.router_driver.RouterDriverActor.reply(RouterDriverActor.scala:45) [vamp-core.jar:0.7.9.1]
at io.vamp.common.akka.ReplyActor$$anonfun$receive$1.applyOrElse(ReplyActor.scala:19) [vamp-core.jar:0.7.9.1]
at akka.actor.Actor$class.aroundReceive(Actor.scala:467) [vamp-core.jar:0.7.9.1]
at io.vamp.core.router_driver.RouterDriverActor.aroundReceive(RouterDriverActor.scala:33) [vamp-core.jar:0.7.9.1]
at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516) [vamp-core.jar:0.7.9.1]
at akka.actor.ActorCell.invoke(ActorCell.scala:487) [vamp-core.jar:0.7.9.1]
at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:238) [vamp-core.jar:0.7.9.1]
at akka.dispatch.Mailbox.run(Mailbox.scala:220) [vamp-core.jar:0.7.9.1]
at akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:397) [vamp-core.jar:0.7.9.1]
at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) [vamp-core.jar:0.7.9.1]
at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) [vamp-core.jar:0.7.9.1]
at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) [vamp-core.jar:0.7.9.1]
at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) [vamp-core.jar:0.7.9.1]
Caused by: java.net.ConnectException: Connection refused: localhost/127.0.0.1:10001 to http://localhost:10001/v1/info
at com.ning.http.client.providers.netty.NettyConnectListener.operationComplete(NettyConnectListener.java:103) ~[vamp-core.jar:0.7.9.1]
... 25 common frames omitted
Caused by: java.net.ConnectException: Connection refused: localhost/127.0.0.1:10001
at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) ~[na:1.8.0_51]
at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717) ~[na:1.8.0_51]
at org.jboss.netty.channel.socket.nio.NioClientBoss.connect(NioClientBoss.java:152) ~[vamp-core.jar:0.7.9.1]
at org.jboss.netty.channel.socket.nio.NioClientBoss.processSelectedKeys(NioClientBoss.java:105) ~[vamp-core.jar:0.7.9.1]
at org.jboss.netty.channel.socket.nio.NioClientBoss.process(NioClientBoss.java:79) ~[vamp-core.jar:0.7.9.1]
at org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:337) ~[vamp-core.jar:0.7.9.1]
at org.jboss.netty.channel.socket.nio.NioClientBoss.run(NioClientBoss.java:42) ~[vamp-core.jar:0.7.9.1]
at org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) ~[vamp-core.jar:0.7.9.1]
at org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42) ~[vamp-core.jar:0.7.9.1]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) ~[na:1.8.0_51]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) ~[na:1.8.0_51]
at java.lang.Thread.run(Thread.java:745) ~[na:1.8.0_51]
The vamp-router logfile (/var/log/vamp-router/vamp-router.log) show that vamp-router starts again and again.
10:05:41.922 main INFO ==>
██╗ ██╗ █████╗ ███╗ ███╗██████╗
██║ ██║██╔══██╗████╗ ████║██╔══██╗
██║ ██║███████║██╔████╔██║██████╔╝
╚██╗ ██╔╝██╔══██║██║╚██╔╝██║██╔═══╝
╚████╔╝ ██║ ██║██║ ╚═╝ ██║██║
╚═══╝ ╚═╝ ╚═╝╚═╝ ╚═╝╚═╝
router
version 0.7.9
by magnetic.io
10:05:41.923 main NOTI ==> Attempting to load config at /usr/share/vamp-router/configuration
10:05:41.925 main NOTI ==> Pidfile exists at /usr/share/vamp-router/data/haproxy-private.pid, proceeding...
10:05:41.928 main CRIT ==> Error while reloading haproxy: exit status 1
10:05:41.948 main INFO ==>
██╗ ██╗ █████╗ ███╗ ███╗██████╗
██║ ██║██╔══██╗████╗ ████║██╔══██╗
██║ ██║███████║██╔████╔██║██████╔╝
╚██╗ ██╔╝██╔══██║██║╚██╔╝██║██╔═══╝
╚████╔╝ ██║ ██║██║ ╚═╝ ██║██║
╚═══╝ ╚═╝ ╚═╝╚═╝ ╚═╝╚═╝
router
version 0.7.9
by magnetic.io
10:05:41.949 main NOTI ==> Attempting to load config at /usr/share/vamp-router/configuration
10:05:41.953 main NOTI ==> Pidfile exists at /usr/share/vamp-router/data/h
@zutherb
The problem of the constant restarting is due to the service scripts that continuously tries to start Vamp Router after it fails. I think Vamp Router fails because either Router or Haproxy cannot write the necessary files to disk.
What we can do to pinpoint the problem is the following:
Stop the vamp router service script: sudo service vamp-router stop
Now start it by hand and see what happens
sudo /usr/share/vamp-router/vamp-router
If that doesn't help, remove these files:
rm /usr/share/vamp-router/configuration/haproxy_new.cfg
rm /usr/share/vamp-router/data/haproxy.stats.sock
rm /usr/share/vamp-router/data/haproxy-private.pid
``
thx @tnolet that works, but when i'm using sudo service vamp-router stop to start the vamp-router again i have the same behaviour again.
@zutherb this is due to the fact that now some files are owned by root and the 'service' command tries to use the vamp-router user. This user cannot read/write the files owned by root so it fails. I'm not sure how we can dodge this in an elegant manner. Once files are owned by root it is hard to switch back.
Hi @zutherb I heard this was still an issue.
1 .You can easily configure logging for Haproxy by just putting an extra line in the Haproxy template Vamp Router uses. Haproxy uses syslog or rsyslog for logging. Just put this in the global section of the haproxy_config.template file that ships with Vamp Router. The file should be at /usr/share/vamp-router/configuration/templates/. You should have a syslog or rsyslog daemon on port 514 on 127.0.0.1
log 127.0.0.1 local0
I have a hard time debugging it from here. Is there a way you can snaphot your VM and upload somewhere so I can see what's happening?
Hi Tim,
I try to upload the vm to google drive. I will send you an invention when the upload is done.
I did the following steps to active the logging:
https://www.percona.com/blog/2014/10/03/haproxy-give-me-some-logs-on-centos-6-5/
The Log is then avaible under /var/log/haproxy.log
I got the following output:
Sep 8 09:52:13 localhost haproxy[31106]: { "timestamp" : 08/Sep/2015:09:52:13.033, "frontend" : "b017c339-9406-48aa-ac90-4307e7f6343a_mongodb_27017", "method" : "", "captured_request_headers" : "", "captures_response_headers" : "" }
Sep 8 09:52:13 localhost haproxy[31106]: { "timestamp" : 08/Sep/2015:09:52:13.034, "frontend" : "cebe1679-e15d-41d7-a29f-c03bbcfd33ee_mongodb_27017", "method" : "", "captured_request_headers" : "", "captures_response_headers" : "" }
Sep 8 09:52:13 localhost haproxy[31106]: { "timestamp" : 08/Sep/2015:09:52:13.034, "frontend" : "b017c339-9406-48aa-ac90-4307e7f6343a_mongodb_27017::mongodb", "method" : "", "captured_request_headers" : "", "captures_response_headers" : "" }
Sep 8 09:52:14 localhost haproxy[31106]: { "timestamp" : 08/Sep/2015:09:52:14.540, "frontend" : "b4d4a437-2a0f-4489-be9e-7118ec3237c1_mongodb_27017", "method" : "", "captured_request_headers" : "", "captures_response_headers" : "" }
Sep 8 09:52:16 localhost haproxy[31106]: { "timestamp" : 08/Sep/2015:09:52:16.425, "frontend" : "b4d4a437-2a0f-4489-be9e-7118ec3237c1_mongodb_27017", "method" : "", "captured_request_headers" : "", "captures_response_headers" : "" }
While trying to connect the mongodb:
berndzuther@cc ~/development/appstash/vamp/vagrant (master●)$ mongo 172.31.1.11:33006
MongoDB shell version: 3.0.6
connecting to: 172.31.1.11:33006/test
Server has startup warnings:
2015-09-08T16:18:36.719+0000 I CONTROL [initandlisten] ** WARNING: You are running this process as the root user, which is not recommended.
2015-09-08T16:18:36.719+0000 I CONTROL [initandlisten]
2015-09-08T16:18:36.719+0000 I CONTROL [initandlisten]
2015-09-08T16:18:36.719+0000 I CONTROL [initandlisten] ** WARNING: /sys/kernel/mm/transparent_hugepage/enabled is 'always'.
2015-09-08T16:18:36.719+0000 I CONTROL [initandlisten] ** We suggest setting it to 'never'
2015-09-08T16:18:36.719+0000 I CONTROL [initandlisten]
2015-09-08T16:18:36.720+0000 I CONTROL [initandlisten] ** WARNING: /sys/kernel/mm/transparent_hugepage/defrag is 'always'.
2015-09-08T16:18:36.720+0000 I CONTROL [initandlisten] ** We suggest setting it to 'never'
2015-09-08T16:18:36.720+0000 I CONTROL [initandlisten]
> ^C
bye
berndzuther@cc ~/development/appstash/vamp/vagrant (master●)$ mongo 172.31.1.11:33006
MongoDB shell version: 3.0.6
connecting to: 172.31.1.11:33006/test
2015-09-08T18:49:24.715+0200 I NETWORK DBClientCursor::init call() failed
2015-09-08T18:49:24.720+0200 E QUERY Error: DBClientBase::findN: transport error: 172.31.1.11:33006 ns: admin.$cmd query: { whatsmyuri: 1 }
at connect (src/mongo/shell/mongo.js:181:14)
at (connect):1:6 at src/mongo/shell/mongo.js:181
exception: connect failed
berndzuther@cc ~/development/appstash/vamp/vagrant (master●)$ mongo 172.31.1.11:33006
MongoDB shell version: 3.0.6
connecting to: 172.31.1.11:33006/test
2015-09-08T18:50:00.926+0200 I NETWORK DBClientCursor::init call() failed
2015-09-08T18:50:00.928+0200 E QUERY Error: DBClientBase::findN: transport error: 172.31.1.11:33006 ns: admin.$cmd query: { whatsmyuri: 1 }
at connect (src/mongo/shell/mongo.js:181:14)
at (connect):1:6 at src/mongo/shell/mongo.js:181
exception: connect failed
When i looked to the haproxy log i notice that haproxy try to connected multiple mongodb instances. So i checked the routes with the rest api http://172.31.1.11:10001/v1/routes
[
{
"name": "cebe1679-e15d-41d7-a29f-c03bbcfd33ee_9050",
"port": 9050,
"protocol": "tcp",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "catalog.ports.port",
"weight": 100,
"servers":
[
{
"name": "cebe1679-e15d-41d7-a29f-c03bbcfd33ee_9050",
"host": "localhost",
"port": 33000
}
]
}
]
},
{
"name": "cebe1679-e15d-41d7-a29f-c03bbcfd33ee_catalog_80",
"port": 33000,
"protocol": "http",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "catalog",
"weight": 100,
"servers":
[
{
"name": "a7ca3d5b6ae0f571519c46e90ff04e728ebe4788",
"host": "172.31.2.12",
"port": 31199
}
]
}
]
},
{
"name": "cebe1679-e15d-41d7-a29f-c03bbcfd33ee_checkout_8080",
"port": 33001,
"protocol": "http",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "checkout",
"weight": 100,
"servers":
[
{
"name": "3a7a866906fcc8d57e570529a8322b3a310daf3c",
"host": "172.31.2.13",
"port": 31009
}
]
}
]
},
{
"name": "cebe1679-e15d-41d7-a29f-c03bbcfd33ee_navigation_18090",
"port": 33003,
"protocol": "http",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "navigation",
"weight": 100,
"servers":
[
{
"name": "9d383f9c62101eaad169a606f6b7df6f633689ef",
"host": "172.31.2.13",
"port": 31053
}
]
}
]
},
{
"name": "cebe1679-e15d-41d7-a29f-c03bbcfd33ee_redis_6379",
"port": 33005,
"protocol": "tcp",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "redis",
"weight": 100,
"servers":
[
{
"name": "ae455150a13722cfe13a6c7a937f1335ae83190d",
"host": "172.31.2.12",
"port": 31435
}
]
}
]
},
{
"name": "cebe1679-e15d-41d7-a29f-c03bbcfd33ee_product_18080",
"port": 33002,
"protocol": "http",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "product",
"weight": 100,
"servers":
[
{
"name": "d7101f706052c48fd5ea07711437c501c7b48a3e",
"host": "172.31.2.13",
"port": 31656
}
]
}
]
},
{
"name": "cebe1679-e15d-41d7-a29f-c03bbcfd33ee_cart_18100",
"port": 33004,
"protocol": "http",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "cart",
"weight": 100,
"servers":
[
{
"name": "8ce645fcd48685809cb5f71b96f275a12f42b005",
"host": "172.31.2.11",
"port": 31491
}
]
}
]
},
{
"name": "cebe1679-e15d-41d7-a29f-c03bbcfd33ee_mongodb_27017",
"port": 33006,
"protocol": "tcp",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "mongodb",
"weight": 100,
"servers":
[
{
"name": "54406b9ac88840ea9ff9988604f9ffa7792bd34f",
"host": "172.31.2.12",
"port": 31375
}
]
}
]
},
{
"name": "9b766960-3680-49c5-8395-5c05e4ac5e6d_9050",
"port": 9050,
"protocol": "tcp",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "catalog.ports.port",
"weight": 100,
"servers":
[
{
"name": "9b766960-3680-49c5-8395-5c05e4ac5e6d_9050",
"host": "172.31.1.11",
"port": 33000
}
]
}
]
},
{
"name": "9b766960-3680-49c5-8395-5c05e4ac5e6d_navigation_18090",
"port": 33003,
"protocol": "http",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "navigation",
"weight": 100,
"servers":
[
{
"name": "2c3468acb8f4c1d90d6b0df20e077815537a9a09",
"host": "172.31.2.11",
"port": 31650
}
]
}
]
},
{
"name": "9b766960-3680-49c5-8395-5c05e4ac5e6d_checkout_8080",
"port": 33001,
"protocol": "http",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "checkout",
"weight": 100,
"servers":
[
{
"name": "a2d6cef22c51c85d9d78cf6da04d019623dccff9",
"host": "172.31.2.11",
"port": 31338
}
]
}
]
},
{
"name": "9b766960-3680-49c5-8395-5c05e4ac5e6d_catalog_80",
"port": 33000,
"protocol": "http",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "catalog",
"weight": 100,
"servers":
[
{
"name": "433750cd0a85c8ea01fab15292259b046265f931",
"host": "172.31.2.11",
"port": 31652
}
]
}
]
},
{
"name": "9b766960-3680-49c5-8395-5c05e4ac5e6d_cart_18100",
"port": 33004,
"protocol": "http",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "cart",
"weight": 100,
"servers":
[
{
"name": "932ba659b6df8522e92656874b8c3177e428fed8",
"host": "172.31.2.11",
"port": 31598
}
]
}
]
},
{
"name": "9b766960-3680-49c5-8395-5c05e4ac5e6d_mongodb_27017",
"port": 33006,
"protocol": "tcp",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "mongodb",
"weight": 100,
"servers":
[
{
"name": "30ffce15d779dc85b16e7e16cae6637b5b6dfe83",
"host": "172.31.2.11",
"port": 31031
}
]
}
]
},
{
"name": "9b766960-3680-49c5-8395-5c05e4ac5e6d_product_18080",
"port": 33002,
"protocol": "http",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "product",
"weight": 100,
"servers":
[
{
"name": "7e326e5166c9bc6f81ab893f8d57ebc77189e219",
"host": "172.31.2.11",
"port": 31622
}
]
}
]
},
{
"name": "9b766960-3680-49c5-8395-5c05e4ac5e6d_redis_6379",
"port": 33005,
"protocol": "tcp",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "redis",
"weight": 100,
"servers":
[
{
"name": "a48675044bfd9eccb572633bc7db0d9dc19e4bc8",
"host": "172.31.2.12",
"port": 31305
}
]
}
]
},
{
"name": "cdee9d2d-8856-4144-811a-e0731056255b_mongodb_27017",
"port": 33006,
"protocol": "tcp",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "mongodb",
"weight": 100,
"servers":
[
{
"name": "a50c9b0aec6eb14969127a89fe65d4ff7c65841b",
"host": "172.31.2.11",
"port": 31208
}
]
}
]
},
{
"name": "cdee9d2d-8856-4144-811a-e0731056255b_cart_18100",
"port": 33004,
"protocol": "http",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "cart",
"weight": 100,
"servers":
[
{
"name": "cafff5b83ba6efe05a6c27ea0f0b42746068384e",
"host": "172.31.2.13",
"port": 31304
}
]
}
]
},
{
"name": "cdee9d2d-8856-4144-811a-e0731056255b_product_18080",
"port": 33002,
"protocol": "http",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "product",
"weight": 100,
"servers":
[
{
"name": "60a47e9baa75c60376613a69864a103553d383f1",
"host": "172.31.2.11",
"port": 31466
}
]
}
]
},
{
"name": "cdee9d2d-8856-4144-811a-e0731056255b_checkout_8080",
"port": 33001,
"protocol": "http",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "checkout",
"weight": 100,
"servers":
[
{
"name": "db1763d79983efadefc4cbc82ed97b35c5dfe69",
"host": "172.31.2.11",
"port": 31852
}
]
}
]
},
{
"name": "cdee9d2d-8856-4144-811a-e0731056255b_9050",
"port": 9050,
"protocol": "tcp",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "catalog.ports.port",
"weight": 100,
"servers":
[
{
"name": "cdee9d2d-8856-4144-811a-e0731056255b_9050",
"host": "172.31.1.11",
"port": 33000
}
]
}
]
},
{
"name": "cdee9d2d-8856-4144-811a-e0731056255b_redis_6379",
"port": 33005,
"protocol": "tcp",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "redis",
"weight": 100,
"servers":
[
{
"name": "b47078fc6733e11c600eaa563692f7094fee1d16",
"host": "172.31.2.12",
"port": 31765
}
]
}
]
},
{
"name": "cdee9d2d-8856-4144-811a-e0731056255b_navigation_18090",
"port": 33003,
"protocol": "http",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "navigation",
"weight": 100,
"servers":
[
{
"name": "b90c7e73e6d9537d984e43af8212c26cab01d62c",
"host": "172.31.2.11",
"port": 31077
}
]
}
]
},
{
"name": "cdee9d2d-8856-4144-811a-e0731056255b_catalog_80",
"port": 33000,
"protocol": "http",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "catalog",
"weight": 100,
"servers":
[
{
"name": "a9dd24652f4a25ed1c980ee8ffbe5b7a172d397",
"host": "172.31.2.12",
"port": 31186
}
]
}
]
},
{
"name": "b4d4a437-2a0f-4489-be9e-7118ec3237c1_redis_6379",
"port": 33005,
"protocol": "tcp",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "redis",
"weight": 100,
"servers":
[
{
"name": "c33e8a1cc12f259e26f8839b7771d9caa3c49661",
"host": "172.31.2.11",
"port": 31609
}
]
}
]
},
{
"name": "b4d4a437-2a0f-4489-be9e-7118ec3237c1_mongodb_27017",
"port": 33006,
"protocol": "tcp",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "mongodb",
"weight": 100,
"servers":
[
{
"name": "90dcd245565752c85712c84a2bdcaa5c5f4f5813",
"host": "172.31.2.12",
"port": 31910
}
]
}
]
},
{
"name": "b4d4a437-2a0f-4489-be9e-7118ec3237c1_cart_18100",
"port": 33004,
"protocol": "http",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "cart",
"weight": 100,
"servers":
[
{
"name": "d96571279bfd13d7aae6b82d91812a0bb1bfda64",
"host": "172.31.2.11",
"port": 31527
}
]
}
]
},
{
"name": "b4d4a437-2a0f-4489-be9e-7118ec3237c1_checkout_8080",
"port": 33001,
"protocol": "http",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "checkout",
"weight": 100,
"servers":
[
{
"name": "5940db53b7e336cedcf30b36971771bff0500bc2",
"host": "172.31.2.13",
"port": 31448
}
]
}
]
},
{
"name": "b4d4a437-2a0f-4489-be9e-7118ec3237c1_product_18080",
"port": 33002,
"protocol": "http",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "product",
"weight": 100,
"servers":
[
{
"name": "fcccda9d9a78f8fcd8e1af994526002c57ecd9f1",
"host": "172.31.2.12",
"port": 31102
}
]
}
]
},
{
"name": "b4d4a437-2a0f-4489-be9e-7118ec3237c1_navigation_18090",
"port": 33003,
"protocol": "http",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "navigation",
"weight": 100,
"servers":
[
{
"name": "178df2e513f49e942a3e1a29180a8f457f8e8aac",
"host": "172.31.2.11",
"port": 31676
}
]
}
]
},
{
"name": "b4d4a437-2a0f-4489-be9e-7118ec3237c1_9050",
"port": 9050,
"protocol": "tcp",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "catalog.ports.port",
"weight": 100,
"servers":
[
{
"name": "b4d4a437-2a0f-4489-be9e-7118ec3237c1_9050",
"host": "172.31.1.11",
"port": 33000
}
]
}
]
},
{
"name": "b4d4a437-2a0f-4489-be9e-7118ec3237c1_catalog_80",
"port": 33000,
"protocol": "http",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "catalog",
"weight": 100,
"servers":
[
{
"name": "389f8ed829e54a6ca41279663e126cc0fba99e5",
"host": "172.31.2.13",
"port": 31458
}
]
}
]
},
{
"name": "b017c339-9406-48aa-ac90-4307e7f6343a_9050",
"port": 9050,
"protocol": "tcp",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "catalog.port",
"weight": 100,
"servers":
[
{
"name": "b017c339-9406-48aa-ac90-4307e7f6343a_9050",
"host": "172.31.1.11",
"port": 33000
}
]
}
]
},
{
"name": "b017c339-9406-48aa-ac90-4307e7f6343a_catalog_80",
"port": 33000,
"protocol": "http",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "catalog",
"weight": 100,
"servers":
[
{
"name": "888fac5307018583257f83b97a8324cf4f63279a",
"host": "172.31.2.12",
"port": 31305
}
]
}
]
},
{
"name": "b017c339-9406-48aa-ac90-4307e7f6343a_product_18080",
"port": 33002,
"protocol": "http",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "product",
"weight": 100,
"servers":
[
{
"name": "6f0c1359cff78fd679427a9e3a05ef78e74ae8fe",
"host": "172.31.2.12",
"port": 31188
}
]
}
]
},
{
"name": "b017c339-9406-48aa-ac90-4307e7f6343a_navigation_18090",
"port": 33003,
"protocol": "http",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "navigation",
"weight": 100,
"servers":
[
{
"name": "443c24f1197b8a5caf0a73a96b2f6e5285f7f8af",
"host": "172.31.2.11",
"port": 31799
}
]
}
]
},
{
"name": "b017c339-9406-48aa-ac90-4307e7f6343a_cart_18100",
"port": 33004,
"protocol": "http",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "cart",
"weight": 100,
"servers":
[
{
"name": "1174248e3ab5b52a9f3d3332eeff8e112d140518",
"host": "172.31.2.11",
"port": 31312
}
]
}
]
},
{
"name": "b017c339-9406-48aa-ac90-4307e7f6343a_redis_6379",
"port": 33005,
"protocol": "tcp",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "redis",
"weight": 100,
"servers":
[
{
"name": "7ff24e9ba4cafa7598f581e1c013e627b43640ec",
"host": "172.31.2.11",
"port": 31289
}
]
}
]
},
{
"name": "b017c339-9406-48aa-ac90-4307e7f6343a_mongodb_27017",
"port": 33006,
"protocol": "tcp",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "mongodb",
"weight": 100,
"servers":
[
{
"name": "1d249a608eb7de5ec0a26179bd666fcd71f1a308",
"host": "172.31.2.14",
"port": 31575
}
]
}
]
},
{
"name": "b017c339-9406-48aa-ac90-4307e7f6343a_checkout_8080",
"port": 33001,
"protocol": "http",
"httpQuota": { },
"tcpQuota": { },
"filters": [ ],
"services":
[
{
"name": "checkout",
"weight": 100,
"servers":
[
{
"name": "87e1e0f376ff3ebb04b294901749bb6967459442",
"host": "172.31.2.11",
"port": 31324
}
]
}
]
}
]
As you can see there are multiple routes for the port 30006:
cebe1679-e15d-41d7-a29f-c03bbcfd33ee_mongodb_27017
9b766960-3680-49c5-8395-5c05e4ac5e6d_mongodb_27017
cdee9d2d-8856-4144-811a-e0731056255b_mongodb_27017
b4d4a437-2a0f-4489-be9e-7118ec3237c1_mongodb_27017
b017c339-9406-48aa-ac90-4307e7f6343a_mongodb_27017
It seems the the router does not delete the old configuration of an old deployment When looking to the deployments call http://172.31.1.11:8081/api/v1/deployments only b017c339-9406-48aa-ac90-4307e7f6343a should deployed.
[
{
"name": "b017c339-9406-48aa-ac90-4307e7f6343a",
"endpoints":
{
"catalog.port": "9050"
},
"clusters":
{
"cart":
{
"services":
[
{
"state":
{
"name": "Deployed",
"started_at": "2015-09-08T09:20:47.312-07:00"
},
"breed":
{
"name": "cart",
"deployable": "docker://zutherb/cart-service:latest",
"ports":
{
"port": "18100/http"
},
"environment_variables":
{
"REDIS_PORT_6379_TCP_ADDR": "$redis.host",
"REDIS_PORT_6379_TCP_PORT": "$redis.ports.port"
},
"constants": { },
"dependencies":
{
"redis": "redis"
}
},
"environment_variables":
{
"REDIS_PORT_6379_TCP_ADDR": "172.31.1.11",
"REDIS_PORT_6379_TCP_PORT": "33005"
},
"scale":
{
"cpu": 0.1,
"memory": 128,
"instances": 1
},
"routing":
{
"weight": 100,
"filters": [ ]
},
"servers":
[
{
"name": "b017c339-9406-48aa-ac90-4307e7f6343a_cart.3d680f6c-5645-11e5-884c-56847afe9799",
"host": "172.31.2.11",
"ports":
{
"18100": 31312
},
"deployed": true
}
],
"dependencies":
{
"redis": "redis"
},
"dialects": { }
}
],
"routes":
{
"18100": 33004
},
"dialects": { }
},
"catalog":
{
"services":
[
{
"state":
{
"name": "Deployed",
"started_at": "2015-09-08T09:20:47.307-07:00"
},
"breed":
{
"name": "catalog",
"deployable": "docker://zutherb/catalog-frontend:latest",
"ports":
{
"port": "80/http"
},
"environment_variables":
{
"CHECKOUT_DESIGN": "standard",
"NAVIGATION_PORT_18090_TCP_PORT": "$navigation.ports.port",
"PRODUCT_PORT_18080_TCP_ADDR": "$product.host",
"PRODUCT_PORT_18080_TCP_PORT": "$product.ports.port",
"CART_PORT_18100_TCP_ADDR": "$cart.host",
"CART_PORT_18100_TCP_PORT": "$cart.ports.port",
"SHOP_PORT_8080_TCP_ADDR": "$checkout.host",
"NAVIGATION_PORT_18090_TCP_ADDR": "$navigation.host",
"SHOP_PORT_8080_TCP_PORT": "$checkout.ports.port"
},
"constants": { },
"dependencies":
{
"product": "product",
"navigation": "navigation",
"cart": "cart",
"checkout": "checkout"
}
},
"environment_variables":
{
"CHECKOUT_DESIGN": "standard",
"NAVIGATION_PORT_18090_TCP_PORT": "33003",
"PRODUCT_PORT_18080_TCP_ADDR": "172.31.1.11",
"PRODUCT_PORT_18080_TCP_PORT": "33002",
"CART_PORT_18100_TCP_ADDR": "172.31.1.11",
"CART_PORT_18100_TCP_PORT": "33004",
"SHOP_PORT_8080_TCP_ADDR": "172.31.1.11",
"NAVIGATION_PORT_18090_TCP_ADDR": "172.31.1.11",
"SHOP_PORT_8080_TCP_PORT": "33001"
},
"scale":
{
"cpu": 0.1,
"memory": 128,
"instances": 1
},
"routing":
{
"weight": 100,
"filters": [ ]
},
"servers":
[
{
"name": "b017c339-9406-48aa-ac90-4307e7f6343a_catalog.47d288a0-5645-11e5-884c-56847afe9799",
"host": "172.31.2.12",
"ports":
{
"80": 31305
},
"deployed": true
}
],
"dependencies":
{
"product": "product",
"navigation": "navigation",
"cart": "cart",
"checkout": "checkout"
},
"dialects": { }
}
],
"routes":
{
"80": 33000
},
"dialects": { }
},
"checkout":
{
"services":
[
{
"state":
{
"name": "Deployed",
"started_at": "2015-09-08T09:50:51.512-07:00"
},
"breed":
{
"name": "checkout",
"deployable": "docker://zutherb/monolithic-shop:latest",
"ports":
{
"port": "8080/http"
},
"environment_variables":
{
"CART_PORT_18100_TCP_ADDR": "$cart.host",
"CART_PORT_18100_TCP_PORT": "$cart.ports.port",
"MONGODB_PORT_27017_TCP_ADDR": "$mongodb.host",
"MONGODB_PORT_27017_TCP_PORT": "$mongodb.ports.port"
},
"constants": { },
"dependencies":
{
"mongodb": "mongodb",
"cart": "cart"
}
},
"environment_variables":
{
"CART_PORT_18100_TCP_ADDR": "172.31.1.11",
"CART_PORT_18100_TCP_PORT": "33004",
"MONGODB_PORT_27017_TCP_ADDR": "172.31.1.11",
"MONGODB_PORT_27017_TCP_PORT": "33006"
},
"scale":
{
"cpu": 0.3,
"memory": 256,
"instances": 1
},
"routing":
{
"weight": 100,
"filters": [ ]
},
"servers":
[
{
"name": "b017c339-9406-48aa-ac90-4307e7f6343a_checkout.bf6e0d95-5649-11e5-884c-56847afe9799",
"host": "172.31.2.11",
"ports":
{
"8080": 31324
},
"deployed": true
}
],
"dependencies":
{
"mongodb": "mongodb",
"cart": "cart"
},
"dialects": { }
}
],
"routes":
{
"8080": 33001
},
"dialects": { }
},
"navigation":
{
"services":
[
{
"state":
{
"name": "Deployed",
"started_at": "2015-09-08T09:20:47.307-07:00"
},
"breed":
{
"name": "navigation",
"deployable": "docker://magneticio/navigation-service:latest",
"ports":
{
"port": "18090/http"
},
"environment_variables":
{
"MONGODB_PORT_27017_TCP_ADDR": "$mongodb.host",
"MONGODB_PORT_27017_TCP_PORT": "$mongodb.ports.port"
},
"constants": { },
"dependencies":
{
"mongodb": "mongodb"
}
},
"environment_variables":
{
"MONGODB_PORT_27017_TCP_ADDR": "172.31.1.11",
"MONGODB_PORT_27017_TCP_PORT": "33006"
},
"scale":
{
"cpu": 0.1,
"memory": 128,
"instances": 1
},
"routing":
{
"weight": 100,
"filters": [ ]
},
"servers":
[
{
"name": "b017c339-9406-48aa-ac90-4307e7f6343a_navigation.49084982-5645-11e5-884c-56847afe9799",
"host": "172.31.2.11",
"ports":
{
"18090": 31799
},
"deployed": true
}
],
"dependencies":
{
"mongodb": "mongodb"
},
"dialects":
{
"marathon":
{
"cmd": "export PORT=18090 && /navigation-0.6/bin/navigation"
}
}
}
],
"routes":
{
"18090": 33003
},
"dialects": { }
},
"redis":
{
"services":
[
{
"state":
{
"name": "Deployed",
"started_at": "2015-09-08T09:20:47.312-07:00"
},
"breed":
{
"name": "redis",
"deployable": "docker://redis:latest",
"ports":
{
"port": "6379/tcp"
},
"environment_variables": { },
"constants": { },
"dependencies": { }
},
"environment_variables": { },
"scale":
{
"cpu": 0.1,
"memory": 128,
"instances": 1
},
"routing":
{
"weight": 100,
"filters": [ ]
},
"servers":
[
{
"name": "b017c339-9406-48aa-ac90-4307e7f6343a_redis.44d5000f-5645-11e5-884c-56847afe9799",
"host": "172.31.2.11",
"ports":
{
"6379": 31289
},
"deployed": true
}
],
"dependencies": { },
"dialects": { }
}
],
"routes":
{
"6379": 33005
},
"dialects": { }
},
"mongodb":
{
"services":
[
{
"state":
{
"name": "Deployed",
"started_at": "2015-09-08T09:50:41.556-07:00"
},
"breed":
{
"name": "mongodb",
"deployable": "docker://mongo:latest",
"ports":
{
"port": "27017/tcp"
},
"environment_variables": { },
"constants": { },
"dependencies": { }
},
"environment_variables": { },
"scale":
{
"cpu": 0.1,
"memory": 128,
"instances": 1
},
"routing":
{
"weight": 100,
"filters": [ ]
},
"servers":
[
{
"name": "b017c339-9406-48aa-ac90-4307e7f6343a_mongodb.ba0ec744-5649-11e5-884c-56847afe9799",
"host": "172.31.2.14",
"ports":
{
"27017": 31575
},
"deployed": true
}
],
"dependencies": { },
"dialects":
{
"marathon":
{
"cmd": "mongod --smallfiles"
}
}
}
],
"routes":
{
"27017": 33006
},
"dialects": { }
},
"product":
{
"services":
[
{
"state":
{
"name": "Deployed",
"started_at": "2015-09-08T09:20:47.307-07:00"
},
"breed":
{
"name": "product",
"deployable": "docker://zutherb/product-service",
"ports":
{
"port": "18080/http"
},
"environment_variables":
{
"MONGODB_PORT_27017_TCP_ADDR": "$mongodb.host",
"MONGODB_PORT_27017_TCP_PORT": "$mongodb.ports.port"
},
"constants": { },
"dependencies":
{
"mongodb": "mongodb"
}
},
"environment_variables":
{
"MONGODB_PORT_27017_TCP_ADDR": "172.31.1.11",
"MONGODB_PORT_27017_TCP_PORT": "33006"
},
"scale":
{
"cpu": 0.1,
"memory": 128,
"instances": 1
},
"routing":
{
"weight": 100,
"filters": [ ]
},
"servers":
[
{
"name": "b017c339-9406-48aa-ac90-4307e7f6343a_product.44a4a13e-5645-11e5-884c-56847afe9799",
"host": "172.31.2.12",
"ports":
{
"18080": 31188
},
"deployed": true
}
],
"dependencies":
{
"mongodb": "mongodb"
},
"dialects": { }
}
],
"routes":
{
"18080": 33002
},
"dialects": { }
}
},
"ports":
{
"navigation.port": "33003",
"cart.port": "33004",
"checkout.port": "33001",
"catalog.port": "33000",
"mongodb.port": "33006",
"redis.port": "33005",
"product.port": "33002"
},
"environment_variables": { },
"hosts":
{
"checkout": "172.31.1.11",
"catalog": "172.31.1.11",
"cart": "172.31.1.11",
"navigation": "172.31.1.11",
"product": "172.31.1.11",
"mongodb": "172.31.1.11",
"redis": "172.31.1.11"
}
}
]
Is there a way to reset the router?
@tnolet
I unlock you for the vm.
My Assumption seems to be correct. After i deleted the following to files in /usr/share/vamp-router/configuration the haproxy works as excepted even that the external port 9050 is not started:
haproxy_new.cfg
vamp_router.json
@zutherb
It seems like indeed Vamp Core and Router were out of sync. You can always reset Vamp completely by using the /api/v1/reset command on the Vamp Core API. See http://vamp.io/documentation/api-reference/debug/
Or you can just reset Vamp Router by posting an empty config to the /v1/config endpoint of Vamp Router API, i.e.:
POST /v1/config
{
"frontends" : [],
"backends": [],
"routes" : []
}
I got the VM downloaded and started. I just need a username + password now.
@tnolet thx, username + password is vagrant + vagrant
@zutherb I can see the machine booting, but I guess its missing some Puppet files as it throws errors about missing those files. When trying to log in it seems the user is either incorrect or not setup:
Tims-MacBook-Pro:VirtualBox VMs tim$ ssh -p 2222 vagrant@127.0.0.1
The authenticity of host '[127.0.0.1]:2222 ([127.0.0.1]:2222)' can't be established.
RSA key fingerprint is 95:88:a8:f9:8f:a1:d1:b6:9b:a6:a3:76:d9:06:9b:68.
Are you sure you want to continue connecting (yes/no)? yes
Warning: Permanently added '[127.0.0.1]:2222' (RSA) to the list of known hosts.
vagrant@127.0.0.1's password:
Permission denied, please try again.
@tnolet hmm ... normally puppet is only need for provisioning the vm.
Can you try to looking with the vagrant inscure key?
https://github.com/mitchellh/vagrant/blob/master/keys/vagrant
@zutherb That works! Great!
@zutherb
There are a couple of issues with the way this particular setup. Luckily, they are all pretty easy to fix:
You are running HAProxy as root or Haproxy was at some point in the past started as root.Vamp Router (when started with 'service vamp-router start') runs under the 'vamp-router' user. You can see this in /etc/init/vamp-router.conf
exec sudo -u vamp-router bin/vamp-router
This means all .sock files and .pid files that Haproxy writes are not accessible to Vamp Router and therefore it cannot start. I fixed this in your VMBox image by first killing haproxy and then chowning all files to vamp-router:
cd /usr/share/vamp-router/
sudo chown -R vamp-router:vamp-router *
I also removed the config files as you did mentioned earlier. We now have a clean start.
I then startup Vamp Router and all is well:
sudo service vamp-router start
with the result of haproxy AND vamp-router running as the same user.
vagrant@master1:/usr/share/vamp-router$ ps -ef | grep vamp-router
root 9856 1 0 05:49 ? 00:00:00 sudo -u vamp-router bin/vamp-router
vamp-ro+ 9858 9856 0 05:49 ? 00:00:00 bash bin/vamp-router
vamp-ro+ 9859 9858 61 05:49 ? 00:00:01 ./vamp-router
vamp-ro+ 9864 1 0 05:49 ? 00:00:00 /usr/sbin/haproxy -f /usr/share/vamp-router/configuration/haproxy_new.cfg -p /usr/share/vamp-router/data/haproxy-privat
You can of course ignore this and just start Vamp Router as root. This also allows you to run HAproxy as root and bind to port 80 for instance. This is done by either editing the /etc/init/vamp-router.conf file or just using something like the following command to start vamp-router and push it to the background.
sudo /usr/share/vamp-router/vamp-router > /dev/null 2>&1 &
@tnolet thx a lot it works.
added to the FAQ.
|
gharchive/issue
| 2015-08-07T15:23:17 |
2025-04-01T06:39:28.939200
|
{
"authors": [
"plamola",
"tnolet",
"zutherb"
],
"repo": "magneticio/vamp-router",
"url": "https://github.com/magneticio/vamp-router/issues/22",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
276854359
|
fork() Does Not Occur During Run Tests
When performing run tests in travis CI, the processes are not being forked. This results in reduced test coverage of the logic performed after fork() runs, confirmed by reviewing Coveralls results.
I suspect this is due to limitations with Travis CI in order to prevent fork bombs.
What we'll do to fix this is check to see if the test_run CLI flag is set, and if so, continue running the contained logic.
Found this exception when running coverage tests:
Traceback (most recent call last):
File "inquisition.py", line 173, in <module>
main()
File "inquisition.py", line 150, in main
anatomize.startAnatomizer()
File "/home/travis/build/magneticstain/Inquisition/lib/anatomize/Anatomize.py", line 150, in startAnatomizer
numLogsBetweenTrackingUpdate=numLogsBetweenTrackingUpdate)
File "/home/travis/build/magneticstain/Inquisition/lib/anatomize/Parser.py", line 523, in pollLogFile
every_n=numLogsBetweenTrackingUpdate):
File "/home/travis/virtualenv/python3.4.6/lib/python3.4/site-packages/pygtail/core.py", line 81, in __init__
[int(line.strip()) for line in offset_fh]
ValueError: need more than 0 values to unpack
I found another bug in my last commit where new process spawn even when in test mode. I fixed that, and now it seems that it has fixed this exception as well.
Still not sure how, but I suspect it's some sort of race condition between parser processes. Fixing this so that all logic runs in a single process during test mode has resolved it during local testing.
https://sentry.io/carlsonet/inquisition/issues/410604874/
https://sentry.io/carlsonet/inquisition/issues/410604750/
|
gharchive/issue
| 2017-11-26T20:35:48 |
2025-04-01T06:39:28.961930
|
{
"authors": [
"magneticstain"
],
"repo": "magneticstain/Inquisition",
"url": "https://github.com/magneticstain/Inquisition/issues/69",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1682344389
|
Add tokenizer to encode input text and decode predicted image from logits
Tokenizer to encode text using BartTokenizer. It decodes Bart decoder output of image tokens to image using VQGAN decoder.
Currently getting an error when trying the tokenizer:
‘BartTokenizer’ object has no attribute ‘to’, for running the line tokenizer.to(device)
Not working with cpu nor cuda.
Trying to see if there's an easy fix.
Let me fix it
|
gharchive/pull-request
| 2023-04-25T02:54:40 |
2025-04-01T06:39:28.980804
|
{
"authors": [
"ArjunKrishna3367",
"mahalrs"
],
"repo": "mahalrs/newsgen",
"url": "https://github.com/mahalrs/newsgen/pull/21",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1909959745
|
#CleanCode No inline js and use configuration variables
@mai-soup The thing with inline js is - just don't do it. It has so many drawbacks:
A bad separation of concerns because the view shouldn't know about the logic
It's hard to read, and even harder to maintain because it's inline instead of a dynamic environment setting linked in a configuration file
Debugging hell, if something goes wrong
....
LoansView.vue
// if the loan is due in 7 days or less, show a button to extend the loan
button(v-if="(new Date(loan.returnDate) - new Date()) < 1000 * 60 * 60 * 24 * 7" @click="doExtend(loan)") Extend
The idea here would be using a function call from JS and configure this value in your .env.
instead of environment variables, decided to assign the loan-related durations on a per-library basis (#113)
|
gharchive/issue
| 2023-09-23T18:25:48 |
2025-04-01T06:39:29.011347
|
{
"authors": [
"Dr4gon",
"mai-soup"
],
"repo": "mai-soup/porch-reads-club",
"url": "https://github.com/mai-soup/porch-reads-club/issues/77",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
283931148
|
fix/use promise to await for archiver to finalize before applying sha…
…256sum
r? @krishnaIndia
(maidsafe_highfive has picked a reviewer for you, use r? to override)
Draft based on this PR: https://github.com/maidsafe/safe_examples/releases/tag/untagged-437234bec79b2cf8b510
|
gharchive/pull-request
| 2017-12-21T16:15:31 |
2025-04-01T06:39:29.015909
|
{
"authors": [
"hunterlester",
"maidsafe-highfive"
],
"repo": "maidsafe/safe_examples",
"url": "https://github.com/maidsafe/safe_examples/pull/339",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
59371552
|
Fixing two iOS display issues
Mail app on iOS 8 the default body padding is 20px. This makes the template rather skinny. Most modern email templates have a smaller left and right body padding of 10px.
20px padding: http://cl.ly/image/0x2b3e2P1i2b
10px padding: http://cl.ly/image/1p2h1b2Z0m1E
Mail app on iOS 8, Helvetica's 600 weight font looks wonky. The letter-spacing and font-weight looks off.
600 weight: http://cl.ly/image/0x2b3e2P1i2b
800 weight: http://cl.ly/image/1p2h1b2Z0m1E
Changed this in a few places ad13d824800bcd968870e1714cc1dea465ce1c1c
Thanks
|
gharchive/pull-request
| 2015-03-01T02:04:28 |
2025-04-01T06:39:29.078091
|
{
"authors": [
"blakeperdue",
"leemunroe"
],
"repo": "mailgun/transactional-email-templates",
"url": "https://github.com/mailgun/transactional-email-templates/pull/10",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
422599216
|
Feature request: Web browser notification
It would be nice to have mailhog capable to display notifications upon receiving messages. This would save clicking on mailhog tab every time to see if a message is received.
Which browser? Here my experience on a Mac:
Firefox: It seems to work. It even plays a bing sound.
Safari: Shows the notifcation, but no sound. (dunno why not playing sound)
Chrome: No notification, no sound. (dunno why)
Screenshot: Mailhog is not the active tab, but receives the notification.
@peterdd Chrome, Linux, Version 72.0.3626.121 (Official Build) (64-bit)
Maybe I have the reason: https://developer.mozilla.org/en-US/docs/Web/API/notificationSecure context
"This feature is available only in secure contexts (HTTPS), in some or all supporting browsers."
So in the latest Firefox 67 it is not working anymore - unless I change in about:config the dom.webnotifications.allowinsecure to false when you use mailhog just local for testing without TLS.
(notification works again, with sound)
In Chrome 74 I fiddled with chrome://flags
#unsafely-treat-insecure-origin-as-secure (enabled and insert address of your mailhog running)
and
#enable-message-center-new-style-notification (enabled)
to get them working for a local http ip address. ( http://10.0.0.x:8025 )
(notifications shown, no sound, don't know why)
Hi,
With Firefox 74, even this does not seem to work.
So in the latest Firefox 67 it is not working anymore - unless I change in about:config the dom.webnotifications.allowinsecure to false when you use mailhog just local for testing without TLS.
Any other way to make this work now?
Sorry, I meant set
dom.webnotifications.allowinsecure to true
With Firefox 74 was able to get the notification alerts on mac again. New seems the little speech bubble icon in the address bar. Click and allow it in the popup dialog and you get the other icon as shown in screnshot.
@peterdd Perfect! Both changes done and it works on Windows 10 & Firefox 74! Thanks a lot!
|
gharchive/issue
| 2019-03-19T08:26:13 |
2025-04-01T06:39:29.085503
|
{
"authors": [
"KoolPal",
"OnkelTem",
"peterdd"
],
"repo": "mailhog/MailHog",
"url": "https://github.com/mailhog/MailHog/issues/243",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2134649440
|
getting Reference to property 'player' in closure requires explicit use of 'self' to make capture semantics explicit
While running in ios, we are getting the following error:/Users/prachhhi/Documents/yes/app/AppV2/node_modules/react-native-video-trim/ios/VideoTrimmerViewController.swift:254:16 Reference to property 'player' in closure requires explicit use of 'self' to make capture semantics explicit
which version of this library and RN are you using?
Have you run Pod install? can you try this in new project, because with the above errors I'll definitely will see it and never publish :)
RN version: 0.72.4
react-native-video-trim version: 1.0.10
yes we have ran pod install.
Yes correct and apologies for not adding full context, the same code is running in another system so we suspect something wrong with xcode settings. But as this issue is coming with library only, I thought of asking for help here.
Can you click each of the error then screenshot and show me exact location for each error?
This file contains all the errors
Let me push a fix for this, I think this is about backward compatibility, it's really fine on my end
Sure, thanks a lot, yes I am also able to run on another system with latest xcode so looks like that only.
Hey @maitrungduc1410, Yep this worked, thanks for this, now I am thinking why I didnt just patched on my own and raised a PR lol.
Again thanks for quickly responding and helping.
|
gharchive/issue
| 2024-02-14T15:47:24 |
2025-04-01T06:39:29.093156
|
{
"authors": [
"amosaxe",
"maitrungduc1410"
],
"repo": "maitrungduc1410/react-native-video-trim",
"url": "https://github.com/maitrungduc1410/react-native-video-trim/issues/37",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2616599487
|
Request support for A1 PRO MAX
Log message
2024-10-24 21:23:05.224 WARNING (MainThread) [custom_components.tuya_local.config_flow] Device matches otu_r1o1_lock with quality of 33%. DPS: {"updated_at": 1729797769.2820706, "9": "high", "16": 0, "19": 1, "33": false, "36": 1, "46": true, "47": true, "62": 1, "78": true}
2024-10-24 21:23:05.224 WARNING (MainThread) [custom_components.tuya_local.config_flow] Include the previous log message with any new device request to https://github.com/make-all/tuya-local/issues/
DPS information
{
"result": {
"model": "{\"modelId\":\"000003hp6s\",\"services\":[{\"actions\":[],\"code\":\"\",\"description\":\"\",\"events\":[],\"name\":\"默认服务\",\"properties\":[{\"abilityId\":1,\"accessMode\":\"rw\",\"code\":\"unlock_method_create\",\"description\":\"【添加开锁方式】设备在线后,添加开锁方式。支持添加指密卡和人脸,以及支持管理员标记。\",\"extensions\":{\"iconName\":\"icon-a_up\",\"attribute\":\"1792\"},\"name\":\"添加开门方式\",\"typeSpec\":{\"type\":\"raw\",\"maxlen\":128}},{\"abilityId\":2,\"accessMode\":\"rw\",\"code\":\"unlock_method_delete\",\"description\":\"【删除开锁方式】设备在线后,删除开锁方式。支持删除指密卡和人脸。\",\"extensions\":{\"iconName\":\"icon-dp_rabbish\",\"attribute\":\"1792\"},\"name\":\"删除开门方式\",\"typeSpec\":{\"type\":\"raw\",\"maxlen\":128}},{\"abilityId\":3,\"accessMode\":\"rw\",\"code\":\"unlock_method_modify\",\"description\":\"【修改开锁方式】设备在线后,支持修改开锁方式的名称,以及密码的内容修改。\",\"extensions\":{\"iconName\":\"icon-edit\",\"attribute\":\"1792\"},\"name\":\"修改开门方式\",\"typeSpec\":{\"type\":\"raw\",\"maxlen\":128}},{\"abilityId\":9,\"accessMode\":\"ro\",\"code\":\"battery_state\",\"description\":\"【设备状态】用于电量以状态等级显示。范围高中低。(与DP8二选一)\",\"extensions\":{\"iconName\":\"icon-dp_battery\",\"attribute\":\"1280\",\"trigger\":\"direct\"},\"name\":\"电量状态\",\"typeSpec\":{\"type\":\"enum\",\"range\":[\"high\",\"medium\",\"low\",\"poweroff\"]}},{\"abilityId\":16,\"accessMode\":\"ro\",\"code\":\"unlock_key\",\"description\":\"【上报开锁记录】若设备本地支持钥匙开锁检测,则用于设备上报机械钥匙开锁的记录。\",\"extensions\":{\"iconName\":\"icon-dp_lock\",\"attribute\":\"1280\",\"trigger\":\"direct\"},\"name\":\"本地手动 开锁\",\"typeSpec\":{\"type\":\"value\",\"max\":999,\"min\":0,\"scale\":0,\"step\":1,\"unit\":\"\"}},{\"abilityId\":19,\"accessMode\":\"ro\",\"code\":\"unlock_ble\",\"description\":\"【上报开锁记录】用于设备上报手机蓝牙开锁时的记录。\",\"extensions\":{\"iconName\":\"icon-dp_lock\",\"attribute\":\"1280\",\"trigger\":\"direct\"},\"name\":\"蓝牙解锁记录\",\"typeSpec\":{\"type\":\"value\",\"max\":999,\"min\":0,\"scale\":0,\"step\":1}},{\"abilityId\":20,\"accessMode\":\"rw\",\"code\":\"lock_record\",\"description\":\"【上报关锁记录】用于设备上报门锁关锁的记录。\",\"extensions\":{\"iconName\":\"icon-dp_lock\",\"attribute\":\"256\"},\"name\":\"关锁记录\",\"typeSpec\":{\"type\":\"raw\",\"maxlen\":128}},{\"abilityId\":21,\"accessMode\":\"ro\",\"code\":\"alarm_lock\",\"description\":\"【告警记录】本地发生告警时,用于上报设备的告警记录。对应值说明:0=指纹试错报警、1=密码试错报警、2=卡试错报警、3=人脸试错报警、4=假锁(锁舌卡住)、5=高温报警、6=超时未关门、7=电子锁舌未弹出、8=防撬报警 、9=钥匙插入、10=低电报警、11=电量耗尽报警、12=震动报警、13=布防报警,如设备只用到告警dp中的几项,多余的内容也不需要删除。如删除则mcu在上报时的序号需要调整。建议开发者不要改动告警中的枚举值。如遇特殊情况,请在pm的指导下,进行增、删,以及配置多语言等操作\",\"extensions\":{\"iconName\":\"icon-dp_warming\",\"attribute\":\"1280\",\"trigger\":\"direct\"},\"name\":\"门锁告警\",\"typeSpec\":{\"type\":\"enum\",\"range\":[\"low_battery\",\"power_off\"]}},{\"abilityId\":31,\"accessMode\":\"rw\",\"code\":\"beep_volume\",\"description\":\"【音量】用于设置设备本地导航语音音量大小。\",\"extensions\":{\"iconName\":\"icon-sound\",\"attribute\":\"1280\"},\"name\":\"提示音量\",\"typeSpec\":{\"type\":\"enum\",\"range\":[\"mute\",\"normal\"]}},{\"abilityId\":33,\"accessMode\":\"rw\",\"code\":\"automatic_lock\",\"description\":\"【自动落锁】用于设置自动落锁功能的开关。关闭后即设备不会自动落锁,此时也可作为常开模式。\",\"extensions\":{\"iconName\":\"icon-power\",\"attribute\":\"1280\",\"trigger\":\"direct\"},\"name\":\"自动落锁开关\",\"typeSpec\":{\"type\":\"bool\"}},{\"abilityId\":36,\"accessMode\":\"rw\",\"code\":\"auto_lock_time\",\"description\":\"【自动落锁】用于设置设备解锁后到执行自动落锁的延迟时间。设置范围1s~30min。\",\"extensions\":{\"iconName\":\"icon-dp_time3\",\"attribute\":\"1280\",\"trigger\":\"direct\"},\"name\":\"自动落锁延时设置\",\"typeSpec\":{\"type\":\"value\",\"max\":1800,\"min\":1,\"scale\":0,\"step\":1}},{\"abilityId\":44,\"accessMode\":\"rw\",\"code\":\"rtc_lock\",\"description\":\"【时钟RTC】若设备本地支持RTC功能,则选择该DP,用于校准云端和本地的时间。\",\"extensions\":{\"iconName\":\"icon-dp_time3\",\"attribute\":\"1280\",\"trigger\":\"direct\"},\"name\":\"硬件时钟RTC\",\"typeSpec\":{\"type\":\"bool\"}},{\"abilityId\":46,\"accessMode\":\"rw\",\"code\":\"manual_lock\",\"description\":\"【落锁操作】用于触发设备执行落锁操作,可用于自动落锁和手动落锁。\",\"extensions\":{\"iconName\":\"icon-dp_lock\",\"attribute\":\"1280\",\"trigger\":\"direct\"},\"name\":\"手动落锁\",\"typeSpec\":{\"type\":\"bool\"}},{\"abilityId\":47,\"accessMode\":\"ro\",\"code\":\"lock_motor_state\",\"description\":\"【设备状态】用于上报设备离合开关的状态,即锁的开关状态。\",\"extensions\":{\"iconName\":\"icon-dp_lock\",\"attribute\":\"1280\",\"trigger\":\"direct\"},\"name\":\"落锁状态\",\"typeSpec\":{\"type\":\"bool\"}},{\"abilityId\":54,\"accessMode\":\"rw\",\"code\":\"synch_method\",\"description\":\"【同步开锁方式】用于手机同步设备本地所有的开锁方式。\",\"extensions\":{\"iconName\":\"icon-dp_upload\",\"attribute\":\"1792\"},\"name\":\"同步开门方式(全量同步)\",\"typeSpec\":{\"type\":\"raw\",\"maxlen\":128}},{\"abilityId\":61,\"accessMode\":\"rw\",\"code\":\"remote_no_dp_key\",\"description\":\"【远程开锁】用于触发设备远程开锁操作。\",\"extensions\":{\"iconName\":\"icon-dp_lock\",\"attribute\":\"1280\"},\"name\":\"远程开门(带校验码)\",\"typeSpec\":{\"type\":\"raw\",\"maxlen\":128}},{\"abilityId\":62,\"accessMode\":\"ro\",\"code\":\"unlock_phone_remote\",\"description\":\"【上报开锁记录】用于设备上报手机远程开锁的记录。\",\"extensions\":{\"iconName\":\"icon-shouji\",\"attribute\":\"1280\",\"trigger\":\"direct\"},\"name\":\"远程手机解锁\",\"typeSpec\":{\"type\":\"value\",\"max\":999,\"min\":0,\"scale\":0,\"step\":1}},{\"abilityId\":63,\"accessMode\":\"ro\",\"code\":\"unlock_voice_remote\",\"description\":\"【上报开锁记录】用于设备上报语音语音开锁的记录。\",\"extensions\":{\"iconName\":\"icon-dp_mic\",\"attribute\":\"1280\",\"trigger\":\"direct\"},\"name\":\"远程语音解锁\",\"typeSpec\":{\"type\":\"value\",\"max\":999,\"min\":0,\"scale\":0,\"step\":1}},{\"abilityId\":68,\"accessMode\":\"rw\",\"code\":\"special_function\",\"description\":\"【特殊功能】可根据自身需求用于设置其他特殊功能的开关。对应文案可在IoT平台进行修改。\",\"extensions\":{\"iconName\":\"icon-dp_mode\",\"attribute\":\"1280\",\"trigger\":\"direct\"},\"name\":\"校准\",\"typeSpec\":{\"type\":\"enum\",\"range\":[\"function1\",\"function2\",\"function3\",\"function4\"]}},{\"abilityId\":69,\"accessMode\":\"rw\",\"code\":\"record\",\"description\":\"【获取开锁记录】用于手机主动获取设备本地存储的开门记录。\",\"extensions\":{\"iconName\":\"icon-dp_book\",\"attribute\":\"1280\"},\"name\":\"获取门锁记录\",\"typeSpec\":{\"type\":\"raw\",\"maxlen\":128}},{\"abilityId\":70,\"accessMode\":\"rw\",\"code\":\"check_code_set\",\"description\":\"【蓝牙开锁】用于配置设备执行蓝牙解锁时需要的校验信息。\",\"extensions\":{\"iconName\":\"icon-dp_mode\",\"attribute\":\"1280\"},\"name\":\"配置蓝牙解锁校验码\",\"typeSpec\":{\"type\":\"raw\",\"maxlen\":128}},{\"abilityId\":71,\"accessMode\":\"rw\",\"code\":\"ble_unlock_check\",\"description\":\"【蓝牙开锁】用于触发设备执行蓝牙解锁。触发源可以是手机、配件或靠近解锁。\",\"extensions\":{\"iconName\":\"icon-dp_lock\",\"attribute\":\"1280\"},\"name\":\"蓝牙解锁(带校验码)\",\"typeSpec\":{\"type\":\"raw\",\"maxlen\":128}},{\"abilityId\":72,\"accessMode\":\"ro\",\"code\":\"unlock_record_check\",\"description\":\"【上报开锁记录】用于设备上报配件开锁及开关锁的记录。需要包含设备端具体触发开关锁的方式。\",\"extensions\":{\"iconName\":\"icon-tongji2\",\"attribute\":\"1280\"},\"name\":\"(带校验码)开关锁记录\",\"typeSpec\":{\"type\":\"raw\",\"maxlen\":128}},{\"abilityId\":73,\"accessMode\":\"rw\",\"code\":\"remote_pd_setkey_check\",\"description\":\"【远程开锁】用于配置远程开锁时需要的密钥。若设备需要进行远程开锁,该DP必选。\",\"extensions\":{\"iconName\":\"icon-dp_lock\",\"attribute\":\"1792\"},\"name\":\"配置远程解锁含校验\",\"typeSpec\":{\"type\":\"raw\",\"maxlen\":128}},{\"abilityId\":78,\"accessMode\":\"rw\",\"code\":\"special_control\",\"description\":\"【特殊控制】可根据自身需求用于触发其他特殊控制能力。对应文案可在IoT平台进行修改。\",\"extensions\":{\"iconName\":\"icon-dp_power2\",\"attribute\":\"1280\"},\"name\":\"开关方向\",\"typeSpec\":{\"type\":\"bool\"}}]}]}"
},
"success": true,
"t": 1730035720957,
"tid": "617139f4946711efad5172ecb0bd0dcf"
}
Product ID
rlyxv7pe
Product Name
A1 PRO MAX
Information about how the device functions
No response
What is an A1 PRO MAX?
I can did it changing some optionals:
`name: Door lock
products:
id: rlyxv7pe
name: Raykube A1 Pro Max
primary_entity:
entity: lock
dps:
id: 1
type: string
optional: true
name: unlock_method_create
id: 2
type: string
optional: true
name: unlock_method_delete
id: 3
type: string
optional: true
name: unlock_method_modify
id: 16
type: integer
optional: true
persist: false
name: unlock_key
id: 19
type: integer
optional: true
persist: false
name: unlock_ble
id: 20
type: string
optional: true
name: lock_record
id: 46
type: boolean
optional: true
name: lock
id: 47
type: boolean
optional: true
name: open
readonly: true
id: 54
type: string
optional: true
name: sync_method
id: 61
type: string
optional: true
name: remote_no_dp_key
id: 62
type: integer
optional: true
persist: false
name: unlock_app
id: 63
type: integer
optional: true
persist: false
name: unlock_voice
id: 69
type: string
optional: true
name: record
id: 70
type: string
optional: true
name: check_code_set
id: 71
type: string
optional: true
name: ble_unlock_check
id: 72
type: string
optional: true
name: unlock_record_check
id: 73
type: string
optional: true
name: remote_pd_setkey_check
secondary_entities:
entity: sensor
class: battery
category: diagnostic
dps:
id: 9
type: string
name: sensor
unit: "%"
mapping:
dps_val: high
value: 90
dps_val: medium
value: 60
dps_val: low
value: 30
dps_val: poweroff
value: 0
entity: event
name: Alert
dps:
id: 21
type: string
name: event
optional: true
mapping:
dps_val: low_battery
value: low_battery
dps_val: power_off
value: power_off
entity: switch
translation_key: keytone
category: config
dps:
id: 31
type: string
optional: true
name: switch
mapping:
dps_val: normal
value: true
dps_val: mute
value: false
entity: switch
name: Auto-lock
icon: "mdi:lock-clock"
category: config
dps:
id: 33
type: boolean
optional: true
name: switch
entity: number
name: Auto lock time
category: config
icon: "mdi:lock-clock"
dps:
id: 36
type: integer
optional: true
name: value
unit: s
range:
min: 1
max: 1800
entity: button
name: Sync clock
icon: "mdi:clock"
category: config
dps:
id: 44
type: boolean
name: button
optional: true
entity: select
name: Configuration
icon: "mdi:function"
category: config
dps:
id: 68
type: string
optional: true
name: option
mapping:
dps_val: function1
value: Recalibrate
dps_val: function2
value: Unlock more
dps_val: function3
value: Keep retracted
dps_val: function4
value: Add force
entity: switch
name: Change direction
category: config
dps:
id: 78
type: boolean
optional: true
name: switch
`
Making things optional blindly does not help, as it will end up false matching too many other devices.
It is better that you send the logs you are getting using the original config, so the minimal set of optional dps can be worked out.
|
gharchive/issue
| 2024-10-27T13:29:38 |
2025-04-01T06:39:29.127188
|
{
"authors": [
"ChrisDes",
"make-all"
],
"repo": "make-all/tuya-local",
"url": "https://github.com/make-all/tuya-local/issues/2439",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1162669660
|
Fix coreJVM/initialCommands to run
Changes
Fix coreJVM/initialCommands to run
Codecov Report
Merging #379 (4d34971) into main (6246719) will not change coverage.
The diff coverage is n/a.
@@ Coverage Diff @@
## main #379 +/- ##
==========================================
Coverage 100.00% 100.00%
==========================================
Files 66 56 -10
Lines 2666 1848 -818
Branches 241 88 -153
==========================================
- Hits 2666 1848 -818
Impacted Files
Coverage Δ
packages/recheck/src/lib/env.ts
packages/recheck/src/browser.ts
packages/recheck/src/lib/java.ts
...ges/eslint-plugin-redos/src/rules/no-vulnerable.ts
packages/recheck/src/lib/pure.ts
packages/recheck/src/lib/exe.ts
packages/recheck/src/lib/native.ts
packages/recheck/src/main.ts
packages/recheck/src/lib/worker-pool.ts
packages/recheck/src/lib/agent.ts
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 6246719...4d34971. Read the comment docs.
|
gharchive/pull-request
| 2022-03-08T13:38:44 |
2025-04-01T06:39:29.147984
|
{
"authors": [
"codecov-commenter",
"makenowjust"
],
"repo": "makenowjust-labs/recheck",
"url": "https://github.com/makenowjust-labs/recheck/pull/379",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
903425615
|
To add working steps in T shape calculator
Is your feature request related to a problem? Please describe.
A clear and concise description of what the problem is. Ex. I'm always frustrated when [...]
Describe the solution you'd like
A clear and concise description of what you want to happen.
Hi, I want to work on this issue I will start working on it as soon as I get assigned!! I am a part of GSSoC'21.
Kindly assign this issue to me!
Discord Username : Abhijeet Sinha(P)
Discord Tag: #4018
@abhijeet141 I'm already working on issue #3924. Please close this issue
|
gharchive/issue
| 2021-05-27T08:53:12 |
2025-04-01T06:39:29.181386
|
{
"authors": [
"abhijeet141",
"geekymeeky"
],
"repo": "makesmatheasy/makesmatheasy",
"url": "https://github.com/makesmatheasy/makesmatheasy/issues/3931",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
128794142
|
Static pages - Cible WEB and HIDDEN displayed in Mobile
Just tried to create a static page with Cible = Web.
It is displayed in Rando but also in Mobile.
Just tried to create a HIDDEN static page and this one is also displayed in Geotrek-mobile.
There's indeed an issue in the condition. It still tests the old value for target and should be updated.
|
gharchive/issue
| 2016-01-26T11:21:13 |
2025-04-01T06:39:29.188586
|
{
"authors": [
"SBats",
"camillemonchicourt"
],
"repo": "makinacorpus/Geotrek-mobile",
"url": "https://github.com/makinacorpus/Geotrek-mobile/issues/182",
"license": "BSD-2-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
1941778458
|
Fix removal of script contexts
This should fix #79.
Okay i found following issue:
the system script_remove_synchronizer handles removing the contexts of the script.
https://github.com/makspll/bevy_mod_scripting/blob/32eba12bc5e2d87ab05cfbf4a2dd14063d3842a2/bevy_mod_scripting_core/src/systems.rs#L94-L103
The problem here is, that .remove_context expects the script ID, as context_entities uses it as the Key.
https://github.com/makspll/bevy_mod_scripting/blob/32eba12bc5e2d87ab05cfbf4a2dd14063d3842a2/bevy_mod_scripting_core/src/hosts.rs#L267-L271
Which results in a very unreliable removing of the scripts. especially once the entities have multiple scripts and not just one.
I have updated the script_remove_synchronizer system, to loop through all context_entities and find those with a matching entity id.
The problem I see with this solution is that we have to loop through everything, as we can't break after one found match. because each entity could have multiple scripts, so we need to loop through all of them to find all.
https://github.com/makspll/bevy_mod_scripting/blob/745c20bb9e22d18761da7c50e96ee7f4766e790f/bevy_mod_scripting_core/src/systems.rs#L94-L112
I just added some changes to the context_entities.
I would love to hear some feedback from you, I personally think it makes more sense to use the entity as a key, instead of the script ID.
Not sure how to fix the test, as it comes from a crate rust_out?
running 7 tests
test src/lib.rs - (line 252) ... ignored
test src/lib.rs - (line 65) ... ignored
test src/lib.rs - (line 268) - compile ... FAILED
test src/lib.rs - (line 208) ... ok
test src/lib.rs - (line 117) ... ok
test src/lib.rs - (line 139) ... ok
test src/lib.rs - (line 179) ... ok
failures:
---- src/lib.rs - (line 268) stdout ----
error[E0601]: `main` function not found in crate `rust_out`
--> src/lib.rs:287:2
|
21 | }
| ^ consider adding a `main` function to `src/lib.rs`
error: aborting due to previous error
For more information about this error, try `rustc --explain E0601`.
Couldn't compile the test.
failures:
src/lib.rs - (line 268)
test result: FAILED. 4 passed; 1 failed; 2 ignored; 0 measured; 0 filtered out; finished in 0.71s
I see, haven't considered that!
In that case I'll quickly revert back to how it was before!
Merge Request is back at only fixing the bug in removing script context
No worries, much appreciated! Just one code style comment and I am happy!
|
gharchive/pull-request
| 2023-10-13T11:35:55 |
2025-04-01T06:39:29.196561
|
{
"authors": [
"makspll",
"zwazel"
],
"repo": "makspll/bevy_mod_scripting",
"url": "https://github.com/makspll/bevy_mod_scripting/pull/81",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
505790810
|
(要望)アケコン方式の復活について
概要
ゲームコントローラーの操作モーションについて、v0.8.7時代のアケコンっぽい動きも使いたい事があるのを確認済み。
指針
やるならやるけど、以下の設計問題に答えられないうちは着手しない。
トリガー系のボタン押したときにリアクションできる事
可能ならホントにアケコン系のコントローラが刺さっているのか検証して切り替えること
通常のゲームコントローラ入力でアケコンの動きするのはエセモーションキャプチャのコンセプトからだいぶ離れてしまうため
moved to #522
|
gharchive/issue
| 2019-10-11T11:14:11 |
2025-04-01T06:39:29.201111
|
{
"authors": [
"malaybaku"
],
"repo": "malaybaku/VMagicMirror",
"url": "https://github.com/malaybaku/VMagicMirror/issues/116",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
484868850
|
License clarification
Hi, does license "BSD" mean the BSD-3-clause (https://opensource.org/licenses/BSD-3-Clause), or one of the other BSD licenses? Thanks!
Good point. We're currently building an app and are working on proper attribution. So far, the only (indirect) dependency that we have that doesn't have an SPDX-compatible license id is readline. It would really help if your package were using one of the identifiers from the SPDX licenses list.
I second that.
We have automatic validations to ensure that all of our project dependencies have a license that matches one of those allowed by our legal department. Unfortunately, "BSD" is considered too vague to be added to that list, which means that unless it is clarified, we'll have to either switch to another similar library that has a clear license, or implement the features we need on our own.
Would it be possible to use one of the SPDX identifiers instead?
https://spdx.org/licenses/
Thanks!
Same for us. We're totally willing to live with not having a valid SPDX identifier, but knowing what BSD flavor this package is meant to be published as would be great.
|
gharchive/issue
| 2019-08-24T22:32:49 |
2025-04-01T06:39:29.204581
|
{
"authors": [
"ariel11",
"bartvanandel",
"fbeaudoincoveo",
"lnqs"
],
"repo": "maleck13/readline",
"url": "https://github.com/maleck13/readline/issues/26",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2647168236
|
Once completed, the system will not reset
if (secondsRemaining == 0)
{
_cancellationTokenSource.Cancel();
return;
}
The return shouldn't be there, because it returns before you call the ResetView() method
Thanks for pointing this out - and fixing it with PR #3 🥳
|
gharchive/issue
| 2024-11-10T11:54:41 |
2025-04-01T06:39:29.219643
|
{
"authors": [
"BartaBlzs",
"mallibone"
],
"repo": "mallibone/MauiUI2022ProgressButton",
"url": "https://github.com/mallibone/MauiUI2022ProgressButton/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
228782667
|
Scrollbars are rejected in MaterialUI Tables
When used in a material UI TableBody, it seems that MaterialUI rejects anything that may alter their table's architecture throught the DOM. Material UI forbids using anything other than TableHeader or TableBody as Table children, and any inside of TableBody or Table Header (per HTML5 standars, this is not permitted either)
render as a div.
Is there a way to assign Scrollbars to the TableBody without using the so that way the Table architecture is unchanged?
Here is the code
`<TableBody
displayRowCheckbox={this.state.showCheckboxes}
deselectOnClickaway={this.state.deselectOnClickaway}
showRowHover={this.state.showRowHover}>
<Scrollbars //not permitted. it renders as a <div>
autoHide={false}
style={ScrollBarsStyle}>
{tableData.map( (row, index) => (
<TableRow key={index}>
<TableRowColumn>{row.name}</TableRowColumn>
<TableRowColumn>{row.type}</TableRowColumn>
<TableRowColumn>{row.owner}</TableRowColumn>
</TableRow>
))}
</Scrollbars>
</TableBody>`
UPDATE
The only way it can work is by adding the entire table inside the Scrollbars tags. But it disables the fixedHeader.
This can be accomplished by using two material-ui tables. One for the header and one for the rows.
<Table>
<TableHeader>
...
</TableHeader>
</Table>
<Scrollbars>
<Table>
<TableBody>
...
</TableBody>
</Table>
</Scrollbars>
Yup! That did it! Thank you!
But what if you need the columns to be the same width?
Then give the both sets of columns a style property e.g. style={{ width: '10%' }}
Unideal solution. What if you want the header to be fixed on vertical scroll, however to scroll with the content on horizontal scroll.
How would the horizontal scrolling work in this case? @justinmasse did you get this to work?
@justinmasse just put both tables inside a horizontally scrollable div
|
gharchive/issue
| 2017-05-15T17:13:22 |
2025-04-01T06:39:29.229332
|
{
"authors": [
"TassosD",
"alexlukelevy",
"carlosafw",
"justinmasse",
"lotusms",
"sarahshuffle"
],
"repo": "malte-wessel/react-custom-scrollbars",
"url": "https://github.com/malte-wessel/react-custom-scrollbars/issues/158",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.