id
stringlengths 4
10
| text
stringlengths 4
2.14M
| source
stringclasses 2
values | created
timestamp[s]date 2001-05-16 21:05:09
2025-01-01 03:38:30
| added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| metadata
dict |
---|---|---|---|---|---|
2380402370
|
Support DataAnnotations?
I searched everywhere and I couldn't find any documentation on how to enable DataAnnotations being generated on the file. It see that it's generating XML Documentation but not generating DataAnnotations.
I was expecting like the xsd.exe to see MaxLength, RegularExpression, Description attributes on everything.
Is there a way to turn these on?
Sorry but LinqToXsd does not support those. It was originally started as a strongly-typed wrapper API around the Linq to XML API (think XDocument, XElement etc), and it's evolved that way. Adding that in is definitely possible though, but the code generator does not generate these attributes at the moment.
If System.Component.DataAnnotations are a hard requirement, I recommend using XmlSchemaClassGenerator - it generates code that's very close to what old-school xsd.exe gives you and also supports emitting DataAnnotations. LinqToXSD has its own validation mechanism that does not use DataAnnotations (it generates its own TypeValidator classes for validation).
|
gharchive/issue
| 2024-06-28T13:08:22 |
2025-04-01T06:39:29.284618
|
{
"authors": [
"jhancock-taxa",
"mamift"
],
"repo": "mamift/LinqToXsdCore",
"url": "https://github.com/mamift/LinqToXsdCore/issues/66",
"license": "MS-PL",
"license_type": "permissive",
"license_source": "github-api"
}
|
2070358291
|
fix: correct url to repo in README
Fix incorrect URL in README
Thank you very much.
|
gharchive/pull-request
| 2024-01-08T12:29:37 |
2025-04-01T06:39:29.285473
|
{
"authors": [
"mampfes",
"swoga"
],
"repo": "mampfes/ha_epex_spot",
"url": "https://github.com/mampfes/ha_epex_spot/pull/86",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
171718690
|
[update] SassからStylusへの移行をするべきか?
StylusはCSSのプリプロセッサとして後発なだけあって、Sassと同等の機能を持っていたり、それ以上の機能を持っている可能性がある。
現状のfunctionやmixinと同等の機能を維持できて、使いやすさや導入コストを考えて移行するメリットが十分にあるのか?
機能性であればStylusかもしれないけど、デファクトスタンダートなSassや、主流になっていく可能性のあるPostCSSとcssnextにしているほうがメリットが大きい。
|
gharchive/issue
| 2016-08-17T17:47:16 |
2025-04-01T06:39:29.294782
|
{
"authors": [
"manabuyasuda"
],
"repo": "manabuyasuda/equip",
"url": "https://github.com/manabuyasuda/equip/issues/81",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
266526322
|
Running the tests frequently hang. Fix tests or underlying problem with the code.
Always hangs on CircleCI and often locally. Seems worse with some of the PRs we have merged than current upstream master.
Here is one example
java.lang.InterruptedException: sleep interrupted
at java.lang.Thread.sleep(Native Method)
at org.jvnet.hudson.test.JenkinsRule.waitForMessage(JenkinsRule.java:1246)
at org.jenkins.plugins.lockableresources.LockStepTest$19.evaluate(LockStepTest.java:735)
at org.jvnet.hudson.test.RestartableJenkinsRule$3.evaluate(RestartableJenkinsRule.java:84)
at org.jvnet.hudson.test.JenkinsRule$2.evaluate(JenkinsRule.java:534)
at org.jvnet.hudson.test.RestartableJenkinsRule.run(RestartableJenkinsRule.java:100)
at org.jvnet.hudson.test.RestartableJenkinsRule.access$100(RestartableJenkinsRule.java:30)
at org.jvnet.hudson.test.RestartableJenkinsRule$1.evaluate(RestartableJenkinsRule.java:64)
at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48)
at org.junit.rules.RunRules.evaluate(RunRules.java:20)
at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
at org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:367)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:274)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:238)
at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:161)
at org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:290)
at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:242)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:121)
interoperabilityOnRestart(org.jenkins.plugins.lockableresources.LockStepTest) Time elapsed: 1.453 sec <<< FAILURE!
java.lang.AssertionError: tried to overwrite p #1 with p #1
at jenkins.model.lazy.AbstractLazyLoadRunMap.load(AbstractLazyLoadRunMap.java:472)
at jenkins.model.lazy.AbstractLazyLoadRunMap.load(AbstractLazyLoadRunMap.java:448)
at jenkins.model.lazy.AbstractLazyLoadRunMap.getByNumber(AbstractLazyLoadRunMap.java:356)
at jenkins.model.lazy.LazyBuildMixIn.getBuildByNumber(LazyBuildMixIn.java:226)
at org.jenkinsci.plugins.workflow.job.WorkflowJob.getBuildByNumber(WorkflowJob.java:215)
at org.jenkins.plugins.lockableresources.LockStepTest$14.evaluate(LockStepTest.java:518)
at org.jvnet.hudson.test.RestartableJenkinsRule$3.evaluate(RestartableJenkinsRule.java:84)
at org.jvnet.hudson.test.JenkinsRule$2.evaluate(JenkinsRule.java:534)
at org.jvnet.hudson.test.RestartableJenkinsRule.run(RestartableJenkinsRule.java:100)
at org.jvnet.hudson.test.RestartableJenkinsRule.access$100(RestartableJenkinsRule.java:30)
at org.jvnet.hudson.test.RestartableJenkinsRule$1.evaluate(RestartableJenkinsRule.java:64)
at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48)
at org.junit.rules.RunRules.evaluate(RunRules.java:20)
at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
at org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:367)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:274)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:238)
at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:161)
at org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:290)
at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:242)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:121)
Running InjectedTest
Tests run: 11, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.957 sec - in InjectedTest
Results :
Flaked tests:
org.jenkins.plugins.lockableresources.LockStepTest.deleteRunningBuildNewBuildClearsLock(org.jenkins.plugins.lockableresources.LockStepTest)
Run 1: LockStepTest.deleteRunningBuildNewBuildClearsLock » Interrupted sleep interrup...
Run 2: PASS
org.jenkins.plugins.lockableresources.LockStepTest.interoperabilityOnRestart(org.jenkins.plugins.lockableresources.LockStepTest)
Run 1: LockStepTest.interoperabilityOnRestart tried to overwrite p #1 with p #1
Run 2: PASS
|
gharchive/issue
| 2017-10-18T14:59:53 |
2025-04-01T06:39:29.301296
|
{
"authors": [
"smootoo"
],
"repo": "manahl/lockable-resources-plugin",
"url": "https://github.com/manahl/lockable-resources-plugin/issues/4",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
111947770
|
Add Hash#key method
Hash#key(value) returns the corresponding key with the given value.
Updated. Added Hash#key? method.
Updated. I followed the #fetch pattern and make #key block variant public for now; however, like you said, I'm not quite sure about this as it is not quite straightforward as a public API.
What do others think about a public Hash#key(value, &block)?
Updated.
Maybe @asterite would have an idea about this?
We are discussing if methods like Hash#fetch and Hash#key (with a block) should be public methods or not. The purpose of these methods are not for public API use but for DRYing related methods, thus their logic might not be as straightforward and may confuse developers at first sight.
Well, I never questioned it for Hash#fetch, there it very well has legit public usages.
Oops, I see.
Well, no other opinions, then let's keep it public for now. Thanks!
Thanks!
|
gharchive/pull-request
| 2015-10-17T06:42:56 |
2025-04-01T06:39:29.305850
|
{
"authors": [
"adlerhsieh",
"jhass"
],
"repo": "manastech/crystal",
"url": "https://github.com/manastech/crystal/pull/1759",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
91503986
|
Fix HTTP::Request#keep_alive? method
Add "upgrade" type of connection to fix Websocket hand shake, that expects "Connection" header to be "Upgrade".
I'm trying to run this
require "http/server"
require "http/server/handlers/websocket_handler"
handlers = [] of HTTP::Handler
handlers << HTTP::LogHandler.new
handlers << HTTP::WebSocketHandler.new do |req|
puts "> #{req}"
end
server = HTTP::Server.new(3000, handlers)
server.listen
But in Chrome it gives me the error "WebSocket connection to 'ws://pandora-102353.nitrousapp.com:3000/' failed: Error during WebSocket handshake: 'Connection' header value must contain 'Upgrade'". Do you have a working example without changing HTTP::Request#keep_alive? method?
This fix worked for me to get WebSockets upgrading properly, but had to be applied to src/http/common.cr
@hugoabonizio @adam12 Sorry for the delay! Your solutions were perfect, but we couldn't merge it because the code moved.
|
gharchive/pull-request
| 2015-06-27T19:13:58 |
2025-04-01T06:39:29.308638
|
{
"authors": [
"adam12",
"asterite",
"hugoabonizio"
],
"repo": "manastech/crystal",
"url": "https://github.com/manastech/crystal/pull/892",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
169074658
|
Implementing State Space Models and simulating the State Space
Updated tests #40 and #41
Process like Ornstein-Uhlenbeck process which can be simulated by specifying the parameters #41 of the process, theta - the mean of the process.
@amitkumarj441 We already have implementation of Gaussian Process, in that you can implement OrnsteinUlhenbeckKernel class which calculates the Ornstein Ulhenbeck covariance function for two points x and y.
Hey @mandar2812 , I'm onto implementing Ornstein Ulhenback covariance function for two points x & y, I already initiated some needed PRs for the above implementation, I'll soon initiate a PR for the same.
@amitkumarj441 : Look at this class RBFKernel to see how to implement kernels in DynaML:
RBFKernel
@mandar2812 Just initiated a PR to generate akka stream of Metropolis Hasting state #50 with breeze implementation of Markov Chain
|
gharchive/issue
| 2016-08-03T08:22:12 |
2025-04-01T06:39:29.311910
|
{
"authors": [
"amitkumarj441",
"mandar2812"
],
"repo": "mandar2812/DynaML",
"url": "https://github.com/mandar2812/DynaML/issues/42",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1727676675
|
Check valid username in installer
Upon following this script:
Open a PowerShell prompt as administrator
Download the installation script installer.ps1 to your desktop
(New-Object net.webclient).DownloadFile('https://raw.githubusercontent.com/mandiant/flare-vm/main/install.ps1',"$([Environment]::GetFolderPath("Desktop"))\install.ps1")
Unblock the installation script by running:
Unblock-File .\install.ps1
Enable script execution by running:
Set-ExecutionPolicy Unrestricted
If you receive an error saying the execution policy is overridden by a policy defined at a more specific scope, you may need to pass a scope in via Set-ExecutionPolicy Unrestricted -Scope CurrentUser to view execution policies for all scopes, type Get-ExecutionPolicy -List
Finally, execute the installer script as follow:
.\install.ps1
I get the following error See Image
The solution is to build the windows 10 VM with a one word user name. I kept making the fresh windows install with user name JOHN DOE instead of a single word user name. This resolved the issue upon package install the paths wheee the package is installed is clear of blanks.
@Ana06 FYSA, adding this check in https://github.com/mandiant/flare-vm/pull/485
See #485, thanks @HuskyHacks!
|
gharchive/issue
| 2023-05-26T13:54:01 |
2025-04-01T06:39:29.316403
|
{
"authors": [
"HuskyHacks",
"elevateman",
"mr-tz"
],
"repo": "mandiant/flare-vm",
"url": "https://github.com/mandiant/flare-vm/issues/452",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
153699606
|
Include all classes (not just components)?
This is a great start. But could it be modified to include all classes? It looks like the tool does parse non-component classes (like services, models, etc.), but doesn't appear to include them in the graph output. I think being able to see all the dependencies would also be quite useful.
@brian428 if you check this output, you can see the blue deps which are classes (services).
Is that what you mean?
Well, yes...partly. Though for some reason I have a number of services and other injected classes that aren't showing up as blue dependencies like you're showing above. Does the graph only pick up providers declared on a given class (rather than singleton (global) or inherited providers)?
Beyond that, I think it would be useful to see other types as well (models, composed/aggregated classes and so on). I realize that could make the graph a lot bigger, so maybe it could be an option. But being able to essentially see the relationships for all imported classes (or at least all imported non-framework classes) could be quite useful to identify issues with module decoupling and organization.
What do you think? Am I making sense? :-)
Yes, the tool crawls only the providers (hence the dependencies feature).
I am open to add this feature. However, I think this has to be behind flags. The developer should be able to choose what she/he wants to generate.
However, my priority for now would be to update the tool so it can handle TS 2.0
Right, that's what I inferred (about providers).
Given that, I'd say two key enhancements might be:
To have it handle global/singleton providers that are configured in the application bootstrap. (Currently, it only seems to handle providers declared on individual components, right?)
To use constructor arguments to determine a component's dependencies rather than relying only on the component's providers. What is shown right now, where only dependencies declared explicitly on the component as providers, is certainly useful. But in many cases, you're dealing with providers that are declared further up a component hierarchy (or bootstrapped at the app level), which are "lost" (to some degree) in the current dependency graph.
I'd have no problem with these options being exposed via flags though.
To be clear, I'm not disparaging what you've done...it's already very useful. As I said, I just think that being able to really see all of the dependencies, across the entire app, would also be very useful. In that light, the constructor params might be a more accurate way to determine the dependencies. Or to go even further, using the imports to truly visualize all of the dependencies.
Thanks!
Sure @brian428 I see your point ^^
Do you think you can send a PR so we can discuss more in details on the implementation?
I will see if I can take a stab at doing this, but it may take some time. Partly because I'm obviously not familiar with how you're actually doing this. And partly because I'm on the hook already for some other PRs on other projects (namely, the angular2-seed project). :-)
|
gharchive/issue
| 2016-05-09T04:02:28 |
2025-04-01T06:39:29.323441
|
{
"authors": [
"brian428",
"manekinekko"
],
"repo": "manekinekko/angular2-dependencies-graph",
"url": "https://github.com/manekinekko/angular2-dependencies-graph/issues/16",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
326647488
|
Error loading discovery document Angular 6.0.3
Using Angular 6.0.3, I am attempting to using this module but keep running into this error. Not sure how to resolve this after about 4 hours of research. Using the Tour of Heroes example from the Angular.io website
app.module.ts
import { BrowserModule } from '@angular/platform-browser';
import { FormsModule } from '@angular/forms';
import { NgModule } from '@angular/core';
import { HttpClientInMemoryWebApiModule } from 'angular-in-memory-web-api';
import { InMemoryDataService } from './services/in-memory-data.service';
import { HttpClientModule } from '@angular/common/http';
import { AppComponent } from './app.component';
import { HeroesComponent } from './heroes/heroes.component';
import { HeroDetailComponent } from './hero-detail/hero-detail.component';
import { MessagesComponent } from './messages/messages.component';
import { AppRoutingModule } from './/app-routing.module';
import { DashboardComponent } from './dashboard/dashboard.component';
import { HeroSearchComponent } from './hero-search/hero-search.component';
import { AppAuthnModule } from './/app-authn.module';
import { HomeComponent } from './home/home.component';
@NgModule({
declarations: [
AppComponent,
HeroesComponent,
HeroDetailComponent,
MessagesComponent,
DashboardComponent,
HeroSearchComponent,
HomeComponent
],
imports: [
BrowserModule,
FormsModule,
AppRoutingModule,
HttpClientModule,
// The HttpClientInMemoryWebApiModule module intercepts HTTP requests
// and returns simulated server responses.
// Remove it when a real server is ready to receive requests.
HttpClientInMemoryWebApiModule.forRoot(InMemoryDataService, {
dataEncapsulation: false
}),
AppAuthnModule
],
providers: [],
bootstrap: [AppComponent]
})
export class AppModule {}
app.component.ts
import { Component } from '@angular/core';
import { OAuthService } from 'angular-oauth2-oidc';
import { JwksValidationHandler } from 'angular-oauth2-oidc';
import { authConfig } from './app-authn.config';
@Component({
selector: 'app-root',
templateUrl: './app.component.html',
styleUrls: ['./app.component.scss']
})
export class AppComponent {
title = 'Tour of Heroes';
constructor(private oauthService: OAuthService) {
this.configureWithNewConfigApi();
}
private configureWithNewConfigApi() {
this.oauthService.configure(authConfig);
this.oauthService.tokenValidationHandler = new JwksValidationHandler();
this.oauthService.loadDiscoveryDocumentAndTryLogin();
}
}
app-authn.module.ts
import { Component } from '@angular/core';
import { OAuthService } from 'angular-oauth2-oidc';
import { JwksValidationHandler } from 'angular-oauth2-oidc';
import { authConfig } from './app-authn.config';
@Component({
selector: 'app-root',
templateUrl: './app.component.html',
styleUrls: ['./app.component.scss']
})
export class AppComponent {
title = 'Tour of Heroes';
constructor(private oauthService: OAuthService) {
this.configureWithNewConfigApi();
}
private configureWithNewConfigApi() {
this.oauthService.configure(authConfig);
this.oauthService.tokenValidationHandler = new JwksValidationHandler();
this.oauthService.loadDiscoveryDocumentAndTryLogin();
}
}
app-authn.config.ts
import { AuthConfig } from 'angular-oauth2-oidc';
export const authConfig: AuthConfig = {
// Url of the Identity Provider
issuer: 'https://steyer-identity-server.azurewebsites.net/identity',
// URL of the SPA to redirect the user to after login
redirectUri: window.location.origin + '/home',
// The SPA's id. The SPA is registered with this id at the auth-server
clientId: 'spa-demo',
// set the scope for the permissions the client should request
// The first three are defined by OIDC. The 4th is a usecase-specific one
scope: 'openid profile email voucher',
showDebugInformation: true
// sessionChecksEnabled: false
};
home.component.ts
import { Component } from '@angular/core';
import { OAuthService } from 'angular-oauth2-oidc';
@Component({
templateUrl: './home.component.html'
})
export class HomeComponent {
constructor(private oauthService: OAuthService) {}
public login() {
this.oauthService.initImplicitFlow();
}
public logoff() {
this.oauthService.logOut();
}
public get name() {
const claims = this.oauthService.getIdentityClaims();
if (!claims) {
return null;
}
return claims;
}
}
home.component.html
<h1 *ngIf="!name">
Hallo
</h1>
<h1 *ngIf="name">
Hallo, {{name}}
</h1>
<button class="btn btn-default" (click)="login()">
Login
</button>
<button class="btn btn-default" (click)="logoff()">
Logout
</button>
<div>
Username/Passwort zum Testen: max/geheim
</div>
Each time I navigate to /home the following error occurs:
error loading discovery document {body: {…}, url: "https://steyer-identity-server.azurewebsites.net/identity/.well-known/openid-configuration", headers: HttpHeaders, status: 404, statusText: "Not Found"}
I can navigate fine to the URL. I am not sure if this is a TLS issue, CORS issue or a problem with the module itself. I do not believe it is a TLS issue since the cert that is served is trusted by the browser and nodejs. I do not believe it is a CORS issues since the author of the identity server would have setup the correct Cross Origin Support. I am running on OSX 10.12.6
Any assistance is appreciated.
Your post is a bit hard to read, making it hard and a bit uncomfortable to help. Could you by any chance use the edit feature for your post, and use the markdown help and preview to format it a bit better?
Updated the formatting in the post.
@thxmike It is about HttpClientInMemoryWebApiModule. If you are using this module, some http call using that module that's why you are getting 404 Error.
If you are remove HttpClientInMemoryWebApiModule your error will gone.
Yeap, I totally missed that. My mistake. My apologies for wasting your time.
I removed it and it works.
@SVRao19056 , I am facing the same issue in Firefox. Can you please share with me how did you solve this issue ?
The error posted before you contained these details:
has no exported member 'JwksValidionHandler'. Did you mean 'JwksValidationHandler'?
So it seems there was a typo (missing a "t" there).
I am facing the below error
Hello, I have a similar issue, please let me know if my configuration data is correct.
this.oauthService.redirectUri = window.location.origin
this.oauthService.clientId = '1_22u25ry3jzqs8ok8cwwg400ogkog4880g67swgswwo4cksgs88'
this.oauthService.dummyClientSecret = '4i0vyay7278k00kkogwkkcsocksw84s89484848guisccogsok'
// this.oauthService.scope = 'code'
this.oauthService.issuer = 'https://learn.deluxus.net/oauth/v2/auth'
this.oauthService.tokenEndpoint = 'https://learn.deluxus.net/oauth/v2/token'
this.oauthService.tokenValidationHandler = new JwksValidationHandler()
// Load Discovery Document and then try to login the user
this.oauthService.loadDiscoveryDocument('https://learn.deluxus.net/oauth/v2/auth').then(() => {
this.oauthService.tryLogin()
})
I found 2 error messages:
error loading discovery document
and
ERROR Error: "Uncaught (in promise): HttpErrorResponse: {"headers":{"normalizedNames":{},"lazyUpdate":null,"headers":{}},"status":0,"statusText":"Unknown Error","url":"https://learn.deluxus.net/oauth/v2/auth","ok":false,"name":"HttpErrorResponse","message":"Http failure response for https://learn.deluxus.net/oauth/v2/auth: 0 Unknown Error","error":{"isTrusted":true}}"
|
gharchive/issue
| 2018-05-25T20:04:19 |
2025-04-01T06:39:29.334212
|
{
"authors": [
"Goppey",
"ibrahimsen",
"iflorespaz",
"jeroenheijmans",
"thxmike"
],
"repo": "manfredsteyer/angular-oauth2-oidc",
"url": "https://github.com/manfredsteyer/angular-oauth2-oidc/issues/335",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1330224827
|
[Enhancement] Pdf functions needed for creation and adding images from list of image file names
editing file app/pdf.go
using this package for pdf creation
create pdf in go
video use of this go package
just a quick example of some possible function names
func GetImageList()[]string{
var imageList []string
// do some stuff here
return imageList
}
func ImagesToPdf(document *pdf, imageList []String){
// some code to add those images to the pdf
}
you can rework this idea however you see fit with the use flags just as long as pdf output works @Yuno-obsessed
|
gharchive/issue
| 2022-08-05T18:09:41 |
2025-04-01T06:39:29.346444
|
{
"authors": [
"4cecoder"
],
"repo": "manga-g/manga-g",
"url": "https://github.com/manga-g/manga-g/issues/38",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2053216744
|
🛑 Mesa Freeworld is down
In e77ee5c, Mesa Freeworld (https://nonfree.eu) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Mesa Freeworld is back up in 7239fd9 after 59 minutes.
|
gharchive/issue
| 2023-12-22T00:42:09 |
2025-04-01T06:39:29.354318
|
{
"authors": [
"boredland"
],
"repo": "manjaro-contrib/upptime",
"url": "https://github.com/manjaro-contrib/upptime/issues/1525",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2308369294
|
🛑 Mesa Freeworld is down
In 791a5bc, Mesa Freeworld (https://nonfree.eu) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Mesa Freeworld is back up in 450e466 after 19 minutes.
|
gharchive/issue
| 2024-05-21T13:53:12 |
2025-04-01T06:39:29.356696
|
{
"authors": [
"boredland"
],
"repo": "manjaro-contrib/upptime",
"url": "https://github.com/manjaro-contrib/upptime/issues/2913",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
129667642
|
Makefile crashes when running with make -j
I think this happens since the io/utils files are specified in each of the required Makefiles, rather than being made once. Multiple threads try to create, e.g., io.o and ends up corrupting the object file. Probably use make -j4 just to be on the safe-side.
This issue is now worse. Previously make would crash for make -j8 or similar; now even make -j2 is enough to crash the code.
This will be the next issue to get fixed. Requires full rewrite of the Makefiles and using non-recursive make.
While this is very unsatisfactory, overhauling all the Makefiles will require some thinking and re-arrangment. Since the user experience is not hampered, will shelve fixing till the next version.
|
gharchive/issue
| 2016-01-29T05:03:29 |
2025-04-01T06:39:29.359508
|
{
"authors": [
"manodeep"
],
"repo": "manodeep/Corrfunc",
"url": "https://github.com/manodeep/Corrfunc/issues/14",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
270409015
|
Add checks for Numpy array endianness to the Python wrappers. Closes…
… #140 and #101.
Thanks! I had two small questions - otherwise this PR is ready to merge.
Merging in. Thanks! 👍
|
gharchive/pull-request
| 2017-11-01T18:40:37 |
2025-04-01T06:39:29.360847
|
{
"authors": [
"lgarrison",
"manodeep"
],
"repo": "manodeep/Corrfunc",
"url": "https://github.com/manodeep/Corrfunc/pull/142",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2443420073
|
Add hovered value label for Histogram plot
The ValueCounts plot shows what the user is hovering before making a selection. This is a quick implementation of a similar behavior for the histogram.
There's a bunch of stuff that could be improved: (I might give it a try before closing this PR)
[ ] Give the label a background to prevent collision with the min/max tick labels
AFAIK, since the labels are done through an SVG, you need to implement this through a rect that matches the text's bounding box
[ ] label following the cursor during selection dragging
also bit complicated due to the events captured by mosaic interactor
[ ] better value formatting
It seems like TODO 2 (mosaic capturing interactions) might be a little tricky to sort out, and I don't want to block this from merging.
I need to think about 1 some more because I'm not totally sure how I'd do that either :) Maybe we append a rect to the tick group and set the background.... we would need to resize the width depending on width of the text (DOM/div would be nice here).
For 3, you can have a look at how I formatted the ticks and we could pick something with less sigfigs but more detail than the axis bounds.
Label background should be done.
Agree that doing the dragging effect the way we'd like might need some further investigation.
|
gharchive/pull-request
| 2024-08-01T20:57:26 |
2025-04-01T06:39:29.497816
|
{
"authors": [
"dvdkouril",
"manzt"
],
"repo": "manzt/quak",
"url": "https://github.com/manzt/quak/pull/32",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
102404523
|
Error when running: "libclang.so: cannot open shared object file: No such file or directory"
First, I ran sudo apt-get install python-yaml and sudo pip install clang. Success.
Then I tried:
[dev@ubuntu:~/cncc (master)] $ ls
cncc examples LICENSE MyClass.cpp MyClass.h README.md util
[dev@ubuntu:~/cncc (master)] $ git log -1
commit 4529cb3536c7cec20ea0bb850d0f95e80cded733
Author: Daniel J. Hofmann <daniel@trvx.org>
Date: Wed Aug 19 14:55:48 2015 +0200
Respect global style file, closes #5
Local style files do not seem to make a lot sense. What we could do is
walk all parent directories like `clang-format` does, though.
[dev@ubuntu:~/cncc (master)] $ cat MyClass.h
class MyClass{
public:
void init(int a, int b);
int loopAlot();
private:
int var1, var2;
};
[dev@ubuntu:~/cncc (master)] $ cat MyClass.cpp
#include "MyClass.h"
void MyClass::init(int a, int b){
var1 = a;
var2 = b;
}
int MyClass::loopAlot(){
int res = this->var1 + var2;
for(int i=0; i<this->var1; ++i){
res = res + this->var2;
}
return res;
}
[dev@ubuntu:~/cncc (master)] $ ./cncc --style=examples/basic.style MyClass.cpp
Traceback (most recent call last):
File "./cncc", line 22, in <module>
index = I.create()
File "/usr/local/lib/python2.7/dist-packages/clang/cindex.py", line 2119, in create
return Index(conf.lib.clang_createIndex(excludeDecls, 0))
File "/usr/local/lib/python2.7/dist-packages/clang/cindex.py", line 141, in __get__
value = self.wrapped(instance)
File "/usr/local/lib/python2.7/dist-packages/clang/cindex.py", line 3429, in lib
lib = self.get_cindex_library()
File "/usr/local/lib/python2.7/dist-packages/clang/cindex.py", line 3460, in get_cindex_library
raise LibclangError(msg)
clang.cindex.LibclangError: libclang.so: cannot open shared object file: No such file or directory. To provide a path to libclang use Config.set_library_path() or Config.set_library_file().
Any idea what I'm doing wrong?
Sorry for not responding earlier, seems like I overlooked the notification.
You do not have a matching libclang native shared library for the Python wrapper that python-clang is supposed to wrap. I found PyPI to not provide wrappers compatible with every clang version. That is, check your clang --version, and then see if PyPI has a matching wrapper for this (I think the one for 3.6 was missing, and e.g. Ubuntu 15.04 comes with Clang 3.6).
That's the main reason this issue was never resolved.
The easiest way to install python-clang is by means of your package manager, e.g. aptitude on Debian-based systems.
Hi, for who have problem with python-clang, make a simbolic link for the same version of native clang and the python-bind:
cd /usr/lib/x86_64-linux-gnu/
sudo ln -s libclang-3.8.so.1 libclang.so
cheers []s
|
gharchive/issue
| 2015-08-21T15:31:00 |
2025-04-01T06:39:29.636215
|
{
"authors": [
"daniel-j-h",
"pandabr",
"stepthom"
],
"repo": "mapbox/cncc",
"url": "https://github.com/mapbox/cncc/issues/6",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
84067781
|
In Mapbox Studio, when adding new layer, can't browse to network drives on Windows
When looking for source files to add as new layers, the browse window only allows you to go to the C root, not to other drives.
Accessing other drives is in the works and should be available with the next release.
In the meantime you could try to mount that drive into a folder on C:, (Technet: Assign a mount point folder path to a drive)
Upcoming release will allow to browse to other drives.
|
gharchive/issue
| 2015-06-02T15:16:08 |
2025-04-01T06:39:29.740286
|
{
"authors": [
"BergWerkGIS",
"bkfunk"
],
"repo": "mapbox/mapbox-studio",
"url": "https://github.com/mapbox/mapbox-studio/issues/1368",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1214939059
|
Renaming the 0.11 branch to next and master to current
We're changing the names of the branches to better represent what they're supposed to refer to, as well as to accomodate some structural changes:
current (Previously master) will be used for the latest version of the last major release,
And next (Previously 0.11) will be used for the next major release in-the-works.
If you're directly referring to these branches in your project, we recommend you switch to the new names so as to prevent issues later on.
This has been done. :-)
Also, branch protection rules have been expanded to cover all branches, which are currently just current and next.
|
gharchive/issue
| 2022-04-25T19:19:36 |
2025-04-01T06:39:29.771965
|
{
"authors": [
"aleokdev",
"bjorn"
],
"repo": "mapeditor/rs-tiled",
"url": "https://github.com/mapeditor/rs-tiled/issues/216",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1004173732
|
A question about memloss
logits=torch.softmax(logits,dim=1)
batch_prob=torch.sum(logits[:,:logits.size(0)],dim=1)
batch_prob=torch.mean(batch_prob)
mem_losses.update(batch_prob.item(),logits.size(0))
why calculate mem_losses use the method above rather than cross-entropy loss?
Thank you very much.
That's is not mem loss, we just use that to track the sum of in-batch probabilities.
Thank you for your reply.But I still can't understand why we need to track the sum of in-batch probabilities rather than loss?and how the in-batch probabilities value changes means that the experimental result is correct.
That can reflect how much the memory bank negatives contribute to the loss optimization.
|
gharchive/issue
| 2021-09-22T11:19:01 |
2025-04-01T06:39:29.797021
|
{
"authors": [
"wang3702",
"zxz0928"
],
"repo": "maple-research-lab/AdCo",
"url": "https://github.com/maple-research-lab/AdCo/issues/12",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
862999958
|
iOS crash when device is locked while animation (zoom/pan) is ongoing
When panning/zooming map and locking device, iOS app with MapLibre SDK crashes occasionally. Likely same issue as this one with Mapbox iOS SDK.
Crash log:
Hardware Model: iPhone13,3
Code Type: ARM-64 (Native)
Role: Foreground
Parent Process: launchd [1]
Date/Time: 2021-04-20 15:14:10.9909 +0300
Launch Time: 2021-04-20 15:11:46.5801 +0300
OS Version: iPhone OS 14.5 (18E5154f)
Release Type: Beta
Baseband Version: 1.60.01-2
Report Version: 104
Exception Type: EXC_BAD_ACCESS (SIGSEGV)
Exception Subtype: KERN_INVALID_ADDRESS at 0x0000000000000000
Termination Signal: Segmentation fault: 11
Termination Reason: Namespace SIGNAL, Code 0xb
Terminating Process: exc handler [6125]
Triggered by Thread: 0
Thread 0 name:
Thread 0 Crashed:
0 libsystem_platform.dylib 0x00000001ded94684 _platform_memmove + 132
1 AGXMetalA14 0x00000001dff7d07c AGX::Texture<(AGXTextureMemoryLayout)3, AGX::G13::Encoders, AGX::G13::Classes>::readRegion(unsigned int, unsigned int, unsigned int, unsigned int, unsigned int, unsigned int, unsigned int, unsigned... + 632 (agxa_texture_template.hpp:2122)
2 AGXMetalA14 0x00000001dff7cdf8 -[AGXA14FamilyTexture getBytes:bytesPerRow:bytesPerImage:fromRegion:mipmapLevel:slice:] + 88 (agxa_texture_objc.mm:654)
3 AppleMetalGLRenderer 0x00000001df2c19a0 gldReadFramebufferData + 1260 (glr_framebuffer.mm:552)
4 GLEngine 0x00000001c7178620 glReadPixels_Exec + 924 (gl_read_pixels.c:329)
5 GLKit 0x00000001c5e074ec -[GLKView snapshot] + 252 (GLKView.m:728)
6 Mapbox 0x00000001038a6bd4 0x1034bc000 + 4107220
7 CoreFoundation 0x0000000192dac9e4 __CFNOTIFICATIONCENTER_IS_CALLING_OUT_TO_AN_OBSERVER__ + 28 (CFNotificationCenter.c:706)
8 CoreFoundation 0x0000000192dac98c ___CFXRegistrationPost_block_invoke + 52 (CFNotificationCenter.c:173)
9 CoreFoundation 0x0000000192dabef8 _CFXRegistrationPost + 440 (CFNotificationCenter.c:198)
10 CoreFoundation 0x0000000192dab8b8 _CFXNotificationPost + 716 (CFNotificationCenter.c:1071)
11 Foundation 0x00000001940a3a3c -[NSNotificationCenter postNotificationName:object:userInfo:] + 64 (NSNotification.m:575)
12 UIKitCore 0x000000019582b808 -[UIApplication _deactivateForReason:notify:] + 1412 (UIApplication.m:1822)
13 UIKitCore 0x0000000194e8af38 __101-[_UISceneLifecycleMultiplexer _evalTransitionToSettings:fromSettings:forceExit:withTransitionStore:]_block_invoke_2 + 988 (_UISceneLifecycleMultiplexer.m:686)
14 UIKitCore 0x00000001953f2978 _UIScenePerformActionsWithLifecycleActionMask + 104 (_UISceneLifecycleState.m:109)
15 UIKitCore 0x0000000194e8aac0 __101-[_UISceneLifecycleMultiplexer _evalTransitionToSettings:fromSettings:forceExit:withTransitionStore:]_block_invoke + 224 (_UISceneLifecycleMultiplexer.m:549)
16 UIKitCore 0x0000000194e8a580 -[_UISceneLifecycleMultiplexer _performBlock:withApplicationOfDeactivationReasons:fromReasons:] + 484 (_UISceneLifecycleMultiplexer.m:501)
17 UIKitCore 0x0000000194e8a8d0 -[_UISceneLifecycleMultiplexer _evalTransitionToSettings:fromSettings:forceExit:withTransitionStore:] + 768 (_UISceneLifecycleMultiplexer.m:548)
18 UIKitCore 0x0000000194e8a10c -[_UISceneLifecycleMultiplexer uiScene:transitionedFromState:withTransitionContext:] + 340 (_UISceneLifecycleMultiplexer.m:454)
19 UIKitCore 0x0000000194e92650 __186-[_UIWindowSceneFBSSceneTransitionContextDrivenLifecycleSettingsDiffAction _performActionsForUIScene:withUpdatedFBSScene:settingsDiff:fromSettings:transitionContext:lifecycleActionType:]_block... + 196 (_UIWindowSceneFBSSceneTransitionContextDrivenLifecycleSettingsDiffAction.m:72)
20 UIKitCore 0x00000001953006e0 +[BSAnimationSettings(UIKit) tryAnimatingWithSettings:actions:completion:] + 892 (BSAnimationSettings+UIKit.m:50)
21 UIKitCore 0x000000019540b788 _UISceneSettingsDiffActionPerformChangesWithTransitionContext + 276 (_UISceneSettingsDiffAction.m:23)
22 UIKitCore 0x0000000194e92348 -[_UIWindowSceneFBSSceneTransitionContextDrivenLifecycleSettingsDiffAction _performActionsForUIScene:withUpdatedFBSScene:settingsDiff:fromSettings:transitionContext:lifecycleActionType:] + 384 (_UIWindowSceneFBSSceneTransitionContextDrivenLifecycleSettingsDiffAction.m:57)
23 UIKitCore 0x0000000194cba1cc __64-[UIScene scene:didUpdateWithDiff:transitionContext:completion:]_block_invoke + 776 (UIScene.m:1419)
24 UIKitCore 0x0000000194cb8b4c -[UIScene _emitSceneSettingsUpdateResponseForCompletion:afterSceneUpdateWork:] + 256 (UIScene.m:1140)
25 UIKitCore 0x0000000194cb9df4 -[UIScene scene:didUpdateWithDiff:transitionContext:completion:] + 248 (UIScene.m:1386)
26 UIKitCore 0x000000019532a1f0 -[UIApplicationSceneClientAgent scene:handleEvent:withCompletion:] + 492 (UIApplicationSceneClientAgent.m:80)
27 FrontBoardServices 0x00000001a284e004 -[FBSScene updater:didUpdateSettings:withDiff:transitionContext:completion:] + 484 (FBSScene.m:547)
28 FrontBoardServices 0x00000001a287a740 __94-[FBSWorkspaceScenesClient _queue_updateScene:withSettings:diff:transitionContext:completion:]_block_invoke_2 + 152 (FBSWorkspaceScenesClient.m:533)
29 FrontBoardServices 0x00000001a285c7f0 -[FBSWorkspace _calloutQueue_executeCalloutFromSource:withBlock:] + 240 (FBSWorkspace.m:355)
30 FrontBoardServices 0x00000001a287a66c __94-[FBSWorkspaceScenesClient _queue_updateScene:withSettings:diff:transitionContext:completion:]_block_invoke + 396 (FBSWorkspaceScenesClient.m:532)
31 libdispatch.dylib 0x0000000192a3e878 _dispatch_client_callout + 20 (object.m:559)
32 libdispatch.dylib 0x0000000192a42368 _dispatch_block_invoke_direct + 268 (queue.c:468)
33 FrontBoardServices 0x00000001a28a13e4 __FBSSERIALQUEUE_IS_CALLING_OUT_TO_A_BLOCK__ + 48 (FBSSerialQueue.m:184)
34 FrontBoardServices 0x00000001a28a1074 -[FBSSerialQueue _targetQueue_performNextIfPossible] + 448 (FBSSerialQueue.m:227)
35 FrontBoardServices 0x00000001a28a15c8 -[FBSSerialQueue _performNextFromRunLoopSource] + 32 (FBSSerialQueue.m:258)
36 CoreFoundation 0x0000000192dcbc98 __CFRUNLOOP_IS_CALLING_OUT_TO_A_SOURCE0_PERFORM_FUNCTION__ + 28 (CFRunLoop.c:1967)
37 CoreFoundation 0x0000000192dcbb94 __CFRunLoopDoSource0 + 208 (CFRunLoop.c:2011)
38 CoreFoundation 0x0000000192dcaf04 __CFRunLoopDoSources0 + 376 (CFRunLoop.c:2056)
39 CoreFoundation 0x0000000192dc4e78 __CFRunLoopRun + 820 (CFRunLoop.c:2925)
40 CoreFoundation 0x0000000192dc4610 CFRunLoopRunSpecific + 600 (CFRunLoop.c:3242)
41 GraphicsServices 0x00000001aa8f2734 GSEventRunModal + 164 (GSEvent.c:2259)
42 UIKitCore 0x000000019583111c -[UIApplication _run] + 1072 (UIApplication.m:3270)
43 UIKitCore 0x000000019583698c UIApplicationMain + 168 (UIApplication.m:4741)
Steps to reproduce
Pan map
While panning, lock device
SDK crashes when unlocked (not every time though)
Expected behavior
No crashes
Actual behavior
App crashes occasionally when opening after described steps
Configuration
Maplibre iOS SDK version: ios-v5.11.0
iOS/macOS versions: iOS 14.5 Beta (18E5154f)
Device/simulator models: iPhone 12 Pro
Xcode version: 12.4
It seems this crash also happens occasionally if SDK is in progress of downloading and rendering new raster tiles when device is locked.
@juhieta If use solution from here instead of verified one it seems the crash does not happen.
|
gharchive/issue
| 2021-04-20T16:33:32 |
2025-04-01T06:39:29.822074
|
{
"authors": [
"degtiarev",
"juhieta"
],
"repo": "maplibre/maplibre-gl-native",
"url": "https://github.com/maplibre/maplibre-gl-native/issues/70",
"license": "BSD-2-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
1498585960
|
Limit work done per frame
Even thought his can be done by newcomers, this issue is HARD
We need to do some work on the main thread:
Receive data from threads or WebWorkers
upload data to GPU
Here for example we process all the data we receive in the event loop:
https://github.com/maxammann/maplibre-rs/blob/2b917e9e0850d95c876bfcf0f0f3c777f56b842d/maplibre/src/stages/populate_tile_store_stage.rs#L41-L43
We could limit this to a certain amount of messages or the message sizes.
🤔 Expected Behavior
No frames should be dropped.
😯 Current Behavior
Sometimes frames are dropped because of uploads.
💁 Possible Solution
Limit the time spend on uploading or other work. This can either be done by measuring time (low level), or by restricting it on a higher level, e.g. the amount of work items.
Steps for this issue:
Check where the most time is spent during the render loop using the Tracy profiler
Reduce it by doing less work per frame.
Closing for now as this has no priority and we should only optimize when needed.
|
gharchive/issue
| 2022-12-15T14:57:01 |
2025-04-01T06:39:29.826292
|
{
"authors": [
"maxammann"
],
"repo": "maplibre/maplibre-rs",
"url": "https://github.com/maplibre/maplibre-rs/issues/229",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
365778562
|
MapR 6.1 compare ecosystem component configuration from Installer and add changes.
MapR 6.1 compare ecosystem component configuration from Installer and add changes.
Done in branch mapr61, needs to be tested properly
|
gharchive/issue
| 2018-10-02T07:49:11 |
2025-04-01T06:39:29.852528
|
{
"authors": [
"devproof"
],
"repo": "mapr-emea/mapr-ansible",
"url": "https://github.com/mapr-emea/mapr-ansible/issues/74",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
203028815
|
Cannont skip a point
After login, I cannot skip this point: http://maproulette.org/map/1434/1284898
KO : Invalid task status supplied.
What is strange is if I logout I can skip it but if I login in the next point, I come back to this one and am blocked again.
Probably the same as #221
|
gharchive/issue
| 2017-01-25T07:11:59 |
2025-04-01T06:39:29.854354
|
{
"authors": [
"mvexel",
"rbovard"
],
"repo": "maproulette/maproulette2",
"url": "https://github.com/maproulette/maproulette2/issues/262",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
459092630
|
Map not zooming with 'animateCamera'
I'm submitting a ... (check one with "x")
[ ] question
[x] any problem or bug report
OS: (check one with "x")
[ ] Android
[ ] iOS
[x] Browser
cordova information: (run $> cordova plugin list)
cordova-plugin-device 2.0.2 "Device"
cordova-plugin-geolocation 4.0.1 "Geolocation"
cordova-plugin-googlemaps 2.6.2 "cordova-plugin-googlemaps"
cordova-plugin-ionic-keyboard 2.1.3 "cordova-plugin-ionic-keyboard"
cordova-plugin-ionic-webview 4.0.1 "cordova-plugin-ionic-webview"
cordova-plugin-splashscreen 5.0.2 "Splashscreen"
cordova-plugin-statusbar 2.4.2 "StatusBar"
cordova-plugin-whitelist 1.3.3 "Whitelist"
If you use @ionic-native/google-maps, please tell the package.json (only @ionic-native/core and @ionic-native/google-maps are fine mostly)
@ionic-native/core : "4.20.0",
@ionic-native/google-maps : "~4.20.0"
Current behavior:
Then trying animate map with increased zoom, map not zooming in. Also, this logic with 'decrease zoom' work correctly, and map zooming out. (sorry for my english :) )
Expected behavior:
Animating map with increased and decreased zoom always.
Screen capture or video record:
Related code, data or error log (please format your code or data):
Map Settings:
controls: {
'compass': false,
'myLocationButton': false,
'myLocation': true, // (blue dot)
'indoorPicker': false,
'zoom': false, // android only
'mapToolbar': false // android only
},
gestures: {
scroll: true,
tilt: false,
zoom: false,
rotate: false
},
preferences: {
zoom: {
minZoom: this.MapZoomLevelMin,
maxZoom: this.MapZoomLevelMax
}
},
Functions:
private zoomIn ( zoom? ) {
let _zoom = this.map.getCameraZoom();
console.log("HomePage/NATIVE @ zoomIn():: current =" + _zoom);
let _cameraPosition: any = {
target: this.map.getCameraPosition().target
};
if ( zoom ) {
_zoom = zoom;
} else {
_zoom++;
}
if ( _zoom >= this.MapZoomLevelMax ) {
console.log("HomePage/NATIVE @ zoomIn() -> zoom already maximum", {
zoom,
_zoom,
max: this.MapZoomLevelMax
});
_zoom = this.MapZoomLevelMax;
}
_cameraPosition.zoom = _zoom;
console.log("HomePage/NATIVE @ zoomIn():: ", {
_cameraPosition, max: this.MapDisplayPositionZoomLevelMax, current: _zoom
});
this.map.animateCamera(_cameraPosition).then(()=>{
console.log("zoomed IN");
});
}
private zoomOut ( zoom? ) {
let _zoom = this.map.getCameraZoom();
let _cameraPosition: any = {
target: this.map.getCameraPosition().target
};
if ( zoom ) {
_zoom = zoom;
} else {
_zoom--;
}
if ( _zoom <= this.MapZoomLevelMin ) {
console.log("HomePage/NATIVE @ zoomOut() -> zoom already minimum", {
zoom,
_zoom,
min: this.MapZoomLevelMin
});
_zoom = this.MapZoomLevelMin;
}
_cameraPosition.zoom = _zoom;
console.log("HomePage/NATIVE @ zoomOut():: ", {
_cameraPosition, min: this.MapZoomLevelMin, current: _zoom
});
this.map.animateCamera(_cameraPosition).then(()=>{
console.log("zoomed OUT");
});
}
Support this plugin activity
I appreicate if you give me a beer :beer: from here
On android platform work correctly.
Browser (Google Maps JavaScript v3) does not support the feature.
In the browser it is impossible to bring the map (zoom in) programmatically?)))
The animateCamera() behaves the same as moveCamera() on browser platform.
The animateCamera() work correctly (with animation) when cameraPosition 'decrease' zoom (a.k.a zoomOut)....
https://github.com/mapsplugin/cordova-plugin-googlemaps-doc/blob/master/v2.6.0/class/Map/animateCameraZoomOut/README.md
|
gharchive/issue
| 2019-06-21T09:04:29 |
2025-04-01T06:39:29.869449
|
{
"authors": [
"Smoke1987",
"wf9a5m75"
],
"repo": "mapsplugin/cordova-plugin-googlemaps",
"url": "https://github.com/mapsplugin/cordova-plugin-googlemaps/issues/2646",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
168434006
|
first pass at map labels on boxes
closes #200
different stylings at different zoom levels
@binx this is awesome! 👍 on the zoom level as well when the labels appear. Is there a way to make those labels links? So instead of the pop-up on click we can add a hover state to the text and make it link to the extract download page?
These look great!
|
gharchive/pull-request
| 2016-07-29T23:36:56 |
2025-04-01T06:39:29.901010
|
{
"authors": [
"binx",
"migurski",
"souperneon"
],
"repo": "mapzen/data-pages",
"url": "https://github.com/mapzen/data-pages/pull/204",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1578540784
|
Add real content
adding real content
hooray
|
gharchive/pull-request
| 2023-02-09T20:23:31 |
2025-04-01T06:39:29.905061
|
{
"authors": [
"pampopik"
],
"repo": "mapzen/write-the-docs-tutorial",
"url": "https://github.com/mapzen/write-the-docs-tutorial/pull/23",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2711383115
|
fix(deps): update module github.com/gardener/cert-management to v0.17.1
This PR contains the following updates:
Package
Type
Update
Change
github.com/gardener/cert-management
require
minor
v0.16.0 -> v0.17.1
Release Notes
gardener/cert-management (github.com/gardener/cert-management)
v0.17.1
Compare Source
[gardener/cert-management]
🐛 Bug Fixes
[OPERATOR] Fix panic if target issuer referenced but not allowed by @MartinWeindel [#371]
Helm Charts
cert-controller-manager: europe-docker.pkg.dev/gardener-project/releases/charts/cert-controller-manager:v0.17.1
Docker Images
cert-management: europe-docker.pkg.dev/gardener-project/releases/cert-controller-manager:v0.17.1
v0.17.0
Compare Source
[gardener/cert-management]
✨ New Features
[USER] Introduce the new Issuer type SelfSigned for creating self-signed certificates. by @RaphaelVogel [#228]
[USER] The certificate resource can now define a duration (the lifetime of the certificate). The issuer (especially Let's Encrypt) may ignore this field. by @marc1404 [#354]
🐛 Bug Fixes
[OPERATOR] Cleanup status for orphan pending certificate resources by @MartinWeindel [#367]
🏃 Others
[DEVELOPER] Use Pebble as an ACME server in the integration tests. by @marc1404 [#339]
Helm Charts
cert-controller-manager: europe-docker.pkg.dev/gardener-project/releases/charts/cert-controller-manager:v0.17.0
Docker Images
cert-management: europe-docker.pkg.dev/gardener-project/releases/cert-controller-manager:v0.17.0
Configuration
📅 Schedule: Branch creation - At any time (no schedule defined), Automerge - At any time (no schedule defined).
🚦 Automerge: Disabled by config. Please merge this manually once you are satisfied.
♻ Rebasing: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox.
🔕 Ignore: Close this PR and you won't be reminded about this update again.
[ ] If you want to rebase/retry this PR, check this box
Release note:
NONE
ℹ Artifact update notice
File name: go.mod
In order to perform the update(s) described in the table above, Renovate ran the go get command, which resulted in the following additional change(s):
8 additional dependencies were updated
Details:
Package
Change
github.com/gardener/etcd-druid
v0.24.1 -> v0.25.0
istio.io/client-go
v1.23.2 -> v1.23.3
k8s.io/kube-openapi
v0.0.0-20240808142205-8e686545bdb8 -> v0.0.0-20240903163716-9e1beecbcb38
github.com/gorilla/websocket
v1.5.0 -> v1.5.1
google.golang.org/genproto/googleapis/api
v0.0.0-20240903143218-8af14fe29dc1 -> v0.0.0-20241015192408-796eee8c2d53
google.golang.org/genproto/googleapis/rpc
v0.0.0-20240903143218-8af14fe29dc1 -> v0.0.0-20241007155032-5fefd90f89a9
google.golang.org/grpc
v1.66.2 -> v1.67.1
k8s.io/gengo/v2
v2.0.0-20240228010128-51d4e06bde70 -> v2.0.0-20240826214909-a7b603a56eb7
|
gharchive/pull-request
| 2024-12-02T10:28:08 |
2025-04-01T06:39:29.923930
|
{
"authors": [
"marc1404"
],
"repo": "marc1404/gardener",
"url": "https://github.com/marc1404/gardener/pull/7",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1192742627
|
🛑 marci4.de is down
In 44998fa, marci4.de (https://marci4.de) was down:
HTTP code: 0
Response time: 0 ms
Resolved: marci4.de is back up in 882e5e6.
|
gharchive/issue
| 2022-04-05T07:29:44 |
2025-04-01T06:39:29.951730
|
{
"authors": [
"marci4server"
],
"repo": "marci4server/uptime",
"url": "https://github.com/marci4server/uptime/issues/164",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
154678410
|
support removing specific attached events
This is a feature I needed to make ResizeSensor.js (as a standalone lib) useful for me at work; its only a small addition and seems like others would find it useful too :)
Basically, this allows you to pass the original event in to .detach methods as well as the element (fully optional) to only remove that event. from the queue.
Thanks!
|
gharchive/pull-request
| 2016-05-13T10:10:53 |
2025-04-01T06:39:29.953003
|
{
"authors": [
"jsdw",
"marcj"
],
"repo": "marcj/css-element-queries",
"url": "https://github.com/marcj/css-element-queries/pull/106",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1360362877
|
local submit directories/sandbox transfer directories and cleanup
Currently jobsub_lite creates directories under $HOME/.jobsub_lite, with submit files, and is the place where condor_transfer_data returns results; which works, but repeated use will eventually run the user out of quota
in their $HOME area.
As we move to jobsub_fetchlog, this becomes increasingly invisible to the users, so it will not occur to
them to clean this up.
We should revisit this, and pick a "standard" location, and make sure files get cleaned up eventually.
Email disucussion included this from Kevin:
After some more thought, while /run/user ($XDG_RUNTIME_DIR) would be convenient, our use would be against spec so we’d need to be careful, especially if dumping logs in there (true of any tmpfs solution). Interactive nodes give /run a few GB and generally have plenty of RAM (depending on what users are doing of course) so maybe it’d be OK…
tmpfs 5.8G 3.0M 5.8G 1% /run
$XDG_CACHE_HOME (default $HOME/.cache) or $XDG_STATE_HOME (default $HOME/.local/state) are probably more appropriate locations, but then we’d need to do cleanup. Maybe that’s not so bad - clean up as we go, and maybe also have a routine run with every jobsub command that looks for old submission dirs that didn’t get cleaned up.
https://specifications.freedesktop.org/basedir-spec/basedir-spec-latest.html#variables
Earlier comment from Kevin:
Maybe we should just delete the local sandbox after successful submission? Fetchlog will re-create the directory if it doesn’t exist, and them I suppose it should delete the directory as well when done. Or just put them in some tmpfs location (/run/user, /run/jobsub?) so we get OS cleanup occasionally.
I think this would work fine -- if we make sure to include the submit file, etc in the files we transfer to the
job, so we get it back at the end, Otherwise the submit file disappears, which could make debugging difficult.
So after discussion I think we're looking at:
making sure all the job files (i.e. the simple.cmd, etc.) get transferred to and from the job
cleaning out the sandbox directory after submission
Copying submit files to current directory, and or a subdirectory of current directory if --no-submit
cleaning out the sandbox directory after condor_transfer_data and copying elsewhere in jobsub_fetchlog
We could also have jobsub_submit and jobsub_fetchlog clean up any more-than-a-week-old submit
directories in case jobsub_submit or jobsub_fetchlog were killed or whatever before they cleaned up.
Then the user still ends up with the submit files, etc. if they do jobsbu_fetchlog.
I'm also leaning to $HOME/.cache/jobsub_lite for the root of the area for these sandbox directories.
Fixed with #115. All the submit directory files are transferrred to the schedd, and submit directory is removed.
Directory is recreated before calling condor_transfer_data in jobsub_fetchlog, and cleaned up again after
jobsub_fetchlog .
|
gharchive/issue
| 2022-09-02T15:47:59 |
2025-04-01T06:39:29.960282
|
{
"authors": [
"marcmengel"
],
"repo": "marcmengel/jobsub_lite",
"url": "https://github.com/marcmengel/jobsub_lite/issues/109",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2010443379
|
🛑 Doctor is down
In 76f93f3, Doctor (https://doctoronline.doctormas.com.pe/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Doctor is back up in ec7f27e after 4 minutes.
|
gharchive/issue
| 2023-11-25T02:41:29 |
2025-04-01T06:39:29.963074
|
{
"authors": [
"marcoadasilvaa"
],
"repo": "marcoadasilvaa/health",
"url": "https://github.com/marcoadasilvaa/health/issues/3009",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
591854344
|
Are there plans to support flutter_web?
as title
I haven't tried it. Does it not work like this? The event bus needs very little to work: Dart Streams and some generic types. See https://github.com/marcojakob/dart-event-bus/blob/master/lib/event_bus.dart
Yes, it doesn't work in flutter_web project, I have n’t even started using it, I just import it at the top of the file, and I got the following error message:
Unable to find modules for some sources, this is usually the result of either a
bad import, a missing dependency in a package (or possibly a dev_dependency
needs to move to a real dependency), or a build failure (if importing a
generated file).
Please check the following imports:
`import 'package:event_bus/event_bus.dart';` from ...
PageWorks.dart at 8:1
Failed after 193ms
How does your pubspec.yaml look like? Here is how to install it: https://pub.dev/packages/event_bus#-installing-tab-
Ok, closing this as we didn't receive any further info.
Sorry, I have forgotten to reply due to the progress of the project in the past few days.
My project uses a lot of other dependent libraries, it seems that the configuration method is the same. I do n’t think there will be any problems in installation, otherwise my "import" statement should be a problem, rather than reporting an error at runtime
However, if other users do not encounter problems, I think it may be related to my computer or environment. I will try again when I have time.
thank you for your reply!
|
gharchive/issue
| 2020-04-01T11:48:44 |
2025-04-01T06:39:29.972995
|
{
"authors": [
"disburden",
"marcojakob"
],
"repo": "marcojakob/dart-event-bus",
"url": "https://github.com/marcojakob/dart-event-bus/issues/34",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1600220709
|
Fix environement
Address 2 comments from https://github.com/hacs/default/pull/1721
done in pr ☝️
|
gharchive/issue
| 2023-02-26T22:10:51 |
2025-04-01T06:39:29.974233
|
{
"authors": [
"marcolivierarsenault"
],
"repo": "marcolivierarsenault/moonraker-home-assistant",
"url": "https://github.com/marcolivierarsenault/moonraker-home-assistant/issues/12",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
869515684
|
Facing issue for explain instance with custom classifier function
exp = explainer.explain_instance(df_val_final.Description[idx],predproba_list,num_features=5, top_labels=2)
While executing the explain instance of LimeTextExplainer, above statement keeps on executing continously with below warning message. Execution stops only if i interrupt the kernel
C:\ProgramData\Anaconda3\lib\site-packages\fastai\torch_core.py:83: UserWarning: Tensor is int32: upgrading to int64; for better performance use int64 input
warn('Tensor is int32: upgrading to int64; for better performance use int64 input')
C:\ProgramData\Anaconda3\lib\site-packages\fastai\torch_core.py:83: UserWarning: Tensor is int32: upgrading to int64; for better performance use int64 input
warn('Tensor is int32: upgrading to int64; for better performance use int64 input')
C:\ProgramData\Anaconda3\lib\site-packages\fastai\torch_core.py:83: UserWarning: Tensor is int32: upgrading to int64; for better performance use int64 input
warn('Tensor is int32: upgrading to int64; for better performance use int64 input')
I want to use my own custom classifier model and hence I wrote a classifier function - predproba_list, which returns a numpy array of predicted probabilties for the classes
Below is the function code
def predproba_list(test1) :
pred = learn_clf.predict(test1)
return np.array(pred[2])
pred[2] vaue is tensor([0.1423, 0.2133, 0.6444]) which i then convert to a numpy array
Can you please advise if the return value of the function is as expected by the explain instance's classifier function, and what could be causing the code to keep on executing without any result
Thanks in advance
Now I am getting the below error,
ValueError: Found input variables with inconsistent numbers of samples: [5000, 1].
5000 is the default value for argument num_samples in function explain_instance() if it is not explicitly defined.
How is the value for num_samples determined if need to set it explicitly
The output should be a 2d array , where columns are prediction probabilities for different labels. If you only have one label, it should still be (n, 1)
|
gharchive/issue
| 2021-04-28T05:38:47 |
2025-04-01T06:39:29.979870
|
{
"authors": [
"marcotcr",
"mayurka"
],
"repo": "marcotcr/lime",
"url": "https://github.com/marcotcr/lime/issues/597",
"license": "bsd-2-clause",
"license_type": "permissive",
"license_source": "bigquery"
}
|
827077992
|
text segfaults
Making text onto no base solid segfaults. Not sure if my fault or upstream yet. There are no tests in CadQuery that cover this behaviour.
Works now, this was probably CadQuery/cadquery#762.
|
gharchive/issue
| 2021-03-10T04:24:51 |
2025-04-01T06:39:29.985996
|
{
"authors": [
"marcus7070"
],
"repo": "marcus7070/cq-flake",
"url": "https://github.com/marcus7070/cq-flake/issues/13",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1219482497
|
Update react.md code sample var name
Since Vault is itself an API object, it is confusing to name an instance of the App object "vault"; why not name it "app"?
Hey (and sorry for the late response)!
Maybe I'm misunderstanding you here, but vault doesn't refer to the app instance here, but the vault instance.
The following code:
const { vault } = useApp();
Is equivalent to:
const app = useApp();
const value = app.vault;
The { vault } syntax extracts a property from the object on the right hand side.
|
gharchive/pull-request
| 2022-04-29T00:11:25 |
2025-04-01T06:39:29.988137
|
{
"authors": [
"codyburleson",
"marcusolsson"
],
"repo": "marcusolsson/obsidian-plugin-docs",
"url": "https://github.com/marcusolsson/obsidian-plugin-docs/pull/46",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
353226016
|
Not working properly
Hello,
I created a dataset of 20 people at two different locations to test the algorithm but unfortunately it is not working. I am taking profile view and 100 frames per person. Can you tell me why it is not working? Also at the different location, i am taking opposite side of a person.
hi @arun121cs
Did you use pre-trained models for pose or gait estimation? Or you train one of them or both from scratch?
I can send you the dataset and code ple guide me
On Fri, Aug 24, 2018, 12:19 AM Marián Margeta notifications@github.com
wrote:
hi @arun121cs https://github.com/arun121cs
Did you use pre-trained models for pose or gait estimation? Or you train
one of them or both from scratch?
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
https://github.com/marian-margeta/gait-recognition/issues/12#issuecomment-415528300,
or mute the thread
https://github.com/notifications/unsubscribe-auth/Ahr4U32v52ElYsWe0brsJJgZNog4jt3bks5uTvklgaJpZM4WI2ih
.
The code might by helpful
Ok. I will send u tomorrow morning. One more thing, I am taking profile
images(video of walking people). Is it enough? Also can you share some of
the videos for testing it more? It will be very helpful. Thanks.
On Fri, Aug 24, 2018, 12:36 AM Marián Margeta notifications@github.com
wrote:
The code might by helpful
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
https://github.com/marian-margeta/gait-recognition/issues/12#issuecomment-415533897,
or mute the thread
https://github.com/notifications/unsubscribe-auth/Ahr4U0b9fiWYDfdWEFhj4CUK0xhtp6KSks5uTv05gaJpZM4WI2ih
.
Also do you have the code to put the object in center?
On Fri, Aug 24, 2018, 12:55 AM Arun Sharma arun121cs@gmail.com wrote:
I am taking 90 images for training the svm and same for testing. But I am
not getting good accuracy.
On Fri, Aug 24, 2018, 12:52 AM Marián Margeta notifications@github.com
wrote:
Yes, video of profile view of walking persons should be enough, but they
have to be centered and in appropriate quality. I used TUM GAID dataset
that have restricted license, so you must contact the TUM to acquire it.
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
https://github.com/marian-margeta/gait-recognition/issues/12#issuecomment-415539828,
or mute the thread
https://github.com/notifications/unsubscribe-auth/Ahr4U_FDQNNwJVtxDr3pMixY2mf09rQFks5uTwDjgaJpZM4WI2ih
.
Unfortunately not. I had just scripts for that that I incrementally changed and now I am not sure if they work at this moment.
What about quantity of images? As I told you I am taking minimum of 90
images for both training and testing (identification).
On Fri, Aug 24, 2018, 1:24 AM Marián Margeta notifications@github.com
wrote:
Unfortunately not. I had just scripts for that that I incrementally
changed and now I am not sure if they work at this moment.
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
https://github.com/marian-margeta/gait-recognition/issues/12#issuecomment-415550476,
or mute the thread
https://github.com/notifications/unsubscribe-auth/Ahr4U4fKnUz3HEilhka4ImPFqobjwnJIks5uTwhwgaJpZM4WI2ih
.
Code for loading models
`import numpy as np
from scipy.misc import imresize, imread
import pandas as pd
from human_pose_nn import HumanPoseIRNetwork
from gait_nn import GaitNetwork
net_pose = HumanPoseIRNetwork()
net_gait = GaitNetwork(recurrent_unit = 'GRU', rnn_layers = 1)
Load pre-trained models
net_pose.restore('./models/Human3.6m.ckpt')
net_gait.restore('./models/H3.6m-GRU-1.ckpt')`
Code for converting frames into tensor:-
`
images = []
images_path = '/home/administrator/Desktop/Video_to_image/images/01'
for image_path in glob.glob(images_path + '/*.jpg'):
img = cv2.imread(image_path)
images.append(cv2.resize(img, (299, 299)))
images_001 = np.array(images)
images_001.shape` [This code is for single person]
Code for getting identification vectors:-
img_list = [images_001, images_004, images_005, images_006, images_007, images_008, images_009, images_010]
features_list = list() for element in img_list: spatial_features = net_pose.feed_forward_features(element) identification_vector = net_gait.feed_forward(spatial_features) #df = pd.DataFrame.from_records(identification_vector) iv =list(identification_vector) iv.pop(1) features_list = features_list + iv
This is how i am getting identification vectors and then converting into dataframe for further classification.
|
gharchive/issue
| 2018-08-23T05:04:30 |
2025-04-01T06:39:30.038731
|
{
"authors": [
"arun121cs",
"marian-margeta"
],
"repo": "marian-margeta/gait-recognition",
"url": "https://github.com/marian-margeta/gait-recognition/issues/12",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
}
|
738524692
|
Difficult times
Finally, 2.0 is released. I'm sorry for the long wait.
As a short disclaimer before you continue reading this: I'm very thankful for Marius creating this library and giving me the opportunity to maintain it. Although you might find some frustration in my statements, I don't want anyone to spread hate towards Marius. I will continue to support him and so should you.
Release 2.0
You'll notice that the 2.0 release isn't hosted on the official ChartJs.Blazor but instead on ChartJs.Blazor.Fork. The reason for that is simple, I cannot publish to ChartJs.Blazor. I pushed everything to the releases branch but the release pipeline was paused quite a while ago and I don't have permission to enable it again. The same goes for the samples thus https://www.iheartblazor.com/ will remain in an old state as long as Marius doesn't return to update it. Of course I tried to publish it normally but Marius hasn't responded to my mail so I thought I'd get it over with now.
Past year
Marius (@mariusmuntean), the owner of the project, has had very little involvement throughout the last year. Both the project maintenance and the development of version 2.0 was done by me since the end of 2019 and I really enjoyed it. However, single-handedly maintaining a library with ~80k downloads as an 18yo without real world experience was challenging at times.
I was the one who put in some horrible features and made some terrible decisions (e.g. the covariant datasets) which made the library unpleasant to use. The goal of version 2.0 was to fix all of my prior mistakes and make the library ready for easy use in most projects. I believe version 2.0 accomplishes that goal but that doesn't mean the library is now finished. We're still only supporting Chart.js 2.9, there are still missing features and bugs (e.g. support for gradients, issues with responsiveness) and we don't have any docs yet.
Future
I will continue to maintain this library by responding to issues and fixing urgent bugs. However, I won't actively develop new features.
I'm still hoping Marius will return and either continue to work on the library or hand the ownership to someone else. I don't feel comfortable searching for new maintainers in the name of the library since I'm "just" a contributor.
However, if you'd like to help develop this library and/or take over it, please tell us and we might get back to you.
When Chart.js 3.0 releases, many people will want to upgrade and so will the people using ChartJs.Blazor. I encourage you to try working with Chart.js 3.0 if you need the new features or the insane performance improvements but it will require some customization and you should probably use your own fork for it (and reference that directly). In the spirit of Open Source, I highly encourage you to contribute those changes back to the library :heart:
As a side-note, I have compiled some of my thoughts in a GitHub project. These are just some points that were important to me and I thought I'd do something public instead of just telling Marius about it.
My journey
Back when I started to work on this library (June 2019), I had very little experience. I had only worked on a handful of terrible school projects, never seen or worked with Open Source and had zero experience with Blazor. As such, most of my changes were bad but for better or worse, they all found their way into this library and now about 80k projects have to suffer from my incompetence. After some time, I got better at programming (of course I'm still nowhere) and started to understand and like Open Source. Suddenly I found myself being the maintainer of this fast growing library in the fast growing Blazor ecosystem. Although stressful (I'm sorry for all the poor issuers that received an unfriendly response), it was a really educational period.
At some point I realized just how big this library got because searching "blazor chart" on google shows this repo as first result. This might be influenced but on DuckDuckGo, it's the 4th entry. Also, ChartJs.Blazor is used in a sample project of an official dotnet repo along with multiple Microsoft employees opening issues on our repo. I thought this was absolutely insane and being the only person actively working on the repo, I didn't want to make all these people use a version that never should've been considered 1.0 in my books. So I got to work on 2.0 and now we're here.
Unfortunately, I have continuously lost interest in this library and it was just about finishing 2.0. A big factor in this is probably that I don't use Blazor myself anymore and I don't think I'll get back to it until .NET 6. Now that version 2.0 is released, I can peacefully slow down my activity here. As I said, I will continue to assist with issues because at the very least, this library has a special place in my heart.
Closing
I'd like to thank everyone who supported me and this library be it opening helpful issues, submitting pull requests or participating in discussions.
Special thanks go to Marius (@mariusmuntean) who has made this all possible in the first place. Without him, ChartJs.Blazor wouldn't exist.
I wish everyone the best
~Joel
Update 24.01.2021
Thank you all for this journey, it's been great. Now, it's time to say goodbye. The last semester of my apprenticeship is about to start and I'd like move on from ChartJs.Blazor. I have contacted Marius multiple times about the state of the library, the 2.0 release and me leaving. I've not heard back from him.
That being said, I wish ChartJs.Blazor the best and it makes me happy to see this community be so helpful and grateful. It's not a big community but we've surpassed 100k downloads recently and I'll gladly look back on this achievement in the future.
As suggested in the comments, I've also contacted Blazorise but I won't pursue that further (if you'd like to, please do).
Goodbye
~Joel
Thank you for your continuous work during the past year @Joelius300. Question: I'm curious if you have been in contact with Microsoft about the project and its future at any point?
@larschristensen20 Thank you for using the library and being so cooperative with your issues :)
Question: I'm curious if you have been in contact with Microsoft about the project and its future at any point?
No, I haven't. Marius started the project, I got into it but it never went beyond being a small project fueled by a few peoples spare time. Why do you ask?
Well as a 18yo this is a good think to put on your CV and talk about in interviews, good job!
@larschristensen20 Thank you for using the library and being so cooperative with your issues :)
Question: I'm curious if you have been in contact with Microsoft about the project and its future at any point?
No, I haven't. Marius started the project, I got into it but it never went beyond being a small project fueled by a few peoples spare time. Why do you ask?
Had a train of thought about them maybe being interested in helping maintain it, but it might very well just have been wishful thinking :-)
Also Blazorise use ChartJS, maybe you could take a look? 😉
Well as a 18yo this is a good thing to put on your CV and talk about in interviews, good job!
Thank you, I definitely will :)
Had a train of thought about them maybe being interested in helping maintain it, but it might very well just have been wishful thinking :-)
I'd also say that's closer to wishful thinking, at least I wouldn't know about similar projects and I also don't know how I would go about asking them. I could see them sponsoring such projects but even that seems highly unlikely to me given that the project still isn't huge.
Also Blazorise use ChartJS, maybe you could take a look? 😉
That's actually a really good pointer, thank you! Maybe they're interested in a collaboration or they could actually supersede our library with theirs 🤔 Both seem like good options. They do have a lot more resources available and ChartJs.Blazor would definitely fit in their system.
Also Blazorise use ChartJS, maybe you could take a look? 😉
That's actually a really good pointer, thank you! Maybe they're interested in a collaboration or they could actually supersede our library with theirs 🤔 Both seem like good options. They do have a lot more resources available and ChartJs.Blazor would definitely fit in their system.
+1 for this, I would love to see this project integrated into theirs
I'll ask them, why not :)
In the end, Marius will have to decide what's going to happen to this library but one of the great things about Open Source is that Blazorise can integrate our library into theirs (almost) however they wish.
Glad you find it helpful. BTW [here is the poorly designed website[(https://jollify.app/) for the app I'm working on.
Thank you so much for maintaining this repository, you've done an incredible job. The 2.0 version seems to fix all the bug remaining in the application I'm currently on, it's really a fantastic news.
The migration procedure from 1.1 to 2.0 is also very precise and straighforward. 👍
I'm very thankfull for your job, and althought I do not program in Blazor very often, I will keep an eye on this project, and I might contribute in the future.
@Joelius300 Thank you so much to your commitment to this project!
Thank you all for the positive comments. I posted an update on the original issue.
TL;DR Goodbye ❤️
Thank you all for the positive comments. I posted an update on the original issue.
TL;DR Goodbye ❤️
@Joelius300
I will continue to maintain this library by responding to issues and fixing urgent bugs. However, I won't actively develop new features.
I highly disagree with this. If the maintainer of this project has abandoned it then someone should probably create a fork of this and actively maintain and update that fork as the new current version. Then this project should be marked as obsolete with a disclaimer that it's no longer being maintained, and put a link to the new project. There are plenty of cases where this has happened before on github.
@Joelius300
I will continue to maintain this library by responding to issues and fixing urgent bugs. However, I won't actively develop new features.
I highly disagree with this. If the maintainer of this project has abandoned it then someone should probably create a fork of this and actively maintain and update that fork as the new current version. Then this project should be marked as obsolete with a disclaimer that it's no longer being maintained, and put a link to the new project. There are plenty of cases where this has happened before on github.
@PoisnFang I thought it worked pretty good for the three months this "mode" was in action. Multiple bugs were fixed and the alternative was leaving the repo for good (which I'm doing now) in which case I wouldn't have fixed any of these bugs..
Now I'm leaving this library and yes, if you fork it and actively maintain it feel free to post about it here in a short comment. But saying "someone" should create a fork and actively maintain that sadly doesn't do the trick.
Do you think it would be better to more clearly highlight that it's unmaintained in the readme?
@PoisnFang I thought it worked pretty good for the three months this "mode" was in action. Multiple bugs were fixed and the alternative was leaving the repo for good (which I'm doing now) in which case I wouldn't have fixed any of these bugs..
Now I'm leaving this library and yes, if you fork it and actively maintain it feel free to post about it here in a short comment. But saying "someone" should create a fork and actively maintain that sadly doesn't do the trick.
Do you think it would be better to more clearly highlight that it's unmaintained in the readme?
But saying "someone" should create a fork and actively maintain that sadly doesn't do the trick.
I agree, if I decide to maintain it in the future then I will post.
Do you think it would be better to more clearly highlight that it's unmaintained in the readme?
Yes please do this.
But saying "someone" should create a fork and actively maintain that sadly doesn't do the trick.
I agree, if I decide to maintain it in the future then I will post.
Do you think it would be better to more clearly highlight that it's unmaintained in the readme?
Yes please do this.
@Joelius300 Please put it at the TOP of the Readme in big bold letters. I do plan on taking this project on to maintain either in a fork or my or custom build for it.
@Joelius300 I have submitted a PR for some changes. I am also keep to take on maintainer role of this repo. Is there a way you can provide me with write access to the repo? Please see the issue I have raised #191
@PoisnFang Are you moving ahead with a fork? I have created a fork and have started making some bug fixes. As per your suggestion, this repo should be marked as non-maintained and point to either my fork or your fork. I can send you the changes I have submitted so far.
I am actively using this library in several commercial and personal projects. I want it to continue. My plans are to build a v3 of this repo targeting v3 of ChartJS, and to create a .NET Core 5 branch.
As everyone in the community has said so far, @Joelius300 @mariusmuntean thank you for your time and commitment to this project. You can rest now. I am able to stand on the shoulders of giants and continue forward.
@pgrimstrup No, I am not actively maintaining a fork of this any more. We should continue with your fork.
Just thought I'd let you all know of a new Blazor ChartJs implementation I encountered recently: https://github.com/erossini/BlazorChartjs Time will tell if it will be actively maintained, but to me that is more important than if the initial feature-set is less.
|
gharchive/issue
| 2020-11-08T18:23:28 |
2025-04-01T06:39:30.113022
|
{
"authors": [
"Joelius300",
"LukeTOBrien",
"PoisnFang",
"axelroy",
"kchristman54",
"larschristensen20",
"pgrimstrup"
],
"repo": "mariusmuntean/ChartJs.Blazor",
"url": "https://github.com/mariusmuntean/ChartJs.Blazor/issues/160",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
823921099
|
Moustache compatibility
Moustache doesn't work - the rendering seems to occur before moustache triggers the replacement
It works - docsify-tabs causes docsify-moustache to fail in the rendering!
|
gharchive/issue
| 2021-03-07T13:39:38 |
2025-04-01T06:39:30.128465
|
{
"authors": [
"markbattistella"
],
"repo": "markbattistella/docsify-charty",
"url": "https://github.com/markbattistella/docsify-charty/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1653076999
|
Is it possible to have the code work on GitHub pages
Thank you very much for this project. I was wondering is it possible to run the code on my own website?
Yes, please following the documentation and the demos there.
|
gharchive/issue
| 2023-04-04T02:51:49 |
2025-04-01T06:39:30.194332
|
{
"authors": [
"Royhowtohack",
"gera2ld"
],
"repo": "markmap/markmap",
"url": "https://github.com/markmap/markmap/issues/161",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2365561464
|
[BUG]
[ ] I have searched for existing issues that already reported this problem and found none
[x] The bug is present in REPL
Describe the bug
there is a typeError when i generate a new markmap, Uncaught (in promise) TypeError: Cannot read properties of undefined (reading 'refreshHook'). i try to find the reason, but This example also reports the same error [https://stackblitz.com/edit/markmap-react?file=src%2Fstyle.css]
how can i fix it , please help me, thank you
me too
me too
来信已收到!顺祝您身体健康!Thank for your email , Good Health and Happy!!
Thanks @Adonis0123
Thank you very much @Adonis0123
|
gharchive/issue
| 2024-06-21T02:15:41 |
2025-04-01T06:39:30.198019
|
{
"authors": [
"flowerhaha",
"gera2ld",
"trry-hub",
"wlk-menglan"
],
"repo": "markmap/markmap",
"url": "https://github.com/markmap/markmap/issues/256",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
120303582
|
Kickstarter csv
Hello guys, just looking to do a school paper - if possible, could someone upload the latest csv file? Noticed a few dropbox links here - though they seem to have disappeared. It would be super appreciated!
Thanks
Here you go. I'll leave this link up for a bit:
https://www.dropbox.com/s/75rxb32xdtipjbd/kickstarter projects.csv.zip?dl=0
One quick note... On Nov 10th of this year, Kickstarter changed their API and it broke part of the way I'm collecting data. So for any projects that have finished on that date or later, they are inaccurately marked in my data as "deleted" and their final backer count and pledged amount are just slightly off.
Hi Ben, that's super helpful! Do you by any chance have any csv data from just before the 10th Nov?
(Just wondering if the 10th Nov bug is possible to fix in the future?)
Yeah, there is 2.5 years worth of data in that CSV. And only projects ending after Nov 10th have the slight inaccuracy.
At some point I am hoping to make a workaround for the API change and fix the data, but I won't have time for a while, unfortunately.
Keep up the good work Ben! ☺ will keep an eye here in case there is development.
Have a good weekend,
Hey benrugg and mrGreenbean,
my name is Richard, I'm a German student and I'm looking for an current data set of kickstarter projects for my master thesis. Actually I have one which ends in September 2015.
@benrugg Here and in another thread I saw, that you provide such data sets, but the download links do not work anymore.
Would it be possible, that you upload the .cvs from above again or, if you fixed the bug, a later one?
This would help me so much!
Best regards!
Richard
Hey @ChckNrrs, sure thing... I just exported the projects (up to date) and put them here temporarily:
https://www.dropbox.com/s/75rxb32xdtipjbd/kickstarter projects.csv.zip?dl=0
(see the note above about the slight inaccuracies for projects ending Nov 10th and later)
Thank you very much @benrugg!
Hello @benrugg and @mrGreenbean ,
As many in this post, I am also working on my master thesis about kickstarter, my bottleneck is the data and I would trully appreciate if you could help me out with the csv file. I had checked the other links but they wont open.
Thanks a lot in advanced!
My best wishes for your projects,
Guillermo
Now that this is coming up so often, I'll try to put some work into an automated export of this data, so I can just post it somewhere, and so we don't have to clog up the github threads...
But for now @gmanueltp, here's the latest dump of project data: https://www.dropbox.com/s/75rxb32xdtipjbd/kickstarter projects.csv.zip?dl=0
Thanks a lot @benrugg I really appreciate it! Keep the good work!!
You're welcome :-) Let me know if you do something cool with it. I love seeing what everyone has been creating/researching with this kickstarter data.
lll do it @benrugg thanks!! Ill be working on it!
Hi @benrugg, I did this project with the help of your data: http://www.axplusb.com/search
I'll update the data soon, thanks a lot for sharing this!
Wow @samronsin, this is is really cool. I love the visualization of that data. Great project!
@benrugg thanks so much for the dataset you have uploaded. Can you provide me with the textual content of the data these campaigns provide. I am a student and would love to do some analysis.
You know what... The info in that dataset is the only content I have. I don't believe you can easily scrape the textual content from Kickstarter.
On Jul 16, 2016, at 2:36 AM, rai1234 notifications@github.com wrote:
@benrugg thanks so much for the dataset you have uploaded. Can you provide me with the textual content of the data these campaigns provide. I am a student and would love to do some analysis.
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub, or mute the thread.
Hi everyone,
Does any of you know any dataset on Kickstarter that also crawled the community data (meaning the country of origin and the number of backers) and/or the number of updates and comments?
Or a dataset with the daily updates on the number of backers and the daily amount raised by a set of projects?
@Rgui - I've got data that shows some of what you're asking for. You can get it at the last dropbox link (above) that I posted.
Hello!
Does anyone have an idea if there is a data set available which includes the comments (text) from all successful campaigns? Thanks!
@vonBlasberg - I don't have that data. If you end up finding it, let me know!
@benrugg Hello, I'm a year-two univiersity business student from Asia, recently I was doing a small class report about data analysis in Kickstarer, I've downloaded the previous dataset you've already uploaded ( covering to 08/2016), I really appreciate your csv dateset, it helps me a lot to build some statistical model, but now I really want the latest csv dataset since there are some new Asia projects coming out. If you could upload the latest csv file or automated export site , I would be very grateful.
Thanks a lot in advanced!
@dragoon3 Sure thing. I just uploaded a new version at that same link above.
@benrugg First of all, thanks a lot! I have recently conducted statistical analysis of this dataset and found many interesting pattern, I would upload my findings after finishing this class report.
I'm really interested in the process of using this code file to get data from the web and build a dataset transformed to csv, could you simply explained how this code file works? (since I'm not familiar with coding, still don't have a clue how to run it)
In addition, can this code file gernerate data including reward level, avg_price, min_price, %backed_two %backed_ten, %backed_hundred, num_comments?
Hi @dragoon3 - I'm definitely curious to see the patterns and analysis you've done with the data. Post that for sure!
The project data that I have is really just pulled directly from Kickstarter. If you wanted to do average price, min price or any other analysis, you'd have to do that in Excel or something. Also, Kickstarter doesn't make any of the detailed backer data available, so the only data is just the basic info on each project.
@benrugg Hello! Thanks for the amazing data! Do you have any up to date data?
I am working on identifying the gender of the person behind each project and for that I need the name of the person, are you pulling this data in any way?
Thank you for all your work!
Hi @marcosmercado - Unfortunately, no... I don't have the name for any project creators. I'm not sure if Kickstarter offers that information easily in their API. Sorry!
Hi @benrugg, amazing that you're keeping track of all the data! Do you happen to have data on the required returns on the projects as well?
@Piedjoo what kind of info are you thinking? The "goal" data for each project is what dollar amount they need to reach for the project to be successful. Were you asking about that or something else?
@benrugg Tnx for your quick reply! What I meant was the return the projects offer. Lets say a project wants to raise $100K in debt and offers a 6% return on it. I would like to use the data to investigate the cost of capital related to crowdfunding.. Thanks!
@Piedjoo ah, yeah, that would be great data. As far as I know, Kickstarter doesn't allow projects to offer debt or pay interest to their backers. I think a couple of the other crowdfunding sites are known for that. (Let me know if I'm wrong!)
Hi guys! @benrugg do you happen to have an updated dataset (in csv format) of the kickstarter projects? I'm writing my master thesis on crowdfunding and I desperately need data... Plus, I've tried to download and open the dropbox files you updated above but excel cracks every time I try to import them, why so?
Hey guys! @benrugg I would be reall happy if you could share the newest Kickstarter database? Also, do you think it is possible to scrape more info from Kickstarter, for example I need to research whether the projects that have video, comments and updates outperform those that do no have as well as the number of friends of founder affects the success of the campaign.And finally make the cross-region comparison
@katekol and @yauhenibankouski - I've updated the shared csv file with the latest data.
If excel is crashing, it's probably because it can't import such a large file. You should be able to find some way of importing only part of it, or using a different app to split the file into pieces.
There isn't an easy way to get other data like video, comments or friends. Sorry!
Where can I find the most updated data set of the kickstarter projects? @benrugg
I'll start doing my thesis on crowdfunding soon and I would really appreciate to have the data available for running the analysis
@marianna240296 - the most updated file is still at the dropbox link in this thread. It's sort of buried now, I guess, so I'll post it again:
https://www.dropbox.com/s/75rxb32xdtipjbd/kickstarter projects.csv.zip?dl=0
Thank you so much for sharing these data @benrugg
Il 21 feb 2017 6:18 PM, "Ben Rugg" notifications@github.com ha scritto:
@marianna240296 https://github.com/marianna240296 - the most updated
file is still at the dropbox link in this thread. It's sort of buried now,
I guess, so I'll post it again:
https://www.dropbox.com/s/75rxb32xdtipjbd/kickstarter%
20projects.csv.zip?dl=0
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
https://github.com/markolson/kickscraper/issues/35#issuecomment-281412351,
or mute the thread
https://github.com/notifications/unsubscribe-auth/AYu6H3aWAKSQWstzAywFeejJKU9w2rTzks5rexx9gaJpZM4GufM6
.
@benrugg Your data set is super helpful. I'm interested in analyzing the stats just for journalism projects (I teach journalism in the US). When I filter the file you posted for journalism as a main or sub category, I find 845 projects -- which is a decent size set but far short of the 4,262 journalism projects that Kickstarter says have been posted (https://www.kickstarter.com/help/stats ... 917 successful, 3,345 failed). Any thoughts there?
Hey @jcsouth -
I just looked into this... it turns out there is an answer, but it's an unsatisfying one. Apparently just a few months after I began collecting data, Kickstarter reworked all their categories. They expanded them to include many more sub categories than they had before (including niche sub categories like "bacon").
When they did that, they also changed the way their API returned data, and unfortunately we never updated our code to match it. It looks like the vast majority of my data incorrectly has main category where it should be sub category. (For example, showing "Childrenswear" as the main category, when it should be "Fashion" as the main and "Childrenswear" as the sub).
This is especially problematic for a few categories, and journalism is one of them. If you search Kickstarter's categories, you can see all main/sub categories. (Go to https://www.kickstarter.com/discover/categories/journalism and then click "Journalism" and scroll down and you can expand each main category one by one).
Journalism has just a few sub categories: Audio, Photo, Print, Video, Web. So in theory you could search my data for projects with each of those in the main category field (and also include the ones with "Journalism" in the main category field), and then you'd have all Journalism projects.
The problem is that "Web" is also a sub category of "Technology". My data has 3787 projects listed as "Web". Some of those are Journalism/Web and the rest are Technology/Web. At this point with the data I have, I don't know of any way to accurately separate them out.
(Also, one small additional source of discrepancy between my data and Kickstarter's is that I didn't start collecting info until June 2013.)
Depending on your purposes, hopefully you can still use the data in some form.
Hi @benrugg ,
Thanks for your hard work. I have a question. When I look at the projects between 2009 and 2013 (excluding 2013), I see only successful projects. Am I right? I just want to be sure about the data I have.
Again thanks a lot
@AccountingResearcher - Yeah, I'm not sure I ever made that clear. Once the scraping system was active (spring or summer 2013), I started collecting all projects (so any data after that is comprehensive). But then I went back and tried to find any past kickstarter projects and add them as well. The huge caveat is that it was only easy to collect successful (and notable/popular) projects from the past. So anything before that time is incomplete.
By the way @jcsouth, I'm working on updating my code to go back and gather accurate main and sub categories for all projects. I think within a day or two I'll have it all, and I'll post the new data.
Thanks, @benrugg. I've been noodling around with different ways to focus my research. The backdrop is that there's been a big shakeout in crowdsourcing platforms for journalism in recent years. A lot of start-ups (Spot.us, Beacon, Contributoria) have all folded, leaving Kickstarter as the default for journalists seeking funding. So that's why I'm interested in looking at funding trends for Kickstarter journalism projects by year since, say, 2013: projects launched, success rates, amounts sought, amounts raised ... I wonder if there was an uptick in success over the past year with all the controversy over "fake news" and "alternative facts." (Might be too soon to measure, but worth a look.) Anyway -- thanks for whatever help you can provide on the data front. I'll credit you, of course, in any academic articles I might write off the data.
@jcsouth, that's a really cool project. I'd love to see what you end up finding. I'll let you know when the data is up-to-date. (It's taking longer because Kickstarter limits the speed at which I can request info).
Ok, I've finished updating all the projects with their correct main/sub categories. @jcsouth, you should now be able to look at just the Journalism projects and the numbers should be a lot closer to what Kickstarter reports.
The csv is at the same link (above).
Thanks again, @benrugg. Sorry for the delayed reply -- grading midterms, etc. I'll wade into this this week. Cheers.
Hi @benrugg! I am currently working on my thesis and would find it super helpful if you could re-post the cvs file here again. I will mention you among the people who helped in the thesis :D
Hi @madelainegur - here's the latest version: https://www.dropbox.com/s/75rxb32xdtipjbd/kickstarter projects.csv.zip?dl=0
@benrugg @madelainegur Hi Ben the column names are kind of confusing. Do we have any information about what each column represents ?
@madelainegur I am working on a kickstarter data for my final project. I am planning to use R and use machine learning to it. How are you planning to approach it ?
Hi, @benrugg (@madelainegur and @Nishantkumark). Ben, I've been meaning to tell you something: I parsed the data that your scraper produced into a spreadsheet and hit a few snags. Apparently, some of data in certain fields have quotation marks and commas -- and so those rows blow up; they won't parse correctly. (If you open the CSV file in Excel, you'll see what I mean.) Could you use a different delimiter -- like a pipe (|) or a tilde (~)? I think that would solve the problem and produce a file that would flow seamlessly into Excel or any DB manager.
@Nishantkumark - a fair amount of the column names are actually just from the project I use the data for - www.jumpkick.me. You can ignore those. The rest should be pretty self-explanatory kickstarter-related fields.
@jcsouth - hmm, that's no good. Excel should be able to handle any kind of delimiting, because any commas or quotes in the values should be escaped. Here's a new copy delimited by a pipe. Hopefully that'll work in Excel...
https://www.dropbox.com/s/1ya53uunkq0ge4d/kickstarter projects - pipe delimited.csv.zip?dl=0
@Nishantkumark Hi there, sorry for the late response. I am studying finance, so my initial plan was to find the factors that influence the success of a campaign. I only chose a specific category and added more data manually, but since I don't really know ho to use R and currently have also no time, I will probably do just a logit regression model. (for now there is a strong correlation between many variables so I will have to think of something else) However I'd like to see your results when you're done.
@madelainegur Hey sure. Can discuss on it little further. Can you mail me at nishant_kumark@yahoo.com
@benrugg hi ben, have you also scraped the data about the project creators? e.g., information about biography, backed_projects_count , created_projects_count , social as mentioned in the WIKI?
|
gharchive/issue
| 2015-12-04T00:21:50 |
2025-04-01T06:39:30.242624
|
{
"authors": [
"AccountingResearcher",
"ChckNrrs",
"Nishantkumark",
"Piedjoo",
"Rgui",
"benrugg",
"dragoon3",
"gmanueltp",
"hannahwen",
"jcsouth",
"katekol",
"madelainegur",
"marcosmercado",
"marianna240296",
"mrGreenbean",
"rai1234",
"samronsin",
"vonBlasberg",
"yauhenibankouski"
],
"repo": "markolson/kickscraper",
"url": "https://github.com/markolson/kickscraper/issues/35",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
67368799
|
Metric Type gets doubled in list while editing
At some point there need to be some fx (layout) imports. In my opinion this should be in the controller class of the fxml classes (because there are fx imports anyway).
fixed. Added some validation
|
gharchive/issue
| 2015-04-09T14:01:33 |
2025-04-01T06:39:30.330576
|
{
"authors": [
"MrMorten"
],
"repo": "markusgumbel/moduro",
"url": "https://github.com/markusgumbel/moduro/issues/16",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1530493502
|
some things are missing like the GUID Struct or CLSID typedef from System.
So some things were not parsed, or transferred.
Apparently GUID has to be defined manually for every projection (see https://github.com/microsoft/win32metadata/issues/1546).
The correct golang version of parse GUID from string. I have checked it with the api D2D1CreateFactory.
type GUID [16]byte
func NewGuidFromString(s string) (out GUID,ok bool){
// input like "06152247-6f50-465a-9245-118bfd3b6007"
// output is binary and hex is "47221506-506f5a46-9245118b-fd3b6007"
if len(s)!=36 || s[8]!='-' || s[13]!='-' || s[18]!='-' || s[23]!='-'{
return GUID{},false
}
_,err:=hex.Decode(out[:4],[]byte(s[:8]))
if err!=nil{
return GUID{},false
}
_,err=hex.Decode(out[4:6],[]byte(s[9:13]))
if err!=nil{
return GUID{},false
}
_,err=hex.Decode(out[6:8],[]byte(s[14:18]))
if err!=nil{
return GUID{},false
}
_,err=hex.Decode(out[8:10],[]byte(s[19:23]))
if err!=nil{
return GUID{},false
}
_,err=hex.Decode(out[10:16],[]byte(s[24:36]))
if err!=nil{
return GUID{},false
}
out[0],out[1],out[2],out[3] = out[3],out[2],out[1],out[0]
out[4],out[5] = out[5],out[4]
out[6],out[7] = out[7],out[6]
return out,true
}
|
gharchive/issue
| 2023-01-12T10:45:43 |
2025-04-01T06:39:30.341823
|
{
"authors": [
"Kreijstal",
"bronze1man",
"emarj"
],
"repo": "marlersoft/win32json",
"url": "https://github.com/marlersoft/win32json/issues/5",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1113484277
|
iframe的属性
iframe的scrolling给auto属性好一些,不然没有滚动条,右侧宽度可以再少点用百分比凑一百或者vw,加个F12监听事件,搞个拖动大小的插件就更完美了
你好,感谢你提出的这个问题。这个项目暂时没有精力维护了,你可以到移步到 https://github.com/ratel-online 提此问题。
|
gharchive/issue
| 2022-01-25T06:30:07 |
2025-04-01T06:39:30.372262
|
{
"authors": [
"marmot-z",
"yanyunpiaomiao"
],
"repo": "marmot-z/js-ratel-client",
"url": "https://github.com/marmot-z/js-ratel-client/issues/8",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2700508031
|
Cannot Find Module - Monorepo
Issue: Quokka Not Recognizing Aliased Paths in Monorepo Setup
Description
I am experiencing an issue where Quokka is unable to resolve aliased paths in my TypeScript monorepo setup. Despite following various troubleshooting steps, the problem persists, and I am unable to import modules using the alias paths defined in my tsconfig.json.
Cannot find module '@/example-service/resources/schema'
Require stack:
- <rootDir>/quokka.js
Environment
Quokka Version: v1.0.671
IDE: Cursor 0.43.5
Operating System: MacOS Sonoma
Node.js Version: 18.20
Project Setup
Monorepo Tool: Yarn, nx
TypeScript Version: 4.95
This is my tsconfig.json
{
"compilerOptions": {
"baseUrl": "./src",
"paths": {
"@dataloaders/": ["../dataloaders/"],
"@/": [""],
"@test/": ["../tests/"]
},
"module": "commonjs",
"target": "es6",
"moduleResolution": "node",
"esModuleInterop": true,
"allowSyntheticDefaultImports": true,
"strict": true,
"skipLibCheck": true
}
Steps to Reproduce
Set up a monorepo using [Yarn Workspaces/Lerna].
Configure tsconfig.json with path aliases as shown above.
Attempt to run a Quokka file that imports a module using an alias path, e.g., import { myModule } from '@/myModule'.
Quokka throws an error: Cannot find module '@/myModule'.
What I've Tried
Installed tsconfig-paths: Ensured tsconfig-paths is installed in the project.
Quokka Configuration: Added the following configuration to .quokka. I've tried a million different permutations of the following. Some configs i have tried a not included:
{
"env": {
"params": {
"runner": "-r tsconfig-paths/register",
"env": "NODE_PATH=./src"
}
}
}
{
"ts": {
"compilerOptions": {
"baseUrl": "./src",
"paths": {
"@dataloaders/*": ["../dataloaders/*"],
"@/*": ["*"],
"@test/*": ["../__tests__/*"]
}
}
},
"env": {
"params": {
"runner": "-r tsconfig-paths/register"
}
}
}
{
"ts": {
"compilerOptions": {
"baseUrl": "./src",
"rootDir": "./src",
"moduleResolution": "node",
"paths": {
"@dataloaders/*": ["../dataloaders/*"],
"@/*": ["*"],
"@test/*": ["../__tests__/*"]
}
}
},
"env": {
"params": {
"env": "NODE_PATH=./src"
}
}
Verified tsconfig.json: Double-checked that baseUrl and paths are correctly set relative to the tsconfig.json file location.
Checked Quokka File Location: Ensured the Quokka file is within the scope of the tsconfig.json.
Restarted Quokka and IDE: Restarted both Quokka and my IDE after making configuration changes.
Used Absolute Paths: Attempted to set NODE_PATH to the src directory.
Checked for Quokka Pro Features: Verified that my Quokka version supports the required features.
Expected Behavior
Quokka should resolve the aliased paths as defined in tsconfig.json and allow importing modules using these paths without errors.
Actual Behavior
Quokka throws an error indicating that it cannot find the module specified by the alias path.
Request for Assistance
I would appreciate any guidance or suggestions on how to resolve this issue. If there are any additional configurations or steps I might have missed, please let me know.
Thank you for your assistance!
Note: Please let me know if you need any further information or if there are specific logs or files I should provide.
Apologies wrong quokka repo
|
gharchive/issue
| 2024-11-28T02:40:16 |
2025-04-01T06:39:30.404135
|
{
"authors": [
"AXSJ"
],
"repo": "marsupialtail/quokka",
"url": "https://github.com/marsupialtail/quokka/issues/62",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2363943495
|
App uses way too much CPU on M1 MacBook Air.
Can you share your Tuneful settings (menu bar, popover and mini player)? I am using M1 Air as well but couldn't reproduce this CPU usage, it's typically near 1% of CPU usage and 50 MB of RAM.
I couldn't reproduce it
i have the same issue on a 2017 touchbar MBP, tuneful is using over 50% CPU
I'm having the same issue on my M2 Mac mini
It is definitely caused by the scrolling song info, when I expand the song info width (so that the number of pixels is high enough for scrolling not to be needed) the CPU usage drops back to more reasonable levels (and increases when reducing it so that it has to start scrolling again).
|
gharchive/issue
| 2024-06-20T09:14:44 |
2025-04-01T06:39:30.436303
|
{
"authors": [
"Cam396",
"Zoobdude",
"ixMarcel",
"martinfekete10"
],
"repo": "martinfekete10/Tuneful",
"url": "https://github.com/martinfekete10/Tuneful/issues/92",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2650041801
|
chore(cleanup): Remove sentiment analysis since it's no longer called from anywhere
Description
The endpoint for sentiment analysis was removed some time ago, which left this as dead code. One of the config.GetInstance() calls remaining after #625 is from sentiment.go. Let's remove all that currently-dead code in a separate PR, so that we have easy access to it if we want to revive it.
Notes for Reviewers
Don't you just love deleting code?
Signed commits
[x] Yes, I signed my commits.
Force-push after rebasing on top of main after merging #625
|
gharchive/pull-request
| 2024-11-11T18:14:45 |
2025-04-01T06:39:30.494761
|
{
"authors": [
"mcamou"
],
"repo": "masa-finance/masa-oracle",
"url": "https://github.com/masa-finance/masa-oracle/pull/626",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1768775733
|
failed to parse group descriptor: EOF
This is directly related to the issue raised in https://github.com/dsoprea/go-ext4/issues/7
Essentially for ext4 filesystems for which featureInCompat64bit is false (64 bit feature is not set), getGroupDescriptor call will result in the below error.
2023-06-22T05:26:59.664+0400 WARN Partition error: filesystem error: unexpected fs error: new ext4 filesystem error: failed to get group Descriptor: failed to parse group descriptor: EOF
Unfortuanately i do not have AMI from which the volume was created but the fix is essentially what was articulated in the above raised issue.
You need to read only half of that struct if the 64-bit feature is not set and not use the rest of the fields. For the next entry you only increment offset by 32, not 64.
So the fix is literally to have 32 byte equivalent of GroupDescriptor and read to it when 64-bit feature is not set.
// GroupDescriptor32 is 32 byte
type GroupDescriptor32 struct {
BlockBitmapLo uint32 `struc:"uint32,little"`
InodeBitmapLo uint32 `struc:"uint32,little"`
InodeTableLo uint32 `struc:"uint32,little"`
FreeBlocksCountLo uint16 `struc:"uint16,little"`
FreeInodesCountLo uint16 `struc:"uint16,little"`
UsedDirsCountLo uint16 `struc:"uint16,little"`
Flags uint16 `struc:"uint16,little"`
ExcludeBitmapLo uint32 `struc:"uint32,little"`
BlockBitmapCsumLo uint16 `struc:"uint16,little"`
InodeBitmapCsumLo uint16 `struc:"uint16,little"`
ItableUnusedLo uint16 `struc:"uint16,little"`
Checksum uint16 `struc:"uint16,little"`
}
Here is the dump from my ext4 formatted volume.
$ sudo dumpe2fs -f /dev/xvdf1
dumpe2fs 1.46.5 (30-Dec-2021)
Filesystem volume name: /
Last mounted on: /
Filesystem UUID: 8cd9967e-f9c0-438f-bebd-a0a7c5886ebc
Filesystem magic number: 0xEF53
Filesystem revision #: 1 (dynamic)
Filesystem features: has_journal ext_attr resize_inode dir_index filetype needs_recovery extent flex_bg sparse_super large_file huge_file uninit_bg dir_nlink extra_isize
Filesystem flags: signed_directory_hash
Default mount options: user_xattr acl
Filesystem state: clean
Errors behavior: Continue
Filesystem OS type: Linux
Inode count: 524288
Block count: 2096635
Reserved block count: 20966
Overhead clusters: 70281
Free blocks: 1395708
Free inodes: 473043
First block: 0
Block size: 4096
Fragment size: 4096
Reserved GDT blocks: 511
Blocks per group: 32768
Fragments per group: 32768
Inodes per group: 8192
Inode blocks per group: 512
Flex block group size: 16
Filesystem created: Mon Jan 15 18:42:02 2018
Last mount time: Wed Jun 21 23:10:14 2023
Last write time: Wed Jun 21 23:10:14 2023
Mount count: 4
Maximum mount count: -1
Last checked: Mon Jan 15 18:42:02 2018
Check interval: 0 (<none>)
Lifetime writes: 455 GB
Reserved blocks uid: 0 (user root)
Reserved blocks gid: 0 (group root)
First inode: 11
Inode size: 256
Required extra isize: 28
Desired extra isize: 28
Journal inode: 8
Default directory hash: half_md4
Directory Hash Seed: 71949b1b-7e27-4206-a570-d27984b3cd37
Journal backup: inode blocks
Journal features: journal_incompat_revoke
Total journal size: 128M
Total journal blocks: 32768
Max transaction length: 32768
Fast commit length: 0
Journal sequence: 0x006fd961
Journal start: 1
@haykbaluyan
Thank you for your issue.
Could you provide how to make ext4-filesystem-32bit ?
|
gharchive/issue
| 2023-06-22T01:34:57 |
2025-04-01T06:39:30.498820
|
{
"authors": [
"haykbaluyan",
"masahiro331"
],
"repo": "masahiro331/go-ext4-filesystem",
"url": "https://github.com/masahiro331/go-ext4-filesystem/issues/8",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
884648260
|
label numbers
Hi,
Your dataset has 2 labels: 0,1. My dataset has 185 labels [0,1..,184]. When i run your code with my dataset i got errors. which part of code should i change for 185labels? Thanks.
Hi you must use one hot encoding
|
gharchive/issue
| 2021-05-10T16:59:42 |
2025-04-01T06:39:30.506384
|
{
"authors": [
"gizemsudekocarslan",
"jghasemi44"
],
"repo": "masashitsubaki/molecularGNN_smiles",
"url": "https://github.com/masashitsubaki/molecularGNN_smiles/issues/7",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1310748527
|
Task . The Job name is displayed in the step bar
进入任务列表时,步骤条处显示对应的Job名称
预期结果:
resolved
|
gharchive/issue
| 2022-07-20T09:35:35 |
2025-04-01T06:39:30.507899
|
{
"authors": [
"LazyEar0"
],
"repo": "masastack/MASA.Scheduler",
"url": "https://github.com/masastack/MASA.Scheduler/issues/60",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1698344409
|
🛑 POH events bot from render hosting is down
In b93812e, POH events bot from render hosting (https://poh-events-bot.onrender.com/ping) was down:
HTTP code: 0
Response time: 0 ms
Resolved: POH events bot from render hosting is back up in 20dc0e0.
|
gharchive/issue
| 2023-05-06T00:54:00 |
2025-04-01T06:39:30.510642
|
{
"authors": [
"masch"
],
"repo": "masch/poh-events-uptime",
"url": "https://github.com/masch/poh-events-uptime/issues/1485",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2433561647
|
🛑 POH events bot from render hosting is down
In 7770ffd, POH events bot from render hosting (https://poh-events-bot.onrender.com/ping) was down:
HTTP code: 0
Response time: 0 ms
Resolved: POH events bot from render hosting is back up in e9f78dd after 10 minutes.
|
gharchive/issue
| 2024-07-27T17:16:55 |
2025-04-01T06:39:30.513104
|
{
"authors": [
"masch"
],
"repo": "masch/poh-events-uptime",
"url": "https://github.com/masch/poh-events-uptime/issues/3363",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1128231360
|
Barcode scanner doesn't work with Galaxy S10/ Galaxy A52s in Google Chrome
Confirm by changing [ ] to [x] below to ensure that it's a bug:
[x] I've gone through the README.md
[x] I've searched for previous similar issues and didn't find any solution
I tried to use the scanner in my Samsung Galaxy 10 and I couldn't make it work. I tested it on Google Chrome, the camera does not get to focus and the image gets blurry though I tried it out with Firefox and it worked fine with the same phone.
I have tested the scanner also on Samsung galaxy A52s and I had the same results with Google Chrome.
Steps to reproduce the behaviour using the demo page.
1- Open Chrome and go to https://qr.maslick.tech/
2- Start scanner
3- Aim camera at book's barcode.
Expected behavior
Scanned barcode text.
Device: [Galaxy S10][Galaxy A52s]
OS: [Android 12]
Browser [Chrome]
Version [98]
Did you try adding Koder to Home screen (Add to Home Screen)?
Does the camera start?
Are there any errors/warnings in the console? Try open Developer Tools on your Laptop and connect to your phone
Can you provide any screenshots?
Can you post an example of the barcode you are trying to scan?
Hi,
Did you try adding Koder to Home screen (Add to Home Screen)?
yes, I tried that. the camera is blurry.
Does the camera start?
yes
Are there any errors/warnings in the console? Try open Chrome Developer Tools on your Laptop and connect to your phone
There are no errors in the console.
Can you provide any screenshots?
I guess your barcode is broken, I can see a couple white dots on some thin bars. Try scanning this code
Here's another one from a real book:
It worked with that code on Chrome. I checked with different books here and if its a normal book barcode then it does not work.
See my comment above
which phone have you used to scan?
iPhone 11 Pro / Chrome Browser
So I presume it's the camera/image quality problem you're facing...
yes, it is strange though that in Firefox the image is good and it works well but not in Chrome :o
We've worked with barcode scanning for a few years, and noticed some strange things:
Android phones have multiple logical cameras (even on some old phones with a single camera)
Chrome may pick a different camera "by default" depending on whether you're running in "standalone" or "fullscreen mode"
there's no way to reliably detect which camera is the not blurry one.
Typically we've seen a blurry camera getting chosen sporadically and our best cure so far is to try to detect the list of cameras and have an array of deviceId of the cameras we've seen working out well.
(The blurry is typically a wide-angle camera)
I have had similar issues on my Galaxy Z Flip 3, im my case i was making a VueJS component from this excellent implementation of zbar and there are a few things i have found.
You need to add support for camera selection the default camera is not always the best it in most cases its a wide angle camera and this doesn't work well for decoding.
The canvas size makes a big difference, I'm my case I made the canvas 600x300 and this gave me a good sweet spot for scanning.
My first test was on the main thread and this worked really well with the exception of slowing down other operations after moving this to a webworker on Android my scans became super slow like 3-4 sec after alot of debugging and adding the frame time to the postmessage from the main thread and passing this back after the detection i could see the processing times for the frame and detection were fast like 10 or 11ms but when checking the original frame time to the time when the message is received back on the main thread there was a delay of 3-4 sec, after alot of playing around I found that if I only send the frame to worker every 20ms then everything runs super smooth full end to end detection of 11-12ms
@maslick thanks for this great implementation with a bit of playing around its extremally fast and quite accurate
@maslick thanks will try increasing the scan rate to see if it has much effect on my devices your correct its probably a bit fast 50fps, 250ms seems slow to me though was there any particular devices that required lower frame rates.
Closing for now...
|
gharchive/issue
| 2022-02-09T08:54:02 |
2025-04-01T06:39:30.532247
|
{
"authors": [
"albi90",
"efedericomedina",
"maslick",
"sesam"
],
"repo": "maslick/koder",
"url": "https://github.com/maslick/koder/issues/64",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2459713054
|
error
[Master HamsterKombat Bot][INFO] 2024-08-11 21:39:24 [Moshe 1] Account ID: 1354016074, Account
detected as bot: False
[Master HamsterKombat Bot][INFO] 2024-08-11 21:39:24 [Moshe 1] Getting account config data...
[Master HamsterKombat Bot][ERROR] 2024-08-11 21:39:25 [Moshe 1] Status code is not 200
[Master HamsterKombat Bot][ERROR] 2024-08-11 21:39:25 [Moshe 1] Response: NOT_FOUND
[Master HamsterKombat Bot][INFO] 2024-08-11 21:39:25 [Moshe 1] Account config version: ae1ZEH61yb8xYU4OOigHqqvH1eta8xx2b3lY75wWppk
[Master HamsterKombat Bot][INFO] 2024-08-11 21:39:25 [Moshe 1] Decoding daily cipher...
[Master HamsterKombat Bot][INFO] 2024-08-11 21:39:25 [Moshe 1] Getting account data...
[Master HamsterKombat Bot][INFO] 2024-08-11 21:39:25 [Moshe 1] Account Balance Coins: 19.78m, Available Taps: 9000, Max Taps: 9000, Total Keys: 135, Balance Keys: 135
https://github.com/masterking32/MasterHamsterKombatBot/commit/1d811126e3004d094e0c1c31a6cc4d57c5ce7be2
|
gharchive/issue
| 2024-08-11T18:40:26 |
2025-04-01T06:39:30.551247
|
{
"authors": [
"masterking32",
"supmyhava"
],
"repo": "masterking32/MasterHamsterKombatBot",
"url": "https://github.com/masterking32/MasterHamsterKombatBot/issues/143",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1260867994
|
🛑 RESTREAM-IR-11-ZirSakht-2 is down
In cd3e52b, RESTREAM-IR-11-ZirSakht-2 ($IR_11) was down:
HTTP code: 0
Response time: 0 ms
Resolved: RESTREAM-IR-11-ZirSakht-2 is back up in 2ab104c.
|
gharchive/issue
| 2022-06-04T18:54:23 |
2025-04-01T06:39:30.553724
|
{
"authors": [
"masterking32"
],
"repo": "masterking32/masterstream_uptime",
"url": "https://github.com/masterking32/masterstream_uptime/issues/1462",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1260985423
|
🛑 RESTREAM-IR-10-ParsOnline-4 is down
In 2e20c39, RESTREAM-IR-10-ParsOnline-4 ($IR_10) was down:
HTTP code: 0
Response time: 0 ms
Resolved: RESTREAM-IR-10-ParsOnline-4 is back up in 5807bf4.
|
gharchive/issue
| 2022-06-05T07:47:13 |
2025-04-01T06:39:30.555980
|
{
"authors": [
"masterking32"
],
"repo": "masterking32/masterstream_uptime",
"url": "https://github.com/masterking32/masterstream_uptime/issues/1512",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1379610906
|
🛑 Iran 11 - ParsOnline 4 is down
In 820e00f, Iran 11 - ParsOnline 4 ($IR_11) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Iran 11 - ParsOnline 4 is back up in 69159de.
|
gharchive/issue
| 2022-09-20T15:28:41 |
2025-04-01T06:39:30.558169
|
{
"authors": [
"masterking32"
],
"repo": "masterking32/masterstream_uptime",
"url": "https://github.com/masterking32/masterstream_uptime/issues/8557",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1939856369
|
Update pom.xml
Update pom-scijava from 33.2.0 to 37.0.0
@tinevez may I ask you to review this?
|
gharchive/pull-request
| 2023-10-12T12:13:54 |
2025-04-01T06:39:30.561137
|
{
"authors": [
"stefanhahmann"
],
"repo": "mastodon-sc/mastodon-graph",
"url": "https://github.com/mastodon-sc/mastodon-graph/pull/15",
"license": "BSD-2-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
1720679024
|
🛑 deces.matchid.io website is down
In 851e3f1, deces.matchid.io website (https://deces.matchid.io) was down:
HTTP code: 0
Response time: 0 ms
Resolved: deces.matchid.io website is back up in 887d8aa.
|
gharchive/issue
| 2023-05-22T22:15:46 |
2025-04-01T06:39:30.624917
|
{
"authors": [
"rhanka"
],
"repo": "matchID-project/uptime",
"url": "https://github.com/matchID-project/uptime/issues/206",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1862769372
|
🛑 dev-deces.matchid.io search API is down
In 670ebe1, dev-deces.matchid.io search API (https://dev-deces.matchid.io/deces/api/v1/search?q=pompidou) was down:
HTTP code: 0
Response time: 0 ms
Resolved: dev-deces.matchid.io search API is back up in 0187ae2 after 633 days, 19 hours, 27 minutes.
|
gharchive/issue
| 2023-08-23T07:50:17 |
2025-04-01T06:39:30.628181
|
{
"authors": [
"rhanka"
],
"repo": "matchID-project/uptime",
"url": "https://github.com/matchID-project/uptime/issues/292",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
434009648
|
GitHub 404 error node-sass
while trying to npm install gulp-sass
it throws:
Cannot download "https://github.com/sass/node-sass/releases/download/v3.13.1/win32-x64-57_binding.node":
HTTP error 404 Not Found.
I tried to google the link and it does not exist, I found the latest binding node though but idk how to install it locally.
Can you try npm install with bit older version of node to see if that works?
|
gharchive/issue
| 2019-04-16T22:17:27 |
2025-04-01T06:39:30.763396
|
{
"authors": [
"Holus",
"abhiomkar"
],
"repo": "material-components/material-components-web-codelabs",
"url": "https://github.com/material-components/material-components-web-codelabs/issues/87",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1233105339
|
HCT produces unexpected RGB values
This is most noticable when the seed color is yellow:
Background color is Primary 98.
I have made a graph that shows the RGB value, with chroma set to 100:
I included CAM16 (that also seems to be broken) and HSL as a reference. These images were made using my C# library, but it shouldn't have any differences, as most of it is just the Java version copy-pasted. I also added every test from the Dart library, which are all passing.
Setting the chroma to a lower value seems to alleviate the problem a bit:
This was made with Flutter, here is the code for that:
import 'package:flutter/material.dart';
import 'package:material_color_utilities/material_color_utilities.dart';
void main() {
runApp(CustomPaint(
painter: ColorsPainter(),
));
}
class ColorsPainter extends CustomPainter {
@override
void paint(Canvas canvas, Size size) {
for (double x = 0; x < 360; x++) {
for (double y = 0; y < 100; y++) {
HctColor hct = HctColor.from(x, 100, y);
canvas.drawCircle(Offset(x, y), 1, Paint()..color = Color(hct.toInt()));
}
}
}
@override
bool shouldRepaint(ColorsPainter oldDelegate) {
return false;
}
}
Flutter code for the CAM16 graph (produces the same result as C# above):
import 'package:flutter/material.dart';
import 'package:material_color_utilities/material_color_utilities.dart';
void main() {
runApp(CustomPaint(
painter: ColorsPainter(),
));
}
class ColorsPainter extends CustomPainter {
@override
void paint(Canvas canvas, Size size) {
for (double x = 0; x < 360; x++) {
for (double y = 0; y < 100; y++) {
Cam16 cam = Cam16.fromJch(y, 100, x);
canvas.drawCircle(
Offset(x, y), 1, Paint()..color = Color(cam.viewedInSRgb));
}
}
}
@override
bool shouldRepaint(ColorsPainter oldDelegate) {
return false;
}
}
After looking around for a while I found Okhsl, which produces very similar results, so this is most likely intended.
|
gharchive/issue
| 2022-05-11T19:42:35 |
2025-04-01T06:39:30.774011
|
{
"authors": [
"albi005"
],
"repo": "material-foundation/material-color-utilities",
"url": "https://github.com/material-foundation/material-color-utilities/issues/40",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
200395361
|
Need the 'id' also in the white spaces ('_')
Hi, I need to have id also in underscores white spaces ('') because I need to work with that when the map complete renders. Do you know where I need to change @mateuszmarkowski ? I think is in:
$row.append(character != '' ?
//if the character is not an underscore (empty space)
(function (naming) {
//so users don't have to specify empty objects
settings.seats[character] = character in settings.seats ? settings.seats[character] : {};
var id = overrideId ? overrideId : naming.getId(character, naming.rows[row], naming.columns[column]);
seats[id] = new seat({
id: id,
label: overrideLabel ?
overrideLabel : naming.getLabel(character, naming.rows[row], naming.columns[column]),
row: row,
column: column,
character: character
});
seatIds.push(id);
return seats[id].node();
})(settings.naming) :
//this is just an empty space (_)
$('<div></div>').addClass('seatCharts-cell seatCharts-space') //// Here???
);
Thanks!
Yes, I think you found the right place, sorry for a late response.
I did it thanks @mateuszmarkowski
|
gharchive/issue
| 2017-01-12T15:23:44 |
2025-04-01T06:39:30.789168
|
{
"authors": [
"marcosicp",
"mateuszmarkowski"
],
"repo": "mateuszmarkowski/jQuery-Seat-Charts",
"url": "https://github.com/mateuszmarkowski/jQuery-Seat-Charts/issues/48",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
152191843
|
What is the ~/bin/bash symlink used for?
I'm just sieving through this project to make my own dotfiles, and i can't figure out what the bash symlink in ~/bin is used for.
It's getting added to PATH in .bash_profile, but what does that enable? Is there a reason why it needs to be added to the path?
After installing bash v4 with homebrew, it still needs to be added to /etc/shell manually, so i'm not sure what value this symlink is adding. Everything seems to work just fine without it??
Really appreciate it!
If I run bash or bash --help or bash -c $cmd I want to use the brew-installed version of bash, not the old Bash v3 that OS X ships with.
I thought as much, but bash --version reports 4.3.42(1)-release even without the bash symlink in ~/bin, which made me question whether it's even needed?
echo $BASH_VERSION will always report the old Bash v3 that ships with OS X, regardless of the symlink being present or not.
It's only until you add /usr/local/bin/bash to /etc/shell and run chsh -s /usr/local/bin/bash that the $BASH_VERSION changes to 4.3.42(1)-release.
I automated that whole bit like so in my brew.sh:
if ! grep '/usr/local/bin/bash' /etc/shells; then
echo '/usr/local/bin/bash' | sudo tee -a /etc/shells;
chsh -s /usr/local/bin/bash;
fi;
@jeffbyrnes Use fgrep -q or grep -F -q, so output does not occur on the if statement. sudo echo '/usr/local/bin/bash' >> /etc/shells, not sure why you would use tee for that.
Ah, nice! @jeffbyrnes you thought about opening a PR for this?
Ah, nice! @jeffbyrnes you thought about opening a PR for this? Also, does that mean that ~/bin/bash isn't needed anymore?
@KingScooty I haven‘t, but I certainly can. Gonna be on the train in about an hour, so I’ll try to remember to open one once I settle in.
@Tatsh
sudo echo '/usr/local/bin/bash' >> /etc/shells — not sure why you would use tee for that.
Because the sudo echo … version doesn’t work. You’re running echo as super user but it’s the >> that needs it.
|
gharchive/issue
| 2016-05-01T10:03:18 |
2025-04-01T06:39:30.810658
|
{
"authors": [
"KingScooty",
"Tatsh",
"jeffbyrnes",
"mathiasbynens"
],
"repo": "mathiasbynens/dotfiles",
"url": "https://github.com/mathiasbynens/dotfiles/issues/656",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
119003858
|
Support Unicode 9 prepended concatenation marks
It’s still a proposal for now, but hey. http://blog.unicode.org/2015/11/new-character-property-for-prepended.html Prepended_Concatenation_Mark
Just curious... it looks like Prepended_Concatenation_Marks were added in Unicode 8 (ref)? I do see mention of these marks in Unicode 9, but I can't tell if any new ones were added.
And on that note... the package.json for this library is still using Unicode 8, shouldn't it be updated for v9?
http://unicode.org/reports/tr29/#Table_Combining_Char_Sequences_and_Grapheme_Clusters
|
gharchive/issue
| 2015-11-26T09:26:13 |
2025-04-01T06:39:30.813917
|
{
"authors": [
"Mottie",
"mathiasbynens"
],
"repo": "mathiasbynens/esrever",
"url": "https://github.com/mathiasbynens/esrever/issues/8",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1000078533
|
Mode tutorial of mode
Mode tutorial of mode with example questions and math contest questions.
http://ec2-34-200-244-39.compute-1.amazonaws.com/boards/5/topics/28/
Better to add a math contest example
|
gharchive/issue
| 2021-09-18T16:39:31 |
2025-04-01T06:39:30.821014
|
{
"authors": [
"mathincubator"
],
"repo": "mathincubator/mathlib",
"url": "https://github.com/mathincubator/mathlib/issues/15",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2554112798
|
\mathcal uses incorrect font in Client side MathJax modes
Issue Summary
As described in https://phabricator.wikimedia.org/T375932 LaTeX input like $\mathcal A$ uses different fonts than it used with MathJax-node
With a preprocessing script the LaTeX input is converted into MathML of this form
<mi>𝒜</mi>
To support MathJax we also added the classes
<mi data-mjx-variant="-tex-calligraphic" mathvariant="script">𝒜</mi>
which might have been contra productive?
Which seems to be recommended. The MathJax rendering looks more like $\mathscr A$
Steps to Reproduce:
Login to https://en.wikipedia.beta.wmflabs.org/
Select Client Side Math rendering in the preferences https://en.wikipedia.beta.wmflabs.org/wiki/Special:Preferences#mw-prefsection-rendering-math
Goto https://en.wikipedia.beta.wmflabs.org/wiki/T375932
See that it works with FF but fails with Chrome.
Technical details:
MathJax Version: 3.2.2
Client OS: Mac OS X 10.8.4
Browser: Firefox 130.0.1
I am using the following MathJax configuration:
https://github.com/wikimedia/mediawiki-extensions-Math/blob/db81c960623d346ca8f43fb91ca3da92f868ea29/modules/ext.math.mathjax.js#L7
( function () {
'use strict';
const extensionAssetsPath = mw.config.get( 'wgExtensionAssetsPath' );
window.MathJax = {
loader: {
// see https://docs.mathjax.org/en/latest/input/mathml.html
load: [ '[mml]/mml3' ],
// see https://docs.mathjax.org/en/latest/options/startup/loader.html
paths: {
mathjax: extensionAssetsPath + '/Math/modules/mathjax/es5'
}
}
};
}() );
and loading MathJax via
<script src="/w/extensions/Math/modules/mathjax/es5/input/mml.js" charset="UTF-8"></script>
<script src="/w/extensions/Math/modules/mathjax/es5/input/mml/extensions/mml3.js" charset="UTF-8"></script>
Supporting information:
Please supply a link to a (live) minimal example page, when possible:
https://en.wikipedia.beta.wmflabs.org/wiki/T375932
If your issue is with the display of the mathematics produced by MathJax, include a screen snapshot that illustrates the problem, when possible.
See also https://phabricator.wikimedia.org/T375932 for more screenshots
Check your browser console window for any error messages, and include them here.
None
Include the MathJax configuration you are using, and the script tag that loads MathJax itself.
load: [ '[mml]/mml3' ],
The underlying problem is that the Unicode Math Alphanumerics block includes only one script variant, and the choice of the variant to use is up to the fonts. Most seem to use a Roundhand style (as in \mathscr), so the Chancery form (\mathcal) is absent. (The STIX2 Math fonts have the Chancery version.)
Since the MathML mathvariant has only one script value, that makes it difficult to implement both \mathscr and \mathcal. MathJax uses data-mjx-variant="-tex-calligraphic" to overcome this difficulty, but since MathML renderers don't know about this attribute, MathJax adds mathvariant="script" to at least get a script variant (whichever one the font implements).
Note, however, that MathJax uses <mi data-mjx-variant="-tex-calligraphic" mathvariant="script">A</mi> not <mi data-mjx-variant="-tex-calligraphic" mathvariant="script">𝒜</mi>. This is because the MathML specification indicates that the Math Alphanumeric values are not to be influenced by CSS styles or mathvariant settings (see end of the third paragraph of the description of mathvariant in the specification):
Note that the appearance of a mathematical alphanumeric symbol character should not be altered by surrounding mathvariant or other style declarations.
That is, if you do
<mi mathvariant="bold">𝔄</mi>
you get the non-bold Fraktur capital A U+1D504, not the bold one at U+1D56C. I suspect the main reason for that is when the mathvariant is on an mstyle element, lik
<mstyle mathvariant="bold">
<mi>x<mi>
<mo>+</mo>
<mi> 𝔄</mi>
</mstyle>
The U+1D504 will not bin inadvertently changed to bold, while the other characters will.
In any case, since you are specifying the character U+1D49C explicitly (not just A), the mathvariant and data-mjx-variant won't change the result. So you get the font's U+1D49C character, which for the MathJax fonts is the Roundhand version, not the Chancery one.
Note that MathML-Core has removed the mathvariant attribute from its specification, except for the mi element, where only mathvariant="normal" is allowed (to avoid the automatic italicization that is done for single-character content). See the mi section of the specification. Unfortunately, that means you can't use <mi data-mjx-variant="-tex-calligraphic" mathvariant="script">A</mi> to get this to work in both MathML-Core and MathJax.
The MathML-Core documentation notes that the issue exists (in the second note of this section of the specification). It suggests two different potential solutions, both being font dependent. The first is to use variant selectors U+FE00 and U+FE01 to specify Chancery and Roundhand respectively. I don't know whether browsers actually implement this or not; testing in Safari, Chrome, and Firefox on MacOS suggests they do not. Also, for <mi>𝒜.</mi>, I get the Chancery version (from STIX2) in Firefox and Chrome, but the Roundhand version in Safari. As I said, it is font dependent. In any case, MathJax doesn't support the variant selectors, either.
The second suggestion is to use font features and the font-variant-alternatives CSS property, but again, that is font dependent, and MathJax doesn't support that, either. I haven't checked what the browser support for that is.
So there seems to be no direct way to reliably get both \mathscr and \mathcal to work in MathML-Core. A cross-browser solution that works with both MathML-Core and MathJax seems out of reach at the moment.
In any case, MathJax is producing the correct output for the MathML you are using.
Thank you for enlightening me! In general, this makes a lot of sense. However, this brings one backdoor to my mind.
So there seems to be no direct way to reliably get both \mathscr and \mathcal to work in MathML-Core. A cross-browser solution that works with both MathML-Core and MathJax seems out of reach at the moment.
We don't have \mathscr as an option. It's not safelisted in the texvc grammar. Is there a way (maybe via an input filter) to make MathJax display \mathcal like \mathsrc?
Is there a way (maybe via an input filter) to make MathJax display \mathcal like \mathsrc?
Well, you could replace the Roundhand script characters by the Chancery ones in MathJax's font definitions. That can be done in the MathJax startup ready() function as follows:
window.MathJax = {
remapChars(v1, v2, base, map, font) {
const c1 = v1.chars;
const c2 = v2.chars;
for (let i = 0; i < 26; i++) {
const data1 = c1[map[i] || base + i] || [];
const data2 = c2[0x41 + i];
if (data1.length === 0) c1[base + i] = data1;
[0, 1, 2].forEach(j => data1[j] = data2[j]);
data1[3] = {...data2[3], f: font, c: String.fromCharCode(0x41 + i)};
}
},
startup: {
ready() {
MathJax.startup.defaultReady();
const variant = MathJax.startup.document.outputJax.font.variant;
const map = {1: 0x212C, 4: 0x2130, 5: 0x2131, 7: 0x210B, 8: 0x2110, 11: 0x2112, 12: 0x2133, 17: 0x211B};
MathJax.config.remapChars(variant.normal, variant['-tex-calligraphic'], 0x1D49C, map, 'C');
MathJax.config.remapChars(variant.normal, variant['-tex-bold-calligraphic'], 0x1D4D0, {}, 'CB');
}
}
};
This replaces the original script (Roundhand) data with the calligraphic (Chancery) character definitions, so that the Unicode Math Alphanumeric script characters will be displayed as Chancery characters. That means the Roundhand variant is no longer available. I think that is what you were asking for.
This exactly matches what I understand was desired. Thank you so much.
|
gharchive/issue
| 2024-09-28T08:40:06 |
2025-04-01T06:39:30.840423
|
{
"authors": [
"dpvc",
"physikerwelt"
],
"repo": "mathjax/MathJax",
"url": "https://github.com/mathjax/MathJax/issues/3292",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
74374779
|
Firefox Writes Global Timeline "Backwards" On Load
Firefox, and only Firefox, seems to disregard the .before() and .after() functions in jQuery, publishing posts from the timeline at the top of the column regardless of their publication order. A better way of filling the timeline is needed here.
This has been completely resolved. Yay!
|
gharchive/issue
| 2015-05-08T14:07:54 |
2025-04-01T06:39:30.862686
|
{
"authors": [
"matigo"
],
"repo": "matigo/nice_social",
"url": "https://github.com/matigo/nice_social/issues/17",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
254772177
|
Compatibility with MO
Not necessarily an issue with zEdit, but when running zEdit through MO v2.0.8.2, I am unable to save the esp due to failure to create directory.
failure to create directory.
Can you expand? There should be no directory created when plugins are saved by zEdit. It just saves plugins into the data directory. MO should detect the saved plugins and move them to overwrite.
http://i.imgur.com/CMDu7Am.png the issue happens even when I try outside of MO as well. I'm unable to save any changes.
Sounds like an issue with file system permissions. The only directory that gets created when files are saved is the directory for backups.
Resolved with Megaloblast on Discord. Fix will be included in the next build. If you're experienced this issue and need a a build before then, grab this (this build only includes a fix for saving plugins, it's otherwise identical to v0.0.1).
|
gharchive/issue
| 2017-09-01T23:29:40 |
2025-04-01T06:39:30.918775
|
{
"authors": [
"matortheeternal",
"megaloblast"
],
"repo": "matortheeternal/zedit",
"url": "https://github.com/matortheeternal/zedit/issues/6",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1525300786
|
Fix panics due to closed channel in some tests
A number of tests start goroutines that write into channels when completed, however the tests would close the channel if it exited early (e.g. due to a timeout).
Instead, make the goroutine that writes into the channel the one that closes it.
This does mean that the goroutines will continue past the end of the test (?), but they did that before too. I am wondering if there is a better way of doing these things...
I suppose the right thing to do is to have the goroutine check a "signal channel" after it completes long-running operations (such as /sync) to make sure that it should continue execution before writing to channels.
I suppose that might make sense for the one goroutine that is multiple statements, but I don't see the benefit for the others (like the one in the example)?
(merging now to unblock fixing flakey tests)
|
gharchive/pull-request
| 2023-01-09T10:23:18 |
2025-04-01T06:39:30.972689
|
{
"authors": [
"erikjohnston"
],
"repo": "matrix-org/complement",
"url": "https://github.com/matrix-org/complement/pull/579",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1124869539
|
Allow talks with no Q&A to disable it
so the bot doesn't shout
@kittykat For WEDF, do any of the talks have Q&A? Currently all JSON-schedule talks will have Q&A disabled; I'd have to add in support for Q&A if it's desired.
@reivilibre Sander confirmed no Q&A rooms
see https://github.com/matrix-org/conference-bot/pull/135
for this year: Q&A availability will be determined by Room ID — D.* and M.test will have Q&A
|
gharchive/issue
| 2022-02-05T08:22:09 |
2025-04-01T06:39:30.975158
|
{
"authors": [
"kittykat",
"reivilibre",
"squahtx"
],
"repo": "matrix-org/conference-bot",
"url": "https://github.com/matrix-org/conference-bot/issues/119",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1908810655
|
ERROR mas_listener::server: error shutting down connection: Socket not connected (os error 107)
I've configured the MAS exactly following the documentation, found the "metrics" issue here and changed "none" to "prometheus", so MAS starts. But always shortly after the start, I'm getting repeated error messages:
2023-09-22T12:18:06.012090Z INFO cli.run.init: mas_cli::commands::server: Starting task worker worker_name="7lvKv8JSRD"
2023-09-22T12:18:06.076241Z INFO cli.run.init: mas_cli::commands::server: Listening on http://[::]:8082 with resources [Discovery, Human, OAuth, Compat, GraphQL { playground: true }, Assets { path: "./share/assets/" }]
2023-09-22T12:18:06.079178Z INFO cli.run.init: mas_cli::commands::server: Listening on http://[::]:8083 with resources [Health]
2023-09-22T12:18:15.781569Z ERROR mas_listener::server: error shutting down connection: Socket not connected (os error 107)
2023-09-22T12:18:25.785975Z ERROR mas_listener::server: error shutting down connection: Socket not connected (os error 107)
2023-09-22T12:18:35.787174Z ERROR mas_listener::server: error shutting down connection: Socket not connected (os error 107)
2023-09-22T12:18:45.788918Z ERROR mas_listener::server: error shutting down connection: Socket not connected (os error 107)
The ports 8082 and 8083 (8080 and 8081 are already used by bridges) are free, and after starting MAS I see that 8082 and 8083 are happily used as listeners from MAS.
So I have no clue what these error messages want to tell me.
Also, when opening Element-Web I'm always getting:
Es gab ein Problem bei der Kommunikation mit dem Heim-Server. Bitte versuche es später erneut.(M_UNRECOGNIZED)
But I guess as long as the errors above are showing up, MAS is not working properly.
Either there is an issue, or something is still undocumented ...
This should probably be clearer in the documentation, but in the section about setting up a reverse proxy in front of it, it mentions that a few Matrix C-S API should be handled by MAS
See the last paragraph here: https://matrix-org.github.io/matrix-authentication-service/setup/reverse-proxy.html#example-nginx-configuration
The errors spitting out are I think unrelated, but something I need to investigate
@sandhose I have a proxy in front, my HAProxy. But no relation to this error, if MAS does not run there's nothing to proxy ;-)
My point is, in your setup, /_matrix/client/v3/login (and a few other endpoints) should not be handled by Synapse, they should be handled by MAS.
This is a requirement for "legacy" (non-OIDC-native clients including Element Web) to work with MAS
@sandhose Indeed the passing of the endpoints to MAS was missing, MAS is now called and the registration / authentication flow looked good up to a point where I grant the permissions for the Element-Web client. Then ... #1864
I think I captured most of the issues you had in other issues?
#1940
#1941
And I think the logs you had were mostly resolved by #1814 and #1775
|
gharchive/issue
| 2023-09-22T12:24:04 |
2025-04-01T06:39:30.983310
|
{
"authors": [
"jacotec",
"sandhose"
],
"repo": "matrix-org/matrix-authentication-service",
"url": "https://github.com/matrix-org/matrix-authentication-service/issues/1835",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
666934671
|
Docker image does not contain register-guest helper
I think it only copies the main matrix-static binary into the image.
I'm not up to date on how such helpers are normally handled in dockerland, would it be loaded into the same image even though the only way to run it would be to mangle the entrypoint at runtime?
|
gharchive/issue
| 2020-07-28T09:46:17 |
2025-04-01T06:39:30.996581
|
{
"authors": [
"michaelkaye",
"t3chguy"
],
"repo": "matrix-org/matrix-static",
"url": "https://github.com/matrix-org/matrix-static/issues/97",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1550568804
|
Undefined function
profile_load_custom_fields in src/Plugin/Infrastructure/MoodleFunctionBasedMatrixUserIdLoader.php is not found. Probably it needs to be escaped and the /user/profile/lib.php needs to be required.
Hi @andremenrath, sorry for not getting to this earlier. Would you want to make a PR for this?
|
gharchive/issue
| 2023-01-20T09:46:39 |
2025-04-01T06:39:30.997758
|
{
"authors": [
"andremenrath",
"axelsimon"
],
"repo": "matrix-org/moodle-mod_matrix",
"url": "https://github.com/matrix-org/moodle-mod_matrix/issues/15",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
137208059
|
Add templateImage parameter
Added a templateImage parameter as discussed in Slack.
This allows Mac OS X to change the styling in the StatusBar to fit the different themes better.
👍
We should add a tutorial plugin to show how to use images too.
Alright, I try to think of one and make a pull request
Pull request: https://github.com/matryer/bitbar-plugins/pull/290
|
gharchive/pull-request
| 2016-02-29T10:08:36 |
2025-04-01T06:39:31.052120
|
{
"authors": [
"ibutra",
"iosdeveloper"
],
"repo": "matryer/bitbar",
"url": "https://github.com/matryer/bitbar/pull/261",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
734980637
|
Autoclose closes "never-close" tags [Issue when loading configs]
Certain tags, like <input />, do not have a closing tag, but autoclose-html closes them anyway (despite them being configured as never-close tags!), in the improper format:
Instead of closing like
<input />
or even just leaving the tag be, it closes the tag like
<input></input>
Which browsers can generally autocorrect, but is probably not best practice...
And yes, <input> is configured as a never-close tag.
Determined this to be a config loading issue, with atom.config.observe. Gonna try to fix.
Seems to be an Atom bug? Submitted Issue #21642 on Atom.
|
gharchive/issue
| 2020-11-03T03:09:23 |
2025-04-01T06:39:31.063037
|
{
"authors": [
"Vbbab"
],
"repo": "mattberkowitz/autoclose-html",
"url": "https://github.com/mattberkowitz/autoclose-html/issues/219",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1077362726
|
Upgrade of this tool
Full upgrade
Closed ZKF-1316
Closes ZKF-1316
|
gharchive/pull-request
| 2021-12-11T02:28:54 |
2025-04-01T06:39:31.088587
|
{
"authors": [
"keriat"
],
"repo": "matter-labs/zksync-tool-locker",
"url": "https://github.com/matter-labs/zksync-tool-locker/pull/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
444133666
|
[MM-15389] Prevent installing 32bit on 64bit and vise versa
This PR configures the Desktop app MSI installers to prevent installation on non-matching-bit systems, i.e. the 32-bit installer won’t install on 64-bit systems.
This issue has been automatically labelled "stale" because it hasn't had recent activity.
A core team member will check in on the status of the PR to help with questions.
Thank you for your contribution!
/cc @jasonblais @hanzei
Closing in favour of a single branch based off of release-test-msi for all MSI related updates to make further updates more manageable. New PR pending.
|
gharchive/pull-request
| 2019-05-14T21:28:37 |
2025-04-01T06:39:31.090600
|
{
"authors": [
"deanwhillier",
"mattermod"
],
"repo": "mattermost/desktop",
"url": "https://github.com/mattermost/desktop/pull/976",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1002505235
|
Added sphinx-tabs extension
This PR adds a maintained public Sphinx extension called sphinx-tabs to enable the ability to develop product documentation content in a tabbed format. We've locked this extension down to the current 3.2.0 version.
When this extension is merged into master, I will close the initial POC of this functionality implemented via https://github.com/mattermost/docs/pull/5042
|
gharchive/pull-request
| 2021-09-21T13:38:39 |
2025-04-01T06:39:31.092485
|
{
"authors": [
"cwarnermm"
],
"repo": "mattermost/docs",
"url": "https://github.com/mattermost/docs/pull/5060",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
827996024
|
MM-31341 Use first metadata info for inline md images
Summary
When an inline image url contains some query strings, the metadata for the image doesn't match the link thus the image was not showing, with this PR if the link does not match the one in the metadata, we will then use the first image available in the metadata to be displayed.
PR also takes care of black images on top of a black background.
Ticket Link
https://mattermost.atlassian.net/browse/MM-31341
https://mattermost.atlassian.net/browse/MM-25811
Building app in separate branch.
Successfully building:
https://circleci.com/gh/mattermost/mattermost-mobile/14347
https://circleci.com/gh/mattermost/mattermost-mobile/14346
Artifact links:
https://14347-70265724-gh.circle-artifacts.com/0/tmp/artifacts/Mattermost_Beta.apk
https://14346-70265724-gh.circle-artifacts.com/0/tmp/artifacts/Mattermost_Beta.ipa
Hi @enahum The image is now being displayed but compared to webapp, the background and sizing look different. Not sure if this is expected.
Mobile
Webapp
@josephbaylon that is actually expected, we have always needed to render the inline images into its own line
|
gharchive/pull-request
| 2021-03-10T16:24:18 |
2025-04-01T06:39:31.098250
|
{
"authors": [
"enahum",
"josephbaylon",
"mattermod"
],
"repo": "mattermost/mattermost-mobile",
"url": "https://github.com/mattermost/mattermost-mobile/pull/5219",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1576251932
|
Android fix
Summary
Fixes android 13 notification permission request and unsigned builds
NONE
/cherry-pick release-2.0
Cherry pick is scheduled.
|
gharchive/pull-request
| 2023-02-08T14:45:56 |
2025-04-01T06:39:31.099641
|
{
"authors": [
"enahum",
"mattermost-build"
],
"repo": "mattermost/mattermost-mobile",
"url": "https://github.com/mattermost/mattermost-mobile/pull/7099",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1631958629
|
MM-49862 - remove no longer needed webapp calls to action and pulsating dots
Summary
This PR removes two no longer needed calls to action: Message Actions pulsating dot and New badge on the forward button. Also adds a condition for the custom status pulsating dot and update status post button where these two are just going to be shown once the user creation time has passed 7 days.
Ticket Link
https://mattermost.atlassian.net/browse/MM-49862
Related Pull Requests
n/a
Screenshots
Before:
After:
The only change for this one is that won't show during the first seven days after user creation
Release Note
NONE
E2E tests not automatically triggered, because PR has no approval yet. Please ask a developer to review and then try again to attach the QA label.
Migrated to https://github.com/mattermost/mattermost-server/pull/22699
|
gharchive/pull-request
| 2023-03-20T11:57:54 |
2025-04-01T06:39:31.162935
|
{
"authors": [
"mattermost-build",
"pvev"
],
"repo": "mattermost/mattermost-webapp",
"url": "https://github.com/mattermost/mattermost-webapp/pull/12374",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
389385812
|
MM-13342: Fix dividers showing when menu section is empty.
Summary
Fix dividers showing when menu section is empty.
This fix is a little ugly - it checks the children-of-children are not
null to work around connected components themselves not being null but
having the actual component within being null.
Ticket Link
https://mattermost.atlassian.net/browse/MM-13342
Checklist
[x] Ran make check-style to check for style errors (required for all pull requests)
[x] Ran make test to ensure unit and component tests passed
[x] Added or updated unit tests (required for all new features)
If you're wanting to avoid adding that extra introspection, perhaps we could actually fix this using CSS? If you look here, we already use .divider + .divider to hide the dividers for empty sections other than the first, so we can add .divider:firstChild to hide that first one as well.
I don't think that will work, as we're not successfully hiding any of the dividers currently with the CSS. This fix targets all the unnecessary ones, not just the first one. Then again my CSS-foo isn't great so maybe I'm misunderstanding it?
Sigh. Looks like this was a duplicate ticket anyway and Asaad has fixed it here https://github.com/mattermost/mattermost-webapp/pull/2152
I've merged and cherry picked that PR, so we could test again now to see if it's still an issue
Looks fixed to me now with Asaad's changes. Closing this PR.
|
gharchive/pull-request
| 2018-12-10T16:38:30 |
2025-04-01T06:39:31.168247
|
{
"authors": [
"grundleborg",
"hmhealey"
],
"repo": "mattermost/mattermost-webapp",
"url": "https://github.com/mattermost/mattermost-webapp/pull/2166",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
448016710
|
M14126 Message Draft Pencil Icon - No text, only file attachment
Summary
Added e2e test to validate message draft with attachment after switching channel.
Ticket Link
Fixes https://github.com/mattermost/mattermost-server/issues/10274
@saturninoabril Could you please review
@composednitin Sure, will do that, thanks
@thekiiingbob could you please review again
@saturninoabril could you please review it again
@composednitin And just sync with master please. :)
|
gharchive/pull-request
| 2019-05-24T07:21:29 |
2025-04-01T06:39:31.170869
|
{
"authors": [
"composednitin",
"saturninoabril",
"thekiiingbob"
],
"repo": "mattermost/mattermost-webapp",
"url": "https://github.com/mattermost/mattermost-webapp/pull/2845",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
476955622
|
[MM-17200] Add render methods for new guest messages
Ticket Link
https://mattermost.atlassian.net/browse/MM-17200
Related Pull Requests
Has server changes
Has redux changes
The ticket only refers to the system post that appears when a guest is added to a channel by another user, but these PRs add a new message to the system post that appears when a guest joins a channel (for example, through an invitation).
@wiersgallak can you please confirm this second change?
Checked that the test failure is caused by mattermost-redux not pointing to the right commit, should be fixed when the redux PR is merged.
Thanks @wiersgallak! Updated
Re-assigned QA review to Rohitesh, as he and Ogi are QA point people on guest accounts.
(Sorry; closed with a bad tab/Enter combo and re-opened. :) )
|
gharchive/pull-request
| 2019-08-05T16:39:50 |
2025-04-01T06:39:31.174585
|
{
"authors": [
"lindalumitchell",
"mgdelacroix"
],
"repo": "mattermost/mattermost-webapp",
"url": "https://github.com/mattermost/mattermost-webapp/pull/3366",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
823111439
|
MM32642 - increase invite others visibility
Summary
MM32642 - increase invite others visibility - These PR adds enhancements to the channels intros to improve the user experience and add more visibility to the call to action for inviting new members to workspaces.
Ticket Link
https://mattermost.atlassian.net/browse/MM-32642
Related Pull Requests
MM-Redux: https://github.com/mattermost/mattermost-redux/pull/1395
Screenshots
@pablovelezvidal @ethervoid I believe this was not included in the cloud branch either. Could you confirm ? If so, this needs to be cherry picked to be included in the March 24th release.
/cherry-pick cloud
Cherry pick is scheduled.
|
gharchive/pull-request
| 2021-03-05T13:51:35 |
2025-04-01T06:39:31.177992
|
{
"authors": [
"ethervoid",
"marianunez",
"mattermod",
"pablovelezvidal"
],
"repo": "mattermost/mattermost-webapp",
"url": "https://github.com/mattermost/mattermost-webapp/pull/7645",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
950993482
|
fix: Subscription title in system console billing is singular
Summary
Fixes a typo. Subscription title should be singular in the System Console billing page for cloud accounts.
Ticket Link
https://mattermost.atlassian.net/browse/MM-30538
Screenshots
Title now reads "Subscription"
Release Note
Fixed System Console billing subscription title to be singular
Hello @neallred,
Thanks for your pull request! A Core Committer will review your pull request soon. For code contributions, you can learn more about the review process here.
|
gharchive/pull-request
| 2021-07-22T19:45:58 |
2025-04-01T06:39:31.180860
|
{
"authors": [
"mattermod",
"neallred"
],
"repo": "mattermost/mattermost-webapp",
"url": "https://github.com/mattermost/mattermost-webapp/pull/8451",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1127452252
|
hide accordion items with status 'OK'
Summary
title says it all. Items that have status 'ok' or 'none' get hidden from the AccordionCard. They still get counted in the overall score.
Ticket Link
n/a
Related Pull Requests
n/a
Screenshots
n/a
Release Note
NONE
@BenCookie95 gentle ping on this one.
|
gharchive/pull-request
| 2022-02-08T16:03:50 |
2025-04-01T06:39:31.182953
|
{
"authors": [
"michelengelen"
],
"repo": "mattermost/mattermost-webapp",
"url": "https://github.com/mattermost/mattermost-webapp/pull/9777",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
290562749
|
Is there any way to get access to the error that's swallowed in readFromSecureStore?
https://github.com/matthewpalmer/Locksmith/blob/49b08bf8fa16ff8d7ce41609a2adbc6b5237f65f/Source/Locksmith.swift#L495
https://github.com/matthewpalmer/Locksmith/blob/49b08bf8fa16ff8d7ce41609a2adbc6b5237f65f/Source/Locksmith.swift#L498
https://github.com/matthewpalmer/Locksmith/blob/49b08bf8fa16ff8d7ce41609a2adbc6b5237f65f/Source/Locksmith.swift#L484
https://github.com/matthewpalmer/Locksmith/blob/49b08bf8fa16ff8d7ce41609a2adbc6b5237f65f/Source/Locksmith.swift#L481
|
gharchive/issue
| 2018-01-22T18:10:31 |
2025-04-01T06:39:31.258264
|
{
"authors": [
"benron",
"slizeray"
],
"repo": "matthewpalmer/Locksmith",
"url": "https://github.com/matthewpalmer/Locksmith/issues/186",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1398278270
|
Cannot extract meter value - Kamstrup Multical 401
I hope you vcan help me. I have a Kamstrup Multical 401 and trying to use your script. I have tried to place de IR head at every possible way at the meter but the following warning keeps appearing. In MQTT i see the topic but no values.
[2022-10-05 21:27:18 daemon.py init: 41 - INFO - initializing daemon]
[2022-10-05 21:27:18 mqtt_handler.py connect: 43 - INFO - Connected to MQTT at: 192.168.2.98:1883]
[2022-10-05 21:27:18 mqtt_handler.py connect: 45 - INFO - with username pi, QoS level = 1 and retain = False]
[2022-10-05 21:27:18 kamstrup_meter.py close: 126 - DEBUG - Closed serial port]
[2022-10-05 21:27:18 kamstrup_meter.py open: 118 - DEBUG - Opened serial port]
[2022-10-05 21:27:20 kamstrup_meter.py rd: 131 - DEBUG - Rx timeout]
[2022-10-05 21:27:20 kamstrup_meter.py readparameter: 198 - WARNING - No response from meter]
[2022-10-05 21:27:22 kamstrup_meter.py rd: 131 - DEBUG - Rx timeout]
[2022-10-05 21:27:22 kamstrup_meter.py readparameter: 198 - WARNING - No response from meter]
[2022-10-05 21:27:24 kamstrup_meter.py rd: 131 - DEBUG - Rx timeout]
[2022-10-05 21:27:24 kamstrup_meter.py readparameter: 198 - WARNING - No response from meter]
[2022-10-05 21:27:26 kamstrup_meter.py rd: 131 - DEBUG - Rx timeout]
[2022-10-05 21:27:26 kamstrup_meter.py readparameter: 198 - WARNING - No response from meter]
[2022-10-05 21:27:26 kamstrup_meter.py close: 126 - DEBUG - Closed serial port]
[2022-10-05 21:27:26 mqtt_handler.py publish: 55 - INFO - Publishing 'kamstrup/values' '{}' to 192.168.2.98:1883]
[2022-10-05 21:27:26 daemon.py run: 84 - INFO - Waiting 1 minute(s) for the next meter readout]
This is my configuration file:
mqtt:
host: 192.168.2.98
port: 1883
client: kamstrup402
topic: kamstrup
qos: 1
retain: False
authentication: True
username: pi
password: [password]
serial_device:
com_port: /dev/serial/by-id/usb-Silicon_Labs_CP2102_USB_to_UART_Bridge_Controller_0001-if00-port0
kamstrup:
parameters:
- energy
- volume
- temp1
- temp2
poll_interval: 1
i hope anyone has an idea what i am doing wrong. Thanks in advance.
Hi,
Are you sure that your meter is "awake"? You can do so by pressing the button. It will remain active for 30 minutes and remains active if a valid IR signal is received.
From: madradrakie @.>
Sent: Wednesday, October 5, 2022 9:36:54 PM
To: matthijsvisser/kamstrup-402-mqtt @.>
Cc: Subscribed @.***>
Subject: [matthijsvisser/kamstrup-402-mqtt] Cannot extract meter value - Kamstrup Multical 401 (Issue #7)
I hope you vcan help me. I have a Kamstrup Multical 401 and trying to use your script. I have tried to place de IR head at every possible way at the meter but the following warning keeps appearing. In MQTT i see the topic but no values.
[2022-10-05 21:27:18 daemon.py init: 41 - INFO - initializing daemon]
[2022-10-05 21:27:18 mqtt_handler.py connect: 43 - INFO - Connected to MQTT at: 192.168.2.98:1883]
[2022-10-05 21:27:18 mqtt_handler.py connect: 45 - INFO - with username pi, QoS level = 1 and retain = False]
[2022-10-05 21:27:18 kamstrup_meter.py close: 126 - DEBUG - Closed serial port]
[2022-10-05 21:27:18 kamstrup_meter.py open: 118 - DEBUG - Opened serial port]
[2022-10-05 21:27:20 kamstrup_meter.py rd: 131 - DEBUG - Rx timeout]
[2022-10-05 21:27:20 kamstrup_meter.py readparameter: 198 - WARNING - No response from meter]
[2022-10-05 21:27:22 kamstrup_meter.py rd: 131 - DEBUG - Rx timeout]
[2022-10-05 21:27:22 kamstrup_meter.py readparameter: 198 - WARNING - No response from meter]
[2022-10-05 21:27:24 kamstrup_meter.py rd: 131 - DEBUG - Rx timeout]
[2022-10-05 21:27:24 kamstrup_meter.py readparameter: 198 - WARNING - No response from meter]
[2022-10-05 21:27:26 kamstrup_meter.py rd: 131 - DEBUG - Rx timeout]
[2022-10-05 21:27:26 kamstrup_meter.py readparameter: 198 - WARNING - No response from meter]
[2022-10-05 21:27:26 kamstrup_meter.py close: 126 - DEBUG - Closed serial port]
[2022-10-05 21:27:26 mqtt_handler.py publish: 55 - INFO - Publishing 'kamstrup/values' '{}' to 192.168.2.98:1883]
[2022-10-05 21:27:26 daemon.py run: 84 - INFO - Waiting 1 minute(s) for the next meter readout]
This is my configuration file:
mqtt:
host: 192.168.2.98
port: 1883
client: kamstrup402
topic: kamstrup
qos: 1
retain: False
authentication: True
username: pi
password: a1connect!
serial_device:
com_port: /dev/serial/by-id/usb-Silicon_Labs_CP2102_USB_to_UART_Bridge_Controller_0001-if00-port0
kamstrup:
parameters:
energy
volume
temp1
temp2
poll_interval: 1
i hope anyone has an idea what i am doing wrong. Thanks in advance.
—
Reply to this email directly, view it on GitHubhttps://github.com/matthijsvisser/kamstrup-402-mqtt/issues/7, or unsubscribehttps://github.com/notifications/unsubscribe-auth/AK4J3SUCARZUZNJO7PEGTJLWBXKFNANCNFSM6AAAAAAQ53QGB4.
You are receiving this because you are subscribed to this thread.Message ID: @.***>
hi matthijs. before i activate the script i press the button on the meter. i assume the 401 is working the same as the 402. i have tried for an hour to find the sweet spot but still no luck. it is also very hard to determine if the IR leds of the head are alligned perfectly
You can try the official software from kamstrup to verify if the meter will respond to those commands
See https://www.kamstrup.com/nl-nl/warmteoplossingen/slimme-warmtemeters/meter-software
You can try the official software from kamstrup to verify if the meter will
respons to those commands
See
https://www.kamstrup.com/nl-nl/warmteoplossingen/slimme-warmtemeters/meter-software
Op wo 5 okt. 2022 21:45 schreef madradrakie @.***>:
hi matthijs. before i activate the script i press the button on the meter.
i assume the 401 is working the same as the 402. i have tried for an hour
to find the sweet spot but still no luck. it is also very hard to determine
if the IR leds of the head are alligned perfectly
—
Reply to this email directly, view it on GitHub
https://github.com/matthijsvisser/kamstrup-402-mqtt/issues/7#issuecomment-1268885487,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/AK4J3SV2ORAOBV5QSAMRRMTWBXLGLANCNFSM6AAAAAAQ53QGB4
.
You are receiving this because you commented.Message ID:
@.***>
i have tried the official software but my meter, the kamstrup 401 is not listed. It starts with the 402. I'am really curious if someone has managed to get it working because thereis a sensor eye present on the 401.
I did have the same problem, but after reading the manual I found out that the communication is different between a M402 and M401. At the moment I can read the values from the meter (m401), but I'm still looking into the possibility to send it via Mqtt (no great coder myself). Maybe you can adapt (USB port) and try this bit of code if your IR is placed correctly and is working?:
`
import serial
from time import sleep
mc401 = serial.Serial(port='/dev/ttyUSB1', bytesize=serial.SEVENBITS, parity=serial.PARITY_EVEN, stopbits=serial.STOPBITS_ONE, timeout=2)
mc401.baudrate = 300
mc401.write(bytes("/#1", 'UTF-8'))
mc401.flush()
sleep(1)
mc401.baudrate = 1200
mc401.flushInput()
data = mc401.read(87).split()
try:
print(data[0], data[1], data[3], data[4], data[5], data[6], data[7])
except IndexError:
pass
`
The main problem that I encountered is that you must send a signal at 300 baudrate, but you can only receive at 1200 baudrate. I have used the signal /#1 as this gives back energy, volume, temp1, temp2, tempdiff.
hello matthijs. Nico to read that you got it working. Can you tell me what you mean by adapt and at which place i should place the code? The IR reader is connected with a raspberry pi. I am still learning everyday :-)
I have same issue with MutliCAL 602. It should be using same protocol "KMP" but it is just echoing the same values I'm sending to it. Any idea?
Hi,
I don't have the 602, so unfortunately I cant help you with that. But here are my thoughts:
If you receive the exact same message as the code sends, than its possible that you having issues with reflections. Which IR head are you using? You can try to put some paper between the IR head and the meter to diffuse the light a little. Or try some more positions with the meter head. When I was testing the code, I experienced that the official tooling was able to handles more 'noise' than my code.
Thanks for the info. I actually noticed that it is echoing the data even if it is not sitting on the kamstrup meter? Should the script wait little before reading the data or maybe flush before reading 🤔Laith…On 7 Dec 2022, at 21.46, Matthijs @.***> wrote:
Hi,
I don't have the 602, so unfortunately I cant help you with that. But here are my thoughts:
If you receive the exact same message as the code sends, than its possible that you having issues with reflections. Which IR head are you using? You can try to put some paper between the IR head and the meter to diffuse the light a little. Or try some more positions with the meter head. When I was testing the code, I experienced that the official tooling was able to handles more 'noise' than my code.
—Reply to this email directly, view it on GitHub, or unsubscribe.You are receiving this because you commented.Message ID: @.***>
I have a similar problem.
We have a Kamstrup MultiCAL 66C. According to the specifications, it works with:
300 baud req /300 baud reply, 1 start bit, 7 data bits, equal parity, 2 stop bits.
How can I change this because if I change this in kamstrup_meter.py to
baud rate = 300,
parity = serial.PARITY_EVEN,
bytesize = serial.SEVENBITS,
I get this error within 4 seconds :
Traceback (most recent call last):
File "/tmp/test/kamstrup-402-mqtt/daemon.py", line 92, in
main()
File "/tmp/test/kamstrup-402-mqtt/daemon.py", line 89, in main
daemon.run()
File "/tmp/test/kamstrup-402-mqtt/daemon.py", line 81, in run
values = self.heat_meter.run()
File "/tmp/test/kamstrup-402-mqtt/kamstrup_meter.py", line 114, in run
value = self.readparameter(int(str(kamstrup_402_params[parameter]), 0))
File "/tmp/test/kamstrup-402-mqtt/kamstrup_meter.py", line 205, in read parameter
elif ((receivedMessage[0] != 0x3f) or
IndexError: byte array index out of range
What do I need to change to make the communication work properly?
It looks like you don't receive any data from the meter since the received Message array is empty.
Sorry verkeerde error. This is the correct error.
Traceback (most recent call last):
File "/opt/test/kamstrup-402-mqtt/daemon.py", line 92, in
main()
File "/opt/test/kamstrup-402-mqtt/daemon.py", line 88, in main
daemon = KamstrupDaemon()
File "/opt/test/kamstrup-402-mqtt/daemon.py", line 69, in init
self.heat_meter = kamstrup(serial_cfg["com_port"], kamstrup_cfg["parameters"])
File "/opt/test/kamstrup-402-mqtt/kamstrup_meter.py", line 97, in init
bytesize = serial.SEVENTBITS,
AttributeError: module 'serial' has no attribute 'SEVENTBITS'
So if no data comes in, the application gives an error?
You are using an invalid parameter name. See
https://pyserial.readthedocs.io/en/latest/pyserial_api.html
It should be serial.sevenbits.
From: Obelix74Git @.>
Sent: Friday, December 16, 2022 10:13:47 PM
To: matthijsvisser/kamstrup-402-mqtt @.>
Cc: Matthijs @.>; Comment @.>
Subject: Re: [matthijsvisser/kamstrup-402-mqtt] Cannot extract meter value - Kamstrup Multical 401 (Issue #7)
Sorry verkeerde error. This is the correct error.
Traceback (most recent call last):
File "/opt/test/kamstrup-402-mqtt/daemon.py", line 92, in
main()
File "/opt/test/kamstrup-402-mqtt/daemon.py", line 88, in main
daemon = KamstrupDaemon()
File "/opt/test/kamstrup-402-mqtt/daemon.py", line 69, in init
self.heat_meter = kamstrup(serial_cfg["com_port"], kamstrup_cfg["parameters"])
File "/opt/test/kamstrup-402-mqtt/kamstrup_meter.py", line 97, in init
bytesize = serial.SEVENTBITS,
AttributeError: module 'serial' has no attribute 'SEVENTBITS'
—
Reply to this email directly, view it on GitHubhttps://github.com/matthijsvisser/kamstrup-402-mqtt/issues/7#issuecomment-1355617360, or unsubscribehttps://github.com/notifications/unsubscribe-auth/AK4J3SUNSEYUU66QVUD3IFDWNTLQXANCNFSM6AAAAAAQ53QGB4.
You are receiving this because you commented.Message ID: @.***>
Did you guys get this working? I have a 401, and put the meter over the reader at the bottom right. It seems to be seated properly, and tried all kinds of positions, but no reading. I also clicked the enter/play button on the reader.
Did you guys get this working? I have a 401, and put the meter over the reader at the bottom right. It seems to be seated properly, and tried all kinds of positions, but no reading. I also clicked the enter/play button on the reader.
For info I am using this reader, as it was suggested by someone. It should be working I think? https://www.aliexpress.com/item/1005003509520122.html
Hello Alekslyse. I got it working with for the 401 meter. I used the following code:
https://github.com/matthijsberk/M401-irreader
See for more information:
https://github.com/matthijsberk/M401-irreader/issues/2
|
gharchive/issue
| 2022-10-05T19:36:44 |
2025-04-01T06:39:31.310224
|
{
"authors": [
"Obelix74Git",
"alekslyse",
"laithsaid",
"madradrakie",
"matthijsberk",
"matthijsvisser"
],
"repo": "matthijsvisser/kamstrup-402-mqtt",
"url": "https://github.com/matthijsvisser/kamstrup-402-mqtt/issues/7",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
96345994
|
x64 Detection using Get-ProcAddress
Within Invoke--Shellcode, things often fail on x86 systems. After some digging, it appears that $64BitCPU returns true; even on x86 systems.
Since $64BitCPU appears to always return true, I followed it down to see how $64BitCPU is being determined.
To test this, I copied Get-Win32Functions, Get-ProcAddress and Get-DelegatedType into a separate script. I then added the logic within Invoke--Shellcode "$IsWow64ProcessAddr = Get-ProcAddress kernel32.dll IsWow64Process". When running this script, it always returns a value on x86, instead of null.
Here is the script:
function local:Get-Win32Functions
{
$Win32Functions = New-Object System.Object
$OpenProcessAddr = Get-ProcAddress kernel32.dll OpenProcess
$OpenProcessDelegate = Get-DelegateType @([UInt32], [Bool], [UInt32]) ([IntPtr])
$OpenProcess = [System.Runtime.InteropServices.Marshal]::GetDelegateForFunctionPointer($OpenProcessAddr, $OpenProcessDelegate)
$Win32Functions | Add-Member NoteProperty -Name OpenProcess -Value $OpenProcess
$VirtualAllocExAddr = Get-ProcAddress kernel32.dll VirtualAllocEx
$VirtualAllocExDelegate = Get-DelegateType @([IntPtr], [IntPtr], [Uint32], [UInt32], [UInt32]) ([IntPtr])
$VirtualAllocEx = [System.Runtime.InteropServices.Marshal]::GetDelegateForFunctionPointer($VirtualAllocExAddr, $VirtualAllocExDelegate)
$Win32Functions | Add-Member NoteProperty -Name VirtualAllocEx -Value $VirtualAllocEx
$WriteProcessMemoryAddr = Get-ProcAddress kernel32.dll WriteProcessMemory
$WriteProcessMemoryDelegate = Get-DelegateType @([IntPtr], [IntPtr], [Byte[]], [UInt32], [UInt32].MakeByRefType()) ([Bool])
$WriteProcessMemory = [System.Runtime.InteropServices.Marshal]::GetDelegateForFunctionPointer($WriteProcessMemoryAddr, $WriteProcessMemoryDelegate)
$Win32Functions | Add-Member NoteProperty -Name WriteProcessMemory -Value $WriteProcessMemory
$CreateRemoteThreadAddr = Get-ProcAddress kernel32.dll CreateRemoteThread
$CreateRemoteThreadDelegate = Get-DelegateType @([IntPtr], [IntPtr], [UInt32], [IntPtr], [IntPtr], [UInt32], [IntPtr]) ([IntPtr])
$CreateRemoteThread = [System.Runtime.InteropServices.Marshal]::GetDelegateForFunctionPointer($CreateRemoteThreadAddr, $CreateRemoteThreadDelegate)
$Win32Functions | Add-Member NoteProperty -Name CreateRemoteThread -Value $CreateRemoteThread
$WaitForSingleObjectAddr = Get-ProcAddress kernel32.dll WaitForSingleObject
$WaitForSingleObjectDelegate = Get-DelegateType @([IntPtr], [UInt32])
$WaitForSingleObject = [System.Runtime.InteropServices.Marshal]::GetDelegateForFunctionPointer($WaitForSingleObjectAddr, $WaitForSingleObjectDelegate)
$Win32Functions | Add-Member NoteProperty -Name WaitForSingleObject -Value $WaitForSingleObject
$CloseHandleAddr = Get-ProcAddress kernel32.dll CloseHandle
$CloseHandleDelegate = Get-DelegateType @([IntPtr]) ([Bool])
$CloseHandle = [System.Runtime.InteropServices.Marshal]::GetDelegateForFunctionPointer($CloseHandleAddr, $CloseHandleDelegate)
$Win32Functions | Add-Member NoteProperty -Name CloseHandle -Value $CloseHandle
$GetLastErrorAddr = Get-ProcAddress kernel32.dll GetLastError
$GetLastErrorDelegate = Get-DelegateType @() ([Uint32])
$GetLastError = [System.Runtime.InteropServices.Marshal]::GetDelegateForFunctionPointer($GetLastErrorAddr, $GetLastErrorDelegate)
$Win32Functions | Add-Member NoteProperty -Name GetLastError -Value $GetLastError
$NtCreateThreadExAddr = Get-ProcAddress NtDll.dll NtCreateThreadEx
$NtCreateThreadExDelegate = Get-DelegateType @([IntPtr].MakeByRefType(), [UInt32], [IntPtr], [IntPtr], [IntPtr], [IntPtr], [Bool], [UInt32], [UInt32], [UInt32], [IntPtr]) ([UInt32])
$NtCreateThreadEx = [System.Runtime.InteropServices.Marshal]::GetDelegateForFunctionPointer($NtCreateThreadExAddr, $NtCreateThreadExDelegate)
$Win32Functions | Add-Member -MemberType NoteProperty -Name NtCreateThreadEx -Value $NtCreateThreadEx
# A valid pointer to IsWow64Process will be returned if CPU is 64-bit
$IsWow64ProcessAddr = Get-ProcAddress kernel32.dll IsWow64Process
if ($IsWow64ProcessAddr)
{
$IsWow64ProcessDelegate = Get-DelegateType @([IntPtr], [Bool].MakeByRefType()) ([Bool])
$IsWow64Process = [System.Runtime.InteropServices.Marshal]::GetDelegateForFunctionPointer($IsWow64ProcessAddr, $IsWow64ProcessDelegate)
$Win32Functions | Add-Member NoteProperty -Name IsWow64Process -Value $IsWow64Process
}
return $Win32Functions
}
function Local:Get-ProcAddress
{
Param
(
[OutputType([IntPtr])]
[Parameter( Position = 0, Mandatory = $True )]
[String]
$Module,
[Parameter( Position = 1, Mandatory = $True )]
[String]
$Procedure
)
# Get a reference to System.dll in the GAC
$SystemAssembly = [AppDomain]::CurrentDomain.GetAssemblies() |
Where-Object { $_.GlobalAssemblyCache -And $_.Location.Split('\\')[-1].Equals('System.dll') }
$UnsafeNativeMethods = $SystemAssembly.GetType('Microsoft.Win32.UnsafeNativeMethods')
# Get a reference to the GetModuleHandle and GetProcAddress methods
$GetModuleHandle = $UnsafeNativeMethods.GetMethod('GetModuleHandle')
$GetProcAddress = $UnsafeNativeMethods.GetMethod('GetProcAddress')
# Get a handle to the module specified
$Kern32Handle = $GetModuleHandle.Invoke($null, @($Module))
$tmpPtr = New-Object IntPtr
$HandleRef = New-Object System.Runtime.InteropServices.HandleRef($tmpPtr, $Kern32Handle)
# Return the address of the function
Write-Output $GetProcAddress.Invoke($null, @([System.Runtime.InteropServices.HandleRef]$HandleRef, $Procedure))
}
function Local:Get-DelegateType
{
Param
(
[OutputType([Type])]
[Parameter( Position = 0)]
[Type[]]
$Parameters = (New-Object Type[](0)),
[Parameter( Position = 1 )]
[Type]
$ReturnType = [Void]
)
$Domain = [AppDomain]::CurrentDomain
$DynAssembly = New-Object System.Reflection.AssemblyName('ReflectedDelegate')
$AssemblyBuilder = $Domain.DefineDynamicAssembly($DynAssembly, [System.Reflection.Emit.AssemblyBuilderAccess]::Run)
$ModuleBuilder = $AssemblyBuilder.DefineDynamicModule('InMemoryModule', $false)
$TypeBuilder = $ModuleBuilder.DefineType('MyDelegateType', 'Class, Public, Sealed, AnsiClass, AutoClass', [System.MulticastDelegate])
$ConstructorBuilder = $TypeBuilder.DefineConstructor('RTSpecialName, HideBySig, Public', [System.Reflection.CallingConventions]::Standard, $Parameters)
$ConstructorBuilder.SetImplementationFlags('Runtime, Managed')
$MethodBuilder = $TypeBuilder.DefineMethod('Invoke', 'Public, HideBySig, NewSlot, Virtual', $ReturnType, $Parameters)
$MethodBuilder.SetImplementationFlags('Runtime, Managed')
Write-Output $TypeBuilder.CreateType()
}
$var = Get-ProcAddress kernel32.dll IsWow64Process
$var
Here is the output:
From my understanding, this should return $null on x86 systems. I'm not sure if this has something to do with the fact that it is a VM. My test environment is a full patched win7 x86 box running on VMware fusion with a Mac host.
Let me know if clarification is needed.
Thanks!
Matt N.
sounds like a pull-request for Get-SystemInfo or { $processor = gwmi
win32_processor; $processor.architecture }
My system info powershell script has some logic you could use for this which could be found https://github.com/L1ghtn1ng/powershell-scripts/blob/master/systemInformation.ps1
@L1ghtn1ng because your logic is based on $env variables, the output will depend on which version of powershell you're running using. e.g. x86 powershell returns "32-bit" on a x64 machine.
Fair enough am still new to powershell and IIRC set has some logic and so does metasploit
Thanks for all the input, folks! I fixed the bug using the AddressWidth field of the Win32_Processor class while being mindful of multiple instances being returned as mentioned in issue #75. Pester tests passed when tested in Win 10 (64-bit) and Windows XP (32-bit).
|
gharchive/issue
| 2015-07-21T15:29:25 |
2025-04-01T06:39:31.323660
|
{
"authors": [
"L1ghtn1ng",
"PowerShellMafia",
"enigma0x3",
"secabstraction"
],
"repo": "mattifestation/PowerSploit",
"url": "https://github.com/mattifestation/PowerSploit/issues/70",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
}
|
154144393
|
For windows using cygwin needed to modify command
/**
* Get the command that is going to be executed.
* @return {String}
*/
Rsync.prototype.command = function() {
var a = this.executable();
var b = this.args().join(' ');
//todo: added chmod manually
b = b + ' --chmod=ug=rwx --chmod=o=rx';
return this.executable() + ' ' + b;
};
See commentin #34
|
gharchive/issue
| 2016-05-11T01:59:05 |
2025-04-01T06:39:31.325093
|
{
"authors": [
"born2net",
"mattijs"
],
"repo": "mattijs/node-rsync",
"url": "https://github.com/mattijs/node-rsync/issues/33",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
68012773
|
Multimodal emission probabilites?
Did you ever extend this package to include HDP priors for Gaussian Mixtures for the emission probabilities? I am trying to locate the acceptable probabilities for emission distributions, but can only find Gaussian.
Thanks
The distributions are mostly in pybasicbayes's distributions.py. To make a mixture that can be plugged into an HMM, you'd also want to use the MixtureDistribution class. Something like this:
from pybasicbayes.models import MixtureDistribution
from pybasicbayes.distributions import Gaussian
obs_distns = \
[MixtureDistribution(
alpha_0=5.,
components=[Gaussian(**obs_hypparams) for _ in xrange(5)])
for state in range(Nmax)]
By default those MixtureDistributions act like weak limit approximations to Dirichlet Process mixture models. If you want to make an HDP model (including multiple mixture models) you'll have to write the hierarchical prat yourself.
You can make mixtures of any of the distributions in pybasicbayes (and it's pretty easy to add new ones).
|
gharchive/issue
| 2015-04-13T05:42:00 |
2025-04-01T06:39:31.327501
|
{
"authors": [
"mathDR",
"mattjj"
],
"repo": "mattjj/pyhsmm",
"url": "https://github.com/mattjj/pyhsmm/issues/44",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
162246706
|
SSR
fixes #53
fixes #64
Coverage remained the same at 87.179% when pulling 27cdadab7d22c8078ff2e80daf7d8a31800006f2 on ssr into 75c2202fac7d15fcfc2b86d134386f6b7b3e030f on master.
Coverage remained the same at 87.179% when pulling 27cdadab7d22c8078ff2e80daf7d8a31800006f2 on ssr into 75c2202fac7d15fcfc2b86d134386f6b7b3e030f on master.
Coverage remained the same at 87.179% when pulling 3994f331702497e952e6197853ac678bce1bad88 on ssr into 75c2202fac7d15fcfc2b86d134386f6b7b3e030f on master.
Coverage remained the same at 87.179% when pulling 85e1a8f8a13d49b08479144c3169f3a27d1ef085 on ssr into 75c2202fac7d15fcfc2b86d134386f6b7b3e030f on master.
Coverage remained the same at 87.179% when pulling eecfe566ea81510585e15fad91e85eb4c56d627c on ssr into 75c2202fac7d15fcfc2b86d134386f6b7b3e030f on master.
|
gharchive/pull-request
| 2016-06-24T23:08:52 |
2025-04-01T06:39:31.333573
|
{
"authors": [
"coveralls",
"mattkrick"
],
"repo": "mattkrick/cashay",
"url": "https://github.com/mattkrick/cashay/pull/73",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
559349472
|
Select returns wrong value when SQL type is DATETIME
I use the following code:
In this code I insert into a database date 2 Jan 2020 and insert works correct as shows SQLite DB Browser.
After that I read data from database using SELECT and expect to get 2 Jan 2020. But I get 1 Jan 0001.
package main
import (
"database/sql"
"fmt"
_ "github.com/mattn/go-sqlite3"
"os"
"time"
)
func InitDummyDB(path string) *sql.DB {
db, err := sql.Open("sqlite3", path)
if err != nil {
panic(err)
}
if db == nil {
panic("db is nil")
}
return db
}
func CreateDummyTable(db *sql.DB) {
sqlTable := `
CREATE TABLE IF NOT EXISTS items (
ID INTEGER PRIMARY KEY, TimeT DATETIME); `
res, err := db.Exec(sqlTable)
if err != nil {
panic(err)
}
fmt.Println(res)
}
func InsertDummyDate(db *sql.DB) {
t1 := time.Date(2020, 1, 2, 3, 0, 0, 0, time.UTC)
request, err := db.Prepare(`INSERT INTO items (TimeT) VALUES (?)`)
if err != nil {
panic(err)
}
_, rerr := request.Exec(t1.String())
if rerr != nil {
panic(rerr)
}
}
func PrintDummyData(db *sql.DB) {
rows, err := db.Query(`SELECT ID, TimeT FROM items ORDER BY ID`)
if err != nil {
panic(err)
}
defer rows.Close()
for rows.Next() {
var id int
var readTime time.Time // 0001-01-01 00:00:00 +0000 UTC will be returned
//var readTime string // string does not help: we get 0001-01-01T00:00:00Z
err = rows.Scan(&id, &readTime)
if err != nil {
panic(err)
}
fmt.Println(id, readTime)
}
}
func main() {
remErr := os.Remove("dummy.db")
if remErr != nil {
fmt.Println(remErr)
}
db := InitDummyDB("dummy.db")
CreateDummyTable(db)
InsertDummyDate(db)
PrintDummyData(db)
}
See https://www.sqlite.org/lang_datefunc.html
When you call time.String(), the output is "2020-01-02 03:00:00 +0000 UTC", which is not one of the supported formats. But if you change it to just request.Exec(t1), this library takes care of it so the resulting read works as expected.
(The problem here is that SQLite itself doesn't really support timestamps as a proper datatype. This library does a best effort to deal with them, but as you can see, it isn't perfect.)
It works now, thank you!
|
gharchive/issue
| 2020-02-03T21:36:42 |
2025-04-01T06:39:31.356293
|
{
"authors": [
"lukinma",
"rittneje"
],
"repo": "mattn/go-sqlite3",
"url": "https://github.com/mattn/go-sqlite3/issues/780",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
546236363
|
added missing mode functionality
the mode option in the dsn did not work as expected, particularly when trying to open a database in read-only mode
@AshfordN Do you have an example of it not working correctly? The mode parameter is handled by SQLite itself and is basically applied on top of SQLITE_OPEN_READWRITE/SQLITE_OPEN_READONLY/SQLITE_OPEN_CREATE. Therefore, this change is not needed.
Well the database isn't actually read-only when you specify mode=ro
Coverage increased (+0.2%) to 51.639% when pulling 76ce3bd1fcd6db5f980a2403bdfd5be7755bcf43 on AshfordN:patch-1 into 53cff3fcebd1e177d04129dc131523635bc45d3d on mattn:master.
Please provide an example code segment where it does not work as expected.
The following code successfully inserts and updates a record in the database, using go1.13.1 on linux/amd64:
package main
import (
"database/sql"
"fmt"
_ "github.com/mattn/go-sqlite3"
)
func main() {
//open database
db, err := sql.Open("sqlite3", "main.db?mode=ro")
if err != nil {
panic(err)
}
//attempt insert
if _, err := db.Exec("INSERT INTO test(col1, col2) VALUES(1, 2)"); err != nil {
fmt.Printf("Insert failed: %w\n", err)
}
//attempt update
if _, err := db.Exec("UPDATE test SET col2 = 4 WHERE col1 = 1"); err != nil {
fmt.Printf("Update failed: %w\n", err)
}
}
See https://www.sqlite.org/uri.html. It should be sql.Open("sqlite3", "file:main.db?mode=ro").
Ok, this works, thanks.
|
gharchive/pull-request
| 2020-01-07T11:38:39 |
2025-04-01T06:39:31.361268
|
{
"authors": [
"AshfordN",
"coveralls",
"rittneje"
],
"repo": "mattn/go-sqlite3",
"url": "https://github.com/mattn/go-sqlite3/pull/772",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
858000635
|
Update amalgamation code
Fixes #926, #927
Codecov Report
Merging #940 (d33341f) into master (ab91e93) will decrease coverage by 0.13%.
The diff coverage is n/a.
@@ Coverage Diff @@
## master #940 +/- ##
==========================================
- Coverage 46.80% 46.67% -0.14%
==========================================
Files 11 11
Lines 1457 1457
==========================================
- Hits 682 680 -2
- Misses 640 641 +1
- Partials 135 136 +1
Impacted Files
Coverage Δ
sqlite3.go
52.84% <0.00%> (-0.23%)
:arrow_down:
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update ab91e93...d33341f. Read the comment docs.
Test fail on Windows since installing go-acc failed. But I tested this PR on Windows. So merging.
Thank you so much for the fast turnaround, really appreciate it! 🥳
@mattn thank you so much for the quick turnaround - author of go-acc here also (by chance). I can try and look into the Windows issue - specifically by adding a windows VM to the CI :)
|
gharchive/pull-request
| 2021-04-14T15:15:28 |
2025-04-01T06:39:31.369946
|
{
"authors": [
"aeneasr",
"codecov-io",
"mattn",
"neilalexander"
],
"repo": "mattn/go-sqlite3",
"url": "https://github.com/mattn/go-sqlite3/pull/940",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1111979087
|
Status text placement
Describe the bug
Status text is off screen. This seems to have started immediately after following the most recent major game update.
To Reproduce
Steps to reproduce the behavior:
Install mod with mod manager
Run game
Expected behavior
Readable status text
Screenshots
Mods installed
jaYzZWjA0a
Extra
Unable to upload save and mod save data after trying to ZIP and upload.
Out of curiosity, what resolution are you using? Also is the "UI layout resolution reference height" (in Settings->Gameplay) different than the resolution's height?
This is 2560x1440p with a reference height of 900. This worked just fine prior to update, however with that being said I did also forget my reference height was so low (I enjoy the larger UI elements).
For now I've set the reference height to 1080 and everything is correctly visible again. I'd still appreciate using 900 but also don't mind just using 1080 as a workaround for this.
Thanks for your time. Feel free to close this, due to this reference heigh workaround :)
I'm definitely going to fix it, I'm just compiling a list of resolutions and ref heights to test before I release versions.
I think this should be fixed with the release from yesterday
Confirming this appears to be resolved. Thanks for your support!
|
gharchive/issue
| 2022-01-23T20:36:17 |
2025-04-01T06:39:31.377887
|
{
"authors": [
"Cringely",
"mattsemar"
],
"repo": "mattsemar/dsp-personal-logistics",
"url": "https://github.com/mattsemar/dsp-personal-logistics/issues/20",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.