id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
404838221
identifier $cont6 unknown? Just for fun, I tried running this code through pythran. The resulting error message is not very helpful :) pythran smallest_enclosing_circle.py CRITICAL: I am in trouble. Your input file does not seem to match Pythran's constraints... E: identifier $cont6 unknown, either because it is an unsupported intrinsic, the input code is faulty, or... pythran is buggy. (line 59) https://gist.github.com/nbecker/43880b6461d5626856d9998999af70b1 I've been trying to debug this. Here is some code that pythran doesn't grok, but I don't see why https://gist.github.com/nbecker/697988e93f1954b182d3a204b432e5df It will compile with the last function _make_circle_two_points commented out. OK, corrected the signature to say #pythran export _make_circle_two_points((float,float) list, (float, float), (float, float)) The 1st argument is a list of 2-tuples of float This still doesn't compile. I think the issue here is perhaps that pythran doesn't know the return type. The function _make_circle_two_points will always return the same type, but I don't think pythran can infer that. Don't know how to fix it. Thanks for narrowing the issue. I'll try to give it a look this week end. @nbecker this proved to be a very tricky one. I spent ~6 hours on this one, but in the end I get a few extra bug fixes in Pythran so, let's say we're good. I still need to format stuff a bit, but I'm on my way. Note: [(int, int)] in a pythran export does not create a list, but a tuple... Thanks for working on this! I've been away for a few days on a XC ski trip. Is there something available for me to test now? On Mon, Feb 04, 2019 at 04:10:30AM -0800, ndbecker wrote: Thanks for working on this! I've been away for a few days on a XC ski trip. Is there something available for me to test now? Sure, test https://github.com/serge-sans-paille/pythran/pull/1200 I've updated https://gist.github.com/nbecker/697988e93f1954b182d3a204b432e5df with the now complete code. The original problem is fixed on master, thanks! But the new code fails to compile with clang. I've tried to look at the code and it seems correct to me. I get clang++: /tmp/tmppiy320f2.cpp In file included from /tmp/tmppiy320f2.cpp:16: In file included from /home/nbecker/.local/lib/python3.7/site-packages/pythran/pythonic/include/__builtin__/max.hpp:5: /home/nbecker/.local/lib/python3.7/site-packages/pythran/pythonic/include/__builtin__/minmax.hpp:34:24: error: ambiguous partial specializations of '__combined<double &, const double &>' typename __combined<T0, T1>::type>::type>::type; ^ /home/nbecker/.local/lib/python3.7/site-packages/pythran/pythonic/include/__builtin__/minmax.hpp:43:20: note: in instantiation of template class '(anonymous namespace)::pythonic::__builtin__::details::minmax<false, double &, const double &>' requested here typename minmax<false, Types...>::result_type>::result_type; ^ /home/nbecker/.local/lib/python3.7/site-packages/pythran/pythonic/include/__builtin__/min.hpp:12:21: note: in instantiation of template class '(anonymous namespace)::pythonic::__builtin__::details::minmax<false, double &, double &, const double &>' requested here typename details::minmax<sizeof...(Types) == 1, Types...>::result_type ^ /home/nbecker/.local/lib/python3.7/site-packages/pythran/pythonic/include/__builtin__/min.hpp:15:18: note: while substituting deduced template arguments into function template 'min' [with Types = <double &, double &, const double &>] DEFINE_FUNCTOR(pythonic::__builtin__, min); ^ /tmp/tmppiy320f2.cpp:768:96: note: while substituting deduced template arguments into function template 'operator()' [with Types = <double &, double &, const double &>] typename pythonic::assignable<decltype((pythonic::operator_::div((pythonic::operator_::add(pythonic::__builtin__::functor::min{}(std::get<0>(a), std::get<0>(b), std::get<0>(c)), pythonic::__builtin__::functor::max{}(std::get<0>(a), std::get<0>(b), std::get<0>(c)))), 2.0)))>::type ox = (pythonic::operator_::div((pythonic::operator_::add(pythonic::__builtin__::functor::min{}(std::get<0>(a), std::get<0>(b), std::get<0>(c)), pythonic::__builtin__::functor::max{}(std::get<0>(a), std::get<0>(b), std::get<0>(c)))), 2.0)); ^ /tmp/tmppiy320f2.cpp:875:89: note: in instantiation of function template specialization '__pythran__smallest_enclosing_circle::make_circumcircle::operator()<(anonymous namespace)::pythonic::types::array<double, 2> &, (anonymous namespace)::pythonic::types::array<double, 2> &, const (anonymous namespace)::pythonic::types::array<double, 2> &>' requested here typename pythonic::assignable<decltype(make_circumcircle()(p, q, r))>::type c = make_circumcircle()(p, q, r); ^ /tmp/tmppiy320f2.cpp:939:17: note: in instantiation of function template specialization '__pythran__smallest_enclosing_circle::_make_circle_two_points::operator()<(anonymous namespace)::pythonic::types::sliced_list<(anonymous namespace)::pythonic::types::array<double, 2>, (anonymous namespace)::pythonic::types::contiguous_slice>, (anonymous namespace)::pythonic::types::array<double, 2> &, (anonymous namespace)::pythonic::types::array<double, 2> &>' requested here c = _make_circle_two_points()(points[pythonic::types::contiguous_slice(pythonic::__builtin__::None,(pythonic::operator_::add(std::get<0>(__tuple0), 1L)))], p, q); ^ /tmp/tmppiy320f2.cpp:969:106: note: in instantiation of function template specialization '__pythran__smallest_enclosing_circle::_make_circle_one_point::operator()<(anonymous namespace)::pythonic::types::list<(anonymous namespace)::pythonic::types::array<double, 2> > &, (anonymous namespace)::pythonic::types::array<double, 2> &>' requested here auto res = __pythran__smallest_enclosing_circle::_make_circle_one_point()(points, p); ^ /home/nbecker/.local/lib/python3.7/site-packages/pythran/pythonic/include/types/combined.hpp:26:8: note: partial specialization matches [with T0 = double &, T1 = const double &] struct __combined<T0, T1> { ^ /home/nbecker/.local/lib/python3.7/site-packages/pythran/pythonic/include/types/combined.hpp:66:8: note: partial specialization matches [with T0 = double, T1 = const double &] struct __combined<T0 &, T1> { ^ /home/nbecker/.local/lib/python3.7/site-packages/pythran/pythonic/include/types/combined.hpp:76:8: note: partial specialization matches [with T0 = double &, T1 = const double] struct __combined<T0, T1 &> { ^ /home/nbecker/.local/lib/python3.7/site-packages/pythran/pythonic/include/types/combined.hpp:81:8: note: partial specialization matches [with T0 = double &, T1 = double] struct __combined<T0, T1 const &> { ^ /home/nbecker/.local/lib/python3.7/site-packages/pythran/pythonic/include/types/combined.hpp:86:8: note: partial specialization matches [with T0 = double, T1 = const double] struct __combined<T0 &, T1 &> { ^ /tmp/tmppiy320f2.cpp:768:96: error: no matching function for call to object of type 'pythonic::__builtin__::functor::min' typename pythonic::assignable<decltype((pythonic::operator_::div((pythonic::operator_::add(pythonic::__builtin__::functor::min{}(std::get<0>(a), std::get<0>(b), std::get<0>(c)), pythonic::__builtin__::functor::max{}(std::get<0>(a), std::get<0>(b), std::get<0>(c)))), 2.0)))>::type ox = (pythonic::operator_::div((pythonic::operator_::add(pythonic::__builtin__::functor::min{}(std::get<0>(a), std::get<0>(b), std::get<0>(c)), pythonic::__builtin__::functor::max{}(std::get<0>(a), std::get<0>(b), std::get<0>(c)))), 2.0)); ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /tmp/tmppiy320f2.cpp:875:89: note: in instantiation of function template specialization '__pythran__smallest_enclosing_circle::make_circumcircle::operator()<(anonymous namespace)::pythonic::types::array<double, 2> &, (anonymous namespace)::pythonic::types::array<double, 2> &, const (anonymous namespace)::pythonic::types::array<double, 2> &>' requested here typename pythonic::assignable<decltype(make_circumcircle()(p, q, r))>::type c = make_circumcircle()(p, q, r); ^ /tmp/tmppiy320f2.cpp:939:17: note: in instantiation of function template specialization '__pythran__smallest_enclosing_circle::_make_circle_two_points::operator()<(anonymous namespace)::pythonic::types::sliced_list<(anonymous namespace)::pythonic::types::array<double, 2>, (anonymous namespace)::pythonic::types::contiguous_slice>, (anonymous namespace)::pythonic::types::array<double, 2> &, (anonymous namespace)::pythonic::types::array<double, 2> &>' requested here c = _make_circle_two_points()(points[pythonic::types::contiguous_slice(pythonic::__builtin__::None,(pythonic::operator_::add(std::get<0>(__tuple0), 1L)))], p, q); ^ /tmp/tmppiy320f2.cpp:969:106: note: in instantiation of function template specialization '__pythran__smallest_enclosing_circle::_make_circle_one_point::operator()<(anonymous namespace)::pythonic::types::list<(anonymous namespace)::pythonic::types::array<double, 2> > &, (anonymous namespace)::pythonic::types::array<double, 2> &>' requested here auto res = __pythran__smallest_enclosing_circle::_make_circle_one_point()(points, p); ^ /home/nbecker/.local/lib/python3.7/site-packages/pythran/pythonic/include/__builtin__/min.hpp:15:3: note: candidate template ignored: substitution failure [with Types = <double &, double &, const double &>]: no matching function for call to 'min' DEFINE_FUNCTOR(pythonic::__builtin__, min); ^ ~~~~~~~~ /home/nbecker/.local/lib/python3.7/site-packages/pythran/pythonic/include/utils/functor.hpp:28:31: note: expanded from macro 'DEFINE_FUNCTOR' #define DEFINE_FUNCTOR(ns, f) DEFINE_FUNCTOR_2(f, ns::f) ^ ~~ /home/nbecker/.local/lib/python3.7/site-packages/pythran/pythonic/include/utils/functor.hpp:14:12: note: expanded from macro 'DEFINE_FUNCTOR_2' auto operator()(Types && ... types) const \ ^ 2 errors generated. WARNING: Compilation error, trying hard to find its origin... CRITICAL: You shall not pass! E: Invalid argument type for function call to `Callable[[T1, T2, T3], ...]`, no overload found, tried: Callable[[Iterable[T0]], T0] Callable[[float, float], float] Callable[[int, int], int] (line 10) On Mon, Feb 4, 2019 at 7:12 AM serge-sans-paille notifications@github.com wrote: On Mon, Feb 04, 2019 at 04:10:30AM -0800, ndbecker wrote: Thanks for working on this! I've been away for a few days on a XC ski trip. Is there something available for me to test now? Sure, test https://github.com/serge-sans-paille/pythran/pull/1200 — You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub https://github.com/serge-sans-paille/pythran/issues/1197#issuecomment-460227486, or mute the thread https://github.com/notifications/unsubscribe-auth/AAHK0LyxH-SPRviCD-q_638Zs6prIyFOks5vKCPIgaJpZM4aagNv . In the process of debug the indentation had gotten messed up in the last gist, I have fixed it. But, this doesn't change the compile error. @nbecker branch updated with a fix!
gharchive/issue
2019-01-30T15:56:20
2025-04-01T04:35:50.639650
{ "authors": [ "nbecker", "serge-sans-paille" ], "repo": "serge-sans-paille/pythran", "url": "https://github.com/serge-sans-paille/pythran/issues/1197", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
99350190
StanfordCoreNLP.annotate will hang with some input text. Here is how StanfordCoreNLP is initialized. // Annotation pipeline configuration var props = new Properties(); props.setProperty("annotators", "tokenize, ssplit, pos, parse, sentiment"); props.setProperty("sutime.binders", "0"); // Make sure the model files are extracted to the current directory from the .jar file. var nlp = new StanfordCoreNLP(props); Annotation annotation = new Annotation(text); nlp.annotate(annotation); The following (long) text input will hang StanfordCoreNLP.annotate call. Process Start in Azure Website ramiramilu&#39 s answer is actually not correct You can run any exe you want ([check this for example](https://github com/projectkudu/kudu/wiki/Kudu-console)) The problem you are hitting is not in running an exe it&#39 s something specific to wkhtml2pdf exe itself That exe uses a bunch of GDI+ calls on Windows for rendering the PDF and that is what&#39 s not allowed in Azure Websites sandbox (the GDI+ calls not running an exe) Using a WebJob won&#39 t help either because WebJobs run in the same context as the site which means under the same sandbox Edit: There is nothing special to how you would launch an external process on Azure Websites than how you would do it normally with C# Again the problem you are facing is with wkhtml2pdf exe and not with the general concept of launching a process Here is a sample that you can try that launches cmd exe and reads what is written on stdout var processStartInfo = new ProcessStartInfo() { Arguments = &quot /c echo \&quot test\&quot &quot FileName = @&quot c:\windows\system32\cmd exe&quot RedirectStandardOutput = true UseShellExecute = false } var process = Process Start(processStartInfo) using (var streamReader = new StreamReader(process StandardOutput BaseStream)) { ViewBag MessageFromExe = streamReader ReadToEnd() } ViewBag MessageFromExe will have the value &quot test&quot you can verify that on your view and you can run that just fine in Azure Websites Close as an old issue
gharchive/issue
2015-08-06T04:24:57
2025-04-01T04:35:50.650638
{ "authors": [ "sergey-tihon", "zxli" ], "repo": "sergey-tihon/Stanford.NLP.NET", "url": "https://github.com/sergey-tihon/Stanford.NLP.NET/issues/21", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
665275844
Статья 12 http://localhost:8000/article/article-12/ Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur. test comment #2 const app = new App(); hello hi
gharchive/issue
2020-07-24T16:19:52
2025-04-01T04:35:50.652839
{ "authors": [ "sergeyyarkov" ], "repo": "sergeyyarkov/devthread.ru", "url": "https://github.com/sergeyyarkov/devthread.ru/issues/25", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
561710894
appsettings.json equivalent to ElasticsearchJsonFormatter() I am using Serilog with a .NET Core 3.1 web application, but I am configuring Serilog via the appsettings.json; however, I am trying to establish how to apply ElasticsearchJsonFormatter() to the "formatter" field of the appsettings.json { "AllowedHosts": "*", "Serilog": { "Using": ["Serilog.Sinks.Console"], "MinimumLevel": { "Default": "Verbose", "Override": { "Microsoft": "Warning", "System": "Warning" } }, "Enrich": [ "FromLogContext", "WithMachineName", "WithProcessId", "WithThreadId" ], "WriteTo": [ { "Name": "Console", "Args": { "formatter": "Serilog.Formatting.Json.JsonFormatter, Serilog" } }, ] } } As you can see I have the standard json formatter, but I don't know what to put for the ElasticsearchJsonFormatter(). Is this even possible, or do I have to reconfigure Serilog to not use appsettings.json Did you ever find a solution for this ? Hi, I installed the nuget package Serilog.Sinks.Elasticsearch and chaged the formatter option in the appsettings.json to Serilog.Formatting.Elasticsearch.ElasticsearchJsonFormatter,Serilog.Formatting.Elasticsearch { "AllowedHosts": "*", "Serilog": { "Using": ["Serilog.Sinks.Console"], "MinimumLevel": { "Default": "Verbose", "Override": { "Microsoft": "Warning", "System": "Warning" } }, "Enrich": [ "FromLogContext", "WithMachineName", "WithProcessId", "WithThreadId" ], "WriteTo": [ { "Name": "Console", "Args": { "formatter": "Serilog.Formatting.Elasticsearch.ElasticsearchJsonFormatter,Serilog.Formatting.Elasticsearch" } }, ] } } Hi all, This seems an old thread and currently it doesn't directly related with this repo. But just want to check if it is possible to set formatProvider in Console with appSettings.json settings. Currently I also tried the previous(☝) post's suggestion, but it doesn't work. Just code-behind setting works. If there is any other solution, would be happy to have. Hi @ardacetinkaya . My suggestion is to rule out one of these two problems first. Serilog is loading its configuration from appsettings.json ? Is any component failing to load? To rule out both of these problems, you need to modify Program.cs in the host project. Some things to check Line 5: Log into Console.Error (stderr on linux) if anything goes wrong with Serilog Line 11 : CreateDefaultBuilder is adding configuration sources , appsetting.json is among them. If you are using something else be sure to add appsettings.json as a configuration source Line 14: logConfig.ReadFrom.Configuration(ctx.Configuration) loads the Serilog configuration the app's configuration that includes appsettings.json (because of line 11) This is a small example in .netcore 3.1 public class Program { public static void Main(string[] args) { Serilog.Debugging.SelfLog.Enable(System.Console.Error); CreateHostBuilder(args).Build().Run(); } public static IHostBuilder CreateHostBuilder(string[] args) => Host .CreateDefaultBuilder(args) .UseSerilog((ctx, logConfig) => { logConfig.ReadFrom.Configuration(ctx.Configuration); }) .ConfigureWebHostDefaults(webBuilder => { webBuilder.UseStartup<Startup>(); }); } Hope this helps!!! @yuudj Thanks it worked
gharchive/issue
2020-02-07T15:35:50
2025-04-01T04:35:50.691274
{ "authors": [ "Vandersteen", "ardacetinkaya", "daz1761", "sumeshes07", "yuudj" ], "repo": "serilog-contrib/serilog-sinks-elasticsearch", "url": "https://github.com/serilog-contrib/serilog-sinks-elasticsearch/issues/314", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2316899849
Support provider name customization Is your feature request related to a problem? Please describe. As an end user, I'd love to be able to customize the name of the Provider Name that can be viewed in the UI dropdown. Describe the solution you'd like I'd love to be able to customize the Provider Name when I register the provider in the service provider, using an optional parameter or an extension method. Additional context As per request from comment in #66 Implemented in #114.
gharchive/issue
2024-05-25T10:10:16
2025-04-01T04:35:50.693745
{ "authors": [ "followynne" ], "repo": "serilog-contrib/serilog-ui", "url": "https://github.com/serilog-contrib/serilog-ui/issues/119", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1765475793
SpanId and TraceId remains similar Hi Team, I am using serilog sink for dotnet 3.1 core application. I am trying to generate log console for capturing trace and span information. I get the traceid not generated based on activity rather I get same value everytime i execute and spanid has | appended with traceid as shown below. {"Timestamp":"2023-06-20T13:47:27.0775037+00:00","Level":"Information","TraceId":"9c351ca5-4cb764e0336735c8","SpanId":"|9c351ca5-4cb764e0336735c8.","Message":"Request finished in 31.6978ms 200 "} Please let me know if I am doing things correctly. Happy to provide more information @Veeraraghavans Can you provide a code snippet with how you've configured the sink and how you are generating your logs? The trace and span IDs you've shown in your example message do not look like the standard values coming from a .NET activity. Thanks @loomis for your comments I fixed it.
gharchive/issue
2023-06-20T13:59:32
2025-04-01T04:35:50.707177
{ "authors": [ "Veeraraghavans", "loomis" ], "repo": "serilog/serilog-sinks-opentelemetry", "url": "https://github.com/serilog/serilog-sinks-opentelemetry/issues/93", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2330087810
Error messages appear in UI after running backup Whenever I run a backup, I get a stream of errors showing up in the bottom right hand corner of the UI with an exclamation point and the following messages: Service method 'items' - file name: backup_service.py - fn name: get - line: 100 Service method 'data' - file name: stats_service.py - fn name: stats - line: 105 The first message appears a lot more frequently than the second. In case it's helpful, I'm running velero client v1.13.2 (to match server) and the velero-ui/api/watchdog package was installed into my cluster via helm. All checks show as OK on the initial login page. Here is the logs from my velero-api pod, where you can see the above errors triggering: E='items', F=/app/service/backup_service.py, L=100 INFO: 10.224.1.142:59518 - "GET /stats/in-progress HTTP/1.1" 400 Bad Request E='items', F=/app/service/backup_service.py, L=100 INFO: 10.224.1.142:59518 - "GET /stats/in-progress HTTP/1.1" 400 Bad Request E='items', F=/app/service/backup_service.py, L=100 INFO: 10.224.1.142:59518 - "GET /stats/in-progress HTTP/1.1" 400 Bad Request E='items', F=/app/service/backup_service.py, L=100 INFO: 10.224.1.142:59518 - "GET /stats/in-progress HTTP/1.1" 400 Bad Request E='items', F=/app/service/backup_service.py, L=100 INFO: 10.224.1.142:59518 - "GET /stats/in-progress HTTP/1.1" 400 Bad Request E='items', F=/app/service/backup_service.py, L=100 INFO: 10.224.1.142:59518 - "GET /stats/in-progress HTTP/1.1" 400 Bad Request E='items', F=/app/service/backup_service.py, L=100 INFO: 10.224.1.142:59518 - "GET /stats/in-progress HTTP/1.1" 400 Bad Request E='items', F=/app/service/backup_service.py, L=100 INFO: 10.224.1.142:59518 - "GET /stats/in-progress HTTP/1.1" 400 Bad Request E='items', F=/app/service/backup_service.py, L=100 INFO: 10.224.1.142:59518 - "GET /stats/in-progress HTTP/1.1" 400 Bad Request E='items', F=/app/service/backup_service.py, L=100 INFO: 10.224.1.142:59518 - "GET /stats/in-progress HTTP/1.1" 400 Bad Request E='items', F=/app/service/backup_service.py, L=100 INFO: 10.224.1.142:59518 - "GET /stats/in-progress HTTP/1.1" 400 Bad Request E='items', F=/app/service/backup_service.py, L=100 INFO: 10.224.1.142:59518 - "GET /stats/in-progress HTTP/1.1" 400 Bad Request INFO: ('10.224.1.142', 34574) - "WebSocket /ws" [accepted] E='items', F=/app/service/backup_service.py, L=100 INFO: 10.224.1.142:46562 - "GET /stats/in-progress HTTP/1.1" 400 Bad Request E='items', F=/app/service/backup_service.py, L=100 INFO: 10.224.1.142:55968 - "GET /stats/in-progress HTTP/1.1" 400 Bad Request E='items', F=/app/service/backup_service.py, L=100 INFO: 10.224.1.142:55968 - "GET /stats/in-progress HTTP/1.1" 400 Bad Request E='items', F=/app/service/backup_service.py, L=102 INFO: 10.224.1.142:55968 - "GET /backup/get HTTP/1.1" 400 Bad Request E='items', F=/app/service/backup_service.py, L=100 INFO: 10.224.1.142:55968 - "GET /stats/in-progress HTTP/1.1" 400 Bad Request INFO: 10.224.1.142:55968 - "GET /restore/get HTTP/1.1" 200 OK E='items', F=/app/service/backup_service.py, L=100 INFO: 10.224.1.142:55968 - "GET /stats/in-progress HTTP/1.1" 400 Bad Request INFO: 10.224.1.142:55968 - "GET /schedule/get HTTP/1.1" 200 OK E='items', F=/app/service/backup_service.py, L=102 E='data', F=/app/service/stats_service.py, L=105 INFO: 10.224.1.142:55968 - "GET /stats/get HTTP/1.1" 400 Bad Request E='items', F=/app/service/backup_service.py, L=100 INFO: 10.224.1.142:55968 - "GET /stats/in-progress HTTP/1.1" 400 Bad Request E='items', F=/app/service/backup_service.py, L=102 INFO: 10.224.1.142:55968 - "GET /backup/get HTTP/1.1" 400 Bad Request INFO: 10.224.1.142:55968 - "GET /restore/get HTTP/1.1" 200 OK INFO: 10.224.1.142:55968 - "GET /schedule/get HTTP/1.1" 200 OK INFO: 10.224.1.142:55968 - "GET /snapshot-location/get HTTP/1.1" 200 OK WARNING: 2024-06-03 03:36:23.347 [service.sc_mapping_service] [/api/v1/sc/change-storage-classes-config-map/get] [admin] 404 Error reading ConfigMap 'change-storage-classes-config' in namespace 'core-test24-velero' INFO: 10.224.1.142:55968 - "GET /sc/change-storage-classes-config-map/get HTTP/1.1" 200 OK E='items', F=/app/service/backup_service.py, L=100 INFO: 10.224.1.142:55968 - "GET /stats/in-progress HTTP/1.1" 400 Bad Request INFO: ('10.224.1.142', 39520) - "WebSocket /ws" [accepted] Connected user via socket: admin E='items', F=/app/service/backup_service.py, L=100 INFO: 10.224.1.142:39506 - "GET /stats/in-progress HTTP/1.1" 400 Bad Request E='items', F=/app/service/backup_service.py, L=100 INFO: 10.224.1.142:39506 - "GET /stats/in-progress HTTP/1.1" 400 Bad Request E='items', F=/app/service/backup_service.py, L=100 INFO: 10.224.1.142:39506 - "GET /stats/in-progress HTTP/1.1" 400 Bad Request E='items', F=/app/service/backup_service.py, L=100 INFO: 10.224.1.142:39506 - "GET /stats/in-progress HTTP/1.1" 400 Bad Request E='items', F=/app/service/backup_service.py, L=100 INFO: 10.224.1.142:39506 - "GET /stats/in-progress HTTP/1.1" 400 Bad Request E='items', F=/app/service/backup_service.py, L=100 INFO: 10.224.1.142:39506 - "GET /stats/in-progress HTTP/1.1" 400 Bad Request E='items', F=/app/service/backup_service.py, L=100 INFO: 10.224.1.142:39506 - "GET /stats/in-progress HTTP/1.1" 400 Bad Request E='items', F=/app/service/backup_service.py, L=100 INFO: 10.224.1.142:39506 - "GET /stats/in-progress HTTP/1.1" 400 Bad Request E='items', F=/app/service/backup_service.py, L=100 INFO: 10.224.1.142:39506 - "GET /stats/in-progress HTTP/1.1" 400 Bad Request E='items', F=/app/service/backup_service.py, L=100 INFO: 10.224.1.142:39506 - "GET /stats/in-progress HTTP/1.1" 400 Bad Request E='items', F=/app/service/backup_service.py, L=100 INFO: 10.224.1.142:39506 - "GET /stats/in-progress HTTP/1.1" 400 Bad Request E='items', F=/app/service/backup_service.py, L=100 INFO: 10.224.1.142:39506 - "GET /stats/in-progress HTTP/1.1" 400 Bad Request E='items', F=/app/service/backup_service.py, L=100 INFO: 10.224.1.142:39506 - "GET /stats/in-progress HTTP/1.1" 400 Bad Request INFO: ('10.224.1.142', 55806) - "WebSocket /ws" [accepted] E='items', F=/app/service/backup_service.py, L=100 INFO: 10.224.1.142:39264 - "GET /stats/in-progress HTTP/1.1" 400 Bad Request INFO: ('10.224.1.142', 50552) - "WebSocket /ws" [accepted] Connected user via socket: admin Connected user via socket: admin E='items', F=/app/service/backup_service.py, L=100 INFO: 10.224.1.142:50564 - "GET /stats/in-progress HTTP/1.1" 400 Bad Request Connected user via socket: admin E='items', F=/app/service/backup_service.py, L=100 INFO: 10.224.1.142:48344 - "GET /stats/in-progress HTTP/1.1" 400 Bad Request INFO: ('10.224.1.142', 48358) - "WebSocket /ws" [accepted] E='items', F=/app/service/backup_service.py, L=100 INFO: 10.224.1.142:48344 - "GET /stats/in-progress HTTP/1.1" 400 Bad Request E='items', F=/app/service/backup_service.py, L=100 INFO: 10.224.1.142:48344 - "GET /stats/in-progress HTTP/1.1" 400 Bad Request E='items', F=/app/service/backup_service.py, L=102 E='data', F=/app/service/stats_service.py, L=105 INFO: 10.224.1.142:48344 - "GET /stats/get HTTP/1.1" 400 Bad Request And by the way - thanks for making this great product available to the community! Actually looks like I might have figured out what's causing it. I was getting the errors above after running a single backup - they weren't showing up when no backups were run, and I later found out that it looked fine after two or more backups were run. Then I ran some test commands in the api pod, see below for results. Note the absence of the items[] array when there is only one backup, it returns a "kind" of "Backup" with only one, whereas zero or 2+ backups returns a "kind" of "BackupList" which has the items array - and which the code in velero-api depends on at those line numbers identified above. With no backups: root@velero-ui-vui-api-7db69577cb-x8mcb:/app# velero backup get -o json { "kind": "BackupList", "apiVersion": "velero.io/v1", "metadata": { "resourceVersion": "222188" }, "items": [] } With only one backup: root@velero-ui-vui-api-7db69577cb-x8mcb:/app# velero backup get -o json -n core-test24-velero { "kind": "Backup", "apiVersion": "velero.io/v1", "metadata": { "name": "test2", "namespace": "core-test24-velero", "uid": "7702996b-848b-4917-a95a-cca9f6853db8", "resourceVersion": "207088", "generation": 6, "creationTimestamp": "2024-06-03T03:20:28Z", "labels": { "velero.io/storage-location": "default" }, "annotations": { "velero.io/resource-timeout": "10m0s", "velero.io/source-cluster-k8s-gitversion": "v1.29.4", "velero.io/source-cluster-k8s-major-version": "1", "velero.io/source-cluster-k8s-minor-version": "29" }, "managedFields": [ { "manager": "velero", "operation": "Update", "apiVersion": "velero.io/v1", "time": "2024-06-03T03:20:28Z", "fieldsType": "FieldsV1", "fieldsV1": { "f:spec": { ".": {}, "f:defaultVolumesToFsBackup": {}, "f:hooks": {}, "f:includedNamespaces": {}, "f:metadata": {}, "f:snapshotVolumes": {}, "f:ttl": {} }, "f:status": {} } }, { "manager": "velero-server", "operation": "Update", "apiVersion": "velero.io/v1", "time": "2024-06-03T03:20:32Z", "fieldsType": "FieldsV1", "fieldsV1": { "f:metadata": { "f:annotations": { ".": {}, "f:velero.io/resource-timeout": {}, "f:velero.io/source-cluster-k8s-gitversion": {}, "f:velero.io/source-cluster-k8s-major-version": {}, "f:velero.io/source-cluster-k8s-minor-version": {} }, "f:labels": { ".": {}, "f:velero.io/storage-location": {} } }, "f:spec": { "f:csiSnapshotTimeout": {}, "f:itemOperationTimeout": {}, "f:snapshotMoveData": {}, "f:storageLocation": {}, "f:volumeSnapshotLocations": {} }, "f:status": { "f:completionTimestamp": {}, "f:expiration": {}, "f:formatVersion": {}, "f:hookStatus": {}, "f:phase": {}, "f:progress": { ".": {}, "f:itemsBackedUp": {}, "f:totalItems": {} }, "f:startTimestamp": {}, "f:version": {} } } } ] }, "spec": { "metadata": {}, "includedNamespaces": [ "core-test24-grafana" ], "snapshotVolumes": true, "ttl": "720h0m0s", "hooks": {}, "storageLocation": "default", "volumeSnapshotLocations": [ "default" ], "defaultVolumesToFsBackup": true, "csiSnapshotTimeout": "10m0s", "itemOperationTimeout": "4h0m0s", "snapshotMoveData": false }, "status": { "version": 1, "formatVersion": "1.1.0", "expiration": "2024-07-03T03:20:28Z", "phase": "Completed", "startTimestamp": "2024-06-03T03:20:28Z", "completionTimestamp": "2024-06-03T03:20:32Z", "progress": { "totalItems": 46, "itemsBackedUp": 46 }, "hookStatus": {} } } With two or more backups: root@velero-ui-vui-api-7db69577cb-x8mcb:/app# velero backup get -o json -n core-test24-velero { "kind": "BackupList", "apiVersion": "velero.io/v1", "metadata": { "resourceVersion": "222678" }, "items": [ { "kind": "Backup", "apiVersion": "velero.io/v1", "metadata": { "name": "test2", "namespace": "core-test24-velero", "uid": "7702996b-848b-4917-a95a-cca9f6853db8", "resourceVersion": "207088", "generation": 6, "creationTimestamp": "2024-06-03T03:20:28Z", "labels": { "velero.io/storage-location": "default" }, "annotations": { "velero.io/resource-timeout": "10m0s", "velero.io/source-cluster-k8s-gitversion": "v1.29.4", "velero.io/source-cluster-k8s-major-version": "1", "velero.io/source-cluster-k8s-minor-version": "29" }, "managedFields": [ { "manager": "velero", "operation": "Update", "apiVersion": "velero.io/v1", "time": "2024-06-03T03:20:28Z", "fieldsType": "FieldsV1", "fieldsV1": { "f:spec": { ".": {}, "f:defaultVolumesToFsBackup": {}, "f:hooks": {}, "f:includedNamespaces": {}, "f:metadata": {}, "f:snapshotVolumes": {}, "f:ttl": {} }, "f:status": {} } }, { "manager": "velero-server", "operation": "Update", "apiVersion": "velero.io/v1", "time": "2024-06-03T03:20:32Z", "fieldsType": "FieldsV1", "fieldsV1": { "f:metadata": { "f:annotations": { ".": {}, "f:velero.io/resource-timeout": {}, "f:velero.io/source-cluster-k8s-gitversion": {}, "f:velero.io/source-cluster-k8s-major-version": {}, "f:velero.io/source-cluster-k8s-minor-version": {} }, "f:labels": { ".": {}, "f:velero.io/storage-location": {} } }, "f:spec": { "f:csiSnapshotTimeout": {}, "f:itemOperationTimeout": {}, "f:snapshotMoveData": {}, "f:storageLocation": {}, "f:volumeSnapshotLocations": {} }, "f:status": { "f:completionTimestamp": {}, "f:expiration": {}, "f:formatVersion": {}, "f:hookStatus": {}, "f:phase": {}, "f:progress": { ".": {}, "f:itemsBackedUp": {}, "f:totalItems": {} }, "f:startTimestamp": {}, "f:version": {} } } } ] }, "spec": { "metadata": {}, "includedNamespaces": [ "core-test24-grafana" ], "snapshotVolumes": true, "ttl": "720h0m0s", "hooks": {}, "storageLocation": "default", "volumeSnapshotLocations": [ "default" ], "defaultVolumesToFsBackup": true, "csiSnapshotTimeout": "10m0s", "itemOperationTimeout": "4h0m0s", "snapshotMoveData": false }, "status": { "version": 1, "formatVersion": "1.1.0", "expiration": "2024-07-03T03:20:28Z", "phase": "Completed", "startTimestamp": "2024-06-03T03:20:28Z", "completionTimestamp": "2024-06-03T03:20:32Z", "progress": { "totalItems": 46, "itemsBackedUp": 46 }, "hookStatus": {} } }, { "kind": "Backup", "apiVersion": "velero.io/v1", "metadata": { "name": "test3", "namespace": "core-test24-velero", "uid": "b554a3ca-f57c-4846-82f0-e68152185a46", "resourceVersion": "222636", "generation": 6, "creationTimestamp": "2024-06-03T03:53:54Z", "labels": { "velero.io/storage-location": "default" }, "annotations": { "velero.io/resource-timeout": "10m0s", "velero.io/source-cluster-k8s-gitversion": "v1.29.4", "velero.io/source-cluster-k8s-major-version": "1", "velero.io/source-cluster-k8s-minor-version": "29" }, "managedFields": [ { "manager": "velero", "operation": "Update", "apiVersion": "velero.io/v1", "time": "2024-06-03T03:53:54Z", "fieldsType": "FieldsV1", "fieldsV1": { "f:spec": { ".": {}, "f:defaultVolumesToFsBackup": {}, "f:hooks": {}, "f:includedNamespaces": {}, "f:includedResources": {}, "f:metadata": {}, "f:snapshotVolumes": {}, "f:ttl": {} }, "f:status": {} } }, { "manager": "velero-server", "operation": "Update", "apiVersion": "velero.io/v1", "time": "2024-06-03T03:53:55Z", "fieldsType": "FieldsV1", "fieldsV1": { "f:metadata": { "f:annotations": { ".": {}, "f:velero.io/resource-timeout": {}, "f:velero.io/source-cluster-k8s-gitversion": {}, "f:velero.io/source-cluster-k8s-major-version": {}, "f:velero.io/source-cluster-k8s-minor-version": {} }, "f:labels": { ".": {}, "f:velero.io/storage-location": {} } }, "f:spec": { "f:csiSnapshotTimeout": {}, "f:itemOperationTimeout": {}, "f:snapshotMoveData": {}, "f:storageLocation": {}, "f:volumeSnapshotLocations": {} }, "f:status": { "f:completionTimestamp": {}, "f:expiration": {}, "f:formatVersion": {}, "f:hookStatus": {}, "f:phase": {}, "f:progress": {}, "f:startTimestamp": {}, "f:version": {} } } } ] }, "spec": { "metadata": {}, "includedNamespaces": [ "core-test24-grafana" ], "includedResources": [ "persistentvolumes" ], "snapshotVolumes": true, "ttl": "720h0m0s", "hooks": {}, "storageLocation": "default", "volumeSnapshotLocations": [ "default" ], "defaultVolumesToFsBackup": true, "csiSnapshotTimeout": "10m0s", "itemOperationTimeout": "4h0m0s", "snapshotMoveData": false }, "status": { "version": 1, "formatVersion": "1.1.0", "expiration": "2024-07-03T03:53:54Z", "phase": "Completed", "startTimestamp": "2024-06-03T03:53:54Z", "completionTimestamp": "2024-06-03T03:53:55Z", "progress": {}, "hookStatus": {} } } ] } Hi Thank you for the report. Yes, there is a small problem with both the backup and the schedules (and possibly some other resources as well) when a single item is present. We will try to fix it as soon as possible.
gharchive/issue
2024-06-03T03:30:01
2025-04-01T04:35:50.721654
{ "authors": [ "davideserio", "dcaputo-harmoni" ], "repo": "seriohub/velero-ui", "url": "https://github.com/seriohub/velero-ui/issues/27", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
525127288
Remove legacy HTML attributes The HTML attributes data-id, data-edit-type and data-edit-field aren't used any more and thus can be deleted. Closed by #109
gharchive/issue
2019-11-19T16:39:57
2025-04-01T04:35:50.733625
{ "authors": [ "inyono", "kulla" ], "repo": "serlo/serlo.org", "url": "https://github.com/serlo/serlo.org/issues/108", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
270914098
Add bash autocompletion to importify commands I heard that optparse-applicative supports bash autocompletion. It would be really great to add such support to importify. As a addition to the issue here is detailed info about completion in optparse-applicative: https://github.com/pcapriotti/optparse-applicative/wiki/Bash-Completion It says that by default, options and commands are always completed, but arguments completion should be enabled.
gharchive/issue
2017-11-03T09:10:01
2025-04-01T04:35:50.735376
{ "authors": [ "ChShersh", "vrom911" ], "repo": "serokell/importify", "url": "https://github.com/serokell/importify/issues/77", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2263701410
🛑 [KaKi87] Mail - SMTP (port 465) is down In f36b5e0, [KaKi87] Mail - SMTP (port 465) (kaki87.net) was down: HTTP code: 0 Response time: 0 ms Resolved: [KaKi87] Mail - SMTP (port 465) is back up in 8c27546 after 10 minutes.
gharchive/issue
2024-04-25T14:10:39
2025-04-01T04:35:50.750992
{ "authors": [ "server-KaTys" ], "repo": "server-KaTys/status", "url": "https://github.com/server-KaTys/status/issues/393", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1562728135
Local environment files are not used The .env.*.local files are not used. I also tried deleting non-local files. But even then it does not work. Is it a bug or do I have to do something to enable support for it? Non-local and also stage dependent env files are used correctly. SST Version: SST v2.0.0-rc.56 Works for me (in rc56). But I think the fix is pretty recent. I have only tried .env.local, not any other variant. Actually, I just tried .env.<stage>.local, and it doesn't work. Not sure if it's intended or not. Not working with both .env.local and .env.*.local in SST v2.0.1 Works if I copy the vars to .env I created a new project with SST v2.0.23, created environment variables with the same name and different values in the .env, .env.local, .env.dev, and .env.dev.local files then tried running yarn build --stage=dev, only .env.dev.local did not work. I think it is because dotenv.config() is not used to read .env.<stage>.local. packages/sst/src/project.ts L141 ProjectContext.provide(project); dotenv.config({ path: path.join(project.paths.root, `.env.${project.config.stage}`), override: true, }); should be ProjectContext.provide(project); dotenv.config({ path: path.join(project.paths.root, `.env.${project.config.stage}`), override: true, }); dotenv.config({ path: path.join(project.paths.root, `.env.${project.config.stage}.local`), override: true, }); Let me ask Dax to take a look and then can you submit a PR with this @moochannel? Yep looks good can you submit a PR? Thanks for merging my PR. I will leave the repository I used for the reproduction here: moochannel/sst-repro-dotenv
gharchive/issue
2023-01-30T16:06:04
2025-04-01T04:35:50.756561
{ "authors": [ "0ptional", "estyrke", "jayair", "manurana", "moochannel", "thdxr" ], "repo": "serverless-stack/sst", "url": "https://github.com/serverless-stack/sst/issues/2397", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2230457047
Randomly getting 'Email Already in use' error while trying to signin I was trying to Sign up using serverpod_auth_email. But while pressing create account button I am getting 'Email already in user' error even though there are no entries in the db. Video: https://github.com/serverpod/serverpod/assets/113328135/ac4c1dfc-dc04-4d37-b441-686b2d8dfc8e My DB fields: server.dart code: auth.AuthConfig.set( auth.AuthConfig( sendValidationEmail: (session, email, validationCode) async { print('email validationCode: $validationCode'); return true; }, sendPasswordResetEmail: (session, userInfo, validationCode) async { print('password reset validationCode: $validationCode'); return true; }, ), ); To Reproduce Steps to reproduce the behavior: Email Authentication flow implementation as in documentation. Using SignInWithEmailButton to input the field Enter fields and press Create Account button. Expected behavior The user should be onboarded if there is no email conflict. But getting the email conflict even thought there are no users in the db. Serverpod versions CLI Version: 1.2.6 Library version: 1.2.6 Platform information Running on Windows 11. Tested on Real device. Hey man, this is a known issue, reported in #1969 and #2051. This is due to this part in the code (I think) which returns "Email already in use', when the authentication does not succeed. It's a little more tricky to set up, but it is possible. Did you run serverpod generate after you added the module to the server https://docs.serverpod.dev/concepts/authentication/setup#server-setup ? Yeah, @BenAuerDev I ran serverpod generate. Infact I shutdown everything, created new db, applied fresh migrations, started server and tried again. But still it is giving the same issue. I remember last day I tried this in my Mac machine and for the first 2 tries it gave this issue, but on the third try I was able to sign in even though I didn't make any changes to the code. But today in windows, I am following the exact steps, but not able to sign in no matter how many times I retries. @AslamThachapalli damn that's bitter :-/. Maybe try this config/project on MAC again to see if it has to do with windows Oh god. I got the error. It was some lack of attention from my end. Used wrong port to connect with the server. I somehow happened to modify the port from 8080 to 8000, and that caused the error. But due to the random 'Email already in use' error, I didn't pay attention to it. I tried another endpoint, not the authentication one, and thus came to know about the blunder. It is very much required to have a meaningful error message while authentication. Closing this issue. Thanks @BenAuerDev for your responses. @AslamThachapalli Awesome :) glad to hear it worked out fine I faced the same issue while working with serverpod and email authentication. My problem was that the SMTP server was not configured correctly.
gharchive/issue
2024-04-08T07:23:03
2025-04-01T04:35:50.850504
{ "authors": [ "AslamThachapalli", "BenAuerDev", "mtariqsiddiqui" ], "repo": "serverpod/serverpod", "url": "https://github.com/serverpod/serverpod/issues/2109", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2302544367
Decouple auth module Is your feature request related to a problem? Please describe. I want to add custom authentication. So I need to add the serverpod_auth_server package to do this. However, it depends on the abandoned (last commit was 2 years ago) firebase_admin package that depends on http 0.13 which is one year old. And the main thing is that I don't even need the google sign in. Describe the solution you'd like Authorization module should be decoupled from the google, apple and other implementations. We can simply create a base auth package or create a separate package for each methods. Describe alternatives you've considered Alternatively, firebase_admin package can be forked to update it's dependencies. Additional context We are planning on doing this, we are currently reworking some auth stuff to make way for this implementation.
gharchive/issue
2024-05-17T11:55:25
2025-04-01T04:35:50.853545
{ "authors": [ "Isakdl", "Maksimka101" ], "repo": "serverpod/serverpod", "url": "https://github.com/serverpod/serverpod/issues/2269", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2312058578
Generated client.dart-code has errors (Target of URI doesn't exist - database_connection.dart) Describe the bug Recently, I used transaction for the first time in my serverpod app. I used it in the implementation of endpoints. Now, the file protocol/client.dart, produced by the command 'serverpod generate' contains the line import 'package:serverpod/src/database/database_connection.dart' as _i4; This produces the error: Target of URI doesn't exist: 'package:serverpod/src/database/database_connection.dart'. Try creating the file referenced by the URI, or try using a URI for a file that does exist. To Reproduce Steps to reproduce the behavior: In my case, the following endpoint implementation caused the error: import 'package:eventer_server/src/generated/event_location.dart'; import 'package:serverpod/serverpod.dart'; class EventLocationEndpoint extends Endpoint { Future create(Session session, EventLocation eventLocation, Transaction? transaction) async { return await EventLocation.db .insertRow(session, eventLocation, transaction: transaction); } Future delete(Session session, EventLocation eventLocation, Transaction? transaction) async { return await EventLocation.db .deleteRow(session, eventLocation, transaction: transaction); } } Expected behavior The generated code should be correct. Serverpod versions CLI Version: 1.2.7 Library version: 1.2.7 Platform information I use viscose on Mac. Hello @hoeverma! The reason you are getting this error is that you have a transaction as an input parameter to an endpoint. This makes it generate on the client side as well. Transactions are a server-only concept since they only make sense in the scope of an open database connection. Once your endpoint has returned the connection for that request to the database will be closed and the transaction would not be based in anything. If you remove transaction from your endpoints parameters the error should go away. Thank you very much, I got it! @SandPod But if I wanted to run a transaction within the current endpoint and also call a method from another endpoint as part of the same transaction, I would have to pass the transaction object as a parameter to that other endpoint, right? @SandPod But if I wanted to run a transaction within the current endpoint and also call a method from another endpoint as part of the same transaction, I would have to pass the transaction object as a parameter to that other endpoint, right? We don't recommend calling endpoints within endpoints. In that case, I would suggest you extract the shared logic from the endpoint and just make it a regular function. Then the transaction could be a parameter to that shared function. @SandPod But if I had to, is this a bad approach to do it ? extension SessionX on Session { UsersEndpoint get usersEndpoint { final usersEndpoint = (server.endpoints .getConnectorByName(UsersEndpoint.endpointConnectorName) ?.endpoint as UsersEndpoint?); if (usersEndpoint == null) { throw Exception( 'Endpoint connector "${UsersEndpoint.endpointConnectorName}" not found'); } return usersEndpoint; } ConfigsEndpoint get configsEndpoint { ... } ... } And then in any other endpoint I can do this: final users = await session.usersEndpoint.getAll(session); @SalahAnwer-dev I'm sorry, but I don't think I understand the motivation behind forcing the business logic to be defined inside of the endpoint. Could you elaborate a bit on what you see the as benefits and at the same time highligt the drawbacks of moving it out? The main benefit that endpoint bring are: They define the API layer of your server (serialization and endpoints definition). Generated client code to make calling the server simple. Authentication is handled by the framework. From the example code you have presented, I don't understand how this is helpful. From my perspective, it seems that you have two, or more, entry points into your server that rely on the same business logic. This should be the perfect opportunity to encapsulate that business logic with a clear interface. It also adds the benefit that if your endpoints evolve in a different direction, the unit of your buisiness logic would still stay the same.
gharchive/issue
2024-05-23T06:25:38
2025-04-01T04:35:50.864060
{ "authors": [ "SalahAnwer-dev", "SandPod", "hoeverma" ], "repo": "serverpod/serverpod", "url": "https://github.com/serverpod/serverpod/issues/2303", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2516755624
Sign in with email (auth module) shows numeric keyboard, but requires text input Haven't verified myself, but got a bug report that sign in with email shows a numeric keyboard for entering the authentication code, but the code has letter too. So, it's impossible to enter the code from the keyboard. Not sure which platforms are affected. @vlidholt i think this is fixed by @BenAuerDev https://github.com/serverpod/serverpod/pull/2455 Closing as resolved.
gharchive/issue
2024-09-10T15:14:55
2025-04-01T04:35:50.866072
{ "authors": [ "SandPod", "klkucaj", "vlidholt" ], "repo": "serverpod/serverpod", "url": "https://github.com/serverpod/serverpod/issues/2735", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1508229840
🛑 sg3.vmes.xyz is down In 0206dfb, sg3.vmes.xyz (sg3.vmes.xyz) was down: HTTP code: 0 Response time: 0 ms Resolved: sg3.vmes.xyz is back up in 0862d81.
gharchive/issue
2022-12-22T17:00:28
2025-04-01T04:35:50.874565
{ "authors": [ "serversshnet" ], "repo": "serversshnet/uptime", "url": "https://github.com/serversshnet/uptime/issues/732", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
695004267
New Approval Controller Framework Summary Clear description or/and business justification of what the task is about. Details Create controller on base of Sample Adapter Source Code as maven module. Acceptance criteria Controller is deployable in APP The issue should be closed within https://github.com/servicecatalog/oscm-app/pull/86
gharchive/issue
2020-09-07T11:26:58
2025-04-01T04:35:50.876809
{ "authors": [ "GoebelL", "kowalczyka" ], "repo": "servicecatalog/oscm-approval", "url": "https://github.com/servicecatalog/oscm-approval/issues/34", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1291139248
🛑 Typeform (Fridges & Freezers) is down In 1467390, Typeform (Fridges & Freezers) (https://servicesite.typeform.com/to/HMORmA) was down: HTTP code: 403 Response time: 18 ms Resolved: Typeform (Fridges & Freezers) is back up in ef6a318. Resolved: Typeform (Fridges & Freezers) is back up in 22be3b3.
gharchive/issue
2022-07-01T09:12:46
2025-04-01T04:35:50.879947
{ "authors": [ "alistairreay" ], "repo": "servicesite/upptime", "url": "https://github.com/servicesite/upptime/issues/198", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1344955096
🛑 Barnsley Repairs is down In 066b7e6, Barnsley Repairs (https://barnsleyrepairs.co.uk) was down: HTTP code: 0 Response time: 0 ms Resolved: Barnsley Repairs is back up in 8763163.
gharchive/issue
2022-08-19T22:50:16
2025-04-01T04:35:50.882294
{ "authors": [ "alistairreay" ], "repo": "servicesite/upptime", "url": "https://github.com/servicesite/upptime/issues/427", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
189253264
Support for macro defined string constants I am trying to make bindings for the OpenFX API and they define some constants using macros: #define kOfxImageEffectPluginApi "OfxImageEffectPluginAPI" #define kOfxImageEffectPluginApiVersion 1 Bindgen seems to have no trouble with the numeric constant, but it ignores the string constant. Would it be possible to add support for macro defined string constants? Patch up in #260, with a bunch of other improvements. That was fast! Thank you :)
gharchive/issue
2016-11-14T23:12:03
2025-04-01T04:35:50.883752
{ "authors": [ "azerupi", "emilio" ], "repo": "servo/rust-bindgen", "url": "https://github.com/servo/rust-bindgen/issues/256", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
140670359
Missing commas in buildbot master config r? @Manishearth @edunham @aneeshusa This change is @bors-servo r+ :pushpin: Commit 9641b6d has been approved by Ms2ger :hourglass: Testing commit 9641b6d with merge fa5b2d6... :sunny: Test successful - travis
gharchive/pull-request
2016-03-14T13:15:52
2025-04-01T04:35:50.904172
{ "authors": [ "Ms2ger", "bors-servo", "larsbergstrom" ], "repo": "servo/saltfs", "url": "https://github.com/servo/saltfs/pull/251", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
410002154
linearize_pretrends() and residualize_covariates() does not respect anticipation Currently, both use supplied long_data and restrict to the pre-treatment period, i.e., cal_time_var < onset_time_var. For the case with anticipation > 0, we'd want this to only use the pre-anticipation period. This is fixed. Both now look at the cal_time_var < onset_time_var - anticipation period, as they should.
gharchive/issue
2019-02-13T21:06:25
2025-04-01T04:35:50.957884
{ "authors": [ "davidnov" ], "repo": "setzler/eventStudy", "url": "https://github.com/setzler/eventStudy/issues/8", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
460705853
[graphene]: add timestamp setting The PyChunkedGraph supports root_id retrieval with a timestamp. Currently, neuroglancer only allows retrieval of the latest root_id. I would like to make this query option available to the user, eg. through a box in the graphene segmentation layer. Proofreading should be disabled when any time but the current time is selected. Do you have typical work flow or can describe a task the user would try to accomplish with this date selection box? I am trying to figure out what the expected behavior would be for modifying the date (set it back and forth several times) while root objects are already selected: Would I load all associated formerParents / laterParents everytime? That could lead to an "explosion" of selected segments after a few iterations. Should I Also, while I can prevent sending split/merge requests through the Neuroglancer interface, the graph server should still check if the request looks reasonable. For example, if the given rootObjectID is the most recent. We combine annotations and segmentation at specific timestamps (materialization). Hence, users run their analysis with an older than the current state of the segmentation. It would be helpful to configure neuroglancer to reflect that state. I agree that there needs to be some specific behavior when it comes to selected root ids. For now, I would just deselect all root ids when the timestamp is changed (and have a popup warning about this?). Mergers and splits are independent of what the root id is and the graph server can still process them. Since the "shown" root ids are not send to the graph server, it has no way of telling what the user saw when making that call. I suggested this freeze because (1) the users decision is not based on the most recent state of the segmentation and (2) the result would not be shown because it is more recent than the selected timestamp. I am basically arguing that it does not make sense from a frontend persepctive to allow the user to make proofreading changes, the graph server is agnostic to that. Implemented in https://github.com/seung-lab/neuroglancer/pull/341
gharchive/issue
2019-06-26T00:37:04
2025-04-01T04:35:50.969028
{ "authors": [ "nkemnitz", "ogewan", "sdorkenw" ], "repo": "seung-lab/neuroglancer", "url": "https://github.com/seung-lab/neuroglancer/issues/284", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2393369640
🛑 media-mstd (Cloudflare CDN, S3) is down In ab65fe9, media-mstd (Cloudflare CDN, S3) (https://media-mstd.seungjin.net/accounts/avatars/109/737/937/659/013/254/original/626c9187e341632b.jpg) was down: HTTP code: 0 Response time: 0 ms Resolved: media-mstd (Cloudflare CDN, S3) is back up in f56dfda after 9 minutes.
gharchive/issue
2024-07-06T03:47:20
2025-04-01T04:35:50.971717
{ "authors": [ "seungjin" ], "repo": "seungjin/uptime", "url": "https://github.com/seungjin/uptime/issues/514", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1514938224
Remove parking_lot dependency now that std's Mutex is faster than before Hi. Do you think we could remove the dependency to parking_lot now that std's Mutex is faster than before? Regards. Why? parking_lot also avoids the poisoning behavior. Oh, I thought it was used for performance reasons. So, removing it would reduce the build time. But if it was used to avoid this behavior, then I guess there's no reason to remove it. parking_lot also avoids the poisoning behavior. You can get same behaviour by simple wrapper: struct Mutex<T>(std::sync::Mutex<T>); impl<T> Mutex<T> { pub fn lock(&self) -> std::sync::MutexGuard<T> { self.0.lock().unwrap_or_else(PoisonError::into_inner) } }
gharchive/issue
2022-12-31T01:17:11
2025-04-01T04:35:50.989080
{ "authors": [ "BratSinot", "antoyo", "sfackler" ], "repo": "sfackler/r2d2", "url": "https://github.com/sfackler/r2d2/issues/138", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
130912865
This adds a list of the users current adoptions to the search sidebar The list is comprised of the names of the drains along with their reverse geocoded address where the name is a link to zoom the map to that drain's location. Addresses #88 @jasonlally @jeanwalshie let me know what you think about this as a first pass, at least. Looks something like: looks great to me. when you click on the link it zooms to the place on the map where the drain is? @jeanwalshie yep!
gharchive/pull-request
2016-02-03T05:00:47
2025-04-01T04:35:50.998233
{ "authors": [ "jeanwalshie", "jszwedko" ], "repo": "sfbrigade/adopt-a-drain", "url": "https://github.com/sfbrigade/adopt-a-drain/pull/135", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
204221841
When will rails_admin 1.1.2 be released? I'd like to use kaminari 1.0.x. Related PR #2810 has already been merged. Would prefer not to be linking to master in our Gemfile. Related ISSUE #2823. Thanks. Please release new version @sferik :) I think it would be great to have it release sooner than later There are always more bugs to be fixed Please? 😄 🚢 https://github.com/sferik/rails_admin/commit/3d12fd1b6ea4ab70f149cd6a70deb3ca18207083 @mshibuya Thank you so much for the release! Much appreciated! On Tue, May 30, 2017 at 8:33 PM Koichi ITO notifications@github.com wrote: @mshibuya https://github.com/mshibuya Thank you so much for the release! — You are receiving this because you commented. Reply to this email directly, view it on GitHub https://github.com/sferik/rails_admin/issues/2826#issuecomment-305072731, or mute the thread https://github.com/notifications/unsubscribe-auth/AAhZg8naMZ4FoxfqCskt7NUceY4NzU6Mks5r_N-LgaJpZM4LyT7v .
gharchive/issue
2017-01-31T07:23:03
2025-04-01T04:35:51.005141
{ "authors": [ "PikachuEXE", "koic", "mshibuya", "pschambacher", "swistaczek" ], "repo": "sferik/rails_admin", "url": "https://github.com/sferik/rails_admin/issues/2826", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1937005
custom field partials I'm using rails admin on several projects, and I made heavy use of the feature of using my own partials for a field just by specifying on config. Now I started a new project with rails 3.1 and the new rails admin (849898ff5bb1f4b2f3774320d7590b8c54946c31), but I found that rails admin doesn't reach to my partials. I followed the docs and tried several ways to do this, but I couldn't get it to work. I see in the logs that rails admin renders its own '_form_field.html.haml' partial. When I put my own '_form_field.html.haml' on rails_admin/main, then I see it renders my form_field partial. OK so far. But it never renders my custom '_my_partial.html.haml' (or '_my_partial.html.erb' either) when I put it on rails_admin/main or any other place. I tried this way: config.model MyModel do edit do field :my_field do partial 'my_partial' end end end But also this way: config.model MyModel do edit do field :my_field do render do bindings[:view].render :partial => 'my_partial', :locals => {:field => self, :form => bindings[:form]} end end end end Also tried specifying the full pathname, and using haml and erb partials. None worked. Rails admin always goes for _form_field.html.haml, either its own or my local one. Also I didn't get any errors that could give some clues. On the list section, 'pretty_value' doesn't seem to work either, but that's maybe unrelated. What am I doing wrong? partials must be in your app/views/rails_admin/main folder. Name it _my_partial.html.haml and reference it like: config.model MyModel do edit do field :my_field do partial :my_partial end end end It should work... It doesn't work for me. It seems like it's ignoring the code block given for the field. I'll dive into your code to figure why, if I have the time. If I don't get the time, I'll simply do the job with some previous version. Will tell you if I find anything. Hello, I had the same issue and was able to get it somewhat functional by defining the partial as you suggested. This works: partial :flight_reservation_details However, I cannot pass locals to it as with the previous implemtation. This breaks it: partial :flight_reservation_details, :locals => { :flight => bindings[:object], :field => self } OK, I have an update which is a bit of a hack but i think does the job. 1.) Edit the fields base class located in "libs/rails_admin/config/fields/base.rb" and add locals to the instance's accessible attributes like this attr_accessor :defined, :order, :locals This will essentially add the locals property to all fields, you can use this as your own little storage variable to pass data to the fields. 2.) Define the property in the config with whatever data you want, like this RailsAdmin.config do |config| config.model SomeModelName do edit do field :whatever do self.locals = {:object => bindings[:object], :some_variable => "something", :something_else => "sure, why not..."} partial :my_custom_partial end end end end 3.) Create the partial name like this "app/views/rails_admin/main/_my_custom_partial.html.erb" 4.) Use your locals property from the partial like this <%= "some variable = " + locals.some_variable + ", something else = " + locals.something_else %> This should print on the page this some variable = something, something else = sure, why not... Now, what do you guys think about adding locals to the base attr_accessor list? Let me know if you like the concept and I'll add a new branch and submit the pull request. Here's another workaround - to access your object inside partial you can use that: field.bindings[:form].options[:parent_builder].object @Rustik form.object should work in any partial. Same issue here do I have to apply jorgeramos patch ? does it change anything if rails admin is installed as a plugin or a gem...I've been trying everythings I have a similar situation, i call the partial successfully in edit section which applies for new and edit records, my partial renders a collection_select with values from one table in my database for users to choose. rails_admin.rb: config.model Tc::TcAgencia do edit do field :tc_tipo_agencia do partial :tc_tipos_agencias_partial end field :tc_region do partial :tc_regiones_partial end end end _tc_tipos_agencias_partial.html.erb: <%= collection_select(:tc_tc_agencia, :tc_tipo_agencia_codigo, Tc::TcTipoAgencia.all, :codigo, :nombre, {:include_blank => true}, {:placeholder=>"Buscar", :style=>"display:none;"}) %> When the user create a new record everything works as expected, but in edit view the collection_select doesn't mark the value saved for that field as selected, it renders the same as the new view. I know that using form.object.[name_of_my_field] is possible to get the saved value but i don't know how to show it as selected in collection_select, it suppose to be taken by the object and rendered as selected without doing anything else. Thanks. @ejmm320 Use the form.select(see Rails doc) I tested the :partial option, no worries whatsoever. render was not configurable, made it configurable again, adding a spec. @bbenezech Thanks a lot!! I fixed the problem using form.select like you said and now the partial behaves the way we expected. This one NOT working with rails_admin-1.1.1. the solution mentioned does not work with list command, (we're trying to implement inline edit)
gharchive/issue
2011-10-18T07:49:33
2025-04-01T04:35:51.020366
{ "authors": [ "Rustik", "amicming", "bbenezech", "ejmm320", "epdp", "jorgeramos", "rodeofly", "shqear93" ], "repo": "sferik/rails_admin", "url": "https://github.com/sferik/rails_admin/issues/787", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1190720887
🛑 Nodo is down In 7ce811e, Nodo (https://nodotecnologico.catamarcacapital.edu.ar/) was down: HTTP code: 0 Response time: 0 ms Resolved: Nodo is back up in f22d089.
gharchive/issue
2022-04-02T19:34:55
2025-04-01T04:35:51.035695
{ "authors": [ "juanml14" ], "repo": "sfvc/upptime", "url": "https://github.com/sfvc/upptime/issues/164", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1762490692
🛑 Civitas Juzgado is down In 5b8d151, Civitas Juzgado (http://civitas.catamarcaciudad.gob.ar:8080/JUZ_CATA/forms/login.jsp) was down: HTTP code: 0 Response time: 0 ms Resolved: Civitas Juzgado is back up in c109e1b.
gharchive/issue
2023-06-18T22:38:01
2025-04-01T04:35:51.038080
{ "authors": [ "juanml14" ], "repo": "sfvc/upptime", "url": "https://github.com/sfvc/upptime/issues/870", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2381826426
Intersecting text taking a while Hi, Firstly, thanks for the awesome library. I am trying to extrude and intersect two pieces of text: const { drawText } = replicad; const main = () => { const john = drawText("JOHN", { startX: 0, startY: 0, }) .sketchOnPlane("XZ") .extrude(50); const mary = drawText("MARY", { startX: 50, startY: 0, }) .sketchOnPlane("YZ") .extrude(50); const combined = john.intersect(mary); return [ combined ] }; In the workbench, however, it takes a while to complete, around ~2-5 minutes which is longer than I expected such an operation to take. Unfortunately, boolean operations with the base library I use (open cascade) are relatively slow. That said, your code run on my machine (4 years old intel macbook pro) in about 10 seconds. Do you use a particular browser / machine? Interesting, that's unfortunate as the library is great otherwise. I tried doing the same thing in JSCad but it doesn't have as nice support for extruding text or directly rendering in the browser. I'm using a Macbook Pro with the Apple M1 Pro, I tried on both Firefox and Chrome and they were both equally slow (ranging from 1 to 5 minutes)
gharchive/issue
2024-06-29T15:56:07
2025-04-01T04:35:51.041849
{ "authors": [ "r-bt", "sgenoud" ], "repo": "sgenoud/replicad", "url": "https://github.com/sgenoud/replicad/issues/159", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2754020507
[Bug] Qwen/QwQ-32B-Preview undefined symbol error Checklist [X] 1. I have searched related issues but cannot get the expected help. [X] 2. The bug has not been fixed in the latest version. [X] 3. Please note that if the bug-related issue you submitted lacks corresponding environment info and a minimal reproducible demo, it will be challenging for us to reproduce and resolve the issue, reducing the likelihood of receiving feedback. [X] 4. If the issue you raised is not a bug but a question, please raise a discussion at https://github.com/sgl-project/sglang/discussions/new/choose Otherwise, it will be closed. [X] 5. Please use English, otherwise it will be closed. Describe the bug $ python -m sglang.launch_server --model-path Qwen/QwQ-32B-Preview 2024-12-21 13:11:42.425868: I tensorflow/core/util/port.cc:153] oneDNN custom operations are on. You may see slightly different numerical results due to floating-point round-off errors from different computation orders. To turn them off, set the environment variable `TF_ENABLE_ONEDNN_OPTS=0`. 2024-12-21 13:11:42.443184: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:477] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered WARNING: All log messages before absl::InitializeLog() is called are written to STDERR E0000 00:00:1734786702.465802 2843 cuda_dnn.cc:8310] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered E0000 00:00:1734786702.472602 2843 cuda_blas.cc:1418] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered 2024-12-21 13:11:42.494050: I tensorflow/core/platform/cpu_feature_guard.cc:210] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations. To enable the following instructions: AVX512F AVX512_VNNI AVX512_BF16 AVX512_FP16 AVX_VNNI, in other operations, rebuild TensorFlow with the appropriate compiler flags. /usr/lib/python3/dist-packages/scipy/__init__.py:146: UserWarning: A NumPy version >=1.17.3 and <1.25.0 is required for this version of SciPy (detected version 1.26.4 warnings.warn(f"A NumPy version >={np_minversion} and <{np_maxversion}" WARNING 12-21 13:11:44 _custom_ops.py:20] Failed to import from vllm._C with ImportError('/home/ubuntu/.local/lib/python3.10/site-packages/vllm/_C.abi3.so: undefined symbol: _ZN5torch3jit17parseSchemaOrNameERKSsb') Traceback (most recent call last): File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main return _run_code(code, main_globals, None, File "/usr/lib/python3.10/runpy.py", line 86, in _run_code exec(code, run_globals) File "/home/ubuntu/.local/lib/python3.10/site-packages/sglang/launch_server.py", line 5, in <module> from sglang.srt.server import launch_server File "/home/ubuntu/.local/lib/python3.10/site-packages/sglang/srt/server.py", line 46, in <module> from sglang.srt.managers.data_parallel_controller import ( File "/home/ubuntu/.local/lib/python3.10/site-packages/sglang/srt/managers/data_parallel_controller.py", line 27, in <module> from sglang.srt.managers.scheduler import run_scheduler_process File "/home/ubuntu/.local/lib/python3.10/site-packages/sglang/srt/managers/scheduler.py", line 63, in <module> from sglang.srt.managers.tp_worker import TpModelWorker File "/home/ubuntu/.local/lib/python3.10/site-packages/sglang/srt/managers/tp_worker.py", line 25, in <module> from sglang.srt.model_executor.model_runner import ModelRunner File "/home/ubuntu/.local/lib/python3.10/site-packages/sglang/srt/model_executor/model_runner.py", line 42, in <module> from sglang.srt.layers.attention.flashinfer_backend import FlashInferAttnBackend File "/home/ubuntu/.local/lib/python3.10/site-packages/sglang/srt/layers/attention/flashinfer_backend.py", line 28, in <module> from flashinfer import ( File "/home/ubuntu/.local/lib/python3.10/site-packages/flashinfer/__init__.py", line 17, in <module> from .activation import gelu_and_mul, gelu_tanh_and_mul, silu_and_mul File "/home/ubuntu/.local/lib/python3.10/site-packages/flashinfer/activation.py", line 32, in <module> raise e File "/home/ubuntu/.local/lib/python3.10/site-packages/flashinfer/activation.py", line 23, in <module> from . import _kernels ImportError: /home/ubuntu/.local/lib/python3.10/site-packages/flashinfer/_kernels.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZN3c106detail23torchInternalAssertFailEPKcS2_jS2_RKSs Reproduction python -m sglang.launch_server --model-path Qwen/QwQ-32B-Preview Environment $ python3 -m sglang.check_env Traceback (most recent call last): File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main return _run_code(code, main_globals, None, File "/usr/lib/python3.10/runpy.py", line 86, in _run_code exec(code, run_globals) File "/home/ubuntu/.local/lib/python3.10/site-packages/sglang/check_env.py", line 213, in <module> check_env() File "/home/ubuntu/.local/lib/python3.10/site-packages/sglang/check_env.py", line 195, in check_env env_info.update(get_package_versions(PACKAGE_LIST)) File "/home/ubuntu/.local/lib/python3.10/site-packages/sglang/check_env.py", line 54, in get_package_versions module = importlib.import_module(package_name) File "/usr/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "<frozen importlib._bootstrap>", line 1050, in _gcd_import File "<frozen importlib._bootstrap>", line 1027, in _find_and_load File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked File "<frozen importlib._bootstrap>", line 688, in _load_unlocked File "<frozen importlib._bootstrap_external>", line 883, in exec_module File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed File "/home/ubuntu/.local/lib/python3.10/site-packages/flashinfer/__init__.py", line 17, in <module> from .activation import gelu_and_mul, gelu_tanh_and_mul, silu_and_mul File "/home/ubuntu/.local/lib/python3.10/site-packages/flashinfer/activation.py", line 32, in <module> raise e File "/home/ubuntu/.local/lib/python3.10/site-packages/flashinfer/activation.py", line 23, in <module> from . import _kernels ImportError: /home/ubuntu/.local/lib/python3.10/site-packages/flashinfer/_kernels.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZN3c106detail23torchInternalAssertFailEPKcS2_jS2_RKSs pip install "sglang[all]" --find-links https://flashinfer.ai/whl/cu121/torch2.4/flashinfer/ --force-reinstall
gharchive/issue
2024-12-21T13:17:13
2025-04-01T04:35:51.047352
{ "authors": [ "HuanzhiMao", "zhyncs" ], "repo": "sgl-project/sglang", "url": "https://github.com/sgl-project/sglang/issues/2539", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2759597947
Refactor Scheduler to improve code organization Motivation When I try to deep into the Zero-Overhead Batch Scheduler , I find is hard to get clear on the scheduling, and is hard to impl a new scheduling policy, so I try to refactor SchedulePolicy,and make it easy to add new policy for me and others. McCabe indicates that the code complexity has exceeded 15 Modifications Move sorting logic into separate static methods for better maintainability Testing: python3 -m sglang.launch_server --model Qwen/Qwen2.5-0.5B-Instruct python3 -m sglang.bench_serving --backend sglang --dataset-name random --num-prompts 500 --random-input 4096 --random-output 2048 Checklist [x] Format your code according to the Contributor Guide. [x] Add unit tests as outlined in the Contributor Guide. [x] Update documentation as needed, including docstrings or example tutorials. related with https://github.com/sgl-project/sglang/pull/2571 cc @merrymercy cc @merrymercy @hnyls2002 if you have time for this PR I would like to optimize for the task in https://github.com/sgl-project/sglang/issues/2273 Further reduce the scheduling overhead of mixed chunked prefill by simplifying the mix_with_running. The current code first constructs a prefill batch and a decode batch and them merge them. A better method can directly construct a whole mixed batch. @libratiger Are you in the slack channel? If you are interested in that, we can chat in more details.
gharchive/pull-request
2024-12-26T10:29:35
2025-04-01T04:35:51.053459
{ "authors": [ "libratiger", "merrymercy" ], "repo": "sgl-project/sglang", "url": "https://github.com/sgl-project/sglang/pull/2593", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
205174234
fix issue #61 “No such module PDFGenerator” added framework dependency Sorry for replying late. Thank you for the help 🙇
gharchive/pull-request
2017-02-03T14:38:51
2025-04-01T04:35:51.062936
{ "authors": [ "mkowalski87", "sgr-ksmt" ], "repo": "sgr-ksmt/PDFGenerator", "url": "https://github.com/sgr-ksmt/PDFGenerator/pull/62", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
107028206
Get a 403 if I use request.setVisibleInDownloadsUi(true); Hi @sgrebnov, I was trying to use your plugin and I was wondering, why you need to use this line: request.setVisibleInDownloadsUi(false); https://github.com/sgrebnov/cordova-plugin-background-download/blob/master/src/android/BackgroundDownload.java#L175 I want to make the downloaded file visible in the Downloads UI, so I set that to true (which is the default value): request.setVisibleInDownloadsUi(true); and I also tried commenting that line. Both times the downloads failed and the reason is 403 (Forbidden). Do you know if there is a way to make the downloads visible in Downloads UI? Thanks in advance and congrats for this excellent plugin. Sorry about this. I don't know how but this is working now. Although, even if I set it to true, the file doesn't appears in the Downloads UI.
gharchive/issue
2015-09-17T17:00:45
2025-04-01T04:35:51.069518
{ "authors": [ "rbournissent" ], "repo": "sgrebnov/cordova-plugin-background-download", "url": "https://github.com/sgrebnov/cordova-plugin-background-download/issues/12", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1299774910
Access specific MetaEdit command with buttons Hi, this is as deep as I can get into MetaEdit from a button: type command action MetaEdit: Run MetaEdit name Update status Because of this, I still have to wade through all these options after clicking: Can't I just pick an option: create new YAML meta-data > status > completed? ever found a solution to this?
gharchive/issue
2022-07-09T21:59:08
2025-04-01T04:35:51.095043
{ "authors": [ "DeutscheGabanna", "Minoo7" ], "repo": "shabegom/buttons", "url": "https://github.com/shabegom/buttons/issues/123", "license": "Unlicense", "license_type": "permissive", "license_source": "github-api" }
2568614336
Attributes and special attributes Karma calculation Karma flag needed for attributes to include karmaSpent. It should be a calculated value for attributes. Maybe I will just calculate each time from the value on the base to the new value.
gharchive/issue
2024-10-06T11:38:17
2025-04-01T04:35:51.154010
{ "authors": [ "shadows-eye" ], "repo": "shadows-eye/sr5-marketplace", "url": "https://github.com/shadows-eye/sr5-marketplace/issues/17", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1753618022
Controllers Hellow Reviewer, My name is @shafiuyushawu and this milestone fulfills the following requirement. For each URL you should create: A route. An action in the correct controller. A view file. View files should include only placeholders written in plain HTML, e.g: Here is a list of posts for a given user Make sure that you added code to handle only the 4 required URLs. @BenMukebo Thanks for the approval, additional index to user resource implement.
gharchive/pull-request
2023-06-12T21:24:06
2025-04-01T04:35:51.228068
{ "authors": [ "shafiuyushawu" ], "repo": "shafiuyushawu/blog-app", "url": "https://github.com/shafiuyushawu/blog-app/pull/4", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
709811312
Added 90's Game Style Retro Buttons Hi, I have added three 90's style retro buttons. Issue: I was not able to give them a border, and I couldn't figure out why. Any help will be appreciated. I wanted to create buttons like this: But I couldn't get a border. One final question: Where did you get the font? I prefer we use Google fonts. If you can find the font on Google fonts please import it. If it's not on there, please link me to look at the license of the font. Either way, we need to include a note about what font we are using for this button (You can see what I mean in Icon Buttons and the note about font awesome). It seems you didn't commit the button as well so I can't really see it on my side. Hey @shahednasser, I'll add the button under special buttons, and do the other tasks related to making a folder and adding the example. I'll also add the fill-btn class and see how that works out, thank you for suggestion there! Regarding the font, it is called 'Diary of an 8-bit Mage', and I got it from FontSpace. If you want, I can even replace it with a Google Font, but I was not able to find a similar one. Anyway, just let me know. On a second thought, I don't think it's necessary to add subtypes like in fill-btn as these buttons will not have different colors for each type (Continue will just be green, reset will just be red, etc...). So just add them each under one retro type. As for the font, can you please link me to where you got it? I need to check if we can use it and how we can inform whoever wants to use the font about how to use it. Awesome, because that was how my initial thought like... and that is why I followed the way Social Buttons were declared in the buttons-examples.js and used the variation property. Also, here's the font link: https://www.fontspace.com/diary-of-an-8-bit-mage-font-f28455 Yea I get what you mean. We still need to add it under "Special" buttons, but other than that you can do it the same way. As for the font it looks like we can use it. We still need to add a note about it. Maybe add it at the end of "How to Use" section for now similar to font awesome, but instead of a code block just add a link to it. We will probably look for another way to add these resources later, but let's do that for now. I'll do that. Also, I found a similar font style on Google Fonts. Here it is: https://fonts.google.com/specimen/Press+Start+2P?query=Press+ I'll make changes to the code! Great then let's use that one it will be easier. Import it in the _retro.less file if you can instead of directly to the website. That way if someone wants to use this button it will be ready in sbuttons.css without any additional imports for them. Great then let's use that one it will be easier. Import it in the _retro.less file if you can instead of directly to the website. That way if someone wants to use this button it will be ready in sbuttons.css without any additional imports for them. In this case, we don't need to add anything in the How to Use section. Where shall I add the Google Font link? Please add it in the less file for the retro button
gharchive/pull-request
2020-09-27T20:35:18
2025-04-01T04:35:51.235755
{ "authors": [ "anonyda", "shahednasser" ], "repo": "shahednasser/sbuttons", "url": "https://github.com/shahednasser/sbuttons/pull/275", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
934585025
README.MD FileNotFound on install When installing 0.1.8 either through pip3 or setup.py it fails with a FileNotFound on "README.MD" Workaround - rename README.md to README.MD. Probably the error might be in setup.py with this line with open("README.MD","r") as file: but I already have put the file name as README.MD , if you could paste the error message for better elaboration?. Thanks $ python3 setup.py install Traceback (most recent call last): File "setup.py", line 3, in with open("README.MD","r") as file: FileNotFoundError: [Errno 2] No such file or directory: 'README.MD' But there is a file at that point in the point tree called README.md (which I renamed to resolve the error)
gharchive/issue
2021-07-01T08:48:52
2025-04-01T04:35:51.250324
{ "authors": [ "RedBeardCleric", "shaikhsajid1111" ], "repo": "shaikhsajid1111/facebook_page_scraper", "url": "https://github.com/shaikhsajid1111/facebook_page_scraper/issues/9", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
906156409
Support playing multiple text tracks at once Have you read the Tutorials? yes Have you read the FAQ and checked for duplicate open issues? yes What version of Shaka Player are you using? v3.1.0 Please ask your question I am working on a project where i use Shaka player as a debug and quality verifier for dash files and have tried to figure out a way to show multiple textTracks of a dash manifest at once. I have looked trough the api documention and the code for some days but have not found a solution and understand that this maybe not something the player is built for but would like to get a tip if u think it should be possible to do in some way. It would be super nice if it could be done with your manifest parser and every thing to build uppon without to many own wierd tweeks and requests to sync it with the player. Thanks in advance :-) We do not currently have any way to play multiple text tracks at the same time, no. It's not too complex of a feature for us to add. It'd require some adjustments to the text engine, and probably a new method on the player to expose the behavior to developers, but not any huge architectural changes. I'll mark this as an enhancement request for now, but I wouldn't necessarily count on us getting to it right away. I don't think anyone else has requested anything like this. In fact, a few weeks ago I had to solve a bug where this was happening by accident! @joeyparrish can you review this request? Thanks! I don't have a solution for this, but I'm open to proposals.
gharchive/issue
2021-05-28T22:31:06
2025-04-01T04:35:51.255268
{ "authors": [ "Sweknas", "avelad", "joeyparrish", "theodab" ], "repo": "shaka-project/shaka-player", "url": "https://github.com/shaka-project/shaka-player/issues/3436", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2534842982
test: Don't fail tests when cast property can't be deleted On some real Cast devices, you can't delete window.cast. This is fine, and shouldn't cause the test to fail. If this happens, skip the test instead. The test, which mocks everything including the platform detection and the Cast API, is still valid on platforms where we can get away with those mocks. @shaka-bot test ce
gharchive/pull-request
2024-09-18T22:13:33
2025-04-01T04:35:51.256467
{ "authors": [ "avelad", "joeyparrish" ], "repo": "shaka-project/shaka-player", "url": "https://github.com/shaka-project/shaka-player/pull/7342", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2711558748
fix: Codec switch reload - apply boundaries correctly Fixes #7595 @shaka-bot test ce @shaka-bot test ce @shaka-bot test ce @shaka-bot test ce @shaka-bot test ce @shaka-bot test ce @shaka-bot test ce
gharchive/pull-request
2024-12-02T11:20:31
2025-04-01T04:35:51.258533
{ "authors": [ "tykus160" ], "repo": "shaka-project/shaka-player", "url": "https://github.com/shaka-project/shaka-player/pull/7700", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1678931504
🛑 Shakedown Social - Uploads is down In 8fb2686, Shakedown Social - Uploads (https://files.shakedown.social/accounts/avatars/109/357/389/718/265/462/original/dd80d2c95e48cfd2.jpeg) was down: HTTP code: 403 Response time: 418 ms Resolved: Shakedown Social - Uploads is back up in 0038586.
gharchive/issue
2023-04-21T18:32:58
2025-04-01T04:35:51.261094
{ "authors": [ "clifff" ], "repo": "shakedown-social/upptime", "url": "https://github.com/shakedown-social/upptime/issues/1263", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1682258504
🛑 Shakedown Social - Uploads is down In f46c347, Shakedown Social - Uploads (https://files.shakedown.social/accounts/avatars/109/357/389/718/265/462/original/dd80d2c95e48cfd2.jpeg) was down: HTTP code: 403 Response time: 598 ms Resolved: Shakedown Social - Uploads is back up in 9cfa4cc.
gharchive/issue
2023-04-25T01:08:36
2025-04-01T04:35:51.263615
{ "authors": [ "clifff" ], "repo": "shakedown-social/upptime", "url": "https://github.com/shakedown-social/upptime/issues/1786", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
267855591
FuckAccount() So..uh.. I messed up an equipment injection and accidentally injected 1,401,001,500 Raven Berets. This causes the FFBE app to crash if you look at equipment under the items menu or if you tap on the head gear equipment box on the unit screen. If FuckAccount() is used to Fuck an account then this would be a method. https://github.com/shalzuth/BraveHaxvius/commit/ce6e863c01e0fe2f3afcf1e6ab1676e1a1f9413d Can sell those items, but might need to in segments (not all at once). Fixed in latest.
gharchive/issue
2017-10-24T00:06:55
2025-04-01T04:35:51.269491
{ "authors": [ "WithinGDR", "shalzuth" ], "repo": "shalzuth/BraveHaxvius", "url": "https://github.com/shalzuth/BraveHaxvius/issues/20", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
398646703
Sync & Download fail in Develop Tools with "HttpError: Not Found" Visual Studio Code Version : [ 1.30.2 ] Code Settings Sync Version : [ 3.2.4 ] Operating System : [ WINDOWS 10 ] Occurs On: [ Sync, Download ] Proxy Enabled: [ No ] Fresh install, Windows 10, recent VSC (1.30.2), recent Sync (3.2.4), settings.json -> "sync.gist": "https://github.com/settings/tokens/<my_id>" syncLocalSettings.json -> "token": "<my_token>" Neither upload nor download works. Fails with: Sync : Error Logged In Console (Help menu > Toggle Developer Tools). Developer Tools log: [Extension Host] HttpError: Not Found at module.exports.e.exports.i.fetch.then.n.text.then.n (C:\Users\obiol\.vscode\extensions\shan.code-settings-sync-3.2.4\out\extension.js:137:45419) at <anonymous> at process._tickCallback (internal/process/next_tick.js:188:7) t.log @ console.ts:134 t._logExtensionHostMessage @ extensionHost.ts:453 (anonymous) @ extensionHost.ts:244 emitTwo @ events.js:126 emit @ events.js:214 emit @ internal/child_process.js:772 _combinedTickCallback @ internal/process/next_tick.js:141 _tickCallback @ internal/process/next_tick.js:180 I am having the exact the same error but I did not mistake one for the other. I am using vs code v1.33.1 and sync extension v3.2.9 Linux Mint 19 Tessa I am also getting this same error on Linux Mint 19. Works fine on my Win10 laptop. Have same issue, token and gist id are correct. So I was able to figure it out. User issue. I was entering the entire gist URL vs just the gist ID. Once I used only the gist ID it worked perfectly. I was entering the entire gist URL vs just the gist ID I wish i could make the documentation or WIKI more clearer for the guys. Most of the issues are raied by miss understanding. I am always open for PR that improve the readme. Also Public Edit in Github Wiki is enabled. Feel free to improve that. where you think there is confusion. I copied github token but not GistId while creating my sync settings. So i lost my GistId and am able to retrieve it by following below steps. Go to your github account Click on your profile and go to 'Your gists' Click on cloud settings, it open several settings Go to settings.json Find and copy value of "sync.gist", this is your GistId You can now download all your uploaded settings Below are few screen shots: i'm getting this problem too, just remove "sync.gist" on settings.json then re-upload again using shift+alt+u and the extention will create a gist for you automatically. I'm getting the same error but my settings.json doesn't have a sync.gist - can I just add it manually? OK solved my problem: my User and Workspace settings had different GistI Ds even though I reset settings several times. I fixed this by recreating my personal access token and updating my "master" settings then uploading. On the target machine I reset everything and started from scratch then checked that my Gist IDs were both set to the new one ==> working!!!! 😄
gharchive/issue
2019-01-13T11:09:48
2025-04-01T04:35:51.296111
{ "authors": [ "Dalot", "aamsur", "asmyshlyaev177", "h3d0", "jf781", "pauloneill", "ravishankarkota2017", "shanalikhan" ], "repo": "shanalikhan/code-settings-sync", "url": "https://github.com/shanalikhan/code-settings-sync/issues/748", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
370518070
How did you create the mask? Two masks which you are using for stop sign how did you create them? Furthermore, here you are considering only the stop sign so only stop sign masks are used but what if I want to use this method for other datasets then I will have to create the masks for specific objects in images? You can just use any stop sign image (e.g. https://en.wikipedia.org/wiki/File:Stop_sign(standard).svg) and use pixel value thresholding to create a mask. We use the mask to constrain the shape and add a regularization term in the loss function to constrain the color. For other datasets you can use the same method, but some object classes may need more complex constraints than only shape and color. Thank you.
gharchive/issue
2018-10-16T09:24:03
2025-04-01T04:35:51.385971
{ "authors": [ "sadafgulshad1", "shangtse" ], "repo": "shangtse/robust-physical-attack", "url": "https://github.com/shangtse/robust-physical-attack/issues/5", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
781844943
isONcorrect published doi The following instructions are for pull requests related to tool entries. If your pull request is about something else they can be ignored. Please describe your pull request below and fill out the check list: Add your description here: Please check whether you could check following boxes. Else it will break the build of the database! [ ] Only (manually) edited long_read_tools_master.csv [ ] Tool/pipeline names were added using only "[a-z],{a-}, [0-9], _, -" characters (e.g. no spaces, slashes or tabs [ ] Add relevant multiple publications separated by ; [ ] Add "NOT-MAINTAINED:: " or "DEPRECIATED:: " to the beginning of the description column of long_read_tools_master.csv [ ] Completed all columns that needs logical values(i.e. TRUE, FALSE) [ ] Have at least one functional category marked as TRUE per tool [ ] Add multiple Programming_Language separated by commas [ ] Check the link of the Source you are noting points to the root of the repository only [ ] Either have peer-reviewed journal DOI or preprint DOI. Leave reference as blank for unpublished tools [ ] Run build_lrs_db.R (optional) Thanks @QGouil .... This information is now added outside the pull request! :) Thanks @QGouil .... This information is now added outside the pull request! :)
gharchive/pull-request
2021-01-08T05:00:23
2025-04-01T04:35:51.390883
{ "authors": [ "QGouil", "shaniAmare" ], "repo": "shaniAmare/long_read_tools", "url": "https://github.com/shaniAmare/long_read_tools/pull/47", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
340085661
How do you create the mention_type_test.txt? I tried to run ./run.sh but it didn't create mention_type_test.txt I get this: Traceback (most recent call last): File "code/Evaluation/convertPredictionToJson.py", line 13, in with open(typeMapFile) as typeF: IOError: [Errno 2] No such file or directory: 'data/intermediate/KBP/rm/type.txt' Traceback (most recent call last): File "code/Evaluation/tune_threshold.py", line 66, in ground_truth = load_labels(indir + '/mention_type_test.txt') File "/workspace/png/CoType-master/code/Evaluation/evaluation.py", line 17, in load_labels with open(file_name) as f: IOError: [Errno 2] No such file or directory: 'data/intermediate/KBP/rm/mention_type_test.txt' It seems like you didn't generate intermediate files properly... They should be generated in the feature generation pipeline.
gharchive/issue
2018-07-11T04:15:08
2025-04-01T04:35:51.393845
{ "authors": [ "33Peng33", "ellenmellon" ], "repo": "shanzhenren/CoType", "url": "https://github.com/shanzhenren/CoType/issues/8", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2458968613
[bug] 修好了但没完全修好 部分图片帖子里面无法预览 修好了
gharchive/issue
2024-08-10T07:52:04
2025-04-01T04:35:51.412082
{ "authors": [ "minyue-yu", "share121" ], "repo": "share121/inter-knot", "url": "https://github.com/share121/inter-knot/issues/734", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
358658657
Collision.RayIntersectsRay treats Rays as lines When finding the intersection point: float s = dets / denominator; float t = dett / denominator; //The points of intersection. Vector3 point1 = ray1.Position + (s * ray1.Direction); Vector3 point2 = ray2.Position + (t * ray2.Direction); both s and t should be non-negative numbers because rays have a starting point. I had to add this code after computing s and t. if (s < 0 || t < 0) { point = Vector3.Zero; return false; } Also, I added the else in this section when the Rays are parallel but not on top of each other: //Lines are parallel. if (MathUtil.IsZero(denominator)) { //Lines are parallel and on top of each other. if (MathUtil.NearEqual(ray2.Position.X, ray1.Position.X) && MathUtil.NearEqual(ray2.Position.Y, ray1.Position.Y) && MathUtil.NearEqual(ray2.Position.Z, ray1.Position.Z)) { point = Vector3.Zero; return true; } else { point = Vector3.Zero; return false; } } Submitted PR for this fix. Could you eyeball the changes to make sure they are correct. I've done my once over and seems to be ok. #1139 - contains your change also.
gharchive/issue
2018-09-10T14:49:39
2025-04-01T04:35:51.458305
{ "authors": [ "h1cks", "mludlum" ], "repo": "sharpdx/SharpDX", "url": "https://github.com/sharpdx/SharpDX/issues/1077", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1708633126
Re-enable autovacuum Also fix the table name for this part of the process. This is a quick fix for tables with no storage parameters but does not fix the issue if autovacuum is enabled but other storage parameters are indeed set. Fixes #82 We could potentially do something like this in the case where storage_parameters are not empty: "ALTER TABLE #{shadow_table} RESET (autovacuum_enabled, toast.autovacuum_enabled); ALTER TABLE #{shadow_table} SET (#{primary_table_storage_parameters});" That would reset autovacuum to the default, then set whatever storage parameters we have saved. @shayonj What do you think? Yeah, I think that makes sense. I will follow up with a spec. We can get this live today.
gharchive/pull-request
2023-05-13T14:52:37
2025-04-01T04:35:51.483233
{ "authors": [ "jfrost", "shayonj" ], "repo": "shayonj/pg-osc", "url": "https://github.com/shayonj/pg-osc/pull/85", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
178219686
Feature/alpine java Needed this for one of my projects so added alpine java with oracle jdk 8 here. Hi @whithajess, thanks for you PR. Alpine support is planned for the third release from now (but it may change depending on KM releases). So I won't merge in master for now. But we have a alpine branch (https://github.com/sheepkiller/kafka-manager-docker/tree/alpine) and an automatic build if you use alpine as tag. It's a second class citizen image for now, I'll review you change later today (timezone: CEST) and give you a more detailed feedback. Ah ok didn't see the branch its similar just without java envs which aren't required and a README update. No problem. The alpine branch is just a test bed for now, mostly because I'd like to drop jdk from the final image. I don't know for now if I use a kind of noop build-only layer which depends on jdk and the install jre, or if I maintain unofficial builds of kafka-manager. BTW, I was unable to lock some time today to review your changes and give you the decent feedback you deserve.
gharchive/pull-request
2016-09-21T01:07:02
2025-04-01T04:35:51.498354
{ "authors": [ "sheepkiller", "whithajess" ], "repo": "sheepkiller/kafka-manager-docker", "url": "https://github.com/sheepkiller/kafka-manager-docker/pull/14", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
187543444
Get pipe tests running on Windows. I used the FIND command which ships with Windows as something to pass data through. The test checks that a POSIX-like find(1) exists and skips the tests if this condition is true because Windows-provided FIND is incompatible with find(1). @nfischer I find ss64 very useful. See FIND. It’s basically like grep(1) (and even the process description in Task Manager mentions grep on modern Windows). An interesting trait of FIND is that it relies on arguments being quoted a specific way, probably for historical reasons. An article which I didn’t fully read but touches on some of those things is “Everyone quotes command line arguments the wrong way”. Basically, on Windows, programs parse the commandline themselves instead of being passed a vector of arguments. Here’s an example: C:\Users\ohnob>ECHO hi | FIND "hi" hi C:\Users\ohnob>ECHO hi | FIND "hey" C:\Users\ohnob>ECHO hi | FIND hey FIND: Parameter format not correct FIND requires its first argument to be passed between double quotes and refuses to run otherwise. It also appears to only support exact substring matching—it doesn’t appear to have a pattern/glob/regex matching option. But for simply testing that piping in shelljs works, it’s sufficient and provided on all Windows machines. However, an alternative approach would be to do something like .exec(common.nodeBinPath + ' -e "process.stdin.pipe(process.stdout)"') except with some logic that would actually mutate the stream. By just shelling out to node instead of a unix-specific built-in, we could get rid of the whole if (process.platform). Want me to rewrite this PR using this idea instead? ^^ Cool. I think I’ve addressed all of your outstanding review comments which was just removing the // Windows-specific comment. Please let me know if I missed anything, otherwise I assume you’re waiting for others to look at this and give feedback (if so, sorry for this bugspam xD). @binki I'll give it another look a bit later and make sure everything is good, and then merge. Would you be able to add a brief explanation of the semantics of windows find in a comment in the code? Something like: Windows `find` is semantically similar to Unix `grep`. It requires that its first argument be surrounded by double quotes, and it only matches literal strings (not regex). Add in whatever you think is useful in making the test-case more readable. Thanks! Oh, I just noticed #525 now. That sounds like a much cleaner way of achieving the goal of this PR in the end. @binki I just released the new shx. Would you be interested in refactoring this to use shx grep instead of find? Then we can get rid of the Windows-specific and unix-specific conditionals. That would be even more readable :smile: Maybe I should rebase to get that ugly FIND stuff out of the history? Maybe I should rebase to get that ugly FIND stuff out of the history? I wouldn't worry about it. We squash commits upon merge, so that should take care of it. LGTM once Travis passes Fixes #525 Someone needs to retrigger the travis build because it “errored” rather than “failed” due to badly timed Mac OS X timeouts ;-).
gharchive/pull-request
2016-11-06T05:32:13
2025-04-01T04:35:51.514123
{ "authors": [ "binki", "nfischer" ], "repo": "shelljs/shelljs", "url": "https://github.com/shelljs/shelljs/pull/550", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
285756191
Enable social sharing Enable twitter, facebook or another social platform to share blog post. By chance, is this being worked on? I really like this theme, but would really love to be able to share blog posts to facebook, twitter and linkedin Thank you.
gharchive/issue
2018-01-03T17:33:35
2025-04-01T04:35:51.516321
{ "authors": [ "alcazar90", "chrisselig" ], "repo": "shenoybr/hugo-goa", "url": "https://github.com/shenoybr/hugo-goa/issues/35", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
503237508
switch to babel loader #7 This PR switch awesome-typescript-loader to babel-loader. But this one doesn't resolve an issue with reducing bundle size. I played with settings and different babel setup, but with no success. From Webpack analyzer report(attached) you can see that semantic-ui-react bundled with all modules(index+88) and I have no idea how to change that. Hi @ivan-nikolaievskyi , thanks for taking time to contribute this! 👍 Do you mind me leaving this PR without merging for sometime? I'd like to look into how to tweak these changes to reduce the bundle size 😊 This PR can be base for your improvements, if you find the way to reduce bundle size. Otherwise there is no sense to merge it at all, because it doesn't add improvements to build process
gharchive/pull-request
2019-10-07T05:18:47
2025-04-01T04:35:51.636647
{ "authors": [ "ivan-nikolaievskyi", "sheshbabu", "vanya829" ], "repo": "sheshbabu/freshlytics", "url": "https://github.com/sheshbabu/freshlytics/pull/16", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
426045865
Add textMapIsSparqlQuery URL parameter to shex-simple This makes it possible to link to a version of shex-simple where users can enter a SPARQL query directly, without having to know the SPARQL extension syntax. Part of our attempts to make the tool a bit easier to use for Wikidata editors not overly familiar with ShEx (T218886). Looks good to me 👍 For posterity's sake, it would be cool to have a stock URL which tested this behavior attached to this PR. Example URL: https://tools.wmflabs.org/shex-simple/wikidata/packages/shex-webapp/doc/shex-simple.html?data=Endpoint: https://query.wikidata.org/sparql&manifest=[]&schemaURL=https%3A%2F%2Fwikidata-shex.wmflabs.org%2Fwiki%2FSpecial%3ASchemaText%2FO2&hideData&textMapIsSparqlQuery
gharchive/pull-request
2019-03-27T15:58:57
2025-04-01T04:35:51.639115
{ "authors": [ "ericprud", "lucaswerkmeister", "micgro42" ], "repo": "shexSpec/shex.js", "url": "https://github.com/shexSpec/shex.js/pull/38", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1926663840
Request to Add Number System Conversion Topics These topics could cover a range of conversions such as binary to decimal, hexadecimal to binary, or any other common conversions related to number systems. I believe that adding these topics would greatly benefit the repository's users who are looking for information and guidance on these fundamental concepts. Could you assign this issue to me? @ZRX-SIGMA I will do this Bro , Let him say, I opened this issue.. @ZRX-SIGMA sorry i thought u were the owner @ZRX-SIGMA sorry i thought u were the owner I think we can share this issue, I can write for python, u can write in any other lang.. i can do C, C++, JS Yupp, That will do. working on it @shhossain #656 done for C, CPP, JS, TS, Lua, Java @shhossain sir review this please @Sarthak950 check your PR. And I am assigning this issue to both @Sarthak950 @ZRX-SIGMA. @shhossain done @ZRX-SIGMA are u working on the python one sir Yess @shhossain #660 check this out @Sarthak950 check my review. i did see anything there @shhossain @shhossain i did not see any thing there @Sarthak950 check #660 @shhossain yes i checked but i am not getting anything @Sarthak950 I've fixed the problem just now. don't need to do anything.
gharchive/issue
2023-10-04T17:21:24
2025-04-01T04:35:51.647024
{ "authors": [ "Sarthak950", "ZRX-SIGMA", "shhossain" ], "repo": "shhossain/computer_science", "url": "https://github.com/shhossain/computer_science/issues/653", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
170188176
Tweaking Fleetup interaction I have a few questions on tweaking. FleetUp doesnt seem to post new operations when posted only upcoming ones. What is the interval for "upcoming" to post and does it re-ping a few times? How can I modify this so say upcoming is 12hrs instead of 1 or whatever it is now. What is the poller interval for notifications because I have a bot on Slack that seems to get notifications 5-6 minutes faster than this, is this adjustable? If so, where? -- Any plans for expansion into posting updated/new FleetUP doctrine/fitting changes? Or responding to a command to spit a link to a fitting? Fleetup right now checks every 30 minutes for new operations and every 2 minutes for upcoming (within 30 minutes). Notifications are based off the cache set by ccp. It's impossible for you to be getting anything faster as the bot checks as soon as the api updates. I could, depends on the demand as I'm not sure how widely used the fleetup features are being used. fleetup is widely used by alot of alliances. its the must have tool for medium-Large sizes corps/alliances
gharchive/issue
2016-08-09T14:57:15
2025-04-01T04:35:51.652733
{ "authors": [ "CokkocZateki", "shibdib", "warriorsoul15" ], "repo": "shibdib/Dramiel", "url": "https://github.com/shibdib/Dramiel/issues/20", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
238020053
银河 web 很快不能用了.... 尊敬的投资者: 为进一步提升交易安全,我公司于2017年6月30日16时起停止提供WEB交易系统,您将无法通过该系统进行证券交易。本次暂停仅针对WEB交易系统,我公司提供的其它交易方式不受影响,为避免影响您的正常交易,建议请您使用我公司提供的其它交易方式(下载地址:http://www.chinastock.com.cn/yhwz/service/download.shtml)。 若您对此有任何意见或建议,请您与您所在当地营业部联系或致电95551。 中国银河证券股份有限公司提醒您:股市有风险,入市需谨慎。 是的,后面会完善下客户端
gharchive/issue
2017-06-23T01:59:47
2025-04-01T04:35:51.654451
{ "authors": [ "4ever911", "shidenggui" ], "repo": "shidenggui/easytrader", "url": "https://github.com/shidenggui/easytrader/issues/210", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
184252448
libOpenCvSharpExtern.so: undefined symbol: _ZTIN2cv16ParallelLoopBodyE OpenCvSharpExtern was compiled successfully, but when I try to use it on a simple example, I get these errors (with MONO_LOG_LEVEL=debug) with the mentions of an undefined symbol. ldd libOpenCvSharpExtern.so doesn't show any links to OpenCV. It looks like a relapse of #85 after the migration of 3.x branch to master. 🤔 please try the following https://ubuntuforums.org/showthread.php?t=2109967 https://github.com/kyamagu/mexopencv/issues/123 Thanks for the answer, I'll return to this in a month or two. Seeing this also on NixOS; more details at https://github.com/shimat/opencvsharp/issues/273#issuecomment-277501280
gharchive/issue
2016-10-20T14:53:51
2025-04-01T04:35:51.658729
{ "authors": [ "Hodapp87", "fedorkov", "shimat" ], "repo": "shimat/opencvsharp", "url": "https://github.com/shimat/opencvsharp/issues/262", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
928748885
Change to development environment to remote container create dockerfile for development on /.devcontainer add poetry in dev container not use virtualenv
gharchive/issue
2021-06-24T00:45:34
2025-04-01T04:35:51.659890
{ "authors": [ "shin-hama" ], "repo": "shin-hama/JunkNoteAPI", "url": "https://github.com/shin-hama/JunkNoteAPI/issues/24", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1242042687
useStore hook returned on Store ref Including a manner by which to consume the resulting Store on the ref allows easy mocking in tests. Usage: Tests: Before After Confirmed after changing to the get trap override that it works:
gharchive/pull-request
2022-05-19T16:41:31
2025-04-01T04:35:51.668374
{ "authors": [ "rkyle35242" ], "repo": "shipt/osmosis", "url": "https://github.com/shipt/osmosis/pull/101", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
56479209
[Feature request] Add role related operations support in cli Would be similar to what shipyard currently has for account related operations roles show roles add-role add role delete-role delete role This is no longer valid in v3. We do not have a custom CLI.
gharchive/issue
2015-02-04T04:36:20
2025-04-01T04:35:51.680348
{ "authors": [ "ehazlett", "isymbo" ], "repo": "shipyard/shipyard", "url": "https://github.com/shipyard/shipyard/issues/415", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2292060610
Questions about sd-scripts and diffusers Hi shirayu, Your work on diffusion models is really impressive! I came across your example_lora_training repo (https://github.com/shirayu/example_lora_training) and as someone just starting to learn about diffusion, I've been working with sd-scripts and diffusers. I'd really appreciate if you could help me with a couple of questions. Do you use regularization images in your training (https://github.com/shirayu/example_lora_training)? I recently tried training a LoRA model using the train_text_to_image_lora.py script from diffusers, with the image captions in a metadata file. However, the results were quite poor. On the other hand, when I tried using the LoRA in the Kohya_ss Web GUI with the image captions txt, I was able to generate my target images successfully. I'm not sure why there's such a difference since I didn't use any regularization images in either case. Could you explain the differences among the train_text_to_image_lora.py, train_dreambooth_lora.py scripts and sd-scripts? Is the Kohya_ss LoRA button equivalent to using dreambooth+caption, but without the need for regularization images (since I didn't input any reg_img)? Also, is it possible to use the train_dreambooth_lora.py script from diffusers without regularization images? Thanks for your help in advance. Hi! The repository was created for the first LoRA experiments and is no longer maintained. I think I probably used regularized images. I have not done learning LoRA with diffusers at all recently. Therefore, I am not able to give proper advice. This is just my personal opinion, but I think sd-scripts is better maintained and you should use it. I have not used dreambooth at all recently, so I am afraid I cannot give you proper advice. Finally, I think that the presence or absence of regularized images has little effect on learning, and that the other parameters are much more important. In fact, I do not use regularized images at all these days. If you need further advice, I recommend that you get it from the respective code community. Thank you.
gharchive/issue
2024-05-13T07:35:01
2025-04-01T04:35:51.686387
{ "authors": [ "elichan5168", "shirayu" ], "repo": "shirayu/tohoku_sd_lora", "url": "https://github.com/shirayu/tohoku_sd_lora/issues/21", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1768077615
[1.19.2 Forge] Crash Ticking World Hi! In my singleplayer world, using the June release of the mod, I get this crash: https://pastebin.com/47Svhjvc Forge version is 43.2.9. It happens with or without shaders. Thanks for your time! And sorry for hitting enter too early lol. what happened to cause this crash? the only way i've had this happen is from an explosion and even then it was super rare I was just flying through my testing world. I assume it was an item on the ground due to faulty world gen. If it helps, I was near a Choice theorem's Overhauled Villages village in a jungle. We also use Terralith for world gen. It happened twice while I was in that area, but did not occur when I reverted to the previous loot beams version. On Thu, Jun 22, 2023, 6:02 AM amo @.***> wrote: what happened to cause this crash? the only way i've had this happen is from an explosion and even then it was super rare — Reply to this email directly, view it on GitHub https://github.com/shiroroku/LootBeams/issues/26#issuecomment-1602441664, or unsubscribe https://github.com/notifications/unsubscribe-auth/AE7XE5T77T23TPIZGQU6HCLXMQQ3PANCNFSM6AAAAAAZPCDK7Y . You are receiving this because you authored the thread.Message ID: @.***> I can try to replicate again and get you a latest.log if you think it might help? On Thu, Jun 22, 2023, 10:30 AM amo @.***> wrote: Yeah previous version didn't have item caching. Looking into it but it seems funky to fix. — Reply to this email directly, view it on GitHub https://github.com/shiroroku/LootBeams/issues/26#issuecomment-1602847673, or unsubscribe https://github.com/notifications/unsubscribe-auth/AE7XE5TPG7AHE4GH4UPJSZLXMRQKHANCNFSM6AAAAAAZPCDK7Y . You are receiving this because you authored the thread.Message ID: @.***>
gharchive/issue
2023-06-21T17:23:06
2025-04-01T04:35:51.694208
{ "authors": [ "amoooooo", "katubug" ], "repo": "shiroroku/LootBeams", "url": "https://github.com/shiroroku/LootBeams/issues/26", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1874747589
Update snmp@8.4 Build snmp@8.4 :beers: @BrewTestBot has triggered a merge.
gharchive/pull-request
2023-08-31T04:03:32
2025-04-01T04:35:51.702786
{ "authors": [ "shivammathur" ], "repo": "shivammathur/homebrew-extensions", "url": "https://github.com/shivammathur/homebrew-extensions/pull/3129", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2501724493
Update php@8.4 Build php@8.4 :beers: @BrewTestBot has triggered a merge.
gharchive/pull-request
2024-09-03T00:26:59
2025-04-01T04:35:51.704110
{ "authors": [ "shivammathur" ], "repo": "shivammathur/homebrew-php", "url": "https://github.com/shivammathur/homebrew-php/pull/3074", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1003352421
Update php@8.1 Build php@8.1 :beers: @BrewTestBot has triggered a merge.
gharchive/pull-request
2021-09-21T21:55:15
2025-04-01T04:35:51.705150
{ "authors": [ "shivammathur" ], "repo": "shivammathur/homebrew-php", "url": "https://github.com/shivammathur/homebrew-php/pull/515", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1023225539
Update php@8.1 Build php@8.1 :beers: @BrewTestBot has triggered a merge.
gharchive/pull-request
2021-10-12T00:05:38
2025-04-01T04:35:51.706175
{ "authors": [ "shivammathur" ], "repo": "shivammathur/homebrew-php", "url": "https://github.com/shivammathur/homebrew-php/pull/559", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1272891336
Update php@8.2 Build php@8.2 :beers: @BrewTestBot has triggered a merge.
gharchive/pull-request
2022-06-16T00:08:27
2025-04-01T04:35:51.707177
{ "authors": [ "shivammathur" ], "repo": "shivammathur/homebrew-php", "url": "https://github.com/shivammathur/homebrew-php/pull/923", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
544770319
add card view to datatable give an option to show a list of cards instead of a datatable. Questions: Would editing still be possible? what functional differences would there be? How would sorting work?
gharchive/issue
2020-01-02T22:15:35
2025-04-01T04:35:51.720342
{ "authors": [ "shliachtx" ], "repo": "shliachtx/lwc-listview", "url": "https://github.com/shliachtx/lwc-listview/issues/9", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2600473878
Feature request idea: automatic retry upon error Seeing errors in the log that state "Scraper error: error sending request for url " and the download stops after that point for the magazine issue. Restarting the download fixes the issue (suspect that rate limiting could fix this). An idea would be to do a retry or N retries of that issue after waiting a bit. Otherwise, loving the software! Implemented in v0.3.4. Configurable with -r/--download-attempts command line parameter.
gharchive/issue
2024-10-20T13:45:53
2025-04-01T04:35:51.721701
{ "authors": [ "EvoDyn", "shloop" ], "repo": "shloop/google-book-scraper", "url": "https://github.com/shloop/google-book-scraper/issues/5", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1709162120
Unit testing Description Describe your changes in detail. Motivation and context Why is this change required? What problem does it solve? If it fixes an open issue, please link to the issue here (if you write fixes #num or closes #num, the issue will be automatically closed when the pull is accepted.) How has this been tested? Please describe in detail how you tested your changes. Include details of your testing environment, and the tests you ran to see how your change affects other areas of the code, etc. Screenshots (if appropriate) Types of changes What types of changes does your code introduce? Put an x in all the boxes that apply: [ ] Bug fix (non-breaking change which fixes an issue) [ ] New feature (non-breaking change which adds functionality) [ ] Breaking change (fix or feature that would cause existing functionality to change) Checklist: Go over all the following points, and put an x in all the boxes that apply. Please, please, please, don't send your pull request until all of the boxes are ticked. Once your pull request is created, it will trigger a build on our continuous integration server to make sure your tests and code style pass. [ ] I have read the CONTRIBUTING document. [ ] My pull request addresses exactly one patch/feature. [ ] I have created a branch for this patch/feature. [ ] Each individual commit in the pull request is meaningful. [ ] If my change requires a change to the documentation, I have updated it accordingly. If you're unsure about any of these, don't hesitate to ask. We're here to help! :tada: This PR is included in version 3.8.1 :tada: The release is available on GitHub release Your semantic-release bot :package::rocket:
gharchive/pull-request
2023-05-15T01:30:42
2025-04-01T04:35:51.728599
{ "authors": [ "shmolf" ], "repo": "shmolf/noted", "url": "https://github.com/shmolf/noted/pull/49", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1090346153
Bad CPU type in executable I'm setting up a new MBP 14'' M1 Pro with macOS 12.0.1 and ran into some problems getting this to work. [08:49:56.928] NightShift[Keyword] Processing complete [08:49:56.930] NightShift[Keyword] Passing output '' to Run Script [08:49:57.011] ERROR: NightShift[Run Script] xattr: [Errno 13] Permission denied: './nightlight' /bin/bash: line 5: ./nightlight: Bad CPU type in executable /bin/bash: line 7: ./nightlight: Bad CPU type in executable [08:49:57.017] NightShift[Run Script] Processing complete [08:49:57.018] NightShift[Run Script] Passing output '' to Run Script I manually changed the permissions and then had to install Rosetta 2 to solve the remaining CPU type error. IIUC some apps prompt for that, which did not happen here. This is more of a heads-up, it might be worth to include a note in the README for the less tech-savvy users. Thank you for maintaining this, I use it every single day! So what you suggest basically boils down to adding a note in the README about maybe changing some permissions if it doesn't work out of the box? And if running on Apple Silicon, also having to install Rosseta 2 if it hasn't been done by the user beforehand? I can do that. Btw. Thank you for maintaining this, I use it every single day! That's great to hear, nice that you find it useful!
gharchive/issue
2021-12-29T08:16:25
2025-04-01T04:35:51.731063
{ "authors": [ "kgrhartlage", "shmulvad" ], "repo": "shmulvad/alfred-nightshift", "url": "https://github.com/shmulvad/alfred-nightshift/issues/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1913983749
Order of Rules Affects Parsing Describe the bug BNF grammar should be independent of the order of the rules as long as it is otherwise well-formed. To Reproduce Take the dna_left_recursive test and reverse the order of the rules. The BNF parser can no longer successfully parse the input string. #[test] fn dna_left_recursive() { let grammar: Grammar = "<base> ::= 'A' | 'C' | 'G' | 'T' <dna> ::= <base> | <dna> <base>" .parse() .unwrap(); let input = "GATTACA"; let parses: Vec<_> = grammar.parse_input(input).map(|a| a.to_string()).collect(); assert_snapshot!(parses.join("\n")); } Thanks for bringing this up! I am not sure yet, but I have a guess. The current bnf::Grammar::parse doesn't have a way to designate the "starting" term. So as is, it assumes the first rule begins with the "starting" term. In this example, this means the "starting" term is <base>, which cannot parse "GATTACA". Sorry if this turns out to be the reason! I meant to mark this strange implicit assumption in the API documentation, but I must have forgotten. That's what I figured was happening, but I figured I'd file a bug so you have a nice reminder to add it to the documentation or whatever you end up doing. :) Thanks!
gharchive/issue
2023-09-26T17:33:46
2025-04-01T04:35:51.734214
{ "authors": [ "CrockAgile", "lambdaknight" ], "repo": "shnewto/bnf", "url": "https://github.com/shnewto/bnf/issues/137", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
293523764
CON-2888: Implemented a language fallback if there is no language file available for the designated language, the en-US file will now be used removed the en-AU language file Coverage decreased (-0.001%) to 4.071% when pulling faf166bc35972cf50069bc5f0d5794472ccdfb9d on CON-2888-translations-should-fallback-to into c7041e165243807b730ac43ba7ee599f8c4323e4 on master.
gharchive/pull-request
2018-02-01T13:08:41
2025-04-01T04:35:51.761054
{ "authors": [ "coveralls", "kernelkiller" ], "repo": "shopgate/theme-ios11", "url": "https://github.com/shopgate/theme-ios11/pull/66", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1645085287
Demo - change default preset to node-server Related to: #90 More details: https://github.com/shopware/frontends/pull/90#issuecomment-1487529143 Acceptance criteria app is building by default with node-server preset when ENV variable is provided with another preset it's used for build Sidenote idea: maybe it could be configured inside nuxt.config.ts file with the default values if env is not presented? also related to #77
gharchive/issue
2023-03-29T05:57:15
2025-04-01T04:35:51.783712
{ "authors": [ "patzick" ], "repo": "shopware/frontends", "url": "https://github.com/shopware/frontends/issues/92", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
489688586
Entity customer is missing discardChanges method When editing a customer in the shopware administration panel and clicking on Cancel, the console prints: [Discard-detail-page-changes Mixin] Could not discard changes for entity with name "customer". and of course the changes that I made to the customer are permanent. Hello @itmuckel and thank you for reporting the bug. I created an issue ticket for this. Here you can see the actual status: NEXT-4842
gharchive/issue
2019-09-05T11:18:11
2025-04-01T04:35:51.785444
{ "authors": [ "itmuckel", "taltholtmann" ], "repo": "shopware/platform", "url": "https://github.com/shopware/platform/issues/166", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
338847687
Allow 0 as ID in state of Order API 1. Why is this change necessary? The stateId is necessary on create of an order. This should not be requiered. Also see the ISSUE SW-21511 I also fixed the typo that states that shipping state is missing instead of billing state The user can register without a stateId/state, why should the API disallow it. 2. What does this change do, exactly? It checks if the stateId is 0 - if so it does not check if the state exists in the list 3. Describe each step to reproduce the issue or behaviour. Create order with stateId 0 fails 4. Please link to the relevant issues (if any). https://issues.shopware.com/issues/SW-21511 5. Which documentation changes (if any) need to be made because of this PR? 6. Checklist [x] I have written tests and verified that they fail without my change [ ] I have squashed any insignificant commits [ ] This change has comments for package types, values, functions, and non-obvious lines of code [x] I have read the contribution requirements and fulfil them. This PR was closed due to inactivity. If this change is still important to you, feel free to create a new pull request. For more information about our contribution guidelines, see https://developers.shopware.com/contributing/contribution-guideline/.
gharchive/pull-request
2018-07-06T08:19:26
2025-04-01T04:35:51.789821
{ "authors": [ "Isengo1989", "shopwareBot" ], "repo": "shopware/shopware", "url": "https://github.com/shopware/shopware/pull/1703", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
150881795
introduce variables for column count in masonry emotions Make column count in masonry emotion worlds configurable, so you can define the column count by setting a LESS variable in your own theme. No BC breaks as the default values are the old ones. see also: http://forum.shopware.com/discussion/34740/override-einkaufswelt-spalten-anpassen-fuer-masonry-in-emotion-less @Phil23 might have a look? Hi @SvenHerrmann, thank you for your commit. The change is a good idea, but we are refactoring the emotion worlds for an upcoming release, so this is no longer relevant. I think the new features will be merged to the 5.2 branch in the next weeks. Stay tuned!
gharchive/pull-request
2016-04-25T14:35:24
2025-04-01T04:35:51.791966
{ "authors": [ "Phil23", "SvenHerrmann", "bcremer" ], "repo": "shopware/shopware", "url": "https://github.com/shopware/shopware/pull/523", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2563120876
Detect offline shotover nodes After introducing ShotoverNodeState to ShotoverNode in https://github.com/shotover/shotover-proxy/pull/1758, we should add a task to detect down shotover nodes and set ShotoverNodeState accordingly. This PR adds a background task check_shotover_peers looping over peer shotover nodes and trying to open a TCP connection to each peer shotover node. If the connection cannot be established within connect_timeout_ms, the peer node is marked as down. connect_timeout_ms is the same configuration used when creating a connection to a destination kafka broker. Each check is delayed for (check_shotover_peers_delay_ms + random(-check_shotover_peers_delay_ms/10, check_shotover_peers_delay_ms/10)) before moving to the next peer shotover node. start_shotover_peers_check is called when the instance of KafkaSinkClusterBuilder is being created and hence is called exactly once. check_shotover_peers is be invoked at all if there's no peer shotover node (i.e., there's only 1 shotover node in the cluster) check_shotover_peers is restarted if the creation of random number generator fails. The next PR will change metadata rewrites to exclude down shotover nodes. The regression benchmark encode_system.local_result_v5_no_compression is for Cassandra and hence a noise.
gharchive/pull-request
2024-10-03T04:53:31
2025-04-01T04:35:51.809802
{ "authors": [ "justinweng-instaclustr" ], "repo": "shotover/shotover-proxy", "url": "https://github.com/shotover/shotover-proxy/pull/1762", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
582570208
Correct Typos This might be the most important commit till now @horiaionut Please change the commit message to something meaningful. Use this https://chris.beams.io/posts/git-commit/ for reference. This is how you can change the commit message https://help.github.com/en/github/committing-changes-to-your-project/changing-a-commit-message
gharchive/pull-request
2020-03-16T20:09:12
2025-04-01T04:35:51.815771
{ "authors": [ "horiaionut", "shresthagrawal" ], "repo": "shresthagrawal/jslibp2p-chat-tutorial", "url": "https://github.com/shresthagrawal/jslibp2p-chat-tutorial/pull/2", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
222345714
ShtrihFiscalPrinter113 readTable writeTable написанны неправильно При чтении/записи из таблицы просиходит ошибка [ShtrihProtocol:getCashiersInfo:117]: Data parameter cannot be null jpos.JposException: Data parameter cannot be null at com.shtrih.jpos.DIOUtils.checkDataNotNull(DIOUtils.java:31) at com.shtrih.jpos.DIOUtils.checkDataMinLength(DIOUtils.java:44) at com.shtrih.jpos.fiscalprinter.directIO.DIOWriteTable.execute(DIOWriteTable.java:29) at com.shtrih.jpos.fiscalprinter.directIO.DirectIOHandler.directIO(DirectIOHandler.java:113) at com.shtrih.jpos.fiscalprinter.FiscalPrinterImpl.directIO(FiscalPrinterImpl.java:2002) at com.shtrih.jpos.fiscalprinter.FiscalPrinterService.directIO(FiscalPrinterService.java:272) at jpos.BaseJposControl.directIO(BaseJposControl.java:461) at com.shtrih.fiscalprinter.ShtrihFiscalPrinter113.directIO(ShtrihFiscalPrinter113.java:1006) at com.shtrih.fiscalprinter.ShtrihFiscalPrinter113.writeTable(ShtrihFiscalPrinter113.java:1156) Связанна она с тем что изначально в функции ShtrihFiscalPrinter113.readTable и ShtrihFiscalPrinter113.readTable параметр data == null Работающая реалищация методов public String readTable(int tableNumber, int rowNumber, int fieldNumber) throws JposException { String[] output = new String[1]; int[] data = new int[]{tableNumber, rowNumber, fieldNumber}; this.directIO(5, data, output); return output[0]; } public void writeTable(int tableNumber, int rowNumber, int fieldNumber, String fieldValue) throws JposException { String[] input = new String[]{fieldValue}; int[] data = new int[]{tableNumber, rowNumber, fieldNumber}; this.directIO(6, data, input); } И да и нет. В виду исторически сложившихся обстоятельств есть 2 версии DIO-команд, выбор производится параметром compatibilityLevel = 0/1 в jpos.xml, текущая реализация написана для compatibilityLevel = 0, Ваша реализация верна для версии 1. Вам нужно в jpos.xml добавить параметр . @nyxiscoo1 Спасибо, заработало
gharchive/issue
2017-04-18T08:53:31
2025-04-01T04:35:51.831501
{ "authors": [ "mifkamaz", "nyxiscoo1" ], "repo": "shtrih-m/javapos_shtrih", "url": "https://github.com/shtrih-m/javapos_shtrih/issues/6", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1638401749
Bug Fix Confuzz Non Reproducible: This is a quick and hacky way to fix, just add all ctest files into the Base Configuration API. Non Deterministic: WIP LGTM! LGTM! Okay. Let's also fix the Non-Deterministic bug in this PR @hwang-pku. What is the current state of that? I plan to look into it more this afternoon. I plan to look into it more this afternoon. Let me know if anything I can help with, we can meet to discuss this if needed tomorrow morning (your evening). I found a bug in this PR. Let me take a look... @hwang-pku I push another 2 commits, one is for the injection fix; the other one is a test in Yarn to test that different configuration API in yarn can still inject correctly. Please let me know if you find this is not sufficient or if you meet any other injection failure.
gharchive/pull-request
2023-03-23T22:01:53
2025-04-01T04:35:51.834601
{ "authors": [ "hwang-pku", "shuaiwang516" ], "repo": "shuaiwang516/fuzz-hadoop", "url": "https://github.com/shuaiwang516/fuzz-hadoop/pull/7", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2165315979
:triangular_flag_on_post: Migrate Avatar 🤔 Nature of this PR [ ] 🐛 bug fix [X] ✨ new feature [ ] 📝 documentation improvement [ ] 📝 demo code improvement [ ] 🚀 component style/interaction improvement [ ] 🏗️ ci/cd improvement [ ] ♻️ refactoring [ ] 🎨 code style optimization [ ] ✅ test cases [ ] 🔀 branch merging [ ] 💡 other 🔗 Related Issue #4 💡 Background and Solution ✅ Pre-merge Checklist ❗️Please self-check and check all options.❗️ [x] Documentation is supplemented or not needed [x] Code demonstration is provided or not needed [x] TypeScript definitions are supplemented or not needed [x] Changelog is provided or not needed Please check and modify your code and commits. do not submit code unrelated to the issue. Each pull request should correspond to a separate feature. By the way, This is one of my mistakes. The Vue version of Avatar is currently undergoing material optimization, so there may be a slightly larger iteration shortly, which will also involve the React version.
gharchive/pull-request
2024-03-03T11:41:10
2025-04-01T04:35:51.845266
{ "authors": [ "Muzych", "higuaifan" ], "repo": "shuimo-design/shuimo-ui-react", "url": "https://github.com/shuimo-design/shuimo-ui-react/pull/5", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
94587551
Added ability to get tagged users (into dev branch) I needed the ability to get tagged users (as I've talked about here: https://github.com/shyambhat/InstagramKit/issues/124) and I've added such functionality. Tagged users can be accessed like media.usersInPhoto. Accessing their X/Y coordinates inside the photo is also supported. Thanks @can16358p. There's some heavy documentation and refactoring work I did on the dev branch. Could you please fix conflicts and make your PR automatically mergable? Okay, laugh at me, but I'm not really good with this. Whatever I did I couldn't do it. You need to pull from the dev branch, fix conflicts on your working branch and then resubmit the PR. Hey there @shyambhat, will the next release include this feature and any estimates on when will it be out? @RyanTLX The PR is not mergable but I'll rewrite this and try to get it in by the end of this week.
gharchive/pull-request
2015-07-12T18:05:26
2025-04-01T04:35:51.930331
{ "authors": [ "RyanTLX", "can16358p", "shyambhat" ], "repo": "shyambhat/InstagramKit", "url": "https://github.com/shyambhat/InstagramKit/pull/126", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1925695766
[Feat]: Home page What feature? Home Page Feature I think that there should be a home page. As every web page should have one. The things I would like to see on the home page are: An introduction to the application What can you do in this application. ✅ I would like to work on this issue. Add screenshots Add screenshots Code of Conduct [X] I agree to follow this project's Code of Conduct I would like to work on this issue. I would like to work on this issue. @Vaibhav91one Assigning you but first you have share the design or reference you will use to built home page. Home Page Reference Design: Navbar Reference Design: Footer: I we have social links we can put them in the footer. And of course I will building them according the theme. These are just design references. @shyamtawli I like the first one. You can start working on it and If you need any help, you can contact me on Twitter hey @shyamtawli I have integrated react router and made a home page and added the necessary routes with Error page handling. Provide feedback and Green light for PR We need more attractive design. @shyamtawli What about color theme. Should I change it. I kept it similar to the original Website https://github.com/shyamtawli/devFind/assets/108279746/a8ead70e-bac4-4ab5-b9cf-92b45f77b1c1 @shyamtawli Here is another version of the home page. The video might lag a little bit. @shyamtawli I showed you the design earlier. You gave it the green light. Later I even modified it. @Vaibhav91one I approved the reference. Like the theme is looking good but we have make it our own. https://github.com/shyamtawli/devFind/assets/108279746/8c001787-a7a0-48c3-a842-d5321b1f9291 @shyamtawli I have made a few changes: animation ✅ background CSS✅ new section✅ theme✅ react router✅ ( I was busy, so I was not able to work on the issue. ) if this still require some enhancements. I would like someone else to build upon my changes. So, that I would not be clinging to the issue. Closing and Will Open new Issue for it.
gharchive/issue
2023-10-04T08:39:43
2025-04-01T04:35:51.940350
{ "authors": [ "Vaibhav91one", "shyamtawli" ], "repo": "shyamtawli/devFind", "url": "https://github.com/shyamtawli/devFind/issues/448", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2459307185
chore(main): release 0.4.0 :robot: I have created a release beep boop 0.4.0 (2024-08-10) Features Add spectral indices computation and calculation functions (3ef8c63) Features generation implemented - automatic and spectral indices (16365e0) Bug Fixes include (6d7e9e5) Remove unnecessary dataclass decorator brackets (28143f3) stubs for spyndex, mlxtend, and autofeat (f89a22a) Dependencies Update dependencies to include spyndex, mlxtend, and autofeat (edd4978) This PR was generated with Release Please. See documentation. :robot: Created releases: v0.4.0 :sunflower:
gharchive/pull-request
2024-08-10T22:05:36
2025-04-01T04:35:51.955192
{ "authors": [ "janezlapajne" ], "repo": "siapy/siapy-lib", "url": "https://github.com/siapy/siapy-lib/pull/99", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
160300027
incompatible version required Maybe this is just a fluke, but testing the package built by Launchpad fails, cf http://cdash-bipop.inrialpes.fr/viewConfigure.php?buildid=20401 This can be reproduced by using ./driver.py --task=siconos_test_deb You may want to change the Ubuntu distrib to 16.04, but it should not matter much May be fixed
gharchive/issue
2016-06-14T22:33:54
2025-04-01T04:35:52.051167
{ "authors": [ "xhub" ], "repo": "siconos/siconos", "url": "https://github.com/siconos/siconos/issues/68", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
149694035
scan error when match="*" func (db *DB) Scan(dataType DataType, cursor []byte, count int, inclusive bool, match string) 这个函数里的match不能为"",如果为""会报错: error parsing regexp: missing argument to repetition operator: * 如果我想穷举所有的key,match怎么设置 看了代码是标准的正则,好,我查一下相关写法
gharchive/issue
2016-04-20T08:12:51
2025-04-01T04:35:52.092124
{ "authors": [ "1fei" ], "repo": "siddontang/ledisdb", "url": "https://github.com/siddontang/ledisdb/issues/237", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
501746669
Add some date formats Add some date formats. #1 Note: Have some short documentation for each method that would be great. I can add them, just leave an example of method documentation then I can update this PR or create a new one adding the rest. @diegoalvis I think the documentation for each method will make the simple library unnecessarily complex. The comments are good enough to show the output of a method. But I realised that you have not added the methods to the README file. Please create a new PR to fix this. Sorry about not updating the readme. Sure I will. @sidhuparas please take a look at this PR #4
gharchive/pull-request
2019-10-02T21:54:21
2025-04-01T04:35:52.121651
{ "authors": [ "diegoalvis", "sidhuparas" ], "repo": "sidhuparas/SimpleDate", "url": "https://github.com/sidhuparas/SimpleDate/pull/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
706712623
Maintain go-kstat Hello @siebenmann Along with @Smithx10 we are working porting nomad to illumos and go-kstat will make things easier for us, we would like to continue working on this package. Would you be willing to transfer the ownership of this repo to illumos organization? We're certainly prepared to accept the repository, move it into the illumos organisation, and then maintain it within the community if you'd like to do that. @jclulow @jclulow @siebenmann, Definitely interested in moving this into the org! I would be happy to see this move to Illumos's hands and keep living on and being useful to people there. I've never done such a transfer in Github before, so I may need some advice on how best to do it. Great! I believe in order to effect a transfer, one GitHub account requires administrative permission over the repository to move, as well as permission to create a repository in the target organisation. I have the latter, so if you can grant me the former then I will try to move it! Hi, @siebenmann! Someone asked me about this again today and I realised I had been busy with other things and had forgotten to follow up. Would you be able to grant me administrative access to this repository so that I can move it into the illumos organisation? If I'm reading Github's documentation correctly, specifically this page, I think the process I need to do is transfer the repository to you, which will let you move it into the Illumos organization. Is this right as far as you know or is there another process/web page/etc I should consult? I suspect it would be better to move it directly from your account to the illumos organisation. If it takes a detour through my account, I don't know if GitHub will continue to redirect people from this URL straight over to the new eventual URL under illumos. If you add me as a collaborator on this repository, with the Admin access level, I believe I can begin the transfer myself -- and I have access to create repositories under illumos, so it should all just work. I've sent the invite now, and I think I can only make you an admin once you're a collaborator (or at least I didn't see an access level option on the invite). I don't think I can make you an Admin of a personal repository. I don't see any option for it, and the Github documentation on permission levels for personal repos doesn't list it as a possibility. Github doesn't immediately disallow an attempt to transfer the repo to 'illumos', although I did not go ahead and click the button to actually start the transfer (which might have made it error out). I can try to go through with that in case it works, or I can transfer it to you; let me know. You're right -- my apologies, most of my GitHub permissions experience is with organisations. If you push the button on a transfer to illumos and that appears to work, that'd be great; otherwise, transfer it to me and I'll transfer it over again. Sorry for the bother! The transfer to illumos failed with an 'I don't have permission to create repos in illumos' error, which I sort of expected, so I'm now transferring it to you. Transfer to illumos is now complete. Thanks, @siebenmann!
gharchive/issue
2020-09-22T21:44:13
2025-04-01T04:35:52.137210
{ "authors": [ "Smithx10", "cneira", "jclulow", "siebenmann" ], "repo": "siebenmann/go-kstat", "url": "https://github.com/siebenmann/go-kstat/issues/2", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
458732124
Connection refused error on backend connecting to Whatsapp After the API connected yo the backend, the backend tries to connect to Whatsapp, then I get "[Errno 111] Connection refused" and the following output: This is my first time trying to use the project. Has anyone run into the same problem? Change https://github.com/sigalor/whatsapp-web-reveng/blob/master/backend/whatsapp.py line 221 to wss: //web.whatsapp.com/ws @Romerito007 Have made the change and it's working now. Thanks.
gharchive/issue
2019-06-20T15:40:23
2025-04-01T04:35:52.152561
{ "authors": [ "AvivBalan", "Romerito007" ], "repo": "sigalor/whatsapp-web-reveng", "url": "https://github.com/sigalor/whatsapp-web-reveng/issues/201", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
978944992
Adding firewall rule for certmanager webhook This is an alternative solution to this PR in Ingress module. Things to note are: Since ingress module is a part of core modules, it make perfect sense to enable it along with SSH or GKE webhooks firewall rules. We are slightly diverging from upstream cert-manager, since their preferred solution was to change the cert-manager webhook port to 10250. LGTM!
gharchive/pull-request
2021-08-25T09:51:11
2025-04-01T04:35:52.183012
{ "authors": [ "angelbarrera92", "nandajavarma" ], "repo": "sighupio/fury-gke-installer", "url": "https://github.com/sighupio/fury-gke-installer/pull/19", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1181293388
Cloud Foundry Buildpack for Collector This is a deployment of the OpenTelemetry Collecter in the format of a Cloud Foundry buildpack. This buildpack supplies the Collector to the app it is used for. When the app is deployed it can run and configure the Collector as a sidecar (as described in the README). This will allow the Collector to observe the given app as well as the whole environment's metrics. Maybe outside of the scope of this PR, but is there anything we can add for CI testing? @jeffreyc-splunk In regards to adding CI testing, I don't think the actual functionality can be automated at this point. The problem is that this requires a Tanzu Application Service (TAS) instance, which currently needs to be created by specific users through a VMware dashboard. I don't believe the current usage agreement would allow for spinning up instances for testing. I'll check to make sure my understanding is correct, but that's why I hadn't implemented anything at this point. @jeffreyc-splunk In regards to adding CI testing, I don't think the actual functionality can be automated at this point. The problem is that this requires a Tanzu Application Service (TAS) instance, which currently needs to be created by specific users through a VMware dashboard. I don't believe the current usage agreement would allow for spinning up instances for testing. I'll check to make sure my understanding is correct, but that's why I hadn't implemented anything at this point. Maybe at a minimum, would it be possible to add a job to just create the buildpack and ensure that the supply script runs correctly? @jeffreyc-splunk In regards to adding CI testing, I don't think the actual functionality can be automated at this point. The problem is that this requires a Tanzu Application Service (TAS) instance, which currently needs to be created by specific users through a VMware dashboard. I don't believe the current usage agreement would allow for spinning up instances for testing. I'll check to make sure my understanding is correct, but that's why I hadn't implemented anything at this point. Maybe at a minimum, would it be possible to add a job to just create the buildpack and ensure that the supply script runs correctly? That should be doable, I'll work on adding that. @jeffreyc-splunk I've added a basic CI test that just makes sure the supply script runs successfully. I wanted to also make sure the buildpack is built successfully, but unfortunately that also requires a cloudfoundry instance to be running, so it isn't feasible. I'm not very familiar with GitHub automation, so please let me know if there's anything to fix!
gharchive/pull-request
2022-03-25T21:40:34
2025-04-01T04:35:52.233892
{ "authors": [ "crobert-1", "jcheng-splunk", "jeffreyc-splunk" ], "repo": "signalfx/splunk-otel-collector", "url": "https://github.com/signalfx/splunk-otel-collector/pull/1404", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2019503119
move discv5 to structured logging transform logging statements from using formating in strings to structured logging, for example: - debug!("Node set to disconnected: {}", node_id) + debug!(%node_id, "Node set to disconnected") This would make it easier to make sense of emitted logs it should be possible to shorten lists of node ids with format!("[{:#}]", list.iter().format(", ")) https://github.com/paradigmxyz/reth/issues/6080#issuecomment-1893329173
gharchive/issue
2023-11-30T21:01:18
2025-04-01T04:35:52.252422
{ "authors": [ "divagant-martian", "emhane" ], "repo": "sigp/discv5", "url": "https://github.com/sigp/discv5/issues/222", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1929600427
Start publishing the cursed token on GitHub Pages Start publishing the token on GitHub Pages (https://sigstore-conformance.github.io/extremely-dangerous-public-oidc-beacon/oidc-token.txt): This makes it a lot easier to find, download and use the token The old artifact upload is preserved so current scripts (so sigstore-conformance 0.7, 0.8) should keep working Fixes #5. Details: All known sigstore-conformance users have updated to v0.8 or are using main: I believe this is safe to merge. The GitHub Pages source is set to "Actions" in GH settings so this should start working right away. I can't see the environment settings but they should be correct by default (there should be a "github-pages" env and main branch should be allowed to deploy). The legacy artifact upload that is currently preserved can be removed once sigstore-conformance uses the new token location and we've seen that it's reliable. At that point I think the workflows in this project can be simplified significantly: the reason for the workflow dispatch dance is that published artifacts are not made available until the workflow finishes -- that limitation likely does not apply to Pages publishing (although it remains to be seen if GitHub is ok with publishing to Pages multiple times from the same workflow) This is a draft until sigstore/sigstore-conformance#102 is merged I supposed we should wait until there's a sigstore-conformance release and until the known users have upgraded as well. note to self: other conformance users have upgraded (or are using main branch), sigstore-js and sigstore-java are still on 0.6 at the time of this comment This bit in sigstore-conformance I'm not sure about: _OIDC_BEACON_WORKFLOW_ID = 55399612 Does the workflow id stay the same? I expect that it does but I can't be sure. Does the workflow id stay the same? I expect that it does but I can't be sure. I think it does, but we can confirm/update with https://docs.github.com/en/rest/actions/workflows?apiVersion=2022-11-28#get-a-workflow if it breaks after merging 🙂 In my experience, GitHub Pages sometimes takes a bit to propagate changes due to caching -- is this a concern w.r.t. the tokens, given their relatively short validity period? Caching on the HTTP header side, or caching on the deployment side? I wouldn't be surprised if GHP serves HTTP caching headers, but we can always ignore those 🙂 Gotcha. I think we can go ahead and deploy this as-is; if we run into consistent issues, then we'll look into another publication mechanism or location. Appears to be working: https://sigstore-conformance.github.io/extremely-dangerous-public-oidc-beacon/oidc-token.txt
gharchive/pull-request
2023-10-06T07:31:19
2025-04-01T04:35:52.292443
{ "authors": [ "jku", "woodruffw" ], "repo": "sigstore-conformance/extremely-dangerous-public-oidc-beacon", "url": "https://github.com/sigstore-conformance/extremely-dangerous-public-oidc-beacon/pull/7", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2202887038
latest tag not set automatically for v2 releases in GCR Description Version I had noticed previously that when we cut a v2 release, the latest tag is not set automatically. https://github.com/sigstore/cosign/issues/3620 noted that after releasing the backport v1 release yesterday, the latest tag was set on the v1 release. We should a) figure out why the v2 releases aren't getting tagged as latest automatically and b) make sure that when we cut a v1 release, we add documentation to manually move the latest tag back to v2 or remove the code from the release-1.13 branch to update the tag. Confirmed this is still happening, I had to manually set latest for https://console.cloud.google.com/gcr/images/projectsigstore/global/cosign@sha256:bed7ba33a8610c1607c16dee696f62bad168814016126abb9da01e9fb7cb2167/details?project=projectsigstore i will take a look at this today
gharchive/issue
2024-03-22T16:36:51
2025-04-01T04:35:52.296076
{ "authors": [ "cpanato", "haydentherapper" ], "repo": "sigstore/cosign", "url": "https://github.com/sigstore/cosign/issues/3621", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }