id
stringlengths 4
10
| text
stringlengths 4
2.14M
| source
stringclasses 2
values | created
timestamp[s]date 2001-05-16 21:05:09
2025-01-01 03:38:30
| added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| metadata
dict |
|---|---|---|---|---|---|
1522863484
|
🛑 TROJAN 🇺🇸 United States USO 1 is down
In 8557070, TROJAN 🇺🇸 United States USO 1 (https://ust-2.opensvr.net/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: TROJAN 🇺🇸 United States USO 1 is back up in a40ded8.
|
gharchive/issue
| 2023-01-06T16:50:59 |
2025-04-01T04:35:44.486402
|
{
"authors": [
"roosterkid"
],
"repo": "roosterkid/opentunnel-status-server",
"url": "https://github.com/roosterkid/opentunnel-status-server/issues/15963",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1627771038
|
🛑 TROJAN 🇺🇸 United States USO 1 is down
In d881a91, TROJAN 🇺🇸 United States USO 1 (https://ust-2.opensvr.net/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: TROJAN 🇺🇸 United States USO 1 is back up in 1881246.
|
gharchive/issue
| 2023-03-16T15:42:35 |
2025-04-01T04:35:44.488844
|
{
"authors": [
"roosterkid"
],
"repo": "roosterkid/opentunnel-status-server",
"url": "https://github.com/roosterkid/opentunnel-status-server/issues/20282",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2035433464
|
🛑 SSH 🇮🇩 Indonesia IDA 1 is down
In 53630c0, SSH 🇮🇩 Indonesia IDA 1 (http://ids-2.optnl.com:8080/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: SSH 🇮🇩 Indonesia IDA 1 is back up in f884f0a.
|
gharchive/issue
| 2023-12-11T11:14:31 |
2025-04-01T04:35:44.491364
|
{
"authors": [
"roosterkid"
],
"repo": "roosterkid/opentunnel-status-server",
"url": "https://github.com/roosterkid/opentunnel-status-server/issues/22132",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2052061905
|
🛑 XRAY 🇬🇧 United Kingdom UKO 1 is down
In 4d1f5ec, XRAY 🇬🇧 United Kingdom UKO 1 (https://ukx-1.openv2ray.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: XRAY 🇬🇧 United Kingdom UKO 1 is back up in 7db231c.
|
gharchive/issue
| 2023-12-21T09:37:28 |
2025-04-01T04:35:44.493791
|
{
"authors": [
"roosterkid"
],
"repo": "roosterkid/opentunnel-status-server",
"url": "https://github.com/roosterkid/opentunnel-status-server/issues/22893",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2056619355
|
🛑 OVPN 🇨🇦 Canada CAO 1 is down
In 4c85fb8, OVPN 🇨🇦 Canada CAO 1 (http://cao-1.optnl.com:8080/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: OVPN 🇨🇦 Canada CAO 1 is back up in e2915af.
|
gharchive/issue
| 2023-12-26T18:30:41 |
2025-04-01T04:35:44.496392
|
{
"authors": [
"roosterkid"
],
"repo": "roosterkid/opentunnel-status-server",
"url": "https://github.com/roosterkid/opentunnel-status-server/issues/23260",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2165866152
|
🛑 V2RAY 🇫🇷 France FR 1 is down
In af758e0, V2RAY 🇫🇷 France FR 1 (https://frv-1.openv2ray.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: V2RAY 🇫🇷 France FR 1 is back up in 3921b3d.
|
gharchive/issue
| 2024-03-04T03:53:53 |
2025-04-01T04:35:44.498806
|
{
"authors": [
"roosterkid"
],
"repo": "roosterkid/opentunnel-status-server",
"url": "https://github.com/roosterkid/opentunnel-status-server/issues/28242",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2177351576
|
🛑 V2RAY 🇫🇷 France FR 1 is down
In 150ba2d, V2RAY 🇫🇷 France FR 1 (https://frv-1.openv2ray.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: V2RAY 🇫🇷 France FR 1 is back up in 8621b4e.
|
gharchive/issue
| 2024-03-09T18:40:45 |
2025-04-01T04:35:44.501177
|
{
"authors": [
"roosterkid"
],
"repo": "roosterkid/opentunnel-status-server",
"url": "https://github.com/roosterkid/opentunnel-status-server/issues/28584",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1048637314
|
🛑 TROJAN 🇺🇸 United States USF 1 is down
In 4bf17e4, TROJAN 🇺🇸 United States USF 1 (https://ust-1.opensvr.net/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: TROJAN 🇺🇸 United States USF 1 is back up in 419799d.
|
gharchive/issue
| 2021-11-09T13:56:43 |
2025-04-01T04:35:44.502706
|
{
"authors": [
"roosterkid"
],
"repo": "roosterkid/opentunnel-status-server",
"url": "https://github.com/roosterkid/opentunnel-status-server/issues/3677",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1094085902
|
🛑 TROJAN 🇮🇩 Indonesia IDJ 1 is down
In 02c4121, TROJAN 🇮🇩 Indonesia IDJ 1 (https://idt-1.opensvr.net/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: TROJAN 🇮🇩 Indonesia IDJ 1 is back up in 94dcde6.
|
gharchive/issue
| 2022-01-05T08:08:57 |
2025-04-01T04:35:44.504302
|
{
"authors": [
"roosterkid"
],
"repo": "roosterkid/opentunnel-status-server",
"url": "https://github.com/roosterkid/opentunnel-status-server/issues/4693",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1187631289
|
🛑 TROJAN 🇸🇬 Singapore SGP 1 is down
In 862c136, TROJAN 🇸🇬 Singapore SGP 1 (https://sgt-2.opensvr.net/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: TROJAN 🇸🇬 Singapore SGP 1 is back up in 899af59.
|
gharchive/issue
| 2022-03-31T07:22:50 |
2025-04-01T04:35:44.506375
|
{
"authors": [
"roosterkid"
],
"repo": "roosterkid/opentunnel-status-server",
"url": "https://github.com/roosterkid/opentunnel-status-server/issues/6284",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1190765737
|
🛑 TROJAN 🇸🇬 Singapore SGA 1 is down
In 3e383f2, TROJAN 🇸🇬 Singapore SGA 1 (https://sgt-3.opensvr.net/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: TROJAN 🇸🇬 Singapore SGA 1 is back up in 6b0640c.
|
gharchive/issue
| 2022-04-02T23:45:20 |
2025-04-01T04:35:44.508010
|
{
"authors": [
"roosterkid"
],
"repo": "roosterkid/opentunnel-status-server",
"url": "https://github.com/roosterkid/opentunnel-status-server/issues/6324",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1410393715
|
🛑 TROJAN 🇺🇸 United States USF 1 is down
In 3647f94, TROJAN 🇺🇸 United States USF 1 (https://ust-1.opensvr.net/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: TROJAN 🇺🇸 United States USF 1 is back up in e659065.
|
gharchive/issue
| 2022-10-16T06:49:19 |
2025-04-01T04:35:44.509599
|
{
"authors": [
"roosterkid"
],
"repo": "roosterkid/opentunnel-status-server",
"url": "https://github.com/roosterkid/opentunnel-status-server/issues/9502",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1291917489
|
RootlyIncidentsPage & RootlyOverviewCard not exported
I've followed the steps to install the rootly plugin
yarn add @rootly/backstage-plugin
but when trying to add RootlyIncidentsPage and RootlyOverviewCard , and runing yarn tsc I get the errors :
Module '"@rootly/backstage-plugin"' has no exported member 'RootlyIncidentsPage'.
Cannot find module '@rootly/backstage-plugin/src/components/Entity' or its corresponding type declarations.
I've tried this on my companies internal backstage, but also I created a new backstage application from scratch to test this out with the same issue.
Not sure if it's relevant, but I'm using node lts/gallium
Thanks @kwent ,
I already have those two lines in my code, as I copied them over from the readme documentation.
I have deleted them and paste them again ( this time from your comment), but it still didn't work.
As far as I can tell, the npmjs package for rootly doesn't export those two objects, and that is why this doesn't work.
This is what I see when I go through my node_modules for @rootly :
import * as _backstage_core_plugin_api from '@backstage/core-plugin-api';
declare const rootlyPlugin: _backstage_core_plugin_api.BackstagePlugin<{
root: _backstage_core_plugin_api.RouteRef<undefined>;
}, {}>;
declare const RootlyPage: () => JSX.Element;
export { RootlyPage, rootlyPlugin };
I'm also attaching a screenshot for sake of clarity :)
The version of the rootly backstage pluging that I seem to be using is "@rootly/backstage-plugin": "^0.1.2"
Hi @avidelatm, could you try 0.1.7 with
import { RootlyIncidentsPage } from '@rootly/backstage-plugin';
import { RootlyOverviewCard } from '@rootly/backstage-plugin';
Regards
I am having the same issue, seems like the types declarations are missing in /dist/index.d.ts
|
gharchive/issue
| 2022-07-02T00:15:43 |
2025-04-01T04:35:44.548936
|
{
"authors": [
"avidelatm",
"drankou",
"kwent"
],
"repo": "rootlyhq/backstage-plugin",
"url": "https://github.com/rootlyhq/backstage-plugin/issues/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2629653948
|
Remove iron workflow and update readme
Due after the next patch release on November 15th.
@mergifyio backport humble
|
gharchive/pull-request
| 2024-11-01T18:41:59 |
2025-04-01T04:35:44.598968
|
{
"authors": [
"christophfroehlich"
],
"repo": "ros-controls/kinematics_interface",
"url": "https://github.com/ros-controls/kinematics_interface/pull/90",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2640884706
|
[JTC] Fix the JTC length_error exceptions in the tests
Fix the exception in JTC tests from time to time
[INFO] [1730930292.973865185] [test_joint_trajectory_controller]: Using 'splines' interpolation method.
[INFO] [1730930292.974352669] [test_joint_trajectory_controller]: Action status changes will be monitored at 20.00 Hz.
terminate called after throwing an instance of 'std::length_error'
what(): basic_string::_M_create
https://github.com/ros-controls/ros2_controllers/actions/runs/11712509515/job/32623343057?pr=1297
Thanks for the fix :) should we backport this?
May be yes!
Do we know why this is suddenly an issue? what has changed?
Do we know why this is suddenly an issue? what has changed?
We are just exiting the test in active state and the LoanedInterfaces are still available and it is throwing this exception upon their destruction(this happens only time to time). I think it should happen in other distros as well, but I don't know why it is not happening.
|
gharchive/pull-request
| 2024-11-07T12:29:39 |
2025-04-01T04:35:44.602536
|
{
"authors": [
"christophfroehlich",
"saikishor"
],
"repo": "ros-controls/ros2_controllers",
"url": "https://github.com/ros-controls/ros2_controllers/pull/1360",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
431411808
|
Unused min/max_payload parameters
This package's launch files expose min_payload and max_payload arguments, which are used to set equally named parameters, but those parameters are not used anywhere.
Those were supported in the master version (which got it from ur_driver), but apparently @Zagitta didn't implement the checks:
https://github.com/ros-industrial/ur_modern_driver/blob/77fa08ae9c846344310d3b50824a7affdc3eda47/include/ur_modern_driver/ros/io_service.h#L65-L70
I added help wanted as this could be a good first issue for someone to tackle.
Did you make any progress here @ben-greenberg?
I have made progress but some further work and testing is still needed. I will be continuing my work on this.
I'm going to unassign this @ben-greenberg.
You're welcome to keep working on it of course.
|
gharchive/issue
| 2019-04-10T09:57:15 |
2025-04-01T04:35:44.629215
|
{
"authors": [
"ben-greenberg",
"gavanderhoorn",
"miguelprada"
],
"repo": "ros-industrial/ur_modern_driver",
"url": "https://github.com/ros-industrial/ur_modern_driver/issues/289",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
497267410
|
fanuc: add Melodic source entries
As per subject.
Thanks for the suggestion @tfoote.
|
gharchive/pull-request
| 2019-09-23T18:53:23 |
2025-04-01T04:35:44.680154
|
{
"authors": [
"gavanderhoorn"
],
"repo": "ros/rosdistro",
"url": "https://github.com/ros/rosdistro/pull/22338",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
1994179289
|
fields2cover: 1.2.1-1 in 'humble/distribution.yaml' [bloom]
Increasing version of package(s) in repository fields2cover to 1.2.1-1:
upstream repository: https://github.com/Fields2Cover/fields2cover.git
release repository: https://github.com/Fields2Cover/fields2cover-release.git
distro file: humble/distribution.yaml
bloom version: 0.10.7
previous version for package: null
@Gonzalo-Mier friendly ping
I don't understand why or how to solve this error. @mjcarroll do you have any insight? Thank you
Typically an entry will have a source, doc and release tag, something like:
filters:
doc:
type: git
url: https://github.com/ros/filters.git
version: ros2
release:
tags:
release: release/humble/{package}/{version}
url: https://github.com/ros2-gbp/filters-release.git
version: 2.1.0-3
source:
test_pull_requests: true
type: git
url: https://github.com/ros/filters.git
version: ros2
status: maintained
Can you update your pull request to at least include source, but maybe additionally doc?
Done. @mjcarroll Thank you for your support!
|
gharchive/pull-request
| 2023-11-15T07:08:07 |
2025-04-01T04:35:44.684546
|
{
"authors": [
"Gonzalo-Mier",
"mjcarroll"
],
"repo": "ros/rosdistro",
"url": "https://github.com/ros/rosdistro/pull/38990",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
2655201865
|
feetech_ros2_driver: 0.1.0-1 in 'humble/distribution.yaml' [bloom]
Increasing version of package(s) in repository feetech_ros2_driver to 0.1.0-1:
upstream repository: https://github.com/JafarAbdi/feetech_ros2_driver.git
release repository: https://github.com/JafarAbdi/feetech_ros2_driver-release.git
distro file: humble/distribution.yaml
bloom version: 0.12.0
previous version for package: null
feetech_ros2_driver
* Add feetech ros2 driver
* Contributors: Jafar Uruç
See the comment in https://github.com/ros/rosdistro/pull/43490#pullrequestreview-2443491346 , which will probably necessitate changes here.
https://github.com/ros/rosdistro/pull/43993 replaces this
|
gharchive/pull-request
| 2024-11-13T11:38:49 |
2025-04-01T04:35:44.688563
|
{
"authors": [
"JafarAbdi",
"clalancette"
],
"repo": "ros/rosdistro",
"url": "https://github.com/ros/rosdistro/pull/43488",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
105364721
|
repository_name: 0.0.2-0 in 'indigo/distribution.yaml' [bloom]
Increasing version of package(s) in repository repository_name to 0.0.2-0:
upstream repository: https://gitlab.uni-koblenz.de/robbie/homer_android_speech.git
release repository: https://gitlab.uni-koblenz.de/robbie/homer_android_speech.git
distro file: indigo/distribution.yaml
bloom version: 0.5.20
previous version for package: null
android_speech_pkg
* big bang
* Contributors: Raphael Memmesheimer
repository_name is a generic name
Changed to homer_android_speech. Don't know where the name came from.
Found it. Copy Paste error from the bloom tutorial.
|
gharchive/pull-request
| 2015-09-08T12:04:40 |
2025-04-01T04:35:44.692141
|
{
"authors": [
"airglow",
"vrabaud"
],
"repo": "ros/rosdistro",
"url": "https://github.com/ros/rosdistro/pull/9405",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
134445235
|
fix mega_genmsg to compile
while trying to build executables for the Native POSIX portion of the FreeRTPS demo, I found that the messages generated by mega_genmsg.py for rcl_interfaces didn't compile. This fixes it.
This isn't part of the binary packaging right? +1
Indeed.
look, I can even start a CI job for this, except it won't run any tests.
http://ci.ros2.org/job/freertps/19/
|
gharchive/pull-request
| 2016-02-18T00:38:05 |
2025-04-01T04:35:44.719783
|
{
"authors": [
"jacquelinekay",
"wjwwood"
],
"repo": "ros2/freertps",
"url": "https://github.com/ros2/freertps/pull/30",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
252804460
|
js files have content undefined
Hi,
when I try uglify, it overwrites all .js files, but they contet is only one single word "undefined".
stripped debug log: cordova -d prepare
No scripts found for hook "before_prepare".
Checking config.xml and package.json for saved platforms that haven't been added to the project
Config.xml and package.json platforms are the same. No pkg.json modification.
Package.json and config.xml platforms are different. Updating config.xml with most current list of platforms.
Failed to require PlatformApi instance for platform "browser". Using polyfill instead.
Checking config.xml for saved plugins that haven't been added to the project
Checking for any plugins added to the project that have not been installed in browser platform
No differences found between plugins added to project and installed in browser platform. Continuing...
Generating config.xml from defaults for platform "browser"
Executing script found in hooks directory for hook "after_prepare": hooks\after_prepare\uglify.js
Running command: "C:\Program Files\nodejs\node.exe" D:\NetBeansProjects\ReturnReporting\hooks\after_prepare\uglify.js D:\NetBeansProjects\ReturnReporting
processFiles err: Error: ENOENT: no such file or directory, scandir 'D:\NetBeansProjects\ReturnReporting\platforms\browser\www\module'
uglifying js file D:\NetBeansProjects\ReturnReporting\platforms\browser\www\controller\BaseController.js
uglifying js file D:\NetBeansProjects\ReturnReporting\platforms\browser\www\controller\Home.controller.js
uglifying js file D:\NetBeansProjects\ReturnReporting\platforms\browser\www\controller\Login.controller.js
uglifying js file D:\NetBeansProjects\ReturnReporting\platforms\browser\www\controller\Settings.controller.js
minifying css file D:\NetBeansProjects\ReturnReporting\platforms\browser\www\css\style.css
uglifying js file D:\NetBeansProjects\ReturnReporting\platforms\browser\www\Component.js
encountered a .xml file, not compressing it
uglifying js file D:\NetBeansProjects\ReturnReporting\platforms\browser\www\confighelper.js
uglifying js file D:\NetBeansProjects\ReturnReporting\platforms\browser\www\controller\App.controller.js
uglifying js file D:\NetBeansProjects\ReturnReporting\platforms\browser\www\cordova.js
uglifying js file D:\NetBeansProjects\ReturnReporting\platforms\browser\www\exec.js
encountered a .html file, not compressing it
encountered a .json file, not compressing it
full debug log: uglify_debug.txt
uglify installed from npm
{ name: 'cordova-uglify',
description: 'Cordova hook that allows you to uglify or minify your apps JavaScript and CSS.',
'dist-tags': { latest: '0.2.8' },
versions:
[ '0.0.1',
'0.0.2',
'0.0.3',
'0.0.4',
'0.0.5',
'0.0.6',
'0.0.7',
'0.0.8',
'0.0.9',
'0.1.0',
'0.1.1',
'0.1.2',
'0.1.3',
'0.1.4',
'0.1.5',
'0.1.6',
'0.1.7',
'0.1.8',
'0.1.9',
'0.2.0',
'0.2.1',
'0.2.2',
'0.2.3',
'0.2.4',
'0.2.5',
'0.2.6',
'0.2.7',
'0.2.8' ],
maintainers: [ 'rossmartin <rmartin311@gmail.com>' ],
time:
{ modified: '2017-01-19T15:51:35.841Z',
created: '2014-06-18T03:56:05.284Z',
'0.0.1': '2014-06-18T03:56:05.284Z',
'0.0.2': '2014-06-18T05:35:27.107Z',
'0.0.3': '2014-06-18T05:42:27.368Z',
'0.0.4': '2014-06-18T05:49:43.994Z',
'0.0.5': '2014-06-18T05:52:58.761Z',
'0.0.6': '2014-06-19T05:21:03.197Z',
'0.0.7': '2014-06-22T22:23:42.514Z',
'0.0.8': '2014-06-23T00:22:46.231Z',
'0.0.9': '2014-11-12T05:09:29.435Z',
'0.1.0': '2014-11-12T05:14:26.594Z',
'0.1.1': '2014-11-12T05:24:53.626Z',
'0.1.2': '2014-12-19T15:56:15.972Z',
'0.1.3': '2014-12-19T16:12:12.345Z',
'0.1.4': '2015-01-12T22:09:19.752Z',
'0.1.5': '2015-04-09T20:55:45.840Z',
'0.1.6': '2015-06-22T14:47:17.993Z',
'0.1.7': '2015-07-11T02:08:10.764Z',
'0.1.8': '2015-07-11T02:31:39.325Z',
'0.1.9': '2015-08-03T21:01:21.157Z',
'0.2.0': '2015-09-30T13:38:50.084Z',
'0.2.1': '2015-10-03T16:06:54.283Z',
'0.2.2': '2015-12-13T22:54:54.170Z',
'0.2.3': '2015-12-16T01:08:51.901Z',
'0.2.4': '2016-07-15T15:12:03.110Z',
'0.2.5': '2016-09-11T17:32:27.083Z',
'0.2.6': '2016-10-24T14:35:28.131Z',
'0.2.7': '2016-11-27T04:43:01.635Z',
'0.2.8': '2017-01-19T15:51:35.841Z' },
homepage: 'https://github.com/rossmartin/cordova-uglify',
keywords: [ 'cordova', 'uglify', 'minify', 'hook', 'hooks' ],
author: 'Ross Martin',
bugs: { url: 'https://github.com/rossmartin/cordova-uglify/issues' },
license: 'MIT',
readmeFilename: 'README.md',
repository:
{ type: 'git',
url: 'git+https://github.com/rossmartin/cordova-uglify.git' },
users:
{ okmogwai: true,
batosai: true,
cnring18: true,
gdevise: true,
ebbm: true },
version: '0.2.8',
dependencies:
{ 'clean-css': '3.4.8',
'ng-annotate': '0.15.4',
shelljs: '^0.7.0',
'uglify-js': '2.4.14' },
scripts:
{ postinstall: 'node scripts/install.js',
postuninstall: 'node scripts/uninstall.js' },
gitHead: '395775651d1b3a5f3625e4fa3d3ec2ca823664b1',
dist:
{ shasum: '5aa9f69797da4c9d51c2f5184f4dcf0510d1b327',
tarball: 'https://registry.npmjs.org/cordova-uglify/-/cordova-uglify-0.2.8.tgz' },
directories: {} }
Sorry for the late response. Can you please provide the contents of your uglify-config.json? I think this issue is isolated to Windows.
Hello,
this is my used config:
{
"alwaysRun": true,
"recursiveFolderSearch": true,
"foldersToProcess": [
"js",
"css",
"img",
"build"
],
"uglifyJsOptions": {
"compress": {
"drop_console": true
},
"fromString": true,
"mangle": false
},
"cleanCssOptions": {
"noAdvanced": true,
"keepSpecialComments": 0
}
}
It is odd that it is modifying files outside of your foldersToProcess - that is how it appears from your debug file. This behavior doesn't happen on OSX or Linux - I'll have to see if I can reproduce your issue on Windows.
I'm not sure this is the same problem, but it may help:
I had the same problem, ie .js files with 'undefined' content.
After debugging uglify.js, the following line
result = UglifyJS.minify (res.src, hookConfig.uglifyJsOptions);
return a "result.code" undefined
an overview before correction:
result = UglifyJS.minify(res.src, hookConfig.uglifyJsOptions);
console.log('result.code ' + result.code);
fs.writeFileSync(file, result.code, 'utf8'); // overwrite the original unminified file
break;`
Correction of the line (removed parameter "hookConfig.uglifyJsOptions" ,fast correction, there is surely better)
result = UglifyJS.minify(res.src);
console.log('result.code ' + result.code);
fs.writeFileSync(file, result.code, 'utf8'); // overwrite the original unminified file
break;
Why?
Updates with npm have pushed new version (uglify and other), the uglifyJsOptions parameters did not match the version of UglifyJs
ps: no more time to spend on this project, it works!
@jeromeXoo Thank you for providing this info. I had some time this evening to look into this issue and it appears that the last PR introduced a few problems. I thought I had tested it after I merged but apparently I didn't have the latest commit.
I've fixed the issue and published a new version (0.2.9) to npm. Let me know if you are still having issues.
Hi,
it crashes on first file:
cordova -d prepare
No scripts found for hook "before_prepare".
Checking config.xml and package.json for saved platforms that haven't been added to the project
Config.xml and package.json platforms are the same. No pkg.json modification.
Package.json and config.xml platforms are different. Updating config.xml with most current list of platforms.
Failed to require PlatformApi instance for platform "browser". Using polyfill instead.
Checking config.xml for saved plugins that haven't been added to the project
Checking for any plugins added to the project that have not been installed in browser platform
No differences found between plugins added to project and installed in browser platform. Continuing...
Generating config.xml from defaults for platform "browser"
Executing script found in hooks directory for hook "after_prepare": hooks\after_prepare\uglify.js
Running command: "C:\Program Files\nodejs\node.exe" D:\NetBeansProjects\ReturnReporting\hooks\after_prepare\uglify.js D:\NetBeansProjects\ReturnReporting
minifying css file D:\NetBeansProjects\ReturnReporting\platforms\browser\www\css\style.css
uglifying js file D:\NetBeansProjects\ReturnReporting\platforms\browser\www\controller\App.controller.js
fs.js:584
return binding.open(pathModule._makeLong(path), stringToFlags(flags), mode);
^
Error: ENOENT: no such file or directory, open 'D:\NetBeansProjects\ReturnReporting\sap.ui.define([
'sap\ui\returnreporting\controller\BaseController',
...
...
at Error (native)
at Object.fs.openSync (fs.js:584:18)
at Object.fs.readFileSync (fs.js:431:33)
at addFile (D:\NetBeansProjects\ReturnReporting\node_modules\uglify-js\tools\node.js:70:22)
at D:\NetBeansProjects\ReturnReporting\node_modules\uglify-js\tools\node.js:81:17
at Array.forEach (native)
at Object.exports.minify (D:\NetBeansProjects\ReturnReporting\node_modules\uglify-js\tools\node.js:79:26)
at compress (D:\NetBeansProjects\ReturnReporting\hooks\after_prepare\uglify.js:127:25)
at D:\NetBeansProjects\ReturnReporting\hooks\after_prepare\uglify.js:94:11
at FSReqWrap.oncomplete (fs.js:82:15)
Command finished with error code 1: C:\Program Files\nodejs\node.exe D:\NetBeansProjects\ReturnReporting\hooks\after_prepare\uglify.js,D:\NetBeansProjects\ReturnReporting
Error: Hook failed with error code 1: D:\NetBeansProjects\ReturnReporting\hooks\after_prepare\uglify.js
my config:
{
"alwaysRun": true,
"recursiveFolderSearch": true,
"foldersToProcess": [
"css",
"controller",
"model",
"view"
],
"uglifyJsOptions": {
"compress": {
"drop_console": true
},
"mangle": false,
"output": {
"code": true
}
},
"cleanCssOptions": {
"noAdvanced": true,
"keepSpecialComments": 0
}
}
cordova uglify installed today by npm
D:\NetBeansProjects\ReturnReporting>npm show cordova-uglify
{ name: 'cordova-uglify',
description: 'Cordova hook that allows you to uglify or minify your apps JavaScript and CSS.',
'dist-tags': { latest: '0.2.9' },
versions:
[ '0.0.1',
'0.0.2',
'0.0.3',
'0.0.4',
'0.0.5',
'0.0.6',
'0.0.7',
'0.0.8',
'0.0.9',
'0.1.0',
'0.1.1',
'0.1.2',
'0.1.3',
'0.1.4',
'0.1.5',
'0.1.6',
'0.1.7',
'0.1.8',
'0.1.9',
'0.2.0',
'0.2.1',
'0.2.2',
'0.2.3',
'0.2.4',
'0.2.5',
'0.2.6',
'0.2.7',
'0.2.8',
'0.2.9' ],
maintainers: [ 'rossmartin <rmartin311@gmail.com>' ],
time:
{ modified: '2017-09-22T02:55:35.360Z',
created: '2014-06-18T03:56:05.284Z',
'0.0.1': '2014-06-18T03:56:05.284Z',
'0.0.2': '2014-06-18T05:35:27.107Z',
'0.0.3': '2014-06-18T05:42:27.368Z',
'0.0.4': '2014-06-18T05:49:43.994Z',
'0.0.5': '2014-06-18T05:52:58.761Z',
'0.0.6': '2014-06-19T05:21:03.197Z',
'0.0.7': '2014-06-22T22:23:42.514Z',
'0.0.8': '2014-06-23T00:22:46.231Z',
'0.0.9': '2014-11-12T05:09:29.435Z',
'0.1.0': '2014-11-12T05:14:26.594Z',
'0.1.1': '2014-11-12T05:24:53.626Z',
'0.1.2': '2014-12-19T15:56:15.972Z',
'0.1.3': '2014-12-19T16:12:12.345Z',
'0.1.4': '2015-01-12T22:09:19.752Z',
'0.1.5': '2015-04-09T20:55:45.840Z',
'0.1.6': '2015-06-22T14:47:17.993Z',
'0.1.7': '2015-07-11T02:08:10.764Z',
'0.1.8': '2015-07-11T02:31:39.325Z',
'0.1.9': '2015-08-03T21:01:21.157Z', 10:42
'0.2.0': '2015-09-30T13:38:50.084Z',
'0.2.1': '2015-10-03T16:06:54.283Z',
'0.2.2': '2015-12-13T22:54:54.170Z',
'0.2.3': '2015-12-16T01:08:51.901Z',
'0.2.4': '2016-07-15T15:12:03.110Z',
'0.2.5': '2016-09-11T17:32:27.083Z',
'0.2.6': '2016-10-24T14:35:28.131Z',
'0.2.7': '2016-11-27T04:43:01.635Z',
'0.2.8': '2017-01-19T15:51:35.841Z',
'0.2.9': '2017-09-22T02:55:35.360Z' },
homepage: 'https://github.com/rossmartin/cordova-uglify',
keywords: [ 'cordova', 'uglify', 'minify', 'hook', 'hooks' ],
author: 'Ross Martin',
bugs: { url: 'https://github.com/rossmartin/cordova-uglify/issues' },
license: 'MIT',
readmeFilename: 'README.md',
repository:
{ type: 'git',
url: 'git+https://github.com/rossmartin/cordova-uglify.git' },
users: { okmogwai: true, cnring18: true, gdevise: true, ebbm: true },
version: '0.2.9',
dependencies:
{ 'clean-css': '3.4.8',
'ng-annotate': '0.15.4',
shelljs: '^0.7.0',
'uglify-js': '3.1.1' },
scripts:
{ postinstall: 'node scripts/install.js',
postuninstall: 'node scripts/uninstall.js' },
gitHead: 'e101c6e9f673e43f2431d2a238c4fdbf301870b2',
dist:
{ shasum: 'af25bcd17b1687d91d60856ad317590901eb3565',
tarball: 'https://registry.npmjs.org/cordova-uglify/-/cordova-uglify-0.2.9.tgz' },
directories: {} }```
@rchovan Thanks for reporting the issue. I have a theatre PC with Windows on it I'll try it this weekend. I'm not experiencing any issues on Ubuntu or Mac OS. I wonder if the issue is because the project is on a separate drive than node/windows.
@rossmartin I'm experiencing the same issue where the JS output files only contain "undefined". I am running the latest version of cordova-uglify 0.3.0 as reported by npm and node v8.9.1 on Windows 10.
@koga73 Thanks for letting me know. Unfortunately my theatre PC that had Windows 10 on it died! Bad motherboard or power supply I think but I don't plan on reviving it. Recently I installed 2 Windows VMs to an external HDD on my Macbook for my day job to test IE11 issues. I'll have to use a VM to test this out and get it fixed. I haven't forgotten about this just haven't had time - it's frustrating because it works fine on Linux and Mac OS.
@rossmartin Did some digging. Turns out your plugin is fine. Issue is we had an old uglify-config.json. Specifically this line was causing it:
"uglifyJsOptions": { "fromString": true }
@koga73 Thanks a bunch for letting me know. I wonder if this is the issue you are having @rchovan ?
I could update the install script to overwrite the uglify-config.json but I think a better solution is to just document it in the readme and output after installing the package. I don't want anyone to lose their config on update.
I have this error too. Using default options from README doesn't help.
Hi all,
started experiencing the same issue.
What is generating the empty file in my case is the following code:
BackgroundGeolocation.on('location', (location) => {
// handle your locations here
// to perform long running operation on iOS
// you need to create background task
BackgroundGeolocation.startTask(taskKey => {
// execute long running task
// eg. ajax post location
// IMPORTANT: task has to be ended by endTask
BackgroundGeolocation.endTask(taskKey);
});
});
If I remove it then it works properly
Running uglijs manually, I get
BackgroundGeolocation.on('location', (position) => {
ERROR: Unexpected token: operator (>)
How could I fix it considering is legitimate js code?
Thanks
I'm trying
https://github.com/mishoo/UglifyJS2/tree/harmony
And it seems to solve the issue.
In uglify hook I changed this line:
var UglifyJS = require(path.join(dependencyPath, 'uglify-js'));
to
var UglifyJS = require(path.join(dependencyPath, 'uglify-es'));
So far seems ok, will test more extensively
@luigi37 I think you may have revealed why others are experiencing this same issue. uglify-js does not work with ES6. I think most transpile their "modern" JS to ES5 before feeding it to uglify-js. It appears that uglify-es might be the way to go moving forward to handle minifying code if it is ES6.
There is also this interesting compress option with uglify-es -
ecma (default: 5) -- Pass 6 or greater to enable compress options that
will transform ES5 code into smaller ES6+ equivalent forms.
Thanks for bringing this up @luigi37. I'm going to give it a go and see how it works out. If it works with the old and new stuff I'll publish a new version today.
Cool, thanks!
@luigi37 I published a new version (0.3.1) that uses uglify-es. I tried the latest code with some old projects that were plain ES5 and a new one with ES6 - I had no issues.
I'll be leaving this issue open for now.
@luigi37 and anyone else, please see #39 about the decision to abandon uglify-es in https://github.com/rossmartin/cordova-uglify/commit/eeb41a851fbf23b44439485885394d9e243fce9c (v0.3.2)
Make sure your cordova project's code is transpiled to ES5 (I think you should even if you don't use this hook).
@rossmartin thank you. I'm not familiar with ES6 enough to confirm if transpiling back to ES5 can affect the functionality.
Could you please confirm?
If so, I'm happy with your solution.
Thanks a lot!
Luigi
@luigi37 if you use babel to transpile your ES6 to ES5 it won't affect functionality. Not all mobile device webviews support ES6 so it would be much safer to transpile.
Thank you.
Is there a way to include the transpile into cordova-uglify to avoid multiple steps?
Thanks
Luigi
@luigi37 That would be a nice feature to have if you're using something like framework 7 & react without using react-scripts. Those that use Ionic it won't be of any benefit.
I'll think about it - this may be a better candidate for a separate hook.
Thanks Ross,
it would be much appreciated.
Specifically I'm using Cordova with Framework7... Anyway it's a specific
cordova plugin which I use which brought me to the ES6 thing...
Luigi
-----------------------------------
Luigi Paiella
luigi@pajella.com luigi@pajella.com**-----------------------------------
2018-04-15 18:25 GMT+02:00 Ross Martin notifications@github.com:
@luigi37 https://github.com/luigi37 That would be a nice feature to
have if you're using something like framework 7 & react without using
react-scripts. Those that use Ionic it won't be of any benefit.
I'll think about it - this may be a better candidate for a separate hook.
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
https://github.com/rossmartin/cordova-uglify/issues/35#issuecomment-381419013,
or mute the thread
https://github.com/notifications/unsubscribe-auth/ABYb8VG7FS5SzsF9CSELk0fACMVn3isDks5to3R5gaJpZM4PCQEX
.
Thanks Ross,
it would be much appreciated.
Specifically I'm using Cordova with Framework7... Anyway it's a specific
cordova plugin which I use which brought me to the ES6 thing...
Luigi
@rossmartin Hello!
The same issue. Works great with uglify-es but with uglify-js one file always has undefined content.
This same issue is happening to me too, however its only happening to three files, all three files are 9 bytes and all say undefined.
I am running an Ionic v1 app, so know Ionic v2 or ES issues to worry about. Just straight old AngularJS and JS.
I'm having the same issue, 1 file only has "undefined"
and I'm not on Windows, please read more in #46
|
gharchive/issue
| 2017-08-25T05:48:56 |
2025-04-01T04:35:44.840641
|
{
"authors": [
"iliakonnov",
"jeromeXoo",
"koga73",
"luigi37",
"rchovan",
"rodrigograca31",
"rolinger",
"rossmartin",
"zankevich"
],
"repo": "rossmartin/cordova-uglify",
"url": "https://github.com/rossmartin/cordova-uglify/issues/35",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
118394506
|
server rendering error
I am getting error when trying to render on the server:
Warning: React can't find the root component node for data-reactid value .ba9ic960ao.1.2. If you're seeing this message, it probably means that you've loaded two copies of React on the page. At this time, only a single copy of React can be loaded at a time.
TypeError: Cannot read property 'firstChild' of undefined
at Object.ReactMount.findComponentRoot
...(and so on)
(Here I created repo to reproduce this issue: https://github.com/olessavluk/r5bug )
I'm using node v5.0.0 & npm v3.4.1. And I have only one veriosn of React installed, because find ./ -name 'react' returns only one result .//node_modules/react.
So I think problem in this file https://github.com/router5/react-router5/blob/master/modules/routeNode.js#L21. Because you are adding listener in constructor and removing it in componentWillUnmout, but unmount never gets executed on the server.
I am quite inexperienced with universal applications so thank you for raising potential issues.
I cloned and ran your example, the error goes away by removing the input element in Timer component. Not sure what that means.
I have created new branch (logall) and put logs everywhere to see what is actually going on.
As you can see, listener executes setState after rendering is complete and this actually causes an error.
:tada: I think I managed to fix it. The problem is not with componentWillUnmount but with suff in component constructor (componentWillMount) which should be in componentDidMount.
I will commit a fix and release it later tonight.
You are talking about this ?) - https://github.com/olessavluk/react-router5/commit/12aad1612002f7d5380c60370e9dbeaffa55376e
Yes, exactly. I also moved the check for the listeners plugin to component did mount so you are not forced to use it on the server side.
1.0.1 has been published
Thank you very much for raising that bug and taking the time to create a repo. I really appreciate. :+1:
Linking it to other similar issues, in case that helps others.
RickWong/react-isomorphic-starterkit#77
denvned/isomorphic-relay-router#5
|
gharchive/issue
| 2015-11-23T14:39:28 |
2025-04-01T04:35:44.892908
|
{
"authors": [
"olessavluk",
"troch"
],
"repo": "router5/react-router5",
"url": "https://github.com/router5/react-router5/issues/11",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
4447111
|
support COPY commands
EM versions of be get_copy_data and put_copy_data are missing.
I'm really interested by this feature.
I was trying to use the "copy_table" method from sequel (via em-pg-sequel) in order to create a fast export to CSV and I was wondering why it was working great in the console but not in EM (ruby process hanging at 100% CPU)...
I willing to help but I have no idea where to start and how difficult it coud be...
|
gharchive/issue
| 2012-05-06T22:49:57 |
2025-04-01T04:35:44.894454
|
{
"authors": [
"flop",
"royaltm"
],
"repo": "royaltm/ruby-em-pg-client",
"url": "https://github.com/royaltm/ruby-em-pg-client/issues/4",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
192587020
|
Added GridFS support
JAVARS-23
OK moved all implementation code into the internal package.
The following internal classes are now public:
MongoClientImpl - so can be constructed from MongoClients
MongoDatabaseImpl - so can be accessed via GridFSBuckets
GridFSBucketImpl - so can be constructed from GridFSBuckets
GridFSAsyncStreamHelper - so can be accessed via AsyncStreamHelper
If you haven't already, make sure you add an exclude for the internal package when generating javadoc.
|
gharchive/pull-request
| 2016-11-30T14:58:15 |
2025-04-01T04:35:44.903902
|
{
"authors": [
"jyemin",
"rozza"
],
"repo": "rozza/mongo-java-driver-reactivestreams",
"url": "https://github.com/rozza/mongo-java-driver-reactivestreams/pull/7",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1694480392
|
🛑 Jay Industries is down
In a78c3ec, Jay Industries (https://jaybrassindustries.com/) was down:
HTTP code: 404
Response time: 17748 ms
Resolved: Jay Industries is back up in 66ebce4.
|
gharchive/issue
| 2023-05-03T17:13:54 |
2025-04-01T04:35:44.906502
|
{
"authors": [
"rpharaniya"
],
"repo": "rpharaniya/websites-uptime-monitor",
"url": "https://github.com/rpharaniya/websites-uptime-monitor/issues/1036",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2102087986
|
🛑 Jay Industries is down
In 1dd54f5, Jay Industries (https://jaybrassindustries.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Jay Industries is back up in 3336b9d after 5 minutes.
|
gharchive/issue
| 2024-01-26T11:47:53 |
2025-04-01T04:35:44.908879
|
{
"authors": [
"rpharaniya"
],
"repo": "rpharaniya/websites-uptime-monitor",
"url": "https://github.com/rpharaniya/websites-uptime-monitor/issues/1734",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2447220554
|
🛑 Silico Rubber Polymers is down
In 6d437c2, Silico Rubber Polymers (https://silicorubberpolymers.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Silico Rubber Polymers is back up in 4fc285e after 15 minutes.
|
gharchive/issue
| 2024-08-04T17:07:02 |
2025-04-01T04:35:44.911240
|
{
"authors": [
"rpharaniya"
],
"repo": "rpharaniya/websites-uptime-monitor",
"url": "https://github.com/rpharaniya/websites-uptime-monitor/issues/2035",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1175018936
|
How to train the model?
Can you please tell me how to train the model onmy own dataset ?
ok,thanks.
|
gharchive/issue
| 2022-03-21T08:02:45 |
2025-04-01T04:35:44.957111
|
{
"authors": [
"zhw2020913"
],
"repo": "rpm1412/LUS_Video_Summarization",
"url": "https://github.com/rpm1412/LUS_Video_Summarization/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
833667577
|
Build error when .NET 6 preview is installed. Any suggestions to correct this is appreciated.
We get the following error when Visual Studio 2019 Preview is installed on our workstation. We tried using global.json to limit to only released versions, but no go. Any thoughts on how to fix this?
global.json
{
"sdk": {
"allowPrerelease": false
}
}
Change your framework target to a supported one, like "net5.0"
Get Outlook for Androidhttps://aka.ms/AAb9ysg
From: Richard Crane @.>
Sent: Wednesday, March 17, 2021 12:21:39 PM
To: rr-wfm/MSBuild.Sdk.SqlProj @.>
Cc: Subscribed @.***>
Subject: [rr-wfm/MSBuild.Sdk.SqlProj] Build error when .NET 6 preview is installed. Any suggestions to correct this is appreciated. (#125)
We get the following error when Visual Studio 2019 Preview is installed on our workstation. We tried using global.json to limit to only released versions, but no go. Any thoughts on how to fix this?
global.json
{
"sdk": {
"allowPrerelease": false
}
}
[image]https://user-images.githubusercontent.com/15870558/111459469-16a3b580-86f1-11eb-9827-5042fda3c946.png
—
You are receiving this because you are subscribed to this thread.
Reply to this email directly, view it on GitHubhttps://github.com/rr-wfm/MSBuild.Sdk.SqlProj/issues/125, or unsubscribehttps://github.com/notifications/unsubscribe-auth/AA7Z3YZRGCSPUBPSAG7TQ7LTECGEHANCNFSM4ZKMP4KQ.
@ChiefInnovator I would expect that global.json to fix this issue. Did you restart Visual Studio after adding the global.json? I believe it doesn't immediately take effect. If that doesn't work, could you share the output of dotnet --info running from the command line in the folder of your project?
I had a similar issue but with "netcoreapp5.0"
Adding the global.json worked for me.
Example of the content:
{
"sdk": {
"version": "3.1.407"
}
}
Find out more here about global.json
@ratynchukD As of version 1.6.0 there should be support for .NET 5 so it would work with any SDK version 5.0. If that is not the case we would like to understand why it doesn't work for you.
We don't support .NET 6 previews just yet. Since .NET 6 is still early in the preview phase I don't expect many people will have an issue with this, but please upvote this issue if you feel this is important to you.
@jmezach This needs to be discussed in my team since we are using the 1.2.0 version. I'm new to the project and might be the first who faced with the issue.
Regarding .NET 6 preview - I don't feel this is important to my team or me.
Thank you!
I don't think there's much more we can do here so I'll go ahead and close this. If you feel there's still an issue here feel free to re-open.
|
gharchive/issue
| 2021-03-17T11:21:20 |
2025-04-01T04:35:44.978670
|
{
"authors": [
"ChiefInnovator",
"ErikEJ",
"jmezach",
"ratynchukD"
],
"repo": "rr-wfm/MSBuild.Sdk.SqlProj",
"url": "https://github.com/rr-wfm/MSBuild.Sdk.SqlProj/issues/125",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
645959307
|
Stack Overflow when use Consumer in dialog
Future<void> _showDialogAddGhiChu(BuildContext context) async {
return showDialog<void>(
context: context,
barrierDismissible: false, // user must tap button!
builder: (_) {
return Consumer((ct, read) {
final clickDate = read(mainStateNotifier).clickDate;
return AddNote(
date: clickDate,
context: ct,
); //magic ^_^
});
},
);
}
error
Another exception was thrown: Stack Overflow
════════ Exception caught by widgets library ═══════════════════════════════════════════════════════
The following StackOverflowError was thrown building Consumer(dirty, dependencies: [ProviderStateOwnerScope], state: _ConsumerState#34eeb):
Stack Overflow
The relevant error-causing widget was:
Consumer file:///Users/tbm98/dev/flutter_app/student-social/lib/presentation/screens/main/main_page.dart:132:16
When the exception was thrown, this was the stack:
#0 MainState.getClickDate (package:studentsocial/presentation/screens/main/main_state.dart:35:3)
#1 MainState.getClickDate (package:studentsocial/presentation/screens/main/main_state.dart:35:32)
#2 MainState.getClickDate (package:studentsocial/presentation/screens/main/main_state.dart:35:32)
#3 MainState.getClickDate (package:studentsocial/presentation/screens/main/main_state.dart:35:32)
#4 MainState.getClickDate (package:studentsocial/presentation/screens/main/main_state.dart:35:32)
...
════════════════════════════════════════════════════════════════════════════════════════════════════
sorry. my getter called itself
|
gharchive/issue
| 2020-06-26T02:18:42 |
2025-04-01T04:35:44.998973
|
{
"authors": [
"tbm98"
],
"repo": "rrousselGit/river_pod",
"url": "https://github.com/rrousselGit/river_pod/issues/15",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1590822366
|
Warn against AsyncNotifier.update misuses
class Example extends AsyncNotifier<int> {
...
void fn() {
state = AsyncLoading(); // KO, "update" will likely never complete
update((data) => ...);
}
void fn() {
update((data) {
state = AsyncLoading(); // OK
...
});
}
void fn() {
update((unused) => ...); // the parameter should be used. Otherwise use AsyncValue.guard
}
}
@rrousselGit
I am trying to update the state of a FamilyAsyncNotifier partially, I just want to know whether I am doing it right. Also in this case of updating a asyncvalue partially, I want to know what's differences in using whenData, AsyncGuard.value and update method . It would be a immense help, if you provide some clarity on this.
update((state) async {
List students =
await ref.read(collegeProvider).getCollegeById(id: this.arg);
return state.copyWith(students: students);
});
With mutations, this is really low-priority. Closing
|
gharchive/issue
| 2023-02-19T19:29:07 |
2025-04-01T04:35:45.001282
|
{
"authors": [
"rrousselGit",
"yknevenky"
],
"repo": "rrousselGit/riverpod",
"url": "https://github.com/rrousselGit/riverpod/issues/2205",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
803344574
|
rmfuse on OSX
Hi,
I really love the idea of rmfuse and I would love to use it on my Mac.
It seems that that pyfuse3 is not supported on OSX:
https://github.com/osxfuse/osxfuse/issues/390
https://github.com/libfuse/pyfuse3/issues/29
This is the error I get when doing pip install rmfuse
Collecting rmfuse
Using cached rmfuse-0.1.1-py3-none-any.whl (8.7 kB)
Collecting rmcl<0.4.0,>=0.3.0
Using cached rmcl-0.3.0-py3-none-any.whl (16 kB)
Collecting rmrl<0.2.0,>=0.1.2
Using cached rmrl-0.1.2-py3-none-any.whl (518 kB)
Collecting pyfuse3<4.0.0,>=3.2.0
Using cached pyfuse3-3.2.0.tar.gz (465 kB)
ERROR: Command errored out with exit status 2:
command: /Users/schneider/anaconda3/bin/python3.8 -c 'import sys, setuptools, tokenize; sys.argv[0] = '"'"'/private/var/folders/d_/755ltxq50klbwlljp6rsqgzc0000gn/T/pip-install-9z2bjy84/pyfuse3/setup.py'"'"'; __file__='"'"'/private/var/folders/d_/755ltxq50klbwlljp6rsqgzc0000gn/T/pip-install-9z2bjy84/pyfuse3/setup.py'"'"';f=getattr(tokenize, '"'"'open'"'"', open)(__file__);code=f.read().replace('"'"'\r\n'"'"', '"'"'\n'"'"');f.close();exec(compile(code, __file__, '"'"'exec'"'"'))' egg_info --egg-base /private/var/folders/d_/755ltxq50klbwlljp6rsqgzc0000gn/T/pip-pip-egg-info-9ju778vx
cwd: /private/var/folders/d_/755ltxq50klbwlljp6rsqgzc0000gn/T/pip-install-9z2bjy84/pyfuse3/
Complete output (4 lines):
Package fuse3 was not found in the pkg-config search path.
Perhaps you should add the directory containing `fuse3.pc'
to the PKG_CONFIG_PATH environment variable
No package 'fuse3' found
----------------------------------------
ERROR: Command errored out with exit status 2: python setup.py egg_info Check the logs for full command output.
Urgh. I was afraid of something like this. I was hoping that a version number > 3.0 on MacFuse indicated that it supported FUSE3, but apparently not. From those issues, I don't hold much hope of MacFuse ever making it to FUSE3.
I don't think there's anything I'm doing that requires FUSE3 specifically; that requirement is coming from the fact that I chose to use pyfuse3, which seemed to be the most active and best documented library. There were a bunch of other bindings I found out there, but the only other one I seriously considered was python-fuse. This is for FUSE2, but I don't know whether it works with MacFuse. If someone would like to check and let me know, we could see if converting to the python-fuse library makes sense. (A simple test would be to pip install fuse-python and then see if this example works.)
All that said, this particular error was because the pyfuse3 module needs to be compiled during installation, and it couldn't find the fuse3 headers. The fuse-python module is the same way, so I don't know if this would also hit the same problem.
Yes I think libfuse2 should work with osx. I tested the example script that you provided (after I successfully installed fuse-python). I get an error,
fuse: no mount point
Traceback (most recent call last):
File "hello.py", line 91, in <module>
main()
File "hello.py", line 88, in main
server.main()
File "/Users/schneider/anaconda3/lib/python3.8/site-packages/fuse.py", line 754, in main
main(**d)
fuse.FuseError: filesystem initialization failed
which seems to be related to the example script, rather than fuse itself. However, since
import fuse
from fuse import Fuse
works. I have no doubt that using libfuse2 would fix the problem for OSX.
However, I suppose that libfuse3 is more stable.... Its a pity that macfuse did not update libfuse....
Thanks for checking. It appears that pyfuse3 forked off from python-llfuse some time ago. That project is now marked as undeveloped, but the API is still pretty similar. It has some hints on building solutions compatible with both libraries. I'll take a look and see if that's feasible here.
Good news -- it seems to work. See #6 for a branch that implements support for llfuse. If you could try it out and let me know if it works with MacFuse, it'd be much appreciated. You'll need llfuse (pip install llfuse). If you're working from git, you can just pull the branch from #6. If not, you should be able to pip install the attached sdist file.
rmfuse-0.1.2a0.tar.gz
That sounds awesome! I will test it today!
It installed successfully!
These steps were necessary:
brew install macfuse
One needs to allow a kernel extension and reboot.
brew install pkg-config
I messed up the onetime code and now get the error:
FileNotFoundError: [Errno 2] No such file or directory: '/Users/schneider/.cache/rmfuse'
Seems to be unrelated to rmfuse.
That is a bug in RMfuse--it stores file metadata in a cache database to avoid hitting the API too often. But I didn't make sure the ~/.cache directory existed before trying to create the db file. I'll get a fix in for that soon, but in the meantime, you ought to be able to fix this with a mkdir ~/.cache on the command line.
thanks for getting back!
I started rmfuse:
(base) schneider@Aarons-MBP ~ % rmfuse remarkable
(base) schneider@Aarons-MBP ~ % cd remarkable
(base) schneider@Aarons-MBP . % ls
ls: .: Function not implemented
(base) schneider@Aarons-MBP . % open .
2021-02-14 12:52:14.104 open[1833:36366] CFURLCopyResourcePropertyForKey failed because it was passed an URL which has no scheme
No application knows how to open . (Error Domain=NSOSStatusErrorDomain Code=-10814 "kLSApplicationNotFoundErr: E.g. no application claims the file" UserInfo={_LSLine=1483, _LSFunction=runEvaluator}).
(base) schneider@Aarons-MBP . % cp ~/Desktop/Chapter_3.pdf ~/remarkable
cp: /Users/schneider/remarkable: Function not implemented
Is there a way to renew the authentication token? I tried to remove the access on my.remarkable.com and I tried to freshly install rmfuse. But now I can't link it again.
(base) schneider@Aarons-MBP ~ % rmfuse remarkable
Traceback (most recent call last):
File "/usr/local/bin/rmfuse", line 8, in <module>
sys.exit(main())
File "/usr/local/lib/python3.9/site-packages/rmfuse/fuse.py", line 452, in main
fuse.main(workers=1)
File "src/fuse_api.pxi", line 327, in llfuse.main
File "src/handlers.pxi", line 80, in llfuse.fuse_getattr
File "src/handlers.pxi", line 81, in llfuse.fuse_getattr
File "/usr/local/lib/python3.9/site-packages/rmfuse/fuselib.py", line 25, in decorated
return trio.run(runner)
File "/usr/local/lib/python3.9/site-packages/trio/_core/_run.py", line 1932, in run
raise runner.main_task_outcome.error
File "/usr/local/lib/python3.9/site-packages/rmfuse/fuselib.py", line 24, in runner
return await afunc(*args, **kw)
File "/usr/local/lib/python3.9/site-packages/rmfuse/fuse.py", line 226, in getattr
return await self._getattr(inode, ctx)
File "/usr/local/lib/python3.9/site-packages/rmfuse/fuse.py", line 195, in _getattr
item = await self.get_by_id(self.get_id(inode))
File "/usr/local/lib/python3.9/site-packages/rmfuse/fuse.py", line 139, in get_by_id
return await Item.get_by_id(id_)
File "/usr/local/lib/python3.9/site-packages/rmcl/items.py", line 46, in get_by_id
return await (await api.get_client()).get_by_id(id_)
File "/usr/local/lib/python3.9/site-packages/rmcl/api.py", line 245, in get_by_id
await self.update_items()
File "/usr/local/lib/python3.9/site-packages/rmcl/api.py", line 212, in update_items
response = await self.request('GET', '/document-storage/json/2/docs')
File "/usr/local/lib/python3.9/site-packages/rmcl/api.py", line 117, in request
await self.renew_token()
File "/usr/local/lib/python3.9/site-packages/rmcl/api.py", line 208, in renew_token
raise AuthError("Can't renew token: {e}".format(
rmcl.exceptions.AuthError: Can't renew token: 401
Second question first: the tokens are stored in ~/.config/rmcl/config.json. Delete that file, and you'll be back to pristine, as far as the API is concerned. Then you can get a new one-time code and get access again.
As for the first problem: it appears that your system is trying to call a filesystem method that isn't implemented in RMfuse yet. (There are quite a few, and I implemented them as I found they were needed in testing.) It's not clear to me which one it is, but if you run rmfuse with the -vv output, you'll get a whole lot of debugging information, including each function being called. If you can capture that output and post it here, it'll help narrow down what the problem is.
Thanks!
Thanks! Here we go:
(rmfuse) schneider@Aarons-MacBook-Pro ~ % rmfuse -vv remarkable
FUSE library version: 2.9.9
unique: 2, opcode: INIT (26), nodeid: 0, insize: 56, pid: 5768
INIT: 7.19
flags=0xee000008
max_readahead=0x00100000
INIT: 7.19
flags=0x00000020
max_readahead=0x00100000
max_write=0x02000000
max_background=0
congestion_threshold=0
unique: 2, success, outsize: 40
unique: 2, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 0
unique: 2, error: -78 (Function not implemented), outsize: 16
unique: 3, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5768
unique: 3, error: -78 (Function not implemented), outsize: 16
unique: 2, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 92
unique: 2, error: -78 (Function not implemented), outsize: 16
unique: 3, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 130
unique: 3, error: -78 (Function not implemented), outsize: 16
unique: 2, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 130
unique: 2, error: -78 (Function not implemented), outsize: 16
unique: 3, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 130
unique: 3, error: -78 (Function not implemented), outsize: 16
unique: 2, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 130
unique: 2, error: -78 (Function not implemented), outsize: 16
unique: 3, opcode: GETATTR (3), nodeid: 1, insize: 56, pid: 89
unique: 3, success, outsize: 136
unique: 2, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5340
unique: 2, error: -78 (Function not implemented), outsize: 16
unique: 3, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 516
unique: 3, error: -78 (Function not implemented), outsize: 16
unique: 4, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 4145
unique: 4, error: -78 (Function not implemented), outsize: 16
unique: 5, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408
unique: 5, error: -78 (Function not implemented), outsize: 16
unique: 6, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 141
unique: 6, error: -78 (Function not implemented), outsize: 16
unique: 7, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 549
unique: 7, error: -78 (Function not implemented), outsize: 16
unique: 8, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 425
unique: 8, error: -78 (Function not implemented), outsize: 16
unique: 9, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 487
unique: 9, error: -78 (Function not implemented), outsize: 16
unique: 10, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 467
unique: 10, error: -78 (Function not implemented), outsize: 16
unique: 11, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 405
unique: 11, error: -78 (Function not implemented), outsize: 16
unique: 2, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 407
unique: 2, error: -78 (Function not implemented), outsize: 16
unique: 12, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 545
unique: 12, error: -78 (Function not implemented), outsize: 16
unique: 13, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 442
unique: 13, error: -78 (Function not implemented), outsize: 16
unique: 14, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 531
unique: 14, error: -78 (Function not implemented), outsize: 16
unique: 15, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408
unique: 15, error: -78 (Function not implemented), outsize: 16
unique: 16, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 89
unique: 16, error: -78 (Function not implemented), outsize: 16
unique: 17, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 387
unique: 17, error: -78 (Function not implemented), outsize: 16
unique: 18, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408
unique: 18, error: -78 (Function not implemented), outsize: 16
unique: 19, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 505
unique: 19, error: -78 (Function not implemented), outsize: 16
unique: 3, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 391
unique: 3, error: -78 (Function not implemented), outsize: 16
unique: 4, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5755
unique: 4, error: -78 (Function not implemented), outsize: 16
unique: 5, opcode: GETXATTR (22), nodeid: 1, insize: 77, pid: 408
unique: 5, error: -78 (Function not implemented), outsize: 16
unique: 6, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408
unique: 6, error: -2 (No such file or directory), outsize: 16
unique: 7, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 387
unique: 7, error: -2 (No such file or directory), outsize: 16
unique: 8, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408
unique: 8, error: -78 (Function not implemented), outsize: 16
unique: 9, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408
unique: 9, error: -78 (Function not implemented), outsize: 16
unique: 10, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408
unique: 10, error: -78 (Function not implemented), outsize: 16
unique: 11, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408
unique: 11, error: -78 (Function not implemented), outsize: 16
unique: 2, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408
unique: 2, error: -78 (Function not implemented), outsize: 16
unique: 12, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 387
unique: 12, error: -2 (No such file or directory), outsize: 16
unique: 13, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408
unique: 13, error: -78 (Function not implemented), outsize: 16
unique: 14, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 387
unique: 14, error: -78 (Function not implemented), outsize: 16
unique: 15, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408
unique: 15, error: -2 (No such file or directory), outsize: 16
unique: 16, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408
unique: 16, error: -2 (No such file or directory), outsize: 16
unique: 17, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408
unique: 17, error: -2 (No such file or directory), outsize: 16
unique: 18, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408
unique: 18, error: -2 (No such file or directory), outsize: 16
unique: 19, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408
unique: 19, error: -2 (No such file or directory), outsize: 16
unique: 3, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5771
unique: 3, error: -78 (Function not implemented), outsize: 16
unique: 4, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5771
unique: 4, error: -78 (Function not implemented), outsize: 16
unique: 5, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5771
unique: 5, error: -78 (Function not implemented), outsize: 16
unique: 6, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5771
unique: 6, error: -78 (Function not implemented), outsize: 16
unique: 7, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5774
unique: 7, error: -78 (Function not implemented), outsize: 16
unique: 8, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5774
unique: 8, error: -78 (Function not implemented), outsize: 16
unique: 9, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5774
unique: 9, error: -78 (Function not implemented), outsize: 16
unique: 10, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5774
unique: 10, error: -78 (Function not implemented), outsize: 16
unique: 11, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5776
unique: 11, error: -78 (Function not implemented), outsize: 16
unique: 2, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5776
unique: 2, error: -78 (Function not implemented), outsize: 16
unique: 12, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5776
unique: 12, error: -78 (Function not implemented), outsize: 16
unique: 13, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5776
unique: 13, error: -78 (Function not implemented), outsize: 16
unique: 14, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 5777
unique: 14, error: -2 (No such file or directory), outsize: 16
unique: 15, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5777
unique: 15, error: -78 (Function not implemented), outsize: 16
unique: 16, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5777
unique: 16, error: -78 (Function not implemented), outsize: 16
unique: 17, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5777
unique: 17, error: -78 (Function not implemented), outsize: 16
unique: 18, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5777
unique: 18, error: -78 (Function not implemented), outsize: 16
unique: 19, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5777
unique: 19, error: -78 (Function not implemented), outsize: 16
unique: 3, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 5777
unique: 3, error: -2 (No such file or directory), outsize: 16
unique: 4, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 5777
unique: 4, error: -2 (No such file or directory), outsize: 16
unique: 5, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5777
unique: 5, error: -78 (Function not implemented), outsize: 16
unique: 6, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5777
unique: 6, error: -78 (Function not implemented), outsize: 16
unique: 7, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5777
unique: 7, error: -78 (Function not implemented), outsize: 16
unique: 8, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5777
unique: 8, error: -78 (Function not implemented), outsize: 16
unique: 9, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5777
unique: 9, error: -78 (Function not implemented), outsize: 16
unique: 10, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 442
unique: 10, error: -2 (No such file or directory), outsize: 16
I see a lot of failed calls to statfs, which hadn't been implemented. I've added a simple implementation in #6. Can you give it a try. Updated sdist file attached, if that helps.
rmfuse-0.1.2a0.tar.gz
Awesome! it works for copying stuff from the fuse drive to the local machine. But right now its giving me an error, when I try to copy a pdf to the fuse drive. Here is a new log:
(rmfuse) schneider@Aarons-MacBook-Pro ~ % rmfuse remarkable -vv
FUSE library version: 2.9.9
unique: 2, opcode: INIT (26), nodeid: 0, insize: 56, pid: 65083
INIT: 7.19
flags=0xee000008
max_readahead=0x00100000
INIT: 7.19
flags=0x00000020
max_readahead=0x00100000
max_write=0x02000000
max_background=0
congestion_threshold=0
unique: 2, success, outsize: 40
unique: 3, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 65083
unique: 3, success, outsize: 96
unique: 2, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 0
unique: 2, success, outsize: 96
unique: 3, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 92
unique: 3, success, outsize: 96
unique: 2, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 130
unique: 2, success, outsize: 96
unique: 3, opcode: GETATTR (3), nodeid: 1, insize: 56, pid: 89
unique: 3, success, outsize: 136
unique: 2, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 130
unique: 2, success, outsize: 96
unique: 3, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 130
unique: 3, success, outsize: 96
unique: 2, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408
unique: 2, success, outsize: 96
unique: 3, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408
unique: 3, success, outsize: 96
unique: 4, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 405
unique: 4, success, outsize: 96
unique: 5, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408
unique: 5, success, outsize: 96
unique: 6, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 407
unique: 6, success, outsize: 96
unique: 7, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 442
unique: 7, success, outsize: 96
unique: 8, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 467
unique: 8, success, outsize: 96
unique: 9, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 425
unique: 9, success, outsize: 96
unique: 10, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 141
unique: 10, success, outsize: 96
unique: 11, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 8572
unique: 11, success, outsize: 96
unique: 12, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 545
unique: 12, success, outsize: 96
unique: 13, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 91827
unique: 13, success, outsize: 96
unique: 14, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 516
unique: 14, success, outsize: 96
unique: 2, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 549
unique: 2, success, outsize: 96
unique: 15, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 89
unique: 15, success, outsize: 96
unique: 16, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 8568
unique: 16, success, outsize: 96
unique: 17, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 8568
unique: 17, success, outsize: 96
unique: 18, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 64910
unique: 18, success, outsize: 96
unique: 19, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 391
unique: 19, success, outsize: 96
unique: 20, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 64914
unique: 20, success, outsize: 96
unique: 21, opcode: GETXATTR (22), nodeid: 1, insize: 77, pid: 408
unique: 21, error: -78 (Function not implemented), outsize: 16
unique: 3, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 7591
unique: 3, success, outsize: 96
unique: 4, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 7570
unique: 4, success, outsize: 96
unique: 5, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408
unique: 5, error: -2 (No such file or directory), outsize: 16
unique: 6, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 387
unique: 6, error: -2 (No such file or directory), outsize: 16
unique: 7, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408
unique: 7, success, outsize: 96
unique: 8, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408
unique: 8, success, outsize: 96
unique: 9, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 387
unique: 9, error: -2 (No such file or directory), outsize: 16
unique: 10, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408
unique: 10, success, outsize: 96
unique: 11, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 442
unique: 11, error: -2 (No such file or directory), outsize: 16
unique: 12, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408
unique: 12, success, outsize: 96
unique: 13, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408
unique: 13, success, outsize: 96
unique: 14, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 387
unique: 14, success, outsize: 96
unique: 2, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408
unique: 2, success, outsize: 96
unique: 15, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408
unique: 15, error: -2 (No such file or directory), outsize: 16
unique: 16, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408
unique: 16, error: -2 (No such file or directory), outsize: 16
unique: 17, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408
unique: 17, error: -2 (No such file or directory), outsize: 16
unique: 18, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408
unique: 18, error: -2 (No such file or directory), outsize: 16
unique: 19, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408
unique: 19, error: -2 (No such file or directory), outsize: 16
unique: 20, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408
unique: 20, error: -2 (No such file or directory), outsize: 16
unique: 21, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408
unique: 21, error: -2 (No such file or directory), outsize: 16
unique: 3, opcode: OPENDIR (27), nodeid: 1, insize: 48, pid: 408
unique: 3, success, outsize: 32
unique: 4, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408
unique: 4, success, outsize: 576
unique: 5, opcode: LOOKUP (1), nodeid: 1, insize: 46, pid: 408
unique: 5, success, outsize: 160
unique: 6, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408
unique: 6, error: -2 (No such file or directory), outsize: 16
unique: 7, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408
unique: 7, error: -2 (No such file or directory), outsize: 16
unique: 8, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408
unique: 8, error: -2 (No such file or directory), outsize: 16
unique: 9, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408
unique: 9, error: -2 (No such file or directory), outsize: 16
unique: 10, opcode: LOOKUP (1), nodeid: 1, insize: 47, pid: 408
unique: 10, success, outsize: 160
unique: 11, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408
unique: 11, error: -2 (No such file or directory), outsize: 16
unique: 12, opcode: LOOKUP (1), nodeid: 1, insize: 49, pid: 408
unique: 12, error: -2 (No such file or directory), outsize: 16
unique: 13, opcode: GETATTR (3), nodeid: 1, insize: 56, pid: 408
unique: 13, success, outsize: 136
unique: 14, opcode: LOOKUP (1), nodeid: 1, insize: 46, pid: 408
unique: 14, success, outsize: 160
unique: 2, opcode: CREATE (35), nodeid: 1, insize: 66, pid: 408
unique: 2, success, outsize: 176
unique: 15, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408
unique: 15, error: -2 (No such file or directory), outsize: 16
unique: 16, opcode: LOOKUP (1), nodeid: 1, insize: 52, pid: 408
unique: 16, error: -2 (No such file or directory), outsize: 16
unique: 17, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408
unique: 17, success, outsize: 160
unique: 18, opcode: FLUSH (25), nodeid: 16, insize: 64, pid: 408
unique: 18, error: -78 (Function not implemented), outsize: 16
unique: 19, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408
unique: 19, error: -2 (No such file or directory), outsize: 16
unique: 20, opcode: RELEASE (18), nodeid: 16, insize: 64, pid: 408
ERROR:rmfuse.fuse:Error: Not a PDF or EPUB file
unique: 20, error: -5 (Input/output error), outsize: 16
unique: 21, opcode: LOOKUP (1), nodeid: 1, insize: 49, pid: 408
unique: 21, success, outsize: 160
unique: 3, opcode: GETATTR (3), nodeid: 16, insize: 56, pid: 408
unique: 3, success, outsize: 136
unique: 4, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 4, error: -2 (No such file or directory), outsize: 16
unique: 5, opcode: LOOKUP (1), nodeid: 1, insize: 52, pid: 408
unique: 5, error: -2 (No such file or directory), outsize: 16
unique: 6, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 6, error: -2 (No such file or directory), outsize: 16
unique: 7, opcode: LOOKUP (1), nodeid: 1, insize: 52, pid: 408
unique: 7, error: -2 (No such file or directory), outsize: 16
unique: 8, opcode: LOOKUP (1), nodeid: 1, insize: 57, pid: 408
unique: 8, success, outsize: 160
unique: 9, opcode: SETXATTR (21), nodeid: 16, insize: 109, pid: 408
unique: 9, error: -78 (Function not implemented), outsize: 16
unique: 10, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 408
unique: 10, error: -2 (No such file or directory), outsize: 16
unique: 11, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 408
unique: 11, error: -2 (No such file or directory), outsize: 16
unique: 12, opcode: LOOKUP (1), nodeid: 1, insize: 46, pid: 408
unique: 12, success, outsize: 160
unique: 13, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408
unique: 13, error: -2 (No such file or directory), outsize: 16
unique: 14, opcode: LOOKUP (1), nodeid: 1, insize: 52, pid: 408
unique: 14, error: -2 (No such file or directory), outsize: 16
unique: 2, opcode: LOOKUP (1), nodeid: 1, insize: 109, pid: 408
unique: 2, success, outsize: 160
unique: 15, opcode: GETATTR (3), nodeid: 1, insize: 56, pid: 408
unique: 15, success, outsize: 136
unique: 16, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408
unique: 16, error: -2 (No such file or directory), outsize: 16
unique: 17, opcode: CREATE (35), nodeid: 1, insize: 68, pid: 408
unique: 17, success, outsize: 176
unique: 18, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408
unique: 18, error: -2 (No such file or directory), outsize: 16
unique: 19, opcode: WRITE (16), nodeid: 17, insize: 4176, pid: 408
unique: 19, success, outsize: 24
unique: 20, opcode: LOOKUP (1), nodeid: 1, insize: 49, pid: 408
unique: 20, success, outsize: 160
unique: 21, opcode: GETATTR (3), nodeid: 17, insize: 56, pid: 408
unique: 21, success, outsize: 136
unique: 3, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 3, error: -2 (No such file or directory), outsize: 16
unique: 4, opcode: RELEASE (18), nodeid: 17, insize: 64, pid: 408
ERROR:rmfuse.fuse:Error: Not a PDF or EPUB file
unique: 4, error: -5 (Input/output error), outsize: 16
unique: 5, opcode: LOOKUP (1), nodeid: 1, insize: 49, pid: 408
unique: 5, success, outsize: 160
unique: 6, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 6, error: -2 (No such file or directory), outsize: 16
unique: 7, opcode: GETATTR (3), nodeid: 1, insize: 56, pid: 408
unique: 7, success, outsize: 136
unique: 8, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408
unique: 8, success, outsize: 160
unique: 9, opcode: LOOKUP (1), nodeid: 1, insize: 52, pid: 408
unique: 9, error: -2 (No such file or directory), outsize: 16
unique: 10, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408
unique: 10, error: -2 (No such file or directory), outsize: 16
unique: 11, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408
unique: 11, success, outsize: 160
unique: 12, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408
unique: 12, error: -2 (No such file or directory), outsize: 16
unique: 13, opcode: LOOKUP (1), nodeid: 1, insize: 46, pid: 408
unique: 13, success, outsize: 160
unique: 14, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408
unique: 14, error: -2 (No such file or directory), outsize: 16
unique: 2, opcode: LOOKUP (1), nodeid: 1, insize: 49, pid: 408
unique: 2, success, outsize: 160
unique: 15, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408
unique: 15, success, outsize: 96
unique: 16, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 16, error: -2 (No such file or directory), outsize: 16
unique: 17, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408
unique: 17, success, outsize: 16
unique: 18, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408
unique: 18, success, outsize: 16
unique: 19, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408
unique: 19, success, outsize: 16
unique: 20, opcode: RELEASEDIR (29), nodeid: 1, insize: 64, pid: 408
unique: 20, error: -78 (Function not implemented), outsize: 16
unique: 21, opcode: GETATTR (3), nodeid: 1, insize: 56, pid: 408
unique: 21, success, outsize: 136
unique: 3, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408
unique: 3, success, outsize: 96
unique: 4, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408
unique: 4, error: -2 (No such file or directory), outsize: 16
unique: 5, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408
unique: 5, error: -2 (No such file or directory), outsize: 16
unique: 6, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408
unique: 6, error: -2 (No such file or directory), outsize: 16
unique: 7, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408
unique: 7, error: -2 (No such file or directory), outsize: 16
unique: 8, opcode: LOOKUP (1), nodeid: 1, insize: 52, pid: 408
unique: 8, error: -2 (No such file or directory), outsize: 16
unique: 9, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 408
unique: 9, error: -2 (No such file or directory), outsize: 16
unique: 10, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 9690
unique: 10, success, outsize: 96
unique: 11, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 408
unique: 11, error: -2 (No such file or directory), outsize: 16
unique: 12, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 9690
unique: 12, success, outsize: 96
unique: 13, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408
unique: 13, error: -2 (No such file or directory), outsize: 16
unique: 14, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 14, error: -2 (No such file or directory), outsize: 16
unique: 2, opcode: LOOKUP (1), nodeid: 1, insize: 62, pid: 9690
unique: 2, error: -2 (No such file or directory), outsize: 16
unique: 15, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 15, error: -2 (No such file or directory), outsize: 16
unique: 16, opcode: LOOKUP (1), nodeid: 1, insize: 57, pid: 9690
unique: 16, error: -2 (No such file or directory), outsize: 16
unique: 17, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 17, error: -2 (No such file or directory), outsize: 16
unique: 18, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408
unique: 18, error: -2 (No such file or directory), outsize: 16
unique: 19, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408
unique: 19, error: -2 (No such file or directory), outsize: 16
unique: 20, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408
unique: 20, error: -2 (No such file or directory), outsize: 16
unique: 21, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408
unique: 21, error: -2 (No such file or directory), outsize: 16
unique: 3, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408
unique: 3, error: -2 (No such file or directory), outsize: 16
unique: 4, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 4, error: -2 (No such file or directory), outsize: 16
unique: 5, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408
unique: 5, error: -2 (No such file or directory), outsize: 16
unique: 6, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408
unique: 6, error: -2 (No such file or directory), outsize: 16
unique: 7, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408
unique: 7, error: -2 (No such file or directory), outsize: 16
unique: 8, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 8, error: -2 (No such file or directory), outsize: 16
unique: 9, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408
unique: 9, error: -2 (No such file or directory), outsize: 16
unique: 10, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408
unique: 10, error: -2 (No such file or directory), outsize: 16
unique: 11, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408
unique: 11, error: -2 (No such file or directory), outsize: 16
unique: 12, opcode: OPENDIR (27), nodeid: 1, insize: 48, pid: 408
unique: 12, success, outsize: 32
unique: 13, opcode: RELEASEDIR (29), nodeid: 1, insize: 64, pid: 408
unique: 13, error: -78 (Function not implemented), outsize: 16
unique: 14, opcode: GETATTR (3), nodeid: 1, insize: 56, pid: 408
unique: 14, success, outsize: 136
unique: 2, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408
unique: 2, error: -2 (No such file or directory), outsize: 16
unique: 15, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 9690
unique: 15, success, outsize: 96
unique: 16, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 9690
unique: 16, success, outsize: 96
unique: 17, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 9690
unique: 17, success, outsize: 96
unique: 18, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408
unique: 18, error: -2 (No such file or directory), outsize: 16
unique: 19, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 9690
unique: 19, error: -2 (No such file or directory), outsize: 16
unique: 20, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 9690
unique: 20, error: -2 (No such file or directory), outsize: 16
unique: 21, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 9690
unique: 21, success, outsize: 96
unique: 3, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408
unique: 3, error: -2 (No such file or directory), outsize: 16
unique: 4, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 9690
unique: 4, error: -2 (No such file or directory), outsize: 16
unique: 5, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 9690
unique: 5, error: -2 (No such file or directory), outsize: 16
unique: 6, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 9690
unique: 6, success, outsize: 96
unique: 7, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408
unique: 7, error: -2 (No such file or directory), outsize: 16
unique: 8, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 9690
unique: 8, success, outsize: 96
unique: 9, opcode: LOOKUP (1), nodeid: 1, insize: 62, pid: 9690
unique: 9, error: -2 (No such file or directory), outsize: 16
unique: 10, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408
unique: 10, error: -2 (No such file or directory), outsize: 16
unique: 11, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 9690
unique: 11, error: -2 (No such file or directory), outsize: 16
unique: 12, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 9690
unique: 12, error: -2 (No such file or directory), outsize: 16
unique: 13, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 9690
unique: 13, success, outsize: 96
unique: 14, opcode: LOOKUP (1), nodeid: 1, insize: 57, pid: 9690
unique: 14, error: -2 (No such file or directory), outsize: 16
unique: 2, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408
unique: 2, error: -2 (No such file or directory), outsize: 16
unique: 15, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 9690
unique: 15, error: -2 (No such file or directory), outsize: 16
unique: 16, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408
unique: 16, error: -2 (No such file or directory), outsize: 16
unique: 17, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 9690
unique: 17, success, outsize: 96
unique: 18, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408
unique: 18, error: -2 (No such file or directory), outsize: 16
unique: 19, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 9690
unique: 19, error: -2 (No such file or directory), outsize: 16
unique: 20, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408
unique: 20, error: -2 (No such file or directory), outsize: 16
unique: 21, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 9690
unique: 21, success, outsize: 96
unique: 3, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 9690
unique: 3, success, outsize: 96
unique: 4, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 9690
unique: 4, success, outsize: 96
unique: 5, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 9690
unique: 5, success, outsize: 96
unique: 6, opcode: OPEN (14), nodeid: 7, insize: 48, pid: 64903
unique: 6, success, outsize: 32
unique: 7, opcode: SETATTR (4), nodeid: 7, insize: 168, pid: 64903
unique: 7, success, outsize: 136
unique: 8, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 9690
unique: 8, success, outsize: 96
unique: 9, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903
INFO:rmcl.items:Rendering <Document "Quick sheets">: 10.0%
INFO:rmcl.items:Rendering <Document "Quick sheets">: 20.0%
INFO:rmcl.items:Rendering <Document "Quick sheets">: 30.0%
INFO:rmcl.items:Rendering <Document "Quick sheets">: 40.0%
INFO:rmcl.items:Rendering <Document "Quick sheets">: 50.0%
INFO:rmcl.items:Rendering <Document "Quick sheets">: 60.0%
INFO:rmcl.items:Rendering <Document "Quick sheets">: 70.0%
INFO:rmcl.items:Rendering <Document "Quick sheets">: 80.0%
INFO:rmcl.items:Rendering <Document "Quick sheets">: 90.0%
INFO:rmcl.items:Rendering <Document "Quick sheets">: 100.0%
INFO:rmrl.render:exported pdf
unique: 9, success, outsize: 4112
unique: 10, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 9690
unique: 10, error: -2 (No such file or directory), outsize: 16
unique: 9, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903
unique: 9, success, outsize: 4112
unique: 11, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 9690
unique: 11, error: -2 (No such file or directory), outsize: 16
unique: 9, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903
unique: 9, success, outsize: 4112
unique: 12, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 9690
unique: 12, success, outsize: 96
unique: 9, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903
unique: 9, success, outsize: 4112
unique: 13, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 9690
unique: 13, error: -2 (No such file or directory), outsize: 16
unique: 9, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903
unique: 9, success, outsize: 4112
unique: 14, opcode: OPEN (14), nodeid: 6, insize: 48, pid: 9690
unique: 14, success, outsize: 32
unique: 9, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903
unique: 9, success, outsize: 4112
unique: 2, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 9690
unique: 2, error: -2 (No such file or directory), outsize: 16
unique: 9, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903
unique: 9, success, outsize: 4112
unique: 15, opcode: RELEASE (18), nodeid: 6, insize: 64, pid: 9690
unique: 15, success, outsize: 16
unique: 9, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903
unique: 9, success, outsize: 4112
unique: 16, opcode: GETATTR (3), nodeid: 6, insize: 56, pid: 9690
unique: 16, success, outsize: 136
unique: 17, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903
unique: 17, success, outsize: 4112
unique: 18, opcode: OPEN (14), nodeid: 6, insize: 48, pid: 9690
unique: 18, success, outsize: 32
unique: 17, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903
unique: 17, success, outsize: 4112
unique: 19, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 9690
unique: 19, success, outsize: 96
unique: 17, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903
unique: 17, success, outsize: 4112
unique: 20, opcode: RELEASE (18), nodeid: 6, insize: 64, pid: 9690
unique: 20, success, outsize: 16
unique: 17, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903
unique: 17, success, outsize: 4112
unique: 17, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903
unique: 17, success, outsize: 2618
unique: 21, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903
unique: 21, success, outsize: 4112
unique: 21, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903
unique: 21, success, outsize: 4112
unique: 21, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903
unique: 21, success, outsize: 4112
unique: 21, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903
unique: 21, success, outsize: 4112
unique: 3, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903
unique: 3, success, outsize: 4112
unique: 3, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903
unique: 3, success, outsize: 4112
unique: 3, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903
unique: 3, success, outsize: 4112
unique: 3, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903
unique: 3, success, outsize: 4112
unique: 4, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903
unique: 4, success, outsize: 4112
unique: 4, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903
unique: 4, success, outsize: 4112
unique: 5, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903
unique: 5, success, outsize: 4112
unique: 5, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903
unique: 5, success, outsize: 4112
unique: 5, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903
unique: 5, success, outsize: 4112
unique: 5, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903
unique: 5, success, outsize: 4112
unique: 5, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903
unique: 5, success, outsize: 4112
unique: 5, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903
unique: 5, success, outsize: 4112
unique: 6, opcode: GETATTR (3), nodeid: 6, insize: 56, pid: 64903
unique: 6, success, outsize: 136
unique: 7, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 408
unique: 7, error: -2 (No such file or directory), outsize: 16
unique: 8, opcode: OPEN (14), nodeid: 6, insize: 48, pid: 64903
unique: 8, success, outsize: 32
unique: 10, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 408
unique: 10, error: -2 (No such file or directory), outsize: 16
unique: 11, opcode: SETATTR (4), nodeid: 6, insize: 168, pid: 64903
unique: 11, success, outsize: 136
unique: 12, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 408
unique: 12, error: -2 (No such file or directory), outsize: 16
unique: 13, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903
INFO:rmcl.items:Rendering <Document "todo">: 10.0%
INFO:rmcl.items:Rendering <Document "todo">: 20.0%
INFO:rmcl.items:Rendering <Document "todo">: 30.0%
INFO:rmcl.items:Rendering <Document "todo">: 40.0%
INFO:rmcl.items:Rendering <Document "todo">: 50.0%
INFO:rmcl.items:Rendering <Document "todo">: 60.0%
INFO:rmcl.items:Rendering <Document "todo">: 70.0%
INFO:rmcl.items:Rendering <Document "todo">: 80.0%
INFO:rmcl.items:Rendering <Document "todo">: 90.0%
INFO:rmcl.items:Rendering <Document "todo">: 100.0%
INFO:rmrl.render:exported pdf
unique: 13, success, outsize: 4112
unique: 14, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 408
unique: 14, error: -2 (No such file or directory), outsize: 16
unique: 13, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903
unique: 13, success, outsize: 4112
unique: 13, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903
unique: 13, success, outsize: 4112
unique: 13, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903
unique: 13, success, outsize: 4112
unique: 13, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903
unique: 13, success, outsize: 4112
unique: 13, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903
unique: 13, success, outsize: 4112
unique: 13, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903
unique: 13, success, outsize: 4112
unique: 13, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903
unique: 13, success, outsize: 4112
unique: 2, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903
unique: 2, success, outsize: 4112
unique: 2, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903
unique: 2, success, outsize: 4112
unique: 2, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903
unique: 2, success, outsize: 4112
unique: 2, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903
unique: 2, success, outsize: 4112
unique: 2, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903
unique: 2, success, outsize: 1225
unique: 15, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903
unique: 15, success, outsize: 4112
unique: 9, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903
unique: 9, success, outsize: 4112
unique: 9, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903
unique: 9, success, outsize: 4112
unique: 9, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903
unique: 9, success, outsize: 4112
unique: 9, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903
unique: 9, success, outsize: 4112
unique: 16, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903
unique: 16, success, outsize: 4112
unique: 16, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903
unique: 16, success, outsize: 4112
unique: 16, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903
unique: 16, success, outsize: 4112
unique: 16, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903
unique: 16, success, outsize: 4112
unique: 16, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903
unique: 16, success, outsize: 4112
unique: 16, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903
unique: 16, success, outsize: 4112
unique: 18, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903
unique: 18, success, outsize: 4112
unique: 18, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903
unique: 18, success, outsize: 4112
unique: 18, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903
unique: 18, success, outsize: 4112
unique: 19, opcode: RELEASE (18), nodeid: 6, insize: 64, pid: 64903
unique: 19, success, outsize: 16
unique: 20, opcode: GETATTR (3), nodeid: 6, insize: 56, pid: 408
unique: 20, success, outsize: 136
unique: 17, opcode: RELEASE (18), nodeid: 7, insize: 64, pid: 64903
unique: 17, success, outsize: 16
unique: 21, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 21, error: -2 (No such file or directory), outsize: 16
unique: 3, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 3, error: -2 (No such file or directory), outsize: 16
unique: 4, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 4, error: -2 (No such file or directory), outsize: 16
unique: 5, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 5, error: -2 (No such file or directory), outsize: 16
unique: 6, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408
unique: 6, error: -2 (No such file or directory), outsize: 16
unique: 7, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408
unique: 7, error: -2 (No such file or directory), outsize: 16
unique: 8, opcode: OPENDIR (27), nodeid: 1, insize: 48, pid: 408
unique: 8, success, outsize: 32
unique: 10, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408
unique: 10, success, outsize: 576
unique: 11, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408
unique: 11, error: -2 (No such file or directory), outsize: 16
unique: 12, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408
unique: 12, error: -2 (No such file or directory), outsize: 16
unique: 14, opcode: LOOKUP (1), nodeid: 1, insize: 49, pid: 408
unique: 14, error: -2 (No such file or directory), outsize: 16
unique: 13, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408
unique: 13, error: -2 (No such file or directory), outsize: 16
unique: 2, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408
unique: 2, error: -2 (No such file or directory), outsize: 16
unique: 15, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 15, error: -2 (No such file or directory), outsize: 16
unique: 9, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 9, error: -2 (No such file or directory), outsize: 16
unique: 16, opcode: GETATTR (3), nodeid: 7, insize: 56, pid: 408
unique: 16, success, outsize: 136
unique: 18, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 408
unique: 18, error: -2 (No such file or directory), outsize: 16
unique: 19, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 408
unique: 19, error: -2 (No such file or directory), outsize: 16
unique: 20, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408
unique: 20, error: -2 (No such file or directory), outsize: 16
unique: 17, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408
unique: 17, error: -2 (No such file or directory), outsize: 16
unique: 21, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408
unique: 21, error: -2 (No such file or directory), outsize: 16
unique: 3, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 3, error: -2 (No such file or directory), outsize: 16
unique: 4, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 4, error: -2 (No such file or directory), outsize: 16
unique: 5, opcode: LOOKUP (1), nodeid: 1, insize: 52, pid: 408
unique: 5, error: -2 (No such file or directory), outsize: 16
unique: 6, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408
unique: 6, error: -2 (No such file or directory), outsize: 16
unique: 7, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408
unique: 7, error: -2 (No such file or directory), outsize: 16
unique: 8, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 8, error: -2 (No such file or directory), outsize: 16
unique: 10, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408
unique: 10, success, outsize: 16
unique: 11, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408
unique: 11, success, outsize: 16
unique: 12, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408
unique: 12, success, outsize: 16
unique: 14, opcode: RELEASEDIR (29), nodeid: 1, insize: 64, pid: 408
unique: 14, error: -78 (Function not implemented), outsize: 16
unique: 13, opcode: GETATTR (3), nodeid: 1, insize: 56, pid: 408
unique: 13, success, outsize: 136
unique: 2, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408
unique: 2, success, outsize: 96
unique: 15, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408
unique: 15, error: -2 (No such file or directory), outsize: 16
unique: 9, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408
unique: 9, error: -2 (No such file or directory), outsize: 16
unique: 16, opcode: OPENDIR (27), nodeid: 1, insize: 48, pid: 408
unique: 16, success, outsize: 32
unique: 18, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408
unique: 18, success, outsize: 576
unique: 19, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408
unique: 19, error: -2 (No such file or directory), outsize: 16
unique: 20, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408
unique: 20, error: -2 (No such file or directory), outsize: 16
unique: 17, opcode: LOOKUP (1), nodeid: 1, insize: 49, pid: 408
unique: 17, error: -2 (No such file or directory), outsize: 16
unique: 21, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408
unique: 21, error: -2 (No such file or directory), outsize: 16
unique: 3, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408
unique: 3, error: -2 (No such file or directory), outsize: 16
unique: 4, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 4, error: -2 (No such file or directory), outsize: 16
unique: 5, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 5, error: -2 (No such file or directory), outsize: 16
unique: 6, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 408
unique: 6, error: -2 (No such file or directory), outsize: 16
unique: 7, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 408
unique: 7, error: -2 (No such file or directory), outsize: 16
unique: 8, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408
unique: 8, error: -2 (No such file or directory), outsize: 16
unique: 10, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408
unique: 10, error: -2 (No such file or directory), outsize: 16
unique: 11, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408
unique: 11, error: -2 (No such file or directory), outsize: 16
unique: 12, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 12, error: -2 (No such file or directory), outsize: 16
unique: 14, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 14, error: -2 (No such file or directory), outsize: 16
unique: 13, opcode: LOOKUP (1), nodeid: 1, insize: 52, pid: 408
unique: 13, error: -2 (No such file or directory), outsize: 16
unique: 2, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408
unique: 2, error: -2 (No such file or directory), outsize: 16
unique: 15, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408
unique: 15, error: -2 (No such file or directory), outsize: 16
unique: 9, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 9, error: -2 (No such file or directory), outsize: 16
unique: 16, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408
unique: 16, success, outsize: 16
unique: 18, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408
unique: 18, success, outsize: 16
unique: 19, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408
unique: 19, success, outsize: 16
unique: 20, opcode: RELEASEDIR (29), nodeid: 1, insize: 64, pid: 408
unique: 20, error: -78 (Function not implemented), outsize: 16
unique: 17, opcode: GETATTR (3), nodeid: 1, insize: 56, pid: 408
unique: 17, success, outsize: 136
unique: 21, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408
unique: 21, success, outsize: 96
unique: 3, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408
unique: 3, error: -2 (No such file or directory), outsize: 16
unique: 4, opcode: LOOKUP (1), nodeid: 1, insize: 95, pid: 408
unique: 4, error: -2 (No such file or directory), outsize: 16
unique: 5, opcode: LOOKUP (1), nodeid: 1, insize: 95, pid: 408
unique: 5, error: -2 (No such file or directory), outsize: 16
unique: 6, opcode: LOOKUP (1), nodeid: 1, insize: 95, pid: 425
unique: 6, error: -2 (No such file or directory), outsize: 16
unique: 7, opcode: LOOKUP (1), nodeid: 1, insize: 95, pid: 425
unique: 7, error: -2 (No such file or directory), outsize: 16
unique: 8, opcode: LOOKUP (1), nodeid: 1, insize: 95, pid: 425
unique: 8, error: -2 (No such file or directory), outsize: 16
unique: 10, opcode: LOOKUP (1), nodeid: 1, insize: 95, pid: 425
unique: 10, error: -2 (No such file or directory), outsize: 16
unique: 11, opcode: LOOKUP (1), nodeid: 1, insize: 95, pid: 425
unique: 11, error: -2 (No such file or directory), outsize: 16
unique: 12, opcode: LOOKUP (1), nodeid: 1, insize: 95, pid: 408
unique: 12, error: -2 (No such file or directory), outsize: 16
unique: 14, opcode: LOOKUP (1), nodeid: 1, insize: 95, pid: 408
unique: 14, error: -2 (No such file or directory), outsize: 16
unique: 13, opcode: LOOKUP (1), nodeid: 1, insize: 95, pid: 408
unique: 13, error: -2 (No such file or directory), outsize: 16
unique: 2, opcode: LOOKUP (1), nodeid: 1, insize: 95, pid: 408
unique: 2, error: -2 (No such file or directory), outsize: 16
unique: 15, opcode: LOOKUP (1), nodeid: 1, insize: 95, pid: 408
unique: 15, error: -2 (No such file or directory), outsize: 16
unique: 9, opcode: LOOKUP (1), nodeid: 1, insize: 95, pid: 408
unique: 9, error: -2 (No such file or directory), outsize: 16
unique: 16, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 65086
unique: 16, error: -2 (No such file or directory), outsize: 16
unique: 18, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 65086
unique: 18, success, outsize: 96
unique: 19, opcode: LOOKUP (1), nodeid: 1, insize: 95, pid: 65086
unique: 19, error: -2 (No such file or directory), outsize: 16
unique: 20, opcode: CREATE (35), nodeid: 1, insize: 111, pid: 65086
unique: 20, success, outsize: 176
unique: 17, opcode: LOOKUP (1), nodeid: 1, insize: 97, pid: 65086
unique: 17, error: -2 (No such file or directory), outsize: 16
unique: 21, opcode: RELEASE (18), nodeid: 18, insize: 64, pid: 65086
ERROR:rmfuse.fuse:Error: Not a PDF or EPUB file
unique: 21, error: -5 (Input/output error), outsize: 16
unique: 3, opcode: GETATTR (3), nodeid: 18, insize: 56, pid: 65086
unique: 3, success, outsize: 136
unique: 4, opcode: LOOKUP (1), nodeid: 1, insize: 97, pid: 65086
unique: 4, error: -2 (No such file or directory), outsize: 16
unique: 5, opcode: LOOKUP (1), nodeid: 1, insize: 97, pid: 65086
unique: 5, error: -2 (No such file or directory), outsize: 16
unique: 7, opcode: LOOKUP (1), nodeid: 1, insize: 97, pid: 65086
unique: 7, error: -2 (No such file or directory), outsize: 16
unique: 8, opcode: GETATTR (3), nodeid: 1, insize: 56, pid: 65086
unique: 8, success, outsize: 136
unique: 10, opcode: CREATE (35), nodeid: 1, insize: 113, pid: 65086
unique: 10, success, outsize: 176
unique: 11, opcode: WRITE (16), nodeid: 19, insize: 4176, pid: 65086
unique: 11, success, outsize: 24
unique: 12, opcode: GETATTR (3), nodeid: 19, insize: 56, pid: 65086
unique: 12, success, outsize: 136
unique: 14, opcode: RELEASE (18), nodeid: 19, insize: 64, pid: 65086
ERROR:rmfuse.fuse:Error: Not a PDF or EPUB file
unique: 14, error: -5 (Input/output error), outsize: 16
unique: 13, opcode: GETATTR (3), nodeid: 1, insize: 56, pid: 65
unique: 13, success, outsize: 136
unique: 2, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408
unique: 2, error: -2 (No such file or directory), outsize: 16
unique: 15, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408
unique: 15, error: -2 (No such file or directory), outsize: 16
unique: 9, opcode: OPENDIR (27), nodeid: 1, insize: 48, pid: 408
unique: 9, success, outsize: 32
unique: 16, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408
unique: 16, success, outsize: 576
unique: 18, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408
unique: 18, error: -2 (No such file or directory), outsize: 16
unique: 19, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408
unique: 19, error: -2 (No such file or directory), outsize: 16
unique: 20, opcode: LOOKUP (1), nodeid: 1, insize: 49, pid: 408
unique: 20, error: -2 (No such file or directory), outsize: 16
unique: 17, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408
unique: 17, error: -2 (No such file or directory), outsize: 16
unique: 21, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408
unique: 21, error: -2 (No such file or directory), outsize: 16
unique: 3, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 3, error: -2 (No such file or directory), outsize: 16
unique: 4, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 4, error: -2 (No such file or directory), outsize: 16
unique: 5, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 408
unique: 5, error: -2 (No such file or directory), outsize: 16
unique: 7, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 408
unique: 7, error: -2 (No such file or directory), outsize: 16
unique: 8, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408
unique: 8, error: -2 (No such file or directory), outsize: 16
unique: 10, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408
unique: 10, error: -2 (No such file or directory), outsize: 16
unique: 11, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408
unique: 11, error: -2 (No such file or directory), outsize: 16
unique: 12, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 12, error: -2 (No such file or directory), outsize: 16
unique: 14, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 14, error: -2 (No such file or directory), outsize: 16
unique: 13, opcode: LOOKUP (1), nodeid: 1, insize: 52, pid: 408
unique: 13, error: -2 (No such file or directory), outsize: 16
unique: 2, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408
unique: 2, error: -2 (No such file or directory), outsize: 16
unique: 15, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408
unique: 15, error: -2 (No such file or directory), outsize: 16
unique: 9, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 9, error: -2 (No such file or directory), outsize: 16
unique: 16, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408
unique: 16, success, outsize: 16
unique: 18, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408
unique: 18, success, outsize: 16
unique: 19, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408
unique: 19, success, outsize: 16
unique: 20, opcode: RELEASEDIR (29), nodeid: 1, insize: 64, pid: 408
unique: 20, error: -78 (Function not implemented), outsize: 16
unique: 17, opcode: GETATTR (3), nodeid: 1, insize: 56, pid: 408
unique: 17, success, outsize: 136
unique: 21, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408
unique: 21, success, outsize: 96
unique: 3, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408
unique: 3, error: -2 (No such file or directory), outsize: 16
unique: 4, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408
unique: 4, error: -2 (No such file or directory), outsize: 16
unique: 5, opcode: OPENDIR (27), nodeid: 1, insize: 48, pid: 408
unique: 5, success, outsize: 32
unique: 7, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408
unique: 7, success, outsize: 576
unique: 8, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408
unique: 8, error: -2 (No such file or directory), outsize: 16
unique: 10, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408
unique: 10, error: -2 (No such file or directory), outsize: 16
unique: 11, opcode: LOOKUP (1), nodeid: 1, insize: 49, pid: 408
unique: 11, error: -2 (No such file or directory), outsize: 16
unique: 12, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408
unique: 12, error: -2 (No such file or directory), outsize: 16
unique: 14, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408
unique: 14, error: -2 (No such file or directory), outsize: 16
unique: 13, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 13, error: -2 (No such file or directory), outsize: 16
unique: 2, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 2, error: -2 (No such file or directory), outsize: 16
unique: 15, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 408
unique: 15, error: -2 (No such file or directory), outsize: 16
unique: 9, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 408
unique: 9, error: -2 (No such file or directory), outsize: 16
unique: 16, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408
unique: 16, error: -2 (No such file or directory), outsize: 16
unique: 18, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408
unique: 18, error: -2 (No such file or directory), outsize: 16
unique: 19, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408
unique: 19, error: -2 (No such file or directory), outsize: 16
unique: 20, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 20, error: -2 (No such file or directory), outsize: 16
unique: 17, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 17, error: -2 (No such file or directory), outsize: 16
unique: 21, opcode: LOOKUP (1), nodeid: 1, insize: 52, pid: 408
unique: 21, error: -2 (No such file or directory), outsize: 16
unique: 3, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408
unique: 3, error: -2 (No such file or directory), outsize: 16
unique: 4, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408
unique: 4, error: -2 (No such file or directory), outsize: 16
unique: 5, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 5, error: -2 (No such file or directory), outsize: 16
unique: 7, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408
unique: 7, success, outsize: 16
unique: 8, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408
unique: 8, success, outsize: 16
unique: 10, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408
unique: 10, success, outsize: 16
unique: 11, opcode: RELEASEDIR (29), nodeid: 1, insize: 64, pid: 408
unique: 11, error: -78 (Function not implemented), outsize: 16
unique: 12, opcode: GETATTR (3), nodeid: 1, insize: 56, pid: 408
unique: 12, success, outsize: 136
unique: 14, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408
unique: 14, success, outsize: 96
unique: 13, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408
unique: 13, error: -2 (No such file or directory), outsize: 16
unique: 2, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408
unique: 2, error: -2 (No such file or directory), outsize: 16
unique: 15, opcode: OPENDIR (27), nodeid: 1, insize: 48, pid: 408
unique: 15, success, outsize: 32
unique: 9, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408
unique: 9, success, outsize: 576
unique: 16, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408
unique: 16, error: -2 (No such file or directory), outsize: 16
unique: 18, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408
unique: 18, error: -2 (No such file or directory), outsize: 16
unique: 19, opcode: LOOKUP (1), nodeid: 1, insize: 49, pid: 408
unique: 19, error: -2 (No such file or directory), outsize: 16
unique: 20, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408
unique: 20, error: -2 (No such file or directory), outsize: 16
unique: 17, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408
unique: 17, error: -2 (No such file or directory), outsize: 16
unique: 21, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 21, error: -2 (No such file or directory), outsize: 16
unique: 3, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 3, error: -2 (No such file or directory), outsize: 16
unique: 4, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 408
unique: 4, error: -2 (No such file or directory), outsize: 16
unique: 5, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 408
unique: 5, error: -2 (No such file or directory), outsize: 16
unique: 7, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408
unique: 7, error: -2 (No such file or directory), outsize: 16
unique: 8, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408
unique: 8, error: -2 (No such file or directory), outsize: 16
unique: 10, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408
unique: 10, error: -2 (No such file or directory), outsize: 16
unique: 11, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 11, error: -2 (No such file or directory), outsize: 16
unique: 12, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 12, error: -2 (No such file or directory), outsize: 16
unique: 14, opcode: LOOKUP (1), nodeid: 1, insize: 52, pid: 408
unique: 14, error: -2 (No such file or directory), outsize: 16
unique: 13, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408
unique: 13, error: -2 (No such file or directory), outsize: 16
unique: 2, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408
unique: 2, error: -2 (No such file or directory), outsize: 16
unique: 15, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408
unique: 15, error: -2 (No such file or directory), outsize: 16
unique: 9, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408
unique: 9, success, outsize: 16
unique: 16, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408
unique: 16, success, outsize: 16
unique: 18, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408
unique: 18, success, outsize: 16
unique: 19, opcode: RELEASEDIR (29), nodeid: 1, insize: 64, pid: 408
unique: 19, error: -78 (Function not implemented), outsize: 16
unique: 20, opcode: GETATTR (3), nodeid: 1, insize: 56, pid: 408
unique: 20, success, outsize: 136
unique: 17, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408
unique: 17, success, outsize: 96
The error seems to be that RMfuse cannot determine that the file you're trying to upload is a PDF or EPUB file. Which type is it? (Note that if it's not one of these, the upload should fail.)
If you're able to send me the file, I can take a look. If not, just the beginning of the file would be enough for me to see if RMfuse should be able to recognize it. You can run head -c 100 file.pdf > file.fragment, for example, to get the first 100 bytes of file.pdf for me.
Hi, I tested it with the same file that I downloaded from the remarkable. It was definitely a pdf file.
Here is again the python error trace when I try to cp:
(rmfuse) schneider@Aarons-MacBook-Pro ~ % cp ~/Desktop/debug.pdf remarkable
cp: remarkable/debug.pdf: fchmod failed: Input/output error
Traceback (most recent call last):
File "/Users/schneider/anaconda3/envs/rmfuse/bin/rmfuse", line 8, in <module>
sys.exit(main())
File "/Users/schneider/anaconda3/envs/rmfuse/lib/python3.9/site-packages/rmfuse/fuse.py", line 470, in main
fuse.main(workers=1)
File "src/fuse_api.pxi", line 327, in llfuse.main
File "src/handlers.pxi", line 136, in llfuse.fuse_setattr
File "src/handlers.pxi", line 137, in llfuse.fuse_setattr
File "/Users/schneider/anaconda3/envs/rmfuse/lib/python3.9/site-packages/rmfuse/fuselib.py", line 25, in decorated
return trio.run(runner)
File "/Users/schneider/anaconda3/envs/rmfuse/lib/python3.9/site-packages/trio/_core/_run.py", line 1932, in run
raise runner.main_task_outcome.error
File "/Users/schneider/anaconda3/envs/rmfuse/lib/python3.9/site-packages/rmfuse/fuselib.py", line 24, in runner
return await afunc(*args, **kw)
File "/Users/schneider/anaconda3/envs/rmfuse/lib/python3.9/site-packages/rmfuse/fuse.py", line 232, in setattr
log.debug(f'setattr called on {await self.get_by_id(self.get_id(inode))!r}')
File "/Users/schneider/anaconda3/envs/rmfuse/lib/python3.9/site-packages/rmfuse/fuse.py", line 139, in get_by_id
return await Item.get_by_id(id_)
File "/Users/schneider/anaconda3/envs/rmfuse/lib/python3.9/site-packages/rmcl/items.py", line 46, in get_by_id
return await (await api.get_client()).get_by_id(id_)
File "/Users/schneider/anaconda3/envs/rmfuse/lib/python3.9/site-packages/rmcl/api.py", line 247, in get_by_id
return self.by_id[id_]
KeyError: 'c369d328-db7f-4673-8588-ac0305b8fbad'
cp: remarkable/debug.pdf: Device not configured
[1] + exit 1 rmfuse remarkable
cp: /Users/schneider/Desktop/debug.pdf: could not copy extended attributes to remarkable/debug.pdf: Device not configured
Here is another error which I get upon renaming a file:
(rmfuse) schneider@Aarons-MacBook-Pro remarkable % mv How\ to\ Debug\ Fortran\ code\ with\ GDB\ -\ Fortran\ Debugging\ -\ Undo.io.pdf debug.pdf
mv: rename How to Debug Fortran code with GDB - Fortran Debugging - Undo.io.pdf to debug.pdf: Input/output error
(rmfuse) schneider@Aarons-MacBook-Pro remarkable % Traceback (most recent call last):
File "/Users/schneider/anaconda3/envs/rmfuse/bin/rmfuse", line 8, in <module>
sys.exit(main())
File "/Users/schneider/anaconda3/envs/rmfuse/lib/python3.9/site-packages/rmfuse/fuse.py", line 470, in main
fuse.main(workers=1)
File "src/fuse_api.pxi", line 327, in llfuse.main
File "src/handlers.pxi", line 263, in llfuse.fuse_rename
File "src/handlers.pxi", line 264, in llfuse.fuse_rename
File "/Users/schneider/anaconda3/envs/rmfuse/lib/python3.9/site-packages/rmfuse/fuselib.py", line 25, in decorated
return trio.run(runner)
File "/Users/schneider/anaconda3/envs/rmfuse/lib/python3.9/site-packages/trio/_core/_run.py", line 1932, in run
raise runner.main_task_outcome.error
File "/Users/schneider/anaconda3/envs/rmfuse/lib/python3.9/site-packages/rmfuse/fuselib.py", line 24, in runner
return await afunc(*args, **kw)
TypeError: rename() missing 1 required positional argument: 'ctx'
Here is the debug.pdf that I copied from the rmcloud to my desktop. And which I tried to upload back to the rm. On the rm this file is called How\ to\ Debug\ Fortran\ code\ with\ GDB\ -\ Fortran\ Debugging\ -\ Undo.io.pdf
debug.pdf
Thanks!
Thanks. I think I missed the tracebacks into the logs above. The good new is, both of these are simple bugs. I've pushed fixes; give it a try. (I swear we're getting closer, even if it doesn't seem that way!)
rmfuse-0.1.2a0.tar.gz
Thanks again for debugging!
We are getting there!
Renaming and deleting files seems to work now. Moving files (either by shell command or by finder) does not yet work.
It often boils down to these errors: RELEASEDIR and LOOKUP.
unique: 17, opcode: RELEASEDIR (29), nodeid: 1, insize: 64, pid: 434
unique: 17, error: -78 (Function not implemented), outsize: 16
unique: 18, opcode: GETATTR (3), nodeid: 1, insize: 56, pid: 434
unique: 18, success, outsize: 136
unique: 19, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 434
unique: 19, success, outsize: 96
unique: 2, opcode: LOOKUP (1), nodeid: 1, insize: 60, pid: 65
unique: 2, error: -2 (No such file or directory), outsize: 16
Both give always errors.
I'm not convinced that those errors are a problem. The LOOKUP is probably result of something on your system looking for a file that doesn't exist, perhaps some metadata directory. The RELEASEDIR isn't implemented, since RMfuse doesn't have anything to do to release a directory. But that shouldn't cause problems. (We can try adding a stub if we can't figure out anything else, but I suspect it wouldn't really help.)
What's the failure mode of moving the file? Does it crash? If so, please provide the traceback. Does it hang? How long did you wait? Does nothing happen? Does this happen when moving files within the RMfuse mount point, moving a file from your system to RMfuse, or moving a file from RMfuse to your system?
sorry for the poor Bugreport. Let me try to improve on your questions:
It oddly depends on the file (and its size). I tested four files:
ms2.pdf (a paper generated with pdflatex), size=961KB
rmfuse_test.pdf (a random pdf printout of this issue CMD + P), size=200KB
Molliere_thesis.pdf (some pdf of a phd thesis, probably also created with pdflatex), size=30.3MB
SA_Chap2_1.pdf (a lecture presentation), size=16.2MB
What's the failure mode of moving the file?
In the file explorer (hereafter finder) I always get this error (regardless of the file):
The operation could not be completed because at least one required object was not found (error -43).
Copying using the terminal succeeds for the smaller documents (ms2.pdf and rmfuse_test.pdf).
In terminal that is moving the file:
cp: Desktop/paper_2/ms2.pdf: could not copy extended attributes to /Users/schneider/remarkable/ms2.pdf: Attribute not found
In terminal that is in control of rmfuse:
Error: Not a PDF or EPUB file
However the two larger documents do not succeed:
(base) schneider@Aarons-MacBook-Pro Desktop % cp Molliere_thesis.pdf ~/remarkable
cp: Molliere_thesis.pdf: could not copy extended attributes to /Users/schneider/remarkable/Molliere_thesis.pdf: Operation not supported
in rmfuse window:
ERROR:rmfuse.fuse:Error: Not a PDF or EPUB file
ERROR:root:Attempt to get non-existant Item 699b20b3-6826-4625-b22a-600354ee4e38
Does it crash? If so, please provide the traceback. Does it hang?
rmfuse does not crash nor hang.
How long did you wait?
a couple of minutes. I also checked with rmapi if the file was copied but just not listed.
Does this happen when moving files within the RMfuse mount point, moving a file from your system to RMfuse, or moving a file from RMfuse to your system?
Moving a file inside rmfuse and moving a file from rmfuse to system gives no error. Only moving files from local to rmfuse.
Thanks, that's very informative. I just don't quite know what it all means. :)
The errors about the extended attributes are expected -- RMfuse doesn't implement them right now. (It might be possible to stuff them into the document metadata at some point.) But it is weird that the error on setting them is different in the two cases. If they get annoying, we could add an option to RMfuse that would pretend to set the attributes, which should get rid of that message.
I'm not sure what's going wrong with the "Not a PDF or EPUB file" error. My best guess is that there's some metadata file that the Mac is trying to create / copy along with the PDF file. The fact that it's happening in both successes and failures suggests that it's not a fatal problem.
The worrisome bit is the "Attempt to get non-existent Item". That means something is going wrong in the upload process. I wonder if we have a race condition -- something is trying to do something to a file while it's in the process of being uploaded. This would slip in a gap of awareness of RMfuse -- it can't find the document in the set of editable new-created documents, since we closed it out to upload. But it also doesn't see it in the list of documents from the Cloud, because the upload hasn't finished. (This would explain why it's only happening to larger files.)
I've built a new version with some more debugging output and an attempt to handle the above case. Please give it a try (with -v) and let me know what happens. One trick I've used for keeping track of what's happening: If you press 'Enter' in the rmfuse terminal between events you're triggering in another terminal, you can better associate the debugging lines with specific actions.
rmfuse-0.1.2a0.tar.gz
Thanks for the explainations!
Copying Molliere_thesis or ms2 to rmfuse using finder (still same error code -43 in finder):
DEBUG:rmfuse.fuse:Created Molliere_thesis for b'Molliere_thesis.pdf', with inode 16 and ID 9dc8cf3f-17ec-40e0-80e1-f4b61ea41289
DEBUG:rmfuse.fuse:Releasing inode 16
ERROR:rmfuse.fuse:Error: Not a PDF or EPUB file (file was Molliere_thesis)
ERROR:root:Attempt to get non-existent Item 9dc8cf3f-17ec-40e0-80e1-f4b61ea41289
The same from command line (only the last couple lines...):
(...)
DEBUG:rmfuse.fuse:Writing to 18 at 30257152, length 4096
DEBUG:rmfuse.fuse:Writing to 18 at 30261248, length 4096
DEBUG:rmfuse.fuse:Writing to 18 at 30265344, length 4096
DEBUG:rmfuse.fuse:Writing to 18 at 30269440, length 4096
DEBUG:rmfuse.fuse:Writing to 18 at 30273536, length 4096
DEBUG:rmfuse.fuse:Writing to 18 at 30277632, length 4096
DEBUG:rmfuse.fuse:Writing to 18 at 30281728, length 4096
DEBUG:rmfuse.fuse:Writing to 18 at 30285824, length 4096
DEBUG:rmfuse.fuse:Writing to 18 at 30289920, length 4096
DEBUG:rmfuse.fuse:Writing to 18 at 30294016, length 4096
DEBUG:rmfuse.fuse:Writing to 18 at 30298112, length 755
DEBUG:rmfuse.fuse:setattr called on <Document "Molliere_thesis">
DEBUG:rmfuse.fuse: Attempting to set mode to 33188
DEBUG:rmfuse.fuse: No changes made
DEBUG:rmfuse.fuse:setattr called on <Document "Molliere_thesis">
DEBUG:rmfuse.fuse: Attempting to set mode to 33188
DEBUG:rmfuse.fuse: No changes made
DEBUG:rmfuse.fuse:Created ._Molliere_thesis for b'._Molliere_thesis.pdf', with inode 19 and ID 870c59c6-47aa-42a1-8599-291f5e2815e9
DEBUG:rmfuse.fuse:Writing to 19 at 0, length 4096
DEBUG:rmfuse.fuse:Releasing inode 19
ERROR:rmfuse.fuse:Error: Not a PDF or EPUB file (file was ._Molliere_thesis)
DEBUG:rmfuse.fuse:Releasing inode 18
ERROR:rmfuse.fuse:Error: Not a PDF or EPUB file (file was Molliere_thesis)
ERROR:root:Attempt to get non-existent Item 4a4313ac-4c86-493d-9f5b-2378666aa553
In contrast here we have ms2.pdf which works:
DEBUG:rmfuse.fuse:Created ms2 for b'ms2.pdf', with inode 20 and ID 6e376b45-4c4b-4e4b-a5b1-34e9679da343
DEBUG:rmfuse.fuse:setattr called on <Document "ms2">
DEBUG:rmfuse.fuse: Attempting to set mode to 33152
DEBUG:rmfuse.fuse: No changes made
DEBUG:rmfuse.fuse:Writing to 20 at 0, length 4096
DEBUG:rmfuse.fuse:Writing to 20 at 4096, length 4096
DEBUG:rmfuse.fuse:Writing to 20 at 8192, length 4096
(...)
DEBUG:rmfuse.fuse:Writing to 20 at 950272, length 4096
DEBUG:rmfuse.fuse:Writing to 20 at 954368, length 4096
DEBUG:rmfuse.fuse:Writing to 20 at 958464, length 2067
DEBUG:rmfuse.fuse:setattr called on <Document "ms2">
DEBUG:rmfuse.fuse: Attempting to set mode to 33188
DEBUG:rmfuse.fuse: No changes made
DEBUG:rmfuse.fuse:setattr called on <Document "ms2">
DEBUG:rmfuse.fuse: Attempting to set mode to 33188
DEBUG:rmfuse.fuse: No changes made
DEBUG:rmfuse.fuse:Created ._ms2 for b'._ms2.pdf', with inode 21 and ID 539b5949-8baa-4336-9c60-5febbb25dfe3
DEBUG:rmfuse.fuse:Writing to 21 at 0, length 4096
DEBUG:rmfuse.fuse:Releasing inode 21
ERROR:rmfuse.fuse:Error: Not a PDF or EPUB file (file was ._ms2)
DEBUG:rmfuse.fuse:Releasing inode 20
DEBUG:rmcl.items:Getting details for <Document "ms2">
DEBUG:rmcl.items:Details for <Document "ms2">: type pdf, size 960531
ERROR:root:Attempt to get non-existent Item 539b5949-8baa-4336-9c60-5febbb25dfe3
From what I can see:
The difference in cp from terminal between Molliere and ms2 boils down to these two lines missing for Molliere_thesis:
DEBUG:rmcl.items:Getting details for <Document "ms2">
DEBUG:rmcl.items:Details for <Document "ms2">: type pdf, size 960531
The difference between a finder copy and a terminal copy seems to be everything in between the first line and the last few lines.
Thanks. I feel like we're inching towards understanding. The thing that has
me worried is this:
ERROR:rmfuse.fuse:Error: Not a PDF or EPUB file (file was Molliere_thesis)
This means that RMfuse couldn't figure out that the file is a PDF document.
Therefore, it couldn't upload the file, and that's why nothing's working.
It looks like there may be two different failures, unfortunately. With
finder, it appears that no data was written to the file. I'm not sure why
that is. (Maybe finder planned on re-opening the file to write the contents
later?) But it makes sense that it failed. But from the command line, we
can see a lot of data did get written, so I'm not sure what went wrong
there.
RMfuse assumes that a PDF file starts with '%PDF'. Can you check whether
this is the case for your problematic files? (The 'head' command is useful
for reading the first n bytes of a file.) If it's not, can you include the
first couple of KB of those files so I can take a look at what's going on
there?
On Sun, Feb 21, 2021 at 11:23 PM Aaron David Schneider <
notifications@github.com> wrote:
Thanks for the explainations!
Copying Molliere_thesis or ms2 to rmfuse using finder (still same error
code -43 in finder):
DEBUG:rmfuse.fuse:Created Molliere_thesis for b'Molliere_thesis.pdf', with inode 16 and ID 9dc8cf3f-17ec-40e0-80e1-f4b61ea41289
DEBUG:rmfuse.fuse:Releasing inode 16
ERROR:rmfuse.fuse:Error: Not a PDF or EPUB file (file was Molliere_thesis)
ERROR:root:Attempt to get non-existent Item 9dc8cf3f-17ec-40e0-80e1-f4b61ea41289
The same from command line (only the last couple lines...):
(...)
DEBUG:rmfuse.fuse:Writing to 18 at 30257152, length 4096
DEBUG:rmfuse.fuse:Writing to 18 at 30261248, length 4096
DEBUG:rmfuse.fuse:Writing to 18 at 30265344, length 4096
DEBUG:rmfuse.fuse:Writing to 18 at 30269440, length 4096
DEBUG:rmfuse.fuse:Writing to 18 at 30273536, length 4096
DEBUG:rmfuse.fuse:Writing to 18 at 30277632, length 4096
DEBUG:rmfuse.fuse:Writing to 18 at 30281728, length 4096
DEBUG:rmfuse.fuse:Writing to 18 at 30285824, length 4096
DEBUG:rmfuse.fuse:Writing to 18 at 30289920, length 4096
DEBUG:rmfuse.fuse:Writing to 18 at 30294016, length 4096
DEBUG:rmfuse.fuse:Writing to 18 at 30298112, length 755
DEBUG:rmfuse.fuse:setattr called on <Document "Molliere_thesis">
DEBUG:rmfuse.fuse: Attempting to set mode to 33188
DEBUG:rmfuse.fuse: No changes made
DEBUG:rmfuse.fuse:setattr called on <Document "Molliere_thesis">
DEBUG:rmfuse.fuse: Attempting to set mode to 33188
DEBUG:rmfuse.fuse: No changes made
DEBUG:rmfuse.fuse:Created ._Molliere_thesis for b'._Molliere_thesis.pdf', with inode 19 and ID 870c59c6-47aa-42a1-8599-291f5e2815e9
DEBUG:rmfuse.fuse:Writing to 19 at 0, length 4096
DEBUG:rmfuse.fuse:Releasing inode 19
ERROR:rmfuse.fuse:Error: Not a PDF or EPUB file (file was ._Molliere_thesis)
DEBUG:rmfuse.fuse:Releasing inode 18
ERROR:rmfuse.fuse:Error: Not a PDF or EPUB file (file was Molliere_thesis)
ERROR:root:Attempt to get non-existent Item 4a4313ac-4c86-493d-9f5b-2378666aa553
In contrast here we have ms2.pdf which works:
DEBUG:rmfuse.fuse:Created ms2 for b'ms2.pdf', with inode 20 and ID 6e376b45-4c4b-4e4b-a5b1-34e9679da343
DEBUG:rmfuse.fuse:setattr called on <Document "ms2">
DEBUG:rmfuse.fuse: Attempting to set mode to 33152
DEBUG:rmfuse.fuse: No changes made
DEBUG:rmfuse.fuse:Writing to 20 at 0, length 4096
DEBUG:rmfuse.fuse:Writing to 20 at 4096, length 4096
DEBUG:rmfuse.fuse:Writing to 20 at 8192, length 4096
(...)
DEBUG:rmfuse.fuse:Writing to 20 at 950272, length 4096
DEBUG:rmfuse.fuse:Writing to 20 at 954368, length 4096
DEBUG:rmfuse.fuse:Writing to 20 at 958464, length 2067
DEBUG:rmfuse.fuse:setattr called on <Document "ms2">
DEBUG:rmfuse.fuse: Attempting to set mode to 33188
DEBUG:rmfuse.fuse: No changes made
DEBUG:rmfuse.fuse:setattr called on <Document "ms2">
DEBUG:rmfuse.fuse: Attempting to set mode to 33188
DEBUG:rmfuse.fuse: No changes made
DEBUG:rmfuse.fuse:Created ._ms2 for b'._ms2.pdf', with inode 21 and ID 539b5949-8baa-4336-9c60-5febbb25dfe3
DEBUG:rmfuse.fuse:Writing to 21 at 0, length 4096
DEBUG:rmfuse.fuse:Releasing inode 21
ERROR:rmfuse.fuse:Error: Not a PDF or EPUB file (file was ._ms2)
DEBUG:rmfuse.fuse:Releasing inode 20
DEBUG:rmcl.items:Getting details for <Document "ms2">
DEBUG:rmcl.items:Details for <Document "ms2">: type pdf, size 960531
ERROR:root:Attempt to get non-existent Item 539b5949-8baa-4336-9c60-5febbb25dfe3
From what I can see:
The difference in cp from terminal between Molliere and ms2 boils down to
these two lines missing for Molliere_thesis:
DEBUG:rmcl.items:Getting details for <Document "ms2">
DEBUG:rmcl.items:Details for <Document "ms2">: type pdf, size 960531
The difference between a finder copy and a terminal copy seems to be
everything in between the first line and the last few lines.
—
You are receiving this because you commented.
Reply to this email directly, view it on GitHub
https://github.com/rschroll/rmfuse/issues/2#issuecomment-783151912, or
unsubscribe
https://github.com/notifications/unsubscribe-auth/AACZ2HZQYXJG545Q2WXCKMLTAIA7DANCNFSM4XISMUSA
.
Thanks for your quick reply!
Here is the head output of Molliere thesis:
%PDF-1.3
%����
1 0 obj
<</CreationDate 8963 0 R /Creator 8962 0 R /ModDate 8963 0 R /Producer 8961 0 R >>
endobj
2 0 obj
<</Count 13/First 7488 0 R /Last 7489 0 R >>
endobj
3 0 obj
<</Contents 5 0 R /MediaBox[ 0 0 595.276 841.89]/Parent 4 0 R /Resources 7 0 R /Rotate 0/Type/Page>>
This would also be a large test pdf (20 MB) - found by googling large test pdf :
https://www.hq.nasa.gov/alsj/a17/A17_FlightPlan.pdf
We could use this file as a benchmark for this issue since this one fails too and I can share it since its an open document.
The bad (or maybe good?) news is that that NASA PDF file works just fine on my test system. This suggests to me that the Mac is doing something I didn't anticipate when writing the file.
Can you check through the 'Writing to X at Y, length Z' messages? Does Y start at zero and always increase? Or does it jump around. (On my system, it increases steadily, but if it doesn't I can imagine that it might break things. Though I don't see how it can break the beginning of the file.)
The only other thought I have is to look at the actual values RMfuse finds in the file. If you're willing to get your hands dirty, try adding a line to rmfuse/fuse.py. Right after the line
log.error(f'Error: Not a PDF or EPUB file (file was {document.name})')
Try adding
log.error(f'File header: {data[:1024]}')
This will print out the first KB of the file as RMfuse got it, and we can see if it's different from what it should be. If you're not comfortable mucking about like this, I can create an instrumented version with this and other debug output later.
I just tested your second idea with the Nasa file.
Upon
(base) schneider@Aarons-MacBook-Pro ~ % cp Downloads/A17_FlightPlan.pdf ~/remarkable
cp: Downloads/A17_FlightPlan.pdf: could not copy extended attributes to /Users/schneider/remarkable/A17_FlightPlan.pdf: Operation not supported
I get the following output in the rmfuse window:
ERROR:rmfuse.fuse:Error: Not a PDF or EPUB file (file was ._A17_FlightPlan)
ERROR:rmfuse.fuse:File header: b'\x00\x05\x16\x07\x00\x02\x00\x00Mac OS X \x00\x02\x00\x00\x00\t\x00\x00\x002\x00\x00\x0e\xb0\x00\x00\x00\x02\x00\x00\x0e\xe2\x00\x00\x01\x1e\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00ATTR\x00\x00\x00\x12\x00\x00\x0e\xe2\x00\x00\x00x\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00'
ERROR:rmfuse.fuse:Error: Not a PDF or EPUB file (file was A17_FlightPlan)
ERROR:rmfuse.fuse:File header: b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00'
ERROR:root:Attempt to get non-existent Item 0e77f5b9-e23f-423b-9fa9-a38a522c8b59
Yeah, it looks like the file it is trying to upload is full of null bytes. That ain't right. (As we saw above, the PDF file should start with '%PDF'.)
Unfortunately, I have no idea why that would be the case. Perhaps the OS decided it would fill the file with zeros first, and then write the actual content? But I don't know why that would be the case, or why that would only happen with large files.
Another possibility would be a bug with llfuse or Macfuse. But again, I don't know why that would only affect large files.
I'll have to puzzle over this for a while.
In happier news, those ._ files are apparently for filesystem metadata. It appears you can turn that off with a special fuse option. If you add fuse_options.add('noappledouble') around here, it might stop complaining about those files. But I don't think it has any effect on the real problem here.
No hurry 👌
In happier news, those ._ files are apparently for filesystem metadata. It appears you can turn that off with a special fuse option. If you add fuse_options.add('noappledouble') around here, it might stop complaining about those files. But I don't think it has any effect on the real problem here.
I can confirm.
No brilliant thoughts so far. Just to confirm: Other than uploading large files, things seem to be working pretty well? I'm thinking that maybe I should merge #6 into master and cut a new release, with the hope that additional users will help us figure out what's going wrong.
How does that sit with you?
I've added MacOS install directions, based on your descriptions here. Can you check that I got the correct? https://github.com/rschroll/rmfuse/blob/release-0.2.0/README.md#macos
Note that if you want to try this out yourself, you'll need to use the --pre option to install a prerelease version of rmfuse with pip.
Thanks!
yes that sounds good!
The new version doesn't start on osx.
(rmfuse) schneider@Aarons-MacBook-Pro ~ % rmfuse ~/remarkable -v
DEBUG:llfuse:Initializing llfuse
DEBUG:llfuse:Calling fuse_mount
DEBUG:llfuse:Calling fuse_lowlevel_new
fuse: unknown option `nodoubleapple'
Traceback (most recent call last):
File "/Users/schneider/anaconda3/envs/rmfuse/bin/rmfuse", line 8, in <module>
sys.exit(main())
File "/Users/schneider/anaconda3/envs/rmfuse/lib/python3.9/site-packages/rmfuse/fuse.py", line 563, in main
fuse.init(fs, options.mountpoint, fuse_options)
File "src/fuse_api.pxi", line 253, in llfuse.init
RuntimeError: fuse_lowlevel_new() failed
The reason seems to be the typo: noappledouble instead of nodoubleapple.
Also:
I can't install it the way you suggest it:
(rmfuse) schneider@Aarons-MacBook-Pro ~ % pip install rmfuse[llfuse]
zsh: no matches found: rmfuse[llfuse]
I suppose pip install llfuse rmfuse should work instead?
On Sun, Feb 28, 2021 at 10:55 PM Aaron David Schneider <
notifications@github.com> wrote:
The new version doesn't start on osx.
The reason seems to be the typo: noappledouble instead of nodoubleapple.
D'oh!
I can't install it the way you suggest it:
(rmfuse) schneider@Aarons-MacBook-Pro ~ % pip install rmfuse[llfuse]
zsh: no matches found: rmfuse[llfuse]
You'll need to add the --pre option, to look for pre-release versions in
pip.
I'll close this issue for now, and we can keep working on the large file upload problem in #18.
|
gharchive/issue
| 2021-02-08T08:50:58 |
2025-04-01T04:35:45.158210
|
{
"authors": [
"AaronDavidSchneider",
"rschroll"
],
"repo": "rschroll/rmfuse",
"url": "https://github.com/rschroll/rmfuse/issues/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
109598598
|
fixed your link
from http://www.codecur.io/us/songdemo to http://codecur.io/us/projects/song-demo
i'm not not mad if you don't accept this merge req.
It was just faster than opening a new issue and describe what i mean.
Awesome project, didn't expect to see a low-pass filter at 4:20am by a humble bundle link ;)
I just misunderstood. I though you meant the link in the repo description, but I see now that the link in the readme also needed to be updated. Thanks for the pull request!
|
gharchive/pull-request
| 2015-10-03T02:45:06 |
2025-04-01T04:35:45.176030
|
{
"authors": [
"AiyionPrime",
"rserota"
],
"repo": "rserota/wad",
"url": "https://github.com/rserota/wad/pull/48",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2022658524
|
Reboot activity categorizations
Right now there's too much present under "Contemporary" to find what you're looking for.
Once TFS drops, I'll revisit the need for this. Many activities will be vaulted, etc.
|
gharchive/issue
| 2023-12-03T18:31:21 |
2025-04-01T04:35:45.252788
|
{
"authors": [
"rslifka"
],
"repo": "rslifka/vault-zero",
"url": "https://github.com/rslifka/vault-zero/issues/27",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
955106870
|
metadata is malformed
how to use RSocketTCPClient#requestResponse?
Steps to Reproduce
import { RSocketClient, JsonSerializers } from "rsocket-core"
import RSocketTCPClient from "rsocket-tcp-client"
const route = "login"
const MIME_TYPE = "application/json"
const client = new RSocketClient({
serializers: JsonSerializers,
setup: {
keepAlive: 10000,
lifetime: 86400000,
dataMimeType: MIME_TYPE,
metadataMimeType: MIME_TYPE,
},
transport: new RSocketTCPClient({
host: "47.101.196.18",
port: 7000,
}),
})
client.connect().subscribe({
onError: error => console.error(error),
onSubscribe: cancel => {
console.log("onSubscribe")
console.log(cancel)
},
onComplete: rsocket => {
rsocket.requestResponse({
data: JSON.stringify({
username: "admin",
password: "123456",
}),
metadata: route,
}).subscribe({
onComplete: data => console.log(data),
onError: e => console.error(e),
})
console.log("onComplete")
}
})
My Environment
RSocket version(s) used:
"rsocket-core": "^0.0.27",
"rsocket-tcp-client": "^0.0.27"
node version:v14.17.3
Hi @yuri-li,
Can you provide some more details about what issues you are encountering? In which way is the observed result differing from your expectations? Is there an error thrown that you do not expect? If so, can you be more specific?
Hi @yuri-li,
Can you provide some more details about what issues you are encountering? In which way is the observed result differing from your expectations? Is there an error thrown that you do not expect? If so, can you be more specific?
hi, @viglucci ,
I am a Java development engineer, good at cloud native. I am afraid of misunderstanding, so the server (springboot.jar) is deployed on the cloud server, the host is real, and the front-end code(typescript) I gave can be run directly.
I ran the back-end code (kotlin coroutines+rsocket+kotest) first, and the test code is as follows:
@viglucci
hi, I found the solution
The front-end code after fixing the bug is as follows:
import { RSocketClient, BufferEncoders, encodeAndAddWellKnownMetadata, MESSAGE_RSOCKET_COMPOSITE_METADATA, MESSAGE_RSOCKET_ROUTING, encodeRoute } from "rsocket-core"
import RSocketTCPClient from "rsocket-tcp-client"
const route = "login"
const client = new RSocketClient({
setup: {
keepAlive: 10000,
lifetime: 86400000,
dataMimeType: "application/json",
metadataMimeType: MESSAGE_RSOCKET_COMPOSITE_METADATA.string,
},
transport: new RSocketTCPClient({
host: "47.101.196.18",
port: 7000,
}, BufferEncoders),
})
client.connect().subscribe({
onError: error => console.error(error),
onComplete: rsocket => {
const routeMetadata = encodeRoute(route)
const metadata = encodeAndAddWellKnownMetadata(
Buffer.alloc(0),
MESSAGE_RSOCKET_ROUTING,
routeMetadata
)
rsocket.requestResponse({
data: Buffer.from(JSON.stringify({
username: "admin",
password: "123456",
})),
metadata,
}).subscribe({
onError: e => console.error(e),
onComplete: payload => console.log((payload.data as Buffer).toString()),
})
console.log("onComplete")
}
})
Hi @yuri-li,
Thanks for following up and providing the solution you found. I believe this issue could be avoided if we had more extensive documentation around composite metadata, as well as more centralized and extensive examples. I'll keep this in mind as we improve the documentation and APIs moving forward.
I'll close this issue since you've come to a solution.
@yuri-li you have incorrect client configurations. Please see these examples to see how to utilize composite metadata -> https://github.com/rsocket/rsocket-js/blob/master/packages/rsocket-examples/src/CompositeMetadataExample.js
Also, here is a working js -> spring-boot-rsocket example https://github.com/OlegDokuka/rsocket-crosslanguage-example
|
gharchive/issue
| 2021-07-28T17:58:01 |
2025-04-01T04:35:45.280495
|
{
"authors": [
"OlegDokuka",
"viglucci",
"yuri-li"
],
"repo": "rsocket/rsocket-js",
"url": "https://github.com/rsocket/rsocket-js/issues/153",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
56555526
|
Add content to homepage of website
Explain the project
This is taken care of.
|
gharchive/issue
| 2015-02-04T17:02:22 |
2025-04-01T04:35:45.281704
|
{
"authors": [
"jenningsanderson",
"rsoden"
],
"repo": "rsoden/osm-history",
"url": "https://github.com/rsoden/osm-history/issues/6",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
324737
|
Netbeans is using "--runner" option when executing 'rake spec'
This problem is described here:
http://www.ruby-forum.com/topic/217187
Netbeans 7.0.1 has a Ruby/Rails Plugin and the issue still exists
|
gharchive/issue
| 2010-09-20T15:24:59 |
2025-04-01T04:35:45.282961
|
{
"authors": [
"PizzaPill",
"bmcardoso"
],
"repo": "rspec/rspec-core",
"url": "https://github.com/rspec/rspec-core/issues/150",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
99131823
|
'block in shellsplit' Error
Getting a random error when trying to run rake:
/Users/GMAC/.rvm/rubies/ruby-2.2.2/bin/ruby -I/Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-core-3.3.2/lib:/Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-support-3.3.0/lib /Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-core-3.3.2/exe/rspec --pattern spec/\*\*\{,/\*/\*\*\}/\*_spec.rb
/Users/GMAC/.rvm/rubies/ruby-2.2.2/lib/ruby/2.2.0/shellwords.rb:75:in `block in shellsplit': Unmatched double quote: "\" Press ? for help" (ArgumentError)
from /Users/GMAC/.rvm/rubies/ruby-2.2.2/lib/ruby/2.2.0/shellwords.rb:73:in `scan'
from /Users/GMAC/.rvm/rubies/ruby-2.2.2/lib/ruby/2.2.0/shellwords.rb:73:in `shellsplit'
from /Users/GMAC/.rvm/rubies/ruby-2.2.2/lib/ruby/2.2.0/shellwords.rb:189:in `shellsplit'
from /Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-core-3.3.2/lib/rspec/core/flat_map.rb:7:in `each'
from /Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-core-3.3.2/lib/rspec/core/flat_map.rb:7:in `flat_map'
from /Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-core-3.3.2/lib/rspec/core/flat_map.rb:7:in `flat_map'
from /Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-core-3.3.2/lib/rspec/core/configuration_options.rb:164:in `args_from_options_file'
from /Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-core-3.3.2/lib/rspec/core/configuration_options.rb:151:in `options_from'
from /Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-core-3.3.2/lib/rspec/core/configuration_options.rb:147:in `global_options'
from /Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-core-3.3.2/lib/rspec/core/configuration_options.rb:118:in `file_options'
from /Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-core-3.3.2/lib/rspec/core/configuration_options.rb:43:in `organize_options'
from /Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-core-3.3.2/lib/rspec/core/configuration_options.rb:13:in `initialize'
from /Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-core-3.3.2/lib/rspec/core/runner.rb:62:in `new'
from /Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-core-3.3.2/lib/rspec/core/runner.rb:62:in `run'
from /Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-core-3.3.2/lib/rspec/core/runner.rb:41:in `invoke'
from /Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-core-3.3.2/exe/rspec:4:in `<main>'
/Users/GMAC/.rvm/rubies/ruby-2.2.2/bin/ruby -I/Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-core-3.3.2/lib:/Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-support-3.3.0/lib /Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-core-3.3.2/exe/rspec --pattern spec/\*\*\{,/\*/\*\*\}/\*_spec.rb failed
Any ideas?
Can you paste the contents of your ~/.rspec file?
Thanks for leading me in the right direction. Apparently created a .rspec file in my root that had my NERDTree menu pasted into it (thats what I get for trying to learn Vim). Deleted the .rspec and everything works again!
|
gharchive/issue
| 2015-08-05T06:08:51 |
2025-04-01T04:35:45.285067
|
{
"authors": [
"gxespino",
"myronmarston"
],
"repo": "rspec/rspec-core",
"url": "https://github.com/rspec/rspec-core/issues/2053",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
208948468
|
Altering captured reference damages spark results.
If you use a variable in dplyr::mutate() against a sparklyr data source the lazy eval captures references to user variables. Changing values of those variables implicitly changes the mutate and changes the values seen in the sparklyr result (which is itself a query). This can be worked around by dropping in dplyr::compute() but it seems like it can produce a lot of incorrect calculations. Below is a small example and a lot information on the versions of everything beeing run. I am assuming the is a sparklyr issue as the query views are failrly different than a number of other dplyr structures, but it could be a dplyr issue.
OSX 10.11.6. Spark installed as described at http://spark.rstudio.com
library('sparklyr')
spark_install(version = "2.0.0")
library('dplyr')
#
# Attaching package: 'dplyr'
# The following objects are masked from 'package:stats':
#
# filter, lag
# The following objects are masked from 'package:base':
#
# intersect, setdiff, setequal, union
library('sparklyr')
R.Version()$version.string
# [1] "R version 3.3.2 (2016-10-31)"
packageVersion('dplyr')
# [1] '0.5.0'
packageVersion('sparklyr')
# [1] '0.5.2'
my_db <- sparklyr::spark_connect(version='2.0.0', master = "local")
class(my_db)
# [1] "spark_connection" "spark_shell_connection" "DBIConnection"
my_db$spark_home
# [1] "/Users/johnmount/Library/Caches/spark/spark-2.0.0-bin-hadoop2.7"
print(my_db)
# $master
# [1] "local[4]"
#
# $method
# [1] "shell"
#
# $app_name
# [1] "sparklyr"
#
# $config
# $config$sparklyr.cores.local
# [1] 4
#
# $config$spark.sql.shuffle.partitions.local
# [1] 4
#
# $config$spark.env.SPARK_LOCAL_IP.local
# [1] "127.0.0.1"
#
# $config$sparklyr.csv.embedded
# [1] "^1.*"
#
# $config$`sparklyr.shell.driver-class-path`
# [1] ""
#
# attr(,"config")
# [1] "default"
# attr(,"file")
# [1] "/Library/Frameworks/R.framework/Versions/3.3/Resources/library/sparklyr/conf/config-template.yml"
#
# $spark_home
# [1] "/Users/johnmount/Library/Caches/spark/spark-2.0.0-bin-hadoop2.7"
#
# $backend
# description class mode text opened
# "->localhost:58964" "sockconn" "wb" "binary" "opened"
# can read can write
# "yes" "yes"
#
# $monitor
# description class mode text opened
# "->localhost:58961" "sockconn" "rb" "binary" "opened"
# can read can write
# "yes" "yes"
#
# $output_file
# [1] "/var/folders/7q/h_jp2vj131g5799gfnpzhdp80000gn/T//Rtmpxkn3Al/filefeb634813b11_spark.log"
#
# $spark_context
# <jobj[5]>
# class org.apache.spark.SparkContext
# org.apache.spark.SparkContext@73d0c0e5
#
# $java_context
# <jobj[6]>
# class org.apache.spark.api.java.JavaSparkContext
# org.apache.spark.api.java.JavaSparkContext@4e527f99
#
# $hive_context
# <jobj[9]>
# class org.apache.spark.sql.SparkSession
# org.apache.spark.sql.SparkSession@1346ee7
#
# attr(,"class")
# [1] "spark_connection" "spark_shell_connection" "DBIConnection"
Expected outcome: s1 has the same value
Observed outcome: changing varaible v changes s1 column.
support <- copy_to(my_db,
data.frame(year=2005:2010),
'support')
v <- 0
s1 <- dplyr::mutate(support,count=v)
print(s1) # print 1
# Source: query [6 x 2]
# Database: spark connection master=local[4] app=sparklyr local=TRUE
#
# year count
# <int> <dbl>
# 1 2005 0
# 2 2006 0
# 3 2007 0
# 4 2008 0
# 5 2009 0
# 6 2010 0
# s1 <- dplyr::compute(s1) # likely work-around
v <- ''
print(s1) # print 2
# Source: query [6 x 2]
# Database: spark connection master=local[4] app=sparklyr local=TRUE
#
# year count
# <int> <chr>
# 1 2005
# 2 2006
# 3 2007
# 4 2008
# 5 2009
# 6 2010
Notice s1 changed its value (like due to lazy evaluation and having captured a reference to v).
version
# _
# platform x86_64-apple-darwin13.4.0
# arch x86_64
# os darwin13.4.0
# system x86_64, darwin13.4.0
# status
# major 3
# minor 3.2
# year 2016
# month 10
# day 31
# svn rev 71607
# language R
# version.string R version 3.3.2 (2016-10-31)
# nickname Sincere Pumpkin Patch
Thanks for the reproducible bug report!
It looks to me like this is a bug in dplyr -- the lazy evaluation performed doesn't capture the value of v when the query is generated, and so every time the query is 'rendered' it looks up what the current value of v is.
Sounds good, I have copied the content over to https://github.com/hadley/dplyr/issues/2455 .
Look like that it this is known in https://github.com/hadley/dplyr/issues/2370 and fixed in the dev versio of dplyr.
|
gharchive/issue
| 2017-02-20T17:56:41 |
2025-04-01T04:35:45.329636
|
{
"authors": [
"JohnMount",
"kevinushey"
],
"repo": "rstudio/sparklyr",
"url": "https://github.com/rstudio/sparklyr/issues/503",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
57287533
|
PDF build failing despite output being written, and PDF cannot be downloaded
My PDF build is marked as "Failed" besides output being written in the end.
It says in the log:
Output written on pypet.pdf (254 pages, 2253138 bytes).
However, when I check the documentation (http://pypet.readthedocs.org/en/latest/) and select to download the PDF, an older version of the documentation from last year is downloaded. Which is also weird, because the build corresponding to the old version is also marked as "Failed".
The failing was due to a Latex Error too deeply nested.
I added the following to the sphinx latex_elements and 'preamble' option to solve this:
\usepackage{enumitem}
\setlistdepth{99}
Downloading an older version was due to caching issues with my browser, sorry :-D
So can be closed :-)
|
gharchive/issue
| 2015-02-11T08:14:36 |
2025-04-01T04:35:45.370617
|
{
"authors": [
"SmokinCaterpillar"
],
"repo": "rtfd/readthedocs.org",
"url": "https://github.com/rtfd/readthedocs.org/issues/1153",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
67393989
|
Translation projects are not found
When I try to add a translation to a RTD project under Admin -> Translations I get the error:
Project https://github.com/tobias47n9e/innsbruck-stereographic/tree/master/docs_int/de was not found
I tried around with other URLs to see if I made a mistake, but at this point I think that the web-form has a bug.
The RTD-project is here: https://readthedocs.org/dashboard/innsbruck-stereographic
This isn't a bug, but it could definitely be done better. The form isn't clear about what it's asking for.
To link a translation, enter the project name as it exists in read the docs -- ie. 'innsbruck-stereographic'
Thank you. That turned out to be rather easy. I now also figured out how to assign different relative directories to different conf.py files. Going multilingual is not as hard as I thought it would be.
|
gharchive/issue
| 2015-04-09T16:03:30 |
2025-04-01T04:35:45.373356
|
{
"authors": [
"agjohnson",
"tobias47n9e"
],
"repo": "rtfd/readthedocs.org",
"url": "https://github.com/rtfd/readthedocs.org/issues/1220",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
199577545
|
Which branch for development
Hi, i'd like to add a new builder for doxygen documentation (but native, not with breath). Since there are a lot of branches like real/relcorp which a far ahead of master, i'd like to know, which branch to choose for development.
Thanks in advance!
Oli
Thanks for your interest. We likely wouldn't include a doxygen builder in code, but we'd be happy to make it easy to implement as a third party builder, with a good API in RTD. We'd also be happy to list the extension in our docs! :)
The main site is deployed from rel, which shouldn't be ahead of master.
Note that there are a few sphinx extensions for integrating doxygen content, notably https://github.com/michaeljones/breathe, you might want to check that out before starting from scratch.
Should I close this issue? The original question was answered (regarding to the branch) :grin:
Do you want to continue the discussion about how to it here? If so, we should change the title of the issue since it will be confusing.
We should probably document how we do deployments and that the rel branch is in production.
|
gharchive/issue
| 2017-01-09T15:20:26 |
2025-04-01T04:35:45.376531
|
{
"authors": [
"alyjak",
"ericholscher",
"humitos",
"zabeloliver"
],
"repo": "rtfd/readthedocs.org",
"url": "https://github.com/rtfd/readthedocs.org/issues/2583",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
538099426
|
App Args: Add tests, fix parsing of LitInt in application arguments
It appears that input.parse::<LitBool>() consumes part of the token
stream, so the subsequent call to input.parse::<LitInt>() does not
behave as expected.
This PR has one possible fix, but there may be a better fix using
fork or step (see https://docs.rs/syn/1.0.11/syn/parse/struct.ParseBuffer.html#method.fork)
Add tests.
Excellent! Thank you.
bors r+
Excellent! Thank you.
bors r+
Great 👍
Bors appears to have failed. What happened?
@richardeoin ah, sorry. I forgot to change the state of the review from "needs changes" to "approved" before sending this PR to bors that's why it didn't get processed by bors.
|
gharchive/pull-request
| 2019-12-15T20:12:02 |
2025-04-01T04:35:45.379807
|
{
"authors": [
"japaric",
"richardeoin"
],
"repo": "rtfm-rs/rtfm-syntax",
"url": "https://github.com/rtfm-rs/rtfm-syntax/pull/21",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
230514185
|
Support for fallback values w/ guaranteed order
This looks hacky... and it is! :-) But I believe it's the simplest way to support fallback values without having to make major changes in the core and affect performance. Fixes #6
Don't worry. I won't be sad if you decline this. :-P But it clearly works.
I don't love the idea of the property names in the value string, but I think this actually pretty robust. I hadn't thought of this approach before, cool stuff! 👍
There's a bit of a tradeoff here in terms of reduced de-duplication, but since this change fully guarantees order I think it's more robust implementation than what exists currently. Also, in practice, it seems that value fallbacks are often merely vendor prefixes, so this would have the benefit of collapsing these into a single atomic class (further compressing the CSS) sort of like https://github.com/rtsao/styletron/issues/104
Since this is related to https://github.com/rtsao/styletron/issues/104 I'd want to make sure this wouldn't cause any problems down the line.
I'm also open to updating the core to facilitating this in a first-class fashion, since admittedly this is pretty hacky and is essentially an abuse of the interface and implementation details.
What about something like:
styletron.injectRawDeclaration(
'color:red',
':hover',
'@media (max-width: 400px)'
);
styletron.injectRawDeclaration(
'color:red;color:blue',
':hover',
'@media (max-width: 400px)'
);
The internal cache structure would need to be changed from:
{
color: {red: 'a'},
pseudos: {
':hover': {
color: {red: 'b'}
}
}
}
to:
{
'color:red': 'a',
pseudos: {
':hover': {
'color:red': 'b'
}
}
}
I'm not sure what the performance implications would be, but in my original benchmarking the former with the original injectDeclaration was faster because there's never any string concatenation performed.
My intuition is that this new cache format would be faster in the case of new styles, but slower in the case of existing styles. Probably worth doing some benchmarking.
I don't love the idea of the property names in the value string
LOL, you're funny. You surly would be high if you'd love it. ;-)
Sure, an injectRawDeclaration is a good idea. But do you mean it should complete the original injectDeclaration with a separate cache or replace it?
But do you mean it should complete the original injectDeclaration with a separate cache or replace it?
I think the new cache format could just replace the original, then both injectDeclaration and injectRawDeclaration could shared the same cache. Additionally, I think at this point it might be worth deprecating the injectDeclaration method, then removing it from the next major version. Nobody really interacts with this low-level API directly so I don't see much value in having two different methods -- and since injectRawDeclaration is lower-level, we might as well support that one only.
I've added injectRawDeclaration to:
styletron-core
styletron-server
styletron-client
And injectStyle now uses injectRawDeclaration instead with full support of fallback values.
Merged with #145
|
gharchive/pull-request
| 2017-05-22T20:45:25 |
2025-04-01T04:35:45.401641
|
{
"authors": [
"faddee",
"rtsao"
],
"repo": "rtsao/styletron",
"url": "https://github.com/rtsao/styletron/pull/133",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1688274380
|
Do i need to write a windows driver to be able to talk to my custom USB HID device?
Im using an ESP32S2 which has native usb support and i want to be able to send reports between a rust program and my custom usb device but so far the only way i can read data from the device is on linux when i give sudo permissions to the device path. In windows i can open the device but when i send a report the rust program cant read it at all.
so far the only way i can read data from the device is on linux when i give sudo permissions to the device path
Not sure about the Windows question, but on Linux this is usually handled with an udev rule, e.g.:
ACTION=="add|change", SUBSYSTEMS=="usb", ATTRS{idVendor}=="04d9", ATTRS{idProduct}=="a052", MODE:="0666"
or
ACTION=="add|change", SUBSYSTEMS=="usb", ATTRS{idVendor}=="04d9", ATTRS{idProduct}=="a052", GROUP:="uucp"
If I think about it, that CO2 meter sample used to work fine for me on Windows, so it might be something specific to your device.
Where could i find that code sample?
|
gharchive/issue
| 2023-04-28T10:02:33 |
2025-04-01T04:35:45.427198
|
{
"authors": [
"Aggeloz",
"lnicola"
],
"repo": "ruabmbua/hidapi-rs",
"url": "https://github.com/ruabmbua/hidapi-rs/issues/112",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
739831733
|
RSpec/FactoryBot/CreateList fails with "undefined method `to_sym' for nil:NilClass"
The Rspec/FactoryBot/CreateList fails with undefined method to_sym for nil:NilClass.
It fails for both styles (create_list, n_times).
Mention the following information in the issue report:
1.2.0 (using Parser 2.7.2.0, rubocop-ast 1.1.1, running on jruby 2.5.7 java)
The following versions are used:
JRuby 9.2.13.0
rubocop 1.2.0
rubocop-rspec 1.42.0
Backtrace:
org/jruby/RubyKernel.java:1009:in `load'
/Users/someuser/.rbenv/versions/jruby-9.2.13.0/bin/rubocop:23:in `<main>'
org/jruby/RubyKernel.java:1009:in `load'
/Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/bundler-2.1.4/lib/bundler/cli/exec.rb:63:in `kernel_load'
/Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/bundler-2.1.4/lib/bundler/cli/exec.rb:28:in `run'
/Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/bundler-2.1.4/lib/bundler/cli.rb:476:in `exec'
/Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/bundler-2.1.4/lib/bundler/vendor/thor/lib/thor/command.rb:27:in `run'
/Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/bundler-2.1.4/lib/bundler/vendor/thor/lib/thor/invocation.rb:127:in `invoke_command'
/Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/bundler-2.1.4/lib/bundler/vendor/thor/lib/thor.rb:399:in `dispatch'
/Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/bundler-2.1.4/lib/bundler/cli.rb:30:in `dispatch'
/Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/bundler-2.1.4/lib/bundler/vendor/thor/lib/thor/base.rb:476:in `start'
/Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/bundler-2.1.4/lib/bundler/cli.rb:24:in `start'
/Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/bundler-2.1.4/exe/bundle:46:in `block in <main>'
/Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/bundler-2.1.4/lib/bundler/friendly_errors.rb:123:in `with_friendly_errors'
/Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/bundler-2.1.4/exe/bundle:34:in `<main>'
org/jruby/RubyKernel.java:1009:in `load'
/Users/someuser/.rbenv/versions/jruby-9.2.13.0/bin/bundle:23:in `<main>'
An error occurred while RSpec/FactoryBot/CreateList cop was inspecting /Users/someuser/dev/someproject/spec/support/helper.rb:12:17.
undefined method `to_sym' for nil:NilClass
/Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/rubocop-1.2.0/lib/rubocop/cop/mixin/configurable_enforced_style.rb:69:in `style'
/Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/rubocop-rspec-1.42.0/lib/rubocop/cop/rspec/factory_bot/create_list.rb:61:in `on_send'
org/jruby/RubyKernel.java:1960:in `public_send'
/Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/rubocop-1.2.0/lib/rubocop/cop/commissioner.rb:100:in `block in trigger_responding_cops'
/Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/rubocop-1.2.0/lib/rubocop/cop/commissioner.rb:166:in `with_cop_error_handling'
/Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/rubocop-1.2.0/lib/rubocop/cop/commissioner.rb:99:in `block in trigger_responding_cops'
org/jruby/RubyArray.java:1809:in `each'
/Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/rubocop-1.2.0/lib/rubocop/cop/commissioner.rb:98:in `trigger_responding_cops'
/Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/rubocop-1.2.0/lib/rubocop/cop/commissioner.rb:69:in `on_send'
/Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/rubocop-ast-1.1.1/lib/rubocop/ast/traversal.rb:158:in `block in on_send'
org/jruby/RubyArray.java:1809:in `each'
org/jruby/RubyEnumerable.java:1258:in `each_with_index'
/Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/rubocop-ast-1.1.1/lib/rubocop/ast/traversal.rb:155:in `on_send'
/Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/rubocop-1.2.0/lib/rubocop/cop/commissioner.rb:71:in `on_send'
/Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/rubocop-ast-1.1.1/lib/rubocop/ast/traversal.rb:153:in `on_def'
/Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/rubocop-1.2.0/lib/rubocop/cop/commissioner.rb:71:in `on_def'
/Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/rubocop-ast-1.1.1/lib/rubocop/ast/traversal.rb:136:in `block in on_dstr'
Line 69 from the rubocop mixin ConfigurableEnforcedStyle and surroundings:
def style
@style ||= begin
s = cop_config[style_parameter_name].to_sym
raise "Unknown style #{s} selected!" unless supported_styles.include?(s)
s
end
end
...
def style_parameter_name
'EnforcedStyle'
end
=> Looks like cop_config[style_parameter_name] is nil.
I think the issue is, that the cop_config only consists of:
{
"Enabled" => true
}
Working example:
{
"Description" => "Use Unix-style line endings.",
"StyleGuide" => "#crlf",
"Enabled" => true,
"VersionAdded" => "0.49",
"EnforcedStyle" => "native",
"SupportedStyles" => ["native", "lf", "crlf"]
}"
rubocop-rspec 1.42 is not compatible with RuboCop 1.2. Please update to a recently released rubocop-rspec 2.0.
@pirj thank you for pointing my to the 2.0 version. Somehow missed it.
|
gharchive/issue
| 2020-11-10T11:33:30 |
2025-04-01T04:35:45.456234
|
{
"authors": [
"ThomasKoppensteiner",
"pirj"
],
"repo": "rubocop-hq/rubocop-rspec",
"url": "https://github.com/rubocop-hq/rubocop-rspec/issues/1086",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
623634184
|
Fix a false positive for Lint/SuppressedException
This PR fixes a false positive for Lint/SuppressedException when empty rescue with comment in def.
% cat example.rb
def foo
do_something
rescue
# noop
end
% bundle exec rubocop --only Lint/SuppressedException
(snip)
Inspecting 1 file
W
Offenses:
example.rb:3:1: W: Lint/SuppressedException: Do not suppress exceptions.
rescue
^^^^^^
1 file inspected, 1 offense detected
The above code should not be warned because AllowComments: true by defualt.
Before submitting the PR make sure the following are checked:
[x] Wrote good commit messages.
[x] Commit message starts with [Fix #issue-number] (if the related issue exists).
[x] Feature branch is up-to-date with master (if not - rebase it).
[x] Squashed related commits together.
[x] Added tests.
[x] Added an entry to the Changelog if the new code introduces user-observable changes. See changelog entry format.
[x] The PR relates to only one subject with a clear title and description in grammatically correct, complete sentences.
[x] Run bundle exec rake default. It executes all tests and RuboCop for itself, and generates the documentation.
Looking forward to this 🎉
Can you make sure that inline comments are also ok? (See #7999) E.g.:
def foo
bar
rescue # ignore
end
Certainly! Also the following code seems to be able to do the same.
begin
bar
rescue # ignore
end
I think that these cases can be extended separately from this PR.
|
gharchive/pull-request
| 2020-05-23T10:57:21 |
2025-04-01T04:35:45.462291
|
{
"authors": [
"koic",
"marcandre"
],
"repo": "rubocop-hq/rubocop",
"url": "https://github.com/rubocop-hq/rubocop/pull/8017",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
713391809
|
Fix #8820 autocorrection for Style/IfWithSemicolon
This change fixes autocorrection for Style/IfWithSemicolon when elsif present, which was reported on #8820
In case of one of more elsif conditions present, the autocorrector will create a full if, elsif, else structure.
Before submitting the PR make sure the following are checked:
[x] Wrote good commit messages.
[x] Commit message starts with [Fix #issue-number] (if the related issue exists).
[x] Feature branch is up-to-date with master (if not - rebase it).
[x] Squashed related commits together.
[x] Added tests.
[x] Added an entry to the Changelog if the new code introduces user-observable changes. See changelog entry format.
[x] The PR relates to only one subject with a clear title and description in grammatically correct, complete sentences.
[x] Run bundle exec rake default. It executes all tests and RuboCop for itself, and generates the documentation.
@adrian-rivera ping :-)
@dvandersluis Would you be interested in taking over this PR and driving it over the finish line?
Sure! I'll take a look in the morning
@bbatsov I don't have access to push to the existing branch, should I create a new one on my fork and open a new PR?
@bbatsov this one should be closed now 😄
Thanks!
|
gharchive/pull-request
| 2020-10-02T06:45:08 |
2025-04-01T04:35:45.467828
|
{
"authors": [
"adrian-rivera",
"bbatsov",
"dvandersluis"
],
"repo": "rubocop-hq/rubocop",
"url": "https://github.com/rubocop-hq/rubocop/pull/8831",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2026522980
|
Make the Rubocop::Minitest::Test suite re-usable by gem consumers
I've been writing custom cops for various jobs for a while now, but I've never had a consistent way to write tests for products that use minitest as the test suite (officially Rubocop only provides rspec helpers) , I started to look around to see what this Gem did and figured it would make sense to make rubocop-minitest provide a minitest class for others that write custom cops for their own code bases.
This does a couple of things
Makes a new Rubocop::Minitest::Test class that includes the code from assert_offense (I don't feel like it needs to be a module), and it also uses declarative testing for people to use.
I moved the "real" test suites in this gem to use the new base class, and kept the code blocks being tested as Minitest::Test
I tested this locally in an app of mine and it works great! So it can allow a consistent way for people to write custom cops and easily test them!
Before submitting the PR make sure the following are checked:
[x] The PR relates to only one subject with a clear title and description in grammatically correct, complete sentences.
[x] Wrote good commit messages.
[-] Commit message starts with [Fix #issue-number] (if the related issue exists).
[x] Feature branch is up-to-date with master (if not - rebase it).
[x] Squashed related commits together.
[x] Added tests.
[x] Ran bundle exec rake default. It executes all tests and runs RuboCop on its own code.
[x] Added an entry (file) to the changelog folder named {change_type}_{change_description}.md if the new code introduces user-observable changes. See changelog entry format for details.
This makes a lot of sense to me and I'd love to use it.
If this is being exposed like this then I think it needs some documetation on how to pass global and cop specific config to a test.
See https://github.com/rubocop/rubocop-minitest/blob/b446022ea09b3f5558df9c0106c9e714c6fc1ec5/test/rubocop/cop/minitest/global_expectations_test.rb#L504-L513 and https://github.com/rubocop/rubocop-minitest/blob/b446022ea09b3f5558df9c0106c9e714c6fc1ec5/test/rubocop/cop/minitest/multiple_assertions_test.rb#L339-L343
@Earlopain Updated the casing of RuboCop, updated the sample docs (both inline and README.md), handling the nil case for @cop
Squashed commits and confirmed CI is passing locally
@koic 👋 Could I ask for a review?
@koic it's been a while, but I'd still like to get this merged
It's been a while since I last looked at this and this may be a dumb question to only come now but can you explain why this is needed?
Specifically I think there already exists what you are looking for: https://github.com/rubocop/rubocop-minitest/issues/155 and https://github.com/rubocop/rubocop-minitest/pull/156
https://github.com/rubocop/rubocop-minitest/blob/master/lib/rubocop/minitest/support.rb
There may be concern about it polluting the main test class but even then you can just make your own and mix that module in yourself.
|
gharchive/pull-request
| 2023-12-05T15:50:29 |
2025-04-01T04:35:45.477476
|
{
"authors": [
"DougEdey",
"Earlopain",
"exterm"
],
"repo": "rubocop/rubocop-minitest",
"url": "https://github.com/rubocop/rubocop-minitest/pull/278",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1594533112
|
Add new InternalAffairs/ProcessedSourceBufferName cop
Follow up https://github.com/rubocop/rubocop-minitest/pull/217#discussion_r1063097686.
This PR adds new InternalAffairs/ProcessedSourceBufferName cop. It enforces the use of processed_source.file_path instead of processed_source.buffer.name.
# bad
processed_source.buffer.name
# good
processed_source.file_path
Before submitting the PR make sure the following are checked:
[x] The PR relates to only one subject with a clear title and description in grammatically correct, complete sentences.
[x] Wrote good commit messages.
[ ] Commit message starts with [Fix #issue-number] (if the related issue exists).
[x] Feature branch is up-to-date with master (if not - rebase it).
[x] Squashed related commits together.
[x] Added tests.
[x] Ran bundle exec rake default. It executes all tests and runs RuboCop on its own code.
[ ] Added an entry (file) to the changelog folder named {change_type}_{change_description}.md if the new code introduces user-observable changes. See changelog entry format for details.
Looks good. Thanks!
|
gharchive/pull-request
| 2023-02-22T06:25:46 |
2025-04-01T04:35:45.482651
|
{
"authors": [
"bbatsov",
"koic"
],
"repo": "rubocop/rubocop",
"url": "https://github.com/rubocop/rubocop/pull/11606",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
427335871
|
Expire/Revalidate a Partner
A diaper bank needs to be able to expire a partner and get them to revalidate their information.
On the partner show page there should be some sort of "expire partner" or similar button which when pushed, sets the partner into needs recertification. It will send a request to the partnerbase app setting their status to recertification required and preventing them from requesting more diapers.
After the partner updates any new information they can submit for approval which changes their status on the diaperbank side ot awaiting review and it will follow as expected.
Hi @seanmarcia, @chaserx and I will work on this.
Thank you for describing the issue so clearly. Would you mind answering my clarification questions below please?
Requested changes
A diaper bank will be able to expire a partner who is currently approved (not a partner who is pending or awaiting approval)?
Are we happy with the words "expire partner" for the button or is there anything else you would prefer?
Are we happy with the status name "recertification_required" or would you prefer something else?
Once the partner's status has been changed to "recertification_required", they will still be able to see their historical diaper requests, but the button "Create new diaper request" will not have any functionality. Does this sound right to you?
Potential additional functionalities
Would the partner, which has received the status "recertification_required" need some kind of alert (an email perhaps) telling them this?
Are there other things that should trigger the "recertification_required" status in addition to the Diaper Bank choosing this? Eg something we might want to automate. I'm not sure of the exact reasons why recertification would be required, but wondering if there's something we already track which makes sense to trigger this status.
Is there any particular field that each partner requiring recertification will be required to update in their application? How will they know this? Again, I'm not sure of the exact use case so apologies if this question or any of the above are redundant.
Thank you very much 😃
Hey Amina! I'm excited you are continuing to contribute!
Let me try and answer your questions!
A diaper bank will be able to expire a partner who is currently approved (not a partner who is pending or awaiting approval)?
Correct!
Are we happy with the words "expire partner" for the button or is there anything else you would prefer?
Nice catch! I think "expire partner" probably conveys the wrong message. How about "re-certify partner"
Are we happy with the status name "recertification_required" or would you prefer something else?
Sounds great!
Once the partner's status has been changed to "recertification_required", they will still be able to see their historical diaper requests, but the button "Create new diaper request" will not have any functionality. Does this sound right to you?
Yep, that sounds accurate!
Potential additional functionalities
5) Would the partner, which has received the status "recertification_required" need some kind of alert (an email perhaps) telling them this?
This is a super good idea! I think there are a few ways this could be handled and I'm not sure if I have a preference. You could have the partner app send them an email letting them know that they need to do a routine recertification of their details so the diaper bank is up to date. You could make a big brightly coloured banner that would show up when they log in. Or both :)
6) Are there other things that should trigger the "recertification_required" status in addition to the Diaper Bank choosing this? Eg something we might want to automate. I'm not sure of the exact reasons why recertification would be required, but wondering if there's something we already track which makes sense to trigger this status.
This is also an excellent idea! The idea behind this is that diaper banks want up to date information on their partners for reporting purposes. It may make sense to have a field and flag in the organization page where they can set a yearly(?) partner recertification date? -- I'd definitely want to check with some of the the diaper bank stakeholders before going forward with this.
7) Is there any particular field that each partner requiring recertification will be required to update in their application? How will they know this? Again, I'm not sure of the exact use case so apologies if this question or any of the above are redundant.
Definitely not redundant! I love the questions and thought you've put into all of this! I don't think a partner will necessarily need to update any fields, if all their organization details haven't changed they can probably just click the button asking to be recertified.
Thanks again for all the amazing work you've done and continue to do!
Hey @Nirvikalpa108 and @chaserx Is this completed?
My apologies if you've been waiting on a review! I'll get to it shortly if it is ready! (Here and on the Partner side!)
Hi @seanmarcia yeah. It fell of my radar for a bit, but. Yes, I think that it's ready for review. @mdworken has weighed. I think I've addressed their concerns.
|
gharchive/issue
| 2019-03-30T22:43:38 |
2025-04-01T04:35:45.524926
|
{
"authors": [
"Nirvikalpa108",
"chaserx",
"seanmarcia"
],
"repo": "rubyforgood/diaper",
"url": "https://github.com/rubyforgood/diaper/issues/773",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1644733012
|
3429 reinviting accepted partners
Checklist:
Checked, although I didn't add documentation
Resolves #3429
Description
This makes it so that if someone has accepted an invitation to the organization, a new one is not sent out when clicking on "Create User" in the partner coworker screen.
The new behavior is being redirected to the coworker screen and a message appearing saying "(name) has already joined the organization
Pros: does intended behavior
Cons: added complexity and an additional check for the user before the invitation is sent
no dependencies
Type of change
New feature (non-breaking change which adds functionality)
This change requires a documentation update (I'm leaving this because I'm not sure it doesn't)
How Has This Been Tested?
Send an invitation to a new user
In a separate browser, open the invitation link and fill out the forms to sign in to the new account (this Is what counts as "accepting" the invitation, and should maybe change?)
In the original browser, invite the same person again (the email is what matters for identification)
verify that the email isn't sent and a message says "(name) has already joined the organization"
Screenshots
@cielf What are the next steps from here? Thanks for the help thus far.
@cielf would this still be an issue with the changes in #3372 ? I think this might conflict with that.
Hmm, the commit with the test didn't get pushed apparently. Before I update the pull request, what simplifications did you have in mind? I've found that all of the logic that is different from before the fix is necessary for the fix to work.
Hmm, the commit with the test didn't get pushed apparently. Before I update the pull request, what simplifications did you have in mind? I've found that all of the logic that is different from before the fix is necessary for the fix to work.
Hey Technical, I am just a fellow contributor! The conversation was quiet so I figured I'd drop my opinion. There's probably a good way to remove the nested if statement, but I only quickly skimmed the code. Good luck!
@technic-void my question was more for @cielf because the concept of "re-inviting a partner user" has changed somewhat in the most recent update of a different PR - I'm not sure this PR is actually needed if the other one is merged.
Hmm. I think they are two different things. 3429 says "If they've already accepted an invitation to this partner, you shouldn't be inviting them again."
Whereas with 3372, we're talking about a new partner/user relationship.
@cielf but in the other PR, we no longer send invites if the user already exists in the DB (regardless of if they belong to the current partner or a different partner, or an organization). So wouldn't that cover this?
See also the comments on the other PR
Adding this to next week's planning discussion in case we aren't both on the same page, and know it in the meantime.
I just checked the behaviour on main, and it does appear that this has been overtaken by other work. Thank you very much for the contribution, @technic-void ! We're going to close it, though.
|
gharchive/pull-request
| 2023-03-28T22:20:06 |
2025-04-01T04:35:45.533405
|
{
"authors": [
"cielf",
"dorner",
"lokisk1155",
"technic-void"
],
"repo": "rubyforgood/human-essentials",
"url": "https://github.com/rubyforgood/human-essentials/pull/3507",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2421330688
|
874 Add strong_migrations gem
🔗 Issue
Resolves https://github.com/rubyforgood/pet-rescue/issues/874
✍️ Description
Added strong_migrations gem
📷 Screenshots/Demos
You know what, I will review a couple PRs now that are close to the finish line. I recall one has migrations. I will let you know once that is done/merged, so we can update the config start_after date in this PR. Just saves people having to re-write migrations in PRs for now.
Re-commenting here because in the past people have not been notified if I request changes...
@egemen-dev I merged two PRs with migrations. Can you please update the config file to use the latest migration on main branch as the start point?
@kasugaijin updated the start_after value with the latest migration timestamp from the main branch.
|
gharchive/pull-request
| 2024-07-21T10:13:42 |
2025-04-01T04:35:45.537810
|
{
"authors": [
"egemen-dev",
"kasugaijin"
],
"repo": "rubyforgood/pet-rescue",
"url": "https://github.com/rubyforgood/pet-rescue/pull/883",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
392460284
|
How can I produce the test bitcoin?
I want to do a trade test between my 2 accounts. But I don't have bitcoin and money. How can I produce the test bitcoin?
@UfukAntep You need to deposit BTC through faucet to this accounts.
|
gharchive/issue
| 2018-12-19T06:35:25 |
2025-04-01T04:35:45.579509
|
{
"authors": [
"UfukAntep",
"mnaichuk"
],
"repo": "rubykube/peatio",
"url": "https://github.com/rubykube/peatio/issues/1909",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2249461829
|
Arrows don't work as expected in Firefox/Archlinux
Arrows UP/DOWN/LEFT/RIGHT, don't Move cell selection box. Instead, they change cell to edit mode.
Browser: Firefox 124.0.1
OS: Archlinux, Gnome (45 & 46)
In Chromium works as expected.
Cheers.
hi, thanks for reporting this. i dont have a copy of arch with me to test this, will check if this is the same on firefox on windows/mac. would it be possible for you to attach a screenrecording from https://ruilisi.github.io/fortune-sheet-demo/?path=/story/features--basic showing this behaviour?
upgraded to 19.4 and works correctly
|
gharchive/issue
| 2024-04-17T23:47:33 |
2025-04-01T04:35:45.634578
|
{
"authors": [
"pezr",
"sanchit3008"
],
"repo": "ruilisi/fortune-sheet",
"url": "https://github.com/ruilisi/fortune-sheet/issues/537",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2286928541
|
create llama@latest community template nextjs-llama3
using create-llama@0.1.1
1 week before it was working now i received errors when calling replicate api, no changes to template just adding api keys and running
using "llamaindex": "0.3.8"
using llamaindex 0.2.10
Had same problem with other class SimpleDirectoryReader.
Attempted import error: 'SimpleDirectoryReader' is not exported from 'llamaindex' (imported as 'SimpleDirectoryReader').
Did i do something wrong with the import? I tried these but same results:
import { SimpleDirectoryReader } from "llamaindex/readers/SimpleDirectoryReader";
import { SimpleDocumentStore, VectorStoreIndex ,SimpleDirectoryReader} from "llamaindex";
@jte0711 This is happening with llamaindex 0.2.10?
|
gharchive/issue
| 2024-05-09T05:35:15 |
2025-04-01T04:35:45.640911
|
{
"authors": [
"jozemario",
"jte0711",
"marcusschiesser"
],
"repo": "run-llama/create-llama",
"url": "https://github.com/run-llama/create-llama/issues/78",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
975673035
|
Quality of Life for Mouse Click Counter Plugin
Could it be possible to add an option in the settings for the Mouse Click Counter plugin that it resets upon every Login? It would be nice to see how many clicks you do in 1 session. I know you can manually reset it for this but having it automated would be nice.
In addition to the above, could you take it one step further to export the number of clicks per session into a notepad? Something like;
Login: 8/20/21 / 1400-1634 / 42124
Date / Time / Clicks
The Mouse Click Counter plugin is a hub plugin, and not maintained by the RuneLite team. Please open this request on their repository instead.
|
gharchive/issue
| 2021-08-20T14:41:35 |
2025-04-01T04:35:45.660745
|
{
"authors": [
"Nightfirecat",
"soccerplayr"
],
"repo": "runelite/runelite",
"url": "https://github.com/runelite/runelite/issues/14047",
"license": "BSD-2-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
327869002
|
Show what needs repairing and what doors needs closing in Pest control
Also some other optimisations would be cool to see. Like if something needs repairing, what doors need closing, what portals are weak against, total points, maybe even set a goal point amount and see your progress towards it.
Please let me know if these need to be separate issues.
Can I get this assigned?
Also, many people think closing doors wastes time. I'm up for discussion on this. It can easily be added because I already have the functionality to highlight all repairables.
Oh and because the points will be visible when in the minigame and on the main island I don't see why a goal feature would fit here. It might be better to add goals to something like #7912
|
gharchive/issue
| 2018-05-30T19:33:10 |
2025-04-01T04:35:45.662370
|
{
"authors": [
"RickertdeH",
"yanikore"
],
"repo": "runelite/runelite",
"url": "https://github.com/runelite/runelite/issues/3500",
"license": "BSD-2-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
338846256
|
Suggestion
It'd be cool if you could add a way to reset the max default zoom option when expand outer zoom is enabled. I like being able to zoom out more, but also like to return to the normal default zoom limit without having to turn off the zoom plugin.
Sounds like a duplicate of #919.
|
gharchive/issue
| 2018-07-06T08:14:12 |
2025-04-01T04:35:45.663319
|
{
"authors": [
"Nightfirecat",
"Runescaper123"
],
"repo": "runelite/runelite",
"url": "https://github.com/runelite/runelite/issues/4197",
"license": "BSD-2-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
302164804
|
Add Superantipoison to Timer Plugin
Noticed today that superantipoison is not supported and it would be nice to have it added please.
Resolved in #1159
|
gharchive/issue
| 2018-03-05T04:00:05 |
2025-04-01T04:35:45.664117
|
{
"authors": [
"Caia-OSRS",
"deathbeam"
],
"repo": "runelite/runelite",
"url": "https://github.com/runelite/runelite/issues/827",
"license": "BSD-2-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
466085230
|
mta: fix Telekinetic solver not moving
The bug happened when the Telekinetic projectile existed on the same game tick as the guardian moving.
Closes #3946
AFAICT this is due to a race with the projectiles being culled at the time the npc update moves the guardian. I can find no reason why this logic existed in the first place.
|
gharchive/pull-request
| 2019-07-10T03:56:04 |
2025-04-01T04:35:45.665358
|
{
"authors": [
"Adam-",
"raiyni"
],
"repo": "runelite/runelite",
"url": "https://github.com/runelite/runelite/pull/9328",
"license": "BSD-2-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
2188433512
|
Usages of klabel{}(_) attribute remain
When working on https://github.com/runtimeverification/k/issues/4035, I noticed that one test exercising the legacy Haskell backend failed after making the mechanical klabel(_), symbol refactorings implied in that issue: https://github.com/runtimeverification/k/tree/master/k-distribution/tests/regression-new/issue-1676-koreBytes
The error from the backend was:
$ make issue-1676-koreBytes
set -o pipefail; (cat 1.test.in 2>/dev/null || true) | /Users/brucecollie/code/k/k-distribution/bin/krun 1.test --output kore --no-exc-wrap --definition ./test-kompiled | diff - 1.test.out
kore-exec: [147057] Error (ErrorException):
Expecting hook 'BYTES.int2bytes' to reduce concrete pattern:
/* T Fn D Spa */
LblInt2Bytes'LParUndsCommUndsCommUndsRParUnds'BYTES-HOOKED'Unds'Bytes'Unds'Int'Unds'Int'Unds'Endianness{}(
/* T Fn D Sfa Cl */ \dv{SortInt{}}("0"),
/* T Fn D Sfa Cl */ \dv{SortInt{}}("0"),
/* T Fn D Sfa Cl */ LbllittleEndianBytes{}()
)
CallStack (from HasCallStack):
error, called at src/Kore/Rewrite/Axiom/EvaluationStrategy.hs:193:26 in kore-0.60.0.0-DddjtXjEFiY3CtblfXDv9H:Kore.Rewrite.Axiom.EvaluationStrategy
Expecting hook 'BYTES.int2bytes' to reduce concrete pattern:
/* T Fn D Spa */
LblInt2Bytes'LParUndsCommUndsCommUndsRParUnds'BYTES-HOOKED'Unds'Bytes'Unds'Int'Unds'Int'Unds'Endianness{}(
/* T Fn D Sfa Cl */ \dv{SortInt{}}("0"),
/* T Fn D Sfa Cl */ \dv{SortInt{}}("0"),
/* T Fn D Sfa Cl */ LbllittleEndianBytes{}()
)
CallStack (from HasCallStack):
error, called at src/Kore/Rewrite/Axiom/EvaluationStrategy.hs:193:26 in kore-0.60.0.0-DddjtXjEFiY3CtblfXDv9H:Kore.Rewrite.Axiom.EvaluationStrategy
Created bug report: kore-exec.tar.gz
[Error] krun: kore-exec ./test-kompiled/definition.kore --module TEST --pattern /var/folders/8r/bzvz46gd0r5_fr11jt23fzyw0000gn/T/.krun-2024-03-15-12-34-50-bXtH0QYDVx/tmp.in.2uQgOtQkNg --output /var/folders/8r/bzvz46gd0r5_fr11jt23fzyw0000gn/T/.krun-2024-03-15-12-34-50-bXtH0QYDVx/result.kore
[Error] krun: Backend crashed during rewriting with exit code 1
0a1,16
> /* T Fn D Sfa Cl */
> Lbl'-LT-'generatedTop'-GT-'{}(
> /* T Fn D Sfa Cl */
> Lbl'-LT-'k'-GT-'{}(
> /* T Fn D Sfa Cl */
> kseq{}(
> /* T Fn D Sfa Cli */
> /* Inj: */ inj{SortBytes{}, SortKItem{}}(
> /* T Fn D Sfa Cl */ \dv{SortBytes{}}("")
> ),
> /* T Fn D Sfa Cl */ dotk{}()
> )
> ),
> /* T Fn D Sfa Cl */
> Lbl'-LT-'generatedCounter'-GT-'{}(/* T Fn D Sfa Cl */ \dv{SortInt{}}("0"))
> )
make: *** [1.test] Error 1
Some digging into the code reveals that this is because the backend treats Endianness and Signedness terms as special-cased builtins, but because these are not K hooked symbols, the backend needs to inspect the klabel{}(_) attribute to check for special-casing.
I have opened a draft fix in https://github.com/runtimeverification/haskell-backend/pull/3741 that simply looks for symbol'Kywd' instead of klabel in these cases (the only such places klabel is used). With this fix applied, I can pass the K regression test suite. However, there is now a dependency loop between the frontend and backend that we need to resolve to merge the fix. If there's a way to first look at klabel, then symbol'Kywd' if there's no klabel, then I think that would address the issue. We would then be able to update the K dependency for the integration tests and immediately remove the backwards compatibility.
As an update, I opened this Frankenstein PR: https://github.com/runtimeverification/haskell-backend/pull/3744
It demonstrates that if https://github.com/runtimeverification/k/pull/4045 is merged along with https://github.com/runtimeverification/haskell-backend/pull/3741, that the integration test suite here is all successful. We still need to figure out exactly how to close this loop though.
|
gharchive/issue
| 2024-03-15T12:43:09 |
2025-04-01T04:35:45.695618
|
{
"authors": [
"Baltoli"
],
"repo": "runtimeverification/haskell-backend",
"url": "https://github.com/runtimeverification/haskell-backend/issues/3742",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
2179140573
|
cheatcodes.md: use buf instead of bufStrict in freshUInt
closes: https://github.com/runtimeverification/evm-semantics/issues/2038
It does save us a simplification step, since that #bufStrict will reduce to #buf. I wonder if we could drop [preserves-definedness] now...
|
gharchive/pull-request
| 2024-03-11T13:29:11 |
2025-04-01T04:35:45.697450
|
{
"authors": [
"PetarMax",
"anvacaru"
],
"repo": "runtimeverification/kontrol",
"url": "https://github.com/runtimeverification/kontrol/pull/430",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
529335776
|
Docs for the actual API?
The docs only seem to explain how to use the example UI.
I imagine most people are only interested in the provided UI as an example. What they will really want to do is apply those techniques to their own apps.
I'm in the process of going through the code in RunwayHub to figure out where the useful stuff happens. It would be really helpful if this project contained a minimal example without all the cruft.
Maybe a Monobehaviour that took a Texture2D input and renderered to another Texture2D?
It's a bit annoying to have to deconstruct a fairly complex example just to do the most basic thing and a minimal example would be a real time saver for others.
Wow. Six months and no reply?
|
gharchive/issue
| 2019-11-27T13:08:31 |
2025-04-01T04:35:45.705799
|
{
"authors": [
"andybak"
],
"repo": "runwayml/RunwayML-for-Unity",
"url": "https://github.com/runwayml/RunwayML-for-Unity/issues/5",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
422965366
|
Pr.wyrm
I've followed this guide
step by step carefully. Yes
I've tested this package remotely and have excluded all broken builds.
Here is the links to the Travis/AppVeyor with status "All passed":
https://ci.appveyor.com/project/a-sevin/hunter/builds/23203978
https://travis-ci.org/a-sevin/hunter/builds/508609964
I have submitted CI configs to https://github.com/ingenue/hunter targeting pkg.template branch,
see this merged pull request https://github.com/ingenue/hunter/pull/348
Testing:
https://ci.appveyor.com/project/ingenue/hunter/builds/23221977
https://travis-ci.org/ingenue/hunter/builds/508924280?utm_source=github_status&utm_medium=notification
update wyrm package to add pybind11 dependency
https://travis-ci.org/a-sevin/hunter/builds/508964942
https://ci.appveyor.com/project/a-sevin/hunter/builds/23224246
Testing:
https://ci.appveyor.com/project/ingenue/hunter/builds/23224721
https://travis-ci.org/ingenue/hunter/builds/508972654?utm_source=github_status&utm_medium=notification
Uploading:
https://ci.appveyor.com/project/ingenue/hunter/builds/23226072
https://travis-ci.org/ingenue/hunter/builds/508991900?utm_source=github_status&utm_medium=notification
Released: https://github.com/ruslo/hunter/releases/tag/v0.23.141
|
gharchive/pull-request
| 2019-03-19T21:36:28 |
2025-04-01T04:35:45.735522
|
{
"authors": [
"a-sevin",
"ruslo"
],
"repo": "ruslo/hunter",
"url": "https://github.com/ruslo/hunter/pull/1790",
"license": "bsd-2-clause",
"license_type": "permissive",
"license_source": "bigquery"
}
|
453600110
|
Update 'ethash' to v0.5.0
I've followed this guide
step by step carefully. [Yes]
I've tested this package remotely and have excluded all broken builds.
Here is the links to the Travis/AppVeyor with status "All passed":
https://ci.appveyor.com/project/chfast/hunter/builds/25124748
https://travis-ci.org/chfast/hunter/builds/542848354
Testing:
https://ci.appveyor.com/project/ingenue/hunter/builds/25151795
https://travis-ci.org/ingenue/hunter/builds/543424429?utm_source=github_status&utm_medium=notification
Released: https://github.com/ruslo/hunter/releases/tag/v0.23.190
|
gharchive/pull-request
| 2019-06-07T16:37:49 |
2025-04-01T04:35:45.740094
|
{
"authors": [
"chfast",
"ruslo"
],
"repo": "ruslo/hunter",
"url": "https://github.com/ruslo/hunter/pull/1893",
"license": "bsd-2-clause",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1408488835
|
Submit PR per: https://github.com/microsoft/vscode/issues/78013
Wondering if you submitted the PR: https://github.com/microsoft/vscode/issues/78013
Thanks.
I never got around to it, as I've been using Jetbrains stuff over VS Code lately. Maybe someday, or feel free to take a stab at it.
Alas, I don't have the expertise. Thanks for the very quick response.
|
gharchive/issue
| 2022-10-13T22:17:57 |
2025-04-01T04:35:45.744500
|
{
"authors": [
"HarjitSi",
"russelldavis"
],
"repo": "russelldavis/vscode-better-pageup-pagedown",
"url": "https://github.com/russelldavis/vscode-better-pageup-pagedown/issues/7",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2473035106
|
private_key_debug_is_obfuscated test is in the wrong crate
This test belongs to secp256k1, not bitcoin. The current situation is fragile and annoying.
We can probably delete the tests if https://github.com/rust-bitcoin/rust-secp256k1/pull/722 merges.
can i take on this issue ?
We can probably delete the tests if rust-bitcoin/rust-secp256k1#722 merges.
It looks like the PR has been merged, so now it's just a matter of removing private_key_debug_is_obfuscated?
Yep, let's do it.
Looks like this is can be closed - fixed by #3517
Thanks man, its rad that you notice these things.
|
gharchive/issue
| 2024-08-19T10:38:33 |
2025-04-01T04:35:45.754558
|
{
"authors": [
"Kixunil",
"apoelstra",
"diogocanut",
"mahmudsudo",
"shinghim",
"tcharding"
],
"repo": "rust-bitcoin/rust-bitcoin",
"url": "https://github.com/rust-bitcoin/rust-bitcoin/issues/3186",
"license": "CC0-1.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2244571368
|
Allow deprecated Params field
I'm not sure why I haven't see this before during the whole test cycle but while running cargo kani --only-codegen we get a bunch of warnings of form:
warning: use of deprecated field consensus::params::Params::pow_limit
We deprecated the pow_limit field but still set it (obviously) in const structs - just shoosh the warning.
Found while investigating the current kani CI failures.
Rebased to pick up the kani fix on master.
cross test fail is network related, I"ll re-run it once the pipeline finishes.
|
gharchive/pull-request
| 2024-04-15T20:37:44 |
2025-04-01T04:35:45.756576
|
{
"authors": [
"tcharding"
],
"repo": "rust-bitcoin/rust-bitcoin",
"url": "https://github.com/rust-bitcoin/rust-bitcoin/pull/2687",
"license": "CC0-1.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
787590844
|
FreeBSD: add libssp
This is required to build some native projects
which is build with -fstack-protector when linked.
Also update sqlite3 version (quarterly is updated)
Needs a rebase.
bors r+
|
gharchive/pull-request
| 2021-01-17T01:26:44 |
2025-04-01T04:35:45.768745
|
{
"authors": [
"junhochoi",
"reitermarkus",
"svenstaro"
],
"repo": "rust-embedded/cross",
"url": "https://github.com/rust-embedded/cross/pull/511",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
704881938
|
WIP: First steps rewrite
A rewrite of chapter 04 and partially chapter 05.
I'm already uploading this now since the cargo-embed GDB stub is blowing up on my machine when trying to debug this chip so I'll probably have to spend some time debugging, however during that time my current work can already be reviewed so feel free to take a look at it!
:warning: Warning :warning:
Pull requests are usually filed against the master branch for this repo, but this one is against rewrite. Please double check that you specified the right target!
I left a few notes asking for resources to link to in chapter 5 in the code, would be nice if someone could come up with some suggestions!
The PR now contains everything that is required for ch04 and 05 and should be ready for review.
Once review of the last parts passes the only thing we'd need is a nice resource for LED matrices and could actually merge this \o/
|
gharchive/pull-request
| 2020-09-19T13:21:50 |
2025-04-01T04:35:45.771084
|
{
"authors": [
"hargoniX",
"rust-highfive"
],
"repo": "rust-embedded/discovery",
"url": "https://github.com/rust-embedded/discovery/pull/271",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
370750536
|
Make digital traits fallible
This makes the digital traits fallible as discussed in #97, #95, #100 and #41 implemented as a clean break.
I'm fine with merging this, but there's discussion in #92 about releasing a last 0.2.x release before merging breaking changes.
Closing in favor of #108 as agreed on #100.
|
gharchive/pull-request
| 2018-10-16T18:52:10 |
2025-04-01T04:35:45.772514
|
{
"authors": [
"eldruin",
"hannobraun"
],
"repo": "rust-embedded/embedded-hal",
"url": "https://github.com/rust-embedded/embedded-hal/pull/105",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
304730516
|
Using Deref in traits
Currently the guidelines say "Only smart pointers implement Deref and DerefMut (C-DEREF)".
In this PR in rand we are considering using this trick:
impl<R: RngCore + ?Sized, T: DerefMut<Target = R>> RngCore for T { ... }
With the comment
This isn't abusing Deref for inheritance like the anti-pattern you listed; it's implementing a trait for anything Deref-able to a type implementing the trait.
Would this be a good exception to the rule?
That RngCore impl does not implement Deref or DerefMut for anything that is not a smart pointer, so the C-DEREF guideline has no bearing. But as a downside you should consider that adding this impl would make RngCore-related error messages much worse. Blanket impls (meaning for T where T is a type parameter) are pretty unusual and should be used sparingly. All in all I would be mildly against adding this impl.
+ use std::ops::DerefMut;
trait RngCore {}
+ impl<R: RngCore + ?Sized, T: DerefMut<Target = R>> RngCore for T {}
fn assert_rng_core<T: RngCore>() {}
fn main() {
assert_rng_core::<String>();
}
Error message without impl
error[E0277]: the trait bound `std::string::String: RngCore` is not satisfied
--> src/main.rs:10:5
|
10 | assert_rng_core::<String>();
| ^^^^^^^^^^^^^^^^^^^^^^^^^ the trait `RngCore` is not implemented for `std::string::String`
|
= note: required by `assert_rng_core`
Error message after adding impl
error[E0277]: the trait bound `str: std::marker::Sized` is not satisfied
--> src/main.rs:10:5
|
10 | assert_rng_core::<String>();
| ^^^^^^^^^^^^^^^^^^^^^^^^^ `str` does not have a constant size known at compile-time
|
= help: the trait `std::marker::Sized` is not implemented for `str`
= note: required because of the requirements on the impl of `RngCore` for `str`
= note: required because of the requirements on the impl of `RngCore` for `std::string::String`
= note: required by `assert_rng_core`
error[E0277]: the trait bound `str: std::ops::DerefMut` is not satisfied
--> src/main.rs:10:5
|
10 | assert_rng_core::<String>();
| ^^^^^^^^^^^^^^^^^^^^^^^^^ the trait `std::ops::DerefMut` is not implemented for `str`
|
= note: required because of the requirements on the impl of `RngCore` for `str`
= note: required because of the requirements on the impl of `RngCore` for `std::string::String`
= note: required by `assert_rng_core`
https://github.com/rust-lang-nursery/rand/pull/291
|
gharchive/issue
| 2018-03-13T11:43:03 |
2025-04-01T04:35:45.795840
|
{
"authors": [
"dtolnay",
"pitdicker"
],
"repo": "rust-lang-nursery/api-guidelines",
"url": "https://github.com/rust-lang-nursery/api-guidelines/issues/158",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
480860896
|
Expose features from rand to avoid compilation breakage on WASM
Current error:
error: target is not supported, for more information see: https://docs.rs/getrandom/#unsupported-targets
--> /home/user/.cargo/registry/src/github.com-1ecc6299db9ec823/getrandom-0.1.9/src/lib.rs:249:9
|
249 | / compile_error!("\
250 | | target is not supported, for more information see: \
251 | | https://docs.rs/getrandom/#unsupported-targets\
252 | | ");
| |___________^
In order to use join! and select! macros, we require the user to enable the
async-await and nightly features. However those features will pull in
rand. In order to work on WASM rand needs to have either the wasm-bindgen
or stdweb feature enabled. Otherwise compile time breakage happens due to the
getrandom crate.
Explanation can be found here: https://docs.rs/getrandom/0.1.9/getrandom/#unsupported-targets
This commit exposes those features through futures-util, allowing the end user to enable them
when compiling on WASM. This means futures does not need to choose which feature to enable
which would take away that decision from the end user.
An alternative would be to detect the wasm32 target in Cargo.toml and to choose
one of the features automatically.
Current approach requires action on part of the user to enable one of those features,
which means it should probably be documented somewhere. Currently the documentation
of join! and select! does not mention the need for the async-await and nightly
features.
I propose we add some documentation changes to this PR before merging. Following a similar
approach to getrandom by generating a compile time error to explain the situation to the user
might be the lowest friction for the end user. eg. when feature async-await is enabled on wasm32, verify
that one of the two required features is enabled as well, if not throw error. I can add that here
if it seems a desirable solution.
I have not tested the emscripten and WASI targets.
On some quick testing this alleviates the compile error on my system, but it would
be good if someone else had a look, or if we had CI testing for WASM.
Ps: another solution would be to see with getrandom why they prefer breaking rather than selecting wasm-bindgen as a default. That would simplify things.
@newpavlov ?
wasm32-unknown-unknown can not make any assumptions about a target on which it will be executed. It can be a browser, Node.JS, smart-contract, game module system, etc. So we simply don't have any "system" entropy source, thus we can not make wasm-bindgen feature enabled by default.
I am not sure if you need wasm-bindgen and stdweb features in this crate. The idea was that users will enable one of those feature only in the top level application crate.
@newpavlov The problem is that futures-util pulls in rand which pulls in getrandom. So the end user doesn't necessarily have getrandom in their Cargo.toml. They will depend on futures however.
I just verified and adding getrandom with the wasm-bindgen feature in my crate fixes the issue without modification to futures. If that is the desired approach, we need to document this!
If that is the desired approach, we need to document this!
Yes, I think it's the desired approach. If you have ideas of how we can improve getrandom docs, feel free to open a PR!
Ok, I filed rust-random/getrandom#89
I still think we should add section about "target platforms" on the readme of futures, and one explaining the "feature flags" too...
I still think we should add section about "target platforms" on the readme of futures
I think having to re-document this in every crate that might transitively depend on getrandom is not worth it. A major reason to pull in semi-standard crates like this instead of rerolling randomness in every crate that needs it is to centralise platform support, it should be well documented by getrandom how to get it working on any odd platforms if it has partial support for them.
(I would not consider wasm32-unknown-unknown a platform that is always expected to work out of the box with std depending crates since it is lacking fundamental OS features like randomness. If there were no_std support for select then I would expect it to be possible to use that on wasm32-unknown-unknown, which reminds me I should open an issue about getting that working.)
@Nemo157 Have you seen the discussion on the getrandom repo? I am proposing some improvements to the docs there. It turns out that this is a features application devs have to put, even if getrandom get's pulled in by a dependency.
I would propose afterwards that we do put a section about WASM in the readme where we explain things like this. I shall close this PR to avoid confusion.
|
gharchive/pull-request
| 2019-08-14T19:57:28 |
2025-04-01T04:35:45.806762
|
{
"authors": [
"Nemo157",
"najamelan",
"newpavlov"
],
"repo": "rust-lang-nursery/futures-rs",
"url": "https://github.com/rust-lang-nursery/futures-rs/pull/1804",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1727890561
|
Expose the Path of Cargo.toml in cargo-metadata output
Problem
There are a few cases in tooling (I am thinking of cases primarily inside of rust-analyzer, but I'm sure there are others) where it'd be very helpful to deterministically determine the crate's root folder (e.g., the folder containing Cargo.toml) in order to present the crate in a meaningful and idiomatic fashion and without resorting to approximations like "search parent folders of the crate root for a Cargo.toml".
For context, both Buck2 and Bazel—which don't enforce structures on the crates that they build—both allow the user to query the build system to determine the file that defines the build target.
Proposed Solution
When emitting the crate root (the Rust file, not the Cargo.toml!), cargo-metadata should also emit the absolute path of the Cargo.toml that defines the crate. In the case of a Cargo workspace, I'd guess that the output should also return the Cargo.toml that defines Cargo workspace if the crate is not downloaded from crates.io.
(I assume this information already exists within Cargo, but it's not exposed and hasn't been asked for before, which is why I didn't see an issue for this request.)
Notes
I'd be happy to add this feature to Cargo's metadata output.
Could you clarify what you are looking for. For example, cargo metadata includes a Package::manifest_path field but I assume you saw that, so I'm trying to understand what is still missing.
Could you clarify what you are looking for. For example, cargo metadata includes a Package::manifest_path field but I assume you saw that, so I'm trying to understand what is still missing.
That's very generous of you to assume that I saw Package::manifest_path, but I unfortunately did not! I completely missed its existence—that's exactly what I'm looking for. My apologies for the noise; I'll close this issue.
|
gharchive/issue
| 2023-05-26T16:08:22 |
2025-04-01T04:35:45.828203
|
{
"authors": [
"davidbarsky",
"epage"
],
"repo": "rust-lang/cargo",
"url": "https://github.com/rust-lang/cargo/issues/12187",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2015144966
|
gc: Determine CLI design for manual cleaning
The current implementation from https://github.com/rust-lang/cargo/pull/12634 exposes a cargo clean gc subcommand to handle manually cleaning cache data. It is not clear what the final CLI design should be (and it is not clear exactly what the user scenarios are for when they would want to take manual control). This issue is tracking for determining what the CLI should look like. There are few different considerations:
What should the actual subcommand be called? There were a few different considerations:
cargo cache is already in use by a third-party command
cargo gc (reserved by a third-party, but unused), was part of early proposals (cargo doesn't handle unstable top-level subcommands very well)
cargo clean gc — the current implementation
cargo clean — just fold the functionality into a single command which handles cleaning caches. @epage has concerns that this is overloading a subcommand used for different types of caches (local vs global, etc.).
cargo maintenance — just an idea stolen from git
How does this evolve with cleaning target directories, global build caches, etc.?
Which flags should be exposed? Currently it exposes a large number of low-level flags. Ideally I would like to only stabilize a small number of higher-level flags. But I don't know what the high-level user scenarios might be, so I'm uncertain exactly what to expose.
For example, one idea is to have a flag for "things that can be recreated" and another for "things that require downloading".
Also bikeshed the names of the options.
Potential sources of prior of other tools with cache management
poetry
docker
git
nix, zig
dune, etc)
Apparently buck2 and Turborepo grow unbounded
Quick scan of brew
autoremove
cleanup
"Removes all downloads more than 120 days old. This can be adjusted with HOMEBREW_CLEANUP_MAX_AGE_DAYS."
HOMEBREW_CLEANUP_PERIODIC_FULL_DAYS (default 30 days)
One complaint that came up was "brew cleanup has not been run in 30 days, running now" ... and then proceeds to run an interminable process in the middle of you attempting to do something else." (mastadon)
We should probably step back and enumerate what the required use cases are and the "if it works" use cases.
I like the way pnpm approaches this,
https://pnpm.io/cli/store
Running pnpm store prune is not harmful and has no side effects on your projects. If future installations require removed packages, pnpm will download them again.
It is best practice to run pnpm store prune occasionally to clean up the store, but not too frequently. Sometimes, unreferenced packages become required again. This could occur when switching branches and installing older dependencies, in which case pnpm would need to re-download all removed packages, briefly slowing down the installation process.
Regarding the concern of deleting crates that might still be in use, I like how rushjs asks you to pass an --unsafe flag to it's purge command,
https://rushjs.io/pages/commands/rush_purge/
--unsafe (UNSAFE!) Also delete shared files such as the package manager
instances stored in the ".rush" folder in the user's home
directory. This is a more aggressive fix that is NOT SAFE to
run in a live environment because it will cause other
concurrent Rush processes to fail.
@juliusl
https://pnpm.io/cli/store
it sounds like pnpm store prune removes everything that isn't referenced, regardless of age or size.
In #13137 I bring up the idea to build on top of the work to track workspaces in #13136 so we pin entries not in current lockfiles.
To extend this to clean up everything has the risk is if a project is transient (e.g. removable media) or moved but a new command wasn't run to register the new location. If its manually done with a command, rather than part of the auto-gc, then that might be reasonable, especially if we swap the logic and have a --execute flag rather than a --dry-run flag.
Regarding the concern of deleting crates that might still be in use, I like how rushjs asks you to pass an --unsafe flag to it's purge command,
I don't think an --unsafe flag is as relevant. For any shared directories, we use filesystem locks to ensure consistent reads/writes.
Any "in use" concerns we have are more about "relevant to the user" and not "file descriptors are open" and is mostly relevant for slow networks/systems and offline usage (I don't want a rarely used dependency being removed just before I go on an airplane to do development offline).
|
gharchive/issue
| 2023-11-28T19:23:28 |
2025-04-01T04:35:45.842025
|
{
"authors": [
"ehuss",
"epage",
"juliusl"
],
"repo": "rust-lang/cargo",
"url": "https://github.com/rust-lang/cargo/issues/13060",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
929334907
|
Failed to update git submodule of dependency
Problem
Cargo can't clone/update a git submodule of dependency, for example redbpf crate, which contains submodule libbpf.
Steps
cargo new cargo_issue_git_submodule
cd cargo_issue_git_submodule
echo 'redbpf = { git = "https://github.com/foniod/redbpf" }' >> Cargo.toml
cargo check
And we will get an error:
error: failed to get `redbpf` as a dependency of package `cargo_issue_git_submodule v0.1.0 (/home/user/cargo_issue_git_submodule)`
Caused by:
failed to load source for dependency `redbpf`
Caused by:
Unable to update https://github.com/foniod/redbpf
Caused by:
failed to update submodule `bpf-sys/libbpf`
Caused by:
object not found - no match for id (5a10cd2060c19959e2c55ee33421baa358c09c0b); class=Odb (9); code=NotFound (-3)
Possible Solution(s)
Update submodule manually in ~/.cargo/git/checkouts/redbpf-HASH/b81f069/bpf-sys/libbpf.
Notes
cargo version: cargo 1.55.0
commit of cargo: 4c27c96645e235d81f6c8dfff03ff9ebaf0ef71d
It looks like the commit 5a10cd2060c19959e2c55ee33421baa358c09c0b does not exist on any branch or tag in the repository https://github.com/redsift/libbpf.git. The way cargo fetches submodules needs to assume that it exists on some branch or tag.
Unfortunately I can't find any APIs in libgit2 that provide a way to fetch a specific commit, so I don't see a direct way this can be supported. One option is to restructure the net.git-fetch-with-cli option so that it uses git submodule update --init instead of git fetch (or pass the direct revision to git fetch which is supported as of git 2.5).
This is something i'm currently running into that i'd really like to see resolved. At the current moment this is making it so that a submodule of one of my dependencies must be on the main/master branch, and that's not exactly doable
|
gharchive/issue
| 2021-06-24T15:14:10 |
2025-04-01T04:35:45.848134
|
{
"authors": [
"daniel-brenot-apcapital",
"ehuss",
"saruman9"
],
"repo": "rust-lang/cargo",
"url": "https://github.com/rust-lang/cargo/issues/9622",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1161797667
|
Update --timings=json output include data similar to the HTML output
Content
This PR attempts to bring the JSON output when using the --timings=json flag more inline with the data that is output when requesting HTML.
Added custom Serialization impls to the Timings, and Unit structs.
Added a report_json function to output the data in a manner similar to the HTML output.
Removed the output of the timing reports on stdout during unit completion when the --timings=json flag was used. That information in now collected in the final report.
Testing
This was my method. There is probably a better way.
Built a release version of Cargo. cargo build --release
In a test project ran cargo clean && <path-to-cargo>/target/release/cargo build --timings=json -Z unstable-options
Verified that JSON output was generated in the <basedir>/target/cargo-timings/ folder.
r? @alexcrichton
(rust-highfive has picked a reviewer for you, use r? to override)
Thanks for the PR, but I'm going to be stepping down from the Cargo team so I'm going to un-assign myself from this. The Cargo team will help review this when they get a chance.
r? @ehuss
|
gharchive/pull-request
| 2022-03-07T18:54:05 |
2025-04-01T04:35:45.853273
|
{
"authors": [
"CraZySacX",
"alexcrichton",
"rust-highfive"
],
"repo": "rust-lang/cargo",
"url": "https://github.com/rust-lang/cargo/pull/10463",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
463718854
|
add env variable to retrieve workspace root path
Closes https://github.com/rust-lang/cargo/issues/7036
r? @matklad @alexcrichton
The current work-around for this is to launch cargo metadata, which is not to pretty: https://github.com/mitsuhiko/insta/commit/943e7bca885140d6253a0ee97127e0f963403eac
The potential gotcha here is that workspace situation is different between locally developed crate, and crate as it exists on crates.io: in the latter case, we remove all workspace structure and just store a flat list of crates.
@matklad
tried to build a crate with this patch, output this snippet:
CARGO_MANIFEST_DIR=/Users/pengtangjin/.cargo/registry/src/github.com-1ecc6299db9ec823/lazy_static-1.3.0
CARGO_WORKSPACE_MANIFEST_DIR=/Users/pengtangjin/.cargo/registry/src/github.com-1ecc6299db9ec823/lazy_static-1.3.0
I thought it was ok, any case dangerous cases I missed?
@matklad
tried to build a crate with this patch, output this snippet:
CARGO_MANIFEST_DIR=xxx/.cargo/registry/src/github.com-1ecc6299db9ec823/lazy_static-1.3.0
CARGO_WORKSPACE_MANIFEST_DIR=xxx/.cargo/registry/src/github.com-1ecc6299db9ec823/lazy_static-1.3.0
I thought it was ok, any case dangerous cases I missed?
Thanks for the PR @csmoe and sorry for the delay in review!
This sort of feature has a bit of history at this point, so I think it's worth paging that in as well. In addition to https://github.com/rust-lang/cargo/issues/7036 there's a lot more discussion on https://github.com/rust-lang/cargo/issues/3946 along with a closed PR and rationale for why the PR was closed.
Could this perhaps address some of the feedback in those threads to indicate why we would merge this instead?
Thanks for reviewing :), the problems wasn't addressed.
So, closed.
|
gharchive/pull-request
| 2019-07-03T12:16:31 |
2025-04-01T04:35:45.859338
|
{
"authors": [
"alexcrichton",
"csmoe",
"matklad"
],
"repo": "rust-lang/cargo",
"url": "https://github.com/rust-lang/cargo/pull/7089",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
616717623
|
Expand error message to explain that a string was found
With opt-level = "3" this previously said:
must be an integer, `z`, or `s`, but found: 3 for ...
The error message doesn't make that super clear.
This should now be a bit more clear.
Fixes #8234
We could even include a bit more saying that "3" should become 3 (either unconditionally or after trying to parse "3" into an integer?
cc @steveklabnik
r? @alexcrichton
(rust_highfive has picked a reviewer for you, use r? to override)
may be we should write quotes also \"3\"
I would agree yeah it probably makes the most sense to also include the surrounding quotes, but otherwise looks great to me!
Adjusted and also fixed the tests.
Now they fail in apt install gcc-multilib
@bors: r+
:pushpin: Commit 09084a365f612192fa026754b448f5b0be231890 has been approved by alexcrichton
:hourglass: Testing commit 09084a365f612192fa026754b448f5b0be231890 with merge db702d3d09de4c1c1284e159090f49b793a7fe4d...
:broken_heart: Test failed - checks-azure
@bors retry
#7858 😦
:hourglass: Testing commit 09084a365f612192fa026754b448f5b0be231890 with merge 55869de80ab8b7bd20cbb9a909e303fc03cf0125...
:sunny: Test successful - checks-azure
Approved by: alexcrichton
Pushing 55869de80ab8b7bd20cbb9a909e303fc03cf0125 to master...
|
gharchive/pull-request
| 2020-05-12T14:48:07 |
2025-04-01T04:35:45.865968
|
{
"authors": [
"alexcrichton",
"badboy",
"bors",
"ehuss",
"rust-highfive",
"shirshak55"
],
"repo": "rust-lang/cargo",
"url": "https://github.com/rust-lang/cargo/pull/8235",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
800744714
|
Backport #9102 to the 1.49 release branch
#9129 but to the branch that matters, this time with the version bump.
Does this look correct?
r? @alexcrichton
(rust-highfive has picked a reviewer for you, use r? to override)
:warning: Warning :warning:
Pull requests are usually filed against the master branch for this repo, but this one is against rust-1.49.0. Please double check that you specified the right target!
Thanks! I don't think it's that worth it to get CI working at this point, so I'm going to hit the merge button on this one. I'll do the publish afterwards as well.
|
gharchive/pull-request
| 2021-02-03T22:15:23 |
2025-04-01T04:35:45.868762
|
{
"authors": [
"Eh2406",
"alexcrichton",
"rust-highfive"
],
"repo": "rust-lang/cargo",
"url": "https://github.com/rust-lang/cargo/pull/9130",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1208540335
|
Add DragonFly umtx_{sleep, wakeup}.
This adds DragonFly's futex functions: https://man.dragonflybsd.org/?command=umtx§ion=2
@bors r+
:pushpin: Commit 9579c1dc5878abe52e7aeef229d69284cdc1543c has been approved by Amanieu
:hourglass: Testing commit 9579c1dc5878abe52e7aeef229d69284cdc1543c with merge 01b49024c26378a78e48358645a119595518e707...
:sunny: Test successful - checks-actions, checks-cirrus-freebsd-11, checks-cirrus-freebsd-12, checks-cirrus-freebsd-13
Approved by: Amanieu
Pushing 01b49024c26378a78e48358645a119595518e707 to master...
|
gharchive/pull-request
| 2022-04-19T15:41:02 |
2025-04-01T04:35:45.880092
|
{
"authors": [
"Amanieu",
"bors",
"m-ou-se"
],
"repo": "rust-lang/libc",
"url": "https://github.com/rust-lang/libc/pull/2763",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1145453582
|
Stacked borrows fails on {ChunksMut,ChunksExactMut}::__iterator_get_unchecked()
Description
fn main() {
let mut arr1 = [0u8; 64];
let arr2 = [0u8; 64];
let mut iter = arr1.chunks_mut(8).zip(arr2.chunks(8));
while let Some((chunk1, chunk2)) = iter.next() {
dbg!(chunk2[0]);
dbg!(chunk1[0]);
iter.next();
dbg!(chunk2[0]);
dbg!(chunk1[0]); // error here
}
}
Running this code in Miri will produce the following output:
[src\main.rs:6] chunk2[0] = 0
[src\main.rs:7] chunk1[0] = 0
[src\main.rs:9] chunk2[0] = 0
error: Undefined Behavior: no item granting read access to tag <1832> at alloc906 found in borrow stack.
--> src\main.rs:10:9
|
10 | dbg!(chunk1[0]); // error here
| ^^^^^^^^^^^^^^^ no item granting read access to tag <1832> at alloc906 found in borrow stack.
|
= help: this indicates a potential bug in the program: it performed an invalid operation, but the rules it violated are still experimental
= help: see https://github.com/rust-lang/unsafe-code-guidelines/blob/master/wip/stacked-borrows.md for further information
With -Zmiri-track-pointer-tag=1832,1585,12319
note: tracking was triggered
--> \library\core\src\slice\iter.rs:1564:19
|
1564 | Self { v: slice, chunk_size: size }
| ^^^^^ created tag 1585
|
= note: inside `std::slice::ChunksMut::<u8>::new` at \library\core\src\slice\iter.rs:1564:19
= note: inside `core::slice::<impl [u8]>::chunks_mut` at \library\core\src\slice\mod.rs:828:9
note: inside `main` at src\main.rs:4:20
--> src\main.rs:4:20
|
4 | let mut iter = arr1.chunks_mut(8).zip(arr2.chunks(8));
| ^^^^^^^^^^^^^^^^^^
note: tracking was triggered
--> src\main.rs:5:21
|
5 | while let Some((chunk1, chunk2)) = iter.next() {
| ^^^^^^ created tag 1832
|
= note: inside `main` at src\main.rs:5:21
[src\main.rs:6] chunk2[0] = 0
[src\main.rs:7] chunk1[0] = 0
note: tracking was triggered
--> \library\core\src\slice\iter.rs:1641:32
|
1641 | let len = cmp::min(self.v.len().unchecked_sub(start), self.chunk_size);
| ^^^^^^^^^^^^ popped tracked tag for item [Unique for <1832>] due to Read access for <1585>
|
= note: inside `<std::slice::ChunksMut<u8> as std::iter::Iterator>::__iterator_get_unchecked` at \library\core\src\slice\iter.rs:1641:32
= note: inside `<std::iter::Zip<std::slice::ChunksMut<u8>, std::slice::Chunks<u8>> as std::iter::adapters::zip::ZipImpl<std::slice::ChunksMut<u8>, std::slice::Chunks<u8>>>::next` at \library\core\src\iter\adapters\zip.rs:278:23
= note: inside `<std::iter::Zip<std::slice::ChunksMut<u8>, std::slice::Chunks<u8>> as std::iter::Iterator>::next` at \library\core\src\iter\adapters\zip.rs:84:9
--> \library\core\src\slice\iter.rs:1642:32
|
1642 | from_raw_parts_mut(self.v.as_mut_ptr().add(start), len)
| ^^^^^^^^^^^^^^^^^^^ created tag 12319
|
= note: inside `<std::slice::ChunksMut<u8> as std::iter::Iterator>::__iterator_get_unchecked` at \library\core\src\slice\iter.rs:1642:32
--> \library\core\src\slice\mod.rs:483:5
|
483 | / pub const fn as_mut_ptr(&mut self) -> *mut T {
484 | | self as *mut [T] as *mut T
485 | | }
| |_____^ popped tracked tag for item [Disabled for <1832>] due to Write access for <12319>
|
= note: inside `core::slice::<impl [u8]>::as_mut_ptr` at \library\core\src\slice\mod.rs:483:5
= note: inside `<std::slice::ChunksMut<u8> as std::iter::Iterator>::__iterator_get_unchecked` at \library\core\src\slice\iter.rs:1642:32
note: inside `main` at src\main.rs:8:9
--> src\main.rs:8:9
|
8 | iter.next();
| ^^^^^^^^^^^
= note
[src\main.rs:9] chunk2[0] = 0
error: Undefined Behavior: no item granting read access to tag <1832> at alloc906 found in borrow stack.
--> src\main.rs:10:9
|
10 | dbg!(chunk1[0]); // error here
| ^^^^^^^^^^^^^^^ no item granting read access to tag <1832> at alloc906 found in borrow stack.
|
= help: this indicates a potential bug in the program: it performed an invalid operation, but the rules it violated are still experimental
= help: see https://github.com/rust-lang/unsafe-code-guidelines/blob/master/wip/stacked-borrows.md for further information
A few things to note here:
Access to the mutable chunk will fail. For example arr1.chunks(8).zip(arr2.chunks_mut(8)) will fail on chunk2[0] instead.
zip() is required. This is probably due to Zip uses __iterator_get_unchecked() internally rather than next().
chunks_mut() and chunks_exact_mut() can both reproduce.
array_chunks_mut() can't reproduce.
Environment
$ rustc --version --verbose
rustc 1.61.0-nightly (45e2c2881 2022-02-20)
binary: rustc
commit-hash: 45e2c2881d11324d610815bfff097e25c412199e
commit-date: 2022-02-20
host: x86_64-pc-windows-msvc
release: 1.61.0-nightly
LLVM version: 14.0.0
$ cargo miri --version
miri 0.1.0 (0db4090 2022-02-12)
There's probably more of these kinds of problems that @saethlin has been looking into recently.
Huge thanks for reporting this @andylizi!
This is definitely an issue with the implementation in core. Perhaps this issue should be moved to rust-lang/rust?
The existing definition of ChunksMut:
https://github.com/rust-lang/rust/blob/1103d2e914b67c18b0deb86073c26c6aefda761d/library/core/src/slice/iter.rs#L1556-L1559
pub struct ChunksMut<'a, T: 'a> {
v: &'a mut [T],
chunk_size: usize,
}
Cannot support the current implementation of __iterator_get_unchecked:
https://github.com/rust-lang/rust/blob/1103d2e914b67c18b0deb86073c26c6aefda761d/library/core/src/slice/iter.rs#L1632-L1644
unsafe fn __iterator_get_unchecked(&mut self, idx: usize) -> Self::Item {
let start = idx * self.chunk_size;
// SAFETY: see comments for `Chunks::__iterator_get_unchecked`.
//
// Also note that the caller also guarantees that we're never called
// with the same index again, and that no other methods that will
// access this subslice are called, so it is valid for the returned
// slice to be mutable.
unsafe {
let len = cmp::min(self.v.len().unchecked_sub(start), self.chunk_size);
from_raw_parts_mut(self.v.as_mut_ptr().add(start), len)
}
}
The problem is that this method is trying to hand out subslices of a &mut [T] while also holding on to the original. Because it wraps a &mut [T], ChunksMut guarantees that its inner slice always shrinks when it returns a slice. It looks like this guarantee was accidentally introduced and accidentally upheld in implementation by avoiding unsafe... until __iterator_get_unchecked was added.
I'm working on a patch that fixes the aliasing problem.
It's perhaps worth noting that the specialization on zip that this exists to support has already earned 4 I-unsound issues, and this aliasing issue isn't due to new code. We/I should have caught this already, but didn't. That's quite annoying.
|
gharchive/issue
| 2022-02-21T08:09:10 |
2025-04-01T04:35:45.898482
|
{
"authors": [
"RalfJung",
"andylizi",
"saethlin"
],
"repo": "rust-lang/miri",
"url": "https://github.com/rust-lang/miri/issues/1980",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1251870966
|
rustup
Cc https://github.com/rust-lang/rust/issues/97486
@bors r+
:pushpin: Commit 7cd5fc3de327b9db96918cc895676f2d94c0a44d has been approved by RalfJung
:hourglass: Testing commit 7cd5fc3de327b9db96918cc895676f2d94c0a44d with merge 5c3e4b6556431c816d05b262a890905a2a50841a...
:sunny: Test successful - checks-actions
Approved by: RalfJung
Pushing 5c3e4b6556431c816d05b262a890905a2a50841a to master...
|
gharchive/pull-request
| 2022-05-29T12:07:01 |
2025-04-01T04:35:45.901608
|
{
"authors": [
"RalfJung",
"bors"
],
"repo": "rust-lang/miri",
"url": "https://github.com/rust-lang/miri/pull/2162",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1428625078
|
Implement condvars for Windows
Adds 3 shims for Windows: SleepConditionVariableSRW, WakeConditionVariable, WakeAllConditionVariable to add support for condvars (which fixes #2628).
Salvaged from what was removed from #2231
I think this panic is caused by a mistake in #2601.
CI failure should be fixed by #2641.
I think this panic is caused by a mistake in #2601.
That PR landed, so please rebase this one. :)
The rebase is done. CI's finally passing :)
Great. :) Please squash the commits a little, then we can land this.
Squashed.
@bors r+
:pushpin: Commit 958ca31d26951c1c1d8be875d704c72efa8aa8b5 has been approved by RalfJung
It is now in the queue for this repository.
:hourglass: Testing commit 958ca31d26951c1c1d8be875d704c72efa8aa8b5 with merge 32c90ffab5663ec5495220bcb85b199327227878...
:sunny: Test successful - checks-actions
Approved by: RalfJung
Pushing 32c90ffab5663ec5495220bcb85b199327227878 to master...
|
gharchive/pull-request
| 2022-10-30T06:05:13 |
2025-04-01T04:35:45.906399
|
{
"authors": [
"DrMeepster",
"RalfJung",
"bors"
],
"repo": "rust-lang/miri",
"url": "https://github.com/rust-lang/miri/pull/2638",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2302862726
|
automata: Fix broken universal start states with sparse DFA
The state IDs were not remapped, which will usually result in an index out of range error.
Is it possible to add a regression test?
Sure, let me give it a try tomorrow (it's late here today).
Hi, I added a test based on examples in Automaton::is_special_state. I'm asserting that the initial state is .is_start_state(), but this seems to be broken for some reason. Could you help figuring out why? I'm afraid I'm not familiar enough with how the start state ranges are being assigned.
Alternatively, I could remove the assert for now and file an issue for it to be fixed later.
It looks like start states were simply being removed with set_no_special_start_states, so this was the intended behavior. I have removed the assert now. Let me know if the tests look good to you.
Hi, mind giving this a review?
Friendly ping, let me know if there's something else I need to address
|
gharchive/pull-request
| 2024-05-17T14:11:29 |
2025-04-01T04:35:45.911445
|
{
"authors": [
"BurntSushi",
"ishitatsuyuki"
],
"repo": "rust-lang/regex",
"url": "https://github.com/rust-lang/regex/pull/1195",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
367545890
|
String::from(“…”) => d”…”
Dynamic Strings are common but using String::From("...") or "...".to_string() can get tedious, so it'll be more productive to write a string litteral d"..." instead. It's more productive, removes the apparent distance between both types of strings and it's syntactically sweeter.
Note that this does not constitute a formal RFC proposal; those are pull requests that follow the template.
Please read the README.md file to see how the process works.
I'm closing this issue in favor of https://internals.rust-lang.org/t/string-from-d/8528.
|
gharchive/issue
| 2018-10-07T11:35:36 |
2025-04-01T04:35:45.913940
|
{
"authors": [
"AmjadHD",
"Centril"
],
"repo": "rust-lang/rfcs",
"url": "https://github.com/rust-lang/rfcs/issues/2559",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
139182339
|
How long should rust-mode keep supporting Emacs 23?
Is anyone using rust-mode regularly with emacs 23? (The last version of emacs 23 was released in January 2012...three and a half years before Rust 1.0 was released, for what it's worth.)
I'm starting to see signs of slight hiccups for people trying to enhance rust-mode. Nothing hard to overcome yet, but it could get more annoying later.
When is it time to cut it off? When emacs 25 comes out? When someone writes a feature that really needs something only available in emacs 24 and above? Sooner than that?
This affects issue #139 as well.
Personally I think we should just drop emacs23 support now.
|
gharchive/issue
| 2016-03-08T05:00:46 |
2025-04-01T04:35:45.944573
|
{
"authors": [
"MicahChalmer",
"nikomatsakis",
"tromey"
],
"repo": "rust-lang/rust-mode",
"url": "https://github.com/rust-lang/rust-mode/issues/137",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
53918525
|
Fails on non-privileged user account.
The previous installer which only installed rust would ask for you to enter a admin password. The new installer which installs both rust and cargo fail instead requiring you to run it as an admin.
The new installer is derived from the old one and I don't believe it ever requested sudo. That said, it might be a good idea for the installer to ask for root if it determines that it can't install. I filed this against rust-installer https://github.com/rust-lang/rust-installer/issues/7
@brson I should I have also said that I run windows 7.
|
gharchive/issue
| 2015-01-09T21:43:30 |
2025-04-01T04:35:45.946385
|
{
"authors": [
"Jexell",
"brson"
],
"repo": "rust-lang/rust-packaging",
"url": "https://github.com/rust-lang/rust-packaging/issues/5",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
51929514
|
rust-lang.org is not rendering correctly (blank)
Most of rust-lang.org is not rendering correctly for me.
Neither with Firefox 34.0.5 nor with Chromium 39.0.2171.95 on Archlinux.
Firefox's console does not show any error but Chrome's one tell me :
Resource interpreted as Font but transferred with MIME type binary/octet-stream: "http://doc.rust-lang.org/Heuristica-Italic.woff"
The MIME type in Firefox is "application/x-font-woff" so I'm not sure it is the issue here.
Just wanted to put here: if you experience the problem, the current official Arch packages for Fira Sans should work.
https://www.archlinux.org/packages/community/any/ttf-fira-sans/
|
gharchive/issue
| 2014-12-14T19:53:25 |
2025-04-01T04:35:45.949824
|
{
"authors": [
"ldng",
"skade"
],
"repo": "rust-lang/rust-www",
"url": "https://github.com/rust-lang/rust-www/issues/75",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
467970524
|
Modules with path still can't be formatted on Windows
#1754 issue is still occuring.
I tried cargo fmt on my tokio checkout and I get several errors like following:
error: couldn't read \\?\D:\tokio\tokio-sync\tests\..\src\mpsc\list.rs: The filename, directory name, or volume label syntax is incorrect. (os error 123)
--> \\?\D:\tokio\tokio-sync\tests\fuzz_list.rs:14:5
|
14 | mod list;
| ^^^^
code for above is:
#[path = "../src/mpsc/list.rs"]
#[allow(warnings)]
mod list;
Can you try the changes in https://github.com/rust-lang/rustfmt/pull/3590 and see if that fixes your problem ? Looks like it doesn't for some users...
It is not building:
--> /home/95th/.cargo/registry/src/github.com-1ecc6299db9ec823/rustc-ap-rustc_target-407.0.0/abi/mod.rs:112:62
|
112 | ["a", ref a..] => dl.aggregate_align = align(a, "a")?,
| ^ expected slice, found &str
|
= note: expected type `&[&str]`
found type `&&str`
error[E0308]: mismatched types
--> /home/95th/.cargo/registry/src/github.com-1ecc6299db9ec823/rustc-ap-rustc_target-407.0.0/abi/mod.rs:113:58
|
113 | ["f32", ref a..] => dl.f32_align = align(a, "f32")?,
| ^ expected slice, found &str
|
= note: expected type `&[&str]`
found type `&&str`
error[E0308]: mismatched types
--> /home/95th/.cargo/registry/src/github.com-1ecc6299db9ec823/rustc-ap-rustc_target-407.0.0/abi/mod.rs:114:58
|
114 | ["f64", ref a..] => dl.f64_align = align(a, "f64")?,
| ^ expected slice, found &str
|
= note: expected type `&[&str]`
found type `&&str`
error[E0308]: mismatched types
--> /home/95th/.cargo/registry/src/github.com-1ecc6299db9ec823/rustc-ap-rustc_target-407.0.0/abi/mod.rs:117:46
|
117 | dl.pointer_align = align(a, p)?;
| ^ expected slice, found &str
|
= note: expected type `&[&str]`
found type `&&str`
error[E0308]: mismatched types
--> /home/95th/.cargo/registry/src/github.com-1ecc6299db9ec823/rustc-ap-rustc_target-407.0.0/abi/mod.rs:127:35
|
127 | let a = align(a, s)?;
| ^ expected slice, found &str
|
= note: expected type `&[&str]`
found type `&&str`
error[E0308]: mismatched types
--> /home/95th/.cargo/registry/src/github.com-1ecc6299db9ec823/rustc-ap-rustc_target-407.0.0/abi/mod.rs:145:35
|
145 | let a = align(a, s)?;
| ^ expected slice, found &str
|
= note: expected type `&[&str]`
found type `&&str`
error: aborting due to 13 previous errors
Hi @scampi, is there any update on this issue please?
I just hit this issue today in my repository. The repository and commit that is failing to rustfmt is:
https://github.com/rojo-rbx/rbx-dom/tree/3448e528766b1558ac76addf5e636df797f180ed
@95th Sorry for the delay, I had little time available for this. The branch is now compiling. Thanks for giving it a try!
This will be fixed in the 2.0 release.
|
gharchive/issue
| 2019-07-15T07:04:07 |
2025-04-01T04:35:46.595006
|
{
"authors": [
"95th",
"LPGhatguy",
"scampi",
"topecongiro"
],
"repo": "rust-lang/rustfmt",
"url": "https://github.com/rust-lang/rustfmt/issues/3688",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
876139041
|
Mask types are missing select implementation
I want to be able to:
fn f(cond: mask32x4, true_v: mask32x4, false_v: mask32x4) -> mask32x4 {
cond.select(true_v, false_v)
}
I was just thinking about this. It's slightly non-trivial because the masks aren't necessarily SIMD types. But no reason it can't be done.
|
gharchive/issue
| 2021-05-05T07:15:37 |
2025-04-01T04:35:46.605386
|
{
"authors": [
"calebzulawski",
"programmerjake"
],
"repo": "rust-lang/stdsimd",
"url": "https://github.com/rust-lang/stdsimd/issues/114",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
599389604
|
Add an environment variable to keep the temp dir
Fixes #66
Published as v0.5.29.
|
gharchive/pull-request
| 2020-04-14T08:27:05 |
2025-04-01T04:35:46.653562
|
{
"authors": [
"phil-opp"
],
"repo": "rust-osdev/cargo-xbuild",
"url": "https://github.com/rust-osdev/cargo-xbuild/pull/67",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2177494796
|
Skip URL check for https://blog.rust.careers
Certificate has expired.
I checked the site to see if I could find an email address to notify that their cert is expired, but didn't find anything from a quick look.
I checked the site to see if I could find an email address to notify that their cert is expired, but didn't find anything from a quick look.
They mention an email on their main site at https://rust.careers/about. I sent them an email a few days ago, let's hope they fix it soon.
Note that this is already the second time that the certificate has expired for that site in the past few months. So I'm fine with keeping the exception even after the certificate is renewed to avoid additional work for us in the future.
|
gharchive/pull-request
| 2024-03-10T02:49:45 |
2025-04-01T04:35:46.656002
|
{
"authors": [
"nicholasbishop",
"phil-opp"
],
"repo": "rust-osdev/homepage",
"url": "https://github.com/rust-osdev/homepage/pull/206",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1393936735
|
Question over stdout mutability
Hello, I'm attempting to use this crate to implement a basic UEFI bootloader but I've encountered an issue that I'm unclear on how to navigate around. The following code
let vendor = system_table.firmware_vendor();
system_table.stdout().output_string(vendor)?;
attempts to retrieve a reference to the firmware vendor from the system table, and then print it to the standard output. However, this code fails to compile due to system_table being borrowed as a mutable reference when calling stdout().
error[E0502]: cannot borrow `system_table` as mutable because it is also borrowed as immutable
--> src\main.rs:26:5
|
25 | let vendor = system_table.firmware_vendor();
| ------------------------------ immutable borrow occurs here
26 | system_table.stdout().output_string(vendor)?;
| ^^^^^^^^^^^^^^^^^^^^^ ------ immutable borrow later used here
| |
| mutable borrow occurs here
Although I've found I can use other means to print the firmware vendor to the standard output, I'd like to avoid having to copy the string data if possible, as it's already present in memory and I'd like to be able to just pass it directly to output_string without overhead.
Is there an intended code pattern I should be using here in order to achieve this, or is there no way with the package API to avoid doing a copy? From what I could tell from peeking at the code, stdout is borrowing the system table reference as mutable, and although it's only touching self.table.stdout, this conceptually invalidates the result of firmware_vendor despite only touching self.table.fw_vendor, but I'm unsure if there's a way for the API to represent this or not.
It's possible to avoid making a copy of the string, but it requires using unsafe.
One way to do it is to use the uefi-services crate, which provides print/println macros (as well as a log implementation). It keeps an internal pointer to the system table, so you can do something like println!("vendor: {}", st.firmware_vendor()) without any borrowing issues. That hides the unsafe from your code, but it is still present.
Another alternative, closer to the example you gave, would be to keep a pointer to the vendor string:
let vendor: *const _ = st.firmware_vendor();
// Safety: we trust the firmware to not invalidate the vendor pointer while calling `output_string`.
unsafe { st.stdout().output_string(&*vendor).unwrap(); }
|
gharchive/issue
| 2022-10-02T23:39:48 |
2025-04-01T04:35:46.660161
|
{
"authors": [
"nicholasbishop",
"ori-sky"
],
"repo": "rust-osdev/uefi-rs",
"url": "https://github.com/rust-osdev/uefi-rs/issues/522",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2201110138
|
Fix some new lints/warnings
Checklist
[ ] Sensible git history (for example, squash "typo" or "fix" commits). See the Rewriting History guide for help.
[ ] Update the changelog (if necessary)
The merge train pipeline fails:
https://github.com/rust-osdev/uefi-rs/actions/runs/8417467100/job/23045999267
@nicholasbishop - unfortunately, I do not have time to look into this.
Looks like the nightly compiler now warns if you import something by name that was already imported by glob. So for example:
use uefi::prelude::*;
use uefi::table::boot::BootServices; // warning: the item `BootServices` is already imported
In a couple places this was also happening due to the prelude glob, e.g. an import of Iterator now warns since it's already imported in the prelude.
Pushed another commit to fix all those cases.
|
gharchive/pull-request
| 2024-03-21T20:23:43 |
2025-04-01T04:35:46.663415
|
{
"authors": [
"nicholasbishop",
"phip1611"
],
"repo": "rust-osdev/uefi-rs",
"url": "https://github.com/rust-osdev/uefi-rs/pull/1103",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
660712875
|
Disable serial tests in CI
CI on master recently failed by crashing QEMU while running the serial protocol tests. It didn't fail on the corresponding PR however, making me think this is bug only occurs sometimes.
For now, I want to check if disabling these tests keeps CI happy.
Rerunning the tests seems to have fixed the issue. I'll investigate if there's any way to prevent them from failing (I'm guessing they timeout during periods of intense activity on GH Actions)
|
gharchive/pull-request
| 2020-07-19T09:12:17 |
2025-04-01T04:35:46.664937
|
{
"authors": [
"GabrielMajeri"
],
"repo": "rust-osdev/uefi-rs",
"url": "https://github.com/rust-osdev/uefi-rs/pull/146",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
833141657
|
Unmaintained dependency - dkjson
Details
The dkjson package required to load the tags to JSONB (instead of HSTORE) is unmaintained (https://github.com/LuaDist/dkjson). Reported on #81, tracking here to avoid losing sight of this.
The dkjson package was suggested by the osm2pgsql project examples (see https://github.com/openstreetmap/osm2pgsql/blob/master/flex-config/places.lua#L5-L7), not sure what other alternatives/options may exist at this time.
Planning to use https://github.com/rxi/json.lua
The rxi project has >1k stars and 245 forks and had relatively recent activity. Their docs mentioned being faster than other options (including dkjson) so did a quick test.
I ran tags.lua to load Colorado 3x w/ dkjson and and 3x w/ rki-json. 4CPU, 8 GB RAM, SSD. Basic command:
osm2pgsql --slim --drop \
--output=flex --style=./style/tags.lua \
--cache 3000 -d $PGOSM_CONN ~/pgosm-data/colorado-latest.osm.pbf
rki-json.lua loaded in an average of 119 seconds vs 136s for dkjson, so ~13% faster with this option.
I looked at a few other options but none were as popular, recently updated, or as simple/straightfoward. I plan to include the json.lua script in the flex-config/ directory with a link back to the project for the ultimate source. Will have to occasionally check on the parent project for updates. The json script can then be included directly (like helpers.lua) without additional prereq to install.
|
gharchive/issue
| 2021-03-16T19:46:47 |
2025-04-01T04:35:46.751663
|
{
"authors": [
"rustprooflabs"
],
"repo": "rustprooflabs/pgosm-flex",
"url": "https://github.com/rustprooflabs/pgosm-flex/issues/88",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1083814307
|
No dense data is shown on dashboard
If I select 3 hours on dashboard, I see data points only every 15min (should be 1min):
It'll now fetch mixed data if history range <= 24h, else sparse
|
gharchive/issue
| 2021-12-18T11:41:06 |
2025-04-01T04:35:46.764342
|
{
"authors": [
"io53",
"laurijamsa"
],
"repo": "ruuvi/com.ruuvi.station.webui",
"url": "https://github.com/ruuvi/com.ruuvi.station.webui/issues/97",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.