deleted by creator
a beautiful robot, dancing alone · showgirls über alles: kylie, angèle · masto · last.fm · listenbrainz · lovekylie
deleted by creator
you probably already found this, but for others who might be curious:
in the settings if you change notification method from websocket to unified push, the UP settings come up, including a server address (which is what they intend to be used) or some air gap mode that i can’t find documented
so per wikipedia and confirmed at MDN, firefox is the only major browser line not to consider certificate transparency at all. and yet it’s the only one that has given me occasional maddening SSL errors that have blocked site access (not always little sites, it’s happened with amazon).
i don’t understand how firefox can be simultaneously the least picky about certificates and the most likely to spuriously decide they’re invalid.
well i feel stupid now for not doing the obvious. but…
Blocked Page
Your organization has blocked access to this page or website.
on the PPA box, this is what it showed me (meanwhile it was attempting to connect to incoming.telemetry.mozilla.org). another symptom of displaying respect for enterprise policies but in fact ignoring them. (as i had mentioned, on this box all of the settings look locked down as they should be, but it’s still attempting to send telemetry.)
thanks, i’ll look again. it’s not that i love the idea of being fingerprinted; i just think that five mylar bags, four tin hats and a partridge in a pear tree won’t save me from that. i need my password manager, and once that’s in, enforcing a generic screen is silly - cow’s out of the barn. but not having the arms race against pocket and telemetry would be a big bonus.
i did try that but the never-dark mode blinded me. i understand the reasoning, but absolute anonymity isn’t my own threat model; i’d like to be able to use themes and resize the window
It exists, it’s called a robots.txt file that the developers can put into place, and then bots like the webarchive crawler will ignore the content.
the internet archive doesn’t respect robots.txt:
Over time we have observed that the robots.txt files that are geared toward search engine crawlers do not necessarily serve our archival purposes.
the only way to stay out of the internet archive is to follow the process they created and hope they agree to remove you. or firewall them.
i made the same migration from markor (files in a folder) to logseq. there’s a lot to be gained - always-preview alone is a game changer - but on mobile the visibility of the keyboard can be fiddly. once in a while you’ll feel like you’re in vi, it has such a mind of its own. but i’m not planning to go back
looks great! the catch for me is that my current host doesn’t have docker support. your dependencies don’t look crazy so in theory i could burst it and install directly to the host environment, but at that point i’m giving myself grocy-level headaches.
reading about docker-capable hosts, i was surprised to see them starting at 1GB RAM - i couldn’t run pac-man in that. what would be a reasonable expectation for kitchenowl?
i haven’t tried the docker route - it seems fairly new. it also doesn’t seem like it would fix the issues i ran into. containerization is great for insulating the app from external dependency hell and environmental variation. but the problems i’ve had involve its own code and logic, and corruption of a sqlite database within its own filesystem; wrapping issues like that in a docker container only makes them harder to solve
appimages just got less easy…
i don’t know which update did it - i think it must have been os-level (i run pop_os, derived from ubuntu) - but appimages silently stopped working. double-click, nothing. finally i looked in the log out of desparation, which said ‘appimages require fuse’.
more accurately, appimages require fuse 2 and the os had just upgraded to fuse 3. the fix is to heat-seek libfuse2, and don’t mess with any other fuse-related package as things can start wrecking themselves:
sudo apt install libfuse2
originally seen on an omgubuntu post
against which kitchen pots have proven surprisingly useful elsewhere. against all odds
to a techie, i’d say, it’s open source and if they ever overpushed politics, they’d find they’d become the fork as the community would fork away.
to a non-techie, i’d say, everyone’s an asshole a different way, but they don’t own the whole place like spez and musk do.
and i wouldn’t argue. let them walk away haters. this platform isn’t ready for everyone to come right now anyway.
woah are those negative or slide? we have a place here that i used to hand a hundred or so negatives at a time, but that was in the before times. they’re still here but don’t really advertise scanning anymore, so not clear on what would be possible
Prerequisites
Installation
cargo install mollysocket
mollysocket
executable if desiredmollysocket
once so that it will emit the default configConfiguration
.config/mollysocket/default-config.toml
and copy it somewhere.config.toml
allowed_endpoints
line withallowed_endpoints = ['*']
. The default 0.0.0.0 config appears to be a bug; this setting controls access to endpoints within the app, not IPs from outside. Leaving the original value causes mollysocket to reject everything.db = './mollysocket.db'
line rather than just having it land wherever you’re sitting.mollysocket.db
that was created on first run (even if it’s already where you’re intending to put it). This is just to make sure the web server creates it and has the correct permissions.Run script
export ROCKET_PORT=8020 export RUST_LOG=info export MOLLY_CONF=/path/to/your/config.toml
Proxy server
/
to your mollysocket server and ROCKET_PORT.Things to know