I’m a robotics researcher. My interests include cybersecurity, repeatable & reproducible research, as well as open source robotics and rust programing.
I fell for it. It took me a minute into the game time to figure what was up and double check today’s date.
I’m using a recent 42" LG OLED TV as a large affordable PC monitor in order to support 4K@120Hz+HDR@10bit, which is great for gaming or content creation that can appreciate the screen real estate. Anything in the proper PC Monitor market similarly sized or even slightly smaller costs way more per screen area and feature parity.
Unfortunately such TVs rarely include anything other than HDMI for digital video input, regardless of the growing trend connecting gaming PCs in the living room, like with fiber optic HDMI cables. I actually went with a GPU with more than one HDMI output so I could display to both TVs in the house simultaneously.
Also, having an API as well as a remote to control my monitor is kind of nice. Enough folks are using LG TVs as monitors for this midsize range that there even open source projects to entirely mimic conventional display behaviors:
I also kind of like using the TV as simple KVMs with less cables. For example with audio, I can independently control volume and mux output to either speakers or multiple Bluetooth devices from the TV, without having fiddle around with repairing Bluetooth peripherals to each PC or gaming console. That’s particularly nice when swapping from playing games on the PC to watching movies on a Chromecast with a friend over two pairs of headphones, while still keeping the house quite for the family. That kind of KVM functionality and connectivity is still kind of a premium feature for modest priced PC monitors. Of course others find their own use cases for hacking the TV remote APIs:
You could get a fiber optic display/HDMI cable, a fiber optic USB cable, and the USB hub, then just move the desktop tower into another room and run the cables through the walls or ceilings to your display setup. Might only be $100 or so cheaper than then a used business thin client, but at least you could still do something 4K 120Hz HDR 12bit over some distance without compromise. E.g:
The first link you listed (viewing the remote community from our local instance) shows 0 subscribers
from the sidebar. From my understanding, no one from our instance is then subscribed to that remote community, so our instance has no reason to index those posts. Although I could be wrong, and it could be that no one from our local instance is subscribed to any community on the remote instance. I’m unsure if only instance federation or community subscription is necessary for merrioring/indexing remote posts.
Another issue is whether the post from a remote instance is colocated on your local instance. It could be the case that your instance never observed the post, as no users on your local instance where first subscribed, to the remote community the post was summited to, before the post was published.
I think there are some user scripts folks are developing via browser extensions, but perhaps it’d be better to mainline the kind of features these ticket are tracking:
Image Transcription: Meme
A photo of an opened semi-trailer unloading a cargo van, with the cargo van rear door open revealing an even smaller blue smart car inside, with each vehicle captioned as “macOS”, “Linux VM” and “Docker” respectively in decreasing font size. Onlookers in the foreground of the photo gawk as a worker opens each vehicle door, revealing a scene like that of russian dolls.
*I’m a human volunteer content transcriber and you could be too! *
That looks neet. Although I suspect this would succumb to the same cross post discoverability issues where URLs pointing to the same video would not match string for string. A better approach might be to facilitate inline embedding of HTML video players into Lemmy using browser extensions, where user scripts could be used to preview youtube links or re-write them to nocookie, allowing the Lemmy web UI to still avoid the use of cross-origin scripts by default.
Found the full transcription for the video from OP author:
Note to self: use
youtube.com
instead ofyoutu.be
for better cross post detection and lemmy integration
For programming tutorials, yep, I also prefer reading documentation instead. Although, it looks like this tutorial these folks put out doesn’t have much of anything you could copy from, like terminal commands, given its a recorded walkthrough in using the graphical web UI. YouTube also now allows for searching the auto or manual transcription text, which is handy when creators always forget to include timestamped chapters.
I suspect this comment was posted to spell out the meme for those unfamiliar, but I wanted to thank you for transcribing it into text for those that also may be blind or visually impaired. With the loss of r/TranscribersOfReddit , I salute your contribution! Please keep at it!
https://www.theverge.com/2023/6/23/23771396/reddit-subreddit-community-transcribers-accessibility
Checking the issues tracker for RES, there’s not yet any mention of lemmy or kbin:
Perhaps you could ask there. I’d also recommend checking out the Lemmy Plugins & Userscripts community:
For posterity, I later stumbled upon the authors original post here:
The community that this was posted from also looks interesting:
Check out this issue:
Do you have any links or references to learn more? Perhaps any github tickets to follow? By bootstrap, are you referring to this?
You can change the color theme from the setting page under the top right drop-down. But it would be nice to have something like Reddit Extension Suite for the default Lemmy UI front end for custom defined CSS.
I think once we get a few more third party clients to explore alternative UIs, folks should have more options for personal preference.
One thing I like about the current web UI already is the low noise in embedded text in the discussion threads. E.g. when I engage my screen reader, all I have to listen to when moving between comments is the post author and post date. Just enough context to understand the TTS engine moving between comments, unlike the old.reddit.com UI that include 5 or 6 different hyperlinked words (parent, context, permalink, etc) that the TTS has two repeat over and over again.
The hover text for icon links should be enough UI context for screen readers, although not all icon links on the current Lemmy UI seem to include hover text meta data, like the permalink chain icon 🔗, while the collapse minimize icon does.
Tagging an image is simply associating a string value to an image pushed to a container registry, as a human readable identifier. Unlike an image ID or image digest sha, an image tag is only loosely associated, and can be remapped later to another image in the same registry repo, e.g
latest
. Untagging is simply removing the tag from the registry, but not necessarily the associated image itself.