I don’t fit in an of these teams, and neither do literally all Linux users I know. Should we have identity crises, or could this be a giant oversimplification?
I don’t fit in an of these teams, and neither do literally all Linux users I know. Should we have identity crises, or could this be a giant oversimplification?
Which compression level are you using? My old server is able to compress flac’s at the highest (and therefore “slowest”) compression level at >50x speed, so bumping the level up shouldn’t be too hard on your CPU.
I’ve been running some external drives on my server for about a year now. In my experience, hard drives with an external power supply suffer less from random disconnects. The specific PC also makes quite a large difference in reliability. My server is just a regular desktop and has very little problem staying connected and powering my 3 external drives. My seedbox is an old laptop, and has been having almost constant problems with random disconnects and power issues. Maybe test how well your framework does with some external drives before committing to the plan?
To change the ownership of the files, you should only have to run sudo chown -R user:group directory
. -R makes chown run recursively, so it will modify the directory and all subdirectories and files. Do note that changing the ownership to plex:plex or something similar would leave your user unable to normally modify the files. My solution to this was to add both my regular user and the plex (in my case jellyfin) user to the same group. That way both users can easily see and modify the files, as long as the group has read/write permissions (the 2nd column of rwx in ls -Al
). If necessary, you can add group permissions with sudo chmod -R g+rw directory
.
On a side note: have you considered using jellyfin? It’s a completely free alternative to plex, which recently received a truly massive update with tons of new features. Some people prefer plex’ overall experience, but I’ve been running jellyfin with almost no complaints.
Small disclaimer: I’m writing from mobile, so the commands might not be 100% correct. Run at your own risk, and NEVER POINT A CHMOD/CHOWN COMMAND AT SYSTEM DIRECTORIES LIKE / OR /USR. That’s one of the easiest ways to completely break your system.
Have you tried the official guide from the jellyfin website?
As for the guide this AI generated: it bothers me that they instruct you to use chocolatey for the *arrs, but still advice you to install docker, qbittorrent and jellyfin manually (all of which have chocolatey packages). I disagree with the comment that external storage would be recommended, as internal storage is generally more reliable (depending on a lot of factors of course). Also, I believe the “adding a library”-section of the jellyfin setup is a bit too short to be of any use, and would recommend referring to the jellyfin docs instead.
This guide also doesn’t explain how to make jellyfin accessible outside of your LAN. Once again, I’d recommend referring to the jellyfin docs if you want to do this.
I personally have only set up qbittorrent, jellyfin and docker (not the *arr suite), so I can’t comment on the completeness of the guide, but I wouldn’t trust it too much (seeing the previous oversights).
And finally, as someone who started their selfhosted server journey on windows: don’t. There is a reason why almost all guides are written for linux, as it is (in my humble opinion) vastly superior for server usage once you get used to it.
didn’t know that was a part of bisexuality
I should probably flee before I get eaten by an army of blahåjar (apparently that’s the correct plural?)
Oh I don’t mind the nitpicking, thanks for the explanation! I (apparently erroneously) thought “demake” and “decompile” were synonyms. Guess I’m one of today’s 10000.
In that case the (now taken down, but forked a gazillion times) portal64 project would be a correct example of a demake, right?
interested in females
Username checks out, though I’m assuming you meant “demakes”?
Anyways, the demake I’m most familiar with is the in-progress Lego island. The YouTuber behind it documented part of the process in vlogs (linked on the GitHub page), so that might be an interesting starting point.
I believe SSD’s don’t actually experience wear when reading data, only when writing. Loading more data from SSD’s shouldn’t cause any premature failure. Overwriting more data each update could cause the drive to fail slightly earlier, but if that’s really that big of a concern, you’d be best of moving to Debian stable (no updates means no SSD writes).
If SSD wear prevention is really that big of a concern, you might be interested in profile-sync-daemon (https://wiki.archlinux.org/title/Profile-sync-daemon). It reduces writes to hard drives by keeping your browser profile in RAM, and only periodically syncing it to disk.
Though I must add that SSD’s wearing out really isn’t that much of an issue with modern drives. With normal usage, a drive will become obsolete long before it actually wears out.
I tried using Linux alternatives to iTunes, but it was always a pain. Even iTunes itself on a separate windows box was more of a hassle than I wanted. I eventually discovered rockbox, which works great with my iPod (5th gen AKA video): it has way more config options and allows me to simply create .m3u playlists and use my own folder structure. If your iPod is supported (https://www.rockbox.org/wiki/IpodPort.html), I’d absolutely recommend Rockbox over other solutions.
If your iPod isn’t supported by Rockbox (like my nano 5th gen), you could probably use strawberry or GTKpod. Both are imperfect, but work “good enough”.
Could it be that the /usr/local/bin directory doesn’t exist? If that’s the case, you’d either have to create it or replace that part of the command with some other directory in your $PATH (make sure to change both occurrences in the command if you decide to go with this latter option). Though I must add that this kind of manual install isn’t great if you want to keep track of installed apps and pending updates, since you’d have to do all of that manually too.
Not OP (OC? Not the person you were helping, you get what I mean), are you sure you meant df -h
? fd -H
seems more useful for to me when trying to find a specific file in a dotfolder, though even that didn’t work on my system. fd
ignores ~/.config
by default, so you need to use fd -u
(which is an alias for fd -I -H
) to find the correct files.
Anyways, from your description it seems like the correct file would be ~/.config/kwinrc
, which exists on my system.
You could look at the awesome-selfhosted list, specifically these two sections:
https://awesome-selfhosted.net/tags/recipe-management.html
https://awesome-selfhosted.net/tags/task-management--to-do-lists.html
I don’t have any experience with any of those, but there might be something that fits your needs.
If the installer is small enough (<650MB I believe), you can upload it to virustotal.com to have it be scanned by ~65 antivirus programs
Ah, it looks like we have a small misunderstanding. I thought you were talking about uncompressed video, which is enormous. This is only used in HDMI cables for example. A 1080p60 uncompressed video is 2.98Gbit/s, or about 1.22 terabytes per hour.
A remux is “uncompressed” in the sense that it isn’t recompressed, or in this case transcoded. A remux is still compressed, just to a lesser degree than a transcode. This means the files are indeed larger, but the quality is also better than transcodes.
To clarify the article’s confusing statement: they claim that remuxes can reduce size by throwing away some audio streams, while keeping the original video. This is true, but the video itself hasn’t gotten any smaller: you are simply throwing away other information.
Remuxes aren’t uncompressed, nor are they losslessly compressed. They’re just a 1:1 direct copy from some other medium (generally blu-rays or DVD’s).
Disclaimer: I have exactly 0 personal experience with eGPU’s.
According to the archwiki:
While some manual configuration (shown below) is needed for most modes of operation, Linux support for eGPUs is generally good.
Yes, some minor formatting changes occur when opening a docx file in libreoffice. Hardly sounds like a deal breaker to me. And yes, you do get a pop-up when saving to docx in libreoffice (with the toggle to disable the pop-ups right there in the message). Microsoft office does the exact same thing when saving to an odt file though:
Once again, if you have to collaborate with office-users (and you cannot deal with the horror of having a different amount of space between the items), just use office online. How many times do I have to repeat myself?
Let me guess you’re someone who works in IT and never had a typical “office job” that includes spending 90% of your time writing reports and pushing spreadsheets around.
This is why you don’t get it, you’re not the typical user of MS Office and you don’t share the same use cases the OP, the article author and myself share.
If you want to use windows, that’s fine. But please don’t share such blatantly ignorant articles, and don’t try to defend them when multiple people point out why it is wrong about so many things.
I probably won’t reply to your next reaction (should there be any) unless you come up with some actual arguments, instead of “the line spacing is broken, you’re out of touch, not me”.
Yes, libreoffice doesn’t really work for live collaboration. But office online is a good solution for that collaboration, and it works in any browser (including Firefox on Linux). Therefore, the author’s conclusion (you need windows to collaborate on word docs) is still wrong.
I personally also believe that WYSIWYG editors are highly overrated: markdown is significantly better for note-taking and similar small documents, and reports would often be better off with LaTeX or something similar. But I understand why the “4 commands is too much hassle to install VirtualBox” crowd might prefer word.
That seems like a good edit, and fair enough. Good to know that there is also room for people who want to use their computer in a non-fanatical way, simply minding our own business.