I’ve never seen a USB-A to A cable in the wild, except recently, where I finally unpacked my SATA/IDE USB adapter from Ugreen.
Goodbye Reddit, Hello Lemmy
I’ve never seen a USB-A to A cable in the wild, except recently, where I finally unpacked my SATA/IDE USB adapter from Ugreen.
Not Nextcloud. livesync is better. Requires CouchDB (Docker available) but with that, its a powerful sync option. Even Settings can be synced
Docker is amazing but not needed. You can compare it to a simpler VM. You can take a docker and run it on any machine. You have an environment that is separate from your host and you and the container can only access it via defined points (volumes and ports).
Imagine you need to run a 2nd Mumble Server. I never set on up but its often that a 2nd instance is not that easy. With docker its easy. The only difference is that you need to use different ports, when you have only one network access or you use a reverse proxy. You can create a 2nd instance to test stuff, without interrupting your productive system. Its a security benefit, because its isolated to some degree and you can remove one easily.
I started using it with MSSQL Server, because I hated how invasive it is on a windows machine, especially I just needed it temporarily to do stuff with it. I’m not a microsoft admin and I know that Servers from Microsoft are a different level. Docker allowed me to start and stop it and remove it very easily. After that I started using it for a lot of and brought my NAS on the next level.
Also one thing worth mentioning are Linux Containerx (LXC). They are in Proxmox but I have less knowledge. It feels more like a full VM than docker but uses less resources. This is the reason why containers in general are more popular. They are less resource hungry than a full VM but have some benefits than running everything on one machine. LXC feels more like a full system, than docker. With docker you rarely get into the system. You may execute some commands, like a create user command or a one time job but don’t access it via a shell from the inside (its possible). LXC on the other hand, you use the shell.
Then use a dummy public repo. You don’t get the nice integration. We used Mantis Bugtracker at work for quite some time but switched to Gitea (a selfhosted Github clone) and are trying to get our testers to use it more.
Also you should keep one thing in mind: needing an additional account, might stop people to use it. With Github I contribute for even smaller ideas. When I need to make an account first, I wont do it. Or when they want to use Discord. I do have Discord but come on…
Making an installation script that will work for all, or even most, OSs and processor architectures can be a lot of work. Are you paying the devs to do it?
I do pay for my software, even when its free, when I like the software and the devs. But if the devs/community think they are something better, then no. I had some where they refused a install script and said something along the lines that if you can’t get it running with the docs, you shouldn’t host the software. Yeah I don’t like such devs. Also when they have enough time to write a documentation, they have enough time to write a script. I even had one project, where the dev refused a correction in the docs, even though it was faulty.
Also you don’t need to write a script for every system. You start with the most used ones. I mean just for testing I would get insane, when I have to repeat some steps over and over. At our company I do write scripts for some things that drive me insane. First I got told, it works this way, this costs too much time to do and it doesn’t sell more software. I just did it and now they thank me for that. Even just an internal tool that I wrote for myself, after I drove insane doing stuff manually, now also customers get.
I don’t speak about software where you pull via git, install some prerequisites and run a script. Not shipping prerequisites can have a legal reason and git pull is just a different way of downloading. It also works to download the tagged source code, instead of cloning but this requires more explanation to less skilled users.
Also with docker I came across some projects where they really butchered it. A docker compose file is my preferred way. I have my file + .env and it works for most containers I come across. It looks clean and feels clean. Running one command for creating a user is a understandable step, to avoid default users.
When you like to hammer in a lot of commands into a shell to install something, do it. I prefer my clean, simple and straight forward install scripts. I don’t need a installation doc that goes over every customization that you can make. I want a setup that works for most users and after that, I can dig around every customization there is, to optimize a software how I want it. Not everything is needed right at start and a default value that most users will use, is enough for the start.
Good thing that there is a ton of software and I can pick the ones where I like their philosophy and support them. That is what I do.
Do you kiss your mother with that mouth? I’m not speaking of this project but in general. Other projects where they does this, not this one.
I hate when the installation takes like 20 steps. Never heard of an installation script and a interactive installation? I’ve installed far more complex software, that gets this but your shitty programm can’t do it?
And they wonder why nearly nobody uses Linux. In Windows nearly all software comes as an executable. Imagine offering a software under windows, where you need to do the setup manually in a shell.
So far so good. The URL is correct, because its the external address. You also don’t need to publish both http and https ports. I only map external https to internal http but you can do https to https. No serious modern browser tries http first and because I always force https anyways, it doesn’t need to be public. Only the reverse proxy may need it, for Let’s Encrypt.
Both UDP aren’t needed for public access. I only have mapped 8096 to my reverse proxy and it works.
Audible + OpenAudible. OpenAudible does “stuff” and you end up with audio files, that you can listen on most devices. Don’t know and care how they do this. Its not free but so is Audible.
When you have an active Audible subscription, you also have access to free Audiobooks. You can download and convert them too. But be aware, that Audible is rate limited. Had downloaded a ton of free audiobooks and after a short limit (maybe 1 hour), I got a long limit for around 24 hours. But I still use Audible. I just have it as a backup and this way I can give my family access to the books I have. But so far my mother only listen to the ones I got for free. I like Science Fiction a lot but my mother not.
A college who I recommended Audiobookshelf, has a subscription from a German only site (Thalia), where apparently the Audiobooks can be downloaded as MP3s. So far I prefer Audible, even with DRM, just because the availability. Not all books I listen to, are available on that site or much later.
A service brought you by a human (usually I see bots do this): !i2p@lemmy.world
Its so you can stay on your instance and your viewer.