Just your normal everyday casual software dev. Nothing to see here.

People can share differing opinions without immediately being on the reverse side. Avoid looking at things as black and white. You can like both waffles and pancakes, just like you can hate both waffles and pancakes.

been trying to lower my social presence on services as of late, may go inactive randomly as a result.

  • 0 Posts
  • 143 Comments
Joined 3 years ago
cake
Cake day: August 15th, 2023

help-circle
  • I’m not PC but, one benefit of using a central server for syncthing is an always on backup that doesn’t require another client device to be on, it also allows for easier creation of new shares.

    For example, with syncthing you can set the “servers” client device to auto approve/accept any shares that are to trusted devices, then when you get a new device, instead of needing to add that device to every device you share on the syncthing network, you only need to add that device to the server and then you can have your other clients connect to the servers share instead of device to device. It’s easier. You can also configure the shares on the server to use encryption by default too, since you don’t really ever need to actually see the files on the server since it’s basically a install and forget style client.

    As an example of what I mean:

    I have 10 different devices that run syncthing, 9 clients and a “server” client. these clients are not always on at the same time, and as such when I change a file, the files can become desynced and cause issues with conflicts. By having a centralized server, as long as the server is on(it always is) and client itself is online, it’s going to always sync. I don’t need to worry about file conflicts between my clients as the server should always have the newest file.

    Then for example say my phone died. Instead of needing to readd every seperate client that the phone needs to share with to the new device, I only need to add the phone as a trusted source on the “server” client via the webui -> click share to that device on every share the phone needs, and then remap the shares to the proper directories on the mobile device. this is vs having to add every device to the phone, and the phone to every device it needs access to ontop of reconfiguring all the shares. It’s simpler, but fair warning does cause a single point of failure if the server goes offline.







  • I run Debian 13 and I will use apt repositories whenever possible and I avoid flatpak with a passion.

    My flowchart is:

    1. App Image if the application supports auto updating or if it’s a temperary app
    2. Apt repository
    3. Deb file
    4. App Image (when previous didn’t apply)
    5. Source
    6. Flatpak

    My main reasoning for it is strictly ease of use. I find flatpaks while I’m sure makes it easier on the developers making it to be super bloaty and take up more system resources, While causing more restrictions and annoyances during configuration due to their enhanced security setup.

    The only time I really don’t use a repo if it’s available is if the program itself updates on its own, or updates super frequently such as Discord, which I got annoyed enough at that I had to make my own update script that to check if there’s an update and then auto update it, because I got sick of the Discord has an update message every other day.





  • Basically many domain providers will hold onto domains for a little while after it expires.

    Some like namecheap also advertise the domain names to peddle-man companies that will somehow buy temporary access to the domain after your extortion recall window expires.

    To continue the namecheap example, when your namecheap domain expires, it gives you a lapse window where you can pay like double the cost of the domain renewal to reclaim it. If you don’t reclaim it during that window they give it to a middleman whom will somehow buy a 2 or 3 months domain lease for it. They will put it on a “site for sale” broker page and will charge you easily 100x what you paid for the domain if you wanted it back.

    I would recommend just keep checking on it every few days to see if it gets released.



  • sadly, it’s a little more complex than just enabling it. The supported self host deployment uses docker, and the docker containers that are available don’t contain the interfaces for voice or video calling as they are not up to date.

    If I understand it right, to enable it would mean you need to either pull the source yourself and run it off of docker, or make a custom docker image using a version of stoat web that contains the ability to do voice calls.

    reading the draft of the linked issue, it looks like the author isn’t doing voice call for the reason that they don’t know the proper way to integrate it into the docker image.

    So to answer it: yes it looks like you can use voice servers on the current self hosted model, but you can’t use pre-existing docker images, and it will require you to manually add the new web UI in and patch where needed.



  • Personally, it seems like it’s trustworthy again. The previous owner of the repo did eventually admit that they authorized the transfer, but, The entire transfer process was extremely sketchy and had no chain of custody or trust. It was just the repository got deleted, and then a few days later showed under a whole blank state again with a user with no profile, no contribution history, and it was just a trust me bro, I knew the original maintainer look I have the keys to prove it.

    The maintainer of the Google Play build of it seems to trust them though, and they are established in the community, plus they archived their sync thing builds again in favor of just using one repo, so it’s likely fine.

    For future people wondering about it as well, it doesn’t help that the new maintainer of the app has deleted every issue that had to do with the migration, so you no longer can research the issue for yourself. The only information you have available to you is the discussion chain listed on the community forums, But any type of issue that they link to were deleted.

    Personally though, I plan on keeping my current version pinned to prior to the transfer until either I’m forced to update due to bugs or I feel comfortable with the current maintainer again. I’m not sure how long that will be.

    For an app that contains very sensitive information, I was not impressed with how the transfer process underwent.



  • I don’t think thats a unfair ask. One local representative in each country seems perfectly fair for me.

    Being said? the user information part? strictly locked to their own content. If the user account is registered in that country they have access. Providers could 100% do that with most operational databases out there. It’s a requirement for stores in order to do payment information. Steam and Epic already do this as it is.

    Should they be able to access that information in the first place is a different discussion, that needs to be had in that corresponding country, but if the country has already decided it needs access to continue, there’s no reason it should have access to all user data. The only thing they really have claim to is their own countries data.


  • my issue with what would happen if this ruling solidifies is the precident that it causes.

    I could not care less about reaction videos, they are really low effort videos that I don’t understand why are so popular.

    My issue entirely is that if the plaintiff wins in this case, it’s effectively saying any type of downloaded video on youtube would classify as circumventing DRM, which would open an avenue aside from a fair use violation for studios to go after content creators for.

    Look at lets plays for example. Those operate almost entirely on fair use clauses. I fear that if we start ruling that recording or downloading videos that your computer is able to decode (as this is all that the youtube downloader is doing, just instead of it going to the client its sending to a file), that means by same principle, recording a video game that contains DRM would also be considered circumventing a DRM. Which would outlaw lets plays.

    This is a very bad precedent regardless of if its just low quality trash reaction videos or not.


  • They are very nice. They share kernelspace so I can understand wanting isolation but, the ability to just throw a base Debian container on, assign it a resource pool and resource allocation, and install a service directly to it, while having it isolated from everything without having to use Docker’s emphereal by design system(which does have its perks but I hate troubleshooting containers on it) or having to use a full VM is nice.

    And yes, by Docker file I would mean either the Docker file or the compose file(usually compose). By straight on the container I mean on the container, My CTs don’t run Docker period, aside from the one that has the primary Docker stack. So I don’t have that layer to worry about on most CT’s

    As for the memory thing, I was just mentioning that Docker does the same thing that containers do if you don’t have enough RAM for what’s been provisioned. The way I had taken that original post is that specifying 2 gigs of RAM to the point the system exhausts it’s ram would cause corruption and the system crashes, which is true but docker falls for the same issue if the system exhausts it’s ram. That’s all I meant by it. Also cgroups sound cool, I gotta say I haven’t messed with them a whole lot. I wish proxmox had a better resource share system to designate a specific group as having X amount of max resources, and then have the CT or vm’s be using those pools.