Just your normal everyday casual software dev. Nothing to see here.

People can share differing opinions without immediately being on the reverse side. Avoid looking at things as black and white. You can like both waffles and pancakes, just like you can hate both waffles and pancakes.

been trying to lower my social presence on services as of late, may go inactive randomly as a result.

  • 0 Posts
  • 146 Comments
Joined 3 years ago
cake
Cake day: August 15th, 2023

help-circle

  • Pika@sh.itjust.workstoLinux@lemmy.mlRTFM
    link
    fedilink
    English
    arrow-up
    8
    arrow-down
    4
    ·
    edit-2
    9 days ago

    this is an absolutely toxic take of the issue. I took OP’s statement as less of a “I won’t read the manual” and more of a “I struggle to be able to read manuals”

    There are many times I had read the manual, and then had to look up the issue further anyway because I either missed the poorly written section, or misunderstood what it was saying.

    If you want a prime example of that, go look at ffmpeg and try to figure out how to select a specific language for subtitles on a video without looking it up online. its via -map as an advanced option, which is described as a parameter to extract specific streams (which also means they would need to map the video and the audio streams since including a -map removes every auto stream). but map doesn’t tell you subtitle tracks are index:s. it does tell you that you can look at stream specifiers for valid search options, which does include s as a type, and lets you know that you can use m for metadata tagging, but you would need to make the connection that the type is s, and the meta data search flag would be m:language:langcode, and you need to make the connection the entire string has to be concated so its index:s:m:language:langcode For someone who is learning ffmpeg and video transcoding, that is not a very good setup. The stream specifiers give a few examples of what the potentials are but, the location where it specifies the types are in a different area than the one where it specifies the metadata keys. At that point just asking online or searching is way easier.

    Note: this is just an issue I have see people come across because ffmpeg is one of the more complicated programs (the man page is over 2300 lines)

    is it in the manual? yes. is someone who doesn’t know how to use ffmpeg and is trying to learn it going to find it? that’s debatable.

    If I was in that situation, my next step would be googling it, and if I couldn’t find it via searching, I would be reaching out to communities. At that point “RTFM” is useless to me.



  • I’m not PC but, one benefit of using a central server for syncthing is an always on backup that doesn’t require another client device to be on, it also allows for easier creation of new shares.

    For example, with syncthing you can set the “servers” client device to auto approve/accept any shares that are to trusted devices, then when you get a new device, instead of needing to add that device to every device you share on the syncthing network, you only need to add that device to the server and then you can have your other clients connect to the servers share instead of device to device. It’s easier. You can also configure the shares on the server to use encryption by default too, since you don’t really ever need to actually see the files on the server since it’s basically a install and forget style client.

    As an example of what I mean:

    I have 10 different devices that run syncthing, 9 clients and a “server” client. these clients are not always on at the same time, and as such when I change a file, the files can become desynced and cause issues with conflicts. By having a centralized server, as long as the server is on(it always is) and client itself is online, it’s going to always sync. I don’t need to worry about file conflicts between my clients as the server should always have the newest file.

    Then for example say my phone died. Instead of needing to readd every seperate client that the phone needs to share with to the new device, I only need to add the phone as a trusted source on the “server” client via the webui -> click share to that device on every share the phone needs, and then remap the shares to the proper directories on the mobile device. this is vs having to add every device to the phone, and the phone to every device it needs access to ontop of reconfiguring all the shares. It’s simpler, but fair warning does cause a single point of failure if the server goes offline.







  • I run Debian 13 and I will use apt repositories whenever possible and I avoid flatpak with a passion.

    My flowchart is:

    1. App Image if the application supports auto updating or if it’s a temperary app
    2. Apt repository
    3. Deb file
    4. App Image (when previous didn’t apply)
    5. Source
    6. Flatpak

    My main reasoning for it is strictly ease of use. I find flatpaks while I’m sure makes it easier on the developers making it to be super bloaty and take up more system resources, While causing more restrictions and annoyances during configuration due to their enhanced security setup.

    The only time I really don’t use a repo if it’s available is if the program itself updates on its own, or updates super frequently such as Discord, which I got annoyed enough at that I had to make my own update script that to check if there’s an update and then auto update it, because I got sick of the Discord has an update message every other day.





  • Basically many domain providers will hold onto domains for a little while after it expires.

    Some like namecheap also advertise the domain names to peddle-man companies that will somehow buy temporary access to the domain after your extortion recall window expires.

    To continue the namecheap example, when your namecheap domain expires, it gives you a lapse window where you can pay like double the cost of the domain renewal to reclaim it. If you don’t reclaim it during that window they give it to a middleman whom will somehow buy a 2 or 3 months domain lease for it. They will put it on a “site for sale” broker page and will charge you easily 100x what you paid for the domain if you wanted it back.

    I would recommend just keep checking on it every few days to see if it gets released.



  • sadly, it’s a little more complex than just enabling it. The supported self host deployment uses docker, and the docker containers that are available don’t contain the interfaces for voice or video calling as they are not up to date.

    If I understand it right, to enable it would mean you need to either pull the source yourself and run it off of docker, or make a custom docker image using a version of stoat web that contains the ability to do voice calls.

    reading the draft of the linked issue, it looks like the author isn’t doing voice call for the reason that they don’t know the proper way to integrate it into the docker image.

    So to answer it: yes it looks like you can use voice servers on the current self hosted model, but you can’t use pre-existing docker images, and it will require you to manually add the new web UI in and patch where needed.



  • Personally, it seems like it’s trustworthy again. The previous owner of the repo did eventually admit that they authorized the transfer, but, The entire transfer process was extremely sketchy and had no chain of custody or trust. It was just the repository got deleted, and then a few days later showed under a whole blank state again with a user with no profile, no contribution history, and it was just a trust me bro, I knew the original maintainer look I have the keys to prove it.

    The maintainer of the Google Play build of it seems to trust them though, and they are established in the community, plus they archived their sync thing builds again in favor of just using one repo, so it’s likely fine.

    For future people wondering about it as well, it doesn’t help that the new maintainer of the app has deleted every issue that had to do with the migration, so you no longer can research the issue for yourself. The only information you have available to you is the discussion chain listed on the community forums, But any type of issue that they link to were deleted.

    Personally though, I plan on keeping my current version pinned to prior to the transfer until either I’m forced to update due to bugs or I feel comfortable with the current maintainer again. I’m not sure how long that will be.

    For an app that contains very sensitive information, I was not impressed with how the transfer process underwent.



  • I don’t think thats a unfair ask. One local representative in each country seems perfectly fair for me.

    Being said? the user information part? strictly locked to their own content. If the user account is registered in that country they have access. Providers could 100% do that with most operational databases out there. It’s a requirement for stores in order to do payment information. Steam and Epic already do this as it is.

    Should they be able to access that information in the first place is a different discussion, that needs to be had in that corresponding country, but if the country has already decided it needs access to continue, there’s no reason it should have access to all user data. The only thing they really have claim to is their own countries data.