• 0 Posts
  • 120 Comments
Joined 5 years ago
cake
Cake day: February 15th, 2021

help-circle
  • The thing is that vi and emacs have existed since long before those other new editors came around.

    What you want is possible to do by configuring your ~/.inputrc (see readline manual page for details), it’s just that the defaults are different because they are from a time when many keyboards didn’t even have arrow keys (and the ones that had them were in non-standard positions) so most of the shortcuts that became standard in those days are completely different than the ones common today. Given that the terminal is meant to emulate old style DEC VT100 terminals (that’s why it’s called terminal “emulator”) it made sense to use those default that people had grown used to.

    Personally, I’ve grown used to Ctrl+a, Ctrl+k, Ctrl+w, Ctrl+e and Ctrl+y …I dont have to reach to wherever the Home key is in whatever keyboard I happen to be using at the moment (specially with modern 75%/60%-sized keyboards today). Or use a combination that also requires shift and having to hold so many keys together. In fact I went the opposite direction and customized my Powershell profile while I’m on windows to keep many of those old shortcuts in the Windows pwsh terminal as well.


  • Ferk@lemmy.mltoLinux@lemmy.mlFSF announces Librephone project
    link
    fedilink
    arrow-up
    1
    arrow-down
    1
    ·
    edit-2
    13 days ago

    Good marketing means achieving an arbitrary limit of what you consider “good” marketing. So it depends on where you set the bar.

    The best marketing necessarily requires some level of unethical behavior, because being honest and saying the whole truth doesn’t sell. Everything has drawbacks and benefits… the better marketing minimizes (or even hides / fails to mention) the drawbacks and emphasizes the benefits, which is a form of deception.


  • I feel it’s a bit like the usability vs security dilemma… you can try to optimize to have both, but then you won’t have as a result neither the most secure system nor the smoothest user-friendly experience, but something in between (you might still consider that “secure” or “usable”, but that just depends on where you set your expectations).

    If you want to maximize marketing then the result won’t be as ethical as it could be, and if you want to maximize ethics then the result won’t be as marketable as it could be.


  • Ferk@lemmy.mltoLinux@lemmy.mlis i2p relevant today?
    link
    fedilink
    arrow-up
    6
    ·
    edit-2
    14 days ago

    I always saw I2P as a more modern and distributed onion-routing alternative to Tor.

    The thing is that people are used to making use of Tor in different ways than the way they use I2P, but you can also have outproxies (ie. exit nodes/relays) in I2P the same way as in Tor… and you can also host a service inside the Tor network without relying on an exit node, like in I2P. It’s just that people only seem to want to host exit nodes for Tor and not so much for I2P, this led to internal communications in I2P being more common (which is a good thing), whereas in Tor it’s common to use it for anonymous access to the clearnet (which strains the network and causes chokepoints, specially with big downloads or torrent sharing). That’s just a matter of usage, not capability.


  • I didn’t downvote you, but it’s unclear what you meant by stating that.

    Depending on how one interprets it, it can be seen as a justification for using “fascist” (since there isn’t a more accurate word) or simply a way to emphasize that the term is inaccurate and shouldn’t be used.

    So I’m not surprised if you get up/down votes from both sides in either direction, specially in a polarizing discussion. Not because of what you said being wrong/right, but because of what they might read between lines.






  • For full independence, why not simply detach development from community?

    You can even have multiple independent communities with multiple independent moderation teams all about the same software.

    As a developer I’ve never needed to engage a particular community on a personal level in order to make a PR to a project… if the technical maintainers want to accept the change, they will, if they won’t then that’s fine, they probably have their reasons. It’s ok to communicate with communities to get feedback, but I’m not making contributions for the social approval, I’m making them when I believe they are useful, and most of the times I write them because I want to have that change myself. If it’s rejected and enough other people are interested in the change, it can be forked. That doesn’t mean I hate the maintainers or that I don’t want the original to exist or anything, it’s not personal.

    But well, I understand that some communities wanna make software and they intertwine development and social relationships. However, if you do this then I don’t see how can independence be a thing. Either separate them and don’t intermix them or mix them and don’t expect them to be separate.




  • On the upside, the end user needs to use up less data for the same content. This is particularly interesting under 4G/5G and restrictive data plans, or when accessing places / servers with weak connection. It helps avoid having to wait for the “buffering” of the video content mid-playback.

    But yes, I agree each iteration has diminishing returns, with a higher bump in requirements. I feel that’s a pattern we see often.


  • It’s not like improper use of “steal” is unheard of, I see all the time people use “I’m gonna steal that” and similar even when it applies to things openly given for free. And considering that it’s quite clear that the MIT allows others to take without sharing back (it’s the main difference with GPL) I’m quite sure the commenter was aware that it wasn’t really theft, yet chose that word probably with the intention to insult the practice, rather than as a fair descriptor.

    So yes, you’re right, it isn’t theft… but I don’t think that was the point of the comment.



  • Compression and efficiency is often a trade-off. H266 is also much slower than AV1, under same conditions. Hopefully there will come more AV1 hw encoders to speed things up… but at least the AV1 decoders are already relatively common.

    Also, the gap between h265 and AV1 is higher than between AV1 and h266. So I’d argue it’s the other way around. AV1 is reported to be capable of ~30-50% bitrate savings over h.265 at the cost of speed. H266 differences with AV1 are minor, it’s reported to get a similar range, but more balanced towards the 50% side and at the cost of even lower speed. I’d say once AV1 encoding hardware is more common and the higher presets for AV1 become viable it’d be a good balance for most cases.

    The thing is that h26x has a consortium of corporations behind with connections and an interest to ensure they can cash in on their investment, so they get a lot of traction to get hw out.



  • It’s actually the lazy way. I only work once, then copy that work everywhere. The copying/syncing is surprisingly easy. If that’s what you call “package management” then I guess doing “package management” saves a lot of work.

    If I had to re-configure my devices to my liking every time I would waste time in repetition, not in an actual improvement. I configured it the way I liked it once already, so I want to be able to simply copy it over easily instead of re-writing it every time for different systems. It’s the same reason why I’ve been reusing my entire /home partition for ages in my desktop, I preserve all my setup even after testing out multiple distros.

    If someone does not customize their defaults much or does not mind re-configuring things all the time, I’m sure for them it would be ok to have different setup on each device… but I prefer working only once and copying it.

    And I didn’t say that bash is the only config I have. Coincidentally, my config does include a config.fish I wrote ages ago (14 years ago apparently). I just don’t use it because most devices don’t have fish so it cannot replace POSIX/Bash… as a result it naturally was left very barebones (probably outdated too) and it’s not as well crafted/featureful as the POSIX/bash one which gets used much more.


  • Manually downloading the same shell scripts on every machine is just doing what the package manager is supposed to do for you

    If you have a package manager available, and what you need is available there, sure. My Synology NAS, my Knulli, my cygwin installs in Windows, my Android device… they are not so easy to have custom shells in (does fish even have a Windows port?).

    I rarely have to manually copy, in many of those environments you can at least git clone, or use existing syncing mechanisms. In the ones that don’t even have that… well, at least copying the config works, I just scp it, not a big deal, it’s not like I have to do that so often… I could even script it to make it automatic if it ever became a problem.

    Also, note that I do not just use things like z straight away… my custom configuration automatically calls z as a fallback when I mistype a directory with cd (or when I intentionally use cd while in a far/wrong location just so I can reach faster/easier)… I have a lot of things customized, the package install would only be the first step.


  • It’s not only clusters… I have my shell configuration even in my Android phone, where I often connect to by ssh. And also in my Kobo, and in my small portable console running Knulli.

    In my case, my shell configuration is structured in some folders where I can add config specific to each location while still sharing the same base.

    Maybe not everything is general, but the things that are general and useful become ingrained in a way that it becomes annoying when you don’t have them. Like specific shortcuts for backwards history search, or even some readline movement shortcuts that apparently are not standard everywhere… or jumping to most ‘frecent’ directory based on a pattern like z does.

    If you don’t mind that those scripts not always work and you have the time to maintain 2 separate sets of configuration and initialization scripts, and aliases, etc. then it’s fine.