• BananaTrifleViolin@lemmy.world
    link
    fedilink
    English
    arrow-up
    220
    arrow-down
    1
    ·
    edit-2
    5 months ago

    This is badly written and ignorant article. Fat32 supports up to 16Tb partition size (depending on cluster size - 2Tb -16Tb).

    Its microsoft’s windows tools that arbitrarily only allow users to create 32Gb partitions, and it is this that is being changed. This is not a change to Fat32, this is a change to windows. 3rd party tools on Windows and other systems like Linux have long offered more options for partition size.

    That its taken to 2024 for Microsoft to fix the command line tool (and still not fix the GUI tools) is ridiculous.

  • Peffse@lemmy.world
    link
    fedilink
    English
    arrow-up
    83
    ·
    5 months ago

    I love how the arstechnica article words it like you will never need FAT32 and it’s silly to consider it.

    I had to download fat32format I don’t know how many times because I needed to format an extra large SD Card or USB drive for some device. Microsoft really shafted exFAT’s adoption with their licensing.

  • /home/pineapplelover@lemm.ee
    link
    fedilink
    English
    arrow-up
    6
    ·
    4 months ago

    I think there was some kind of tool that let you extend it more. I had a 512gb drive on fat32 but it sucked so much I just reformated to ext4 and it performed much better

    • EddoWagt@feddit.nl
      link
      fedilink
      English
      arrow-up
      4
      ·
      4 months ago

      Yeah, GUIFormat can do that. Fat32 has its limitations, but I pretty much always use it as the stuff I use micro SD cards in, require it

    • rdri@lemmy.world
      link
      fedilink
      English
      arrow-up
      8
      arrow-down
      8
      ·
      5 months ago

      Linux still unable to catch up with NTFS when it comes to filename length, sadly. 256 bytes in an era of Unicode is ridiculous.

      • jabjoe@feddit.uk
        link
        fedilink
        English
        arrow-up
        19
        ·
        5 months ago

        NTFS also has a 255 limit, but it’s UTF16, so for unicode, you will get more out of it. High price to pay for UTF16. Windows basically is moving stuff between UTF16 and ASCII all the time. Most apps are ASCII but Windows is natively UTF16. All other modernly maintained OS do UTF8, which “won” unicode.

        The fact that all major Unix (not just Linux) filesystems are to 255 bytes says it’s not a feature in demand.

        I’d much rather have COW subvolume snapshotting and incremental backup of btrfs or zfs. Plus all the other things Linux has over Windows of course.

        • rdri@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          arrow-down
          5
          ·
          5 months ago

          NTFS also has a 255 limit, but it’s UTF16, so for unicode, you will get more out of it.

          I think this is a biased way of putting it. NTFS way is easy to understand and therefore manage. What’s more important is that ASCII basically means English only. I’ve seen enough of such “discrimination” (stuff breaks etc.) based on used language in software/technology and it should end for good.

          All other modernly maintained OS do UTF8, which “won” unicode.

          UTF8 is Unicode. UTF8 symbols can take more than 1 byte.

          Plus all the other things Linux has over Windows of course.

          There are also encryption methods that slash maximum length of each filename even further.

          • jabjoe@feddit.uk
            link
            fedilink
            English
            arrow-up
            5
            ·
            edit-2
            4 months ago

            Of course UTF8 is Unicode. The cool thing about UTF8 is that is ASCII, until it isn’t. It cover all of Unicode, but doesn’t need any bloat if you are just doing latin characters. Plus UTF8 will seamless go through ASCII code and things that understand it do, others just have patches of jibberish, but still work otherwise. It’s a way better approach. Better legacy handling and more efficient packing for latin languages. Which is why it “won” out. UTF16 pretty much only exists in Windows because it’s legacy it will be hard for it to escape.

            LUKS is by far the most common encryption setup on Linux. It’s done at block layer and the filesystem doesn’t know about it. No effect of filename length, or anything else.

            • rdri@lemmy.world
              link
              fedilink
              English
              arrow-up
              2
              arrow-down
              4
              ·
              4 months ago

              None of that helps or discards anything I’ve said above. But it allows to say that NTFS limit can be basically 1024 bytes. Just because you like what UTF-8 offers it doesn’t solve hurdles with Linux limits.

              LUKS is commonly used but not the only one.

              • jabjoe@feddit.uk
                link
                fedilink
                English
                arrow-up
                2
                ·
                edit-2
                4 months ago

                Linus’s VFS is where the 256 limit is hard. Some Linux filesystem, like RaiserFS, go way beyond it. If it was a big deal, it would be patched and widely spread. The magic of Linux, is you can try it yourself, run your own fork and submit patches.

                LUKS is the one to talk about as the others aren’t as good an approach in general. LUKS is the recommended approach.

                Edit: oh and NTFS is 512 bytes. UTF16 = 16bit = 2 bytes. 256*2 = 512

                • rdri@lemmy.world
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  arrow-down
                  1
                  ·
                  4 months ago

                  The magic of Linux, is you can try it yourself, run your own fork and submit patches.

                  Well it should probably go further and offer more of another kind of magic - where stuff works as user expects it to work.

                  As for submitting patches, it sounds like you suggest people play around and touch core parts responsible for file system operations. Such an advice is not going to work for everyone. Open source software is not ideal. It can be ideal in theory, but that’s it.

                  LUKS is the one to talk about as the others aren’t as good an approach in general. LUKS is the recommended approach.

                  It looks like there are enough use cases where some people would not prefer LUKS.

        • rdri@lemmy.world
          link
          fedilink
          English
          arrow-up
          5
          arrow-down
          2
          ·
          5 months ago

          Linux file system is shit? Otherwise I don’t get why you’ve used the “because” word. NTFS is certainly not shit.

          • Vilian@lemmy.ca
            link
            fedilink
            English
            arrow-up
            6
            ·
            edit-2
            5 months ago

            I re-read your comment and i completely misunderstood it sorry it’s 4am

  • OutrageousUmpire@lemmy.world
    link
    fedilink
    English
    arrow-up
    3
    arrow-down
    1
    ·
    5 months ago

    I don’t know how much it matters though? If I try it on my Windows XP machine I’ll still be stuck with the old limit right?

  • MrSoup@lemmy.zip
    link
    fedilink
    English
    arrow-up
    5
    arrow-down
    39
    ·
    edit-2
    5 months ago

    If someone still use win-dos, 4GB per file and 32GB partition cap is what they deserve.