• @BlemboTheThird@lemmy.ca
      link
      fedilink
      235 months ago

      yeah i mean, theres no way lol. even if the tech gets here that quickly there’s 0% chance prices come down significantly on lower capacity drives. these’ll be at least $500 and possibly far far more

  • @pop@lemmy.ml
    link
    fedilink
    325 months ago

    20 TB at that price range could brankrupt some small cloud providers. Selfhosting would be much easier without having to worry about space. IF the price stays the same, but we’ll see.

    • @Blue_Morpho@lemmy.world
      link
      fedilink
      255 months ago

      I don’t think the vast majority use cloud storage because it’s cheap. The vast majority use it because they are unwilling or unable to setup their own.

      Hosting an Internet facing service out of your own house requires constant maintenance for security.

      • @RvTV95XBeo@sh.itjust.works
        link
        fedilink
        75 months ago

        I self host, and drop encrypted backups onto a cloud storage provider. If anything, cheap storage is going to cost me more because I’ll be inclined to back more up.

      • @qjkxbmwvz@startrek.website
        link
        fedilink
        45 months ago

        Right. And if you want to self host with some geographic redundancy, it requires having friends or family with a good Internet connection who are willing to let you have a server at their place. Not impossible, but can be annoying.

        I’m setting up a raspberry pi+HDD at family’s house, with wireguard to my home network. Fun stuff, but it’s not an off-the-shelf solution, especially when you consider that it’s not my Internet access, it’s theirs, so trying to be polite with bandwidth/data caps means it’s a bit kneecapped.

      • hswolf
        link
        fedilink
        35 months ago

        Yeah, there are a lot of security layers checks that big providers handle that most beginners users don’t even know exist.

        I can’t imagine the damage of someone breaching through your server and reaching your personal network.

        Also an extra machine, if you plan to have 99% uptime.

    • @4am@lemm.ee
      link
      fedilink
      155 months ago

      I’d be interested what the wear-leveling and write-cycles look like. $250 for 20TB is half the current price of decent spinning rust, but if they’ll die in a year because they’re part of a Ceph cluster or ZFS array, that’s gonna be a no from me, dawg.

  • @GenderNeutralBro
    link
    English
    285 months ago

    Do I need a 20TB boot drive? No. Do I want it enough to pay $250? Yes, absolutely. I’m running 1TB now and I need to manage my space far more often than I’d like, despite the fact that I keep my multimedia on external mass storage. Also, sometimes the performance of that external HD really is a hindrance. I’d love to just have (almost) everything on my primary volume and never worry about it.

    It’s kind of weird how I have less internal storage today than I did 15 years ago. I mean, it’s like 50 times faster, but still.

    I’m not super-skeptical about the pricing. This stuff can’t stay expensive forever, and 2027 is still a ways off.

    • @Cort@lemmy.world
      link
      fedilink
      75 months ago

      Honestly, if they could just get 8tb ssds down to $200-250 I’d be happy with sata interfaces

  • @Zachariah@lemmy.world
    link
    fedilink
    135 months ago

    I get a sense that the tipping point for the final HDD to SSD transition will be soon. A lot of people are willing to spend a little more for SSD when the capacity-to-price is almost as good as HDD. I think this will first spike demand followed by spiked production followed by a significant drop in price after production ramps up (as long as the companies avoid any economic funny business).

    • @Pechente@feddit.org
      link
      fedilink
      English
      85 months ago

      I build my NAS years ago and the drives are still doing fine. I can’t wait though to upgrade that thing to SSDs. This thing is loud and sucks a lot of power right now.

    • @WolfLink@sh.itjust.works
      link
      fedilink
      35 months ago

      Nah there are still use cases where longevity is most important. You can’t set an SSD in a closet for 20 years and expect to still have your data. HDDs also have longer active life expectancies AIUI.

  • @litchralee@sh.itjust.works
    link
    fedilink
    English
    115 months ago

    It would be amazing if PCIe lanes becomes the predominant limiting factor, rather than drive cost, for building large storage arrays. What a world it would be, when even Epyc and its lanes-for-days proves to be insufficient for large Chia miners err Plex servers uh, Linux ISO mirrors.

  • Estebiu
    link
    fedilink
    75 months ago

    Maybe I should put off building a NAS for the time being.

    • Ioughttamow
      link
      fedilink
      45 months ago

      Yeah…. I think I’ll still need to pick up some drives this year, but I might do less robust of a build out

      • @InputZero@lemmy.ml
        link
        fedilink
        25 months ago

        I’ve switched over to one disk redundancy to stretch out my spinning rust. If SSDs come down just a little bit more it’ll be worth it to replace my array. Just hoping two drives doesn’t die until then.

  • @RobotToaster@mander.xyz
    link
    fedilink
    -45 months ago

    Generally the more layers you add to an SSD the less robust it is. If this is real your data will be corrupt within a week.

    • veroxii
      link
      fedilink
      95 months ago

      I mean you can say the same for spinning magnetic platters. “The more bits you’re trying to squeeze into a fixed size HDD the less robust it is.”

      I’m not saying these guys can do it, but dismissing higher densities of storage out of hand seems a bit glib considering the last 60 years of progress and innovation.

        • pbjamm
          link
          fedilink
          English
          35 months ago

          Should work perfectly with Flatpack apps then!

      • @RobotToaster@mander.xyz
        link
        fedilink
        -15 months ago

        That’s never really been an issue with HDDs as far as I’m aware, although 10k rpm drives were known to be more fragile IIRC. The lower life and robustness of QLC vs SLC flash is well known.

        • @Blue_Morpho@lemmy.world
          link
          fedilink
          35 months ago

          Because drives use ECC and spare sectors to give the illusion of reliability just like QLC.

          Here is reliability vs drive size.

          • @sudo42@lemmy.world
            link
            fedilink
            English
            55 months ago

            Decades ago, a collegue of mine (who once worked in hard drive design) said, “Oh, hard drives stopped reading 1’s and 0’s years ago. Now they compute the probability that the data just read was a 1 or a 0.”

    • @ShortN0te@lemmy.ml
      link
      fedilink
      65 months ago

      That is exactly what ppl like you said when SLC came out and TLC came out and QLC came out…

      Look back now.

    • adr1an
      link
      fedilink
      -15 months ago

      And they use far more energy. Meanwhile spinning disks can sit idle with all of my hoarded data.

      • @Blue_Morpho@lemmy.world
        link
        fedilink
        35 months ago

        Google says hdd idle is higher than SSD. Ssd is higher under load but it’s important to look at total energy used. If the SSD spikes high, but is 10x faster, the total energy used will be less.

      • @Cobrachicken@lemmy.world
        link
        fedilink
        25 months ago

        The Seagate Exos in my aray consume 8-9W each on idle, and 12-13W active. I doubt that a NVME would consume that much in idle, which will be most of the time for data storage.

      • adr1an
        link
        fedilink
        15 months ago

        AFAIK power consumption increases with size on SSDs. And that’s not the case with spinning disks. That’s what I tried to point out, from the perspective of hoarding data (idle disks) bigger sizes are not something to be pursued. Then of course there’s the use case of needing a high volume fast storage (e.g. zfs cache), for which use case these are great!