* [gentoo-user] Seagate hard drives with dual actuators.
@ 2024-11-13 23:10 Dale
2024-11-14 0:46 ` Matt Jolly
` (2 more replies)
0 siblings, 3 replies; 38+ messages in thread
From: Dale @ 2024-11-13 23:10 UTC (permalink / raw
To: Gentoo User
Howdy,
One of my PVs is about 83% full. Time to add more space, soon anyway.
I try not to go past 90%. Anyway, I was looking at hard drives and
noticed something new. I think I saw one a while back but didn't look
into it at the time. I'm looking at 18TB drives, right now. Some new
Seagate drives have dual actuators. Basically, they have two sets of
heads. In theory, if circumstances are right, it could read data twice
as fast. Of course, most of the time that won't be the case but it can
happen often enough to make it get data a little faster. Even a 25% or
30% increase gives Seagate something to brag about. Another sales tool.
Some heavy data users wouldn't mind either.
My question is this. Given they cost about $20 more, from what I've
found anyway, is it worth it? Is there a downside to this new set of
heads being added? I'm thinking a higher failure rate, more risk to
data or something like that. I think this is a fairly new thing, last
couple years or so maybe. We all know how some new things don't work out.
Just looking for thoughts and opinions, facts if someone has some.
Failure rate compared to single actuator drives if there is such data.
My searched didn't help me find anything useful.
Thanks.
Dale
:-) :-)
P. S. My greens are growing like weeds. Usually they ready to pick by
now but having to wait for the tree to be cut down and cut up delayed
that. Should be ready by Christmas, I hope. Oh, planted oats, clover,
kale and some other extra seeds I had in open area. I saw a LARGE buck
deer the other night snacking on the oats. My neighbor would rather see
it in his freezer tho. o_0
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [gentoo-user] Seagate hard drives with dual actuators.
2024-11-13 23:10 [gentoo-user] Seagate hard drives with dual actuators Dale
@ 2024-11-14 0:46 ` Matt Jolly
2024-11-14 13:05 ` Dale
2024-11-14 7:55 ` Wols Lists
2024-11-14 11:21 ` Michael
2 siblings, 1 reply; 38+ messages in thread
From: Matt Jolly @ 2024-11-14 0:46 UTC (permalink / raw
To: gentoo-user
Hi Dale,
> My question is this. Given they cost about $20 more, from what I've
> found anyway, is it worth it? Is there a downside to this new set of
> heads being added? I'm thinking a higher failure rate, more risk to
> data or something like that. I think this is a fairly new thing, last
> couple years or so maybe. We all know how some new things don't work out.
At least one vendor has been trying to sell me on these recently,
claiming higher bandwidth and better ability to seek. I have not yet
actually used these.
> Just looking for thoughts and opinions, facts if someone has some.
> Failure rate compared to single actuator drives if there is such data.
> My searched didn't help me find anything useful.
There's no reason to suspect that these are going to be terrible, they
appear to be the next step on the Seagate roadmap before HAMR drives
hit the market in the coming years.
I haven't seen much on the reliability side of things, however I
wouldn't be too concerned, assuming that there's proper backup in place
- Any other drive (including SSDs) could "just die" for a multitude of
reasons.
> P. S. My greens are growing like weeds. Usually they ready to pick by
> now but having to wait for the tree to be cut down and cut up delayed
> that. Should be ready by Christmas, I hope. Oh, planted oats, clover,
> kale and some other extra seeds I had in open area. I saw a LARGE buck
> deer the other night snacking on the oats. My neighbor would rather see
> it in his freezer tho. o_0
Hopefully there's some left for you!
Cheers,
Matt
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [gentoo-user] Seagate hard drives with dual actuators.
2024-11-13 23:10 [gentoo-user] Seagate hard drives with dual actuators Dale
2024-11-14 0:46 ` Matt Jolly
@ 2024-11-14 7:55 ` Wols Lists
2024-11-14 16:48 ` Dale
2024-11-14 11:21 ` Michael
2 siblings, 1 reply; 38+ messages in thread
From: Wols Lists @ 2024-11-14 7:55 UTC (permalink / raw
To: gentoo-user
On 13/11/2024 23:10, Dale wrote:
> My question is this. Given they cost about $20 more, from what I've
> found anyway, is it worth it? Is there a downside to this new set of
> heads being added? I'm thinking a higher failure rate, more risk to
> data or something like that. I think this is a fairly new thing, last
> couple years or so maybe. We all know how some new things don't work out.
I think this technology has actually been out for a long time. I'm sure
I've heard of it ages ago.
Thing is, it's probably one of those things that's been available in
high-end drives for years, but the cost-benefit ratio has been low so
few people bought them. Now presumably the economics have changed.
If the actuators are mounted opposite each other, then they can't
collide, and presumably can operate completely independent of each
other. The costs of two of them were presumably just deemed not worth it.
An opposite niche (and rather apposite for you) was when I started
buying disk drives. I think my first was a 2GB Bigfoot, followed by a
6GB, and I bought several 18GBs for work. They were "old tech", 5.25"
5200rpm in an era of 3.5" 7500rpm, but their capacities were huge and
cheap. If all you wanted was storage, they were great. Most people
thought the size and speed of the smaller drives was better value, even
if it cost more per meg.
Cheers,
Wol
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [gentoo-user] Seagate hard drives with dual actuators.
2024-11-13 23:10 [gentoo-user] Seagate hard drives with dual actuators Dale
2024-11-14 0:46 ` Matt Jolly
2024-11-14 7:55 ` Wols Lists
@ 2024-11-14 11:21 ` Michael
2024-11-14 17:00 ` Dale
2 siblings, 1 reply; 38+ messages in thread
From: Michael @ 2024-11-14 11:21 UTC (permalink / raw
To: gentoo-user
[-- Attachment #1: Type: text/plain, Size: 4086 bytes --]
On Wednesday 13 November 2024 23:10:10 GMT Dale wrote:
> Howdy,
>
> One of my PVs is about 83% full. Time to add more space, soon anyway.
> I try not to go past 90%. Anyway, I was looking at hard drives and
> noticed something new. I think I saw one a while back but didn't look
> into it at the time. I'm looking at 18TB drives, right now. Some new
> Seagate drives have dual actuators. Basically, they have two sets of
> heads. In theory, if circumstances are right, it could read data twice
> as fast. Of course, most of the time that won't be the case but it can
> happen often enough to make it get data a little faster. Even a 25% or
> 30% increase gives Seagate something to brag about. Another sales tool.
> Some heavy data users wouldn't mind either.
>
> My question is this. Given they cost about $20 more, from what I've
> found anyway, is it worth it? Is there a downside to this new set of
> heads being added? I'm thinking a higher failure rate, more risk to
> data or something like that. I think this is a fairly new thing, last
> couple years or so maybe. We all know how some new things don't work out.
>
> Just looking for thoughts and opinions, facts if someone has some.
> Failure rate compared to single actuator drives if there is such data.
> My searched didn't help me find anything useful.
>
> Thanks.
>
> Dale
>
> :-) :-)
I don't know much about these drives beyond what the OEM claims. From what I
read, I can surmise the following hypotheses:
These drives draw more power from your PSU and although they are filled with
helium to mitigate against higher power/heat, they will require better cooling
at the margin than a conventional drive.
Your system will use dev-libs/libaio to read the whole disk as a single SATA
drive (a SAS port will read it as two separate LUNs). The first 50% of LBAs
will be accessed by the first head and the last 50% by the other head. So
far, so good.
Theoretically, I suspect this creates a higher probability of failure. In the
hypothetical scenario of a large sequential write where both heads are writing
data of a single file, then both heads must succeed in their write operation.
The cumulative probability of success of head A + head B is calculated as
P(A⋂B). As an example, if say the probability of a successful write of each
head is 80%, the cumulative probability of both heads succeeding is only 64%:
0.8 * 0.8 = 0.64
As long as I didn't make any glaring errors, this simplistic thought
experiment assumes all else being equal with a conventional single head drive,
but it never is. The reliability of a conventional non-helium filled drive
may be lower to start with. Seagate claim their Exos 2 reliability is
comparable to other enterprise-grade hard drives, but I don't have any real
world experience to share here. I expect by the time enough reliability
statistics are available, the OEMs would have moved on to different drive
technologies.
When considering buying this drive you could look at the market segment needs
and use cases Seagate/WD could have tried to address by developing and
marketing this technology. These drives are for cloud storage
implementations, where higher IOPS, data density and speed of read/write is
desired, while everything is RAID'ed and backed up. The trade off is power
usage and heat.
Personally, I tend to buy n-1 versions of storage solutions, for the following
reasons:
1. Price per GB is cheaper.
2. Any bad news and rumours about novel failing technologies or unsuitable
implementations (e.g. unmarked SMRs being used in NAS) tend to spread far and
wide over time.
3. High volume sellers start offering discounts for older models.
However, I don't have a need to store the amount of data you do. Most of my
drives stay empty. Here's a 4TB spinning disk with 3 OS and 9 partitions:
~ # gdisk -l /dev/sda | grep TiB
Disk /dev/sda: 7814037168 sectors, 3.6 TiB
Total free space is 6986885052 sectors (3.3 TiB)
HTH
[-- Attachment #2: This is a digitally signed message part. --]
[-- Type: application/pgp-signature, Size: 833 bytes --]
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [gentoo-user] Seagate hard drives with dual actuators.
2024-11-14 0:46 ` Matt Jolly
@ 2024-11-14 13:05 ` Dale
0 siblings, 0 replies; 38+ messages in thread
From: Dale @ 2024-11-14 13:05 UTC (permalink / raw
To: gentoo-user
Matt Jolly wrote:
> Hi Dale,
>
>> My question is this. Given they cost about $20 more, from what I've
>> found anyway, is it worth it? Is there a downside to this new set of
>> heads being added? I'm thinking a higher failure rate, more risk to
>> data or something like that. I think this is a fairly new thing, last
>> couple years or so maybe. We all know how some new things don't work
>> out.
>
> At least one vendor has been trying to sell me on these recently,
> claiming higher bandwidth and better ability to seek. I have not yet
> actually used these.
>
>> Just looking for thoughts and opinions, facts if someone has some.
>> Failure rate compared to single actuator drives if there is such data.
>> My searched didn't help me find anything useful.
>
> There's no reason to suspect that these are going to be terrible, they
> appear to be the next step on the Seagate roadmap before HAMR drives
> hit the market in the coming years.
>
> I haven't seen much on the reliability side of things, however I
> wouldn't be too concerned, assuming that there's proper backup in
> place - Any other drive (including SSDs) could "just die" for a
> multitude of
> reasons.
>
Well, the PV I'm planning to put this on isn't backed up at all. It's
torrent data. If it is lost, hopefully I can get it back again. Still,
I don't want to lose any drive or data if I can help it. It's why I
always get cases that have good cooling. My old Cooler Master HAF-932
had excellent hard drive cooling for the ones mounted on the bottom.
The 5-1/4 bays not so much. While running a bit warmer, it still cooled
good even when filled up. My new Fractal Design Define 7 XL has even
better cooling in a way. The hard drives are not as close to the fans
but it cools all mechanical drives. Anyway, it was something new to me
and I was curious as to what it does. Reading other replies, it sounds
like it is OK but as with anything, there can be issues.
>> P. S. My greens are growing like weeds. Usually they ready to pick by
>> now but having to wait for the tree to be cut down and cut up delayed
>> that. Should be ready by Christmas, I hope. Oh, planted oats, clover,
>> kale and some other extra seeds I had in open area. I saw a LARGE buck
>> deer the other night snacking on the oats. My neighbor would rather see
>> it in his freezer tho. o_0
>
> Hopefully there's some left for you!
>
> Cheers,
>
> Matt
>
>
1/3rd of my garden is for people. Kale, collards, turnip and mustard
greens. That is inside a electric fence to keep the deer out. The open
part outside the fence is a mix of oats, clover and such. I just grow
it to keep the soil erosion down, improve the organic matter and let the
deer have something to munch on, besides what is inside the fence. I
did have to add another spool of fence tape tho. Dang deer got in but
had trouble leaving. He tore up my fence a bit. I got the wires closer
together now so he won't be getting in again. Funny thing, bucks tend
to tear up fences. Does rarely do. It's the antlers. They get hung up
and the shocks makes a deer want to be somewhere else. I can't blame
him tho. Those shocks hurt, bad. Several of them can easily kill a
deer sized animal, even a cow or horse.
Time for my weekly Doctor visit. Reply to others later on.
Dale
:-) :-)
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [gentoo-user] Seagate hard drives with dual actuators.
2024-11-14 7:55 ` Wols Lists
@ 2024-11-14 16:48 ` Dale
2024-11-15 0:18 ` [OT] " Peter Humphrey
0 siblings, 1 reply; 38+ messages in thread
From: Dale @ 2024-11-14 16:48 UTC (permalink / raw
To: gentoo-user
Wols Lists wrote:
> On 13/11/2024 23:10, Dale wrote:
>> My question is this. Given they cost about $20 more, from what I've
>> found anyway, is it worth it? Is there a downside to this new set of
>> heads being added? I'm thinking a higher failure rate, more risk to
>> data or something like that. I think this is a fairly new thing, last
>> couple years or so maybe. We all know how some new things don't work
>> out.
>
> I think this technology has actually been out for a long time. I'm
> sure I've heard of it ages ago.
>
> Thing is, it's probably one of those things that's been available in
> high-end drives for years, but the cost-benefit ratio has been low so
> few people bought them. Now presumably the economics have changed.
>
> If the actuators are mounted opposite each other, then they can't
> collide, and presumably can operate completely independent of each
> other. The costs of two of them were presumably just deemed not worth it.
>
> An opposite niche (and rather apposite for you) was when I started
> buying disk drives. I think my first was a 2GB Bigfoot, followed by a
> 6GB, and I bought several 18GBs for work. They were "old tech", 5.25"
> 5200rpm in an era of 3.5" 7500rpm, but their capacities were huge and
> cheap. If all you wanted was storage, they were great. Most people
> thought the size and speed of the smaller drives was better value,
> even if it cost more per meg.
>
> Cheers,
> Wol
>
>
It may have been available for really expensive hard drives. Those far
above what I'd be willing to pay and wouldn't look for. I do find it
interesting tho. Why not have three actuators? Or four? If they can
all fit in there and not hit each other, it does give the heads more
opportunities to be over needed data. Then the port between drive and
controller, SATA or maybe SAS, likely becomes a bottle neck. Of course,
you can only fit so many in there. Two may be the limit. Drives are
fairly small already. Plus there is the heat problem.
I remember seeing old drives that had I think 14" platters. They had
large motors to spin them. The controller was a separate thing too. I
think their capacity was like 30 or 40MBs or so. It usually took two
people to move one of those things. The company I worked for upgraded
systems twice in the five years I worked for them. Their fastest one
ran at a blazingly fast 25MHz. It was RISC based, I think I spelled
that right. It was a NCR system. We used old Wyse terminals but that
new system was really fast. Except when two or more of those Genicom
line printers were running. I think they were the model 4440 or
something. The printhead didn't move. It was as wide as the page and
printed a whole line at a time, hence the name line printer. They could
print far faster than a person could read. Still, those systems are
nothing compared to what we have today. Heck, my old video card in my
current rig likely has more computing power than the NCR system back
then, which could handle 200 to 300 users at a time. How far we have
came.
Those were the days tho.
Dale
:-) :-)
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [gentoo-user] Seagate hard drives with dual actuators.
2024-11-14 11:21 ` Michael
@ 2024-11-14 17:00 ` Dale
2024-11-14 19:12 ` Michael
0 siblings, 1 reply; 38+ messages in thread
From: Dale @ 2024-11-14 17:00 UTC (permalink / raw
To: gentoo-user
[-- Attachment #1: Type: text/plain, Size: 4822 bytes --]
Michael wrote:
> On Wednesday 13 November 2024 23:10:10 GMT Dale wrote:
>> Howdy,
>>
>> One of my PVs is about 83% full. Time to add more space, soon anyway.
>> I try not to go past 90%. Anyway, I was looking at hard drives and
>> noticed something new. I think I saw one a while back but didn't look
>> into it at the time. I'm looking at 18TB drives, right now. Some new
>> Seagate drives have dual actuators. Basically, they have two sets of
>> heads. In theory, if circumstances are right, it could read data twice
>> as fast. Of course, most of the time that won't be the case but it can
>> happen often enough to make it get data a little faster. Even a 25% or
>> 30% increase gives Seagate something to brag about. Another sales tool.
>> Some heavy data users wouldn't mind either.
>>
>> My question is this. Given they cost about $20 more, from what I've
>> found anyway, is it worth it? Is there a downside to this new set of
>> heads being added? I'm thinking a higher failure rate, more risk to
>> data or something like that. I think this is a fairly new thing, last
>> couple years or so maybe. We all know how some new things don't work out.
>>
>> Just looking for thoughts and opinions, facts if someone has some.
>> Failure rate compared to single actuator drives if there is such data.
>> My searched didn't help me find anything useful.
>>
>> Thanks.
>>
>> Dale
>>
>> :-) :-)
> I don't know much about these drives beyond what the OEM claims. From what I
> read, I can surmise the following hypotheses:
>
> These drives draw more power from your PSU and although they are filled with
> helium to mitigate against higher power/heat, they will require better cooling
> at the margin than a conventional drive.
>
> Your system will use dev-libs/libaio to read the whole disk as a single SATA
> drive (a SAS port will read it as two separate LUNs). The first 50% of LBAs
> will be accessed by the first head and the last 50% by the other head. So
> far, so good.
>
> Theoretically, I suspect this creates a higher probability of failure. In the
> hypothetical scenario of a large sequential write where both heads are writing
> data of a single file, then both heads must succeed in their write operation.
> The cumulative probability of success of head A + head B is calculated as
> P(A⋂B). As an example, if say the probability of a successful write of each
> head is 80%, the cumulative probability of both heads succeeding is only 64%:
>
> 0.8 * 0.8 = 0.64
>
> As long as I didn't make any glaring errors, this simplistic thought
> experiment assumes all else being equal with a conventional single head drive,
> but it never is. The reliability of a conventional non-helium filled drive
> may be lower to start with. Seagate claim their Exos 2 reliability is
> comparable to other enterprise-grade hard drives, but I don't have any real
> world experience to share here. I expect by the time enough reliability
> statistics are available, the OEMs would have moved on to different drive
> technologies.
>
> When considering buying this drive you could look at the market segment needs
> and use cases Seagate/WD could have tried to address by developing and
> marketing this technology. These drives are for cloud storage
> implementations, where higher IOPS, data density and speed of read/write is
> desired, while everything is RAID'ed and backed up. The trade off is power
> usage and heat.
>
> Personally, I tend to buy n-1 versions of storage solutions, for the following
> reasons:
>
> 1. Price per GB is cheaper.
> 2. Any bad news and rumours about novel failing technologies or unsuitable
> implementations (e.g. unmarked SMRs being used in NAS) tend to spread far and
> wide over time.
> 3. High volume sellers start offering discounts for older models.
>
> However, I don't have a need to store the amount of data you do. Most of my
> drives stay empty. Here's a 4TB spinning disk with 3 OS and 9 partitions:
>
> ~ # gdisk -l /dev/sda | grep TiB
> Disk /dev/sda: 7814037168 sectors, 3.6 TiB
> Total free space is 6986885052 sectors (3.3 TiB)
>
> HTH
Sounds like my system may not can even handle one of these. I'm not
sure my SATA ports support that stuff. It sounds like this is not
something I really need anyway. After all, I'm already spanning my data
over three drives. I'm sure some data is coming from each drive. No
way to really know for sure but makes sense.
Do you have a link or something to a place that explains what parts of
the Seagate model number means? I know ST is for Seagate. The size is
next. After that, everything I find is old and outdated. I looked on
the Seagate website to but had no luck. I figure someone made one,
somewhere. A link would be fine.
Thanks.
Dale
:-) :-)
[-- Attachment #2: Type: text/html, Size: 5572 bytes --]
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [gentoo-user] Seagate hard drives with dual actuators.
2024-11-14 17:00 ` Dale
@ 2024-11-14 19:12 ` Michael
2024-11-14 19:51 ` Frank Steinmetzger
2024-11-14 20:33 ` Dale
0 siblings, 2 replies; 38+ messages in thread
From: Michael @ 2024-11-14 19:12 UTC (permalink / raw
To: gentoo-user
[-- Attachment #1: Type: text/plain, Size: 5600 bytes --]
On Thursday 14 November 2024 17:00:07 GMT Dale wrote:
> Michael wrote:
> > On Wednesday 13 November 2024 23:10:10 GMT Dale wrote:
> >> Howdy,
> >>
> >> One of my PVs is about 83% full. Time to add more space, soon anyway.
> >> I try not to go past 90%. Anyway, I was looking at hard drives and
> >> noticed something new. I think I saw one a while back but didn't look
> >> into it at the time. I'm looking at 18TB drives, right now. Some new
> >> Seagate drives have dual actuators. Basically, they have two sets of
> >> heads. In theory, if circumstances are right, it could read data twice
> >> as fast. Of course, most of the time that won't be the case but it can
> >> happen often enough to make it get data a little faster. Even a 25% or
> >> 30% increase gives Seagate something to brag about. Another sales tool.
> >>
> >> Some heavy data users wouldn't mind either.
> >>
> >> My question is this. Given they cost about $20 more, from what I've
> >> found anyway, is it worth it? Is there a downside to this new set of
> >> heads being added? I'm thinking a higher failure rate, more risk to
> >> data or something like that. I think this is a fairly new thing, last
> >> couple years or so maybe. We all know how some new things don't work
> >> out.
> >>
> >> Just looking for thoughts and opinions, facts if someone has some.
> >> Failure rate compared to single actuator drives if there is such data.
> >> My searched didn't help me find anything useful.
> >>
> >> Thanks.
> >>
> >> Dale
> >>
> >> :-) :-)
> >
> > I don't know much about these drives beyond what the OEM claims. From
> > what I read, I can surmise the following hypotheses:
> >
> > These drives draw more power from your PSU and although they are filled
> > with helium to mitigate against higher power/heat, they will require
> > better cooling at the margin than a conventional drive.
> >
> > Your system will use dev-libs/libaio to read the whole disk as a single
> > SATA drive (a SAS port will read it as two separate LUNs). The first 50%
> > of LBAs will be accessed by the first head and the last 50% by the other
> > head. So far, so good.
> >
> > Theoretically, I suspect this creates a higher probability of failure. In
> > the hypothetical scenario of a large sequential write where both heads
> > are writing data of a single file, then both heads must succeed in their
> > write operation. The cumulative probability of success of head A + head B
> > is calculated as P(A⋂B). As an example, if say the probability of a
> > successful write of each head is 80%, the cumulative probability of both
> > heads succeeding is only 64%:
> >
> > 0.8 * 0.8 = 0.64
> >
> > As long as I didn't make any glaring errors, this simplistic thought
> > experiment assumes all else being equal with a conventional single head
> > drive, but it never is. The reliability of a conventional non-helium
> > filled drive may be lower to start with. Seagate claim their Exos 2
> > reliability is comparable to other enterprise-grade hard drives, but I
> > don't have any real world experience to share here. I expect by the time
> > enough reliability statistics are available, the OEMs would have moved on
> > to different drive technologies.
> >
> > When considering buying this drive you could look at the market segment
> > needs and use cases Seagate/WD could have tried to address by developing
> > and marketing this technology. These drives are for cloud storage
> > implementations, where higher IOPS, data density and speed of read/write
> > is
> > desired, while everything is RAID'ed and backed up. The trade off is
> > power
> > usage and heat.
> >
> > Personally, I tend to buy n-1 versions of storage solutions, for the
> > following reasons:
> >
> > 1. Price per GB is cheaper.
> > 2. Any bad news and rumours about novel failing technologies or unsuitable
> > implementations (e.g. unmarked SMRs being used in NAS) tend to spread far
> > and wide over time.
> > 3. High volume sellers start offering discounts for older models.
> >
> > However, I don't have a need to store the amount of data you do. Most of
> > my drives stay empty. Here's a 4TB spinning disk with 3 OS and 9
> > partitions:
> >
> > ~ # gdisk -l /dev/sda | grep TiB
> > Disk /dev/sda: 7814037168 sectors, 3.6 TiB
> > Total free space is 6986885052 sectors (3.3 TiB)
> >
> > HTH
>
> Sounds like my system may not can even handle one of these. I'm not
> sure my SATA ports support that stuff.
I think your PC would handle these fine.
> It sounds like this is not something I really need anyway.
Well, this is more to the point. ;-)
> After all, I'm already spanning my data
> over three drives. I'm sure some data is coming from each drive. No
> way to really know for sure but makes sense.
>
> Do you have a link or something to a place that explains what parts of
> the Seagate model number means? I know ST is for Seagate. The size is
> next. After that, everything I find is old and outdated. I looked on
> the Seagate website to but had no luck. I figure someone made one,
> somewhere. A link would be fine.
This document is from 2011, I don't know if they changed their nomenclature
since then.
https://www.seagate.com/files/staticfiles/docs/pdf/marketing/st-model-number-cheat-sheet-sc504-1-1102us.pdf
> Thanks.
>
> Dale
>
> :-) :-)
The only Seagate 7200RPM disk I have started playing up a month ago. I now
have to replace it. :-(
[-- Attachment #2: This is a digitally signed message part. --]
[-- Type: application/pgp-signature, Size: 833 bytes --]
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [gentoo-user] Seagate hard drives with dual actuators.
2024-11-14 19:12 ` Michael
@ 2024-11-14 19:51 ` Frank Steinmetzger
2024-11-14 19:55 ` Frank Steinmetzger
2024-11-14 20:33 ` Dale
1 sibling, 1 reply; 38+ messages in thread
From: Frank Steinmetzger @ 2024-11-14 19:51 UTC (permalink / raw
To: gentoo-user
Am Donnerstag, 14. November 2024, 20:12:25 Mitteleuropäische Normalzeit
> The only Seagate 7200RPM disk I have started playing up a month ago. I now
> have to replace it. :-(
The German tech bubble has a saying when it’s about Seagate: “Sie geht oder
sie geht nicht”. It plays on the fact that “sie geht” (literally “she runs”¹,
meaning “it works”) sounds very similar to “Seagate”. So the literal joke is
“Either it works or it doesn’t”, and the meta joke is “Seagate or not
Seagate”.
¹ hard disk (“die Festplatte”) has feminine grammatical gender in German
--
Grüße | Greetings | Salut | Qapla’
Just don’t ignore it.
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [gentoo-user] Seagate hard drives with dual actuators.
2024-11-14 19:51 ` Frank Steinmetzger
@ 2024-11-14 19:55 ` Frank Steinmetzger
2024-11-14 23:14 ` Peter Humphrey
0 siblings, 1 reply; 38+ messages in thread
From: Frank Steinmetzger @ 2024-11-14 19:55 UTC (permalink / raw
To: gentoo-user
Am Donnerstag, 14. November 2024, 20:51:32 Mitteleuropäische Normalzeit
schrieb Frank Steinmetzger:
> Am Donnerstag, 14. November 2024, 20:12:25 Mitteleuropäische Normalzeit
>
> > The only Seagate 7200RPM disk I have started playing up a month ago. I
> > now
> > have to replace it. :-(
>
> The German tech bubble has a saying when it’s about Seagate: “Sie geht oder
> sie geht nicht”. It plays on the fact that “sie geht” (literally “she
> runs”¹, meaning “it works”) sounds very similar to “Seagate”. So the
> literal joke is “Either it works or it doesn’t”, and the meta joke is
> “Seagate or not Seagate”.
Lol, writing the above text gave me the strange feeling of having written it
before. So I looked into my archive and I have indeed: in June 2014 *and* in
December 2020. 🫣
--
Grüße | Greetings | Salut | Qapla’
What do brushing teeth and voting have in common?
If you don’t do it, it becomes brown on its own.
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [gentoo-user] Seagate hard drives with dual actuators.
2024-11-14 19:12 ` Michael
2024-11-14 19:51 ` Frank Steinmetzger
@ 2024-11-14 20:33 ` Dale
2024-11-14 20:57 ` Rich Freeman
2024-11-14 22:38 ` Wols Lists
1 sibling, 2 replies; 38+ messages in thread
From: Dale @ 2024-11-14 20:33 UTC (permalink / raw
To: gentoo-user
[-- Attachment #1: Type: text/plain, Size: 9764 bytes --]
Michael wrote:
> On Thursday 14 November 2024 17:00:07 GMT Dale wrote:
>> Michael wrote:
>>> On Wednesday 13 November 2024 23:10:10 GMT Dale wrote:
>>>> Howdy,
>>>>
>>>> One of my PVs is about 83% full. Time to add more space, soon anyway.
>>>> I try not to go past 90%. Anyway, I was looking at hard drives and
>>>> noticed something new. I think I saw one a while back but didn't look
>>>> into it at the time. I'm looking at 18TB drives, right now. Some new
>>>> Seagate drives have dual actuators. Basically, they have two sets of
>>>> heads. In theory, if circumstances are right, it could read data twice
>>>> as fast. Of course, most of the time that won't be the case but it can
>>>> happen often enough to make it get data a little faster. Even a 25% or
>>>> 30% increase gives Seagate something to brag about. Another sales tool.
>>>>
>>>> Some heavy data users wouldn't mind either.
>>>>
>>>> My question is this. Given they cost about $20 more, from what I've
>>>> found anyway, is it worth it? Is there a downside to this new set of
>>>> heads being added? I'm thinking a higher failure rate, more risk to
>>>> data or something like that. I think this is a fairly new thing, last
>>>> couple years or so maybe. We all know how some new things don't work
>>>> out.
>>>>
>>>> Just looking for thoughts and opinions, facts if someone has some.
>>>> Failure rate compared to single actuator drives if there is such data.
>>>> My searched didn't help me find anything useful.
>>>>
>>>> Thanks.
>>>>
>>>> Dale
>>>>
>>>> :-) :-)
>>> I don't know much about these drives beyond what the OEM claims. From
>>> what I read, I can surmise the following hypotheses:
>>>
>>> These drives draw more power from your PSU and although they are filled
>>> with helium to mitigate against higher power/heat, they will require
>>> better cooling at the margin than a conventional drive.
>>>
>>> Your system will use dev-libs/libaio to read the whole disk as a single
>>> SATA drive (a SAS port will read it as two separate LUNs). The first 50%
>>> of LBAs will be accessed by the first head and the last 50% by the other
>>> head. So far, so good.
>>>
>>> Theoretically, I suspect this creates a higher probability of failure. In
>>> the hypothetical scenario of a large sequential write where both heads
>>> are writing data of a single file, then both heads must succeed in their
>>> write operation. The cumulative probability of success of head A + head B
>>> is calculated as P(A⋂B). As an example, if say the probability of a
>>> successful write of each head is 80%, the cumulative probability of both
>>> heads succeeding is only 64%:
>>>
>>> 0.8 * 0.8 = 0.64
>>>
>>> As long as I didn't make any glaring errors, this simplistic thought
>>> experiment assumes all else being equal with a conventional single head
>>> drive, but it never is. The reliability of a conventional non-helium
>>> filled drive may be lower to start with. Seagate claim their Exos 2
>>> reliability is comparable to other enterprise-grade hard drives, but I
>>> don't have any real world experience to share here. I expect by the time
>>> enough reliability statistics are available, the OEMs would have moved on
>>> to different drive technologies.
>>>
>>> When considering buying this drive you could look at the market segment
>>> needs and use cases Seagate/WD could have tried to address by developing
>>> and marketing this technology. These drives are for cloud storage
>>> implementations, where higher IOPS, data density and speed of read/write
>>> is
>>> desired, while everything is RAID'ed and backed up. The trade off is
>>> power
>>> usage and heat.
>>>
>>> Personally, I tend to buy n-1 versions of storage solutions, for the
>>> following reasons:
>>>
>>> 1. Price per GB is cheaper.
>>> 2. Any bad news and rumours about novel failing technologies or unsuitable
>>> implementations (e.g. unmarked SMRs being used in NAS) tend to spread far
>>> and wide over time.
>>> 3. High volume sellers start offering discounts for older models.
>>>
>>> However, I don't have a need to store the amount of data you do. Most of
>>> my drives stay empty. Here's a 4TB spinning disk with 3 OS and 9
>>> partitions:
>>>
>>> ~ # gdisk -l /dev/sda | grep TiB
>>> Disk /dev/sda: 7814037168 sectors, 3.6 TiB
>>> Total free space is 6986885052 sectors (3.3 TiB)
>>>
>>> HTH
>> Sounds like my system may not can even handle one of these. I'm not
>> sure my SATA ports support that stuff.
> I think your PC would handle these fine.
>
>
>> It sounds like this is not something I really need anyway.
> Well, this is more to the point. ;-)
>
>
>> After all, I'm already spanning my data
>> over three drives. I'm sure some data is coming from each drive. No
>> way to really know for sure but makes sense.
>>
>> Do you have a link or something to a place that explains what parts of
>> the Seagate model number means? I know ST is for Seagate. The size is
>> next. After that, everything I find is old and outdated. I looked on
>> the Seagate website to but had no luck. I figure someone made one,
>> somewhere. A link would be fine.
> This document is from 2011, I don't know if they changed their nomenclature
> since then.
>
> https://www.seagate.com/files/staticfiles/docs/pdf/marketing/st-model-number-cheat-sheet-sc504-1-1102us.pdf
>
>
>> Thanks.
>>
>> Dale
>>
>> :-) :-)
> The only Seagate 7200RPM disk I have started playing up a month ago. I now
> have to replace it. :-(
Yea, I found that one too. I see drives with letters that are not
listed under Segment. They got new stuff, or changed letters to trick
folks. I emailed the company I usually buy drives from, they do server
stuff, but haven't heard back yet. Could be, there isn't one for new
drives. Could be they there to make it look like they mean something
but don't, again, to trick folks.
I've had a Seagate, a Maxtor from way back and a Western Digital go
bad. This is one reason I don't knock any drive maker. Any of them can
produce a bad drive. What matters, if they stand behind it and make it
good or not. It's one thing that kinda gets on my nerves about SMR. It
seems, sounds, like they tried to hide it from people to make money.
Thing is, as some learned, they don't do well in a RAID and some other
situations. Heck, they do OK reading but when writing, they can get
real slow when writing a lot of data. Then you have to wait until it
gets done redoing things so that it is complete. I still have that SMR
drive for a backup. It completes the backup pretty quick, if it isn't
much data, but after it is done, it does that bumpy thing for a lot
longer than the copy process does. I wish I never bought that thing.
The one good thing, I can unmount it and unhook the SATA cable while it
finishes. All it needs is power. Still annoying tho.
Think I'll try for a 18TB drive with one actuator. Oh, some info on my
data storage. This doesn't include backup drives.
root@Gentoo-1 / # dfc
FILESYSTEM (=) USED FREE (-) %USED USED AVAILABLE
TOTAL MOUNTED ON
/dev/root [===-----------------] 11.4% 24.6G 348.0G
392.7G /
devtmpfs [--------------------] 0.0% 0.0B 10.0M
10.0M /dev
tmpfs [=-------------------] 0.0% 1.7M 25.1G
25.1G /run
efivarfs [=========-----------] 43.2% 50.3K 72.7K
128.0K +ys/firmware/efi/efivars
shm [=-------------------] 0.0% 136.0K 62.9G
62.9G /dev/shm
/dev/nvme0n1p2 [==------------------] 6.4% 137.5M
9.2G 9.8G /boot
/dev/nvme0n1p4 [=====---------------] 20.9% 18.8G 139.3G
176.1G /var
+v/mapper/home2-home--lv [=====---------------] 21.5% 1.4T
5.7T 7.2T /home
/dev/nvme0n1p1 [=-------------------] 0.0% 152.0K
2.0G 2.0G /efi
+ev/mapper/datavg-datalv [=================---] 83.3% 34.6T 6.9T
41.5T /home/dale/Desktop/Data
tmpfs [=-------------------] 0.0% 4.0K 70.0G
70.0G /var/tmp/portage
tmpfs [=-------------------] 0.0% 44.0K 12.6G
12.6G /run/user/1000
/dev/mapper/crypt [===============-----] 73.9% 34.8T 12.3T
47.1T /home/dale/Desktop/Crypt
/dev/mapper/6tb-main [=============-------] 61.2% 3.3T
2.1T 5.4T /mnt/6tb-main
SUM: [===============-----] 72.7% 74.2T 27.6T
102.0T
root@Gentoo-1 / # pvs -O vg_name
PV VG Fmt Attr PSize PFree
/dev/sde1 datavg lvm2 a-- 12.73t 0
/dev/sdc1 datavg lvm2 a-- 14.55t 0
/dev/sdb1 datavg lvm2 a-- 14.55t 0
/dev/sda1 home2 lvm2 a-- <7.28t 0
/dev/sdf1 vg.crypt lvm2 a-- 16.37t 0
/dev/sdd1 vg.crypt lvm2 a-- 14.55t 0
/dev/sdg1 vg.crypt lvm2 a-- 16.37t 0
root@Gentoo-1 / #
That looks better in a Konsole. Oooo. I'm over 100TBs now. O_O
Dale
:-) :-)
[-- Attachment #2: Type: text/html, Size: 12003 bytes --]
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [gentoo-user] Seagate hard drives with dual actuators.
2024-11-14 20:33 ` Dale
@ 2024-11-14 20:57 ` Rich Freeman
2024-11-14 23:10 ` Dale
2024-11-14 22:38 ` Wols Lists
1 sibling, 1 reply; 38+ messages in thread
From: Rich Freeman @ 2024-11-14 20:57 UTC (permalink / raw
To: gentoo-user
On Thu, Nov 14, 2024 at 3:33 PM Dale <rdalek1967@gmail.com> wrote:
>
>
> I've had a Seagate, a Maxtor from way back and a Western Digital go bad. This is one reason I don't knock any drive maker. Any of them can produce a bad drive.
++
All the consumer drive manufacturers are in a super-price-conscious
market. For the most part their drives work as advertised, and
basically all of them have made a model with an abysmal failure rate
at some point. I think Backblaze still publishes their stats and I
don't think any manufacturer stood out when it comes to these cheaper
consumer drives. The enterprise stuff probably is more reliable, but
for HDD I don't think it is worth it - just use RAID.
I've personally been trying to shift towards solid state. Granted, it
is about double the price of large 5400RPM drives, but the performance
is incomparable. I've also been moving more towards used enterprise
drives with PLP/etc and where I can find the drive endurance info on
the ebay listing. You can typically pay about $50/TB for an
enterprise SSD - either SATA or NVMe. You'll pay a bit more if you
want a high capacity drive (U.2 16+TB or whatever). That's in part
because I've been shifting to Ceph which is pretty IOPS-sensitive.
However, it is nice that when I add/replace a drive the cluster
rebuilds in an hour at most with kinda shocking network IO.
I'll use cheap consumer SATA/M.2 SSDs for OS drives that are easily
reimaged. I'll use higher performance M.2 for gaming rigs (mostly
read-oriented), and back it up. Be aware that the consumer write
benchmarks only work for short bursts and fall to a fraction of the
advertisement for sustained writes - the enterprise write benchmarks
reflect sustained writes and you can run at that speed until the drive
dies.
--
Rich
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [gentoo-user] Seagate hard drives with dual actuators.
2024-11-14 20:33 ` Dale
2024-11-14 20:57 ` Rich Freeman
@ 2024-11-14 22:38 ` Wols Lists
2024-11-15 9:35 ` Michael
1 sibling, 1 reply; 38+ messages in thread
From: Wols Lists @ 2024-11-14 22:38 UTC (permalink / raw
To: gentoo-user
On 14/11/2024 20:33, Dale wrote:
> It's one thing that kinda gets on my nerves about SMR. It seems,
> sounds, like they tried to hide it from people to make money. Thing is,
> as some learned, they don't do well in a RAID and some other
> situations. Heck, they do OK reading but when writing, they can get
> real slow when writing a lot of data. Then you have to wait until it
> gets done redoing things so that it is complete.
Incidentally, when I looked up HAMR (I didn't know what it was) it's
touted as making SMR obsolete. I can see why ...
And dual actuator? I would have thought that would be good for SMR
drives. Not that I have a clue how they work internally, but I would
have thought it made sense to have zones and a streaming log-structured
layout. So when the user is using it, you're filling up the zones, and
then when the drive has "free time", it takes a full zone that has the
largest "freed/dead space" and streams it to the current zone, one
actuator to read and one to write. Indeed, it could possibly do that
while the drive is being used ...
Cheers,
Wol
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [gentoo-user] Seagate hard drives with dual actuators.
2024-11-14 20:57 ` Rich Freeman
@ 2024-11-14 23:10 ` Dale
2024-11-15 0:59 ` Rich Freeman
0 siblings, 1 reply; 38+ messages in thread
From: Dale @ 2024-11-14 23:10 UTC (permalink / raw
To: gentoo-user
Rich Freeman wrote:
> On Thu, Nov 14, 2024 at 3:33 PM Dale <rdalek1967@gmail.com> wrote:
>>
>> I've had a Seagate, a Maxtor from way back and a Western Digital go bad. This is one reason I don't knock any drive maker. Any of them can produce a bad drive.
> ++
>
> All the consumer drive manufacturers are in a super-price-conscious
> market. For the most part their drives work as advertised, and
> basically all of them have made a model with an abysmal failure rate
> at some point. I think Backblaze still publishes their stats and I
> don't think any manufacturer stood out when it comes to these cheaper
> consumer drives. The enterprise stuff probably is more reliable, but
> for HDD I don't think it is worth it - just use RAID.
>
> I've personally been trying to shift towards solid state. Granted, it
> is about double the price of large 5400RPM drives, but the performance
> is incomparable. I've also been moving more towards used enterprise
> drives with PLP/etc and where I can find the drive endurance info on
> the ebay listing. You can typically pay about $50/TB for an
> enterprise SSD - either SATA or NVMe. You'll pay a bit more if you
> want a high capacity drive (U.2 16+TB or whatever). That's in part
> because I've been shifting to Ceph which is pretty IOPS-sensitive.
> However, it is nice that when I add/replace a drive the cluster
> rebuilds in an hour at most with kinda shocking network IO.
>
> I'll use cheap consumer SATA/M.2 SSDs for OS drives that are easily
> reimaged. I'll use higher performance M.2 for gaming rigs (mostly
> read-oriented), and back it up. Be aware that the consumer write
> benchmarks only work for short bursts and fall to a fraction of the
> advertisement for sustained writes - the enterprise write benchmarks
> reflect sustained writes and you can run at that speed until the drive
> dies.
>
The biggest downside to the large drives available now, even if SMART
tells you a drive is failing, you likely won't have time to copy the
data over to a new drive before it fails. On a 18TB drive, using
pvmove, it can take a long time to move data. I think the last time I
did it, it took like two days. Usually SMART is more like a 24 hour
warning. Of course with some failures, you get no warning at all.
SMART can't detect all modes of failure.
Wol has a good point too. What they trying to do is compete with SSD if
they can. SSD, NVME and such are pricey if you need to store bulk
amounts of data. I don't even want to think what it would cost to put
all my 100TBs or so on SSD or NVME drives. WOW!!!
Dale
:-) :-)
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [gentoo-user] Seagate hard drives with dual actuators.
2024-11-14 19:55 ` Frank Steinmetzger
@ 2024-11-14 23:14 ` Peter Humphrey
0 siblings, 0 replies; 38+ messages in thread
From: Peter Humphrey @ 2024-11-14 23:14 UTC (permalink / raw
To: gentoo-user
On Thursday 14 November 2024 19:55:19 GMT Frank Steinmetzger wrote:
> Lol, writing the above text gave me the strange feeling of having written it
> before. So I looked into my archive and I have indeed: in June 2014 *and*
> in December 2020. 🫣
Tiresomely repetitious, then...
:)
--
Regards,
Peter.
^ permalink raw reply [flat|nested] 38+ messages in thread
* [OT] Re: [gentoo-user] Seagate hard drives with dual actuators.
2024-11-14 16:48 ` Dale
@ 2024-11-15 0:18 ` Peter Humphrey
2024-11-15 8:41 ` [gentoo-user] Hollerith (was: Seagate hard drives with dual actuators) karl
2024-11-15 9:51 ` [OT] Re: [gentoo-user] Seagate hard drives with dual actuators Wols Lists
0 siblings, 2 replies; 38+ messages in thread
From: Peter Humphrey @ 2024-11-15 0:18 UTC (permalink / raw
To: gentoo-user
On Thursday 14 November 2024 16:48:56 GMT Dale wrote:
> I remember seeing old drives that had I think 14" platters. They had
> large motors to spin them. The controller was a separate thing too. I
> think their capacity was like 30 or 40MBs or so. It usually took two
> people to move one of those things. The company I worked for upgraded
> systems twice in the five years I worked for them. Their fastest one
> ran at a blazingly fast 25MHz. It was RISC based, I think I spelled
> that right.
"Reduced instruction set computing." Hyphenate as you like.
In the 70s and 80s the national grid control centre in this country used three
2MB disks, any one or two of which could be online at any time. I can't tell
you the platter size, but they were mounted in cabinets about 5' long, 3'6"
tall and 2' wide. Each. Our training included servicing the air circulation
system and its filters. I still remember the aluminium-alloy casings.
The two computers were Ferranti Argus 500 minis, one on-line, one hot standby,
with 16MB of 2-microsecond core store, with the applications written in-house.
The operating system was primitive compared with today, also written by us.
Nothing relocatable; programs and data lived in strictly defined regions. They
drove (I think) 24 analogue vector-drawing displays (no raster to modulate) in
the control room upstairs. They never could display a perfect circle - there
was always a tiny gap. The D/A converters were big beasts. The systems had
been developed from earlier ones in nuclear submarines. We didn't have their
big, circular consoles though.
I hate to think how many miles of 8-hole tape I wound and rewound. Thank
goodness we didn't have to cope with 80-column punched cards (Hollerith?) as
the ivory-tower, batch-processing mainframe people did.
We used to show visitors around, some from American utility companies, who
always asked "yes, nice display driver, but where are your main machines?
Great things be could achieved with assembler and first-class people. More
efficient than just throwing money at a problem until it gives in.
> Those were the days tho.
Indeed, the best. Mind you, nostalgia isn't what it used to be...
Time for bed, said Zebedee.
--
Regards,
Peter.
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [gentoo-user] Seagate hard drives with dual actuators.
2024-11-14 23:10 ` Dale
@ 2024-11-15 0:59 ` Rich Freeman
2024-11-15 5:53 ` Dale
0 siblings, 1 reply; 38+ messages in thread
From: Rich Freeman @ 2024-11-15 0:59 UTC (permalink / raw
To: gentoo-user
On Thu, Nov 14, 2024 at 6:10 PM Dale <rdalek1967@gmail.com> wrote:
>
> The biggest downside to the large drives available now, even if SMART
> tells you a drive is failing, you likely won't have time to copy the
> data over to a new drive before it fails. On a 18TB drive, using
> pvmove, it can take a long time to move data.
Very true. This is why I'm essentially running RAID6. Well, that and
for various reasons you don't want to allow writes to Ceph without at
least one drive worth of redundancy, so having an extra replica means
that you can lose one and remain read-write, and then if you lose a
second during recovery you might be read-only but you still have data
integrity. (Don't want to get into details - it is a Ceph-specific
issue.)
> I don't even want to think what it would cost to put
> all my 100TBs or so on SSD or NVME drives. WOW!!!
# kubectl rook-ceph ceph osd df class ssd
ID CLASS WEIGHT REWEIGHT SIZE RAW USE DATA OMAP
META AVAIL %USE VAR PGS STATUS
8 ssd 6.98630 1.00000 7.0 TiB 1.7 TiB 1.7 TiB 63 MiB 3.9
GiB 5.3 TiB 24.66 1.04 179 up
4 ssd 1.74660 1.00000 1.7 TiB 465 GiB 462 GiB 16 MiB 2.5
GiB 1.3 TiB 25.99 1.10 45 up
12 ssd 1.74660 1.00000 1.7 TiB 547 GiB 545 GiB 30 MiB 2.1
GiB 1.2 TiB 30.57 1.29 52 up
1 ssd 6.98630 1.00000 7.0 TiB 1.7 TiB 1.7 TiB 50 MiB 4.2
GiB 5.3 TiB 24.42 1.03 177 up
5 ssd 6.98630 1.00000 7.0 TiB 1.8 TiB 1.8 TiB 24 MiB 5.0
GiB 5.2 TiB 25.14 1.07 180 up
3 ssd 1.74660 1.00000 1.7 TiB 585 GiB 583 GiB 18 MiB 2.0
GiB 1.2 TiB 32.70 1.39 57 up
21 ssd 1.74660 1.00000 1.7 TiB 470 GiB 468 GiB 27 MiB 1.9
GiB 1.3 TiB 26.26 1.11 52 up
9 ssd 1.74660 1.00000 1.7 TiB 506 GiB 504 GiB 11 MiB 2.0
GiB 1.3 TiB 28.29 1.20 49 up
18 ssd 1.74660 1.00000 1.7 TiB 565 GiB 563 GiB 16 MiB 1.7
GiB 1.2 TiB 31.59 1.34 55 up
10 ssd 1.74660 1.00000 1.7 TiB 490 GiB 489 GiB 28 MiB 1.6
GiB 1.3 TiB 27.42 1.16 53 up
22 ssd 1.74660 1.00000 1.7 TiB 479 GiB 478 GiB 19 MiB 1.7
GiB 1.3 TiB 26.80 1.14 50 up
19 ssd 13.97249 1.00000 14 TiB 2.3 TiB 2.3 TiB 87 MiB 5.2
GiB 12 TiB 16.81 0.71 262 up
TOTAL 49 TiB 12 TiB 12 TiB 388 MiB 34
GiB 37 TiB 23.61
I'm getting there. Granted, at 3+2 erasure coding that's only a bit
over 30TiB usable space.
--
Rich
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [gentoo-user] Seagate hard drives with dual actuators.
2024-11-15 0:59 ` Rich Freeman
@ 2024-11-15 5:53 ` Dale
2024-11-15 10:09 ` Michael
2024-11-15 10:38 ` Frank Steinmetzger
0 siblings, 2 replies; 38+ messages in thread
From: Dale @ 2024-11-15 5:53 UTC (permalink / raw
To: gentoo-user
Rich Freeman wrote:
> On Thu, Nov 14, 2024 at 6:10 PM Dale <rdalek1967@gmail.com> wrote:
>> The biggest downside to the large drives available now, even if SMART
>> tells you a drive is failing, you likely won't have time to copy the
>> data over to a new drive before it fails. On a 18TB drive, using
>> pvmove, it can take a long time to move data.
> Very true. This is why I'm essentially running RAID6. Well, that and
> for various reasons you don't want to allow writes to Ceph without at
> least one drive worth of redundancy, so having an extra replica means
> that you can lose one and remain read-write, and then if you lose a
> second during recovery you might be read-only but you still have data
> integrity. (Don't want to get into details - it is a Ceph-specific
> issue.)
I think I did some math on this once. I'm not positive on this and it
could vary depending on system ability of moving data. I think about
8TB is as large as you want if you get a 24 hour notice from SMART and
see that notice fairly quickly to act on. Anything beyond that and you
may not have enough time to move data, if the data is even good still.
>> I don't even want to think what it would cost to put
>> all my 100TBs or so on SSD or NVME drives. WOW!!!
> # kubectl rook-ceph ceph osd df class ssd
> ID CLASS WEIGHT REWEIGHT SIZE RAW USE DATA OMAP
> META AVAIL %USE VAR PGS STATUS
> 8 ssd 6.98630 1.00000 7.0 TiB 1.7 TiB 1.7 TiB 63 MiB 3.9
> GiB 5.3 TiB 24.66 1.04 179 up
> 4 ssd 1.74660 1.00000 1.7 TiB 465 GiB 462 GiB 16 MiB 2.5
> GiB 1.3 TiB 25.99 1.10 45 up
> 12 ssd 1.74660 1.00000 1.7 TiB 547 GiB 545 GiB 30 MiB 2.1
> GiB 1.2 TiB 30.57 1.29 52 up
> 1 ssd 6.98630 1.00000 7.0 TiB 1.7 TiB 1.7 TiB 50 MiB 4.2
> GiB 5.3 TiB 24.42 1.03 177 up
> 5 ssd 6.98630 1.00000 7.0 TiB 1.8 TiB 1.8 TiB 24 MiB 5.0
> GiB 5.2 TiB 25.14 1.07 180 up
> 3 ssd 1.74660 1.00000 1.7 TiB 585 GiB 583 GiB 18 MiB 2.0
> GiB 1.2 TiB 32.70 1.39 57 up
> 21 ssd 1.74660 1.00000 1.7 TiB 470 GiB 468 GiB 27 MiB 1.9
> GiB 1.3 TiB 26.26 1.11 52 up
> 9 ssd 1.74660 1.00000 1.7 TiB 506 GiB 504 GiB 11 MiB 2.0
> GiB 1.3 TiB 28.29 1.20 49 up
> 18 ssd 1.74660 1.00000 1.7 TiB 565 GiB 563 GiB 16 MiB 1.7
> GiB 1.2 TiB 31.59 1.34 55 up
> 10 ssd 1.74660 1.00000 1.7 TiB 490 GiB 489 GiB 28 MiB 1.6
> GiB 1.3 TiB 27.42 1.16 53 up
> 22 ssd 1.74660 1.00000 1.7 TiB 479 GiB 478 GiB 19 MiB 1.7
> GiB 1.3 TiB 26.80 1.14 50 up
> 19 ssd 13.97249 1.00000 14 TiB 2.3 TiB 2.3 TiB 87 MiB 5.2
> GiB 12 TiB 16.81 0.71 262 up
> TOTAL 49 TiB 12 TiB 12 TiB 388 MiB 34
> GiB 37 TiB 23.61
>
> I'm getting there. Granted, at 3+2 erasure coding that's only a bit
> over 30TiB usable space.
>
The thing about my data, it's mostly large video files. If I were
storing documents or something, then SSD or something would be a good
option. Plus, I mostly write once, then it either sits there a while or
gets read on occasion. I checked and I have some 56,000 videos. That
doesn't include Youtube videos. This is also why I wanted to use that
checksum script.
I do wish there was a easy way to make columns work when we copy and
paste into email. :/
Dale
:-) :-)
^ permalink raw reply [flat|nested] 38+ messages in thread
* [gentoo-user] Hollerith (was: Seagate hard drives with dual actuators)
2024-11-15 0:18 ` [OT] " Peter Humphrey
@ 2024-11-15 8:41 ` karl
2024-11-15 9:51 ` [OT] Re: [gentoo-user] Seagate hard drives with dual actuators Wols Lists
1 sibling, 0 replies; 38+ messages in thread
From: karl @ 2024-11-15 8:41 UTC (permalink / raw
To: gentoo-user
Peter Humphrey:
...
> I hate to think how many miles of 8-hole tape I wound and rewound. Thank
> goodness we didn't have to cope with 80-column punched cards (Hollerith?) as
> the ivory-tower, batch-processing mainframe people did.
...
Hollerith was the founder of the company which later became IBM.
There are two things named as Hollerith, the cards which later
developed into the IBM 80-col. card, and the constants representing
text/strings/characters in early fortran.
https://en.wikipedia.org/wiki/Hollerith
Fortran was my first language, I still have unpunched cards left,
perfect as bookmarks.
///
Old style secondary storage (boxes of punched cards):
https://en.wikipedia.org/wiki/File:Photograph_of_Federal_Records_Center,_Alexandria,_Virginia_(34877725360).jpg
Regards,
/Karl Hammar
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [gentoo-user] Seagate hard drives with dual actuators.
2024-11-14 22:38 ` Wols Lists
@ 2024-11-15 9:35 ` Michael
0 siblings, 0 replies; 38+ messages in thread
From: Michael @ 2024-11-15 9:35 UTC (permalink / raw
To: gentoo-user
[-- Attachment #1: Type: text/plain, Size: 1824 bytes --]
On Thursday 14 November 2024 22:38:59 GMT Wols Lists wrote:
> On 14/11/2024 20:33, Dale wrote:
> > It's one thing that kinda gets on my nerves about SMR. It seems,
> > sounds, like they tried to hide it from people to make money. Thing is,
> > as some learned, they don't do well in a RAID and some other
> > situations. Heck, they do OK reading but when writing, they can get
> > real slow when writing a lot of data. Then you have to wait until it
> > gets done redoing things so that it is complete.
>
> Incidentally, when I looked up HAMR (I didn't know what it was) it's
> touted as making SMR obsolete. I can see why ...
aaaand .... it's gone! LOL! Apparently, HDMR is on the cards to replace
HAMR.
> And dual actuator? I would have thought that would be good for SMR
> drives. Not that I have a clue how they work internally, but I would
> have thought it made sense to have zones and a streaming log-structured
> layout. So when the user is using it, you're filling up the zones, and
> then when the drive has "free time", it takes a full zone that has the
> largest "freed/dead space" and streams it to the current zone, one
> actuator to read and one to write. Indeed, it could possibly do that
> while the drive is being used ...
>
> Cheers,
> Wol
As I understand it the dual actuator drives are like two separate drives
placed inside the same casing, but being accessed in parallel by their
respective head, hence the higher IOPS. The helium allows thinner platters
which makes it possible to have higher total storage capacity within the same
physical size.
The SMR is more complicated in technological terms, with its overlapping
multilayered recording. It achieves higher storage density than conventional
drives, but with lower IOPS on writing operations once their cache is
exhausted.
[-- Attachment #2: This is a digitally signed message part. --]
[-- Type: application/pgp-signature, Size: 833 bytes --]
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [OT] Re: [gentoo-user] Seagate hard drives with dual actuators.
2024-11-15 0:18 ` [OT] " Peter Humphrey
2024-11-15 8:41 ` [gentoo-user] Hollerith (was: Seagate hard drives with dual actuators) karl
@ 2024-11-15 9:51 ` Wols Lists
1 sibling, 0 replies; 38+ messages in thread
From: Wols Lists @ 2024-11-15 9:51 UTC (permalink / raw
To: gentoo-user
On 15/11/2024 00:18, Peter Humphrey wrote:
> In the 70s and 80s the national grid control centre in this country used three
> 2MB disks, any one or two of which could be online at any time. I can't tell
> you the platter size, but they were mounted in cabinets about 5' long, 3'6"
> tall and 2' wide. Each. Our training included servicing the air circulation
> system and its filters. I still remember the aluminium-alloy casings.
When I started in the early 80s, the company I worked for had those 19"
platters. We had one drive cabinet with one fixed and one removable
platter, 16MB each for a 32MB drive. And, iirc, an 80MB drive.
When we got one of those 300MB drives with "19 platters" disk packs, wow!
If anybody thinks those sizes don't quite make sense, I think the
platters actually stored 16MB/side, but one side per pack was used for
tracking data. So the 80BM drive had 3 physical platters (5 usable
sides), and the the 300MB had 10 physical platters (19 usable sides).
I got contracted out to an oil exploration company that had a Perkin
Elmer mainframe and a room of those 300 MB drives ...
And wrote a plotting program - in Fortran - that plotted rock
cross-sections. My programs always forced explicit declaration, but
stuck with the implicit conventions, so when I wanted variables "left"
and "right" I was wondering what to call them. I settled on SINIST and
DEXTER, and of course, when my boss at the client looked at my code he'd
studied latin ... those were the days ...
> Great things be could achieved with assembler and first-class people.
> More efficient than just throwing money at a problem until it gives
> in.
The problem with that is that it forces everyone else to just throw
money at it too ... We're using Oracle and BigQuery at work, and oh my
god do I hanker for Pick. 32 users hammering a 1GB database, with a
MIPS3000 processor in the database server (the same chip that powered
the Playstation 1). Even when the hard disk was going 19 to the dozen,
the users never complained that it was slow - they might have remarked,
because it was noticeable, but not at the "go make a cup of tea" level...
> Indeed, the best. Mind you, nostalgia isn't what it used to be...
Of course...
Cheers,
Wol
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [gentoo-user] Seagate hard drives with dual actuators.
2024-11-15 5:53 ` Dale
@ 2024-11-15 10:09 ` Michael
2024-11-15 11:59 ` Dale
2024-11-15 10:38 ` Frank Steinmetzger
1 sibling, 1 reply; 38+ messages in thread
From: Michael @ 2024-11-15 10:09 UTC (permalink / raw
To: gentoo-user
[-- Attachment #1: Type: text/plain, Size: 778 bytes --]
On Friday 15 November 2024 05:53:53 GMT Dale wrote:
> The thing about my data, it's mostly large video files. If I were
> storing documents or something, then SSD or something would be a good
> option. Plus, I mostly write once, then it either sits there a while or
> gets read on occasion.
For a write once - read often use case, the SMR drives are a good solution.
They were designed for this purpose. Because of their shingled layers they
provide higher storage density than comparable CMR drives.
> I checked and I have some 56,000 videos. That
> doesn't include Youtube videos. This is also why I wanted to use that
> checksum script.
>
> I do wish there was a easy way to make columns work when we copy and
> paste into email. :/
>
> Dale
>
> :-) :-)
[-- Attachment #2: This is a digitally signed message part. --]
[-- Type: application/pgp-signature, Size: 833 bytes --]
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [gentoo-user] Seagate hard drives with dual actuators.
2024-11-15 5:53 ` Dale
2024-11-15 10:09 ` Michael
@ 2024-11-15 10:38 ` Frank Steinmetzger
2024-11-15 12:19 ` Dale
1 sibling, 1 reply; 38+ messages in thread
From: Frank Steinmetzger @ 2024-11-15 10:38 UTC (permalink / raw
To: gentoo-user
Am Freitag, 15. November 2024, 06:53:53 Mitteleuropäische Normalzeit schrieb
Dale:
> Rich Freeman wrote:
> > On Thu, Nov 14, 2024 at 6:10 PM Dale <rdalek1967@gmail.com> wrote:
> >> The biggest downside to the large drives available now, even if SMART
> >> tells you a drive is failing, you likely won't have time to copy the
> >> data over to a new drive before it fails. On a 18TB drive, using
> >> pvmove, it can take a long time to move data.
> >> […]
>
> I think I did some math on this once. I'm not positive on this and it
> could vary depending on system ability of moving data. I think about
> 8TB is as large as you want if you get a 24 hour notice from SMART and
> see that notice fairly quickly to act on. Anything beyond that and you
> may not have enough time to move data, if the data is even good still.
I have 6 TB drives in my NAS, good ol’ WD Reds from before SMR time. When I
scrub them, i.e. read the whole of their data out sequentially at 80 %
capacity (so effectively around 5 TB), it takes 10½ hours. Looks like your
math adds up. Maybe 10 or even 12 TB would also still work in that time
window. Recently I switched from ZFS’s Raid6 to Raid5 because of said 80 %
occupancy and I needed more space, but had neither any free slots left nor
wanted to buy new hardware. Fingers crossed …
> >> I don't even want to think what it would cost to put
> >> all my 100TBs or so on SSD or NVME drives. WOW!!!
> >
> > # kubectl rook-ceph ceph osd df class ssd
> > ID CLASS WEIGHT REWEIGHT SIZE RAW USE DATA OMAP
> > META AVAIL %USE VAR PGS STATUS
> >
> > 8 ssd 6.98630 1.00000 7.0 TiB 1.7 TiB 1.7 TiB 63 MiB 3.9
> >
> > GiB 5.3 TiB 24.66 1.04 179 up
> > […]
>
> I do wish there was a easy way to make columns work when we copy and
> paste into email. :/
For special cases like this I think we wouldn’t mind using HTML mail. Or
simply disable automatic wrapping and use long lines of text for the entire
message. The client can then decide where to wrap.
I know it’s like a religious debate whether to wrap at <80 columns (please
don’t start one here), but there is actually an automatism for this: if you
end the line with a space, you can still wrap you text statically at <80
columns, but the space tells the client that it may wrap here or not. I forgot
the name of it though, I learned about it in the mutt user group.
For me it’s easier: as I use vim in mutt. I usually let it do the wrapping for
me (including the mechanism I described). But I can disable wrapping on-the-
fly, so I can paste longer terminal output.
> Dale
>
> :-) :-)
--
Grüße | Greetings | Salut | Qapla’
Feet are fat, lumpy hands with short, stubby fingers.
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [gentoo-user] Seagate hard drives with dual actuators.
2024-11-15 10:09 ` Michael
@ 2024-11-15 11:59 ` Dale
2024-11-15 15:35 ` Michael
0 siblings, 1 reply; 38+ messages in thread
From: Dale @ 2024-11-15 11:59 UTC (permalink / raw
To: gentoo-user
[-- Attachment #1: Type: text/plain, Size: 2388 bytes --]
Michael wrote:
> On Friday 15 November 2024 05:53:53 GMT Dale wrote:
>
>> The thing about my data, it's mostly large video files. If I were
>> storing documents or something, then SSD or something would be a good
>> option. Plus, I mostly write once, then it either sits there a while or
>> gets read on occasion.
> For a write once - read often use case, the SMR drives are a good solution.
> They were designed for this purpose. Because of their shingled layers they
> provide higher storage density than comparable CMR drives.
>
>
True but I don't like when I'm told a write is done, it kinda isn't. I
recall a while back I reorganized some stuff, mostly renamed directories
but also moved some files. Some were Youtube videos. It took about 30
minutes to update the data on the SMR backup drive. The part I see
anyway. It sat there for a hour at least doing that bumpy thing before
it finally finished. I realize if I just turn the drive off, the data
is still there. Still, I don't like it appearing to be done when it
really is still working on it. Another thing, I may switch to RAID one
of these days. If I do, that drive isn't a good option.
When I update my backups, I start the one I do with my NAS setup first.
Then I start the home directory backup with the SMR drive. I then
backup everything else I backup on other drives. I do that so that I
can leave the SMR drive at least powered on while it does it's bumpy
thing and I do other backups. Quite often, the SMR drive is the last
one I put back in the safe. That bumpy thing can take quite a while at
times.
This is the drive model I'm looking at. ST18000NM000J I don't think it
has its own encryption or anything and is a drive that will work in my
situation. I don't know what the last bits mean tho but saw somewhere
that 000 means no encryption. Not sure on the J part on the end. I did
find where I bought other sizes of drives with the same bit on the end.
They seem to work fine, so far anyway. I order my drives from here:
https://serverpartdeals.com/
I've bought what falls into the used category and them have only a few
power up hours on them. A lot of the time, it is less than 10 hours. I
have seen 3 or 4 hours or less with SMART a few times.
Now to get my money saved up. Won't be long. Another $200 but lots of
data.
Dale
:-) :-)
[-- Attachment #2: Type: text/html, Size: 3360 bytes --]
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [gentoo-user] Seagate hard drives with dual actuators.
2024-11-15 10:38 ` Frank Steinmetzger
@ 2024-11-15 12:19 ` Dale
0 siblings, 0 replies; 38+ messages in thread
From: Dale @ 2024-11-15 12:19 UTC (permalink / raw
To: gentoo-user
Frank Steinmetzger wrote:
> Am Freitag, 15. November 2024, 06:53:53 Mitteleuropäische Normalzeit schrieb
> Dale:
>> Rich Freeman wrote:
>>> On Thu, Nov 14, 2024 at 6:10 PM Dale <rdalek1967@gmail.com> wrote:
>>>> The biggest downside to the large drives available now, even if SMART
>>>> tells you a drive is failing, you likely won't have time to copy the
>>>> data over to a new drive before it fails. On a 18TB drive, using
>>>> pvmove, it can take a long time to move data.
>>>> […]
>> I think I did some math on this once. I'm not positive on this and it
>> could vary depending on system ability of moving data. I think about
>> 8TB is as large as you want if you get a 24 hour notice from SMART and
>> see that notice fairly quickly to act on. Anything beyond that and you
>> may not have enough time to move data, if the data is even good still.
> I have 6 TB drives in my NAS, good ol’ WD Reds from before SMR time. When I
> scrub them, i.e. read the whole of their data out sequentially at 80 %
> capacity (so effectively around 5 TB), it takes 10½ hours. Looks like your
> math adds up. Maybe 10 or even 12 TB would also still work in that time
> window. Recently I switched from ZFS’s Raid6 to Raid5 because of said 80 %
> occupancy and I needed more space, but had neither any free slots left nor
> wanted to buy new hardware. Fingers crossed …
>
Given the notice could be sent when one is not looking and may not be
seen for a while, I'd error on smaller not larger. Some systems could
be faster than others as well. Since I use LVM so much, I time pvmove.
That is what I would use anyway. If one uses some other tool, it would
be good to time with that tool. It could be faster or slower. I
suspect the drive would likely be the limit but one needs to test to be
sure.
>>>> I don't even want to think what it would cost to put
>>>> all my 100TBs or so on SSD or NVME drives. WOW!!!
>>> # kubectl rook-ceph ceph osd df class ssd
>>> ID CLASS WEIGHT REWEIGHT SIZE RAW USE DATA OMAP
>>> META AVAIL %USE VAR PGS STATUS
>>>
>>> 8 ssd 6.98630 1.00000 7.0 TiB 1.7 TiB 1.7 TiB 63 MiB 3.9
>>>
>>> GiB 5.3 TiB 24.66 1.04 179 up
>>> […]
>> I do wish there was a easy way to make columns work when we copy and
>> paste into email. :/
> For special cases like this I think we wouldn’t mind using HTML mail. Or
> simply disable automatic wrapping and use long lines of text for the entire
> message. The client can then decide where to wrap.
>
> I know it’s like a religious debate whether to wrap at <80 columns (please
> don’t start one here), but there is actually an automatism for this: if you
> end the line with a space, you can still wrap you text statically at <80
> columns, but the space tells the client that it may wrap here or not. I forgot
> the name of it though, I learned about it in the mutt user group.
>
> For me it’s easier: as I use vim in mutt. I usually let it do the wrapping for
> me (including the mechanism I described). But I can disable wrapping on-the-
> fly, so I can paste longer terminal output.
>
>> Dale
>>
>> :-) :-)
I've tried to make things look right before and most of the time, it
either gets worse or does something really weird, it double wrapped
once. I never sent that double wrapped one but it was weird looking. I
have Seamonkey set to convert to text only to anything sent to a
@gentoo.org address. Even if I compose it in HTML, it will convert it
to text only when sending. I have thought of using screenshots,
images. The bad thing there, can't copy and paste it. Plus, some don't
like large emails and images can't make one larger than usual.
It seems one can't win for losing. LOL
Dale
:-) :-)
P. S. I feel like I am forgetting to mention something. My poor
brain. :-(
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [gentoo-user] Seagate hard drives with dual actuators.
2024-11-15 11:59 ` Dale
@ 2024-11-15 15:35 ` Michael
2024-11-15 16:36 ` Dale
2024-11-15 22:13 ` Rich Freeman
0 siblings, 2 replies; 38+ messages in thread
From: Michael @ 2024-11-15 15:35 UTC (permalink / raw
To: gentoo-user
[-- Attachment #1: Type: text/plain, Size: 3584 bytes --]
On Friday 15 November 2024 11:59:34 GMT Dale wrote:
> Michael wrote:
> > On Friday 15 November 2024 05:53:53 GMT Dale wrote:
> >> The thing about my data, it's mostly large video files. If I were
> >> storing documents or something, then SSD or something would be a good
> >> option. Plus, I mostly write once, then it either sits there a while or
> >> gets read on occasion.
> >
> > For a write once - read often use case, the SMR drives are a good
> > solution.
> > They were designed for this purpose. Because of their shingled layers
> > they
> > provide higher storage density than comparable CMR drives.
>
> True but I don't like when I'm told a write is done, it kinda isn't. I
> recall a while back I reorganized some stuff, mostly renamed directories
> but also moved some files. Some were Youtube videos. It took about 30
> minutes to update the data on the SMR backup drive. The part I see
> anyway.
Right there is your problem, "... SMR backup drive". SMRs are best suited to
sequential writes. With repeat random writes they go into a read-modify-write
cycle and slow down.
Consequently, they are well suited to storage of media files, archiving data
long term and such write-once read-often applications. They are not suited to
heavy transactional loads and frequently overwritten data.
> It sat there for a hour at least doing that bumpy thing before
> it finally finished. I realize if I just turn the drive off, the data
> is still there. Still, I don't like it appearing to be done when it
> really is still working on it.
SMR drives have to read a whole band of shingled tracks, modify the small
region where the data has changed and then write the whole band of tracks back
on the disk in one go. The onboard cache on drive managed SMRs (DM-SMR) is
meant to hide this from the OS by queuing up writes before writing them on the
disk in a sequential stream, but if you keep hammering it with many random
writes you will soon exhaust the onboard cache and performance then becomes
glacial.
Host managed SMRs (HM-SMR) require the OS and FS to be aware of the need for
sequential writes and manage submitted data sympathetically to this limitation
of the SMR drive, by queuing up random writes in batches and submitting these
as a sequential stream.
I understand the ext4-lazy option and some patches on btrfs have improved
performance of these filesystems on SMR drivers, but perhaps f2fs will perform
better? :-/
> Another thing, I may switch to RAID one
> of these days. If I do, that drive isn't a good option.
Ugh! RAID striping will combine shingled bands across drives. A random write
on one drive will cause other drives to read-modify-write bands. Whatever
speed benefit is meant to be derived from striping will be reversed. On a NAS
application, where many users could be accessing the storage simultaneously
trying to save their interwebs downloads, etc., the SMR performance will nose
dive.
> When I update my backups, I start the one I do with my NAS setup first.
> Then I start the home directory backup with the SMR drive. I then
> backup everything else I backup on other drives. I do that so that I
> can leave the SMR drive at least powered on while it does it's bumpy
> thing and I do other backups. Quite often, the SMR drive is the last
> one I put back in the safe. That bumpy thing can take quite a while at
> times.
Instead of using the SMR for your /home fs backup, you would do better if you
repurposed it for media files and document backups which do not change as
frequently.
[-- Attachment #2: This is a digitally signed message part. --]
[-- Type: application/pgp-signature, Size: 833 bytes --]
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [gentoo-user] Seagate hard drives with dual actuators.
2024-11-15 15:35 ` Michael
@ 2024-11-15 16:36 ` Dale
2024-11-15 22:13 ` Rich Freeman
1 sibling, 0 replies; 38+ messages in thread
From: Dale @ 2024-11-15 16:36 UTC (permalink / raw
To: gentoo-user
[-- Attachment #1: Type: text/plain, Size: 5495 bytes --]
Michael wrote:
> On Friday 15 November 2024 11:59:34 GMT Dale wrote:
>> Michael wrote:
>>> On Friday 15 November 2024 05:53:53 GMT Dale wrote:
>>>> The thing about my data, it's mostly large video files. If I were
>>>> storing documents or something, then SSD or something would be a good
>>>> option. Plus, I mostly write once, then it either sits there a while or
>>>> gets read on occasion.
>>> For a write once - read often use case, the SMR drives are a good
>>> solution.
>>> They were designed for this purpose. Because of their shingled layers
>>> they
>>> provide higher storage density than comparable CMR drives.
>> True but I don't like when I'm told a write is done, it kinda isn't. I
>> recall a while back I reorganized some stuff, mostly renamed directories
>> but also moved some files. Some were Youtube videos. It took about 30
>> minutes to update the data on the SMR backup drive. The part I see
>> anyway.
> Right there is your problem, "... SMR backup drive". SMRs are best suited to
> sequential writes. With repeat random writes they go into a read-modify-write
> cycle and slow down.
>
> Consequently, they are well suited to storage of media files, archiving data
> long term and such write-once read-often applications. They are not suited to
> heavy transactional loads and frequently overwritten data.
>
All true. This was mentioned by Rich I think way back when I started a
thread about this drive constantly bumping. I feel the heads moving is
what the bump is. Until then, I had no idea it was a SMR drive. I'd
never heard of them before.
>> It sat there for a hour at least doing that bumpy thing before
>> it finally finished. I realize if I just turn the drive off, the data
>> is still there. Still, I don't like it appearing to be done when it
>> really is still working on it.
> SMR drives have to read a whole band of shingled tracks, modify the small
> region where the data has changed and then write the whole band of tracks back
> on the disk in one go. The onboard cache on drive managed SMRs (DM-SMR) is
> meant to hide this from the OS by queuing up writes before writing them on the
> disk in a sequential stream, but if you keep hammering it with many random
> writes you will soon exhaust the onboard cache and performance then becomes
> glacial.
>
> Host managed SMRs (HM-SMR) require the OS and FS to be aware of the need for
> sequential writes and manage submitted data sympathetically to this limitation
> of the SMR drive, by queuing up random writes in batches and submitting these
> as a sequential stream.
>
> I understand the ext4-lazy option and some patches on btrfs have improved
> performance of these filesystems on SMR drivers, but perhaps f2fs will perform
> better? :-/
>
And that is exactly how it works. It is fast at first, what I see
anyway, but once that buffer/cache fills up, leap year. It slows by
half or more usually. The more that gets sent its way, the worse it
gets it seems, watching progress from rsync.
>> Another thing, I may switch to RAID one
>> of these days. If I do, that drive isn't a good option.
> Ugh! RAID striping will combine shingled bands across drives. A random write
> on one drive will cause other drives to read-modify-write bands. Whatever
> speed benefit is meant to be derived from striping will be reversed. On a NAS
> application, where many users could be accessing the storage simultaneously
> trying to save their interwebs downloads, etc., the SMR performance will nose
> dive.
>
I marked the drive itself with a marker that it is a SMR drive. I'd
never put that thing in a RAID setup or anything like RAID for that
matter. I really don't want it in a LVM setup either. It will always
run as a single drive and for nothing that I need to handle heavy writes
most of the time.
>> When I update my backups, I start the one I do with my NAS setup first.
>> Then I start the home directory backup with the SMR drive. I then
>> backup everything else I backup on other drives. I do that so that I
>> can leave the SMR drive at least powered on while it does it's bumpy
>> thing and I do other backups. Quite often, the SMR drive is the last
>> one I put back in the safe. That bumpy thing can take quite a while at
>> times.
> Instead of using the SMR for your /home fs backup, you would do better if you
> repurposed it for media files and document backups which do not change as
> frequently.
Well, usually my home backup has only small changes. Most of it is
config files or my emails. I do add new videos on occasion but they
stream pretty well to new spots. What made it hard that time tho, I
moved a lot of files around and renamed things, same as moving to the
file system I guess. That slowed things down a lot. I only use it
because I only do backups once a week and it is a nice sized drive with
plenty of room for home. Otherwise, I'd buy a better drive.
If I had known it was a SMR drive before I bought it, I would have
bought something else even if it cost a little more. That is one thing
I like about the company I buy from, they sell mostly drives that are
used in server type systems. When I asked, they said they don't stock
any SMR drives. They can special order them for a big customer but they
don't stock them. Plus, they have good deals and stand behind what they
sell too. ;-)
Now to figure out what I'm going to get into today.
Dale
:-) :-)
[-- Attachment #2: Type: text/html, Size: 8155 bytes --]
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [gentoo-user] Seagate hard drives with dual actuators.
2024-11-15 15:35 ` Michael
2024-11-15 16:36 ` Dale
@ 2024-11-15 22:13 ` Rich Freeman
2024-11-16 11:02 ` Michael
1 sibling, 1 reply; 38+ messages in thread
From: Rich Freeman @ 2024-11-15 22:13 UTC (permalink / raw
To: gentoo-user
On Fri, Nov 15, 2024 at 10:35 AM Michael <confabulate@kintzios.com> wrote:
>
> Host managed SMRs (HM-SMR) require the OS and FS to be aware of the need for
> sequential writes and manage submitted data sympathetically to this limitation
> of the SMR drive, by queuing up random writes in batches and submitting these
> as a sequential stream.
>
> I understand the ext4-lazy option and some patches on btrfs have improved
> performance of these filesystems on SMR drivers, but perhaps f2fs will perform
> better? :-/
IMO a host-managed solution is likely to be the only thing that will
work reliably. If the drive supports discard/trim MAYBE a dumber
drive might be able to be used with the right filesystem. Even if
you're doing "write-once" workloads any kind of metadata change is
going to cause random writes unless the filesystem was designed for
SMR. Ideally you'd store metadata on a non-SMR device, though it
isn't strictly necessary with a log-based approach.
If the SMR drive tries really hard to not look like an SMR drive and
doesn't support discard/trim then even an SMR-aware solution probably
won't be able to use it effectively. The drive is going to keep doing
read-before-write cycles to preserve data even if there is nothing
useful to preserve.
--
Rich
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [gentoo-user] Seagate hard drives with dual actuators.
2024-11-15 22:13 ` Rich Freeman
@ 2024-11-16 11:02 ` Michael
2024-11-16 14:36 ` Rich Freeman
0 siblings, 1 reply; 38+ messages in thread
From: Michael @ 2024-11-16 11:02 UTC (permalink / raw
To: gentoo-user
[-- Attachment #1: Type: text/plain, Size: 3287 bytes --]
On Friday 15 November 2024 22:13:27 GMT Rich Freeman wrote:
> On Fri, Nov 15, 2024 at 10:35 AM Michael <confabulate@kintzios.com> wrote:
> > Host managed SMRs (HM-SMR) require the OS and FS to be aware of the need
> > for sequential writes and manage submitted data sympathetically to this
> > limitation of the SMR drive, by queuing up random writes in batches and
> > submitting these as a sequential stream.
> >
> > I understand the ext4-lazy option and some patches on btrfs have improved
> > performance of these filesystems on SMR drivers, but perhaps f2fs will
> > perform better? :-/
>
> IMO a host-managed solution is likely to be the only thing that will
> work reliably.
I think HM-SMRs may be used in commercial applications and DM-SMRs fed to the
unaware retail consumer, not that we can tell without the OEMs providing this
rather vital piece of information. I assume (simplistically) with DM-SMRs the
discard-garbage collection is managed wholly by the onboard drive controller,
while with HM-SMRs the OS will signal the drive to start trimming when the
workload is low in order to distribute the timing overheads to the system's
idle time.
> If the drive supports discard/trim MAYBE a dumber
> drive might be able to be used with the right filesystem. Even if
> you're doing "write-once" workloads any kind of metadata change is
> going to cause random writes unless the filesystem was designed for
> SMR. Ideally you'd store metadata on a non-SMR device, though it
> isn't strictly necessary with a log-based approach.
I've considered the flash storage trim operation to be loosely like the SMR
behaviour of read-modify-write and similarly susceptible to write-
amplification, but with SSD the write speed is so much faster it doesn't cause
the same noticeable slowdown as with SMR.
> If the SMR drive tries really hard to not look like an SMR drive and
> doesn't support discard/trim then even an SMR-aware solution probably
> won't be able to use it effectively. The drive is going to keep doing
> read-before-write cycles to preserve data even if there is nothing
> useful to preserve.
>
> --
> Rich
Sure, trimming nudges by the OS won't have an effect on the DM-SMR drive since
small changes to shingled bands are managed internally by the onboard STL
controller and the persistent SMR cache. However, there are some things a
filesystem driver can still provide to improve performance. For example,
compression, caching-batching to facilitate sequential writing directly in
whole bands rather than random small regions within a band, etc. In the case
of ext4 'lazy writeback journaling' the allocated journaling space is
increased upfront and the metadata are written once only with a mapping
inserted in memory to join the location of the metadata in the journal itself
instead of writing it twice. The two links at the bottom of this page explain
it better:
https://www.usenix.org/conference/fast17/technical-sessions/presentation/
aghayev
I'm not sure if this happens automatically, or if some mkfs and mount
option(s) need to be called upon. With up to 30x faster write speed, perhaps
this is something Dale may want to investigate and experiment with.
[-- Attachment #2: This is a digitally signed message part. --]
[-- Type: application/pgp-signature, Size: 833 bytes --]
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [gentoo-user] Seagate hard drives with dual actuators.
2024-11-16 11:02 ` Michael
@ 2024-11-16 14:36 ` Rich Freeman
2024-11-16 19:47 ` Michael
0 siblings, 1 reply; 38+ messages in thread
From: Rich Freeman @ 2024-11-16 14:36 UTC (permalink / raw
To: gentoo-user
On Sat, Nov 16, 2024 at 6:02 AM Michael <confabulate@kintzios.com> wrote:
>
> I assume (simplistically) with DM-SMRs the
> discard-garbage collection is managed wholly by the onboard drive controller,
> while with HM-SMRs the OS will signal the drive to start trimming when the
> workload is low in order to distribute the timing overheads to the system's
> idle time.
I'll admit I haven't looked into the details as I have no need for SMR
and there aren't any good FOSS solutions for using it that I'm aware
of (just a few that might be slightly less terrible). However, this
doesn't seem correct for two reasons:
First, I'm not sure why HM-SMR would even need a discard function.
The discard command is used to tell a drive that a block is safe to
overwrite without preservation. A host-managed SMR drive doesn't need
to know what data is disposable and what data is not. It simply needs
to write data when the host instructs it to do so, destroying other
data in the process, and it is the host's job to not destroy anything
it cares about. If a write requires a prior read, then the host needs
to first do the read, then adjust the written data appropriately so
that nothing is lost.
Second, there is no reason that any drive of any kind (SMR or SSD)
NEEDS to do discard/trim operations when the drive is idle, because
discard/trim is entirely a metadata operation that doesn't require IO
with the drive data itself. Now, some drives might CHOOSE to
implement it that way, but they don't have to. On an SSD, a discard
command does not mean that the drive needs to erase or move any data
at all. It just means that if there is a subsequent erase that would
impact that block, it isn't necessary to first read the data and
re-write it afterwards. A discard could be implemented entirely in
non-volatile metadata storage, such as with a bitmap. For a DM-SMR
using flash for this purpose would make a lot of sense - you wouldn't
need much of it.
This is probably why you have endless arguing online about whether
discard/trim is helpful for SSDs. It completely depends on how the
drive implements the command. The drives I've owned can discard
blocks without any impact on IO, but I've heard some have a terrible
impact on IO. It is just like how you can complete the same sort
operation in seconds or hours depending on how dumb your sorting
algorithm is.
In any case, to really take advantage of SMR the OS needs to
understand exactly how to structure its writes so as to not take a
penalty, and that requires information about the implementation of the
storage that isn't visible in a DM-SMR. Sure, some designs will do
better on SMR even without this information, but I don't think they'll
ever be all that efficient. It is no different from putting f2fs on a
flash drive with a brain-dead discard implementation - even if the OS
does all its discards in nice consolidated contiguous operations it
doesn't mean that the drive will handle that in milliseconds instead
of just blocking all IO for an hour - sure, the drive COULD do the
operation quickly, but that doesn't mean that the firmware designers
didn't just ignore the simplest use case in favor of just optimizing
around the assumption that NTFS is the only filesystem in the world.
--
Rich
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [gentoo-user] Seagate hard drives with dual actuators.
2024-11-16 14:36 ` Rich Freeman
@ 2024-11-16 19:47 ` Michael
2024-11-16 20:13 ` Rich Freeman
0 siblings, 1 reply; 38+ messages in thread
From: Michael @ 2024-11-16 19:47 UTC (permalink / raw
To: gentoo-user
[-- Attachment #1: Type: text/plain, Size: 6391 bytes --]
On Saturday 16 November 2024 14:36:02 GMT Rich Freeman wrote:
> On Sat, Nov 16, 2024 at 6:02 AM Michael <confabulate@kintzios.com> wrote:
> > I assume (simplistically) with DM-SMRs the
> > discard-garbage collection is managed wholly by the onboard drive
> > controller, while with HM-SMRs the OS will signal the drive to start
> > trimming when the workload is low in order to distribute the timing
> > overheads to the system's idle time.
>
> I'll admit I haven't looked into the details as I have no need for SMR
> and there aren't any good FOSS solutions for using it that I'm aware
> of (just a few that might be slightly less terrible). However, this
> doesn't seem correct for two reasons:
>
> First, I'm not sure why HM-SMR would even need a discard function.
> The discard command is used to tell a drive that a block is safe to
> overwrite without preservation. A host-managed SMR drive doesn't need
> to know what data is disposable and what data is not. It simply needs
> to write data when the host instructs it to do so, destroying other
> data in the process, and it is the host's job to not destroy anything
> it cares about. If a write requires a prior read, then the host needs
> to first do the read, then adjust the written data appropriately so
> that nothing is lost.
As I understand it from reading various articles, the constraint of having to
write sequentially a whole band when a random block changes within a band
works the same on both HM-SMR and the more common DM-SMR. What differs with
HM-SMR instructions is the host is meant to take over the management of random
writes and submit these as sequential whole band streams to the drive to be
committed without a read-modify-write penalty. I suppose for the host to have
to read the whole band first from the drive, modify it and then submit it to
the drive to write it as a whole band will be faster than letting the drive
manage this operation internally and getting its internal cache full. This
will not absolve the drive firmware from having to manage its own trim
operations and the impact metadata changes could have on the drive, but some
timing optimisation is perhaps reasonable. I can't recall where I read this
bit - perhaps some presentation on XFS or ext4 - not sure.
> Second, there is no reason that any drive of any kind (SMR or SSD)
> NEEDS to do discard/trim operations when the drive is idle, because
> discard/trim is entirely a metadata operation that doesn't require IO
> with the drive data itself. Now, some drives might CHOOSE to
> implement it that way, but they don't have to. On an SSD, a discard
> command does not mean that the drive needs to erase or move any data
> at all. It just means that if there is a subsequent erase that would
> impact that block, it isn't necessary to first read the data and
> re-write it afterwards. A discard could be implemented entirely in
> non-volatile metadata storage, such as with a bitmap. For a DM-SMR
> using flash for this purpose would make a lot of sense - you wouldn't
> need much of it.
I don't know if SMRs use flash to record their STL status and data allocation
between their persistent cache and shingled storage space. I would think yes,
or at least they ought to. Without metadata written to different media, for
such a small random write to take place atomically a whole SMR band will be
read, modified in memory, written to a new temporary location and finally
overwrite the original SMR band.
> This is probably why you have endless arguing online about whether
> discard/trim is helpful for SSDs. It completely depends on how the
> drive implements the command. The drives I've owned can discard
> blocks without any impact on IO, but I've heard some have a terrible
> impact on IO. It is just like how you can complete the same sort
> operation in seconds or hours depending on how dumb your sorting
> algorithm is.
I have an old OCZ which would increase IO latency to many seconds if not
minutes whenever trim was running, to the point where users started
complaining I had 'broken' their PC. As if I would do such a thing. LOL!
Never mind trying to write anything, reading from the disk would take ages and
the drive IO LED on the case stayed on for many minutes while TRIM was
running. I reformatted with btrfs, overprovisioned enough spare capacity and
reduced the cron job for trim to once a month, which stopped them complaining.
I don't know if the firmware was trying to write zeros to the drive
deterministically, instead of just de-allocating the trimmed blocks.
> In any case, to really take advantage of SMR the OS needs to
> understand exactly how to structure its writes so as to not take a
> penalty, and that requires information about the implementation of the
> storage that isn't visible in a DM-SMR.
Yes, I think all the OS can do is seek to minimise random writes and from what
I read a SMR-friendlier fs will try to do this.
> Sure, some designs will do
> better on SMR even without this information, but I don't think they'll
> ever be all that efficient. It is no different from putting f2fs on a
> flash drive with a brain-dead discard implementation - even if the OS
> does all its discards in nice consolidated contiguous operations it
> doesn't mean that the drive will handle that in milliseconds instead
> of just blocking all IO for an hour - sure, the drive COULD do the
> operation quickly, but that doesn't mean that the firmware designers
> didn't just ignore the simplest use case in favor of just optimizing
> around the assumption that NTFS is the only filesystem in the world.
For all I know consumer grade USB sticks with their cheap controller chips use
no wear levelling at all:
https://support-en.sandisk.com/app/answers/detailweb/a_id/25185/~/learn-about-trim-support-for-usb-flash%2C-memory-cards%2C-and-ssd-on-windows-and
Consequently, all flash friendly fs can do is perhaps compress and write in
batched mode to minimise write ops.
I can see where an SMR drive would be a suitable solution for storing media
files, but I don't know if the shingled bands would cause leakage due to their
proximity and eventually start losing data. I haven't seen any reliability
reports on this technology.
[-- Attachment #2: This is a digitally signed message part. --]
[-- Type: application/pgp-signature, Size: 833 bytes --]
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [gentoo-user] Seagate hard drives with dual actuators.
2024-11-16 19:47 ` Michael
@ 2024-11-16 20:13 ` Rich Freeman
2024-11-16 23:21 ` Wol
2024-11-17 11:22 ` Michael
0 siblings, 2 replies; 38+ messages in thread
From: Rich Freeman @ 2024-11-16 20:13 UTC (permalink / raw
To: gentoo-user
On Sat, Nov 16, 2024 at 2:47 PM Michael <confabulate@kintzios.com> wrote:
>
> As I understand it from reading various articles, the constraint of having to
> write sequentially a whole band when a random block changes within a band
> works the same on both HM-SMR and the more common DM-SMR.
Correct.
> What differs with
> HM-SMR instructions is the host is meant to take over the management of random
> writes and submit these as sequential whole band streams to the drive to be
> committed without a read-modify-write penalty. I suppose for the host to have
> to read the whole band first from the drive, modify it and then submit it to
> the drive to write it as a whole band will be faster than letting the drive
> manage this operation internally and getting its internal cache full.
I doubt this would be any faster with a host-managed drive. The same
pattern of writes is going to incur the same penalties.
The idea of a host-managed drive is to avoid the random writes in the
first place, and the need to do the random reads. For this to work
the host has to know where the boundaries of the various regions are
and where it is safe to begin writes in each region.
Sure, a host could just use software to make the host-managed drive
behave the same as a drive-managed drive, but there isn't much benefit
there. You'd want to use a log-based storage system/etc to just avoid
the random writes entirely. You might not even want to use a POSIX
filesystem on it.
> This
> will not absolve the drive firmware from having to manage its own trim
> operations and the impact metadata changes could have on the drive, but some
> timing optimisation is perhaps reasonable.
Why would a host-managed SMR drive have ANY trim operations? What
does trimming even mean on a host-managed drive?
Trimming is the act of telling the drive that it is safe to delete a
block without preserving it. A host-managed drive shouldn't need to
be concerned with preserving any data during a write operation. If it
is told to write something, it will just overwrite the data in the
subsequent overlapping cylinders.
Trimming is helpful with drive-managed SMR because if the drive isn't
told to trim a block that is about to be overwritten due to a write to
a different block, then the drive needs to first read, and then
rewrite the block. Trimming tells the drive that this step can be
skipped, assuming ALL the blocks in that region have been trimmed.
> I don't know if SMRs use flash to record their STL status and data allocation
> between their persistent cache and shingled storage space. I would think yes,
> or at least they ought to. Without metadata written to different media, for
> such a small random write to take place atomically a whole SMR band will be
> read, modified in memory, written to a new temporary location and finally
> overwrite the original SMR band.
Well, drive-managed SMR drives typically have CMR regions for data
caching, and they could also be used to store the bitmap. Cheap
drives might not support trim at all, and would just preserve all data
on write. After all, it isn't performance that is driving the
decision to sneak SMR into consumer drives. Flash would be the most
sensible way to do it though.
--
Rich
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [gentoo-user] Seagate hard drives with dual actuators.
2024-11-16 20:13 ` Rich Freeman
@ 2024-11-16 23:21 ` Wol
2024-11-17 11:22 ` Michael
1 sibling, 0 replies; 38+ messages in thread
From: Wol @ 2024-11-16 23:21 UTC (permalink / raw
To: gentoo-user
On 16/11/2024 20:13, Rich Freeman wrote:
> Well, drive-managed SMR drives typically have CMR regions for data
> caching, and they could also be used to store the bitmap. Cheap
> drives might not support trim at all, and would just preserve all data
> on write. After all, it isn't performance that is driving the
> decision to sneak SMR into consumer drives. Flash would be the most
> sensible way to do it though.
I would have thought the best way for a host-managed drive to avoid
masses of read-write would simply be to stream the data (files) to an
SMR region, and the metadata (directory structure) to SMR.
That way, if you store the block list in the directory, you just drop
data blocks, and if you keep track of which directory contents are
stored in which SMR block, you can simply recover the space by copying
the directory(ies) to a new block.
Cheers,
Wol
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [gentoo-user] Seagate hard drives with dual actuators.
2024-11-16 20:13 ` Rich Freeman
2024-11-16 23:21 ` Wol
@ 2024-11-17 11:22 ` Michael
2024-11-17 21:26 ` Rich Freeman
1 sibling, 1 reply; 38+ messages in thread
From: Michael @ 2024-11-17 11:22 UTC (permalink / raw
To: gentoo-user
[-- Attachment #1: Type: text/plain, Size: 3207 bytes --]
On Saturday 16 November 2024 20:13:30 GMT Rich Freeman wrote:
> On Sat, Nov 16, 2024 at 2:47 PM Michael <confabulate@kintzios.com> wrote:
> > What differs with
> > HM-SMR instructions is the host is meant to take over the management of
> > random writes and submit these as sequential whole band streams to the
> > drive to be committed without a read-modify-write penalty. I suppose for
> > the host to have to read the whole band first from the drive, modify it
> > and then submit it to the drive to write it as a whole band will be
> > faster than letting the drive manage this operation internally and
> > getting its internal cache full.
>
> I doubt this would be any faster with a host-managed drive. The same
> pattern of writes is going to incur the same penalties.
>
> The idea of a host-managed drive is to avoid the random writes in the
> first place, and the need to do the random reads. For this to work
> the host has to know where the boundaries of the various regions are
> and where it is safe to begin writes in each region.
The random reads do not incur a time penalty, it is the R-M-W ops that cost
time. The host don't need to know where bands start and finish, only needs to
submit data in whole sequential streams, so they can be written directly to
the disk as in a CMR. As long as data and metadata are submitted and written
directly, the SMR would be alike a CMR in terms of its performance.
> Sure, a host could just use software to make the host-managed drive
> behave the same as a drive-managed drive, but there isn't much benefit
> there. You'd want to use a log-based storage system/etc to just avoid
> the random writes entirely. You might not even want to use a POSIX
> filesystem on it.
>
> > This
> > will not absolve the drive firmware from having to manage its own trim
> > operations and the impact metadata changes could have on the drive, but
> > some timing optimisation is perhaps reasonable.
>
> Why would a host-managed SMR drive have ANY trim operations? What
> does trimming even mean on a host-managed drive?
>
> Trimming is the act of telling the drive that it is safe to delete a
> block without preserving it. A host-managed drive shouldn't need to
> be concerned with preserving any data during a write operation. If it
> is told to write something, it will just overwrite the data in the
> subsequent overlapping cylinders.
I assumed, may be wrongly, there is still an STL function performed by the
controller on HM-SMRs, to de-allocate deleted data bands whenever files are
deleted, perform secure data deletions via its firmware, etc. However, I can
see if this is managed at the fs journal layer the drive controller could be
dumb in this respect. Perhaps what I had read referred to HM-SMR 'aware'
drives, which may behave as DM-SMRs depending on the OS capability.
It would be interesting to see how different fs types perform on DM-SMRs.
Looking at used drives on ebay they are rather still rather pricey for me to
splash out on one of them, but since my PVR 14 year old WD SATA 2 refuses to
die they may get cheaper by the time I need a replacement.
[-- Attachment #2: This is a digitally signed message part. --]
[-- Type: application/pgp-signature, Size: 833 bytes --]
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [gentoo-user] Seagate hard drives with dual actuators.
2024-11-17 11:22 ` Michael
@ 2024-11-17 21:26 ` Rich Freeman
2024-11-17 23:04 ` Jack
0 siblings, 1 reply; 38+ messages in thread
From: Rich Freeman @ 2024-11-17 21:26 UTC (permalink / raw
To: gentoo-user
On Sun, Nov 17, 2024 at 6:22 AM Michael <confabulate@kintzios.com> wrote:
>
> On Saturday 16 November 2024 20:13:30 GMT Rich Freeman wrote:
> >
> > The idea of a host-managed drive is to avoid the random writes in the
> > first place, and the need to do the random reads. For this to work
> > the host has to know where the boundaries of the various regions are
> > and where it is safe to begin writes in each region.
>
> The random reads do not incur a time penalty, it is the R-M-W ops that cost
> time.
We're saying the same thing. If you don't preserve data that you
overwrite, then there is no need to read anything. Random reads are
the same speed on CMR and SMR, but not doing a read is faster than
doing a read on either platform, and any read on an HDD is very slow.
> The host don't need to know where bands start and finish, only needs to
> submit data in whole sequential streams, so they can be written directly to
> the disk as in a CMR. As long as data and metadata are submitted and written
> directly, the SMR would be alike a CMR in terms of its performance.
Again, we're saying the same thing, but making different assumptions
about how HM-SMR is implemented.
SMR can be appended to without penalty, just like tape. In order to
append and not overwrite, the host needs to know where the boundaries
of the SMR domains are.
> I assumed, may be wrongly, there is still an STL function performed by the
> controller on HM-SMRs, to de-allocate deleted data bands whenever files are
> deleted, perform secure data deletions via its firmware, etc. However, I can
> see if this is managed at the fs journal layer the drive controller could be
> dumb in this respect.
Honestly, I don't know exactly what commands an HM-SMR implements, and
since I doubt I'll ever use one, I can't really be bothered to look
them up. The whole point of a HM-SMR drive is that the drive just
does exactly what the host does, and doesn't try to shield the host
from the details of how SMR works. That's why they can be used
without performance penalties. They're just very destructive to data
if they aren't used correctly.
> It would be interesting to see how different fs types perform on DM-SMRs.
Not that interesting, for me personally. That's like asking how well
different filesystems would perform on tape. If I'm storing data on
tape, I'll use an algorithm designed to work on tape, and a tape drive
that actually has a command set that doesn't try to pretend that it is
useful for random writes. SMR is pretty analogous to tape, with the
benefit of being as fast as CMR for random reads.
If anything I've been trying to migrate away from HDD entirely. NVMe
will always be more expensive I'm sure but the density and endurance
are continuing to improve, and of course the speed is incomparable.
Cost is only a few times more. Biggest challenge is the lanes but
used workstations seem to be a way to get around that.
--
Rich
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [gentoo-user] Seagate hard drives with dual actuators.
2024-11-17 21:26 ` Rich Freeman
@ 2024-11-17 23:04 ` Jack
2024-11-18 0:23 ` Rich Freeman
0 siblings, 1 reply; 38+ messages in thread
From: Jack @ 2024-11-17 23:04 UTC (permalink / raw
To: gentoo-user
On 2024.11.17 16:26, Rich Freeman wrote:
> On Sun, Nov 17, 2024 at 6:22 AM Michael <confabulate@kintzios.com>
> wrote:
> > On Saturday 16 November 2024 20:13:30 GMT Rich Freeman wrote:
[snip ....]
> > It would be interesting to see how different fs types perform on
> DM-SMRs.
>
> Not that interesting, for me personally. That's like asking how well
> different filesystems would perform on tape. If I'm storing data on
> tape, I'll use an algorithm designed to work on tape, and a tape
> drive that actually has a command set that doesn't try to pretend
> that it is useful for random writes.
What about DEC-Tape? :-) (https://en.wikipedia.org/wiki/DECtape) (I
may even have a few left in a closet somewhere, if only I could find
someone to read them.)
> SMR is pretty analogous to tape, with the benefit of being as fast as
> CMR for random reads.
[snip ....]
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [gentoo-user] Seagate hard drives with dual actuators.
2024-11-17 23:04 ` Jack
@ 2024-11-18 0:23 ` Rich Freeman
2024-11-18 2:32 ` Matt Jolly
0 siblings, 1 reply; 38+ messages in thread
From: Rich Freeman @ 2024-11-18 0:23 UTC (permalink / raw
To: gentoo-user
On Sun, Nov 17, 2024 at 6:04 PM Jack <ostroffjh@users.sourceforge.net> wrote:
>
> What about DEC-Tape? :-) (https://en.wikipedia.org/wiki/DECtape) (I
> may even have a few left in a closet somewhere, if only I could find
> someone to read them.)
>
LTO is pretty much the only sensible choice these days as I understand
it. I've looked into it for backup but you need to store a LOT of
data for it to make sense. The issue is that the drives are just
super-expensive. You can get much older generation drives used for
reasonable prices, but then the tapes have a very low capacity but
they aren't that cheap, so your cost per TB is pretty high, and of
course you have the inconvenience of long backup times and lots of
tape changes. The newer generation drives are very reasonable in
terms of cost per TB, but the drives themselves cost thousands of
dollars. Unless you're archiving hundreds of TB it is cheaper to just
buy lots of USB3 hard drives at $15/TB, and then you get the random IO
performance as a bonus. The main downside to HDD at smaller scales is
that the drives themselves are more fragile, but that is mostly if
you're dropping them - in terms of storage conditions tape needs
better care than many appreciate for it to remain reliable.
For my offline onsite backups I just use a pair of USB3 drives on ZFS
right now. For actual storage I'm trying to buy U.2 NVMe for future
expansion, but I don't have a pressing need for that until HDDs die or
I need more space. Never makes sense to buy more than you need since
all this stuff gets cheaper with time...
--
Rich
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [gentoo-user] Seagate hard drives with dual actuators.
2024-11-18 0:23 ` Rich Freeman
@ 2024-11-18 2:32 ` Matt Jolly
0 siblings, 0 replies; 38+ messages in thread
From: Matt Jolly @ 2024-11-18 2:32 UTC (permalink / raw
To: gentoo-user
Hi,
> LTO is pretty much the only sensible choice these days as I understand
> it.
That's really the case, for bulk storage of any type you need to be able
to tier a lot of it offsite/offline. I'm responsible for a tape library
with a robot arm and about 13 drives raging from LTO7 through to LTO9.
> I've looked into it for backup but you need to store a LOT of
> data for it to make sense. The issue is that the drives are just
> super-expensive. You can get much older generation drives used for
> reasonable prices, but then the tapes have a very low capacity but
> they aren't that cheap, so your cost per TB is pretty high, and of
> course you have the inconvenience of long backup times and lots of
> tape changes.
The 7s are on their way out atm, so I'd expect to start seeing more pop
up for sale secondhand.
If you're doing tape though, 3:2:1 still applies, and you also (ideally)
want two different manufacturer drives writing to two different
manufacturer tapes to mitigate against issues like 'oh I got a firmware
update and my tape drive started writing garbage'.
> The newer generation drives are very reasonable in
> terms of cost per TB, but the drives themselves cost thousands of
> dollars. Unless you're archiving hundreds of TB it is cheaper to just
> buy lots of USB3 hard drives at $15/TB, and then you get the random IO
> performance as a bonus.
... but you have to swap out hundreds of USB drives which, especially
on the cheap side, are likely to be significantly less robust than
tape carts over time.
> The main downside to HDD at smaller scales is
> that the drives themselves are more fragile, but that is mostly if
> you're dropping them - in terms of storage conditions tape needs
> better care than many appreciate for it to remain reliable.
It's really all downsides at the home scale. Your best bet is often
ensuring that you have an offsite (s3 + restic can be decent)
backup of your essential data and otherwise assuming that everything
else could be lost at any time and just not caring about it.
Anecdotally rsync (or s3...) to a cheap Hetzner VPS can be cost
effective for smaller datasets / backups. My 100T server at
home though? If it dies I lose everything non-critical on it!
Cheers,
Matt
^ permalink raw reply [flat|nested] 38+ messages in thread
end of thread, other threads:[~2024-11-18 2:32 UTC | newest]
Thread overview: 38+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2024-11-13 23:10 [gentoo-user] Seagate hard drives with dual actuators Dale
2024-11-14 0:46 ` Matt Jolly
2024-11-14 13:05 ` Dale
2024-11-14 7:55 ` Wols Lists
2024-11-14 16:48 ` Dale
2024-11-15 0:18 ` [OT] " Peter Humphrey
2024-11-15 8:41 ` [gentoo-user] Hollerith (was: Seagate hard drives with dual actuators) karl
2024-11-15 9:51 ` [OT] Re: [gentoo-user] Seagate hard drives with dual actuators Wols Lists
2024-11-14 11:21 ` Michael
2024-11-14 17:00 ` Dale
2024-11-14 19:12 ` Michael
2024-11-14 19:51 ` Frank Steinmetzger
2024-11-14 19:55 ` Frank Steinmetzger
2024-11-14 23:14 ` Peter Humphrey
2024-11-14 20:33 ` Dale
2024-11-14 20:57 ` Rich Freeman
2024-11-14 23:10 ` Dale
2024-11-15 0:59 ` Rich Freeman
2024-11-15 5:53 ` Dale
2024-11-15 10:09 ` Michael
2024-11-15 11:59 ` Dale
2024-11-15 15:35 ` Michael
2024-11-15 16:36 ` Dale
2024-11-15 22:13 ` Rich Freeman
2024-11-16 11:02 ` Michael
2024-11-16 14:36 ` Rich Freeman
2024-11-16 19:47 ` Michael
2024-11-16 20:13 ` Rich Freeman
2024-11-16 23:21 ` Wol
2024-11-17 11:22 ` Michael
2024-11-17 21:26 ` Rich Freeman
2024-11-17 23:04 ` Jack
2024-11-18 0:23 ` Rich Freeman
2024-11-18 2:32 ` Matt Jolly
2024-11-15 10:38 ` Frank Steinmetzger
2024-11-15 12:19 ` Dale
2024-11-14 22:38 ` Wols Lists
2024-11-15 9:35 ` Michael
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox