I had an ADATA SSD die on me recently after less than 1 month power-on time with negligible write throughput. Tons of reallocated sectors and it failed by refusing any further writes (after having previously caused some unexplained FS corruption a day earlier). (thread)
-
Show this thread
-
At first I thought this was just a fluke/shitty SSD, but then I remembered something. As part of a hack, for the last 10 days of uptime or so I had a process writing a number to a file a few times per second. Maybe this caused the early death?
1 reply 0 retweets 2 likesShow this thread -
If the FTL was utterly terrible, that could've caused it to rewrite the same flash block every time. And after wearing out a block it would go to an alternate... and if those are limited (e.g. indexed by some kind of hash scheme), it would eventually run out of alternates.
1 reply 0 retweets 2 likesShow this thread -
Clearly a poor quality SSD implementation in that case... but I wonder if this is what actually happened, as opposed to just a random early failure dud. Perhaps SSD endurance tests should test this use case. Instead of writing over the entire drive, rewrite a single sector.
1 reply 0 retweets 2 likesShow this thread -
I bet we'd see wildly different results for different brands, and many deaths well before the specified total drive write endurance. Perhaps a good metric for how good the firmware/FTL algorithm is.
2 replies 0 retweets 7 likesShow this thread -
There are plenty of use cases that have this "rewrite one sector repeatedly" pattern - e.g. md-raid metadata includes an event count that basically increases every time the array is written to. I wonder if anyone else has seen this kind of early death and attributed it to this.
2 replies 0 retweets 12 likesShow this thread
ohhhh but wait till you hear my stories about spinning rust... 
Loading seems to be taking a while.
Twitter may be over capacity or experiencing a momentary hiccup. Try again or visit Twitter Status for more information.