1 |
Am Wed, Jun 30, 2021 at 09:00:29PM +0100 schrieb antlists: |
2 |
|
3 |
> > I reached 80 % usage (which is the recommended maximum for ZFS) and am |
4 |
> > now evaluating my options for the coming years. |
5 |
> > 1) Reduce use of space by re-encoding. My payload is mainly movies, among |
6 |
> > which are 3 TB of DVDs which can be shrunk by at least ⅔ by re-encoding. |
7 |
> > → this takes time and computing effort, but is a long-term goal anyway. |
8 |
> > 2) Replace all drives with bigger ones. There are three counter arguments: |
9 |
> > • 1000 € for four 10 TB drives (the biggest size available w/o helium) |
10 |
> > • they are only available with 7200 rpm (more power, noise and heat) |
11 |
> > • I am left with four perfectly fine 6 TB drives |
12 |
> > 3) Go for 4+2 RaidZ2. This requires a bigger case (with new PSU due to |
13 |
> > different form factor) and a SATA expansion card b/c the Mobo only has |
14 |
> > six connectors (I need at least one more for the system drive), costing |
15 |
> > 250 € plus drives. |
16 |
> > 4) Convert to RaidZ1. Gain space of one drive at the cost of resilience. I |
17 |
> > can live with the latter; the server only runs occasionally and not for |
18 |
> > very long at a time. *** This option brings me to my question above, |
19 |
> > because it is easy to achieve and costs no €€€. |
20 |
> > |
21 |
> 5) Dunno if this is possible but ... replace one 6TB by a 12TB (any reason |
22 |
> you don't like Helium?) |
23 |
|
24 |
It is technically impossible to keep it in. It will diffuse through the case |
25 |
eventually, because the atoms are smaller than in any other material. AFAIK, |
26 |
the drive will still work, but suffer reduced performance. But I’m not sure. |
27 |
And who knows how it will behave in a RAID once that happens. Big capacity |
28 |
drives need longer to rebuild, which increases the probability of failure |
29 |
during RAID rebuild. I think that’s why companies tend to stick to smaller |
30 |
drivers (2 TB or so). |
31 |
|
32 |
I also prefer slow spinning disks (power consumption and noise). |
33 |
There is exactly one HDD model with 10 TB, no helium and 5400 rpm (naturally |
34 |
non-SMR). It’s a “WD Red Desktop Mainstream Kit”, which in itself sounds |
35 |
like an oxymoron. |
36 |
|
37 |
> and raid-0 two of the remaining 6's together. Dunno anything about what |
38 |
> the raidZ's are but I presume this would give you 12TB of mirrored |
39 |
> storage. It would also only use 3 slots, |
40 |
|
41 |
The approach in itself sounds interesting – if I already had the drives. But |
42 |
I don’t, and so I’d have to pay 350 € to get 2 TB more effective storage, |
43 |
while still losing one level of redundancy. :-/ |
44 |
|
45 |
> so you could use the 4th for eg your videos, and back them up on external |
46 |
> storage ie the drive you've just removed :-) |
47 |
|
48 |
Unfortunately, I can come up with many reasons against this approach. |
49 |
- I don’t like the idea of using two different file systems for one single |
50 |
purpose, because sooner or later one will fill up and the data “spills |
51 |
over” to the other FS. Just the idea that this might happen bugs me and |
52 |
I’d always have to think of it when I copy something over, starting with |
53 |
the decision on which FS to use in the first place. ;-) |
54 |
- I don’t want to have to deal with making backups to external media. I’d |
55 |
have to hook them up regularly and maintain and run a backup tool. |
56 |
- I don’t want to rely on external storage in general. That’s what I bought |
57 |
the NAS. :o) |
58 |
|
59 |
I used external drives exclusively until I had the NAS. Over time, the |
60 |
biggest drive of the day became too small and I bought a bigger one as my |
61 |
wallet allowed (sounds like the story Dale told some time ago). The first |
62 |
had 500 G, then 1 T, then 3 T. Each one has its own power adapter, usually |
63 |
incompatible with anything else I have. It needs a reliable USB connection, |
64 |
yadda yadda. The files will inevitably become cluttered and dispersed and I |
65 |
need to keep tabs on what was where. No thanks. :) |
66 |
|
67 |
I do have a hot swap bay in my PC for bare drives¹. But SATA connectors |
68 |
aren’t made for many physical connect-disconnect cycles. |
69 |
|
70 |
¹ https://en.sharkoon.com/product//12640 |
71 |
|
72 |
> (The raid-0, I'd probably stripe rather than linear for performance.) |
73 |
|
74 |
When I did some research over the last days I read that ZFS distributes |
75 |
writes across all vdevs of a pool depending on their individual fill state. |
76 |
So one doesn’t really have control over linear vs. striped anyway. |
77 |
|
78 |
|
79 |
Dang, I wanted to go to bed 1½ hours ago. Instead I composed mails. :) |
80 |
|
81 |
-- |
82 |
Grüße | Greetings | Qapla’ |
83 |
Please do not share anything from, with or about me on any social network. |
84 |
|
85 |
There is so much sand in Northern Africa that if it were spread out over the |
86 |
globe it would completely cover the Sahara Desert. |