Date: Fri, 19 May 2017 07:34:36 +0000 From: kc atgb <kisscoolandthegangbang@hotmail.fr> To: "freebsd-fs@freebsd.org" <freebsd-fs@freebsd.org> Subject: Re: Different size after zfs send receive Message-ID: <AMSPR05MB148026492ACA46AD8D2A15BA0E50@AMSPR05MB148.eurprd05.prod.outlook.com> In-Reply-To: <58A6B47B-2992-4BB8-A80E-44F74EAE93B2@longcount.org> References: <DBXPR05MB157C1956B267EA6BE59F570A0E40@DBXPR05MB157.eurprd05.prod.outlook.com> <58A6B47B-2992-4BB8-A80E-44F74EAE93B2@longcount.org>
next in thread | previous in thread | raw e-mail | index | archive | help
Le Thu, 18 May 2017 21:53:23 +0000,
Mark Saad <nonesuch@longcount.org> a =E9crit :
Hi,
I see what you are talking about I thing. You refer to "raid" splitting, ri=
ght ? In this case this is something in the "internals" of the raid system.=
Isn't
zfs list suppose to report raw data sizes (without metadata, checksums, ...=
) ?=20
I don't really think it is related to what I'm refering.=20
Look, for the same pool configuration (one 4 disks raidz1 vdev) with the sa=
me disks and the same data, it reports for storage/usrobj
5819085888 before backup and
5820359616 after restore to the recreated pool.=20
Even for pools with one single disk vdev (again same disks, same configurat=
ion, same data as above...) for the same dataset
5675081728 in backup1 disk and
5675188224 in backup2
The difference isn't so big but the numbers differ and I would imagine numb=
ers to be the same.=20
K.
> Hi kc=20
> This has to do with how data blocks are replicated when stored on a rai=
dzN . Moving them to a mirror removes replicated blocks . This is way over
> simplified but imagine you store a file of 10gb on a raidz1 . The system =
splits the file into smaller chunks; of say 1mb , and stores one extra chun=
k for
> each chunk that us striped around the raidz1 . Storing on a mirror is jus=
t write the chunk once on each disk . However with a mirror since you only =
see 1/2
> the number of disks you never see the extra chunks in the used field .=20
>=20
> Hope this helps .=20
>=20
> ---
> Mark Saad | nonesuch@longcount.org
>=20
> > On May 18, 2017, at 3:36 PM, kc atgb <kisscoolandthegangbang@hotmail.fr=
> wrote:
> >=20
> > Hi,
> >=20
> > Some days ago I had a need to backup my current pool and restore it aft=
er pool destroy and create.=20
> >=20
> > The pool in my home server is a raidz1 with 4 disks. To backup this poo=
l I grabbed two 4TB disks (single disk pools) to have a double backup (I ha=
ve just
> > one sata port left I can use to plug a disk).=20
> >=20
> > The whole process of backup and restore went well as I can say. But loo=
king at the size reported by zfs list make me a little bit curious.=20
> >=20
> > storage/datas/ISO =
35420869824 381747995136 35420726976 /datas/ISO
> > storage/datas/ISO@backup_send =
142848 - 35420726976 -
> > storage/datas/ISO@backup_sync =
0 - 35420726976 -
> >=20
> > b1/datas/ISO 354=
39308800 2176300351488 35439210496 /datas/ISO
> > b1/datas/ISO@backup_send =
98304 - 35439210496 -
> > b1/datas/ISO@backup_sync =
0 - 35439210496 -
> >=20
> > b2/datas/ISO 354=
39308800 2176298991616 35439210496 /datas/ISO
> > b2/datas/ISO@backup_send =
98304 - 35439210496 -
> > b2/datas/ISO@backup_sync =
0 - 35439210496 -
> >=20
> > storage/datas/ISO =
35421024576 381303470016 35420715072 /datas/ISO
> > storage/datas/ISO@backup_send =
142848 - 35420715072 -
> > storage/datas/ISO@backup_sync =
11904 - 35420715072 -
> >=20
> >=20
> > storage/usrobj =
5819085888 381747995136 5816276544 legacy
> > storage/usrobj@create =
166656 - 214272 -
> > storage/usrobj@backup_send =
2642688 - 5816228928 -
> > storage/usrobj@backup_sync =
0 - 5816276544 -
> >=20
> > b1/usrobj 56=
75081728 2176300351488 5673222144 legacy
> > b1/usrobj@create =
114688 - 147456 -
> > b1/usrobj@backup_send =
1744896 - 5673222144 -
> > b1/usrobj@backup_sync =
0 - 5673222144 -
> >=20
> > b2/usrobj 56=
75188224 2176298991616 5673328640 legacy
> > b2/usrobj@create =
114688 - 147456 -
> > b2/usrobj@backup_send =
1744896 - 5673328640 -
> > b2/usrobj@backup_sync =
0 - 5673328640 -
> >=20
> > storage/usrobj =
5820359616 381303470016 5815098048 legacy
> > storage/usrobj@create =
166656 - 214272 -
> > storage/usrobj@backup_send =
2535552 - 5815098048 -
> > storage/usrobj@backup_sync =
11904 - 5815098048 -
> >=20
> > As you can see the numbers are different for each pool (the initial rai=
dz1, backup1 disk, backup2 disk and new raidz1). I mean in the USED column.=
I have
> > nearly all my datasets in the same situation (those with fixed data tha=
t have not changed between the beginning of the process and now). backup1 a=
nd backup2
> > are identical disks with exactly the same configurations and have diffe=
rent numbers. I used the same commands for all my transfers except the name=
of the
> > destination pool.=20
> >=20
> > So, I wonder what can cause these differences ? Is it something I have =
to worry about ? Can I consider this as a normal behavior ?=20
> >=20
> > Thanks for your enlightments,
> > K.
> > _______________________________________________
> > freebsd-fs@freebsd.org mailing list
> > https://lists.freebsd.org/mailman/listinfo/freebsd-fs
> > To unsubscribe, send any mail to "freebsd-fs-unsubscribe@freebsd.org"
>=20
Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?AMSPR05MB148026492ACA46AD8D2A15BA0E50>
