ZFS Install Guide

Revision as of 22:43, 18 March 2014 by Fearedbliss (Talk)

Disclaimers

  • This guide will only show you how to install Funtoo/ZFS on a 64 bit system!
  • This guide is also using new technologies such as bliss-boot/initramfs/kernel. If you would like to use boot-update and or genkernel, please check out the ZFS Install Guide's Addendum.
  • This guide is a work in progress.

Introduction

In this guide we will show you how to install Funtoo on Native ZFS. The tutorial is meant to be an "overlay" over the Regular Funtoo Installation. Follow the normal guide during parts that are omitted.

Introduction to ZFS

Since ZFS is a new technology for Linux, it can be helpful to understand some of its benefits, particularly in comparison to BTRFS, another popular next-generation Linux filesystem:

  • On Linux, the ZFS code can be updated independently of the kernel to obtain the latest fixes. btrfs is exclusive to Linux and you need to build the latest kernel sources to get the latest fixes.
  • ZFS is supported on multiple platforms. The platforms with the best support are Solaris, FreeBSD and Linux. Other platforms with varying degrees of support are NetBSD, Mac OS X and Windows. btrfs is exclusive to Linux.
  • ZFS has the Adaptive Replacement Cache replacement algorithm while btrfs uses the Linux kernel's Last Recently Used replacement algorithm. The former often has an overwhelmingly superior hit rate, which means fewer disk accesses.
  • ZFS has the ZFS Intent Log and SLOG devices, which accelerates small synchronous write performance.
  • ZFS handles internal fragmentation gracefully, such that you can fill it until 100%. Internal fragmentation in btrfs can make btrfs think it is full at 10%. Btrfs has no automatic rebalancing code, so it requires a manual rebalance to correct it.
  • ZFS has raidz, which is like RAID 5/6 (or a hypothetical RAID 7 that supports 3 parity disks), except it does not suffer from the RAID write hole issue thanks to its use of CoW and a variable stripe size. btrfs gained integrated RAID 5/6 functionality in Linux 3.9. However, its implementation uses a stripe cache that can only partially mitigate the effect of the RAID write hole.
  • ZFS send/receive implementation supports incremental update when doing backups. btrfs' send/receive implementation requires sending the entire snapshot.
  • ZFS supports data deduplication, which is a memory hog and only works well for specialized workloads. btrfs has no equivalent.
  • ZFS datasets have a hierarchical namespace while btrfs subvolumes have a flat namespace.
  • ZFS has the ability to create virtual block devices called zvols in its namespace. btrfs has no equivalent and must rely on the loop device for this functionality, which is cumbersome.

The only area where btrfs is ahead of ZFS is in the area of small file efficiency. btrfs supports a feature called block suballocation, which enables it to store small files far more efficiently than ZFS. It is possible to use another filesystem (e.g. reiserfs) on top of a ZFS zvol to obtain similar benefits (with arguably better data integrity) when dealing with many small files (e.g. the portage tree).

For a quick tour of ZFS and have a big picture of its common operations you can consult the page ZFS Fun.


Video Tutorial

As a companion to the installation instructions below, a YouTube video tutorial is now available:

Downloading the ISO (With ZFS)

In order for us to install Funtoo on ZFS, you will need an environment that already provides the ZFS tools. Therefore we will download a customized version of System Rescue CD with ZFS included.

Name: sysresccd-4.0.1_zfs_0.6.2.iso  (545 MB)
Release Date: 2014-02-25
md5sum 01f4e6929247d54db77ab7be4d156d85


Download System Rescue CD with ZFS

Creating a bootable USB from ISO (From a Linux Environment)

Now we will download the iso, plug in our flash drive, and install sysresccd into the drive using the default sysresccd install script:

Download the iso to your home directory
# cd ~/
# [Download ISO]

Make a temporary directory to mount loop the iso
# mkdir /tmp/loop

Mount the iso
# mount -o ro,loop ~/sysresccd-4.0.1_zfs_0.6.2.iso /tmp/loop

Run the usb installer
# /tmp/loop/usb_inst.sh

That should be all you need to do to get your flash drive working.

Warning

Make sure the USB you want to use is unmounted before running the script, or it will fail. The script will partition, and mount your usb to /mnt/backup automatically.

Booting the ISO

Warning

Select the "Alternate 64 bit kernel (altker64)" option in the System Rescue CD Boot Menu. The ZFS modules have been built specifically for this kernel rather than the standard kernel. If you select a different kernel, you will get a fail to load module stack error message.

Creating partitions

There are two ways to partition your disk: You can use your entire drive and let ZFS automatically partition it, or you can do it manually.

We will be showing you how to partition it manually because if you do so, you get to create your own layout, you get to have your own separate /boot partition (Which is nice since not every bootloader supports booting from ZFS pools), and you get to boot into RAID10, RAID5 (RAIDZ) pools and any other layouts without any hassles due to you having a separate /boot partition.

gdisk (GPT Style)

A Fresh Start:

First lets make sure that the disk is completely wiped from any previous disk labels and partitions. We will also assume that /dev/sda is the target drive.

# sgdisk -Z /dev/sda

Warning

This is a destructive operation and the program will not ask you for confirmation! Make sure you really don't want anything on this disk.

Now that we have a clean drive, we will create the new layout.

First open up the application:

# gdisk /dev/sda

Create Partition 1 (boot):

Command: n ↵
Partition Number: 
First sector: 
Last sector: +250M ↵
Hex Code: 

Create Partition 2 (BIOS Boot Partition):

Command: n ↵
Partition Number: 
First sector: 
Last sector: +32M ↵
Hex Code: EF02 ↵

Create Partition 3 (swap):

Command: n ↵
Partition Number: 
First sector: 
Last sector: +4G ↵
Hex Code: 8200 ↵

Note

A rule of thumb for swap size is either 2 x RAM or RAM + 1.

Create Partition 4 (ZFS):

Command: n ↵
Partition Number: 
First sector: 
Last sector: 
Hex Code: bf00 ↵

Command: p ↵

Number  Start (sector)    End (sector)  Size       Code  Name   
1            2048          514047   250.0 MiB   8300  Linux filesystem   
2          514048          579583   32.0 MiB    EF02  BIOS boot partition   
3          579584         8968191   4.0 GiB     8200  Linux swap   
4         8968192      1048575966   495.7 GiB   BF00  Solaris root

Command: w ↵

Format your /boot partition

# mkfs.ext2 -m 1 /dev/sda1

Encryption (Optional)

If you want encryption, then create your encrypted vault(s) now by doing the following:

# cryptsetup luksFormat /dev/sda4
# cryptsetup luksOpen /dev/sda4 vault_0

Warning

On some machines, a combination of ZFS and LUKS has caused instability and system crashes.

Create the zpool

We will first create the pool. The pool will be named `tank` and the disk will be aligned to 4K sectors:

# zpool create -f -o ashift=12 -o cachefile= -O compression=lz4 -O normalization=formD -m none -R /mnt/funtoo tank /dev/sda4

Important

If you are using encrypted root, change /dev/sda4 to /dev/mapper/vault_0.


Important

If you have a previous pool that you would like to import, you can do:

# zpool import -f -o cachefile= -R /mnt/funtoo <pool_name>

Create the zfs datasets

We will now create some datasets. For this installation, we will create a small but future proof amount of datasets. We will have a dataset for the OS (/). We will also show you how to create some optional datasets: /home, /var, /usr/src, and /usr/portage.

Create some empty containers for organization purposes, and make the dataset that will hold /
# zfs create tank/funtoo
# zfs create -o mountpoint=/ tank/funtoo/root

Optional, but recommended datasets: /home
# zfs create -o mountpoint=/home tank/funtoo/home

Optional datasets: /usr/src, /usr/portage/
# zfs create -o mountpoint=/usr/src tank/funtoo/src
# zfs create -o mountpoint=/usr/portage -o compression=off tank/funtoo/portage


Format your swap

# mkswap -f /dev/sda3
# swapon /dev/sda3

Now we will continue to install funtoo.

Installing Funtoo

Pre-Chroot

Go into the directory that you will chroot into
# cd /mnt/funtoo

Make a boot folder and mount your boot drive
# mkdir boot
# mount /dev/sda1 boot

Now download and extract the Funtoo stage3 ...

Once you've extracted the stage3, do a few more preparations and chroot into your new funtoo environment:

Bind the kernel related directories
# mount -t proc none proc
# mount --rbind /dev dev
# mount --rbind /sys sys

Copy network settings
# cp -f /etc/resolv.conf etc

Make the zfs folder in 'etc' and copy your zpool.cache
# mkdir etc/zfs
# cp /etc/zfs/zpool.cache etc/zfs

Chroot into Funtoo
# env -i HOME=/root TERM=$TERM chroot . bash -l

In Chroot

Create a symbolic link to your mountpoints
# ln -sf /proc/mounts /etc/mtab

Sync your tree
# emerge --sync

Add filesystems to /etc/fstab

Before we install our kernel in the next section, we will edit the /etc/fstab file and make sure our /boot line is correct. This is because portage needs to know where our /boot is so that it can install the files there.

Edit /etc/fstab:

# <fs>                  <mountpoint>    <type>          <opts>          <dump/pass>

/dev/sda1               /boot           ext2            defaults        0 2
/dev/sda3               none            swap            sw              0 0

Kernel Configuration

To speed this step up, we will install a pre-configured/compiled kernel called bliss-kernel. This kernel already has the correct configurations for ZFS and a variety of other scenarios. It's a vanilla kernel from kernel.org without any external patches.

To install sys-kernel/bliss-kernel type the following:

# emerge bliss-kernel

Now make sure that your /usr/src/linux symlink is pointing to this kernel by typing the following:

# eselect kernel list
Available kernel symlink targets:
[1]   linux-3.12.13-KS.02 *

You should see a star next to the version you installed. In this case it was 3.12.13-KS.02. If it's not set, you can type eselect kernel set #.

Installing the ZFS userspace tools and kernel modules

Emerge sys-fs/zfs. This package will bring in sys-kernel/spl, and sys-fs/zfs-kmod as its dependencies:

# emerge zfs

Check to make sure that the zfs tools are working. The zpool.cache file that you copied before should be displayed.

# zpool status
# zfs list

If everything worked, continue.

Installing & Configuring the Bootloader

First we will install GRUB 2:

# emerge grub

Now install grub to the drive itself (not a partition):

# grub-install /dev/sda

You should receive the following message:

Installation finished. No error reported.

Checking your /boot folder should reveal a fully loaded grub directory:

# ls -l /boot/grub
total 2520
-rw-r--r-- 1 root root    1024 Jan  4 16:09 grubenv
drwxr-xr-x 2 root root    8192 Jan 12 14:29 i386-pc
drwxr-xr-x 2 root root    4096 Jan 12 14:28 locale
-rw-r--r-- 1 root root 2555597 Feb  4 11:50 unifont.pf2

Automatic bootloader configuration with bliss-boot

This is a new program that is designed to generate a simple, human-readable/editable, configuration file for a variety of bootloaders. It currently supports grub2, extlinux, and lilo. For the purposes of this guide, we will show you how to generate a GRUB 2 configuration.

First install it via the following command:

# emerge bliss-boot

In order to generate our bootloader configuration file, we will first configure bliss-boot so that it knows what we want. The 'bliss-boot' configuration file is located in /etc/bliss-boot/conf.py. Open that file and make sure that the following variables are set appropriately:

# This should be set to the bootloader you installed earlier: (grub2, extlinux, and lilo are the available options)
bootloader = "grub2"

# This should be set to the kernel you installed earlier
default = "3.12.13-KS.02" 

Scroll all the way down until you find 'kernels'. You will need to add the kernels and the options you want for these kernels.

For normal zfs installations

kernel = {
    '3.12.13-KS.02' : 'root=tank/funtoo/root quiet',
}

For encrypted zfs installations

If you are using encryption then you would let the initramfs know:

  • How do you want to decrypt the drive? (enc_type=)
  • pass = will ask for passphrase directly
  • key = a plain unencrypted key file
  • key_gpg = an encrypted key file
  • Where is the encrypted drive?" (enc_root=)
  • Where is the root pool after it has been decrypted?" (root=)
kernel = {
    '3.12.13-KS.02' : 'root=tank/funtoo/root enc_root=/dev/sda4 enc_type=pass quiet',
}

Warning

Add the 'refresh' option to the end of the line the first time you boot your system. This is because ZFS's cache might have the wrong encryption vault_#s. Once a new zpool.cache is generated, and your system is booted, regenerate your initramfs, and remove the 'refresh' line.

Generate the configuration

Now that we have configure our /etc/bliss-boot/conf.py file, we can generate our config. Simply run the following command:

# bliss-boot

Important

For safety reasons, bliss-boot doesn't automatically overwrite /boot/grub/grub.cfg but rather generates the configuration file in your current directory. After you check the file, move it to /boot/grub/grub.cfg.

Create the initramfs

If you are encrypting your drives, then add the "luks" use flag to your package.use before emerging:

# echo "sys-kernel/bliss-initramfs luks" >> /etc/portage/package.use

Now install the program and run it:

# emerge bliss-initramfs

You can either run it without any parameters to get an interactive menu
or you can pass the parameters directly. 1 = zfs, 6 = encrypted zfs, and the kernel name.
# bliss-initramfs 1 3.12.13-KS.02

Your file will be in the current directory.

Moving the initrd into the correct location

Place the generated file into your /boot/kernels/3.12.13-KS.02/ folder. For bliss-boot, the file needs to be called 'initrd' rather than 'initrd-3.12.13-KS.02'.

# mv initrd-3.12.13-KS.02 /boot/kernels/3.12.13-KS.02/initrd

Final configuration

Add the zfs tools to openrc

# rc-config add zfs boot

Clean up and reboot

We are almost done, just clean up, set our root password, and unmount whatever we mounted and get out:

Delete the stage3 tarball that you downloaded earlier so it doesn't take up space.
# cd /
# rm stage3-latest.tar.xz

Set your root password
# passwd
>> Enter your password, you won't see what you are writing (for security reasons), but it is there!

Get out of the chroot environment
# exit

Unmount all the kernel filesystem stuff and boot (if you have a separate /boot)
# umount -l proc dev sys boot

Turn off the swap
# swapoff /dev/sda3

Reboot
# reboot

Warning

Do not export the zpool. Just reboot. ZFS's zpool.cache file is shared between reboots and exporting the pool will cause problems. Exporting the pool is only meant when you want to move the zpool from one machine to another.

Important

Don't forget to set your root password as stated above before exiting chroot and rebooting. If you don't set the root password, you won't be able to log into your new system.

and that should be enough to get your system to boot on ZFS.

After reboot

Forgot to reset password?

System Rescue CD

If you aren't using bliss-initramfs, then you can reboot back into your sysresccd and reset through there by mounting your drive, chrooting, and then typing passwd.

Example:

# zpool import -f -o cachefile= -R /mnt/funtoo tank
# chroot /mnt/funtoo bash -l
# passwd
# exit
# reboot

Using bliss-initramfs

If you forgot to reset your password and are using bliss-initramfs, you can add the su option to your bootloader parameters and the initramfs will throw you into the rootfs of your drive. In there you can run 'passwd' and then type 'exit'. Once you type 'exit', the initramfs will continue to boot your system as normal.

Create initial ZFS Snapshot

Continue to set up anything you need in terms of /etc configurations. Once you have everything the way you like it, take a snapshot of your system. You will be using this snapshot to revert back to this state if anything ever happens to your system down the road. The snapshots are cheap, and almost instant.

To take the snapshot of your system, type the following:

# zfs snapshot -r tank@install

To see if your snapshot was taken, type:

# zfs list -t snapshot

If your machine ever fails and you need to get back to this state, just type (This will only revert your / dataset while keeping the rest of your data intact):

# zfs rollback tank/funtoo/root@install

Important

For a detailed overview, presentation of ZFS' capabilities, as well as usage examples, please refer to the ZFS Fun page.

Troubleshooting

Starting from scratch

If your installation has gotten screwed up for whatever reason and you need a fresh restart, you can do the following from sysresccd to start fresh:

Destroy the pool and any snapshots and datasets it has
# zpool destroy -R -f tank

This deletes the files from /dev/sda1 so that even after we zap, recreating the drive in the exact sector
position and size will not give us access to the old files in this partition.
# mkfs.ext2 /dev/sda1
# sgdisk -Z /dev/sda

Now start the guide again :).