Chinaunix首页 | 论坛 | 博客
  • 博客访问: 591720
  • 博文数量: 118
  • 博客积分: 2114
  • 博客等级: 大尉
  • 技术积分: 1275
  • 用 户 组: 普通用户
  • 注册时间: 2009-03-10 00:02
文章分类

全部博文(118)

文章存档

2019年(1)

2018年(4)

2017年(1)

2016年(6)

2015年(1)

2014年(1)

2013年(5)

2012年(4)

2011年(17)

2010年(13)

2009年(65)

分类:

2010-06-04 16:39:07

Diagnosing Potential Problems
  • General commands for diagnosing hardware problems are:
 zpool status
 zpool status -v
 fmdump
 fmdump -ev or fmdump -eV
 format or rmformat
  • Identify hardware problems with the zpool status commands. If a pool is in the DEGRADED state, use the zpool status command to identify if a disk is unavailable. For example:
  # zpool status -x
  pool: zeepool
 state: DEGRADED
status: One or more devices could not be opened.  Sufficient replicas exist for
        the pool to continue functioning in a degraded state.
action: Attach the missing device and online it using 'zpool online'.
   see: 
 scrub: resilver completed after 0h12m with 0 errors on Thu Aug 28 09:29:43 2008
config:

        NAME                 STATE     READ WRITE CKSUM
        zeepool              DEGRADED     0     0     0
          mirror             DEGRADED     0     0     0
            c1t2d0           ONLINE       0     0     0
            spare            DEGRADED     0     0     0
              c2t1d0         UNAVAIL      0     0     0  cannot open
              c2t3d0         ONLINE       0     0     0
        spares
          c1t3d0             AVAIL   
          c2t3d0             INUSE     currently in use

errors: No known data errors
  • See the disk replacement example to recover from a failed disk.
  • Identify potential data corruption with the zpool status -v command. If only one file is corrupted, then you might choose to deal with it directly, without needing to restore the entire pool.
# zpool status -v rpool
  pool: rpool
 state: DEGRADED
status: One or more devices has experienced an error resulting in data
        corruption. Applications may be affected.

action: Restore the file in question if possible. Otherwise restore the
        entire pool from backup.

see: 
scrub: scrub completed after 0h2m with 1 errors on Tue Mar 11 13:12:42 2008

config:
       NAME        STATE     READ WRITE CKSUM
       rpool    DEGRADED     0     0     9
         c2t0d0s0  DEGRADED     0     0     9

errors: Permanent errors have been detected in the following files:
           /mnt/root/lib/amd64/libc.so.1
  • Display the list of suspected faulty devices using the fmdump command. It is also useful to know the diagnosis engines available on your system and how busy they have been, which is obtained via the fmstat command. Similarly, fmadm will show the status of the diagnosis engines. You can also see that there are 4 diagnosis engines which are appropriate to devices and ZFS: disk-transport, io-retire, zfs-diagnosis, and zfs-retire. Check your OS release for the available FMA diagnosis engine capability.
# fmdump
TIME                 UUID                                 SUNW-MSG-ID
Aug 18 18:32:48.1940 940422d6-03fb-4ea0-b012-aec91b8dafd3 ZFS-8000-D3
Aug 21 06:46:18.5264 692476c6-a4fa-4f24-e6ba-8edf6f10702b ZFS-8000-D3
Aug 21 06:46:18.7312 45848a75-eae5-66fe-a8ba-f8b8f81deae7 ZFS-8000-D3
# fmstat
module             ev_recv ev_acpt wait  svc_t  %w  %b  open solve  memsz  bufsz
cpumem-retire            0       0  0.0    0.0   0   0     0     0      0      0
disk-transport           0       0  0.0   55.9   0   0     0     0    32b      0
eft                      0       0  0.0    0.0   0   0     0     0   1.2M      0
fabric-xlate             0       0  0.0    0.0   0   0     0     0      0      0
fmd-self-diagnosis       0       0  0.0    0.0   0   0     0     0      0      0
io-retire                0       0  0.0    0.0   0   0     0     0      0      0
snmp-trapgen             0       0  0.0    0.0   0   0     0     0    32b      0
sysevent-transport       0       0  0.0 4501.8   0   0     0     0      0      0
syslog-msgs              0       0  0.0    0.0   0   0     0     0      0      0
zfs-diagnosis            0       0  0.0    0.0   0   0     0     0      0      0
zfs-retire               0       0  0.0    0.0   0   0     0     0      0      0
# fmadm config
MODULE                   VERSION STATUS  DESCRIPTION
cpumem-retire            1.1     active  CPU/Memory Retire Agent
disk-transport           1.0     active  Disk Transport Agent
eft                      1.16    active  eft diagnosis engine
fabric-xlate             1.0     active  Fabric Ereport Translater
fmd-self-diagnosis       1.0     active  Fault Manager Self-Diagnosis
io-retire                2.0     active  I/O Retire Agent
snmp-trapgen             1.0     active  SNMP Trap Generation Agent
sysevent-transport       1.0     active  SysEvent Transport Agent
syslog-msgs              1.0     active  Syslog Messaging Agent
zfs-diagnosis            1.0     active  ZFS Diagnosis Engine
zfs-retire               1.0     active  ZFS Retire Agent
  • Display more details about potential hardware problems by examining the error reports with fmdump -ev. Display even more details with fmdump -eV.
# fmdump -eV
TIME                           CLASS
Aug 18 2008 18:32:35.186159293 ereport.fs.zfs.vdev.open_failed
nvlist version: 0
        class = ereport.fs.zfs.vdev.open_failed
        ena = 0xd3229ac5100401
        detector = (embedded nvlist)
        nvlist version: 0
                version = 0x0
                scheme = zfs
                pool = 0x4540c565343f39c2
                vdev = 0xcba57455fe08750b
        (end detector)

        pool = whoo
        pool_guid = 0x4540c565343f39c2
        pool_context = 1
        pool_failmode = wait
        vdev_guid = 0xcba57455fe08750b
        vdev_type = disk
        vdev_path = /dev/ramdisk/rdx
        parent_guid = 0x4540c565343f39c2
        parent_type = root
        prev_state = 0x1
        __ttl = 0x1
        __tod = 0x48aa22b3 0xb1890bd
  • If expected devices can't be displayed with the format or rmformat utility, then those devices won't be visible to ZFS.

[] Disk Replacement Example

Replacing Devices in a Pool

  • In the following example, we are assuming that c4t60060160C166120099E5419F6C29DC11d0s6 is the faulty disk and it is replaced with c4t60060160C16612006A4583D66C29DC11d0s6 in the z-mirror pool.
    # zpool status z-mirror
      pool: z-mirror
     state: ONLINE
     scrub: resilver completed with 0 errors on Tue Sep 11 09:05:44 2007
    config:
    
            NAME                                         STATE     READ WRITE CKSUM
            z-mirror                                     ONLINE       0     0     0
              mirror                                     ONLINE       0     0     0
                c4t60060160C166120064F22DA86C29DC11d0s6  ONLINE       0     0     0
                c4t60060160C166120099E5419F6C29DC11d0s6  FAULTED      0     0     0
  • If you are replacing the faulted disk with a replacement disk in the same physical location, physically replace the new disk.
  • Let ZFS know that the faulted disk has been replaced by using this syntax: zpool replace [-f] [new_device]
    # zpool replace z-mirror c4t60060160C166120099E5419F6C29DC11d0s6 c4t60060160C16612006A4583D66C29DC11d0s6

    If you are replacing a disk in the same physical location, then you only need identify the original device. For example:

    # zpool replace z-mirror c4t60060160C166120099E5419F6C29DC11d0s6
  • Review the pool status.
    # zpool status z-mirror
      pool: z-mirror
     state: ONLINE
     scrub: resilver completed with 0 errors on Tue Sep 11 09:08:44 2007
    config:
             NAME                                         STATE     READ WRITE CKSUM
             z-mirror                                     ONLINE       0     0     0
              mirror                                      ONLINE       0     0     0
                c4t60060160C166120064F22DA86C29DC11d0s6   ONLINE       0     0     0
                c4t60060160C16612006A4583D66C29DC11d0s6   ONLINE       0     0     0
    
    errors: No known data errors
  • If necessary, clear any existing errors that resulted from the faulted disk.
    # zpool clear z-mirror
  • Scrub the pool to ensure the new device is functioning properly.
    # zpool scrub z-mirror

[] Notify FMA That Device Replacement is Complete

Follow these steps after the device is replaced, the pool errors are cleared, and the pool is scrubbed.

  • Review the FMA/zfs diagnostic counters from the previous fault.
    # fmstat
  • Reset the ZFS diagnostic counters and determine whether any new fault activity is occurring.
    # fmadm reset zfs-diagnosis
    # fmadm reset zfs-retire
    # fmstat
  • Determine the FMA fault event from the failed device.
    # fmadm faulty -a
    --------------- ------------------------------------  -------------- ---------
    TIME            EVENT-ID                              MSG-ID         SEVERITY
    --------------- ------------------------------------  -------------- ---------
    Jul 17 11:03:56 378924d1-840b-c4dd-c8e2-a5491d4047ff  ZFS-8000-D3    Major    
    ...
    Fault class : fault.fs.zfs.device
    Affects     : zfs://pool=rzpool/vdev=70f7855d9f673fcc
                      faulted but still in service
    Problem in  : zfs://pool=rzpool/vdev=70f7855d9f673fcc
                      faulted but still in service
    ...
  • Let FMA know that the ZFS fault event is cleared.
    # fmadm repair zfs://pool=rzpool/vdev=70f7855d9f673fcc
    fmadm: recorded repair to zfs://pool=rzpool/vdev=70f7855d9f673fcc
  • Confirm that no new faults have occurred.
    # fmadm faulty

[] Solving Mirrored {Root} Pool Problems (zpool attach)

  • If you cannot attach a disk to create a mirrored root or non-root pool with the zpool attach command, and you see messages similar to the following:
# zpool attach rpool c1t1d0s0 c1t0d0s0
cannot attach c1t0d0s0 to c1t1d0s0: new device must be a single disk
  • You might be running into CR 6852962, which was been seen in an LDOM environment.
  • If the problem is not 6852962 and the system is booted under a virtualization product, make sure the devices are accessible by ZFS outside of the virtualization product.

[] Creating a Pool or Attaching a Disk to a Pool (I/O error)

If you attempt to create a pool or attach a disk or a disk slice to a existing pool and you see the following error:

# zpool attach rpool c4t0d0s0 c4t1d0s0
cannot open '/dev/dsk/c4t1d0s0': I/O error

This error means that the disk slice doesn't have any disk space allocated to it or possibly that a Solaris fdisk partition and the slice doesn't exist on an x86 system. Use the format utility to allocate disk space to a slice. If the x86 system doesn't have a Solaris fdisk partition, use the fdisk utility to create one.

[] Panic/Reboot/Pool Import Problems

During the boot process, each pool must be opened, which means that pool failures might cause a system to enter into a panic-reboot loop. In order to recover from this situation, ZFS must be informed not to look for any pools on startup.

[] Boot From Milestone=None Recovery Method

  • Boot to the none milestone by using the -m milestone=none boot option.
ok boot -m milestone=none
  • Remount your root file system as writable.
  • Rename or move the /etc/zfs/zpool.cache file to another location.
   These actions cause ZFS to forget that any pools exist on the system, preventing it from trying to access the bad pool causing the 
   problem. If you have multiple pools on the system, do these additional steps:
   * Determine which pool might have issues by using the fmdump -eV command to display the pools with reported fatal errors.
   * Import the pools one-by-one, skipping the pools that are having issues, as described in the fmdump output.
  • If the system is back up, issue the svcadm milestone all command.

[] Boot From OpenSolaris Live CD Recovery Method

If you are running a Solaris SXCE or Solaris 10 release, you might be able to boot from the OpenSolaris Live CD and fix whatever is causing the pool import to fail.

  • Boot from the OpenSolaris Live CD
  • Import the pool
  • Resolve the issue that causes the pool import to fail, such as replace a failed disk
  • Export the pool
  • Boot from the original Solaris release
  • Import the pool

[] Changing Disk Capacity Sizes

  • You can use the autoexpand property to expand a disk's size in the Nevada release, build 117. For example, two 16-Gbyte disks in a mirrored pool are replaced with two 72-Gbyte disks. The autoexpand property is enabled after the disk replacements to expand the full LUN sizes.
# zpool list pool
NAME   SIZE   USED  AVAIL    CAP  HEALTH  ALTROOT
pool  16.8G  76.5K  16.7G     0%  ONLINE  -
# zpool replace pool c1t16d0 c1t1d0
# zpool replace pool c1t17d0 c1t2d0
# zpool list pool
NAME   SIZE   USED  AVAIL    CAP  HEALTH  ALTROOT
pool  16.8G  88.5K  16.7G     0%  ONLINE  -
# zpool set autoexpand=on pool
# zpool list pool
NAME   SIZE   USED  AVAIL    CAP  HEALTH  ALTROOT
pool  68.2G   117K  68.2G     0%  ONLINE  -
  • If you resize a LUN from a storage array and the zpool status command doesn't display the LUN's expected capacity, export and import the pool to see expected capacity. This problem is related to CRs 6475340/6606879, fixed in the Nevada release, build 117.
  • If zpool status doesn't display the array's LUN expected capacity, confirm that the expected capacity is visible from the format utility. For example, the format output below shows that one LUN is configured as 931.01 Gbytes and one is configured as 931.01 Mbytes.
          2. c6t600A0B800049F93C0000030A48B3EA2Cd0 
             /scsi_vhci/ssd@g600a0b800049f93c0000030a48b3ea2c
          3. c6t600A0B800049F93C0000030D48B3EAB6d0 
             /scsi_vhci/ssd@g600a0b800049f93c0000030d48b3eab6
  • You will need to reconfigure the array's LUN capacity with the array sizing tool to correct this sizing problem.
  • When the LUN sizes are resolved, export and import the pool if the pool has already been created with these LUNs.

[] Resolving ZFS Disk Space Issues

  • The zpool list command reports physical disk capacity in the pool and zfs list command reports usable space that is available to file systems, which is disk space minus ZFS redundancy metadata overhead, if any. If the zpool list command reports additional capacity, but the zfs list command says the available file system is 0, then the file system is full.
  • We recommend using the zpool list and zfs list commands rather than the df command to identify available pool space and available file system space. The df doesn't understand descendent datasets, whether snapshots exist, nor is df dedup-aware. If any ZFS properties, such as compression and quotas, are set on file systems, reconciling the space consumption that is reported by df might be difficult.
  • For more information about the ZFS deduplication feature impacts space accounting, see .
  • The zpool list and zpool import command output has changed starting in Nevada, builds 125-129. For more information, see .

[] Resolving Software Problems

[] System hangs or application hangs with failmode=continue (6808756)

  • Description
    • Disk failure might result in ZFS pool I/O operations being erroneously blocked that results in an affected application hanging.
    • A system or application hang can occur even when the pool property 'failmode' is set to 'continue' in the event of a device failure.
  • Solaris Releases
    • SPARC systems that run Solaris 10 (patch 137137-09) or x86 systems that run Solaris 10 (patch 137138-09)
    • SPARC or x86 systems that run Open Solaris releases, builds snv_77 and above.
  • Symptoms
    • System or application hangs.
    • All I/O access to the affected ZFS storage pool is blocked.
  • Resolution
  1. Correct the disk failure, which might include physically replacing a failed disk Then, let ZFS know the disk is replaced.
    # zpool replace pool-name device-name
  2. Clear the errors to bring the pool back to an operational state.
    # zpool clear pool-name

[] RAID-Z Checksum Errors in Nevada Builds, 120-123

Update: This bug is fixed in Nevada, build 124.

  • A bug (6869090) in NV build 120 causes checksum errors in RAID-Z configurations as follows:
    • Data corruption on a RAID-Z system of any sort (raidz1, raidz2, raidz3) can lead to spurious checksum errors being reported on devices that were not used as part of the reconstruction. These errors are harmless and can be cleared safely (zpool clear ).
    • A far more serious problem with single-parity RAID-Z configurations can lead to data corruption. This data corruption is recoverable as long as no additional data corruption or drive failure occurs. That is to say, data is fine provided no additional problems occur. The problem is present on all RAIDZ-1 configurations that use an odd number of children (disks) for example, 4+1, or 6+1. Note that RAIDZ-1 configurations with an even number of children (for example, 3+1), raidz2, and raidz3 are unaffected.
  • Mirrored storage pools are not impacted by this bug.
  • Solaris 10 releases are not impacted by this bug.
  • OpenSolaris releases are not impacted by this bug unless you have updated to build 120-123.
  • The recommended course of action is to roll back to build snv_119 or earlier. If for some reason this is impossible, please email zfs-discuss, and we can discuss the best course of action for you. After rolling back, initiate a scrub. ZFS will identify and correct these errors, but if enough accumulate it will (incorrectly) identify drives as faulty (which they likely aren't). You can clear these failures (zpool clear ).
  • Without rolling back, repeated scrubs will eventually remove all traces of the data corruption. You may need to clear checksum failures as they're identified to ensure that enough drives remain online.

[] Unsupported CIFS properties in Solaris 10 Releases

  • The Solaris 10 10/08 release includes modifications to support the Solaris CIFS environment as described in zfs.1m. However, the CIFS features are not supported in the Solaris 10 release. Therefore, these properties are set to read-only values. If you attempt to reset the CIFS-related properties, you will see a message similar to the following:
# zfs set utf8only=on rpool
cannot set property for 'rpool': 'utf8only' is readonly

[] Cache Device Support Starts in the Solaris 10 10/09 Release

  • Cache devices are supported starting in the Solaris 10 10/09 release. Starting in the Solaris 10 10/08 release, cache device support is identified as available by using the zpool upgrade -v command. For example:
# zpool upgrade -v
This system is currently running ZFS pool version 10.

The following versions are supported:

VER  DESCRIPTION
---  --------------------------------------------------------
 1   Initial ZFS version
 2   Ditto blocks (replicated metadata)
 3   Hot spares and double parity RAID-Z
 4   zpool history
 5   Compression using the gzip algorithm
 6   bootfs pool property
 7   Separate intent log devices
 8   Delegated administration
 9   refquota and refreservation properties
 10  Cache devices
  • However, cache devices are not supported in Solaris 10 releases prior to the Solaris 10 10/09 release.
  • If you attempt to add a cache device to a ZFS storage pool when the pool is created, the following message is displayed:
# zpool create pool mirror c0t1d0 c0t2d0 cache c0t3d0
cannot create 'pool': operation not supported on this type of pool
  • If you attempt to add a cache device to a ZFS storage pool after the pool is created, the following message is displayed:
# zpool create pool mirror c0t1d0 c0t2d0
# zpool add pool cache c0t3d0
cannot add to 'pool': pool must be upgraded to add these vdevs

[] Resolving NFS Problems

  • Current Solaris releases run NFSv4 by default, which provides access to ZFS style ACLs. If you are sharing data with Linux systems or older Solaris system, which might be running NFSv3, the NFSv4 ACL info will be available but not viewable or modifiable.
  • If you are sharing UIDs/GIDs across systems, make sure the UIDs/GIDs are translated correctly. If you see "nobody" instead of a UID, then check that the NFSMAPID_DOMAIN is set correctly.
  • Check basic network connectivity and NFS sharing first before testing ACL behavior. Create and share a test file system with no ACLs first and see if the data is accessible on the NFS client.
  • Use getent to check that hostnames and usernames are available across systems.
  • Current OpenSolaris releases provide the mirror mount feature, which means that when file systems are created on the NFS server, the NFS client can automatically discover these newly created file systems within their existing mount of a parent file system. This feature is not available in Solaris 10 releases.
  • Current OpenSolaris releases require fully-qualified hostnames when you use the sharenfs property to share ZFS file systems.

[] Pool Migration Issues

  • Moving a ZFS storage pool from a FreeBSD system to a Solaris system:
    • A ZFS storage pool that is created on a FreeBSD system is created on a disk's p* devices with an EFI label. On a FreeBSD system, you can boot from a disk that has an EFI label.
    • A ZFS storage pool that is created on a Solaris system is either created on a disk's d* devices with an EFI label or if the pool is used for booting, the pool must be created on disk slices (s* devices) with an SMI label. You cannot boot from an disk that has an EFI label in Solaris releases.
    • If you are migrating a pool from a FreeBSD system that was used for booting, you will need to unset the bootfs property before the migration if the migrated pool will no longer be used for booting on the Solaris system. An existing bug in the Solaris releases does not allow you to unset the bootfs property on a pool that has disks with an EFI label. If you don't unset the bootfs property before the migration, you will not be able to add whole disks to this pool.
    • A workaround is to boot the system under FreeBSD's Fixit mode and use the zpool set bootfs= pool-name command to unset the bootfs property. Then, import the pool on a Solaris system.
    • A pool that is used for booting on a FreeBSD system can't be used for booting on a Solaris system. You would need to relabel the disks with SMI labels and allocate space in the disk slices. This would cause any data on the existing pool to be removed.
  • A larger issue is that a ZFS storage pool should not be created on a disk's p* (partition) devices. In Solaris releases, the p* devices represent the larger fdisk partitions. These partitions are available for a Solaris installation and other possible OSes. A ZFS storage pool should be created on a disk's d* devices that represent the whole disk. Creating a pool on a disk's s* devices represent a slice within the disk and should be used when creating a ZFS root pool or for some other specialized purpose.
  • The problem with creating a storage pool on a disk's p* devices is that you could conceivably create pools with overlapping disk partitions, such as creating one pool with c0t1d0p0 and another pool with c0t1d0, for example. This overlap could cause data loss or a system crash. This is CR 6884447.

[] ZFS Installation Issues

[] Review Solaris Installation Requirements

  • 768 Mbytes is the minimum amount of memory required to install a ZFS root file system
  • 1 Gbyte of memory is recommended for better overall ZFS performance
  • At least 16 Gbytes of disk space is recommended
  • ZFS boot and install support is provided starting in the Solaris 10 10/08 release, SXCE, build 90, and the OpenSolaris releases

[] Before You Start

  • Due to an existing boot limitation, disks intended for a bootable ZFS root pool must be created with disk slices and must be labeled with a VTOC (SMI) disk label.
  • If you relabel EFI-labeled disks with VTOC labels, be sure that the desired disk space for the root pool is in the disk slice that will be used to create the bootable ZFS pool.

[] Solaris/ ZFS Initial Installation

  • For the OpenSolaris releases, a ZFS root file system is installed by default and there is no option to choose another type of root file system.
  • For SXCE, build 90 and Solaris 10 10/08 releases, you can only install a ZFS root file system from the text installer.
  • You cannot use the standard upgrade option to install or migrate to a ZFS root file system. For information on creating ZFS flash archive in the Solaris 10 release by installing patches, go to .
  • Starting in SXCE, build 90 and the Solaris 10 10/08 release, you can use Solaris LiveUpgrade to migrate a UFS root file system to a ZFS root file system.
  • Access the text installer as follows:
    • On SPARC based system, use the following syntax from the Solaris installation DVD or the network:
 ok boot cdrom - text
 ok boot net - text
**On an x86 based system, select the text-mode install option when presented.
  • If you accidentally start the GUI install method, do the following:
 * Exit the GUI installer
 * Expand the terminal window to 80 x 24
 * Unset the DISPLAY, like this:
   # DISPLAY=
   # export DISPLAY
   # install-solaris

[] ZFS Root Pool Recommendations and Requirements

  • During an initial installation, select two disks to create a mirrored root pool. Or, you can also attach a disk to create a mirrored root pool after installation. See the ZFS Administration Guide for details.
  • Disks intended for the root pool must contain a slice and have an SMI label. Disks with EFI labels cannot be booted. Several factors are at work here, including BIOS support for booting from EFI labeled disks.
  • Minimum root pool version is 10.
  • If you attach another disk to create a mirrored root pool later, make sure you specify a bootable slice and not the whole disk because the latter may try to install an EFI label.
  • You cannot use a RAID-Z configuration for a root pool. Only single-disk pools or pools with mirrored disks are supported. You will see the following message if you attempt to use an unsupported pool for the root pool:
ERROR: ZFS pool  does not support boot environments
  • Root pools cannot have a separate log device. For example:
# zpool add -f rpool log c0t6d0s0
cannot add to 'rpool': root pool can not have multiple vdevs or separate logs
  • The lzjb compression property is supported for root pools but the other compression types are not supported.
  • Keep a second ZFS BE for recovery and patching purposes. This process is safer than patching a live boot environment for two reasons:
    • You're applying the patches to an inactive boot environment, which is safer than patching a live environment
    • Your previous environment isn't modified at all, which makes it a secure environment to go back to if the new one fails
  • The patching or upgrading process is as follows:
    • Clone your current boot environment, using either LiveUpgrade or beadm (beadm is part of the OpenSolaris release).
    • Apply the patches or perform the upgrade to the clone BE.
    • Activate and boot the clone BE.
    • If the patched or upgraded boot environment fails, reboot the old one.
  • Keep root pool snapshots on a remote system. See the steps below for details.

[] Solaris Live Upgrade Migration Scenarios

  • You can use the Solaris Live Upgrade feature to migrate a UFS root file system to a ZFS root file system.
  • You can't use Solaris Live Upgrade to migrate a ZFS boot environment (BE) to UFS BE.
  • You can't use Solaris Live Upgrade to migrate non-root or shared UFS file systems to ZFS file systems.

[] Review LU Requirements

  • You must be running the SXCE, build 90 release or the Solaris 10 10/08 release to use LU to migrate a UFS root file system to a ZFS root file system.
  • You must create a ZFS storage pool that contains disk slices before the LU migration.
  • The pool must exist either on a disk slice or on disk slices that are mirrored, but not on a RAID-Z configuration or on a nonredundant configuration of multiple disks. If you attempt to use an unsupported pool configuration during a Live Upgrade migration, you will see a message similar to the following:
ERROR: ZFS  does not support boot environments
  • If you see this message, then either the pool doesn't exist or its an unsupported configuration.

[] Live Upgrade Issues

  • The Solaris installation GUI's standard-upgrade option is not available for migrating from a UFS to a ZFS root file system. To migrate from a UFS file system, you must use Solaris Live Upgrade. You cannot use Solaris Live Upgrade to create a UFS BE from a ZFS BE.
  • Do not rename your ZFS BEs with the zfs rename command because the Solaris Live Upgrade feature is unaware of the name change. Subsequent commands, such as ludelete, will fail. In fact, do not rename your ZFS pools or file systems if you have existing BEs that you want to continue to use.
  • Solaris Live Upgrade creates the datasets for the BE and ZFS volumes for the swap area and dump device but does not account for any existing dataset property modifications. Thus, if you want a dataset property enabled in the new BE, you must set the property before the lucreate operation. For example:
# zfs set compression=on rpool/ROOT
  • When creating an alternative BE that is a clone of the primary BE, you cannot use the -f, -x, -y, -Y, and -z options to include or exclude files from the primary BE. You can still use the inclusion and exclusion option set in the following cases:

UFS -> UFS UFS -> ZFS ZFS -> ZFS (different pool)

  • Although you can use Solaris Live Upgrade to upgrade your UFS root file system to a ZFS root file system, you cannot use Solaris Live Upgrade to upgrade non-root or shared file systems.
  • On a SPARC system that runs the Solaris 10 5/09 release, set the BOOT_MENU_FILE variable before activating the ZFS BE with luactivate, due to CR 6824589.
# BOOT_MENU_FILE="menu.lst"
# export BOOT_MENU_FILE
  • LiveUpgrade is not supported on a system in diagnostic mode, where the diag switch is set to true.
[] Live Upgrade Issues (Solaris 10 10/09)

Review the following information before upgrading to the Solaris 10 10/09 release.

  • If you have a separate /var dataset in a previous Solaris 10 release and you want to use luupgrade to upgrade to the Solaris 10 10/09 release, and you have applied patch 121430-42 (on SPARC systems) or 121431-43 (on x86 systems), apply IDR143154-01 (SPARC) or IDR143155-01 (x86) to avoid the issue described in CR 6884728. Symptoms of this bug are that entries for the ZFS BE and ZFS /var dataset are incorrectly placed in the /etc/vfstab file. The workaround is to boot in maintenance mode and remove the erroneous vfstab entries.
  • If you attempt to use lucreate and you have a ZFS dataset with a mountpoint set to legacy in a non-global zone, the luactivate might fail. This is CR 6837400.
  • If the luactivate fails, you might need to reinstall the boot blocks to boot from the previous BE.
[] Live Upgrade Problem (Starting in Nevada, build 125)
  • Starting in build 125, device naming changes means that attempts to use Live Upgrade on a mirrored root pool cause Live Upgrade to fail. Error messages are similar to the following:
# luactivate b126
System has findroot enabled GRUB
ERROR: Unable to determine the configuration of the current boot environment . 
  • Supported OpenSolaris releases are not impacted by the build 125 device naming changes.
  • If you are considering this release for the ZFS log device removal feature, then also consider that you will not be able to patch or upgrade the ZFS root dataset in a mirrored root pool in this release with Solaris Live Upgrade unless you apply one of the workarounds. Unmirrored root pools are not impacted.
  • Until CR 6894189 integrates, any Live Upgrade of a mirrored root after build 125 will fail. See the workaround described in CR 6894189, or detach your secondary root pool disk, run the Live Upgrade operation, and then re-attach your secondary root pool disk.

[] Live Upgrade with Zones

Review the following supported ZFS and zones configurations. These configurations are upgradeable and patchable.

[] Migrate a UFS Root File System with Zones Installed to a ZFS Root File System

This ZFS zone root configuration can be upgraded or patched. See the ZFS Administration Guide for information about supported zones configurations that can be upgraded or patched in the Solaris 10 release.

  1. Upgrade the system to the Solaris 10 10/08 release if it is running a previous Solaris 10 release.
  2. Create the root pool.
    # zpool create rpool mirror c1t0d0s0 c1t1d0s0
  3. Confirm that the zones from the UFS environment are booted.
  4. Create the new boot environment.
    # lucreate -n S10BE2 -p rpool
  5. Activate the new boot environment.
  6. Reboot the system.
  7. Migrate the UFS zones to the ZFS BE.
    1. Boot the zones.
    2. Create another BE within the pool.
      # lucreate S10BE3
    3. Activate the new boot environment.
      # luactivate S10BE3
    4. Reboot the system.
      # init 6
  8. Resolve any potential mount point problems, due to a Solaris Live Upgrade bug.
    1. Review the zfs list output and look for any temporary mount points.
      # zfs list -r -o name,mountpoint rpool/ROOT/s10u6
          
          NAME                               MOUNTPOINT
          rpool/ROOT/s10u6                   /.alt.tmp.b-VP.mnt/
          rpool/ROOT/s10u6/zones             /.alt.tmp.b-VP.mnt//zones
          rpool/ROOT/s10u6/zones/zonerootA   /.alt.tmp.b-VP.mnt/zones/zonerootA
      The mount point for the root ZFS BE (rpool/ROOT/s10u6) should be /.
    2. Reset the mount points for the ZFS BE and its datasets.
      # zfs inherit -r mountpoint rpool/ROOT/s10u6
      # zfs set mountpoint=/ rpool/ROOT/s10u6
    3. Reboot the system. When the option is presented to boot a specific boot environment, either in the GRUB menu or at the OpenBoot Prom prompt, select the boot environment whose mount points were just corrected.
[] Configure a ZFS Root File System With Zone Roots on ZFS

Set up a ZFS root file system and ZFS zone root configuration that can be upgraded or patched. In this configuration, the ZFS zone roots are created as ZFS datasets.

  1. Install the system with a ZFS root, either by using the interactive initial installation method or the Solaris JumpStart installation method.
  2. Boot the system from the newly-created root pool.
  3. Create a dataset for grouping the zone roots.
    zfs create -o canmount=noauto rpool/ROOT/S10be/zones
    Setting the noauto value for the canmount property prevents the dataset from being mounted other than by the explicit action of Solaris Live Upgrade and system startup code.
  4. Mount the newly-created zones container dataset.
    # zfs mount rpool/ROOT/S10be/zones
    The dataset is mounted at /zones.
  5. Create and mount a dataset for each zone root.
    # zfs create -o canmount=noauto rpool/ROOT/S10be/zones/zonerootA
    # zfs mount rpool/ROOT/S10be/zones/zonerootA
  6. Set the appropriate permissions on the zone root directory.
    # chmod 700 /zones/zonerootA
  7. Configure the zone, setting the zone path as follows:
    # zonecfg -z zoneA
        zoneA: No such zone configured
        Use 'create' to begin configuring a new zone.
        zonecfg:zoneA> create
        zonecfg:zoneA> set zonepath=/zones/zonerootA
  8. Install the zone.
    # zoneadm -z zoneA install
  9. Boot the zone.
    # zoneadm -z zoneA boot
[] Upgrade or Patch a ZFS Root File System With Zone Roots on ZFS

Upgrade or patch a ZFS root file system with zone roots on ZFS. These updates can either be a system upgrade or the application of patches.

  1. Create the boot environment to upgrade or patch.
    # lucreate -n newBE
    The existing boot environment, including all the zones, are cloned. New datasets are created for each dataset in the original boot environment. The new datasets are created in the same pool as the current root pool.
  2. Select one of the following to upgrade the system or apply patches to the new boot environment.
    • Upgrade the system.
      # luupgrade -u -n newBE -s /net/install/export/s10u7/latest
      Where the -s option is the location of a Solaris installation medium.
    • Apply patches to the new boot environment.
      # luupgrade -t -n newBE -t -s /patchdir 139147-02 157347-14
  3. Activate the new boot environment after the updates to the new boot environment are complete.
    # luactivate newBE
  4. Boot from newly-activated boot environment.
    # init 6
  5. Resolve any potential mount point problems, due to a Solaris Live Upgrade bug.
    1. Review the zfs list output and look for any temporary mount points.
      # zfs list -r -o name,mountpoint rpool/ROOT/newBE
          
          NAME                               MOUNTPOINT
          rpool/ROOT/newBE                   /.alt.tmp.b-VP.mnt/
          rpool/ROOT/newBE/zones             /.alt.tmp.b-VP.mnt//zones
          rpool/ROOT/newBE/zones/zonerootA   /.alt.tmp.b-VP.mnt/zones/zonerootA
          
      The mount point for the root ZFS BE (rpool/ROOT/newBE) should be /.
    2. Reset the mount points for the ZFS BE and its datasets.
      # zfs inherit -r mountpoint rpool/ROOT/newBE
      # zfs set mountpoint=/ rpool/ROOT/newBE
    3. Reboot the system. When the option is presented to boot a specific boot environment, either in the GRUB menu or at the OpenBoot Prom prompt, select the boot environment whose mount points were just corrected.

    [] Recover from BE Removal Failure (ludelete)

    • If you use ludelete to remove an unwanted BE and it fails with messages similar to the following:
    # ludelete -f c0t1d0s0
    System has findroot enabled GRUB
    Updating GRUB menu default setting
    Changing GRUB menu default setting to <0>
    ERROR: Failed to copy file  to top level dataset for BE 
    ERROR: Unable to delete GRUB menu entry for deleted boot environment .
    Unable to delete boot environment.
    
    • You might be running into the following bugs: 6718038, 6715220, 6743529
    • The workaround is as follows:
    1. Edit /usr/lib/lu/lulib and in line 2934, replace the following text:
      lulib_copy_to_top_dataset "$BE_NAME" "$ldme_menu" "/${BOOT_MENU}"
       with this text:
      
      lulib_copy_to_top_dataset `/usr/sbin/lucurr` "$ldme_menu" "/${BOOT_MENU}"
    2. Rerun the ludelete operation.

    [] ZFS Root Pool and Boot Issues

    • The boot process can be slow if the boot archive is updated or a dump device has changed. Be patient.
    • CR 6704717 – Do not place offline the primary disk in a mirrored ZFS root configuration. If you do need to offline or detach a mirrored root disk for replacement, then boot from another mirrored disk in the pool.
    • CR 6668666 - If you attach a disk to create a mirrored root pool after an initial installation, you will need to apply the boot blocks to the secondary disks. For example:
      sparc# installboot -F zfs /usr/platform/`uname -i`/lib/fs/zfs/bootblk /dev/rdsk/c0t1d0s0
      x86# installgrub /boot/grub/stage1 /boot/grub/stage2 /dev/rdsk/c0t1d0s0 
    
    • CR 6741743 - The boot -L command doesn't work if you migrated from a UFS root file system. Copy the bootlst command to the correct location. This CR is fixed in the Solaris 10 5/09 release.

    [] ZFS Boot Error Messages

    • CR 2164779 - Ignore the following krtld messages from the boot -Z command. They are harmless:
    krtld: Ignoring invalid kernel option -Z.
    krtld: Unused kernel arguments: `rpool/ROOT/zfs1008BE'.
    

    [] Recover from Lost Root Password or Similar Problem

    If you need to recover the root password or some similar problem that prevents successful login in a ZFS root environment, you will need to boot failsafe mode or boot from alternate media, depending on the severity of the error. In addition, the OpenSolaris release doesn't support failsafe mode.

    Select either of the following recovery methods:

    1. On a Solaris 10 system or Nevada system, boot failsafe mode by using the following steps:
      • Boot failsafe mode.
      ok boot -F failsafe
      • Mount the ZFS BE on /a when prompted:
      .
      .
      .
      ROOT/zfsBE was found on rpool.
      Do you wish to have it mounted read-write on /a? [y,n,?] y
      mounting rpool on /a
      Starting shell.
      • Change to the /a/etc directory.
      # cd /a/etc
      • Correct the passwd or shadow file.
      # vi passwd
      • Reboot the system.
      # init 6
    2. On an OpenSolaris system, boot from alternate media by using the following steps:
      • Boot from an installation CD or from the network:
      ok boot cdrom -s
      ok boot net -s

      If you don't use the -s option, you will need to exit the installation program.

      • Import the root pool and specify an alternate mount point:
      # zpool import -R /a rpool
      • Mount the ZFS BE specifically because canmount is set to noauto by default.
      # zfs mount rpool/ROOT/zfsBE
      • Change to the /a/etc directory.
      # cd /a/etc
      • Correct the passwd or shadow file.
      # vi shadow
      • Reboot the system.
      # init 6

    [] Resolving ZFS Mount Point Problems That Prevent Successful Booting

    The best way to change the active boot environment is to use the luactivate command. If booting the active environment fails, due to a bad patch or a configuration error, the only way to boot a different environment is by selecting that environment at boot time. You can select an alternate BE from the GRUB menu on an x86 based system or by booting it explicitly from the PROM on an SPARC based system.

    Due to a bug in the Live Upgrade feature, the non-active boot environment might fail to boot because the ZFS datasets or the zone's ZFS dataset in the boot environment has an invalid mount point.

    The same bug also prevents the BE from mounting if it has a separate /var dataset.

    The mount points can be corrected by taking the following steps.

    [] Resolve ZFS Mount Point Problems

    1. Boot the system from a failsafe archive.
    2. Import the pool.
      # zpool import rpool
    3. Review the zfs list output after the pool is imported, looking for incorrect temporary mount points. For example:
      # zfs list -r -o name,mountpoint rpool/ROOT/s10u6
          
          NAME                               MOUNTPOINT
          rpool/ROOT/s10u6                   /.alt.tmp.b-VP.mnt/
          rpool/ROOT/s10u6/zones             /.alt.tmp.b-VP.mnt//zones
          rpool/ROOT/s10u6/zones/zonerootA   /.alt.tmp.b-VP.mnt/zones/zonerootA
      The mount point for the root BE (rpool/ROOT/s10u6) should be /.
    4. Reset the mount points for the ZFS BE and its datasets.
      # zfs inherit -r mountpoint rpool/ROOT/s10u6
      # zfs set mountpoint=/ rpool/ROOT/s10u6
    5. Reboot the system. When the option is presented to boot a specific boot environment, either in the GRUB menu or at the OpenBoot Prom prompt, select the boot environment whose mount points were just corrected.

    [] Boot From a Alternate Disk in a Mirrored ZFS Root Pool

    You can boot from different devices in a mirrored ZFS root pool.

    • Identify the device pathnames for the alternate disks in the mirrored root pool by reviewing the zpool status output. In the example output, disks are c0t0d0s0 and c0t1d0s0.
    # zpool status
      pool: rpool
     state: ONLINE
     scrub: resilver completed after 0h6m with 0 errors on Thu Sep 11 10:55:28 2008
    config:
    
            NAME          STATE     READ WRITE CKSUM
            rpool         ONLINE       0     0     0
              mirror      ONLINE       0     0     0
                c0t0d0s0  ONLINE       0     0     0
                c0t1d0s0  ONLINE       0     0     0
    
    errors: No known data errors
    • If you attached the second disk in the mirror configuration after an initial installation, apply the bootblocks after the second disk has resilvered. For example, on a SPARC system:
    # installboot -F zfs /usr/platform/`uname -i`/lib/fs/zfs/bootblk /dev/rdsk/c0t1d0s0
    • Depending on the hardware configuration, you might need to update the OpenBoot PROM configuration or the BIOS to specify a different boot device. For example, on a SPARC system:
    ok setenv boot-device /pci@7c0/pci@0/pci@1/pci@0,2/LSILogic,sas@2/disk@1
    ok boot

    [] Activating a BE Fails (OpenSolaris Releases Prior to 101a)

    • If you attached a second disk to your ZFS root pool and that disk has an EFI label, booting the BE will fail with a messages similar to the following:
    # beadm activate opensolaris-2
    Unable to activate opensolaris-2.
    Unknown external error.
    • ZFS root pool disks must contain a VTOC label. Starting in build 101a, you will be warned about adding a disk with an EFI label to the root pool.
    • See the steps below to detach and relabel the disk with a VTOC label. These steps are also applicable to the Solaris Nevada (SXCE) and Solaris 10 releases.
    1. Detach the disk. For example:
      # zpool detach rpool c0t0d0s0
    2. Relabel the disk.
    3. # format -e c0t1d0s0
      format> label
      [0] SMI Label
      [1] EFI Label
      Specify Label type[1]: 0
      Ready to label disk, continue? yes
      format> quit
      
      Make sure all the disk space is in s0. The relabeling process might go back to the default sizing so check to see that all the disk space is where you want it.
  6. Attach the disk. For example:
  7. # zpool attach rpool c0t0d0s0 c0t1d0s0
  8. Confirm that the newly attached disk has resilvered completely by using the zpool status command to watch the progress.
    # zpool status
      pool: rpool
     state: ONLINE
     scrub: resilver completed after 0h8m with 0 errors on Mon Jan 26 10:39:11 2009
    config:
    
            NAME          STATE     READ WRITE CKSUM
            rpool         ONLINE       0     0     0
              mirror      ONLINE       0     0     0
                c0t0d0s0  ONLINE       0     0     0  67.9M resilvered
                c0t1d0s0  ONLINE       0     0     0  6.55G resilvered
    
  9. Install the bootblock on the newly attached disk.
    x86# installgrub /boot/grub/stage1 /boot/grub/stage2 /dev/rdsk/c0t1d0s0 
  10. Confirm that you can boot from the newly attached disk by selecting this disk from the BIOS level.

[] ZFS Root Pool Recovery

[] Complete Solaris ZFS Root Pool Recovery

The section describes how to create and restore root pool snapshots starting with Solaris 10 10/08 and recent SXCE releases. The following issues can complicate the root pool snapshot process:

  • CR 6462803 - Fixed in the SXCE (Nevada, build 111) release but still open in the Solaris 10 5/09 release, and means that a recursive snapshot of the root pool can fail if more than one BE exists. The workaround is to create a recursive snapshot of the root pool prior to the creation of additional BEs.
  • CR 6794452 - Fixed in the SXCE (Nevada, build 107) release and the Solaris 10 5/09 release, but still open in the Solaris 10 10/08 release, which means that you must send and receive individual root pool dataset snapshots. Apply kernel patch 139555-08 for SPARC and 139556-08 for x86 to resolve this CR in the Solaris 10 10/08 release. The resolution of this CR is to use the zfs receive -u option when restoring the root pool snapshots even when sending and receiving the entire recursive root pool snapshot.

Different ways exist to send and receive root pool snapshots:

  • You can send the snapshots to be stored as files in a pool on a remote system. Confirm that the remote snapshot files can be received on the local system in a test pool. See the first procedure below.
  • You can send the snapshots to be stored in a pool on a remote system. You might need to enable ssh for root on the remote system if you send the snapshots as root. During a root pool recovery, you will also need to enable remote services so that you can use rsh to receive the snapshots back to local system while booted from the miniroot. See the second procedure below.

Validating remotely stored snapshots as files or snapshots is an important step in root pool recovery. In either method, snapshots should be recreated on a routine basis, such as when the pool configuration changes or the Solaris OS is upgraded.

The following procedures have been tested with one ZFS BE.

[] Create Root Pool Snapshots (Stored Remotely as Files)

Create root pool snapshots to be stored as files in a pool on a remote system for recovery purposes. For example:

  1. Create space on a remote system to store the snapshots.
    remote# zfs create rpool/snaps
    remote# zfs list
    NAME          USED  AVAIL  REFER  MOUNTPOINT
    rpool         108K  8.24G    19K  /rpool
    rpool/snaps    18K  8.24G    18K  /rpool/snaps
    
  2. Share the space to the local system.
    remote# zfs set sharenfs='rw=local-system,root=local-system' rpool/snaps
    # share
    -@rpool/snaps   /rpool/snaps   sec=sys,rw=local-system,root=local-system   "" 
    
    If you are running a current OpenSolaris release, you will need to provide the fully qualified domain name for the local-system, in this example.
  3. Create a recursive snapshot of the root pool.
    local# zfs snapshot -r rpool@0316
    local# zpool set listsnapshots=on rpool
    local# zfs list
    NAME                           USED  AVAIL  REFER  MOUNTPOINT
    rpool                         17.6G   116G    67K  /rpool
    rpool@0316                        0      -    67K  -
    rpool/ROOT                    5.43G   116G    21K  legacy
    rpool/ROOT@0316                   0      -    21K  -
    rpool/ROOT/osolBE             5.43G   116G  4.68G  /
    rpool/ROOT/osolBE@install      773M      -  3.89G  -
    rpool/ROOT/osolBE@0316         290K      -  4.68G  -
    rpool/dump                    4.00G   116G  4.00G  -
    rpool/dump@0316                   0      -  4.00G  -
    rpool/export                  69.5K   116G    23K  /export
    rpool/export@0316                 0      -    23K  -
    rpool/export/home             46.5K   116G    23K  /export/home
    rpool/export/home@0316            0      -    23K  -
    rpool/export/home/admin       23.5K   116G  23.5K  /export/home/admin
    rpool/export/home/admin@0316      0      -  23.5K  -
    rpool/swap                    8.20G   124G  15.2M  -
    rpool/swap@0316                   0      -  15.2M  -
    
  4. Send the individual snapshots to the remote system. This method is required for a Solaris 10 10/08 system that does not have the above mentioned kernel patch level. Furthermore, this method provides the following advantages:
    • Allows recovery of individual datasets rather than the entire root pool should that be desired.
    • Does not send the dump and swap device snapshots, which is not necessary and adds time to the recursive method. But it does require manual creation of these datasets as outlined in the steps below.
    local# zfs send -Rv rpool/ROOT/@0316 > /net/remote-system/rpool/snaps/rpoolzfsBE.0316
    local# zfs send -Rv rpool/export@0316 > /net/remote-system/rpool/snaps/rpoolexport.0316
    

    If you are running the SXCE build 107 or the Solaris 10 5/09 or later release, you can send the entire recursive snapshot.

    # zfs send -Rv rpool@0316 > /net/remote-system/rpool/snaps/rpool.0316
[] Recreate Root Pool and Restore Root Pool Snapshots (Stored Remotely as Files)

In this scenario, assume the following conditions:

  • ZFS root pool cannot be recovered
  • ZFS root pool snapshots are stored on a remote system and shared over NFS
  • The system is booted from an equivalent Solaris release to the root pool version so that the Solaris release and the pool version match. Otherwise, you will need to add the -o version=version-number property option and value when you recreate the root pool in step 4 below.

All the steps below are performed on the local system.

  1. Boot from CD/DVD or the network.
    ok boot net

    or

    ok boot cdrom

    Then, exit out of the installation program.

  2. Mount the remote snapshot dataset.
    # mount -F nfs remote-system:/rpool/snaps /mnt
  3. If the root pool disk is replaced and does not contain a disk label that is usable by ZFS, you will have to relabel the disk. For more information, see .
  4. Recreate the root pool. For example:
    # zpool create -f -o failmode=continue -R /a -m legacy -o cachefile=/etc/zfs/zpool.cache rpool c1t1d0s0 
    
  5. If you had to replace or relabel the disk, then you might need to reinstall the boot blocks. For example:
    sparc# installboot -F zfs /usr/platform/`uname -i`/lib/fs/zfs/bootblk /dev/rdsk/c1t1d0s0
    x86# installgrub /boot/grub/stage1 /boot/grub/stage2 /dev/rdsk/c1t1d0s0
  6. Select one of the following if you are running the Solaris 10 10/08, Solaris 10 5/09, or the Nevada, build 107 or later release:
    • Solaris 10 10/08 or Nevada, prior to build 107: Receive the individual root pool snapshots. This step might take some time. For example:
      # cat /mnt/rpoolzfsBE.0316 | zfs receive -Fd rpool
      # cat /mnt/rpoolexport.0316 | zfs receive -Fd rpool
      
      Go to the next step.
    • Solaris 10 10/08 with above-mentioned kernel patch, Solaris 10 5/09 or later, Nevada, build 107 or later: Receive the recursive root pool snapshot. This step might take some time. For example:
      # cat /mnt/rpool.0316 | zfs receive -Fdu rpool

      Using the -u option means that the restored archive is not mounted when the zfs receive completes. The -u option is available starting in the Solaris 10 5/09 release.

      • If you want to modify something in the BE, you will need to explicitly mount them like this:
        # zfs mount rpool/ROOT/osolBE
        # zfs mount rpool/ROOT/osolBE/var
      • Then, mount everything in the pool that is not part of a BE.
        # zfs mount -a rpool

        Other BEs are not mounted since they all have canmount=noauto, which suppresses mounting when the zfs mount -a is done.

  7. Verify that the root pool datasets are restored:
    # zfs list
    NAME                           USED  AVAIL  REFER  MOUNTPOINT
    rpool                         17.6G   116G    67K  /rpool
    rpool@0316                        0      -    67K  -
    rpool/ROOT                    5.44G   116G    21K  legacy
    rpool/ROOT@0316                   0      -    21K  -
    rpool/ROOT/osolBE             5.44G   116G  4.68G  /
    rpool/ROOT/osolBE@install      773M      -  3.89G  -
    rpool/ROOT/osolBE@0316        3.69M      -  4.68G  -
    rpool/dump                    4.00G   116G  4.00G  -
    rpool/dump@0316                   0      -  4.00G  -
    rpool/export                   106K   116G    23K  /export
    rpool/export@0316               18K      -    23K  -
    rpool/export/home             64.5K   116G    23K  /export/home
    rpool/export/home@0316          18K      -    23K  -
    rpool/export/home/admin       23.5K   116G  23.5K  /export/home/admin
    rpool/export/home/admin@0316      0      -  23.5K  -
    rpool/swap                    8.20G   124G  15.2M  -
    rpool/swap@0316                   0      -  15.2M  -
    
  8. Set the bootfs property on the root pool BE.
    # zpool set bootfs=rpool/ROOT/osolBE rpool
  9. Recreate the dump device. This step is not necessary when using the recursive restore method because it is recreated with the rpool.
    # zfs create -V 2G rpool/dump
  10. Recreate the swap device. This step is not necessary when using the recursive restore method because it is recreated with the rpool.
    SPARC# zfs create -V 2G -b 8k rpool/swap
    x86# zfs create -V 2G -b 4k rpool/swap
  11. Reboot the system.
    # init 6
[] Create Root Pool Snapshots (Stored Remotely as Snapshots)

Create root pool snapshots and send them as snapshots to a pool on a remote system for recovery purposes. The remote system must be configured to allow ssh as root.

  1. Create a recursive root pool snapshot. For example:
    local# zfs snapshot -r rpool@0901
    local# zfs list -t all
    NAME                        USED  AVAIL  REFER  MOUNTPOINT
    rpool                      8.88G  58.1G    21K  legacy
    rpool@0901                     0      -    21K  -
    rpool/ROOT                 4.87G  58.1G    21K  legacy
    rpool/ROOT@0901                0      -    21K  -
    rpool/ROOT/zfs1009BE       4.87G  58.1G  4.87G  legacy
    rpool/ROOT/zfs1009BE@0901   519K      -  4.87G  -
    rpool/dump                 2.00G  58.1G  2.00G  -
    rpool/dump@0901                0      -  2.00G  -
    rpool/swap                 2.00G  60.1G    16K  -
    rpool/swap@0901                0      -    16K  -
  2. Send the root pool snapshots to a pool on a remote system. Only the root pool snapshots needed to recreate the BE environment are sent.
    local# zfs send rpool/ROOT@0901 | ssh remote-system zfs receive -Fdu tank
    Password: 
    zfs send rpool/ROOT/zfs1009BE@0901 | ssh remote-system zfs receive -Fdu tank
    Password:
    
  3. Confirm that the root pool snapshots are sent to the remote system.
    remote# zfs list -r tank
    tank                       4.87G   129G    23K  /tank
    tank/ROOT                  4.87G   129G    23K  /tank/ROOT
    tank/ROOT@0901               18K      -    21K  -
    tank/ROOT/zfs1009BE        4.87G   129G  4.87G  /tank/ROOT/zfs1009BE
    tank/ROOT/zfs1009BE@0901       0      -  4.87G  -
[] Recreate Root Pool and Restore Root Pool Snapshots (Stored Remotely as Snapshot)

In this scenario, assume the following conditions:

  • ZFS root pool cannot be recovered
  • ZFS root pool snapshots are stored on a remote system
  • The system is booted from an equivalent Solaris release to the root pool version so that the Solaris release and the pool version match. Otherwise, you will need to add the -o version=version-number property option and value when you recreate the root pool in step 4 below.
  • Remote system must enable rsh access. Or, configure ssh access in the miniroot on the local system and enable ssh root access on the remote system.
  1. Boot the local system from CD/DVD or the network.
    ok boot net

    or

    ok boot cdrom

    Then, exit out of the installation program.

  2. Configure the remote system to allow network services and rsh access.
    remote# netservices open
    remote# vi /.rhosts
    remote# vi /etc/hosts.equiv
  3. If the root pool disk is replaced and does not contain a disk label that is usable by ZFS, you will have to relabel the disk. For more information, see .
  4. Recreate the root pool. For example:
    local# zpool create -f -o failmode=continue -R /a -m legacy -o cachefile=/etc/zfs/zpool.cache rpool c1t1d0s0 
    
  5. If you had to replace or relabel the disk, then you might need to reinstall the boot blocks. For example:
    local-sparc# installboot -F zfs /usr/platform/`uname -i`/lib/fs/zfs/bootblk /dev/rdsk/c1t1d0s0
    local-x86# installgrub /boot/grub/stage1 /boot/grub/stage2 /dev/rdsk/c1t1d0s0
  6. Receive the snapshots from the remote system. For example:
    local# rsh remote-system zfs send tank/ROOT@0901 | zfs receive -Fdu rpool
    local# rsh remote-system zfs send tank/ROOT/zfs1009BE@0901 | zfs receive -Fdu rpool
    This step might take some time.
  7. Confirm that the snapshots are received from the remote system.
    local# zfs list
    NAME                        USED  AVAIL  REFER  MOUNTPOINT
    rpool                      4.87G  62.1G    21K  legacy
    rpool/ROOT                 4.87G  62.1G    21K  legacy
    rpool/ROOT@0901                0      -    21K  -
    rpool/ROOT/zfs1009BE       4.87G  62.1G  4.87G  legacy
    rpool/ROOT/zfs1009BE@0901      0      -  4.87G  -
  8. If you want to modify something in the BE, you will need to explicitly mount them like this:
    local# zfs mount rpool/ROOT/zfs1009BE
    local# zfs mount rpool/ROOT/zfs1009BE/var

    Then, mount everything in the pool that is not part of a BE.

    local# zfs mount -a rpool

    Other BEs are not mounted since they all have canmount=noauto, which suppresses mounting when the zfs mount -a is done.

  9. Set the bootfs property on the root pool BE.
    local# zpool set bootfs=rpool/ROOT/zfs1009BE rpool
  10. Recreate the dump device. This step is not necessary when using the recursive restore method because it is recreated with the rpool.
    local# zfs create -V 2G rpool/dump
  11. Recreate the swap device. This step is not necessary when using the recursive restore method because it is recreated with the rpool.
    local-SPARC# zfs create -V 2G -b 8k rpool/swap
    local-x86# zfs create -V 2G -b 4k rpool/swap
  12. Reboot the system.
    local# init 6
  13. Confirm that the root pool datasets are available. For example:
    local# zfs list -t all
    NAME                        USED  AVAIL  REFER  MOUNTPOINT
    rpool                      8.88G  58.1G    21K  legacy
    rpool@0901                     0      -    21K  -
    rpool/ROOT                 4.87G  58.1G    21K  legacy
    rpool/ROOT@0901                0      -    21K  -
    rpool/ROOT/zfs1009BE       4.87G  58.1G  4.87G  legacy
    rpool/ROOT/zfs1009BE@0901   519K      -  4.87G  -
    rpool/dump                 2.00G  58.1G  2.00G  -
    rpool/swap                 2.00G  60.1G    16K  -
    

[] Replacing/Relabeling the Root Pool Disk

You might need to replace a disk in the root pool for the following reasons:

  • The root pool is too small and you want to replace it with a larger disk
  • The root pool disk is failing. If the disk is failing so that the system won't boot, you'll need to boot from an alternate media, such as a CD or the network, before you replace the root pool disk.

Part of recovering the root pool might be to replace or relabel the root pool disk. Follow the steps below to relabel and replace the root pool disk.

  1. Physically attach the replacement disk.
  2. If the replacement disk has an EFI label, the fdisk output looks similar to the following on an x86 system.
    # fdisk /dev/rdsk/c1t1d0p0
    selecting c1t1d0p0
      Total disk size is 8924 cylinders
                 Cylinder size is 16065 (512 byte) blocks
    
                                                   Cylinders
          Partition   Status    Type          Start   End   Length    %
          =========   ======    ============  =====   ===   ======   ===
              1                 EFI               0  8924    8925    100
    .
    .
    .
    Enter Selection: 6

    Use fdisk to change this to a Solaris partition.

  3. Select one of the following to create a Solaris fdisk partition for a disk on an x86 system or create an SMI label for a disk on a SPARC system.
    • On an x86 system, create a Solaris fdisk partition that can be used for booting by selecting 1=SOLARIS2. You can create a Solaris partition by using the fdisk -B option that creates one Solaris partition that uses the whole disk. Beware that the following command uses the whole disk.
      # fdisk -B /dev/rdsk/c1t1d0p0

      Display the newly created Solaris partition. For example:

       Total disk size is 8924 cylinders
                   Cylinder size is 16065 (512 byte) blocks
      
                                                     Cylinders
            Partition   Status    Type          Start   End   Length    %
            =========   ======    ============  =====   ===   ======   ===
                1       Active    Solaris2          1  8923    8923    100
      .
      .
      .
      Enter Selection: 6
    • On a SPARC based system, make sure you have an SMI label. Use the format -e command to determine if the disk label is EFI or SMI and relabel the disk, if necessary. In the output below, the disk label includes sectors and not cylinders. This is an EFI label.
      # format -e
      Searching for disks...done
      AVAILABLE DISK SELECTIONS:
             0. c1t0d0 
                /pci@8,600000/SUNW,qlc@4/fp@0,0/ssd@w21000004cf7fac8a,0
             1. c1t1d0 
                /pci@8,600000/SUNW,qlc@4/fp@0,0/ssd@w21000004cf7fad21,0
      Specify disk (enter its number): 1
      selecting c1t1d0
      [disk formatted]
      format> p
      partition> p
      Current partition table (original):
      Total disk sectors available: 71116541 + 16384 (reserved sectors)
      
      Part      Tag    Flag     First Sector        Size        Last Sector
        0        usr    wm                34      33.91GB         71116541    
        1 unassigned    wm                 0          0              0    
        2 unassigned    wm                 0          0              0    
        3 unassigned    wm                 0          0              0    
        4 unassigned    wm                 0          0              0    
        5 unassigned    wm                 0          0              0    
        6 unassigned    wm                 0          0              0    
        7 unassigned    wm                 0          0              0    
        8   reserved    wm          71116542       8.00MB         71132925    
      partition> label
      [0] SMI Label
      [1] EFI Label
      Specify Label type[1]: 0
      Auto configuration via format.dat[no]? 
      Auto configuration via generic SCSI-2[no]? 
      partition> 
  4. Create a slice in the Solaris partition for the root pool. Creating a slice on x86 and SPARC system is similar except that an x86 system has a slice 8. In the example below, a slice 0 is created and the disk space is allocated to slice 0 on an x86 system. For a SPARC system, just ignore the slice 8 input.
    # format
    Specify disk (enter its number): 1
    selecting c1t1d0
    [disk formatted]
    FORMAT MENU:
            disk       - select a disk
            type       - select (define) a disk type
            partition  - select (define) a partition table
            current    - describe the current disk
            format     - format and analyze the disk
            fdisk      - run the fdisk program
    .
    .
    .
    format> p
    PARTITION MENU:
            0      - change `0' partition
            1      - change `1' partition
            2      - change `2' partition
            3      - change `3' partition
            4      - change `4' partition
            5      - change `5' partition
            6      - change `6' partition
            7      - change `7' partition
            select - select a predefined table
            modify - modify a predefined partition table
            name   - name the current table
            print  - display the current table
            label  - write partition map and label to the disk
            ! - execute , then return
            quit
    partition> p
    Current partition table (original):
    Total disk cylinders available: 8921 + 2 (reserved cylinders)
    
    Part      Tag    Flag     Cylinders        Size            Blocks
      0 unassigned    wm       0               0         (0/0/0)            0
      1 unassigned    wm       0               0         (0/0/0)            0
      2     backup    wu       0 - 8920       68.34GB    (8921/0/0) 143315865
      3 unassigned    wm       0               0         (0/0/0)            0
      4 unassigned    wm       0               0         (0/0/0)            0
      5 unassigned    wm       0               0         (0/0/0)            0
      6 unassigned    wm       0               0         (0/0/0)            0
      7 unassigned    wm       0               0         (0/0/0)            0
      8       boot    wu       0 -    0        7.84MB    (1/0/0)        16065
      9 unassigned    wm       0               0         (0/0/0)            0
    partition> modify
    Select partitioning base:
            0. Current partition table (original)
            1. All Free Hog
    Choose base (enter number) [0]? 1
    
    Part      Tag    Flag     Cylinders        Size            Blocks
      0       root    wm       0               0         (0/0/0)            0
      1       swap    wu       0               0         (0/0/0)            0
      2     backup    wu       0 - 8920       68.34GB    (8921/0/0) 143315865
      3 unassigned    wm       0               0         (0/0/0)            0
      4 unassigned    wm       0               0         (0/0/0)            0
      5 unassigned    wm       0               0         (0/0/0)            0
      6        usr    wm       0               0         (0/0/0)            0
      7 unassigned    wm       0               0         (0/0/0)            0
      8       boot    wu       0 -    0        7.84MB    (1/0/0)        16065
      9 alternates    wm       0               0         (0/0/0)            0
    
    Do you wish to continue creating a new partition
    table based on above table[yes]? 
    Free Hog partition[6]? 0
    Enter size of partition '1' [0b, 0c, 0.00mb, 0.00gb]: 
    Enter size of partition '3' [0b, 0c, 0.00mb, 0.00gb]: 
    Enter size of partition '4' [0b, 0c, 0.00mb, 0.00gb]: 
    Enter size of partition '5' [0b, 0c, 0.00mb, 0.00gb]: 
    Enter size of partition '6' [0b, 0c, 0.00mb, 0.00gb]: 
    Enter size of partition '7' [0b, 0c, 0.00mb, 0.00gb]: 
    
    Part      Tag    Flag     Cylinders        Size            Blocks
      0       root    wm       1 - 8920       68.33GB    (8920/0/0) 143299800
      1       swap    wu       0               0         (0/0/0)            0
      2     backup    wu       0 - 8920       68.34GB    (8921/0/0) 143315865
      3 unassigned    wm       0               0         (0/0/0)            0
      4 unassigned    wm       0               0         (0/0/0)            0
      5 unassigned    wm       0               0         (0/0/0)            0
      6        usr    wm       0               0         (0/0/0)            0
      7 unassigned    wm       0               0         (0/0/0)            0
      8       boot    wu       0 -    0        7.84MB    (1/0/0)        16065
      9 alternates    wm       0               0         (0/0/0)            0
    
    Okay to make this the current partition table[yes]? 
    Enter table name (remember quotes): "disk1"
    
    Ready to label disk, continue? yes
    partition> q
    format> q
    
  5. Attach the replacement disk.
    # zpool attach rpool c1t0d0s0 c1t1d0s0
    Please be sure to invoke installgrub(1M) to make 'c1t1d0s0' bootable.
  6. Check the resilvering status of the newly attached disk.
    # zpool status rpool
  7. After the disk resilvering is complete, install the boot blocks.
    x86# installgrub /boot/grub/stage1 /boot/grub/stage2 /dev/rdsk/c1t1d0s0
    sparc# installboot -F zfs /usr/platform/`uname -i`/lib/fs/zfs/bootblk /dev/rdsk/c1t1d0s0
  8. Confirm that you can boot from the replacement disk.
  9. Detach the smaller or unneeded disk.
    # zpool detach rpool c1t0d0s0
  10. [] Rolling Back a Root Pool Snapshot From a Failsafe Boot

    This procedure assumes that existing root pool snapshots are available. In this example, the root pool snapshots are available on the local system.

    # zfs snapshot -r rpool@0730
    # zfs list
    NAME                        USED  AVAIL  REFER  MOUNTPOINT
    rpool                      6.87G  60.1G    37K  /rpool
    rpool@0730                   18K      -    37K  -
    rpool/ROOT                 3.87G  60.1G    18K  legacy
    rpool/ROOT@0730                0      -    18K  -
    rpool/ROOT/zfs1008BE       3.87G  60.1G  3.82G  /
    rpool/ROOT/zfs1008BE@0730  52.3M      -  3.81G  -
    rpool/dump                 1.00G  60.1G  1.00G  -
    rpool/dump@0730              16K      -  1.00G  -
    rpool/export                 52K  60.1G    19K  /export
    rpool/export@0730            15K      -    19K  -
    rpool/export/home            18K  60.1G    18K  /export/home
    rpool/export/home@0730         0      -    18K  -
    rpool/swap                 2.00G  62.1G    16K  -
    rpool/swap@0730                0      -    16K  -
    
    1. Shutdown the system and boot failsafe mode.
      ok boot -F failsafe
      Multiple OS instances were found. To check and mount one of them
      read-write under /a, select it from the following list. To not mount
      any, select 'q'.
      
        1  /dev/dsk/c0t1d0s0             Solaris 10 xx SPARC
        2  rpool:5907401335443048350     ROOT/zfs1008
      
      Please select a device to be mounted (q for none) [?,??,q]: 2
      mounting rpool on /a
      
    2. Roll back the individual root pool snapshots.
      # zfs rollback rpool@0730
      # zfs rollback rpool/ROOT@0730
      # zfs rollback rpool/ROOT/zfs1008BE@0730
      # zfs rollback rpool/export@0730
      .
      .
      .
      Current ZFS snapshot rollback behavior is that recursive snapshots are not rolled back with the -r option. You must roll back the individual snapshots from the recursive snapshot.
    3. Reboot back to multiuser mode.
      # init 6
    4. [] Primary Mirror Disk in a ZFS Root Pool is Unavailable or Fails

      • If the primary disk in the pool fails, you might need to boot from the secondary disk by specifying the boot path. For example, on a SPARC system, a devalias is available to boot from the second disk as disk1.
      ok boot disk1
      • In some cases, you might need to remove the failed boot disk from the system to boot successfully from the other disk in the mirror.
      • If the system supports hot-plugging you can attempt to do a live replacement of a failed disk. On some systems, you might have to offline and unconfigure the disk before you physically replace it.
      # zpool offline rpool c0t0d0s0
      # cfgadm -c unconfigure c1::dsk/c0t0d0
      • Physically replace the primary disk. For example, c0t0d0s0.
      • Put a VTOC label on the new disk with format -e.
      • Reconfigure the disk and bring it online, if necessary. For example:
      # cfgadm -c configure c1::dsk/c0t0d0
      # zpool online rpool c0t0d0
      • Let ZFS know the primary disk was physically replaced at the same location.
      # zpool replace rpool c0t0d0s0
      • If the zpool replace step fails, detach and attach the primary mirror disk:
      # zpool detach rpool c0t0d0s0
      # zpool attach rpool c0t1d0s0 c0t0d0s0
      • Confirm that the disk is available and the disk is resilvered.
      # zpool status rpool
      • After the resilvering is complete, replace the bootblocks on the primary disk.
      SPARC# installboot -F zfs /usr/platform/`uname -i`/lib/fs/zfs/bootblk /dev/rdsk/c0t0d0s0
      x86# installgrub /boot/grub/stage1 /boot/grub/stage2 /dev/rdsk/c0t0d0s0
      • Confirm that you can boot from the primary disk.

      [] ZFS Swap and Dump Devices

      During an initial installation or a Live Upgrade migration, a swap volume and dump volume are created. The default sizes of the swap and dump volumes that are created by the Solaris installation program are as follows:

      • Swap volume size is calculated as half the size of physical memory, generally in the 512 MB to 2 GB range.
      • Dump volume size is calculated by the kernel based on dumpadm information and the size of physical memory.

      You can adjust the sizes of your swap and dump volumes in a JumpStart profile or during an initial installation to sizes of your choosing as long as the new sizes support system operation.

      # zfs list
      NAME                   USED  AVAIL  REFER  MOUNTPOINT
      rpool                 5.66G  27.6G  21.5K  /rpool
      rpool/ROOT            4.65G  27.6G    18K  /rpool/ROOT
      rpool/ROOT/zfs1008BE  4.65G  27.6G  4.65G  /
      rpool/dump             515M  27.6G   515M  -
      rpool/swap             513M  28.1G    16K  -
      

      [] Resizing ZFS Swap and Dump Devices

      • You can adjust the size of your swap and dump volumes during an initial installation.
      • You can create and size your swap and dump volumes before you do a Solaris Live Upgrade operation. ZFS dump volume performance is better when the volume is created with a 128-Kbyte block size. In SXCE, build 102, ZFS dump volumes are automatically created with a 128-Kbyte block size (CR 6725698). For example:
      # zpool create rpool mirror c0t0d0s0 c0t1d0s0
      /* The Solaris 10 10/08 dump creation syntax would be:
      # zfs create -V 2G -b 128k rpool/dump
      /* The SXCE build 102 dump creation syntax would be:
      # zfs create -V 2G rpool/dump
      SPARC# zfs create -V 2G -b 8k rpool/swap
      x86# zfs create -V 2G -b 4k rpool/swap
      
      • Activate your swap device, if necessary:
      # swap -a /dev/zvol/dsk/rpool/swap
      • Add the swap entry to the /etc/vfstab file similar to the following, if necessary:
      /dev/zvol/dsk/rpool/swap        -               -               swap    -       no
      • Activate your dump device, if necessary:
      # dumpadm -d /dev/zvol/dsk/rpool/dump
      • Solaris Live Upgrade does not resize existing swap and dump volumes. You can reset the volsize property of the swap and dump devices after a system is installed. For example:
      # zfs set volsize=2G rpool/dump
      # zfs get volsize rpool/dump
      NAME        PROPERTY  VALUE       SOURCE
      rpool/dump  volsize   2G          -
      
      • You can adjust the size of the swap and dump volumes in a JumpStart profile by using profile syntax similar to the following:
      install_type initial_install
      cluster SUNWCXall
      pool rpool 16g 2g 2g c0t0d0s0
      
      In this profile, the 2g and 2g entries set the size of the swap area and dump device as 2 Gbytes and 2 Gbytes, respectively.
      
      • You can adjust the size of your dump volume, but it might take some time, depending on the size of the dump volume. For example:
      # zfs set volsize=2G rpool/dump
      # zfs get volsize rpool/dump
      NAME        PROPERTY  VALUE       SOURCE
      rpool/dump  volsize   2G          -
      [] Adjusting the Size of the Swap Volume on an Active System

      If you need to adjust the size of the swap volume after installation on an active system, review the following steps. See CR 6765386 for more information.

      1. If your swap device is in use, then you might not be able to delete it. Check to see if the swap area is in use. For example:
        # swap -l
        swapfile                 dev    swaplo   blocks     free
        /dev/zvol/dsk/rpool/swap 182,2         8  4194296  4194296
        In the above output, blocks == free, so the swap device is not actually being used.
      2. If the swap area is not is use, remove the swap area. For example:
        # swap -d /dev/zvol/dsk/rpool/swap
      3. Confirm that the swap area is removed.
        # swap -l
        No swap devices configured
      4. Resize the swap volume. For example:
        # zfs set volsize=1G rpool/swap
      5. Activate the swap area.
        # swap -a /dev/zvol/dsk/rpool/swap
        # swap -l
        swapfile                 dev    swaplo   blocks     free
        /dev/zvol/dsk/rpool/swap 182,2         8  2097136  2097136

        The swap -a attempt might fail if the swap area is already listed in /etc/vfstab or is in use by Live Upgrade. In this case, use the swapadd feature instead.

        # /sbin/swapadd
        # swap -l
        swapfile                 dev  swaplo blocks   free
        /dev/zvol/dsk/rpool/swap 256,1      16 2097136 2097136
        

      If you need to add swap space but removing an existing swap device is difficult on a busy system, add another swap volume. For example:

      # zfs create -V 2G rpool/swap1
      # swap -a /dev/zvol/dsk/rpool/swap1
      # swap -l
      swapfile             dev  swaplo blocks   free
      /dev/zvol/dsk/rpool/swap 256,1      16 2097136 2097136
      /dev/zvol/dsk/rpool/swap1 256,5      16 4194288 4194288
      Add an entry for the second swap volume to the /etc/vfstab file.

      [] Destroying a Pool With an Active Dump/Swap Device

      If you want to destroy a ZFS root pool that is no longer needed, but it still has an active dump device and swap area, you'll need to use the dumpadm and swap commands to remove the dump device and swap area. Then, use these commands to establish a new dump device and swap area.

      # dumpadm -d swap
      # dumpadm -d none
      < destroy the root pool >
      # swap -a 
      # dumpadm -d swap
      
阅读(4288) | 评论(1) | 转发(0) |
给主人留下些什么吧!~~

chinaunix网友2010-06-07 16:35:55

很好