summaryrefslogtreecommitdiffstats
path: root/share/man/man4
diff options
context:
space:
mode:
authorjsing <jsing@openbsd.org>2012-04-06 15:42:56 +0000
committerjsing <jsing@openbsd.org>2012-04-06 15:42:56 +0000
commitad711cad4a6fab4bc1dc1244f3eb4c314e1bb89a (patch)
tree5903ab6af69124e95c07adad1c7bef8c252e115a /share/man/man4
parentRegen. (diff)
downloadwireguard-openbsd-ad711cad4a6fab4bc1dc1244f3eb4c314e1bb89a.tar.xz
wireguard-openbsd-ad711cad4a6fab4bc1dc1244f3eb4c314e1bb89a.zip
Remove raidframe related references.
Diffstat (limited to 'share/man/man4')
-rw-r--r--share/man/man4/Makefile4
-rw-r--r--share/man/man4/options.417
-rw-r--r--share/man/man4/raid.4353
3 files changed, 4 insertions, 370 deletions
diff --git a/share/man/man4/Makefile b/share/man/man4/Makefile
index 34c40460325..579d637ecdc 100644
--- a/share/man/man4/Makefile
+++ b/share/man/man4/Makefile
@@ -1,4 +1,4 @@
-# $OpenBSD: Makefile,v 1.531 2011/12/21 23:12:02 miod Exp $
+# $OpenBSD: Makefile,v 1.532 2012/04/06 15:42:56 jsing Exp $
MAN= aac.4 ac97.4 acphy.4 \
acpi.4 acpiac.4 acpiasus.4 acpibat.4 acpibtn.4 acpicpu.4 acpidock.4 \
@@ -44,7 +44,7 @@ MAN= aac.4 ac97.4 acphy.4 \
owtemp.4 pcagpio.4 pcaled.4 pcdisplay.4 pchb.4 pci.4 pcib.4 pcfadc.4 \
pcfiic.4 pciide.4 pckbc.4 pckbd.4 pcmcia.4 pcn.4 pcppi.4 pcscp.4 \
pf.4 pflog.4 pflow.4 pfsync.4 pgt.4 piixpm.4 pim.4 \
- pms.4 ppb.4 ppp.4 pppoe.4 pty.4 puc.4 pwdog.4 qsphy.4 radio.4 raid.4 \
+ pms.4 ppb.4 ppp.4 pppoe.4 pty.4 puc.4 pwdog.4 qsphy.4 radio.4 \
ral.4 random.4 ray.4 rd.4 rdac.4 re.4 rdcphy.4 rgephy.4 ricohrtc.4 \
rl.4 rlphy.4 route.4 rsu.4 rt.4 rum.4 run.4 \
rtfps.4 rtii.4 rtw.4 safe.4 safte.4 san.4 sbt.4 sbus.4 schsio.4 \
diff --git a/share/man/man4/options.4 b/share/man/man4/options.4
index dd7961431d7..e3ae9e6dade 100644
--- a/share/man/man4/options.4
+++ b/share/man/man4/options.4
@@ -1,4 +1,4 @@
-.\" $OpenBSD: options.4,v 1.225 2012/03/28 07:29:17 jmc Exp $
+.\" $OpenBSD: options.4,v 1.226 2012/04/06 15:42:56 jsing Exp $
.\" $NetBSD: options.4,v 1.21 1997/06/25 03:13:00 thorpej Exp $
.\"
.\" Copyright (c) 1998 Theo de Raadt
@@ -34,7 +34,7 @@
.\" THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
.\"
.\"
-.Dd $Mdocdate: March 28 2012 $
+.Dd $Mdocdate: April 6 2012 $
.Dt OPTIONS 4
.Os
.Sh NAME
@@ -174,11 +174,6 @@ control and observe another process.
See
.Xr ptrace 2
for details.
-.It Cd option RAIDDEBUG
-Be verbose on what RAIDframe does.
-See
-.Xr raid 4
-for details.
.It Cd option SMALL_KERNEL
Removes some features and some optimizations from the kernel to reduce the
size of the resulting kernel binary.
@@ -430,14 +425,6 @@ Makes the boot process more verbose for PCI peripherals
(vendor names and other information is printed, etc.).
.It Cd option PCMCIAVERBOSE
Makes the boot process more verbose for PCMCIA peripherals.
-.It Cd option RAID_AUTOCONFIG
-Adds support for auto-configuring the RAIDframe devices during the kernel
-initialization.
-See
-.Xr raid 4
-and
-.Xr raidctl 8
-for details.
.It Cd option USER_LDT
Enable userland manipulation of per-process
Local Descriptor Table (LDT) entries;
diff --git a/share/man/man4/raid.4 b/share/man/man4/raid.4
deleted file mode 100644
index 788ad9d2d8e..00000000000
--- a/share/man/man4/raid.4
+++ /dev/null
@@ -1,353 +0,0 @@
-.\" $OpenBSD: raid.4,v 1.32 2011/10/06 22:22:11 jmc Exp $
-.\" $NetBSD: raid.4,v 1.20 2001/09/22 16:03:58 wiz Exp $
-.\"
-.\" Copyright (c) 1998 The NetBSD Foundation, Inc.
-.\" All rights reserved.
-.\"
-.\" This code is derived from software contributed to The NetBSD Foundation
-.\" by Greg Oster
-.\"
-.\" Redistribution and use in source and binary forms, with or without
-.\" modification, are permitted provided that the following conditions
-.\" are met:
-.\" 1. Redistributions of source code must retain the above copyright
-.\" notice, this list of conditions and the following disclaimer.
-.\" 2. Redistributions in binary form must reproduce the above copyright
-.\" notice, this list of conditions and the following disclaimer in the
-.\" documentation and/or other materials provided with the distribution.
-.\"
-.\" THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
-.\" ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
-.\" TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
-.\" PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
-.\" BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-.\" CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-.\" SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-.\" INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-.\" CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-.\" ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-.\" POSSIBILITY OF SUCH DAMAGE.
-.\"
-.\"
-.\" Copyright (c) 1995 Carnegie-Mellon University.
-.\" All rights reserved.
-.\"
-.\" Author: Mark Holland
-.\"
-.\" Permission to use, copy, modify and distribute this software and
-.\" its documentation is hereby granted, provided that both the copyright
-.\" notice and this permission notice appear in all copies of the
-.\" software, derivative works or modified versions, and any portions
-.\" thereof, and that both notices appear in supporting documentation.
-.\"
-.\" CARNEGIE MELLON ALLOWS FREE USE OF THIS SOFTWARE IN ITS "AS IS"
-.\" CONDITION. CARNEGIE MELLON DISCLAIMS ANY LIABILITY OF ANY KIND
-.\" FOR ANY DAMAGES WHATSOEVER RESULTING FROM THE USE OF THIS SOFTWARE.
-.\"
-.\" Carnegie Mellon requests users of this software to return to
-.\"
-.\" Software Distribution Coordinator or Software.Distribution@CS.CMU.EDU
-.\" School of Computer Science
-.\" Carnegie Mellon University
-.\" Pittsburgh PA 15213-3890
-.\"
-.\" any improvements or extensions that they make and grant Carnegie the
-.\" rights to redistribute these changes.
-.\"
-.Dd $Mdocdate: October 6 2011 $
-.Dt RAID 4
-.Os
-.Sh NAME
-.Nm raid
-.Nd RAIDframe disk driver
-.Sh SYNOPSIS
-.Cd "pseudo-device raid" Op Ar count
-.Sh DESCRIPTION
-The
-.Nm
-driver provides RAID 0, 1, 4, and 5 (and more!) capabilities to
-.Ox .
-This
-document assumes that the reader has at least some familiarity with RAID
-and RAID concepts.
-The reader is also assumed to know how to configure
-disks and pseudo-devices into kernels, how to generate kernels, and how
-to partition disks.
-.Pp
-RAIDframe provides a number of different RAID levels including:
-.Bl -tag -width indent
-.It RAID 0
-provides simple data striping across the components.
-.It RAID 1
-provides mirroring.
-.It RAID 4
-provides data striping across the components, with parity
-stored on a dedicated drive (in this case, the last component).
-.It RAID 5
-provides data striping across the components, with parity
-distributed across all the components.
-.El
-.Pp
-There are a wide variety of other RAID levels supported by RAIDframe,
-including Even-Odd parity, RAID level 5 with rotated sparing, Chained
-declustering, and Interleaved declustering.
-The reader is referred to the RAIDframe documentation mentioned in the
-.Sx HISTORY
-section for more detail on these various RAID configurations.
-.Pp
-Depending on the parity level configured, the device driver can
-support the failure of component drives.
-The number of failures allowed depends on the parity level selected.
-If the driver is able to handle drive failures, and a drive does fail,
-then the system is operating in "degraded mode".
-In this mode, all missing data must be reconstructed from the data and
-parity present on the other components.
-This results in much slower data accesses, but does mean that a failure
-need not bring the system to a complete halt.
-.Pp
-The RAID driver supports and enforces the use of
-.Sq component labels .
-A
-.Sq component label
-contains important information about the component, including a
-user-specified serial number, the row and column of that component in
-the RAID set, and whether the data (and parity) on the component is
-.Sq clean .
-If the driver determines that the labels are very inconsistent with
-respect to each other (e.g. two or more serial numbers do not match)
-or that the component label is not consistent with its assigned place
-in the set (e.g., the component label claims the component should be
-the 3rd one of a 6-disk set, but the RAID set has it as the 3rd component
-in a 5-disk set) then the device will fail to configure.
-If the driver determines that exactly one component label seems to be
-incorrect, and the RAID set is being configured as a set that supports
-a single failure, then the RAID set will be allowed to configure, but
-the incorrectly labeled component will be marked as
-.Sq failed ,
-and the RAID set will begin operation in degraded mode.
-If all of the components are consistent among themselves, the RAID set
-will configure normally.
-.Pp
-Component labels are also used to support the auto-detection and
-auto-configuration of RAID sets.
-A RAID set can be flagged as auto-configurable, in which case it will be
-configured automatically during the kernel boot process.
-RAID filesystems which are
-automatically configured are also eligible to be the root filesystem.
-There is currently no support for booting a kernel directly from a RAID
-set.
-To use a RAID set as the root filesystem, a kernel is usually
-obtained from a small non-RAID partition, after which any
-auto-configuring RAID set can be used for the root filesystem.
-See
-.Xr raidctl 8
-for more information on auto-configuration of RAID sets.
-.Pp
-The driver supports
-.Sq hot spares ,
-disks which are on-line, but are not actively used in an existing
-filesystem.
-Should a disk fail, the driver is capable of reconstructing
-the failed disk onto a hot spare or back onto a replacement drive.
-If the components are hot swapable, the failed disk can then be
-removed, a new disk put in its place, and a copyback operation
-performed.
-The copyback operation, as its name indicates, will copy
-the reconstructed data from the hot spare to the previously failed
-(and now replaced) disk.
-Hot spares can also be hot-added using
-.Xr raidctl 8 .
-.Pp
-If a component cannot be detected when the RAID device is configured,
-that component will be simply marked as 'failed'.
-.Pp
-The user-land utility for doing all
-.Nm
-configuration and other operations
-is
-.Xr raidctl 8 .
-Most importantly,
-.Xr raidctl 8
-must be used with the
-.Fl i
-option to initialize all RAID sets.
-In particular, this initialization includes re-building the parity data.
-This rebuilding of parity data is also required when either a) a new RAID
-device is brought up for the first time or b) after an un-clean shutdown of a
-RAID device.
-By using the
-.Fl P
-option to
-.Xr raidctl 8 ,
-and performing this on-demand recomputation of all parity
-before doing a
-.Xr fsck 8
-or a
-.Xr newfs 8 ,
-filesystem integrity and parity integrity can be ensured.
-It bears repeating again that parity recomputation is
-.Ar required
-before any filesystems are created or used on the RAID device.
-If the parity is not correct, then missing data cannot be correctly recovered.
-.Pp
-RAID levels may be combined in a hierarchical fashion.
-For example, a RAID 0 device can be constructed out of a number of RAID 5
-devices (which, in turn, may be constructed out of the physical disks,
-or of other RAID devices).
-.Pp
-It is important that drives be hard-coded at their respective
-addresses (i.e., not left free-floating, where a drive with SCSI ID of
-4 can end up as
-.Pa /dev/sd0c )
-for well-behaved functioning of the RAID device.
-This is true for all types of drives, including IDE, HP-IB, etc.
-For normal SCSI drives, for example, the following can be used
-to fix the device addresses:
-.Bd -unfilled -offset indent
-sd0 at scsibus0 target 0 # SCSI disk drives
-sd1 at scsibus0 target 1 # SCSI disk drives
-sd2 at scsibus0 target 2 # SCSI disk drives
-sd3 at scsibus0 target 3 # SCSI disk drives
-sd4 at scsibus0 target 4 # SCSI disk drives
-sd5 at scsibus0 target 5 # SCSI disk drives
-sd6 at scsibus0 target 6 # SCSI disk drives
-.Ed
-.Pp
-See
-.Xr sd 4
-for more information.
-The rationale for fixing the device addresses is as follows:
-Consider a system with three SCSI drives at SCSI ID's 4, 5, and 6,
-and which map to components
-.Pa /dev/sd0e , /dev/sd1e ,
-and
-.Pa /dev/sd2e
-of a RAID 5 set.
-If the drive with SCSI ID 5 fails, and the system reboots, the old
-.Pa /dev/sd2e
-will show up as
-.Pa /dev/sd1e .
-The RAID driver is able to detect that component positions have changed, and
-will not allow normal configuration.
-If the device addresses are hard
-coded, however, the RAID driver would detect that the middle component
-is unavailable, and bring the RAID 5 set up in degraded mode.
-Note that the auto-detection and auto-configuration code does not care
-about where the components live.
-The auto-configuration code will
-correctly configure a device even after any number of the components
-have been re-arranged.
-.Pp
-The first step to using the
-.Nm
-driver is to ensure that it is suitably configured in the kernel.
-This is done by adding a line similar to:
-.Bd -unfilled -offset indent
-pseudo-device raid 4 # RAIDframe disk device
-.Ed
-.Pp
-to the kernel configuration file.
-The
-.Sq count
-argument (
-.Sq 4 ,
-in this case), specifies the number of RAIDframe drivers to configure.
-To turn on component auto-detection and auto-configuration of RAID
-sets, simply add:
-.Bd -unfilled -offset indent
-option RAID_AUTOCONFIG
-.Ed
-.Pp
-to the kernel configuration file.
-.Pp
-All component partitions must be of the type
-.Dv FS_BSDFFS
-(e.g., 4.2BSD) or
-.Dv FS_RAID
-(e.g., RAID).
-The use of the latter is strongly encouraged, and is
-required if auto-configuration of the RAID set is desired.
-Since RAIDframe leaves room for disklabels, RAID components can be simply
-raw disks, or partitions which use an entire disk.
-Note that some platforms (such as SUN) do not allow using the FS_RAID
-partition type.
-On these platforms, the
-.Nm
-driver can still auto-configure from FS_BSDFFS partitions.
-.Pp
-A more detailed treatment of actually using a
-.Nm
-device is found in
-.Xr raidctl 8 .
-It is highly recommended that the steps to reconstruct, copyback, and
-re-compute parity are well understood by the system administrator(s)
-.Ar before
-a component failure.
-Doing the wrong thing when a component fails may result in data loss.
-.Pp
-Additional debug information can be sent to the console by specifying:
-.Bd -unfilled -offset indent
-option RAIDDEBUG
-.Ed
-.Sh FILES
-.Bl -tag -width /dev/XXrXraidX -compact
-.It Pa /dev/{,r}raid*
-.Nm
-device special files.
-.El
-.Sh SEE ALSO
-.Xr sd 4 ,
-.Xr wd 4 ,
-.Xr config 8 ,
-.Xr fsck 8 ,
-.Xr MAKEDEV 8 ,
-.Xr mount 8 ,
-.Xr newfs 8 ,
-.Xr raidctl 8
-.Sh HISTORY
-The
-.Nm
-driver in
-.Ox
-is a port of RAIDframe, a framework for rapid prototyping of RAID
-structures developed by the folks at the Parallel Data Laboratory at
-Carnegie Mellon University (CMU).
-RAIDframe, as originally distributed
-by CMU, provides a RAID simulator for a number of different
-architectures, and a user-level device driver and a kernel device
-driver for Digital UNIX.
-The
-.Nm
-driver is a kernelized version of RAIDframe v1.1.
-.Pp
-A more complete description of the internals and functionality of
-RAIDframe is found in the paper "RAIDframe: A Rapid Prototyping Tool
-for RAID Systems", by William V. Courtright II, Garth Gibson, Mark
-Holland, LeAnn Neal Reilly, and Jim Zelenka, and published by the
-Parallel Data Laboratory of Carnegie Mellon University.
-The
-.Nm
-driver first appeared in
-.Nx 1.4
-from where it was ported to
-.Ox 2.5 .
-.Sh CAVEATS
-Certain RAID levels (1, 4, 5, 6, and others) can protect against some
-data loss due to component failure.
-However the loss of two components of a RAID 4 or 5 system, or the loss
-of a single component of a RAID 0 system, will result in the entire
-filesystems on that RAID device being lost.
-RAID is
-.Ar NOT
-a substitute for good backup practices.
-.Pp
-Recomputation of parity
-.Ar MUST
-be performed whenever there is a chance that it may have been
-compromised.
-This includes after system crashes, or before a RAID
-device has been used for the first time.
-Failure to keep parity correct will be catastrophic should a component
-ever fail -- it is better to use RAID 0 and get the additional space and
-speed, than it is to use parity, but not keep the parity correct.
-At least with RAID 0 there is no perception of increased data security.