Gitweb:
http://git.fedorahosted.org/git/?p=gfs2-utils.git;a=commitdiff;h=34fa4222...
Commit: 34fa4222bdb012193ac593c6f188456c0609a425
Parent: 6727d5614a6fa0444f1e618656abd1dd6bdf495c
Author: Andrew Price <anprice(a)redhat.com>
AuthorDate: Tue Jul 23 11:34:24 2013 +0100
Committer: Andrew Price <anprice(a)redhat.com>
CommitterDate: Tue Jul 23 11:37:05 2013 +0100
gfs2-utils: Update docs
Update README.build to reflect the new dependencies and remove
references to gfs_controld.
Also remove usage.txt which mostly contained out of date information.
The rest is covered in the gfs2 man page and other docs.
Signed-off-by: Andrew Price <anprice(a)redhat.com>
---
README.build | 13 +----
doc/usage.txt | 165 ---------------------------------------------------------
2 files changed, 2 insertions(+), 176 deletions(-)
diff --git a/README.build b/README.build
index 1f73b1d..2aad6ab 100644
--- a/README.build
+++ b/README.build
@@ -12,18 +12,9 @@ Plus the following libraries:
bison
flex
zlib
+ libblkid
check (optional, enables the test suite)
-By default gfs_controld is not built as it is not required in Fedora 17 and
-later and cannot build on it. To re-enable it for older systems, use
---enable-gfs_controld when running the configure script (see below). The
-library dependencies for gfs_controld are only required when this option is
-given. The extra libraries upon which gfs_controld depends are:
-
- clusterlib
- corosynclib
- openaislib
-
To build gfs2-utils, run the following commands:
./autogen.sh
@@ -38,4 +29,4 @@ To install gfs2-utils, run:
make install
-See also README.contributing for details on submitting patches.
+See also doc/README.contributing for details on submitting patches.
diff --git a/doc/usage.txt b/doc/usage.txt
deleted file mode 100644
index 2ad5091..0000000
--- a/doc/usage.txt
+++ /dev/null
@@ -1,165 +0,0 @@
-How to install and run GFS2.
-
-Refer to the cluster project page for the latest information.
-http://sources.redhat.com/cluster/
-
-
-Install
--------
-
-Install a Linux kernel with GFS2, DLM, HOTPLUG, LBD, CONFIGFS, and
-optionally IPV6.
-
-Install openais
- get the latest "whitetank" (stable) release from
-
http://openais.org/
- or
- svn checkout
http://svn.osdl.org/openais
- cd openais/branches/whitetank
- make; make install DESTDIR=/
-
-Install gfs/dlm/fencing/etc components
- get the latest cluster-2.xx.yy tarball from
-
ftp://sources.redhat.com/pub/cluster/
- or
- cvs -d :pserver:cvs@sources.redhat.com:/cvs/cluster login cvs
- cvs -d :pserver:cvs@sources.redhat.com:/cvs/cluster checkout cluster
- the password is "cvs"
- cd cluster
- ./configure --kernel_src=/path/to/kernel
- make install
-
- NOTE: On 64-bit systems, you will usually need to add '--libdir=/usr/lib64'
- to the configure line.
-
-Install LVM2/CLVM (optional)
- cvs -d :pserver:cvs@sources.redhat.com:/cvs/lvm2 login cvs
- cvs -d :pserver:cvs@sources.redhat.com:/cvs/lvm2 checkout LVM2
- cvs -d :pserver:cvs@sources.redhat.com:/cvs/lvm2
- the password is "cvs"
- cd LVM2
- ./configure --with-clvmd=cman --with-cluster=shared
- make; make install
-
- NOTE: On 64-bit systems, you will usually need to add '--libdir=/usr/lib64'
- to the configure line.
-
-
- .... or alternatively, just get the packages from your friendly,
-neighbourhood distro, e.g. the gfs2-utils and cman packages from
-Fedora.
-
-Configuration
--------------
-
-Create /etc/cluster/cluster.conf and copy it to all nodes.
-
- The format and content of cluster.conf has changed little since the
- last generation of the software. See old example here:
-
http://sources.redhat.com/cluster/doc/usage.txt
- The one change you will need to make is to add nodeids for all nodes
- in the cluster. These are now mandatory. eg:
-
- <clusternode name="node12.mycluster.mycompany.com" votes="1"
nodeid="12">
-
- If you already have a cluster.conf file with no nodeids in it, then you can
- use the 'ccs_tool addnodeids' command to add them.
-
-
-Example cluster.conf
---------------------
-
-This is a basic cluster.conf file that requires manual fencing. The node
-names should resolve to the address on the network interface you want to
-use for openais/cman/dlm communication.
-
-<?xml version="1.0"?>
-<cluster name="alpha" config_version="1">
-
-<clusternodes>
-<clusternode name="node01" nodeid="1">
- <fence>
- </fence>
-</clusternode>
-
-<clusternode name="node02" nodeid="2">
- <fence>
- </fence>
-</clusternode>
-
-<clusternode name="node03" nodeid="3">
- <fence>
- </fence>
-</clusternode>
-</clusternodes>
-
-<fencedevices>
-</fencedevices>
-
-</cluster>
-
-
-Startup procedure
------------------
-
-Run these commands on each cluster node:
-
-> mount -t configfs none /sys/kernel/config
-> ccsd
-> cman_tool join
-> groupd
-> fenced
-> fence_tool join
-> dlm_controld
-> gfs_controld
-> clvmd (optional)
-> mkfs -t gfs2 -p lock_dlm -t <clustername>:<fsname> -j <#journals>
<blockdev>
-> mount -t gfs2 [-v] <blockdev> <mountpoint>
-
-Notes:
-- replace "gfs2" with "gfs" above to use gfs1 instead of gfs2
-- <clustername> in mkfs should match the one in cluster.conf.
-- <fsname> in mkfs is any name you pick, each fs must have a different name.
-- <#journals> in mkfs should be greater than or equal to the number of nodes
- that you want to mount this fs, each node uses a separate journal.
-- To avoid unnecessary fencing when starting the cluster, it's best for
- all nodes to join the cluster (complete cman_tool join) before any
- of them do fence_tool join.
-- The cman_tool "status" and "nodes" options show the status and
members
- of the cluster.
-- The group_tool command shows the status of fencing, dlm and gfs groups
- that the local node is part of.
-- The "cman" init script can be used for starting everything up through
- gfs_controld in the list above.
-
-
-Shutdown procedure
-------------------
-
-Run these commands on each cluster node:
-
-> umount [-v] <mountpoint>
-> fence_tool leave
-> cman_tool leave
-
-
-Converting from GFS1 to GFS2
-----------------------------
-
-If you have GFS1 filesystems that you need to convert to GFS2, follow
-this procedure:
-
-1. Back up your entire filesystem first.
- e.g. cp /dev/your_vg/lvol0 /your_gfs_backup
-
-2. Run fsck to ensure filesystem integrity.
- e.g. gfs2_fsck /dev/your_vg/lvol0
-
-3. Make sure the filesystem is not mounted from any node.
- e.g. for i in `grep "<clusternode name" /etc/cluster/cluster.conf | cut
-d '"' -f2` ; do ssh $i "mount | grep gfs" ; done
-
-4. Make sure you have the latest software versions.
-
-5. Run gfs2_convert <blockdev> from one of the nodes.
- e.g. gfs2_convert /dev/your_vg/lvol0
-