From 808753d4c0e87e200da8b6e2e37a84d7279f190f Mon Sep 17 00:00:00 2001
From: Manuel Franceschini <manuel.franceschini@gmail.com>
Date: Wed, 30 Apr 2008 12:51:07 +0000
Subject: [PATCH] Remove deprecated disk templates from doc

Since local_raid1 and remote_raid1 are deprecated they are removed
from the docs. This patch removes some old documentation sections
and bumps the documented version from 1.2 to 1.3.

Reviewed-by: iustinp
---
 doc/admin.sgml   | 46 +++++++----------------------------
 doc/install.sgml | 62 +++++++++++-------------------------------------
 2 files changed, 22 insertions(+), 86 deletions(-)

diff --git a/doc/admin.sgml b/doc/admin.sgml
index ccb885879..dd43da38f 100644
--- a/doc/admin.sgml
+++ b/doc/admin.sgml
@@ -4,7 +4,7 @@
   <articleinfo>
     <title>Ganeti administrator's guide</title>
   </articleinfo>
-  <para>Documents Ganeti version 1.2</para>
+  <para>Documents Ganeti version 1.3</para>
   <sect1>
     <title>Introduction</title>
 
@@ -173,19 +173,10 @@
         </varlistentry>
 
         <varlistentry>
-          <term>local_raid1</term>
-          <listitem>
-            <para>A local mirror is set between LVM devices to back the
-              instance. This provides some redundancy for the instance's
-              data.</para>
-          </listitem>
-        </varlistentry>
-
-        <varlistentry>
-          <term>remote_raid1</term>
+          <term>drbd</term>
           <listitem>
             <simpara><emphasis role="strong">Note:</emphasis> This is only
-              valid for multi-node clusters using drbd 0.7.</simpara>
+              valid for multi-node clusters using drbd 8.0.x</simpara>
             <simpara>
               A mirror is set between the local node and a remote one, which
               must be specified with the second value of the --node option. Use
@@ -195,29 +186,12 @@
           </listitem>
         </varlistentry>
 
-        <varlistentry>
-          <term>drbd</term>
-          <listitem>
-            <simpara><emphasis role="strong">Note:</emphasis> This is only
-              valid for multi-node clusters using drbd 8.0.</simpara>
-            <simpara>
-              This is similar to the
-              <replaceable>remote_raid1</replaceable> option, but uses
-              new features in drbd 8 to simplify the device
-              stack. From a user's point of view, this will improve
-              the speed of the <command>replace-disks</command>
-              command and (in future versions) provide more
-              functionality.
-            </simpara>
-          </listitem>
-        </varlistentry>
-
       </variablelist>
 
       <para>
         For example if you want to create an highly available instance use the
-        remote_raid1 or drbd disk templates:
-        <synopsis>gnt-instance add -n <replaceable>TARGET_NODE</replaceable><optional>:<replaceable>SECONDARY_NODE</replaceable></optional> -o <replaceable>OS_TYPE</replaceable> -t remote_raid1 \
+        drbd disk templates:
+        <synopsis>gnt-instance add -n <replaceable>TARGET_NODE</replaceable><optional>:<replaceable>SECONDARY_NODE</replaceable></optional> -o <replaceable>OS_TYPE</replaceable> -t drbd \
   <replaceable>INSTANCE_NAME</replaceable></synopsis>
 
       <para>
@@ -329,14 +303,10 @@
         failed, or you plan to remove a node from your cluster, and
         you failed over all its instances, but it's still secondary
         for some? The solution here is to replace the instance disks,
-        changing the secondary node. This is done in two ways, depending on the disk template type. For <literal>remote_raid1</literal>:
-
-        <synopsis>gnt-instance replace-disks <option>-n <replaceable>NEW_SECONDARY</replaceable></option> <replaceable>INSTANCE_NAME</replaceable></synopsis>
-
-        and for <literal>drbd</literal>:
-        <synopsis>gnt-instance replace-disks <option>-s</option> <option>-n <replaceable>NEW_SECONDARY</replaceable></option> <replaceable>INSTANCE_NAME</replaceable></synopsis>
+        changing the secondary node:
+        <synopsis>gnt-instance replace-disks <option>-s</option> <option>--new-secondary <replaceable>NODE</replaceable></option> <replaceable>INSTANCE_NAME</replaceable></synopsis>
 
-        This process is a bit longer, but involves no instance
+        This process is a bit long, but involves no instance
         downtime, and at the end of it the instance has changed its
         secondary node, to which it can if necessary be failed over.
       </para>
diff --git a/doc/install.sgml b/doc/install.sgml
index 445c3b4ae..af9d4e1e0 100644
--- a/doc/install.sgml
+++ b/doc/install.sgml
@@ -314,13 +314,8 @@ ln -s initrd.img-2.6.18-5-xen-686 initrd-2.6-xenU
       </para>
 
       <para>
-        Supported DRBD versions: the <literal>0.7</literal> series
-        <emphasis role="strong">or</emphasis>
-        <literal>8.0.7</literal>. It's recommended to have at least
-        version <literal>0.7.24</literal> if you use
-        <command>udev</command> since older versions have a bug
-        related to device discovery which can be triggered in cases of
-        hard drive failure.
+        Supported DRBD versions: <literal>8.0.x</literal>.
+        It's recommended to have at least version <literal>8.0.7</literal>.
       </para>
 
       <para>
@@ -336,36 +331,20 @@ ln -s initrd.img-2.6.18-5-xen-686 initrd-2.6-xenU
         you have the DRBD utils installed and the module in your
         kernel you're fine. Please check that your system is
         configured to load the module at every boot, and that it
-        passes the following option to the module (for
-        <literal>0.7.x</literal>:
-        <computeroutput>minor_count=64</computeroutput> (this will
-        allow you to use up to 32 instances per node) or for
-        <literal>8.0.x</literal> you can use up to
-        <constant>255</constant>
-        (i.e. <computeroutput>minor_count=255</computeroutput>, but
-        for most clusters <constant>128</constant> should be enough).
+        passes the following option to the module
+        <computeroutput>minor_count=255</computeroutput>. This will
+        allow you to use up to 128 instances per node (for most clusters
+        <constant>128 </constant> should be enough, though).
       </para>
 
       <formalpara><title>Debian</title>
         <para>
-         You can just install (build) the DRBD 0.7 module with the
+         You can just install (build) the DRBD 8.0.x module with the
          following commands (make sure you are running the Xen
          kernel):
         </para>
       </formalpara>
 
-      <screen>
-apt-get install drbd0.7-module-source drbd0.7-utils
-m-a update
-m-a a-i drbd0.7
-echo drbd minor_count=64 >> /etc/modules
-modprobe drbd minor_count=64
-      </screen>
-      <para>
-        or for using DRBD <literal>8.x</literal> from the etch
-        backports (note: you need at least 8.0.7, older version have
-        a bug that breaks ganeti's usage of drbd):
-      </para>
       <screen>
 apt-get install -t etch-backports drbd8-module-source drbd8-utils
 m-a update
@@ -376,7 +355,7 @@ modprobe drbd minor_count=128
 
       <para>
         It is also recommended that you comment out the default
-        resources in the <filename>/etc/dbrd.conf</filename> file, so
+        resources in the <filename>/etc/drbd.conf</filename> file, so
         that the init script doesn't try to configure any drbd
         devices. You can do this by prefixing all
         <literal>resource</literal> lines in the file with the keyword
@@ -427,19 +406,9 @@ skip resource "r1" {
           url="ftp://ftp.inr.ac.ru/ip-routing/iputils-current.tar.gz">arping</ulink>
           (part of iputils package)</simpara>
         </listitem>
-        <listitem>
-          <simpara><ulink
-          url="http://www.kernel.org/pub/linux/utils/raid/mdadm/">mdadm</ulink>
-          (Linux Software Raid tools)</simpara>
-        </listitem>
         <listitem>
           <simpara><ulink url="http://www.python.org">Python 2.4</ulink></simpara>
         </listitem>
-        <listitem>
-          <simpara><ulink url="http://twistedmatrix.com/">Python
-          Twisted library</ulink> - the core library is
-          enough</simpara>
-        </listitem>
         <listitem>
           <simpara><ulink
           url="http://pyopenssl.sourceforge.net/">Python OpenSSL
@@ -472,8 +441,7 @@ skip resource "r1" {
       </formalpara>
       <screen>
 # apt-get install lvm2 ssh bridge-utils iproute iputils-arping \
-  python2.4 python-twisted-core python-pyopenssl openssl \
-  mdadm python-pyparsing python-simplejson
+  python2.4 python-pyopenssl openssl python-pyparsing python-simplejson
       </screen>
 
     </sect2>
@@ -834,8 +802,7 @@ node1.example.com 197404 197404   2047  1896   125     0     0
         This step shows how to setup a virtual instance with either
         non-mirrored disks (<computeroutput>plain</computeroutput>) or
         with network mirrored disks
-        (<computeroutput>remote_raid1</computeroutput> for drbd 0.7
-        and <computeroutput>drbd</computeroutput> for drbd 8.x).  All
+        (<computeroutput>drbd</computeroutput>).  All
         commands need to be executed on the Ganeti master node (the
         one on which <computeroutput>gnt-cluster init</computeroutput>
         was run).  Verify that the OS scripts are present on all
@@ -872,14 +839,13 @@ creating os for instance inst1.example.com on node node1.example.com
       <para>
         To create a network mirrored instance, change the argument to
         the <option>-t</option> option from <literal>plain</literal>
-        to <literal>remote_raid1</literal> (drbd 0.7) or
-        <literal>drbd</literal> (drbd 8.0) and specify the node on
+        to <literal>drbd</literal> and specify the node on
         which the mirror should reside with the second value of the
         <option>--node</option> option, like this:
       </para>
 
       <screen>
-# gnt-instance add -t remote_raid1 -n node1:node2 -o debian-etch instance2
+# gnt-instance add -t drbd -n node1:node2 -o debian-etch instance2
 * creating instance disks...
 adding instance instance2 to cluster config
 Waiting for instance instance1 to sync disks.
@@ -919,8 +885,8 @@ creating os for instance instance2 on node node1.example.com
 
       <para>
         To failover an instance to its secondary node (only possible
-        in <literal>remote_raid1</literal> or <literal>drbd</literal>
-        disk templates), use <computeroutput>gnt-instance failover
+        with <literal>drbd</literal> disk templates), use
+        <computeroutput>gnt-instance failover
         <replaceable>INSTANCENAME</replaceable></computeroutput>.
       </para>
 
-- 
GitLab