#!/usr/bin/python
-
+#
# Copyright (c) 2003 Intel Corporation
# All rights reserved.
#
import os, sys
import string
import popen2
-
+import time
from Exceptions import *
import utils
import BootServerRequest
-
+import BootAPI
import ModelOptions
def Run( vars, log ):
if( len(INSTALL_BLOCK_DEVICES) == 0 ):
raise ValueError, "INSTALL_BLOCK_DEVICES is empty"
- ROOT_SIZE= vars["ROOT_SIZE"]
+ # use vs_ROOT_SIZE or lxc_ROOT_SIZE as appropriate
+ varname=vars['virt']+"_ROOT_SIZE"
+ ROOT_SIZE= vars[varname]
if ROOT_SIZE == "" or ROOT_SIZE == 0:
raise ValueError, "ROOT_SIZE invalid"
if NODE_MODEL_OPTIONS & ModelOptions.RAWDISK:
VSERVERS_SIZE= "-1"
- if "VSERVER_SIZE" in vars:
+ if "VSERVERS_SIZE" in vars:
VSERVERS_SIZE= vars["VSERVERS_SIZE"]
if VSERVERS_SIZE == "" or VSERVERS_SIZE == 0:
raise ValueError, "VSERVERS_SIZE"
utils.sysexec_noerr( "lvremove -f %s" % PARTITIONS["swap"], log )
utils.sysexec_noerr( "lvremove -f %s" % PARTITIONS["vservers"], log )
utils.sysexec_noerr( "vgchange -an", log )
- utils.sysexec_noerr( "vgremove planetlab", log )
+ utils.sysexec_noerr( "vgremove -f planetlab", log )
log.write( "Running vgscan for devices\n" )
utils.sysexec_noerr( "vgscan", log )
# list of devices to be used with vgcreate
vg_device_list= ""
- # initialize the physical volumes
+ # get partitions
+ partitions = []
for device in used_devices:
-
part_path= get_partition_path_from_device( device, vars, log )
-
+ partitions.append(part_path)
+
+ # create raid partition
+ raid_partition = create_raid_partition(partitions, vars, log)
+ if raid_partition != None:
+ partitions = [raid_partition]
+ log.write("PARTITIONS %s\n" % str(partitions))
+ # initialize the physical volumes
+ for part_path in partitions:
if not create_lvm_physical_volume( part_path, vars, log ):
raise BootManagerException, "Could not create lvm physical volume " \
"on partition %s" % part_path
-
vg_device_list = vg_device_list + " " + part_path
# create an lvm volume group
# create swap logical volume
utils.sysexec( "lvcreate -L%s -nswap planetlab" % SWAP_SIZE, log )
- # create root logical volume
- utils.sysexec( "lvcreate -L%s -nroot planetlab" % ROOT_SIZE, log )
-
- if vars['NODE_MODEL_OPTIONS'] & ModelOptions.RAWDISK and VSERVERS_SIZE != "-1":
- utils.sysexec( "lvcreate -L%s -nvservers planetlab" % VSERVERS_SIZE, log )
+ # check if we want a separate partition for VMs
+ one_partition = (ROOT_SIZE == "-1")
+ if (one_partition):
remaining_extents= get_remaining_extents_on_vg( vars, log )
- utils.sysexec( "lvcreate -l%s -nrawdisk planetlab" % remaining_extents, log )
+ utils.sysexec( "lvcreate -l%s -nroot planetlab" % remaining_extents, log )
else:
- # create vservers logical volume with all remaining space
- # first, we need to get the number of remaining extents we can use
- remaining_extents= get_remaining_extents_on_vg( vars, log )
-
- utils.sysexec( "lvcreate -l%s -nvservers planetlab" % remaining_extents, log )
+ utils.sysexec( "lvcreate -L%s -nroot planetlab" % ROOT_SIZE, log )
+ if vars['NODE_MODEL_OPTIONS'] & ModelOptions.RAWDISK and VSERVERS_SIZE != "-1":
+ utils.sysexec( "lvcreate -L%s -nvservers planetlab" % VSERVERS_SIZE, log )
+ remaining_extents= get_remaining_extents_on_vg( vars, log )
+ utils.sysexec( "lvcreate -l%s -nrawdisk planetlab" % remaining_extents, log )
+ else:
+ # create vservers logical volume with all remaining space
+ # first, we need to get the number of remaining extents we can use
+ remaining_extents= get_remaining_extents_on_vg( vars, log )
+ utils.sysexec( "lvcreate -l%s -nvservers planetlab" % remaining_extents, log )
# activate volume group (should already be active)
#utils.sysexec( TEMP_PATH + "vgchange -ay planetlab", log )
# make swap
- utils.sysexec( "mkswap %s" % PARTITIONS["swap"], log )
+ utils.sysexec( "mkswap -f %s" % PARTITIONS["swap"], log )
# check if badhd option has been set
option = ''
# reserved-blocks-percentages
filesystems = {"root":5,"vservers":0}
- # make the file systems
- for fs in filesystems.keys():
- # get the reserved blocks percentage
- rbp = filesystems[fs]
- devname = PARTITIONS[fs]
+ # ROOT filesystem is always with ext2
+ fs = 'root'
+ rbp = filesystems[fs]
+ devname = PARTITIONS[fs]
+ log.write("formatting %s partition (%s)%s.\n" % (fs,devname,txt))
+ utils.sysexec( "mkfs.ext2 -q %s -m %d -j %s" % (option,rbp,devname), log )
+ # disable time/count based filesystems checks
+ utils.sysexec_noerr( "tune2fs -c -1 -i 0 %s" % devname, log)
+
+ # VSERVER filesystem with btrfs to support snapshoting and stuff
+ fs = 'vservers'
+ rbp = filesystems[fs]
+ devname = PARTITIONS[fs]
+ if vars['virt']=='vs':
log.write("formatting %s partition (%s)%s.\n" % (fs,devname,txt))
utils.sysexec( "mkfs.ext2 -q %s -m %d -j %s" % (option,rbp,devname), log )
+ # disable time/count based filesystems checks
+ utils.sysexec_noerr( "tune2fs -c -1 -i 0 %s" % devname, log)
+ elif (not one_partition):
+ log.write("formatting %s btrfs partition (%s).\n" % (fs,devname))
+ # early BootCD's seem to come with a version of mkfs.btrfs that does not support -f
+ # let's check for that before invoking it
+ mkfs="mkfs.btrfs"
+ if os.system("mkfs.btrfs --help 2>&1 | grep force")==0:
+ mkfs+=" -f"
+ mkfs+=" %s"%devname
+ utils.sysexec( mkfs, log )
+ # as of 2013/02 it looks like there's not yet an option to set fsck frequency with btrfs
# save the list of block devices in the log
log.write( "Block devices used (in lvm): %s\n" % repr(used_devices))
return 1
-
+import parted
def single_partition_device( device, vars, log ):
"""
initialize a disk by removing the old partition tables,
return 1 if sucessful, 0 otherwise
"""
- import parted
+ # two forms, depending on which version of pyparted we have
+ # v1 does not have a 'version' method
+ # v2 and above does, but to make it worse,
+ # parted-3.4 on f14 has parted.version broken and raises SystemError
+ try:
+ parted.version()
+ return single_partition_device_2_x (device, vars, log)
+ except AttributeError:
+ # old parted does not have version at all
+ return single_partition_device_1_x (device, vars, log)
+ except SystemError:
+ # let's assume this is >=2
+ return single_partition_device_2_x (device, vars, log)
+ except:
+ raise
+
+def single_partition_device_1_x ( device, vars, log):
lvm_flag= parted.partition_flag_get_by_name('lvm')
try:
+ log.write("Using pyparted 1.x\n")
# wipe the old partition table
utils.sysexec( "dd if=/dev/zero of=%s bs=512 count=1" % device, log )
+def single_partition_device_2_x ( device, vars, log):
+ try:
+ log.write("Using pyparted 2.x\n")
+
+ # Thierry june 2012 -- for disks larger than 2TB
+ # calling this with part_type='msdos' would fail at the maximizePartition stage
+ # create a new partition table
+ def partition_table (device, part_type, fs_type):
+ # wipe the old partition table
+ utils.sysexec( "dd if=/dev/zero of=%s bs=512 count=1" % device, log )
+ # get the device
+ dev= parted.Device(device)
+ disk = parted.freshDisk(dev,part_type)
+ # create one big partition on each block device
+ constraint= parted.constraint.Constraint (device=dev)
+ geometry = parted.geometry.Geometry (device=dev, start=0, end=1)
+ fs = parted.filesystem.FileSystem (type=fs_type,geometry=geometry)
+ new_part= parted.partition.Partition (disk, type=parted.PARTITION_NORMAL,
+ fs=fs, geometry=geometry)
+ # make it an lvm partition
+ new_part.setFlag(parted.PARTITION_LVM)
+ # actually add the partition to the disk
+ disk.addPartition(new_part, constraint)
+ disk.maximizePartition(new_part,constraint)
+ disk.commit()
+ log.write ("Current disk for %s - partition type %s\n%s\n"%(device,part_type,disk))
+ log.write ("Current dev for %s\n%s\n"%(device,dev))
+ del disk
+
+ try:
+ partition_table (device, 'msdos', 'ext2')
+ except:
+ partition_table (device, 'gpt', 'ext2')
+
+ except Exception, e:
+ log.write( "Exception inside single_partition_device_2_x : %s\n" % str(e) )
+ import traceback
+ traceback.print_exc(file=log)
+ return 0
+
+ return 1
+
+
+
def create_lvm_physical_volume( part_path, vars, log ):
"""
make the specificed partition a lvm physical volume.
return 1
+def create_raid_partition(partitions, vars, log):
+ """
+ create raid array using specified partitions.
+ """
+ raid_part = None
+ raid_enabled = False
+ node_tags = BootAPI.call_api_function( vars, "GetNodeTags",
+ ({'node_id': vars['NODE_ID']},))
+ for node_tag in node_tags:
+ if node_tag['tagname'] == 'raid_enabled' and \
+ node_tag['value'] == '1':
+ raid_enabled = True
+ break
+ if not raid_enabled:
+ return raid_part
+
+ try:
+ log.write( "Software raid enabled.\n" )
+ # wipe everything
+ utils.sysexec_noerr("mdadm --stop /dev/md0", log)
+ time.sleep(1)
+ for part_path in partitions:
+ utils.sysexec_noerr("mdadm --zero-superblock %s " % part_path, log)
+
+ # assume each partiton is on a separate disk
+ num_parts = len(partitions)
+ if num_parts < 2:
+ log.write( "Not enough disks for raid. Found: %s\n" % partitions )
+ raise BootManagerException("Not enough disks for raid. Found: %s\n" % partitions)
+ if num_parts == 2:
+ lvl = 1
+ else:
+ lvl = 5
+
+ # make the array
+ part_list = " ".join(partitions)
+ raid_part = "/dev/md0"
+ cmd = "mdadm --create %(raid_part)s --chunk=128 --level=raid%(lvl)s " % locals() + \
+ "--raid-devices=%(num_parts)s %(part_list)s" % locals()
+ utils.sysexec(cmd, log)
+
+ except BootManagerException, e:
+ log.write("create_raid_partition failed.\n")
+ raid_part = None
+
+ return raid_part
+
def get_partition_path_from_device( device, vars, log ):
"""