table of contents
other versions
- wheezy 2.5.2-1+deb7u2
gnt-instance(8) | Version 2.5.2 | gnt-instance(8) |
Name¶
gnt-instance - Ganeti instance administrationSynopsis¶
gnt-instance {command} [arguments...]DESCRIPTION¶
The gnt-instance command is used for instance administration in the Ganeti system.COMMANDS¶
Creation/removal/querying¶
ADD¶
add {-t|--disk-template {diskless | file | plain | drbd}} {--disk= N: {size=VAL | adopt= LV}[,vg=VG][,metavg=VG][,mode=ro|rw]| {-s|--os-size} SIZE} [--no-ip-check] [--no-name-check] [--no-start] [--no-install] [--net= N [:options...] | --no-nics] [{-B|--backend-parameters} BEPARAMS] [{-H|--hypervisor-parameters} HYPERVISOR [: option=value... ]] [{-O|--os-parameters} param=value... ] [--file-storage-dir dir_path] [--file-driver {loop | blktap}] {{-n|--node} node[:secondary-node] | {-I|--iallocator} name} {{-o|--os-type} os-type} [--submit] { instance} Creates a new instance on the specified host. The instance argument must be in DNS, but depending on the bridge/routing setup, need not be in the same network as the nodes in the cluster. The disk option specifies the parameters for the disks of the instance. The numbering of disks starts at zero, and at least one disk needs to be passed. For each disk, either the size or the adoption source needs to be given, and optionally the access mode (read-only or the default of read-write) and the LVM volume group can also be specified (via the vg key). For DRBD devices, a different VG can be specified for the metadata device using the metavg key. The size is interpreted (when no unit is given) in mebibytes. You can also use one of the suffixes m, g or t to specify the exact the units used; these suffixes map to mebibytes, gibibytes and tebibytes. When using the adopt key in the disk definition, Ganeti will reuse those volumes (instead of creating new ones) as the instance's disks. Ganeti will rename these volumes to the standard format, and (without installing the OS) will use them as-is for the instance. This allows migrating instances from non-managed mode (e.q. plain KVM with LVM) to being managed via Ganeti. Note that this works only for the `plain' disk template (see below for template details). Alternatively, a single-disk instance can be created via the -s option which takes a single argument, the size of the disk. This is similar to the Ganeti 1.2 version (but will only create one disk). The minimum disk specification is therefore --disk 0:size=20G (or -s 20G when using the -s option), and a three-disk instance can be specified as --disk 0:size=20G --disk 1:size=4G --disk 2:size=100G. The --no-ip-check skips the checks that are done to see if the instance's IP is not already alive (i.e. reachable from the master node). The --no-name-check skips the check for the instance name via the resolver (e.g. in DNS or /etc/hosts, depending on your setup). Since the name check is used to compute the IP address, if you pass this option you must also pass the --no-ip-check option. If you don't wat the instance to automatically start after creation, this is possible via the --no-start option. This will leave the instance down until a subsequent gnt-instance start command. The NICs of the instances can be specified via the --net option. By default, one NIC is created for the instance, with a random MAC, and set up according the the cluster level nic parameters. Each NIC can take these parameters (all optional):
- mac
- either a value or 'generate' to generate a new unique MAC
- ip
- specifies the IP address assigned to the instance from the Ganeti side (this is not necessarily what the instance will use, but what the node expects the instance to use)
- mode
- specifies the connection mode for this nic: routed or bridged.
- link
- in bridged mode specifies the bridge to attach this NIC to, in routed mode it's intended to differentiate between different routing tables/instance groups (but the meaning is dependent on the network script, see gnt-cluster(8) for more details)
- memory
- the memory size of the instance; as usual, suffixes can be used to denote the unit, otherwise the value is taken in mebibites
- vcpus
- the number of VCPUs to assign to the instance (if this value makes sense for the hypervisor)
- auto_balance
- whether the instance is considered in the N+1 cluster checks (enough redundancy in the cluster to survive a node failure)
- boot_order
- Valid for the Xen HVM and KVM hypervisors.
A string value denoting the boot order. This has different meaning for the Xen
HVM hypervisor and for the KVM one.
For Xen HVM, The boot order is a string of letters listing the boot devices,
with valid device letters being:
- a
- floppy drive
- c
- hard disk
- d
- CDROM drive
- n
- network boot (PXE)
- blockdev_prefix
- Valid for the Xen HVM and PVM hypervisors.
Relevant to non-pvops guest kernels, in which the disk device names are given by
the host. Allows one to specify 'xvd', which helps run Red Hat based
installers, driven by anaconda.
- floppy_image_path
- Valid for the KVM hypervisor.
The path to a floppy disk image to attach to the instance. This is useful to
install Windows operating systems on Virt/IO disks because you can specify
here the floppy for the drivers at installation time.
- cdrom_image_path
- Valid for the Xen HVM and KVM hypervisors.
The path to a CDROM image to attach to the instance.
- cdrom2_image_path
- Valid for the KVM hypervisor.
The path to a second CDROM image to attach to the instance. NOTE: This
image can't be used to boot the system. To do that you have to use the
'cdrom_image_path' option.
- nic_type
- Valid for the Xen HVM and KVM hypervisors.
This parameter determines the way the network cards are presented to the
instance. The possible options are:
- •
- rtl8139 (default for Xen HVM) (HVM & KVM)
- •
- ne2k_isa (HVM & KVM)
- •
- ne2k_pci (HVM & KVM)
- •
- i82551 (KVM)
- •
- i82557b (KVM)
- •
- i82559er (KVM)
- •
- pcnet (KVM)
- •
- e1000 (KVM)
- •
- paravirtual (default for KVM) (HVM & KVM)
- disk_type
- Valid for the Xen HVM and KVM hypervisors.
This parameter determines the way the disks are presented to the instance. The
possible options are:
- •
- ioemu [default] (HVM & KVM)
- •
- ide (HVM & KVM)
- •
- scsi (KVM)
- •
- sd (KVM)
- •
- mtd (KVM)
- •
- pflash (KVM)
- cdrom_disk_type
- Valid for the KVM hypervisor.
This parameter determines the way the cdroms disks are presented to the
instance. The default behavior is to get the same value of the eariler
parameter (disk_type). The possible options are:
- •
- paravirtual
- •
- ide
- •
- scsi
- •
- sd
- •
- mtd
- •
- pflash
- vnc_bind_address
- Valid for the Xen HVM and KVM hypervisors.
Specifies the address that the VNC listener for this instance should bind to.
Valid values are IPv4 addresses. Use the address 0.0.0.0 to bind to all
available interfaces (this is the default) or specify the address of one of
the interfaces on the node to restrict listening to that interface.
- vnc_tls
- Valid for the KVM hypervisor.
A boolean option that controls whether the VNC connection is secured with
TLS.
- vnc_x509_path
- Valid for the KVM hypervisor.
If vnc_tls is enabled, this options specifies the path to the x509
certificate to use.
- vnc_x509_verify
- Valid for the KVM hypervisor.
- spice_bind
- Valid for the KVM hypervisor.
Specifies the address or interface on which the SPICE server will listen. Valid
values are:
- •
- IPv4 addresses, including 0.0.0.0 and 127.0.0.1
- •
- IPv6 addresses, including :: and ::1
- •
- names of network interfaces
- spice_ip_version
- Valid for the KVM hypervisor.
Specifies which version of the IP protocol should be used by the SPICE server.
It is mainly intended to be used for specifying what kind of IP addresses should
be used if a network interface with both IPv4 and IPv6 addresses is specified
via the spice_bind parameter. In this case, if the
spice_ip_version parameter is not used, the default IP version of the
cluster will be used.
- acpi
- Valid for the Xen HVM and KVM hypervisors.
A boolean option that specifies if the hypervisor should enable ACPI support for
this instance. By default, ACPI is disabled.
- pae
- Valid for the Xen HVM and KVM hypervisors.
A boolean option that specifies if the hypervisor should enabled PAE support for
this instance. The default is false, disabling PAE support.
- use_localtime
- Valid for the Xen HVM and KVM hypervisors.
A boolean option that specifies if the instance should be started with its clock
set to the localtime of the machine (when true) or to the UTC (When false).
The default is false, which is useful for Linux/Unix machines; for Windows
OSes, it is recommended to enable this parameter.
- kernel_path
- Valid for the Xen PVM and KVM hypervisors.
This option specifies the path (on the node) to the kernel to boot the instance
with. Xen PVM instances always require this, while for KVM if this option is
empty, it will cause the machine to load the kernel from its disks (and the
boot will be done accordingly to boot_order).
- kernel_args
- Valid for the Xen PVM and KVM hypervisors.
This options specifies extra arguments to the kernel that will be loaded.
device. This is always used for Xen PVM, while for KVM it is only used if the
kernel_path option is also specified.
The default setting for this value is simply "ro", which mounts
the root disk (initially) in read-only one. For example, setting this to
single will cause the instance to start in single-user mode.
- initrd_path
- Valid for the Xen PVM and KVM hypervisors.
This option specifies the path (on the node) to the initrd to boot the instance
with. Xen PVM instances can use this always, while for KVM if this option is
only used if the kernel_path option is also specified. You can pass
here either an absolute filename (the path to the initrd) if you want to use
an initrd, or use the format no_initrd_path for no initrd.
- root_path
- Valid for the Xen PVM and KVM hypervisors.
This options specifies the name of the root device. This is always needed for
Xen PVM, while for KVM it is only used if the kernel_path option is
also specified.
Please note, that if this setting is an empty string and the hypervisor is Xen
it will not be written to the Xen configuration file
- serial_console
- Valid for the KVM hypervisor.
This boolean option specifies whether to emulate a serial console for the
instance.
- disk_cache
- Valid for the KVM hypervisor.
The disk cache mode. It can be either default to not pass any cache option to
KVM, or one of the KVM cache modes: none (for direct I/O), writethrough (to
use the host cache but report completion to the guest only when the host has
committed the changes to disk) or writeback (to use the host cache and report
completion as soon as the data is in the host cache). Note that there are
special considerations for the cache mode depending on version of KVM used and
disk type (always raw file under Ganeti), please refer to the KVM
documentation for more details.
- security_model
- Valid for the KVM hypervisor.
The security model for kvm. Currently one of none, user or
pool. Under none, the default, nothing is done and instances are
run as the Ganeti daemon user (normally root).
Under user kvm will drop privileges and become the user specified by the
security_domain parameter.
Under pool a global cluster pool of users will be used, making sure no
two instances share the same user on the same node. (this mode is not
implemented yet)
- security_domain
- Valid for the KVM hypervisor.
Under security model user the username to run the instance under. It must
be a valid username existing on the host.
Cannot be set under security model none or pool.
- kvm_flag
- Valid for the KVM hypervisor.
If enabled the -enable-kvm flag is passed to kvm. If disabled
-disable-kvm is passed. If unset no flag is passed, and the default running
mode for your kvm binary will be used.
- mem_path
- Valid for the KVM hypervisor.
This option passes the -mem-path argument to kvm with the path (on the node) to
the mount point of the hugetlbfs file system, along with the -mem-prealloc
argument too.
- use_chroot
- Valid for the KVM hypervisor.
This boolean option determines wether to run the KVM instance in a chroot
directory.
If it is set to true, an empty directory is created before starting the
instance and its path is passed via the -chroot flag to kvm. The directory is
removed when the instance is stopped.
It is set to false by default.
- migration_downtime
- Valid for the KVM hypervisor.
The maximum amount of time (in ms) a KVM instance is allowed to be frozen during
a live migration, in order to copy dirty memory pages. Default value is 30ms,
but you may need to increase this value for busy instances.
This option is only effective with kvm versions >= 87 and qemu-kvm versions
>= 0.11.0.
- cpu_mask
- Valid for the LXC hypervisor.
The processes belonging to the given instance are only scheduled on the
specified CPUs.
The parameter format is a comma-separated list of CPU IDs or CPU ID ranges. The
ranges are defined by a lower and higher boundary, separated by a dash. The
boundaries are inclusive.
- usb_mouse
- Valid for the KVM hypervisor.
This option specifies the usb mouse type to be used. It can be "mouse"
or "tablet". When using VNC it's recommended to set it to
"tablet".
- keymap
- Valid for the KVM hypervisor.
This option specifies the keyboard mapping to be used. It is only needed when
using the VNC console. For example: "fr" or "en-gb".
- reboot_behavior
- Valid for Xen PVM, Xen HVM and KVM hypervisors.
Normally if an instance reboots, the hypervisor will restart it. If this option
is set to exit, the hypervisor will treat a reboot as a shutdown
instead.
It is set to reboot by default.
The -O (--os-parameters) option allows customisation of the OS
parameters. The actual parameter names and values depends on the OS being
used, but the syntax is the same key=value. For example, setting a
hypothetical dhcp parameter to yes can be achieved by:
gnt-instance add -O dhcp=yes ... The -I (--iallocator) option specifies the instance allocator plugin to use. If you pass in this option the allocator will select nodes for this instance automatically, so you don't need to pass them with the -n option. For more information please refer to the instance allocator documentation. The -t (--disk-template) options specifies the disk layout type for the instance. The available choices are:
- diskless
- This creates an instance with no disks. Its useful for testing only (or other special cases).
- file
- Disk devices will be regular files.
- plain
- Disk devices will be logical volumes.
- drbd
- Disk devices will be drbd (version 8.x) on top of lvm volumes.
- loop
- Kernel loopback driver. This driver uses loopback devices to access the filesystem within the file. However, running I/O intensive applications in your instance using the loop driver might result in slowdowns. Furthermore, if you use the loopback driver consider increasing the maximum amount of loopback devices (on most systems it's 8) using the max_loop param.
- blktap
- The blktap driver (for Xen hypervisors). In order to be able to use the blktap driver you should check if the 'blktapctrl' user space disk agent is running (usually automatically started via xend). This user-level disk I/O interface has the advantage of better performance. Especially if you use a network file system (e.g. NFS) to store your instances this is the recommended choice.
# gnt-instance add -t file --disk 0:size=30g -B memory=512 -o debian-etch \
-n node1.example.com --file-storage-dir=mysubdir instance1.example.com
# gnt-instance add -t plain --disk 0:size=30g -B memory=512 -o debian-etch \
-n node1.example.com instance1.example.com
# gnt-instance add -t plain --disk 0:size=30g --disk 1:size=100g,vg=san \
-B memory=512 -o debian-etch -n node1.example.com instance1.example.com
# gnt-instance add -t drbd --disk 0:size=30g -B memory=512 -o debian-etch \
-n node1.example.com:node2.example.com instance2.example.com
BATCH-CREATE¶
batch-create {instances_file.json} This command (similar to the Ganeti 1.2 batcher tool) submits multiple instance creation jobs based on a definition file. The instance configurations do not encompass all the possible options for the add command, but only a subset. The instance file should be a valid-formed JSON file, containing a dictionary with instance name and instance parameters. The accepted parameters are:- disk_size
- The size of the disks of the instance.
- disk_template
- The disk template to use for the instance, the same as in the add command.
- backend
- A dictionary of backend parameters.
- hypervisor
- A dictionary with a single key (the hypervisor name), and as value the hypervisor options. If not passed, the default hypervisor and hypervisor options will be inherited.
- mac, ip, mode, link
- Specifications for the one NIC that will be created for the instance. 'bridge' is also accepted as a backwards compatibile key.
- nics
- List of nics that will be created for the instance. Each entry should be a dict, with mac, ip, mode and link as possible keys. Please don't provide the "mac, ip, mode, link" parent keys if you use this method for specifying nics.
- primary_node, secondary_node
- The primary and optionally the secondary node to use for the instance (in case an iallocator script is not used).
- iallocator
- Instead of specifying the nodes, an iallocator script can be used to automatically compute them.
- start
- whether to start the instance
- ip_check
- Skip the check for already-in-use instance; see the description in the add command for details.
- name_check
- Skip the name check for instances; see the description in the add command for details.
- file_storage_dir, file_driver
- Configuration for the file disk type, see the add command for details.
{
"instance3": {
"template": "drbd",
"os": "debootstrap",
"disk_size": ["25G"],
"iallocator": "dumb"
},
"instance5": {
"template": "drbd",
"os": "debootstrap",
"disk_size": ["25G"],
"iallocator": "dumb",
"hypervisor": "xen-hvm",
"hvparams": {"acpi": true},
"backend": {"memory": 512}
}
} The command will display the job id for each submitted instance, as follows:
# gnt-instance batch-create instances.json
instance3: 11224
instance5: 11225
REMOVE¶
remove [--ignore-failures] [--shutdown-timeout=N] [--submit] [--force] { instance} Remove an instance. This will remove all data from the instance and there is no way back. If you are not sure if you use an instance again, use shutdown first and leave it in the shutdown state for a while. The --ignore-failures option will cause the removal to proceed even in the presence of errors during the removal of the instance (e.g. during the shutdown or the disk removal). If this option is not given, the command will stop at the first error. The --shutdown-timeout is used to specify how much time to wait before forcing the shutdown (e.g. xm destroy in Xen, killing the kvm process for KVM, etc.). By default two minutes are given to each instance to stop. The --submit option is used to send the job to the master daemon but not wait for its completion. The job ID will be shown so that it can be examined via gnt-job info. The --force option is used to skip the interactive confirmation. Example:# gnt-instance remove instance1.example.com
LIST¶
list [--no-headers] [--separator= SEPARATOR] [--units=UNITS] [-v] [{-o|--output} [+]FIELD,...] [--filter] [instance...] Shows the currently configured instances with memory usage, disk usage, the node they are running on, and their run status. The --no-headers option will skip the initial header line. The --separator option takes an argument which denotes what will be used between the output fields. Both these options are to help scripting. The units used to display the numeric values in the output varies, depending on the options given. By default, the values will be formatted in the most appropriate unit. If the --separator option is given, then the values are shown in mebibytes to allow parsing by scripts. In both cases, the --units option can be used to enforce a given output unit. The -v option activates verbose mode, which changes the display of special field states (see ganeti(7)). The -o (--output) option takes a comma-separated list of output fields. The available fields and their meaning are:- admin_state
- Desired state of instance (if set, the instance should be up)
- be/auto_balance
- The "auto_balance" backend parameter
- be/memory
- The "memory" backend parameter
- be/vcpus
- The "vcpus" backend parameter
- beparams
- Backend parameters (merged)
- bridge
- Bridge of 1st network interface
- console
- Instance console information
- ctime
- Creation timestamp
- custom_beparams
- Custom backend parameters
- custom_hvparams
- Custom hypervisor parameters
- custom_nicparams
- Custom network interface parameters
- custom_osparams
- Custom operating system parameters
- disk.count
- Number of disks
- disk.size/0
- Disk size of 1st disk
- disk.size/1
- Disk size of 2nd disk
- disk.size/2
- Disk size of 3rd disk
- disk.size/3
- Disk size of 4th disk
- disk.size/4
- Disk size of 5th disk
- disk.size/5
- Disk size of 6th disk
- disk.size/6
- Disk size of 7th disk
- disk.size/7
- Disk size of 8th disk
- disk.size/8
- Disk size of 9th disk
- disk.size/9
- Disk size of 10th disk
- disk.size/10
- Disk size of 11th disk
- disk.size/11
- Disk size of 12th disk
- disk.size/12
- Disk size of 13th disk
- disk.size/13
- Disk size of 14th disk
- disk.size/14
- Disk size of 15th disk
- disk.size/15
- Disk size of 16th disk
- disk.sizes
- List of disk sizes
- disk_template
- Instance disk template
- disk_usage
- Total disk space used by instance on each of its nodes; this is not the disk size visible to the instance, but the usage on the node
- hv/acpi
- The "acpi" hypervisor parameter
- hv/blockdev_prefix
- The "blockdev_prefix" hypervisor parameter
- hv/boot_order
- The "boot_order" hypervisor parameter
- hv/bootloader_args
- The "bootloader_args" hypervisor parameter
- hv/bootloader_path
- The "bootloader_path" hypervisor parameter
- hv/cdrom2_image_path
- The "cdrom2_image_path" hypervisor parameter
- hv/cdrom_disk_type
- The "cdrom_disk_type" hypervisor parameter
- hv/cdrom_image_path
- The "cdrom_image_path" hypervisor parameter
- hv/cpu_mask
- The "cpu_mask" hypervisor parameter
- hv/device_model
- The "device_model" hypervisor parameter
- hv/disk_cache
- The "disk_cache" hypervisor parameter
- hv/disk_type
- The "disk_type" hypervisor parameter
- hv/floppy_image_path
- The "floppy_image_path" hypervisor parameter
- hv/init_script
- The "init_script" hypervisor parameter
- hv/initrd_path
- The "initrd_path" hypervisor parameter
- hv/kernel_args
- The "kernel_args" hypervisor parameter
- hv/kernel_path
- The "kernel_path" hypervisor parameter
- hv/keymap
- The "keymap" hypervisor parameter
- hv/kvm_flag
- The "kvm_flag" hypervisor parameter
- hv/mem_path
- The "mem_path" hypervisor parameter
- hv/migration_downtime
- The "migration_downtime" hypervisor parameter
- hv/nic_type
- The "nic_type" hypervisor parameter
- hv/pae
- The "pae" hypervisor parameter
- hv/reboot_behavior
- The "reboot_behavior" hypervisor parameter
- hv/root_path
- The "root_path" hypervisor parameter
- hv/security_domain
- The "security_domain" hypervisor parameter
- hv/security_model
- The "security_model" hypervisor parameter
- hv/serial_console
- The "serial_console" hypervisor parameter
- hv/spice_bind
- The "spice_bind" hypervisor parameter
- hv/spice_ip_version
- The "spice_ip_version" hypervisor parameter
- hv/usb_mouse
- The "usb_mouse" hypervisor parameter
- hv/use_bootloader
- The "use_bootloader" hypervisor parameter
- hv/use_chroot
- The "use_chroot" hypervisor parameter
- hv/use_localtime
- The "use_localtime" hypervisor parameter
- hv/vhost_net
- The "vhost_net" hypervisor parameter
- hv/vnc_bind_address
- The "vnc_bind_address" hypervisor parameter
- hv/vnc_password_file
- The "vnc_password_file" hypervisor parameter
- hv/vnc_tls
- The "vnc_tls" hypervisor parameter
- hv/vnc_x509_path
- The "vnc_x509_path" hypervisor parameter
- hv/vnc_x509_verify
- The "vnc_x509_verify" hypervisor parameter
- hvparams
- Hypervisor parameters (merged)
- hypervisor
- Hypervisor name
- ip
- IP address of 1st network interface
- mac
- MAC address of 1st network interface
- mtime
- Modification timestamp
- name
- Instance name
- network_port
- Instance network port if available (e.g. for VNC console)
- nic.bridge/0
- Bridge of 1st network interface
- nic.bridge/1
- Bridge of 2nd network interface
- nic.bridge/2
- Bridge of 3rd network interface
- nic.bridge/3
- Bridge of 4th network interface
- nic.bridge/4
- Bridge of 5th network interface
- nic.bridge/5
- Bridge of 6th network interface
- nic.bridge/6
- Bridge of 7th network interface
- nic.bridge/7
- Bridge of 8th network interface
- nic.bridges
- List containing each network interface's bridge
- nic.count
- Number of network interfaces
- nic.ip/0
- IP address of 1st network interface
- nic.ip/1
- IP address of 2nd network interface
- nic.ip/2
- IP address of 3rd network interface
- nic.ip/3
- IP address of 4th network interface
- nic.ip/4
- IP address of 5th network interface
- nic.ip/5
- IP address of 6th network interface
- nic.ip/6
- IP address of 7th network interface
- nic.ip/7
- IP address of 8th network interface
- nic.ips
- List containing each network interface's IP address
- nic.link/0
- Link of 1st network interface
- nic.link/1
- Link of 2nd network interface
- nic.link/2
- Link of 3rd network interface
- nic.link/3
- Link of 4th network interface
- nic.link/4
- Link of 5th network interface
- nic.link/5
- Link of 6th network interface
- nic.link/6
- Link of 7th network interface
- nic.link/7
- Link of 8th network interface
- nic.links
- List containing each network interface's link
- nic.mac/0
- MAC address of 1st network interface
- nic.mac/1
- MAC address of 2nd network interface
- nic.mac/2
- MAC address of 3rd network interface
- nic.mac/3
- MAC address of 4th network interface
- nic.mac/4
- MAC address of 5th network interface
- nic.mac/5
- MAC address of 6th network interface
- nic.mac/6
- MAC address of 7th network interface
- nic.mac/7
- MAC address of 8th network interface
- nic.macs
- List containing each network interface's MAC address
- nic.mode/0
- Mode of 1st network interface
- nic.mode/1
- Mode of 2nd network interface
- nic.mode/2
- Mode of 3rd network interface
- nic.mode/3
- Mode of 4th network interface
- nic.mode/4
- Mode of 5th network interface
- nic.mode/5
- Mode of 6th network interface
- nic.mode/6
- Mode of 7th network interface
- nic.mode/7
- Mode of 8th network interface
- nic.modes
- List containing each network interface's mode
- nic_link
- Link of 1st network interface
- nic_mode
- Mode of 1st network interface
- oper_ram
- Actual memory usage as seen by hypervisor
- oper_state
- Actual state of instance
- oper_vcpus
- Actual number of VCPUs as seen by hypervisor
- os
- Operating system
- osparams
- Operating system parameters (merged)
- pnode
- Primary node
- pnode.group
- Primary node's group
- pnode.group.uuid
- Primary node's group UUID
- sda_size
- Disk size of 1st disk
- sdb_size
- Disk size of 2nd disk
- serial_no
- Instance object serial number, incremented on each modification
- snodes
- Secondary nodes; usually this will just be one node
- snodes.group
- Node groups of secondary nodes
- snodes.group.uuid
- Node group UUIDs of secondary nodes
- status
- Instance status; "running" if instance is set to be running and actually is, "ADMIN_down" if instance is stopped and is not running, "ERROR_wrongnode" if instance running, but not on its designated primary node, "ERROR_up" if instance should be stopped, but is actually running, "ERROR_down" if instance should run, but doesn't, "ERROR_nodedown" if instance's primary node is down, "ERROR_nodeoffline" if instance's primary node is marked offline
- tags
- Tags
- uuid
- Instance UUID
- vcpus
- The "vcpus" backend parameter
LIST-FIELDS¶
list-fields [field...] Lists available fields for instances.INFO¶
info [-s | --static] [--roman] {--all | instance} Show detailed information about the given instance(s). This is different from list as it shows detailed data about the instance's disks (especially useful for the drbd disk template). If the option -s is used, only information available in the configuration file is returned, without querying nodes, making the operation faster. Use the --all to get info about all instances, rather than explicitly passing the ones you're interested in. The --roman option can be used to cause envy among people who like ancient cultures, but are stuck with non-latin-friendly cluster virtualization technologies.MODIFY¶
modify [{-H|--hypervisor-parameters} HYPERVISOR_PARAMETERS] [{-B|--backend-parameters} BACKEND_PARAMETERS] [--net add [:options] | --net remove | --net N:options] [--disk add:size= SIZE[,vg=VG][,metavg=VG] | --disk remove |--disk N:mode=MODE] [{-t|--disk-template} plain | {-t|--disk-template} drbd -n new_secondary] [--no-wait-for-sync] [--os-type= OS [--force-variant]] [{-O|--os-parameters} param=value... ] [--submit] { instance} Modifies the memory size, number of vcpus, ip address, MAC address and/or nic parameters for an instance. It can also add and remove disks and NICs to/from the instance. Note that you need to give at least one of the arguments, otherwise the command complains. The -H (--hypervisor-parameters), -B (--backend-parameters) and -O (--os-parameters) options specifies hypervisor, backend and OS parameter options in the form of name=value[,...]. For details which options can be specified, see the add command. The -t (--disk-template) option will change the disk template of the instance. Currently only conversions between the plain and drbd disk templates are supported, and the instance must be stopped before attempting the conversion. When changing from the plain to the drbd disk template, a new secondary node must be specified via the -n option. The option --no-wait-for-sync can be used when converting to the drbd template in order to make the instance available for startup before DRBD has finished resyncing. The --disk add:size=SIZE option adds a disk to the instance. The optional vg=VG option specifies LVM volume group other than default vg to create the disk on. For DRBD disks, the metavg=VG option specifies the volume group for the metadata device. The --disk remove option will remove the last disk of the instance. The --disk N:mode=MODE option will change the mode of the Nth disk of the instance between read-only ( ro) and read-write ( rw). The --net add:options option will add a new NIC to the instance. The available options are the same as in the add command (mac, ip, link, mode). The --net remove will remove the last NIC of the instance, while the --net N:options option will change the parameters of the Nth instance NIC. The option -o (--os-type) will change the OS name for the instance (without reinstallation). In case an OS variant is specified that is not found, then by default the modification is refused, unless --force-variant is passed. An invalid OS will also be refused, unless the --force option is given. The --submit option is used to send the job to the master daemon but not wait for its completion. The job ID will be shown so that it can be examined via gnt-job info. All the changes take effect at the next restart. If the instance is running, there is no effect on the instance.
REINSTALL¶
reinstall [{-o|--os-type} os-type] [--select-os] [-f force] [--force-multiple] [--instance | --node | --primary | --secondary | --all] [{-O|--os-parameters} OS_PARAMETERS] [--submit] { instance...} Reinstalls the operating system on the given instance(s). The instance(s) must be stopped when running this command. If the -o (--os-type) is specified, the operating system is changed. The --select-os option switches to an interactive OS reinstall. The user is prompted to select the OS template from the list of available OS templates. OS parameters can be overridden using -O (--os-parameters) (more documentation for this option under the add command). Since this is a potentially dangerous command, the user will be required to confirm this action, unless the -f flag is passed. When multiple instances are selected (either by passing multiple arguments or by using the --node, --primary, --secondary or --all options), the user must pass the --force-multiple options to skip the interactive confirmation. The --submit option is used to send the job to the master daemon but not wait for its completion. The job ID will be shown so that it can be examined via gnt-job info.RENAME¶
rename [--no-ip-check] [--no-name-check] [--submit] { instance} {new_name} Renames the given instance. The instance must be stopped when running this command. The requirements for the new name are the same as for adding an instance: the new name must be resolvable and the IP it resolves to must not be reachable (in order to prevent duplicate IPs the next time the instance is started). The IP test can be skipped if the --no-ip-check option is passed. The --no-name-check skips the check for the new instance name via the resolver (e.g. in DNS or /etc/hosts, depending on your setup) and that the resolved name matches the provided name. Since the name check is used to compute the IP address, if you pass this option you must also pass the --no-ip-check option. The --submit option is used to send the job to the master daemon but not wait for its completion. The job ID will be shown so that it can be examined via gnt-job info.Starting/stopping/connecting to console¶
STARTUP¶
startup [--force] [--ignore-offline] [--force-multiple] [--no-remember] [--instance | --node | --primary | --secondary | --all | --tags | --node-tags | --pri-node-tags | --sec-node-tags] [{-H|--hypervisor-parameters} key=value...] [{-B|--backend-parameters} key=value...] [--submit] [--paused] { name...} Starts one or more instances, depending on the following options. The four available modes are:- --instance
- will start the instances given as arguments (at least one argument required); this is the default selection
- --node
- will start the instances who have the given node as either primary or secondary
- --primary
- will start all instances whose primary node is in the list of nodes passed as arguments (at least one node required)
- --secondary
- will start all instances whose secondary node is in the list of nodes passed as arguments (at least one node required)
- --all
- will start all instances in the cluster (no arguments accepted)
- --tags
- will start all instances in the cluster with the tags given as arguments
- --node-tags
- will start all instances in the cluster on nodes with the tags given as arguments
- --pri-node-tags
- will start all instances in the cluster on primary nodes with the tags given as arguments
- --sec-node-tags
- will start all instances in the cluster on secondary nodes with the tags given as arguments
# gnt-instance start -H kernel_args="single" instance1
# gnt-instance start -B memory=2048 instance2 The first form will start the instance instance1 in single-user mode, and the instance instance2 with 2GB of RAM (this time only, unless that is the actual instance memory size already). Note that the values override the instance parameters (and not extend them): an instance with "kernel_args=ro" when started with -H kernel_args=single will result in "single", not "ro single". The --submit option is used to send the job to the master daemon but not wait for its completion. The job ID will be shown so that it can be examined via gnt-job info. The --paused option is only valid for Xen and kvm hypervisors. This pauses the instance at the start of bootup, awaiting gnt-instance console to unpause it, allowing the entire boot process to be monitored for debugging. Example:
# gnt-instance start instance1.example.com
# gnt-instance start --node node1.example.com node2.example.com
# gnt-instance start --all
SHUTDOWN¶
shutdown [--timeout= N] [--force-multiple] [--ignore-offline] [--no-remember] [--instance | --node | --primary | --secondary | --all | --tags | --node-tags | --pri-node-tags | --sec-node-tags] [--submit] { name...} Stops one or more instances. If the instance cannot be cleanly stopped during a hardcoded interval (currently 2 minutes), it will forcibly stop the instance (equivalent to switching off the power on a physical machine). The --timeout is used to specify how much time to wait before forcing the shutdown (e.g. xm destroy in Xen, killing the kvm process for KVM, etc.). By default two minutes are given to each instance to stop. The --instance, --node, --primary, --secondary, --all, --tags, --node-tags, --pri-node-tags and --sec-node-tags options are similar as for the startup command and they influence the actual instances being shutdown. The --submit option is used to send the job to the master daemon but not wait for its completion. The job ID will be shown so that it can be examined via gnt-job info. --ignore-offline can be used to ignore offline primary nodes and force the instance to be marked as stopped. This option should be used with care as it can lead to an inconsistent cluster state. The --no-remember option will perform the shutdown but not change the state of the instance in the configuration file (if it was running before, Ganeti will still thinks it needs to be running). This can be useful for a cluster-wide shutdown, where some instances are marked as up and some as down, and you don't want to change the running state: you just need to disable the watcher, shutdown all instances with --no-remember, and when the watcher is activated again it will restore the correct runtime state for all instances. Example:# gnt-instance shutdown instance1.example.com
# gnt-instance shutdown --all
REBOOT¶
reboot [{-t|--type} REBOOT-TYPE] [--ignore-secondaries] [--shutdown-timeout= N] [--force-multiple] [--instance | --node | --primary | --secondary | --all | --tags | --node-tags | --pri-node-tags | --sec-node-tags] [--submit] [ name...] Reboots one or more instances. The type of reboot depends on the value of -t (--type). A soft reboot does a hypervisor reboot, a hard reboot does a instance stop, recreates the hypervisor config for the instance and starts the instance. A full reboot does the equivalent of gnt-instance shutdown && gnt-instance startup. The default is hard reboot. For the hard reboot the option --ignore-secondaries ignores errors for the secondary node while re-assembling the instance disks. The --instance, --node, --primary, --secondary, --all, --tags, --node-tags, --pri-node-tags and --sec-node-tags options are similar as for the startup command and they influence the actual instances being rebooted. The --shutdown-timeout is used to specify how much time to wait before forcing the shutdown (xm destroy in xen, killing the kvm process, for kvm). By default two minutes are given to each instance to stop. The --force-multiple will skip the interactive confirmation in the case the more than one instance will be affected. Example:# gnt-instance reboot instance1.example.com
# gnt-instance reboot --type=full instance1.example.com
CONSOLE¶
console [--show-cmd] {instance} Connects to the console of the given instance. If the instance is not up, an error is returned. Use the --show-cmd option to display the command instead of executing it. For HVM instances, this will attempt to connect to the serial console of the instance. To connect to the virtualized "physical" console of a HVM instance, use a VNC client with the connection info from the info command. For Xen/kvm instances, if the instance is paused, this attempts to unpause the instance after waiting a few seconds for the connection to the console to be made. Example:# gnt-instance console instance1.example.com
Disk management¶
REPLACE-DISKS¶
replace-disks [--submit] [--early-release] {-p} [--disks idx] {instance} replace-disks [--submit] [--early-release] {-s} [--disks idx] {instance} replace-disks [--submit] [--early-release] {--iallocator name | --new-secondary NODE} {instance} replace-disks [--submit] [--early-release] {--auto} { instance} This command is a generalized form for replacing disks. It is currently only valid for the mirrored (DRBD) disk template. The first form (when passing the -p option) will replace the disks on the primary, while the second form (when passing the -s option will replace the disks on the secondary node. For these two cases (as the node doesn't change), it is possible to only run the replace for a subset of the disks, using the option --disks which takes a list of comma-delimited disk indices (zero-based), e.g. 0,2 to replace only the first and third disks. The third form (when passing either the --iallocator or the --new-secondary option) is designed to change secondary node of the instance. Specifying --iallocator makes the new secondary be selected automatically by the specified allocator plugin, otherwise the new secondary node will be the one chosen manually via the --new-secondary option. The fourth form (when using --auto) will automatically determine which disks of an instance are faulty and replace them within the same node. The --auto option works only when an instance has only faulty disks on either the primary or secondary node; it doesn't work when both sides have faulty disks. The --submit option is used to send the job to the master daemon but not wait for its completion. The job ID will be shown so that it can be examined via gnt-job info. The --early-release changes the code so that the old storage on secondary node(s) is removed early (before the resync is completed) and the internal Ganeti locks for the current (and new, if any) secondary node are also released, thus allowing more parallelism in the cluster operation. This should be used only when recovering from a disk failure on the current secondary (thus the old storage is already broken) or when the storage on the primary node is known to be fine (thus we won't need the old storage for potential recovery). Note that it is not possible to select an offline or drained node as a new secondary.ACTIVATE-DISKS¶
activate-disks [--submit] [--ignore-size] { instance} Activates the block devices of the given instance. If successful, the command will show the location and name of the block devices:node1.example.com:disk/0:/dev/drbd0
node1.example.com:disk/1:/dev/drbd1 In this example, node1.example.com is the name of the node on which the devices have been activated. The disk/0 and disk/1 are the Ganeti-names of the instance disks; how they are visible inside the instance is hypervisor-specific. /dev/drbd0 and /dev/drbd1 are the actual block devices as visible on the node. The --submit option is used to send the job to the master daemon but not wait for its completion. The job ID will be shown so that it can be examined via gnt-job info. The --ignore-size option can be used to activate disks ignoring the currently configured size in Ganeti. This can be used in cases where the configuration has gotten out of sync with the real-world (e.g. after a partially-failed grow-disk operation or due to rounding in LVM devices). This should not be used in normal cases, but only when activate-disks fails without it. Note that it is safe to run this command while the instance is already running.
DEACTIVATE-DISKS¶
deactivate-disks [-f] [--submit] {instance} De-activates the block devices of the given instance. Note that if you run this command for an instance with a drbd disk template, while it is running, it will not be able to shutdown the block devices on the primary node, but it will shutdown the block devices on the secondary nodes, thus breaking the replication. The -f/--force option will skip checks that the instance is down; in case the hypervisor is confused and we can't talk to it, normally Ganeti will refuse to deactivate the disks, but with this option passed it will skip this check and directly try to deactivate the disks. This can still fail due to the instance actually running or other issues. The --submit option is used to send the job to the master daemon but not wait for its completion. The job ID will be shown so that it can be examined via gnt-job info.GROW-DISK¶
grow-disk [--no-wait-for-sync] [--submit] { instance} {disk} { amount} Grows an instance's disk. This is only possible for instances having a plain or drbd disk template. Note that this command only change the block device size; it will not grow the actual filesystems, partitions, etc. that live on that disk. Usually, you will need to:- 1.
- use gnt-instance grow-disk
- 2.
- reboot the instance (later, at a convenient time)
- 3.
- use a filesystem resizer, such as ext2online(8) or xfs_growfs(8) to resize the filesystem, or use fdisk(8) to change the partition table on the disk
# gnt-instance grow-disk instance1.example.com 0 16g Also note that disk shrinking is not supported; use gnt-backup export and then gnt-backup import to reduce the disk size of an instance.
RECREATE-DISKS¶
- recreate-disks [--submit] [--disks=indices] [-n node1:[node2]]
- {instance}
Recovery¶
FAILOVER¶
failover [-f] [--ignore-consistency] [--shutdown-timeout= N] [--submit] { instance} Failover will stop the instance (if running), change its primary node, and if it was originally running it will start it again (on the new primary). This only works for instances with drbd template (in which case you can only fail to the secondary node) and for externally mirrored templates (shared storage) (which can change to any other node). Normally the failover will check the consistency of the disks before failing over the instance. If you are trying to migrate instances off a dead node, this will fail. Use the --ignore-consistency option for this purpose. Note that this option can be dangerous as errors in shutting down the instance will be ignored, resulting in possibly having the instance running on two machines in parallel (on disconnected DRBD drives). The --shutdown-timeout is used to specify how much time to wait before forcing the shutdown (xm destroy in xen, killing the kvm process, for kvm). By default two minutes are given to each instance to stop. The --submit option is used to send the job to the master daemon but not wait for its completion. The job ID will be shown so that it can be examined via gnt-job info. Example:# gnt-instance failover instance1.example.com
MIGRATE¶
migrate [-f] {--cleanup} {instance} migrate [-f] [--allow-failover] [--non-live] [--migration-mode=live|non-live] { instance} Migrate will move the instance to its secondary node without shutdown. It only works for instances having the drbd8 disk template type. The migration command needs a perfectly healthy instance, as we rely on the dual-master capability of drbd8 and the disks of the instance are not allowed to be degraded. The --non-live and --migration-mode=non-live options will switch (for the hypervisors that support it) between a "fully live" (i.e. the interruption is as minimal as possible) migration and one in which the instance is frozen, its state saved and transported to the remote node, and then resumed there. This all depends on the hypervisor support for two different methods. In any case, it is not an error to pass this parameter (it will just be ignored if the hypervisor doesn't support it). The option --migration-mode=live option will request a fully-live migration. The default, when neither option is passed, depends on the hypervisor parameters (and can be viewed with the gnt-cluster info command). If the --cleanup option is passed, the operation changes from migration to attempting recovery from a failed previous migration. In this mode, Ganeti checks if the instance runs on the correct node (and updates its configuration if not) and ensures the instances's disks are configured correctly. In this mode, the --non-live option is ignored. The option -f will skip the prompting for confirmation. If --allow-failover is specified it tries to fallback to failover if it already can determine that a migration wont work (i.e. if the instance is shutdown). Please note that the fallback will not happen during execution. If a migration fails during execution it still fails. Example (and expected output):# gnt-instance migrate instance1
Migrate will happen to the instance instance1. Note that migration is
**experimental** in this version. This might impact the instance if
anything goes wrong. Continue?
y/[n]/?: y
* checking disk consistency between source and target
* ensuring the target is in secondary mode
* changing disks into dual-master mode
- INFO: Waiting for instance instance1 to sync disks.
- INFO: Instance instance1's disks are in sync.
* migrating instance to node2.example.com
* changing the instance's disks on source node to secondary
- INFO: Waiting for instance instance1 to sync disks.
- INFO: Instance instance1's disks are in sync.
* changing the instance's disks to single-master
#
MOVE¶
move [-f] [--ignore-consistency] [-n node] [--shutdown-timeout= N] [--submit] {instance} Move will move the instance to an arbitrary node in the cluster. This works only for instances having a plain or file disk template. Note that since this operation is done via data copy, it will take a long time for big disks (similar to replace-disks for a drbd instance). The --shutdown-timeout is used to specify how much time to wait before forcing the shutdown (e.g. xm destroy in XEN, killing the kvm process for KVM, etc.). By default two minutes are given to each instance to stop. The --ignore-consistency option will make Ganeti ignore any errors in trying to shutdown the instance on its node; useful if the hypervisor is broken and you want to recuperate the data. The --submit option is used to send the job to the master daemon but not wait for its completion. The job ID will be shown so that it can be examined via gnt-job info. Example:# gnt-instance move -n node3.example.com instance1.example.com
CHANGE-GROUP¶
change-group [--iallocator NAME] [--to GROUP...] {instance} This command moves an instance to another node group. The move is calculated by an iallocator, either given on the command line or as a cluster default. If no specific destination groups are specified using --to, all groups except the one containing the instance are considered. Example:# gnt-instance change-group -I hail --to rack2 inst1.example.com
TAGS¶
ADD-TAGS¶
add-tags [--from file] {instancename} { tag...} Add tags to the given instance. If any of the tags contains invalid characters, the entire operation will abort. If the --from option is given, the list of tags will be extended with the contents of that file (each line becomes a tag). In this case, there is not need to pass tags on the command line (if you do, both sources will be used). A file name of - will be interpreted as stdin.LIST-TAGS¶
list-tags {instancename} List the tags of the given instance.REMOVE-TAGS¶
remove-tags [--from file] {instancename} { tag...} Remove tags from the given instance. If any of the tags are not existing on the node, the entire operation will abort. If the --from option is given, the list of tags to be removed will be extended with the contents of that file (each line becomes a tag). In this case, there is not need to pass tags on the command line (if you do, tags from both sources will be removed). A file name of - will be interpreted as stdin.REPORTING BUGS¶
Report bugs to project website (http://code.google.com/p/ganeti/) or contact the developers using the Ganeti mailing list (ganeti@googlegroups.com).SEE ALSO¶
Ganeti overview and specifications: ganeti(7) (general overview), ganeti-os-interface(7) (guest OS definitions). Ganeti commands: gnt-cluster(8) (cluster-wide commands), gnt-job(8) (job-related commands), gnt-node(8) (node-related commands), gnt-instance(8) (instance commands), gnt-os(8) (guest OS commands), gnt-group(8) (node group commands), gnt-backup(8) (instance import/export commands), gnt-debug(8) (debug commands). Ganeti daemons: ganeti-watcher(8) (automatic instance restarter), ganeti-cleaner(8) (job queue cleaner), ganeti-noded(8) (node daemon), ganeti-masterd(8) (master daemon), ganeti-rapi(8) (remote API daemon). Ganeti htools: htools(1) (generic binary), hbal(1) (cluster balancer), hspace(1) (capacity calculation), hail(1) (IAllocator plugin), hscan(1) (data gatherer from remote clusters).COPYRIGHT¶
Copyright (C) 2006, 2007, 2008, 2009, 2010, 2011 Google Inc. Permission is granted to copy, distribute and/or modify under the terms of the GNU General Public License as published by the Free Software Foundation; either version 2 of the License, or (at your option) any later version. On Debian systems, the complete text of the GNU General Public License can be found in /usr/share/common-licenses/GPL.Ganeti |