194 lines
5.7 KiB
Ruby
194 lines
5.7 KiB
Ruby
#
|
|
# Author:: Kyle Bader <kyle.bader@dreamhost.com>
|
|
# Cookbook Name:: ceph
|
|
# Recipe:: osd
|
|
#
|
|
# Copyright 2011, DreamHost Web Hosting
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
# you may not use this file except in compliance with the License.
|
|
# You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
|
|
# this recipe allows bootstrapping new osds, with help from mon
|
|
# Sample environment:
|
|
# #knife node edit ceph1
|
|
# "osd_devices": [
|
|
# {
|
|
# "device": "/dev/sdc"
|
|
# },
|
|
# {
|
|
# "device": "/dev/sdd",
|
|
# "dmcrypt": true,
|
|
# "journal": "/dev/sdd"
|
|
# }
|
|
# ]
|
|
|
|
include_recipe 'ceph::_common'
|
|
include_recipe 'ceph::osd_install'
|
|
include_recipe 'ceph::conf'
|
|
|
|
package 'gdisk' do
|
|
action :upgrade
|
|
end
|
|
|
|
package 'cryptsetup' do
|
|
action :upgrade
|
|
only_if { node['dmcrypt'] }
|
|
end
|
|
|
|
service_type = node['ceph']['osd']['init_style']
|
|
|
|
directory '/var/lib/ceph/bootstrap-osd' do
|
|
owner 'root'
|
|
group 'root'
|
|
mode '0755'
|
|
end
|
|
|
|
# TODO: cluster name
|
|
cluster = 'ceph'
|
|
|
|
execute 'format bootstrap-osd as keyring' do
|
|
command lazy { "ceph-authtool '/var/lib/ceph/bootstrap-osd/#{cluster}.keyring' --create-keyring --name=client.bootstrap-osd --add-key='#{osd_secret}'" }
|
|
creates "/var/lib/ceph/bootstrap-osd/#{cluster}.keyring"
|
|
only_if { osd_secret }
|
|
end
|
|
|
|
node_osds = node['ceph']['osd_devices']
|
|
|
|
if node_osds.nil? or node_osds.empty?
|
|
osd_device = []
|
|
ssd_disk = ssd_device
|
|
ssd_index = 0
|
|
# search normal osd device
|
|
node['block_device'].each do |device|
|
|
device_hash = Hash.new
|
|
device_name = device[0]
|
|
if device_name.include?"sd"
|
|
# whether the storage device is in use
|
|
device_ssd_flag = Mixlib::ShellOut.new("cat /sys/block/#{device_name}/queue/rotational").run_command.stdout.strip
|
|
device_partion_num = Mixlib::ShellOut.new("cat /proc/partitions | grep #{device_name} -c").run_command.stdout.strip
|
|
if device_partion_num == "1" and device_ssd_flag == "1"
|
|
device_hash['device'] = "/dev/#{device_name}"
|
|
unless ssd_disk.empty?
|
|
ssd_index = (ssd_index >= ssd_disk.length ? 0 : ssd_index)
|
|
ssd_partion = nil
|
|
while ssd_partion.nil?
|
|
if ssd_index >= ssd_disk.length
|
|
break
|
|
end
|
|
ssd_partion = create_disk_partion(ssd_disk[ssd_index])
|
|
ssd_index = ssd_index + 1
|
|
end
|
|
ssd_index = ssd_index + 1
|
|
end
|
|
device_hash['journal'] = ssd_partion unless ssd_partion.nil?
|
|
end
|
|
osd_device << device_hash unless device_hash.empty?
|
|
else
|
|
next
|
|
end
|
|
node.normal['ceph']['osd_devices'] = osd_device
|
|
node.save
|
|
node_osds = osd_device
|
|
Log.info("osd_devices are #{node['ceph']['osd_devices']}")
|
|
end
|
|
end
|
|
|
|
if crowbar?
|
|
node['crowbar']['disks'].each do |disk, _data|
|
|
execute "ceph-disk-prepare #{disk}" do
|
|
command "ceph-disk-prepare /dev/#{disk}"
|
|
only_if { node['crowbar']['disks'][disk]['usage'] == 'Storage' }
|
|
notifies :run, 'execute[udev trigger]', :immediately
|
|
end
|
|
|
|
ruby_block "set disk usage for #{disk}" do
|
|
block do
|
|
node.set['crowbar']['disks'][disk]['usage'] = 'ceph-osd'
|
|
node.save
|
|
end
|
|
end
|
|
end
|
|
|
|
execute 'udev trigger' do
|
|
command 'udevadm trigger --subsystem-match=block --action=add'
|
|
action :nothing
|
|
end
|
|
else
|
|
# Calling ceph-disk-prepare is sufficient for deploying an OSD
|
|
# After ceph-disk-prepare finishes, the new device will be caught
|
|
# by udev which will run ceph-disk-activate on it (udev will map
|
|
# the devices if dm-crypt is used).
|
|
# IMPORTANT:
|
|
# - Always use the default path for OSD (i.e. /var/lib/ceph/
|
|
# osd/$cluster-$id)
|
|
# - $cluster should always be ceph
|
|
# - The --dmcrypt option will be available starting w/ Cuttlefish
|
|
if node_osds
|
|
devices = node_osds
|
|
|
|
devices = Hash[(0...devices.size).zip devices] unless devices.kind_of? Hash
|
|
|
|
devices.each do |index, osd_device|
|
|
unless osd_device['status'].nil?
|
|
Log.info("osd: osd_device #{osd_device} has already been setup.")
|
|
next
|
|
end
|
|
|
|
directory osd_device['device'] do # ~FC022
|
|
owner 'root'
|
|
group 'root'
|
|
recursive true
|
|
only_if { osd_device['type'] == 'directory' }
|
|
end
|
|
|
|
dmcrypt = osd_device['encrypted'] == true ? '--dmcrypt' : ''
|
|
|
|
execute "ceph-disk-prepare on #{osd_device['device']}" do
|
|
command "ceph-disk-prepare #{dmcrypt} #{osd_device['device']} #{osd_device['journal']}"
|
|
action :run
|
|
notifies :create, "ruby_block[save osd_device status #{index}]", :immediately
|
|
end
|
|
|
|
execute "ceph-disk-activate #{osd_device['device']}" do
|
|
only_if { osd_device['type'] == 'directory' }
|
|
end
|
|
|
|
# we add this status to the node env
|
|
# so that we can implement recreate
|
|
# and/or delete functionalities in the
|
|
# future.
|
|
ruby_block "save osd_device status #{index}" do
|
|
block do
|
|
node.normal['ceph']['osd_devices'][index]['status'] = 'deployed'
|
|
node.save
|
|
end
|
|
action :nothing
|
|
end
|
|
end
|
|
else
|
|
Log.info('node["ceph"]["osd_devices"] empty')
|
|
end
|
|
end
|
|
|
|
service 'ceph_osd' do
|
|
case service_type
|
|
when 'upstart'
|
|
service_name 'ceph-osd-all-starter'
|
|
provider Chef::Provider::Service::Upstart
|
|
else
|
|
service_name 'ceph'
|
|
end
|
|
action [:enable, :start]
|
|
supports :restart => true
|
|
subscribes :restart, resources('template[/etc/ceph/ceph.conf]')
|
|
end
|