Evacuate Fails 'Invalid state of instance files' using Ceph Ephemeral RBD
Affects | Status | Importance | Assigned to | Milestone | |
---|---|---|---|---|---|
OpenStack Compute (nova) |
Fix Released
|
Medium
|
Feilong Wang | ||
Icehouse |
Fix Released
|
Undecided
|
Unassigned | ||
Juno |
Fix Released
|
Undecided
|
Unassigned |
Bug Description
Greetings,
We can't seem to be able to evacuate instances from a failed compute node using shared storage. We are using Ceph Ephemeral RBD as the storage medium.
Steps to reproduce:
nova evacuate --on-shared-storage 6e2081ec-
or
POST to http://
{"evacuate"
Here is what shows up in the logs:
180>Jul 10 20:36:48 node-24 nova-nova.
<179>Jul 10 20:36:48 node-24 nova-nova.
Traceback (most recent call last):
File "/usr/lib/
yield
File "/usr/lib/
_("Invalid state of instance files on shared"
InvalidSharedSt
<179>Jul 10 20:36:49 node-24 nova-oslo.
Traceback (most recent call last):
File "/usr/lib/
incoming.
File "/usr/lib/
return self._do_
File "/usr/lib/
result = getattr(endpoint, method)(ctxt, **new_args)
File "/usr/lib/
return function(self, context, *args, **kwargs)
File "/usr/lib/
return func(*args, **kwargs)
File "/usr/lib/
payload)
File "/usr/lib/
six.
File "/usr/lib/
return f(self, context, *args, **kw)
File "/usr/lib/
pass
File "/usr/lib/
six.
File "/usr/lib/
return function(self, context, *args, **kwargs)
File "/usr/lib/
function(self, context, *args, **kwargs)
File "/usr/lib/
e, sys.exc_info())
File "/usr/lib/
six.
File "/usr/lib/
return function(self, context, *args, **kwargs)
File "/usr/lib/
_("Invalid state of instance files on shared"
InvalidSharedSt
Changed in nova: | |
status: | New → Confirmed |
importance: | Undecided → Low |
Changed in nova: | |
assignee: | nobody → Fei Long Wang (flwang) |
Changed in nova: | |
importance: | Low → Medium |
tags: | added: juno-backport-potential |
tags: | added: cts |
tags: | removed: juno-backport-potential |
Changed in nova: | |
milestone: | none → kilo-1 |
status: | Fix Committed → Fix Released |
Changed in nova: | |
milestone: | kilo-1 → 2015.1.0 |
Was able to complete a workaround by
1. Edit nova.instances and replace all references of old node to destination node
2. reset-status of instance to active
3. Issue a hard-reboot to the instance
This will re-create the xml and console log on the destination node and boot the instance using the existing Ceph RBD