[Openstack] [Swift] Unexplained 404s
Shrinand Javadekar
shrinand at maginatics.com
Wed May 25 06:51:20 UTC 2016
I was able to reproduce the issue with some manual intervention on the
same 1 node setup.
1. Using swift-get-nodes, I found the exact order of nodes in which
Swift was going to attempt to write an object.
2. Then I manually unmounted the primary and first handoff disk.
3. Then I wrote the object using the swift command line tool. It
succeeded. As expected, the object went into the second handoff node.
4. Then I remounted all the disks.
5. Then used the swift command line tool to issue a 'stat' on that
object. It failed with a 404 and logs showed that only the primary
node and first handoff node were accessed.
6. I then bumped up request_node_count to 4 and restarted the proxy server.
7. The same 'stat' command that failed in step 5 above now succeeded.
The logs showed that the object was accessed from the second handoff
node.
For now, we could live with adding request_node_count == disk count.
But it would be great to find if this can happen in non-single replica
setups too and why request_node_count affects reads and not writes.
Regardless, ClayG, you've been remarkable!! Thanks a ton!
-Shri
On Tue, May 24, 2016 at 10:46 PM, Mark Kirkwood
<mark.kirkwood at catalyst.net.nz> wrote:
> On 25/05/16 15:10, Clay Gerrard wrote:
>>
>> At the risk of repeating myself:
>>
>> On Tue, May 24, 2016 at 5:30 PM, Clay Gerrard <clay.gerrard at gmail.com
>> <mailto:clay.gerrard at gmail.com>> wrote:
>>
>>
>> This inconsistency in search depth based on the per-worker error
>> limiting may be something worth looking into generally - but it's
>> probably mostly hidden on clusters that are going 3 or more nodes
>> deep into handoffs in the default case.
>>
>>
>> I just don't know how much of an issue it is for clusters where each
>> device in the nodes iter is going to be on a separate node, probably in
>> independent failure domais - and you've got to not only wiff multiple
>> primaries but multiple handoffs too. More over - you'd have to have *all*
>> the replicas get past request_node_count or else one of the earlier replicas
>> is going to service the read anyway without ever noticing the write that
>> landed further out. Or even if my theory about error limiting maybe
>> allowing a PUT to go deeper than the request_node_count holds...
>>
>> Obviously the data got on second handoff device somehow - it be
>> interesting to see the transaction logs for the write that did that - but
>> there's a lot of ifs in there and I'm not sure it's an issue except in the
>> single replica single node case. Still sorta interesting I guess...
>>
>>
>
> Yeah, single replica is pretty special (maybe useful for teasing out some
> unusual bugs to examine).
>
> I'll have another look at a slow system I have here (3 replicas, 3 nodes
> each with 6 devices or thereabouts) that *was* exhibiting 404's when trying
> to read just created containers or objects. *If* it is not immediately
> misconfigured and if I can still reproduce the behaviour I'll post (new
> thread to avoid confusion).
>
> regards
>
> Mark
More information about the Openstack
mailing list