[Lustre-discuss] LBUG ASSERTION(lock->l_resource != NULL) failed

Brock Palen brockp at umich.edu
Wed Jan 14 16:30:30 PST 2009


Gah!  Ok no problem.

No risk of data loss right?  And is there anyway to 'limp along' till  
an outage without rebooting OST's?

Thanks for the insight!

Brock Palen
www.umich.edu/~brockp
Center for Advanced Computing
brockp at umich.edu
(734)936-1985



On Jan 14, 2009, at 7:27 PM, Cliff White wrote:

> Brock Palen wrote:
>> I am having servers LBUG on a regular basis, Clients are running   
>> 1.6.6 patchless on RHEL4,  servers are running RHEL4 with 1.6.5.1   
>> RPM's from the download page.  All connection is over Ethernet,    
>> Servers are x4600's.
>
> This looks like bug 16496, which is fixed in 1.6.6. You should upgrade
> your servers to 1.6.6
> cliffw
>
>> The OSS that BUG'd has in its log:
>> Jan 13 16:35:39 oss2 kernel: LustreError: 10243:0:(ldlm_lock.c:  
>> 430:__ldlm_handle2lock()) ASSERTION(lock->l_resource != NULL) failed
>> Jan 13 16:35:39 oss2 kernel: LustreError: 10243:0:(tracefile.c:  
>> 432:libcfs_assertion_failed()) LBUG
>> Jan 13 16:35:39 oss2 kernel: Lustre: 10243:0:(linux-debug.c:  
>> 167:libcfs_debug_dumpstack()) showing stack for process 10243
>> Jan 13 16:35:39 oss2 kernel: ldlm_cn_08    R  running task        
>> 0  10243      1         10244  7776 (L-TLB)
>> Jan 13 16:35:39 oss2 kernel: 0000000000000000 ffffffffa0414629   
>> 00000103d83c7e00 0000000000000000
>> Jan 13 16:35:39 oss2 kernel:        00000101f8c88d40  
>> ffffffffa021445e  00000103e315dd98 0000000000000001
>> Jan 13 16:35:39 oss2 kernel:        00000101f3993ea0 0000000000000000
>> Jan 13 16:35:39 oss2 kernel: Call Trace:<ffffffffa0414629>  
>> {:ptlrpc:ptlrpc_server_handle_request+2457}
>> Jan 13 16:35:39 oss2 kernel:        <ffffffffa021445e>  
>> {:libcfs:lcw_update_time+30} <ffffffff80133855>{__wake_up_common+67}
>> Jan 13 16:35:39 oss2 kernel:        <ffffffffa0416d05>  
>> {:ptlrpc:ptlrpc_main+3989} <ffffffffa0415270>  
>> {:ptlrpc:ptlrpc_retry_rqbds+0}
>> Jan 13 16:35:39 oss2 kernel:        <ffffffffa0415270>  
>> {:ptlrpc:ptlrpc_retry_rqbds+0} <ffffffffa0415270>  
>> {:ptlrpc:ptlrpc_retry_rqbds+0}
>> Jan 13 16:35:39 oss2 kernel:        <ffffffff80110de3>{child_rip 
>> +8}  <ffffffffa0415d70>{:ptlrpc:ptlrpc_main+0}
>> Jan 13 16:35:39 oss2 kernel:        <ffffffff80110ddb>{child_rip+0}
>> Jan 13 16:35:40 oss2 kernel: LustreError: dumping log to /tmp/ 
>> lustre- log.1231882539.10243
>> At the same time a client (nyx346) lost contact with that oss, and  
>> is  never allowed to reconnect.
>> Client /var/log/message:
>> Jan 13 16:37:20 nyx346 kernel: Lustre: nobackup-OST000d-  
>> osc-000001022c2a7800: Connection to service nobackup-OST000d via  
>> nid  10.164.3.245 at tcp was lost; in progress operations using this  
>> service  will wait for recovery to complete.Jan 13 16:37:20 nyx346  
>> kernel:  Lustre: Skipped 6 previous similar messagesJan 13  
>> 16:37:20 nyx346  kernel: LustreError: 3889:0:(ldlm_request.c: 
>> 996:ldlm_cli_cancel_req ()) Got rc -11 from cancel RPC: canceling  
>> anywayJan 13 16:37:20  nyx346 kernel: LustreError: 3889:0: 
>> (ldlm_request.c: 1605:ldlm_cli_cancel_list())  
>> ldlm_cli_cancel_list: -11Jan 13 16:37:20  nyx346 kernel:  
>> LustreError: 11-0: an error occurred while  communicating with  
>> 10.164.3.245 at tcp. The ost_connect operation failed  with -16Jan 13  
>> 16:37:20 nyx346 kernel: LustreError: Skipped 10  previous similar  
>> messages
>> Jan 13 16:37:45 nyx346 kernel: Lustre: 3849:0:(import.c:  
>> 410:import_select_connection()) nobackup-OST000d-  
>> osc-000001022c2a7800: tried all connections, increasing latency to 7s
>> Even now the server(OSS) is refusing connection to OST00d,  with  
>> the  message:
>> Lustre: 9631:0:(ldlm_lib.c:760:target_handle_connect()) nobackup-  
>> OST000d: refuse reconnection from 145a1ec5-07ef-  
>> f7eb-0ca9-2a2b6503e0cd at 10.164.1.90@tcp to 0x00000103d5ce7000;  
>> still  busy with 2 active RPCs
>> If I reboot the OSS, the OST's on it go though recovery like  
>> normal,  and then the client is fine.
>> Network looks clean, found one machine with lots of dropped  
>> packets  between the servers, but that is not the client in question.
>> Thank you!  If it happens again, and I find any other data I will  
>> let  you know.
>> Brock Palen
>> www.umich.edu/~brockp
>> Center for Advanced Computing
>> brockp at umich.edu
>> (734)936-1985
>> _______________________________________________
>> Lustre-discuss mailing list
>> Lustre-discuss at lists.lustre.org
>> http://lists.lustre.org/mailman/listinfo/lustre-discuss
>
>
>




More information about the lustre-discuss mailing list