[lustre-discuss] problems to mount MDS and MDT

Colin Faber cfaber at gmail.com
Mon May 17 12:50:03 PDT 2021


It appears part of the debug data is missing (the part before you posted
it), Can you try again, lctl dk > /dev/null to clear it then try your mount
and grab the debug again?

On Mon, May 17, 2021 at 1:35 PM Abdeslam Tahari <abeslam at gmail.com> wrote:

> Thank you Colin
>
> No i don't have iptables or rules
>
> firewalled is stopped selinux disabled as well
>  iptables -L
> Chain INPUT (policy ACCEPT)
> target     prot opt source               destination
>
> Chain FORWARD (policy ACCEPT)
> target     prot opt source               destination
>
> Chain OUTPUT (policy ACCEPT)
> target     prot opt source               destination
>
>
> Regards
>
>
> Regards
>
> Le lun. 17 mai 2021 à 21:29, Colin Faber <cfaber at gmail.com> a écrit :
>
>> Firewall rules dealing with localhost?
>>
>> On Mon, May 17, 2021 at 11:33 AM Abdeslam Tahari via lustre-discuss <
>> lustre-discuss at lists.lustre.org> wrote:
>>
>>> Hello
>>>
>>> i have a problem to mount the mds/mdt luster, it wont mount at all and
>>> there is no message errors at the console
>>>
>>> -it does not show errors or messages while mounting it
>>>
>>> here are some debug file logs
>>>
>>>
>>> i specify it is a new project that i am doing.
>>>
>>> the version and packages of luter installed:
>>> kmod-lustre-2.12.5-1.el7.x86_64
>>> kernel-devel-3.10.0-1127.8.2.el7_lustre.x86_64
>>> lustre-2.12.5-1.el7.x86_64
>>> lustre-resource-agents-2.12.5-1.el7.x86_64
>>> kernel-3.10.0-1160.2.1.el7_lustre.x86_64
>>> kernel-debuginfo-common-x86_64-3.10.0-1160.2.1.el7_lustre.x86_64
>>> kmod-lustre-osd-ldiskfs-2.12.5-1.el7.x86_64
>>> kernel-3.10.0-1127.8.2.el7_lustre.x86_64
>>> lustre-osd-ldiskfs-mount-2.12.5-1.el7.x86_64
>>>
>>>
>>>
>>> the system(os) Centos 7
>>>
>>> the kernel
>>> Linux lustre-mds1 3.10.0-1127.8.2.el7_lustre.x86_64
>>>  cat /etc/redhat-release
>>>
>>>
>>> when i mount the luster file-system it wont show up and no errors
>>>
>>> mount -t lustre /dev/sda /mds
>>>
>>> lctl dl  does not show up
>>>
>>> df -h   no mount point for /dev/sda
>>>
>>>
>>> lctl dl
>>>
>>> shows this:
>>> lctl dl
>>>   0 UP osd-ldiskfs lustre-MDT0000-osd lustre-MDT0000-osd_UUID 3
>>>   2 UP mgc MGC10.0.1.70 at tcp 57e06c2d-5294-f034-fd95-460cee4f92b7 4
>>>   3 UP mds MDS MDS_uuid 2
>>>
>>>
>>> but unfortunately it disappears after 03 seconds
>>>
>>> lctl  dl shows nothing
>>>
>>> lctl dk
>>>
>>> shows this debug output
>>>
>>>
>>> 00000020:00000080:18.0:1621276062.004338:0:13403:0:(obd_config.c:1128:class_process_config())
>>> processing cmd: cf006
>>> 00000020:00000080:18.0:1621276062.004341:0:13403:0:(obd_config.c:1147:class_process_config())
>>> removing mappings for uuid MGC10.0.1.70 at tcp_0
>>> 00000020:01000004:18.0:1621276062.004346:0:13403:0:(obd_mount.c:661:lustre_put_lsi())
>>> put ffff9bbbf91d5800 1
>>> 00000020:00000080:18.0:1621276062.004351:0:13403:0:(genops.c:1501:class_disconnect())
>>> disconnect: cookie 0x256dd92fc5bf929c
>>> 00000020:00000080:18.0:1621276062.004354:0:13403:0:(genops.c:1024:class_export_put())
>>> final put ffff9bbf3e66a400/lustre-MDT0000-osd_UUID
>>> 00000020:01000000:18.0:1621276062.004361:0:13403:0:(obd_config.c:2100:class_manual_cleanup())
>>> Manual cleanup of lustre-MDT0000-osd (flags='')
>>> 00000020:00000080:18.0:1621276062.004368:0:821:0:(genops.c:974:class_export_destroy())
>>> destroying export ffff9bbf3e66a400/lustre-MDT0000-osd_UUID for
>>> lustre-MDT0000-osd
>>> 00000020:00000080:18.0:1621276062.004376:0:13403:0:(obd_config.c:1128:class_process_config())
>>> processing cmd: cf004
>>> 00000020:00000080:18.0:1621276062.004379:0:13403:0:(obd_config.c:659:class_cleanup())
>>> lustre-MDT0000-osd: forcing exports to disconnect: 0/0
>>> 00000020:00080000:18.0:1621276062.004382:0:13403:0:(genops.c:1590:class_disconnect_exports())
>>> OBD device 0 (ffff9bbf47141080) has no exports
>>> 00000020:00000080:18.0:1621276062.004788:0:13403:0:(obd_config.c:1128:class_process_config())
>>> processing cmd: cf002
>>> 00000020:00000080:18.0:1621276062.004791:0:13403:0:(obd_config.c:589:class_detach())
>>> detach on obd lustre-MDT0000-osd (uuid lustre-MDT0000-osd_UUID)
>>> 00000020:00000080:18.0:1621276062.004794:0:13403:0:(genops.c:1024:class_export_put())
>>> final put ffff9bbf48800c00/lustre-MDT0000-osd_UUID
>>> 00000020:00000080:18.0:1621276062.004796:0:13403:0:(genops.c:974:class_export_destroy())
>>> destroying export ffff9bbf48800c00/lustre-MDT0000-osd_UUID for
>>> lustre-MDT0000-osd
>>> 00000020:01000000:18.0:1621276062.004799:0:13403:0:(genops.c:481:class_free_dev())
>>> finishing cleanup of obd lustre-MDT0000-osd (lustre-MDT0000-osd_UUID)
>>> 00000020:01000004:18.0:1621276062.450759:0:13403:0:(obd_mount.c:605:lustre_free_lsi())
>>> Freeing lsi ffff9bbbf91d6800
>>> 00000020:01000000:18.0:1621276062.450805:0:13403:0:(obd_config.c:2100:class_manual_cleanup())
>>> Manual cleanup of MDS (flags='F')
>>> 00000020:00000080:18.0:1621276062.450806:0:13403:0:(obd_config.c:1128:class_process_config())
>>> processing cmd: cf004
>>> 00000020:00000080:18.0:1621276062.450807:0:13403:0:(obd_config.c:659:class_cleanup())
>>> MDS: forcing exports to disconnect: 0/0
>>> 00000020:00080000:18.0:1621276062.450809:0:13403:0:(genops.c:1590:class_disconnect_exports())
>>> OBD device 3 (ffff9bbf43fdd280) has no exports
>>> 00000020:00000080:58.0F:1621276062.490781:0:13403:0:(obd_config.c:1128:class_process_config())
>>> processing cmd: cf002
>>> 00000020:00000080:58.0:1621276062.490787:0:13403:0:(obd_config.c:589:class_detach())
>>> detach on obd MDS (uuid MDS_uuid)
>>> 00000020:00000080:58.0:1621276062.490788:0:13403:0:(genops.c:1024:class_export_put())
>>> final put ffff9bbf3e668800/MDS_uuid
>>> 00000020:00000080:58.0:1621276062.490790:0:13403:0:(genops.c:974:class_export_destroy())
>>> destroying export ffff9bbf3e668800/MDS_uuid for MDS
>>> 00000020:01000000:58.0:1621276062.490791:0:13403:0:(genops.c:481:class_free_dev())
>>> finishing cleanup of obd MDS (MDS_uuid)
>>> 00000020:02000400:58.0:1621276062.490877:0:13403:0:(obd_mount_server.c:1642:server_put_super())
>>> server umount lustre-MDT0000 complete
>>> 00000400:02020000:42.0:1621276086.284109:0:5400:0:(acceptor.c:321:lnet_accept())
>>> 120-3: Refusing connection from 127.0.0.1 for 127.0.0.1 at tcp: No
>>> matching NI
>>> 00000800:00020000:6.0:1621276086.284152:0:5383:0:(socklnd_cb.c:1817:ksocknal_recv_hello())
>>> Error -104 reading HELLO from 127.0.0.1
>>> 00000400:02020000:6.0:1621276086.284174:0:5383:0:(acceptor.c:127:lnet_connect_console_error())
>>> 11b-b: Connection to 127.0.0.1 at tcp at host 127.0.0.1 on port 988 was
>>> reset: is it running a compatible version of Lustre and is 127.0.0.1 at tcp
>>> one of its NIDs?
>>> 00000800:00000100:6.0:1621276086.284189:0:5383:0:(socklnd_cb.c:438:ksocknal_txlist_done())
>>> Deleting packet type 2 len 0 10.0.1.70 at tcp->127.0.0.1 at tcp
>>> 00000800:00000100:34.0:1621276136.363882:0:5401:0:(socklnd_cb.c:979:ksocknal_launch_packet())
>>> No usable routes to 12345-127.0.0.1 at tcp
>>> 00000400:02020000:42.0:1621276186.440095:0:5400:0:(acceptor.c:321:lnet_accept())
>>> 120-3: Refusing connection from 127.0.0.1 for 127.0.0.1 at tcp: No
>>> matching NI
>>> 00000800:00020000:44.0:1621276186.446533:0:5386:0:(socklnd_cb.c:1817:ksocknal_recv_hello())
>>> Error -104 reading HELLO from 127.0.0.1
>>> 00000400:02020000:44.0:1621276186.452996:0:5386:0:(acceptor.c:127:lnet_connect_console_error())
>>> 11b-b: Connection to 127.0.0.1 at tcp at host 127.0.0.1 on port 988 was
>>> reset: is it running a compatible version of Lustre and is 127.0.0.1 at tcp
>>> one of its NIDs?
>>> 00000800:00000100:44.0:1621276186.461433:0:5386:0:(socklnd_cb.c:438:ksocknal_txlist_done())
>>> Deleting packet type 2 len 0 10.0.1.70 at tcp->127.0.0.1 at tcp
>>> Debug log: 872 lines, 872 kept, 0 dropped, 0 bad.
>>>
>>>
>>>
>>> I just cant find out any help would be very appreciated
>>>
>>>
>>> Thanks all
>>>
>>>
>>>
>>>
>>>
>>>
>>> --
>>> Tahari.Abdeslam
>>> _______________________________________________
>>> lustre-discuss mailing list
>>> lustre-discuss at lists.lustre.org
>>> http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
>>>
>>
>
> --
> Tahari.Abdeslam
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.lustre.org/pipermail/lustre-discuss-lustre.org/attachments/20210517/2adc6c81/attachment-0001.html>


More information about the lustre-discuss mailing list