[lustre-discuss] lustre-discuss Digest, Vol 182, Issue 12

Abdeslam Tahari abeslam at gmail.com
Wed May 19 04:45:00 PDT 2021


Hello Ms Megan

Thank you for the reply and your help

I have checked the lctl ping
it seems to be ok the result
 lctl ping 10.0.1.70
12345-0 at lo
12345-10.0.1.70 at tcp


the ping is good it is always ok .

the problem is when i mount the luster file system

mount -t lustre /dev/sda /mds

i have the following output
 lctl dl
  0 UP osd-ldiskfs lustre-MDT0000-osd lustre-MDT0000-osd_UUID 3
  2 UP mgc MGC10.0.1.70 at tcp 3ec79ce9-5167-9661-9bd6-0b897fcc42f2 4
  3 UP mds MDS MDS_uuid 2


if i execute the command for the second time i will have no output at all
and the filesystem in reality is not mounted

i think but i am not sure it complains about the UUID of the MDT

from the output of the

lctl dk
00000100:00080000:78.0:1621365812.955564:0:84913:0:(pinger.c:413:ptlrpc_pinger_del_import())
removing pingable import
lustre-MDT0000-lwp-MDT0000_UUID->lustre-MDT0000_UUID
00000100:00080000:78.0:1621365812.955567:0:84913:0:(import.c:86:import_set_state_nolock())
ffff9b985701b800 lustre-MDT0000_UUID: changing import state from DISCONN to
CLOSED
*00000100:00080000:78.0:1621365812.955571:0:84913:0:(import.c:157:ptlrpc_deactivate_import_nolock())
setting import lustre-MDT0000_UUID INVALID*
10000000:01000000:78.0:1621365812.965420:0:84913:0:(mgc_request.c:151:config_log_put())
dropping config log lustre-mdtir

Kind regards


Le mer. 19 mai 2021 à 03:15, Ms. Megan Larko via lustre-discuss <
lustre-discuss at lists.lustre.org> a écrit :

> Hello Tahari,
> What is the result of "lctl ping 10.0.1.70 at tcp_0" from the box on which
> you are trying to mount the Lustre File System?   Is the ping successful
> and then fails after 03 seconds? If yes, you may wish to check the
> /etc/lnet.conf file for Lustre LNet path "discovery"  (1 allows LNet
> discovery while 0 does not), and drop_asym_route (0 disallows
> asymmetrical routing while 1 permits it).   I have worked with a few
> complex networks in which we chose to turn off LNet discovery and specify,
> via /etc/lnet.conf, the routes.  On one system the asymmetrical routing (we
> have 16 LNet boxes between the system and the Lustre storage) seemed to be
> a problem, but we couldn't pin it to any particular box.  On that system
> disallowing asymmetrical routing seemed to help maintain LNet/Lustre
> connectivity.
>
> One may check the lctl ping to narrow down net connectivity from other
> possibilities.
>
> Cheers,
> megan
>
> On Mon, May 17, 2021 at 3:50 PM <lustre-discuss-request at lists.lustre.org>
> wrote:
>
>> Send lustre-discuss mailing list submissions to
>>         lustre-discuss at lists.lustre.org
>>
>> To subscribe or unsubscribe via the World Wide Web, visit
>>         http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
>> or, via email, send a message with subject or body 'help' to
>>         lustre-discuss-request at lists.lustre.org
>>
>> You can reach the person managing the list at
>>         lustre-discuss-owner at lists.lustre.org
>>
>> When replying, please edit your Subject line so it is more specific
>> than "Re: Contents of lustre-discuss digest..."
>>
>>
>> Today's Topics:
>>
>>    1. Re: problems to mount MDS and MDT (Abdeslam Tahari)
>>    2. Re: problems to mount MDS and MDT (Colin Faber)
>>
>>
>> ----------------------------------------------------------------------
>>
>> Message: 1
>> Date: Mon, 17 May 2021 21:35:34 +0200
>> From: Abdeslam Tahari <abeslam at gmail.com>
>> To: Colin Faber <cfaber at gmail.com>
>> Cc: lustre-discuss <lustre-discuss at lists.lustre.org>
>> Subject: Re: [lustre-discuss] problems to mount MDS and MDT
>> Message-ID:
>>         <CA+LuYSL9_TTcHopwHYbFRosZNgUFK=
>> bxeCePEn5DzZD+QXnwiQ at mail.gmail.com>
>> Content-Type: text/plain; charset="utf-8"
>>
>> Thank you Colin
>>
>> No i don't have iptables or rules
>>
>> firewalled is stopped selinux disabled as well
>>  iptables -L
>> Chain INPUT (policy ACCEPT)
>> target     prot opt source               destination
>>
>> Chain FORWARD (policy ACCEPT)
>> target     prot opt source               destination
>>
>> Chain OUTPUT (policy ACCEPT)
>> target     prot opt source               destination
>>
>>
>> Regards
>>
>>
>> Regards
>>
>> Le lun. 17 mai 2021 ? 21:29, Colin Faber <cfaber at gmail.com> a ?crit :
>>
>> > Firewall rules dealing with localhost?
>> >
>> > On Mon, May 17, 2021 at 11:33 AM Abdeslam Tahari via lustre-discuss <
>> > lustre-discuss at lists.lustre.org> wrote:
>> >
>> >> Hello
>> >>
>> >> i have a problem to mount the mds/mdt luster, it wont mount at all and
>> >> there is no message errors at the console
>> >>
>> >> -it does not show errors or messages while mounting it
>> >>
>> >> here are some debug file logs
>> >>
>> >>
>> >> i specify it is a new project that i am doing.
>> >>
>> >> the version and packages of luter installed:
>> >> kmod-lustre-2.12.5-1.el7.x86_64
>> >> kernel-devel-3.10.0-1127.8.2.el7_lustre.x86_64
>> >> lustre-2.12.5-1.el7.x86_64
>> >> lustre-resource-agents-2.12.5-1.el7.x86_64
>> >> kernel-3.10.0-1160.2.1.el7_lustre.x86_64
>> >> kernel-debuginfo-common-x86_64-3.10.0-1160.2.1.el7_lustre.x86_64
>> >> kmod-lustre-osd-ldiskfs-2.12.5-1.el7.x86_64
>> >> kernel-3.10.0-1127.8.2.el7_lustre.x86_64
>> >> lustre-osd-ldiskfs-mount-2.12.5-1.el7.x86_64
>> >>
>> >>
>> >>
>> >> the system(os) Centos 7
>> >>
>> >> the kernel
>> >> Linux lustre-mds1 3.10.0-1127.8.2.el7_lustre.x86_64
>> >>  cat /etc/redhat-release
>> >>
>> >>
>> >> when i mount the luster file-system it wont show up and no errors
>> >>
>> >> mount -t lustre /dev/sda /mds
>> >>
>> >> lctl dl  does not show up
>> >>
>> >> df -h   no mount point for /dev/sda
>> >>
>> >>
>> >> lctl dl
>> >>
>> >> shows this:
>> >> lctl dl
>> >>   0 UP osd-ldiskfs lustre-MDT0000-osd lustre-MDT0000-osd_UUID 3
>> >>   2 UP mgc MGC10.0.1.70 at tcp 57e06c2d-5294-f034-fd95-460cee4f92b7 4
>> >>   3 UP mds MDS MDS_uuid 2
>> >>
>> >>
>> >> but unfortunately it disappears after 03 seconds
>> >>
>> >> lctl  dl shows nothing
>> >>
>> >> lctl dk
>> >>
>> >> shows this debug output
>> >>
>> >>
>> >>
>> 00000020:00000080:18.0:1621276062.004338:0:13403:0:(obd_config.c:1128:class_process_config())
>> >> processing cmd: cf006
>> >>
>> 00000020:00000080:18.0:1621276062.004341:0:13403:0:(obd_config.c:1147:class_process_config())
>> >> removing mappings for uuid MGC10.0.1.70 at tcp_0
>> >>
>> 00000020:01000004:18.0:1621276062.004346:0:13403:0:(obd_mount.c:661:lustre_put_lsi())
>> >> put ffff9bbbf91d5800 1
>> >>
>> 00000020:00000080:18.0:1621276062.004351:0:13403:0:(genops.c:1501:class_disconnect())
>> >> disconnect: cookie 0x256dd92fc5bf929c
>> >>
>> 00000020:00000080:18.0:1621276062.004354:0:13403:0:(genops.c:1024:class_export_put())
>> >> final put ffff9bbf3e66a400/lustre-MDT0000-osd_UUID
>> >>
>> 00000020:01000000:18.0:1621276062.004361:0:13403:0:(obd_config.c:2100:class_manual_cleanup())
>> >> Manual cleanup of lustre-MDT0000-osd (flags='')
>> >>
>> 00000020:00000080:18.0:1621276062.004368:0:821:0:(genops.c:974:class_export_destroy())
>> >> destroying export ffff9bbf3e66a400/lustre-MDT0000-osd_UUID for
>> >> lustre-MDT0000-osd
>> >>
>> 00000020:00000080:18.0:1621276062.004376:0:13403:0:(obd_config.c:1128:class_process_config())
>> >> processing cmd: cf004
>> >>
>> 00000020:00000080:18.0:1621276062.004379:0:13403:0:(obd_config.c:659:class_cleanup())
>> >> lustre-MDT0000-osd: forcing exports to disconnect: 0/0
>> >>
>> 00000020:00080000:18.0:1621276062.004382:0:13403:0:(genops.c:1590:class_disconnect_exports())
>> >> OBD device 0 (ffff9bbf47141080) has no exports
>> >>
>> 00000020:00000080:18.0:1621276062.004788:0:13403:0:(obd_config.c:1128:class_process_config())
>> >> processing cmd: cf002
>> >>
>> 00000020:00000080:18.0:1621276062.004791:0:13403:0:(obd_config.c:589:class_detach())
>> >> detach on obd lustre-MDT0000-osd (uuid lustre-MDT0000-osd_UUID)
>> >>
>> 00000020:00000080:18.0:1621276062.004794:0:13403:0:(genops.c:1024:class_export_put())
>> >> final put ffff9bbf48800c00/lustre-MDT0000-osd_UUID
>> >>
>> 00000020:00000080:18.0:1621276062.004796:0:13403:0:(genops.c:974:class_export_destroy())
>> >> destroying export ffff9bbf48800c00/lustre-MDT0000-osd_UUID for
>> >> lustre-MDT0000-osd
>> >>
>> 00000020:01000000:18.0:1621276062.004799:0:13403:0:(genops.c:481:class_free_dev())
>> >> finishing cleanup of obd lustre-MDT0000-osd (lustre-MDT0000-osd_UUID)
>> >>
>> 00000020:01000004:18.0:1621276062.450759:0:13403:0:(obd_mount.c:605:lustre_free_lsi())
>> >> Freeing lsi ffff9bbbf91d6800
>> >>
>> 00000020:01000000:18.0:1621276062.450805:0:13403:0:(obd_config.c:2100:class_manual_cleanup())
>> >> Manual cleanup of MDS (flags='F')
>> >>
>> 00000020:00000080:18.0:1621276062.450806:0:13403:0:(obd_config.c:1128:class_process_config())
>> >> processing cmd: cf004
>> >>
>> 00000020:00000080:18.0:1621276062.450807:0:13403:0:(obd_config.c:659:class_cleanup())
>> >> MDS: forcing exports to disconnect: 0/0
>> >>
>> 00000020:00080000:18.0:1621276062.450809:0:13403:0:(genops.c:1590:class_disconnect_exports())
>> >> OBD device 3 (ffff9bbf43fdd280) has no exports
>> >>
>> 00000020:00000080:58.0F:1621276062.490781:0:13403:0:(obd_config.c:1128:class_process_config())
>> >> processing cmd: cf002
>> >>
>> 00000020:00000080:58.0:1621276062.490787:0:13403:0:(obd_config.c:589:class_detach())
>> >> detach on obd MDS (uuid MDS_uuid)
>> >>
>> 00000020:00000080:58.0:1621276062.490788:0:13403:0:(genops.c:1024:class_export_put())
>> >> final put ffff9bbf3e668800/MDS_uuid
>> >>
>> 00000020:00000080:58.0:1621276062.490790:0:13403:0:(genops.c:974:class_export_destroy())
>> >> destroying export ffff9bbf3e668800/MDS_uuid for MDS
>> >>
>> 00000020:01000000:58.0:1621276062.490791:0:13403:0:(genops.c:481:class_free_dev())
>> >> finishing cleanup of obd MDS (MDS_uuid)
>> >>
>> 00000020:02000400:58.0:1621276062.490877:0:13403:0:(obd_mount_server.c:1642:server_put_super())
>> >> server umount lustre-MDT0000 complete
>> >>
>> 00000400:02020000:42.0:1621276086.284109:0:5400:0:(acceptor.c:321:lnet_accept())
>> >> 120-3: Refusing connection from 127.0.0.1 for 127.0.0.1 at tcp: No
>> matching
>> >> NI
>> >>
>> 00000800:00020000:6.0:1621276086.284152:0:5383:0:(socklnd_cb.c:1817:ksocknal_recv_hello())
>> >> Error -104 reading HELLO from 127.0.0.1
>> >>
>> 00000400:02020000:6.0:1621276086.284174:0:5383:0:(acceptor.c:127:lnet_connect_console_error())
>> >> 11b-b: Connection to 127.0.0.1 at tcp at host 127.0.0.1 on port 988 was
>> >> reset: is it running a compatible version of Lustre and is
>> 127.0.0.1 at tcp
>> >> one of its NIDs?
>> >>
>> 00000800:00000100:6.0:1621276086.284189:0:5383:0:(socklnd_cb.c:438:ksocknal_txlist_done())
>> >> Deleting packet type 2 len 0 10.0.1.70 at tcp->127.0.0.1 at tcp
>> >>
>> 00000800:00000100:34.0:1621276136.363882:0:5401:0:(socklnd_cb.c:979:ksocknal_launch_packet())
>> >> No usable routes to 12345-127.0.0.1 at tcp
>> >>
>> 00000400:02020000:42.0:1621276186.440095:0:5400:0:(acceptor.c:321:lnet_accept())
>> >> 120-3: Refusing connection from 127.0.0.1 for 127.0.0.1 at tcp: No
>> matching
>> >> NI
>> >>
>> 00000800:00020000:44.0:1621276186.446533:0:5386:0:(socklnd_cb.c:1817:ksocknal_recv_hello())
>> >> Error -104 reading HELLO from 127.0.0.1
>> >>
>> 00000400:02020000:44.0:1621276186.452996:0:5386:0:(acceptor.c:127:lnet_connect_console_error())
>> >> 11b-b: Connection to 127.0.0.1 at tcp at host 127.0.0.1 on port 988 was
>> >> reset: is it running a compatible version of Lustre and is
>> 127.0.0.1 at tcp
>> >> one of its NIDs?
>> >>
>> 00000800:00000100:44.0:1621276186.461433:0:5386:0:(socklnd_cb.c:438:ksocknal_txlist_done())
>> >> Deleting packet type 2 len 0 10.0.1.70 at tcp->127.0.0.1 at tcp
>> >> Debug log: 872 lines, 872 kept, 0 dropped, 0 bad.
>> >>
>> >>
>> >>
>> >> I just cant find out any help would be very appreciated
>> >>
>> >>
>> >> Thanks all
>> >>
>> >>
>> >>
>> >>
>> >>
>> >>
>> >> --
>> >> Tahari.Abdeslam
>> >> _______________________________________________
>> >> lustre-discuss mailing list
>> >> lustre-discuss at lists.lustre.org
>> >> http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
>> >>
>> >
>>
>> --
>> Tahari.Abdeslam
>> -------------- next part --------------
>> An HTML attachment was scrubbed...
>> URL: <
>> http://lists.lustre.org/pipermail/lustre-discuss-lustre.org/attachments/20210517/1decdc97/attachment-0001.html
>> >
>>
>> ------------------------------
>>
>> Message: 2
>> Date: Mon, 17 May 2021 13:50:03 -0600
>> From: Colin Faber <cfaber at gmail.com>
>> To: Abdeslam Tahari <abeslam at gmail.com>
>> Cc: lustre-discuss <lustre-discuss at lists.lustre.org>
>> Subject: Re: [lustre-discuss] problems to mount MDS and MDT
>> Message-ID:
>>         <CAJcXmB=T884j=
>> 5N8nhWspFBvNS+nAOoMa9b8xJUdhXT-fBoysw at mail.gmail.com>
>> Content-Type: text/plain; charset="utf-8"
>>
>> It appears part of the debug data is missing (the part before you posted
>> it), Can you try again, lctl dk > /dev/null to clear it then try your
>> mount
>> and grab the debug again?
>>
>> On Mon, May 17, 2021 at 1:35 PM Abdeslam Tahari <abeslam at gmail.com>
>> wrote:
>>
>> > Thank you Colin
>> >
>> > No i don't have iptables or rules
>> >
>> > firewalled is stopped selinux disabled as well
>> >  iptables -L
>> > Chain INPUT (policy ACCEPT)
>> > target     prot opt source               destination
>> >
>> > Chain FORWARD (policy ACCEPT)
>> > target     prot opt source               destination
>> >
>> > Chain OUTPUT (policy ACCEPT)
>> > target     prot opt source               destination
>> >
>> >
>> > Regards
>> >
>> >
>> > Regards
>> >
>> > Le lun. 17 mai 2021 ? 21:29, Colin Faber <cfaber at gmail.com> a ?crit :
>> >
>> >> Firewall rules dealing with localhost?
>> >>
>> >> On Mon, May 17, 2021 at 11:33 AM Abdeslam Tahari via lustre-discuss <
>> >> lustre-discuss at lists.lustre.org> wrote:
>> >>
>> >>> Hello
>> >>>
>> >>> i have a problem to mount the mds/mdt luster, it wont mount at all and
>> >>> there is no message errors at the console
>> >>>
>> >>> -it does not show errors or messages while mounting it
>> >>>
>> >>> here are some debug file logs
>> >>>
>> >>>
>> >>> i specify it is a new project that i am doing.
>> >>>
>> >>> the version and packages of luter installed:
>> >>> kmod-lustre-2.12.5-1.el7.x86_64
>> >>> kernel-devel-3.10.0-1127.8.2.el7_lustre.x86_64
>> >>> lustre-2.12.5-1.el7.x86_64
>> >>> lustre-resource-agents-2.12.5-1.el7.x86_64
>> >>> kernel-3.10.0-1160.2.1.el7_lustre.x86_64
>> >>> kernel-debuginfo-common-x86_64-3.10.0-1160.2.1.el7_lustre.x86_64
>> >>> kmod-lustre-osd-ldiskfs-2.12.5-1.el7.x86_64
>> >>> kernel-3.10.0-1127.8.2.el7_lustre.x86_64
>> >>> lustre-osd-ldiskfs-mount-2.12.5-1.el7.x86_64
>> >>>
>> >>>
>> >>>
>> >>> the system(os) Centos 7
>> >>>
>> >>> the kernel
>> >>> Linux lustre-mds1 3.10.0-1127.8.2.el7_lustre.x86_64
>> >>>  cat /etc/redhat-release
>> >>>
>> >>>
>> >>> when i mount the luster file-system it wont show up and no errors
>> >>>
>> >>> mount -t lustre /dev/sda /mds
>> >>>
>> >>> lctl dl  does not show up
>> >>>
>> >>> df -h   no mount point for /dev/sda
>> >>>
>> >>>
>> >>> lctl dl
>> >>>
>> >>> shows this:
>> >>> lctl dl
>> >>>   0 UP osd-ldiskfs lustre-MDT0000-osd lustre-MDT0000-osd_UUID 3
>> >>>   2 UP mgc MGC10.0.1.70 at tcp 57e06c2d-5294-f034-fd95-460cee4f92b7 4
>> >>>   3 UP mds MDS MDS_uuid 2
>> >>>
>> >>>
>> >>> but unfortunately it disappears after 03 seconds
>> >>>
>> >>> lctl  dl shows nothing
>> >>>
>> >>> lctl dk
>> >>>
>> >>> shows this debug output
>> >>>
>> >>>
>> >>>
>> 00000020:00000080:18.0:1621276062.004338:0:13403:0:(obd_config.c:1128:class_process_config())
>> >>> processing cmd: cf006
>> >>>
>> 00000020:00000080:18.0:1621276062.004341:0:13403:0:(obd_config.c:1147:class_process_config())
>> >>> removing mappings for uuid MGC10.0.1.70 at tcp_0
>> >>>
>> 00000020:01000004:18.0:1621276062.004346:0:13403:0:(obd_mount.c:661:lustre_put_lsi())
>> >>> put ffff9bbbf91d5800 1
>> >>>
>> 00000020:00000080:18.0:1621276062.004351:0:13403:0:(genops.c:1501:class_disconnect())
>> >>> disconnect: cookie 0x256dd92fc5bf929c
>> >>>
>> 00000020:00000080:18.0:1621276062.004354:0:13403:0:(genops.c:1024:class_export_put())
>> >>> final put ffff9bbf3e66a400/lustre-MDT0000-osd_UUID
>> >>>
>> 00000020:01000000:18.0:1621276062.004361:0:13403:0:(obd_config.c:2100:class_manual_cleanup())
>> >>> Manual cleanup of lustre-MDT0000-osd (flags='')
>> >>>
>> 00000020:00000080:18.0:1621276062.004368:0:821:0:(genops.c:974:class_export_destroy())
>> >>> destroying export ffff9bbf3e66a400/lustre-MDT0000-osd_UUID for
>> >>> lustre-MDT0000-osd
>> >>>
>> 00000020:00000080:18.0:1621276062.004376:0:13403:0:(obd_config.c:1128:class_process_config())
>> >>> processing cmd: cf004
>> >>>
>> 00000020:00000080:18.0:1621276062.004379:0:13403:0:(obd_config.c:659:class_cleanup())
>> >>> lustre-MDT0000-osd: forcing exports to disconnect: 0/0
>> >>>
>> 00000020:00080000:18.0:1621276062.004382:0:13403:0:(genops.c:1590:class_disconnect_exports())
>> >>> OBD device 0 (ffff9bbf47141080) has no exports
>> >>>
>> 00000020:00000080:18.0:1621276062.004788:0:13403:0:(obd_config.c:1128:class_process_config())
>> >>> processing cmd: cf002
>> >>>
>> 00000020:00000080:18.0:1621276062.004791:0:13403:0:(obd_config.c:589:class_detach())
>> >>> detach on obd lustre-MDT0000-osd (uuid lustre-MDT0000-osd_UUID)
>> >>>
>> 00000020:00000080:18.0:1621276062.004794:0:13403:0:(genops.c:1024:class_export_put())
>> >>> final put ffff9bbf48800c00/lustre-MDT0000-osd_UUID
>> >>>
>> 00000020:00000080:18.0:1621276062.004796:0:13403:0:(genops.c:974:class_export_destroy())
>> >>> destroying export ffff9bbf48800c00/lustre-MDT0000-osd_UUID for
>> >>> lustre-MDT0000-osd
>> >>>
>> 00000020:01000000:18.0:1621276062.004799:0:13403:0:(genops.c:481:class_free_dev())
>> >>> finishing cleanup of obd lustre-MDT0000-osd (lustre-MDT0000-osd_UUID)
>> >>>
>> 00000020:01000004:18.0:1621276062.450759:0:13403:0:(obd_mount.c:605:lustre_free_lsi())
>> >>> Freeing lsi ffff9bbbf91d6800
>> >>>
>> 00000020:01000000:18.0:1621276062.450805:0:13403:0:(obd_config.c:2100:class_manual_cleanup())
>> >>> Manual cleanup of MDS (flags='F')
>> >>>
>> 00000020:00000080:18.0:1621276062.450806:0:13403:0:(obd_config.c:1128:class_process_config())
>> >>> processing cmd: cf004
>> >>>
>> 00000020:00000080:18.0:1621276062.450807:0:13403:0:(obd_config.c:659:class_cleanup())
>> >>> MDS: forcing exports to disconnect: 0/0
>> >>>
>> 00000020:00080000:18.0:1621276062.450809:0:13403:0:(genops.c:1590:class_disconnect_exports())
>> >>> OBD device 3 (ffff9bbf43fdd280) has no exports
>> >>>
>> 00000020:00000080:58.0F:1621276062.490781:0:13403:0:(obd_config.c:1128:class_process_config())
>> >>> processing cmd: cf002
>> >>>
>> 00000020:00000080:58.0:1621276062.490787:0:13403:0:(obd_config.c:589:class_detach())
>> >>> detach on obd MDS (uuid MDS_uuid)
>> >>>
>> 00000020:00000080:58.0:1621276062.490788:0:13403:0:(genops.c:1024:class_export_put())
>> >>> final put ffff9bbf3e668800/MDS_uuid
>> >>>
>> 00000020:00000080:58.0:1621276062.490790:0:13403:0:(genops.c:974:class_export_destroy())
>> >>> destroying export ffff9bbf3e668800/MDS_uuid for MDS
>> >>>
>> 00000020:01000000:58.0:1621276062.490791:0:13403:0:(genops.c:481:class_free_dev())
>> >>> finishing cleanup of obd MDS (MDS_uuid)
>> >>>
>> 00000020:02000400:58.0:1621276062.490877:0:13403:0:(obd_mount_server.c:1642:server_put_super())
>> >>> server umount lustre-MDT0000 complete
>> >>>
>> 00000400:02020000:42.0:1621276086.284109:0:5400:0:(acceptor.c:321:lnet_accept())
>> >>> 120-3: Refusing connection from 127.0.0.1 for 127.0.0.1 at tcp: No
>> >>> matching NI
>> >>>
>> 00000800:00020000:6.0:1621276086.284152:0:5383:0:(socklnd_cb.c:1817:ksocknal_recv_hello())
>> >>> Error -104 reading HELLO from 127.0.0.1
>> >>>
>> 00000400:02020000:6.0:1621276086.284174:0:5383:0:(acceptor.c:127:lnet_connect_console_error())
>> >>> 11b-b: Connection to 127.0.0.1 at tcp at host 127.0.0.1 on port 988 was
>> >>> reset: is it running a compatible version of Lustre and is
>> 127.0.0.1 at tcp
>> >>> one of its NIDs?
>> >>>
>> 00000800:00000100:6.0:1621276086.284189:0:5383:0:(socklnd_cb.c:438:ksocknal_txlist_done())
>> >>> Deleting packet type 2 len 0 10.0.1.70 at tcp->127.0.0.1 at tcp
>> >>>
>> 00000800:00000100:34.0:1621276136.363882:0:5401:0:(socklnd_cb.c:979:ksocknal_launch_packet())
>> >>> No usable routes to 12345-127.0.0.1 at tcp
>> >>>
>> 00000400:02020000:42.0:1621276186.440095:0:5400:0:(acceptor.c:321:lnet_accept())
>> >>> 120-3: Refusing connection from 127.0.0.1 for 127.0.0.1 at tcp: No
>> >>> matching NI
>> >>>
>> 00000800:00020000:44.0:1621276186.446533:0:5386:0:(socklnd_cb.c:1817:ksocknal_recv_hello())
>> >>> Error -104 reading HELLO from 127.0.0.1
>> >>>
>> 00000400:02020000:44.0:1621276186.452996:0:5386:0:(acceptor.c:127:lnet_connect_console_error())
>> >>> 11b-b: Connection to 127.0.0.1 at tcp at host 127.0.0.1 on port 988 was
>> >>> reset: is it running a compatible version of Lustre and is
>> 127.0.0.1 at tcp
>> >>> one of its NIDs?
>> >>>
>> 00000800:00000100:44.0:1621276186.461433:0:5386:0:(socklnd_cb.c:438:ksocknal_txlist_done())
>> >>> Deleting packet type 2 len 0 10.0.1.70 at tcp->127.0.0.1 at tcp
>> >>> Debug log: 872 lines, 872 kept, 0 dropped, 0 bad.
>> >>>
>> >>>
>> >>>
>> >>> I just cant find out any help would be very appreciated
>> >>>
>> >>>
>> >>> Thanks all
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>
>> >>> --
>> >>> Tahari.Abdeslam
>> >>> _______________________________________________
>> >>> lustre-discuss mailing list
>> >>> lustre-discuss at lists.lustre.org
>> >>> http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
>> >>>
>> >>
>> >
>> > --
>> > Tahari.Abdeslam
>> >
>> -------------- next part --------------
>> An HTML attachment was scrubbed...
>> URL: <
>> http://lists.lustre.org/pipermail/lustre-discuss-lustre.org/attachments/20210517/2adc6c81/attachment.html
>> >
>>
>> ------------------------------
>>
>> Subject: Digest Footer
>>
>> _______________________________________________
>> lustre-discuss mailing list
>> lustre-discuss at lists.lustre.org
>> http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
>>
>>
>> ------------------------------
>>
>> End of lustre-discuss Digest, Vol 182, Issue 12
>> ***********************************************
>>
> _______________________________________________
> lustre-discuss mailing list
> lustre-discuss at lists.lustre.org
> http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
>


-- 
Tahari.Abdeslam
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.lustre.org/pipermail/lustre-discuss-lustre.org/attachments/20210519/a627cc1d/attachment-0001.html>


More information about the lustre-discuss mailing list