[lustre-discuss] lustre-discuss Digest, Vol 182, Issue 12

Abdeslam Tahari abeslam at gmail.com
Wed May 19 12:44:21 PDT 2021


Hello Ms Megan



I am happy it is resolved


it was a problem of UUID


I will post later on the solution+ problem



Cheers

Le 19/05/2021 à 13:45, Abdeslam Tahari a écrit :
> Hello Ms Megan
>
> Thank you for the reply and your help
>
> I have checked the lctl ping
> it seems to be ok the result
>  lctl ping 10.0.1.70
> 12345-0 at lo
> 12345-10.0.1.70 at tcp
>
>
> the ping is good it is always ok .
>
> the problem is when i mount the luster file system
>
> mount -t lustre /dev/sda /mds
>
> i have the following output
>  lctl dl
>   0 UP osd-ldiskfs lustre-MDT0000-osd lustre-MDT0000-osd_UUID 3
>   2 UP mgc MGC10.0.1.70 at tcp 3ec79ce9-5167-9661-9bd6-0b897fcc42f2 4
>   3 UP mds MDS MDS_uuid 2
>
>
> if i execute the command for the second time i will have no output at all
> and the filesystem in reality is not mounted
>
> i think but i am not sure it complains about the UUID of the MDT
>
> from the output of the
>
> lctl dk
> 00000100:00080000:78.0:1621365812.955564:0:84913:0:(pinger.c:413:ptlrpc_pinger_del_import()) 
> removing pingable import 
> lustre-MDT0000-lwp-MDT0000_UUID->lustre-MDT0000_UUID
> 00000100:00080000:78.0:1621365812.955567:0:84913:0:(import.c:86:import_set_state_nolock()) 
> ffff9b985701b800 lustre-MDT0000_UUID: changing import state from 
> DISCONN to CLOSED
> *00000100:00080000:78.0:1621365812.955571:0:84913:0:(import.c:157:ptlrpc_deactivate_import_nolock()) 
> setting import lustre-MDT0000_UUID INVALID*
> 10000000:01000000:78.0:1621365812.965420:0:84913:0:(mgc_request.c:151:config_log_put()) 
> dropping config log lustre-mdtir
>
> Kind regards
>
>
> Le mer. 19 mai 2021 à 03:15, Ms. Megan Larko via lustre-discuss 
> <lustre-discuss at lists.lustre.org 
> <mailto:lustre-discuss at lists.lustre.org>> a écrit :
>
>     Hello Tahari,
>     What is the result of "lctl ping 10.0.1.70 at tcp_0" from the box on
>     which you are trying to mount the Lustre File System?   Is the
>     ping successful and then fails after 03 seconds? If yes, you may
>     wish to check the /etc/lnet.conf file for Lustre LNet path
>     "discovery"  (1 allows LNet discovery while 0 does not), and
>     drop_asym_route (0 disallows asymmetrical routing while 1 permits
>     it).   I have worked with a few complex networks in which we chose
>     to turn off LNet discovery and specify, via /etc/lnet.conf, the
>     routes.  On one system the asymmetrical routing (we have 16 LNet
>     boxes between the system and the Lustre storage) seemed to be a
>     problem, but we couldn't pin it to any particular box.  On that
>     system disallowing asymmetrical routing seemed to help maintain
>     LNet/Lustre connectivity.
>
>     One may check the lctl ping to narrow down net connectivity from
>     other possibilities.
>
>     Cheers,
>     megan
>
>     On Mon, May 17, 2021 at 3:50 PM
>     <lustre-discuss-request at lists.lustre.org
>     <mailto:lustre-discuss-request at lists.lustre.org>> wrote:
>
>         Send lustre-discuss mailing list submissions to
>         lustre-discuss at lists.lustre.org
>         <mailto:lustre-discuss at lists.lustre.org>
>
>         To subscribe or unsubscribe via the World Wide Web, visit
>         http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
>         <http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org>
>         or, via email, send a message with subject or body 'help' to
>         lustre-discuss-request at lists.lustre.org
>         <mailto:lustre-discuss-request at lists.lustre.org>
>
>         You can reach the person managing the list at
>         lustre-discuss-owner at lists.lustre.org
>         <mailto:lustre-discuss-owner at lists.lustre.org>
>
>         When replying, please edit your Subject line so it is more
>         specific
>         than "Re: Contents of lustre-discuss digest..."
>
>
>         Today's Topics:
>
>            1. Re: problems to mount MDS and MDT (Abdeslam Tahari)
>            2. Re: problems to mount MDS and MDT (Colin Faber)
>
>
>         ----------------------------------------------------------------------
>
>         Message: 1
>         Date: Mon, 17 May 2021 21:35:34 +0200
>         From: Abdeslam Tahari <abeslam at gmail.com
>         <mailto:abeslam at gmail.com>>
>         To: Colin Faber <cfaber at gmail.com <mailto:cfaber at gmail.com>>
>         Cc: lustre-discuss <lustre-discuss at lists.lustre.org
>         <mailto:lustre-discuss at lists.lustre.org>>
>         Subject: Re: [lustre-discuss] problems to mount MDS and MDT
>         Message-ID:
>                
>         <CA+LuYSL9_TTcHopwHYbFRosZNgUFK=bxeCePEn5DzZD+QXnwiQ at mail.gmail.com
>         <mailto:bxeCePEn5DzZD%2BQXnwiQ at mail.gmail.com>>
>         Content-Type: text/plain; charset="utf-8"
>
>         Thank you Colin
>
>         No i don't have iptables or rules
>
>         firewalled is stopped selinux disabled as well
>          iptables -L
>         Chain INPUT (policy ACCEPT)
>         target     prot opt source               destination
>
>         Chain FORWARD (policy ACCEPT)
>         target     prot opt source               destination
>
>         Chain OUTPUT (policy ACCEPT)
>         target     prot opt source               destination
>
>
>         Regards
>
>
>         Regards
>
>         Le lun. 17 mai 2021 ? 21:29, Colin Faber <cfaber at gmail.com
>         <mailto:cfaber at gmail.com>> a ?crit :
>
>         > Firewall rules dealing with localhost?
>         >
>         > On Mon, May 17, 2021 at 11:33 AM Abdeslam Tahari via
>         lustre-discuss <
>         > lustre-discuss at lists.lustre.org
>         <mailto:lustre-discuss at lists.lustre.org>> wrote:
>         >
>         >> Hello
>         >>
>         >> i have a problem to mount the mds/mdt luster, it wont mount
>         at all and
>         >> there is no message errors at the console
>         >>
>         >> -it does not show errors or messages while mounting it
>         >>
>         >> here are some debug file logs
>         >>
>         >>
>         >> i specify it is a new project that i am doing.
>         >>
>         >> the version and packages of luter installed:
>         >> kmod-lustre-2.12.5-1.el7.x86_64
>         >> kernel-devel-3.10.0-1127.8.2.el7_lustre.x86_64
>         >> lustre-2.12.5-1.el7.x86_64
>         >> lustre-resource-agents-2.12.5-1.el7.x86_64
>         >> kernel-3.10.0-1160.2.1.el7_lustre.x86_64
>         >>
>         kernel-debuginfo-common-x86_64-3.10.0-1160.2.1.el7_lustre.x86_64
>         >> kmod-lustre-osd-ldiskfs-2.12.5-1.el7.x86_64
>         >> kernel-3.10.0-1127.8.2.el7_lustre.x86_64
>         >> lustre-osd-ldiskfs-mount-2.12.5-1.el7.x86_64
>         >>
>         >>
>         >>
>         >> the system(os) Centos 7
>         >>
>         >> the kernel
>         >> Linux lustre-mds1 3.10.0-1127.8.2.el7_lustre.x86_64
>         >>  cat /etc/redhat-release
>         >>
>         >>
>         >> when i mount the luster file-system it wont show up and no
>         errors
>         >>
>         >> mount -t lustre /dev/sda /mds
>         >>
>         >> lctl dl  does not show up
>         >>
>         >> df -h   no mount point for /dev/sda
>         >>
>         >>
>         >> lctl dl
>         >>
>         >> shows this:
>         >> lctl dl
>         >>   0 UP osd-ldiskfs lustre-MDT0000-osd lustre-MDT0000-osd_UUID 3
>         >>   2 UP mgc MGC10.0.1.70 at tcp
>         57e06c2d-5294-f034-fd95-460cee4f92b7 4
>         >>   3 UP mds MDS MDS_uuid 2
>         >>
>         >>
>         >> but unfortunately it disappears after 03 seconds
>         >>
>         >> lctl  dl shows nothing
>         >>
>         >> lctl dk
>         >>
>         >> shows this debug output
>         >>
>         >>
>         >>
>         00000020:00000080:18.0:1621276062.004338:0:13403:0:(obd_config.c:1128:class_process_config())
>         >> processing cmd: cf006
>         >>
>         00000020:00000080:18.0:1621276062.004341:0:13403:0:(obd_config.c:1147:class_process_config())
>         >> removing mappings for uuid MGC10.0.1.70 at tcp_0
>         >>
>         00000020:01000004:18.0:1621276062.004346:0:13403:0:(obd_mount.c:661:lustre_put_lsi())
>         >> put ffff9bbbf91d5800 1
>         >>
>         00000020:00000080:18.0:1621276062.004351:0:13403:0:(genops.c:1501:class_disconnect())
>         >> disconnect: cookie 0x256dd92fc5bf929c
>         >>
>         00000020:00000080:18.0:1621276062.004354:0:13403:0:(genops.c:1024:class_export_put())
>         >> final put ffff9bbf3e66a400/lustre-MDT0000-osd_UUID
>         >>
>         00000020:01000000:18.0:1621276062.004361:0:13403:0:(obd_config.c:2100:class_manual_cleanup())
>         >> Manual cleanup of lustre-MDT0000-osd (flags='')
>         >>
>         00000020:00000080:18.0:1621276062.004368:0:821:0:(genops.c:974:class_export_destroy())
>         >> destroying export ffff9bbf3e66a400/lustre-MDT0000-osd_UUID for
>         >> lustre-MDT0000-osd
>         >>
>         00000020:00000080:18.0:1621276062.004376:0:13403:0:(obd_config.c:1128:class_process_config())
>         >> processing cmd: cf004
>         >>
>         00000020:00000080:18.0:1621276062.004379:0:13403:0:(obd_config.c:659:class_cleanup())
>         >> lustre-MDT0000-osd: forcing exports to disconnect: 0/0
>         >>
>         00000020:00080000:18.0:1621276062.004382:0:13403:0:(genops.c:1590:class_disconnect_exports())
>         >> OBD device 0 (ffff9bbf47141080) has no exports
>         >>
>         00000020:00000080:18.0:1621276062.004788:0:13403:0:(obd_config.c:1128:class_process_config())
>         >> processing cmd: cf002
>         >>
>         00000020:00000080:18.0:1621276062.004791:0:13403:0:(obd_config.c:589:class_detach())
>         >> detach on obd lustre-MDT0000-osd (uuid lustre-MDT0000-osd_UUID)
>         >>
>         00000020:00000080:18.0:1621276062.004794:0:13403:0:(genops.c:1024:class_export_put())
>         >> final put ffff9bbf48800c00/lustre-MDT0000-osd_UUID
>         >>
>         00000020:00000080:18.0:1621276062.004796:0:13403:0:(genops.c:974:class_export_destroy())
>         >> destroying export ffff9bbf48800c00/lustre-MDT0000-osd_UUID for
>         >> lustre-MDT0000-osd
>         >>
>         00000020:01000000:18.0:1621276062.004799:0:13403:0:(genops.c:481:class_free_dev())
>         >> finishing cleanup of obd lustre-MDT0000-osd
>         (lustre-MDT0000-osd_UUID)
>         >>
>         00000020:01000004:18.0:1621276062.450759:0:13403:0:(obd_mount.c:605:lustre_free_lsi())
>         >> Freeing lsi ffff9bbbf91d6800
>         >>
>         00000020:01000000:18.0:1621276062.450805:0:13403:0:(obd_config.c:2100:class_manual_cleanup())
>         >> Manual cleanup of MDS (flags='F')
>         >>
>         00000020:00000080:18.0:1621276062.450806:0:13403:0:(obd_config.c:1128:class_process_config())
>         >> processing cmd: cf004
>         >>
>         00000020:00000080:18.0:1621276062.450807:0:13403:0:(obd_config.c:659:class_cleanup())
>         >> MDS: forcing exports to disconnect: 0/0
>         >>
>         00000020:00080000:18.0:1621276062.450809:0:13403:0:(genops.c:1590:class_disconnect_exports())
>         >> OBD device 3 (ffff9bbf43fdd280) has no exports
>         >>
>         00000020:00000080:58.0F:1621276062.490781:0:13403:0:(obd_config.c:1128:class_process_config())
>         >> processing cmd: cf002
>         >>
>         00000020:00000080:58.0:1621276062.490787:0:13403:0:(obd_config.c:589:class_detach())
>         >> detach on obd MDS (uuid MDS_uuid)
>         >>
>         00000020:00000080:58.0:1621276062.490788:0:13403:0:(genops.c:1024:class_export_put())
>         >> final put ffff9bbf3e668800/MDS_uuid
>         >>
>         00000020:00000080:58.0:1621276062.490790:0:13403:0:(genops.c:974:class_export_destroy())
>         >> destroying export ffff9bbf3e668800/MDS_uuid for MDS
>         >>
>         00000020:01000000:58.0:1621276062.490791:0:13403:0:(genops.c:481:class_free_dev())
>         >> finishing cleanup of obd MDS (MDS_uuid)
>         >>
>         00000020:02000400:58.0:1621276062.490877:0:13403:0:(obd_mount_server.c:1642:server_put_super())
>         >> server umount lustre-MDT0000 complete
>         >>
>         00000400:02020000:42.0:1621276086.284109:0:5400:0:(acceptor.c:321:lnet_accept())
>         >> 120-3: Refusing connection from 127.0.0.1 for
>         127.0.0.1 at tcp: No matching
>         >> NI
>         >>
>         00000800:00020000:6.0:1621276086.284152:0:5383:0:(socklnd_cb.c:1817:ksocknal_recv_hello())
>         >> Error -104 reading HELLO from 127.0.0.1
>         >>
>         00000400:02020000:6.0:1621276086.284174:0:5383:0:(acceptor.c:127:lnet_connect_console_error())
>         >> 11b-b: Connection to 127.0.0.1 at tcp at host 127.0.0.1 on
>         port 988 was
>         >> reset: is it running a compatible version of Lustre and is
>         127.0.0.1 at tcp
>         >> one of its NIDs?
>         >>
>         00000800:00000100:6.0:1621276086.284189:0:5383:0:(socklnd_cb.c:438:ksocknal_txlist_done())
>         >> Deleting packet type 2 len 0 10.0.1.70 at tcp->127.0.0.1 at tcp
>         >>
>         00000800:00000100:34.0:1621276136.363882:0:5401:0:(socklnd_cb.c:979:ksocknal_launch_packet())
>         >> No usable routes to 12345-127.0.0.1 at tcp
>         >>
>         00000400:02020000:42.0:1621276186.440095:0:5400:0:(acceptor.c:321:lnet_accept())
>         >> 120-3: Refusing connection from 127.0.0.1 for
>         127.0.0.1 at tcp: No matching
>         >> NI
>         >>
>         00000800:00020000:44.0:1621276186.446533:0:5386:0:(socklnd_cb.c:1817:ksocknal_recv_hello())
>         >> Error -104 reading HELLO from 127.0.0.1
>         >>
>         00000400:02020000:44.0:1621276186.452996:0:5386:0:(acceptor.c:127:lnet_connect_console_error())
>         >> 11b-b: Connection to 127.0.0.1 at tcp at host 127.0.0.1 on
>         port 988 was
>         >> reset: is it running a compatible version of Lustre and is
>         127.0.0.1 at tcp
>         >> one of its NIDs?
>         >>
>         00000800:00000100:44.0:1621276186.461433:0:5386:0:(socklnd_cb.c:438:ksocknal_txlist_done())
>         >> Deleting packet type 2 len 0 10.0.1.70 at tcp->127.0.0.1 at tcp
>         >> Debug log: 872 lines, 872 kept, 0 dropped, 0 bad.
>         >>
>         >>
>         >>
>         >> I just cant find out any help would be very appreciated
>         >>
>         >>
>         >> Thanks all
>         >>
>         >>
>         >>
>         >>
>         >>
>         >>
>         >> --
>         >> Tahari.Abdeslam
>         >> _______________________________________________
>         >> lustre-discuss mailing list
>         >> lustre-discuss at lists.lustre.org
>         <mailto:lustre-discuss at lists.lustre.org>
>         >>
>         http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
>         <http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org>
>         >>
>         >
>
>         -- 
>         Tahari.Abdeslam
>         -------------- next part --------------
>         An HTML attachment was scrubbed...
>         URL:
>         <http://lists.lustre.org/pipermail/lustre-discuss-lustre.org/attachments/20210517/1decdc97/attachment-0001.html
>         <http://lists.lustre.org/pipermail/lustre-discuss-lustre.org/attachments/20210517/1decdc97/attachment-0001.html>>
>
>         ------------------------------
>
>         Message: 2
>         Date: Mon, 17 May 2021 13:50:03 -0600
>         From: Colin Faber <cfaber at gmail.com <mailto:cfaber at gmail.com>>
>         To: Abdeslam Tahari <abeslam at gmail.com <mailto:abeslam at gmail.com>>
>         Cc: lustre-discuss <lustre-discuss at lists.lustre.org
>         <mailto:lustre-discuss at lists.lustre.org>>
>         Subject: Re: [lustre-discuss] problems to mount MDS and MDT
>         Message-ID:
>                
>         <CAJcXmB=T884j=5N8nhWspFBvNS+nAOoMa9b8xJUdhXT-fBoysw at mail.gmail.com
>         <mailto:5N8nhWspFBvNS%2BnAOoMa9b8xJUdhXT-fBoysw at mail.gmail.com>>
>         Content-Type: text/plain; charset="utf-8"
>
>         It appears part of the debug data is missing (the part before
>         you posted
>         it), Can you try again, lctl dk > /dev/null to clear it then
>         try your mount
>         and grab the debug again?
>
>         On Mon, May 17, 2021 at 1:35 PM Abdeslam Tahari
>         <abeslam at gmail.com <mailto:abeslam at gmail.com>> wrote:
>
>         > Thank you Colin
>         >
>         > No i don't have iptables or rules
>         >
>         > firewalled is stopped selinux disabled as well
>         >  iptables -L
>         > Chain INPUT (policy ACCEPT)
>         > target     prot opt source               destination
>         >
>         > Chain FORWARD (policy ACCEPT)
>         > target     prot opt source               destination
>         >
>         > Chain OUTPUT (policy ACCEPT)
>         > target     prot opt source               destination
>         >
>         >
>         > Regards
>         >
>         >
>         > Regards
>         >
>         > Le lun. 17 mai 2021 ? 21:29, Colin Faber <cfaber at gmail.com
>         <mailto:cfaber at gmail.com>> a ?crit :
>         >
>         >> Firewall rules dealing with localhost?
>         >>
>         >> On Mon, May 17, 2021 at 11:33 AM Abdeslam Tahari via
>         lustre-discuss <
>         >> lustre-discuss at lists.lustre.org
>         <mailto:lustre-discuss at lists.lustre.org>> wrote:
>         >>
>         >>> Hello
>         >>>
>         >>> i have a problem to mount the mds/mdt luster, it wont
>         mount at all and
>         >>> there is no message errors at the console
>         >>>
>         >>> -it does not show errors or messages while mounting it
>         >>>
>         >>> here are some debug file logs
>         >>>
>         >>>
>         >>> i specify it is a new project that i am doing.
>         >>>
>         >>> the version and packages of luter installed:
>         >>> kmod-lustre-2.12.5-1.el7.x86_64
>         >>> kernel-devel-3.10.0-1127.8.2.el7_lustre.x86_64
>         >>> lustre-2.12.5-1.el7.x86_64
>         >>> lustre-resource-agents-2.12.5-1.el7.x86_64
>         >>> kernel-3.10.0-1160.2.1.el7_lustre.x86_64
>         >>>
>         kernel-debuginfo-common-x86_64-3.10.0-1160.2.1.el7_lustre.x86_64
>         >>> kmod-lustre-osd-ldiskfs-2.12.5-1.el7.x86_64
>         >>> kernel-3.10.0-1127.8.2.el7_lustre.x86_64
>         >>> lustre-osd-ldiskfs-mount-2.12.5-1.el7.x86_64
>         >>>
>         >>>
>         >>>
>         >>> the system(os) Centos 7
>         >>>
>         >>> the kernel
>         >>> Linux lustre-mds1 3.10.0-1127.8.2.el7_lustre.x86_64
>         >>>  cat /etc/redhat-release
>         >>>
>         >>>
>         >>> when i mount the luster file-system it wont show up and no
>         errors
>         >>>
>         >>> mount -t lustre /dev/sda /mds
>         >>>
>         >>> lctl dl  does not show up
>         >>>
>         >>> df -h   no mount point for /dev/sda
>         >>>
>         >>>
>         >>> lctl dl
>         >>>
>         >>> shows this:
>         >>> lctl dl
>         >>>   0 UP osd-ldiskfs lustre-MDT0000-osd
>         lustre-MDT0000-osd_UUID 3
>         >>>   2 UP mgc MGC10.0.1.70 at tcp
>         57e06c2d-5294-f034-fd95-460cee4f92b7 4
>         >>>   3 UP mds MDS MDS_uuid 2
>         >>>
>         >>>
>         >>> but unfortunately it disappears after 03 seconds
>         >>>
>         >>> lctl  dl shows nothing
>         >>>
>         >>> lctl dk
>         >>>
>         >>> shows this debug output
>         >>>
>         >>>
>         >>>
>         00000020:00000080:18.0:1621276062.004338:0:13403:0:(obd_config.c:1128:class_process_config())
>         >>> processing cmd: cf006
>         >>>
>         00000020:00000080:18.0:1621276062.004341:0:13403:0:(obd_config.c:1147:class_process_config())
>         >>> removing mappings for uuid MGC10.0.1.70 at tcp_0
>         >>>
>         00000020:01000004:18.0:1621276062.004346:0:13403:0:(obd_mount.c:661:lustre_put_lsi())
>         >>> put ffff9bbbf91d5800 1
>         >>>
>         00000020:00000080:18.0:1621276062.004351:0:13403:0:(genops.c:1501:class_disconnect())
>         >>> disconnect: cookie 0x256dd92fc5bf929c
>         >>>
>         00000020:00000080:18.0:1621276062.004354:0:13403:0:(genops.c:1024:class_export_put())
>         >>> final put ffff9bbf3e66a400/lustre-MDT0000-osd_UUID
>         >>>
>         00000020:01000000:18.0:1621276062.004361:0:13403:0:(obd_config.c:2100:class_manual_cleanup())
>         >>> Manual cleanup of lustre-MDT0000-osd (flags='')
>         >>>
>         00000020:00000080:18.0:1621276062.004368:0:821:0:(genops.c:974:class_export_destroy())
>         >>> destroying export ffff9bbf3e66a400/lustre-MDT0000-osd_UUID for
>         >>> lustre-MDT0000-osd
>         >>>
>         00000020:00000080:18.0:1621276062.004376:0:13403:0:(obd_config.c:1128:class_process_config())
>         >>> processing cmd: cf004
>         >>>
>         00000020:00000080:18.0:1621276062.004379:0:13403:0:(obd_config.c:659:class_cleanup())
>         >>> lustre-MDT0000-osd: forcing exports to disconnect: 0/0
>         >>>
>         00000020:00080000:18.0:1621276062.004382:0:13403:0:(genops.c:1590:class_disconnect_exports())
>         >>> OBD device 0 (ffff9bbf47141080) has no exports
>         >>>
>         00000020:00000080:18.0:1621276062.004788:0:13403:0:(obd_config.c:1128:class_process_config())
>         >>> processing cmd: cf002
>         >>>
>         00000020:00000080:18.0:1621276062.004791:0:13403:0:(obd_config.c:589:class_detach())
>         >>> detach on obd lustre-MDT0000-osd (uuid
>         lustre-MDT0000-osd_UUID)
>         >>>
>         00000020:00000080:18.0:1621276062.004794:0:13403:0:(genops.c:1024:class_export_put())
>         >>> final put ffff9bbf48800c00/lustre-MDT0000-osd_UUID
>         >>>
>         00000020:00000080:18.0:1621276062.004796:0:13403:0:(genops.c:974:class_export_destroy())
>         >>> destroying export ffff9bbf48800c00/lustre-MDT0000-osd_UUID for
>         >>> lustre-MDT0000-osd
>         >>>
>         00000020:01000000:18.0:1621276062.004799:0:13403:0:(genops.c:481:class_free_dev())
>         >>> finishing cleanup of obd lustre-MDT0000-osd
>         (lustre-MDT0000-osd_UUID)
>         >>>
>         00000020:01000004:18.0:1621276062.450759:0:13403:0:(obd_mount.c:605:lustre_free_lsi())
>         >>> Freeing lsi ffff9bbbf91d6800
>         >>>
>         00000020:01000000:18.0:1621276062.450805:0:13403:0:(obd_config.c:2100:class_manual_cleanup())
>         >>> Manual cleanup of MDS (flags='F')
>         >>>
>         00000020:00000080:18.0:1621276062.450806:0:13403:0:(obd_config.c:1128:class_process_config())
>         >>> processing cmd: cf004
>         >>>
>         00000020:00000080:18.0:1621276062.450807:0:13403:0:(obd_config.c:659:class_cleanup())
>         >>> MDS: forcing exports to disconnect: 0/0
>         >>>
>         00000020:00080000:18.0:1621276062.450809:0:13403:0:(genops.c:1590:class_disconnect_exports())
>         >>> OBD device 3 (ffff9bbf43fdd280) has no exports
>         >>>
>         00000020:00000080:58.0F:1621276062.490781:0:13403:0:(obd_config.c:1128:class_process_config())
>         >>> processing cmd: cf002
>         >>>
>         00000020:00000080:58.0:1621276062.490787:0:13403:0:(obd_config.c:589:class_detach())
>         >>> detach on obd MDS (uuid MDS_uuid)
>         >>>
>         00000020:00000080:58.0:1621276062.490788:0:13403:0:(genops.c:1024:class_export_put())
>         >>> final put ffff9bbf3e668800/MDS_uuid
>         >>>
>         00000020:00000080:58.0:1621276062.490790:0:13403:0:(genops.c:974:class_export_destroy())
>         >>> destroying export ffff9bbf3e668800/MDS_uuid for MDS
>         >>>
>         00000020:01000000:58.0:1621276062.490791:0:13403:0:(genops.c:481:class_free_dev())
>         >>> finishing cleanup of obd MDS (MDS_uuid)
>         >>>
>         00000020:02000400:58.0:1621276062.490877:0:13403:0:(obd_mount_server.c:1642:server_put_super())
>         >>> server umount lustre-MDT0000 complete
>         >>>
>         00000400:02020000:42.0:1621276086.284109:0:5400:0:(acceptor.c:321:lnet_accept())
>         >>> 120-3: Refusing connection from 127.0.0.1 for
>         127.0.0.1 at tcp: No
>         >>> matching NI
>         >>>
>         00000800:00020000:6.0:1621276086.284152:0:5383:0:(socklnd_cb.c:1817:ksocknal_recv_hello())
>         >>> Error -104 reading HELLO from 127.0.0.1
>         >>>
>         00000400:02020000:6.0:1621276086.284174:0:5383:0:(acceptor.c:127:lnet_connect_console_error())
>         >>> 11b-b: Connection to 127.0.0.1 at tcp at host 127.0.0.1 on
>         port 988 was
>         >>> reset: is it running a compatible version of Lustre and is
>         127.0.0.1 at tcp
>         >>> one of its NIDs?
>         >>>
>         00000800:00000100:6.0:1621276086.284189:0:5383:0:(socklnd_cb.c:438:ksocknal_txlist_done())
>         >>> Deleting packet type 2 len 0 10.0.1.70 at tcp->127.0.0.1 at tcp
>         >>>
>         00000800:00000100:34.0:1621276136.363882:0:5401:0:(socklnd_cb.c:979:ksocknal_launch_packet())
>         >>> No usable routes to 12345-127.0.0.1 at tcp
>         >>>
>         00000400:02020000:42.0:1621276186.440095:0:5400:0:(acceptor.c:321:lnet_accept())
>         >>> 120-3: Refusing connection from 127.0.0.1 for
>         127.0.0.1 at tcp: No
>         >>> matching NI
>         >>>
>         00000800:00020000:44.0:1621276186.446533:0:5386:0:(socklnd_cb.c:1817:ksocknal_recv_hello())
>         >>> Error -104 reading HELLO from 127.0.0.1
>         >>>
>         00000400:02020000:44.0:1621276186.452996:0:5386:0:(acceptor.c:127:lnet_connect_console_error())
>         >>> 11b-b: Connection to 127.0.0.1 at tcp at host 127.0.0.1 on
>         port 988 was
>         >>> reset: is it running a compatible version of Lustre and is
>         127.0.0.1 at tcp
>         >>> one of its NIDs?
>         >>>
>         00000800:00000100:44.0:1621276186.461433:0:5386:0:(socklnd_cb.c:438:ksocknal_txlist_done())
>         >>> Deleting packet type 2 len 0 10.0.1.70 at tcp->127.0.0.1 at tcp
>         >>> Debug log: 872 lines, 872 kept, 0 dropped, 0 bad.
>         >>>
>         >>>
>         >>>
>         >>> I just cant find out any help would be very appreciated
>         >>>
>         >>>
>         >>> Thanks all
>         >>>
>         >>>
>         >>>
>         >>>
>         >>>
>         >>>
>         >>> --
>         >>> Tahari.Abdeslam
>         >>> _______________________________________________
>         >>> lustre-discuss mailing list
>         >>> lustre-discuss at lists.lustre.org
>         <mailto:lustre-discuss at lists.lustre.org>
>         >>>
>         http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
>         <http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org>
>         >>>
>         >>
>         >
>         > --
>         > Tahari.Abdeslam
>         >
>         -------------- next part --------------
>         An HTML attachment was scrubbed...
>         URL:
>         <http://lists.lustre.org/pipermail/lustre-discuss-lustre.org/attachments/20210517/2adc6c81/attachment.html
>         <http://lists.lustre.org/pipermail/lustre-discuss-lustre.org/attachments/20210517/2adc6c81/attachment.html>>
>
>         ------------------------------
>
>         Subject: Digest Footer
>
>         _______________________________________________
>         lustre-discuss mailing list
>         lustre-discuss at lists.lustre.org
>         <mailto:lustre-discuss at lists.lustre.org>
>         http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
>         <http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org>
>
>
>         ------------------------------
>
>         End of lustre-discuss Digest, Vol 182, Issue 12
>         ***********************************************
>
>     _______________________________________________
>     lustre-discuss mailing list
>     lustre-discuss at lists.lustre.org
>     <mailto:lustre-discuss at lists.lustre.org>
>     http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
>     <http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org>
>
>
>
> -- 
> Tahari.Abdeslam
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.lustre.org/pipermail/lustre-discuss-lustre.org/attachments/20210519/b768a741/attachment-0001.html>


More information about the lustre-discuss mailing list