#ceph IRC Log

Index

IRC Log for 2016-09-07

Timestamps are in GMT/BST.

[0:01] * wjw-freebsd (~wjw@smtp.digiware.nl) Quit (Ping timeout: 480 seconds)
[0:04] * rendar (~I@host197-114-dynamic.57-82-r.retail.telecomitalia.it) Quit (Quit: std::lower_bound + std::less_equal *works* with a vector without duplicates!)
[0:17] * danieagle (~Daniel@200-171-231-229.customer.telesp.net.br) Quit (Quit: Obrigado por Tudo! :-) inte+ :-))
[0:17] * sudocat1 (~dibarra@192.185.1.20) Quit (Ping timeout: 480 seconds)
[0:18] * Kingrat (~shiny@2605:6000:1526:4063:d44d:3add:dc13:51ec) has joined #ceph
[0:21] * N3X15 (~lobstar@108.61.122.72) Quit ()
[0:21] * Kyso_ (~offer@m95-153-32-10.cust.tele2.ee) has joined #ceph
[0:32] * wjw-freebsd (~wjw@smtp.digiware.nl) has joined #ceph
[0:34] * wjw-freebsd2 (~wjw@smtp.digiware.nl) Quit (Ping timeout: 480 seconds)
[0:48] * johnavp1989 (~jpetrini@pool-100-14-10-2.phlapa.fios.verizon.net) has joined #ceph
[0:48] <- *johnavp1989* To prove that you are human, please enter the result of 8+3
[0:50] * vasu (~vasu@c-73-231-60-138.hsd1.ca.comcast.net) Quit (Quit: Leaving)
[0:51] * Kyso_ (~offer@m95-153-32-10.cust.tele2.ee) Quit ()
[0:56] * mLegion (~Neon@209.222.77.220) has joined #ceph
[0:57] * srk (~Siva@cpe-70-113-23-93.austin.res.rr.com) has joined #ceph
[0:58] * stiopa (~stiopa@cpc73832-dals21-2-0-cust453.20-2.cable.virginm.net) Quit (Ping timeout: 480 seconds)
[1:03] * Discovery (~Discovery@109.235.52.5) Quit (Read error: Connection reset by peer)
[1:09] * kwork_ (~quassel@bnc.ee) has joined #ceph
[1:10] * fsimonce (~simon@host203-44-dynamic.183-80-r.retail.telecomitalia.it) Quit (Ping timeout: 480 seconds)
[1:11] * kwork (~quassel@bnc.ee) Quit (Ping timeout: 480 seconds)
[1:17] * srk (~Siva@cpe-70-113-23-93.austin.res.rr.com) Quit (Ping timeout: 480 seconds)
[1:22] * andreww (~xarses@c-73-202-191-48.hsd1.ca.comcast.net) has joined #ceph
[1:23] * blizzow (~jburns@50-243-148-102-static.hfc.comcastbusiness.net) Quit (Ping timeout: 480 seconds)
[1:25] * AndroUser2 (~androirc@107.170.0.159) has joined #ceph
[1:26] * mLegion (~Neon@209.222.77.220) Quit ()
[1:29] * AndroUser2 (~androirc@107.170.0.159) Quit (Remote host closed the connection)
[1:33] * ircolle (~Adium@2601:285:201:633a:19f:e2d7:59a2:44a1) Quit (Quit: Leaving.)
[1:33] * scg (~zscg@181.122.4.166) Quit (Ping timeout: 480 seconds)
[1:41] * wjw-freebsd2 (~wjw@smtp.digiware.nl) has joined #ceph
[1:42] * _mrp (~mrp@93-86-57-229.dynamic.isp.telekom.rs) Quit (Quit: My Mac has gone to sleep. ZZZzzz???)
[1:44] * wjw-freebsd (~wjw@smtp.digiware.nl) Quit (Ping timeout: 480 seconds)
[1:44] * wjw-freebsd (~wjw@smtp.digiware.nl) has joined #ceph
[1:44] * wjw-freebsd2 (~wjw@smtp.digiware.nl) Quit (Read error: Connection reset by peer)
[1:48] * srk (~Siva@2605:6000:ed04:ce00:c534:b089:8e72:1a8a) has joined #ceph
[1:48] * Jeffrey4l_ (~Jeffrey@110.252.41.177) has joined #ceph
[1:53] * n0x1d (~xolotl@185.65.134.74) has joined #ceph
[1:55] * johnavp1989 (~jpetrini@pool-100-14-10-2.phlapa.fios.verizon.net) Quit (Quit: Leaving.)
[1:55] * johnavp1989 (~jpetrini@pool-100-14-10-2.phlapa.fios.verizon.net) has joined #ceph
[1:55] <- *johnavp1989* To prove that you are human, please enter the result of 8+3
[1:56] * oms101_ (~oms101@p20030057EA04AE00C6D987FFFE4339A1.dip0.t-ipconnect.de) Quit (Ping timeout: 480 seconds)
[1:58] * Jeffrey4l_ (~Jeffrey@110.252.41.177) Quit (Ping timeout: 480 seconds)
[2:00] * srk_ (~Siva@cpe-70-113-23-93.austin.res.rr.com) has joined #ceph
[2:02] * wushudoin (~wushudoin@2601:646:8200:c9f0:2ab2:bdff:fe0b:a6ee) Quit (Ping timeout: 480 seconds)
[2:04] * oms101_ (~oms101@p20030057EA035E00C6D987FFFE4339A1.dip0.t-ipconnect.de) has joined #ceph
[2:05] * srk (~Siva@2605:6000:ed04:ce00:c534:b089:8e72:1a8a) Quit (Ping timeout: 480 seconds)
[2:06] * wes_dillingham (~wes_dilli@209-6-222-74.c3-0.hdp-ubr1.sbo-hdp.ma.cable.rcn.com) has joined #ceph
[2:11] * doppelgrau (~doppelgra@dslb-088-072-094-200.088.072.pools.vodafone-ip.de) Quit (Quit: doppelgrau)
[2:15] * AndroUser2 (~androirc@107.170.0.159) has joined #ceph
[2:18] * AndroUser2 (~androirc@107.170.0.159) Quit (Remote host closed the connection)
[2:19] * kristen (~kristen@134.134.139.76) Quit (Quit: Leaving)
[2:22] * AndroUser2 (~androirc@107.170.0.159) has joined #ceph
[2:23] * n0x1d (~xolotl@185.65.134.74) Quit ()
[2:31] * AndroUser2 (~androirc@107.170.0.159) Quit (Ping timeout: 480 seconds)
[2:44] * huangjun (~kvirc@113.57.168.154) has joined #ceph
[2:44] * srk_ (~Siva@cpe-70-113-23-93.austin.res.rr.com) Quit (Ping timeout: 480 seconds)
[2:46] * davidzlap (~Adium@2605:e000:1313:8003:9025:525b:2aea:f861) Quit (Quit: Leaving.)
[2:47] * Jeffrey4l_ (~Jeffrey@110.252.41.177) has joined #ceph
[2:49] * kuku (~kuku@119.93.91.136) has joined #ceph
[3:04] * vbellur (~vijay@2601:18f:700:55b0:5e51:4fff:fee8:6a5c) has joined #ceph
[3:23] * jfaj (~jan@p4FC5BE73.dip0.t-ipconnect.de) has joined #ceph
[3:30] * jfaj__ (~jan@p4FC2528F.dip0.t-ipconnect.de) Quit (Ping timeout: 480 seconds)
[3:43] * kefu (~kefu@114.92.125.128) has joined #ceph
[3:47] * elfurbe (~Adam@saint.uits.arizona.edu) Quit (Quit: Textual IRC Client: www.textualapp.com)
[3:55] * EinstCrazy (~EinstCraz@58.247.119.250) has joined #ceph
[3:55] * srk_ (~Siva@cpe-70-113-23-93.austin.res.rr.com) has joined #ceph
[3:57] * huangjun|2 (~kvirc@113.57.168.154) has joined #ceph
[3:57] * huangjun (~kvirc@113.57.168.154) Quit (Read error: Connection reset by peer)
[3:58] * EinstCrazy (~EinstCraz@58.247.119.250) Quit (Remote host closed the connection)
[4:03] * EinstCrazy (~EinstCraz@58.247.119.250) has joined #ceph
[4:06] * shyu (~Frank@218.241.172.114) has joined #ceph
[4:07] * wes_dillingham (~wes_dilli@209-6-222-74.c3-0.hdp-ubr1.sbo-hdp.ma.cable.rcn.com) Quit (Quit: wes_dillingham)
[4:12] * LegalResale (~LegalResa@66.165.126.130) has joined #ceph
[4:17] * bene2 (~bene@2601:193:4101:f410:ea2a:eaff:fe08:3c7a) Quit (Quit: Konversation terminated!)
[4:23] * flisky (~Thunderbi@106.38.61.187) has joined #ceph
[4:26] * davidzlap (~Adium@cpe-172-91-154-245.socal.res.rr.com) has joined #ceph
[4:29] * srk_ (~Siva@cpe-70-113-23-93.austin.res.rr.com) Quit (Read error: Connection reset by peer)
[4:29] * srk_ (~Siva@cpe-70-113-23-93.austin.res.rr.com) has joined #ceph
[4:40] * johnavp1989 (~jpetrini@pool-100-14-10-2.phlapa.fios.verizon.net) Quit (Ping timeout: 480 seconds)
[4:52] * vicente (~~vicente@125-227-238-55.HINET-IP.hinet.net) has joined #ceph
[4:53] * johnavp1989 (~jpetrini@pool-100-14-10-2.phlapa.fios.verizon.net) has joined #ceph
[4:53] <- *johnavp1989* To prove that you are human, please enter the result of 8+3
[4:56] * davidzlap (~Adium@cpe-172-91-154-245.socal.res.rr.com) Quit (Quit: Leaving.)
[4:59] * ebersnack (~oftc-webi@CPEf81a67d09b8e-CM78cd8e7b49fd.cpe.net.cable.rogers.com) has joined #ceph
[4:59] <ebersnack> does anyone have a good doc on non-trivial CRUSH maps? (other than the online doc and The Paper)
[5:00] * skorgu_ (skorgu@pylon.skorgu.net) Quit (Remote host closed the connection)
[5:06] * johnavp1989 (~jpetrini@pool-100-14-10-2.phlapa.fios.verizon.net) Quit (Ping timeout: 480 seconds)
[5:13] * srk_ (~Siva@cpe-70-113-23-93.austin.res.rr.com) Quit (Read error: Connection reset by peer)
[5:14] * srk (~Siva@cpe-70-113-23-93.austin.res.rr.com) has joined #ceph
[5:16] * neurodrone__ (~neurodron@pool-100-35-225-168.nwrknj.fios.verizon.net) Quit (Quit: neurodrone__)
[5:18] * andreww (~xarses@c-73-202-191-48.hsd1.ca.comcast.net) Quit (Remote host closed the connection)
[5:23] <evilrob> so I'm not going to bang on it all night, but I've got 5 OSDs that last logged "done with init, starting boot process" still show "down" in "ceph osd tree" but they're not logging anyting
[5:24] <evilrob> I rebuilt one from scratch completely (well, I didn't make a new filesystem, but I blew away everything on it)
[5:24] <evilrob> so I'm confused
[5:28] * Vacuum_ (~Vacuum@i59F79635.versanet.de) has joined #ceph
[5:32] * davidzlap (~Adium@cpe-172-91-154-245.socal.res.rr.com) has joined #ceph
[5:34] * srk (~Siva@cpe-70-113-23-93.austin.res.rr.com) Quit (Ping timeout: 480 seconds)
[5:35] * Vacuum__ (~Vacuum@88.130.195.185) Quit (Ping timeout: 480 seconds)
[5:36] * davidzlap (~Adium@cpe-172-91-154-245.socal.res.rr.com) Quit ()
[5:39] * skorgu (skorgu@pylon.skorgu.net) has joined #ceph
[5:40] <evilrob> sigh... it's that we didn't pass through hammer on our way up. somehow these were 0.80.11 OSDs.
[5:44] * vimal (~vikumar@114.143.162.59) has joined #ceph
[5:48] * skorgu_ (skorgu@pylon.skorgu.net) has joined #ceph
[5:57] * davidzlap (~Adium@2605:e000:1313:8003:959c:6f04:2e8c:11df) has joined #ceph
[6:01] * skorgu (skorgu@pylon.skorgu.net) Quit (Quit: leaving)
[6:01] * ebersnack (~oftc-webi@CPEf81a67d09b8e-CM78cd8e7b49fd.cpe.net.cable.rogers.com) Quit (Quit: Page closed)
[6:03] * walcubi_ (~walcubi@87.149.186.180) has joined #ceph
[6:05] <evilrob> so we had infernalis mon data on disk and giant osd data. hammer mon code didn't know what to do, so I had to start hammer OSDs, upgrade the package to infernalis, start those monitors (thank god they can still talk to each other) and run that way while I'm recovering. then I'll do my 3rd node.
[6:10] * walcubi (~walcubi@p5795BC82.dip0.t-ipconnect.de) Quit (Ping timeout: 480 seconds)
[6:12] * vimal (~vikumar@114.143.162.59) Quit (Quit: Leaving)
[6:15] * wjw-freebsd (~wjw@smtp.digiware.nl) Quit (Ping timeout: 480 seconds)
[6:20] * davidzlap (~Adium@2605:e000:1313:8003:959c:6f04:2e8c:11df) Quit (Quit: Leaving.)
[6:23] * pvh_sa (~pvh@169-0-184-226.ip.afrihost.co.za) has joined #ceph
[6:31] * vimal (~vikumar@121.244.87.116) has joined #ceph
[6:33] * jclm (~jclm@ip68-96-196-245.lv.lv.cox.net) has joined #ceph
[6:33] * jclm (~jclm@ip68-96-196-245.lv.lv.cox.net) Quit ()
[6:34] * rdas (~rdas@125.212.212.209) has joined #ceph
[6:34] * rdas (~rdas@125.212.212.209) Quit (Read error: Connection reset by peer)
[6:39] * tritonx (~zviratko@81-248-47-212.rev.cloud.scaleway.com) has joined #ceph
[6:40] * pvh_sa (~pvh@169-0-184-226.ip.afrihost.co.za) Quit (Ping timeout: 480 seconds)
[6:40] * kefu (~kefu@114.92.125.128) Quit (Quit: My Mac has gone to sleep. ZZZzzz???)
[6:41] * pvh_sa (~pvh@169-0-184-226.ip.afrihost.co.za) has joined #ceph
[6:42] * rdas (~rdas@125.212.212.209) has joined #ceph
[6:46] * ronrib (~boswortr@45.32.242.135) Quit (Ping timeout: 480 seconds)
[6:50] * ivve (~zed@cust-gw-11.se.zetup.net) has joined #ceph
[6:52] * vikhyat (~vumrao@121.244.87.116) has joined #ceph
[6:53] * swami1 (~swami@49.38.3.250) has joined #ceph
[7:00] * Jeffrey4l__ (~Jeffrey@110.252.60.139) has joined #ceph
[7:01] * whatevsz (~quassel@b9168e43.cgn.dg-w.de) has joined #ceph
[7:03] * pvh_sa (~pvh@169-0-184-226.ip.afrihost.co.za) Quit (Ping timeout: 480 seconds)
[7:04] * kefu (~kefu@114.92.125.128) has joined #ceph
[7:06] * xarses (~xarses@c-73-202-191-48.hsd1.ca.comcast.net) has joined #ceph
[7:06] * TomasCZ (~TomasCZ@yes.tenlab.net) Quit (Quit: Leaving)
[7:07] * rotbeard (~redbeard@185.32.80.238) has joined #ceph
[7:08] * Jeffrey4l_ (~Jeffrey@110.252.41.177) Quit (Ping timeout: 480 seconds)
[7:09] * tritonx (~zviratko@81-248-47-212.rev.cloud.scaleway.com) Quit ()
[7:09] * rdas (~rdas@125.212.212.209) Quit (Quit: Leaving)
[7:19] * pvh_sa (~pvh@169-0-184-226.ip.afrihost.co.za) has joined #ceph
[7:23] * nastidon (~Quackie@108.61.123.82) has joined #ceph
[7:23] * skorgu_ (skorgu@pylon.skorgu.net) Quit (synthon.oftc.net oxygen.oftc.net)
[7:23] * LegalResale (~LegalResa@66.165.126.130) Quit (synthon.oftc.net oxygen.oftc.net)
[7:23] * wkennington (~wkenningt@c-71-204-170-241.hsd1.ca.comcast.net) Quit (synthon.oftc.net oxygen.oftc.net)
[7:23] * pdrakeweb (~pdrakeweb@cpe-71-74-153-111.neo.res.rr.com) Quit (synthon.oftc.net oxygen.oftc.net)
[7:23] * squizzi (~squizzi@107.13.237.240) Quit (synthon.oftc.net oxygen.oftc.net)
[7:23] * nhm (~nhm@c-50-171-139-246.hsd1.mn.comcast.net) Quit (synthon.oftc.net oxygen.oftc.net)
[7:23] * KindOne (kindone@0001a7db.user.oftc.net) Quit (synthon.oftc.net oxygen.oftc.net)
[7:23] * dneary (~dneary@nat-pool-bos-u.redhat.com) Quit (synthon.oftc.net oxygen.oftc.net)
[7:23] * badone (~badone@66.187.239.16) Quit (synthon.oftc.net oxygen.oftc.net)
[7:23] * KindOne (kindone@h248.177.190.173.ip.windstream.net) has joined #ceph
[7:25] * skorgu_ (skorgu@pylon.skorgu.net) has joined #ceph
[7:25] * LegalResale (~LegalResa@66.165.126.130) has joined #ceph
[7:25] * wkennington (~wkenningt@c-71-204-170-241.hsd1.ca.comcast.net) has joined #ceph
[7:25] * pdrakeweb (~pdrakeweb@cpe-71-74-153-111.neo.res.rr.com) has joined #ceph
[7:25] * squizzi (~squizzi@107.13.237.240) has joined #ceph
[7:25] * nhm (~nhm@c-50-171-139-246.hsd1.mn.comcast.net) has joined #ceph
[7:25] * dneary (~dneary@nat-pool-bos-u.redhat.com) has joined #ceph
[7:25] * badone (~badone@66.187.239.16) has joined #ceph
[7:25] * ChanServ sets mode +v nhm
[7:26] * LegalResale (~LegalResa@66.165.126.130) Quit (Ping timeout: 481 seconds)
[7:30] * EinstCrazy (~EinstCraz@58.247.119.250) Quit (Remote host closed the connection)
[7:31] * EinstCrazy (~EinstCraz@58.247.119.250) has joined #ceph
[7:34] * kefu (~kefu@114.92.125.128) Quit (Quit: My Mac has gone to sleep. ZZZzzz???)
[7:34] * ronrib (~boswortr@45.32.242.135) has joined #ceph
[7:34] <ivve> anyone knows if "osd deep scrub interval" needs to be set on monitors as well?
[7:35] <ivve> after setting the value, i run a deepscrub but it's still scheduled at 1 week from now, i have increased to 4weeks
[7:36] * LegalResale (~LegalResa@66.165.126.130) has joined #ceph
[7:38] * vimal (~vikumar@121.244.87.116) Quit (Ping timeout: 480 seconds)
[7:41] * coreping (~Michael_G@n1.coreping.org) has joined #ceph
[7:49] * vimal (~vikumar@121.244.87.124) has joined #ceph
[7:51] * kefu (~kefu@114.92.125.128) has joined #ceph
[7:53] * nastidon (~Quackie@108.61.123.82) Quit ()
[7:55] * Be-El (~blinke@nat-router.computational.bio.uni-giessen.de) has joined #ceph
[7:55] * vimal (~vikumar@121.244.87.124) Quit (Quit: Leaving)
[7:56] * stiopa (~stiopa@cpc73832-dals21-2-0-cust453.20-2.cable.virginm.net) has joined #ceph
[7:59] * vimal (~vikumar@121.244.87.116) has joined #ceph
[8:02] * Skaag (~lunix@cpe-172-91-77-84.socal.res.rr.com) has joined #ceph
[8:05] * kuku (~kuku@119.93.91.136) Quit (Remote host closed the connection)
[8:05] * kuku (~kuku@119.93.91.136) has joined #ceph
[8:06] * kuku (~kuku@119.93.91.136) Quit (Remote host closed the connection)
[8:06] * EinstCrazy (~EinstCraz@58.247.119.250) Quit (Remote host closed the connection)
[8:06] * kuku (~kuku@119.93.91.136) has joined #ceph
[8:07] * stiopa (~stiopa@cpc73832-dals21-2-0-cust453.20-2.cable.virginm.net) Quit (Ping timeout: 480 seconds)
[8:09] * kuku (~kuku@119.93.91.136) Quit (Remote host closed the connection)
[8:10] * EinstCrazy (~EinstCraz@58.247.117.134) has joined #ceph
[8:10] * kuku (~kuku@119.93.91.136) has joined #ceph
[8:14] * derjohn_mob (~aj@p578b6aa1.dip0.t-ipconnect.de) has joined #ceph
[8:15] * Maariu5_ (~Xylios@142.4.205.238) has joined #ceph
[8:29] * rendar (~I@host5-183-dynamic.46-79-r.retail.telecomitalia.it) has joined #ceph
[8:38] * wenduo (~wenduo@218.30.116.7) has joined #ceph
[8:41] * rakeshgm (~rakesh@121.244.87.117) has joined #ceph
[8:41] * treenerd_ (~gsulzberg@cpe90-146-148-47.liwest.at) has joined #ceph
[8:45] * Maariu5_ (~Xylios@142.4.205.238) Quit ()
[8:48] * doppelgrau (~doppelgra@dslb-088-072-094-200.088.072.pools.vodafone-ip.de) has joined #ceph
[8:48] * karnan (~karnan@121.244.87.117) has joined #ceph
[8:54] * EinstCrazy (~EinstCraz@58.247.117.134) Quit (Remote host closed the connection)
[8:54] * moegyver (~moe@212.85.78.250) has joined #ceph
[8:59] * Hemanth (~hkumar_@121.244.87.117) has joined #ceph
[9:00] * ashah (~ashah@121.244.87.117) has joined #ceph
[9:00] * Skaag (~lunix@cpe-172-91-77-84.socal.res.rr.com) Quit (Quit: Leaving.)
[9:03] * EinstCrazy (~EinstCraz@58.247.117.134) has joined #ceph
[9:15] <ivve> anyone knows if "osd deep scrub interval" needs to be set on monitors as well? after setting the value, i run a deepscrub but pgs are still scheduled at 1 week from now, i have increased to 4weeks
[9:15] * bara (~bara@nat-pool-brq-t.redhat.com) has joined #ceph
[9:17] <Jeeves_> Don't you need to restart your osds for it to take effect?
[9:18] * dugravot61 (~dugravot6@nat-persul-plg.wifi.univ-lorraine.fr) has joined #ceph
[9:19] * fsimonce (~simon@host145-64-dynamic.52-79-r.retail.telecomitalia.it) has joined #ceph
[9:22] * derjohn_mob (~aj@p578b6aa1.dip0.t-ipconnect.de) Quit (Ping timeout: 480 seconds)
[9:22] * nilez (~nilez@96.44.143.234) Quit (Ping timeout: 480 seconds)
[9:22] * nilez (~nilez@96.44.144.194) has joined #ceph
[9:23] <ivve> unsure
[9:23] <ivve> however it seems that deepscrubs have something to do with regular scrubs as well
[9:23] * dugravot6 (~dugravot6@l-p-dn-in-4a.lionnois.site.univ-lorraine.fr) Quit (Ping timeout: 480 seconds)
[9:24] <ivve> thinking about your question, values like osd recovery or backfill doesn't require a restart of the osd
[9:24] <ivve> but maybe scrubbing values do..
[9:24] <ivve> usually documentation will tell you if it does require a restart in order to take effect
[9:27] * EinstCrazy (~EinstCraz@58.247.117.134) Quit (Remote host closed the connection)
[9:32] <Be-El> or ceph tell osd.* injectargs will complain if a value cannot be changed on a running osd
[9:36] * ledgr (~ledgr@78.57.252.56) has joined #ceph
[9:36] * EinstCrazy (~EinstCraz@58.247.117.134) has joined #ceph
[9:36] <ivve> yeah
[9:36] <ivve> it works
[9:37] <ivve> however
[9:37] * karnan (~karnan@121.244.87.117) Quit (Ping timeout: 480 seconds)
[9:37] <ivve> after changing deep scrub interval to ~2million
[9:37] <ivve> ceph pg dump all | grep active | awk '{ print $20}' | sort -k1 | uniq -c
[9:37] <ivve> looks like this
[9:37] <ivve> 1789 2016-08-31
[9:37] <ivve> 9 2016-09-01
[9:37] <ivve> 29 2016-09-02
[9:37] <ivve> 42 2016-09-05
[9:37] <ivve> 5162 2016-09-06
[9:37] <ivve> 1293 2016-09-07
[9:38] <ivve> i have been running this for ~24 hours
[9:38] <ivve> */5 * * * * `ceph pg dump | egrep '^[0-9]\.[0-9a-f]{1,4}' | sort -k20 | awk '{ print $1 }' | head -1 | sed 's/^/ceph pg deep-scrub /g'`
[9:39] <ivve> but from what i can understand regular scrubs have something to do with deepscrubs when it comes to inverval
[9:39] <ivve> do i need to increase regular interval scrubs to "minimum of deepscrub interval" ?
[9:41] <ivve> by my calculations all "manual" deepscrubs or automatic for that matter, should be placed around 2016-10-X ?
[9:41] <ivve> but they are just sent 7 days forward
[9:41] <ivve> (scheduled)
[9:44] * dugravot61 (~dugravot6@nat-persul-plg.wifi.univ-lorraine.fr) Quit (Quit: Leaving.)
[9:45] * karnan (~karnan@121.244.87.117) has joined #ceph
[9:46] * Dw_Sn (~Dw_Sn@00020a72.user.oftc.net) has joined #ceph
[9:47] * swami1 (~swami@49.38.3.250) Quit (Ping timeout: 480 seconds)
[9:55] * TMM (~hp@dhcp-077-248-009-229.chello.nl) Quit (Quit: Ex-Chat)
[9:55] * derjohn_mob (~aj@46.189.28.52) has joined #ceph
[9:55] * Nicho1as (~nicho1as@00022427.user.oftc.net) has joined #ceph
[9:55] * T1w (~jens@node3.survey-it.dk) has joined #ceph
[9:56] * flisky1 (~Thunderbi@106.38.61.181) has joined #ceph
[9:56] * flisky1 (~Thunderbi@106.38.61.181) Quit ()
[9:56] <Nicho1as> is it possible for a ceph cluster to have degraded pgs even if it has got no objects? I haven't written any object on my newly created ceph cluster that I made with ceph-deploy
[9:57] <Nicho1as> I just increased the pool size from 2 to 3 (the number of same objects per PG)
[9:57] <Nicho1as> I've tried increasing pg_num and pgp_num in vain.
[9:58] <Nicho1as> pgmap v684: 256 pgs, 1 pools, 0 bytes data, 0 objects; 225 MB used, 5145 GB / 5147 GB avail; 236 active+clean; 14 active+remapped; 6 active+undersized+degraded;
[9:59] <ivve> Nicho1as: yes if your crushmap can't map the pgs
[10:00] <ivve> for example. you have rules that make 3 replicas and the leaf is host, but only have 2 hosts. it wants to map on a thirds host that doesn't exists
[10:00] <ivve> and will say it is degraded
[10:00] <ivve> put this into a pastebin:
[10:00] <ivve> ceph osd pool ls detail
[10:00] <ivve> ceph osd tree
[10:01] * flisky (~Thunderbi@106.38.61.187) Quit (Ping timeout: 480 seconds)
[10:01] * sleinen (~Adium@2001:620:0:2d:a65e:60ff:fedb:f305) has joined #ceph
[10:02] * sleinen1 (~Adium@macsl.switch.ch) has joined #ceph
[10:03] * EinstCrazy (~EinstCraz@58.247.117.134) Quit (Remote host closed the connection)
[10:03] <Nicho1as> ivve: http://hastebin.com/ginahivahu.sm
[10:03] <Nicho1as> Thanks
[10:03] * sleinen2 (~Adium@macsl.switch.ch) has joined #ceph
[10:03] * sleinen1 (~Adium@macsl.switch.ch) Quit (Read error: Connection reset by peer)
[10:04] * Lokta (~Lokta@193.164.231.98) has joined #ceph
[10:04] <ivve> yeah 3 replicas might give you problems there
[10:05] <ivve> if you paste in "ceph osd crush rule dump"
[10:05] * hroussea (~hroussea@000200d7.user.oftc.net) Quit (Remote host closed the connection)
[10:05] * branto (~branto@ip-78-102-208-181.net.upcbroadband.cz) has joined #ceph
[10:05] <ivve> you probably have a chooseleaf on host
[10:05] * hroussea (~hroussea@goyavier-10ge-vl69-4-2.par01.moulticast.net) has joined #ceph
[10:05] <Nicho1as> it does
[10:06] <Nicho1as> the second one in 3 steps
[10:06] <ivve> some pgs, not all will not be able to locate themselves in 3 different hosts
[10:06] <ivve> since they are of different size
[10:06] <Nicho1as> http://hastebin.com/kasodedoqo.sm
[10:07] <ivve> basically, osd.4 is the "issue"
[10:08] <ivve> its huge
[10:08] <ivve> and the host peta-s-227 host is tiny
[10:09] <ivve> reducing to 2 size might not even fix the issue
[10:09] * sleinen (~Adium@2001:620:0:2d:a65e:60ff:fedb:f305) Quit (Ping timeout: 480 seconds)
[10:09] <Nicho1as> Thank you for the description; I appreciate it
[10:10] <ivve> np :)
[10:12] * EinstCrazy (~EinstCraz@58.247.117.134) has joined #ceph
[10:12] <Nicho1as> I might have to say it didn't have any issues when it was first created with pool size 2
[10:14] <ivve> ah there you go
[10:14] * swami1 (~swami@49.44.57.243) has joined #ceph
[10:14] <ivve> Be-El: documentation doesn't say anything about "osd scrub max interval", do you know if it affects deep scrubs as well?
[10:14] <ivve> this might be my issue
[10:14] <ivve> nice my max is lower than my deep scrub interval
[10:14] <Be-El> ivve: sorry, no clue
[10:15] <ivve> "osd_deep_scrub_interval": "2.4192e+06"
[10:15] <ivve> "osd_scrub_min_interval": "86400"
[10:15] <ivve> and lastly
[10:15] <ivve> "osd_scrub_max_interval": "604800"
[10:16] <ivve> default for deepscrub is 604800 as well
[10:16] <ivve> i will test :)
[10:17] <ivve> and let you know
[10:18] <ivve> this surely is an issue in large clusters, as in size large
[10:18] <ivve> datasize
[10:18] <ivve> 0.5PB trying to do a few 1000x deepscrubs in 1 day, even with 1per OSD this means up to 40-100+ at once
[10:19] <ivve> getting like 40k read IOPS
[10:19] * sleinen2 (~Adium@macsl.switch.ch) Quit (Quit: Leaving.)
[10:20] * Tene (~tene@173.13.139.236) Quit (Ping timeout: 480 seconds)
[10:23] * nils_ (~nils_@doomstreet.collins.kg) Quit (Quit: This computer has gone to sleep)
[10:24] * Tene (~tene@173.13.139.236) has joined #ceph
[10:27] <doppelgrau> ivve: scrub-logik is: at a normal scrub look if the deep scrub is overdue, in that case do a deep scrub
[10:29] <doppelgrau> ivve: since the ???last deep scrub??? is for all PGs created at the same time equal (at the beginning), you can either wait very long until the small random faktor for normal scrubs distribute the deep scrubs or use an existing script
[10:29] <doppelgrau> ivve: like https://www.formann.de/2015/05/cronjob-to-enable-timed-deep-scrubbing-in-a-ceph-cluster/
[10:31] * ledgr (~ledgr@78.57.252.56) Quit (Remote host closed the connection)
[10:32] * ledgr (~ledgr@78.57.252.56) has joined #ceph
[10:34] * EinstCra_ (~EinstCraz@58.247.117.134) has joined #ceph
[10:34] * Pulp (~Pulp@195.50.221.63) Quit (Read error: Connection reset by peer)
[10:35] * wkennington (~wkenningt@c-71-204-170-241.hsd1.ca.comcast.net) Quit (Quit: Leaving)
[10:37] * i_m (~ivan.miro@heater.s.infra-it.ru) has joined #ceph
[10:38] * jakekosberg (~Bored@142.4.205.238) has joined #ceph
[10:39] * _mrp (~mrp@93-86-57-229.dynamic.isp.telekom.rs) has joined #ceph
[10:40] * _mrp (~mrp@93-86-57-229.dynamic.isp.telekom.rs) Quit ()
[10:40] * ledgr (~ledgr@78.57.252.56) Quit (Ping timeout: 480 seconds)
[10:41] * EinstCrazy (~EinstCraz@58.247.117.134) Quit (Ping timeout: 480 seconds)
[10:42] * saintpablo (~saintpabl@gw01.mhitp.dk) has joined #ceph
[10:48] * rraja (~rraja@121.244.87.117) Quit (Ping timeout: 480 seconds)
[10:52] <ivve> doppelgrau: yeah i know but i like the stamp to be 4 week into the future instead of 1week
[10:52] <ivve> for the system deepscrub to function
[10:52] <ivve> i want it to work automatically, and with a script "help it even out" when trouble arises
[10:53] <ivve> however, im doing manual deeps right now
[10:53] <ivve> but they are only stamped overdue for 1week aead
[10:53] * DanFoster (~Daniel@2a00:1ee0:3:1337:99a:5e22:ea51:c3a) has joined #ceph
[10:54] * TMM (~hp@185.5.121.201) has joined #ceph
[10:55] * EinstCra_ (~EinstCraz@58.247.117.134) Quit (Remote host closed the connection)
[10:55] * Nicho1as (~nicho1as@00022427.user.oftc.net) Quit (Quit: A man from the Far East; using WeeChat 1.5)
[10:55] <doppelgrau> ivve: which stamp do you mean? ceph pg dump show me only the past scrubs/stamps
[10:56] <ivve> one sec
[10:56] * EinstCrazy (~EinstCraz@58.247.117.134) has joined #ceph
[10:56] <ivve> dumped all in format plain
[10:56] <ivve> 1.1cdd 3913 0 0 0 0 16406306816 6952 6952 active+clean 2016-09-07 10:30:28.939639 69938'178906 69938:172501 [101,51] 101 [101,51] 101 69938'178906 2016-09-07 10:30:28.939163 69938'178906 2016-09-07 10:30:17.054012
[10:57] <ivve> this one is scrubbed today
[10:57] <ivve> few minutes ago
[10:57] <ivve> here is another one scrubb a while ago
[10:58] <ivve> dumped all in format plain
[10:58] <ivve> 1.1e82 3827 0 0 0 0 16046206976 6761 6761 active+clean+scrubbing+deep 2016-09-07 10:42:11.480355 69938'168725 69938:162951 [19,65] 19 [19,65] 1969938'150538 2016-08-31 10:39:31.184296 43678'124157 2016-08-10 10:37:27.954213
[10:58] <ivve> oh but wait
[10:58] <ivve> :)
[10:59] <ivve> how can i easily know when the next scrub is due?
[11:01] * rraja (~rraja@125.16.34.66) has joined #ceph
[11:01] <doppelgrau> ivve: cant really for scrub, sometime between min_scrub_interval and max_scrub_intervall + radom deviation (forgott the parameter name)
[11:03] <doppelgrau> ivve: mext deep scrub about timestamp + osd_deep_scrub_interval (more or less, next scrub that happens after that, but usually scrub happens once a day)
[11:03] <ivve> i see
[11:03] <ivve> so in order to spread them out
[11:03] <ivve> i set a script to force scrubs manually
[11:03] <ivve> once for the period
[11:03] <ivve> then it should be fine?
[11:08] * jakekosberg (~Bored@142.4.205.238) Quit ()
[11:12] * walcubi_ is now known as walcubi
[11:14] <doppelgrau> ivve: yes
[11:15] <ivve> so no real good way to predict when the next scrub is incoming
[11:15] <ivve> *deepscrub
[11:15] <doppelgrau> ivve: and if you want the deep scrubs happen at a specific time (e.g. during the night) you lett a script run at these times and set osd_deep_scrub_interval greater than the time the script needs to deep scrub all
[11:16] * rraja (~rraja@125.16.34.66) Quit (Ping timeout: 480 seconds)
[11:16] <doppelgrau> ivve: next deep scrub happend timestamp + osd_deep_scrub_interval + random(0-(osd_scrub_max_interval-1) more or less :)
[11:19] * yanzheng (~zhyan@125.70.22.186) has joined #ceph
[11:19] * dgurtner (~dgurtner@84-73-130-19.dclient.hispeed.ch) has joined #ceph
[11:22] * sep (~sep@95.62-50-191.enivest.net) Quit (Read error: Connection reset by peer)
[11:23] <ivve> cool
[11:23] <ivve> thanks
[11:23] * _mrp (~mrp@93-86-57-229.dynamic.isp.telekom.rs) has joined #ceph
[11:24] * doppelgrau (~doppelgra@dslb-088-072-094-200.088.072.pools.vodafone-ip.de) Quit (Quit: doppelgrau)
[11:24] * sep (~sep@2a04:2740:1ab:1::2) has joined #ceph
[11:26] * _mrp (~mrp@93-86-57-229.dynamic.isp.telekom.rs) Quit ()
[11:27] * sep (~sep@2a04:2740:1ab:1::2) Quit (Read error: No route to host)
[11:28] * ledgr (~ledgr@78.57.252.56) has joined #ceph
[11:31] * fdmanana (~fdmanana@2001:8a0:6e0c:6601:4d30:1cb6:3a49:7398) has joined #ceph
[11:31] * hgichon (~hgichon@112.220.91.130) Quit (Ping timeout: 480 seconds)
[11:33] * _mrp (~mrp@93-86-57-229.dynamic.isp.telekom.rs) has joined #ceph
[11:36] * _mrp (~mrp@93-86-57-229.dynamic.isp.telekom.rs) Quit ()
[11:37] * EinstCrazy (~EinstCraz@58.247.117.134) Quit (Remote host closed the connection)
[11:38] * _mrp (~mrp@93-86-57-229.dynamic.isp.telekom.rs) has joined #ceph
[11:39] * EinstCrazy (~EinstCraz@58.247.119.250) has joined #ceph
[11:40] * _mrp (~mrp@93-86-57-229.dynamic.isp.telekom.rs) Quit ()
[11:41] * BennyRene (~BennyRene@host-89-241-119-65.as13285.net) has joined #ceph
[11:42] * BennyRene (~BennyRene@host-89-241-119-65.as13285.net) has left #ceph
[11:44] * SurfMaths (~theghost9@tor.offend.me.uk) has joined #ceph
[11:46] * sleinen (~Adium@macsl.switch.ch) has joined #ceph
[11:50] * wjw-freebsd (~wjw@smtp.digiware.nl) has joined #ceph
[11:52] * ledgr (~ledgr@78.57.252.56) Quit (Remote host closed the connection)
[11:52] * ledgr (~ledgr@88-119-196-104.static.zebra.lt) has joined #ceph
[11:57] * sep (~sep@2a04:2740:1ab:1::2) has joined #ceph
[12:02] * hoonetorg (~hoonetorg@77.119.226.254.static.drei.at) Quit (Ping timeout: 480 seconds)
[12:03] * doppelgrau (~doppelgra@132.252.235.172) has joined #ceph
[12:03] * kuku (~kuku@119.93.91.136) Quit (Remote host closed the connection)
[12:05] * qman__ (~rohroh@2600:3c00::f03c:91ff:fe69:92af) Quit (Remote host closed the connection)
[12:07] * karnan (~karnan@121.244.87.117) Quit (Ping timeout: 480 seconds)
[12:10] * kefu is now known as kefu|afk
[12:10] * huangjun|2 (~kvirc@113.57.168.154) Quit (Quit: KVIrc 4.2.0 Equilibrium http://www.kvirc.net/)
[12:12] <sep> going to have a meeting with PHB's tomorrow. deciding the future of storage here. have been running a ceph lab for a few months but i can hardly show them ssh and ceph osd tree without them running screaming at the closest expencive vendor that guarantee that they never have to learn anything. :) </rant> ;; are there any free and open slideshows that presents ceph and it's features strengths and weaknesses that is allowed to use ? of the more recent kind
[12:12] <sep> ? have found some for older ceph, that i can use parts of . but it would be nice to have something more recent
[12:14] * SurfMaths (~theghost9@tor.offend.me.uk) Quit ()
[12:14] * vimal is now known as vimal|brb
[12:15] <IcePic> there was a fun youtube that explained it, which just a slight bit of humor (catpics) in it
[12:17] <IcePic> https://www.youtube.com/watch?v=QBkH1g4DuKE
[12:17] <IcePic> that one
[12:17] <T1w> sep: I'd say you are allowed to use whatever you can find that suits your purpose for internal presentations
[12:18] <T1w> also other peoples published presentations/figures/slides
[12:18] * i_m (~ivan.miro@heater.s.infra-it.ru) Quit (Ping timeout: 480 seconds)
[12:18] <T1w> that you can find anywhere online
[12:19] * rdas (~rdas@125.212.212.209) has joined #ceph
[12:19] * karnan (~karnan@121.244.87.117) has joined #ceph
[12:20] <sep> then there are lots to choose from. thanks
[12:21] * kefu|afk (~kefu@114.92.125.128) Quit (Quit: My Mac has gone to sleep. ZZZzzz???)
[12:22] <sep> i fnid many that talk about ceph advantages. do you know of any that address ceph weaknesses (without it beeing a storage spaces or netapp slide) iow objectivly :)
[12:24] <IcePic> in reality, lots of enterprises think they can buy blame shifting.
[12:24] * DV_ (~veillard@2001:41d0:a:f29f::1) has joined #ceph
[12:24] <IcePic> All large (well..) network FSs ive managed have had issues, have gotten into corner cases and acted up at some point. In more or less all of those situations, the local guy gets to do a lot of work
[12:25] <IcePic> some of them had decent support contracts, but they just send some consultant that knows as much as you do and that attended the same education you might have been to and does the best he can
[12:26] * qman (~rohroh@2600:3c00::f03c:91ff:fe69:92af) has joined #ceph
[12:26] <IcePic> so in my experience, trying to buy a contract that allows the local storage people to keep sleeping while the fs cluster fails just wont work, even if PHBs might think you can
[12:28] * DV (~veillard@2001:41d0:a:f29f::1) Quit (Ping timeout: 480 seconds)
[12:30] * i_m (~ivan.miro@2001:470:1f0b:1a08:5ef:dd4b:132e:d39d) has joined #ceph
[12:31] * Kurt (~Adium@2001:628:1:5:c1e1:5b6e:f4a4:e94c) Quit (Quit: Leaving.)
[12:32] * kefu (~kefu@114.92.125.128) has joined #ceph
[12:33] <IcePic> so all else being equal, I prefer not to pay for contracts which dont really give tech guys what they would like, and if I am to do most or all of the heavy lifting anyhow, we might aswell not pay someone else for it.
[12:34] <IcePic> not that there aren't other good commercial products, but ceph does a lot of the things right in terms of redundancy and scaling.
[12:39] * i_m (~ivan.miro@2001:470:1f0b:1a08:5ef:dd4b:132e:d39d) Quit (Ping timeout: 480 seconds)
[12:40] * wjw-freebsd (~wjw@smtp.digiware.nl) Quit (Ping timeout: 480 seconds)
[12:41] * zdyxry (~oftc-webi@18.41.110.36.static.bjtelecom.net) has joined #ceph
[12:41] * zdyxry (~oftc-webi@18.41.110.36.static.bjtelecom.net) has left #ceph
[12:42] * dgurtner (~dgurtner@84-73-130-19.dclient.hispeed.ch) Quit (Ping timeout: 480 seconds)
[12:43] * appleq (~appleq@86.57.157.66) has joined #ceph
[12:44] * appleq (~appleq@86.57.157.66) Quit (Quit: Leaving)
[12:45] <doppelgrau> sep: what about a nice setup of cephdash for visual, some applikation running on ceph and then pull out a powercord to show how resillient it is
[12:46] <doppelgrau> "it even continues to work when I doo THIS"
[12:47] <sep> have it planned basically.
[12:48] * ledgr_ (~ledgr@78.57.252.56) has joined #ceph
[12:55] * ledgr (~ledgr@88-119-196-104.static.zebra.lt) Quit (Ping timeout: 480 seconds)
[12:55] * yuelongguang_ (~chatzilla@114.134.84.144) has joined #ceph
[12:57] * rakeshgm (~rakesh@121.244.87.117) Quit (Quit: Leaving)
[12:57] * srk (~Siva@cpe-70-113-23-93.austin.res.rr.com) has joined #ceph
[12:58] * lifeboy (~roland@196.45.25.14) Quit (Quit: Ex-Chat)
[13:00] * _mrp (~mrp@82.117.199.26) has joined #ceph
[13:01] * yuelongguang (~chatzilla@114.134.84.144) Quit (Ping timeout: 480 seconds)
[13:01] * yuelongguang_ is now known as yuelongguang
[13:02] * rdas (~rdas@125.212.212.209) Quit (Quit: Leaving)
[13:03] * ira (~ira@c-24-34-255-34.hsd1.ma.comcast.net) has joined #ceph
[13:07] * coreping (~Michael_G@n1.coreping.org) Quit (Quit: WeeChat 0.4.2)
[13:08] * nils_ (~nils_@doomstreet.collins.kg) has joined #ceph
[13:11] * neurodrone_ (~neurodron@pool-100-35-225-168.nwrknj.fios.verizon.net) has joined #ceph
[13:11] * neurodrone_ (~neurodron@pool-100-35-225-168.nwrknj.fios.verizon.net) Quit ()
[13:12] * hoopy (~pico@185.80.50.33) has joined #ceph
[13:13] * vicente (~~vicente@125-227-238-55.HINET-IP.hinet.net) Quit (Quit: Leaving)
[13:14] * ledgr_ (~ledgr@78.57.252.56) Quit (Remote host closed the connection)
[13:14] * srk (~Siva@cpe-70-113-23-93.austin.res.rr.com) Quit (Read error: Connection reset by peer)
[13:14] * coreping (~Michael_G@n1.coreping.org) has joined #ceph
[13:15] * rraja (~rraja@121.244.87.117) has joined #ceph
[13:17] * srk (~Siva@cpe-70-113-23-93.austin.res.rr.com) has joined #ceph
[13:17] * DanFoster (~Daniel@2a00:1ee0:3:1337:99a:5e22:ea51:c3a) Quit (Quit: Leaving)
[13:20] * karnan (~karnan@121.244.87.117) Quit (Ping timeout: 480 seconds)
[13:21] * nils_ (~nils_@doomstreet.collins.kg) Quit (Ping timeout: 480 seconds)
[13:26] * srk (~Siva@cpe-70-113-23-93.austin.res.rr.com) Quit (Ping timeout: 480 seconds)
[13:29] * karnan (~karnan@121.244.87.117) has joined #ceph
[13:29] * i_m (~ivan.miro@109.188.124.6) has joined #ceph
[13:31] <jcsp> is Christian Eichelmann on this channel by any chance? I'm getting bounces trying to send an email to him.
[13:34] * vimal|brb is now known as vimal
[13:34] * swami2 (~swami@49.38.1.152) has joined #ceph
[13:36] * evelu (~erwan@2a01:e34:eecb:7400:4eeb:42ff:fedc:8ac) has joined #ceph
[13:38] * i_m1 (~ivan.miro@94.25.168.142) has joined #ceph
[13:38] * spgriffi_ (~spgriffin@66-46-246-206.dedicated.allstream.net) Quit (Ping timeout: 480 seconds)
[13:38] * swami1 (~swami@49.44.57.243) Quit (Ping timeout: 480 seconds)
[13:39] * i_m (~ivan.miro@109.188.124.6) Quit (Ping timeout: 480 seconds)
[13:39] * sleinen (~Adium@macsl.switch.ch) Quit (Quit: Leaving.)
[13:42] * hoopy (~pico@185.80.50.33) Quit ()
[13:42] * sankarshan (~sankarsha@121.244.87.117) has joined #ceph
[13:49] * spgriffinjr (~spgriffin@66.46.246.206) has joined #ceph
[13:50] * rakeshgm (~rakesh@125.16.34.66) has joined #ceph
[13:52] * ledgr (~ledgr@78.57.252.56) has joined #ceph
[13:56] * bniver (~bniver@pool-98-110-180-234.bstnma.fios.verizon.net) has joined #ceph
[14:02] * rotbeard (~redbeard@185.32.80.238) Quit (Quit: Leaving)
[14:02] * kefu (~kefu@114.92.125.128) Quit (Quit: My Mac has gone to sleep. ZZZzzz???)
[14:03] * ashah (~ashah@121.244.87.117) Quit (Quit: Leaving)
[14:03] * karnan (~karnan@121.244.87.117) Quit (Remote host closed the connection)
[14:04] * bara (~bara@nat-pool-brq-t.redhat.com) Quit (Quit: Bye guys! (??????????????????? ?????????)
[14:06] * bara (~bara@213.175.37.12) has joined #ceph
[14:09] * matx (~Mraedis@108.61.123.87) has joined #ceph
[14:09] * dgurtner (~dgurtner@84-73-130-19.dclient.hispeed.ch) has joined #ceph
[14:11] * Racpatel (~Racpatel@c-73-194-155-223.hsd1.nj.comcast.net) has joined #ceph
[14:21] * EinstCrazy (~EinstCraz@58.247.119.250) Quit (Remote host closed the connection)
[14:23] * hoonetorg (~hoonetorg@77.119.226.254.static.drei.at) has joined #ceph
[14:24] * Arvydas (~oftc-webi@185.84.63.241) has joined #ceph
[14:26] * srk (~Siva@cpe-70-113-23-93.austin.res.rr.com) has joined #ceph
[14:26] * ircolle (~Adium@2601:285:201:633a:7d2c:a008:914c:acde) has joined #ceph
[14:28] * zhen (~Thunderbi@43.255.178.227) has joined #ceph
[14:32] * Racpatel (~Racpatel@c-73-194-155-223.hsd1.nj.comcast.net) Quit (Quit: Leaving)
[14:32] * Racpatel (~Racpatel@2601:87:3:31e3:4e34:88ff:fe87:9abf) has joined #ceph
[14:34] * T1w (~jens@node3.survey-it.dk) Quit (Ping timeout: 480 seconds)
[14:36] * danieagle (~Daniel@200-171-230-208.customer.telesp.net.br) has joined #ceph
[14:39] * neurodrone_ (~neurodron@pool-100-35-225-168.nwrknj.fios.verizon.net) has joined #ceph
[14:39] * matx (~Mraedis@108.61.123.87) Quit ()
[14:41] * rakeshgm (~rakesh@125.16.34.66) Quit (Remote host closed the connection)
[14:45] * o0c (~o0c@chris.any.mx) has joined #ceph
[14:45] * srk (~Siva@cpe-70-113-23-93.austin.res.rr.com) Quit (Ping timeout: 480 seconds)
[14:48] * yuelongguang (~chatzilla@114.134.84.144) Quit (Ping timeout: 480 seconds)
[14:48] * MrBy_ (~MrBy@85.115.23.38) has joined #ceph
[14:48] * MrBy (~MrBy@85.115.23.42) Quit (Read error: Connection reset by peer)
[14:49] * MrBy__ (~MrBy@85.115.23.42) has joined #ceph
[14:51] * srk (~Siva@cpe-70-113-23-93.austin.res.rr.com) has joined #ceph
[14:53] * vimal (~vikumar@121.244.87.116) Quit (Quit: Leaving)
[14:54] * Nicho1as (~nicho1as@00022427.user.oftc.net) has joined #ceph
[14:54] * ledgr (~ledgr@78.57.252.56) Quit (Remote host closed the connection)
[14:55] * ledgr (~ledgr@88-119-196-104.static.zebra.lt) has joined #ceph
[14:55] * swami2 (~swami@49.38.1.152) Quit (Quit: Leaving.)
[14:56] * MrBy_ (~MrBy@85.115.23.38) Quit (Ping timeout: 480 seconds)
[15:01] * lmb_ (~Lars@2a02:8109:8100:1d2c:2ad2:44ff:fedf:3318) Quit (Ping timeout: 480 seconds)
[15:02] * mattbenjamin (~mbenjamin@12.118.3.106) has joined #ceph
[15:10] * ircolle (~Adium@2601:285:201:633a:7d2c:a008:914c:acde) Quit (Quit: Leaving.)
[15:12] * lmb_ (~Lars@ip5b404bab.dynamic.kabel-deutschland.de) has joined #ceph
[15:12] * Racpatel (~Racpatel@2601:87:3:31e3:4e34:88ff:fe87:9abf) Quit (Ping timeout: 480 seconds)
[15:14] * Discovery (~Discovery@109.235.52.3) has joined #ceph
[15:21] * Racpatel (~Racpatel@2601:87:3:31e3:4e34:88ff:fe87:9abf) has joined #ceph
[15:22] * kuku (~kuku@112.203.52.190) has joined #ceph
[15:23] * delaf (~delaf@legendary.xserve.fr) has joined #ceph
[15:23] * delaf (~delaf@legendary.xserve.fr) has left #ceph
[15:28] * mhack (~mhack@nat-pool-bos-t.redhat.com) has joined #ceph
[15:30] * bara (~bara@213.175.37.12) Quit (Ping timeout: 480 seconds)
[15:40] * bara (~bara@nat-pool-brq-t.redhat.com) has joined #ceph
[15:41] * johnavp1989 (~jpetrini@8.39.115.8) has joined #ceph
[15:41] <- *johnavp1989* To prove that you are human, please enter the result of 8+3
[15:43] * walcubi (~walcubi@87.149.186.180) Quit (Remote host closed the connection)
[15:45] * srk (~Siva@cpe-70-113-23-93.austin.res.rr.com) Quit (Read error: Connection reset by peer)
[15:45] * sankarshan (~sankarsha@121.244.87.117) Quit (Ping timeout: 480 seconds)
[15:47] * Sigma (~cheese^@199.254.238.44) has joined #ceph
[15:49] * saintpablo (~saintpabl@gw01.mhitp.dk) Quit (Ping timeout: 480 seconds)
[15:54] * CephFan1 (~textual@96-31-67-206.static.hvvc.us) has joined #ceph
[15:56] * wes_dillingham (~wes_dilli@140.247.242.44) has joined #ceph
[16:02] * CephFan1 (~textual@96-31-67-206.static.hvvc.us) Quit (Ping timeout: 480 seconds)
[16:02] * ircolle (~Adium@2601:285:201:633a:7d2c:a008:914c:acde) has joined #ceph
[16:03] * rwheeler (~rwheeler@pool-173-48-195-215.bstnma.fios.verizon.net) Quit (Quit: Leaving)
[16:05] * bene2 (~bene@nat-pool-bos-t.redhat.com) has joined #ceph
[16:06] * Arvydas (~oftc-webi@185.84.63.241) Quit (Ping timeout: 480 seconds)
[16:06] * oliveiradan2 (~doliveira@67.214.238.80) Quit (Ping timeout: 480 seconds)
[16:11] * Lokta (~Lokta@193.164.231.98) Quit (Ping timeout: 480 seconds)
[16:12] <jprins_> Hi everyone. I'm testing Ceph, 10.2.2, so far everything fine. But now I installed a NFS gateway to export some filesystem to a Linux client on a closed network. The NFS gateway mounts the CephFS and exports the mounted tree using NFS.
[16:12] * valeech (~valeech@pool-96-247-203-33.clppva.fios.verizon.net) Quit (Quit: valeech)
[16:13] <jprins_> It looks like this construction is really very very slow. Has anyone here done this before?
[16:14] * Hemanth (~hkumar_@121.244.87.117) Quit (Ping timeout: 480 seconds)
[16:17] * Sigma (~cheese^@199.254.238.44) Quit ()
[16:18] * xarses (~xarses@c-73-202-191-48.hsd1.ca.comcast.net) Quit (Ping timeout: 480 seconds)
[16:25] * xinli (~charleyst@32.97.110.50) has joined #ceph
[16:29] * Discovery (~Discovery@109.235.52.3) Quit (Read error: Connection reset by peer)
[16:30] * EinstCrazy (~EinstCraz@58.39.76.182) has joined #ceph
[16:34] <SamYaple> jprins_: NFS is going to do a whole lotta fsyncs which is in general very slow. Without a specific problem mentioned, the behaviour you describe sounds very normal
[16:35] * danielsj (~Kyso_@exit0.radia.tor-relays.net) has joined #ceph
[16:38] * lifeboy (~roland@196.45.25.14) has joined #ceph
[16:40] * joshd1 (~jdurgin@2602:30a:c089:2b0:b137:188e:d216:76bc) has joined #ceph
[16:41] * xarses (~xarses@64.124.158.3) has joined #ceph
[16:41] * vikhyat (~vumrao@121.244.87.116) Quit (Quit: Leaving)
[16:42] <Be-El> jprins_: you may want to try nfs-ganesha, which has a native libcephfs plugin
[16:43] <Be-El> in any case nfs and cephfs do not mix very well due to the different coherency/data integrity moels
[16:44] * kefu (~kefu@114.92.125.128) has joined #ceph
[16:46] * hoonetorg (~hoonetorg@77.119.226.254.static.drei.at) Quit (Ping timeout: 480 seconds)
[16:50] * pvh_sa (~pvh@169-0-184-226.ip.afrihost.co.za) Quit (Ping timeout: 480 seconds)
[16:53] * ivve (~zed@cust-gw-11.se.zetup.net) Quit (Ping timeout: 480 seconds)
[16:56] * hoonetorg (~hoonetorg@77.119.226.254.static.drei.at) has joined #ceph
[16:56] * xinli (~charleyst@32.97.110.50) Quit (Remote host closed the connection)
[16:56] * xinli (~charleyst@32.97.110.50) has joined #ceph
[16:59] * Dw_Sn_ (~Dw_Sn@forbin.ichec.ie) has joined #ceph
[17:01] * EinstCrazy (~EinstCraz@58.39.76.182) Quit (Remote host closed the connection)
[17:02] * yanzheng (~zhyan@125.70.22.186) Quit (Quit: This computer has gone to sleep)
[17:04] * rwheeler (~rwheeler@nat-pool-bos-t.redhat.com) has joined #ceph
[17:05] * danielsj (~Kyso_@exit0.radia.tor-relays.net) Quit ()
[17:07] * Dw_Sn (~Dw_Sn@00020a72.user.oftc.net) Quit (Ping timeout: 480 seconds)
[17:09] <IcePic> you need to start lying at some level, and claim "sure thing boss, the data is safe, move along" even though it isnt.
[17:12] * sebastian-w (~quassel@212.218.8.138) Quit (Remote host closed the connection)
[17:12] * sebastian-w (~quassel@212.218.8.139) has joined #ceph
[17:16] * ledgr (~ledgr@88-119-196-104.static.zebra.lt) Quit (Remote host closed the connection)
[17:16] * wushudoin (~wushudoin@38.140.108.2) has joined #ceph
[17:18] * blizzow (~jburns@50-243-148-102-static.hfc.comcastbusiness.net) has joined #ceph
[17:19] * bara (~bara@nat-pool-brq-t.redhat.com) Quit (Quit: Bye guys! (??????????????????? ?????????)
[17:19] * kuku (~kuku@112.203.52.190) Quit (Read error: Connection reset by peer)
[17:20] * kuku (~kuku@112.203.52.190) has joined #ceph
[17:21] * mykola (~Mikolaj@91.245.78.210) has joined #ceph
[17:27] * valeech (~valeech@50-205-143-162-static.hfc.comcastbusiness.net) has joined #ceph
[17:27] * EinstCrazy (~EinstCraz@58.39.76.182) has joined #ceph
[17:27] * TMM (~hp@185.5.121.201) Quit (Quit: Ex-Chat)
[17:32] * vbellur (~vijay@2601:18f:700:55b0:5e51:4fff:fee8:6a5c) Quit (Ping timeout: 480 seconds)
[17:32] * rraja (~rraja@121.244.87.117) Quit (Ping timeout: 480 seconds)
[17:32] * dclayton (~dclayton@64.202.160.88) has joined #ceph
[17:32] <dclayton> hi all, do you know if ceph support gzip compression on asset downloads?
[17:38] * F|1nt (~F|1nt@host37-211.lan-isdn.imaginet.fr) has joined #ceph
[17:39] * kuku (~kuku@112.203.52.190) Quit (Remote host closed the connection)
[17:40] * haplo37 (~haplo37@107-190-37-90.cpe.teksavvy.com) has joined #ceph
[17:43] * sudocat (~dibarra@192.185.1.19) has joined #ceph
[17:44] * rraja (~rraja@125.16.34.66) has joined #ceph
[17:47] * fritchie (~oftc-webi@199.168.151.102) has joined #ceph
[17:47] <fritchie> I am having trouble with 'radosgw-admin key rm --uid=blah:swift', I get he error: could not remove key: unable to parse request, user info was not populated
[17:48] <fritchie> the subuser has been removed, the key is still there though
[17:48] * karnan (~karnan@106.51.136.211) has joined #ceph
[17:50] * kefu (~kefu@114.92.125.128) Quit (Quit: My Mac has gone to sleep. ZZZzzz???)
[17:51] * kefu (~kefu@114.92.125.128) has joined #ceph
[17:51] * sudocat (~dibarra@192.185.1.19) Quit (Ping timeout: 480 seconds)
[17:56] * fritchie (~oftc-webi@199.168.151.102) Quit (Quit: Page closed)
[17:58] * valeech (~valeech@50-205-143-162-static.hfc.comcastbusiness.net) Quit (Quit: valeech)
[17:59] * zhen (~Thunderbi@43.255.178.227) Quit (Remote host closed the connection)
[17:59] * ChanServ sets mode +o ircolle
[17:59] * ChanServ sets mode +v joao
[17:59] * ChanServ sets mode +v leseb
[17:59] * msudnick (~michael@24-212-218-207.cable.teksavvy.com) has joined #ceph
[17:59] * Racpatel (~Racpatel@2601:87:3:31e3:4e34:88ff:fe87:9abf) Quit (Quit: Leaving)
[17:59] * Racpatel (~Racpatel@2601:87:3:31e3::2433) has joined #ceph
[17:59] * kefu is now known as kefu|afk
[18:00] * vbellur (~vijay@nat-pool-bos-u.redhat.com) has joined #ceph
[18:03] * BrianA (~BrianA@4.53.142.178) has joined #ceph
[18:03] * pvh_sa (~pvh@169-0-184-226.ip.afrihost.co.za) has joined #ceph
[18:05] * EinstCrazy (~EinstCraz@58.39.76.182) Quit (Remote host closed the connection)
[18:07] * EinstCrazy (~EinstCraz@58.39.76.182) has joined #ceph
[18:08] * Dw_Sn_ (~Dw_Sn@forbin.ichec.ie) Quit (Quit: leaving)
[18:08] * BrianA (~BrianA@4.53.142.178) Quit (Read error: Connection reset by peer)
[18:08] * F|1nt (~F|1nt@host37-211.lan-isdn.imaginet.fr) Quit (Quit: Be back later ...)
[18:13] * kristen (~kristen@134.134.139.77) has joined #ceph
[18:16] * EinstCrazy (~EinstCraz@58.39.76.182) Quit (Remote host closed the connection)
[18:18] * kristen (~kristen@134.134.139.77) Quit (Remote host closed the connection)
[18:21] * davidzlap (~Adium@2605:e000:1313:8003:992d:901f:dd23:80b) has joined #ceph
[18:24] * moegyver (~moe@212.85.78.250) Quit (Ping timeout: 480 seconds)
[18:24] * rraja (~rraja@125.16.34.66) Quit (Ping timeout: 480 seconds)
[18:29] * ivveh (~ivveh@c83-254-11-200.bredband.comhem.se) has joined #ceph
[18:30] * dclayton (~dclayton@64.202.160.88) Quit ()
[18:32] * ircolle (~Adium@2601:285:201:633a:7d2c:a008:914c:acde) Quit (Quit: Leaving.)
[18:35] * kristen (~kristen@134.134.139.83) has joined #ceph
[18:36] * vasu (~vasu@c-73-231-60-138.hsd1.ca.comcast.net) has joined #ceph
[18:38] * Be-El (~blinke@nat-router.computational.bio.uni-giessen.de) Quit (Quit: Leaving.)
[18:41] * doppelgrau (~doppelgra@132.252.235.172) Quit (Quit: Leaving.)
[18:42] * rakeshgm (~rakesh@106.51.26.134) has joined #ceph
[18:50] * msudnick (~michael@24-212-218-207.cable.teksavvy.com) Quit (Remote host closed the connection)
[18:52] * msudnick (~michael@24-212-218-207.cable.teksavvy.com) has joined #ceph
[18:53] * KindOne (kindone@0001a7db.user.oftc.net) Quit (Ping timeout: 480 seconds)
[18:55] * Miouge (~Miouge@78-62-227-225.static.zebra.lt) has joined #ceph
[18:55] * oliveiradan (~doliveira@137.65.133.10) has joined #ceph
[18:55] * branto (~branto@ip-78-102-208-181.net.upcbroadband.cz) Quit (Quit: Leaving.)
[18:57] * Nicho1as (~nicho1as@00022427.user.oftc.net) Quit (Quit: A man from the Far East; using WeeChat 1.5)
[18:58] * vbellur (~vijay@nat-pool-bos-u.redhat.com) Quit (Ping timeout: 480 seconds)
[19:00] * rakeshgm (~rakesh@106.51.26.134) Quit (Quit: Leaving)
[19:01] * ivveh (~ivveh@c83-254-11-200.bredband.comhem.se) Quit ()
[19:04] * derjohn_mob (~aj@46.189.28.52) Quit (Remote host closed the connection)
[19:04] * wenduo (~wenduo@218.30.116.7) Quit (Ping timeout: 480 seconds)
[19:09] * srk (~Siva@32.97.110.53) has joined #ceph
[19:10] * rraja (~rraja@121.244.87.117) has joined #ceph
[19:10] * vbellur (~vijay@nat-pool-bos-t.redhat.com) has joined #ceph
[19:14] <ron-slc> We have been using CephFS, since pre-Argonaut. Our cluster has followed the typical upgrade path, we have only skipped from Hammer -> Jewel.
[19:14] <ron-slc> On Jewel we are getting many seg-faults in the ceph-fuse module.
[19:15] <ron-slc> Is it recommended to run any sort of FSCK or cephfs-data-scan operation?
[19:15] <ron-slc> We also did the Touch procedure , as recommended in a previous version upgrade
[19:16] * joshd1 (~jdurgin@2602:30a:c089:2b0:b137:188e:d216:76bc) Quit (Quit: Leaving.)
[19:16] <ron-slc> And upon Jewel upgrade the: 'cephfs-data-scan tmap_upgrade'
[19:22] <btaylor> should ceph-osd handle mounting the osd???s after a reboot, or should i be adding the dev???s to /etc/fstab?
[19:24] * sleinen (~Adium@2001:620:0:69::102) has joined #ceph
[19:24] * doppelgrau (~doppelgra@dslb-088-072-094-200.088.072.pools.vodafone-ip.de) has joined #ceph
[19:25] <ron-slc> btaylor - I believe since Firefly; ceph-osd process attempts to mount volumes. You can simulate w/o rebooting, by: 1) stop OSD 2) manually unmount the osd's volume, 3) start the OSD
[19:25] <btaylor> k, just wondering because ceph-osd always fails when i reboot
[19:25] <btaylor> but works if i add them to the fstab and reboot
[19:27] <ron-slc> In our cluster (not using ceph-deploy) : In the ceph.conf, we specify the devs = /dev/disk/...... and osd journal = /dev/disk/...... (using the partitions wwn-xxxxx)
[19:27] <ron-slc> This is the link from [osd.##] to the auto-mounting of DEV and finding the Journal device/partition.
[19:29] <ron-slc> Example:
[19:29] <ron-slc> [osd.10]
[19:29] <ron-slc> host = osdsrv01
[19:29] <ron-slc> devs = /dev/disk/by-id/wwn-0x6001e68482c3000018722f9613526fe8-part2
[19:29] <ron-slc> osd journal = /dev/disk/by-id/wwn-0x5e48a97100012207-part5
[19:30] <btaylor> hmm interesting
[19:31] <btaylor> i did use ceph deploy. so i guess that step doesn???t happen.
[19:33] <ron-slc> btaylor - Possibly. Ceph deploy, I believe; attempts to use Sym links in the /var/lib/ceph/osd/ceph-### dirs, to the OSD, and Journal
[19:33] <ron-slc> There are multiple ways to skin that cat.
[19:33] <ron-slc> I'm a control freak, and use the ceph.conf method.
[19:34] <ron-slc> If you are in a hurry to get services up, you can likely manually mount the OSD. And then start the ceph-osd
[19:34] <ron-slc> When ceph-osd sees the volume is already mounted, it will use it.
[19:34] <btaylor> i saw there was a symlink to teh journal but the osd-# seems mounted.
[19:34] <rkeene> I have a daemon that generates ceph.conf and another that adds OSDs to it
[19:35] * BrianA (~BrianA@fw-rw.shutterfly.com) has joined #ceph
[19:35] <ron-slc> rkeene: using the devs = method?
[19:35] <rkeene> (If you plug in a new disk, the daemon will put a filesystem on it and add it to ceph, if you add a disk that already is from this Ceph cluster it'll start using it)
[19:35] * Miouge (~Miouge@78-62-227-225.static.zebra.lt) Quit (Quit: Miouge)
[19:35] <rkeene> ron-slc, No, there's no reason to give ceph-osd the device since it talks to the filesystem, not the device
[19:36] <ron-slc> Interesting. I'll have to test with that!
[19:36] <rkeene> (Unless you're playing with bluestore, which is experimental -- and I haven't experimented with it)
[19:36] <ron-slc> yea, I haven't had time for bluestore, yet
[19:38] * luckz (~Enikma@tor-exit1.sjc02.svwh.net) has joined #ceph
[19:39] * if_gaga0 (~if_gaga0@91.207.132.67) has joined #ceph
[19:40] * Miouge (~Miouge@78-62-227-225.static.zebra.lt) has joined #ceph
[19:42] * Miouge (~Miouge@78-62-227-225.static.zebra.lt) Quit ()
[19:42] * if_gaga0 is now known as if_gagarin0
[19:43] * TomasCZ (~TomasCZ@yes.tenlab.net) has joined #ceph
[19:44] * Miouge (~Miouge@78-62-227-225.static.zebra.lt) has joined #ceph
[19:45] <if_gagarin0> hello all! can somebody help me with troobleshoting, why radosgw don't listen fcgi socket
[19:45] * georgem (~Adium@206.108.127.16) has joined #ceph
[19:46] <if_gagarin0> i spent about 3 hours for debugging this problem http://paste.openstack.org/show/567583/
[19:46] <if_gagarin0> but have no idea, why i can't see listen/bind system calls in strace output for radosgw...
[19:48] <if_gagarin0> after runs radosgw, they don't listen :9000 tcp port for apache proxy requests
[19:48] <if_gagarin0> sorry about my bad english
[19:49] * Miouge (~Miouge@78-62-227-225.static.zebra.lt) Quit (Quit: Miouge)
[19:50] * sleinen (~Adium@2001:620:0:69::102) Quit (Ping timeout: 480 seconds)
[19:54] * whatevsz (~quassel@b9168e43.cgn.dg-w.de) Quit (Remote host closed the connection)
[19:57] * Miouge (~Miouge@78-62-227-225.static.zebra.lt) has joined #ceph
[19:57] * msudnick (~michael@24-212-218-207.cable.teksavvy.com) has left #ceph
[20:00] * Miouge (~Miouge@78-62-227-225.static.zebra.lt) Quit ()
[20:01] * shaunm (~shaunm@ms-208-102-105-216.gsm.cbwireless.com) has joined #ceph
[20:01] * srk (~Siva@32.97.110.53) Quit (Ping timeout: 480 seconds)
[20:04] * elfurbe (~Adam@saint.uits.arizona.edu) has joined #ceph
[20:07] * Heebie (~thebert@dub-bdtn-office-r1.net.digiweb.ie) Quit (Read error: Connection reset by peer)
[20:08] * luckz (~Enikma@tor-exit1.sjc02.svwh.net) Quit ()
[20:08] * Heebie (~thebert@dub-bdtn-office-r1.net.digiweb.ie) has joined #ceph
[20:12] * bene2 (~bene@nat-pool-bos-t.redhat.com) Quit (Quit: Konversation terminated!)
[20:21] * sudocat (~dibarra@192.185.1.20) has joined #ceph
[20:22] * ChrisHolcombe (~chris@c-73-180-29-35.hsd1.or.comcast.net) has joined #ceph
[20:25] <elfurbe> So, I upgraded my ceph cluster from infernalis to jewel, but now none of my OSDs will start and the error is "** ERROR: unable to open OSD superblock on /var/lib/ceph/osd/ceph-0: (2) No such file or directory". The filesystems are mounted, looks fine to me, not sure what's going on there. Anyone have some protips?
[20:27] <ron-slc> Verify that your /dev/sd### is owned by ceph
[20:27] * cholcombe (~chris@2001:67c:1562:8007::aac:40f1) Quit (Ping timeout: 480 seconds)
[20:28] * xinli (~charleyst@32.97.110.50) Quit ()
[20:28] <elfurbe> They are
[20:28] <ron-slc> Verify that all files under the mounted OSD, in /var/lib/ceph/osd/ceph-###/ are owned by ceph:ceph
[20:29] <ron-slc> (This should have been done for the Infernalis upgrade, I believe.)
[20:29] <elfurbe> That's all owned by root
[20:29] <elfurbe> So I should just chown all of that to ceph:ceph?
[20:29] <ron-slc> With Jewel, your ceph-osd should now be running as Ceph user. So that entire OSD mount needs to be owned by ceph:ceph
[20:30] <elfurbe> Makes sense
[20:30] <ron-slc> Yes. On my production cluster, this took a VERY long time per OSD Millions of files per OSD. You'll likely want to get several SCREENS/TMUX running in parallel, on the hosts's OSD's
[20:31] <elfurbe> Ha, rgr
[20:32] <ron-slc> elfurbe: happy travels
[20:32] <elfurbe> Hey, thanks for the help!
[20:32] <ron-slc> np!
[20:33] <elfurbe> Oh, also, were there changes in the osd automounting as well?
[20:33] <elfurbe> None of the OSD fses were mounted at boot
[20:33] <elfurbe> I thought they used to
[20:34] <ron-slc> I can only speak for a non "ceph-deploy" cluster. My cluster uses the ceph.conf, to define the devs= and osd journal= params.
[20:34] <btaylor> tried adding in [osd.0] osd uuid = <UUID>, still not ???automount'
[20:34] <ron-slc> These continued to function, as they were manually defined. I know that ceph-deploy uses a slightly different method.
[20:34] <btaylor> got uuid from /dev/disk/by-uuid/...
[20:34] <elfurbe> @ron-slc: rgr that, I'll poke around
[20:34] <ron-slc> KK!
[20:35] * rendar (~I@host5-183-dynamic.46-79-r.retail.telecomitalia.it) Quit (Ping timeout: 480 seconds)
[20:40] * srk (~Siva@32.97.110.53) has joined #ceph
[20:41] * rraja (~rraja@121.244.87.117) Quit (Ping timeout: 480 seconds)
[20:42] <ron-slc> btaylor: sorry didn't see you snuck in. I'm unsure on the osd uuid = , personally. As I instead use the SAS/SCSI wwn- as in: evs = /dev/disk/by-id/wwn-0x6001e68482c3000018722f9613526fe8-part2
[20:43] <ron-slc> s/evs/devs/
[20:43] <btaylor> ron-slc: pulled it out of http://docs.ceph.com/docs/master/rados/configuration/osd-config-ref/ but was just thinking about it. if i have to manually add the uuid or disk dev into ceph.conf, then i might as well add it into /etc/fstab, right?
[20:44] * pvh_sa_ (~pvh@169-0-182-84.ip.afrihost.co.za) has joined #ceph
[20:44] <btaylor> and even in the description for the osd data setting right below osd uuid it says ??? The path to the OSDs data. You must create the directory when deploying Ceph. You should mount a drive for OSD data at this mount point. We do not recommend changing the default.???
[20:44] <ron-slc> btaylor: Either method would work! Long ago I used the fstab method, and things were great.. But now that ceph does a pretty good job or mounting, I prefer the non-fstab method(s)
[20:45] <btaylor> ???you should mount the drive??? is what stood out to me
[20:46] <ron-slc> agreed. :) every cluster is a little different. Based on when it was deployed. Mine has been around 1 year before Argonaut.. VERY VERY old. so I've always done things the "hard" way.
[20:47] <ron-slc> And then you get into differences like Operating Systems, and then which versions of that OS.. etc.
[20:48] <ron-slc> so things like mounting are always slightly different. :)
[20:50] * pvh_sa (~pvh@169-0-184-226.ip.afrihost.co.za) Quit (Ping timeout: 480 seconds)
[20:50] * treenerd_ (~gsulzberg@cpe90-146-148-47.liwest.at) Quit (Read error: No route to host)
[20:51] * valeech (~valeech@pool-96-247-203-33.clppva.fios.verizon.net) has joined #ceph
[20:51] * sleinen (~Adium@2001:620:1000:3:a65e:60ff:fedb:f305) has joined #ceph
[20:51] * lifeboy (~roland@196.45.25.14) Quit (Quit: Ex-Chat)
[20:53] * karnan (~karnan@106.51.136.211) Quit (Remote host closed the connection)
[20:54] * ledgr (~ledgr@78.57.252.56) has joined #ceph
[20:59] * kefu|afk (~kefu@114.92.125.128) Quit (Quit: My Mac has gone to sleep. ZZZzzz???)
[21:00] <btaylor> ron-slc: true.
[21:01] <btaylor> just means i???ll have to add that part into my automation..
[21:01] * rendar (~I@host5-183-dynamic.46-79-r.retail.telecomitalia.it) has joined #ceph
[21:07] * ledgr (~ledgr@78.57.252.56) Quit (Remote host closed the connection)
[21:07] * ledgr (~ledgr@88-119-196-104.static.zebra.lt) has joined #ceph
[21:08] * rwheeler (~rwheeler@nat-pool-bos-t.redhat.com) Quit (Quit: Leaving)
[21:11] * _mrp (~mrp@82.117.199.26) Quit (Quit: My Mac has gone to sleep. ZZZzzz???)
[21:11] * _mrp (~mrp@82.117.199.26) has joined #ceph
[21:13] * doppelgrau (~doppelgra@dslb-088-072-094-200.088.072.pools.vodafone-ip.de) Quit (Quit: doppelgrau)
[21:17] <mnaser_> does anyone have ideas about 10.2.3 release date?
[21:17] * mnaser_ is now known as mnaser
[21:18] <mnaser> i got bit by this.. http://tracker.ceph.com/issues/16672
[21:18] <mnaser> and we're putting our upgrade on a halt for now
[21:18] <mnaser> rebuilding osds as we go seems like a good way to ruin our disks :D
[21:19] * _mrp (~mrp@82.117.199.26) Quit (Ping timeout: 480 seconds)
[21:22] * jimbo_insa (~oftc-webi@129.162.189.102) has joined #ceph
[21:24] * bene2 (~bene@nat-pool-bos-t.redhat.com) has joined #ceph
[21:27] <jimbo_insa> So first time here, not sure the way this channel works, but I have a question about configuring a cephfs cluster. In particular, with the ceph mon daemons. Running jewel release on centos7
[21:29] * vata (~vata@96.127.202.136) has joined #ceph
[21:29] <mnaser> you should just ask your question :>
[21:30] <jimbo_insa> I'm setting up 4 ceph storage nodes dual connected to a private network, and my corporate network. On the private side, I have compute nodes mounting the ceph volume, over infiniband. The private network clients fstab mount use the private side IP of the ceph monitors
[21:31] * cyphase (~cyphase@000134f2.user.oftc.net) Quit (Ping timeout: 480 seconds)
[21:32] <jimbo_insa> However, some clients also need to mount the volume from our corporate network, but even thought the ceph storage/mds nodes are dual connected, they cant mount using the corporate side network addresses, and must use the private addresses of the ceph nodes. Is there any way to get the mds daemons to listen on all interfaces like the osd/mds daemons do?
[21:33] * cyphase (~cyphase@c-50-148-131-137.hsd1.ca.comcast.net) has joined #ceph
[21:33] * Uniju1 (~skrblr@46.166.188.236) has joined #ceph
[21:35] <jimbo_insa> It seems like its not possible. I was thinking of having two of my mds daemons added using their private side ip, and the third added using its corporate side ip. this way the corporate nodes could mount using that corporate ip, as the cluster is actually NAT from the main network via our cluster head node
[21:35] <jimbo_insa> so mon, not mds daemons
[21:36] <if_gagarin0>
[21:41] * vata (~vata@96.127.202.136) Quit (Quit: Leaving.)
[21:42] * vata (~vata@96.127.202.136) has joined #ceph
[21:48] * squizzi (~squizzi@107.13.237.240) Quit (Quit: bye)
[21:54] * davidzlap1 (~Adium@cpe-172-91-154-245.socal.res.rr.com) has joined #ceph
[21:55] * davidzlap (~Adium@2605:e000:1313:8003:992d:901f:dd23:80b) Quit (Ping timeout: 480 seconds)
[21:59] * TMM (~hp@dhcp-077-248-009-229.chello.nl) has joined #ceph
[22:02] * bene2 (~bene@nat-pool-bos-t.redhat.com) Quit (Ping timeout: 480 seconds)
[22:03] * Uniju1 (~skrblr@46.166.188.236) Quit ()
[22:14] * bauruine (~bauruine@2a01:4f8:130:8285:fefe::36) Quit (Quit: ZNC - http://znc.in)
[22:15] * BennyRene2016 (~BennyRene@host-89-241-119-65.as13285.net) has joined #ceph
[22:17] * georgem (~Adium@206.108.127.16) Quit (Quit: Leaving.)
[22:21] * mattbenjamin (~mbenjamin@12.118.3.106) Quit (Quit: Leaving.)
[22:22] * KeeperOfTheSoul (~SurfMaths@80.255.3.122) has joined #ceph
[22:26] * dgurtner (~dgurtner@84-73-130-19.dclient.hispeed.ch) Quit (Ping timeout: 480 seconds)
[22:26] * vbellur (~vijay@nat-pool-bos-t.redhat.com) Quit (Quit: Leaving.)
[22:27] * mykola (~Mikolaj@91.245.78.210) Quit (Quit: away)
[22:28] * bauruine (~bauruine@2a01:4f8:130:8285:fefe::36) has joined #ceph
[22:32] * vata (~vata@96.127.202.136) Quit (Quit: Leaving.)
[22:34] * vata (~vata@96.127.202.136) has joined #ceph
[22:38] <jimbo_insa> can the ceph monitor daemons listen on more than one ip?
[22:38] * brians (~brian@80.111.114.175) Quit (Quit: Textual IRC Client: www.textualapp.com)
[22:41] * vvv (~vvv@1.54.212.20) has joined #ceph
[22:41] * vvv (~vvv@1.54.212.20) has left #ceph
[22:45] * wak-work (~wak-work@2620:15c:2c5:3:3938:b41d:15b2:ea22) Quit (Remote host closed the connection)
[22:45] * wak-work (~wak-work@2620:15c:2c5:3:64:234a:77af:bd28) has joined #ceph
[22:46] <jimbo_insa> Does anyone actually try to assist in this channel?
[22:47] * wak-work (~wak-work@2620:15c:2c5:3:64:234a:77af:bd28) Quit (Remote host closed the connection)
[22:47] * wak-work (~wak-work@2620:15c:2c5:3:64:234a:77af:bd28) has joined #ceph
[22:47] * bene2 (~bene@nat-pool-bos-t.redhat.com) has joined #ceph
[22:49] * wak-work (~wak-work@2620:15c:2c5:3:64:234a:77af:bd28) Quit (Remote host closed the connection)
[22:50] * wak-work (~wak-work@2620:15c:2c5:3:64:234a:77af:bd28) has joined #ceph
[22:51] * ledgr (~ledgr@88-119-196-104.static.zebra.lt) Quit (Remote host closed the connection)
[22:51] * KeeperOfTheSoul (~SurfMaths@80.255.3.122) Quit ()
[22:55] * ledgr (~ledgr@78.57.252.56) has joined #ceph
[23:07] * wes_dillingham (~wes_dilli@140.247.242.44) Quit (Ping timeout: 480 seconds)
[23:11] * ledgr (~ledgr@78.57.252.56) Quit (Remote host closed the connection)
[23:15] * nils_ (~nils_@doomstreet.collins.kg) has joined #ceph
[23:19] * ffilzwin (~ffilz@c-76-115-190-27.hsd1.or.comcast.net) Quit (Quit: Leaving)
[23:19] * northrup (~northrup@75-146-11-137-Nashville.hfc.comcastbusiness.net) has joined #ceph
[23:19] <northrup> Is there a way to change who the active fsmon is?
[23:20] <northrup> that doesn't take the cephFS system down?
[23:22] * srk (~Siva@32.97.110.53) Quit (Ping timeout: 480 seconds)
[23:26] * johnavp1989 (~jpetrini@8.39.115.8) Quit (Ping timeout: 480 seconds)
[23:28] * srk (~Siva@32.97.110.53) has joined #ceph
[23:30] * ffilzwin (~ffilz@c-76-115-190-27.hsd1.or.comcast.net) has joined #ceph
[23:33] * mhack (~mhack@nat-pool-bos-t.redhat.com) Quit (Remote host closed the connection)
[23:38] <northrup> does anyone in here know if it's possible to change the CephFS 'fsmap' active node without taking down time?
[23:38] * Megadawg (~Superdawg@ec2-54-243-59-20.compute-1.amazonaws.com) has left #ceph
[23:39] <northrup> anyone from InkTank?
[23:46] * toastydeath (~toast@pool-71-255-253-39.washdc.fios.verizon.net) Quit (Ping timeout: 480 seconds)
[23:49] * haplo37 (~haplo37@107-190-37-90.cpe.teksavvy.com) Quit (Read error: Connection reset by peer)
[23:50] * sleinen (~Adium@2001:620:1000:3:a65e:60ff:fedb:f305) Quit (Quit: Leaving.)
[23:57] * fsimonce (~simon@host145-64-dynamic.52-79-r.retail.telecomitalia.it) Quit (Remote host closed the connection)

These logs were automatically created by CephLogBot on irc.oftc.net using the Java IRC LogBot.