From owner-freebsd-fs@freebsd.org Tue Mar 8 03:56:23 2016 Return-Path: Delivered-To: freebsd-fs@mailman.ysv.freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:1900:2254:206a::19:1]) by mailman.ysv.freebsd.org (Postfix) with ESMTP id 89035AC74E3; Tue, 8 Mar 2016 03:56:23 +0000 (UTC) (envelope-from fred.fliu@gmail.com) Received: from mail-lb0-x22a.google.com (mail-lb0-x22a.google.com [IPv6:2a00:1450:4010:c04::22a]) (using TLSv1.2 with cipher ECDHE-RSA-AES128-GCM-SHA256 (128/128 bits)) (Client CN "smtp.gmail.com", Issuer "Google Internet Authority G2" (verified OK)) by mx1.freebsd.org (Postfix) with ESMTPS id 1457D764; Tue, 8 Mar 2016 03:56:23 +0000 (UTC) (envelope-from fred.fliu@gmail.com) Received: by mail-lb0-x22a.google.com with SMTP id k15so3620913lbg.0; Mon, 07 Mar 2016 19:56:22 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20120113; h=mime-version:in-reply-to:references:date:message-id:subject:from:to :cc; bh=OaFv1WIP9IS0uwcRqw1S9Sn0qI/rckbXP42M6r28SYU=; b=oGeVYvu0J03Ao//K4hcpZuM9nOLK38PIHyZlLYcgPDAM7kjE5fgsRSFwFRKa1hwtFo +lepalhJXSTfpo7X3RG2yhNYikpREkzZD+kF/LrWySvSdQVXc4vC+A1Ovw1dby9BaUDT QVJEJ+51WvJB6L0zVFEvOtyquK9LjCz0MjcKOU0YaP7xkh4sYmfwFgQ9hyU/8eq5zWV2 BC6ZwG2GYSMIXL5IbfncxYB2geXtNqWtBFnmySycoVTP27N2qUR9Gb7TAxVCOjzTG8an 9T5ZLEF1v1Ve3HP1eoQacQblbGBrIUML+8U46MXe31jcuiXjho5MA47Zf8R8sTFUqIap U/4w== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20130820; h=x-gm-message-state:mime-version:in-reply-to:references:date :message-id:subject:from:to:cc; bh=OaFv1WIP9IS0uwcRqw1S9Sn0qI/rckbXP42M6r28SYU=; b=Ksm3drqv4GhDm7pyES63WZJXsxJwzAyfEE9zRftRHvJ5BhTmCzOdiPbssyWJT51s3/ B6yyb32YP40jbGIExuMPsLNY831msu5i+VNLMZDr6/XuPvbil5PHiRrjovik7zQjuaa+ Qvo2VHpi3yB5xtYM/KccuLjUnLzH4t7kXXCWYMzWVCUoZgruyynPg5bw/I1vC4ifVsbL RN5fYV0dYZjUM28QeMv8FiKSIFeRlP9J6SpKXQEndEpQpMEJejexopfgs87Rjoz5XnqK pkfx+yRmhP/OiXAbQJxb/H6OEJhXkIZPvJtmU1d+522zH1DXCQ1T+dEvKfXf/xgFoUVI TpCg== X-Gm-Message-State: AD7BkJI4G93ofocIaYuYHibIaoGVlrPQrOE4zG0U7s7dNePl89gEqpCi5937GQzC55Rh9PPBkbzz+bRAD+xEPA== MIME-Version: 1.0 X-Received: by 10.25.161.131 with SMTP id k125mr9032768lfe.83.1457409380752; Mon, 07 Mar 2016 19:56:20 -0800 (PST) Received: by 10.25.20.164 with HTTP; Mon, 7 Mar 2016 19:56:20 -0800 (PST) In-Reply-To: References: <95563acb-d27b-4d4b-b8f3-afeb87a3d599@me.com> <56D87784.4090103@broken.net> <5158F354-9636-4031-9536-E99450F312B3@RichardElling.com> <6E2B77D1-E0CA-4901-A6BD-6A22C07536B3@gmail.com> Date: Tue, 8 Mar 2016 11:56:20 +0800 Message-ID: Subject: Re: [zfs] [developer] Re: [smartos-discuss] an interesting survey -- the zpool with most disks you have ever built From: Fred Liu To: illumos-zfs Cc: "smartos-discuss@lists.smartos.org" , developer@lists.open-zfs.org, developer , illumos-developer , omnios-discuss , Discussion list for OpenIndiana , "zfs-discuss@list.zfsonlinux.org" , "freebsd-fs@FreeBSD.org" , "zfs-devel@freebsd.org" X-Mailman-Approved-At: Tue, 08 Mar 2016 05:03:16 +0000 Content-Type: text/plain; charset=UTF-8 X-Content-Filtered-By: Mailman/MimeDel 2.1.21 X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.21 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Tue, 08 Mar 2016 03:56:23 -0000 2016-03-08 4:55 GMT+08:00 Liam Slusser : > I don't have a 2000 drive array (thats amazing!) but I do have two 280 > drive arrays which are in production. Here are the generic stats: > > server setup: > OpenIndiana oi_151 > 1 server rack > Dell r720xd 64g ram with mirrored 250g boot disks > 5 x LSI 9207-8e dualport SAS pci-e host bus adapters > Intel 10g fibre ethernet (dual port) > 2 x SSD for log cache > 2 x SSD for cache > 23 x Dell MD1200 with 3T,4T, or 6T NLSAS disks (a mix of Toshiba, Western > Digital, and Seagate drives - basically whatever Dell sends) > > zpool setup: > 23 x 12-disk raidz2 glued together. 276 total disks. Basically each new > 12 disk MD1200 is a new raidz2 added to the pool. > > Total size: ~797T > > We have an identical server which we replicate changes via zfs snapshots > every few minutes. The whole setup as been up and running for a few years > now, no issues. As we run low on space we purchase two additional MD1200 > shelfs (one for each system) and add the new raidz2 into pool on-the-fly. > > The only real issues we've had is sometimes a disk fails in such a way > (think Monty Python and the holy grail i'm not dead yet) where the disk > hasn't failed but is timing out and slows the whole array to a standstill > until we can manual find and remove the disk. Other problems are once a > disk has been replaced sometimes the resilver process can take > an eternity. We have also found the snapshot replication process can > interfere with the resilver process - resilver gets stuck at 99% and never > ends - so we end up stopping or only doing one replication a day until the > resilver process is done. > > The last helpful hint I have was lowering all the drive timeouts, see > http://everycity.co.uk/alasdair/2011/05/adjusting-drive-timeouts-with-mdb-on-solaris-or-openindiana/ > for info. > > [Fred]: zpool wiith 280 drives in production is pretty big! I think 2000 > drives were just in test. It is true that huge pools have lots of operation > challenges. I have met the similar sluggish issue caused by a > will-die disk. Just curious, what is the cluster software implemented in http://everycity.co.uk/alasdair/2011/05/adjusting-drive-timeouts-with-mdb-on-solaris-or-openindiana/ ? Thanks. Fred > > > >>> >>> >> > *illumos-zfs* | Archives > > | > Modify > > Your Subscription >