From owner-freebsd-fs@FreeBSD.ORG Sun Oct 17 02:10:16 2010 Return-Path: Delivered-To: freebsd-fs@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id 393F2106564A; Sun, 17 Oct 2010 02:10:16 +0000 (UTC) (envelope-from fjwcash@gmail.com) Received: from mail-gx0-f182.google.com (mail-gx0-f182.google.com [209.85.161.182]) by mx1.freebsd.org (Postfix) with ESMTP id D504A8FC12; Sun, 17 Oct 2010 02:10:15 +0000 (UTC) Received: by gxk4 with SMTP id 4so1053271gxk.13 for ; Sat, 16 Oct 2010 19:10:14 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=gamma; h=domainkey-signature:mime-version:received:received:in-reply-to :references:date:message-id:subject:from:to:cc:content-type :content-transfer-encoding; bh=PHTphSreKPZtTip+oHwKeIRYDXlQl9GEIrYSfmZkNi8=; b=uO8xOUwF5353fu6p07s2ssOv8VTofzGzJnru4WUT9+bKn5pTfM0JvW3aa0cslLGJA6 R5RfskoGeoP1R5Mh3CbX85SoXxG/ppcLYdhLT1A7eXpsRn+8N/8q2Pg1btyqtDL9ohor /yqOBDhyztBl4si/XtXEjql+ZwHXHozsO2PPk= DomainKey-Signature: a=rsa-sha1; c=nofws; d=gmail.com; s=gamma; h=mime-version:in-reply-to:references:date:message-id:subject:from:to :cc:content-type:content-transfer-encoding; b=ut75IgvTrFSbKNTDt8waMyECneMoipOUW5lCrvOG36iE1pS2gAnH2gqv2F4Fstdoqo mCKq1RTGp5ytUl/A7CIUh8khp8lYaHPGVaJSTeJr83r7gRty9J+b400ClnqeTC75OTVs UURSjSTP007oLFBHCnkb3S5Y5WPRjZjFvT+ts= MIME-Version: 1.0 Received: by 10.90.26.6 with SMTP id 6mr975948agz.101.1287281414704; Sat, 16 Oct 2010 19:10:14 -0700 (PDT) Received: by 10.90.56.10 with HTTP; Sat, 16 Oct 2010 19:10:14 -0700 (PDT) In-Reply-To: <20101016222833.GA6765@garage.freebsd.pl> References: <20101016222833.GA6765@garage.freebsd.pl> Date: Sat, 16 Oct 2010 19:10:14 -0700 Message-ID: From: Freddie Cash To: Pawel Jakub Dawidek Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: quoted-printable Cc: freebsd-fs@freebsd.org Subject: Re: Converting a non-HAST ZFS pool to a HAST pool X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Sun, 17 Oct 2010 02:10:16 -0000 On Sat, Oct 16, 2010 at 3:28 PM, Pawel Jakub Dawidek wrot= e: > On Fri, Oct 15, 2010 at 11:37:34AM -0700, Freddie Cash wrote: >> Has anyone looked into, attempted, or considered converting a non-HAST >> ZFS pool configuration into a HAST one? =C2=A0While the pool is live and >> the server is in use. =C2=A0Would it even be possible? >> >> For example, would the following work (in a pool with a single raidz2 >> vdev, where the underlying GEOM provider is glabel) >> =C2=A0 - zpool offline 1 drive =C2=A0(pool is now running degraded) >> =C2=A0 - configure hastd in master mode with a single provider using the >> "offline" disk (hast metadata takes the place of glabel metadata) > > HAST metadata takes much more space than glabel metadata. The latter > takes only one sector, while the former depends on provider size, but we > have to keep entire extent bitmap there, so definitely more than one > sector. Hrm, darn. So close. :) >> =C2=A0 - zpool replace the offline drive using the /dev/hast/whatever pr= ovider >> =C2=A0 - wait for ZFS to resilver the device >> =C2=A0 - repeat for each disk in the vdev >> >> Is it possible to change the hast configuration and reload hastd to >> pick up new providers, without restarting it? =C2=A0Without losing the >> /dev/hast devices? =C2=A0Or would it require a restart, which would caus= e >> the /dev/hast/* devices to disappear, thus b0rking the pool? > > It is possible in 8-STABLE. Good to know. >> Just wondering how hard it's going to be to move from our current >> setup (backup server 1 rsync'd to backup server 2) to a HAST setup >> (server 1 hast master with server 2 hast slave). >> >> If the above won't work, then I guess I'll have to grab a third >> storage box, configure that as a hast box, copy the data over, destroy >> server 2, configure it as a hast box, and let them sync. > > I'm afraid you have to backup your pool, recreate it on HAST provider > and recover. That's what I figured, and why I thought I'd post before attempting it. :) Thanks for the info. --=20 Freddie Cash fjwcash@gmail.com From owner-freebsd-fs@FreeBSD.ORG Sun Oct 17 03:39:00 2010 Return-Path: Delivered-To: freebsd-fs@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id 22566106564A for ; Sun, 17 Oct 2010 03:39:00 +0000 (UTC) (envelope-from jhellenthal@gmail.com) Received: from mail-gx0-f182.google.com (mail-gx0-f182.google.com [209.85.161.182]) by mx1.freebsd.org (Postfix) with ESMTP id C88088FC08 for ; Sun, 17 Oct 2010 03:38:59 +0000 (UTC) Received: by gxk4 with SMTP id 4so1065520gxk.13 for ; Sat, 16 Oct 2010 20:38:59 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=gamma; h=domainkey-signature:received:received:sender:message-id:date:from :organization:user-agent:mime-version:to:cc:subject:references :in-reply-to:x-enigmail-version:content-type :content-transfer-encoding; bh=ZI3iWXj00uFBuN9a7zEFYhkMq4RQSxZ/GL0mkT/5P70=; b=VpngejTfhgz3XsrYKlbFhm+gPq7ql6RE0XYJn7dvg/Ih6KcoTKz8ua9DnsaEyKRb8y zLRh9Vmzi8KZtu0a96QxnoRLBCR8BzSlMJq+fXrwmghG3WpcMhTg1dm4jGMyJ4lpiuNs OtOM4G3u0+rsnLVcOLFMKV0ApBOHVi3GpY/jk= DomainKey-Signature: a=rsa-sha1; c=nofws; d=gmail.com; s=gamma; h=sender:message-id:date:from:organization:user-agent:mime-version:to :cc:subject:references:in-reply-to:x-enigmail-version:content-type :content-transfer-encoding; b=sw/lqXZ7iWin9+obB+hf58zSpPdcEHL/INF/RCM4nq422mYJE7vrsydlLPy6SXEL5P eCu6dkRFAD5mGsHQri0uqtP0LK+Qx5q71n1KluSFjQjTCsOuekwQ9LEszqX8RfTRtTiL B237j8TxkCcFjQXzcYb5BXVxMlKLVJ3CT5GOg= Received: by 10.236.102.145 with SMTP id d17mr3931643yhg.29.1287286738887; Sat, 16 Oct 2010 20:38:58 -0700 (PDT) Received: from centel.dataix.local ([99.181.158.202]) by mx.google.com with ESMTPS id j64sm7054954yha.24.2010.10.16.20.38.56 (version=SSLv3 cipher=RC4-MD5); Sat, 16 Oct 2010 20:38:57 -0700 (PDT) Sender: "J. Hellenthal" Message-ID: <4CBA6FCF.3080301@DataIX.net> Date: Sat, 16 Oct 2010 23:38:55 -0400 From: jhell Organization: http://www.DataIX.net User-Agent: Mozilla/5.0 (X11; U; FreeBSD i386; en-US; rv:1.9.2.9) Gecko/20100917 Lightning/1.0b1 Thunderbird MIME-Version: 1.0 To: David Samms References: <201010151909.o9FJ9cZf065459@lurza.secnetix.de> In-Reply-To: X-Enigmail-Version: 1.1.2 Content-Type: text/plain; charset=ISO-8859-1 Content-Transfer-Encoding: 7bit Cc: freebsd-fs@freebsd.org Subject: Re: ZFS trouble: unbelievably large files created X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Sun, 17 Oct 2010 03:39:00 -0000 On 10/15/2010 18:44, David Samms wrote: > On 10/15/10 15:09, Oliver Fromme wrote: >> > ls -kls shawxp/var/amavis/.spamBAD >> > total 13352131 >> > 771 -rw------- 1 110 110 684032 Oct 15 12:02 >> > auto-whitelist >> > 1 -rw------- 1 110 110 40 Oct 15 12:37 >> bayes.lock >> > 1284 -rw------- 1 110 110 1294336 Oct 15 12:38 >> bayes_seen >> > 4229 -rw------- 1 110 110 4227072 Oct 15 12:38 >> bayes_toks >> > 5025103 -rw------- 1 110 110 553184002048 Oct 15 12:38 >> > bayes_toks.expire3515 >> > 8320745 -rw------- 1 110 110 140743122878464 Oct 15 12:14 >> > bayes_toks.expire97254 >> >> Ok, so those files are so-called "sparse" files, i.e. they >> contain holes that don't actually occupy disk space. >> >> The numbers in the first column indicate the amount of >> physical disk space allocated (in KB). That's about 5 GB >> for the first file and 8 GB for the second (this is also >> consistent with the "total" value in the first line of the >> ls output, i.e. about 13 GB). >> >> That's still quite big, but certainly not in the TB range. >> I do not know why amavis creates such large sparse files, >> though. > > Under UFS I don't observe amavis creating the sparse files. The problem > is fairly repeatable, not on demand, but occurs several times a day. > High disk activity is a sure sign amavis is creating large files. Task > shuts down normaly and cpu load is low. > > This is a production server but I have moved all critical customer jails > back to UFS so can do testing if anyone has any ideas of what to look for. > I do not believe this to be a problem with amavis/amavisd, lets get that out of the way. Why I say this is because I have experienced the same problem on a server with UFS and p5-Mail-Spamassassin. I was logged into that server doing some work and there was plenty of disk space to do the work as it consisted of copying a ton of space eating files to the file-system where this happened. After operations were done, cleaned up the disk properly restoring space back for normal usage and then logged out. I came back the next day only to find my INBOX full of error messages and other junk from that server saying things like can't write to disk and such error messages. What it turned out to be was the same files that you had mentioned above but in my ~/.spamassassin/.bayes*. One of the files just like yours had grown to eat up the entire amount of disk space that was left besides the 8% that was reserved. I tried to find the offending email messages that might have caused this to happen but in some odd ~7 hours total time did not come up with a conclusion and could not afford to waste further time on that issue. IMHO this has to do with bayes and the usage of dbm, or db4* and something in a messages header or body that triggers it to happen but do not have enough time or conclusive evidence yet to suggest anything further. Ultimately I removed the offending files and went about business as usual until I see it happen again. Also tuned the quota for myself to be a much lower amount disk space so I should get an email as soon as it happens again without effecting the rest of operations on that disk. Regards, -- jhell,v From owner-freebsd-fs@FreeBSD.ORG Sun Oct 17 05:19:36 2010 Return-Path: Delivered-To: freebsd-fs@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id 53E6F106566C for ; Sun, 17 Oct 2010 05:19:36 +0000 (UTC) (envelope-from sfourman@gmail.com) Received: from mail-qy0-f175.google.com (mail-qy0-f175.google.com [209.85.216.175]) by mx1.freebsd.org (Postfix) with ESMTP id 0E06D8FC0A for ; Sun, 17 Oct 2010 05:19:35 +0000 (UTC) Received: by qyk31 with SMTP id 31so2809679qyk.13 for ; Sat, 16 Oct 2010 22:19:35 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=gamma; h=domainkey-signature:mime-version:received:received:in-reply-to :references:date:message-id:subject:from:to:cc:content-type; bh=SaOaOtHUoqcdFRfFExAROU5b2AzQW4bVmWuNG+I+AN8=; b=yC2ghr97BaFuB8aq5HmfyIvq6/Z4XnqhuskBe7Ote4CyqxcruvimG/Uq+HJ/uS2oCb 5t7RvP6ONuf0jj0cqqvjIfTs6WVKlS8ShWaRoqrNOTdLhVEhIL9cuqTYMXOeGE6TdzmH l+F7+2e2Q09BORRlJ9W+efI3NmWC6ew8RokCM= DomainKey-Signature: a=rsa-sha1; c=nofws; d=gmail.com; s=gamma; h=mime-version:in-reply-to:references:date:message-id:subject:from:to :cc:content-type; b=hgW1YuBvTBJx+8SnbuGFOQmIPD6Ys+yDRObf2UPKOqww5C5ofNMyYk3TSXb3CJOQDo rr7g/5Jea5WUgSFOudnK+2RAjC66Vqn155CQxCPH3Ph5wRmFJKKoVaAoQGvHDA0znpwo 6IeqPQyA656xBPf3Ecjj+L754n1npmZli2gRo= MIME-Version: 1.0 Received: by 10.229.223.149 with SMTP id ik21mr2441563qcb.224.1287292773548; Sat, 16 Oct 2010 22:19:33 -0700 (PDT) Received: by 10.229.240.72 with HTTP; Sat, 16 Oct 2010 22:19:33 -0700 (PDT) In-Reply-To: References: Date: Sun, 17 Oct 2010 00:19:33 -0500 Message-ID: From: "Sam Fourman Jr." To: David Samms Content-Type: text/plain; charset=ISO-8859-1 Cc: freebsd-fs@freebsd.org Subject: Re: ZFS trouble: unbelievably large files created X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Sun, 17 Oct 2010 05:19:36 -0000 On Fri, Oct 15, 2010 at 12:07 PM, David Samms wrote: > I am running into issues with ZFS where perl, specifically amavisd running > spamassassin creates unbelievably large files, as in files 100x the size of > the hard disk. > I can confirm that this is a problem with ZFS, I saw it on a amd64 FreeBSD 8.0 install using samba as a fileserver, the clients are windows xp. a 14TB file would be created every time we would copy a file directly from USB storage(plugged into a xp client) and drag and drop it on a ZFS backed samba share. I never could fix the problem... so I reformatted and installed RELENG_8, and the problem never came back. -- Sam Fourman Jr. Fourman Networks http://www.fourmannetworks.com From owner-freebsd-fs@FreeBSD.ORG Sun Oct 17 15:34:30 2010 Return-Path: Delivered-To: freebsd-fs@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id 28AF4106566B; Sun, 17 Oct 2010 15:34:30 +0000 (UTC) (envelope-from to.my.trociny@gmail.com) Received: from mail-fx0-f54.google.com (mail-fx0-f54.google.com [209.85.161.54]) by mx1.freebsd.org (Postfix) with ESMTP id 7FE0B8FC17; Sun, 17 Oct 2010 15:34:28 +0000 (UTC) Received: by fxm12 with SMTP id 12so28073fxm.13 for ; Sun, 17 Oct 2010 08:34:28 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=gamma; h=domainkey-signature:received:received:from:to:cc:subject:date :message-id:user-agent:mime-version:content-type; bh=ZXe6PeCo52qEmn+ye+vlZmceQkmbkhfFOPjcqmEhbu0=; b=gQjyBW2S2oGR9ae88R9FoJJ9stSyOoq5leFdxf7X+6Tk/jNUNhhEL66TC47iWgmU3i KTag9Rbhjgdi21+fBLm+2gCQ7u+mSSn05rUcSEJyGB961ZF4VjFlbIVfCnhlWMDwSxmC HEfTEfX8zV+pOxAXiYSCG3SuyDnAz73PN2S5g= DomainKey-Signature: a=rsa-sha1; c=nofws; d=gmail.com; s=gamma; h=from:to:cc:subject:date:message-id:user-agent:mime-version :content-type; b=QeTrKT/FyyPeGX6ZH11DafFDXfVaOhVfB3VJfqlaHLSj9S2OGzU9EXODWbIO+hkMOe 7EiDBJFWCFR0bsGfRx4nbBY58PXp1l+kCOCGOK4INSwVlyv/6vH/T3kAJ+3Cak8Fmjx0 +y5Duy3NI/fDe87OsYlpWOBQb+sNBI/7uLj68= Received: by 10.223.1.133 with SMTP id 5mr378367faf.18.1287329668174; Sun, 17 Oct 2010 08:34:28 -0700 (PDT) Received: from localhost ([95.69.174.185]) by mx.google.com with ESMTPS id s18sm3718576fax.34.2010.10.17.08.34.26 (version=TLSv1/SSLv3 cipher=RC4-MD5); Sun, 17 Oct 2010 08:34:27 -0700 (PDT) From: Mikolaj Golub To: freebsd-fs@freebsd.org Date: Sun, 17 Oct 2010 18:34:24 +0300 Message-ID: <86zkuc4yv3.fsf@kopusha.home.net> User-Agent: Gnus/5.13 (Gnus v5.13) Emacs/23.2 (berkeley-unix) MIME-Version: 1.0 Content-Type: multipart/mixed; boundary="=-=-=" Cc: Pawel Jakub Dawidek Subject: hastd: wrong address is logged when cancelling half-open connection X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Sun, 17 Oct 2010 15:34:30 -0000 --=-=-= Hi, It looks like there is a typo in the listen_accept(): when half-open connection is cancelled hastd reports a wrong address being cancelled. See the attached patch. -- Mikolaj Golub --=-=-= Content-Type: text/x-patch Content-Disposition: inline; filename=hastd.cancel.patch Index: sbin/hastd/hastd.c =================================================================== --- sbin/hastd/hastd.c (revision 213978) +++ sbin/hastd/hastd.c (working copy) @@ -527,7 +527,7 @@ listen_accept(void) } else if (res->hr_remotein != NULL) { char oaddr[256]; - proto_remote_address(conn, oaddr, sizeof(oaddr)); + proto_remote_address(res->hr_remotein, oaddr, sizeof(oaddr)); pjdlog_debug(1, "Canceling half-open connection from %s on connection from %s.", oaddr, raddr); --=-=-=-- From owner-freebsd-fs@FreeBSD.ORG Sun Oct 17 15:47:52 2010 Return-Path: Delivered-To: freebsd-fs@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id 4104410656C4 for ; Sun, 17 Oct 2010 15:47:52 +0000 (UTC) (envelope-from pjd@garage.freebsd.pl) Received: from mail.garage.freebsd.pl (60.wheelsystems.com [83.12.187.60]) by mx1.freebsd.org (Postfix) with ESMTP id E19998FC1C for ; Sun, 17 Oct 2010 15:47:49 +0000 (UTC) Received: by mail.garage.freebsd.pl (Postfix, from userid 65534) id 50E3345E98; Sun, 17 Oct 2010 17:47:48 +0200 (CEST) Received: from localhost (chello089073192049.chello.pl [89.73.192.49]) (using TLSv1 with cipher DHE-RSA-AES256-SHA (256/256 bits)) (No client certificate requested) by mail.garage.freebsd.pl (Postfix) with ESMTP id 0D1E245E87; Sun, 17 Oct 2010 17:47:42 +0200 (CEST) Date: Sun, 17 Oct 2010 17:47:10 +0200 From: Pawel Jakub Dawidek To: Mikolaj Golub Message-ID: <20101017154710.GD6765@garage.freebsd.pl> References: <86zkuc4yv3.fsf@kopusha.home.net> Mime-Version: 1.0 Content-Type: multipart/signed; micalg=pgp-sha1; protocol="application/pgp-signature"; boundary="oj4kGyHlBMXGt3Le" Content-Disposition: inline In-Reply-To: <86zkuc4yv3.fsf@kopusha.home.net> User-Agent: Mutt/1.4.2.3i X-PGP-Key-URL: http://people.freebsd.org/~pjd/pjd.asc X-OS: FreeBSD 9.0-CURRENT amd64 X-Spam-Checker-Version: SpamAssassin 3.0.4 (2005-06-05) on mail.garage.freebsd.pl X-Spam-Level: X-Spam-Status: No, score=-0.6 required=4.5 tests=BAYES_00,RCVD_IN_SORBS_DUL autolearn=no version=3.0.4 Cc: freebsd-fs@freebsd.org Subject: Re: hastd: wrong address is logged when cancelling half-open connection X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Sun, 17 Oct 2010 15:47:52 -0000 --oj4kGyHlBMXGt3Le Content-Type: text/plain; charset=us-ascii Content-Disposition: inline Content-Transfer-Encoding: quoted-printable On Sun, Oct 17, 2010 at 06:34:24PM +0300, Mikolaj Golub wrote: > Hi, >=20 > It looks like there is a typo in the listen_accept(): when half-open > connection is cancelled hastd reports a wrong address being cancelled. Se= e the > attached patch. Committed, thanks! --=20 Pawel Jakub Dawidek http://www.wheelsystems.com pjd@FreeBSD.org http://www.FreeBSD.org FreeBSD committer Am I Evil? Yes, I Am! --oj4kGyHlBMXGt3Le Content-Type: application/pgp-signature Content-Disposition: inline -----BEGIN PGP SIGNATURE----- Version: GnuPG v2.0.14 (FreeBSD) iEYEARECAAYFAky7Gn0ACgkQForvXbEpPzTHLgCfRTSxSjxSyslXQMnKgNZOEi0M rEYAnR7/q8t8DZ4LaCyPcFIGipbkEC2w =2wt4 -----END PGP SIGNATURE----- --oj4kGyHlBMXGt3Le-- From owner-freebsd-fs@FreeBSD.ORG Mon Oct 18 02:34:34 2010 Return-Path: Delivered-To: freebsd-fs@FreeBSD.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id E08B11065674 for ; Mon, 18 Oct 2010 02:34:34 +0000 (UTC) (envelope-from dougb@FreeBSD.org) Received: from mail2.fluidhosting.com (mx23.fluidhosting.com [204.14.89.6]) by mx1.freebsd.org (Postfix) with ESMTP id 6FF9A8FC12 for ; Mon, 18 Oct 2010 02:34:34 +0000 (UTC) Received: (qmail 26446 invoked by uid 399); 18 Oct 2010 02:15:47 -0000 Received: from localhost (HELO laptop.dougb.net) (dougb@dougbarton.us@127.0.0.1) by localhost with ESMTPAM; 18 Oct 2010 02:15:47 -0000 X-Originating-IP: 127.0.0.1 X-Sender: dougb@dougbarton.us Date: Sun, 17 Oct 2010 19:15:45 -0700 (PDT) From: Doug Barton To: freebsd-fs@FreeBSD.org Message-ID: User-Agent: Alpine 2.00 (BSF 1167 2008-08-23) X-message-flag: Outlook -- Not just for spreading viruses anymore! OpenPGP: id=1A1ABC84 Organization: http://SupersetSolutions.com/ MIME-Version: 1.0 Content-Type: TEXT/PLAIN; format=flowed; charset=US-ASCII Cc: Subject: Is there a way to fsck ext[23]fs? X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Mon, 18 Oct 2010 02:34:35 -0000 I have a shared partition that is formatted with ext3, and after a recent crash when I tried rebooting it came up dirty and wouldn't mount. Some brief looking through man pages didn't reveal anything useful, so I fsck'ed it in linux; but obviously that's not an optimal solution. I'll be happy with an RTFM if someone can point me to the right M. :) Doug From owner-freebsd-fs@FreeBSD.ORG Mon Oct 18 03:11:16 2010 Return-Path: Delivered-To: freebsd-fs@FreeBSD.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id 570881065675 for ; Mon, 18 Oct 2010 03:11:16 +0000 (UTC) (envelope-from spawk@acm.poly.edu) Received: from acm.poly.edu (acm.poly.edu [128.238.9.200]) by mx1.freebsd.org (Postfix) with ESMTP id E97D28FC0C for ; Mon, 18 Oct 2010 03:11:15 +0000 (UTC) Received: (qmail 45445 invoked from network); 18 Oct 2010 02:44:35 -0000 Received: from unknown (HELO ?192.168.0.2?) (spawk@96.224.221.101) by acm.poly.edu with CAMELLIA256-SHA encrypted SMTP; 18 Oct 2010 02:44:35 -0000 Message-ID: <4CBBB467.6000601@acm.poly.edu> Date: Sun, 17 Oct 2010 22:43:51 -0400 From: Boris Kochergin User-Agent: Mozilla/5.0 (X11; U; FreeBSD i386; en-US; rv:1.9.2.9) Gecko/20101011 Thunderbird/3.1.4 MIME-Version: 1.0 To: Doug Barton References: In-Reply-To: Content-Type: text/plain; charset=ISO-8859-1; format=flowed Content-Transfer-Encoding: 7bit Cc: freebsd-fs@FreeBSD.org Subject: Re: Is there a way to fsck ext[23]fs? X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Mon, 18 Oct 2010 03:11:16 -0000 On 10/17/10 22:15, Doug Barton wrote: > I have a shared partition that is formatted with ext3, and after a > recent crash when I tried rebooting it came up dirty and wouldn't > mount. Some brief looking through man pages didn't reveal anything > useful, so I fsck'ed it in linux; but obviously that's not an optimal > solution. > > I'll be happy with an RTFM if someone can point me to the right M. :) > > > Doug > _______________________________________________ > freebsd-fs@freebsd.org mailing list > http://lists.freebsd.org/mailman/listinfo/freebsd-fs > To unsubscribe, send any mail to "freebsd-fs-unsubscribe@freebsd.org" There are fsck programs for ext2/3/4 in /usr/ports/sysutils/e2fsprogs/. -Boris From owner-freebsd-fs@FreeBSD.ORG Mon Oct 18 05:23:42 2010 Return-Path: Delivered-To: freebsd-fs@FreeBSD.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id 6A524106566C for ; Mon, 18 Oct 2010 05:23:42 +0000 (UTC) (envelope-from dougb@FreeBSD.org) Received: from mail2.fluidhosting.com (mx23.fluidhosting.com [204.14.89.6]) by mx1.freebsd.org (Postfix) with ESMTP id EC8398FC1A for ; Mon, 18 Oct 2010 05:23:41 +0000 (UTC) Received: (qmail 28395 invoked by uid 399); 18 Oct 2010 05:23:40 -0000 Received: from localhost (HELO laptop.dougb.net) (dougb@dougbarton.us@127.0.0.1) by localhost with ESMTPAM; 18 Oct 2010 05:23:40 -0000 X-Originating-IP: 127.0.0.1 X-Sender: dougb@dougbarton.us Date: Sun, 17 Oct 2010 22:23:35 -0700 (PDT) From: Doug Barton To: Boris Kochergin In-Reply-To: <4CBBB467.6000601@acm.poly.edu> Message-ID: References: <4CBBB467.6000601@acm.poly.edu> User-Agent: Alpine 2.00 (BSF 1167 2008-08-23) X-message-flag: Outlook -- Not just for spreading viruses anymore! OpenPGP: id=1A1ABC84 Organization: http://SupersetSolutions.com/ MIME-Version: 1.0 Content-Type: TEXT/PLAIN; charset=US-ASCII; format=flowed Cc: freebsd-fs@FreeBSD.org Subject: Re: Is there a way to fsck ext[23]fs? X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Mon, 18 Oct 2010 05:23:42 -0000 -----BEGIN PGP SIGNED MESSAGE----- Hash: SHA256 On Sun, 17 Oct 2010, Boris Kochergin wrote: > There are fsck programs for ext2/3/4 in /usr/ports/sysutils/e2fsprogs/. Most execellent, thanks! :) Doug -----BEGIN PGP SIGNATURE----- Version: GnuPG v2.0.16 (FreeBSD) iQEcBAEBCAAGBQJMu9naAAoJEFzGhvEaGryEBikH/iNQ7T/73m0LOVCPOiqTPvNK wBliu3zSeCA8fPoodsBFbdchXVXvseJfkl5g6eUEU0Fd2+I3MBVX73ngZrKusMJp ZEy3ar0yTZCHd69jdtJ5yd0UaDLvn5J9jzI2/2L1Mx4ToZg9wobFpkKU4uLwRnQ6 lqZ13zSsLl+sNEPlHj1hG77rDI2OoPegwfpEzSL6ubNTIkH51I1c3zwk6hXau//u SBrTucvknyuP2/oZEX9d0KV+H8KjTWj8MaR8merI5pFT1rle4IpjRPm1EuHAqV0H uiGPqprv3Oblcbc5TIg0uNtz32wyb+dm/w+6rB4/UXqS0S1MOsno+UwhP7kHQfM= =kMXu -----END PGP SIGNATURE----- From owner-freebsd-fs@FreeBSD.ORG Mon Oct 18 10:13:36 2010 Return-Path: Delivered-To: freebsd-fs@FreeBSD.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id 44AC41065672 for ; Mon, 18 Oct 2010 10:13:36 +0000 (UTC) (envelope-from 000.fbsd@quip.cz) Received: from elsa.codelab.cz (elsa.codelab.cz [94.124.105.4]) by mx1.freebsd.org (Postfix) with ESMTP id F41838FC0C for ; Mon, 18 Oct 2010 10:13:35 +0000 (UTC) Received: from elsa.codelab.cz (localhost.codelab.cz [127.0.0.1]) by elsa.codelab.cz (Postfix) with ESMTP id BAF5A19E02E; Mon, 18 Oct 2010 12:13:33 +0200 (CEST) Received: from [10.43.15.14] (cage.codelab.cz [94.124.105.27]) (using TLSv1 with cipher DHE-RSA-AES256-SHA (256/256 bits)) (No client certificate requested) by elsa.codelab.cz (Postfix) with ESMTPSA id 1351A19E02A; Mon, 18 Oct 2010 12:13:31 +0200 (CEST) Message-ID: <4CBC1DAE.4070800@quip.cz> Date: Mon, 18 Oct 2010 12:13:02 +0200 From: Miroslav Lachman <000.fbsd@quip.cz> User-Agent: Mozilla/5.0 (Windows; U; Windows NT 5.0; en-US; rv:1.9.1.13) Gecko/20100914 SeaMonkey/2.0.8 MIME-Version: 1.0 To: Doug Barton References: <4CBBB467.6000601@acm.poly.edu> In-Reply-To: Content-Type: text/plain; charset=ISO-8859-1; format=flowed Content-Transfer-Encoding: 7bit Cc: freebsd-fs@FreeBSD.org Subject: Re: Is there a way to fsck ext[23]fs? X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Mon, 18 Oct 2010 10:13:36 -0000 Doug Barton wrote: > -----BEGIN PGP SIGNED MESSAGE----- > Hash: SHA256 > > > On Sun, 17 Oct 2010, Boris Kochergin wrote: > >> There are fsck programs for ext2/3/4 in /usr/ports/sysutils/e2fsprogs/. > > Most execellent, thanks! Ext2 can be fscked by fsck in base of 8.1: /sbin/fsck_ext2fs Ext[234] with e2fsprogs: /usr/local/sbin/fsck.ext2 /usr/local/sbin/fsck.ext3 /usr/local/sbin/fsck.ext4 /usr/local/sbin/fsck.ext4dev Miroslav Lachman From owner-freebsd-fs@FreeBSD.ORG Mon Oct 18 11:03:22 2010 Return-Path: Delivered-To: freebsd-fs@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id 2F1121065670; Mon, 18 Oct 2010 11:03:22 +0000 (UTC) (envelope-from petefrench@ticketswitch.com) Received: from constantine.ticketswitch.com (constantine.ticketswitch.com [IPv6:2002:57e0:1d4e:1::3]) by mx1.freebsd.org (Postfix) with ESMTP id E88A28FC1E; Mon, 18 Oct 2010 11:03:21 +0000 (UTC) Received: from dilbert.rattatosk ([10.64.50.6] helo=dilbert.ticketswitch.com) by constantine.ticketswitch.com with esmtps (TLSv1:AES256-SHA:256) (Exim 4.69 (FreeBSD)) (envelope-from ) id 1P7nUu-0009Qi-Md; Mon, 18 Oct 2010 12:03:20 +0100 Received: from petefrench by dilbert.ticketswitch.com with local (Exim 4.72 (FreeBSD)) (envelope-from ) id 1P7nUu-000Fy0-LZ; Mon, 18 Oct 2010 12:03:20 +0100 Date: Mon, 18 Oct 2010 12:03:20 +0100 Message-Id: To: fjwcash@gmail.com, pjd@FreeBSD.org In-Reply-To: <20101016222833.GA6765@garage.freebsd.pl> From: Pete French Cc: freebsd-fs@freebsd.org Subject: Re: Converting a non-HAST ZFS pool to a HAST pool X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Mon, 18 Oct 2010 11:03:22 -0000 > HAST metadata takes much more space than glabel metadata. The latter > takes only one sector, while the former depends on provider size, but we > have to keep entire extent bitmap there, so definitely more than one > sector. I am attempting to do the same thing - convert a ZFS pool over to using hast. I have a current setup using gmirror, ggated and a pair of labeled drives. Luckily it seems that the meta data used by gmirror+glabel is biiger than that used by hast - so I should be able to do this without recreating the pool. I only mention it here because I know that the gmirror+ggate is a fairly common setup. If you start with labeled drives then you will be able to migrate to hast without any downtime. -pete. From owner-freebsd-fs@FreeBSD.ORG Mon Oct 18 11:06:57 2010 Return-Path: Delivered-To: freebsd-fs@FreeBSD.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id E12E010656A3 for ; Mon, 18 Oct 2010 11:06:57 +0000 (UTC) (envelope-from owner-bugmaster@FreeBSD.org) Received: from freefall.freebsd.org (freefall.freebsd.org [IPv6:2001:4f8:fff6::28]) by mx1.freebsd.org (Postfix) with ESMTP id B36508FC25 for ; Mon, 18 Oct 2010 11:06:57 +0000 (UTC) Received: from freefall.freebsd.org (localhost [127.0.0.1]) by freefall.freebsd.org (8.14.4/8.14.4) with ESMTP id o9IB6vd5029321 for ; Mon, 18 Oct 2010 11:06:57 GMT (envelope-from owner-bugmaster@FreeBSD.org) Received: (from gnats@localhost) by freefall.freebsd.org (8.14.4/8.14.4/Submit) id o9IB6v16029319 for freebsd-fs@FreeBSD.org; Mon, 18 Oct 2010 11:06:57 GMT (envelope-from owner-bugmaster@FreeBSD.org) Date: Mon, 18 Oct 2010 11:06:57 GMT Message-Id: <201010181106.o9IB6v16029319@freefall.freebsd.org> X-Authentication-Warning: freefall.freebsd.org: gnats set sender to owner-bugmaster@FreeBSD.org using -f From: FreeBSD bugmaster To: freebsd-fs@FreeBSD.org Cc: Subject: Current problem reports assigned to freebsd-fs@FreeBSD.org X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Mon, 18 Oct 2010 11:06:58 -0000 Note: to view an individual PR, use: http://www.freebsd.org/cgi/query-pr.cgi?pr=(number). The following is a listing of current problems submitted by FreeBSD users. These represent problem reports covering all versions including experimental development code and obsolete releases. S Tracker Resp. Description -------------------------------------------------------------------------------- o kern/151111 fs [zfs] vnodes leakage during zfs unmount o kern/151082 fs [zfs] [patch] sappend-flaged files on ZFS not working f kern/150910 fs [nfs] wsize=16384 on udp nfs mount unusable o kern/150796 fs [panic] [suj] [ufs] [softupdates] Panic on portbuild o kern/150503 fs [zfs] ZFS disks are UNAVAIL and corrupted after reboot o kern/150501 fs [zfs] ZFS vdev failure vdev.bad_label on amd64 o kern/150390 fs [zfs] zfs deadlock when arcmsr reports drive faulted o kern/150336 fs [nfs] mountd/nfsd became confused; refused to reload n o kern/150207 fs zpool(1): zpool import -d /dev tries to open weird dev o kern/149855 fs [gvinum] growfs causes fsck to report errors in Filesy o kern/149495 fs [zfs] chflags sappend on zfs not working right o kern/149173 fs [patch] [zfs] make OpenSolaris installa o kern/149022 fs [hang] File system operations hangs with suspfs state o kern/149015 fs [zfs] [patch] misc fixes for ZFS code to build on Glib o kern/149014 fs [zfs] [patch] declarations in ZFS libraries/utilities o kern/149013 fs [zfs] [patch] make ZFS makefiles use the libraries fro o kern/148504 fs [zfs] ZFS' zpool does not allow replacing drives to be o kern/148490 fs [zfs]: zpool attach - resilver bidirectionally, and re o kern/148368 fs [zfs] ZFS hanging forever on 8.1-PRERELEASE o bin/148296 fs [zfs] [loader] [patch] Very slow probe in /usr/src/sys o kern/148204 fs [nfs] UDP NFS causes overload o kern/148138 fs [zfs] zfs raidz pool commands freeze o kern/147903 fs [zfs] [panic] Kernel panics on faulty zfs device o kern/147881 fs [zfs] [patch] ZFS "sharenfs" doesn't allow different " o kern/147790 fs [zfs] zfs set acl(mode|inherit) fails on existing zfs o kern/147560 fs [zfs] [boot] Booting 8.1-PRERELEASE raidz system take o kern/147420 fs [ufs] [panic] ufs_dirbad, nullfs, jail panic (corrupt o kern/147292 fs [nfs] [patch] readahead missing in nfs client options o kern/146941 fs [zfs] [panic] Kernel Double Fault - Happens constantly o kern/146786 fs [zfs] zpool import hangs with checksum errors o kern/146708 fs [ufs] [panic] Kernel panic in softdep_disk_write_compl o kern/146528 fs [zfs] Severe memory leak in ZFS on i386 o kern/146502 fs [nfs] FreeBSD 8 NFS Client Connection to Server o kern/146375 fs [nfs] [patch] Typos in macro variables names in sys/fs s kern/145712 fs [zfs] cannot offline two drives in a raidz2 configurat o kern/145411 fs [xfs] [panic] Kernel panics shortly after mounting an o bin/145309 fs bsdlabel: Editing disk label invalidates the whole dev o kern/145272 fs [zfs] [panic] Panic during boot when accessing zfs on o kern/145246 fs [ufs] dirhash in 7.3 gratuitously frees hashes when it o kern/145238 fs [zfs] [panic] kernel panic on zpool clear tank o kern/145229 fs [zfs] Vast differences in ZFS ARC behavior between 8.0 o kern/145189 fs [nfs] nfsd performs abysmally under load o kern/144929 fs [ufs] [lor] vfs_bio.c + ufs_dirhash.c o kern/144458 fs [nfs] [patch] nfsd fails as a kld p kern/144447 fs [zfs] sharenfs fsunshare() & fsshare_main() non functi o kern/144416 fs [panic] Kernel panic on online filesystem optimization s kern/144415 fs [zfs] [panic] kernel panics on boot after zfs crash o kern/144234 fs [zfs] Cannot boot machine with recent gptzfsboot code o kern/143825 fs [nfs] [panic] Kernel panic on NFS client o bin/143572 fs [zfs] zpool(1): [patch] The verbose output from iostat o kern/143345 fs [ext2fs] [patch] extfs minor header cleanups to better o kern/143212 fs [nfs] NFSv4 client strange work ... o kern/143184 fs [zfs] [lor] zfs/bufwait LOR o kern/142924 fs [ext2fs] [patch] Small cleanup for the inode struct in o kern/142914 fs [zfs] ZFS performance degradation over time o kern/142878 fs [zfs] [vfs] lock order reversal o kern/142597 fs [ext2fs] ext2fs does not work on filesystems with real o kern/142489 fs [zfs] [lor] allproc/zfs LOR o kern/142466 fs Update 7.2 -> 8.0 on Raid 1 ends with screwed raid [re o kern/142401 fs [ntfs] [patch] Minor updates to NTFS from NetBSD o kern/142306 fs [zfs] [panic] ZFS drive (from OSX Leopard) causes two o kern/142068 fs [ufs] BSD labels are got deleted spontaneously o kern/141897 fs [msdosfs] [panic] Kernel panic. msdofs: file name leng o kern/141463 fs [nfs] [panic] Frequent kernel panics after upgrade fro o kern/141305 fs [zfs] FreeBSD ZFS+sendfile severe performance issues ( o kern/141091 fs [patch] [nullfs] fix panics with DIAGNOSTIC enabled o kern/141086 fs [nfs] [panic] panic("nfs: bioread, not dir") on FreeBS o kern/141010 fs [zfs] "zfs scrub" fails when backed by files in UFS2 o kern/140888 fs [zfs] boot fail from zfs root while the pool resilveri o kern/140661 fs [zfs] [patch] /boot/loader fails to work on a GPT/ZFS- o kern/140640 fs [zfs] snapshot crash o kern/140134 fs [msdosfs] write and fsck destroy filesystem integrity o kern/140068 fs [smbfs] [patch] smbfs does not allow semicolon in file o kern/139725 fs [zfs] zdb(1) dumps core on i386 when examining zpool c o kern/139715 fs [zfs] vfs.numvnodes leak on busy zfs o bin/139651 fs [nfs] mount(8): read-only remount of NFS volume does n o kern/139597 fs [patch] [tmpfs] tmpfs initializes va_gen but doesn't u o kern/139564 fs [zfs] [panic] 8.0-RC1 - Fatal trap 12 at end of shutdo o kern/139407 fs [smbfs] [panic] smb mount causes system crash if remot o kern/139363 fs [nfs] diskless root nfs mount from non FreeBSD server o kern/138790 fs [zfs] ZFS ceases caching when mem demand is high o kern/138662 fs [panic] ffs_blkfree: freeing free block o kern/138421 fs [ufs] [patch] remove UFS label limitations o kern/138202 fs mount_msdosfs(1) see only 2Gb o kern/136968 fs [ufs] [lor] ufs/bufwait/ufs (open) o kern/136945 fs [ufs] [lor] filedesc structure/ufs (poll) o kern/136944 fs [ffs] [lor] bufwait/snaplk (fsync) o kern/136873 fs [ntfs] Missing directories/files on NTFS volume o kern/136865 fs [nfs] [patch] NFS exports atomic and on-the-fly atomic o kern/136470 fs [nfs] Cannot mount / in read-only, over NFS o kern/135667 fs [lor] LORs causing ufs filesystem corruption on XEN Do o kern/135546 fs [zfs] zfs.ko module doesn't ignore zpool.cache filenam o kern/135469 fs [ufs] [panic] kernel crash on md operation in ufs_dirb o kern/135050 fs [zfs] ZFS clears/hides disk errors on reboot o kern/134491 fs [zfs] Hot spares are rather cold... o kern/133676 fs [smbfs] [panic] umount -f'ing a vnode-based memory dis o kern/133614 fs [panic] panic: ffs_truncate: read-only filesystem o kern/133174 fs [msdosfs] [patch] msdosfs must support utf-encoded int o kern/132960 fs [ufs] [panic] panic:ffs_blkfree: freeing free frag o kern/132397 fs reboot causes filesystem corruption (failure to sync b o kern/132331 fs [ufs] [lor] LOR ufs and syncer o kern/132237 fs [msdosfs] msdosfs has problems to read MSDOS Floppy o kern/132145 fs [panic] File System Hard Crashes o kern/131441 fs [unionfs] [nullfs] unionfs and/or nullfs not combineab o kern/131360 fs [nfs] poor scaling behavior of the NFS server under lo o kern/131342 fs [nfs] mounting/unmounting of disks causes NFS to fail o bin/131341 fs makefs: error "Bad file descriptor" on the mount poin o kern/130920 fs [msdosfs] cp(1) takes 100% CPU time while copying file o kern/130210 fs [nullfs] Error by check nullfs o kern/129760 fs [nfs] after 'umount -f' of a stale NFS share FreeBSD l o kern/129488 fs [smbfs] Kernel "bug" when using smbfs in smbfs_smb.c: o kern/129231 fs [ufs] [patch] New UFS mount (norandom) option - mostly o kern/129152 fs [panic] non-userfriendly panic when trying to mount(8) o kern/129059 fs [zfs] [patch] ZFS bootloader whitelistable via WITHOUT f kern/128829 fs smbd(8) causes periodic panic on 7-RELEASE o kern/127787 fs [lor] [ufs] Three LORs: vfslock/devfs/vfslock, ufs/vfs o bin/127270 fs fsck_msdosfs(8) may crash if BytesPerSec is zero o kern/127029 fs [panic] mount(8): trying to mount a write protected zi o kern/126287 fs [ufs] [panic] Kernel panics while mounting an UFS file o kern/125895 fs [ffs] [panic] kernel: panic: ffs_blkfree: freeing free s kern/125738 fs [zfs] [request] SHA256 acceleration in ZFS p kern/124621 fs [ext3] [patch] Cannot mount ext2fs partition f bin/124424 fs [zfs] zfs(8): zfs list -r shows strange snapshots' siz o kern/123939 fs [msdosfs] corrupts new files o kern/122380 fs [ffs] ffs_valloc:dup alloc (Soekris 4801/7.0/USB Flash o bin/122172 fs [fs]: amd(8) automount daemon dies on 6.3-STABLE i386, o bin/121898 fs [nullfs] pwd(1)/getcwd(2) fails with Permission denied o bin/121779 fs [ufs] snapinfo(8) (and related tools?) only work for t o bin/121366 fs [zfs] [patch] Automatic disk scrubbing from periodic(8 o bin/121072 fs [smbfs] mount_smbfs(8) cannot normally convert the cha f kern/120991 fs [panic] [ffs] [snapshot] System crashes when manipulat o kern/120483 fs [ntfs] [patch] NTFS filesystem locking changes o kern/120482 fs [ntfs] [patch] Sync style changes between NetBSD and F f kern/119735 fs [zfs] geli + ZFS + samba starting on boot panics 7.0-B o kern/118912 fs [2tb] disk sizing/geometry problem with large array o kern/118713 fs [minidump] [patch] Display media size required for a k o bin/118249 fs mv(1): moving a directory changes its mtime o kern/118107 fs [ntfs] [panic] Kernel panic when accessing a file at N o kern/117954 fs [ufs] dirhash on very large directories blocks the mac o bin/117315 fs [smbfs] mount_smbfs(8) and related options can't mount o kern/117314 fs [ntfs] Long-filename only NTFS fs'es cause kernel pani o kern/117158 fs [zfs] zpool scrub causes panic if geli vdevs detach on o bin/116980 fs [msdosfs] [patch] mount_msdosfs(8) resets some flags f o conf/116931 fs lack of fsck_cd9660 prevents mounting iso images with p kern/116608 fs [msdosfs] [patch] msdosfs fails to check mount options o kern/116583 fs [ffs] [hang] System freezes for short time when using o kern/116170 fs [panic] Kernel panic when mounting /tmp f kern/115645 fs [ffs] [snapshots] [panic] lockmgr: thread 0xc4c00d80, o bin/115361 fs [zfs] mount(8) gets into a state where it won't set/un o kern/114955 fs [cd9660] [patch] [request] support for mask,dirmask,ui o kern/114847 fs [ntfs] [patch] [request] dirmask support for NTFS ala o kern/114676 fs [ufs] snapshot creation panics: snapacct_ufs2: bad blo o bin/114468 fs [patch] [request] add -d option to umount(8) to detach o kern/113852 fs [smbfs] smbfs does not properly implement DFS referral o bin/113838 fs [patch] [request] mount(8): add support for relative p o bin/113049 fs [patch] [request] make quot(8) use getopt(3) and show o kern/112658 fs [smbfs] [patch] smbfs and caching problems (resolves b o kern/111843 fs [msdosfs] Long Names of files are incorrectly created o kern/111782 fs [ufs] dump(8) fails horribly for large filesystems s bin/111146 fs [2tb] fsck(8) fails on 6T filesystem o kern/109024 fs [msdosfs] [iconv] mount_msdosfs: msdosfs_iconv: Operat o kern/109010 fs [msdosfs] can't mv directory within fat32 file system o bin/107829 fs [2TB] fdisk(8): invalid boundary checking in fdisk / w o kern/106107 fs [ufs] left-over fsck_snapshot after unfinished backgro o kern/106030 fs [ufs] [panic] panic in ufs from geom when a dead disk o kern/104406 fs [ufs] Processes get stuck in "ufs" state under persist o kern/104133 fs [ext2fs] EXT2FS module corrupts EXT2/3 filesystems o kern/103035 fs [ntfs] Directories in NTFS mounted disc images appear o kern/101324 fs [smbfs] smbfs sometimes not case sensitive when it's s o kern/99290 fs [ntfs] mount_ntfs ignorant of cluster sizes s bin/97498 fs [request] newfs(8) has no option to clear the first 12 o kern/97377 fs [ntfs] [patch] syntax cleanup for ntfs_ihash.c o kern/95222 fs [iso9660] File sections on ISO9660 level 3 CDs ignored o kern/94849 fs [ufs] rename on UFS filesystem is not atomic o bin/94810 fs fsck(8) incorrectly reports 'file system marked clean' o kern/94769 fs [ufs] Multiple file deletions on multi-snapshotted fil o kern/94733 fs [smbfs] smbfs may cause double unlock o bin/94635 fs snapinfo(8)/libufs only works for disk-backed filesyst o kern/93942 fs [vfs] [patch] panic: ufs_dirbad: bad dir (patch from D o kern/92272 fs [ffs] [hang] Filling a filesystem while creating a sna f kern/91568 fs [ufs] [panic] writing to UFS/softupdates DVD media in o kern/91134 fs [smbfs] [patch] Preserve access and modification time a kern/90815 fs [smbfs] [patch] SMBFS with character conversions somet o kern/88657 fs [smbfs] windows client hang when browsing a samba shar o kern/88555 fs [panic] ffs_blkfree: freeing free frag on AMD 64 o kern/88266 fs [smbfs] smbfs does not implement UIO_NOCOPY and sendfi o bin/87966 fs [patch] newfs(8): introduce -A flag for newfs to enabl o kern/87859 fs [smbfs] System reboot while umount smbfs. o kern/86587 fs [msdosfs] rm -r /PATH fails with lots of small files o bin/85494 fs fsck_ffs: unchecked use of cg_inosused macro etc. o kern/85326 fs [smbfs] [panic] saving a file via samba to an overquot o kern/84589 fs [2TB] 5.4-STABLE unresponsive during background fsck 2 o kern/80088 fs [smbfs] Incorrect file time setting on NTFS mounted vi o bin/74779 fs Background-fsck checks one filesystem twice and omits o kern/73484 fs [ntfs] Kernel panic when doing `ls` from the client si o bin/73019 fs [ufs] fsck_ufs(8) cannot alloc 607016868 bytes for ino o kern/71774 fs [ntfs] NTFS cannot "see" files on a WinXP filesystem o bin/70600 fs fsck(8) throws files away when it can't grow lost+foun o kern/68978 fs [panic] [ufs] crashes with failing hard disk, loose po o kern/65920 fs [nwfs] Mounted Netware filesystem behaves strange o kern/65901 fs [smbfs] [patch] smbfs fails fsx write/truncate-down/tr o kern/61503 fs [smbfs] mount_smbfs does not work as non-root o kern/55617 fs [smbfs] Accessing an nsmb-mounted drive via a smb expo o kern/51685 fs [hang] Unbounded inode allocation causes kernel to loc o kern/51583 fs [nullfs] [patch] allow to work with devices and socket o kern/36566 fs [smbfs] System reboot with dead smb mount and umount o kern/33464 fs [ufs] soft update inconsistencies after system crash o bin/27687 fs fsck(8) wrapper is not properly passing options to fsc o kern/18874 fs [2TB] 32bit NFS servers export wrong negative values t 209 problems total. From owner-freebsd-fs@FreeBSD.ORG Mon Oct 18 12:30:14 2010 Return-Path: Delivered-To: freebsd-fs@hub.freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id 582111065675 for ; Mon, 18 Oct 2010 12:30:14 +0000 (UTC) (envelope-from gnats@FreeBSD.org) Received: from freefall.freebsd.org (freefall.freebsd.org [IPv6:2001:4f8:fff6::28]) by mx1.freebsd.org (Postfix) with ESMTP id 46B428FC0C for ; Mon, 18 Oct 2010 12:30:14 +0000 (UTC) Received: from freefall.freebsd.org (localhost [127.0.0.1]) by freefall.freebsd.org (8.14.4/8.14.4) with ESMTP id o9ICUE6B016478 for ; Mon, 18 Oct 2010 12:30:14 GMT (envelope-from gnats@freefall.freebsd.org) Received: (from gnats@localhost) by freefall.freebsd.org (8.14.4/8.14.4/Submit) id o9ICUEAi016474; Mon, 18 Oct 2010 12:30:14 GMT (envelope-from gnats) Date: Mon, 18 Oct 2010 12:30:14 GMT Message-Id: <201010181230.o9ICUEAi016474@freefall.freebsd.org> To: freebsd-fs@FreeBSD.org From: Gleb Smirnoff Cc: Subject: Re: misc/151111: vnodes leakage during zfs unmount X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list Reply-To: Gleb Smirnoff List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Mon, 18 Oct 2010 12:30:14 -0000 The following reply was made to PR kern/151111; it has been noted by GNATS. From: Gleb Smirnoff To: "Oleg A. Mamontov" Cc: kib@FreeBSD.org, pjd@FreeBSD.org, freebsd-gnats-submit@FreeBSD.org Subject: Re: misc/151111: vnodes leakage during zfs unmount Date: Mon, 18 Oct 2010 16:11:55 +0400 On Thu, Sep 30, 2010 at 11:56:18AM +0000, Oleg A. Mamontov wrote: O> for i in `jot 1000`; do zfs mount tank; printf '%4d ' $i; sysctl vfs.numvnodes; zfs umount tank; done O> ######################## O> 1 vfs.numvnodes: 708 O> 2 vfs.numvnodes: 709 O> 3 vfs.numvnodes: 710 O> 4 vfs.numvnodes: 711 O> 5 vfs.numvnodes: 712 O> O> .. O> O> 995 vfs.numvnodes: 1702 O> 996 vfs.numvnodes: 1703 O> 997 vfs.numvnodes: 1704 O> 998 vfs.numvnodes: 1705 O> 999 vfs.numvnodes: 1706 O> 1000 vfs.numvnodes: 1707 Here is my lame investigation of the problem. In the zfs_domount() function we've got the following code: /* Grab extra reference. */ VERIFY(VFS_ROOT(vfsp, LK_EXCLUSIVE, &vp) == 0); VOP_UNLOCK(vp, 0); I suppose this code is expected to put an extra reference on the vfsp->mnt_vnodecovered vnode. Do I mistake here? If I don't then this is the source of leak. Debugging shows that zfs_zget(), called subsequently from zfs_root(), does not find an existing znode/vnode and allocates a new one, see at the end of zfs_zget() function. This vnode gots a reference and is forgotten. Sorry, if I am being mistaken :) -- Totus tuus, Glebius. From owner-freebsd-fs@FreeBSD.ORG Mon Oct 18 14:29:15 2010 Return-Path: Delivered-To: freebsd-fs@hub.freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id 1252A1065679; Mon, 18 Oct 2010 14:29:15 +0000 (UTC) (envelope-from jhb@FreeBSD.org) Received: from freefall.freebsd.org (freefall.freebsd.org [IPv6:2001:4f8:fff6::28]) by mx1.freebsd.org (Postfix) with ESMTP id DB7BF8FC13; Mon, 18 Oct 2010 14:29:14 +0000 (UTC) Received: from freefall.freebsd.org (localhost [127.0.0.1]) by freefall.freebsd.org (8.14.4/8.14.4) with ESMTP id o9IETEUx041602; Mon, 18 Oct 2010 14:29:14 GMT (envelope-from jhb@freefall.freebsd.org) Received: (from jhb@localhost) by freefall.freebsd.org (8.14.4/8.14.4/Submit) id o9IETEYu041598; Mon, 18 Oct 2010 14:29:14 GMT (envelope-from jhb) Date: Mon, 18 Oct 2010 14:29:14 GMT Message-Id: <201010181429.o9IETEYu041598@freefall.freebsd.org> To: gcooper@FreeBSD.org, jhb@FreeBSD.org, freebsd-fs@FreeBSD.org From: jhb@FreeBSD.org Cc: Subject: Re: kern/129059: [zfs] [patch] ZFS bootloader whitelistable via WITHOUT_CDDL X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Mon, 18 Oct 2010 14:29:15 -0000 Synopsis: [zfs] [patch] ZFS bootloader whitelistable via WITHOUT_CDDL State-Changed-From-To: open->closed State-Changed-By: jhb State-Changed-When: Mon Oct 18 14:28:43 UTC 2010 State-Changed-Why: Fix merged to 7 and 8 a while back. http://www.freebsd.org/cgi/query-pr.cgi?pr=129059 From owner-freebsd-fs@FreeBSD.ORG Mon Oct 18 17:54:10 2010 Return-Path: Delivered-To: freebsd-fs@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id E1FFE106564A for ; Mon, 18 Oct 2010 17:54:10 +0000 (UTC) (envelope-from jerk_alert@hotmail.com) Received: from snt0-omc4-s18.snt0.hotmail.com (snt0-omc4-s18.snt0.hotmail.com [65.55.90.221]) by mx1.freebsd.org (Postfix) with ESMTP id B797E8FC0C for ; Mon, 18 Oct 2010 17:54:10 +0000 (UTC) Received: from SNT137-W42 ([65.55.90.199]) by snt0-omc4-s18.snt0.hotmail.com with Microsoft SMTPSVC(6.0.3790.4675); Mon, 18 Oct 2010 10:42:10 -0700 Message-ID: X-Originating-IP: [67.161.61.253] From: Bonnie Guerelli To: Date: Mon, 18 Oct 2010 17:42:09 +0000 Importance: Normal MIME-Version: 1.0 X-OriginalArrivalTime: 18 Oct 2010 17:42:10.0025 (UTC) FILETIME=[CA479590:01CB6EEB] Content-Type: text/plain; charset="iso-8859-1" Content-Transfer-Encoding: quoted-printable X-Content-Filtered-By: Mailman/MimeDel 2.1.5 Subject: Fatal trap 12 page fault error on 8.0, 8.1, 8-STABLE 64bit from ZFS RAIDZ2 volume X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Mon, 18 Oct 2010 17:54:11 -0000 Hi all=2C I posted quite a bit about the problems I've been having with a RAIDZ2 volu= me at http://forums.freebsd.org/showthread.php?p=3D105831 but am still nowh= ere towards getting this volume back online. It was created with 8.0 RELEAS= E 64bit and ZFS v13 on a QC6600 system with 8GB ram. In a nutshell=2C when attempting to run import=2C export=2C mount=2C scrub= =2C etc.=2C I receive a 'Fatal trap 12: page fault while in kernel mode' er= ror. This comes typically within a few seconds of issuing any of these comm= ands=2C but sometimes scrub has gone approx 6 hours=2C almost to completio= n. When running zdb=2C I've gotten 'Assertion failed' errors. Any help would be much obliged=2C Thanks=2C -bg = From owner-freebsd-fs@FreeBSD.ORG Mon Oct 18 22:26:50 2010 Return-Path: Delivered-To: freebsd-fs@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id E5B221065670 for ; Mon, 18 Oct 2010 22:26:50 +0000 (UTC) (envelope-from pjd@garage.freebsd.pl) Received: from mail.garage.freebsd.pl (60.wheelsystems.com [83.12.187.60]) by mx1.freebsd.org (Postfix) with ESMTP id 900DF8FC0A for ; Mon, 18 Oct 2010 22:26:49 +0000 (UTC) Received: by mail.garage.freebsd.pl (Postfix, from userid 65534) id AD79645C98; Tue, 19 Oct 2010 00:26:47 +0200 (CEST) Received: from localhost (chello089073192049.chello.pl [89.73.192.49]) (using TLSv1 with cipher DHE-RSA-AES256-SHA (256/256 bits)) (No client certificate requested) by mail.garage.freebsd.pl (Postfix) with ESMTP id 2FFCC45C9B; Tue, 19 Oct 2010 00:26:42 +0200 (CEST) Date: Tue, 19 Oct 2010 00:26:11 +0200 From: Pawel Jakub Dawidek To: Pete French Message-ID: <20101018222611.GC2375@garage.freebsd.pl> References: <20101016222833.GA6765@garage.freebsd.pl> Mime-Version: 1.0 Content-Type: multipart/signed; micalg=pgp-sha1; protocol="application/pgp-signature"; boundary="IpbVkmxF4tDyP/Kb" Content-Disposition: inline In-Reply-To: User-Agent: Mutt/1.4.2.3i X-PGP-Key-URL: http://people.freebsd.org/~pjd/pjd.asc X-OS: FreeBSD 9.0-CURRENT amd64 X-Spam-Checker-Version: SpamAssassin 3.0.4 (2005-06-05) on mail.garage.freebsd.pl X-Spam-Level: X-Spam-Status: No, score=-0.6 required=4.5 tests=BAYES_00,RCVD_IN_SORBS_DUL autolearn=no version=3.0.4 Cc: freebsd-fs@freebsd.org Subject: Re: Converting a non-HAST ZFS pool to a HAST pool X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Mon, 18 Oct 2010 22:26:51 -0000 --IpbVkmxF4tDyP/Kb Content-Type: text/plain; charset=us-ascii Content-Disposition: inline Content-Transfer-Encoding: quoted-printable On Mon, Oct 18, 2010 at 12:03:20PM +0100, Pete French wrote: > > HAST metadata takes much more space than glabel metadata. The latter > > takes only one sector, while the former depends on provider size, but we > > have to keep entire extent bitmap there, so definitely more than one > > sector. >=20 > I am attempting to do the same thing - convert a ZFS pool over to using > hast. I have a current setup using gmirror, ggated and a pair of labeled > drives. Luckily it seems that the meta data used by gmirror+glabel is bii= ger > than that used by hast - so I should be able to do this without recreating > the pool. Maybe I'm missing something, but how do you guys check this? Both gmirror and glabel use only single sector for its metadata, so 1kB in total. HAST metadata takes 4kB plus activemap size, which depends on provider size. --=20 Pawel Jakub Dawidek http://www.wheelsystems.com pjd@FreeBSD.org http://www.FreeBSD.org FreeBSD committer Am I Evil? Yes, I Am! --IpbVkmxF4tDyP/Kb Content-Type: application/pgp-signature Content-Disposition: inline -----BEGIN PGP SIGNATURE----- Version: GnuPG v2.0.14 (FreeBSD) iEYEARECAAYFAky8yYIACgkQForvXbEpPzR6oQCfYqUPm/GLrBeFlLt4Kswh3/87 gqQAnj7+ez3YYEK/zHqWCwxS4cojkl9U =QEin -----END PGP SIGNATURE----- --IpbVkmxF4tDyP/Kb-- From owner-freebsd-fs@FreeBSD.ORG Tue Oct 19 04:53:42 2010 Return-Path: Delivered-To: freebsd-fs@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id 2DCFF10656CD; Tue, 19 Oct 2010 04:53:42 +0000 (UTC) (envelope-from zbeeble@gmail.com) Received: from mail-gw0-f54.google.com (mail-gw0-f54.google.com [74.125.83.54]) by mx1.freebsd.org (Postfix) with ESMTP id CA3CC8FC08; Tue, 19 Oct 2010 04:53:40 +0000 (UTC) Received: by gwb20 with SMTP id 20so948780gwb.13 for ; Mon, 18 Oct 2010 21:53:40 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=gamma; h=domainkey-signature:mime-version:received:received:in-reply-to :references:date:message-id:subject:from:to:cc:content-type; bh=TfuOzPBpAQdwER/m5XETh3VMIxD0CUt5yAgd0b5y4Ic=; b=QdTT8HwvxrCQCXI033FUqbCCzTozgBAXzoQ3q83vCikwAsw4vp3SDBiaSim/m0ver2 SRcZqfvROi68Xr6duZolIhe6Jd2hn8mOCNpGkXbR2ESf5e5M/L2pt+5CQXAFyci5zXEG flr0ezWukG+FPGVve5d0zfeFoEqKl61HjNk+c= DomainKey-Signature: a=rsa-sha1; c=nofws; d=gmail.com; s=gamma; h=mime-version:in-reply-to:references:date:message-id:subject:from:to :cc:content-type; b=NlvZP0i1EYTyVyf9KR+V5cvH9rpMtOvWNqWHfxBDzvivBHH/SMih4IKOI/H8UzaEtr RlSProYRy/ee+CYYERvSsHw166BHsaTUF7xUqMEMthSWNrFsN7Q+vo2ughMfmtvA0fH5 AMZgL42iwaV1Rko8bHSY+Gj/kJ0y91LBubPtQ= MIME-Version: 1.0 Received: by 10.150.52.18 with SMTP id z18mr8279445ybz.189.1287464018301; Mon, 18 Oct 2010 21:53:38 -0700 (PDT) Received: by 10.151.82.1 with HTTP; Mon, 18 Oct 2010 21:53:38 -0700 (PDT) In-Reply-To: <20101018222611.GC2375@garage.freebsd.pl> References: <20101016222833.GA6765@garage.freebsd.pl> <20101018222611.GC2375@garage.freebsd.pl> Date: Tue, 19 Oct 2010 00:53:38 -0400 Message-ID: From: Zaphod Beeblebrox To: Pawel Jakub Dawidek Content-Type: text/plain; charset=ISO-8859-1 Cc: freebsd-fs@freebsd.org Subject: Re: Converting a non-HAST ZFS pool to a HAST pool X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Tue, 19 Oct 2010 04:53:42 -0000 I'm wondering if I'm missing something here --- because I'm wondering if running HAST under ZFS isn't a step backwards. My quick read of HAST seems to indicate that it's going to manage two disks and present them as one disk to ZFS. The design problem with this (especially since we're talking a _lot_ of network (and memory) transfers involved) is data corruption --- the idea that ZFS protects data better when it can determine one disk has it right while another disk has it wrong (as it can when it manages the two disks). Wouldn't it be better to just have network (iscsi-like) spools attached to ZFS? Individual spools could still fail. What am I missing? Is there a better description of HAST than the FreeBSD wiki page? From owner-freebsd-fs@FreeBSD.ORG Tue Oct 19 08:12:43 2010 Return-Path: Delivered-To: fs@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id 140AA1065670 for ; Tue, 19 Oct 2010 08:12:43 +0000 (UTC) (envelope-from wjw@digiware.nl) Received: from mail.digiware.nl (mail.ip6.digiware.nl [IPv6:2001:4cb8:1:106::2]) by mx1.freebsd.org (Postfix) with ESMTP id 9FBE98FC08 for ; Tue, 19 Oct 2010 08:12:42 +0000 (UTC) Received: from localhost (localhost.digiware.nl [127.0.0.1]) by mail.digiware.nl (Postfix) with ESMTP id 1D4D8153434 for ; Tue, 19 Oct 2010 10:12:41 +0200 (CEST) X-Virus-Scanned: amavisd-new at digiware.nl Received: from mail.digiware.nl ([127.0.0.1]) by localhost (rack1.digiware.nl [127.0.0.1]) (amavisd-new, port 10024) with ESMTP id iyDqCQ1qniy8 for ; Tue, 19 Oct 2010 10:12:39 +0200 (CEST) Received: from [127.0.0.1] (opteron [192.168.10.67]) (using TLSv1 with cipher DHE-RSA-CAMELLIA256-SHA (256/256 bits)) (No client certificate requested) by mail.digiware.nl (Postfix) with ESMTPSA id 134CB153433 for ; Tue, 19 Oct 2010 10:12:39 +0200 (CEST) Message-ID: <4CBD52F4.2010105@digiware.nl> Date: Tue, 19 Oct 2010 10:12:36 +0200 From: Willem Jan Withagen User-Agent: Mozilla/5.0 (Windows; U; Windows NT 6.1; en-US; rv:1.9.2.9) Gecko/20100915 Lightning/1.0b2 Thunderbird/3.1.4 MIME-Version: 1.0 To: fs@freebsd.org Content-Type: text/plain; charset=ISO-8859-1; format=flowed Content-Transfer-Encoding: 7bit Cc: Subject: ZFS destroy snapshot does not work X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Tue, 19 Oct 2010 08:12:43 -0000 Hi, Probably due to too many reboots when the system was livelocked. But now I have the following problem. /sbin/zfs destroy -r zfsraid/backups@Monday cannot destroy 'zfsraid/backups@Monday': dataset does not exist no snapshots destroyed /sbin/zfs snapshot -r zfsraid/backups@Monday cannot create snapshot 'zfsraid/backups@Monday': dataset is busy no snapshots were created /sbin/zfs destroy -r zfsraid/home@$DATE cannot destroy 'zfsraid/home@Monday': dataset does not exist no snapshots destroyed /sbin/zfs snapshot -r zfsraid/home@Monday cannot create snapshot 'zfsraid/home@Monday': dataset is busy no snapshots were created Where this used to work(tm).... I can iterate over all filesystems and then destroy/create a backup. Except for one: # zfs list -r zfsraid/home/trouble NAME USED AVAIL REFER MOUNTPOINT zfsraid/home/trouble 149G 3.49T 148G none zfsraid/home/trouble@Sunday 1.06G - 148G - # zfs destroy -r zfsraid/home/trouble cannot destroy 'zfsraid/home/trouble@Sunday': dataset already exists Exit 1 # zfs destroy zfsraid/home/trouble@Sunday cannot destroy 'zfsraid/home/trouble@Sunday': dataset already exists Exit 1 # zfs destroy -r zfsraid/home/trouble@Sunday cannot destroy 'zfsraid/home/trouble@Sunday': snapshot is cloned no snapshots destroyed But of this last fact I do not remember cloning anything. Nor can I find any suggestions as to the filesystem being cloned..... So how do I get ride of this rouge filesystem. There is something like zdb, but the manual page is purposely vague on how to use that... Thanx, --WjW From owner-freebsd-fs@FreeBSD.ORG Tue Oct 19 11:49:16 2010 Return-Path: Delivered-To: freebsd-fs@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id CB692106564A; Tue, 19 Oct 2010 11:49:16 +0000 (UTC) (envelope-from petefrench@ticketswitch.com) Received: from constantine.ticketswitch.com (constantine.ticketswitch.com [IPv6:2002:57e0:1d4e:1::3]) by mx1.freebsd.org (Postfix) with ESMTP id 8DA0D8FC0C; Tue, 19 Oct 2010 11:49:16 +0000 (UTC) Received: from dilbert.rattatosk ([10.64.50.6] helo=dilbert.ticketswitch.com) by constantine.ticketswitch.com with esmtps (TLSv1:AES256-SHA:256) (Exim 4.69 (FreeBSD)) (envelope-from ) id 1P8Agt-000Hh1-7d; Tue, 19 Oct 2010 12:49:15 +0100 Received: from petefrench by dilbert.ticketswitch.com with local (Exim 4.72 (FreeBSD)) (envelope-from ) id 1P8Agt-0001Ob-6I; Tue, 19 Oct 2010 12:49:15 +0100 Date: Tue, 19 Oct 2010 12:49:15 +0100 Message-Id: To: pjd@FreeBSD.org In-Reply-To: <20101018222611.GC2375@garage.freebsd.pl> From: Pete French Cc: freebsd-fs@freebsd.org Subject: Re: Converting a non-HAST ZFS pool to a HAST pool X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Tue, 19 Oct 2010 11:49:16 -0000 > Maybe I'm missing something, but how do you guys check this? Both > gmirror and glabel use only single sector for its metadata, so 1kB in > total. HAST metadata takes 4kB plus activemap size, which depends on > provider size. I checked it by disconnecting one half of the mirror - which is on a separate machine in my case remember - and then setting that machine up with hast. Created the hast device and did a 'diskinfo' on the resulting device. Compared that to a 'diskinfo' on the original provider. Here are the results, for the bare disk, the hast device, the mirrored device, and partition 'a' on the mirrored device. I wrote a disklabel there when it was UFSA, in case you were wondering. diskinfo /dev/da0s1 /dev/da0s1 512 68718411776 134215648 0 16384 16447 255 32 diskinfo /dev/hast/serp0 /dev/hast/serp0 512 68718403584 134215632 0 0 diskinfo /dev/mirror/mysql0 /dev/mirror/mysql0 512 68718410752 134215646 0 0 diskinfo /dev/mirror/mysql0a /dev/mirror/mysql0a 512 68718402560 134215630 0 8192 8354 255 63 So my current zpool uses /dev/mirror/mysql0a which is 68718402560 bytes, and when I move to hast I will be using /dev/hast/serp0 which is 68718403584 bytes - i.e. 1024 bytes larger. Note that if it wasn't for the BSD style disklabel it would be larger, and wouldnt work. -pete. From owner-freebsd-fs@FreeBSD.ORG Tue Oct 19 11:54:04 2010 Return-Path: Delivered-To: freebsd-fs@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id AA011106564A; Tue, 19 Oct 2010 11:54:04 +0000 (UTC) (envelope-from petefrench@ticketswitch.com) Received: from constantine.ticketswitch.com (constantine.ticketswitch.com [IPv6:2002:57e0:1d4e:1::3]) by mx1.freebsd.org (Postfix) with ESMTP id 6DF278FC1F; Tue, 19 Oct 2010 11:54:04 +0000 (UTC) Received: from dilbert.rattatosk ([10.64.50.6] helo=dilbert.ticketswitch.com) by constantine.ticketswitch.com with esmtps (TLSv1:AES256-SHA:256) (Exim 4.69 (FreeBSD)) (envelope-from ) id 1P8AlX-000Hid-LF; Tue, 19 Oct 2010 12:54:03 +0100 Received: from petefrench by dilbert.ticketswitch.com with local (Exim 4.72 (FreeBSD)) (envelope-from ) id 1P8AlX-0001P4-KI; Tue, 19 Oct 2010 12:54:03 +0100 Date: Tue, 19 Oct 2010 12:54:03 +0100 Message-Id: To: pjd@freebsd.org, zbeeble@gmail.com In-Reply-To: From: Pete French Cc: freebsd-fs@freebsd.org Subject: Re: Converting a non-HAST ZFS pool to a HAST pool X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Tue, 19 Oct 2010 11:54:04 -0000 > I'm wondering if I'm missing something here --- because I'm wondering > if running HAST under ZFS isn't a step backwards. > > My quick read of HAST seems to indicate that it's going to manage two > disks and present them as one disk to ZFS. The design problem with > this (especially since we're talking a _lot_ of network (and memory) > transfers involved) is data corruption --- the idea that ZFS protects > data better when it can determine one disk has it right while another > disk has it wrong (as it can when it manages the two disks). > > Wouldn't it be better to just have network (iscsi-like) spools > attached to ZFS? Individual spools could still fail. What am I > missing? Is there a better description of HAST than the FreeBSD wiki > page? I guess in theory the answer there is 'yes' - but have you actually tried it in practice ? I did this for a while as an experiment using ggated - the problem is that when the remote fails (for example) it doesnt signal up to ZFS properly and instead of ZFS seeing a failed driev it just locks up. iSCSI had similar issues. Note that this test was a while ago, and the situation may have improved - but that was the way I did it at the time I set this up. Interstingly, I am considering a hybrid - using ZFS as a mirror over a pair of HAST devices. My servers have hardwar RAID in them and a pair of drives each, so I could split those, remove the hardware RAID, and run 4 drives as 2 hast devices with ZFS on top. That soulds like the best soultuon to me - but you do then have the issue of doubling your network bandwidth required. -pete. From owner-freebsd-fs@FreeBSD.ORG Tue Oct 19 14:52:54 2010 Return-Path: Delivered-To: freebsd-fs@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id 8E0121065675 for ; Tue, 19 Oct 2010 14:52:54 +0000 (UTC) (envelope-from kpielorz_lst@tdx.co.uk) Received: from mail.tdx.com (mail.tdx.com [62.13.128.18]) by mx1.freebsd.org (Postfix) with ESMTP id 0B0988FC15 for ; Tue, 19 Oct 2010 14:52:53 +0000 (UTC) Received: from HexaDeca64.dmpriest.net.uk (HPQuadro64.dmpriest.net.uk [62.13.130.30]) (authenticated bits=0) by mail.tdx.com (8.14.3/8.14.3/Kp) with ESMTP id o9JEqqqs016620 (version=TLSv1/SSLv3 cipher=DHE-DSS-AES256-SHA bits=256 verify=NO) for ; Tue, 19 Oct 2010 15:52:52 +0100 (BST) Date: Tue, 19 Oct 2010 15:52:52 +0100 From: Karl Pielorz To: freebsd-fs@freebsd.org Message-ID: X-Mailer: Mulberry/4.0.8 (Win32) MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii; format=flowed Content-Transfer-Encoding: 7bit Content-Disposition: inline Subject: ZFS 'read-only' device / pool scan / import? X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Tue, 19 Oct 2010 14:52:54 -0000 Hi, On FreeBSD if I bring the system up 'single user' - the first time I do (for example): " zpool status " There's a pause, a flurry of disk activity - and the system appears to import any pools it finds, and the status appears. I'd guess at this point - some data is written to the disks? - Is there any way of avoiding that, i.e. a kind of "If you were to import the pools/display a status, what would you show, without actually writing any data?" (or importing it) - i.e. to ensure the devices are opened read only? The reason I ask is we recently had a pool that had a mishap. The backing RAID controller dropped a drive (we were using them in JBOD). This happened cleanly. The system got shutdown, and I think the wrong drive was replaced. When it came up the RAID controller 'collapsed' the device list (so there was no gap where the old drive was) - that, plus the wrong drive being replaced meant when we looked at the system we had: " pool: vol state: UNAVAIL status: One or more devices could not be used because the label is missing or invalid. There are insufficient replicas for the pool to continue functioning. " A number of devices were listed as 'corrupted data' - some devices were listed twice as members of the pool - i.e. pretty screwed up. 'undoing' the damage and restarting the server - just threw up the same status. I'm wondering if through the action of having the pool imported/mounted etc. - ZFS has actually *written* to the drives that were available that other drives aren't available / corrupt - and basically, because that info was written, and check-summed correctly - it takes that as gospel now, rather than actually 're-checking' the drives (or is simply unable to re-check the drives - because the meta data has been changed from the previous boot). If you see what I mean :) In the end, we're fortunate - we have backups (and they're currently restoring now) - but I was just interested in if you 'attempt' to mount/import a messed up pool - it could potentially screwup any chances of mounting that pool cleanly again, even if you were to 'undo' the hardware changes. I have a feeling that a zpool import or 'initial' zpool status has to be a read/write operation (i.e. would fail anyway if you could magically make the underlying devices read-only?) -Kp From owner-freebsd-fs@FreeBSD.ORG Tue Oct 19 14:54:27 2010 Return-Path: Delivered-To: freebsd-fs@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id 888081065694; Tue, 19 Oct 2010 14:54:27 +0000 (UTC) (envelope-from gleb.kurtsou@gmail.com) Received: from mail-ew0-f54.google.com (mail-ew0-f54.google.com [209.85.215.54]) by mx1.freebsd.org (Postfix) with ESMTP id DF23A8FC0C; Tue, 19 Oct 2010 14:54:26 +0000 (UTC) Received: by ewy21 with SMTP id 21so1614588ewy.13 for ; Tue, 19 Oct 2010 07:54:25 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=gamma; h=domainkey-signature:received:received:date:from:to:cc:subject :message-id:references:mime-version:content-type:content-disposition :in-reply-to:user-agent; bh=Y5Pqt/jkAsvAwonqoTbu1VTPENJOABSvs1OPE7dq6/M=; b=JfFV0tEg/vwxa8riazfkqK5oUws/1xOAOyD47T5JyDc/QXKrad5jVCYC97tQZ9ZYHw vKpewZe7PohCIpSSuex15oyqTfTEDqlzKCvYxNe3Ve7UEYiqYo23KLZ8ZqyCxNybR5ay DLzsl6hIpR/Ns+u5O9tX6AA6t/M4vHumT1qoA= DomainKey-Signature: a=rsa-sha1; c=nofws; d=gmail.com; s=gamma; h=date:from:to:cc:subject:message-id:references:mime-version :content-type:content-disposition:in-reply-to:user-agent; b=XYcmxj61clt5XTqQRSfbsBEUdqWPICEmcCsWgb30YyW74dOCoXbWHzHUrYRtnED2WR s9L2eAuBklzGmZwLMqMiqBO4/jXEMOpZdLXScUBCm39N0jeW6zm0q7B1NKwxuvXTiI4S ShAGsCWpZCoUQH3PrU6cLpP8Ks+sBrBTaO48I= Received: by 10.213.26.81 with SMTP id d17mr419868ebc.42.1287499656732; Tue, 19 Oct 2010 07:47:36 -0700 (PDT) Received: from localhost (lan-78-157-92-5.vln.skynet.lt [78.157.92.5]) by mx.google.com with ESMTPS id v56sm8619645eeh.2.2010.10.19.07.47.35 (version=TLSv1/SSLv3 cipher=RC4-MD5); Tue, 19 Oct 2010 07:47:35 -0700 (PDT) Date: Tue, 19 Oct 2010 17:47:33 +0300 From: Gleb Kurtsou To: Gleb Smirnoff Message-ID: <20101019144733.GA2672@tops> References: <201010181230.o9ICUEAi016474@freefall.freebsd.org> <20101019143601.GA1982@tops> MIME-Version: 1.0 Content-Type: text/plain; charset=utf-8 Content-Disposition: inline In-Reply-To: <20101019143601.GA1982@tops> User-Agent: Mutt/1.5.20 (2009-06-14) Cc: freebsd-fs@FreeBSD.org Subject: Re: misc/151111: vnodes leakage during zfs unmount X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Tue, 19 Oct 2010 14:54:27 -0000 On (19/10/2010 17:36), Gleb Kurtsou wrote: > On (18/10/2010 12:30), Gleb Smirnoff wrote: > > The following reply was made to PR kern/151111; it has been noted by GNATS. > > > > From: Gleb Smirnoff > > To: "Oleg A. Mamontov" > > Cc: kib@FreeBSD.org, pjd@FreeBSD.org, freebsd-gnats-submit@FreeBSD.org > > Subject: Re: misc/151111: vnodes leakage during zfs unmount > > Date: Mon, 18 Oct 2010 16:11:55 +0400 > > > > On Thu, Sep 30, 2010 at 11:56:18AM +0000, Oleg A. Mamontov wrote: > > O> for i in `jot 1000`; do zfs mount tank; printf '%4d ' $i; sysctl vfs.numvnodes; zfs umount tank; done > > O> ######################## > > O> 1 vfs.numvnodes: 708 > > O> 2 vfs.numvnodes: 709 > > O> 3 vfs.numvnodes: 710 > > O> 4 vfs.numvnodes: 711 > > O> 5 vfs.numvnodes: 712 > > O> > > O> .. > > O> > > O> 995 vfs.numvnodes: 1702 > > O> 996 vfs.numvnodes: 1703 > > O> 997 vfs.numvnodes: 1704 > > O> 998 vfs.numvnodes: 1705 > > O> 999 vfs.numvnodes: 1706 > > O> 1000 vfs.numvnodes: 1707 > > In my tests zfs vnodes are getting properly gc'ed. Most likely maxvnodes > limit is not hit during the test, thus vnodes are not reclaimed. > > Try making all available vnodes used, e.g. by running > 'find /usr >/dev/null' before the test. Running the test you'll see > vfs.numvnodes going up and down. > > It's more visible after lowering kern.maxvnodes. I've used > kern.maxvnodes=1000. Default value on my system is ~127000. Please ignore my previous email. There is indeed a leak. > > Here is my lame investigation of the problem. In the zfs_domount() function > > we've got the following code: > > > > /* Grab extra reference. */ > > VERIFY(VFS_ROOT(vfsp, LK_EXCLUSIVE, &vp) == 0); > > VOP_UNLOCK(vp, 0); > > > > I suppose this code is expected to put an extra reference on the > > vfsp->mnt_vnodecovered vnode. Do I mistake here? If I don't then this is > > the source of leak. > > > > Debugging shows that zfs_zget(), called subsequently from zfs_root(), > > does not find an existing znode/vnode and allocates a new one, see at the > > end of zfs_zget() function. This vnode gots a reference and is forgotten. > > > > Sorry, if I am being mistaken :) > > > > -- > > Totus tuus, Glebius. > > _______________________________________________ > > freebsd-fs@freebsd.org mailing list > > http://lists.freebsd.org/mailman/listinfo/freebsd-fs > > To unsubscribe, send any mail to "freebsd-fs-unsubscribe@freebsd.org" From owner-freebsd-fs@FreeBSD.ORG Tue Oct 19 15:11:32 2010 Return-Path: Delivered-To: freebsd-fs@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id 5CE31106564A; Tue, 19 Oct 2010 15:11:32 +0000 (UTC) (envelope-from gleb.kurtsou@gmail.com) Received: from mail-pz0-f54.google.com (mail-pz0-f54.google.com [209.85.210.54]) by mx1.freebsd.org (Postfix) with ESMTP id 25E988FC15; Tue, 19 Oct 2010 15:11:31 +0000 (UTC) Received: by pzk2 with SMTP id 2so547171pzk.13 for ; Tue, 19 Oct 2010 08:11:31 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=gamma; h=domainkey-signature:received:received:date:from:to:cc:subject :message-id:references:mime-version:content-type:content-disposition :in-reply-to:user-agent; bh=aPFJf82vRousVSfZxRCmS1A6CYqpC+RChUxV8QvwNvM=; b=bX7CObLR0xqXJXsGELZiLFXxILaYz0NkikB94W6Ew91HtiPI3sPqi9bVZru9MHp9m0 vsoIEdd/FI5PTzPAEeFxA/lsIgzriKd/aiVsNyVjbS3w4cELnGS7ynnmQtLxjbT1yBz7 sLXtNLAig0JIp3BqMnmiFOmkbwmCaBgN+6y6E= DomainKey-Signature: a=rsa-sha1; c=nofws; d=gmail.com; s=gamma; h=date:from:to:cc:subject:message-id:references:mime-version :content-type:content-disposition:in-reply-to:user-agent; b=fsHD/9oxiRygpfNP7Mw0kHKOtDbuv9WAIolrPjElRSZ+pje/eTH7U8sBKoEbedJf0V ZCTs7+1Y4ThLKt+oDvMlhTGBz71eGBScSndlufq/s3WlykuNOyW6Krng/SldFhTRaBw5 AJNwYl20uyq/c/1iNxYSUXqRxgP+CgtUYINco= Received: by 10.14.37.10 with SMTP id x10mr3735334eea.30.1287498965715; Tue, 19 Oct 2010 07:36:05 -0700 (PDT) Received: from localhost (lan-78-157-92-5.vln.skynet.lt [78.157.92.5]) by mx.google.com with ESMTPS id w20sm248052eeh.6.2010.10.19.07.36.03 (version=TLSv1/SSLv3 cipher=RC4-MD5); Tue, 19 Oct 2010 07:36:03 -0700 (PDT) Date: Tue, 19 Oct 2010 17:36:01 +0300 From: Gleb Kurtsou To: Gleb Smirnoff Message-ID: <20101019143601.GA1982@tops> References: <201010181230.o9ICUEAi016474@freefall.freebsd.org> MIME-Version: 1.0 Content-Type: text/plain; charset=utf-8 Content-Disposition: inline In-Reply-To: <201010181230.o9ICUEAi016474@freefall.freebsd.org> User-Agent: Mutt/1.5.20 (2009-06-14) Cc: freebsd-fs@FreeBSD.org Subject: Re: misc/151111: vnodes leakage during zfs unmount X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Tue, 19 Oct 2010 15:11:32 -0000 On (18/10/2010 12:30), Gleb Smirnoff wrote: > The following reply was made to PR kern/151111; it has been noted by GNATS. > > From: Gleb Smirnoff > To: "Oleg A. Mamontov" > Cc: kib@FreeBSD.org, pjd@FreeBSD.org, freebsd-gnats-submit@FreeBSD.org > Subject: Re: misc/151111: vnodes leakage during zfs unmount > Date: Mon, 18 Oct 2010 16:11:55 +0400 > > On Thu, Sep 30, 2010 at 11:56:18AM +0000, Oleg A. Mamontov wrote: > O> for i in `jot 1000`; do zfs mount tank; printf '%4d ' $i; sysctl vfs.numvnodes; zfs umount tank; done > O> ######################## > O> 1 vfs.numvnodes: 708 > O> 2 vfs.numvnodes: 709 > O> 3 vfs.numvnodes: 710 > O> 4 vfs.numvnodes: 711 > O> 5 vfs.numvnodes: 712 > O> > O> .. > O> > O> 995 vfs.numvnodes: 1702 > O> 996 vfs.numvnodes: 1703 > O> 997 vfs.numvnodes: 1704 > O> 998 vfs.numvnodes: 1705 > O> 999 vfs.numvnodes: 1706 > O> 1000 vfs.numvnodes: 1707 In my tests zfs vnodes are getting properly gc'ed. Most likely maxvnodes limit is not hit during the test, thus vnodes are not reclaimed. Try making all available vnodes used, e.g. by running 'find /usr >/dev/null' before the test. Running the test you'll see vfs.numvnodes going up and down. It's more visible after lowering kern.maxvnodes. I've used kern.maxvnodes=1000. Default value on my system is ~127000. > Here is my lame investigation of the problem. In the zfs_domount() function > we've got the following code: > > /* Grab extra reference. */ > VERIFY(VFS_ROOT(vfsp, LK_EXCLUSIVE, &vp) == 0); > VOP_UNLOCK(vp, 0); > > I suppose this code is expected to put an extra reference on the > vfsp->mnt_vnodecovered vnode. Do I mistake here? If I don't then this is > the source of leak. > > Debugging shows that zfs_zget(), called subsequently from zfs_root(), > does not find an existing znode/vnode and allocates a new one, see at the > end of zfs_zget() function. This vnode gots a reference and is forgotten. > > Sorry, if I am being mistaken :) > > -- > Totus tuus, Glebius. > _______________________________________________ > freebsd-fs@freebsd.org mailing list > http://lists.freebsd.org/mailman/listinfo/freebsd-fs > To unsubscribe, send any mail to "freebsd-fs-unsubscribe@freebsd.org" From owner-freebsd-fs@FreeBSD.ORG Tue Oct 19 15:16:03 2010 Return-Path: Delivered-To: freebsd-fs@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id EB7D21065672 for ; Tue, 19 Oct 2010 15:16:03 +0000 (UTC) (envelope-from jdc@koitsu.dyndns.org) Received: from qmta04.emeryville.ca.mail.comcast.net (qmta04.emeryville.ca.mail.comcast.net [76.96.30.40]) by mx1.freebsd.org (Postfix) with ESMTP id D03EE8FC0A for ; Tue, 19 Oct 2010 15:16:03 +0000 (UTC) Received: from omta03.emeryville.ca.mail.comcast.net ([76.96.30.27]) by qmta04.emeryville.ca.mail.comcast.net with comcast id LcQN1f0080b6N64A4fG3Dc; Tue, 19 Oct 2010 15:16:03 +0000 Received: from koitsu.dyndns.org ([98.248.41.155]) by omta03.emeryville.ca.mail.comcast.net with comcast id LfG21f0063LrwQ28PfG273; Tue, 19 Oct 2010 15:16:03 +0000 Received: by icarus.home.lan (Postfix, from userid 1000) id 0F17C9B418; Tue, 19 Oct 2010 08:16:02 -0700 (PDT) Date: Tue, 19 Oct 2010 08:16:02 -0700 From: Jeremy Chadwick To: Karl Pielorz Message-ID: <20101019151602.GA61733@icarus.home.lan> References: MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: User-Agent: Mutt/1.5.21 (2010-09-15) Cc: freebsd-fs@freebsd.org Subject: Re: ZFS 'read-only' device / pool scan / import? X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Tue, 19 Oct 2010 15:16:04 -0000 On Tue, Oct 19, 2010 at 03:52:52PM +0100, Karl Pielorz wrote: > On FreeBSD if I bring the system up 'single user' - the first time I > do (for example): > > " > zpool status > " > > There's a pause, a flurry of disk activity - and the system appears > to import any pools it finds, and the status appears. > > I'd guess at this point - some data is written to the disks? - Is > there any way of avoiding that, i.e. a kind of "If you were to > import the pools/display a status, what would you show, without > actually writing any data?" (or importing it) - i.e. to ensure the > devices are opened read only? The activity you see is almost certainly the result of kernel modules being loaded dynamically and so on, plus disk tasting and metadata analysis. This might be different if you had opensolaris_load="yes" and zfs_load="yes" in /boot/loader.conf, but I'm guessing that isn't the case. BTW, in single-user, I always do the following before doing *any* ZFS-related work. This is assuming the system uses UFS for /, /var, /tmp, and /usr. mount -t ufs -a /etc/rc.d/hostid start /etc/rc.d/zfs start > The reason I ask is we recently had a pool that had a mishap. The > backing RAID controller dropped a drive (we were using them in > JBOD). This happened cleanly. The system got shutdown, and I think > the wrong drive was replaced. > > When it came up the RAID controller 'collapsed' the device list (so > there was no gap where the old drive was) - that, plus the wrong > drive being replaced meant when we looked at the system we had: > > " > pool: vol > state: UNAVAIL > status: One or more devices could not be used because the label is missing > or invalid. There are insufficient replicas for the pool to > continue functioning. > " > > A number of devices were listed as 'corrupted data' - some devices > were listed twice as members of the pool - i.e. pretty screwed up. > > 'undoing' the damage and restarting the server - just threw up the > same status. > > I'm wondering if through the action of having the pool > imported/mounted etc. - ZFS has actually *written* to the drives > that were available that other drives aren't available / corrupt - > and basically, because that info was written, and check-summed > correctly - it takes that as gospel now, rather than actually > 're-checking' the drives (or is simply unable to re-check the drives > - because the meta data has been changed from the previous boot). > > If you see what I mean :) > > In the end, we're fortunate - we have backups (and they're currently > restoring now) - but I was just interested in if you 'attempt' to > mount/import a messed up pool - it could potentially screwup any > chances of mounting that pool cleanly again, even if you were to > 'undo' the hardware changes. > > I have a feeling that a zpool import or 'initial' zpool status has > to be a read/write operation (i.e. would fail anyway if you could > magically make the underlying devices read-only?) Experts here might be able to help, but you're really going to need to provide every little detail, in chronological order. What commands were done, what output was seen, what physical actions took place, etc.. 1) Restoring from backups is probably your best bet (IMHO; this is what I would do as well). 2) Take inventory of your hardware; get disk serial numbers (smartctl can show you this) and so on, and correlate them with specific device IDs (adaX, daX, etc.). I also label my drive bays with which device ID they associate with. 3) You didn't disclose what kind of ZFS pool setup you have. I'm guessing raidz1. Or is it a pool of mirrors? Or raidz2? Or...? It obviously matters. 4) If it's raidz1, I'm guessing you're hurt, and here's why: the disk falling off the bus during shut-down (or whatever happened -- meaning the incident that occurred *prior* to the wrong disk being replaced) would almost certainly have resulted in ZFS saying "hey! the array is degraded! Replace the disk! If anything else happens until that disk is replaced, you'll experience data loss!" Then somehow the wrong disk was replaced. At this point you have 1 disk which may be broken/wrong/whatever (the one which disappeared from the bus during the system shut-down), and one disk which is brand new and needs to be resilvered... so you're basically down to 1 disk. raidz1 isn't going to help you in this case: you can lose 1 disk, regardless of situation, period. There's a lot of other things I could add to the item list here (probably reach 9 or 10 if I tried), but in general the above sounds like its what happened. raidz2 would have been able to save you in this situation, but would require at least 4 disks. -- | Jeremy Chadwick jdc@parodius.com | | Parodius Networking http://www.parodius.com/ | | UNIX Systems Administrator Mountain View, CA, USA | | Making life hard for others since 1977. PGP: 4BD6C0CB | From owner-freebsd-fs@FreeBSD.ORG Tue Oct 19 15:30:43 2010 Return-Path: Delivered-To: freebsd-fs@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id 500CA106564A for ; Tue, 19 Oct 2010 15:30:43 +0000 (UTC) (envelope-from kpielorz_lst@tdx.co.uk) Received: from mail.tdx.com (mail.tdx.com [62.13.128.18]) by mx1.freebsd.org (Postfix) with ESMTP id D20F08FC0A for ; Tue, 19 Oct 2010 15:30:42 +0000 (UTC) Received: from HexaDeca64.dmpriest.net.uk (HPQuadro64.dmpriest.net.uk [62.13.130.30]) (authenticated bits=0) by mail.tdx.com (8.14.3/8.14.3/Kp) with ESMTP id o9JFUfcj020790 (version=TLSv1/SSLv3 cipher=DHE-DSS-AES256-SHA bits=256 verify=NO); Tue, 19 Oct 2010 16:30:41 +0100 (BST) Date: Tue, 19 Oct 2010 16:30:41 +0100 From: Karl Pielorz To: Jeremy Chadwick Message-ID: <7BEF90D9F4D4CB985F3573C3@HexaDeca64.dmpriest.net.uk> In-Reply-To: <20101019151602.GA61733@icarus.home.lan> References: <20101019151602.GA61733@icarus.home.lan> X-Mailer: Mulberry/4.0.8 (Win32) MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii; format=flowed Content-Transfer-Encoding: 7bit Content-Disposition: inline Cc: freebsd-fs@freebsd.org Subject: Re: ZFS 'read-only' device / pool scan / import? X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Tue, 19 Oct 2010 15:30:43 -0000 --On 19 October 2010 08:16 -0700 Jeremy Chadwick wrote: > Experts here might be able to help, but you're really going to need to > provide every little detail, in chronological order. What commands were > done, what output was seen, what physical actions took place, etc.. > > 1) Restoring from backups is probably your best bet (IMHO; this is what I > would do as well). I didn't provide much detail - as there isn't much detail left to provide (the pools been destroyed / rebuilt) - how it got messed up is almost certainly a case of human error / controller 'oddity' with failed devices [which is now suitably noted for that machine!]... It was more a 'for future reference' kind of question - does attempting to import a pool (or even running something as simple as a 'zfs status' when ZFS has not been 'loaded') actually write to the disks? i.e. could it cause a pool that is currently 'messed up' to become permanently 'messed up' - because ZFS will change metadata on the pool, if 'at the time' it deems devices to be faulted / corrupt etc. - And, if it does - is there any way of doing a 'test mount/import' (i.e. with the underlying devices only being opened 'read only' - or does [as I suspect] ZFS *need* r/w access to those devices as part of the work to actually import/mount. > There's a lot of other things I could add to the item list here > (probably reach 9 or 10 if I tried), but in general the above sounds > like its what happened. raidz2 would have been able to save you in this > situation, but would require at least 4 disks. It was RAIDZ2 - it got totally screwed: " vol UNAVAIL 0 0 0 insufficient replicas raidz2 UNAVAIL 0 0 0 insufficient replicas da3 FAULTED 0 0 0 corrupted data da4 FAULTED 0 0 0 corrupted data da5 FAULTED 0 0 0 corrupted data da6 FAULTED 0 0 0 corrupted data da7 FAULTED 0 0 0 corrupted data da8 FAULTED 0 0 0 corrupted data raidz2 UNAVAIL 0 0 0 insufficient replicas da1 ONLINE 0 0 0 da2 ONLINE 0 0 0 da9 FAULTED 0 0 0 corrupted data da10 FAULTED 0 0 0 corrupted data da11 FAULTED 0 0 0 corrupted data da11 ONLINE 0 0 0 " As there is such a large aspect of human error (and controller behaviour), I don't think it's worth digging into any deeper. It's the first pool we've ever "lost" under ZFS, and like I said a combination of the controller collapsing devices, and humans replacing wrong disks, 'twas doomed to fail from the start. We've replaced failed drives on this system before - but never rebooted after a failure, before a replacement - and never replaced the wrong drive :) Definitely a good advert for backups though :) -Karl From owner-freebsd-fs@FreeBSD.ORG Tue Oct 19 20:30:49 2010 Return-Path: Delivered-To: freebsd-fs@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id B7F68106566B for ; Tue, 19 Oct 2010 20:30:49 +0000 (UTC) (envelope-from wjw@digiware.nl) Received: from mail.digiware.nl (mail.ip6.digiware.nl [IPv6:2001:4cb8:1:106::2]) by mx1.freebsd.org (Postfix) with ESMTP id 4D70E8FC0A for ; Tue, 19 Oct 2010 20:30:49 +0000 (UTC) Received: from localhost (localhost.digiware.nl [127.0.0.1]) by mail.digiware.nl (Postfix) with ESMTP id C0E67153434; Tue, 19 Oct 2010 22:30:47 +0200 (CEST) X-Virus-Scanned: amavisd-new at digiware.nl Received: from mail.digiware.nl ([127.0.0.1]) by localhost (rack1.digiware.nl [127.0.0.1]) (amavisd-new, port 10024) with ESMTP id Tqo7D3N6K-pz; Tue, 19 Oct 2010 22:30:45 +0200 (CEST) Received: from [127.0.0.1] (opteron [192.168.10.67]) (using TLSv1 with cipher DHE-RSA-CAMELLIA256-SHA (256/256 bits)) (No client certificate requested) by mail.digiware.nl (Postfix) with ESMTPSA id 69C6A153433; Tue, 19 Oct 2010 22:30:45 +0200 (CEST) Message-ID: <4CBDFFF6.5080701@digiware.nl> Date: Tue, 19 Oct 2010 22:30:46 +0200 From: Willem Jan Withagen User-Agent: Mozilla/5.0 (Windows; U; Windows NT 6.1; en-US; rv:1.9.2.9) Gecko/20100915 Lightning/1.0b2 Thunderbird/3.1.4 MIME-Version: 1.0 To: Karl Pielorz References: <20101019151602.GA61733@icarus.home.lan> <7BEF90D9F4D4CB985F3573C3@HexaDeca64.dmpriest.net.uk> In-Reply-To: <7BEF90D9F4D4CB985F3573C3@HexaDeca64.dmpriest.net.uk> Content-Type: text/plain; charset=ISO-8859-1; format=flowed Content-Transfer-Encoding: 7bit Cc: freebsd-fs@freebsd.org Subject: Re: ZFS 'read-only' device / pool scan / import? X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Tue, 19 Oct 2010 20:30:49 -0000 On 2010-10-19 17:30, Karl Pielorz wrote: > As there is such a large aspect of human error (and controller > behaviour), I don't think it's worth digging into any deeper. It's the > first pool we've ever "lost" under ZFS, and like I said a combination of > the controller collapsing devices, and humans replacing wrong disks, > 'twas doomed to fail from the start. > > We've replaced failed drives on this system before - but never rebooted > after a failure, before a replacement - and never replaced the wrong > drive :) > > Definitely a good advert for backups though :) I'm running my ZFS stuff on a 3ware and an areca controller, and they once in a while forgot their order of disks during booting. (the 3ware got fixed by a bios upgrade) The areca just keeps reordering no matter how hard you like to tell it otherwise. But GPT really proves useful since reallocation of disks does not result in a different device in the gpt directory. eg.: pool: zroot state: ONLINE scrub: none requested config: NAME STATE READ WRITE CKSUM zroot ONLINE 0 0 0 mirror ONLINE 0 0 0 gpt/root4 ONLINE 0 0 0 gpt/root6 ONLINE 0 0 0 I could even migrate a disk from the 3ware controller to the std SATA interfaces without losing the the gpt-label. --WjW From owner-freebsd-fs@FreeBSD.ORG Tue Oct 19 21:03:12 2010 Return-Path: Delivered-To: freebsd-fs@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id A413A106566B for ; Tue, 19 Oct 2010 21:03:12 +0000 (UTC) (envelope-from fjwcash@gmail.com) Received: from mail-gw0-f54.google.com (mail-gw0-f54.google.com [74.125.83.54]) by mx1.freebsd.org (Postfix) with ESMTP id 5DE368FC14 for ; Tue, 19 Oct 2010 21:03:11 +0000 (UTC) Received: by gwb20 with SMTP id 20so1599425gwb.13 for ; Tue, 19 Oct 2010 14:03:11 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=gamma; h=domainkey-signature:mime-version:received:received:in-reply-to :references:date:message-id:subject:from:to:content-type :content-transfer-encoding; bh=UyqnUJdYpPgJmj1VoENyfHaPPe9fTIE4x/B80CjLF+4=; b=IA2/LKgE4hdsRT0DrEuUqLMmh5gmbf5NGLPrMv4ApVuYd0612gsJoEiiJEWnK+D51I hxUfu12oI5WMToqoNW/0o3sOVet1RMTAdbsiWE4rkSvZ2qtvwaElqx9ZEhi1h95pnTbk fgFTJoD10CTsVJueJJtAOZxSKjVJKAEzKEaUA= DomainKey-Signature: a=rsa-sha1; c=nofws; d=gmail.com; s=gamma; h=mime-version:in-reply-to:references:date:message-id:subject:from:to :content-type:content-transfer-encoding; b=bUQ1ztv/wS2S2UkPTunXWjk7xx29oATeLY/QcHjEmLtsr9V5sAxSwOQSt8HEo+FPpU R5RXAgPQaKaUMis4w1jpQz7Mk8j9LjmLsBQeLR5mi+kjAbYpVLC08ZdsAafoa6zStukq 8gQchy1/ys8QfwUVnu/X+75MILabaEKLdtCVA= MIME-Version: 1.0 Received: by 10.90.103.2 with SMTP id a2mr3033528agc.73.1287522191374; Tue, 19 Oct 2010 14:03:11 -0700 (PDT) Received: by 10.90.56.10 with HTTP; Tue, 19 Oct 2010 14:03:11 -0700 (PDT) In-Reply-To: <4CBDFFF6.5080701@digiware.nl> References: <20101019151602.GA61733@icarus.home.lan> <7BEF90D9F4D4CB985F3573C3@HexaDeca64.dmpriest.net.uk> <4CBDFFF6.5080701@digiware.nl> Date: Tue, 19 Oct 2010 14:03:11 -0700 Message-ID: From: Freddie Cash To: freebsd-fs@freebsd.org Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: quoted-printable Subject: Re: ZFS 'read-only' device / pool scan / import? X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Tue, 19 Oct 2010 21:03:12 -0000 On Tue, Oct 19, 2010 at 1:30 PM, Willem Jan Withagen wrot= e: > On 2010-10-19 17:30, Karl Pielorz wrote: > >> As there is such a large aspect of human error (and controller >> behaviour), I don't think it's worth digging into any deeper. It's the >> first pool we've ever "lost" under ZFS, and like I said a combination of >> the controller collapsing devices, and humans replacing wrong disks, >> 'twas doomed to fail from the start. >> >> We've replaced failed drives on this system before - but never rebooted >> after a failure, before a replacement - and never replaced the wrong >> drive :) >> >> Definitely a good advert for backups though :) > > I'm running my ZFS stuff on a 3ware and an areca controller, and they onc= e > in a while forgot their order of disks during booting. > (the 3ware got fixed by a bios upgrade) > The areca just keeps reordering no matter how hard you like to tell it > otherwise. > > But GPT really proves useful since reallocation of disks does not result = in > a different device in the gpt directory. > > eg.: > =C2=A0pool: zroot > =C2=A0state: ONLINE > =C2=A0scrub: none requested > config: > > =C2=A0 =C2=A0 =C2=A0 =C2=A0NAME =C2=A0 =C2=A0 =C2=A0 =C2=A0 =C2=A0 STATE = =C2=A0 =C2=A0 READ WRITE CKSUM > =C2=A0 =C2=A0 =C2=A0 =C2=A0zroot =C2=A0 =C2=A0 =C2=A0 =C2=A0 =C2=A0ONLINE= =C2=A0 =C2=A0 =C2=A0 0 =C2=A0 =C2=A0 0 =C2=A0 =C2=A0 0 > =C2=A0 =C2=A0 =C2=A0 =C2=A0 =C2=A0mirror =C2=A0 =C2=A0 =C2=A0 ONLINE =C2= =A0 =C2=A0 =C2=A0 0 =C2=A0 =C2=A0 0 =C2=A0 =C2=A0 0 > =C2=A0 =C2=A0 =C2=A0 =C2=A0 =C2=A0 =C2=A0gpt/root4 =C2=A0ONLINE =C2=A0 = =C2=A0 =C2=A0 0 =C2=A0 =C2=A0 0 =C2=A0 =C2=A0 0 > =C2=A0 =C2=A0 =C2=A0 =C2=A0 =C2=A0 =C2=A0gpt/root6 =C2=A0ONLINE =C2=A0 = =C2=A0 =C2=A0 0 =C2=A0 =C2=A0 0 =C2=A0 =C2=A0 0 > > I could even migrate a disk from the 3ware controller to the std SATA > interfaces without losing the the gpt-label. While using labelled devices (be it glabel, gpt label, or whatever) certainly helps keep things ordered and working. However, ZFS also labels the devices in the pool. A simple "zpool export poolname" followed by a "zpool import poolname" will scan the metadata on the drives, find all the devices in the pool, re-order things internally, and carry on. Back when I started with ZFS, I used unlabelled drives (also on 3Ware controllers) and made the mistake once of booting with a failed drive removed. Pool came up faulted saying all the drives after the missing one were also faulted. Thought I lost the whole pool. However, some digging online showed the export/import info, and I was able to continue on without losing any data. --=20 Freddie Cash fjwcash@gmail.com From owner-freebsd-fs@FreeBSD.ORG Tue Oct 19 21:58:32 2010 Return-Path: Delivered-To: freebsd-fs@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id 15CB5106564A for ; Tue, 19 Oct 2010 21:58:32 +0000 (UTC) (envelope-from wjw@digiware.nl) Received: from mail.digiware.nl (mail.ip6.digiware.nl [IPv6:2001:4cb8:1:106::2]) by mx1.freebsd.org (Postfix) with ESMTP id 9DDD18FC14 for ; Tue, 19 Oct 2010 21:58:31 +0000 (UTC) Received: from localhost (localhost.digiware.nl [127.0.0.1]) by mail.digiware.nl (Postfix) with ESMTP id E5AF1153434; Tue, 19 Oct 2010 23:58:30 +0200 (CEST) X-Virus-Scanned: amavisd-new at digiware.nl Received: from mail.digiware.nl ([127.0.0.1]) by localhost (rack1.digiware.nl [127.0.0.1]) (amavisd-new, port 10024) with ESMTP id OlwzFulwcqrV; Tue, 19 Oct 2010 23:58:28 +0200 (CEST) Received: from [127.0.0.1] (opteron [192.168.10.67]) (using TLSv1 with cipher DHE-RSA-CAMELLIA256-SHA (256/256 bits)) (No client certificate requested) by mail.digiware.nl (Postfix) with ESMTPSA id B8452153433; Tue, 19 Oct 2010 23:58:28 +0200 (CEST) Message-ID: <4CBE1485.2070400@digiware.nl> Date: Tue, 19 Oct 2010 23:58:29 +0200 From: Willem Jan Withagen User-Agent: Mozilla/5.0 (Windows; U; Windows NT 6.1; en-US; rv:1.9.2.9) Gecko/20100915 Lightning/1.0b2 Thunderbird/3.1.4 MIME-Version: 1.0 To: Freddie Cash References: <20101019151602.GA61733@icarus.home.lan> <7BEF90D9F4D4CB985F3573C3@HexaDeca64.dmpriest.net.uk> <4CBDFFF6.5080701@digiware.nl> In-Reply-To: Content-Type: text/plain; charset=UTF-8; format=flowed Content-Transfer-Encoding: 7bit Cc: freebsd-fs@freebsd.org Subject: Re: ZFS 'read-only' device / pool scan / import? X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Tue, 19 Oct 2010 21:58:32 -0000 On 2010-10-19 23:03, Freddie Cash wrote: > While using labelled devices (be it glabel, gpt label, or whatever) > certainly helps keep things ordered and working. > > However, ZFS also labels the devices in the pool. A simple "zpool > export poolname" followed by a "zpool import poolname" will scan the > metadata on the drives, find all the devices in the pool, re-order > things internally, and carry on. > > Back when I started with ZFS, I used unlabelled drives (also on 3Ware > controllers) and made the mistake once of booting with a failed drive > removed. Pool came up faulted saying all the drives after the missing > one were also faulted. Thought I lost the whole pool. However, some > digging online showed the export/import info, and I was able to > continue on without losing any data. At the moment I have two 10Tb systems available so I'm experimenting a lot to get a feeling on how easy recovering is. So your comments are well appreciated. And your suggestion is that although things looked really messed up, just export/import fixed the lot.... The GPT trick is sort of selfdocumenting, because it is otherwise real easy to get lost with 10 disks and 4 flash-devices. :( And just because I yanked a single log disk (the wrong one) I ended up with a corrupt zraid. So the least lesson for me here (again) has been that disk handling requires the utmost care. Trouble is just around the corner. --WjW From owner-freebsd-fs@FreeBSD.ORG Tue Oct 19 22:12:23 2010 Return-Path: Delivered-To: freebsd-fs@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id 0E4F71065696 for ; Tue, 19 Oct 2010 22:12:23 +0000 (UTC) (envelope-from andspyr@it.teithe.gr) Received: from alpha.it.teithe.gr (alpha.it.teithe.gr [195.251.240.232]) by mx1.freebsd.org (Postfix) with ESMTP id 630998FC16 for ; Tue, 19 Oct 2010 22:12:21 +0000 (UTC) Received: from localhost (babel.noc.teithe.gr [195.251.240.240]) by alpha.it.teithe.gr (8.14.3/8.14.3/Debian-5+lenny1) with ESMTP id o9JLqWeC023865 for ; Wed, 20 Oct 2010 00:52:32 +0300 Received: from cust-206-13.on4.ontelecoms.gr (cust-206-13.on4.ontelecoms.gr [92.118.206.13]) by webmail.teithe.gr (Horde Framework) with HTTP; Wed, 20 Oct 2010 00:52:35 +0300 Message-ID: <20101020005235.544110onceyxojs3@webmail.teithe.gr> Date: Wed, 20 Oct 2010 00:52:35 +0300 From: "Spiros Andreou" To: freebsd-fs@freebsd.org MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8; DelSp="Yes"; format="flowed" Content-Disposition: inline Content-Transfer-Encoding: 7bit User-Agent: Internet Messaging Program (IMP) H3 (4.2) Subject: hast_vbox_current X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Tue, 19 Oct 2010 22:12:23 -0000 hello there I have two virtual systems (hasta hastb that running freebsd current) using Vbox and with real ip hasta 192.168.1.40 hastb 192.168.1.41 virtual ip 192.168.1.41 I did setup hast like http://wiki.freebsd.org/HAST With normal halt it works. The problem is that when the master (hasta) is going down unexpectedly (eg unplug the cable or power off) the slave (hastb) isn't becoming master . As a result,the slave doesn't mount the file system, but it changes the shared ip . Any suggestion ? ty. -- Spiros Andreou -- ' `. |a_ a | \<_)_/ /( )\ |\`> < /\ \_|=='|_/ Linux Counter ID: 408991 If you can't see the above message please inform me. ---------------------------------------------------------------- This message was sent using IMP, the Internet Messaging Program. From owner-freebsd-fs@FreeBSD.ORG Wed Oct 20 09:04:50 2010 Return-Path: Delivered-To: freebsd-fs@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id 747F41065672 for ; Wed, 20 Oct 2010 09:04:49 +0000 (UTC) (envelope-from kpielorz_lst@tdx.co.uk) Received: from mail.tdx.com (mail.tdx.com [62.13.128.18]) by mx1.freebsd.org (Postfix) with ESMTP id 5F71A8FC1A for ; Wed, 20 Oct 2010 09:04:48 +0000 (UTC) Received: from HexaDeca64.dmpriest.net.uk (HPQuadro64.dmpriest.net.uk [62.13.130.30]) (authenticated bits=0) by mail.tdx.com (8.14.3/8.14.3/Kp) with ESMTP id o9K94eBO009884 (version=TLSv1/SSLv3 cipher=DHE-DSS-AES256-SHA bits=256 verify=NO); Wed, 20 Oct 2010 10:04:40 +0100 (BST) Date: Wed, 20 Oct 2010 10:04:37 +0100 From: Karl Pielorz To: Willem Jan Withagen , Freddie Cash Message-ID: <79DCDDF23AE4C12D5A222B8B@HexaDeca64.dmpriest.net.uk> In-Reply-To: <4CBE1485.2070400@digiware.nl> References: <20101019151602.GA61733@icarus.home.lan> <7BEF90D9F4D4CB985F3573C3@HexaDeca64.dmpriest.net.uk> <4CBDFFF6.5080701@digiware.nl> <4CBE1485.2070400@digiware.nl> X-Mailer: Mulberry/4.0.8 (Win32) MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii; format=flowed Content-Transfer-Encoding: 7bit Content-Disposition: inline Cc: freebsd-fs@freebsd.org Subject: Re: ZFS 'read-only' device / pool scan / import? X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Wed, 20 Oct 2010 09:04:50 -0000 --On 19 October 2010 23:58 +0200 Willem Jan Withagen wrote: > And your suggestion is that although things looked really messed up, just > export/import fixed the lot.... I've had messed up pools fixed up before by shutting down, 'rejigging' the drives and restarting - but I've didn't think of exporting/importing [afaik the FreeBSD code goes and 'looks' for all the drives anyway at startup - i.e. it doesn't have something akin to zfs.cache] - so I kind of didn't see the point... I also didn't think you could export a faulted pool :) [or, as the original thread started - that once you'd tried to import a faulted pool, the on-drive metadata gets change, possibly corrupted - or at the very least it makes a note of the fact the missing drives are, 'missing' :) > The GPT trick is sort of selfdocumenting, because it is otherwise real > easy to get lost with 10 disks and 4 flash-devices. :( > And just because I yanked a single log disk (the wrong one) I ended up > with a corrupt zraid. Yeah, we'll definitely start looking at GPT - I've kind of always shied away from it before as we've always just used raw disks, and I think there was a paranoia about other stuff 'clobbering' the GPT metadata. That was probably a few years ago now though - so definitely worth a re-visit. > So the least lesson for me here (again) has been that disk handling > requires the utmost care. Trouble is just around the corner. A very valid point - for all the wonder and security ZFS brings, you *still* need to be careful with data! I'm also starting to disfavour RAIDZx vs. mirrors - mirrors are so much easier to work with (performance aside) and the ability to just 'buddy up' a possibly failing disk, or change redundancy on a volume is very nice. I guess a lot depends on if you have the bays for it :) I guess a counter argument is that if you've designed the system correctly you shouldn't want or need to go affecting a volumes redundancy! :) -Karl From owner-freebsd-fs@FreeBSD.ORG Wed Oct 20 10:35:54 2010 Return-Path: Delivered-To: freebsd-fs@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id 288ED106564A for ; Wed, 20 Oct 2010 10:35:54 +0000 (UTC) (envelope-from petefrench@ticketswitch.com) Received: from constantine.ticketswitch.com (constantine.ticketswitch.com [IPv6:2002:57e0:1d4e:1::3]) by mx1.freebsd.org (Postfix) with ESMTP id E08CF8FC14 for ; Wed, 20 Oct 2010 10:35:53 +0000 (UTC) Received: from dilbert.rattatosk ([10.64.50.6] helo=dilbert.ticketswitch.com) by constantine.ticketswitch.com with esmtps (TLSv1:AES256-SHA:256) (Exim 4.69 (FreeBSD)) (envelope-from ) id 1P8W1Q-0007Mk-9b; Wed, 20 Oct 2010 11:35:52 +0100 Received: from petefrench by dilbert.ticketswitch.com with local (Exim 4.72 (FreeBSD)) (envelope-from ) id 1P8W1Q-0004Y2-7Z; Wed, 20 Oct 2010 11:35:52 +0100 Date: Wed, 20 Oct 2010 11:35:52 +0100 Message-Id: To: fjwcash@gmail.com, kpielorz_lst@tdx.co.uk, wjw@digiware.nl In-Reply-To: <79DCDDF23AE4C12D5A222B8B@HexaDeca64.dmpriest.net.uk> From: Pete French Cc: freebsd-fs@freebsd.org Subject: Re: ZFS 'read-only' device / pool scan / import? X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Wed, 20 Oct 2010 10:35:54 -0000 > the FreeBSD code goes and 'looks' for all the drives anyway at startup - > i.e. it doesn't have something akin to zfs.cache] - so I kind of didn't see > the point... Theres a file /boot/zfs/zpool.cache which is, I think, the equivalent > I also didn't think you could export a faulted pool :) [or, as the original > thread started - that once you'd tried to import a faulted pool, the > on-drive metadata gets change, possibly corrupted - or at the very least it > makes a note of the fact the missing drives are, 'missing' :) You can do a forced export - as long as the system doesnt think the pool is imported then an import will go looking. I made a similar mess of my drives the first time I changed discs oover, and I also didnt know about the import/export trick. But since then I have used iit every time I have got into trouble and it has always worked for me. cheers, -pete. From owner-freebsd-fs@FreeBSD.ORG Thu Oct 21 15:43:43 2010 Return-Path: Delivered-To: freebsd-fs@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id EAE461065672 for ; Thu, 21 Oct 2010 15:43:43 +0000 (UTC) (envelope-from fjwcash@gmail.com) Received: from mail-gx0-f182.google.com (mail-gx0-f182.google.com [209.85.161.182]) by mx1.freebsd.org (Postfix) with ESMTP id A77618FC14 for ; Thu, 21 Oct 2010 15:43:43 +0000 (UTC) Received: by gxk2 with SMTP id 2so132338gxk.13 for ; Thu, 21 Oct 2010 08:43:42 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=gamma; h=domainkey-signature:mime-version:received:received:in-reply-to :references:date:message-id:subject:from:to:cc:content-type; bh=oTJr/q+Nh3eoFY+SBUlW+PMcZi0Dh+dHfH/1YqKxQCo=; b=ItJ3VylBcg2zVnW5CvvOPdd7HG1NC/keEz/X7FKkrNmWJXoIm7CxKXkgjmtlGABIoW F7I3bg9Wpk9asXGNjGEiMUBgRWISf+JQb8nW8CpAROk4GWey2AZSi0cDjSCAf9LzqEhG uSHwd0K2X8o3456S4OKyVpNWgEE1qEHxRoPSo= DomainKey-Signature: a=rsa-sha1; c=nofws; d=gmail.com; s=gamma; h=mime-version:in-reply-to:references:date:message-id:subject:from:to :cc:content-type; b=oRCCh9vCvxSDrrnvJ+LMb7tShOgKpYE0aHO1ZmDoys7pcq5ll760VRwgupBZPOHjSs KoXnVqvOa25Sf22fbfehAkv+LhYgqQ2J0xYNfW52q49CG83XuGUiaTdkmj47ihgDwFm4 irsFdI9059CCt/WU8IploLZQmhEid7vfPYC8A= MIME-Version: 1.0 Received: by 10.90.93.8 with SMTP id q8mr1454538agb.163.1287675821369; Thu, 21 Oct 2010 08:43:41 -0700 (PDT) Received: by 10.90.56.10 with HTTP; Thu, 21 Oct 2010 08:43:41 -0700 (PDT) In-Reply-To: <4CAE798D.6040905@xpam.de> References: <4CAE6C3A.5070509@rcn.com> <4CAE798D.6040905@xpam.de> Date: Thu, 21 Oct 2010 08:43:41 -0700 Message-ID: From: Freddie Cash To: Adam Nowacki Content-Type: text/plain; charset=UTF-8 Cc: freebsd-fs@freebsd.org Subject: Re: ZFS and Samsung Spinpoint F4 4K sector drive X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Thu, 21 Oct 2010 15:43:44 -0000 On Thu, Oct 7, 2010 at 6:53 PM, Adam Nowacki wrote: > Nothing user friendly but if you're willing to modify sources and rebuild > kernel there is a one line tweak. > > /usr/src/sys/cddl/contrib/opensolaris/uts/common/fs/zfs/vdev_geom.c find: > *ashift = highbit(MAX(pp->sectorsize, SPA_MINBLOCKSIZE)) - 1; > and modify to something like: > *ashift = highbit(MAX(MAX(4096, pp->sectorsize), SPA_MINBLOCKSIZE)) - 1; > > I'm using this for 3 months with 20 2TB 4kb sector WDC disks (in 2 raidz2 > arrays of 10) without any issues. Writes go at 300MB/s. Does this patch work the same was as the following: http://www.solarismen.de/archives/5-Solaris-and-the-new-4K-Sector-Disks-e.g.-WDxxEARS-Part-2.html They both deal with setting ashift to 4096 bytes, but are in two very different places in the code. -- Freddie Cash fjwcash@gmail.com From owner-freebsd-fs@FreeBSD.ORG Thu Oct 21 21:40:10 2010 Return-Path: Delivered-To: freebsd-fs@hub.freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id CDCBD106566C for ; Thu, 21 Oct 2010 21:40:10 +0000 (UTC) (envelope-from gnats@FreeBSD.org) Received: from freefall.freebsd.org (freefall.freebsd.org [IPv6:2001:4f8:fff6::28]) by mx1.freebsd.org (Postfix) with ESMTP id A1DBB8FC14 for ; Thu, 21 Oct 2010 21:40:10 +0000 (UTC) Received: from freefall.freebsd.org (localhost [127.0.0.1]) by freefall.freebsd.org (8.14.4/8.14.4) with ESMTP id o9LLeAtc063746 for ; Thu, 21 Oct 2010 21:40:10 GMT (envelope-from gnats@freefall.freebsd.org) Received: (from gnats@localhost) by freefall.freebsd.org (8.14.4/8.14.4/Submit) id o9LLeA72063745; Thu, 21 Oct 2010 21:40:10 GMT (envelope-from gnats) Date: Thu, 21 Oct 2010 21:40:10 GMT Message-Id: <201010212140.o9LLeA72063745@freefall.freebsd.org> To: freebsd-fs@FreeBSD.org From: Richard Conto Cc: Subject: Re: kern/147881: [zfs] [patch] ZFS "sharenfs" doesn't allow different "exports" options for different hosts X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list Reply-To: Richard Conto List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Thu, 21 Oct 2010 21:40:10 -0000 The following reply was made to PR kern/147881; it has been noted by GNATS. From: Richard Conto To: Martin Birgmeier Cc: bug-followup@FreeBSD.org, Richard Conto Subject: Re: kern/147881: [zfs] [patch] ZFS "sharenfs" doesn't allow different "exports" options for different hosts Date: Thu, 21 Oct 2010 17:10:19 -0400 My patch is IPv4/IPv6/DNS agnostic - it ought to work either way. NFSv4 = might be a problem due to the syntax differences in /etc/exports = (/etc/zfs/exports). All my patch does is use ';' as a special character to allow multiple = values (like you want) for each export. Each value is prefixed by the = mount point and written to /etc/zfs/exports. I'm still running it - but it's for a home file server where I want to = restrict "root" access from certain VPNs. Given that all it does is = change the output to /etc/zfs/exports, which is handled by the regular = NFS export mechanism, I don't think that there should be any significant = difference between home and production/enterprise use - unless NFS = exports keel over when there's hundreds of mount points (easy to do with = ZFS) rather than tens. My home set up is exporting only 25 mount = points. --- Richard On Oct 15, 2010, at 3:15 AM, Martin Birgmeier wrote: > Please implement this or a similar patch. I need this in my dual = IPv4/IPv6 setup, where it is necessary to have something like >=20 > /tank/vol -network 192.168.0.0 -mask 255.255.0.0 > /tank/vol -network fec0:0:0:1::/56 >=20 > in /etc/zfs/exports in order to get both IPv4 and IPv6 clients to = work. >=20 > Regards, >=20 > Martin From owner-freebsd-fs@FreeBSD.ORG Fri Oct 22 13:34:51 2010 Return-Path: Delivered-To: fs@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id 7CB5E1065670 for ; Fri, 22 Oct 2010 13:34:51 +0000 (UTC) (envelope-from jhellenthal@gmail.com) Received: from mail-gx0-f182.google.com (mail-gx0-f182.google.com [209.85.161.182]) by mx1.freebsd.org (Postfix) with ESMTP id 308B68FC1C for ; Fri, 22 Oct 2010 13:34:50 +0000 (UTC) Received: by gxk2 with SMTP id 2so619336gxk.13 for ; Fri, 22 Oct 2010 06:34:50 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=gamma; h=domainkey-signature:received:received:sender:message-id:date:from :organization:user-agent:mime-version:to:cc:subject:references :in-reply-to:x-enigmail-version:content-type :content-transfer-encoding; bh=0k/2z4GkvemSYO4lsHaN8p6m+Pletcri3jY9RyGcU1c=; b=BuuX7FjcBUYho349I9vqYk+0K/S+XzrwbgJt1NuIYoZ5bytxESKbrq2lrlnVyiW2yI WdOO1jYXYRikxVvlIESkkMOJ80Gls7S8/Xi274klPHtdtXEQrQ/xB095AT56YUZlGjwL g7ke9k+XC1XFFTgaMoHjSe2dZWd3h9O+vCsE4= DomainKey-Signature: a=rsa-sha1; c=nofws; d=gmail.com; s=gamma; h=sender:message-id:date:from:organization:user-agent:mime-version:to :cc:subject:references:in-reply-to:x-enigmail-version:content-type :content-transfer-encoding; b=vF7xZ4vtXHTM+ThLlStkt4kxNPEyQ8+8fUk27Qg4d7RSefr5IvsrsBSHqc4WSFjgwd ymRo8VZG9ItTkj6Oca+k4vSWUQmYw0RqMDB68UJ5WpThO40zkQENxnyrslNLMLTpU8Xq e6pzflJlTgYTlMxk9esuAvV7VWu7qkr00Rxrk= Received: by 10.229.189.74 with SMTP id dd10mr2225752qcb.73.1287752833553; Fri, 22 Oct 2010 06:07:13 -0700 (PDT) Received: from centel.dataix.local (adsl-99-181-136-243.dsl.klmzmi.sbcglobal.net [99.181.136.243]) by mx.google.com with ESMTPS id m7sm2555958qck.37.2010.10.22.06.07.11 (version=SSLv3 cipher=RC4-MD5); Fri, 22 Oct 2010 06:07:12 -0700 (PDT) Sender: "J. Hellenthal" Message-ID: <4CC18C7E.702@DataIX.net> Date: Fri, 22 Oct 2010 09:07:10 -0400 From: jhell Organization: http://www.DataIX.net User-Agent: Mozilla/5.0 (X11; U; FreeBSD i386; en-US; rv:1.9.2.11) Gecko/20101021 Lightning/1.0b1 Thunderbird MIME-Version: 1.0 To: Willem Jan Withagen References: <4CBD52F4.2010105@digiware.nl> In-Reply-To: <4CBD52F4.2010105@digiware.nl> X-Enigmail-Version: 1.1.2 Content-Type: text/plain; charset=ISO-8859-1 Content-Transfer-Encoding: 7bit Cc: fs@freebsd.org Subject: Re: ZFS destroy snapshot does not work X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Fri, 22 Oct 2010 13:34:51 -0000 On 10/19/2010 04:12, Willem Jan Withagen wrote: > Hi, > > Probably due to too many reboots when the system was livelocked. > But now I have the following problem. > > /sbin/zfs destroy -r zfsraid/backups@Monday > cannot destroy 'zfsraid/backups@Monday': dataset does not exist > no snapshots destroyed > /sbin/zfs snapshot -r zfsraid/backups@Monday > cannot create snapshot 'zfsraid/backups@Monday': dataset is busy > no snapshots were created > > /sbin/zfs destroy -r zfsraid/home@$DATE > cannot destroy 'zfsraid/home@Monday': dataset does not exist > no snapshots destroyed > /sbin/zfs snapshot -r zfsraid/home@Monday > cannot create snapshot 'zfsraid/home@Monday': dataset is busy > no snapshots were created > > Where this used to work(tm).... > > I can iterate over all filesystems and then destroy/create a backup. > > Except for one: > # zfs list -r zfsraid/home/trouble > NAME USED AVAIL REFER MOUNTPOINT > zfsraid/home/trouble 149G 3.49T 148G none > zfsraid/home/trouble@Sunday 1.06G - 148G - > > # zfs destroy -r zfsraid/home/trouble > cannot destroy 'zfsraid/home/trouble@Sunday': dataset already exists > Exit 1 > # zfs destroy zfsraid/home/trouble@Sunday > cannot destroy 'zfsraid/home/trouble@Sunday': dataset already exists > Exit 1 > # zfs destroy -r zfsraid/home/trouble@Sunday > cannot destroy 'zfsraid/home/trouble@Sunday': snapshot is cloned > no snapshots destroyed > > But of this last fact I do not remember cloning anything. Nor can I find > any suggestions as to the filesystem being cloned..... > > So how do I get ride of this rouge filesystem. > > There is something like zdb, but the manual page is purposely vague on > how to use that... > > Thanx, > --WjW > Can you give a ( zfs list -t snapshot ) here please? -- jhell,v From owner-freebsd-fs@FreeBSD.ORG Fri Oct 22 13:47:12 2010 Return-Path: Delivered-To: freebsd-fs@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id 74783106566C; Fri, 22 Oct 2010 13:47:12 +0000 (UTC) (envelope-from to.my.trociny@gmail.com) Received: from mail-bw0-f54.google.com (mail-bw0-f54.google.com [209.85.214.54]) by mx1.freebsd.org (Postfix) with ESMTP id BFC468FC18; Fri, 22 Oct 2010 13:47:11 +0000 (UTC) Received: by bwz3 with SMTP id 3so1225324bwz.13 for ; Fri, 22 Oct 2010 06:47:10 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=gamma; h=domainkey-signature:received:received:from:to:cc:subject:date :message-id:user-agent:mime-version:content-type; bh=9Cg8H+hHfZn80jBWSOPyyZZDOAd1AI3lXtzKsp0GVD8=; b=mKXfhyf6xLPFraE/arqiNWtPZmBOtsTtfAr1pZ2zNF2L152nRPZFvh4MAVsauHyBSP txPxGk106Qsl4kSXcti9hMluzbsrIYZMymXdqvApfOY06d8HhOwgihre+PK8pDJgG4OT 4Sj3yaWjTaHcbQPTb9BUFhsIJhhmuuMV/+5EA= DomainKey-Signature: a=rsa-sha1; c=nofws; d=gmail.com; s=gamma; h=from:to:cc:subject:date:message-id:user-agent:mime-version :content-type; b=CO/J8wP4K/7ErcyYixVKEEsk75/jBOOFPv7xmeOWBdqUodz1Y/OPzeVhmAdzR4JDzq CAxgeMmqwYPazBuygZYIcV87ch1BEULXaF2G+7ut0KlGISExU6ZakTMP9u2al+okPrMw hfo2XpfCDIrwuv3+nGdQGozzTdqRTq3iZv4/0= Received: by 10.204.63.9 with SMTP id z9mr2135814bkh.66.1287755230554; Fri, 22 Oct 2010 06:47:10 -0700 (PDT) Received: from localhost ([95.69.174.185]) by mx.google.com with ESMTPS id k4sm1422090faa.32.2010.10.22.06.47.08 (version=TLSv1/SSLv3 cipher=RC4-MD5); Fri, 22 Oct 2010 06:47:09 -0700 (PDT) From: Mikolaj Golub To: freebsd-fs@freebsd.org Date: Fri, 22 Oct 2010 16:47:08 +0300 Message-ID: <8662wuibkz.fsf@kopusha.home.net> User-Agent: Gnus/5.13 (Gnus v5.13) Emacs/23.2 (berkeley-unix) MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Cc: Pawel Jakub Dawidek Subject: hastmon: cluster monitoring daemon X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Fri, 22 Oct 2010 13:47:12 -0000 Hi, Some time ago when studding HAST code I realized that only small part of the code is HAST specific while the large part is a generic code that can be reused for a project where you need to implement network nodes and communication between them. So to prove this I implemented a simple cluster monitoring daemon that can be used e.g. to monitor HAST :-) and do automatic failover. http://code.google.com/p/hastmon/wiki/README -- Mikolaj Golub From owner-freebsd-fs@FreeBSD.ORG Fri Oct 22 20:19:36 2010 Return-Path: Delivered-To: fs@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id 411B3106566C for ; Fri, 22 Oct 2010 20:19:36 +0000 (UTC) (envelope-from wjw@digiware.nl) Received: from mail.digiware.nl (mail.ip6.digiware.nl [IPv6:2001:4cb8:1:106::2]) by mx1.freebsd.org (Postfix) with ESMTP id CEA318FC15 for ; Fri, 22 Oct 2010 20:19:35 +0000 (UTC) Received: from localhost (localhost.digiware.nl [127.0.0.1]) by mail.digiware.nl (Postfix) with ESMTP id ACEB4153434; Fri, 22 Oct 2010 22:19:34 +0200 (CEST) X-Virus-Scanned: amavisd-new at digiware.nl Received: from mail.digiware.nl ([127.0.0.1]) by localhost (rack1.digiware.nl [127.0.0.1]) (amavisd-new, port 10024) with ESMTP id kPKg9vZkGW3n; Fri, 22 Oct 2010 22:19:32 +0200 (CEST) Received: from [127.0.0.1] (vaio [192.168.10.10]) (using TLSv1 with cipher DHE-RSA-CAMELLIA256-SHA (256/256 bits)) (No client certificate requested) by mail.digiware.nl (Postfix) with ESMTPSA id 64851153433; Fri, 22 Oct 2010 22:19:32 +0200 (CEST) Message-ID: <4CC1F1D9.6040802@digiware.nl> Date: Fri, 22 Oct 2010 22:19:37 +0200 From: Willem Jan Withagen User-Agent: Mozilla/5.0 (Windows; U; Windows NT 6.1; en-US; rv:1.9.2.11) Gecko/20101013 Lightning/1.0b2 Thunderbird/3.1.5 MIME-Version: 1.0 To: jhell References: <4CBD52F4.2010105@digiware.nl> <4CC18C7E.702@DataIX.net> In-Reply-To: <4CC18C7E.702@DataIX.net> X-Enigmail-Version: 1.1.1 Content-Type: text/plain; charset=ISO-8859-1 Content-Transfer-Encoding: 7bit Cc: fs@freebsd.org Subject: Re: ZFS destroy snapshot does not work X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Fri, 22 Oct 2010 20:19:36 -0000 On 22-10-2010 15:07, jhell wrote: > On 10/19/2010 04:12, Willem Jan Withagen wrote: >> >> But of this last fact I do not remember cloning anything. Nor can I find >> any suggestions as to the filesystem being cloned..... >> >> So how do I get ride of this rouge filesystem. > > Can you give a ( zfs list -t snapshot ) here please? Sorry, I had to destroy the filesystem this morning, since the storage server on which the actual data is, needs to go back after the weekend. So I needed to transfer the data to a consistent new fs. Fortunatly I is again consistent :) --WjW From owner-freebsd-fs@FreeBSD.ORG Fri Oct 22 23:12:10 2010 Return-Path: Delivered-To: fs@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id E1880106564A for ; Fri, 22 Oct 2010 23:12:10 +0000 (UTC) (envelope-from gene@nttmcl.com) Received: from mx2.nttmcl.com (MX2.nttmcl.com [216.69.68.200]) by mx1.freebsd.org (Postfix) with ESMTP id C2A5F8FC12 for ; Fri, 22 Oct 2010 23:12:10 +0000 (UTC) Received: from localhost (localhost [127.0.0.1]) by mx2.nttmcl.com (Postfix) with ESMTP id 04C724DDA5 for ; Fri, 22 Oct 2010 15:53:24 -0700 (PDT) X-Spam-Flag: NO X-Spam-Score: -3.911 X-Spam-Level: X-Spam-Status: No, score=-3.911 tagged_above=-999 required=5 tests=[ALL_TRUSTED=-1.8, AWL=0.332, BAYES_00=-2.599, SUBJECT_FUZZY_TION=0.156] autolearn=no Received: from mx2.nttmcl.com ([127.0.0.1]) by localhost (mx2.nttmcl.com [127.0.0.1]) (amavisd-new, port 10024) with ESMTP id arG5Sgp3ZtYw for ; Fri, 22 Oct 2010 15:53:23 -0700 (PDT) Received: from [216.69.70.67] (dyn-v6-67.nttmcl.com [216.69.70.67]) by mx2.nttmcl.com (Postfix) with ESMTPS id 7ED9C4DD8C for ; Fri, 22 Oct 2010 15:53:23 -0700 (PDT) Message-ID: <4CC215B4.3050607@nttmcl.com> Date: Fri, 22 Oct 2010 15:52:36 -0700 From: "Eugene M. Kim" User-Agent: Mozilla/5.0 (Windows; U; Windows NT 6.0; en-US; rv:1.9.2.11) Gecko/20101013 Lightning/1.0b2 Thunderbird/3.1.5 MIME-Version: 1.0 To: fs@freebsd.org X-Enigmail-Version: 1.1.1 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 7bit Cc: Subject: ZFS: Parallel I/O to vdevs that appear to be separate physical disks but really are partitions X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Fri, 22 Oct 2010 23:12:11 -0000 -----BEGIN PGP SIGNED MESSAGE----- Hash: SHA1 Greetings, I run a FreeBSD guest in VMware ESXi with a 10GB zpool. Lately the originally provisioned 10GB proved insufficient, and I would like to provision another 10GB virtual disk and add it to the zpool as a top-level vdev. The original and new virtual disks are from the same physical pool (a RAID-5 array), but appears to be separate physical disks to the FreeBSD guest (da0 and da1). I am afraid that the ZFS would schedule I/O to the two virtual disks in parallel thinking that the pool performance will improve, while the performance would actually suffer due to seeking back and forth between two regions of one physical pool. 1. Will ZFS schedule parallel access to different top-level vdevs? 2. If so, how can I suggest ZFS that certain vdevs should be treated not as separate physical disks but as partitions on the same physical disk? Regards, Eugene -----BEGIN PGP SIGNATURE----- Version: GnuPG v1.4.10 (MingW32) Comment: Using GnuPG with Mozilla - http://enigmail.mozdev.org/ iQIcBAEBAgAGBQJMwhWzAAoJEHSb/9O6h+h00RQP/3qZh063VpiOA4eT78WJyGtQ I4hVdp6ehwgKJV3wcNDWCkSlTguzQcIijVngKFEqVV2HWSgVZcXg8SMeW4yEkbTz uGGfTI8yNbncQo4J6LtJMG6rj0jhORwPVCvDlM8DiwYzNOdfqYgDf7y8qZogOQOf FErbxaq/D5i3tryoIv+5oyfPQ9LzlmCHl5WhOshOLD0eH/MyGW1MyJY8CDlpYRDl zhMtNMiYNox1QCuZXrK3KvQgbdPMSnCq8VkwHpBPhUHc78q3dZOpQT1erblvQrhv YQDNHBmuevoMiicmpBwYuBW8M6MTfMuIZ/r6lwF0mTUMm0Lwprn8wPZJgJ2F4Axq J1JrCyHYZq+GfU654bCMu+CxhptQf0/Ojytx2Ubvif6L9Afm4GFZO6iVuHvrRVob EmAfpEOXSCOV87BzAp+z3W5RDKDaK7bRFdqZmBEf/StmoLnVEbbxdoqUNM+IvXBu DimW3JZENHBv8046fRJXiQs7KLsBdBEcgdyxJpjjztaIo6mFt6FtHP0RzMCPTLr/ uXBGKgTC9LBOy7lXIFl6JmulMQ7u3hBXzlGtAxMWU/jxAyMm/LypsY1V1k9pnBUb Pxcbv8uv+UO0J43Yiw/t+hGEQ7IFtF0gNSU/q5r9C8nMAdJj2+GhGsO8sJyTTu8J Qg7/z8GY1OlXDfHm95zR =fh8Y -----END PGP SIGNATURE----- From owner-freebsd-fs@FreeBSD.ORG Sat Oct 23 07:47:53 2010 Return-Path: Delivered-To: freebsd-fs@hub.freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id 8B5BF106564A; Sat, 23 Oct 2010 07:47:53 +0000 (UTC) (envelope-from linimon@FreeBSD.org) Received: from freefall.freebsd.org (freefall.freebsd.org [IPv6:2001:4f8:fff6::28]) by mx1.freebsd.org (Postfix) with ESMTP id 632008FC2C; Sat, 23 Oct 2010 07:47:53 +0000 (UTC) Received: from freefall.freebsd.org (localhost [127.0.0.1]) by freefall.freebsd.org (8.14.4/8.14.4) with ESMTP id o9N7lrCE020600; Sat, 23 Oct 2010 07:47:53 GMT (envelope-from linimon@freefall.freebsd.org) Received: (from linimon@localhost) by freefall.freebsd.org (8.14.4/8.14.4/Submit) id o9N7lrLi020596; Sat, 23 Oct 2010 07:47:53 GMT (envelope-from linimon) Date: Sat, 23 Oct 2010 07:47:53 GMT Message-Id: <201010230747.o9N7lrLi020596@freefall.freebsd.org> To: linimon@FreeBSD.org, freebsd-bugs@FreeBSD.org, freebsd-fs@FreeBSD.org From: linimon@FreeBSD.org Cc: Subject: Re: kern/151226: [zfs] can't delete zfs snapshot X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Sat, 23 Oct 2010 07:47:53 -0000 Old Synopsis: can't delete zfs snapshot New Synopsis: [zfs] can't delete zfs snapshot Responsible-Changed-From-To: freebsd-bugs->freebsd-fs Responsible-Changed-By: linimon Responsible-Changed-When: Sat Oct 23 07:47:40 UTC 2010 Responsible-Changed-Why: Over to maintainer(s). http://www.freebsd.org/cgi/query-pr.cgi?pr=151226 From owner-freebsd-fs@FreeBSD.ORG Sat Oct 23 07:56:12 2010 Return-Path: Delivered-To: freebsd-fs@hub.freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id EC8911065670; Sat, 23 Oct 2010 07:56:12 +0000 (UTC) (envelope-from linimon@FreeBSD.org) Received: from freefall.freebsd.org (freefall.freebsd.org [IPv6:2001:4f8:fff6::28]) by mx1.freebsd.org (Postfix) with ESMTP id B03838FC08; Sat, 23 Oct 2010 07:56:12 +0000 (UTC) Received: from freefall.freebsd.org (localhost [127.0.0.1]) by freefall.freebsd.org (8.14.4/8.14.4) with ESMTP id o9N7uCie030565; Sat, 23 Oct 2010 07:56:12 GMT (envelope-from linimon@freefall.freebsd.org) Received: (from linimon@localhost) by freefall.freebsd.org (8.14.4/8.14.4/Submit) id o9N7uCHA030561; Sat, 23 Oct 2010 07:56:12 GMT (envelope-from linimon) Date: Sat, 23 Oct 2010 07:56:12 GMT Message-Id: <201010230756.o9N7uCHA030561@freefall.freebsd.org> To: linimon@FreeBSD.org, freebsd-bugs@FreeBSD.org, freebsd-fs@FreeBSD.org From: linimon@FreeBSD.org Cc: Subject: Re: kern/151629: [fs] [patch] Skip empty directory entries during name lookup (unionfs, dirent_exists()) X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Sat, 23 Oct 2010 07:56:13 -0000 Synopsis: [fs] [patch] Skip empty directory entries during name lookup (unionfs, dirent_exists()) Responsible-Changed-From-To: freebsd-bugs->freebsd-fs Responsible-Changed-By: linimon Responsible-Changed-When: Sat Oct 23 07:56:02 UTC 2010 Responsible-Changed-Why: Over to maintainer(s). http://www.freebsd.org/cgi/query-pr.cgi?pr=151629 From owner-freebsd-fs@FreeBSD.ORG Sat Oct 23 07:57:59 2010 Return-Path: Delivered-To: freebsd-fs@hub.freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id 86329106566C; Sat, 23 Oct 2010 07:57:59 +0000 (UTC) (envelope-from linimon@FreeBSD.org) Received: from freefall.freebsd.org (freefall.freebsd.org [IPv6:2001:4f8:fff6::28]) by mx1.freebsd.org (Postfix) with ESMTP id 5D2FB8FC12; Sat, 23 Oct 2010 07:57:59 +0000 (UTC) Received: from freefall.freebsd.org (localhost [127.0.0.1]) by freefall.freebsd.org (8.14.4/8.14.4) with ESMTP id o9N7vxr0030861; Sat, 23 Oct 2010 07:57:59 GMT (envelope-from linimon@freefall.freebsd.org) Received: (from linimon@localhost) by freefall.freebsd.org (8.14.4/8.14.4/Submit) id o9N7vxL1030857; Sat, 23 Oct 2010 07:57:59 GMT (envelope-from linimon) Date: Sat, 23 Oct 2010 07:57:59 GMT Message-Id: <201010230757.o9N7vxL1030857@freefall.freebsd.org> To: linimon@FreeBSD.org, freebsd-bugs@FreeBSD.org, freebsd-fs@FreeBSD.org From: linimon@FreeBSD.org Cc: Subject: Re: kern/151648: [zfs] disk wait bug X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Sat, 23 Oct 2010 07:57:59 -0000 Synopsis: [zfs] disk wait bug Responsible-Changed-From-To: freebsd-bugs->freebsd-fs Responsible-Changed-By: linimon Responsible-Changed-When: Sat Oct 23 07:57:51 UTC 2010 Responsible-Changed-Why: Over to maintainer(s). http://www.freebsd.org/cgi/query-pr.cgi?pr=151648 From owner-freebsd-fs@FreeBSD.ORG Sat Oct 23 08:55:40 2010 Return-Path: Delivered-To: freebsd-fs@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id 89A87106564A for ; Sat, 23 Oct 2010 08:55:40 +0000 (UTC) (envelope-from kpielorz_lst@tdx.co.uk) Received: from mail.tdx.com (mail.tdx.com [62.13.128.18]) by mx1.freebsd.org (Postfix) with ESMTP id 0C48B8FC0A for ; Sat, 23 Oct 2010 08:55:39 +0000 (UTC) Received: from Octa64 (octa64.tdx.co.uk [62.13.130.232]) (authenticated bits=0) by mail.tdx.com (8.14.3/8.14.3/Kp) with ESMTP id o9N8tb2U082533; Sat, 23 Oct 2010 09:55:37 +0100 (BST) Date: Sat, 23 Oct 2010 09:56:43 +0100 From: Karl Pielorz To: Pete French , fjwcash@gmail.com, wjw@digiware.nl Message-ID: In-Reply-To: References: X-Mailer: Mulberry/4.0.8 (Win32) MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii; format=flowed Content-Transfer-Encoding: 7bit Content-Disposition: inline Cc: freebsd-fs@freebsd.org Subject: Re: ZFS 'read-only' device / pool scan / import? X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Sat, 23 Oct 2010 08:55:40 -0000 --On 20 October 2010 11:35 +0100 Pete French wrote: > You can do a forced export - as long as the system doesnt think > the pool is imported then an import will go looking. I made a > similar mess of my drives the first time I changed discs oover, and I > also didnt know about the import/export trick. But since then I > have used iit every time I have got into trouble and it has always > worked for me. As a follow up to this - we tried to reproduce the original problem once we'd restored the data. We removed a drive - rebooted, and got 'a mess': " NAME STATE READ WRITE CKSUM vol UNAVAIL 0 0 0 insufficient replicas mirror ONLINE 0 0 0 da1 ONLINE 0 0 0 da2 ONLINE 0 0 0 da3 UNAVAIL 0 0 0 corrupted data mirror UNAVAIL 0 0 0 insufficient replicas da4 FAULTED 0 0 0 corrupted data da5 FAULTED 0 0 0 corrupted data da6 FAULTED 0 0 0 corrupted data mirror UNAVAIL 0 0 0 insufficient replicas da7 FAULTED 0 0 0 corrupted data da8 FAULTED 0 0 0 corrupted data da9 FAULTED 0 0 0 corrupted data mirror UNAVAIL 0 0 0 insufficient replicas da10 FAULTED 0 0 0 corrupted data da11 FAULTED 0 0 0 corrupted data " Doing a zpool export, then a zpool import (which required '-f' to get past "cannot import 'vol': pool may be in use from other system use '-f' to import anyway") Seems to have done the trick: " vol ONLINE mirror ONLINE da1 ONLINE da2 ONLINE da4 ONLINE mirror ONLINE da5 ONLINE da6 ONLINE da7 ONLINE mirror ONLINE da8 ONLINE da9 ONLINE da10 ONLINE mirror ONLINE da11 ONLINE da12 ONLINE " We now just have to re-attach the missing da3 to the last mirror set... Thanks to all who replied - this has now all been written up in that machines notes, Regards, -Karl From owner-freebsd-fs@FreeBSD.ORG Sat Oct 23 08:59:13 2010 Return-Path: Delivered-To: fs@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id 14B8C106564A for ; Sat, 23 Oct 2010 08:59:13 +0000 (UTC) (envelope-from pjd@garage.freebsd.pl) Received: from mail.garage.freebsd.pl (60.wheelsystems.com [83.12.187.60]) by mx1.freebsd.org (Postfix) with ESMTP id B22418FC0C for ; Sat, 23 Oct 2010 08:59:12 +0000 (UTC) Received: by mail.garage.freebsd.pl (Postfix, from userid 65534) id 210C245CD9; Sat, 23 Oct 2010 10:59:11 +0200 (CEST) Received: from localhost (chello089073192049.chello.pl [89.73.192.49]) (using TLSv1 with cipher DHE-RSA-AES256-SHA (256/256 bits)) (No client certificate requested) by mail.garage.freebsd.pl (Postfix) with ESMTP id 7100345C99; Sat, 23 Oct 2010 10:59:05 +0200 (CEST) Date: Sat, 23 Oct 2010 10:58:32 +0200 From: Pawel Jakub Dawidek To: "Eugene M. Kim" Message-ID: <20101023085832.GD1742@garage.freebsd.pl> References: <4CC215B4.3050607@nttmcl.com> Mime-Version: 1.0 Content-Type: multipart/signed; micalg=pgp-sha1; protocol="application/pgp-signature"; boundary="NklN7DEeGtkPCoo3" Content-Disposition: inline In-Reply-To: <4CC215B4.3050607@nttmcl.com> User-Agent: Mutt/1.4.2.3i X-PGP-Key-URL: http://people.freebsd.org/~pjd/pjd.asc X-OS: FreeBSD 9.0-CURRENT amd64 X-Spam-Checker-Version: SpamAssassin 3.0.4 (2005-06-05) on mail.garage.freebsd.pl X-Spam-Level: X-Spam-Status: No, score=-0.6 required=4.5 tests=BAYES_00,RCVD_IN_SORBS_DUL autolearn=no version=3.0.4 Cc: fs@freebsd.org Subject: Re: ZFS: Parallel I/O to vdevs that appear to be separate physical disks but really are partitions X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Sat, 23 Oct 2010 08:59:13 -0000 --NklN7DEeGtkPCoo3 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline Content-Transfer-Encoding: quoted-printable On Fri, Oct 22, 2010 at 03:52:36PM -0700, Eugene M. Kim wrote: >=20 > -----BEGIN PGP SIGNED MESSAGE----- > Hash: SHA1 > =20 > Greetings, >=20 > I run a FreeBSD guest in VMware ESXi with a 10GB zpool. Lately the > originally provisioned 10GB proved insufficient, and I would like to > provision another 10GB virtual disk and add it to the zpool as a > top-level vdev. >=20 > The original and new virtual disks are from the same physical pool (a > RAID-5 array), but appears to be separate physical disks to the > FreeBSD guest (da0 and da1). I am afraid that the ZFS would schedule > I/O to the two virtual disks in parallel thinking that the pool > performance will improve, while the performance would actually suffer > due to seeking back and forth between two regions of one physical pool. >=20 > 1. Will ZFS schedule parallel access to different top-level vdevs? Yes, ZFS stripes requests over all top level vdevs. > 2. If so, how can I suggest ZFS that certain vdevs should be treated > not as separate physical disks but as partitions on the same physical > disk? Nope, that's not possible. Instead of adding next 10GB disk maybe it is possible to grow the one you have? ZFS should be able to grow automatically. --=20 Pawel Jakub Dawidek http://www.wheelsystems.com pjd@FreeBSD.org http://www.FreeBSD.org FreeBSD committer Am I Evil? Yes, I Am! --NklN7DEeGtkPCoo3 Content-Type: application/pgp-signature Content-Disposition: inline -----BEGIN PGP SIGNATURE----- Version: GnuPG v2.0.14 (FreeBSD) iEYEARECAAYFAkzCo7cACgkQForvXbEpPzT/YACcDq6bxtlR7oLaRFUZeaQ5nDzM X9gAoK36XBJcZsabNBwbKn8+I06l367T =RNLw -----END PGP SIGNATURE----- --NklN7DEeGtkPCoo3-- From owner-freebsd-fs@FreeBSD.ORG Sat Oct 23 10:03:52 2010 Return-Path: Delivered-To: freebsd-fs@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id CB7161065672 for ; Sat, 23 Oct 2010 10:03:52 +0000 (UTC) (envelope-from pjd@garage.freebsd.pl) Received: from mail.garage.freebsd.pl (60.wheelsystems.com [83.12.187.60]) by mx1.freebsd.org (Postfix) with ESMTP id 3E6D18FC14 for ; Sat, 23 Oct 2010 10:03:44 +0000 (UTC) Received: by mail.garage.freebsd.pl (Postfix, from userid 65534) id 2F7FD45C9F; Sat, 23 Oct 2010 12:03:43 +0200 (CEST) Received: from localhost (chello089073192049.chello.pl [89.73.192.49]) (using TLSv1 with cipher DHE-RSA-AES256-SHA (256/256 bits)) (No client certificate requested) by mail.garage.freebsd.pl (Postfix) with ESMTP id 3BB3345E13; Sat, 23 Oct 2010 12:03:37 +0200 (CEST) Date: Sat, 23 Oct 2010 12:03:03 +0200 From: Pawel Jakub Dawidek To: Karl Pielorz Message-ID: <20101023100303.GE1742@garage.freebsd.pl> References: Mime-Version: 1.0 Content-Type: multipart/signed; micalg=pgp-sha1; protocol="application/pgp-signature"; boundary="cPi+lWm09sJ+d57q" Content-Disposition: inline In-Reply-To: User-Agent: Mutt/1.4.2.3i X-PGP-Key-URL: http://people.freebsd.org/~pjd/pjd.asc X-OS: FreeBSD 9.0-CURRENT amd64 X-Spam-Checker-Version: SpamAssassin 3.0.4 (2005-06-05) on mail.garage.freebsd.pl X-Spam-Level: X-Spam-Status: No, score=-0.6 required=4.5 tests=BAYES_00,RCVD_IN_SORBS_DUL autolearn=no version=3.0.4 Cc: freebsd-fs@freebsd.org Subject: Re: ZFS 'read-only' device / pool scan / import? X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Sat, 23 Oct 2010 10:03:52 -0000 --cPi+lWm09sJ+d57q Content-Type: text/plain; charset=us-ascii Content-Disposition: inline Content-Transfer-Encoding: quoted-printable On Sat, Oct 23, 2010 at 09:56:43AM +0100, Karl Pielorz wrote: >=20 > --On 20 October 2010 11:35 +0100 Pete French =20 > wrote: >=20 > >You can do a forced export - as long as the system doesnt think > >the pool is imported then an import will go looking. I made a > >similar mess of my drives the first time I changed discs oover, and I > >also didnt know about the import/export trick. But since then I > >have used iit every time I have got into trouble and it has always > >worked for me. >=20 > As a follow up to this - we tried to reproduce the original problem once= =20 > we'd restored the data. We removed a drive - rebooted, and got 'a mess': >=20 > " > NAME STATE READ WRITE CKSUM > vol UNAVAIL 0 0 0 insufficient replicas > mirror ONLINE 0 0 0 > da1 ONLINE 0 0 0 > da2 ONLINE 0 0 0 > da3 UNAVAIL 0 0 0 corrupted data > mirror UNAVAIL 0 0 0 insufficient replicas > da4 FAULTED 0 0 0 corrupted data > da5 FAULTED 0 0 0 corrupted data > da6 FAULTED 0 0 0 corrupted data > mirror UNAVAIL 0 0 0 insufficient replicas > da7 FAULTED 0 0 0 corrupted data > da8 FAULTED 0 0 0 corrupted data > da9 FAULTED 0 0 0 corrupted data > mirror UNAVAIL 0 0 0 insufficient replicas > da10 FAULTED 0 0 0 corrupted data > da11 FAULTED 0 0 0 corrupted data > " Your analysis is incorrect, I'm afraid. ZFS in FreeBSD can cope really well with disks name changes. Your problem is somewhere else. My guess is that you booted your machine in single-user mode and you didn't run: # /etc/rc.d/hostid start ZFS has protection for SAN/NAS environments where it keeps hostid in its metadata, so when you try to import the pool on different machine by accident it will warn you and will require -f option if you really know what you are doing. The rc.d/hostid script is there to configure hostid for your system. It is not executed in single-user mode, so hostid of your system is not set properly. I do agree that running /etc/rc.d/hostid in single-user mode should be documented better. > Doing a zpool export, then a zpool import (which required '-f' to get pas= t=20 > "cannot import 'vol': pool may be in use from other system use '-f' to=20 > import anyway") Yes, -f is required, because hostid is now different (uninitialized), but really you don't need export/import cycle if you first set your hostid right. If you could repeat your test, but executing the following command in the following order once you boot into single-user mode: # /etc/rc.d/hostid start # zpool status > Seems to have done the trick: [...] When you now boot in multi-user mode I guess you will have the same problem, because now ZFS stored invalid (uninitialized) hostid in its metadata. As for read-only imports. This is not possible in the ZFS version we have now, but such functionality is available in most recent ZFS, which is not yet in FreeBSD. --=20 Pawel Jakub Dawidek http://www.wheelsystems.com pjd@FreeBSD.org http://www.FreeBSD.org FreeBSD committer Am I Evil? Yes, I Am! --cPi+lWm09sJ+d57q Content-Type: application/pgp-signature Content-Disposition: inline -----BEGIN PGP SIGNATURE----- Version: GnuPG v2.0.14 (FreeBSD) iEYEARECAAYFAkzCstcACgkQForvXbEpPzRApACdHTQFEAUFX+KKVr19JMeI4iYA 9p8AoM+edvhy3j2TEp0STxORmGyV9h76 =Omqk -----END PGP SIGNATURE----- --cPi+lWm09sJ+d57q-- From owner-freebsd-fs@FreeBSD.ORG Sat Oct 23 11:00:22 2010 Return-Path: Delivered-To: freebsd-fs@hub.freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id 29EA51065679 for ; Sat, 23 Oct 2010 11:00:22 +0000 (UTC) (envelope-from gnats@FreeBSD.org) Received: from freefall.freebsd.org (freefall.freebsd.org [IPv6:2001:4f8:fff6::28]) by mx1.freebsd.org (Postfix) with ESMTP id CBFFA8FC13 for ; Sat, 23 Oct 2010 11:00:21 +0000 (UTC) Received: from freefall.freebsd.org (localhost [127.0.0.1]) by freefall.freebsd.org (8.14.4/8.14.4) with ESMTP id o9NB0LXT022654 for ; Sat, 23 Oct 2010 11:00:21 GMT (envelope-from gnats@freefall.freebsd.org) Received: (from gnats@localhost) by freefall.freebsd.org (8.14.4/8.14.4/Submit) id o9NB0LYX022606; Sat, 23 Oct 2010 11:00:21 GMT (envelope-from gnats) Date: Sat, 23 Oct 2010 11:00:21 GMT Message-Id: <201010231100.o9NB0LYX022606@freefall.freebsd.org> To: freebsd-fs@FreeBSD.org From: Andriy Gapon Cc: Subject: Re: kern/151648: [zfs] disk wait bug X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list Reply-To: Andriy Gapon List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Sat, 23 Oct 2010 11:00:22 -0000 The following reply was made to PR kern/151648; it has been noted by GNATS. From: Andriy Gapon To: bug-followup@FreeBSD.org, b.nowicki@admin.net.pl Cc: Subject: Re: kern/151648: [zfs] disk wait bug Date: Sat, 23 Oct 2010 11:58:59 +0300 Please provide procstat -kk -a output. -- Andriy Gapon From owner-freebsd-fs@FreeBSD.ORG Sat Oct 23 11:59:48 2010 Return-Path: Delivered-To: freebsd-fs@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id E2126106566C; Sat, 23 Oct 2010 11:59:48 +0000 (UTC) (envelope-from kpielorz_lst@tdx.co.uk) Received: from mail.tdx.com (mail.tdx.com [62.13.128.18]) by mx1.freebsd.org (Postfix) with ESMTP id 64FA38FC1F; Sat, 23 Oct 2010 11:59:48 +0000 (UTC) Received: from Octa64 (octa64.tdx.co.uk [62.13.130.232]) (authenticated bits=0) by mail.tdx.com (8.14.3/8.14.3/Kp) with ESMTP id o9NBxl7v097542; Sat, 23 Oct 2010 12:59:47 +0100 (BST) Date: Sat, 23 Oct 2010 13:00:53 +0100 From: Karl Pielorz To: Pawel Jakub Dawidek Message-ID: <9C1A2093A817726315BAA098@Octa64> In-Reply-To: <20101023100303.GE1742@garage.freebsd.pl> References: <20101023100303.GE1742@garage.freebsd.pl> X-Mailer: Mulberry/4.0.8 (Win32) MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii; format=flowed Content-Transfer-Encoding: 7bit Content-Disposition: inline Cc: freebsd-fs@freebsd.org Subject: Re: ZFS 'read-only' device / pool scan / import? X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Sat, 23 Oct 2010 11:59:49 -0000 --On 23 October 2010 12:03 +0200 Pawel Jakub Dawidek wrote: > As for read-only imports. This is not possible in the ZFS version we > have now, but such functionality is available in most recent ZFS, which > is not yet in FreeBSD. Thanks, that just answered my original question perfectly :) At least we know the mess we got into is recoverable, and numerous lessons have been learnt - and that read only imports (which won't change anything on the pool) should, hopefully be with us at some point :) Regards, -Karl From owner-freebsd-fs@FreeBSD.ORG Sat Oct 23 16:10:13 2010 Return-Path: Delivered-To: freebsd-fs@hub.freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id 98F261065760 for ; Sat, 23 Oct 2010 16:10:13 +0000 (UTC) (envelope-from gnats@FreeBSD.org) Received: from freefall.freebsd.org (freefall.freebsd.org [IPv6:2001:4f8:fff6::28]) by mx1.freebsd.org (Postfix) with ESMTP id 6C69D8FC16 for ; Sat, 23 Oct 2010 16:10:13 +0000 (UTC) Received: from freefall.freebsd.org (localhost [127.0.0.1]) by freefall.freebsd.org (8.14.4/8.14.4) with ESMTP id o9NGADLx044164 for ; Sat, 23 Oct 2010 16:10:13 GMT (envelope-from gnats@freefall.freebsd.org) Received: (from gnats@localhost) by freefall.freebsd.org (8.14.4/8.14.4/Submit) id o9NGADp7044163; Sat, 23 Oct 2010 16:10:13 GMT (envelope-from gnats) Date: Sat, 23 Oct 2010 16:10:13 GMT Message-Id: <201010231610.o9NGADp7044163@freefall.freebsd.org> To: freebsd-fs@FreeBSD.org From: "Bartosz Nowicki" Cc: Subject: Re: kern/151648: [zfs] disk wait bug X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list Reply-To: Bartosz Nowicki List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Sat, 23 Oct 2010 16:10:13 -0000 The following reply was made to PR kern/151648; it has been noted by GNATS. From: "Bartosz Nowicki" To: "Andriy Gapon" , bug-followup@freebsd.org Cc: Subject: Re: kern/151648: [zfs] disk wait bug Date: Sat, 23 Oct 2010 17:43:36 +0200 Dnia 23-10-2010 o 10:58:59 Andriy Gapon napisał(a): > Please provide procstat -kk -a output. I would rather not reveal all processes (production OS, many user accounts), but discovered that it's not only when using nginx. Problem also appears when trying to access /usr/home directory: # truss -o ls.truss -f ls /usr/home 95580: __sysctl(0x7fffffffdde0,0x2,0x7fffffffddfc,0x7fffffffddf0,0x0,0x0) = 0 (0x0) 95580: mmap(0x0,656,PROT_READ|PROT_WRITE,MAP_ANON,-1,0x0) = 34365190144 (0x800533000) 95580: munmap(0x800533000,656) = 0 (0x0) 95580: __sysctl(0x7fffffffde50,0x2,0x80063d6c8,0x7fffffffde48,0x0,0x0) = 0 (0x0) 95580: mmap(0x0,32768,PROT_READ|PROT_WRITE,MAP_PRIVATE|MAP_ANON,-1,0x0) = 34365190144 (0x800533000) 95580: issetugid(0x800534015,0x80052eb04,0x800649ef0,0x800649ec0,0x5511,0x0) = 0 (0x0) 95580: open("/etc/libmap.conf",O_RDONLY,0666) ERR#2 'No such file or directory' 95580: open("/var/run/ld-elf.so.hints",O_RDONLY,057) = 3 (0x3) 95580: read(3,"Ehnt\^A\0\0\0\M^@\0\0\0\M^S\0\0"...,128) = 128 (0x80) 95580: lseek(3,0x80,SEEK_SET) = 128 (0x80) 95580: read(3,"/lib:/usr/lib:/usr/lib/compat:/u"...,147) = 147 (0x93) 95580: close(3) = 0 (0x0) 95580: access("/lib/libutil.so.8",0) = 0 (0x0) 95580: open("/lib/libutil.so.8",O_RDONLY,030752640) = 3 (0x3) 95580: fstat(3,{ mode=-r--r--r-- ,inode=4120,size=62688,blksize=62976 }) = 0 (0x0) 95580: pread(0x3,0x80063c580,0x1000,0x0,0x101010101010101,0x8080808080808080) = 4096 (0x1000) 95580: mmap(0x0,1110016,PROT_NONE,MAP_PRIVATE|MAP_ANON|MAP_NOCORE,-1,0x0) = 34366332928 (0x80064a000) 95580: mmap(0x80064a000,53248,PROT_READ|PROT_EXEC,MAP_PRIVATE|MAP_FIXED|MAP_NOCORE,3,0x0) = 34366332928 (0x80064a000) 95580: mmap(0x800756000,12288,PROT_READ|PROT_WRITE,MAP_PRIVATE|MAP_FIXED,3,0xc000) = 34367430656 (0x800756000) 95580: close(3) = 0 (0x0) 95580: access("/lib/libncurses.so.8",0) = 0 (0x0) 95580: open("/lib/libncurses.so.8",O_RDONLY,030752640) = 3 (0x3) 95580: fstat(3,{ mode=-r--r--r-- ,inode=4114,size=313128,blksize=131072 }) = 0 (0x0) 95580: pread(0x3,0x80063c580,0x1000,0x0,0x101010101010101,0x8080808080808080) = 4096 (0x1000) 95580: mmap(0x0,1359872,PROT_NONE,MAP_PRIVATE|MAP_ANON|MAP_NOCORE,-1,0x0) = 34367442944 (0x800759000) 95580: mmap(0x800759000,270336,PROT_READ|PROT_EXEC,MAP_PRIVATE|MAP_FIXED|MAP_NOCORE,3,0x0) = 34367442944 (0x800759000) 95580: mmap(0x80089b000,36864,PROT_READ|PROT_WRITE,MAP_PRIVATE|MAP_FIXED,3,0x42000) = 34368761856 (0x80089b000) 95580: mprotect(0x8008a4000,4096,PROT_READ|PROT_WRITE) = 0 (0x0) 95580: close(3) = 0 (0x0) 95580: access("/lib/libc.so.7",0) = 0 (0x0) 95580: open("/lib/libc.so.7",O_RDONLY,030752640) = 3 (0x3) 95580: fstat(3,{ mode=-r--r--r-- ,inode=4104,size=1254032,blksize=131072 }) = 0 (0x0) 95580: pread(0x3,0x80063c580,0x1000,0x0,0x101010101010101,0x8080808080808080) = 4096 (0x1000) 95580: mmap(0x0,2330624,PROT_NONE,MAP_PRIVATE|MAP_ANON|MAP_NOCORE,-1,0x0) = 34368802816 (0x8008a5000) 95580: mmap(0x8008a5000,1036288,PROT_READ|PROT_EXEC,MAP_PRIVATE|MAP_FIXED|MAP_NOCORE,3,0x0) = 34368802816 (0x8008a5000) 95580: mmap(0x800aa2000,131072,PROT_READ|PROT_WRITE,MAP_PRIVATE|MAP_FIXED,3,0xfd000) = 34370887680 (0x800aa2000) 95580: mprotect(0x800ac2000,114688,PROT_READ|PROT_WRITE) = 0 (0x0) 95580: close(3) = 0 (0x0) 95580: sysarch(0x81,0x7fffffffded0,0x800538188,0x0,0xffffffffffa8c550,0x8008c0e78) = 0 (0x0) 95580: mmap(0x0,1408,PROT_READ|PROT_WRITE,MAP_ANON,-1,0x0) = 34365222912 (0x80053b000) 95580: munmap(0x80053b000,1408) = 0 (0x0) 95580: mmap(0x0,4240,PROT_READ|PROT_WRITE,MAP_ANON,-1,0x0) = 34365222912 (0x80053b000) 95580: munmap(0x80053b000,4240) = 0 (0x0) 95580: mmap(0x0,10096,PROT_READ|PROT_WRITE,MAP_ANON,-1,0x0) = 34365222912 (0x80053b000) 95580: munmap(0x80053b000,10096) = 0 (0x0) 95580: mmap(0x0,43696,PROT_READ|PROT_WRITE,MAP_ANON,-1,0x0) = 34365222912 (0x80053b000) 95580: munmap(0x80053b000,43696) = 0 (0x0) 95580: sigprocmask(SIG_BLOCK,SIGHUP|SIGINT|SIGQUIT|SIGKILL|SIGPIPE|SIGALRM|SIGTERM|SIGURG|SIGSTOP|SIGTSTP|SIGCONT|SIGCHLD|SIGTTIN|SIGTTOU|SIGIO|SIGXCPU|SIGXFSZ|SIGVTALRM|SIGPROF|SIGWINCH|SIGINFO|SIGUSR1|SIGUSR2,0x0) = 0 (0x0) 95580: sigprocmask(SIG_SETMASK,0x0,0x0) = 0 (0x0) 95580: __sysctl(0x7fffffffde60,0x2,0x506b00,0x7fffffffde58,0x0,0x0) = 0 (0x0) 95580: sigprocmask(SIG_BLOCK,SIGHUP|SIGINT|SIGQUIT|SIGKILL|SIGPIPE|SIGALRM|SIGTERM|SIGURG|SIGSTOP|SIGTSTP|SIGCONT|SIGCHLD|SIGTTIN|SIGTTOU|SIGIO|SIGXCPU|SIGXFSZ|SIGVTALRM|SIGPROF|SIGWINCH|SIGINFO|SIGUSR1|SIGUSR2,0x0) = 0 (0x0) 95580: sigprocmask(SIG_SETMASK,0x0,0x0) = 0 (0x0) 95580: sigprocmask(SIG_BLOCK,SIGHUP|SIGINT|SIGQUIT|SIGKILL|SIGPIPE|SIGALRM|SIGTERM|SIGURG|SIGSTOP|SIGTSTP|SIGCONT|SIGCHLD|SIGTTIN|SIGTTOU|SIGIO|SIGXCPU|SIGXFSZ|SIGVTALRM|SIGPROF|SIGWINCH|SIGINFO|SIGUSR1|SIGUSR2,0x0) = 0 (0x0) 95580: sigprocmask(SIG_SETMASK,0x0,0x0) = 0 (0x0) 95580: sigprocmask(SIG_BLOCK,SIGHUP|SIGINT|SIGQUIT|SIGKILL|SIGPIPE|SIGALRM|SIGTERM|SIGURG|SIGSTOP|SIGTSTP|SIGCONT|SIGCHLD|SIGTTIN|SIGTTOU|SIGIO|SIGXCPU|SIGXFSZ|SIGVTALRM|SIGPROF|SIGWINCH|SIGINFO|SIGUSR1|SIGUSR2,0x0) = 0 (0x0) 95580: sigprocmask(SIG_SETMASK,0x0,0x0) = 0 (0x0) 95580: open("/usr/share/locale/pl_PL.ISO8859-2/LC_COLLATE",O_RDONLY,0666) = 3 (0x3) 95580: fstat(3,{ mode=-r--r--r-- ,inode=73575,size=4642,blksize=5120 }) = 0 (0x0) 95580: __sysctl(0x7fffffffd4d0,0x2,0x800ac85c8,0x7fffffffd4c8,0x0,0x0) = 0 (0x0) 95580: __sysctl(0x7fffffffd400,0x2,0x7fffffffd390,0x7fffffffd3f8,0x800994c60,0xc) = 0 (0x0) 95580: __sysctl(0x7fffffffd390,0x2,0x800ac87d0,0x7fffffffd450,0x0,0x0) = 0 (0x0) 95580: readlink("/etc/malloc.conf",0x7fffffffd4f0,1024) ERR#2 'No such file or directory' 95580: issetugid(0x800993919,0x7fffffffd4f0,0xffffffffffffffff,0x0,0x2,0x0) = 0 (0x0) 95580: break(0x600000) = 0 (0x0) 95580: __sysctl(0x7fffffffd780,0x2,0x7fffffffd79c,0x7fffffffd790,0x0,0x0) = 0 (0x0) 95580: mmap(0x0,2097152,PROT_READ|PROT_WRITE,MAP_PRIVATE|MAP_ANON,-1,0x0) = 34371133440 (0x800ade000) 95580: mmap(0x800cde000,1187840,PROT_READ|PROT_WRITE,MAP_PRIVATE|MAP_ANON,-1,0x0) = 34373230592 (0x800cde000) 95580: munmap(0x800ade000,1187840) = 0 (0x0) 95580: read(3,"1.2\n\0\0\0\0\0\0\0\0\0\^A\0\0\0"...,5120) = 4642 (0x1222) 95580: close(3) = 0 (0x0) 95580: open("/usr/share/locale/pl_PL.ISO8859-2/LC_CTYPE",O_RDONLY,0666) = 3 (0x3) 95580: fstat(3,{ mode=-r--r--r-- ,inode=75239,size=3168,blksize=4096 }) = 0 (0x0) 95580: fstat(3,{ mode=-r--r--r-- ,inode=75239,size=3168,blksize=4096 }) = 0 (0x0) 95580: lseek(3,0x0,SEEK_CUR) = 0 (0x0) 95580: lseek(3,0x0,SEEK_SET) = 0 (0x0) 95580: read(3,"RuneMag1NONE\0\0\0\0\0\0\0\0\0\0"...,4096) = 3168 (0xc60) 95580: close(3) = 0 (0x0) 95580: open("/usr/share/locale/pl_PL.ISO8859-2/LC_MONETARY",O_RDONLY,013720646000) = 3 (0x3) 95580: fstat(3,{ mode=-r--r--r-- ,inode=75444,size=35,blksize=4096 }) = 0 (0x0) 95580: read(3,"PLN \nz\M-3\n,\n \n3;3\n\n-\n2\n"...,35) = 35 (0x23) 95580: close(3) = 0 (0x0) 95580: open("/usr/share/locale/pl_PL.ISO8859-2/LC_NUMERIC",O_RDONLY,00) = 3 (0x3) 95580: fstat(3,{ mode=-r--r--r-- ,inode=75762,size=8,blksize=4096 }) = 0 (0x0) 95580: read(3,",\n \n3;3\n",8) = 8 (0x8) 95580: close(3) = 0 (0x0) 95580: open("/usr/share/locale/pl_PL.ISO8859-2/LC_MESSAGES",O_RDONLY,036764463000) = 3 (0x3) 95580: fstat(3,{ mode=-r--r--r-- ,inode=75608,size=26,blksize=4096 }) = 0 (0x0) 95580: read(3,"^[tTyY].*\n^[nN].*\ntak\nnie\n",26) = 26 (0x1a) 95580: close(3) = 0 (0x0) 95580: ioctl(1,TIOCGETA,0xffffdf10) = 0 (0x0) 95580: ioctl(1,TIOCGWINSZ,0xffffdf70) = 0 (0x0) 95580: getuid() = 0 (0x0) 95580: stat("/usr/home",{ mode=drwxr-xr-x ,inode=3,size=269,blksize=17408 }) = 0 (0x0) 95580: open(".",O_RDONLY,00) = 3 (0x3) 95580: fchdir(0x3,0x19,0x0,0x2,0x0,0x0) = 0 (0x0) 95580: stat("/usr/home",{ mode=drwxr-xr-x ,inode=3,size=269,blksize=17408 }) # mount mydevil on / (zfs, local) devfs on /dev (devfs, local, multilabel) mydevil/tmp on /tmp (zfs, local, nosuid) mydevil/usr on /usr (zfs, local) mydevil/usr/home on /usr/home (zfs, local) mydevil/usr/ports on /usr/ports (zfs, local, nosuid) mydevil/usr/ports/distfiles on /usr/ports/distfiles (zfs, local, noexec, nosuid) mydevil/usr/ports/packages on /usr/ports/packages (zfs, local, noexec, nosuid) mydevil/usr/src on /usr/src (zfs, local, noexec, nosuid) mydevil/var on /var (zfs, local) mydevil/var/crash on /var/crash (zfs, local, noexec, nosuid) mydevil/var/db on /var/db (zfs, local, noexec, nosuid) mydevil/var/db/pkg on /var/db/pkg (zfs, local, nosuid) mydevil/var/empty on /var/empty (zfs, local, noexec, nosuid, read-only) mydevil/var/mail on /var/mail (zfs, local, noexec, nosuid) mydevil/var/run on /var/run (zfs, local, noexec, nosuid) mydevil/var/tmp on /var/tmp (zfs, local, nosuid) devfs on /var/named/dev (devfs, local, multilabel) mydevil/var/log on /var/log (zfs, local, noexec, nosuid) truss and ls: # ps lp 95579 95580 UID PID PPID CPU PRI NI VSZ RSS MWCHAN STAT TT TIME COMMAND 0 95579 95562 0 55 10 5784 1236 wait IN+ 117 0:00,01 truss -o ls.truss -f ls /usr/home 0 95580 95579 0 55 10 8200 1432 zfs DNX 117 0:00,00 ls /usr/home # procstat -kk 95579 PID TID COMM TDNAME KSTACK 95579 102312 truss - mi_switch+0x176 sleepq_catch_signals+0x309 sleepq_wait_sig+0xc _sleep+0x25d kern_wait+0x6ef wait4+0x33 syscall+0x1cd Xfast_syscall+0xe2 # procstat -kk 95580 PID TID COMM TDNAME KSTACK 95580 101995 ls - mi_switch+0x176 sleepq_wait+0x3b __lockmgr_args+0x642 vop_stdlock+0x39 VOP_LOCK1_APV+0x46 _vn_lock+0x44 zfs_root+0x7c lookup+0x996 namei+0x518 vn_open_cred+0x390 kern_openat+0x165 syscall+0x1cd Xfast_syscall+0xe2 If it isn't enough I will send full output. -- mydevil.net - niebiańskie konta shell From owner-freebsd-fs@FreeBSD.ORG Sat Oct 23 16:36:13 2010 Return-Path: Delivered-To: freebsd-fs@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id 5A02C106566C for ; Sat, 23 Oct 2010 16:36:13 +0000 (UTC) (envelope-from mcdouga9@egr.msu.edu) Received: from mail.egr.msu.edu (gribble.egr.msu.edu [35.9.37.169]) by mx1.freebsd.org (Postfix) with ESMTP id 157728FC1C for ; Sat, 23 Oct 2010 16:36:12 +0000 (UTC) Received: from gribble (localhost [127.0.0.1]) by mail.egr.msu.edu (Postfix) with ESMTP id 1CAA6ABC42 for ; Sat, 23 Oct 2010 12:19:40 -0400 (EDT) X-Virus-Scanned: amavisd-new at egr.msu.edu Received: from mail.egr.msu.edu ([127.0.0.1]) by gribble (gribble.egr.msu.edu [127.0.0.1]) (amavisd-new, port 10024) with ESMTP id x05VwkKuW+uD for ; Sat, 23 Oct 2010 12:19:39 -0400 (EDT) Received: from [35.9.36.248] (mcdouga9-tun.egr.msu.edu [35.9.36.248]) (using TLSv1 with cipher DHE-RSA-CAMELLIA256-SHA (256/256 bits)) (No client certificate requested) (Authenticated sender: mcdouga9) by mail.egr.msu.edu (Postfix) with ESMTPSA id CAAD1ABC3C for ; Sat, 23 Oct 2010 12:19:39 -0400 (EDT) Message-ID: <4CC30B22.7090507@egr.msu.edu> Date: Sat, 23 Oct 2010 12:19:46 -0400 From: Adam McDougall User-Agent: Mozilla/5.0 (Windows; U; Windows NT 6.1; en-US; rv:1.9.2.11) Gecko/20101013 Lightning/1.0b2 Thunderbird/3.1.5 MIME-Version: 1.0 To: freebsd-fs@freebsd.org References: <4CC215B4.3050607@nttmcl.com> <20101023085832.GD1742@garage.freebsd.pl> In-Reply-To: <20101023085832.GD1742@garage.freebsd.pl> Content-Type: text/plain; charset=ISO-8859-1; format=flowed Content-Transfer-Encoding: 7bit Subject: Re: ZFS: Parallel I/O to vdevs that appear to be separate physical disks but really are partitions X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Sat, 23 Oct 2010 16:36:13 -0000 On 10/23/2010 4:58 AM, Pawel Jakub Dawidek wrote: > On Fri, Oct 22, 2010 at 03:52:36PM -0700, Eugene M. Kim wrote: >> -----BEGIN PGP SIGNED MESSAGE----- >> Hash: SHA1 >> >> Greetings, >> >> I run a FreeBSD guest in VMware ESXi with a 10GB zpool. Lately the >> originally provisioned 10GB proved insufficient, and I would like to >> provision another 10GB virtual disk and add it to the zpool as a >> top-level vdev. >> >> The original and new virtual disks are from the same physical pool (a >> RAID-5 array), but appears to be separate physical disks to the >> FreeBSD guest (da0 and da1). I am afraid that the ZFS would schedule >> I/O to the two virtual disks in parallel thinking that the pool >> performance will improve, while the performance would actually suffer >> due to seeking back and forth between two regions of one physical pool. >> >> 1. Will ZFS schedule parallel access to different top-level vdevs? > Yes, ZFS stripes requests over all top level vdevs. > >> 2. If so, how can I suggest ZFS that certain vdevs should be treated >> not as separate physical disks but as partitions on the same physical >> disk? > Nope, that's not possible. > > Instead of adding next 10GB disk maybe it is possible to grow the one > you have? ZFS should be able to grow automatically. > What about creating a new 20gb "disk" then using zpool replace, export and import it to see the new size, and then retiring the old 10gb "disk"? From owner-freebsd-fs@FreeBSD.ORG Sat Oct 23 19:33:20 2010 Return-Path: Delivered-To: fs@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id 8E459106564A for ; Sat, 23 Oct 2010 19:33:20 +0000 (UTC) (envelope-from toasty@dragondata.com) Received: from mail-yw0-f54.google.com (mail-yw0-f54.google.com [209.85.213.54]) by mx1.freebsd.org (Postfix) with ESMTP id 50FFD8FC18 for ; Sat, 23 Oct 2010 19:33:20 +0000 (UTC) Received: by ywc21 with SMTP id 21so413559ywc.13 for ; Sat, 23 Oct 2010 12:33:19 -0700 (PDT) Received: by 10.42.71.70 with SMTP id i6mr3564876icj.200.1287860786595; Sat, 23 Oct 2010 12:06:26 -0700 (PDT) Received: from vpn177.ord02.your.org (vpn177.ord02.your.org [204.9.55.177]) by mx.google.com with ESMTPS id u6sm5229974ibd.12.2010.10.23.12.06.25 (version=TLSv1/SSLv3 cipher=RC4-MD5); Sat, 23 Oct 2010 12:06:25 -0700 (PDT) Mime-Version: 1.0 (Apple Message framework v1081) Content-Type: text/plain; charset=us-ascii From: Kevin Day In-Reply-To: <4CC215B4.3050607@nttmcl.com> Date: Sat, 23 Oct 2010 14:06:23 -0500 Content-Transfer-Encoding: quoted-printable Message-Id: <86C8DC50-9DE0-42B3-8A57-63AB4D095E6D@dragondata.com> References: <4CC215B4.3050607@nttmcl.com> To: Eugene M. Kim X-Mailer: Apple Mail (2.1081) Cc: fs@freebsd.org Subject: Re: ZFS: Parallel I/O to vdevs that appear to be separate physical disks but really are partitions X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Sat, 23 Oct 2010 19:33:20 -0000 On Oct 22, 2010, at 5:52 PM, Eugene M. Kim wrote: >=20 > -----BEGIN PGP SIGNED MESSAGE----- > Hash: SHA1 >=20 > Greetings, >=20 > I run a FreeBSD guest in VMware ESXi with a 10GB zpool. Lately the > originally provisioned 10GB proved insufficient, and I would like to > provision another 10GB virtual disk and add it to the zpool as a > top-level vdev. >=20 > The original and new virtual disks are from the same physical pool (a > RAID-5 array), but appears to be separate physical disks to the > FreeBSD guest (da0 and da1). I am afraid that the ZFS would schedule > I/O to the two virtual disks in parallel thinking that the pool > performance will improve, while the performance would actually suffer > due to seeking back and forth between two regions of one physical = pool. Just to chime in with a bit of practical experience here... It really = won't measurably matter for most workloads. FreeBSD will schedule I/O = separately, but VMWare will recombine them on the hypervisor and = reschedule them as best it can knowing the real hardware layout. It doesn't work well if you're running two mirrored disks, where your OS = might try round-robin'ing the requests between what it thinks are two = identical drives that will seek independently. But, if you've only got = one place you can possibly look for the data, you really have no choice = where you're going to ask to read it from. So the OS issues a bunch of = requests as needed, and VMWare will reorder them the best it can.=20 On occasion, where VMWare is connected to a very very large SAN or local = storage with many (48+) drives, we've even seen small performance = increases by giving FreeBSD several small disks and using vinum or ccd = to stripe between them. If FreeBSD thinks there are a half dozen drives, = there are times where it will allow more outstanding I/O requests at a = time, and VMWare can reschedule at its whim. -- Kevin