Skip site navigation (1)Skip section navigation (2)
Date:      Sat, 17 Jun 2017 05:16:22 +0000
From:      "Caza, Aaron" <Aaron.Caza@ca.weatherford.com>
To:        "freebsd-hackers@freebsd.org" <freebsd-hackers@freebsd.org>
Subject:   Re: FreeBSD10 Stable + ZFS + PostgreSQL + SSD performance drop < 24 hours
Message-ID:  <4561529b83ce4270b09aa0e3b12f299f@BLUPR58MB002.032d.mgd.msft.net>

next in thread | raw e-mail | index | archive | help

Regarding this issue, I've now conducted testing using merely a FreeBSD 10.=
3 Stable amd64 GENERIC kernel and using dd to read a large file.  The follo=
wing is a log, taken hourly, of the degradation which occurred at just over=
 9 hours of uptime.  As the original is quite large, I've removed some sect=
ions; however, these can be supplied if desired.
Supplied are the initial dmesg and zpool status, logged only on startup, fo=
llowed by uptime, uname -a, and zfs-stats -a output, each of which are logg=
ed hourly.


Copyright (c) 1992-2017 The FreeBSD Project.
Copyright (c) 1979, 1980, 1983, 1986, 1988, 1989, 1991, 1992, 1993, 1994
The Regents of the University of California. All rights reserved.
FreeBSD is a registered trademark of The FreeBSD Foundation.
FreeBSD 10.3-STABLE #0 r319701: Mon Jun 12 19:23:44 UTC 2017
    root@releng1.nyi.freebsd.org:/usr/obj/usr/src/sys/GENERIC amd64
FreeBSD clang version 3.4.1 (tags/RELEASE_34/dot1-final 208032) 20140512
CPU: Intel(R) Xeon(R) CPU E31240 @ 3.30GHz (3292.60-MHz K8-class CPU)
  Origin=3D"GenuineIntel"  Id=3D0x206a7  Family=3D0x6  Model=3D0x2a  Steppi=
ng=3D7
  Features=3D0xbfebfbff<FPU,VME,DE,PSE,TSC,MSR,PAE,MCE,CX8,APIC,SEP,MTRR,PG=
E,MCA,CMOV,PAT,PSE36,CLFLUSH,DTS,ACPI,MMX,FXSR,SSE,SSE2,SS,HTT,TM,PBE>
  Features2=3D0x1dbae3ff<SSE3,PCLMULQDQ,DTES64,MON,DS_CPL,VMX,SMX,EST,TM2,S=
SSE3,CX16,xTPR,PDCM,PCID,SSE4.1,SSE4.2,x2APIC,POPCNT,TSCDLT,XSAVE,OSXSAVE,A=
VX>
  AMD Features=3D0x28100800<SYSCALL,NX,RDTSCP,LM>
  AMD Features2=3D0x1<LAHF>
  XSAVE Features=3D0x1<XSAVEOPT>
  VT-x: PAT,HLT,MTF,PAUSE,EPT,UG,VPID
  TSC: P-state invariant, performance statistics
real memory  =3D 8589934592 (8192 MB)
avail memory =3D 8219299840 (7838 MB)
Event timer "LAPIC" quality 600
ACPI APIC Table: <SUPERM SMCI--MB>
FreeBSD/SMP: Multiprocessor System Detected: 8 CPUs
FreeBSD/SMP: 1 package(s) x 4 core(s) x 2 SMT threads
 cpu0 (BSP): APIC ID:  0
 cpu1 (AP): APIC ID:  1
 cpu2 (AP): APIC ID:  2
 cpu3 (AP): APIC ID:  3
 cpu4 (AP): APIC ID:  4
 cpu5 (AP): APIC ID:  5
 cpu6 (AP): APIC ID:  6
 cpu7 (AP): APIC ID:  7
random: <Software, Yarrow> initialized
ioapic0 <Version 2.0> irqs 0-23 on motherboard
kbd1 at kbdmux0
cryptosoft0: <software crypto> on motherboard
acpi0: <SUPERM SMCI--MB> on motherboard
acpi0: Power Button (fixed)
cpu0: <ACPI CPU> on acpi0
cpu1: <ACPI CPU> on acpi0
cpu2: <ACPI CPU> on acpi0
cpu3: <ACPI CPU> on acpi0
cpu4: <ACPI CPU> on acpi0
cpu5: <ACPI CPU> on acpi0
cpu6: <ACPI CPU> on acpi0
cpu7: <ACPI CPU> on acpi0
attimer0: <AT timer> port 0x40-0x43 irq 0 on acpi0
Timecounter "i8254" frequency 1193182 Hz quality 0
Event timer "i8254" frequency 1193182 Hz quality 100
atrtc0: <AT realtime clock> port 0x70-0x71 irq 8 on acpi0
Event timer "RTC" frequency 32768 Hz quality 0
hpet0: <High Precision Event Timer> iomem 0xfed00000-0xfed003ff on acpi0
Timecounter "HPET" frequency 14318180 Hz quality 950
Event timer "HPET" frequency 14318180 Hz quality 550
Timecounter "ACPI-fast" frequency 3579545 Hz quality 900
acpi_timer0: <24-bit timer at 3.579545MHz> port 0x408-0x40b on acpi0
pcib0: <ACPI Host-PCI bridge> port 0xcf8-0xcff on acpi0
pci0: <ACPI PCI bus> on pcib0
em0: <Intel(R) PRO/1000 Network Connection 7.6.1-k> port 0xf020-0xf03f mem =
0xfba00000-0xfba1ffff,0xfba24000-0xfba24fff irq 20 at device 25.0 on pci0
em0: Using an MSI interrupt
em0: Ethernet address: 00:25:90:76:6b:41
ehci0: <Intel Cougar Point USB 2.0 controller> mem 0xfba23000-0xfba233ff ir=
q 16 at device 26.0 on pci0
usbus0: EHCI version 1.0
usbus0 on ehci0
pcib1: <ACPI PCI-PCI bridge> irq 17 at device 28.0 on pci0
pci1: <ACPI PCI bus> on pcib1
pcib2: <ACPI PCI-PCI bridge> irq 17 at device 28.4 on pci0
pci2: <ACPI PCI bus> on pcib2
em1: <Intel(R) PRO/1000 Network Connection 7.6.1-k> port 0xe000-0xe01f mem =
0xfb900000-0xfb91ffff,0xfb920000-0xfb923fff irq 16 at device 0.0 on pci2
em1: Using MSIX interrupts with 3 vectors
em1: Ethernet address: 00:25:90:76:6b:40
ehci1: <Intel Cougar Point USB 2.0 controller> mem 0xfba22000-0xfba223ff ir=
q 23 at device 29.0 on pci0
usbus1: EHCI version 1.0
usbus1 on ehci1
pcib3: <ACPI PCI-PCI bridge> at device 30.0 on pci0
pci3: <ACPI PCI bus> on pcib3
vgapci0: <VGA-compatible display> mem 0xfe000000-0xfe7fffff,0xfb800000-0xfb=
803fff,0xfb000000-0xfb7fffff irq 23 at device 3.0 on pci3
vgapci0: Boot video device
isab0: <PCI-ISA bridge> at device 31.0 on pci0
isa0: <ISA bus> on isab0
ahci0: <Intel Cougar Point AHCI SATA controller> port 0xf070-0xf077,0xf060-=
0xf063,0xf050-0xf057,0xf040-0xf043,0xf000-0xf01f mem 0xfba21000-0xfba217ff =
irq 19 at device 31.2 on pci0
ahci0: AHCI v1.30 with 6 6Gbps ports, Port Multiplier not supported
ahcich0: <AHCI channel> at channel 0 on ahci0
ahcich1: <AHCI channel> at channel 1 on ahci0
ahciem0: <AHCI enclosure management bridge> on ahci0
acpi_button0: <Power Button> on acpi0
atkbdc0: <Keyboard controller (i8042)> port 0x60,0x64 irq 1 on acpi0
atkbd0: <AT Keyboard> irq 1 on atkbdc0
kbd0 at atkbd0
atkbd0: [GIANT-LOCKED]
psm0: <PS/2 Mouse> irq 12 on atkbdc0
psm0: [GIANT-LOCKED]
psm0: model IntelliMouse Explorer, device ID 4
uart0: <16550 or compatible> port 0x3f8-0x3ff irq 4 flags 0x10 on acpi0
orm0: <ISA Option ROMs> at iomem 0xc0000-0xc7fff,0xc8000-0xc8fff on isa0
sc0: <System console> at flags 0x100 on isa0
sc0: VGA <16 virtual consoles, flags=3D0x300>
vga0: <Generic ISA VGA> at port 0x3c0-0x3df iomem 0xa0000-0xbffff on isa0
ppc0: cannot reserve I/O port range
est0: <Enhanced SpeedStep Frequency Control> on cpu0
est1: <Enhanced SpeedStep Frequency Control> on cpu1
est2: <Enhanced SpeedStep Frequency Control> on cpu2
est3: <Enhanced SpeedStep Frequency Control> on cpu3
est4: <Enhanced SpeedStep Frequency Control> on cpu4
est5: <Enhanced SpeedStep Frequency Control> on cpu5
est6: <Enhanced SpeedStep Frequency Control> on cpu6
est7: <Enhanced SpeedStep Frequency Control> on cpu7
ZFS filesystem version: 5
ZFS storage pool version: features support (5000)
Timecounters tick every 1.000 msec
md0: Preloaded image </boot/mfsroot> 17686528 bytes at 0xffffffff81daa1b8
random: unblocking device.
usbus0: 480Mbps High Speed USB v2.0
usbus1: 480Mbps High Speed USB v2.0
ugen0.1: <Intel EHCI root HUB> at usbus0
uhub0: <Intel EHCI root HUB, class 9/0, rev 2.00/1.00, addr 1> on usbus0
ugen1.1: <Intel EHCI root HUB> at usbus1
uhub1: <Intel EHCI root HUB, class 9/0, rev 2.00/1.00, addr 1> on usbus1
ada0 at ahcich0 bus 0 scbus0 target 0 lun 0
ada0: <Samsung SSD 850 PRO 256GB EXM03B6Q> ACS-2 ATA SATA 3.x device
ada0: Serial Number S39KNB0HB00482Y
ada0: 600.000MB/s transfers (SATA 3.x, UDMA6, PIO 512bytes)
ada0: Command Queueing enabled
ada0: 244198MB (500118192 512 byte sectors)
ada0: quirks=3D0x1<4K>
ada1 at ahcich1 bus 0 scbus1 target 0 lun 0
ada1: <Samsung SSD 850 PRO 256GB EXM03B6Q> ACS-2 ATA SATA 3.x device
ada1: Serial Number S39KNB0HB00473Z
ada1: 600.000MB/s transfers (SATA 3.x, UDMA6, PIO 512bytes)
ada1: Command Queueing enabled
ada1: 244198MB (500118192 512 byte sectors)
ada1: quirks=3D0x1<4K>
ses0 at ahciem0 bus 0 scbus2 target 0 lun 0
ses0: <AHCI SGPIO Enclosure 1.00 0001> SEMB S-E-S 2.00 device
ses0: SEMB SES Device
SMP: AP CPU #1 Launched!
SMP: AP CPU #6 Launched!
SMP: AP CPU #3 Launched!
SMP: AP CPU #5 Launched!
SMP: AP CPU #2 Launched!
SMP: AP CPU #4 Launched!
SMP: AP CPU #7 Launched!
Timecounter "TSC-low" frequency 1646298306 Hz quality 1000
Root mount waiting for: usbus1 usbus0
uhub1: 2 ports with 2 removable, self powered
uhub0: 2 ports with 2 removable, self powered
Root mount waiting for: usbus1 usbus0
ugen1.2: <vendor 0x8087 product 0x0024> at usbus1
uhub2: <vendor 0x8087 product 0x0024, class 9/0, rev 2.00/0.00, addr 2> on =
usbus1
ugen0.2: <vendor 0x8087 product 0x0024> at usbus0
uhub3: <vendor 0x8087 product 0x0024, class 9/0, rev 2.00/0.00, addr 2> on =
usbus0
Root mount waiting for: usbus1 usbus0
uhub2: 6 ports with 6 removable, self powered
uhub3: 6 ports with 6 removable, self powered
ugen1.3: <Weatherford SPD> at usbus1
Trying to mount root from ufs:/dev/md0 []...
bridge0: Ethernet address: 02:5d:9c:c3:f4:00
bridge0: link state changed to UP
em0: promiscuous mode enabled
em1: promiscuous mode enabled
em0: link state changed to UP

Zpool Status:
  pool: wwbase
 state: ONLINE
  scan: scrub repaired 0 in 0h1m with 0 errors on Sat Jun 10 18:01:26 2017
config:

NAME                                          STATE     READ WRITE CKSUM
wwbase                                        ONLINE       0     0     0
  mirror-0                                    ONLINE       0     0     0
    gpt/83b2ce34-4b9f-11e7-8d12-002590766b41  ONLINE       0     0     0
    gpt/8ec0d395-4b9f-11e7-8d12-002590766b41  ONLINE       0     0     0

errors: No known data errors

---------------------------------------------------------------------------=
----

Testing SSD performance @ Fri Jun 16 19:00:00 UTC 2017

FreeBSD xyz.test 10.3-STABLE FreeBSD 10.3-STABLE #0 r319701: Mon Jun 12 19:=
23:44 UTC 2017     root@releng1.nyi.freebsd.org:/usr/obj/usr/src/sys/GENERI=
C  amd64

 7:00PM  up 12 mins, 1 user, load averages: 0.00, 0.04, 0.07

Starting 'dd' test of large file...please wait
16000+0 records in
16000+0 records out
16777216000 bytes transferred in 18.844918 secs (890278004 bytes/sec)


------------------------------------------------------------------------
ZFS Subsystem ReportFri Jun 16 19:00:18 2017
------------------------------------------------------------------------

System Information:

Kernel Version:1003514 (osreldate)
Hardware Platform:amd64
Processor Architecture:amd64

ZFS Storage pool Version:5000
ZFS Filesystem Version:5

FreeBSD 10.3-STABLE #0 r319701: Mon Jun 12 19:23:44 UTC 2017 root
 7:00PM  up 13 mins, 1 user, load averages: 2.27, 0.57, 0.25

------------------------------------------------------------------------

System Memory:

0.19%14.90MiB Active,0.38%29.62MiB Inact
4.31%339.96MiB Wired,0.00%0 Cache
95.13%7.33GiB Free,0.00%4.00KiB Gap

Real Installed:8.00GiB
Real Available:99.18%7.93GiB
Real Managed:97.11%7.71GiB

Logical Total:8.00GiB
Logical Used:8.02%656.83MiB
Logical Free:91.98%7.36GiB

Kernel Memory:102.71MiB
Data:73.04%75.02MiB
Text:26.96%27.69MiB

Kernel Memory Map:7.71GiB
Size:2.48%196.06MiB
Free:97.52%7.51GiB

------------------------------------------------------------------------

ARC Summary: (HEALTHY)
Memory Throttle Count:0

ARC Misc:
Deleted:2.06m
Recycle Misses:0
Mutex Misses:10.06k
Evict Skips:1.49m

ARC Size:103.44%52.76MiB
Target Size: (Adaptive)100.00%51.00MiB
Min Size (Hard Limit):98.04%50.00MiB
Max Size (High Water):1:151.00MiB

ARC Size Breakdown:
Recently Used Cache Size:89.95%47.45MiB
Frequently Used Cache Size:10.05%5.30MiB

ARC Hash Breakdown:
Elements Max:11.92k
Elements Current:73.83%8.80k
Collisions:19.65k
Chain Max:2
Chains:25

------------------------------------------------------------------------

ARC Efficiency:4.06m
Cache Hit Ratio:49.05%1.99m
Cache Miss Ratio:50.95%2.07m
Actual Hit Ratio:49.05%1.99m

Data Demand Efficiency:99.91%1.94m
Data Prefetch Efficiency:0.01%2.05m

CACHE HITS BY CACHE LIST:
  Anonymously Used:0.01%159
  Most Recently Used:99.82%1.99m
  Most Frequently Used:0.16%3.28k
  Most Recently Used Ghost:0.00%9
  Most Frequently Used Ghost:0.00%83

CACHE HITS BY DATA TYPE:
  Demand Data:97.40%1.94m
  Prefetch Data:0.01%151
  Demand Metadata:2.59%51.44k
  Prefetch Metadata:0.01%100

CACHE MISSES BY DATA TYPE:
  Demand Data:0.09%1.76k
  Prefetch Data:99.07%2.05m
  Demand Metadata:0.07%1.50k
  Prefetch Metadata:0.77%16.01k

------------------------------------------------------------------------

L2ARC is disabled

------------------------------------------------------------------------

File-Level Prefetch: (HEALTHY)

DMU Efficiency:46.70k
Hit Ratio:34.40%16.06k
Miss Ratio:65.60%30.64k

Colinear:0
  Hit Ratio:100.00%0
  Miss Ratio:100.00%0

Stride:0
  Hit Ratio:100.00%0
  Miss Ratio:100.00%0

DMU Misc:
Reclaim:0
  Successes:100.00%0
  Failures:100.00%0

Streams:0
  +Resets:100.00%0
  -Resets:100.00%0
  Bogus:0

------------------------------------------------------------------------

VDEV Cache Summary:16.52k
Hit Ratio:1.65%272
Miss Ratio:93.47%15.44k
Delegations:4.88%806

------------------------------------------------------------------------

ZFS Tunables (sysctl):
kern.maxusers                           843
vm.kmem_size                            8273297408
vm.kmem_size_scale                      1
vm.kmem_size_min                        0
vm.kmem_size_max                        1319413950874
vfs.zfs.trim.max_interval               1
vfs.zfs.trim.timeout                    30
vfs.zfs.trim.txg_delay                  32
vfs.zfs.trim.enabled                    0
vfs.zfs.vol.unmap_enabled               1
vfs.zfs.vol.mode                        1
vfs.zfs.version.zpl                     5
vfs.zfs.version.spa                     5000
vfs.zfs.version.acl                     1
vfs.zfs.version.ioctl                   7
vfs.zfs.debug                           0
vfs.zfs.super_owner                     0
vfs.zfs.sync_pass_rewrite               2
vfs.zfs.sync_pass_dont_compress         5
vfs.zfs.sync_pass_deferred_free         2
vfs.zfs.zio.dva_throttle_enabled        1
vfs.zfs.zio.exclude_metadata            0
vfs.zfs.zio.use_uma                     1
vfs.zfs.cache_flush_disable             0
vfs.zfs.zil_replay_disable              0
vfs.zfs.min_auto_ashift                 9
vfs.zfs.max_auto_ashift                 13
vfs.zfs.vdev.trim_max_pending           10000
vfs.zfs.vdev.bio_delete_disable         0
vfs.zfs.vdev.bio_flush_disable          0
vfs.zfs.vdev.queue_depth_pct            1000
vfs.zfs.vdev.write_gap_limit            4096
vfs.zfs.vdev.read_gap_limit             32768
vfs.zfs.vdev.aggregation_limit          131072
vfs.zfs.vdev.trim_max_active            64
vfs.zfs.vdev.trim_min_active            1
vfs.zfs.vdev.scrub_max_active           2
vfs.zfs.vdev.scrub_min_active           1
vfs.zfs.vdev.async_write_max_active     10
vfs.zfs.vdev.async_write_min_active     1
vfs.zfs.vdev.async_read_max_active      3
vfs.zfs.vdev.async_read_min_active      1
vfs.zfs.vdev.sync_write_max_active      10
vfs.zfs.vdev.sync_write_min_active      10
vfs.zfs.vdev.sync_read_max_active       10
vfs.zfs.vdev.sync_read_min_active       10
vfs.zfs.vdev.max_active                 1000
vfs.zfs.vdev.async_write_active_max_dirty_percent60
vfs.zfs.vdev.async_write_active_min_dirty_percent30
vfs.zfs.vdev.mirror.non_rotating_seek_inc1
vfs.zfs.vdev.mirror.non_rotating_inc    0
vfs.zfs.vdev.mirror.rotating_seek_offset1048576
vfs.zfs.vdev.mirror.rotating_seek_inc   5
vfs.zfs.vdev.mirror.rotating_inc        0
vfs.zfs.vdev.trim_on_init               1
vfs.zfs.vdev.cache.bshift               16
vfs.zfs.vdev.cache.size                 10485760
vfs.zfs.vdev.cache.max                  16384
vfs.zfs.vdev.metaslabs_per_vdev         200
vfs.zfs.txg.timeout                     5
vfs.zfs.space_map_blksz                 4096
vfs.zfs.spa_min_slop                    134217728
vfs.zfs.spa_slop_shift                  5
vfs.zfs.spa_asize_inflation             24
vfs.zfs.deadman_enabled                 1
vfs.zfs.deadman_checktime_ms            5000
vfs.zfs.deadman_synctime_ms             1000000
vfs.zfs.debug_flags                     0
vfs.zfs.debugflags                      0
vfs.zfs.recover                         0
vfs.zfs.spa_load_verify_data            1
vfs.zfs.spa_load_verify_metadata        1
vfs.zfs.spa_load_verify_maxinflight     10000
vfs.zfs.ccw_retry_interval              300
vfs.zfs.check_hostid                    1
vfs.zfs.mg_fragmentation_threshold      85
vfs.zfs.mg_noalloc_threshold            0
vfs.zfs.condense_pct                    200
vfs.zfs.metaslab.bias_enabled           1
vfs.zfs.metaslab.lba_weighting_enabled  1
vfs.zfs.metaslab.fragmentation_factor_enabled1
vfs.zfs.metaslab.preload_enabled        1
vfs.zfs.metaslab.preload_limit          3
vfs.zfs.metaslab.unload_delay           8
vfs.zfs.metaslab.load_pct               50
vfs.zfs.metaslab.min_alloc_size         33554432
vfs.zfs.metaslab.df_free_pct            4
vfs.zfs.metaslab.df_alloc_threshold     131072
vfs.zfs.metaslab.debug_unload           0
vfs.zfs.metaslab.debug_load             0
vfs.zfs.metaslab.fragmentation_threshold70
vfs.zfs.metaslab.gang_bang              16777217
vfs.zfs.free_bpobj_enabled              1
vfs.zfs.free_max_blocks                 -1
vfs.zfs.no_scrub_prefetch               0
vfs.zfs.no_scrub_io                     0
vfs.zfs.resilver_min_time_ms            3000
vfs.zfs.free_min_time_ms                1000
vfs.zfs.scan_min_time_ms                1000
vfs.zfs.scan_idle                       50
vfs.zfs.scrub_delay                     4
vfs.zfs.resilver_delay                  2
vfs.zfs.top_maxinflight                 32
vfs.zfs.zfetch.array_rd_sz              1048576
vfs.zfs.zfetch.max_distance             8388608
vfs.zfs.zfetch.min_sec_reap             2
vfs.zfs.zfetch.max_streams              8
vfs.zfs.prefetch_disable                0
vfs.zfs.delay_scale                     500000
vfs.zfs.delay_min_dirty_percent         60
vfs.zfs.dirty_data_sync                 67108864
vfs.zfs.dirty_data_max_percent          10
vfs.zfs.dirty_data_max_max              4294967296
vfs.zfs.dirty_data_max                  851961036
vfs.zfs.max_recordsize                  1048576
vfs.zfs.send_holes_without_birth_time   1
vfs.zfs.mdcomp_disable                  0
vfs.zfs.nopwrite_enabled                1
vfs.zfs.dedup.prefetch                  1
vfs.zfs.l2c_only_size                   0
vfs.zfs.mfu_ghost_data_esize            0
vfs.zfs.mfu_ghost_metadata_esize        39650304
vfs.zfs.mfu_ghost_size                  39650304
vfs.zfs.mfu_data_esize                  0
vfs.zfs.mfu_metadata_esize              2785280
vfs.zfs.mfu_size                        2839040
vfs.zfs.mru_ghost_data_esize            253952
vfs.zfs.mru_ghost_metadata_esize        7712768
vfs.zfs.mru_ghost_size                  7966720
vfs.zfs.mru_data_esize                  40481280
vfs.zfs.mru_metadata_esize              3874816
vfs.zfs.mru_size                        49604608
vfs.zfs.anon_data_esize                 0
vfs.zfs.anon_metadata_esize             0
vfs.zfs.anon_size                       28672
vfs.zfs.l2arc_norw                      1
vfs.zfs.l2arc_feed_again                1
vfs.zfs.l2arc_noprefetch                1
vfs.zfs.l2arc_feed_min_ms               200
vfs.zfs.l2arc_feed_secs                 1
vfs.zfs.l2arc_headroom                  2
vfs.zfs.l2arc_write_boost               8388608
vfs.zfs.l2arc_write_max                 8388608
vfs.zfs.arc_meta_limit                  13369344
vfs.zfs.arc_free_target                 14047
vfs.zfs.compressed_arc_enabled          1
vfs.zfs.arc_shrink_shift                7
vfs.zfs.arc_average_blocksize           8192
vfs.zfs.arc_min                         52428800
vfs.zfs.arc_max                         53477376

------------------------------------------------------------------------


SSD performance testing completed @ Fri Jun 16 19:00:19 UTC 2017

---------------------------------------------------------------------------=
----




      This section was removed in the interests of brevity but can be
      supplied is required.




---------------------------------------------------------------------------=
----

Testing SSD performance @ Sat Jun 17 03:00:00 UTC 2017

FreeBSD xyz.test 10.3-STABLE FreeBSD 10.3-STABLE #0 r319701: Mon Jun 12 19:=
23:44 UTC 2017     root@releng1.nyi.freebsd.org:/usr/obj/usr/src/sys/GENERI=
C  amd64

 3:00AM  up  8:12, 0 users, load averages: 0.00, 0.00, 0.00

Starting 'dd' test of large file...please wait
16000+0 records in
16000+0 records out
16777216000 bytes transferred in 18.995611 secs (883215382 bytes/sec)


------------------------------------------------------------------------
ZFS Subsystem ReportSat Jun 17 03:00:19 2017
------------------------------------------------------------------------

System Information:

Kernel Version:1003514 (osreldate)
Hardware Platform:amd64
Processor Architecture:amd64

ZFS Storage pool Version:5000
ZFS Filesystem Version:5

FreeBSD 10.3-STABLE #0 r319701: Mon Jun 12 19:23:44 UTC 2017 root
 3:00AM  up  8:13, 0 users, load averages: 3.41, 0.80, 0.29

------------------------------------------------------------------------

System Memory:

0.04%2.78MiB Active,0.55%43.71MiB Inact
4.59%361.96MiB Wired,0.01%632.00KiB Cache
94.82%7.31GiB Free,0.00%4.00KiB Gap

Real Installed:8.00GiB
Real Available:99.18%7.93GiB
Real Managed:97.11%7.71GiB

Logical Total:8.00GiB
Logical Used:8.14%666.72MiB
Logical Free:91.86%7.35GiB

Kernel Memory:103.60MiB
Data:73.27%75.91MiB
Text:26.73%27.69MiB

Kernel Memory Map:7.71GiB
Size:2.66%210.04MiB
Free:97.34%7.50GiB

------------------------------------------------------------------------

ARC Summary: (HEALTHY)
Memory Throttle Count:0

ARC Misc:
Deleted:18.57m
Recycle Misses:0
Mutex Misses:91.37k
Evict Skips:13.33m

ARC Size:110.62%56.41MiB
Target Size: (Adaptive)100.00%51.00MiB
Min Size (Hard Limit):98.04%50.00MiB
Max Size (High Water):1:151.00MiB

ARC Size Breakdown:
Recently Used Cache Size:84.43%47.63MiB
Frequently Used Cache Size:15.57%8.78MiB

ARC Hash Breakdown:
Elements Max:11.93k
Elements Current:80.43%9.59k
Collisions:193.31k
Chain Max:3
Chains:30

------------------------------------------------------------------------

ARC Efficiency:36.57m
Cache Hit Ratio:49.18%17.98m
Cache Miss Ratio:50.82%18.58m
Actual Hit Ratio:49.18%17.98m

Data Demand Efficiency:99.96%17.53m
Data Prefetch Efficiency:0.00%18.43m

CACHE HITS BY CACHE LIST:
  Anonymously Used:0.00%588
  Most Recently Used:99.83%17.95m
  Most Frequently Used:0.16%29.16k
  Most Recently Used Ghost:0.00%56
  Most Frequently Used Ghost:0.00%263

CACHE HITS BY DATA TYPE:
  Demand Data:97.46%17.53m
  Prefetch Data:0.00%633
  Demand Metadata:2.54%456.26k
  Prefetch Metadata:0.00%275

CACHE MISSES BY DATA TYPE:
  Demand Data:0.04%7.74k
  Prefetch Data:99.15%18.43m
  Demand Metadata:0.03%5.77k
  Prefetch Metadata:0.77%143.89k

------------------------------------------------------------------------

L2ARC is disabled

------------------------------------------------------------------------

File-Level Prefetch: (HEALTHY)

DMU Efficiency:597.55k
Hit Ratio:24.22%144.70k
Miss Ratio:75.78%452.85k

Colinear:0
  Hit Ratio:100.00%0
  Miss Ratio:100.00%0

Stride:0
  Hit Ratio:100.00%0
  Miss Ratio:100.00%0

DMU Misc:
Reclaim:0
  Successes:100.00%0
  Failures:100.00%0

Streams:0
  +Resets:100.00%0
  -Resets:100.00%0
  Bogus:0

------------------------------------------------------------------------

VDEV Cache Summary:138.31k
Hit Ratio:0.57%794
Miss Ratio:98.82%136.67k
Delegations:0.61%840

------------------------------------------------------------------------

ZFS Tunables (sysctl):
kern.maxusers                           843
vm.kmem_size                            8273297408
vm.kmem_size_scale                      1
vm.kmem_size_min                        0
vm.kmem_size_max                        1319413950874
vfs.zfs.trim.max_interval               1
vfs.zfs.trim.timeout                    30
vfs.zfs.trim.txg_delay                  32
vfs.zfs.trim.enabled                    0
vfs.zfs.vol.unmap_enabled               1
vfs.zfs.vol.mode                        1
vfs.zfs.version.zpl                     5
vfs.zfs.version.spa                     5000
vfs.zfs.version.acl                     1
vfs.zfs.version.ioctl                   7
vfs.zfs.debug                           0
vfs.zfs.super_owner                     0
vfs.zfs.sync_pass_rewrite               2
vfs.zfs.sync_pass_dont_compress         5
vfs.zfs.sync_pass_deferred_free         2
vfs.zfs.zio.dva_throttle_enabled        1
vfs.zfs.zio.exclude_metadata            0
vfs.zfs.zio.use_uma                     1
vfs.zfs.cache_flush_disable             0
vfs.zfs.zil_replay_disable              0
vfs.zfs.min_auto_ashift                 9
vfs.zfs.max_auto_ashift                 13
vfs.zfs.vdev.trim_max_pending           10000
vfs.zfs.vdev.bio_delete_disable         0
vfs.zfs.vdev.bio_flush_disable          0
vfs.zfs.vdev.queue_depth_pct            1000
vfs.zfs.vdev.write_gap_limit            4096
vfs.zfs.vdev.read_gap_limit             32768
vfs.zfs.vdev.aggregation_limit          131072
vfs.zfs.vdev.trim_max_active            64
vfs.zfs.vdev.trim_min_active            1
vfs.zfs.vdev.scrub_max_active           2
vfs.zfs.vdev.scrub_min_active           1
vfs.zfs.vdev.async_write_max_active     10
vfs.zfs.vdev.async_write_min_active     1
vfs.zfs.vdev.async_read_max_active      3
vfs.zfs.vdev.async_read_min_active      1
vfs.zfs.vdev.sync_write_max_active      10
vfs.zfs.vdev.sync_write_min_active      10
vfs.zfs.vdev.sync_read_max_active       10
vfs.zfs.vdev.sync_read_min_active       10
vfs.zfs.vdev.max_active                 1000
vfs.zfs.vdev.async_write_active_max_dirty_percent60
vfs.zfs.vdev.async_write_active_min_dirty_percent30
vfs.zfs.vdev.mirror.non_rotating_seek_inc1
vfs.zfs.vdev.mirror.non_rotating_inc    0
vfs.zfs.vdev.mirror.rotating_seek_offset1048576
vfs.zfs.vdev.mirror.rotating_seek_inc   5
vfs.zfs.vdev.mirror.rotating_inc        0
vfs.zfs.vdev.trim_on_init               1
vfs.zfs.vdev.cache.bshift               16
vfs.zfs.vdev.cache.size                 10485760
vfs.zfs.vdev.cache.max                  16384
vfs.zfs.vdev.metaslabs_per_vdev         200
vfs.zfs.txg.timeout                     5
vfs.zfs.space_map_blksz                 4096
vfs.zfs.spa_min_slop                    134217728
vfs.zfs.spa_slop_shift                  5
vfs.zfs.spa_asize_inflation             24
vfs.zfs.deadman_enabled                 1
vfs.zfs.deadman_checktime_ms            5000
vfs.zfs.deadman_synctime_ms             1000000
vfs.zfs.debug_flags                     0
vfs.zfs.debugflags                      0
vfs.zfs.recover                         0
vfs.zfs.spa_load_verify_data            1
vfs.zfs.spa_load_verify_metadata        1
vfs.zfs.spa_load_verify_maxinflight     10000
vfs.zfs.ccw_retry_interval              300
vfs.zfs.check_hostid                    1
vfs.zfs.mg_fragmentation_threshold      85
vfs.zfs.mg_noalloc_threshold            0
vfs.zfs.condense_pct                    200
vfs.zfs.metaslab.bias_enabled           1
vfs.zfs.metaslab.lba_weighting_enabled  1
vfs.zfs.metaslab.fragmentation_factor_enabled1
vfs.zfs.metaslab.preload_enabled        1
vfs.zfs.metaslab.preload_limit          3
vfs.zfs.metaslab.unload_delay           8
vfs.zfs.metaslab.load_pct               50
vfs.zfs.metaslab.min_alloc_size         33554432
vfs.zfs.metaslab.df_free_pct            4
vfs.zfs.metaslab.df_alloc_threshold     131072
vfs.zfs.metaslab.debug_unload           0
vfs.zfs.metaslab.debug_load             0
vfs.zfs.metaslab.fragmentation_threshold70
vfs.zfs.metaslab.gang_bang              16777217
vfs.zfs.free_bpobj_enabled              1
vfs.zfs.free_max_blocks                 -1
vfs.zfs.no_scrub_prefetch               0
vfs.zfs.no_scrub_io                     0
vfs.zfs.resilver_min_time_ms            3000
vfs.zfs.free_min_time_ms                1000
vfs.zfs.scan_min_time_ms                1000
vfs.zfs.scan_idle                       50
vfs.zfs.scrub_delay                     4
vfs.zfs.resilver_delay                  2
vfs.zfs.top_maxinflight                 32
vfs.zfs.zfetch.array_rd_sz              1048576
vfs.zfs.zfetch.max_distance             8388608
vfs.zfs.zfetch.min_sec_reap             2
vfs.zfs.zfetch.max_streams              8
vfs.zfs.prefetch_disable                0
vfs.zfs.delay_scale                     500000
vfs.zfs.delay_min_dirty_percent         60
vfs.zfs.dirty_data_sync                 67108864
vfs.zfs.dirty_data_max_percent          10
vfs.zfs.dirty_data_max_max              4294967296
vfs.zfs.dirty_data_max                  851961036
vfs.zfs.max_recordsize                  1048576
vfs.zfs.send_holes_without_birth_time   1
vfs.zfs.mdcomp_disable                  0
vfs.zfs.nopwrite_enabled                1
vfs.zfs.dedup.prefetch                  1
vfs.zfs.l2c_only_size                   0
vfs.zfs.mfu_ghost_data_esize            0
vfs.zfs.mfu_ghost_metadata_esize        41795584
vfs.zfs.mfu_ghost_size                  41795584
vfs.zfs.mfu_data_esize                  0
vfs.zfs.mfu_metadata_esize              2670592
vfs.zfs.mfu_size                        2719744
vfs.zfs.mru_ghost_data_esize            1622016
vfs.zfs.mru_ghost_metadata_esize        6819840
vfs.zfs.mru_ghost_size                  8441856
vfs.zfs.mru_data_esize                  42278912
vfs.zfs.mru_metadata_esize              4341760
vfs.zfs.mru_size                        52709376
vfs.zfs.anon_data_esize                 0
vfs.zfs.anon_metadata_esize             0
vfs.zfs.anon_size                       147456
vfs.zfs.l2arc_norw                      1
vfs.zfs.l2arc_feed_again                1
vfs.zfs.l2arc_noprefetch                1
vfs.zfs.l2arc_feed_min_ms               200
vfs.zfs.l2arc_feed_secs                 1
vfs.zfs.l2arc_headroom                  2
vfs.zfs.l2arc_write_boost               8388608
vfs.zfs.l2arc_write_max                 8388608
vfs.zfs.arc_meta_limit                  13369344
vfs.zfs.arc_free_target                 14047
vfs.zfs.compressed_arc_enabled          1
vfs.zfs.arc_shrink_shift                7
vfs.zfs.arc_average_blocksize           8192
vfs.zfs.arc_min                         52428800
vfs.zfs.arc_max                         53477376

------------------------------------------------------------------------



SSD performance testing completed @ Sat Jun 17 03:00:19 UTC 2017

---------------------------------------------------------------------------=
----

Testing SSD performance @ Sat Jun 17 04:00:00 UTC 2017

FreeBSD xyz.test 10.3-STABLE FreeBSD 10.3-STABLE #0 r319701: Mon Jun 12 19:=
23:44 UTC 2017     root@releng1.nyi.freebsd.org:/usr/obj/usr/src/sys/GENERI=
C  amd64

 4:00AM  up  9:12, 0 users, load averages: 0.00, 0.00, 0.00

Starting 'dd' test of large file...please wait
16000+0 records in
16000+0 records out
16777216000 bytes transferred in 268.165167 secs (62562995 bytes/sec)


------------------------------------------------------------------------
ZFS Subsystem ReportSat Jun 17 04:04:28 2017
------------------------------------------------------------------------

System Information:

Kernel Version:1003514 (osreldate)
Hardware Platform:amd64
Processor Architecture:amd64

ZFS Storage pool Version:5000
ZFS Filesystem Version:5

FreeBSD 10.3-STABLE #0 r319701: Mon Jun 12 19:23:44 UTC 2017 root
 4:04AM  up  9:17, 0 users, load averages: 1.05, 0.67, 0.30

------------------------------------------------------------------------

System Memory:

0.04%2.80MiB Active,0.60%47.30MiB Inact
5.90%465.19MiB Wired,0.07%5.36MiB Cache
93.40%7.20GiB Free,0.00%4.00KiB Gap

Real Installed:8.00GiB
Real Available:99.18%7.93GiB
Real Managed:97.11%7.71GiB

Logical Total:8.00GiB
Logical Used:9.40%769.96MiB
Logical Free:90.60%7.25GiB

Kernel Memory:108.62MiB
Data:74.51%80.93MiB
Text:25.49%27.69MiB

Kernel Memory Map:7.71GiB
Size:2.68%211.31MiB
Free:97.32%7.50GiB

------------------------------------------------------------------------

ARC Summary: (HEALTHY)
Memory Throttle Count:0

ARC Misc:
Deleted:20.64m
Recycle Misses:0
Mutex Misses:682.38k
Evict Skips:2.33b

ARC Size:141.09%71.95MiB
Target Size: (Adaptive)100.00%51.00MiB
Min Size (Hard Limit):98.04%50.00MiB
Max Size (High Water):1:151.00MiB

ARC Size Breakdown:
Recently Used Cache Size:66.20%47.63MiB
Frequently Used Cache Size:33.80%24.32MiB

ARC Hash Breakdown:
Elements Max:11.93k
Elements Current:41.76%4.98k
Collisions:204.57k
Chain Max:3
Chains:7

------------------------------------------------------------------------

ARC Efficiency:40.75m
Cache Hit Ratio:49.30%20.09m
Cache Miss Ratio:50.70%20.66m
Actual Hit Ratio:49.29%20.08m

Data Demand Efficiency:99.96%19.58m
Data Prefetch Efficiency:0.00%20.47m

CACHE HITS BY CACHE LIST:
  Anonymously Used:0.02%3.46k
  Most Recently Used:99.63%20.02m
  Most Frequently Used:0.33%66.92k
  Most Recently Used Ghost:0.01%2.68k
  Most Frequently Used Ghost:0.00%604

CACHE HITS BY DATA TYPE:
  Demand Data:97.43%19.57m
  Prefetch Data:0.00%639
  Demand Metadata:2.54%510.40k
  Prefetch Metadata:0.03%6.10k

CACHE MISSES BY DATA TYPE:
  Demand Data:0.04%8.56k
  Prefetch Data:99.10%20.47m
  Demand Metadata:0.08%16.61k
  Prefetch Metadata:0.78%160.96k

------------------------------------------------------------------------

L2ARC is disabled

------------------------------------------------------------------------

File-Level Prefetch: (HEALTHY)

DMU Efficiency:775.18k
Hit Ratio:20.75%160.83k
Miss Ratio:79.25%614.36k

Colinear:0
  Hit Ratio:100.00%0
  Miss Ratio:100.00%0

Stride:0
  Hit Ratio:100.00%0
  Miss Ratio:100.00%0

DMU Misc:
Reclaim:0
  Successes:100.00%0
  Failures:100.00%0

Streams:0
  +Resets:100.00%0
  -Resets:100.00%0
  Bogus:0

------------------------------------------------------------------------

VDEV Cache Summary:156.84k
Hit Ratio:1.80%2.83k
Miss Ratio:97.53%152.96k
Delegations:0.67%1.05k

------------------------------------------------------------------------

ZFS Tunables (sysctl):
kern.maxusers                           843
vm.kmem_size                            8273297408
vm.kmem_size_scale                      1
vm.kmem_size_min                        0
vm.kmem_size_max                        1319413950874
vfs.zfs.trim.max_interval               1
vfs.zfs.trim.timeout                    30
vfs.zfs.trim.txg_delay                  32
vfs.zfs.trim.enabled                    0
vfs.zfs.vol.unmap_enabled               1
vfs.zfs.vol.mode                        1
vfs.zfs.version.zpl                     5
vfs.zfs.version.spa                     5000
vfs.zfs.version.acl                     1
vfs.zfs.version.ioctl                   7
vfs.zfs.debug                           0
vfs.zfs.super_owner                     0
vfs.zfs.sync_pass_rewrite               2
vfs.zfs.sync_pass_dont_compress         5
vfs.zfs.sync_pass_deferred_free         2
vfs.zfs.zio.dva_throttle_enabled        1
vfs.zfs.zio.exclude_metadata            0
vfs.zfs.zio.use_uma                     1
vfs.zfs.cache_flush_disable             0
vfs.zfs.zil_replay_disable              0
vfs.zfs.min_auto_ashift                 9
vfs.zfs.max_auto_ashift                 13
vfs.zfs.vdev.trim_max_pending           10000
vfs.zfs.vdev.bio_delete_disable         0
vfs.zfs.vdev.bio_flush_disable          0
vfs.zfs.vdev.queue_depth_pct            1000
vfs.zfs.vdev.write_gap_limit            4096
vfs.zfs.vdev.read_gap_limit             32768
vfs.zfs.vdev.aggregation_limit          131072
vfs.zfs.vdev.trim_max_active            64
vfs.zfs.vdev.trim_min_active            1
vfs.zfs.vdev.scrub_max_active           2
vfs.zfs.vdev.scrub_min_active           1
vfs.zfs.vdev.async_write_max_active     10
vfs.zfs.vdev.async_write_min_active     1
vfs.zfs.vdev.async_read_max_active      3
vfs.zfs.vdev.async_read_min_active      1
vfs.zfs.vdev.sync_write_max_active      10
vfs.zfs.vdev.sync_write_min_active      10
vfs.zfs.vdev.sync_read_max_active       10
vfs.zfs.vdev.sync_read_min_active       10
vfs.zfs.vdev.max_active                 1000
vfs.zfs.vdev.async_write_active_max_dirty_percent60
vfs.zfs.vdev.async_write_active_min_dirty_percent30
vfs.zfs.vdev.mirror.non_rotating_seek_inc1
vfs.zfs.vdev.mirror.non_rotating_inc    0
vfs.zfs.vdev.mirror.rotating_seek_offset1048576
vfs.zfs.vdev.mirror.rotating_seek_inc   5
vfs.zfs.vdev.mirror.rotating_inc        0
vfs.zfs.vdev.trim_on_init               1
vfs.zfs.vdev.cache.bshift               16
vfs.zfs.vdev.cache.size                 10485760
vfs.zfs.vdev.cache.max                  16384
vfs.zfs.vdev.metaslabs_per_vdev         200
vfs.zfs.txg.timeout                     5
vfs.zfs.space_map_blksz                 4096
vfs.zfs.spa_min_slop                    134217728
vfs.zfs.spa_slop_shift                  5
vfs.zfs.spa_asize_inflation             24
vfs.zfs.deadman_enabled                 1
vfs.zfs.deadman_checktime_ms            5000
vfs.zfs.deadman_synctime_ms             1000000
vfs.zfs.debug_flags                     0
vfs.zfs.debugflags                      0
vfs.zfs.recover                         0
vfs.zfs.spa_load_verify_data            1
vfs.zfs.spa_load_verify_metadata        1
vfs.zfs.spa_load_verify_maxinflight     10000
vfs.zfs.ccw_retry_interval              300
vfs.zfs.check_hostid                    1
vfs.zfs.mg_fragmentation_threshold      85
vfs.zfs.mg_noalloc_threshold            0
vfs.zfs.condense_pct                    200
vfs.zfs.metaslab.bias_enabled           1
vfs.zfs.metaslab.lba_weighting_enabled  1
vfs.zfs.metaslab.fragmentation_factor_enabled1
vfs.zfs.metaslab.preload_enabled        1
vfs.zfs.metaslab.preload_limit          3
vfs.zfs.metaslab.unload_delay           8
vfs.zfs.metaslab.load_pct               50
vfs.zfs.metaslab.min_alloc_size         33554432
vfs.zfs.metaslab.df_free_pct            4
vfs.zfs.metaslab.df_alloc_threshold     131072
vfs.zfs.metaslab.debug_unload           0
vfs.zfs.metaslab.debug_load             0
vfs.zfs.metaslab.fragmentation_threshold70
vfs.zfs.metaslab.gang_bang              16777217
vfs.zfs.free_bpobj_enabled              1
vfs.zfs.free_max_blocks                 -1
vfs.zfs.no_scrub_prefetch               0
vfs.zfs.no_scrub_io                     0
vfs.zfs.resilver_min_time_ms            3000
vfs.zfs.free_min_time_ms                1000
vfs.zfs.scan_min_time_ms                1000
vfs.zfs.scan_idle                       50
vfs.zfs.scrub_delay                     4
vfs.zfs.resilver_delay                  2
vfs.zfs.top_maxinflight                 32
vfs.zfs.zfetch.array_rd_sz              1048576
vfs.zfs.zfetch.max_distance             8388608
vfs.zfs.zfetch.min_sec_reap             2
vfs.zfs.zfetch.max_streams              8
vfs.zfs.prefetch_disable                0
vfs.zfs.delay_scale                     500000
vfs.zfs.delay_min_dirty_percent         60
vfs.zfs.dirty_data_sync                 67108864
vfs.zfs.dirty_data_max_percent          10
vfs.zfs.dirty_data_max_max              4294967296
vfs.zfs.dirty_data_max                  851961036
vfs.zfs.max_recordsize                  1048576
vfs.zfs.send_holes_without_birth_time   1
vfs.zfs.mdcomp_disable                  0
vfs.zfs.nopwrite_enabled                1
vfs.zfs.dedup.prefetch                  1
vfs.zfs.l2c_only_size                   0
vfs.zfs.mfu_ghost_data_esize            1933312
vfs.zfs.mfu_ghost_metadata_esize        45613056
vfs.zfs.mfu_ghost_size                  47546368
vfs.zfs.mfu_data_esize                  0
vfs.zfs.mfu_metadata_esize              475136
vfs.zfs.mfu_size                        1127936
vfs.zfs.mru_ghost_data_esize            4771840
vfs.zfs.mru_ghost_metadata_esize        1060864
vfs.zfs.mru_ghost_size                  5832704
vfs.zfs.mru_data_esize                  0
vfs.zfs.mru_metadata_esize              0
vfs.zfs.mru_size                        26115584
vfs.zfs.anon_data_esize                 0
vfs.zfs.anon_metadata_esize             0
vfs.zfs.anon_size                       147456
vfs.zfs.l2arc_norw                      1
vfs.zfs.l2arc_feed_again                1
vfs.zfs.l2arc_noprefetch                1
vfs.zfs.l2arc_feed_min_ms               200
vfs.zfs.l2arc_feed_secs                 1
vfs.zfs.l2arc_headroom                  2
vfs.zfs.l2arc_write_boost               8388608
vfs.zfs.l2arc_write_max                 8388608
vfs.zfs.arc_meta_limit                  13369344
vfs.zfs.arc_free_target                 14047
vfs.zfs.compressed_arc_enabled          1
vfs.zfs.arc_shrink_shift                7
vfs.zfs.arc_average_blocksize           8192
vfs.zfs.arc_min                         52428800
vfs.zfs.arc_max                         53477376

------------------------------------------------------------------------


SSD performance testing completed @ Sat Jun 17 04:04:28 UTC 2017

---------------------------------------------------------------------------=
----


Hopefully, the above proves useful to help track down this issue.

--
Aaron
This message may contain confidential and privileged information. If it has=
 been sent to you in error, please reply to advise the sender of the error =
and then immediately delete it. If you are not the intended recipient, do n=
ot read, copy, disclose or otherwise use this message. The sender disclaims=
 any liability for such unauthorized use. PLEASE NOTE that all incoming e-m=
ails sent to Weatherford e-mail accounts will be archived and may be scanne=
d by us and/or by external service providers to detect and prevent threats =
to our systems, investigate illegal or inappropriate behavior, and/or elimi=
nate unsolicited promotional e-mails (spam). This process could result in d=
eletion of a legitimate e-mail before it is read by its intended recipient =
at our organization. Moreover, based on the scanning results, the full text=
 of e-mails and attachments may be made available to Weatherford security a=
nd other personnel for review and appropriate action. If you have any conce=
rns about this process, please contact us at dataprivacy@weatherford.com.



Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?4561529b83ce4270b09aa0e3b12f299f>