Is there any special requirement for zfs other then these 4 rpms? Since all
kmod addon rpms for rhel collected in elrepo. Wouldn't it be better to add
zfs to elrepo rather then sl-addon?
2013.03.22. 20:30, "Pat Riehecky" <[log in to unmask]> ezt írta:

> The forums are 'unofficial' so go ahead and put whatever you feel is
> appropriate.
>
> Pat
>
> On 03/22/2013 02:08 PM, Brown, Chris (GE Healthcare) wrote:
>
>> All,
>>
>> The preliminary thought process on ZOL was to include it in SL addons for
>> SL 6.4.
>> This will give the SL community a chance to shake it down and provide
>> thoughts/feedback on ZOL.
>> Additionally this provides further permeation of the ZOL project beyond
>> us (GEHC HELiOS), LLNL, Ubuntu, other distros which have integrated it and
>> its existing user base.
>> Linux is and has been in dire need of a file system like ZFS and its
>> features and maturity level. Thus far in our evaluation and testing ZOL has
>> been nothing short of impressive.
>>
>> If things go well with ZOL for SL the next step would be promote ZFS into
>> SL 6.5 directly.
>>
>>
>> Pat,
>> Let me know if it would be ok to additionally post the below to the
>> unofficial SL forums and scientific-linux-users.
>>
>>  FAQ for Questions, Comments, and concerns with ZOL in SL <<
>>>>
>>> *QuickLinks*
>> - Main ZFS on Linux Website: http://zfsonlinux.org/
>> - ZOL FAQ: http://zfsonlinux.org/faq.html
>> - SPL Source Repository: https://github.com/zfsonlinux/**spl<https://github.com/zfsonlinux/spl>
>> - ZFS Source Repository: https://github.com/zfsonlinux/**zfs<https://github.com/zfsonlinux/zfs>
>> - ZFS Announce: https://groups.google.com/a/**zfsonlinux.org/forum/?**
>> fromgroups#!forum/zfs-announce<https://groups.google.com/a/zfsonlinux.org/forum/?fromgroups#!forum/zfs-announce>
>> - ZFS_Best_Practices_Guide (written for Solaris but most things still
>> apply): http://www.solarisinternals.**com/wiki/index.php/ZFS_Best_**
>> Practices_Guide<http://www.solarisinternals.com/wiki/index.php/ZFS_Best_Practices_Guide>
>> - ZFS_Evil_Tuning_Guide (written for Solaris but most things still
>> apply): http://www.solarisinternals.**com/wiki/index.php/ZFS_Evil_**
>> Tuning_Guide<http://www.solarisinternals.com/wiki/index.php/ZFS_Evil_Tuning_Guide>
>>
>> For Providing feedback or general use and installation questions use the
>> following mailing lists
>>    --> scientific-linux-users
>>    Or
>>    --> Existing zfs-discuss (https://groups.google.com/a/**
>> zfsonlinux.org/forum/?**fromgroups#!forum/zfs-discuss<https://groups.google.com/a/zfsonlinux.org/forum/?fromgroups#!forum/zfs-discuss>
>> )
>>
>> - For developmental type questions use the following existing ZOL mailing
>> list:
>>    --> https://groups.google.com/a/**zfsonlinux.org/forum/?**
>> fromgroups#!forum/zfs-devel<https://groups.google.com/a/zfsonlinux.org/forum/?fromgroups#!forum/zfs-devel>
>>
>> - To report SPL bugs:
>>    --> https://github.com/zfsonlinux/**spl/issues<https://github.com/zfsonlinux/spl/issues>
>>
>> - To report ZFS bugs:
>>    --> https://github.com/zfsonlinux/**zfs/issues<https://github.com/zfsonlinux/zfs/issues>
>>
>> *Some general best practices to note with ZOL*
>> - SSD ZIL and L2ARC devices help performance in general
>> - Always try and use low latency SSD devices!
>> - Multipath your disks using multiple HBA whenever possible
>> - Something to note in ZOL was this change in rc12:
>> --> https://github.com/zfsonlinux/**zfs/commit/**
>> 920dd524fb2997225d4b1ac180bcbc**14b045fda6<https://github.com/zfsonlinux/zfs/commit/920dd524fb2997225d4b1ac180bcbc14b045fda6>
>> --> Translation: How ZOL better handles and avoids the situation with the
>> BSD and native Solaris ZFS described here: http://www.nex7.com/node/12
>> - Try to limit pools to no more than 48 disks
>> - If using ZOL to store KVM Virtual Machine images I have found the
>> following setup yields the best performance:
>> --> Using ZVOL's formatted with ext4 with a larger blocksize yield the
>> best performance when combined with KVM and RAW thick or thin provisioned
>> file backed disks.
>> --> Create a zvol as follows (Example): zfs create -V 100G -o
>> volblocksize=64K das0/foo
>> --> After that a simple mkfs.ext4 -L <zvolname> /dev/das0/foo
>> --> mount command (Example) mount /dev/das0/foo /some/mount/point -o
>> noatime
>> --> /some/mount/point is exported via NFS v3
>> --> Feel free to enable NFS async for additional performance however with
>> the understanding on the implications of doing so
>> --> Additionally the qemu/kvm VM disk cache policy set to none with the
>> IO policy set to threaded.
>>
>> *ZFS Examples*
>> # Limit ZFS ARC else it defaults to total system ram size minus 1GB
>> # Example 64GB
>> # Create and add the following to /etc/modprobe.d/zfs.conf --> options
>> zfs zfs_arc_max=68719476736
>>
>> ZFS Zpool Creation Commands (examples are using dm-multipath disks)
>> # EX: Create a 24 Disk Raidz2 (Raid6) pool
>> zpool create das0 raidz2 mpathb mpatha mpathc mpathd mpathe mpathf mpathg
>> mpathh mpathi mpathj mpathk mpathl mpathm mpathn mpatho mpathp mpathq
>> mpathr mpaths mpatht mpathu mpathv mpathw mpathx
>>
>> # EX: Create a 24 Disk Striped Mirrors (Raid10) pool
>> zpool create das0 mirror mpathb mpatha mirror mpathc mpathd mirror mpathe
>> mpathf mirror mpathg mpathh mirror mpathi mpathj mirror mpathk mpathl
>> mirror mpathm mpathn mirror mpatho mpathp mirror mpathq mpathr mirror
>> mpaths mpatht mirror mpathu mpathv mirror mpathw mpathx
>>
>> # EX: Create a 24 Disk Striped Mirrors (Raid10) pool with ashift option
>> # Also set ashift=12 this is required when dealing with Advanced Format
>> drives
>> # Using it with Non AF drives can give a performance boost with some
>> workloads
>> # Using it does decrease overall pool capacity
>> zpool create -o ashift=12 das0 mirror mpathb mpatha mirror mpathc mpathd
>> mirror mpathe mpathf mirror mpathg mpathh mirror mpathi mpathj mirror
>> mpathk mpathl mirror mpathm mpathn mirror mpatho mpathp mirror mpathq
>> mpathr mirror mpaths mpatht mirror mpathu mpathv mirror mpathw mpathx
>>
>> Zpool autoexpand option
>> # Can be specified at pool creation time
>> # Can be set at anytime
>> # Needed if you want to add drives to an existing pool
>> # At pool creation time: -o expand=on
>> # After pool creation: zpool set expand=on
>>
>> Add two SSD striped together as Read cache (l2arc) to a pool
>> zpool add das0 cache /dev/<disk-by-path> /dev/<disk-by-path>
>>
>> Add two SSD striped together as Write cache ZFS intent Log (ZIL) to a pool
>> Add striped zil devices:  zpool add das0 log /dev/<disk-by-path>
>> /dev/<disk-by-path>
>>
>> Create a ZFS Filesystem
>> # EX: create ZFS filesystem named foo in pool das0
>> zfs create das0/foo
>>
>> Create a ZFS zvol
>> # EX: create zvol named foov in pool das0
>> zfs create -V 100G das0/foo
>>
>> Create sparse ZFS zvol with custom blocksize
>> zfs create -s -V 500G -o volblocksize=64K das0/foo
>>
>> Grow a ZFS zvol
>> # EX: grow foov on pool das0 to 500G from 100G
>> # Note if zvol is formatted with and FS use that FS tool to grow the FS
>> after resize
>> # EX: resize2fs (ext4/ext3) xfs_growfs (xfs)
>> zfs set volsize=500G das0/foov
>>
>> Export and Import a zpool
>> # EX: zpool created from devices listed in /dev/mapper and /dev/disk/by-id
>> zpool export
>> zpool import -f -d /dev/mapper -d /dev/disk/by-id -a
>>
>> - Chris
>>
>> -----Original Message-----
>> From: owner-scientific-linux-devel@**listserv.fnal.gov<[log in to unmask]>[mailto:
>> owner-scientific-**[log in to unmask]<[log in to unmask]>]
>> On Behalf Of Andras Horvath
>> Sent: Friday, March 22, 2013 12:29 PM
>> To: Pat Riehecky
>> Cc: Brown, Chris (GE Healthcare); SCIENTIFIC-LINUX-DEVEL@**
>> LISTSERV.FNAL.GOV <[log in to unmask]>
>> Subject: Re: [SCIENTIFIC-LINUX-DEVEL] Initial ZFS On Linux (ZOL) YAML for
>> SL-addons
>>
>> The zfs package did install fine now.
>>
>> Thanks!
>>
>>
>> On Fri, 22 Mar 2013 12:16:34 -0500
>> Pat Riehecky <[log in to unmask]> wrote:
>>
>>  The sl-addons repo has a fixed zfs-modules-dkms package now.
>>>
>>> you will want to run a 'yum clean expire-cache' to catch the updates.
>>>
>>> Since DKMS is in EPEL6 where it is actively maintained, adding it to
>>> SL Addons would create work on maintenance without adding much.
>>>
>>> It does make loading some packages more complex,
>>>
>>> Pat
>>>
>>>
>>> On 03/22/2013 12:00 PM, Andras Horvath wrote:
>>>
>>>> Hi,
>>>>
>>>> I've just tried to install ZFS on a recently update 6.4 system (64
>>>> bit). Yum complains about dependency error, see below.
>>>>
>>>> Even manually installing spl-modules-dkms doesn't help, though it's
>>>> complaining about this one.
>>>>
>>>> BTW, is it normal that the dkms package is available only from Epel
>>>> repo and not the main one?
>>>>
>>>>
>>>> Thanks,
>>>> Andras
>>>>
>>>>
>>>> # lsb_release -d
>>>> Description:    Scientific Linux release 6.4 (Carbon)
>>>>
>>>> # yum install --disableplugin=fastestmirror --enablerepo=sl-addons
>>>> zfs Loaded plugins: protectbase, refresh-packagekit, security
>>>> 82 packages excluded due to repository protections
>>>> Setting up Install Process
>>>> Resolving Dependencies
>>>> --> Running transaction check
>>>> ---> Package zfs.x86_64 0:0.6.0-rc14.el6 will be installed
>>>> --> Processing Dependency: zfs-modules for package:
>>>> zfs-0.6.0-rc14.el6.x86_64 --> Running transaction check
>>>> ---> Package zfs-modules-dkms.noarch 0:0.6.0-rc14.el6 will be
>>>> installed --> Processing Dependency: spl-modules-dkms = X for
>>>> package: zfs-modules-dkms-0.6.0-rc14.**el6.noarch --> Finished
>>>> Dependency Resolution Error: Package:
>>>> zfs-modules-dkms-0.6.0-rc14.**el6.noarch (sl-addons) Requires:
>>>> spl-modules-dkms = X Installed:
>>>> spl-modules-dkms-0.6.0-rc14.**el6.noarch (@sl-addons)
>>>> spl-modules-dkms = 0.6.0-rc14.el6 You could try using --skip-broken
>>>> to work around the problem
>>>>
>>>>
>>>> On Fri, 22 Mar 2013 11:16:14 -0500
>>>> Pat Riehecky <[log in to unmask]> wrote:
>>>>
>>>>  The HELiOS folks, Connie, and I have been communication about this
>>>>> for some time.  So the quick turn around on our end is not
>>>>> unexpected.
>>>>>
>>>>> I've just now posted the packages to sl-addons (x86_64 ONLY).
>>>>>
>>>>> Thanks Chris!
>>>>>
>>>>> Pat
>>>>>
>>>>> On 03/22/2013 11:08 AM, Brown, Chris (GE Healthcare) wrote:
>>>>>
>>>>>> Team,
>>>>>>
>>>>>> Attached is the YAML for the initial ZFS inclusion into SL addons.
>>>>>>
>>>>>> *Release Explanation*
>>>>>>
>>>>>> ZFS on Linux (ZOL) is currently manually added to HELiOS
>>>>>> (Healthcare Enterprise Linux Operating System).
>>>>>>
>>>>>> HELiOS is a spin of Scientific Linux created and maintained by the
>>>>>> GE Healthcare Compute Systems team (CST).
>>>>>>
>>>>>> HELiOS strives to be as upstream as possible thus including ZOL
>>>>>> into SL helps us better maintain upstream purity in HELiOS.
>>>>>>
>>>>>> Including ZOL in SL also allows the rest of the SL community to
>>>>>> benefit from our work with ZOL.
>>>>>>
>>>>>> **Release Notes**
>>>>>>
>>>>>> Core ZFS development started in 2001 with ZFS being Officially
>>>>>> released by Sun in 2004.
>>>>>>
>>>>>> Testing and evaluation of ZOL by CST has shown better performance,
>>>>>> scalability, and stability, then BTRFS.
>>>>>>
>>>>>> As a result of the maturity of core ZFS, ZOL inherits ZFS features
>>>>>> which are many years ahead of BTRFS.
>>>>>>
>>>>>> Testing by CST with ZOL on a proper hardware setup with proper SSD
>>>>>> ZIL/L2ARC devices has shown ZOL to yield better performance then
>>>>>> both BTRFS and native Solaris ZFS.
>>>>>>
>>>>>> Also performance tests of ZOL by CST has yielded better
>>>>>> performance results then running the equivalent  tests on
>>>>>> Sun/Oracle or Nexenta ZFS storage appliances.
>>>>>>
>>>>>> Additional testing by CST of ZOL has shown that when combined with
>>>>>> GlusterFS the results are a very powerful, redundant and almost
>>>>>> infinitely scalable storage solution.
>>>>>>
>>>>>> ZOL is the work of Lawrence Livermore National Laboratory (LLNL)
>>>>>> under Contract No. DE-AC52-07NA27344 (Contract 44) between the
>>>>>> U.S. Department of Energy (DOE) and Lawrence Livermore National
>>>>>> Security, LLC (LLNS) for the operation of LLNL.
>>>>>>
>>>>>> regards,
>>>>>>
>>>>>> *Chris Brown*
>>>>>> *GE Healthcare Technologies
>>>>>> **/Compute Systems Architect/*
>>>>>>
>>>>>>
>>>
>
> --
> Pat Riehecky
>
> Scientific Linux developer
> http://www.scientificlinux.**org/ <http://www.scientificlinux.org/>
>