Is there any special requirement for zfs other then these 4 rpms? Since all kmod addon rpms for rhel collected in elrepo. Wouldn't it be better to add zfs to elrepo rather then sl-addon? 2013.03.22. 20:30, "Pat Riehecky" <[log in to unmask]> ezt írta: > The forums are 'unofficial' so go ahead and put whatever you feel is > appropriate. > > Pat > > On 03/22/2013 02:08 PM, Brown, Chris (GE Healthcare) wrote: > >> All, >> >> The preliminary thought process on ZOL was to include it in SL addons for >> SL 6.4. >> This will give the SL community a chance to shake it down and provide >> thoughts/feedback on ZOL. >> Additionally this provides further permeation of the ZOL project beyond >> us (GEHC HELiOS), LLNL, Ubuntu, other distros which have integrated it and >> its existing user base. >> Linux is and has been in dire need of a file system like ZFS and its >> features and maturity level. Thus far in our evaluation and testing ZOL has >> been nothing short of impressive. >> >> If things go well with ZOL for SL the next step would be promote ZFS into >> SL 6.5 directly. >> >> >> Pat, >> Let me know if it would be ok to additionally post the below to the >> unofficial SL forums and scientific-linux-users. >> >> FAQ for Questions, Comments, and concerns with ZOL in SL << >>>> >>> *QuickLinks* >> - Main ZFS on Linux Website: http://zfsonlinux.org/ >> - ZOL FAQ: http://zfsonlinux.org/faq.html >> - SPL Source Repository: https://github.com/zfsonlinux/**spl<https://github.com/zfsonlinux/spl> >> - ZFS Source Repository: https://github.com/zfsonlinux/**zfs<https://github.com/zfsonlinux/zfs> >> - ZFS Announce: https://groups.google.com/a/**zfsonlinux.org/forum/?** >> fromgroups#!forum/zfs-announce<https://groups.google.com/a/zfsonlinux.org/forum/?fromgroups#!forum/zfs-announce> >> - ZFS_Best_Practices_Guide (written for Solaris but most things still >> apply): http://www.solarisinternals.**com/wiki/index.php/ZFS_Best_** >> Practices_Guide<http://www.solarisinternals.com/wiki/index.php/ZFS_Best_Practices_Guide> >> - ZFS_Evil_Tuning_Guide (written for Solaris but most things still >> apply): http://www.solarisinternals.**com/wiki/index.php/ZFS_Evil_** >> Tuning_Guide<http://www.solarisinternals.com/wiki/index.php/ZFS_Evil_Tuning_Guide> >> >> For Providing feedback or general use and installation questions use the >> following mailing lists >> --> scientific-linux-users >> Or >> --> Existing zfs-discuss (https://groups.google.com/a/** >> zfsonlinux.org/forum/?**fromgroups#!forum/zfs-discuss<https://groups.google.com/a/zfsonlinux.org/forum/?fromgroups#!forum/zfs-discuss> >> ) >> >> - For developmental type questions use the following existing ZOL mailing >> list: >> --> https://groups.google.com/a/**zfsonlinux.org/forum/?** >> fromgroups#!forum/zfs-devel<https://groups.google.com/a/zfsonlinux.org/forum/?fromgroups#!forum/zfs-devel> >> >> - To report SPL bugs: >> --> https://github.com/zfsonlinux/**spl/issues<https://github.com/zfsonlinux/spl/issues> >> >> - To report ZFS bugs: >> --> https://github.com/zfsonlinux/**zfs/issues<https://github.com/zfsonlinux/zfs/issues> >> >> *Some general best practices to note with ZOL* >> - SSD ZIL and L2ARC devices help performance in general >> - Always try and use low latency SSD devices! >> - Multipath your disks using multiple HBA whenever possible >> - Something to note in ZOL was this change in rc12: >> --> https://github.com/zfsonlinux/**zfs/commit/** >> 920dd524fb2997225d4b1ac180bcbc**14b045fda6<https://github.com/zfsonlinux/zfs/commit/920dd524fb2997225d4b1ac180bcbc14b045fda6> >> --> Translation: How ZOL better handles and avoids the situation with the >> BSD and native Solaris ZFS described here: http://www.nex7.com/node/12 >> - Try to limit pools to no more than 48 disks >> - If using ZOL to store KVM Virtual Machine images I have found the >> following setup yields the best performance: >> --> Using ZVOL's formatted with ext4 with a larger blocksize yield the >> best performance when combined with KVM and RAW thick or thin provisioned >> file backed disks. >> --> Create a zvol as follows (Example): zfs create -V 100G -o >> volblocksize=64K das0/foo >> --> After that a simple mkfs.ext4 -L <zvolname> /dev/das0/foo >> --> mount command (Example) mount /dev/das0/foo /some/mount/point -o >> noatime >> --> /some/mount/point is exported via NFS v3 >> --> Feel free to enable NFS async for additional performance however with >> the understanding on the implications of doing so >> --> Additionally the qemu/kvm VM disk cache policy set to none with the >> IO policy set to threaded. >> >> *ZFS Examples* >> # Limit ZFS ARC else it defaults to total system ram size minus 1GB >> # Example 64GB >> # Create and add the following to /etc/modprobe.d/zfs.conf --> options >> zfs zfs_arc_max=68719476736 >> >> ZFS Zpool Creation Commands (examples are using dm-multipath disks) >> # EX: Create a 24 Disk Raidz2 (Raid6) pool >> zpool create das0 raidz2 mpathb mpatha mpathc mpathd mpathe mpathf mpathg >> mpathh mpathi mpathj mpathk mpathl mpathm mpathn mpatho mpathp mpathq >> mpathr mpaths mpatht mpathu mpathv mpathw mpathx >> >> # EX: Create a 24 Disk Striped Mirrors (Raid10) pool >> zpool create das0 mirror mpathb mpatha mirror mpathc mpathd mirror mpathe >> mpathf mirror mpathg mpathh mirror mpathi mpathj mirror mpathk mpathl >> mirror mpathm mpathn mirror mpatho mpathp mirror mpathq mpathr mirror >> mpaths mpatht mirror mpathu mpathv mirror mpathw mpathx >> >> # EX: Create a 24 Disk Striped Mirrors (Raid10) pool with ashift option >> # Also set ashift=12 this is required when dealing with Advanced Format >> drives >> # Using it with Non AF drives can give a performance boost with some >> workloads >> # Using it does decrease overall pool capacity >> zpool create -o ashift=12 das0 mirror mpathb mpatha mirror mpathc mpathd >> mirror mpathe mpathf mirror mpathg mpathh mirror mpathi mpathj mirror >> mpathk mpathl mirror mpathm mpathn mirror mpatho mpathp mirror mpathq >> mpathr mirror mpaths mpatht mirror mpathu mpathv mirror mpathw mpathx >> >> Zpool autoexpand option >> # Can be specified at pool creation time >> # Can be set at anytime >> # Needed if you want to add drives to an existing pool >> # At pool creation time: -o expand=on >> # After pool creation: zpool set expand=on >> >> Add two SSD striped together as Read cache (l2arc) to a pool >> zpool add das0 cache /dev/<disk-by-path> /dev/<disk-by-path> >> >> Add two SSD striped together as Write cache ZFS intent Log (ZIL) to a pool >> Add striped zil devices: zpool add das0 log /dev/<disk-by-path> >> /dev/<disk-by-path> >> >> Create a ZFS Filesystem >> # EX: create ZFS filesystem named foo in pool das0 >> zfs create das0/foo >> >> Create a ZFS zvol >> # EX: create zvol named foov in pool das0 >> zfs create -V 100G das0/foo >> >> Create sparse ZFS zvol with custom blocksize >> zfs create -s -V 500G -o volblocksize=64K das0/foo >> >> Grow a ZFS zvol >> # EX: grow foov on pool das0 to 500G from 100G >> # Note if zvol is formatted with and FS use that FS tool to grow the FS >> after resize >> # EX: resize2fs (ext4/ext3) xfs_growfs (xfs) >> zfs set volsize=500G das0/foov >> >> Export and Import a zpool >> # EX: zpool created from devices listed in /dev/mapper and /dev/disk/by-id >> zpool export >> zpool import -f -d /dev/mapper -d /dev/disk/by-id -a >> >> - Chris >> >> -----Original Message----- >> From: owner-scientific-linux-devel@**listserv.fnal.gov<[log in to unmask]>[mailto: >> owner-scientific-**[log in to unmask]<[log in to unmask]>] >> On Behalf Of Andras Horvath >> Sent: Friday, March 22, 2013 12:29 PM >> To: Pat Riehecky >> Cc: Brown, Chris (GE Healthcare); SCIENTIFIC-LINUX-DEVEL@** >> LISTSERV.FNAL.GOV <[log in to unmask]> >> Subject: Re: [SCIENTIFIC-LINUX-DEVEL] Initial ZFS On Linux (ZOL) YAML for >> SL-addons >> >> The zfs package did install fine now. >> >> Thanks! >> >> >> On Fri, 22 Mar 2013 12:16:34 -0500 >> Pat Riehecky <[log in to unmask]> wrote: >> >> The sl-addons repo has a fixed zfs-modules-dkms package now. >>> >>> you will want to run a 'yum clean expire-cache' to catch the updates. >>> >>> Since DKMS is in EPEL6 where it is actively maintained, adding it to >>> SL Addons would create work on maintenance without adding much. >>> >>> It does make loading some packages more complex, >>> >>> Pat >>> >>> >>> On 03/22/2013 12:00 PM, Andras Horvath wrote: >>> >>>> Hi, >>>> >>>> I've just tried to install ZFS on a recently update 6.4 system (64 >>>> bit). Yum complains about dependency error, see below. >>>> >>>> Even manually installing spl-modules-dkms doesn't help, though it's >>>> complaining about this one. >>>> >>>> BTW, is it normal that the dkms package is available only from Epel >>>> repo and not the main one? >>>> >>>> >>>> Thanks, >>>> Andras >>>> >>>> >>>> # lsb_release -d >>>> Description: Scientific Linux release 6.4 (Carbon) >>>> >>>> # yum install --disableplugin=fastestmirror --enablerepo=sl-addons >>>> zfs Loaded plugins: protectbase, refresh-packagekit, security >>>> 82 packages excluded due to repository protections >>>> Setting up Install Process >>>> Resolving Dependencies >>>> --> Running transaction check >>>> ---> Package zfs.x86_64 0:0.6.0-rc14.el6 will be installed >>>> --> Processing Dependency: zfs-modules for package: >>>> zfs-0.6.0-rc14.el6.x86_64 --> Running transaction check >>>> ---> Package zfs-modules-dkms.noarch 0:0.6.0-rc14.el6 will be >>>> installed --> Processing Dependency: spl-modules-dkms = X for >>>> package: zfs-modules-dkms-0.6.0-rc14.**el6.noarch --> Finished >>>> Dependency Resolution Error: Package: >>>> zfs-modules-dkms-0.6.0-rc14.**el6.noarch (sl-addons) Requires: >>>> spl-modules-dkms = X Installed: >>>> spl-modules-dkms-0.6.0-rc14.**el6.noarch (@sl-addons) >>>> spl-modules-dkms = 0.6.0-rc14.el6 You could try using --skip-broken >>>> to work around the problem >>>> >>>> >>>> On Fri, 22 Mar 2013 11:16:14 -0500 >>>> Pat Riehecky <[log in to unmask]> wrote: >>>> >>>> The HELiOS folks, Connie, and I have been communication about this >>>>> for some time. So the quick turn around on our end is not >>>>> unexpected. >>>>> >>>>> I've just now posted the packages to sl-addons (x86_64 ONLY). >>>>> >>>>> Thanks Chris! >>>>> >>>>> Pat >>>>> >>>>> On 03/22/2013 11:08 AM, Brown, Chris (GE Healthcare) wrote: >>>>> >>>>>> Team, >>>>>> >>>>>> Attached is the YAML for the initial ZFS inclusion into SL addons. >>>>>> >>>>>> *Release Explanation* >>>>>> >>>>>> ZFS on Linux (ZOL) is currently manually added to HELiOS >>>>>> (Healthcare Enterprise Linux Operating System). >>>>>> >>>>>> HELiOS is a spin of Scientific Linux created and maintained by the >>>>>> GE Healthcare Compute Systems team (CST). >>>>>> >>>>>> HELiOS strives to be as upstream as possible thus including ZOL >>>>>> into SL helps us better maintain upstream purity in HELiOS. >>>>>> >>>>>> Including ZOL in SL also allows the rest of the SL community to >>>>>> benefit from our work with ZOL. >>>>>> >>>>>> **Release Notes** >>>>>> >>>>>> Core ZFS development started in 2001 with ZFS being Officially >>>>>> released by Sun in 2004. >>>>>> >>>>>> Testing and evaluation of ZOL by CST has shown better performance, >>>>>> scalability, and stability, then BTRFS. >>>>>> >>>>>> As a result of the maturity of core ZFS, ZOL inherits ZFS features >>>>>> which are many years ahead of BTRFS. >>>>>> >>>>>> Testing by CST with ZOL on a proper hardware setup with proper SSD >>>>>> ZIL/L2ARC devices has shown ZOL to yield better performance then >>>>>> both BTRFS and native Solaris ZFS. >>>>>> >>>>>> Also performance tests of ZOL by CST has yielded better >>>>>> performance results then running the equivalent tests on >>>>>> Sun/Oracle or Nexenta ZFS storage appliances. >>>>>> >>>>>> Additional testing by CST of ZOL has shown that when combined with >>>>>> GlusterFS the results are a very powerful, redundant and almost >>>>>> infinitely scalable storage solution. >>>>>> >>>>>> ZOL is the work of Lawrence Livermore National Laboratory (LLNL) >>>>>> under Contract No. DE-AC52-07NA27344 (Contract 44) between the >>>>>> U.S. Department of Energy (DOE) and Lawrence Livermore National >>>>>> Security, LLC (LLNS) for the operation of LLNL. >>>>>> >>>>>> regards, >>>>>> >>>>>> *Chris Brown* >>>>>> *GE Healthcare Technologies >>>>>> **/Compute Systems Architect/* >>>>>> >>>>>> >>> > > -- > Pat Riehecky > > Scientific Linux developer > http://www.scientificlinux.**org/ <http://www.scientificlinux.org/> >