From 5327fb1c0fdc25c60e880940f01071b3b8d2e4b6 Mon Sep 17 00:00:00 2001 From: Chris Gianelloni Date: Thu, 6 Apr 2017 16:40:49 -0400 Subject: [PATCH] Version bump to 2.10.0 --- CHANGELOG.md | 51 +++++++++++++++++++++++++++++++++++++++++++++++++++ metadata.json | 30 ++++-------------------------- metadata.rb | 34 ++-------------------------------- 3 files changed, 57 insertions(+), 58 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 285c47d..3825d96 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,13 +1,21 @@ hadoop CHANGELOG =============== +v2.10.0 (Apr 6, 2017) +--------------------- + +- Update tests ( Issues: #314 #318 ) +- Support HDP 2.6.0.3-8 ( Issue: #319 ) + v2.9.0 (Dec 13, 2016) --------------------- + - Update CDH/HDP default versions ( Issue: #310 ) - Support HDP 2.5.3.0 ( Issue: #311 ) v2.8.0 (Dec 6, 2016) -------------------- + - Updates to testing configurations ( Issues: #301 #302 #304 ) - Update HDP versions when short versions given ( Issue: #303 ) - Support 2015 and 2016 Amazon Linux AMIs ( Issue: #305 ) @@ -17,12 +25,14 @@ v2.8.0 (Dec 6, 2016) v2.7.1 (Oct 25, 2016) --------------------- + - Switch to new Ruby hash syntax ( Issue #297) - Add non-cookbook files to chefignore ( Issue #298) - Support for Spark on IOP ( Issue #299 ) v2.7.0 (Oct 17, 2016) --------------------- + - Pull in code from forks ( Issue #282 ) - Use correct License file ( Issue #283 ) - Set distribution defaults in attributes instead of recipes ( Issue #284 ) @@ -36,6 +46,7 @@ v2.7.0 (Oct 17, 2016) v2.6.0 (Sep 9, 2016) -------------------- + - Support for IBM Open Platform ( Issue: #226 ) - Support IOP 4.2.0.0 (and others) ( Issue: #278 ) - HDP: Remove 2.2.1.0 and 2.2.6.3 and add 2.3.6.0 ( Issue: #279 ) @@ -43,6 +54,7 @@ v2.6.0 (Sep 9, 2016) v2.5.0 (Sep 1, 2016) -------------------- + - Remove `policy_rcd` and use `dpkg_autostart` cookbook ( Issue: #272 ) - Update testing framework ( Issue: #273 ) - Use major_platform_version for centos5 or HDP >= 2.3 ( Issue: #275 ) @@ -50,10 +62,12 @@ v2.5.0 (Sep 1, 2016) v2.4.1 (Aug 1, 2016) -------------------- + - Support for HDP 2.4.2.0 ( Issue: #270 ) v2.4.0 (Jul 27, 2016) --------------------- + - Update ext.js download URL to Cloudera, per @jeremyhahn ( Issue: #265 ) - Restrict Gem versions on older Ruby ( Issue: #266 ) - Set ZOOPIDFILE ( Issues: #267 [COOK-105](https://issues.cask.co/browse/COOK-105) ) @@ -61,22 +75,26 @@ v2.4.0 (Jul 27, 2016) v2.3.3 (Jun 16, 2016) --------------------- + - Setting YARN_LOG_DIR is a redundant test, use text-based ( Issue: #261 ) - ZooKeeper filesystem objects should use zookeeper group ( Issues: #262 [COOK-100](https://issues.cask.co/browse/COOK-100) ) - Use upstream ulimit cookbook for testing ( Issue: #263 ) v2.3.2 (May 6, 2016) -------------------- + - Improve test coverage ( Issue: #258 ) - HDP 2.2+ log directories are not modified on Ubuntu ( Issues: #259 [COOK-96](https://issues.cask.co/browse/COOK-96) ) v2.3.1 (Apr 19, 2016) --------------------- + - Allow overriding distribution_version at default ( Issues: #256 [COOK-93](https://issues.cask.co/browse/COOK-93) ) - Set SPARK_DIST_CLASSPATH and redirect STDERR to logs ( Issue: #257 ) v2.3.0 (Apr 12, 2016) --------------------- + - Change spark-defaults from XML to .conf property file ( Issue: #241 ) - Update default versions: HDP 2.3.4.7, CDH 5.6.0 ( Issue: #242 ) - Support HDP 2.3.4.7 and 2.4.0.0 ( Issue: #250 ) @@ -87,6 +105,7 @@ v2.3.0 (Apr 12, 2016) v2.2.1 (Feb 24, 2016) --------------------- + - Support for HDP 2.3.4.0 per @kriszentner ( Issue: #243 ) - Style updates ( Issues: #244 #247 ) - Support for Bigtop 1.0.0 relocated repositories ( Issue: #245 ) @@ -94,6 +113,7 @@ v2.2.1 (Feb 24, 2016) v2.2.0 (Dec 16, 2015) --------------------- + - Add Code Climate badge ( Issue: #232 ) - Syntax fix in metadata.rb per @mrjefftang ( Issue: #234 ) - Fix up ImmutableMash/Array for older Chef per @mrjefftang ( Issue: #235 ) @@ -104,6 +124,7 @@ v2.2.0 (Dec 16, 2015) v2.1.0 (Dec 7, 2015) -------------------- + - Add support for Apache Storm ( Issue: #223 ) - Support Bigtop 1.0.0 ( Issue: #224 ) - Update minimum `apt` cookbook dependency ( Issue: #227 ) @@ -115,32 +136,38 @@ v2.1.0 (Dec 7, 2015) v2.0.9 (Sep 16, 2015) --------------------- + - Support later HDP 2.1 and HDP 2.2 updates on Ubuntu ( Issue: #225 ) v2.0.8 (Sep 15, 2015) --------------------- + - Fix Hive init scripts, per @QuentinFra ( Issue: #220 ) - Correct JSVC_HOME for HDP 2.0 ( Issues: #221 [COOK-70](https://issues.cask.co/browse/COOK-70) ) - Support HDP 2.2.8.0 ( Issue: #222 ) v2.0.7 (Aug 21, 2015) --------------------- + - Fix Hive sql_connector jar on Ubuntu ( Issues: #216 [COOK-65](https://issues.cask.co/browse/COOK-65) ) - Style updates ( Issue: #217 ) - Set Yarn increment-allocation appropriately for Fair Scheduler ( Issues: #218 [COOK-67](https://issues.cask.co/browse/COOK-67) ) v2.0.6 (Jul 30, 2015) --------------------- + - Fix Spark CONF_DIR ( Issue: #215 ) v2.0.5 (Jul 30, 2015) --------------------- + - Support HDP 2.2.6.3 ( Issue: #212 ) - Keep HADOOP_CLASSPATH before Tez's CLASSPATH ( Issue: #213 ) - Support HDP 2.3.0.0 ( Issue: #214 ) v2.0.4 (Jul 23, 2015) --------------------- + - Fix ChefSpec ( Issue: #207 ) - Support HDP 2.1.15.0, 2.2.4.4, and 2.2.6.0 ( Issue: #208 ) - HiveServer2 process fix per @jsh2134 ( Issue: #210 ) @@ -148,22 +175,26 @@ v2.0.4 (Jul 23, 2015) v2.0.3 (Jun 25, 2015) --------------------- + - Config files should be root owned ( Issue: #204 ) - Fix disable THP Compaction ( Issues: #205 [COOK-57](https://issues.cask.co/browse/COOK-57) ) - Fix init for EXE_ARGS ending in ampersand ( Issues: #206 [COOK-59](https://issues.cask.co/browse/COOK-59) ) v2.0.2 (Jun 12, 2015) --------------------- + - Don't make /etc/default files executable ( Issue: #201 ) - Remove Vagrantfile ( Issue: #202 ) - Fix Ubuntu init ( Issue: #203 ) v2.0.1 (Jun 9, 2015) -------------------- + - Supply /etc/default/hbase for hbase binary ( Issue: #200 ) v2.0.0 (Jun 8, 2015) -------------------- + - Transparent Hugepages are not universally available, per @jdecello and @taverentech ( Issue: #156 ) - Support HDP 2.2.4.2 repo ( Issues: #160 #186 ) - Fix YARN/Hive/Oozie PATHs for HDP 2.2 ( Issue: #161 ) @@ -196,42 +227,50 @@ v2.0.0 (Jun 8, 2015) v1.13.1 (Apr 15, 2015) ----------------------- + - Fix YARN AM staging dir ( Issues: #157 [COOK-30](https://issues.cask.co/browse/COOK-30) ) - Support HDP 2.0.13.0 and bump HDP-UTILS to 1.1.0.20 ( Issue: #158 ) - Document issue tracker location ( Issues: #159 [COOK-32](https://issues.cask.co/browse/COOK-32) ) v1.13.0 (Mar 31, 2015) ---------------------- + - Enable system tuning ( Issue: #148 ) - Test against more Ruby versions ( Issue: #153 ) - Fix guard on mapreduce.jobhistory.done-dir ( Issue: #154 ) v1.12.0 (Mar 20, 2015) ---------------------- + - Support yarn.app.mapreduce.am.staging-dir ( Issue: #150 ) - Support mapreduce.jobhistory.done-dir and mapreduce.jobhistory.intermediate-done-dir ( Issue: #151 ) - Tests for #135 and #150 ( Issue: #152 ) v1.11.2 (Mar 9, 2015) --------------------- + - Prefix internal recipes with underscore ( Issue: #147 ) - Fix Java 7 check ( Issues: #149 [COOK-27](https://issues.cask.co/browse/COOK-27) ) v1.11.1 (Feb 27, 2015) ---------------------- + - Packaging fix v1.11.0 (Feb 27, 2015) ---------------------- + - Stop packages from auto-starting on install ( Issues: #145 [COOK-26](https://issues.cask.co/browse/COOK-26) ) - Fail fast on invalid distribution ( Issues: #146 [COOK-25](https://issues.cask.co/browse/COOK-25) ) v1.10.1 (Feb 24, 2015) ---------------------- + - HDP Repo fix ( Issues: #144 [COOK-24](https://issues.cask.co/browse/COOK-24) ) v1.10.0 (Feb 24, 2015) ---------------------- + - Enforce Java 7 or higher on CDH 5.3 ( Issues: #140 [COOK-18](https://issues.cask.co/browse/COOK-18) ) - Default `hive.metastore.uris` ( Issues: #141 [COOK-19](https://issues.cask.co/browse/COOK-19) ) - HDP 2.2 support ( Issues: #142 [COOK-16](https://issues.cask.co/browse/COOK-16) ) @@ -239,20 +278,24 @@ v1.10.0 (Feb 24, 2015) v1.9.2 (Jan 8, 2015) -------------------- + - Defaults for log4j ( Issue: #139 ) v1.9.1 (Dec 9, 2014) -------------------- + - Spark tests for #129 ( Issue: #133 ) - Improve *_LOG_DIR symlink handling ( Issue: #134 ) - Fix PATH to `jsvc` in `/etc/default/hadoop` ( Issue: #135 ) v1.9.0 (Dec 8, 2014) -------------------- + - Tez support from @mandrews ( Issues: #127 #132 ) v1.8.1 (Dec 8, 2014) -------------------- + - Ubuntu Trusty support for CDH5 ( Issue: #128 ) - Spark MLib requires `libgfortran.so.3` ( Issue: #129 ) - Simplify `container-executor.cfg` ( Issue: #130 ) @@ -260,6 +303,7 @@ v1.8.1 (Dec 8, 2014) v1.8.0 (Nov 24, 2014) --------------------- + - Opportunistic creation of `hive.exec.local.scratchdir` ( Issue: #117 ) - Only use `hadoop::repo` for Hive ( Issue: #120 ) - More Oozie tests ( Issue: #121 ) @@ -271,11 +315,13 @@ v1.8.0 (Nov 24, 2014) v1.7.1 (Nov 5, 2014) -------------------- + - Initial Oozie tests ( Issue: #118 ) - Hotfix symlink log dirs ( Issue: #119 ) v1.7.0 (Nov 5, 2014) -------------------- + - Use Java 7 by default ( Issue: #108 [COOK-5](https://issues.cask.co/browse/COOK-5) ) - Use HDP 2.1 by default ( Issue: #109 ) - Update tests ( Issues: #110 #111 #114 #115 #116 ) @@ -283,14 +329,17 @@ v1.7.0 (Nov 5, 2014) v1.6.1 (Oct 16, 2014) --------------------- + - Update Bigtop to `0.8.0` release ( Issues: #106 #107 [COOK-1](https://issues.cask.co/browse/COOK-1) ) v1.6.0 (Oct 16, 2014) --------------------- + - Add Bigtop support ( Issue: #105 [COOK-1](https://issues.cask.co/browse/COOK-1) ) v1.5.0 (Sep 25, 2014) --------------------- + This release adds Flume support to the cookbook. - Update test-kitchen to use more recipes ( Issue: #95 ) @@ -300,11 +349,13 @@ This release adds Flume support to the cookbook. v1.4.1 (Sep 18, 2014) --------------------- + - Add `zookeeper` group after package installs ( Issue: #96 ) - Code consistency updates ( Issue: #97 ) v1.4.0 (Sep 18, 2014) --------------------- + - Support Amazon Linux ( Issues: #84 #90 ) - Remove addition of `zookeeper` user/group ( Issue: #87 ) - Add support for HDP 2.1.5.0 ( Issue: #88 ) diff --git a/metadata.json b/metadata.json index f592b88..4856078 100644 --- a/metadata.json +++ b/metadata.json @@ -4,7 +4,7 @@ "long_description": "# hadoop cookbook\n\n[![Cookbook Version](http://img.shields.io/cookbook/v/hadoop.svg)](https://supermarket.chef.io/cookbooks/hadoop)\n[![Apache License 2.0](http://img.shields.io/badge/license-apache%202.0-green.svg)](http://opensource.org/licenses/Apache-2.0)\n[![Build Status](http://img.shields.io/travis/caskdata/hadoop_cookbook.svg)](http://travis-ci.org/caskdata/hadoop_cookbook)\n[![Code Climate](https://codeclimate.com/github/caskdata/hadoop_cookbook/badges/gpa.svg)](https://codeclimate.com/github/caskdata/hadoop_cookbook)\n\n# Requirements\n\nThis cookbook may work on earlier versions, but these are the minimal tested versions.\n\n* Chef 11.4.0+\n* CentOS 6.4+\n* Debian 6.0+\n* Ubuntu 12.04+\n\nThis cookbook assumes that you have a working Java installation. It has been tested using version `1.21.2` of the `java` cookbook, using Oracle JDK 7. If you plan on using Hive with a database other than the embedded Derby, you will need to provide it and set it up prior to starting Hive Metastore service.\n\n# Usage\n\nThis cookbook is designed to be used with a wrapper cookbook or a role with settings for configuring Hadoop. The services should work out of the box on a single host, but little validation is done that you have made a working Hadoop configuration. The cookbook is attribute-driven and is suitable for use via either `chef-client` or `chef-solo` since it does not use any server-based functionality. The cookbook defines service definitions for each Hadoop service, but it does not enable or start them, by default.\n\nFor more information, read the [Wrapping this cookbook](https://github.com/caskdata/hadoop_cookbook/wiki/Wrapping-this-cookbook) wiki entry.\n\n# Attributes\n\nAttributes for this cookbook define the configuration files for Hadoop and its various services. Hadoop configuration files are XML files, with name/value property pairs. The attribute name determines which file the property is placed and the property name. The attribute value is the property value. The attribute `hadoop['core_site']['fs.defaultFS']` will configure a property named `fs.defaultFS` in `core-site.xml` in `hadoop['conf_dir']`. All attribute values are taken as-is and only minimal configuration checking is done on values. It is up to the user to provide a valid configuration for your cluster.\n\nAttribute Tree | File | Location\n-------------- | ---- | --------\nflume['flume_conf'] | flume.conf | `flume['conf_dir']`\nhadoop['capacity_scheduler'] | capacity-scheduler.xml | `hadoop['conf_dir']`\nhadoop['container_executor'] | container-executor.cfg | `hadoop['conf_dir']`\nhadoop['core_site'] | core-site.xml | `hadoop['conf_dir']`\nhadoop['fair_scheduler'] | fair-scheduler.xml | `hadoop['conf_dir']`\nhadoop['hadoop_env'] | hadoop-env.sh | `hadoop['conf_dir']`\nhadoop['hadoop_metrics'] | hadoop-metrics.properties | `hadoop['conf_dir']`\nhadoop['hadoop_policy'] | hadoop-policy.xml | `hadoop['conf_dir']`\nhadoop['hdfs_site'] | hdfs-site.xml | `hadoop['conf_dir']`\nhadoop['log4j'] | log4j.properties | `hadoop['conf_dir']`\nhadoop['mapred_env'] | mapred-env.sh | `hadoop['conf_dir']`\nhadoop['mapred_site'] | mapred-site.xml | `hadoop['conf_dir']`\nhadoop['yarn_env'] | yarn-env.sh | `hadoop['conf_dir']`\nhadoop['yarn_site'] | yarn-site.xml | `hadoop['conf_dir']`\nhbase['hadoop_metrics'] | hadoop-metrics.properties | `hbase['conf_dir']`\nhbase['hbase_env'] | hbase-env.sh | `hbase['conf_dir']`\nhbase['hbase_policy'] | hbase-policy.xml | `hbase['conf_dir']`\nhbase['hbase_site'] | hbase-site.xml | `hbase['conf_dir']`\nhbase['jaas'] | jaas.conf | `hbase['conf_dir']`\nhbase['log4j'] | log4j.properties | `hbase['conf_dir']`\nhive['hive_env'] | hive-env.sh | `hive['conf_dir']`\nhive['hive_site'] | hive-site.xml | `hive['conf_dir']`\nhive['jaas'] | jaas.conf | `hive['conf_dir']`\noozie['oozie_env'] | oozie-env.sh | `oozie['conf_dir']`\noozie['oozie_site'] | oozie-site.xml | `oozie['conf_dir']`\nspark['log4j'] | log4j.properties | `spark['conf_dir']`\nspark['metrics'] | metrics.properties | `spark['conf_dir']`\nspark['spark_env'] | spark-env.sh | `spark['conf_dir']`\nstorm['storm_env'] | storm-env.sh | `storm['conf_dir']`\nstorm['storm_env'] | storm_env.ini | `storm['conf_dir']`\nstorm['storm_conf'] | storm.yaml | `storm['conf_dir']`\ntez['tez_env'] | tez-env.sh | `tez['conf_dir']`\ntez['tez_site'] | tez-site.xml | `tez['conf_dir']`\nzookeeper['jaas'] | jaas.conf | `zookeeper['conf_dir']`\nzookeeper['log4j'] | log4j.properties | `zookeeper['conf_dir']`\nzookeeper['zoocfg'] | zoo.cfg | `zookeeper['conf_dir']`\n\n## Distribution Attributes\n\n* `hadoop['distribution']` - Specifies which Hadoop distribution to use, currently supported: cdh, hdp, bigtop. Default `hdp`\n* `hadoop['distribution_version']` - Specifies which version of `hadoop['distribution']` to use. Default `2.0` if `hadoop['distribution']` is `hdp`, `5` if `hadoop['distribution']` is `cdh`, and `0.8.0` if `hadoop['distribution']` is `bigtop`. It can also be set to `develop` when `hadoop['distribution']` is `bigtop` to allow installing from development repos without gpg validation.\n\n### APT-specific settings\n\n* `hadoop['apt_repo_url']` - Provide an alternate apt installation source location. If you change this attribute, you are expected to provide a path to a working repo for the `hadoop['distribution']` used. Default: `nil`\n* `hadoop['apt_repo_key_url']` - Provide an alternative apt repository key source location. Default `nil`\n\n### RPM-specific settings\n\n* `hadoop['yum_repo_url']` - Provide an alternate yum installation source location. If you change this attribute, you are expected to provide a path to a working repo for the `hadoop['distribution']` used. Default: `nil`\n* `hadoop['yum_repo_key_url']` - Provide an alternative yum repository key source location. Default `nil`\n\n## Global Configuration Attributes\n\n* `hadoop['conf_dir']` - The directory used inside `/etc/hadoop` and used via the alternatives system. Default `conf.chef`\n* `hbase['conf_dir']` - The directory used inside `/etc/hbase` and used via the alternatives system. Default `conf.chef`\n* `hive['conf_dir']` - The directory used inside `/etc/hive` and used via the alternatives system. Default `conf.chef`\n* `oozie['conf_dir']` - The directory used inside `/etc/oozie` and used via the alternatives system. Default `conf.chef`\n* `tez['conf_dir']` - The directory used inside `/etc/tez` and used via the alternatives system. Default `conf.chef`\n* `spark['conf_dir']` - The directory used inside `/etc/spark` and used via the alternatives system. Default `conf.chef`\n* `storm['conf_dir']` - The directory used inside `/etc/storm` and used via the alternatives system. Default `conf.chef`\n* `zookeeper['conf_dir']` - The directory used inside `/etc/zookeeper` and used via the alternatives system. Default `conf.chef`\n\n## Default Attributes\n\n* `hadoop['core_site']['fs.defaultFS']` - Sets URI to HDFS NameNode. Default `hdfs://localhost`\n* `hadoop['yarn_site']['yarn.resourcemanager.hostname']` - Sets hostname of YARN ResourceManager. Default `localhost`\n* `hive['hive_site']['javax.jdo.option.ConnectionURL']` - Sets JDBC URL. Default `jdbc:derby:;databaseName=/var/lib/hive/metastore/metastore_db;create=true`\n* `hive['hive_site']['javax.jdo.option.ConnectionDriverName']` - Sets JDBC Driver. Default `org.apache.derby.jdbc.EmbeddedDriver`\n\n# Recipes\n\n* `default.rb` - Sets up configuration and `hadoop-client` packages.\n* `hadoop_hdfs_checkconfig` - Ensures the HDFS configuration meets required parameters.\n* `hadoop_hdfs_datanode` - Sets up an HDFS DataNode.\n* `hadoop_hdfs_ha_checkconfig` - Ensures the HDFS configuration meets requirements for High Availability.\n* `hadoop_hdfs_journalnode` - Sets up an HDFS JournalNode.\n* `hadoop_hdfs_namenode` - Sets up an HDFS NameNode.\n* `hadoop_hdfs_secondarynamenode` - Sets up an HDFS Secondary NameNode.\n* `hadoop_hdfs_zkfc` - Sets up HDFS Failover Controller, required for automated NameNode failover.\n* `hadoop_yarn_nodemanager` - Sets up a YARN NodeManager.\n* `hadoop_yarn_proxyserver` - Sets up a YARN Web Proxy.\n* `hadoop_yarn_resourcemanager` - Sets up a YARN ResourceManager.\n* `hbase` - Sets up configuration and `hbase` packages.\n* `hbase_checkconfig` - Ensures the HBase configuration meets required parameters.\n* `hbase_master` - Sets up an HBase Master.\n* `hbase_regionserver` - Sets up an HBase RegionServer.\n* `hbase_rest` - Sets up an HBase REST interface.\n* `hbase_thrift` - Sets up an HBase Thrift interface.\n* `hive` - Sets up configuration and `hive` packages.\n* `hive_metastore` - Sets up Hive Metastore metadata repository.\n* `hive_server` - Sets up a Hive Thrift service.\n* `hive_server2` - Sets up a Hive Thrift service with Kerberos and multi-client concurrency support.\n* `oozie` - Sets up an Oozie server.\n* `oozie_client` - Sets up an Oozie client.\n* `pig` - Installs pig interpreter.\n* `repo` - Sets up package manager repositories for specified `hadoop['distribution']`\n* `spark` - Sets up configuration and `spark-core` packages.\n* `spark_master` - Sets up a Spark Master.\n* `spark_worker` - Sets up a Spark Worker.\n* `storm` - Sets up `storm` package.\n* `storm_nimbus` - Setups a Storm Nimbus server.\n* `storm_supervisor` - Setups a Storm Supervisor server.\n* `storm_ui` - Setups a Storm UI server.\n* `tez` - Sets up configuration and `tez` packages.\n* `zookeeper` - Sets up `zookeeper` package.\n* `zookeeper_server` - Sets up a ZooKeeper server.\n\n# Author\n\nAuthor:: Cask Data, Inc. ()\n\n# Testing\n\nThis cookbook has several ways to test it. It includes code tests, which are done using `foodcritic`, `rubocop`, and `chefspec`.\nIt, also, includes functionality testing, provided by `kitchen`.\n\n```text\nrake chefspec # Run RSpec code examples\nrake foodcritic # Foodcritic linter\nrake integration # Run Test Kitchen integration tests\nrake metadata # Create metadata.json from metadata.rb\nrake rubocop # Ruby style guide linter\nrake share # Share cookbook to community site\n```\n\n# License\n\nLicensed under the Apache License, Version 2.0 (the \"License\"); you may not use this software except in compliance with the License. You may obtain a copy of the License at\n\nhttp://www.apache.org/licenses/LICENSE-2.0\n\nUnless required by applicable law or agreed to in writing, software distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License.\n", "maintainer": "Cask Data, Inc.", "maintainer_email": "ops@cask.co", - "license": "Apache 2.0", + "license": "Apache-2.0", "platforms": { "amazon": ">= 0.0.0", "centos": ">= 0.0.0", @@ -22,35 +22,13 @@ "ulimit": ">= 0.0.0" }, "recommendations": { - "java": "~> 1.40" + }, "suggestions": { }, "conflicting": { - "apache_hadoop": ">= 0.0.0", - "apache_spark": ">= 0.0.0", - "apache_zookeeper": ">= 0.0.0", - "hadoop_cluster": ">= 0.0.0", - "hadoop_cluster_rpm": ">= 0.0.0", - "hadoop_for_hbase": ">= 0.0.0", - "hadoop_spark": ">= 0.0.0", - "hbase": ">= 0.0.0", - "hbase_cluster": ">= 0.0.0", - "hive": ">= 0.0.0", - "pig": ">= 0.0.0", - "spark": ">= 0.0.0", - "spark-jobserver": ">= 0.0.0", - "spark-platform": ">= 0.0.0", - "storm": ">= 0.0.0", - "storm-cluster": ">= 0.0.0", - "storm-platform": ">= 0.0.0", - "tez": ">= 0.0.0", - "zookeeper": ">= 0.0.0", - "zookeeperd": ">= 0.0.0", - "zookeeper_cluster": ">= 0.0.0", - "zookeeper-cluster": ">= 0.0.0", - "zookeeper-platform": ">= 0.0.0" + }, "providing": { @@ -67,7 +45,7 @@ "recipes": { }, - "version": "2.9.0", + "version": "2.10.0", "source_url": "https://github.com/caskdata/hadoop_cookbook", "issues_url": "https://issues.cask.co/browse/COOK/component/10600" } diff --git a/metadata.rb b/metadata.rb index 83a55b8..3583b80 100644 --- a/metadata.rb +++ b/metadata.rb @@ -1,10 +1,10 @@ name 'hadoop' maintainer 'Cask Data, Inc.' maintainer_email 'ops@cask.co' -license 'Apache 2.0' +license 'Apache-2.0' description 'Installs/Configures Hadoop (HDFS/YARN/MRv2), HBase, Hive, Flume, Oozie, Pig, Spark, Storm, Tez, and ZooKeeper' long_description IO.read(File.join(File.dirname(__FILE__), 'README.md')) -version '2.9.0' +version '2.10.0' depends 'yum', '>= 3.0' depends 'apt', '>= 2.1.2' @@ -13,8 +13,6 @@ depends cb end -recommends 'java', '~> 1.40' # ~FC053 - %w( amazon centos @@ -26,33 +24,5 @@ supports os end -%w( - apache_hadoop - apache_spark - apache_zookeeper - hadoop_cluster - hadoop_cluster_rpm - hadoop_for_hbase - hadoop_spark - hbase - hbase_cluster - hive - pig - spark - spark-jobserver - spark-platform - storm - storm-cluster - storm-platform - tez - zookeeper - zookeeperd - zookeeper_cluster - zookeeper-cluster - zookeeper-platform -).each do |cb| - conflicts cb -end - source_url 'https://github.com/caskdata/hadoop_cookbook' if respond_to?(:source_url) issues_url 'https://issues.cask.co/browse/COOK/component/10600' if respond_to?(:issues_url)