HADOOP-14876. Create downstream developer docs from the compatibility guidelines
(cherry picked from commit bfd588789a879b8583ea4abd59f4f5843c5ac285)
This commit is contained in:
parent
1c8d334287
commit
4a70e75386
|
@ -63,6 +63,14 @@ when the various labels are appropriate. As a general rule, all new interfaces
|
||||||
and APIs should have the most limited labels (e.g. Private Unstable) that will
|
and APIs should have the most limited labels (e.g. Private Unstable) that will
|
||||||
not inhibit the intent of the interface or API.
|
not inhibit the intent of the interface or API.
|
||||||
|
|
||||||
|
### Structure
|
||||||
|
|
||||||
|
This document is arranged in sections according to the various compatibility
|
||||||
|
concerns. Within each section an introductory text explains what compatibility
|
||||||
|
means in that section, why it's important, and what the intent to support
|
||||||
|
compatibility is. The subsequent "Policy" section then sets forth in specific
|
||||||
|
terms what the governing policy is.
|
||||||
|
|
||||||
### Notational Conventions
|
### Notational Conventions
|
||||||
|
|
||||||
The key words "MUST" "MUST NOT", "REQUIRED", "SHALL", "SHALL NOT", "SHOULD",
|
The key words "MUST" "MUST NOT", "REQUIRED", "SHALL", "SHALL NOT", "SHOULD",
|
||||||
|
@ -77,12 +85,18 @@ flagged for removal. The standard meaning of the annotation is that the
|
||||||
API element should not be used and may be removed in a later version.
|
API element should not be used and may be removed in a later version.
|
||||||
|
|
||||||
In all cases removing an element from an API is an incompatible
|
In all cases removing an element from an API is an incompatible
|
||||||
change. In the case of [Stable](./InterfaceClassification.html#Stable) APIs,
|
change. The stability of the element SHALL determine when such a change is
|
||||||
the change cannot be made between minor releases within the same major
|
permissible. A [Stable](./InterfaceClassification.html#Stable) element MUST
|
||||||
version. In addition, to allow consumers of the API time to adapt to the change,
|
be marked as deprecated for a full major release before it can be removed and
|
||||||
the API element to be removed should be marked as deprecated for a full major
|
SHALL NOT be removed in a minor or maintenance release. An
|
||||||
release before it is removed. For example, if a method is marked as deprecated
|
[Evolving](./InterfaceClassification.html#Evolving) element MUST be marked as
|
||||||
in Hadoop 2.8, it cannot be removed until Hadoop 4.0.
|
deprecated for a full minor release before it can be removed and SHALL NOT be
|
||||||
|
removed during a maintenance release. An
|
||||||
|
[Unstable](./InterfaceClassification.html#Unstable) element MAY be removed at
|
||||||
|
any time. When possible an [Unstable](./InterfaceClassification.html#Unstable)
|
||||||
|
element SHOULD be marked as deprecated for at least one release before being
|
||||||
|
removed. For example, if a method is marked as deprecated in Hadoop 2.8, it
|
||||||
|
cannot be removed until Hadoop 4.0.
|
||||||
|
|
||||||
### Policy
|
### Policy
|
||||||
|
|
||||||
|
@ -141,7 +155,7 @@ in hand.
|
||||||
#### Semantic compatibility
|
#### Semantic compatibility
|
||||||
|
|
||||||
Apache Hadoop strives to ensure that the behavior of APIs remains consistent
|
Apache Hadoop strives to ensure that the behavior of APIs remains consistent
|
||||||
over versions, though changes for correctness may result in changes in
|
across releases, though changes for correctness may result in changes in
|
||||||
behavior. API behavior SHALL be specified by the JavaDoc API documentation
|
behavior. API behavior SHALL be specified by the JavaDoc API documentation
|
||||||
where present and complete. When JavaDoc API documentation is not available,
|
where present and complete. When JavaDoc API documentation is not available,
|
||||||
behavior SHALL be specified by the behavior expected by the related unit tests.
|
behavior SHALL be specified by the behavior expected by the related unit tests.
|
||||||
|
@ -229,8 +243,8 @@ transports, such as SSL. Upgrading a service from SSLv2 to SSLv3 may break
|
||||||
existing SSLv2 clients. The minimum supported major version of any transports
|
existing SSLv2 clients. The minimum supported major version of any transports
|
||||||
MUST not increase across minor releases within a major version.
|
MUST not increase across minor releases within a major version.
|
||||||
|
|
||||||
Service ports are considered as part of the transport mechanism. Fixed
|
Service ports are considered as part of the transport mechanism. Default
|
||||||
service port numbers MUST be kept consistent to prevent breaking clients.
|
service port numbers must be kept consistent to prevent breaking clients.
|
||||||
|
|
||||||
#### Policy
|
#### Policy
|
||||||
|
|
||||||
|
@ -281,9 +295,8 @@ according to the following:
|
||||||
* Client-Server compatibility MUST be maintained so as to allow upgrading individual components without upgrading others. For example, upgrade HDFS from version 2.1.0 to 2.2.0 without upgrading MapReduce.
|
* Client-Server compatibility MUST be maintained so as to allow upgrading individual components without upgrading others. For example, upgrade HDFS from version 2.1.0 to 2.2.0 without upgrading MapReduce.
|
||||||
* Server-Server compatibility MUST be maintained so as to allow mixed versions within an active cluster so the cluster may be upgraded without downtime in a rolling fashion.
|
* Server-Server compatibility MUST be maintained so as to allow mixed versions within an active cluster so the cluster may be upgraded without downtime in a rolling fashion.
|
||||||
|
|
||||||
New transport mechanisms MUST only be introduced with minor or major version
|
Existing transport mechanisms MUST continue to be supported across
|
||||||
changes. Existing transport mechanisms MUST continue to be supported across
|
minor versions within a major version. Default service port numbers MUST remain
|
||||||
minor versions within a major version. Service port numbers MUST remain
|
|
||||||
consistent across minor version numbers within a major version.
|
consistent across minor version numbers within a major version.
|
||||||
|
|
||||||
### REST APIs
|
### REST APIs
|
||||||
|
@ -312,7 +325,8 @@ The Hadoop REST APIs SHALL be considered
|
||||||
numbers, the Hadoop REST APIs SHALL be considered
|
numbers, the Hadoop REST APIs SHALL be considered
|
||||||
[Public](./InterfaceClassification.html#Public) and
|
[Public](./InterfaceClassification.html#Public) and
|
||||||
[Stable](./InterfaceClassification.html#Stable), i.e. no incompatible changes
|
[Stable](./InterfaceClassification.html#Stable), i.e. no incompatible changes
|
||||||
are allowed to within an API version number.
|
are allowed to within an API version number. A REST API version must be labeled
|
||||||
|
as deprecated for a full major release before it can be removed.
|
||||||
|
|
||||||
### Log Output
|
### Log Output
|
||||||
|
|
||||||
|
@ -325,7 +339,9 @@ use cases are also supported.
|
||||||
|
|
||||||
All log output SHALL be considered
|
All log output SHALL be considered
|
||||||
[Public](./InterfaceClassification.html#Public) and
|
[Public](./InterfaceClassification.html#Public) and
|
||||||
[Evolving](./InterfaceClassification.html#Evolving).
|
[Unstable](./InterfaceClassification.html#Unstable). For log output, an
|
||||||
|
incompatible change is one that renders a parser unable to find or recognize
|
||||||
|
a line of log output.
|
||||||
|
|
||||||
### Audit Log Output
|
### Audit Log Output
|
||||||
|
|
||||||
|
@ -400,37 +416,42 @@ server's jhist file format, SHALL be considered
|
||||||
|
|
||||||
HDFS persists metadata (the image and edit logs) in a private file format.
|
HDFS persists metadata (the image and edit logs) in a private file format.
|
||||||
Incompatible changes to either the format or the metadata prevent subsequent
|
Incompatible changes to either the format or the metadata prevent subsequent
|
||||||
releases from reading older metadata. Incompatible changes MUST include a
|
releases from reading older metadata. Incompatible changes must include a
|
||||||
process by which existing metadata may be upgraded. Changes SHALL be
|
process by which existing metadata may be upgraded.
|
||||||
allowed to require more than one upgrade. Incompatible changes MUST result in
|
|
||||||
the metadata version number being incremented.
|
|
||||||
|
|
||||||
Depending on the degree of incompatibility in the changes, the following
|
Depending on the degree of incompatibility in the changes, the following
|
||||||
potential scenarios can arise:
|
potential scenarios can arise:
|
||||||
|
|
||||||
* Automatic: The image upgrades automatically, no need for an explicit "upgrade".
|
* Automatic: The image upgrades automatically, no need for an explicit "upgrade".
|
||||||
* Direct: The image is upgradable, but might require one explicit release "upgrade".
|
* Direct: The image is upgradeable, but might require one explicit release "upgrade".
|
||||||
* Indirect: The image is upgradable, but might require upgrading to intermediate release(s) first.
|
* Indirect: The image is upgradeable, but might require upgrading to intermediate release(s) first.
|
||||||
* Not upgradeable: The image is not upgradeable.
|
* Not upgradeable: The image is not upgradeable.
|
||||||
|
|
||||||
HDFS data nodes store data in a private directory structure. The schema of that
|
HDFS data nodes store data in a private directory structure. Incompatible
|
||||||
directory structure must remain stable to retain compatibility.
|
changes to the directory structure may prevent older releases from accessing
|
||||||
|
stored data. Incompatible changes must include a process by which existing
|
||||||
|
data directories may be upgraded.
|
||||||
|
|
||||||
###### Policy
|
###### Policy
|
||||||
|
|
||||||
The HDFS metadata format SHALL be considered
|
The HDFS metadata format SHALL be considered
|
||||||
[Private](./InterfaceClassification.html#Private) and
|
[Private](./InterfaceClassification.html#Private) and
|
||||||
[Evolving](./InterfaceClassification.html#Evolving). Incompatible
|
[Evolving](./InterfaceClassification.html#Evolving). Incompatible
|
||||||
changes MUST include a process by which existing metada may be upgraded. The
|
changes MUST include a process by which existing metadata may be upgraded.
|
||||||
upgrade process MUST allow the cluster metadata to be rolled back to the older
|
The upgrade process SHALL be allowed to require more than one upgrade.
|
||||||
version and its older disk format. The rollback MUST restore the original data
|
The upgrade process MUST allow the cluster metadata to be
|
||||||
but is not REQUIRED to restore the updated data. Any incompatible change
|
rolled back to the older version and its older disk format. The rollback
|
||||||
to the format MUST result in the major version number of the schema being
|
MUST restore the original data but is not REQUIRED to restore the updated
|
||||||
incremented.
|
data. Any incompatible change to the format MUST result in the major version
|
||||||
|
number of the schema being incremented.
|
||||||
|
|
||||||
The data node directory format SHALL be considered
|
The data node directory format SHALL be considered
|
||||||
[Private](./InterfaceClassification.html#Private) and
|
[Private](./InterfaceClassification.html#Private) and
|
||||||
[Evolving](./InterfaceClassification.html#Evolving).
|
[Evolving](./InterfaceClassification.html#Evolving). Incompatible
|
||||||
|
changes MUST include a process by which existing data directories may be
|
||||||
|
upgraded. The upgrade process SHALL be allowed to require more than one upgrade.
|
||||||
|
The upgrade process MUST allow the data directories to be
|
||||||
|
rolled back to the older layout.
|
||||||
|
|
||||||
##### AWS S3A Guard Metadata
|
##### AWS S3A Guard Metadata
|
||||||
|
|
||||||
|
@ -533,15 +554,42 @@ The Hadoop Web UI SHALL be considered
|
||||||
[Public](./InterfaceClassification.html#Public) and
|
[Public](./InterfaceClassification.html#Public) and
|
||||||
[Unstable](./InterfaceClassification.html#Unstable).
|
[Unstable](./InterfaceClassification.html#Unstable).
|
||||||
|
|
||||||
|
### Functional Compatibility
|
||||||
|
|
||||||
|
Users depend on the behavior of a Hadoop cluster remaining consistent across
|
||||||
|
releases. Changes which cause unexpectedly different behaviors from the cluster
|
||||||
|
can lead to frustration and long adoption cycles. No new configuration should
|
||||||
|
be added which changes the behavior of an existing
|
||||||
|
cluster, assuming the cluster's configuration files remain unchanged. For any
|
||||||
|
new settings that are defined, care should be taken to ensure that the new
|
||||||
|
setting does not change the behavior of existing clusters.
|
||||||
|
|
||||||
|
#### Policy
|
||||||
|
|
||||||
|
Changes to existing functionality MUST NOT change the default behavior or the
|
||||||
|
meaning of existing configuration settings between maintenance releases within
|
||||||
|
the same minor version, regardless of whether the changes arise from changes
|
||||||
|
to the system or logic or to internal or external default configuration values.
|
||||||
|
|
||||||
|
Changes to existing functionality SHOULD NOT change the default behavior or the
|
||||||
|
meaning of existing configuration settings between minor releases within
|
||||||
|
the same major version, though changes, such as to fix correctness or
|
||||||
|
security issues, may require incompatible behavioral changes. Where possible
|
||||||
|
such behavioral changes SHOULD be off by default.
|
||||||
|
|
||||||
### Hadoop Configuration Files
|
### Hadoop Configuration Files
|
||||||
|
|
||||||
Users use Hadoop-defined properties to configure and provide hints to Hadoop and
|
Users use Hadoop-defined properties to configure and provide hints to Hadoop and
|
||||||
custom properties to pass information to jobs. Users are encouraged to avoid
|
custom properties to pass information to jobs. Users are encouraged to avoid
|
||||||
using custom configuration property names that conflict with the namespace of
|
using custom configuration property names that conflict with the namespace of
|
||||||
Hadoop-defined properties and should avoid using any prefixes used by Hadoop,
|
Hadoop-defined properties and should avoid using any prefixes used by Hadoop,
|
||||||
e.g. hadoop, io, ipc, fs, net, file, ftp, s3, kfs, ha, file, dfs, mapred,
|
e.g. hadoop, io, ipc, fs, net, file, ftp, kfs, ha, file, dfs, mapred,
|
||||||
mapreduce, and yarn.
|
mapreduce, and yarn.
|
||||||
|
|
||||||
|
In addition to properties files, Hadoop uses other configuration files to
|
||||||
|
set system behavior, such as the fair scheduler configuration file or the
|
||||||
|
resource profiles configuration file.
|
||||||
|
|
||||||
#### Policy
|
#### Policy
|
||||||
|
|
||||||
Hadoop-defined properties (names and meanings) SHALL be considered
|
Hadoop-defined properties (names and meanings) SHALL be considered
|
||||||
|
@ -552,6 +600,14 @@ across major versions. Default values of Hadoop-defined properties SHALL be
|
||||||
considered [Public](./InterfaceClassification.html#Public) and
|
considered [Public](./InterfaceClassification.html#Public) and
|
||||||
[Evolving](./InterfaceClassification.html#Evolving).
|
[Evolving](./InterfaceClassification.html#Evolving).
|
||||||
|
|
||||||
|
Hadoop configuration files that are not governed by the above rules about
|
||||||
|
Hadoop-defined properties SHALL be considered
|
||||||
|
[Public](./InterfaceClassification.html#Public) and
|
||||||
|
[Stable](./InterfaceClassification.html#Stable). The definition of an
|
||||||
|
incompatible change depends on the particular configuration file format, but
|
||||||
|
the general rule is that a compatible change will allow a configuration
|
||||||
|
file that was valid before the change to remain valid after the change.
|
||||||
|
|
||||||
### Log4j Configuration Files
|
### Log4j Configuration Files
|
||||||
|
|
||||||
The log output produced by Hadoop daemons and CLIs is governed by a set of
|
The log output produced by Hadoop daemons and CLIs is governed by a set of
|
||||||
|
@ -568,11 +624,11 @@ All Log4j configurations SHALL be considered
|
||||||
### Directory Structure
|
### Directory Structure
|
||||||
|
|
||||||
Source code, artifacts (source and tests), user logs, configuration files,
|
Source code, artifacts (source and tests), user logs, configuration files,
|
||||||
output, and job history are all stored on disk either local file system or HDFS.
|
output, and job history are all stored on disk on either the local file system
|
||||||
Changing the directory structure of these user-accessible files can break
|
or HDFS. Changing the directory structure of these user-accessible files can
|
||||||
compatibility, even in cases where the original path is preserved via symbolic
|
break compatibility, even in cases where the original path is preserved via
|
||||||
links (such as when the path is accessed by a servlet that is configured to
|
symbolic links (such as when the path is accessed by a servlet that is
|
||||||
not follow symbolic links).
|
configured to not follow symbolic links).
|
||||||
|
|
||||||
#### Policy
|
#### Policy
|
||||||
|
|
||||||
|
@ -680,11 +736,11 @@ upgrading other dependent software components.
|
||||||
#### Policies
|
#### Policies
|
||||||
|
|
||||||
* Hardware
|
* Hardware
|
||||||
* Architecture: The community has no plans to restrict Hadoop to specific architectures, but can have family-specific optimizations.
|
* Architecture: Intel and AMD are the processor architectures currently supported by the community. The community has no plans to restrict Hadoop to specific architectures, but MAY have family-specific optimizations. Support for any processor architecture SHOULD NOT be dropped without first being documented as deprecated for a full major release and MUST NOT be dropped without first being deprecated for at least a full minor release.
|
||||||
* Minimum resources: While there are no guarantees on the minimum resources required by Hadoop daemons, the developer community SHOULD avoid increasing requirements within a minor release.
|
* Minimum resources: While there are no guarantees on the minimum resources required by Hadoop daemons, the developer community SHOULD avoid increasing requirements within a minor release.
|
||||||
* Operating Systems: The community SHOULD maintain the same minimum OS requirements (OS kernel versions) within a minor release. Currently GNU/Linux and Microsoft Windows are the OSes officially supported by the community, while Apache Hadoop is known to work reasonably well on other OSes such as Apple MacOSX, Solaris, etc.
|
* Operating Systems: The community SHOULD maintain the same minimum OS requirements (OS kernel versions) within a minor release. Currently GNU/Linux and Microsoft Windows are the OSes officially supported by the community, while Apache Hadoop is known to work reasonably well on other OSes such as Apple MacOSX, Solaris, etc. Support for any OS SHOULD NOT be dropped without first being documented as deprecated for a full major release and MUST NOT be dropped without first being deprecated for at least a full minor release.
|
||||||
* The JVM requirements SHALL NOT change across minor releases within the same major release unless the JVM version in question becomes unsupported. The JVM version requirement MAY be different for different operating systems or even operating system releases.
|
* The JVM requirements SHALL NOT change across minor releases within the same major release unless the JVM version in question becomes unsupported. The JVM version requirement MAY be different for different operating systems or even operating system releases.
|
||||||
* File systems supported by Hadoop, e.g. through the HDFS FileSystem API, SHOULD not become unsupported between minor releases within a major version unless a migration path to an alternate client implementation is available.
|
* File systems supported by Hadoop, e.g. through the FileSystem API, SHOULD not become unsupported between minor releases within a major version unless a migration path to an alternate client implementation is available.
|
||||||
|
|
||||||
References
|
References
|
||||||
----------
|
----------
|
||||||
|
|
|
@ -0,0 +1,432 @@
|
||||||
|
<!---
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License. See accompanying LICENSE file.
|
||||||
|
-->
|
||||||
|
|
||||||
|
Apache Hadoop Downstream Developer's Guide
|
||||||
|
==========================================
|
||||||
|
|
||||||
|
<!-- MACRO{toc|fromDepth=0|toDepth=3} -->
|
||||||
|
|
||||||
|
Purpose
|
||||||
|
-------
|
||||||
|
|
||||||
|
The point of this document is to provide downstream developers with a clear
|
||||||
|
reference for what to expect when building applications against the Hadoop
|
||||||
|
source base. This document is primarily a distillation of the
|
||||||
|
[Hadoop Compatibility Guidelines](./Compatibility.html) and hence focuses on
|
||||||
|
what the compatibility guarantees are for the various Hadoop interfaces
|
||||||
|
across releases.
|
||||||
|
|
||||||
|
### Target Audience
|
||||||
|
|
||||||
|
The target audience for this document is any developer working on a project
|
||||||
|
or application that builds or depends on Apache Hadoop, whether the
|
||||||
|
dependency is on the source code itself, a build artifact, or interacting
|
||||||
|
with a running system.
|
||||||
|
|
||||||
|
Hadoop Releases
|
||||||
|
---------------
|
||||||
|
|
||||||
|
The Hadoop development community periodically produces new Hadoop releases to
|
||||||
|
introduce new functionality and fix existing issues. Realeses fall into three
|
||||||
|
categories:
|
||||||
|
|
||||||
|
* Major: a major release will typically include significant new functionality and generally represents the largest upgrade compatibility risk. A major release increments the first number of the release version, e.g. going from 2.8.2 to 3.0.0.
|
||||||
|
* Minor: a minor release will typically include some new functionality as well as fixes for some notable issues. A minor release should not pose much upgrade risk in most cases. A minor release increments the middle number of release version, e.g. going from 2.8.2 to 2.9.0.
|
||||||
|
* Maintenance: a maintenance release should not include any new functionality. The purpose of a maintenance release is to resolve a set of issues that are deemed by the developer community to be significant enough to be worth pushing a new release to address them. Maintenance releases should pose very little upgrade risk. A maintenance release increments the final number in the release version, e.g. going from 2.8.2 to 2.8.3.
|
||||||
|
|
||||||
|
Consuming Hadoop APIs
|
||||||
|
---------------------
|
||||||
|
|
||||||
|
When writing software that calls methods or uses classes that belong to Apache
|
||||||
|
Hadoop, developers should adhere to the following guidelines. Failure to adhere
|
||||||
|
to the guidelines may result in problems transitioning from one Hadoop release
|
||||||
|
to another.
|
||||||
|
|
||||||
|
### Privacy
|
||||||
|
|
||||||
|
Packages, classes, and methods may be annotated with an audience annotation.
|
||||||
|
The three privacy levels are:
|
||||||
|
[Public](./InterfaceClassification.html#Public),
|
||||||
|
[Limited-Private](./InterfaceClassification.html#Limited-Private), and
|
||||||
|
[Private](./InterfaceClassification.html#Private). Downstream developers should
|
||||||
|
only use packages, classes, methods, and fields that are marked as
|
||||||
|
[Public](./InterfaceClassification.html#Public). Packages, classes, and methods
|
||||||
|
that are not marked as [Public](./InterfaceClassification.html#Public) are
|
||||||
|
considered internal to Hadoop and are intended only for consumption by other
|
||||||
|
components of Hadoop.
|
||||||
|
|
||||||
|
If an element has an annotation that conflicts with it's containing element's
|
||||||
|
annotation, then the most restrictive annotation takes precedence. For example,
|
||||||
|
If a [Private](./InterfaceClassification.html#Private) method is contained
|
||||||
|
in a [Public](./InterfaceClassification.html#Public) class, then the method
|
||||||
|
should be treated as [Private](./InterfaceClassification.html#Private). If a
|
||||||
|
[Public](./InterfaceClassification.html#Public) method is contained in a
|
||||||
|
[Private](./InterfaceClassification.html#Private) class, the method should be
|
||||||
|
treated as [Private](./InterfaceClassification.html#Private).
|
||||||
|
|
||||||
|
If a method has no privacy annotation, then it inherits its privacy from its
|
||||||
|
class. If a class has no privacy, it inherits its privacy from its package.
|
||||||
|
If a package has no privacy, it should be assumed to be
|
||||||
|
[Private](./InterfaceClassification.html#Private).
|
||||||
|
|
||||||
|
### Stability
|
||||||
|
|
||||||
|
Packages, classes, and methods may be annotated with a stability annotation.
|
||||||
|
There are three classes of stability:
|
||||||
|
[Stable](./InterfaceClassification.html#Stable),
|
||||||
|
[Evolving](./InterfaceClassification.html#Evolving), and
|
||||||
|
[Unstable](./InterfaceClassification.html#Unstable). The stability annotations
|
||||||
|
determine when
|
||||||
|
[incompatible changes](./InterfaceClassification.html#Change-Compatibility)
|
||||||
|
are allowed to be made. [Stable](./InterfaceClassification.html#Stable) means
|
||||||
|
that no incompatible changes are allowed between major releases.
|
||||||
|
[Evolving](./InterfaceClassification.html#Evolving) means no incompatible
|
||||||
|
changes are allowed between minor releases.
|
||||||
|
[Unstable](./InterfaceClassification.html#Unstable) means that incompatible
|
||||||
|
changes are allowed at any time. As a downstream developer, it is best to
|
||||||
|
avoid [Unstable](./InterfaceClassification.html#Unstable) APIs and where
|
||||||
|
possible to prefer [Stable](./InterfaceClassification.html#Stable) ones.
|
||||||
|
|
||||||
|
If a method has no stability annotation, then it inherits its stability from its
|
||||||
|
class. If a class has no stability, it inherits its stability from its package.
|
||||||
|
If a package has no stability, it should be assumed to be
|
||||||
|
[Unstable](./InterfaceClassification.html#Unstable).
|
||||||
|
|
||||||
|
|
||||||
|
#### Releases and Stability
|
||||||
|
|
||||||
|
Per the above rules on API stability, new releases are allowed to change APIs
|
||||||
|
as follows:
|
||||||
|
|
||||||
|
| Release Type | Stable API Changes | Evolving API Changes | Unstable API Changes |
|
||||||
|
|:---- |:---- |:---- |:---- |
|
||||||
|
| Major | Allowed | Allowed | Allowed |
|
||||||
|
| Minor | Not Allowed | Allowed | Allowed |
|
||||||
|
| Maintenance | Not Allowed | Not Allowed | Allowed |
|
||||||
|
|
||||||
|
Note that a major release is *allowed* to break compatibility of any API, even
|
||||||
|
though the Hadoop developer community strives to maintain compatibility as much
|
||||||
|
as possible, even across major releases. Note also that an
|
||||||
|
[Unstable](./InterfaceClassification.html#Unstable) API may change at any time
|
||||||
|
without notice.
|
||||||
|
|
||||||
|
### Deprecation
|
||||||
|
|
||||||
|
Classes or methods that are annotated as @Deprecated are no longer safe to use.
|
||||||
|
The deprecated element should continue to function, but may and likely will be
|
||||||
|
removed in a subsequent release. The stability annotation
|
||||||
|
will determine the earliest release when the deprecated element can be removed.
|
||||||
|
A [Stable](./InterfaceClassification.html#Stable) element cannot be removed
|
||||||
|
until the next major release. An
|
||||||
|
[Evolving](./InterfaceClassification.html#Evolving) element cannot be removed
|
||||||
|
until the next minor release. An
|
||||||
|
[Unstable](./InterfaceClassification.html#Unstable) element may be removed at
|
||||||
|
any time and will typically not be marked as deprecated before it is removed.
|
||||||
|
[Stable](./InterfaceClassification.html#Stable) and
|
||||||
|
[Evolving](./InterfaceClassification.html#Evolving) elements must be marked as
|
||||||
|
deprecated for a full major or minor release (respectively) before they can be
|
||||||
|
removed. For example, if a [Stable](./InterfaceClassification.html#Stable)
|
||||||
|
is marked as deprecated in Hadoop 3.1, it cannot be removed until Hadoop 5.0.
|
||||||
|
|
||||||
|
### Semantic Compatibility
|
||||||
|
|
||||||
|
The Apache Hadoop developer community strives to ensure that the behavior of
|
||||||
|
APIs remains consistent across releases, though changes for correctness may
|
||||||
|
result in changes in behavior. The API JavaDocs are considered the primary
|
||||||
|
authority for the expected behavior of an API. In cases where the
|
||||||
|
JavaDocs are insufficient or missing, the unit tests are considered the
|
||||||
|
fallback authority for expected behavior. Where unit tests are not present,
|
||||||
|
the intended behavior should be inferred from the naming. As much as possible
|
||||||
|
downstream developers should avoid looking at the source code for the API
|
||||||
|
itself to determine expected behavior as that approach can create dependencies
|
||||||
|
on implementation details that are not expressly held as expected behavior by
|
||||||
|
the Hadoop development community.
|
||||||
|
|
||||||
|
In cases where the JavaDocs are insufficient to infer expected behavior,
|
||||||
|
downstream developers are strongly encouraged to file a Hadoop JIRA to request
|
||||||
|
the JavaDocs be added or improved.
|
||||||
|
|
||||||
|
Be aware that fixes done for correctness reasons may cause changes to the
|
||||||
|
expected behavior of an API, though such changes are expected to be
|
||||||
|
accompanied by documentation that clarifies the new behavior.
|
||||||
|
|
||||||
|
The Apache Hadoop developer community tries to maintain binary compatibility
|
||||||
|
for end user applications across releases. Ideally no updates to applications
|
||||||
|
will be required when upgrading to a new Hadoop release, assuming the
|
||||||
|
application does not use [Private](./InterfaceClassification.html#Private),
|
||||||
|
[Limited-Private](./InterfaceClassification.html#Limited-Private), or
|
||||||
|
[Unstable](./InterfaceClassification.html#Unstable) APIs. MapReduce applications
|
||||||
|
in particular are guaranteed binary compatibility across releases.
|
||||||
|
|
||||||
|
### Compatibility Issues
|
||||||
|
|
||||||
|
The [Hadoop Compatibility Specification](./Compatibility.html) states the
|
||||||
|
standards that the Hadoop development community is expected to uphold, but for
|
||||||
|
various reasons, the source code may not live up to the ideals of the
|
||||||
|
[compatibility specification](./Compatibility.html).
|
||||||
|
|
||||||
|
Two common issues that a downstream developer will encounter are:
|
||||||
|
|
||||||
|
1. APIs that are needed for application development aren't
|
||||||
|
[Public](./InterfaceClassification.html#Public)
|
||||||
|
2. A [Public](./InterfaceClassification.html#Public) API on which a downstream
|
||||||
|
application depends is changed unexpectedly and incompatibly.
|
||||||
|
|
||||||
|
In both of these cases, downstream developers are strongly encouraged to raise
|
||||||
|
the issues with the Hadoop developer community either by sending an email to
|
||||||
|
the appropriate
|
||||||
|
[developer mailing list](https://hadoop.apache.org/mailing_lists.html) or
|
||||||
|
[filing a JIRA](https://hadoop.apache.org/issue_tracking.html) or both. The
|
||||||
|
developer community appreciates the feedback.
|
||||||
|
|
||||||
|
Downstream developers are encouraged to reach out to the Hadoop development
|
||||||
|
community in any case when they encounter an issue while developing an
|
||||||
|
application against Hadoop. Odds are good that if it's an issue for one
|
||||||
|
developer, it's an issue that numerous developers have or will encounter.
|
||||||
|
|
||||||
|
Using the FileSystem API
|
||||||
|
------------------------
|
||||||
|
|
||||||
|
In the specific case of working with streams in Hadoop, e.g.
|
||||||
|
`FSDataOutputStream`, an application can programmatically query for the
|
||||||
|
capabilities of the stream using the methods of the
|
||||||
|
[StreamCapabilities](http://hadoop.apache.org/docs/current/api/org/apache/hadoop/fs/StreamCapabilities.html)
|
||||||
|
class. Dynamically adjusting to stream capabilities can make an applcation
|
||||||
|
more robust in the face of changing implementations and environments.
|
||||||
|
|
||||||
|
Consuming Hadoop REST APIs
|
||||||
|
--------------------------
|
||||||
|
|
||||||
|
The Hadoop REST APIs are a primary interface for a variety of downstream
|
||||||
|
and internal applications and services. To support REST clients, the Hadoop
|
||||||
|
REST APIs are versioned and will not change incompatibly within a version.
|
||||||
|
Both the endpoint itself along with the list of supported parameters and the
|
||||||
|
output from the endpoint are prohibited from changing incompatibly within a
|
||||||
|
REST endpoint version. Note, however, that introducing new fields and other
|
||||||
|
additive changes are considered compatible changes, so any consumer of the
|
||||||
|
REST API should be flexible enough to ignore unknown fields.
|
||||||
|
|
||||||
|
The REST API version is a single number and has no relationship with the Hadoop
|
||||||
|
version number. The version number is encoded in the endpoint URL prefixed
|
||||||
|
with a 'v', for example 'v1'. A new REST endpoint version may only be
|
||||||
|
introduced with a minor or major release. A REST endpoint version may only be
|
||||||
|
removed after being labeled as deprecated for a full major release.
|
||||||
|
|
||||||
|
Consuming Hadoop Output
|
||||||
|
-----------------------
|
||||||
|
|
||||||
|
Hadoop produces a variety of outputs that could conceivably be consumed by
|
||||||
|
application clients or downstream libraries. When consuming output from Hadoop,
|
||||||
|
please consider the following:
|
||||||
|
|
||||||
|
* Hadoop log output is not expected to change with a maintenance release unless it resolves a correctness issue. While log output can be consumed by software directly, it is intended primarily for a human reader.
|
||||||
|
* Hadoop produces audit logs for a variety of operations. The audit logs are intended to be machine readable, though the addition of new records and fields are considered to be compatible changes. Any consumer of the audit logs should allow for unexpected records and fields. The audit log format must not change incompatibly between major releases.
|
||||||
|
* Metrics data produced by Hadoop is mostly intended for automated consumption. The metrics format may not change in an incompatible way between major releases, but new records and fields can be compatibly added at any time. Consumers of the metrics data should allow for unknown records and fields.
|
||||||
|
|
||||||
|
Consuming Hadoop Data
|
||||||
|
---------------------
|
||||||
|
|
||||||
|
Binary file formats used by Hadoop to store data, such as sequence files, HAR
|
||||||
|
files, etc, are guaranteed to remain compatible between minor releases. In
|
||||||
|
addition, in cases where changes are made between major releases,
|
||||||
|
both backward and forward compatibility must be maintained. Note that only the
|
||||||
|
sequence file format is guaranteed not to change incompatibly, not the
|
||||||
|
serialized classes that are contained therein.
|
||||||
|
|
||||||
|
In addition to the data produced by operations, Hadoop maintains its state
|
||||||
|
information in a variety of data stores in various formats, such as the HDFS
|
||||||
|
metadata store, the YARN resource manager state store, and the YARN federation
|
||||||
|
state store. All Hadoop internal data stores are considered internal and
|
||||||
|
[Private](./InterfaceClassification.html#Private) to Hadoop. Downstream
|
||||||
|
developers should not attempt to consume data from the Hadoop state store as
|
||||||
|
the data and/or data format may change unpredictably.
|
||||||
|
|
||||||
|
Automating Operations with the Hadoop CLI
|
||||||
|
-----------------------------------------
|
||||||
|
|
||||||
|
The set of tools that make up the Hadoop command-line interface are intended
|
||||||
|
both for consumption by end users and by downstream developers who are
|
||||||
|
creating tools that execute the CLI tools and parse the output. For this reason
|
||||||
|
the Hadoop CLI tools are treated like an interface and held stable between
|
||||||
|
major releases. Between major releases, no CLI tool options will be removed or
|
||||||
|
change semantically. The output from CLI tools will likewise remain the same
|
||||||
|
within a major version number. Note that any change to CLI tool output is
|
||||||
|
considered an incompatible change, so between major version, the CLI output
|
||||||
|
will not change. Note that the CLI tool output is distinct from
|
||||||
|
the log output produced by the CLI tools. Log output is not intended for
|
||||||
|
automated consumption and may change at any time.
|
||||||
|
|
||||||
|
Consuming the Hadoop Web UI
|
||||||
|
---------------------------
|
||||||
|
|
||||||
|
The web UIs that are exposed by Hadoop are for human consumption only. Scraping
|
||||||
|
the UIs for data is not a supported use case. No effort is made to ensure any
|
||||||
|
kind of compatibility between the data displayed in any of the web UIs
|
||||||
|
across releases.
|
||||||
|
|
||||||
|
Working with Hadoop configurations
|
||||||
|
----------------------------------
|
||||||
|
|
||||||
|
Hadoop uses two primary forms of configuration files: XML configuration files
|
||||||
|
and logging configuration files.
|
||||||
|
|
||||||
|
### XML Configuration Files
|
||||||
|
|
||||||
|
The XML configuration files contain a set of properties as name-value pairs.
|
||||||
|
The names and meanings of the properties are defined by Hadoop and are
|
||||||
|
guaranteed to be stable across minor releases. A property can only be removed
|
||||||
|
in a major release and only if it has been marked as deprecated for at least a
|
||||||
|
full major release. Most properties have a default value that will be used if
|
||||||
|
the property is not explicitly set in the XML configuration files. The default
|
||||||
|
property values will not be changed during a maintenance release. For details
|
||||||
|
about the properties supported by the various Hadoop components,
|
||||||
|
see the component documentation.
|
||||||
|
|
||||||
|
Downstream developers and users can add their own properties into the XML
|
||||||
|
configuration files for use by their tools and applications. While Hadoop
|
||||||
|
makes no formal restrictions about defining new properties, a new property
|
||||||
|
that conflicts with a property defined by Hadoop can lead to unexpected and
|
||||||
|
undesirable results. Users are encouraged to avoid using custom configuration
|
||||||
|
property names that conflict with the namespace of Hadoop-defined properties
|
||||||
|
and thus should avoid using any prefixes used by Hadoop,
|
||||||
|
e.g. hadoop, io, ipc, fs, net, file, ftp, kfs, ha, file, dfs, mapred,
|
||||||
|
mapreduce, and yarn.
|
||||||
|
|
||||||
|
### Logging Configuration Files
|
||||||
|
|
||||||
|
The log output produced by Hadoop daemons and CLIs is governed by a set of
|
||||||
|
configuration files. These files control the minimum level of log message that
|
||||||
|
will be output by the various components of Hadoop, as well as where and how
|
||||||
|
those messages are stored. Between minor releases no changes will be made to
|
||||||
|
the log configuration that reduce, eliminate, or redirect the log messages.
|
||||||
|
|
||||||
|
### Other Configuration Files
|
||||||
|
|
||||||
|
Hadoop makes use of a number of other types of configuration files in a variety
|
||||||
|
of formats, such as the JSON resource profiles configuration or the XML fair
|
||||||
|
scheduler configuration. No incompatible changes will be introduced to the
|
||||||
|
configuration file formats within a minor release. Even between minor releases
|
||||||
|
incompatible configuration file format changes will be avoided if possible.
|
||||||
|
|
||||||
|
Using and Consuming Hadoop Artifacts
|
||||||
|
------------------------------------
|
||||||
|
|
||||||
|
### Source and Configuration Files
|
||||||
|
|
||||||
|
As a downstream developer or consumer of Hadoop, it's possible to access all
|
||||||
|
elements of the Hadoop platform, including source code, configuration files,
|
||||||
|
build artifacts, etc. While the open nature of the platform allows it,
|
||||||
|
developers should not create dependencies on these internal details of Hadoop
|
||||||
|
as they may change at any time. The Hadoop development community will attempt,
|
||||||
|
however, to keep the existing structure stable within a major version.
|
||||||
|
|
||||||
|
The location and general structure of the Hadoop configuration files, job
|
||||||
|
history information (as consumed by the job history server), and logs files
|
||||||
|
generated by Hadoop will be maintained across maintenance releases.
|
||||||
|
|
||||||
|
### Build Artifacts
|
||||||
|
|
||||||
|
The build artifacts produced by the Hadoop build process, e.g. JAR files, are
|
||||||
|
subject to change at any time and should not be treated as reliable, except
|
||||||
|
for the client artifacts. Client artifacts and their contents will remain
|
||||||
|
compatible within a major release. It is the goal of the Hadoop development
|
||||||
|
community to allow application code to continue to function unchanged across
|
||||||
|
minor releases and, whenever possible, across major releases. The current list
|
||||||
|
of client artifacts is as follows:
|
||||||
|
|
||||||
|
* hadoop-client
|
||||||
|
* hadoop-client-api
|
||||||
|
* hadoop-client-minicluster
|
||||||
|
* hadoop-client-runtime
|
||||||
|
* hadoop-hdfs-client
|
||||||
|
* hadoop-hdfs-native-client
|
||||||
|
* hadoop-mapreduce-client-app
|
||||||
|
* hadoop-mapreduce-client-common
|
||||||
|
* hadoop-mapreduce-client-core
|
||||||
|
* hadoop-mapreduce-client-jobclient
|
||||||
|
* hadoop-mapreduce-client-nativetask
|
||||||
|
* hadoop-yarn-client
|
||||||
|
|
||||||
|
### Environment Variables
|
||||||
|
|
||||||
|
Some Hadoop components receive information through environment variables. For
|
||||||
|
example, the ```HADOOP_OPTS``` environment variable is interpreted by most
|
||||||
|
Hadoop processes as a string of additional JVM arguments to be used when
|
||||||
|
starting a new JVM. Between minor releases the way Hadoop interprets environment
|
||||||
|
variables will not change in an incompatible way. In other words, the same value
|
||||||
|
placed into the same variable should produce the same result for all Hadoop
|
||||||
|
releases within the same major version.
|
||||||
|
|
||||||
|
### Library Dependencies
|
||||||
|
|
||||||
|
Hadoop relies on a large number of third-party libraries for its operation. As
|
||||||
|
much as possible the Hadoop developer community works to hide these dependencies
|
||||||
|
from downstream developers. Some common libraries, such as Guava, could cause
|
||||||
|
significant compatibility issues between Hadoop and downstream applications if
|
||||||
|
those dependencies were exposed downstream. Nonetheless Hadoop does expose some
|
||||||
|
of its dependencies, especially prior to Hadoop 3. No new dependency
|
||||||
|
will be exposed by Hadoop via the client artifacts between major releases.
|
||||||
|
|
||||||
|
A common downstream anti-pattern is to use the output of ```hadoop classpath```
|
||||||
|
to set the downstream application's classpath or add all third-party JARs
|
||||||
|
included with Hadoop to the downstream application's classpath. This practice
|
||||||
|
creates a tight coupling between the downstream application and Hadoop's
|
||||||
|
third-party dependencies, which leads to a fragile application that is hard to
|
||||||
|
maintain as Hadoop's dependencies change. This practice is strongly discouraged.
|
||||||
|
|
||||||
|
Hadoop depends on the Java virtual machine for its operation, which can
|
||||||
|
impact downstream applications. To minimize disruption, the minimum supported
|
||||||
|
version of the JVM will not change between major releases of Hadoop. In the
|
||||||
|
event that the current minimum supported JVM version becomes unsupported
|
||||||
|
between major releases, the minimum supported JVM version may be changed in a
|
||||||
|
minor release.
|
||||||
|
|
||||||
|
Hadoop also includes several native components, including compression, the
|
||||||
|
container executor binary, and various native integrations. These native
|
||||||
|
components introduce a set of native dependencies for Hadoop. The set of
|
||||||
|
native dependencies can change in a minor release, but the Hadoop developer
|
||||||
|
community will try to limit any dependency version changes to minor version
|
||||||
|
changes as much as possible.
|
||||||
|
|
||||||
|
### Hardware and OS Dependencies
|
||||||
|
|
||||||
|
Hadoop is currently supported by the Hadoop developer community on Linux and
|
||||||
|
Windows running on x86 and AMD processors. These OSes and processors are likely
|
||||||
|
to remain supported for the foreseeable future. In the event that support plans
|
||||||
|
change, the OS or processor to be dropped will be documented as deprecated
|
||||||
|
for at least a full minor release, but ideally a full major release, before
|
||||||
|
actually being dropped. Hadoop may function on other OSes and processor
|
||||||
|
architectures, but the community may not be able to provide assistance in the
|
||||||
|
event of issues.
|
||||||
|
|
||||||
|
There are no guarantees on how the minimum resources required by Hadoop daemons
|
||||||
|
will change between releases, even maintenance releases. Nonetheless, the
|
||||||
|
Hadoop developer community will try to avoid increasing the requirements within
|
||||||
|
a minor release.
|
||||||
|
|
||||||
|
Any file systems supported Hadoop, such as through the FileSystem API, will
|
||||||
|
in most cases continue to be supported throughout a major release. The only
|
||||||
|
case where support for a file system can be dropped within a major version is
|
||||||
|
if a clean migration path to an alternate client implementation is provided.
|
||||||
|
|
||||||
|
Questions
|
||||||
|
---------
|
||||||
|
|
||||||
|
For question about developing applications and projects against Apache Hadoop,
|
||||||
|
please contact the developer mailing list for the relevant component(s):
|
||||||
|
|
||||||
|
* [dev-common](mailto:dev-common@apache.org)
|
||||||
|
* [dev-hdfs](mailto:dev-hdfs@apache.org)
|
||||||
|
* [dev-mapreduce](mailto:dev-mapreduce@apache.org)
|
||||||
|
* [dev-yarn](mailto:dev-yarn@apache.org)
|
|
@ -52,7 +52,8 @@
|
||||||
<item name="Cluster Setup" href="hadoop-project-dist/hadoop-common/ClusterSetup.html"/>
|
<item name="Cluster Setup" href="hadoop-project-dist/hadoop-common/ClusterSetup.html"/>
|
||||||
<item name="Commands Reference" href="hadoop-project-dist/hadoop-common/CommandsManual.html"/>
|
<item name="Commands Reference" href="hadoop-project-dist/hadoop-common/CommandsManual.html"/>
|
||||||
<item name="FileSystem Shell" href="hadoop-project-dist/hadoop-common/FileSystemShell.html"/>
|
<item name="FileSystem Shell" href="hadoop-project-dist/hadoop-common/FileSystemShell.html"/>
|
||||||
<item name="Compatibility" href="hadoop-project-dist/hadoop-common/Compatibility.html"/>
|
<item name="Compatibility Specification" href="hadoop-project-dist/hadoop-common/Compatibility.html"/>
|
||||||
|
<item name="Downstream Developer's Guide" href="hadoop-project-dist/hadoop-common/DownstreamDev.html"/>
|
||||||
<item name="Interface Classification" href="hadoop-project-dist/hadoop-common/InterfaceClassification.html"/>
|
<item name="Interface Classification" href="hadoop-project-dist/hadoop-common/InterfaceClassification.html"/>
|
||||||
<item name="FileSystem Specification"
|
<item name="FileSystem Specification"
|
||||||
href="hadoop-project-dist/hadoop-common/filesystem/index.html"/>
|
href="hadoop-project-dist/hadoop-common/filesystem/index.html"/>
|
||||||
|
|
Loading…
Reference in New Issue