hbase/hbase-protocol
Ted Malaska e95358a7fc HBASE-14181 Add Spark DataFrame DataSource to HBase-Spark Module
Signed-off-by: Sean Busbey <busbey@cloudera.com>
2015-09-07 09:44:10 -05:00
..
src/main HBASE-14181 Add Spark DataFrame DataSource to HBase-Spark Module 2015-09-07 09:44:10 -05:00
README.txt HBASE-13325 Protocol Buffers 2.5 no longer available for download on code.google.com 2015-03-25 14:05:34 -07:00
pom.xml HBASE-14260 skip javadoc generation for the hbase-protocol generated classes 2015-08-20 21:14:06 -05:00

README.txt

These are the protobuf definition files used by hbase. The produced java
classes are generated into src/main/java/org/apache/hadoop/hbase/protobuf/generated
and then checked in.  The reasoning is that they change infrequently.

To regenerate the classes after making definition file changes, ensure first that
the protobuf protoc tool is in your $PATH (You may need to download it and build
it first; its part of the protobuf package obtainable from here: 
https://github.com/google/protobuf/releases/tag/v2.5.0).

HBase uses hadoop-maven-plugins:protoc goal to invoke the protoc command. You can 
compile the protoc definitions by invoking maven with profile compile-protobuf or 
passing in compile-protobuf property. 

mvn compile -Dcompile-protobuf
or
mvn compile -Pcompile-protobuf

You may also want to define protoc.path for the protoc binary

mvn compile -Dcompile-protobuf -Dprotoc.path=/opt/local/bin/protoc

If you have added a new proto file, you should add it to the pom.xml file first.
Other modules also support the maven profile.

After you've done the above, check it in and then check it in (or post a patch
on a JIRA with your definition file changes and the generated files).