svn merge -c 1589405 merging from trunk to branch-2 to fix:HADOOP-10527. Fix incorrect return code and allow more retries on EINTR.

git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/branches/branch-2@1589407 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
Kihwal Lee 2014-04-23 13:28:55 +00:00
parent c5450fd0fe
commit f4bba31c92
4 changed files with 62 additions and 59 deletions

View File

@ -107,6 +107,9 @@ Release 2.4.1 - UNRELEASED
HADOOP-10522. JniBasedUnixGroupMapping mishandles errors. (kihwal) HADOOP-10522. JniBasedUnixGroupMapping mishandles errors. (kihwal)
HADOOP-10527. Fix incorrect return code and allow more retries on EINTR.
(kihwal)
Release 2.4.0 - 2014-04-07 Release 2.4.0 - 2014-04-07
INCOMPATIBLE CHANGES INCOMPATIBLE CHANGES

View File

@ -124,10 +124,8 @@ Java_org_apache_hadoop_security_JniBasedUnixGroupsMapping_getGroupsForUser
if (ret == ENOENT) { if (ret == ENOENT) {
jgroups = (*env)->NewObjectArray(env, 0, g_string_clazz, NULL); jgroups = (*env)->NewObjectArray(env, 0, g_string_clazz, NULL);
} else { // handle other errors } else { // handle other errors
char buf[128]; (*env)->Throw(env, newRuntimeException(env,
snprintf(buf, sizeof(buf), "getgrouplist: error looking up user. %d (%s)", "getgrouplist: error looking up user. %d (%s)", ret, terror(ret)));
ret, terror(ret));
THROW(env, "java/lang/RuntimeException", buf);
} }
goto done; goto done;
} }
@ -142,10 +140,8 @@ Java_org_apache_hadoop_security_JniBasedUnixGroupsMapping_getGroupsForUser
if (ret == ENOMEM) { if (ret == ENOMEM) {
THROW(env, "java/lang/OutOfMemoryError", NULL); THROW(env, "java/lang/OutOfMemoryError", NULL);
} else { } else {
char buf[128]; (*env)->Throw(env, newRuntimeException(env,
snprintf(buf, sizeof(buf), "getgrouplist: error looking up groups. %d (%s)", "getgrouplist: error looking up group. %d (%s)", ret, terror(ret)));
ret, terror(ret));
THROW(env, "java/lang/RuntimeException", buf);
} }
goto done; goto done;
} }

View File

@ -27,25 +27,24 @@
#include <string.h> #include <string.h>
#include <unistd.h> #include <unistd.h>
#define MAX_GROUP_LOOKUP_TRIES 5 // Assuming the average of user name length of 15 bytes,
// 8KB buffer will be large enough for a group with about 500 members.
// 2MB buffer will be large enough for a group with about 130K members.
#define INITIAL_GROUP_BUFFER_SIZE (8*1024)
#define MAX_GROUP_BUFFER_SIZE (2*1024*1024)
struct hadoop_group_info *hadoop_group_info_alloc(void) struct hadoop_group_info *hadoop_group_info_alloc(void)
{ {
struct hadoop_group_info *ginfo; struct hadoop_group_info *ginfo;
size_t buf_sz;
char *buf; char *buf;
ginfo = calloc(1, sizeof(struct hadoop_group_info)); ginfo = calloc(1, sizeof(struct hadoop_group_info));
buf_sz = sysconf(_SC_GETGR_R_SIZE_MAX); buf = malloc(INITIAL_GROUP_BUFFER_SIZE);
if (buf_sz < 1024) {
buf_sz = 1024;
}
buf = malloc(buf_sz);
if (!buf) { if (!buf) {
free(ginfo); free(ginfo);
return NULL; return NULL;
} }
ginfo->buf_sz = buf_sz; ginfo->buf_sz = INITIAL_GROUP_BUFFER_SIZE;
ginfo->buf = buf; ginfo->buf = buf;
return ginfo; return ginfo;
} }
@ -86,48 +85,49 @@ static int getgrgid_error_translate(int err)
int hadoop_group_info_fetch(struct hadoop_group_info *ginfo, gid_t gid) int hadoop_group_info_fetch(struct hadoop_group_info *ginfo, gid_t gid)
{ {
struct group *group; struct group *group;
int ret, i; int ret;
size_t buf_sz; size_t buf_sz;
char *nbuf; char *nbuf;
hadoop_group_info_clear(ginfo); hadoop_group_info_clear(ginfo);
for (i = 0, ret = 0; i < MAX_GROUP_LOOKUP_TRIES; i++) { for (;;) {
// If the previous call returned ERANGE, increase the buffer size // On success, the following call returns 0 and group is set to non-NULL.
if (ret == ERANGE) {
buf_sz = ginfo->buf_sz * 2;
nbuf = realloc(ginfo->buf, buf_sz);
if (!nbuf) {
return ENOMEM;
}
ginfo->buf = nbuf;
ginfo->buf_sz = buf_sz;
}
// The following call returns errno. Reading the global errno wihtout
// locking is not thread-safe.
group = NULL; group = NULL;
ret = getgrgid_r(gid, &ginfo->group, ginfo->buf, ret = getgrgid_r(gid, &ginfo->group, ginfo->buf,
ginfo->buf_sz, &group); ginfo->buf_sz, &group);
switch(ret) { switch(ret) {
case 0: case 0:
if (!group) { if (!group) {
// The underlying library likely has a bug. // Not found.
return EIO; return ENOENT;
} }
// Found.
return 0; return 0;
case EINTR: case EINTR:
case ERANGE:
// Retry on these errors.
// EINTR: a signal was handled and this thread was allowed to continue. // EINTR: a signal was handled and this thread was allowed to continue.
break;
case ERANGE:
// ERANGE: the buffer was not big enough. // ERANGE: the buffer was not big enough.
if (ginfo->buf_sz == MAX_GROUP_BUFFER_SIZE) {
// Already tried with the max size.
return ENOMEM;
}
buf_sz = ginfo->buf_sz * 2;
if (buf_sz > MAX_GROUP_BUFFER_SIZE) {
buf_sz = MAX_GROUP_BUFFER_SIZE;
}
nbuf = realloc(ginfo->buf, buf_sz);
if (!nbuf) {
return ENOMEM;
}
ginfo->buf = nbuf;
ginfo->buf_sz = buf_sz;
break; break;
default: default:
// Lookup failed. // Lookup failed.
return getgrgid_error_translate(ret); return getgrgid_error_translate(ret);
} }
} }
// Did not succeed after the retries. Return the last error.
return getgrgid_error_translate(ret);
} }
#ifdef GROUP_TESTING #ifdef GROUP_TESTING

View File

@ -28,7 +28,10 @@
#include <unistd.h> #include <unistd.h>
#define INITIAL_GIDS_SIZE 32 #define INITIAL_GIDS_SIZE 32
#define MAX_USER_LOOKUP_TRIES 5 // 1KB buffer should be large enough to store a passwd record in most
// cases, but it can get bigger if each field is maximally used. The
// max is defined to avoid buggy libraries making us run out of memory.
#define MAX_USER_BUFFER_SIZE (32*1024)
struct hadoop_user_info *hadoop_user_info_alloc(void) struct hadoop_user_info *hadoop_user_info_alloc(void)
{ {
@ -96,48 +99,49 @@ int hadoop_user_info_fetch(struct hadoop_user_info *uinfo,
const char *username) const char *username)
{ {
struct passwd *pwd; struct passwd *pwd;
int ret, i; int ret;
size_t buf_sz; size_t buf_sz;
char *nbuf; char *nbuf;
hadoop_user_info_clear(uinfo); hadoop_user_info_clear(uinfo);
for (i = 0, ret = 0; i < MAX_USER_LOOKUP_TRIES; i++) { for (;;) {
// If the previous call returned ERANGE, increase the buffer size // On success, the following call returns 0 and pwd is set to non-NULL.
if (ret == ERANGE) {
buf_sz = uinfo->buf_sz * 2;
nbuf = realloc(uinfo->buf, buf_sz);
if (!nbuf) {
return ENOMEM;
}
uinfo->buf = nbuf;
uinfo->buf_sz = buf_sz;
}
// The following call returns errno. Reading the global errno wihtout
// locking is not thread-safe.
pwd = NULL; pwd = NULL;
ret = getpwnam_r(username, &uinfo->pwd, uinfo->buf, ret = getpwnam_r(username, &uinfo->pwd, uinfo->buf,
uinfo->buf_sz, &pwd); uinfo->buf_sz, &pwd);
switch(ret) { switch(ret) {
case 0: case 0:
if (!pwd) { if (!pwd) {
// The underlying library likely has a bug. // Not found.
return EIO; return ENOENT;
} }
// Found.
return 0; return 0;
case EINTR: case EINTR:
case ERANGE:
// Retry on these errors.
// EINTR: a signal was handled and this thread was allowed to continue. // EINTR: a signal was handled and this thread was allowed to continue.
break;
case ERANGE:
// ERANGE: the buffer was not big enough. // ERANGE: the buffer was not big enough.
if (uinfo->buf_sz == MAX_USER_BUFFER_SIZE) {
// Already tried with the max size.
return ENOMEM;
}
buf_sz = uinfo->buf_sz * 2;
if (buf_sz > MAX_USER_BUFFER_SIZE) {
buf_sz = MAX_USER_BUFFER_SIZE;
}
nbuf = realloc(uinfo->buf, buf_sz);
if (!nbuf) {
return ENOMEM;
}
uinfo->buf = nbuf;
uinfo->buf_sz = buf_sz;
break; break;
default: default:
// Lookup failed. // Lookup failed.
return getpwnam_error_translate(ret); return getpwnam_error_translate(ret);
} }
} }
// Did not succeed after the retries. Return the last error.
return getpwnam_error_translate(ret);
} }
static int put_primary_gid_first(struct hadoop_user_info *uinfo) static int put_primary_gid_first(struct hadoop_user_info *uinfo)