Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[#3432] improvement(test): Add kerberos authentication IT for HDFS cluster #3435

Merged
merged 27 commits into from
May 23, 2024
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
27 commits
Select commit Hold shift + click to select a range
fd390d0
Add kerberos IT
yuqi1129 May 17, 2024
780e375
Fix
yuqi1129 May 18, 2024
9c0a2ed
Fix test error
yuqi1129 May 18, 2024
3a6a31b
Merge branch 'main' of github.com:datastrato/graviton into issue_3432
yuqi1129 May 20, 2024
3f06416
Add fix.
yuqi1129 May 20, 2024
85e6b73
fix
yuqi1129 May 20, 2024
66640fc
fix
yuqi1129 May 20, 2024
01dbf60
Fix
yuqi1129 May 20, 2024
d6960c8
Merge branch 'main' into issue_3432
yuqi1129 May 20, 2024
e71dde3
Remove unused code.
yuqi1129 May 21, 2024
5a26aa1
Merge remote-tracking branch 'me/issue_3432' into issue_3432
yuqi1129 May 21, 2024
6488d03
Merge branch 'main' into issue_3432
yuqi1129 May 21, 2024
8aa9b4d
optimize code.
yuqi1129 May 21, 2024
4609efb
Merge remote-tracking branch 'me/issue_3432' into issue_3432
yuqi1129 May 21, 2024
30d4c60
Fix mistake
yuqi1129 May 21, 2024
965f63a
Revert the code that check status of `show databases` for Hive contai…
yuqi1129 May 21, 2024
0f8051c
Merge branch 'main' into issue_3432
yuqi1129 May 21, 2024
6919203
Merge main and rebase the code.
yuqi1129 May 22, 2024
ac82116
Merge remote-tracking branch 'me/issue_3432' into issue_3432
yuqi1129 May 22, 2024
2cb6eb6
Resolve comments
yuqi1129 May 22, 2024
cea7331
Merge main and resolve conflicts
yuqi1129 May 22, 2024
83d661a
Remove used environment
yuqi1129 May 22, 2024
64dbfc4
Fix style
yuqi1129 May 22, 2024
3f3fd1f
Merge branch 'main' into issue_3432
yuqi1129 May 22, 2024
7a2b2a3
Increase retry number.
yuqi1129 May 23, 2024
b7b06a7
Merge remote-tracking branch 'me/issue_3432' into issue_3432
yuqi1129 May 23, 2024
2089149
fix
yuqi1129 May 23, 2024
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
1 change: 1 addition & 0 deletions catalogs/catalog-hadoop/build.gradle.kts
Original file line number Diff line number Diff line change
Expand Up @@ -99,6 +99,7 @@ tasks.test {

doFirst {
environment("GRAVITINO_CI_HIVE_DOCKER_IMAGE", "datastrato/gravitino-ci-hive:0.1.12")
environment("GRAVITINO_CI_KERBEROS_HIVE_DOCKER_IMAGE", "datastrato/gravitino-ci-kerberos-hive:0.1.0")
}

val init = project.extra.get("initIntegrationTest") as (Test) -> Unit
Expand Down
Original file line number Diff line number Diff line change
@@ -0,0 +1,120 @@
/*
* Copyright 2024 Datastrato Pvt Ltd.
* This software is licensed under the Apache License version 2.
*/

package com.datastrato.gravitino.catalog.hadoop.integration.test;

import com.datastrato.gravitino.integration.test.container.ContainerSuite;
import com.datastrato.gravitino.integration.test.container.HiveContainer;
import java.io.File;
import java.io.IOException;
import java.nio.charset.StandardCharsets;
import java.nio.file.Files;
import java.security.PrivilegedAction;
import org.apache.commons.io.FileUtils;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.security.AccessControlException;
import org.apache.hadoop.security.UserGroupInformation;
import org.junit.jupiter.api.AfterAll;
import org.junit.jupiter.api.Assertions;
import org.junit.jupiter.api.BeforeAll;
import org.junit.jupiter.api.Tag;
import org.junit.jupiter.api.Test;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;

@Tag("gravitino-docker-it")
public class HDFSKerberosIT {
private static final Logger LOG = LoggerFactory.getLogger(HDFSKerberosIT.class);

private static final ContainerSuite containerSuite = ContainerSuite.getInstance();
private static final String CLIENT_PRINCIPAL = "cli@HADOOPKRB";
private static UserGroupInformation clientUGI;

private static String keytabPath;

@BeforeAll
public static void setup() throws IOException {
containerSuite.startKerberosHiveContainer();

File baseDir = new File(System.getProperty("java.io.tmpdir"));
File file = Files.createTempDirectory(baseDir.toPath(), "test").toFile();
file.deleteOnExit();

// Copy the keytab and krb5.conf from the container
keytabPath = file.getAbsolutePath() + "/client.keytab";
containerSuite
.getKerberosHiveContainer()
.getContainer()
.copyFileFromContainer("/etc/admin.keytab", keytabPath);

String krb5TmpPath = file.getAbsolutePath() + "/krb5.conf_tmp";
String krb5Path = file.getAbsolutePath() + "/krb5.conf";
containerSuite
.getKerberosHiveContainer()
.getContainer()
.copyFileFromContainer("/etc/krb5.conf", krb5TmpPath);

// Modify the krb5.conf and change the kdc and admin_server to the container IP
String ip = containerSuite.getKerberosHiveContainer().getContainerIpAddress();
String content = FileUtils.readFileToString(new File(krb5TmpPath), StandardCharsets.UTF_8);
content = content.replace("kdc = localhost:88", "kdc = " + ip + ":88");
content = content.replace("admin_server = localhost", "admin_server = " + ip + ":749");
FileUtils.write(new File(krb5Path), content, StandardCharsets.UTF_8);

LOG.info("Kerberos kdc config:\n{}", content);

System.setProperty("java.security.krb5.conf", krb5Path);
System.setProperty("sun.security.krb5.debug", "true");
}

@AfterAll
public static void tearDown() {
// Reset the UGI
UserGroupInformation.reset();

// Clean up the kerberos configuration
System.clearProperty("java.security.krb5.conf");
System.clearProperty("sun.security.krb5.debug");
}

@Test
public void testKerberosHDFS() throws IOException {
Configuration conf = new Configuration();
conf.set("fs.defaultFS", defaultBaseLocation());
conf.setBoolean("fs.hdfs.impl.disable.cache", true);
conf.set("hadoop.security.authentication", "kerberos");

UserGroupInformation.setConfiguration(conf);
clientUGI = UserGroupInformation.loginUserFromKeytabAndReturnUGI(CLIENT_PRINCIPAL, keytabPath);
PrivilegedAction<?> action =
(PrivilegedAction)
() -> {
try {
FileSystem fs = FileSystem.get(conf);
Path path = new Path("/");
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

You need to create a file and check the file owner to see if it is expected.

If kerberos is enabled, the user should be "cli" from my understanding.

Assertions.assertTrue(fs.exists(path));
return null;
} catch (IOException e) {
throw new RuntimeException(e);
}
};

clientUGI.doAs(action);

// Clear UGI, It will throw exception
UserGroupInformation.reset();
Exception e = Assertions.assertThrows(Exception.class, action::run);
Assertions.assertInstanceOf(AccessControlException.class, e.getCause());
}

private static String defaultBaseLocation() {
return String.format(
"hdfs://%s:%d/user/",
containerSuite.getKerberosHiveContainer().getContainerIpAddress(),
HiveContainer.HDFS_DEFAULTFS_PORT);
}
}
Original file line number Diff line number Diff line change
Expand Up @@ -36,7 +36,7 @@
public abstract class BaseContainer implements AutoCloseable {
public static final Logger LOG = LoggerFactory.getLogger(BaseContainer.class);
// Host name of the container
private final String hostName;
protected final String hostName;
// Exposed ports of the container
private final Set<Integer> ports;
// Files to mount in the container
Expand Down Expand Up @@ -153,6 +153,7 @@ protected abstract static class Builder<
SELF extends Builder<SELF, CONTAINER>, CONTAINER extends BaseContainer> {
protected String image;
protected String hostName;
protected boolean kerberosEnabled;
protected Set<Integer> exposePorts = ImmutableSet.of();
protected Map<String, String> extraHosts = ImmutableMap.of();
protected Map<String, String> filesToMount = ImmutableMap.of();
Expand Down Expand Up @@ -201,6 +202,11 @@ public SELF withNetwork(Network network) {
return self;
}

public SELF withKerberosEnabled(boolean kerberosEnabled) {
this.kerberosEnabled = kerberosEnabled;
return self;
}

public abstract CONTAINER build();
}
}
Original file line number Diff line number Diff line change
Expand Up @@ -42,6 +42,7 @@ public class ContainerSuite implements Closeable {
private static volatile TrinoITContainers trinoITContainers;
private static volatile KafkaContainer kafkaContainer;
private static volatile DorisContainer dorisContainer;
private static volatile HiveContainer kerberosHiveContainer;

private static volatile MySQLContainer mySQLContainer;
private static volatile MySQLContainer mySQLVersion5Container;
Expand Down Expand Up @@ -101,6 +102,25 @@ public void startHiveContainer() {
}
}

public void startKerberosHiveContainer() {
if (kerberosHiveContainer == null) {
synchronized (ContainerSuite.class) {
if (kerberosHiveContainer == null) {
// Start Hive container
HiveContainer.Builder hiveBuilder =
HiveContainer.builder()
.withHostName("gravitino-ci-kerberos-hive")
.withEnvVars(ImmutableMap.<String, String>builder().build())
.withKerberosEnabled(true)
.withNetwork(network);
HiveContainer container = closer.register(hiveBuilder.build());
container.start();
kerberosHiveContainer = container;
}
}
}
}

public void startTrinoContainer(
String trinoConfDir,
String trinoConnectorLibDir,
Expand Down Expand Up @@ -284,6 +304,10 @@ public HiveContainer getHiveContainer() {
return hiveContainer;
}

public HiveContainer getKerberosHiveContainer() {
return kerberosHiveContainer;
}

public DorisContainer getDorisContainer() {
return dorisContainer;
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -25,6 +25,8 @@ public class HiveContainer extends BaseContainer {
public static final Logger LOG = LoggerFactory.getLogger(HiveContainer.class);

public static final String DEFAULT_IMAGE = System.getenv("GRAVITINO_CI_HIVE_DOCKER_IMAGE");
public static final String KERBEROS_IMAGE =
System.getenv("GRAVITINO_CI_KERBEROS_HIVE_DOCKER_IMAGE");
public static final String HOST_NAME = "gravitino-ci-hive";
private static final int MYSQL_PORT = 3306;
public static final int HDFS_DEFAULTFS_PORT = 9000;
Expand All @@ -51,13 +53,13 @@ protected HiveContainer(
@Override
protected void setupContainer() {
super.setupContainer();
withLogConsumer(new PrintingContainerLog(format("%-14s| ", "HiveContainer")));
withLogConsumer(new PrintingContainerLog(format("%-14s| ", "HiveContainer-" + hostName)));
}

@Override
public void start() {
super.start();
Preconditions.check("Hive container startup failed!", checkContainerStatus(10));
Preconditions.check("Hive container startup failed!", checkContainerStatus(15));
}

@Override
Expand Down Expand Up @@ -88,8 +90,8 @@ private void copyHiveLog() {
@Override
protected boolean checkContainerStatus(int retryLimit) {
await()
.atMost(100, TimeUnit.SECONDS)
.pollInterval(100 / retryLimit, TimeUnit.SECONDS)
.atMost(150, TimeUnit.SECONDS)
.pollInterval(150 / retryLimit, TimeUnit.SECONDS)
.until(
() -> {
try {
Expand All @@ -113,40 +115,37 @@ protected boolean checkContainerStatus(int retryLimit) {
return false;
});

String sql = "show databases";
final String showDatabaseSQL = "show databases";
await()
.atMost(30, TimeUnit.SECONDS)
.pollInterval(30 / retryLimit, TimeUnit.SECONDS)
.until(
() -> {
try {
Container.ExecResult result = executeInContainer("hive", "-e", sql);
Container.ExecResult result = executeInContainer("hive", "-e", showDatabaseSQL);
if (result.getStdout().contains("default")) {
return true;
}
} catch (Exception e) {
LOG.error("Failed to execute sql: {}", sql, e);
LOG.error("Failed to execute sql: {}", showDatabaseSQL, e);
}
return false;
});

final String createTableSQL =
"CREATE TABLE IF NOT EXISTS default.employee ( eid int, name String, "
+ "salary String, destination String) ";
await()
.atMost(30, TimeUnit.SECONDS)
.pollInterval(30 / retryLimit, TimeUnit.SECONDS)
.until(
() -> {
try {
Container.ExecResult result =
executeInContainer(
"hive",
"-e",
"CREATE TABLE IF NOT EXISTS default.employee ( eid int, name String, "
+ "salary String, destination String) ");
Container.ExecResult result = executeInContainer("hive", "-e", createTableSQL);
if (result.getExitCode() == 0) {
return true;
}
} catch (Exception e) {
LOG.error("Failed to execute sql: {}", sql, e);
LOG.error("Failed to execute sql: {}", createTableSQL, e);
}
return false;
});
Expand Down Expand Up @@ -183,7 +182,13 @@ private Builder() {
@Override
public HiveContainer build() {
return new HiveContainer(
image, hostName, exposePorts, extraHosts, filesToMount, envVars, network);
kerberosEnabled ? KERBEROS_IMAGE : image,
kerberosEnabled ? "kerberos-" + hostName : hostName,
exposePorts,
extraHosts,
filesToMount,
envVars,
network);
}
}
}
Loading