Skip to content

Commit

Permalink
HBASE-24380 : Provide WAL splitting journal logging (#1860)
Browse files Browse the repository at this point in the history
Signed-off-by: Andrew Purtell <apurtell@apache.org>
  • Loading branch information
virajjasani authored Jun 19, 2020
1 parent 6b62c0b commit 3c31981
Show file tree
Hide file tree
Showing 6 changed files with 92 additions and 39 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -57,7 +57,7 @@ public AbstractRecoveredEditsOutputSink(WALSplitter walSplitter,
* @return a writer that wraps a {@link WALProvider.Writer} and its Path. Caller should close.
*/
protected RecoveredEditsWriter createRecoveredEditsWriter(TableName tableName, byte[] region,
long seqId) throws IOException {
long seqId) throws IOException {
Path regionEditsPath = getRegionSplitEditsPath(tableName, region, seqId,
walSplitter.getFileBeingSplit().getPath().getName(), walSplitter.getTmpDirName(),
walSplitter.conf);
Expand All @@ -70,27 +70,35 @@ protected RecoveredEditsWriter createRecoveredEditsWriter(TableName tableName, b
}
}
WALProvider.Writer w = walSplitter.createWriter(regionEditsPath);
LOG.info("Creating recovered edits writer path={}", regionEditsPath);
final String msg = "Creating recovered edits writer path=" + regionEditsPath;
LOG.info(msg);
updateStatusWithMsg(msg);
return new RecoveredEditsWriter(region, regionEditsPath, w, seqId);
}

protected Path closeRecoveredEditsWriter(RecoveredEditsWriter editsWriter,
List<IOException> thrown) throws IOException {
List<IOException> thrown) throws IOException {
try {
editsWriter.writer.close();
} catch (IOException ioe) {
LOG.error("Could not close recovered edits at {}", editsWriter.path, ioe);
final String errorMsg = "Could not close recovered edits at " + editsWriter.path;
LOG.error(errorMsg, ioe);
updateStatusWithMsg(errorMsg);
thrown.add(ioe);
return null;
}
LOG.info("Closed recovered edits writer path={} (wrote {} edits, skipped {} edits in {} ms",
editsWriter.path, editsWriter.editsWritten, editsWriter.editsSkipped,
editsWriter.nanosSpent / 1000 / 1000);
final String msg = "Closed recovered edits writer path=" + editsWriter.path + " (wrote "
+ editsWriter.editsWritten + " edits, skipped " + editsWriter.editsSkipped + " edits in " + (
editsWriter.nanosSpent / 1000 / 1000) + " ms)";
LOG.info(msg);
updateStatusWithMsg(msg);
if (editsWriter.editsWritten == 0) {
// just remove the empty recovered.edits file
if (walSplitter.walFS.exists(editsWriter.path) &&
!walSplitter.walFS.delete(editsWriter.path, false)) {
LOG.warn("Failed deleting empty {}", editsWriter.path);
if (walSplitter.walFS.exists(editsWriter.path)
&& !walSplitter.walFS.delete(editsWriter.path, false)) {
final String errorMsg = "Failed deleting empty " + editsWriter.path;
LOG.warn(errorMsg);
updateStatusWithMsg(errorMsg);
throw new IOException("Failed deleting empty " + editsWriter.path);
}
return null;
Expand All @@ -107,13 +115,20 @@ protected Path closeRecoveredEditsWriter(RecoveredEditsWriter editsWriter,
// TestHLogSplit#testThreading is an example.
if (walSplitter.walFS.exists(editsWriter.path)) {
if (!walSplitter.walFS.rename(editsWriter.path, dst)) {
throw new IOException(
"Failed renaming recovered edits " + editsWriter.path + " to " + dst);
final String errorMsg =
"Failed renaming recovered edits " + editsWriter.path + " to " + dst;
updateStatusWithMsg(errorMsg);
throw new IOException(errorMsg);
}
LOG.info("Rename recovered edits {} to {}", editsWriter.path, dst);
final String renameEditMsg = "Rename recovered edits " + editsWriter.path + " to " + dst;
LOG.info(renameEditMsg);
updateStatusWithMsg(renameEditMsg);
}
} catch (IOException ioe) {
LOG.error("Could not rename recovered edits {} to {}", editsWriter.path, dst, ioe);
final String errorMsg = "Could not rename recovered edits " + editsWriter.path
+ " to " + dst;
LOG.error(errorMsg, ioe);
updateStatusWithMsg(errorMsg);
thrown.add(ioe);
return null;
}
Expand Down Expand Up @@ -216,26 +231,33 @@ private void incrementNanoTime(long nanos) {

void writeRegionEntries(List<WAL.Entry> entries) throws IOException {
long startTime = System.nanoTime();
try {
int editsCount = 0;
for (WAL.Entry logEntry : entries) {
filterCellByStore(logEntry);
if (!logEntry.getEdit().isEmpty()) {
int editsCount = 0;
for (WAL.Entry logEntry : entries) {
filterCellByStore(logEntry);
if (!logEntry.getEdit().isEmpty()) {
try {
writer.append(logEntry);
updateRegionMaximumEditLogSeqNum(logEntry);
editsCount++;
} else {
incrementSkippedEdits(1);
} catch (IOException e) {
logAndThrowWriterAppendFailure(logEntry, e);
}
updateRegionMaximumEditLogSeqNum(logEntry);
editsCount++;
} else {
incrementSkippedEdits(1);
}
// Pass along summary statistics
incrementEdits(editsCount);
incrementNanoTime(System.nanoTime() - startTime);
} catch (IOException e) {
e = e instanceof RemoteException ? ((RemoteException) e).unwrapRemoteException() : e;
LOG.error(HBaseMarkers.FATAL, "Got while writing log entry to log", e);
throw e;
}
// Pass along summary statistics
incrementEdits(editsCount);
incrementNanoTime(System.nanoTime() - startTime);
}

private void logAndThrowWriterAppendFailure(WAL.Entry logEntry, IOException e)
throws IOException {
e = e instanceof RemoteException ? ((RemoteException) e).unwrapRemoteException() : e;
final String errorMsg = "Failed to write log entry " + logEntry.toString() + " to log";
LOG.error(HBaseMarkers.FATAL, errorMsg, e);
updateStatusWithMsg(errorMsg);
throw e;
}

private void filterCellByStore(WAL.Entry logEntry) {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -57,7 +57,8 @@ public BoundedRecoveredEditsOutputSink(WALSplitter walSplitter,
}

@Override
public void append(EntryBuffers.RegionEntryBuffer buffer) throws IOException {
public void append(EntryBuffers.RegionEntryBuffer buffer)
throws IOException {
List<WAL.Entry> entries = buffer.entryBuffer;
if (entries.isEmpty()) {
LOG.warn("got an empty buffer, skipping");
Expand Down Expand Up @@ -86,7 +87,7 @@ public void append(EntryBuffers.RegionEntryBuffer buffer) throws IOException {
public List<Path> close() throws IOException {
boolean isSuccessful = true;
try {
isSuccessful &= finishWriterThreads();
isSuccessful = finishWriterThreads();
} finally {
isSuccessful &= writeRemainingEntryBuffers();
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -131,7 +131,7 @@ void append(RegionEntryBuffer buffer) throws IOException {
public List<Path> close() throws IOException {
boolean isSuccessful = true;
try {
isSuccessful &= finishWriterThreads();
isSuccessful = finishWriterThreads();
} finally {
isSuccessful &= writeRemainingEntryBuffers();
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -29,6 +29,7 @@
import java.util.concurrent.atomic.AtomicLong;

import org.apache.hadoop.fs.Path;
import org.apache.hadoop.hbase.monitoring.MonitoredTask;
import org.apache.hadoop.hbase.util.CancelableProgressable;
import org.apache.hadoop.hbase.util.Threads;
import org.apache.yetus.audience.InterfaceAudience;
Expand Down Expand Up @@ -61,6 +62,8 @@ abstract class OutputSink {
*/
protected final List<Path> splits = new ArrayList<>();

protected MonitoredTask status = null;

/**
* Used when close this output sink.
*/
Expand All @@ -81,6 +84,10 @@ void setReporter(CancelableProgressable reporter) {
this.reporter = reporter;
}

void setStatus(MonitoredTask status) {
this.status = status;
}

/**
* Start the threads that will pump data from the entryBuffers to the output files.
*/
Expand Down Expand Up @@ -117,7 +124,9 @@ boolean finishWriterThreads() throws IOException {
}
}
controller.checkForErrors();
LOG.info("{} split writer threads finished", this.writerThreads.size());
final String msg = this.writerThreads.size() + " split writer threads finished";
LOG.info(msg);
updateStatusWithMsg(msg);
return (!progressFailed);
}

Expand All @@ -132,6 +141,7 @@ long getTotalSkippedEdits() {

/**
* @param buffer A buffer of some number of edits for a given region.
* @throws IOException For any IO errors
*/
abstract void append(EntryBuffers.RegionEntryBuffer buffer) throws IOException;

Expand All @@ -154,6 +164,16 @@ long getTotalSkippedEdits() {
*/
abstract boolean keepRegionEvent(WAL.Entry entry);

/**
* Set status message in {@link MonitoredTask} instance that is set in this OutputSink
* @param msg message to update the status with
*/
protected final void updateStatusWithMsg(String msg) {
if (status != null) {
status.setStatus(msg);
}
}

public static class WriterThread extends Thread {
private volatile boolean shouldStop = false;
private WALSplitter.PipelineController controller;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -28,6 +28,7 @@
import java.util.concurrent.Future;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.hbase.TableName;
import org.apache.hadoop.hbase.monitoring.MonitoredTask;
import org.apache.hadoop.hbase.util.Bytes;
import org.apache.hadoop.io.MultipleIOException;
import org.apache.yetus.audience.InterfaceAudience;
Expand All @@ -54,7 +55,8 @@ public RecoveredEditsOutputSink(WALSplitter walSplitter,
}

@Override
public void append(EntryBuffers.RegionEntryBuffer buffer) throws IOException {
public void append(EntryBuffers.RegionEntryBuffer buffer)
throws IOException {
List<WAL.Entry> entries = buffer.entryBuffer;
if (entries.isEmpty()) {
LOG.warn("got an empty buffer, skipping");
Expand All @@ -74,7 +76,7 @@ public void append(EntryBuffers.RegionEntryBuffer buffer) throws IOException {
* @return null if this region shouldn't output any logs
*/
private RecoveredEditsWriter getRecoveredEditsWriter(TableName tableName, byte[] region,
long seqId) throws IOException {
long seqId) throws IOException {
RecoveredEditsWriter ret = writers.get(Bytes.toString(region));
if (ret != null) {
return ret;
Expand All @@ -92,7 +94,7 @@ private RecoveredEditsWriter getRecoveredEditsWriter(TableName tableName, byte[]
public List<Path> close() throws IOException {
boolean isSuccessful = true;
try {
isSuccessful &= finishWriterThreads();
isSuccessful = finishWriterThreads();
} finally {
isSuccessful &= closeWriters();
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -269,14 +269,15 @@ boolean splitLogFile(FileStatus logfile, CancelableProgressable reporter) throws

status = TaskMonitor.get().createStatus(
"Splitting log file " + logfile.getPath() + "into a temporary staging area.");
status.enableStatusJournal(true);
Reader logFileReader = null;
this.fileBeingSplit = logfile;
long startTS = EnvironmentEdgeManager.currentTime();
try {
long logLength = logfile.getLen();
LOG.info("Splitting WAL={}, size={} ({} bytes)", logPath, StringUtils.humanSize(logLength),
logLength);
status.setStatus("Opening log file");
status.setStatus("Opening log file " + logPath);
if (reporter != null && !reporter.progress()) {
progressFailed = true;
return false;
Expand All @@ -291,6 +292,7 @@ boolean splitLogFile(FileStatus logfile, CancelableProgressable reporter) throws
int numOpenedFilesBeforeReporting = conf.getInt("hbase.splitlog.report.openedfiles", 3);
int numOpenedFilesLastCheck = 0;
outputSink.setReporter(reporter);
outputSink.setStatus(status);
outputSink.startWriterThreads();
outputSinkStarted = true;
Entry entry;
Expand Down Expand Up @@ -375,7 +377,9 @@ boolean splitLogFile(FileStatus logfile, CancelableProgressable reporter) throws
e = e instanceof RemoteException ? ((RemoteException) e).unwrapRemoteException() : e;
throw e;
} finally {
LOG.debug("Finishing writing output logs and closing down");
final String log = "Finishing writing output logs and closing down";
LOG.debug(log);
status.setStatus(log);
try {
if (null != logFileReader) {
logFileReader.close();
Expand All @@ -400,6 +404,10 @@ boolean splitLogFile(FileStatus logfile, CancelableProgressable reporter) throws
", corrupted=" + isCorrupted + ", progress failed=" + progressFailed;
LOG.info(msg);
status.markComplete(msg);
if (LOG.isDebugEnabled()) {
LOG.debug("WAL split completed for {} , Journal Log: {}", logPath,
status.prettyPrintJournal());
}
}
}
return !progressFailed;
Expand Down

0 comments on commit 3c31981

Please sign in to comment.