ebulk: new features and fixes

- minor fixes in storage inputs
- retry on failed connections
- resume split operation for download
- control on first ingestion / partial ingestion
- report messages when storage ingestion finishes
- better control on exception and messages
parent 1e8783af
...@@ -151,13 +151,19 @@ function updateConfigFile { ...@@ -151,13 +151,19 @@ function updateConfigFile {
echo echo
OPTION="" OPTION=""
if [ "$ADVANCED" = true ] ; then if [ "$ADVANCED" = true ] ; then
echo "[INFO] If you want to edit the configuration file by yourself (advanced) please type CONFIG, otherwise press enter to continue." echo "[INFO] If you want to edit the configuration file by yourself (advanced) please type YES, otherwise press enter to continue."
read -e OPTION echo -n "** Open configuration file? ('YES' to confirm): "
if [ "$OPTION" = "CONFIG" ] ; then read -e OPTION
if [ "$OPTION" = "YES" ] ; then
CUSTOM=true CUSTOM=true
fi fi
fi fi
if [ "$OPTION" != "CONFIG" ] ; then if [ "$OPTION" != "YES" ] ; then
if [ "$ADVANCED" = true ] ; then
echo
echo "** Please enter the information needed to configure the storage **"
echo
fi
$PARAMETER_FUNCTION $PARAMETER_FUNCTION
if [ "$ADVANCED" = false ] ; then if [ "$ADVANCED" = false ] ; then
echo "** If you need a more advanced storage configuration, you can run the tool with the parameter --advanced **" echo "** If you need a more advanced storage configuration, you can run the tool with the parameter --advanced **"
...@@ -231,7 +237,7 @@ function runProcess { ...@@ -231,7 +237,7 @@ function runProcess {
if [ ! -d $LOG_DIR ]; then if [ ! -d $LOG_DIR ]; then
mkdir $LOG_DIR 2>/dev/null mkdir $LOG_DIR 2>/dev/null
fi fi
if [ -z "$RESUME_STORAGE_INGESTION" ]; then if [ "$RESTART_STORAGE_INGESTION" ]; then
rm -f ${DIFF_FILE} 2>/dev/null rm -f ${DIFF_FILE} 2>/dev/null
fi fi
$embulk run -I $TOOL_PATH/embulk-wendelin-dataset-tool/lib $FILE $DIFF_COMMAND 2> "$LOG_DIR/error.log" || { $embulk run -I $TOOL_PATH/embulk-wendelin-dataset-tool/lib $FILE $DIFF_COMMAND 2> "$LOG_DIR/error.log" || {
...@@ -322,9 +328,7 @@ function checkStoragePlugin { ...@@ -322,9 +328,7 @@ function checkStoragePlugin {
if [ "$STORAGE_GEM" != "" ] ; then if [ "$STORAGE_GEM" != "" ] ; then
echo -n "[INFO] Checking if '$STORAGE' plugin is installed... " echo -n "[INFO] Checking if '$STORAGE' plugin is installed... "
if [ "$CUSTOM" = false ] ; then if [ "$CUSTOM" = false ] ; then
sleep 1
echo -e "${GREEN}OK${NC}" echo -e "${GREEN}OK${NC}"
sleep 1
else else
$embulk gem list 2>/dev/null | grep -q "$STORAGE_GEM" 2>/dev/null $embulk gem list 2>/dev/null | grep -q "$STORAGE_GEM" 2>/dev/null
if [ $? == 0 ]; then if [ $? == 0 ]; then
...@@ -386,12 +390,15 @@ function askFTPparameters { ...@@ -386,12 +390,15 @@ function askFTPparameters {
echo -e "${ORANGE}[ERROR] Empty host.${NC}" echo -e "${ORANGE}[ERROR] Empty host.${NC}"
exit exit
fi fi
FTP_HOST="${FTP_HOST/ftp:\/\//$empty}"
if [[ $FTP_HOST == *"/"* ]]; then
echo -e "${ORANGE}[ERROR] Please, enter only the ftp host, without '/' or path. Path will be requested after.${NC}"
exit
fi
echo "Path prefix:" echo "Path prefix:"
echo "* (e.g. /mydata/sample/dir/) *" echo "* (e.g. /mydata/sample/dir/) *"
read -e FTP_PATH read -e FTP_PATH
if [ "$FTP_PATH" = "" ] ; then FTP_PATH="/$FTP_PATH"
FTP_PATH="/"
fi
echo "FTP User:" echo "FTP User:"
echo "* you can leave this input empty and anonymous authentication will be used *" echo "* you can leave this input empty and anonymous authentication will be used *"
read -e FTP_USER read -e FTP_USER
...@@ -493,7 +500,7 @@ while [ "$1" != "" ]; do ...@@ -493,7 +500,7 @@ while [ "$1" != "" ]; do
;; ;;
-a | --advanced ) ADVANCED=true -a | --advanced ) ADVANCED=true
;; ;;
-rs | --resume ) RESUME_STORAGE_INGESTION=true -rs | --restart ) RESTART_STORAGE_INGESTION=true
;; ;;
-dc | --discard-changes ) DISCARD_CHANGES=true -dc | --discard-changes ) DISCARD_CHANGES=true
;; ;;
......
...@@ -9,9 +9,12 @@ class DatasetUtils ...@@ -9,9 +9,12 @@ class DatasetUtils
DATASET_TEMP_REPORT_FILE = ".temp-dataset-task-report" DATASET_TEMP_REPORT_FILE = ".temp-dataset-task-report"
DATASET_COMPLETED_FILE = ".dataset-completed" DATASET_COMPLETED_FILE = ".dataset-completed"
RESUME_OPERATION_FILE = ".resume-operation" RESUME_OPERATION_FILE = ".resume-operation"
INITIAL_INGESTION_FILE = ".initial-ingestion" PARTIAL_INGESTION_FILE = ".partial-ingestion"
STAGED_FILE = ".staged" STAGED_FILE = ".staged"
DISCARD_CHANGES_FILE = ".discard-changes" DISCARD_CHANGES_FILE = ".discard-changes"
SPLIT_FILE = ".split-operation"
SPLIT_CONTROL_FILE = ".control-split-operation"
FIRST_INGESTION_FILE = ".first-ingestion"
RUN_DONE = "done" RUN_DONE = "done"
RUN_ERROR = "error" RUN_ERROR = "error"
...@@ -19,6 +22,7 @@ class DatasetUtils ...@@ -19,6 +22,7 @@ class DatasetUtils
DELETE = "DELETE" DELETE = "DELETE"
RENAME = "RENAME" RENAME = "RENAME"
INGESTION = "ingestion" INGESTION = "ingestion"
DOWNLOAD = "download"
ADD = "add" ADD = "add"
REMOVE = "remove" REMOVE = "remove"
STATUS_NEW = "new" STATUS_NEW = "new"
...@@ -45,15 +49,18 @@ class DatasetUtils ...@@ -45,15 +49,18 @@ class DatasetUtils
DATE_FORMAT = "%Y-%m-%d-%H-%M-%S" DATE_FORMAT = "%Y-%m-%d-%H-%M-%S"
def initialize(data_set_directory) def initialize(data_set_directory)
@data_set_directory = data_set_directory @data_set_directory = appendSlashTo(data_set_directory)
@logger = LogManager.instance() @logger = LogManager.instance()
@task_report_file = @data_set_directory + DATASET_REPORT_FILE @task_report_file = @data_set_directory + DATASET_REPORT_FILE
@temp_report_file = @data_set_directory + DATASET_TEMP_REPORT_FILE @temp_report_file = @data_set_directory + DATASET_TEMP_REPORT_FILE
@completed_file = @data_set_directory + DATASET_COMPLETED_FILE @completed_file = @data_set_directory + DATASET_COMPLETED_FILE
@resume_operation_file = @data_set_directory + RESUME_OPERATION_FILE @resume_operation_file = @data_set_directory + RESUME_OPERATION_FILE
@initial_ingestion_file = @data_set_directory + INITIAL_INGESTION_FILE @partial_ingestion_file = @data_set_directory + PARTIAL_INGESTION_FILE
@first_ingestion_file = @data_set_directory + FIRST_INGESTION_FILE
@staged_file = @data_set_directory + STAGED_FILE @staged_file = @data_set_directory + STAGED_FILE
@discard_changes_file = @data_set_directory + DISCARD_CHANGES_FILE @discard_changes_file = @data_set_directory + DISCARD_CHANGES_FILE
@split_file = @data_set_directory + SPLIT_FILE
@split_control_file = @data_set_directory + SPLIT_CONTROL_FILE
end end
def getLocalPaths(paths) def getLocalPaths(paths)
...@@ -207,7 +214,7 @@ class DatasetUtils ...@@ -207,7 +214,7 @@ class DatasetUtils
end end
end end
def showTaskErrors(failed_tasks) def showTaskErrors(failed_tasks, storage=FALSE)
puts puts
@logger.error("The following files could not be processed. Please check the details in the log file: " + @logger.getLogPath(), print=TRUE) @logger.error("The following files could not be processed. Please check the details in the log file: " + @logger.getLogPath(), print=TRUE)
if failed_tasks.length > 15 if failed_tasks.length > 15
...@@ -217,10 +224,77 @@ class DatasetUtils ...@@ -217,10 +224,77 @@ class DatasetUtils
else else
@logger.error(failed_tasks, print=TRUE) @logger.error(failed_tasks, print=TRUE)
end end
@logger.info("You can retry the operation for those files.", print=TRUE) if storage
puts
@logger.info("You can retry the operation for those files by running your command again with parameter --restart .", print=TRUE)
else
@logger.info("You can retry the operation for those files.", print=TRUE)
end
puts puts
end end
def splitOperationControlFileExist(reference)
file_reference = reference.gsub('/', '__')
return File.exist?(@split_control_file + "__" + file_reference)
end
def createSplitOperationControlFile(reference)
file_reference = reference.gsub('/', '__')
File.open(@split_control_file + "__" + file_reference, 'w') {}
end
def deleteSplitOperationControlFile(reference=nil)
if reference.nil?
Dir.glob("#{@split_control_file}*").each { |file| File.delete(file) }
else
file_reference = reference.gsub('/', '__')
File.delete(@split_control_file + "__" + file_reference) if File.exist?(@split_control_file + "__" + file_reference)
end
end
def deleteSplitOperationFile(reference=nil)
if reference.nil?
Dir.glob("#{@split_file}*").each { |file| File.delete(file) }
else
file_reference = reference.gsub('/', '__')
File.delete(@split_file + "__" + file_reference) if File.exist?(@split_file + "__" + file_reference)
end
end
def saveSplitOperation(operation, reference, eof, hash, chunk_size)
file_reference = reference.gsub('/', '__')
record = [operation, reference, eof, hash, Integer(chunk_size)].join(RECORD_SEPARATOR)
File.open(@split_file + "__" + file_reference, 'w') { |file| file.puts(record) }
deleteSplitOperationControlFile(reference)
end
def splitOperationFileExist(reference)
file_reference = reference.gsub('/', '__')
return File.exist?(@split_file + "__" + file_reference)
end
def getLastSplitOperation(operation, reference, hash, chunk_size)
file_reference = reference.gsub('/', '__')
return 0 if not File.exist?(@split_file + "__" + file_reference)
record = File.open(@split_file + "__" + file_reference).read.chomp.split(RECORD_SEPARATOR)
if record[0] == operation && record[1] == reference && record[3] == hash && record[4] == Integer(chunk_size).to_s && record[2] != EOF
# discard if user interrupted (ctrl+c) the operation
if splitOperationControlFileExist(file_reference)
@logger.warn("Previous split operation attempt for file #{reference} was interrupt by user (aborted tool execution), it will be restarted.", print=TRUE)
deleteSplitOperationFile(file_reference)
deleteSplitOperationControlFile(file_reference)
return 0
end
createSplitOperationControlFile(file_reference)
return record[2].to_i
end
return 0
rescue Exception => e
@logger.error("An error occurred in getLastSplitOperation method:" + e.to_s)
@logger.error(e.backtrace)
return 0
end
def deleteDiscardChangesFile() def deleteDiscardChangesFile()
File.delete(@discard_changes_file) if File.exist?(@discard_changes_file) File.delete(@discard_changes_file) if File.exist?(@discard_changes_file)
end end
...@@ -255,16 +329,28 @@ class DatasetUtils ...@@ -255,16 +329,28 @@ class DatasetUtils
end end
end end
def deleteInitialIngestionFile() def deleteFirstIngestionFile()
File.delete(@initial_ingestion_file) if File.exist?(@initial_ingestion_file) File.delete(@first_ingestion_file) if File.exist?(@first_ingestion_file)
end
def createFirstIngestionFile()
File.open(@first_ingestion_file, 'w') {}
end
def firstIngestionFileExist()
return File.exist?(@first_ingestion_file)
end
def deletePartialIngestionFile()
File.delete(@partial_ingestion_file) if File.exist?(@partial_ingestion_file)
end end
def createInitialIngestionFile() def createPartialIngestionFile()
File.open(@initial_ingestion_file, 'w') {} File.open(@partial_ingestion_file, 'w') {}
end end
def initialIngestionFileExist() def partialIngestionFileExist()
return File.exist?(@initial_ingestion_file) return File.exist?(@partial_ingestion_file)
end end
def stagedFileExist() def stagedFileExist()
......
...@@ -22,7 +22,7 @@ module Embulk ...@@ -22,7 +22,7 @@ module Embulk
] ]
def self.status(task, push=FALSE) def self.status(task, push=FALSE)
partial_ingestion = @dataset_utils.initialIngestionFileExist() partial_ingestion = @dataset_utils.partialIngestionFileExist()
staged_changes, untracked_changes = @dataset_utils.getLocalChanges(task['paths'], task['data_set'], staged=TRUE, partial_ingestion=partial_ingestion) staged_changes, untracked_changes = @dataset_utils.getLocalChanges(task['paths'], task['data_set'], staged=TRUE, partial_ingestion=partial_ingestion)
staged = (not staged_changes.empty?) staged = (not staged_changes.empty?)
task['paths'] = staged ? staged_changes : untracked_changes task['paths'] = staged ? staged_changes : untracked_changes
...@@ -96,7 +96,7 @@ module Embulk ...@@ -96,7 +96,7 @@ module Embulk
@status = @status == "" ? FALSE : @status @status = @status == "" ? FALSE : @status
@dataset_utils.deleteDiscardChangesFile() @dataset_utils.deleteDiscardChangesFile()
if @status if @status
if not @dataset_utils.initialIngestionFileExist() if not @dataset_utils.partialIngestionFileExist()
if not @dataset_utils.reportFileExist() if not @dataset_utils.reportFileExist()
puts puts
@logger.error("The dataset directory does not contain a valid dataset.", print=TRUE) @logger.error("The dataset directory does not contain a valid dataset.", print=TRUE)
...@@ -123,11 +123,12 @@ module Embulk ...@@ -123,11 +123,12 @@ module Embulk
@logger.abortExecution() @logger.abortExecution()
end end
task['data_streams'] = data_stream_dict["result"] task['data_streams'] = data_stream_dict["result"]
first_ingestion = task['data_streams'].length == 0
if not @dataset_utils.reportFileExist() if not @dataset_utils.reportFileExist()
@dataset_utils.createInitialIngestionFile() @dataset_utils.createPartialIngestionFile()
@dataset_utils.createFirstIngestionFile() if first_ingestion
else else
if not @dataset_utils.initialIngestionFileExist() if not @dataset_utils.partialIngestionFileExist()
@logger.info("Checking local dataset...", print=TRUE) @logger.info("Checking local dataset...", print=TRUE)
if not @dataset_utils.reportUpToDate(data_stream_dict, @data_set) if not @dataset_utils.reportUpToDate(data_stream_dict, @data_set)
puts puts
...@@ -188,12 +189,17 @@ module Embulk ...@@ -188,12 +189,17 @@ module Embulk
task_reports = yield(task, columns, count) task_reports = yield(task, columns, count)
next_config_diff = task_reports.map{|hash| hash[DatasetUtils::RUN_DONE]}.flatten.compact next_config_diff = task_reports.map{|hash| hash[DatasetUtils::RUN_DONE]}.flatten.compact
@dataset_utils.showTaskReport(next_config_diff) @dataset_utils.showTaskReport(next_config_diff)
element_output = @dataset_utils.initialIngestionFileExist() ? "new file" : "change" element_output = @dataset_utils.partialIngestionFileExist() ? "new file" : "change"
@logger.info("#{next_config_diff.length} #{element_output}(s) ingested.", print=TRUE) @logger.info("#{next_config_diff.length} #{element_output}(s) ingested.", print=TRUE)
if(next_config_diff.length == count) if(next_config_diff.length == count)
@logger.info("Dataset successfully ingested.", print=TRUE) @logger.info("Dataset successfully ingested.", print=TRUE)
@wendelin.increaseDatasetVersion(@data_set) @wendelin.increaseDatasetVersion(@data_set)
@dataset_utils.deleteStagedFile() @dataset_utils.deleteStagedFile()
if @dataset_utils.firstIngestionFileExist()
@dataset_utils.createCompletedFile()
@dataset_utils.deletePartialIngestionFile()
@dataset_utils.deleteFirstIngestionFile()
end
else else
failed_tasks = task_reports.map{|hash| hash[DatasetUtils::RUN_ERROR] || hash[DatasetUtils::RUN_ABORTED] }.flatten.compact failed_tasks = task_reports.map{|hash| hash[DatasetUtils::RUN_ERROR] || hash[DatasetUtils::RUN_ABORTED] }.flatten.compact
@dataset_utils.showTaskErrors(failed_tasks) @dataset_utils.showTaskErrors(failed_tasks)
...@@ -221,6 +227,7 @@ module Embulk ...@@ -221,6 +227,7 @@ module Embulk
hash = file_dict["hash"] hash = file_dict["hash"]
delete = hash == DatasetUtils::DELETE delete = hash == DatasetUtils::DELETE
rename = file_dict["status"] == DatasetUtils::STATUS_RENAMED rename = file_dict["status"] == DatasetUtils::STATUS_RENAMED
split = FALSE
if size == "" and hash == "" #new file if size == "" and hash == "" #new file
size = File.size(path) size = File.size(path)
hash = @dataset_utils.getHash(path) hash = @dataset_utils.getHash(path)
...@@ -229,10 +236,17 @@ module Embulk ...@@ -229,10 +236,17 @@ module Embulk
filename, extension, reference = @dataset_utils.getPathInfo(path, @dataset) filename, extension, reference = @dataset_utils.getPathInfo(path, @dataset)
operation = rename ? DatasetUtils::RENAME : DatasetUtils::INGESTION operation = rename ? DatasetUtils::RENAME : DatasetUtils::INGESTION
@dataset_utils.saveCurrentOperation(operation, reference, new_reference) @dataset_utils.saveCurrentOperation(operation, reference, new_reference)
each_chunk(path, filename, extension, size, hash, schema[1..-1].map{|elm| elm.name}, @chunk_size, delete, new_reference) do |entry| resume_split = @dataset_utils.splitOperationFileExist(reference) ? @dataset_utils.getLastSplitOperation(operation, reference, hash, @chunk_size) : 0
each_chunk(path, filename, extension, size, hash, schema[1..-1].map{|elm| elm.name}, @chunk_size, delete, new_reference, resume_split) do |entry|
@dataset_utils.createSplitOperationControlFile(reference) if split
@page_builder.add(entry) @page_builder.add(entry)
if ! delete && ! rename && entry[5] != ""
split = TRUE
@dataset_utils.saveSplitOperation(operation, reference, entry[5], hash, @chunk_size)
end
end end
@page_builder.finish @page_builder.finish
@dataset_utils.deleteSplitOperationFile(reference) if split
rescue java.lang.OutOfMemoryError rescue java.lang.OutOfMemoryError
@logger.logOutOfMemoryError(path) @logger.logOutOfMemoryError(path)
return_value = DatasetUtils::RUN_ABORTED return_value = DatasetUtils::RUN_ABORTED
...@@ -244,6 +258,7 @@ module Embulk ...@@ -244,6 +258,7 @@ module Embulk
else else
return_value = DatasetUtils::RUN_DONE return_value = DatasetUtils::RUN_DONE
end end
@dataset_utils.deleteSplitOperationControlFile(reference) if return_value != DatasetUtils::RUN_DONE
if return_value == DatasetUtils::RUN_DONE if return_value == DatasetUtils::RUN_DONE
if delete if delete
if @dataset_utils.reportFileExist() if @dataset_utils.reportFileExist()
...@@ -261,7 +276,7 @@ module Embulk ...@@ -261,7 +276,7 @@ module Embulk
private private
def each_chunk(path, filename, extension, size, hash, fields, chunk_size=DatasetUtils::CHUNK_SIZE, delete=FALSE, new_reference=FALSE) def each_chunk(path, filename, extension, size, hash, fields, chunk_size=DatasetUtils::CHUNK_SIZE, delete=FALSE, new_reference=FALSE, resume_split=0)
if delete if delete
File.delete(path) if File.exist?(path) File.delete(path) if File.exist?(path)
values = [@supplier, @dataset, filename, extension, "", DatasetUtils::DELETE, "", ""] values = [@supplier, @dataset, filename, extension, "", DatasetUtils::DELETE, "", ""]
...@@ -278,22 +293,28 @@ module Embulk ...@@ -278,22 +293,28 @@ module Embulk
while true while true
data = next_byte data = next_byte
if not next_byte if not next_byte
if first # this means this is an empty file if first # this means this is an empty file
values = [@supplier, @dataset, filename, extension, "", "", size, hash] values = [@supplier, @dataset, filename, extension, "", "", size, hash]
yield(values) yield(values)
end end
break break
end
if resume_split > 0
@logger.info("Resuming interrupted split ingestion", print=TRUE)
first = FALSE
file_object.seek(chunk_size*resume_split+resume_split) # ignore ingested part
next_byte = file_object.read(1)
data = next_byte
npart = resume_split
resume_split = 0
end end
data += file_object.read(chunk_size) data += file_object.read(chunk_size)
next_byte = file_object.read(1) next_byte = file_object.read(1)
if not next_byte if not next_byte
eof = DatasetUtils::EOF eof = first ? "" : DatasetUtils::EOF
if first # this means that the whole file will be ingested at once (not split)
eof = ""
end
else else
npart += 1 npart += 1
eof = npart.to_s.rjust(3, "0") eof = npart.to_s.rjust(3, "0")
end end
content = Base64.encode64(data) content = Base64.encode64(data)
values = [@supplier, @dataset, filename, extension, content, eof, size, hash] values = [@supplier, @dataset, filename, extension, content, eof, size, hash]
......
...@@ -68,6 +68,8 @@ module Embulk ...@@ -68,6 +68,8 @@ module Embulk
@logger.info("Your downloaded dataset is already up to date.", print=TRUE) @logger.info("Your downloaded dataset is already up to date.", print=TRUE)
end end
when DOWNLOAD when DOWNLOAD
@dataset_utils.deleteSplitOperationControlFile()
@dataset_utils.deleteSplitOperationFile()
@logger.info("Checking remote files and posible local conflicts...", print=TRUE) @logger.info("Checking remote files and posible local conflicts...", print=TRUE)
self.warnConflicts(task['data_streams'], task['data_set']) self.warnConflicts(task['data_streams'], task['data_set'])
@dataset_utils.deleteCompletedFile() @dataset_utils.deleteCompletedFile()
...@@ -112,7 +114,7 @@ module Embulk ...@@ -112,7 +114,7 @@ module Embulk
task['data_set_directory'] = @dataset_utils.appendSlashTo(@output_path) task['data_set_directory'] = @dataset_utils.appendSlashTo(@output_path)
@data_set_directory = task['data_set_directory'] @data_set_directory = task['data_set_directory']
@dataset_utils = DatasetUtils.new(@data_set_directory) @dataset_utils = DatasetUtils.new(@data_set_directory)
if @dataset_utils.reportFileExist() && @dataset_utils.completedFileExist() && @dataset_utils.discardChangesFileExist() && ! @dataset_utils.initialIngestionFileExist() if @dataset_utils.reportFileExist() && @dataset_utils.completedFileExist() && @dataset_utils.discardChangesFileExist() && ! @dataset_utils.partialIngestionFileExist()
task['discard_changes'] = @dataset_utils.discardChangesFileExist() task['discard_changes'] = @dataset_utils.discardChangesFileExist()
local_changes = @dataset_utils.getRemoteFileListForDiscardLocalChanges([], @data_set, check_changes=TRUE) local_changes = @dataset_utils.getRemoteFileListForDiscardLocalChanges([], @data_set, check_changes=TRUE)
if local_changes.empty? if local_changes.empty?
...@@ -159,7 +161,7 @@ module Embulk ...@@ -159,7 +161,7 @@ module Embulk
puts puts
self.askUserForAction(task, action=UPDATE) self.askUserForAction(task, action=UPDATE)
end end
elsif not @dataset_utils.initialIngestionFileExist() elsif not @dataset_utils.partialIngestionFileExist()
puts puts
@logger.info("There was a previous attempt to download this dataset but it did not finish successfully.", print=TRUE) @logger.info("There was a previous attempt to download this dataset but it did not finish successfully.", print=TRUE)
@logger.info("What do you want to do?", print=TRUE) @logger.info("What do you want to do?", print=TRUE)
...@@ -195,7 +197,7 @@ module Embulk ...@@ -195,7 +197,7 @@ module Embulk
end end
@dataset_utils.createReportFile() @dataset_utils.createReportFile()
end end
@dataset_utils.deleteInitialIngestionFile() @dataset_utils.deletePartialIngestionFile()
@dataset_utils.deleteDiscardChangesFile() @dataset_utils.deleteDiscardChangesFile()
columns = [ columns = [
Column.new(0, "reference", :string), Column.new(0, "reference", :string),
...@@ -254,34 +256,42 @@ module Embulk ...@@ -254,34 +256,42 @@ module Embulk
def run def run
data_stream = task['data_streams'][@index] data_stream = task['data_streams'][@index]
id = data_stream["id"] id = data_stream["id"]
ref = data_stream["reference"] reference = data_stream["reference"]
size = data_stream["size"] size = data_stream["size"]
hash = data_stream["hash"] hash = data_stream["hash"]
renamed = data_stream["status"] == DatasetUtils::STATUS_RENAMED renamed = data_stream["status"] == DatasetUtils::STATUS_RENAMED
deleted = hash.to_s == DatasetUtils::DELETE deleted = hash.to_s == DatasetUtils::DELETE
begin begin
if deleted if deleted
entry = [ref, "", @data_set, DatasetUtils::DELETE, renamed] entry = [reference, "", @data_set, DatasetUtils::DELETE, renamed]
page_builder.add(entry) page_builder.add(entry)
elsif renamed elsif renamed
new_reference = data_stream["new_reference"] new_reference = data_stream["new_reference"]
entry = [ref, new_reference, @data_set, TRUE, renamed] entry = [reference, new_reference, @data_set, TRUE, renamed]
page_builder.add(entry) page_builder.add(entry)
else else
@logger.info("Discarding local change on '#{data_stream["path"]}'", print=TRUE) if task['discard_changes'] @logger.info("Discarding local change on '#{data_stream["path"]}'", print=TRUE) if task['discard_changes']
@logger.info("Getting content from remote #{ref}", print=TRUE) @logger.info("Getting content from remote #{reference}", print=TRUE)
n_chunk = 0 @logger.info("Downloading...", print=TRUE)
@wendelin.eachDataStreamContentChunk(id, @chunk_size) do |chunk| resume_split = @dataset_utils.splitOperationFileExist(reference) ? @dataset_utils.getLastSplitOperation(DatasetUtils::DOWNLOAD, reference, hash, @chunk_size) : 0
n_chunk = resume_split == 0 ? 0 : resume_split+1
split = n_chunk > 0
@logger.info("Resuming interrupted split download...", print=TRUE) if split
@wendelin.eachDataStreamContentChunk(id, @chunk_size, n_chunk) do |chunk|
content = chunk.nil? || chunk.empty? ? "" : Base64.encode64(chunk) content = chunk.nil? || chunk.empty? ? "" : Base64.encode64(chunk)
begin_of_file = n_chunk == 0 begin_of_file = n_chunk == 0
entry = [ref, content, @data_set, begin_of_file, renamed] split = n_chunk > 0
@dataset_utils.createSplitOperationControlFile(reference) if split
entry = [reference, content, @data_set, begin_of_file, renamed]
page_builder.add(entry) page_builder.add(entry)
@dataset_utils.saveSplitOperation(DatasetUtils::DOWNLOAD, reference, n_chunk, hash, @chunk_size) if split
n_chunk += 1 n_chunk += 1
end end
end end
page_builder.finish page_builder.finish
@dataset_utils.deleteSplitOperationFile(reference) if split
rescue java.lang.OutOfMemoryError rescue java.lang.OutOfMemoryError
@logger.logOutOfMemoryError(ref) @logger.logOutOfMemoryError(reference)
return_value = DatasetUtils::RUN_ABORTED return_value = DatasetUtils::RUN_ABORTED
rescue Exception => e rescue Exception => e
@logger.error(e.to_s, print=TRUE) @logger.error(e.to_s, print=TRUE)
...@@ -291,14 +301,15 @@ module Embulk ...@@ -291,14 +301,15 @@ module Embulk
else else
return_value = DatasetUtils::RUN_DONE return_value = DatasetUtils::RUN_DONE
end end
@dataset_utils.deleteSplitOperationControlFile(reference) if return_value != DatasetUtils::RUN_DONE
if return_value == DatasetUtils::RUN_DONE if return_value == DatasetUtils::RUN_DONE
if deleted if deleted
@dataset_utils.deleteFromReport(ref, return_value) @dataset_utils.deleteFromReport(reference, return_value)
else else
@dataset_utils.addToReport(ref, return_value, size, hash, task['data_set'], new_reference) @dataset_utils.addToReport(reference, return_value, size, hash, task['data_set'], new_reference)
end end
end end
return {return_value => ref} return {return_value => reference}
end end
end end
end end
......
...@@ -18,18 +18,31 @@ module Embulk ...@@ -18,18 +18,31 @@ module Embulk
"data_set" => config.param("data_set", :string, default: nil), "data_set" => config.param("data_set", :string, default: nil),
"erp5_base_url" => config.param("erp5_base_url", :string, default: nil) "erp5_base_url" => config.param("erp5_base_url", :string, default: nil)
} }
storage_ingestion = ! task["type_input"]
task_reports = yield(task) task_reports = yield(task)
next_config_diff = {} next_config_diff = {}
@logger = LogManager.instance() @logger = LogManager.instance()
@dataset_utils = DatasetUtils.new(Dir.pwd)
if task_reports.length > 0 if task_reports.length > 0
@logger.info("Your ingested files will be available in the site in a few minutes. Thank for your patience.", print=TRUE) @logger.info("Your ingested files will be available in the site in a few minutes. Thank for your patience.", print=TRUE)
# if ingestion was done from a storage different than filesystem, increase dataset version if storage_ingestion
if not task["type_input"] and task["data_set"] and task["erp5_base_url"] done_tasks = task_reports.map{|hash| hash[DatasetUtils::RUN_DONE]}.flatten.compact
@wendelin = WendelinClient.new(task["erp5_base_url"], task["user"], task["password"]) @dataset_utils.showTaskReport(done_tasks)
@wendelin.increaseDatasetVersion(task["data_set"]) @logger.info("#{done_tasks.length} new file(s) ingested.", print=TRUE)
if(done_tasks.length == count)
@logger.info("Dataset successfully ingested.", print=TRUE)
@wendelin = WendelinClient.new(task["erp5_base_url"], task["user"], task["password"])
@wendelin.increaseDatasetVersion(task["data_set"])
else
failed_tasks = task_reports.map{|hash| hash[DatasetUtils::RUN_ERROR] || hash[DatasetUtils::RUN_ABORTED] }.flatten.compact
@dataset_utils.showTaskErrors(failed_tasks, storage=TRUE)
end
end end
else else
@logger.info("No new files where processed for ingestion.", print=TRUE) @logger.info("No new files found for ingestion.", print=TRUE)
if storage_ingestion
@logger.info("You can restart the ingestion from this storage by running your command again with parameter --restart .", print=TRUE)
end
end end
return next_config_diff return next_config_diff
end end
...@@ -43,11 +56,9 @@ module Embulk ...@@ -43,11 +56,9 @@ module Embulk
@wendelin = WendelinClient.new(@erp5_url, @user, @password) @wendelin = WendelinClient.new(@erp5_url, @user, @password)
end end
def close
end
def add(page) def add(page)
page.each do |record| page.each do |record|
@return_value = DatasetUtils::RUN_ERROR
supplier = (record[0].nil? || record[0].empty?) ? "default" : record[0] supplier = (record[0].nil? || record[0].empty?) ? "default" : record[0]
dataset = (record[1].nil? || record[1].empty?) ? "default" : record[1] dataset = (record[1].nil? || record[1].empty?) ? "default" : record[1]
filename = record[2] filename = record[2]
...@@ -58,25 +69,30 @@ module Embulk ...@@ -58,25 +69,30 @@ module Embulk
hash = record[7] hash = record[7]
begin begin
if eof == DatasetUtils::DELETE if eof == DatasetUtils::DELETE
reference = [dataset, filename, extension].join(DatasetUtils::REFERENCE_SEPARATOR) @reference = [dataset, filename, extension].join(DatasetUtils::REFERENCE_SEPARATOR)
@wendelin.delete(reference) @wendelin.delete(@reference)
elsif eof == DatasetUtils::RENAME elsif eof == DatasetUtils::RENAME
reference = [dataset, filename, extension].join(DatasetUtils::REFERENCE_SEPARATOR) @reference = [dataset, filename, extension].join(DatasetUtils::REFERENCE_SEPARATOR)
@wendelin.rename(reference, record[4].to_s) @wendelin.rename(@reference, record[4].to_s)
else else
reference = [supplier, dataset, filename, extension, eof, size, hash].join(DatasetUtils::REFERENCE_SEPARATOR) @reference = [supplier, dataset, filename, extension, eof, size, hash].join(DatasetUtils::REFERENCE_SEPARATOR)
split = eof != "" split = eof != ""
if not @wendelin.ingest(reference, data_chunk, split) ingestion_response = @wendelin.ingest(@reference, data_chunk, split)
raise "could not ingest" if not ingestion_response["success"]
raise ingestion_response["message"]
end end
end end
rescue Exception => e rescue Exception => e
raise e
@logger.error(e.backtrace) @logger.error(e.backtrace)
raise e
end end
@return_value = DatasetUtils::RUN_DONE
end end
end end
def close
end
def finish def finish
end end
...@@ -84,7 +100,7 @@ module Embulk ...@@ -84,7 +100,7 @@ module Embulk
end end
def commit def commit
task_report = {} task_report = {@return_value => @reference}
return task_report return task_report
end end
end end
......
require_relative '../filelogger' require_relative '../filelogger'
require_relative '../dataset_utils' require_relative '../dataset_utils'
require_relative '../wendelin_client'
class Index class Index
include Singleton include Singleton
...@@ -53,8 +54,10 @@ module Embulk ...@@ -53,8 +54,10 @@ module Embulk
def run(file_input) def run(file_input)
@index = Index.instance().get() @index = Index.instance().get()
@logger = LogManager.instance() @logger = LogManager.instance()
split = FALSE
while file = file_input.next_file while file = file_input.next_file
begin begin
@dataset_utils = DatasetUtils.new(Dir.pwd)
metadata_file = Dir.pwd + METADATA_FILE_NAME metadata_file = Dir.pwd + METADATA_FILE_NAME
metadata = File.open(metadata_file) {|f| f.readline} if File.exist?(metadata_file) metadata = File.open(metadata_file) {|f| f.readline} if File.exist?(metadata_file)
File.delete(metadata_file) if File.exist?(metadata_file) File.delete(metadata_file) if File.exist?(metadata_file)
...@@ -72,43 +75,65 @@ module Embulk ...@@ -72,43 +75,65 @@ module Embulk
filename = "file_from_#{task['input_plugin']}_#{task['date']}" filename = "file_from_#{task['input_plugin']}_#{task['date']}"
extension = @index.to_s.rjust(3, "0") extension = @index.to_s.rjust(3, "0")
end end
each_chunk(file, filename.chomp, extension.chomp, task['chunk_size']) do |record| reference = [task['data_set'], filename, extension.chomp].join(DatasetUtils::REFERENCE_SEPARATOR)
resume_split = @dataset_utils.splitOperationFileExist(reference) ? @dataset_utils.getLastSplitOperation(DatasetUtils::INGESTION, reference, "", task['chunk_size']) : 0
each_chunk(file, filename.chomp, extension.chomp, task['chunk_size'], resume_split) do |record|
@dataset_utils.createSplitOperationControlFile(reference) if split
@page_builder.add(record) @page_builder.add(record)
if record[5] != ""
split = TRUE
@dataset_utils.saveSplitOperation(DatasetUtils::INGESTION, reference, record[5], "", task['chunk_size'])
end
end end
@page_builder.finish @page_builder.finish
Index.instance().increase() Index.instance().increase()
@dataset_utils.deleteSplitOperationFile(reference) if split
rescue java.lang.OutOfMemoryError rescue java.lang.OutOfMemoryError
@logger.logOutOfMemoryError(path) @logger.logOutOfMemoryError(reference)
@dataset_utils.deleteSplitOperationControlFile(reference)
@logger.abortExecution() @logger.abortExecution()
rescue Exception => e rescue Exception => e
@logger.error("An error occurred during file ingestion: " + e.to_s, print=TRUE) @logger.error("An error occurred during file ingestion: " + e.to_s, print=TRUE)
@logger.error(e.backtrace) @logger.error(e.backtrace)
puts "[INFO] For more detailed information, please refer to the log file: " + @logger.getLogPath() puts "[INFO] For more detailed information, please refer to the log file: " + @logger.getLogPath()
@dataset_utils.deleteSplitOperationControlFile(reference)
@logger.abortExecution(error=FALSE) if (e.to_s.include? '401' or e.to_s.include? 'Unauthorized')
raise e if not e.to_s.include? WendelinClient::HTTP_REFERENCE_EXIST
end end
end end
end end
private private
def each_chunk(file, filename, extension, chunk_size=DatasetUtils::CHUNK_SIZE) def each_chunk(stream, filename, extension, chunk_size=DatasetUtils::CHUNK_SIZE, resume_split=0)
npart = 0 npart = 0
next_byte = file.read(1) next_byte = stream.read(1)
first = TRUE first = TRUE
while true while true
data = next_byte data = next_byte
if not next_byte if not next_byte
if first if first
# this means this is an empty file # this means this is an empty stream
values = [task['supplier'], task['data_set'], filename, extension, "", "", "", ""] values = [task['supplier'], task['data_set'], filename, extension, "", "", "", ""]
yield(values) yield(values)
end end
break break
end end
data += file.read(chunk_size) if resume_split > 0
next_byte = file.read(1) @logger.warn("Large files resuming could take some minutes", print=TRUE) if resume_split > 10
@logger.info("Resuming interrupted split ingestion...", print=TRUE)
first = FALSE
discardIngestedPart(stream, resume_split, chunk_size)
next_byte = stream.read(1)
data = next_byte
npart = resume_split
resume_split = 0
end
data += stream.read(chunk_size)
next_byte = stream.read(1)
if not next_byte if not next_byte
eof = DatasetUtils::EOF eof = DatasetUtils::EOF
if first if first
# this means that the whole file will be ingested at once (not split) # this means that the whole stream will be ingested at once (not split)
eof = "" eof = ""
end end
else else
...@@ -122,6 +147,16 @@ module Embulk ...@@ -122,6 +147,16 @@ module Embulk
end end
end end
def discardIngestedPart(stream, resume_split, chunk_size)
read = 0
while read < resume_split
stream.read(chunk_size)
read += 1
sleep 1
end
stream.read(resume_split-1)
end
end end
end end
end end
...@@ -7,6 +7,14 @@ require_relative 'filelogger' ...@@ -7,6 +7,14 @@ require_relative 'filelogger'
# class representing a Wendelin client # class representing a Wendelin client
class WendelinClient class WendelinClient
HTTP_MESSAGE_401 = "Unauthorized access. Please check your user credentials and try again."
HTTP_MESSAGE_5XX = "Internal Server Error: if the error persists, please contact the administrator."
HTTP_MESSAGE_OTHER = "Sorry, an error ocurred. If the error persists, please contact the administrator."
HTTP_MESSAGE_NOT_2XX = "HTTP-NOT-OK"
HTTP_MESSAGE_EXCEPTION = "HTTP-ERROR"
HTTP_MEMORY_ERROR = "MEMORY-ERROR"
HTTP_REFERENCE_EXIST = "REFERENCE-EXIST"
def initialize(erp5_url, user, password) def initialize(erp5_url, user, password)
@erp5_url = erp5_url @erp5_url = erp5_url
@user = user @user = user
...@@ -44,7 +52,7 @@ class WendelinClient ...@@ -44,7 +52,7 @@ class WendelinClient
checkReferenceChars(reference) checkReferenceChars(reference)
uri = URI(URI.escape("#{@erp5_url}/ERP5Site_invalidateIngestionObjects?reference=#{reference}")) uri = URI(URI.escape("#{@erp5_url}/ERP5Site_invalidateIngestionObjects?reference=#{reference}"))
res = handleRequest(uri) res = handleRequest(uri)
if res == FALSE if res["success"] == FALSE
@logger.abortExecution() @logger.abortExecution()
end end
@logger.info("Remote file successfully ingested.", print=TRUE) @logger.info("Remote file successfully ingested.", print=TRUE)
...@@ -56,7 +64,7 @@ class WendelinClient ...@@ -56,7 +64,7 @@ class WendelinClient
checkReferenceChars(new_reference) checkReferenceChars(new_reference)
uri = URI(URI.escape("#{@erp5_url}/ERP5Site_renameIngestion?reference=#{reference}&new_reference=#{new_reference}")) uri = URI(URI.escape("#{@erp5_url}/ERP5Site_renameIngestion?reference=#{reference}&new_reference=#{new_reference}"))
res = handleRequest(uri) res = handleRequest(uri)
if res == FALSE if res["success"] == FALSE
@logger.abortExecution() @logger.abortExecution()
end end
@logger.info("Remote file successfully renamed.", print=TRUE) @logger.info("Remote file successfully renamed.", print=TRUE)
...@@ -67,8 +75,8 @@ class WendelinClient ...@@ -67,8 +75,8 @@ class WendelinClient
@logger.warn("Could not increase data set version because dataset reference is empty.") @logger.warn("Could not increase data set version because dataset reference is empty.")
else else
@logger.info("Increasing dataset version") @logger.info("Increasing dataset version")
uri = URI(URI.escape("#{@erp5_url}/ERP5Site_increaseDatasetVersion?reference=#{reference}"))
begin begin
uri = URI(URI.escape("#{@erp5_url}/ERP5Site_increaseDatasetVersion?reference=#{reference}"))
res = open(uri, http_basic_authentication: [@user, @password]).read res = open(uri, http_basic_authentication: [@user, @password]).read
rescue Exception => e rescue Exception => e
@logger.error("An error occurred while increasing dataset version: " + e.to_s) @logger.error("An error occurred while increasing dataset version: " + e.to_s)
...@@ -87,46 +95,77 @@ class WendelinClient ...@@ -87,46 +95,77 @@ class WendelinClient
@logger.info("There is another ingestion already done for the pair dataset-filename. Reference "\ @logger.info("There is another ingestion already done for the pair dataset-filename. Reference "\
+ reference, print=TRUE) + reference, print=TRUE)
@logger.info("Rename your file or download the full dataset to make local changes.", print=TRUE) @logger.info("Rename your file or download the full dataset to make local changes.", print=TRUE)
return FALSE return {"success"=>FALSE, "message"=>HTTP_REFERENCE_EXIST}
end end
checkReferenceChars(reference)
uri = URI(URI.escape("#{@erp5_url}/ingest?reference=#{reference}")) uri = URI(URI.escape("#{@erp5_url}/ingest?reference=#{reference}"))
n_retry = 0
response = handleRequest(uri, reference, data_chunk) response = handleRequest(uri, reference, data_chunk)
if response == FALSE while response["success"] == FALSE && n_retry < 10
return FALSE if response["message"] != HTTP_MESSAGE_EXCEPTION && response["message"] != HTTP_MESSAGE_NOT_2XX
return response
else
n_retry += 1
@logger.info("Retrying #{n_retry}/10...", print=TRUE)
sleep 30
response = handleRequest(uri, reference, data_chunk)
end
end
if response["success"] == FALSE
return response
end end
@logger.info("Record successfully ingested.", print=TRUE) @logger.info("Record successfully ingested.", print=TRUE)
@last_ingestion = Time.new @last_ingestion = Time.new
return TRUE return {"success"=>TRUE, "message"=>"success"}
end end
def eachDataStreamContentChunk(id, chunk_size) def eachDataStreamContentChunk(id, chunk_size, n_chunk=0)
uri = URI(URI.escape("#{@erp5_url}#{id}/getData")) n_part = n_chunk
@logger.info("Downloading...", print=TRUE) done = FALSE
first = TRUE first = TRUE
res = open(uri, http_basic_authentication: [@user, @password]) { while not done
|content| start_offset = n_part*chunk_size
while true end_offset = n_part*chunk_size+chunk_size
chunk = content.read(chunk_size) uri = URI(URI.escape("#{@erp5_url}getDataStreamChunk?id=#{id}&start_offset=#{start_offset}&end_offset=#{end_offset}"))
if chunk.nil? success = FALSE
if first n_retry = 0
yield chunk while ! success && n_retry < 10
end begin
@logger.info("Done", print=TRUE) res = open(uri, http_basic_authentication: [@user, @password]) {
break |content|
end chunk = content.read()
first = FALSE if chunk.nil? || chunk.empty?
yield chunk if first
yield chunk
end
@logger.info("Done", print=TRUE)
done = TRUE
else
first = FALSE
n_part += 1
yield chunk
end
}
success = TRUE
rescue Exception => e
exception = e
@logger.error("Error downloading data: " + e.to_s, print=TRUE)
@logger.error(e.backtrace)
n_retry += 1
@logger.info("Retrying #{n_retry}/10...", print=TRUE)
sleep 30
end end
} end
end raise exception if not success
end
end
def getDataStreams(data_set_reference) def getDataStreams(data_set_reference)
uri = URI(URI.escape("#{@erp5_url}getDataStreamList?data_set_reference=#{data_set_reference}")) uri = URI(URI.escape("#{@erp5_url}getDataStreamList?data_set_reference=#{data_set_reference}"))
str = handleRequest(uri) response = handleRequest(uri)
if str == FALSE if response["success"] == FALSE
@logger.abortExecution() @logger.abortExecution()
end end
str = response["message"]
if not str.nil? if not str.nil?
str.gsub!(/(\,)(\S)/, "\\1 \\2") str.gsub!(/(\,)(\S)/, "\\1 \\2")
return YAML::load(str) return YAML::load(str)
...@@ -145,11 +184,14 @@ class WendelinClient ...@@ -145,11 +184,14 @@ class WendelinClient
req.set_form_data('data_chunk' => data_chunk) req.set_form_data('data_chunk' => data_chunk)
rescue java.lang.OutOfMemoryError rescue java.lang.OutOfMemoryError
@logger.logOutOfMemoryError(reference) @logger.logOutOfMemoryError(reference)
return FALSE return {"success"=>FALSE, "message"=>HTTP_MEMORY_ERROR}
rescue Exception => e
@logger.error("Error setting form data: " + e.to_s, print=TRUE)
@logger.error(e.backtrace)
return {"success"=>FALSE, "message"=>HTTP_MESSAGE_EXCEPTION}
end end
@logger.info("Sending record:'#{reference}'...", print=TRUE) @logger.info("Sending record:'#{reference}'...", print=TRUE)
end end
begin begin
res = Net::HTTP.start(uri.hostname, uri.port, res = Net::HTTP.start(uri.hostname, uri.port,
:use_ssl => (uri.scheme == 'https'), :use_ssl => (uri.scheme == 'https'),
...@@ -158,25 +200,28 @@ class WendelinClient ...@@ -158,25 +200,28 @@ class WendelinClient
) do |http| ) do |http|
http.request(req) http.request(req)
end end
rescue Exception => e rescue Exception => e
@logger.error("HTTP ERROR: " + e.to_s, print=TRUE) @logger.error("HTTP ERROR: " + e.to_s, print=TRUE)
@logger.error(e.backtrace) @logger.error(e.backtrace)
return FALSE return {"success"=>FALSE, "message"=>HTTP_MESSAGE_EXCEPTION}
else else
if res.kind_of?(Net::HTTPSuccess) # res.code is 2XX if res.kind_of?(Net::HTTPSuccess) # res.code is 2XX
@logger.info("Done") @logger.info("Done")
return res.body return {"success"=>TRUE, "message"=>res.body}
else else
@logger.error("HTTP FAIL - code: #{res.code}", print=TRUE) @logger.error("HTTP FAIL - code: #{res.code}", print=TRUE)
if res.code == '500' or res.code == '502' or res.code == '503' if res.code == '500' or res.code == '502' or res.code == '503'
@logger.error("Internal Server Error: if the error persists, please contact the administrator.", print=TRUE) @logger.error(HTTP_MESSAGE_5XX, print=TRUE)
elsif res.code == '401' elsif res.code == '401'
@logger.error("Unauthorized access. Please check your user credentials and try again.", print=TRUE) @logger.error(HTTP_MESSAGE_401, print=TRUE)
@logger.abortExecution()
elsif res.code == '400'
@logger.error(HTTP_MESSAGE_400, print=TRUE)
@logger.abortExecution() @logger.abortExecution()
else else
@logger.error("Sorry, an error ocurred. If the error persists, please contact the administrator.", print=TRUE) @logger.error(HTTP_MESSAGE_OTHER, print=TRUE)
end end
return FALSE return {"success"=>FALSE, "message"=>HTTP_MESSAGE_NOT_2XX}
end end
end end
end end
......
...@@ -32,6 +32,6 @@ options: ...@@ -32,6 +32,6 @@ options:
-s, --storage <storage> Uses the selected input storage from this set: [http, ftp, s3] -s, --storage <storage> Uses the selected input storage from this set: [http, ftp, s3]
-cs, --custom-storage Allows user to set a new input storage. -cs, --custom-storage Allows user to set a new input storage.
-a, --advanced Allows to edit the Embulk cofiguration file of the input storage -a, --advanced Allows to edit the Embulk cofiguration file of the input storage
-dc, --discard-changes Discards local changes by checking the remote dataset -dc, --discard-changes Discards local changes by checking the remote dataset
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment