updater: Drop the support for BBOTA v1 and v2.
The script support for BBOTA v1 and v2 has been dropped in commit
8fad03e7712082eb880ffaaffb69eb13252ce220 (platform/build).
Bug: 33694730
Test: Apply an incremental with the new updater.
Test: recovery_component_test
Change-Id: I038b1bf8d10f030cab8ec0aa6ee565c5a9545dfd
diff --git a/updater/blockimg.cpp b/updater/blockimg.cpp
index 12ca151..b3fe455 100644
--- a/updater/blockimg.cpp
+++ b/updater/blockimg.cpp
@@ -429,46 +429,11 @@
uint8_t* patch_start;
};
-// Do a source/target load for move/bsdiff/imgdiff in version 1.
-// We expect to parse the remainder of the parameter tokens as:
-//
-// <src_range> <tgt_range>
-//
-// The source range is loaded into the provided buffer, reallocating
-// it to make it larger if necessary.
-
-static int LoadSrcTgtVersion1(CommandParameters& params, RangeSet& tgt, size_t& src_blocks,
- std::vector<uint8_t>& buffer, int fd) {
-
- if (params.cpos + 1 >= params.tokens.size()) {
- LOG(ERROR) << "invalid parameters";
- return -1;
- }
-
- // <src_range>
- RangeSet src = parse_range(params.tokens[params.cpos++]);
-
- // <tgt_range>
- tgt = parse_range(params.tokens[params.cpos++]);
-
- allocate(src.size * BLOCKSIZE, buffer);
- int rc = ReadBlocks(src, buffer, fd);
- src_blocks = src.size;
-
- return rc;
-}
-
// Print the hash in hex for corrupted source blocks (excluding the stashed blocks which is
// handled separately).
static void PrintHashForCorruptedSourceBlocks(const CommandParameters& params,
const std::vector<uint8_t>& buffer) {
LOG(INFO) << "unexpected contents of source blocks in cmd:\n" << params.cmdline;
- if (params.version < 3) {
- // TODO handle version 1,2
- LOG(WARNING) << "version number " << params.version << " is not supported to print hashes";
- return;
- }
-
CHECK(params.tokens[0] == "move" || params.tokens[0] == "bsdiff" ||
params.tokens[0] == "imgdiff");
@@ -896,49 +861,46 @@
}
static int SaveStash(CommandParameters& params, const std::string& base,
- std::vector<uint8_t>& buffer, int fd, bool usehash) {
+ std::vector<uint8_t>& buffer, int fd) {
+ // <stash_id> <src_range>
+ if (params.cpos + 1 >= params.tokens.size()) {
+ LOG(ERROR) << "missing id and/or src range fields in stash command";
+ return -1;
+ }
- // <stash_id> <src_range>
- if (params.cpos + 1 >= params.tokens.size()) {
- LOG(ERROR) << "missing id and/or src range fields in stash command";
- return -1;
- }
- const std::string& id = params.tokens[params.cpos++];
+ const std::string& id = params.tokens[params.cpos++];
+ size_t blocks = 0;
+ if (LoadStash(params, base, id, true, &blocks, buffer, false) == 0) {
+ // Stash file already exists and has expected contents. Do not read from source again, as the
+ // source may have been already overwritten during a previous attempt.
+ return 0;
+ }
- size_t blocks = 0;
- if (usehash && LoadStash(params, base, id, true, &blocks, buffer, false) == 0) {
- // Stash file already exists and has expected contents. Do not
- // read from source again, as the source may have been already
- // overwritten during a previous attempt.
- return 0;
- }
+ RangeSet src = parse_range(params.tokens[params.cpos++]);
- RangeSet src = parse_range(params.tokens[params.cpos++]);
+ allocate(src.size * BLOCKSIZE, buffer);
+ if (ReadBlocks(src, buffer, fd) == -1) {
+ return -1;
+ }
+ blocks = src.size;
+ stash_map[id] = src;
- allocate(src.size * BLOCKSIZE, buffer);
- if (ReadBlocks(src, buffer, fd) == -1) {
- return -1;
- }
- blocks = src.size;
- stash_map[id] = src;
+ if (VerifyBlocks(id, buffer, blocks, true) != 0) {
+ // Source blocks have unexpected contents. If we actually need this data later, this is an
+ // unrecoverable error. However, the command that uses the data may have already completed
+ // previously, so the possible failure will occur during source block verification.
+ LOG(ERROR) << "failed to load source blocks for stash " << id;
+ return 0;
+ }
- if (usehash && VerifyBlocks(id, buffer, blocks, true) != 0) {
- // Source blocks have unexpected contents. If we actually need this
- // data later, this is an unrecoverable error. However, the command
- // that uses the data may have already completed previously, so the
- // possible failure will occur during source block verification.
- LOG(ERROR) << "failed to load source blocks for stash " << id;
- return 0;
- }
+ // In verify mode, we don't need to stash any blocks.
+ if (!params.canwrite) {
+ return 0;
+ }
- // In verify mode, we don't need to stash any blocks.
- if (!params.canwrite && usehash) {
- return 0;
- }
-
- LOG(INFO) << "stashing " << blocks << " blocks to " << id;
- params.stashed += blocks;
- return WriteStash(base, id, blocks, buffer, false, nullptr);
+ LOG(INFO) << "stashing " << blocks << " blocks to " << id;
+ params.stashed += blocks;
+ return WriteStash(base, id, blocks, buffer, false, nullptr);
}
static int FreeStash(const std::string& base, const std::string& id) {
@@ -1061,26 +1023,35 @@
return 0;
}
-// Do a source/target load for move/bsdiff/imgdiff in version 3.
-//
-// Parameters are the same as for LoadSrcTgtVersion2, except for 'onehash', which
-// tells the function whether to expect separate source and targe block hashes, or
-// if they are both the same and only one hash should be expected, and
-// 'isunresumable', which receives a non-zero value if block verification fails in
-// a way that the update cannot be resumed anymore.
-//
-// If the function is unable to load the necessary blocks or their contents don't
-// match the hashes, the return value is -1 and the command should be aborted.
-//
-// If the return value is 1, the command has already been completed according to
-// the contents of the target blocks, and should not be performed again.
-//
-// If the return value is 0, source blocks have expected content and the command
-// can be performed.
-
+/**
+ * Do a source/target load for move/bsdiff/imgdiff in version 3.
+ *
+ * We expect to parse the remainder of the parameter tokens as one of:
+ *
+ * <tgt_range> <src_block_count> <src_range>
+ * (loads data from source image only)
+ *
+ * <tgt_range> <src_block_count> - <[stash_id:stash_range] ...>
+ * (loads data from stashes only)
+ *
+ * <tgt_range> <src_block_count> <src_range> <src_loc> <[stash_id:stash_range] ...>
+ * (loads data from both source image and stashes)
+ *
+ * Parameters are the same as for LoadSrcTgtVersion2, except for 'onehash', which tells the function
+ * whether to expect separate source and targe block hashes, or if they are both the same and only
+ * one hash should be expected, and 'isunresumable', which receives a non-zero value if block
+ * verification fails in a way that the update cannot be resumed anymore.
+ *
+ * If the function is unable to load the necessary blocks or their contents don't match the hashes,
+ * the return value is -1 and the command should be aborted.
+ *
+ * If the return value is 1, the command has already been completed according to the contents of the
+ * target blocks, and should not be performed again.
+ *
+ * If the return value is 0, source blocks have expected content and the command can be performed.
+ */
static int LoadSrcTgtVersion3(CommandParameters& params, RangeSet& tgt, size_t& src_blocks,
- bool onehash, bool& overlap) {
-
+ bool onehash, bool& overlap) {
if (params.cpos >= params.tokens.size()) {
LOG(ERROR) << "missing source hash";
return -1;
@@ -1099,8 +1070,8 @@
tgthash = params.tokens[params.cpos++];
}
- if (LoadSrcTgtVersion2(params, tgt, src_blocks, params.buffer, params.fd,
- params.stashbase, &overlap) == -1) {
+ if (LoadSrcTgtVersion2(params, tgt, src_blocks, params.buffer, params.fd, params.stashbase,
+ &overlap) == -1) {
return -1;
}
@@ -1111,7 +1082,7 @@
}
if (VerifyBlocks(tgthash, tgtbuffer, tgt.size, false) == 0) {
- // Target blocks already have expected content, command should be skipped
+ // Target blocks already have expected content, command should be skipped.
return 1;
}
@@ -1130,13 +1101,13 @@
}
params.stashed += src_blocks;
- // Can be deleted when the write has completed
+ // Can be deleted when the write has completed.
if (!stash_exists) {
params.freestash = srchash;
}
}
- // Source blocks have expected content, command can proceed
+ // Source blocks have expected content, command can proceed.
return 0;
}
@@ -1148,7 +1119,7 @@
return 0;
}
- // Valid source data not available, update cannot be resumed
+ // Valid source data not available, update cannot be resumed.
LOG(ERROR) << "partition has unexpected contents";
PrintHashForCorruptedSourceBlocks(params, params.buffer);
@@ -1158,57 +1129,46 @@
}
static int PerformCommandMove(CommandParameters& params) {
- size_t blocks = 0;
- bool overlap = false;
- int status = 0;
- RangeSet tgt;
+ size_t blocks = 0;
+ bool overlap = false;
+ RangeSet tgt;
+ int status = LoadSrcTgtVersion3(params, tgt, blocks, true, overlap);
- if (params.version == 1) {
- status = LoadSrcTgtVersion1(params, tgt, blocks, params.buffer, params.fd);
- } else if (params.version == 2) {
- status = LoadSrcTgtVersion2(params, tgt, blocks, params.buffer, params.fd,
- params.stashbase, nullptr);
- } else if (params.version >= 3) {
- status = LoadSrcTgtVersion3(params, tgt, blocks, true, overlap);
- }
+ if (status == -1) {
+ LOG(ERROR) << "failed to read blocks for move";
+ return -1;
+ }
- if (status == -1) {
- LOG(ERROR) << "failed to read blocks for move";
- return -1;
- }
+ if (status == 0) {
+ params.foundwrites = true;
+ } else if (params.foundwrites) {
+ LOG(WARNING) << "warning: commands executed out of order [" << params.cmdname << "]";
+ }
+ if (params.canwrite) {
if (status == 0) {
- params.foundwrites = true;
- } else if (params.foundwrites) {
- LOG(WARNING) << "warning: commands executed out of order [" << params.cmdname << "]";
+ LOG(INFO) << " moving " << blocks << " blocks";
+
+ if (WriteBlocks(tgt, params.buffer, params.fd) == -1) {
+ return -1;
+ }
+ } else {
+ LOG(INFO) << "skipping " << blocks << " already moved blocks";
}
+ }
- if (params.canwrite) {
- if (status == 0) {
- LOG(INFO) << " moving " << blocks << " blocks";
+ if (!params.freestash.empty()) {
+ FreeStash(params.stashbase, params.freestash);
+ params.freestash.clear();
+ }
- if (WriteBlocks(tgt, params.buffer, params.fd) == -1) {
- return -1;
- }
- } else {
- LOG(INFO) << "skipping " << blocks << " already moved blocks";
- }
+ params.written += tgt.size;
- }
-
- if (!params.freestash.empty()) {
- FreeStash(params.stashbase, params.freestash);
- params.freestash.clear();
- }
-
- params.written += tgt.size;
-
- return 0;
+ return 0;
}
static int PerformCommandStash(CommandParameters& params) {
- return SaveStash(params, params.stashbase, params.buffer, params.fd,
- (params.version >= 3));
+ return SaveStash(params, params.stashbase, params.buffer, params.fd);
}
static int PerformCommandFree(CommandParameters& params) {
@@ -1337,15 +1297,7 @@
RangeSet tgt;
size_t blocks = 0;
bool overlap = false;
- int status = 0;
- if (params.version == 1) {
- status = LoadSrcTgtVersion1(params, tgt, blocks, params.buffer, params.fd);
- } else if (params.version == 2) {
- status = LoadSrcTgtVersion2(params, tgt, blocks, params.buffer, params.fd,
- params.stashbase, nullptr);
- } else if (params.version >= 3) {
- status = LoadSrcTgtVersion3(params, tgt, blocks, false, overlap);
- }
+ int status = LoadSrcTgtVersion3(params, tgt, blocks, false, overlap);
if (status == -1) {
LOG(ERROR) << "failed to read blocks for diff";
@@ -1471,301 +1423,281 @@
static Value* PerformBlockImageUpdate(const char* name, State* state,
const std::vector<std::unique_ptr<Expr>>& argv,
const Command* commands, size_t cmdcount, bool dryrun) {
- CommandParameters params = {};
- params.canwrite = !dryrun;
+ CommandParameters params = {};
+ params.canwrite = !dryrun;
- LOG(INFO) << "performing " << (dryrun ? "verification" : "update");
- if (state->is_retry) {
- is_retry = true;
- LOG(INFO) << "This update is a retry.";
+ LOG(INFO) << "performing " << (dryrun ? "verification" : "update");
+ if (state->is_retry) {
+ is_retry = true;
+ LOG(INFO) << "This update is a retry.";
+ }
+ if (argv.size() != 4) {
+ ErrorAbort(state, kArgsParsingFailure, "block_image_update expects 4 arguments, got %zu",
+ argv.size());
+ return StringValue("");
+ }
+
+ std::vector<std::unique_ptr<Value>> args;
+ if (!ReadValueArgs(state, argv, &args)) {
+ return nullptr;
+ }
+
+ const Value* blockdev_filename = args[0].get();
+ const Value* transfer_list_value = args[1].get();
+ const Value* new_data_fn = args[2].get();
+ const Value* patch_data_fn = args[3].get();
+
+ if (blockdev_filename->type != VAL_STRING) {
+ ErrorAbort(state, kArgsParsingFailure, "blockdev_filename argument to %s must be string", name);
+ return StringValue("");
+ }
+ if (transfer_list_value->type != VAL_BLOB) {
+ ErrorAbort(state, kArgsParsingFailure, "transfer_list argument to %s must be blob", name);
+ return StringValue("");
+ }
+ if (new_data_fn->type != VAL_STRING) {
+ ErrorAbort(state, kArgsParsingFailure, "new_data_fn argument to %s must be string", name);
+ return StringValue("");
+ }
+ if (patch_data_fn->type != VAL_STRING) {
+ ErrorAbort(state, kArgsParsingFailure, "patch_data_fn argument to %s must be string", name);
+ return StringValue("");
+ }
+
+ UpdaterInfo* ui = static_cast<UpdaterInfo*>(state->cookie);
+ if (ui == nullptr) {
+ return StringValue("");
+ }
+
+ FILE* cmd_pipe = ui->cmd_pipe;
+ ZipArchiveHandle za = ui->package_zip;
+
+ if (cmd_pipe == nullptr || za == nullptr) {
+ return StringValue("");
+ }
+
+ ZipString path_data(patch_data_fn->data.c_str());
+ ZipEntry patch_entry;
+ if (FindEntry(za, path_data, &patch_entry) != 0) {
+ LOG(ERROR) << name << "(): no file \"" << patch_data_fn->data << "\" in package";
+ return StringValue("");
+ }
+
+ params.patch_start = ui->package_zip_addr + patch_entry.offset;
+ ZipString new_data(new_data_fn->data.c_str());
+ ZipEntry new_entry;
+ if (FindEntry(za, new_data, &new_entry) != 0) {
+ LOG(ERROR) << name << "(): no file \"" << new_data_fn->data << "\" in package";
+ return StringValue("");
+ }
+
+ params.fd.reset(TEMP_FAILURE_RETRY(ota_open(blockdev_filename->data.c_str(), O_RDWR)));
+ if (params.fd == -1) {
+ PLOG(ERROR) << "open \"" << blockdev_filename->data << "\" failed";
+ return StringValue("");
+ }
+
+ if (params.canwrite) {
+ params.nti.za = za;
+ params.nti.entry = new_entry;
+
+ pthread_mutex_init(¶ms.nti.mu, nullptr);
+ pthread_cond_init(¶ms.nti.cv, nullptr);
+ pthread_attr_t attr;
+ pthread_attr_init(&attr);
+ pthread_attr_setdetachstate(&attr, PTHREAD_CREATE_JOINABLE);
+
+ int error = pthread_create(¶ms.thread, &attr, unzip_new_data, ¶ms.nti);
+ if (error != 0) {
+ PLOG(ERROR) << "pthread_create failed";
+ return StringValue("");
}
- if (argv.size() != 4) {
- ErrorAbort(state, kArgsParsingFailure, "block_image_update expects 4 arguments, got %zu",
- argv.size());
- return StringValue("");
+ }
+
+ std::vector<std::string> lines = android::base::Split(transfer_list_value->data, "\n");
+ if (lines.size() < 2) {
+ ErrorAbort(state, kArgsParsingFailure, "too few lines in the transfer list [%zd]\n",
+ lines.size());
+ return StringValue("");
+ }
+
+ // First line in transfer list is the version number.
+ if (!android::base::ParseInt(lines[0], ¶ms.version, 3, 4)) {
+ LOG(ERROR) << "unexpected transfer list version [" << lines[0] << "]";
+ return StringValue("");
+ }
+
+ LOG(INFO) << "blockimg version is " << params.version;
+
+ // Second line in transfer list is the total number of blocks we expect to write.
+ size_t total_blocks;
+ if (!android::base::ParseUint(lines[1], &total_blocks)) {
+ ErrorAbort(state, kArgsParsingFailure, "unexpected block count [%s]\n", lines[1].c_str());
+ return StringValue("");
+ }
+
+ if (total_blocks == 0) {
+ return StringValue("t");
+ }
+
+ size_t start = 2;
+ if (lines.size() < 4) {
+ ErrorAbort(state, kArgsParsingFailure, "too few lines in the transfer list [%zu]\n",
+ lines.size());
+ return StringValue("");
+ }
+
+ // Third line is how many stash entries are needed simultaneously.
+ LOG(INFO) << "maximum stash entries " << lines[2];
+
+ // Fourth line is the maximum number of blocks that will be stashed simultaneously
+ size_t stash_max_blocks;
+ if (!android::base::ParseUint(lines[3], &stash_max_blocks)) {
+ ErrorAbort(state, kArgsParsingFailure, "unexpected maximum stash blocks [%s]\n",
+ lines[3].c_str());
+ return StringValue("");
+ }
+
+ int res = CreateStash(state, stash_max_blocks, blockdev_filename->data, params.stashbase);
+ if (res == -1) {
+ return StringValue("");
+ }
+
+ params.createdstash = res;
+
+ start += 2;
+
+ // Build a map of the available commands
+ std::unordered_map<std::string, const Command*> cmd_map;
+ for (size_t i = 0; i < cmdcount; ++i) {
+ if (cmd_map.find(commands[i].name) != cmd_map.end()) {
+ LOG(ERROR) << "Error: command [" << commands[i].name << "] already exists in the cmd map.";
+ return StringValue(strdup(""));
+ }
+ cmd_map[commands[i].name] = &commands[i];
+ }
+
+ int rc = -1;
+
+ // Subsequent lines are all individual transfer commands
+ for (auto it = lines.cbegin() + start; it != lines.cend(); it++) {
+ const std::string& line(*it);
+ if (line.empty()) continue;
+
+ params.tokens = android::base::Split(line, " ");
+ params.cpos = 0;
+ params.cmdname = params.tokens[params.cpos++].c_str();
+ params.cmdline = line.c_str();
+
+ if (cmd_map.find(params.cmdname) == cmd_map.end()) {
+ LOG(ERROR) << "unexpected command [" << params.cmdname << "]";
+ goto pbiudone;
}
- std::vector<std::unique_ptr<Value>> args;
- if (!ReadValueArgs(state, argv, &args)) {
- return nullptr;
- }
+ const Command* cmd = cmd_map[params.cmdname];
- const Value* blockdev_filename = args[0].get();
- const Value* transfer_list_value = args[1].get();
- const Value* new_data_fn = args[2].get();
- const Value* patch_data_fn = args[3].get();
-
- if (blockdev_filename->type != VAL_STRING) {
- ErrorAbort(state, kArgsParsingFailure, "blockdev_filename argument to %s must be string",
- name);
- return StringValue("");
- }
- if (transfer_list_value->type != VAL_BLOB) {
- ErrorAbort(state, kArgsParsingFailure, "transfer_list argument to %s must be blob", name);
- return StringValue("");
- }
- if (new_data_fn->type != VAL_STRING) {
- ErrorAbort(state, kArgsParsingFailure, "new_data_fn argument to %s must be string", name);
- return StringValue("");
- }
- if (patch_data_fn->type != VAL_STRING) {
- ErrorAbort(state, kArgsParsingFailure, "patch_data_fn argument to %s must be string",
- name);
- return StringValue("");
- }
-
- UpdaterInfo* ui = static_cast<UpdaterInfo*>(state->cookie);
- if (ui == nullptr) {
- return StringValue("");
- }
-
- FILE* cmd_pipe = ui->cmd_pipe;
- ZipArchiveHandle za = ui->package_zip;
-
- if (cmd_pipe == nullptr || za == nullptr) {
- return StringValue("");
- }
-
- ZipString path_data(patch_data_fn->data.c_str());
- ZipEntry patch_entry;
- if (FindEntry(za, path_data, &patch_entry) != 0) {
- LOG(ERROR) << name << "(): no file \"" << patch_data_fn->data << "\" in package";
- return StringValue("");
- }
-
- params.patch_start = ui->package_zip_addr + patch_entry.offset;
- ZipString new_data(new_data_fn->data.c_str());
- ZipEntry new_entry;
- if (FindEntry(za, new_data, &new_entry) != 0) {
- LOG(ERROR) << name << "(): no file \"" << new_data_fn->data << "\" in package";
- return StringValue("");
- }
-
- params.fd.reset(TEMP_FAILURE_RETRY(ota_open(blockdev_filename->data.c_str(), O_RDWR)));
- if (params.fd == -1) {
- PLOG(ERROR) << "open \"" << blockdev_filename->data << "\" failed";
- return StringValue("");
+ if (cmd->f != nullptr && cmd->f(params) == -1) {
+ LOG(ERROR) << "failed to execute command [" << line << "]";
+ goto pbiudone;
}
if (params.canwrite) {
- params.nti.za = za;
- params.nti.entry = new_entry;
-
- pthread_mutex_init(¶ms.nti.mu, nullptr);
- pthread_cond_init(¶ms.nti.cv, nullptr);
- pthread_attr_t attr;
- pthread_attr_init(&attr);
- pthread_attr_setdetachstate(&attr, PTHREAD_CREATE_JOINABLE);
-
- int error = pthread_create(¶ms.thread, &attr, unzip_new_data, ¶ms.nti);
- if (error != 0) {
- PLOG(ERROR) << "pthread_create failed";
- return StringValue("");
- }
- }
-
- std::vector<std::string> lines = android::base::Split(transfer_list_value->data, "\n");
- if (lines.size() < 2) {
- ErrorAbort(state, kArgsParsingFailure, "too few lines in the transfer list [%zd]\n",
- lines.size());
- return StringValue("");
- }
-
- // First line in transfer list is the version number
- if (!android::base::ParseInt(lines[0], ¶ms.version, 1, 4)) {
- LOG(ERROR) << "unexpected transfer list version [" << lines[0] << "]";
- return StringValue("");
- }
-
- LOG(INFO) << "blockimg version is " << params.version;
-
- // Second line in transfer list is the total number of blocks we expect to write
- size_t total_blocks;
- if (!android::base::ParseUint(lines[1], &total_blocks)) {
- ErrorAbort(state, kArgsParsingFailure, "unexpected block count [%s]\n", lines[1].c_str());
- return StringValue("");
- }
-
- if (total_blocks == 0) {
- return StringValue("t");
- }
-
- size_t start = 2;
- if (params.version >= 2) {
- if (lines.size() < 4) {
- ErrorAbort(state, kArgsParsingFailure, "too few lines in the transfer list [%zu]\n",
- lines.size());
- return StringValue("");
- }
-
- // Third line is how many stash entries are needed simultaneously
- LOG(INFO) << "maximum stash entries " << lines[2];
-
- // Fourth line is the maximum number of blocks that will be stashed simultaneously
- size_t stash_max_blocks;
- if (!android::base::ParseUint(lines[3], &stash_max_blocks)) {
- ErrorAbort(state, kArgsParsingFailure, "unexpected maximum stash blocks [%s]\n",
- lines[3].c_str());
- return StringValue("");
- }
-
- int res = CreateStash(state, stash_max_blocks, blockdev_filename->data, params.stashbase);
- if (res == -1) {
- return StringValue("");
- }
-
- params.createdstash = res;
-
- start += 2;
- }
-
- // Build a map of the available commands
- std::unordered_map<std::string, const Command*> cmd_map;
- for (size_t i = 0; i < cmdcount; ++i) {
- if (cmd_map.find(commands[i].name) != cmd_map.end()) {
- LOG(ERROR) << "Error: command [" << commands[i].name
- << "] already exists in the cmd map.";
- return StringValue(strdup(""));
- }
- cmd_map[commands[i].name] = &commands[i];
- }
-
- int rc = -1;
-
- // Subsequent lines are all individual transfer commands
- for (auto it = lines.cbegin() + start; it != lines.cend(); it++) {
- const std::string& line(*it);
- if (line.empty()) continue;
-
- params.tokens = android::base::Split(line, " ");
- params.cpos = 0;
- params.cmdname = params.tokens[params.cpos++].c_str();
- params.cmdline = line.c_str();
-
- if (cmd_map.find(params.cmdname) == cmd_map.end()) {
- LOG(ERROR) << "unexpected command [" << params.cmdname << "]";
- goto pbiudone;
- }
-
- const Command* cmd = cmd_map[params.cmdname];
-
- if (cmd->f != nullptr && cmd->f(params) == -1) {
- LOG(ERROR) << "failed to execute command [" << line << "]";
- goto pbiudone;
- }
-
- if (params.canwrite) {
- if (ota_fsync(params.fd) == -1) {
- failure_type = kFsyncFailure;
- PLOG(ERROR) << "fsync failed";
- goto pbiudone;
- }
- fprintf(cmd_pipe, "set_progress %.4f\n",
- static_cast<double>(params.written) / total_blocks);
- fflush(cmd_pipe);
- }
- }
-
- if (params.canwrite) {
- pthread_join(params.thread, nullptr);
-
- LOG(INFO) << "wrote " << params.written << " blocks; expected " << total_blocks;
- LOG(INFO) << "stashed " << params.stashed << " blocks";
- LOG(INFO) << "max alloc needed was " << params.buffer.size();
-
- const char* partition = strrchr(blockdev_filename->data.c_str(), '/');
- if (partition != nullptr && *(partition + 1) != 0) {
- fprintf(cmd_pipe, "log bytes_written_%s: %zu\n", partition + 1,
- params.written * BLOCKSIZE);
- fprintf(cmd_pipe, "log bytes_stashed_%s: %zu\n", partition + 1,
- params.stashed * BLOCKSIZE);
- fflush(cmd_pipe);
- }
- // Delete stash only after successfully completing the update, as it
- // may contain blocks needed to complete the update later.
- DeleteStash(params.stashbase);
- } else {
- LOG(INFO) << "verified partition contents; update may be resumed";
- }
-
- rc = 0;
-
-pbiudone:
- if (ota_fsync(params.fd) == -1) {
+ if (ota_fsync(params.fd) == -1) {
failure_type = kFsyncFailure;
PLOG(ERROR) << "fsync failed";
+ goto pbiudone;
+ }
+ fprintf(cmd_pipe, "set_progress %.4f\n", static_cast<double>(params.written) / total_blocks);
+ fflush(cmd_pipe);
}
- // params.fd will be automatically closed because it's a unique_fd.
+ }
- // Only delete the stash if the update cannot be resumed, or it's
- // a verification run and we created the stash.
- if (params.isunresumable || (!params.canwrite && params.createdstash)) {
- DeleteStash(params.stashbase);
+ if (params.canwrite) {
+ pthread_join(params.thread, nullptr);
+
+ LOG(INFO) << "wrote " << params.written << " blocks; expected " << total_blocks;
+ LOG(INFO) << "stashed " << params.stashed << " blocks";
+ LOG(INFO) << "max alloc needed was " << params.buffer.size();
+
+ const char* partition = strrchr(blockdev_filename->data.c_str(), '/');
+ if (partition != nullptr && *(partition + 1) != 0) {
+ fprintf(cmd_pipe, "log bytes_written_%s: %zu\n", partition + 1, params.written * BLOCKSIZE);
+ fprintf(cmd_pipe, "log bytes_stashed_%s: %zu\n", partition + 1, params.stashed * BLOCKSIZE);
+ fflush(cmd_pipe);
}
+ // Delete stash only after successfully completing the update, as it may contain blocks needed
+ // to complete the update later.
+ DeleteStash(params.stashbase);
+ } else {
+ LOG(INFO) << "verified partition contents; update may be resumed";
+ }
- if (failure_type != kNoCause && state->cause_code == kNoCause) {
- state->cause_code = failure_type;
- }
+ rc = 0;
- return StringValue(rc == 0 ? "t" : "");
+pbiudone:
+ if (ota_fsync(params.fd) == -1) {
+ failure_type = kFsyncFailure;
+ PLOG(ERROR) << "fsync failed";
+ }
+ // params.fd will be automatically closed because it's a unique_fd.
+
+ // Only delete the stash if the update cannot be resumed, or it's a verification run and we
+ // created the stash.
+ if (params.isunresumable || (!params.canwrite && params.createdstash)) {
+ DeleteStash(params.stashbase);
+ }
+
+ if (failure_type != kNoCause && state->cause_code == kNoCause) {
+ state->cause_code = failure_type;
+ }
+
+ return StringValue(rc == 0 ? "t" : "");
}
-// The transfer list is a text file containing commands to
-// transfer data from one place to another on the target
-// partition. We parse it and execute the commands in order:
-//
-// zero [rangeset]
-// - fill the indicated blocks with zeros
-//
-// new [rangeset]
-// - fill the blocks with data read from the new_data file
-//
-// erase [rangeset]
-// - mark the given blocks as empty
-//
-// move <...>
-// bsdiff <patchstart> <patchlen> <...>
-// imgdiff <patchstart> <patchlen> <...>
-// - read the source blocks, apply a patch (or not in the
-// case of move), write result to target blocks. bsdiff or
-// imgdiff specifies the type of patch; move means no patch
-// at all.
-//
-// The format of <...> differs between versions 1 and 2;
-// see the LoadSrcTgtVersion{1,2}() functions for a
-// description of what's expected.
-//
-// stash <stash_id> <src_range>
-// - (version 2+ only) load the given source range and stash
-// the data in the given slot of the stash table.
-//
-// free <stash_id>
-// - (version 3+ only) free the given stash data.
-//
-// The creator of the transfer list will guarantee that no block
-// is read (ie, used as the source for a patch or move) after it
-// has been written.
-//
-// In version 2, the creator will guarantee that a given stash is
-// loaded (with a stash command) before it's used in a
-// move/bsdiff/imgdiff command.
-//
-// Within one command the source and target ranges may overlap so
-// in general we need to read the entire source into memory before
-// writing anything to the target blocks.
-//
-// All the patch data is concatenated into one patch_data file in
-// the update package. It must be stored uncompressed because we
-// memory-map it in directly from the archive. (Since patches are
-// already compressed, we lose very little by not compressing
-// their concatenation.)
-//
-// In version 3, commands that read data from the partition (i.e.
-// move/bsdiff/imgdiff/stash) have one or more additional hashes
-// before the range parameters, which are used to check if the
-// command has already been completed and verify the integrity of
-// the source data.
-
+/**
+ * The transfer list is a text file containing commands to transfer data from one place to another
+ * on the target partition. We parse it and execute the commands in order:
+ *
+ * zero [rangeset]
+ * - Fill the indicated blocks with zeros.
+ *
+ * new [rangeset]
+ * - Fill the blocks with data read from the new_data file.
+ *
+ * erase [rangeset]
+ * - Mark the given blocks as empty.
+ *
+ * move <...>
+ * bsdiff <patchstart> <patchlen> <...>
+ * imgdiff <patchstart> <patchlen> <...>
+ * - Read the source blocks, apply a patch (or not in the case of move), write result to target
+ * blocks. bsdiff or imgdiff specifies the type of patch; move means no patch at all.
+ *
+ * See the comments in LoadSrcTgtVersion3() for a description of the <...> format.
+ *
+ * stash <stash_id> <src_range>
+ * - Load the given source range and stash the data in the given slot of the stash table.
+ *
+ * free <stash_id>
+ * - Free the given stash data.
+ *
+ * The creator of the transfer list will guarantee that no block is read (ie, used as the source for
+ * a patch or move) after it has been written.
+ *
+ * The creator will guarantee that a given stash is loaded (with a stash command) before it's used
+ * in a move/bsdiff/imgdiff command.
+ *
+ * Within one command the source and target ranges may overlap so in general we need to read the
+ * entire source into memory before writing anything to the target blocks.
+ *
+ * All the patch data is concatenated into one patch_data file in the update package. It must be
+ * stored uncompressed because we memory-map it in directly from the archive. (Since patches are
+ * already compressed, we lose very little by not compressing their concatenation.)
+ *
+ * Commands that read data from the partition (i.e. move/bsdiff/imgdiff/stash) have one or more
+ * additional hashes before the range parameters, which are used to check if the command has already
+ * been completed and verify the integrity of the source data.
+ */
Value* BlockImageVerifyFn(const char* name, State* state,
const std::vector<std::unique_ptr<Expr>>& argv) {
// Commands which are not tested are set to nullptr to skip them completely