Index: full_update_generator.cc |
diff --git a/full_update_generator.cc b/full_update_generator.cc |
new file mode 100644 |
index 0000000000000000000000000000000000000000..0140107f8818faa1e22e35c78080ddebb1ee137b |
--- /dev/null |
+++ b/full_update_generator.cc |
@@ -0,0 +1,197 @@ |
+// Copyright (c) 2010 The Chromium OS Authors. All rights reserved. |
+// Use of this source code is governed by a BSD-style license that can be |
+// found in the LICENSE file. |
+ |
+#include "update_engine/full_update_generator.h" |
+ |
+#include <inttypes.h> |
+#include <fcntl.h> |
+ |
+#include <tr1/memory> |
+ |
+#include <base/string_util.h> |
+ |
+#include "update_engine/bzip.h" |
+#include "update_engine/utils.h" |
+ |
+using std::deque; |
+using std::min; |
+using std::max; |
+using std::string; |
+using std::tr1::shared_ptr; |
+using std::vector; |
+ |
+namespace chromeos_update_engine { |
+ |
+namespace { |
+ |
+// This class encapsulates a full update chunk processing thread. The processor |
+// reads a chunk of data from the input file descriptor and compresses it. The |
+// processor needs to be started through Start() then waited on through Wait(). |
+class ChunkProcessor { |
+ public: |
+ // Read a chunk of |size| bytes from |fd| starting at offset |offset|. |
+ ChunkProcessor(int fd, off_t offset, size_t size) |
+ : thread_(NULL), |
+ fd_(fd), |
+ offset_(offset), |
+ buffer_in_(size) {} |
+ ~ChunkProcessor() { Wait(); } |
+ |
+ off_t offset() const { return offset_; } |
+ const vector<char>& buffer_in() const { return buffer_in_; } |
+ const vector<char>& buffer_compressed() const { return buffer_compressed_; } |
+ |
+ // Starts the processor. Returns true on success, false on failure. |
+ bool Start(); |
+ |
+ // Waits for the processor to complete. Returns true on success, false on |
+ // failure. |
+ bool Wait(); |
+ |
+ bool ShouldCompress() const { |
+ return buffer_compressed_.size() < buffer_in_.size(); |
+ } |
+ |
+ private: |
+ // Reads the input data into |buffer_in_| and compresses it into |
+ // |buffer_compressed_|. Returns true on success, false otherwise. |
+ bool ReadAndCompress(); |
+ static gpointer ReadAndCompressThread(gpointer data); |
+ |
+ GThread* thread_; |
+ int fd_; |
+ off_t offset_; |
+ vector<char> buffer_in_; |
+ vector<char> buffer_compressed_; |
+ |
+ DISALLOW_COPY_AND_ASSIGN(ChunkProcessor); |
+}; |
+ |
+bool ChunkProcessor::Start() { |
+ thread_ = g_thread_create(ReadAndCompressThread, this, TRUE, NULL); |
+ TEST_AND_RETURN_FALSE(thread_ != NULL); |
+ return true; |
+} |
+ |
+bool ChunkProcessor::Wait() { |
+ if (!thread_) { |
+ return false; |
+ } |
+ gpointer result = g_thread_join(thread_); |
+ thread_ = NULL; |
+ TEST_AND_RETURN_FALSE(result == this); |
+ return true; |
+} |
+ |
+gpointer ChunkProcessor::ReadAndCompressThread(gpointer data) { |
+ return |
+ reinterpret_cast<ChunkProcessor*>(data)->ReadAndCompress() ? data : NULL; |
+} |
+ |
+bool ChunkProcessor::ReadAndCompress() { |
+ ssize_t bytes_read = -1; |
+ TEST_AND_RETURN_FALSE(utils::PReadAll(fd_, |
+ buffer_in_.data(), |
+ buffer_in_.size(), |
+ offset_, |
+ &bytes_read)); |
+ TEST_AND_RETURN_FALSE(bytes_read == static_cast<ssize_t>(buffer_in_.size())); |
+ TEST_AND_RETURN_FALSE(BzipCompress(buffer_in_, &buffer_compressed_)); |
+ return true; |
+} |
+ |
+} // namespace |
+ |
+bool FullUpdateGenerator::Run( |
+ Graph* graph, |
+ const std::string& new_kernel_part, |
+ const std::string& new_image, |
+ off_t image_size, |
+ int fd, |
+ off_t* data_file_size, |
+ off_t chunk_size, |
+ off_t block_size, |
+ vector<DeltaArchiveManifest_InstallOperation>* kernel_ops, |
+ std::vector<Vertex::Index>* final_order) { |
+ TEST_AND_RETURN_FALSE(chunk_size > 0); |
+ TEST_AND_RETURN_FALSE((chunk_size % block_size) == 0); |
+ |
+ size_t max_threads = max(sysconf(_SC_NPROCESSORS_ONLN), 4L); |
+ LOG(INFO) << "Max threads: " << max_threads; |
+ |
+ // Get the sizes early in the function, so we can fail fast if the user |
+ // passed us bad paths. |
+ TEST_AND_RETURN_FALSE(image_size >= 0 && |
+ image_size <= utils::FileSize(new_image)); |
+ const off_t kernel_size = utils::FileSize(new_kernel_part); |
+ TEST_AND_RETURN_FALSE(kernel_size >= 0); |
+ |
+ off_t part_sizes[] = { image_size, kernel_size }; |
+ string paths[] = { new_image, new_kernel_part }; |
+ |
+ for (int partition = 0; partition < 2; ++partition) { |
+ const string& path = paths[partition]; |
+ LOG(INFO) << "compressing " << path; |
+ |
+ int in_fd = open(path.c_str(), O_RDONLY, 0); |
+ TEST_AND_RETURN_FALSE(in_fd >= 0); |
+ ScopedFdCloser in_fd_closer(&in_fd); |
+ |
+ deque<shared_ptr<ChunkProcessor> > threads; |
+ |
+ off_t bytes_left = part_sizes[partition], counter = 0, offset = 0; |
+ while (bytes_left > 0 || !threads.empty()) { |
+ // Check and start new chunk processors if possible. |
+ while (threads.size() < max_threads && bytes_left > 0) { |
+ shared_ptr<ChunkProcessor> processor( |
+ new ChunkProcessor(in_fd, offset, min(bytes_left, chunk_size))); |
+ threads.push_back(processor); |
+ TEST_AND_RETURN_FALSE(processor->Start()); |
+ bytes_left -= chunk_size; |
+ offset += chunk_size; |
+ } |
+ |
+ // Need to wait for a chunk processor to complete and process its ouput |
+ // before spawning new processors. |
+ shared_ptr<ChunkProcessor> processor = threads.front(); |
+ threads.pop_front(); |
adlr
2010/11/08 21:27:08
might be faster to wait on all of the threads, but
petkov
2010/11/08 21:45:21
Maybe... We could also make the separate threads w
|
+ TEST_AND_RETURN_FALSE(processor->Wait()); |
+ |
+ DeltaArchiveManifest_InstallOperation* op = NULL; |
+ if (partition == 0) { |
+ graph->resize(graph->size() + 1); |
+ graph->back().file_name = |
+ StringPrintf("<rootfs-operation-%" PRIi64 ">", counter++); |
+ op = &graph->back().op; |
+ final_order->push_back(graph->size() - 1); |
+ } else { |
+ kernel_ops->resize(kernel_ops->size() + 1); |
+ op = &kernel_ops->back(); |
+ } |
+ |
+ const bool compress = processor->ShouldCompress(); |
+ const vector<char>& use_buf = |
+ compress ? processor->buffer_compressed() : processor->buffer_in(); |
+ op->set_type(compress ? |
+ DeltaArchiveManifest_InstallOperation_Type_REPLACE_BZ : |
+ DeltaArchiveManifest_InstallOperation_Type_REPLACE); |
+ op->set_data_offset(*data_file_size); |
+ TEST_AND_RETURN_FALSE(utils::WriteAll(fd, &use_buf[0], use_buf.size())); |
+ *data_file_size += use_buf.size(); |
+ op->set_data_length(use_buf.size()); |
+ Extent* dst_extent = op->add_dst_extents(); |
+ dst_extent->set_start_block(processor->offset() / block_size); |
+ dst_extent->set_num_blocks(chunk_size / block_size); |
+ |
+ LOG(INFO) |
+ << StringPrintf("%.1f", |
+ processor->offset() * 100.0 / part_sizes[partition]) |
+ << "% complete (output size: " << *data_file_size << ")"; |
+ } |
+ } |
+ |
+ return true; |
+} |
+ |
+} // namespace chromeos_update_engine |