| Index: base/utf_offset_string_conversions.cc
|
| ===================================================================
|
| --- base/utf_offset_string_conversions.cc (revision 31511)
|
| +++ base/utf_offset_string_conversions.cc (working copy)
|
| @@ -2,149 +2,29 @@
|
| // Use of this source code is governed by a BSD-style license that can be
|
| // found in the LICENSE file.
|
|
|
| -#include "base/utf_string_conversions.h"
|
| +#include "base/utf_offset_string_conversions.h"
|
|
|
| -#include <vector>
|
| +#include "base/string_piece.h"
|
| +#include "base/utf_string_conversion_utils.h"
|
|
|
| -#include "base/basictypes.h"
|
| -#include "base/logging.h"
|
| -#include "base/string_util.h"
|
| -#include "base/third_party/icu/icu_utf.h"
|
| +using base::PrepareForUTF16Or32Output;
|
| +using base::ReadUnicodeCharacter;
|
| +using base::WriteUnicodeCharacter;
|
|
|
| -namespace {
|
| -
|
| -inline bool IsValidCodepoint(uint32 code_point) {
|
| - // Excludes the surrogate code points ([0xD800, 0xDFFF]) and
|
| - // codepoints larger than 0x10FFFF (the highest codepoint allowed).
|
| - // Non-characters and unassigned codepoints are allowed.
|
| - return code_point < 0xD800u ||
|
| - (code_point >= 0xE000u && code_point <= 0x10FFFFu);
|
| -}
|
| -
|
| -// ReadUnicodeCharacter --------------------------------------------------------
|
| -
|
| -// Reads a UTF-8 stream, placing the next code point into the given output
|
| -// |*code_point|. |src| represents the entire string to read, and |*char_index|
|
| -// is the character offset within the string to start reading at. |*char_index|
|
| -// will be updated to index the last character read, such that incrementing it
|
| -// (as in a for loop) will take the reader to the next character.
|
| -//
|
| -// Returns true on success. On false, |*code_point| will be invalid.
|
| -bool ReadUnicodeCharacter(const char* src, int32 src_len,
|
| - int32* char_index, uint32* code_point_out) {
|
| - // U8_NEXT expects to be able to use -1 to signal an error, so we must
|
| - // use a signed type for code_point. But this function returns false
|
| - // on error anyway, so code_point_out is unsigned.
|
| - int32 code_point;
|
| - CBU8_NEXT(src, *char_index, src_len, code_point);
|
| - *code_point_out = static_cast<uint32>(code_point);
|
| -
|
| - // The ICU macro above moves to the next char, we want to point to the last
|
| - // char consumed.
|
| - (*char_index)--;
|
| -
|
| - // Validate the decoded value.
|
| - return IsValidCodepoint(code_point);
|
| -}
|
| -
|
| -// Reads a UTF-16 character. The usage is the same as the 8-bit version above.
|
| -bool ReadUnicodeCharacter(const char16* src, int32 src_len,
|
| - int32* char_index, uint32* code_point) {
|
| - if (CBU16_IS_SURROGATE(src[*char_index])) {
|
| - if (!CBU16_IS_SURROGATE_LEAD(src[*char_index]) ||
|
| - *char_index + 1 >= src_len ||
|
| - !CBU16_IS_TRAIL(src[*char_index + 1])) {
|
| - // Invalid surrogate pair.
|
| - return false;
|
| - }
|
| -
|
| - // Valid surrogate pair.
|
| - *code_point = CBU16_GET_SUPPLEMENTARY(src[*char_index],
|
| - src[*char_index + 1]);
|
| - (*char_index)++;
|
| - } else {
|
| - // Not a surrogate, just one 16-bit word.
|
| - *code_point = src[*char_index];
|
| - }
|
| -
|
| - return IsValidCodepoint(*code_point);
|
| -}
|
| -
|
| -#if defined(WCHAR_T_IS_UTF32)
|
| -// Reads UTF-32 character. The usage is the same as the 8-bit version above.
|
| -bool ReadUnicodeCharacter(const wchar_t* src, int32 src_len,
|
| - int32* char_index, uint32* code_point) {
|
| - // Conversion is easy since the source is 32-bit.
|
| - *code_point = src[*char_index];
|
| -
|
| - // Validate the value.
|
| - return IsValidCodepoint(*code_point);
|
| -}
|
| -#endif // defined(WCHAR_T_IS_UTF32)
|
| -
|
| -// WriteUnicodeCharacter -------------------------------------------------------
|
| -
|
| -// Appends a UTF-8 character to the given 8-bit string. Returns the number of
|
| -// bytes written.
|
| -size_t WriteUnicodeCharacter(uint32 code_point, std::string* output) {
|
| - if (code_point <= 0x7f) {
|
| - // Fast path the common case of one byte.
|
| - output->push_back(code_point);
|
| - return 1;
|
| - }
|
| -
|
| - // CBU8_APPEND_UNSAFE can append up to 4 bytes.
|
| - size_t char_offset = output->length();
|
| - size_t original_char_offset = char_offset;
|
| - output->resize(char_offset + CBU8_MAX_LENGTH);
|
| -
|
| - CBU8_APPEND_UNSAFE(&(*output)[0], char_offset, code_point);
|
| -
|
| - // CBU8_APPEND_UNSAFE will advance our pointer past the inserted character, so
|
| - // it will represent the new length of the string.
|
| - output->resize(char_offset);
|
| - return char_offset - original_char_offset;
|
| -}
|
| -
|
| -// Appends the given code point as a UTF-16 character to the given 16-bit
|
| -// string. Returns the number of 16-bit values written.
|
| -size_t WriteUnicodeCharacter(uint32 code_point, string16* output) {
|
| - if (CBU16_LENGTH(code_point) == 1) {
|
| - // Thie code point is in the Basic Multilingual Plane (BMP).
|
| - output->push_back(static_cast<char16>(code_point));
|
| - return 1;
|
| - }
|
| - // Non-BMP characters use a double-character encoding.
|
| - size_t char_offset = output->length();
|
| - output->resize(char_offset + CBU16_MAX_LENGTH);
|
| - CBU16_APPEND_UNSAFE(&(*output)[0], char_offset, code_point);
|
| - return CBU16_MAX_LENGTH;
|
| -}
|
| -
|
| -#if defined(WCHAR_T_IS_UTF32)
|
| -// Appends the given UTF-32 character to the given 32-bit string. Returns the
|
| -// number of 32-bit values written.
|
| -inline size_t WriteUnicodeCharacter(uint32 code_point, std::wstring* output) {
|
| - // This is the easy case, just append the character.
|
| - output->push_back(code_point);
|
| - return 1;
|
| -}
|
| -#endif // defined(WCHAR_T_IS_UTF32)
|
| -
|
| // Generalized Unicode converter -----------------------------------------------
|
|
|
| // Converts the given source Unicode character type to the given destination
|
| // Unicode character type as a STL string. The given input buffer and size
|
| // determine the source, and the given output STL string will be replaced by
|
| // the result.
|
| -template<typename SRC_CHAR, typename DEST_STRING>
|
| +template<typename SRC_CHAR>
|
| bool ConvertUnicode(const SRC_CHAR* src,
|
| size_t src_len,
|
| - DEST_STRING* output,
|
| + std::wstring* output,
|
| size_t* offset_for_adjustment) {
|
| size_t output_offset =
|
| (offset_for_adjustment && *offset_for_adjustment < src_len) ?
|
| - *offset_for_adjustment : DEST_STRING::npos;
|
| + *offset_for_adjustment : std::wstring::npos;
|
|
|
| // ICU requires 32-bit numbers.
|
| bool success = true;
|
| @@ -160,14 +40,14 @@
|
| // in place of an invalid codepoint.
|
| success = false;
|
| }
|
| - if ((output_offset != DEST_STRING::npos) &&
|
| + if ((output_offset != std::wstring::npos) &&
|
| (*offset_for_adjustment > original_i)) {
|
| // NOTE: ReadUnicodeCharacter() adjusts |i| to point _at_ the last
|
| // character read, not after it (so that incrementing it in the loop
|
| // increment will place it at the right location), so we need to account
|
| // for that in determining the amount that was read.
|
| if (*offset_for_adjustment <= static_cast<size_t>(i))
|
| - output_offset = DEST_STRING::npos;
|
| + output_offset = std::wstring::npos;
|
| else
|
| output_offset += chars_written - (i - original_i + 1);
|
| }
|
| @@ -178,69 +58,14 @@
|
| return success;
|
| }
|
|
|
| -// Guesses the length of the output in UTF-8 in bytes, clears that output
|
| -// string, and reserves that amount of space. We assume that the input
|
| -// character types are unsigned, which will be true for UTF-16 and -32 on our
|
| -// systems.
|
| -template<typename CHAR>
|
| -void PrepareForUTF8Output(const CHAR* src,
|
| - size_t src_len,
|
| - std::string* output) {
|
| - output->clear();
|
| - if (src_len == 0)
|
| - return;
|
| - if (src[0] < 0x80) {
|
| - // Assume that the entire input will be ASCII.
|
| - output->reserve(src_len);
|
| - } else {
|
| - // Assume that the entire input is non-ASCII and will have 3 bytes per char.
|
| - output->reserve(src_len * 3);
|
| - }
|
| -}
|
| -
|
| -// Prepares an output buffer (containing either UTF-16 or -32 data) given some
|
| -// UTF-8 input that will be converted to it. See PrepareForUTF8Output().
|
| -template<typename STRING>
|
| -void PrepareForUTF16Or32Output(const char* src,
|
| - size_t src_len,
|
| - STRING* output) {
|
| - output->clear();
|
| - if (src_len == 0)
|
| - return;
|
| - if (static_cast<unsigned char>(src[0]) < 0x80) {
|
| - // Assume the input is all ASCII, which means 1:1 correspondence.
|
| - output->reserve(src_len);
|
| - } else {
|
| - // Otherwise assume that the UTF-8 sequences will have 2 bytes for each
|
| - // character.
|
| - output->reserve(src_len / 2);
|
| - }
|
| -}
|
| -
|
| -} // namespace
|
| -
|
| // UTF-8 <-> Wide --------------------------------------------------------------
|
|
|
| -bool WideToUTF8(const wchar_t* src, size_t src_len, std::string* output) {
|
| - PrepareForUTF8Output(src, src_len, output);
|
| - return ConvertUnicode<wchar_t, std::string>(src, src_len, output, NULL);
|
| -}
|
| -
|
| -std::string WideToUTF8(const std::wstring& wide) {
|
| - std::string ret;
|
| - // Ignore the success flag of this call, it will do the best it can for
|
| - // invalid input, which is what we want here.
|
| - WideToUTF8(wide.data(), wide.length(), &ret);
|
| - return ret;
|
| -}
|
| -
|
| bool UTF8ToWideAndAdjustOffset(const char* src,
|
| size_t src_len,
|
| std::wstring* output,
|
| size_t* offset_for_adjustment) {
|
| PrepareForUTF16Or32Output(src, src_len, output);
|
| - return ConvertUnicode<char, std::wstring>(src, src_len, output,
|
| - offset_for_adjustment);
|
| + return ConvertUnicode(src, src_len, output, offset_for_adjustment);
|
| }
|
|
|
| std::wstring UTF8ToWideAndAdjustOffset(const base::StringPiece& utf8,
|
| @@ -256,15 +81,6 @@
|
| #if defined(WCHAR_T_IS_UTF16)
|
|
|
| // When wide == UTF-16, then conversions are a NOP.
|
| -bool WideToUTF16(const wchar_t* src, size_t src_len, string16* output) {
|
| - output->assign(src, src_len);
|
| - return true;
|
| -}
|
| -
|
| -string16 WideToUTF16(const std::wstring& wide) {
|
| - return wide;
|
| -}
|
| -
|
| bool UTF16ToWideAndAdjustOffset(const char16* src,
|
| size_t src_len,
|
| std::wstring* output,
|
| @@ -284,20 +100,6 @@
|
|
|
| #elif defined(WCHAR_T_IS_UTF32)
|
|
|
| -bool WideToUTF16(const wchar_t* src, size_t src_len, string16* output) {
|
| - output->clear();
|
| - // Assume that normally we won't have any non-BMP characters so the counts
|
| - // will be the same.
|
| - output->reserve(src_len);
|
| - return ConvertUnicode<wchar_t, string16>(src, src_len, output, NULL);
|
| -}
|
| -
|
| -string16 WideToUTF16(const std::wstring& wide) {
|
| - string16 ret;
|
| - WideToUTF16(wide.data(), wide.length(), &ret);
|
| - return ret;
|
| -}
|
| -
|
| bool UTF16ToWideAndAdjustOffset(const char16* src,
|
| size_t src_len,
|
| std::wstring* output,
|
| @@ -306,8 +108,7 @@
|
| // Assume that normally we won't have any non-BMP characters so the counts
|
| // will be the same.
|
| output->reserve(src_len);
|
| - return ConvertUnicode<char16, std::wstring>(src, src_len, output,
|
| - offset_for_adjustment);
|
| + return ConvertUnicode(src, src_len, output, offset_for_adjustment);
|
| }
|
|
|
| std::wstring UTF16ToWideAndAdjustOffset(const string16& utf16,
|
| @@ -319,54 +120,3 @@
|
| }
|
|
|
| #endif // defined(WCHAR_T_IS_UTF32)
|
| -
|
| -// UTF16 <-> UTF8 --------------------------------------------------------------
|
| -
|
| -#if defined(WCHAR_T_IS_UTF32)
|
| -
|
| -bool UTF8ToUTF16(const char* src, size_t src_len, string16* output) {
|
| - PrepareForUTF16Or32Output(src, src_len, output);
|
| - return ConvertUnicode<char, string16>(src, src_len, output, NULL);
|
| -}
|
| -
|
| -string16 UTF8ToUTF16(const std::string& utf8) {
|
| - string16 ret;
|
| - // Ignore the success flag of this call, it will do the best it can for
|
| - // invalid input, which is what we want here.
|
| - UTF8ToUTF16(utf8.data(), utf8.length(), &ret);
|
| - return ret;
|
| -}
|
| -
|
| -bool UTF16ToUTF8(const char16* src, size_t src_len, std::string* output) {
|
| - PrepareForUTF8Output(src, src_len, output);
|
| - return ConvertUnicode<char16, std::string>(src, src_len, output, NULL);
|
| -}
|
| -
|
| -std::string UTF16ToUTF8(const string16& utf16) {
|
| - std::string ret;
|
| - // Ignore the success flag of this call, it will do the best it can for
|
| - // invalid input, which is what we want here.
|
| - UTF16ToUTF8(utf16.data(), utf16.length(), &ret);
|
| - return ret;
|
| -}
|
| -
|
| -#elif defined(WCHAR_T_IS_UTF16)
|
| -// Easy case since we can use the "wide" versions we already wrote above.
|
| -
|
| -bool UTF8ToUTF16(const char* src, size_t src_len, string16* output) {
|
| - return UTF8ToWide(src, src_len, output);
|
| -}
|
| -
|
| -string16 UTF8ToUTF16(const std::string& utf8) {
|
| - return UTF8ToWide(utf8);
|
| -}
|
| -
|
| -bool UTF16ToUTF8(const char16* src, size_t src_len, std::string* output) {
|
| - return WideToUTF8(src, src_len, output);
|
| -}
|
| -
|
| -std::string UTF16ToUTF8(const string16& utf16) {
|
| - return WideToUTF8(utf16);
|
| -}
|
| -
|
| -#endif
|
|
|