HELLO·Android
系统源代码
IT资讯
技术文章
我的收藏
注册
登录
-
我收藏的文章
创建代码块
我的代码块
我的账号
Nougat 7.0
|
7.0.0_r31
下载
查看原文件
收藏
根目录
prebuilts
tools
darwin-x86_64
protoc
include
google
protobuf
repeated_field.h
// Protocol Buffers - Google's data interchange format // Copyright 2008 Google Inc. All rights reserved. // https://developers.google.com/protocol-buffers/ // // Redistribution and use in source and binary forms, with or without // modification, are permitted provided that the following conditions are // met: // // * Redistributions of source code must retain the above copyright // notice, this list of conditions and the following disclaimer. // * Redistributions in binary form must reproduce the above // copyright notice, this list of conditions and the following disclaimer // in the documentation and/or other materials provided with the // distribution. // * Neither the name of Google Inc. nor the names of its // contributors may be used to endorse or promote products derived from // this software without specific prior written permission. // // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS // "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT // LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR // A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT // OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, // SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT // LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, // DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. // Author: kenton@google.com (Kenton Varda) // Based on original Protocol Buffers design by // Sanjay Ghemawat, Jeff Dean, and others. // // RepeatedField and RepeatedPtrField are used by generated protocol message // classes to manipulate repeated fields. These classes are very similar to // STL's vector, but include a number of optimizations found to be useful // specifically in the case of Protocol Buffers. RepeatedPtrField is // particularly different from STL vector as it manages ownership of the // pointers that it contains. // // Typically, clients should not need to access RepeatedField objects directly, // but should instead use the accessor functions generated automatically by the // protocol compiler. #ifndef GOOGLE_PROTOBUF_REPEATED_FIELD_H__ #define GOOGLE_PROTOBUF_REPEATED_FIELD_H__ #ifdef _MSC_VER // This is required for min/max on VS2013 only. #include
#endif #include
#include
#include
#include
#include
#include
#include
#include
#include
namespace google { namespace upb { namespace google_opensource { class GMR_Handlers; } // namespace google_opensource } // namespace upb namespace protobuf { class Message; namespace internal { static const int kMinRepeatedFieldAllocationSize = 4; // A utility function for logging that doesn't need any template types. void LogIndexOutOfBounds(int index, int size); template
inline int CalculateReserve(Iter begin, Iter end, std::forward_iterator_tag) { return std::distance(begin, end); } template
inline int CalculateReserve(Iter /*begin*/, Iter /*end*/, std::input_iterator_tag /*unused*/) { return -1; } template
inline int CalculateReserve(Iter begin, Iter end) { typedef typename std::iterator_traits
::iterator_category Category; return CalculateReserve(begin, end, Category()); } } // namespace internal // RepeatedField is used to represent repeated fields of a primitive type (in // other words, everything except strings and nested Messages). Most users will // not ever use a RepeatedField directly; they will use the get-by-index, // set-by-index, and add accessors that are generated for all repeated fields. template
class RepeatedField { public: RepeatedField(); explicit RepeatedField(Arena* arena); RepeatedField(const RepeatedField& other); template
RepeatedField(Iter begin, const Iter& end); ~RepeatedField(); RepeatedField& operator=(const RepeatedField& other); bool empty() const; int size() const; const Element& Get(int index) const; Element* Mutable(int index); void Set(int index, const Element& value); void Add(const Element& value); Element* Add(); // Remove the last element in the array. void RemoveLast(); // Extract elements with indices in "[start .. start+num-1]". // Copy them into "elements[0 .. num-1]" if "elements" is not NULL. // Caution: implementation also moves elements with indices [start+num ..]. // Calling this routine inside a loop can cause quadratic behavior. void ExtractSubrange(int start, int num, Element* elements); void Clear(); void MergeFrom(const RepeatedField& other); void CopyFrom(const RepeatedField& other); // Reserve space to expand the field to at least the given size. If the // array is grown, it will always be at least doubled in size. void Reserve(int new_size); // Resize the RepeatedField to a new, smaller size. This is O(1). void Truncate(int new_size); void AddAlreadyReserved(const Element& value); Element* AddAlreadyReserved(); int Capacity() const; // Like STL resize. Uses value to fill appended elements. // Like Truncate() if new_size <= size(), otherwise this is // O(new_size - size()). void Resize(int new_size, const Element& value); // Gets the underlying array. This pointer is possibly invalidated by // any add or remove operation. Element* mutable_data(); const Element* data() const; // Swap entire contents with "other". If they are separate arenas then, copies // data between each other. void Swap(RepeatedField* other); // Swap entire contents with "other". Should be called only if the caller can // guarantee that both repeated fields are on the same arena or are on the // heap. Swapping between different arenas is disallowed and caught by a // GOOGLE_DCHECK (see API docs for details). void UnsafeArenaSwap(RepeatedField* other); // Swap two elements. void SwapElements(int index1, int index2); // STL-like iterator support typedef Element* iterator; typedef const Element* const_iterator; typedef Element value_type; typedef value_type& reference; typedef const value_type& const_reference; typedef value_type* pointer; typedef const value_type* const_pointer; typedef int size_type; typedef ptrdiff_t difference_type; iterator begin(); const_iterator begin() const; const_iterator cbegin() const; iterator end(); const_iterator end() const; const_iterator cend() const; // Reverse iterator support typedef std::reverse_iterator
const_reverse_iterator; typedef std::reverse_iterator
reverse_iterator; reverse_iterator rbegin() { return reverse_iterator(end()); } const_reverse_iterator rbegin() const { return const_reverse_iterator(end()); } reverse_iterator rend() { return reverse_iterator(begin()); } const_reverse_iterator rend() const { return const_reverse_iterator(begin()); } // Returns the number of bytes used by the repeated field, excluding // sizeof(*this) int SpaceUsedExcludingSelf() const; // Removes the element referenced by position. // // Returns an iterator to the element immediately following the removed // element. // // Invalidates all iterators at or after the removed element, including end(). iterator erase(const_iterator position); // Removes the elements in the range [first, last). // // Returns an iterator to the element immediately following the removed range. // // Invalidates all iterators at or after the removed range, including end(). iterator erase(const_iterator first, const_iterator last); // Get the Arena on which this RepeatedField stores its elements. ::google::protobuf::Arena* GetArena() const { return GetArenaNoVirtual(); } private: static const int kInitialSize = 0; // A note on the representation here (see also comment below for // RepeatedPtrFieldBase's struct Rep): // // We maintain the same sizeof(RepeatedField) as before we added arena support // so that we do not degrade performance by bloating memory usage. Directly // adding an arena_ element to RepeatedField is quite costly. By using // indirection in this way, we keep the same size when the RepeatedField is // empty (common case), and add only an 8-byte header to the elements array // when non-empty. We make sure to place the size fields directly in the // RepeatedField class to avoid costly cache misses due to the indirection. int current_size_; int total_size_; struct Rep { Arena* arena; Element elements[1]; }; // We can not use sizeof(Rep) - sizeof(Element) due to the trailing padding on // the struct. We can not use sizeof(Arena*) as well because there might be // a "gap" after the field arena and before the field elements (e.g., when // Element is double and pointer is 32bit). static const size_t kRepHeaderSize; // Contains arena ptr and the elements array. We also keep the invariant that // if rep_ is NULL, then arena is NULL. Rep* rep_; friend class Arena; typedef void InternalArenaConstructable_; // Move the contents of |from| into |to|, possibly clobbering |from| in the // process. For primitive types this is just a memcpy(), but it could be // specialized for non-primitive types to, say, swap each element instead. void MoveArray(Element* to, Element* from, int size); // Copy the elements of |from| into |to|. void CopyArray(Element* to, const Element* from, int size); inline void InternalSwap(RepeatedField* other); // Internal helper expected by Arena methods. inline Arena* GetArenaNoVirtual() const { return (rep_ == NULL) ? NULL : rep_->arena; } // Internal helper to delete all elements and deallocate the storage. // If Element has a trivial destructor (for example, if it's a fundamental // type, like int32), the loop will be removed by the optimizer. void InternalDeallocate(Rep* rep, int size) { if (rep != NULL) { Element* e = &rep->elements[0]; Element* limit = &rep->elements[size]; for (; e < limit; e++) { e->Element::~Element(); } if (rep->arena == NULL) { delete[] reinterpret_cast
(rep); } } } }; template
const size_t RepeatedField
::kRepHeaderSize = reinterpret_cast
(&reinterpret_cast
(16)->elements[0]) - 16; namespace internal { template
class RepeatedPtrIterator; template
class RepeatedPtrOverPtrsIterator; } // namespace internal namespace internal { // This is a helper template to copy an array of elements effeciently when they // have a trivial copy constructor, and correctly otherwise. This really // shouldn't be necessary, but our compiler doesn't optimize std::copy very // effectively. template
::value> struct ElementCopier { void operator()(Element* to, const Element* from, int array_size); }; } // namespace internal namespace internal { // type-traits helper for RepeatedPtrFieldBase: we only want to invoke // arena-related "copy if on different arena" behavior if the necessary methods // exist on the contained type. In particular, we rely on MergeFrom() existing // as a general proxy for the fact that a copy will work, and we also provide a // specific override for string*. template
struct TypeImplementsMergeBehavior { typedef char HasMerge; typedef long HasNoMerge; // We accept either of: // - void MergeFrom(const T& other) // - bool MergeFrom(const T& other) // // We mangle these names a bit to avoid compatibility issues in 'unclean' // include environments that may have, e.g., "#define test ..." (yes, this // exists). template
struct CheckType; template
static HasMerge Check( CheckType
*); template
static HasMerge Check( CheckType
*); template
static HasNoMerge Check(...); // Resovles to either google::protobuf::internal::true_type or google::protobuf::internal::false_type. typedef google::protobuf::internal::integral_constant
(0)) == sizeof(HasMerge))> type; }; template<> struct TypeImplementsMergeBehavior< ::std::string > { typedef google::protobuf::internal::true_type type; }; // This is the common base class for RepeatedPtrFields. It deals only in void* // pointers. Users should not use this interface directly. // // The methods of this interface correspond to the methods of RepeatedPtrField, // but may have a template argument called TypeHandler. Its signature is: // class TypeHandler { // public: // typedef MyType Type; // static Type* New(); // static void Delete(Type*); // static void Clear(Type*); // static void Merge(const Type& from, Type* to); // // // Only needs to be implemented if SpaceUsedExcludingSelf() is called. // static int SpaceUsed(const Type&); // }; class LIBPROTOBUF_EXPORT RepeatedPtrFieldBase { protected: // The reflection implementation needs to call protected methods directly, // reinterpreting pointers as being to Message instead of a specific Message // subclass. friend class GeneratedMessageReflection; // ExtensionSet stores repeated message extensions as // RepeatedPtrField
, but non-lite ExtensionSets need to // implement SpaceUsed(), and thus need to call SpaceUsedExcludingSelf() // reinterpreting MessageLite as Message. ExtensionSet also needs to make // use of AddFromCleared(), which is not part of the public interface. friend class ExtensionSet; // The MapFieldBase implementation needs to call protected methods directly, // reinterpreting pointers as being to Message instead of a specific Message // subclass. friend class MapFieldBase; // To parse directly into a proto2 generated class, the upb class GMR_Handlers // needs to be able to modify a RepeatedPtrFieldBase directly. friend class upb::google_opensource::GMR_Handlers; RepeatedPtrFieldBase(); explicit RepeatedPtrFieldBase(::google::protobuf::Arena* arena); ~RepeatedPtrFieldBase() {} // Must be called from destructor. template
void Destroy(); bool empty() const; int size() const; template
const typename TypeHandler::Type& Get(int index) const; template
typename TypeHandler::Type* Mutable(int index); template
void Delete(int index); template
typename TypeHandler::Type* Add(typename TypeHandler::Type* prototype = NULL); template
void RemoveLast(); template
void Clear(); template
void MergeFrom(const RepeatedPtrFieldBase& other); template
void CopyFrom(const RepeatedPtrFieldBase& other); void CloseGap(int start, int num); void Reserve(int new_size); int Capacity() const; // Used for constructing iterators. void* const* raw_data() const; void** raw_mutable_data() const; template
typename TypeHandler::Type** mutable_data(); template
const typename TypeHandler::Type* const* data() const; template
GOOGLE_ATTRIBUTE_ALWAYS_INLINE void Swap(RepeatedPtrFieldBase* other); void SwapElements(int index1, int index2); template
int SpaceUsedExcludingSelf() const; // Advanced memory management -------------------------------------- // Like Add(), but if there are no cleared objects to use, returns NULL. template
typename TypeHandler::Type* AddFromCleared(); template
void AddAllocated(typename TypeHandler::Type* value) { typename TypeImplementsMergeBehavior
::type t; AddAllocatedInternal
(value, t); } template
void UnsafeArenaAddAllocated(typename TypeHandler::Type* value); template
typename TypeHandler::Type* ReleaseLast() { typename TypeImplementsMergeBehavior
::type t; return ReleaseLastInternal
(t); } // Releases last element and returns it, but does not do out-of-arena copy. // And just returns the raw pointer to the contained element in the arena. template
typename TypeHandler::Type* UnsafeArenaReleaseLast(); int ClearedCount() const; template
void AddCleared(typename TypeHandler::Type* value); template
typename TypeHandler::Type* ReleaseCleared(); protected: inline void InternalSwap(RepeatedPtrFieldBase* other); template
void AddAllocatedInternal(typename TypeHandler::Type* value, google::protobuf::internal::true_type); template
void AddAllocatedInternal(typename TypeHandler::Type* value, google::protobuf::internal::false_type); template
GOOGLE_ATTRIBUTE_NOINLINE void AddAllocatedSlowWithCopy(typename TypeHandler::Type* value, Arena* value_arena, Arena* my_arena); template
GOOGLE_ATTRIBUTE_NOINLINE void AddAllocatedSlowWithoutCopy(typename TypeHandler::Type* value); template
typename TypeHandler::Type* ReleaseLastInternal(google::protobuf::internal::true_type); template
typename TypeHandler::Type* ReleaseLastInternal(google::protobuf::internal::false_type); template
GOOGLE_ATTRIBUTE_NOINLINE void SwapFallback(RepeatedPtrFieldBase* other); inline Arena* GetArenaNoVirtual() const { return arena_; } private: static const int kInitialSize = 0; // A few notes on internal representation: // // We use an indirected approach, with struct Rep, to keep // sizeof(RepeatedPtrFieldBase) equivalent to what it was before arena support // was added, namely, 3 8-byte machine words on x86-64. An instance of Rep is // allocated only when the repeated field is non-empty, and it is a // dynamically-sized struct (the header is directly followed by elements[]). // We place arena_ and current_size_ directly in the object to avoid cache // misses due to the indirection, because these fields are checked frequently. // Placing all fields directly in the RepeatedPtrFieldBase instance costs // significant performance for memory-sensitive workloads. Arena* arena_; int current_size_; int total_size_; struct Rep { int allocated_size; void* elements[1]; }; static const size_t kRepHeaderSize = sizeof(Rep) - sizeof(void*); // Contains arena ptr and the elements array. We also keep the invariant that // if rep_ is NULL, then arena is NULL. Rep* rep_; template
static inline typename TypeHandler::Type* cast(void* element) { return reinterpret_cast
(element); } template
static inline const typename TypeHandler::Type* cast(const void* element) { return reinterpret_cast
(element); } // Non-templated inner function to avoid code duplication. Takes a function // pointer to the type-specific (templated) inner allocate/merge loop. void MergeFromInternal( const RepeatedPtrFieldBase& other, void (RepeatedPtrFieldBase::*inner_loop)(void**, void**, int, int)); template
void MergeFromInnerLoop( void** our_elems, void** other_elems, int length, int already_allocated); // Internal helper: extend array space if necessary to contain |extend_amount| // more elements, and return a pointer to the element immediately following // the old list of elements. This interface factors out common behavior from // Reserve() and MergeFrom() to reduce code size. |extend_amount| must be > 0. void** InternalExtend(int extend_amount); GOOGLE_DISALLOW_EVIL_CONSTRUCTORS(RepeatedPtrFieldBase); }; template
class GenericTypeHandler { public: typedef GenericType Type; static inline GenericType* New(Arena* arena) { return ::google::protobuf::Arena::CreateMaybeMessage
( arena, static_cast
(0)); } // We force NewFromPrototype() and Delete() to be non-inline to reduce code // size: else, several other methods get inlined copies of message types' // constructors and destructors. GOOGLE_ATTRIBUTE_NOINLINE static GenericType* NewFromPrototype( const GenericType* prototype, ::google::protobuf::Arena* arena = NULL); GOOGLE_ATTRIBUTE_NOINLINE static void Delete(GenericType* value, Arena* arena); static inline ::google::protobuf::Arena* GetArena(GenericType* value) { return ::google::protobuf::Arena::GetArena
(value); } static inline void* GetMaybeArenaPointer(GenericType* value) { return ::google::protobuf::Arena::GetArena
(value); } static inline void Clear(GenericType* value) { value->Clear(); } GOOGLE_ATTRIBUTE_NOINLINE static void Merge(const GenericType& from, GenericType* to); static inline int SpaceUsed(const GenericType& value) { return value.SpaceUsed(); } static inline const Type& default_instance() { return Type::default_instance(); } }; template
GenericType* GenericTypeHandler
::NewFromPrototype( const GenericType* /* prototype */, ::google::protobuf::Arena* arena) { return New(arena); } template
void GenericTypeHandler
::Delete(GenericType* value, Arena* arena) { if (arena == NULL) { delete value; } } template
void GenericTypeHandler
::Merge(const GenericType& from, GenericType* to) { to->MergeFrom(from); } // NewFromPrototype() and Merge() cannot be defined here; if they're declared // inline the compiler will complain about not matching GOOGLE_ATTRIBUTE_NOINLINE // above, and if not, compilation will result in multiple definitions. These // are therefore declared as specializations here and defined in // message_lite.cc. template<> MessageLite* GenericTypeHandler
::NewFromPrototype( const MessageLite* prototype, google::protobuf::Arena* arena); template<> inline google::protobuf::Arena* GenericTypeHandler
::GetArena( MessageLite* value) { return value->GetArena(); } template<> inline void* GenericTypeHandler
::GetMaybeArenaPointer( MessageLite* value) { return value->GetMaybeArenaPointer(); } template <> void GenericTypeHandler
::Merge(const MessageLite& from, MessageLite* to); template<> inline void GenericTypeHandler
::Clear(string* value) { value->clear(); } template<> void GenericTypeHandler
::Merge(const string& from, string* to); // Declarations of the specialization as we cannot define them here, as the // header that defines ProtocolMessage depends on types defined in this header. #define DECLARE_SPECIALIZATIONS_FOR_BASE_PROTO_TYPES(TypeName) \ template<> \ TypeName* GenericTypeHandler
::NewFromPrototype( \ const TypeName* prototype, google::protobuf::Arena* arena); \ template<> \ google::protobuf::Arena* GenericTypeHandler
::GetArena( \ TypeName* value); \ template<> \ void* GenericTypeHandler
::GetMaybeArenaPointer( \ TypeName* value); // Message specialization bodies defined in message.cc. This split is necessary // to allow proto2-lite (which includes this header) to be independent of // Message. DECLARE_SPECIALIZATIONS_FOR_BASE_PROTO_TYPES(Message) #undef DECLARE_SPECIALIZATIONS_FOR_BASE_PROTO_TYPES template <> inline const MessageLite& GenericTypeHandler
::default_instance() { // Yes, the behavior of the code is undefined, but this function is only // called when we're already deep into the world of undefined, because the // caller called Get(index) out of bounds. MessageLite* null = NULL; return *null; } template <> inline const Message& GenericTypeHandler
::default_instance() { // Yes, the behavior of the code is undefined, but this function is only // called when we're already deep into the world of undefined, because the // caller called Get(index) out of bounds. Message* null = NULL; return *null; } // HACK: If a class is declared as DLL-exported in MSVC, it insists on // generating copies of all its methods -- even inline ones -- to include // in the DLL. But SpaceUsed() calls StringSpaceUsedExcludingSelf() which // isn't in the lite library, therefore the lite library cannot link if // StringTypeHandler is exported. So, we factor out StringTypeHandlerBase, // export that, then make StringTypeHandler be a subclass which is NOT // exported. // TODO(kenton): Now that StringSpaceUsedExcludingSelf() is in the lite // library, this can be cleaned up. class LIBPROTOBUF_EXPORT StringTypeHandlerBase { public: typedef string Type; static inline string* New(Arena* arena) { return Arena::Create
(arena); } static inline string* NewFromPrototype(const string*, ::google::protobuf::Arena* arena) { return New(arena); } static inline ::google::protobuf::Arena* GetArena(string*) { return NULL; } static inline void* GetMaybeArenaPointer(string* /* value */) { return NULL; } static inline void Delete(string* value, Arena* arena) { if (arena == NULL) { delete value; } } static inline void Clear(string* value) { value->clear(); } static inline void Merge(const string& from, string* to) { *to = from; } static inline const Type& default_instance() { return ::google::protobuf::internal::GetEmptyString(); } }; class StringTypeHandler : public StringTypeHandlerBase { public: static int SpaceUsed(const string& value) { return static_cast
(sizeof(value)) + StringSpaceUsedExcludingSelf(value); } }; } // namespace internal // RepeatedPtrField is like RepeatedField, but used for repeated strings or // Messages. template
class RepeatedPtrField : public internal::RepeatedPtrFieldBase { public: RepeatedPtrField(); explicit RepeatedPtrField(::google::protobuf::Arena* arena); RepeatedPtrField(const RepeatedPtrField& other); template
RepeatedPtrField(Iter begin, const Iter& end); ~RepeatedPtrField(); RepeatedPtrField& operator=(const RepeatedPtrField& other); bool empty() const; int size() const; const Element& Get(int index) const; Element* Mutable(int index); Element* Add(); // Remove the last element in the array. // Ownership of the element is retained by the array. void RemoveLast(); // Delete elements with indices in the range [start .. start+num-1]. // Caution: implementation moves all elements with indices [start+num .. ]. // Calling this routine inside a loop can cause quadratic behavior. void DeleteSubrange(int start, int num); void Clear(); void MergeFrom(const RepeatedPtrField& other); void CopyFrom(const RepeatedPtrField& other); // Reserve space to expand the field to at least the given size. This only // resizes the pointer array; it doesn't allocate any objects. If the // array is grown, it will always be at least doubled in size. void Reserve(int new_size); int Capacity() const; // Gets the underlying array. This pointer is possibly invalidated by // any add or remove operation. Element** mutable_data(); const Element* const* data() const; // Swap entire contents with "other". If they are on separate arenas, then // copies data. void Swap(RepeatedPtrField* other); // Swap entire contents with "other". Caller should guarantee that either both // fields are on the same arena or both are on the heap. Swapping between // different arenas with this function is disallowed and is caught via // GOOGLE_DCHECK. void UnsafeArenaSwap(RepeatedPtrField* other); // Swap two elements. void SwapElements(int index1, int index2); // STL-like iterator support typedef internal::RepeatedPtrIterator
iterator; typedef internal::RepeatedPtrIterator
const_iterator; typedef Element value_type; typedef value_type& reference; typedef const value_type& const_reference; typedef value_type* pointer; typedef const value_type* const_pointer; typedef int size_type; typedef ptrdiff_t difference_type; iterator begin(); const_iterator begin() const; const_iterator cbegin() const; iterator end(); const_iterator end() const; const_iterator cend() const; // Reverse iterator support typedef std::reverse_iterator
const_reverse_iterator; typedef std::reverse_iterator
reverse_iterator; reverse_iterator rbegin() { return reverse_iterator(end()); } const_reverse_iterator rbegin() const { return const_reverse_iterator(end()); } reverse_iterator rend() { return reverse_iterator(begin()); } const_reverse_iterator rend() const { return const_reverse_iterator(begin()); } // Custom STL-like iterator that iterates over and returns the underlying // pointers to Element rather than Element itself. typedef internal::RepeatedPtrOverPtrsIterator
pointer_iterator; typedef internal::RepeatedPtrOverPtrsIterator
const_pointer_iterator; pointer_iterator pointer_begin(); const_pointer_iterator pointer_begin() const; pointer_iterator pointer_end(); const_pointer_iterator pointer_end() const; // Returns (an estimate of) the number of bytes used by the repeated field, // excluding sizeof(*this). int SpaceUsedExcludingSelf() const; // Advanced memory management -------------------------------------- // When hardcore memory management becomes necessary -- as it sometimes // does here at Google -- the following methods may be useful. // Add an already-allocated object, passing ownership to the // RepeatedPtrField. // // Note that some special behavior occurs with respect to arenas: // // (i) if this field holds submessages, the new submessage will be copied if // the original is in an arena and this RepeatedPtrField is either in a // different arena, or on the heap. // (ii) if this field holds strings, the passed-in string *must* be // heap-allocated, not arena-allocated. There is no way to dynamically check // this at runtime, so User Beware. void AddAllocated(Element* value); // Remove the last element and return it, passing ownership to the caller. // Requires: size() > 0 // // If this RepeatedPtrField is on an arena, an object copy is required to pass // ownership back to the user (for compatible semantics). Use // UnsafeArenaReleaseLast() if this behavior is undesired. Element* ReleaseLast(); // Add an already-allocated object, skipping arena-ownership checks. The user // must guarantee that the given object is in the same arena as this // RepeatedPtrField. // It is also useful in legacy code that uses temporary ownership to avoid // copies. Example: // RepeatedPtrField
temp_field; // temp_field.AddAllocated(new T); // ... // Do something with temp_field // temp_field.ExtractSubrange(0, temp_field.size(), NULL); // If you put temp_field on the arena this fails, because the ownership // transfers to the arena at the "AddAllocated" call and is not released // anymore causing a double delete. UnsafeArenaAddAllocated prevents this. void UnsafeArenaAddAllocated(Element* value); // Remove the last element and return it. Works only when operating on an // arena. The returned pointer is to the original object in the arena, hence // has the arena's lifetime. // Requires: current_size_ > 0 Element* UnsafeArenaReleaseLast(); // Extract elements with indices in the range "[start .. start+num-1]". // The caller assumes ownership of the extracted elements and is responsible // for deleting them when they are no longer needed. // If "elements" is non-NULL, then pointers to the extracted elements // are stored in "elements[0 .. num-1]" for the convenience of the caller. // If "elements" is NULL, then the caller must use some other mechanism // to perform any further operations (like deletion) on these elements. // Caution: implementation also moves elements with indices [start+num ..]. // Calling this routine inside a loop can cause quadratic behavior. // // Memory copying behavior is identical to ReleaseLast(), described above: if // this RepeatedPtrField is on an arena, an object copy is performed for each // returned element, so that all returned element pointers are to // heap-allocated copies. If this copy is not desired, the user should call // UnsafeArenaExtractSubrange(). void ExtractSubrange(int start, int num, Element** elements); // Identical to ExtractSubrange() described above, except that when this // repeated field is on an arena, no object copies are performed. Instead, the // raw object pointers are returned. Thus, if on an arena, the returned // objects must not be freed, because they will not be heap-allocated objects. void UnsafeArenaExtractSubrange(int start, int num, Element** elements); // When elements are removed by calls to RemoveLast() or Clear(), they // are not actually freed. Instead, they are cleared and kept so that // they can be reused later. This can save lots of CPU time when // repeatedly reusing a protocol message for similar purposes. // // Hardcore programs may choose to manipulate these cleared objects // to better optimize memory management using the following routines. // Get the number of cleared objects that are currently being kept // around for reuse. int ClearedCount() const; // Add an element to the pool of cleared objects, passing ownership to // the RepeatedPtrField. The element must be cleared prior to calling // this method. // // This method cannot be called when the repeated field is on an arena or when // |value| is; both cases will trigger a GOOGLE_DCHECK-failure. void AddCleared(Element* value); // Remove a single element from the cleared pool and return it, passing // ownership to the caller. The element is guaranteed to be cleared. // Requires: ClearedCount() > 0 // // // This method cannot be called when the repeated field is on an arena; doing // so will trigger a GOOGLE_DCHECK-failure. Element* ReleaseCleared(); // Removes the element referenced by position. // // Returns an iterator to the element immediately following the removed // element. // // Invalidates all iterators at or after the removed element, including end(). iterator erase(const_iterator position); // Removes the elements in the range [first, last). // // Returns an iterator to the element immediately following the removed range. // // Invalidates all iterators at or after the removed range, including end(). iterator erase(const_iterator first, const_iterator last); // Gets the arena on which this RepeatedPtrField stores its elements. ::google::protobuf::Arena* GetArena() const { return GetArenaNoVirtual(); } protected: // Note: RepeatedPtrField SHOULD NOT be subclassed by users. We only // subclass it in one place as a hack for compatibility with proto1. The // subclass needs to know about TypeHandler in order to call protected // methods on RepeatedPtrFieldBase. class TypeHandler; // Internal arena accessor expected by helpers in Arena. inline Arena* GetArenaNoVirtual() const; private: // Implementations for ExtractSubrange(). The copying behavior must be // included only if the type supports the necessary operations (e.g., // MergeFrom()), so we must resolve this at compile time. ExtractSubrange() // uses SFINAE to choose one of the below implementations. void ExtractSubrangeInternal(int start, int num, Element** elements, google::protobuf::internal::true_type); void ExtractSubrangeInternal(int start, int num, Element** elements, google::protobuf::internal::false_type); friend class Arena; typedef void InternalArenaConstructable_; }; // implementation ==================================================== template
inline RepeatedField
::RepeatedField() : current_size_(0), total_size_(0), rep_(NULL) { } template
inline RepeatedField
::RepeatedField(Arena* arena) : current_size_(0), total_size_(0), rep_(NULL) { // In case arena is NULL, then we do not create rep_, as code has an invariant // `rep_ == NULL then arena == NULL`. if (arena != NULL) { rep_ = reinterpret_cast
( ::google::protobuf::Arena::CreateArray
(arena, kRepHeaderSize)); rep_->arena = arena; } } template
inline RepeatedField
::RepeatedField(const RepeatedField& other) : current_size_(0), total_size_(0), rep_(NULL) { CopyFrom(other); } template
template
RepeatedField
::RepeatedField(Iter begin, const Iter& end) : current_size_(0), total_size_(0), rep_(NULL) { int reserve = internal::CalculateReserve(begin, end); if (reserve != -1) { Reserve(reserve); for (; begin != end; ++begin) { AddAlreadyReserved(*begin); } } else { for (; begin != end; ++begin) { Add(*begin); } } } template
RepeatedField
::~RepeatedField() { // See explanation in Reserve(): we need to invoke destructors here for the // case that Element has a non-trivial destructor. InternalDeallocate(rep_, total_size_); } template
inline RepeatedField
& RepeatedField
::operator=(const RepeatedField& other) { if (this != &other) CopyFrom(other); return *this; } template
inline bool RepeatedField
::empty() const { return current_size_ == 0; } template
inline int RepeatedField
::size() const { return current_size_; } template
inline int RepeatedField
::Capacity() const { return total_size_; } template
inline void RepeatedField
::AddAlreadyReserved(const Element& value) { GOOGLE_DCHECK_LT(current_size_, total_size_); rep_->elements[current_size_++] = value; } template
inline Element* RepeatedField
::AddAlreadyReserved() { GOOGLE_DCHECK_LT(current_size_, total_size_); return &rep_->elements[current_size_++]; } template
inline void RepeatedField
::Resize(int new_size, const Element& value) { GOOGLE_DCHECK_GE(new_size, 0); if (new_size > current_size_) { Reserve(new_size); std::fill(&rep_->elements[current_size_], &rep_->elements[new_size], value); } current_size_ = new_size; } template
inline const Element& RepeatedField
::Get(int index) const { GOOGLE_DCHECK_GE(index, 0); GOOGLE_DCHECK_LT(index, current_size_); return rep_->elements[index]; } template
inline Element* RepeatedField
::Mutable(int index) { GOOGLE_DCHECK_GE(index, 0); GOOGLE_DCHECK_LT(index, current_size_); return &rep_->elements[index]; } template
inline void RepeatedField
::Set(int index, const Element& value) { GOOGLE_DCHECK_GE(index, 0); GOOGLE_DCHECK_LT(index, current_size_); rep_->elements[index] = value; } template
inline void RepeatedField
::Add(const Element& value) { if (current_size_ == total_size_) Reserve(total_size_ + 1); rep_->elements[current_size_++] = value; } template
inline Element* RepeatedField
::Add() { if (current_size_ == total_size_) Reserve(total_size_ + 1); return &rep_->elements[current_size_++]; } template
inline void RepeatedField
::RemoveLast() { GOOGLE_DCHECK_GT(current_size_, 0); current_size_--; } template
void RepeatedField
::ExtractSubrange( int start, int num, Element* elements) { GOOGLE_DCHECK_GE(start, 0); GOOGLE_DCHECK_GE(num, 0); GOOGLE_DCHECK_LE(start + num, this->current_size_); // Save the values of the removed elements if requested. if (elements != NULL) { for (int i = 0; i < num; ++i) elements[i] = this->Get(i + start); } // Slide remaining elements down to fill the gap. if (num > 0) { for (int i = start + num; i < this->current_size_; ++i) this->Set(i - num, this->Get(i)); this->Truncate(this->current_size_ - num); } } template
inline void RepeatedField
::Clear() { current_size_ = 0; } template
inline void RepeatedField
::MergeFrom(const RepeatedField& other) { GOOGLE_CHECK_NE(&other, this); if (other.current_size_ != 0) { Reserve(current_size_ + other.current_size_); CopyArray(rep_->elements + current_size_, other.rep_->elements, other.current_size_); current_size_ += other.current_size_; } } template
inline void RepeatedField
::CopyFrom(const RepeatedField& other) { if (&other == this) return; Clear(); MergeFrom(other); } template
inline typename RepeatedField
::iterator RepeatedField
::erase( const_iterator position) { return erase(position, position + 1); } template
inline typename RepeatedField
::iterator RepeatedField
::erase( const_iterator first, const_iterator last) { size_type first_offset = first - cbegin(); if (first != last) { Truncate(std::copy(last, cend(), begin() + first_offset) - cbegin()); } return begin() + first_offset; } template
inline Element* RepeatedField
::mutable_data() { return rep_ ? rep_->elements : NULL; } template
inline const Element* RepeatedField
::data() const { return rep_ ? rep_->elements : NULL; } template
inline void RepeatedField
::InternalSwap(RepeatedField* other) { std::swap(rep_, other->rep_); std::swap(current_size_, other->current_size_); std::swap(total_size_, other->total_size_); } template
void RepeatedField
::Swap(RepeatedField* other) { if (this == other) return; if (GetArenaNoVirtual() == other->GetArenaNoVirtual()) { InternalSwap(other); } else { RepeatedField
temp(other->GetArenaNoVirtual()); temp.MergeFrom(*this); CopyFrom(*other); other->UnsafeArenaSwap(&temp); } } template
void RepeatedField
::UnsafeArenaSwap(RepeatedField* other) { if (this == other) return; GOOGLE_DCHECK(GetArenaNoVirtual() == other->GetArenaNoVirtual()); InternalSwap(other); } template
void RepeatedField
::SwapElements(int index1, int index2) { using std::swap; // enable ADL with fallback swap(rep_->elements[index1], rep_->elements[index2]); } template
inline typename RepeatedField
::iterator RepeatedField
::begin() { return rep_ ? rep_->elements : NULL; } template
inline typename RepeatedField
::const_iterator RepeatedField
::begin() const { return rep_ ? rep_->elements : NULL; } template
inline typename RepeatedField
::const_iterator RepeatedField
::cbegin() const { return rep_ ? rep_->elements : NULL; } template
inline typename RepeatedField
::iterator RepeatedField
::end() { return rep_ ? rep_->elements + current_size_ : NULL; } template
inline typename RepeatedField
::const_iterator RepeatedField
::end() const { return rep_ ? rep_->elements + current_size_ : NULL; } template
inline typename RepeatedField
::const_iterator RepeatedField
::cend() const { return rep_ ? rep_->elements + current_size_ : NULL; } template
inline int RepeatedField
::SpaceUsedExcludingSelf() const { return rep_ ? (total_size_ * sizeof(Element) + kRepHeaderSize) : 0; } // Avoid inlining of Reserve(): new, copy, and delete[] lead to a significant // amount of code bloat. template
void RepeatedField
::Reserve(int new_size) { if (total_size_ >= new_size) return; Rep* old_rep = rep_; Arena* arena = GetArenaNoVirtual(); new_size = std::max(google::protobuf::internal::kMinRepeatedFieldAllocationSize, std::max(total_size_ * 2, new_size)); GOOGLE_CHECK_LE(static_cast
(new_size), (std::numeric_limits
::max() - kRepHeaderSize) / sizeof(Element)) << "Requested size is too large to fit into size_t."; if (arena == NULL) { rep_ = reinterpret_cast
( new char[kRepHeaderSize + sizeof(Element) * new_size]); } else { rep_ = reinterpret_cast
( ::google::protobuf::Arena::CreateArray
(arena, kRepHeaderSize + sizeof(Element) * new_size)); } rep_->arena = arena; int old_total_size = total_size_; total_size_ = new_size; // Invoke placement-new on newly allocated elements. We shouldn't have to do // this, since Element is supposed to be POD, but a previous version of this // code allocated storage with "new Element[size]" and some code uses // RepeatedField with non-POD types, relying on constructor invocation. If // Element has a trivial constructor (e.g., int32), gcc (tested with -O2) // completely removes this loop because the loop body is empty, so this has no // effect unless its side-effects are required for correctness. // Note that we do this before MoveArray() below because Element's copy // assignment implementation will want an initialized instance first. Element* e = &rep_->elements[0]; Element* limit = &rep_->elements[total_size_]; for (; e < limit; e++) { new (e) Element(); } if (current_size_ > 0) { MoveArray(rep_->elements, old_rep->elements, current_size_); } // Likewise, we need to invoke destructors on the old array. InternalDeallocate(old_rep, old_total_size); } template
inline void RepeatedField
::Truncate(int new_size) { GOOGLE_DCHECK_LE(new_size, current_size_); if (current_size_ > 0) { current_size_ = new_size; } } template
inline void RepeatedField
::MoveArray( Element* to, Element* from, int array_size) { CopyArray(to, from, array_size); } template
inline void RepeatedField
::CopyArray( Element* to, const Element* from, int array_size) { internal::ElementCopier
()(to, from, array_size); } namespace internal { template
void ElementCopier
::operator()( Element* to, const Element* from, int array_size) { std::copy(from, from + array_size, to); } template
struct ElementCopier
{ void operator()(Element* to, const Element* from, int array_size) { memcpy(to, from, array_size * sizeof(Element)); } }; } // namespace internal // ------------------------------------------------------------------- namespace internal { inline RepeatedPtrFieldBase::RepeatedPtrFieldBase() : arena_(NULL), current_size_(0), total_size_(0), rep_(NULL) { } inline RepeatedPtrFieldBase::RepeatedPtrFieldBase(::google::protobuf::Arena* arena) : arena_(arena), current_size_(0), total_size_(0), rep_(NULL) { } template
void RepeatedPtrFieldBase::Destroy() { if (rep_ != NULL) { for (int i = 0; i < rep_->allocated_size; i++) { TypeHandler::Delete(cast
(rep_->elements[i]), arena_); } if (arena_ == NULL) { delete [] reinterpret_cast
(rep_); } } rep_ = NULL; } template
inline void RepeatedPtrFieldBase::Swap(RepeatedPtrFieldBase* other) { if (other->GetArenaNoVirtual() == GetArenaNoVirtual()) { InternalSwap(other); } else { SwapFallback
(other); } } template
void RepeatedPtrFieldBase::SwapFallback(RepeatedPtrFieldBase* other) { GOOGLE_DCHECK(other->GetArenaNoVirtual() != GetArenaNoVirtual()); // Copy semantics in this case. We try to improve efficiency by placing the // temporary on |other|'s arena so that messages are copied cross-arena only // once, not twice. RepeatedPtrFieldBase temp(other->GetArenaNoVirtual()); temp.MergeFrom
(*this); this->Clear
(); this->MergeFrom
(*other); other->Clear
(); other->InternalSwap(&temp); temp.Destroy
(); // Frees rep_ if `other` had no arena. } inline bool RepeatedPtrFieldBase::empty() const { return current_size_ == 0; } inline int RepeatedPtrFieldBase::size() const { return current_size_; } template
inline const typename TypeHandler::Type& RepeatedPtrFieldBase::Get(int index) const { GOOGLE_DCHECK_GE(index, 0); GOOGLE_DCHECK_LT(index, current_size_); return *cast
(rep_->elements[index]); } template
inline typename TypeHandler::Type* RepeatedPtrFieldBase::Mutable(int index) { GOOGLE_DCHECK_GE(index, 0); GOOGLE_DCHECK_LT(index, current_size_); return cast
(rep_->elements[index]); } template
inline void RepeatedPtrFieldBase::Delete(int index) { GOOGLE_DCHECK_GE(index, 0); GOOGLE_DCHECK_LT(index, current_size_); TypeHandler::Delete(cast
(rep_->elements[index]), arena_); } template
inline typename TypeHandler::Type* RepeatedPtrFieldBase::Add( typename TypeHandler::Type* prototype) { if (rep_ != NULL && current_size_ < rep_->allocated_size) { return cast
(rep_->elements[current_size_++]); } if (!rep_ || rep_->allocated_size == total_size_) { Reserve(total_size_ + 1); } ++rep_->allocated_size; typename TypeHandler::Type* result = TypeHandler::NewFromPrototype(prototype, arena_); rep_->elements[current_size_++] = result; return result; } template
inline void RepeatedPtrFieldBase::RemoveLast() { GOOGLE_DCHECK_GT(current_size_, 0); TypeHandler::Clear(cast
(rep_->elements[--current_size_])); } template
void RepeatedPtrFieldBase::Clear() { const int n = current_size_; GOOGLE_DCHECK_GE(n, 0); if (n > 0) { void* const* elements = rep_->elements; int i = 0; do { TypeHandler::Clear(cast
(elements[i++])); } while (i < n); current_size_ = 0; } } // To avoid unnecessary code duplication and reduce binary size, we use a // layered approach to implementing MergeFrom(). The toplevel method is // templated, so we get a small thunk per concrete message type in the binary. // This calls a shared implementation with most of the logic, passing a function // pointer to another type-specific piece of code that calls the object-allocate // and merge handlers. template
inline void RepeatedPtrFieldBase::MergeFrom(const RepeatedPtrFieldBase& other) { GOOGLE_DCHECK_NE(&other, this); if (other.current_size_ == 0) return; MergeFromInternal( other, &RepeatedPtrFieldBase::MergeFromInnerLoop
); } inline void RepeatedPtrFieldBase::MergeFromInternal( const RepeatedPtrFieldBase& other, void (RepeatedPtrFieldBase::*inner_loop)(void**, void**, int, int)) { // Note: wrapper has already guaranteed that other.rep_ != NULL here. int other_size = other.current_size_; void** other_elements = other.rep_->elements; void** new_elements = InternalExtend(other_size); int allocated_elems = rep_->allocated_size - current_size_; (this->*inner_loop)(new_elements, other_elements, other_size, allocated_elems); current_size_ += other_size; if (rep_->allocated_size < current_size_) { rep_->allocated_size = current_size_; } } // Merges other_elems to our_elems. template
void RepeatedPtrFieldBase::MergeFromInnerLoop( void** our_elems, void** other_elems, int length, int already_allocated) { // Split into two loops, over ranges [0, allocated) and [allocated, length), // to avoid a branch within the loop. for (int i = 0; i < already_allocated && i < length; i++) { // Already allocated: use existing element. typename TypeHandler::Type* other_elem = reinterpret_cast
(other_elems[i]); typename TypeHandler::Type* new_elem = reinterpret_cast
(our_elems[i]); TypeHandler::Merge(*other_elem, new_elem); } Arena* arena = GetArenaNoVirtual(); for (int i = already_allocated; i < length; i++) { // Not allocated: alloc a new element first, then merge it. typename TypeHandler::Type* other_elem = reinterpret_cast
(other_elems[i]); typename TypeHandler::Type* new_elem = TypeHandler::NewFromPrototype(other_elem, arena); TypeHandler::Merge(*other_elem, new_elem); our_elems[i] = new_elem; } } template
inline void RepeatedPtrFieldBase::CopyFrom(const RepeatedPtrFieldBase& other) { if (&other == this) return; RepeatedPtrFieldBase::Clear
(); RepeatedPtrFieldBase::MergeFrom
(other); } inline int RepeatedPtrFieldBase::Capacity() const { return total_size_; } inline void* const* RepeatedPtrFieldBase::raw_data() const { return rep_ ? rep_->elements : NULL; } inline void** RepeatedPtrFieldBase::raw_mutable_data() const { return rep_ ? const_cast
(rep_->elements) : NULL; } template
inline typename TypeHandler::Type** RepeatedPtrFieldBase::mutable_data() { // TODO(kenton): Breaks C++ aliasing rules. We should probably remove this // method entirely. return reinterpret_cast
(raw_mutable_data()); } template
inline const typename TypeHandler::Type* const* RepeatedPtrFieldBase::data() const { // TODO(kenton): Breaks C++ aliasing rules. We should probably remove this // method entirely. return reinterpret_cast
(raw_data()); } inline void RepeatedPtrFieldBase::SwapElements(int index1, int index2) { using std::swap; // enable ADL with fallback swap(rep_->elements[index1], rep_->elements[index2]); } template
inline int RepeatedPtrFieldBase::SpaceUsedExcludingSelf() const { int allocated_bytes = total_size_ * sizeof(void*); if (rep_ != NULL) { for (int i = 0; i < rep_->allocated_size; ++i) { allocated_bytes += TypeHandler::SpaceUsed( *cast
(rep_->elements[i])); } allocated_bytes += kRepHeaderSize; } return allocated_bytes; } template
inline typename TypeHandler::Type* RepeatedPtrFieldBase::AddFromCleared() { if (rep_ != NULL && current_size_ < rep_->allocated_size) { return cast
(rep_->elements[current_size_++]); } else { return NULL; } } // AddAllocated version that implements arena-safe copying behavior. template
void RepeatedPtrFieldBase::AddAllocatedInternal( typename TypeHandler::Type* value, google::protobuf::internal::true_type) { Arena* element_arena = reinterpret_cast
( TypeHandler::GetMaybeArenaPointer(value)); Arena* arena = GetArenaNoVirtual(); if (arena == element_arena && rep_ && rep_->allocated_size < total_size_) { // Fast path: underlying arena representation (tagged pointer) is equal to // our arena pointer, and we can add to array without resizing it (at least // one slot that is not allocated). void** elems = rep_->elements; if (current_size_ < rep_->allocated_size) { // Make space at [current] by moving first allocated element to end of // allocated list. elems[rep_->allocated_size] = elems[current_size_]; } elems[current_size_] = value; current_size_ = current_size_ + 1; rep_->allocated_size = rep_->allocated_size + 1; return; } else { AddAllocatedSlowWithCopy
( value, TypeHandler::GetArena(value), arena); } } // Slowpath handles all cases, copying if necessary. template