1662 lines
42 KiB
C++
1662 lines
42 KiB
C++
/* -*- Mode:C++; c-file-style:"gnu"; indent-tabs-mode:nil; -*- */
|
|
/*
|
|
* Copyright (c) 2005,2006,2007 INRIA
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License version 2 as
|
|
* published by the Free Software Foundation;
|
|
*
|
|
* This program is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with this program; if not, write to the Free Software
|
|
* Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
|
|
*
|
|
* Author: Mathieu Lacage <mathieu.lacage@sophia.inria.fr>
|
|
*/
|
|
#include "buffer.h"
|
|
#include "ns3/assert.h"
|
|
#include "ns3/log.h"
|
|
#include "ns3/fatal-error.h"
|
|
#include "ns3/test.h"
|
|
#include "ns3/random-variable.h"
|
|
#include <iomanip>
|
|
#include <iostream>
|
|
|
|
NS_LOG_COMPONENT_DEFINE ("Buffer");
|
|
|
|
#define LOG_INTERNAL_STATE(y) \
|
|
NS_LOG_LOGIC (y << "start="<<m_start<<", end="<<m_end<<", zero start="<<m_zeroAreaStart<< \
|
|
", zero end="<<m_zeroAreaEnd<<", count="<<m_data->m_count<<", size="<<m_data->m_size<< \
|
|
", dirty start="<<m_data->m_dirtyStart<<", dirty end="<<m_data->m_dirtyEnd)
|
|
|
|
#ifdef BUFFER_HEURISTICS
|
|
#define HEURISTICS(x) x
|
|
#else
|
|
#define HEURISTICS(x)
|
|
#endif
|
|
|
|
namespace {
|
|
|
|
static struct Zeroes
|
|
{
|
|
Zeroes ()
|
|
: size (1000)
|
|
{
|
|
memset (buffer, 0, size);
|
|
}
|
|
char buffer[1000];
|
|
const uint32_t size;
|
|
} g_zeroes;
|
|
|
|
}
|
|
|
|
//#define PRINT_STATS 1
|
|
|
|
namespace ns3 {
|
|
|
|
/**
|
|
* This data structure is variable-sized through its last member whose size
|
|
* is determined at allocation time and stored in the m_size field.
|
|
*
|
|
* The so-called "dirty area" describes the area in the buffer which
|
|
* has been reserved and used by a user. Multiple Buffer instances
|
|
* may reference the same BufferData object instance and may
|
|
* reference different parts of the underlying byte buffer. The
|
|
* "dirty area" is union of all the areas referenced by the Buffer
|
|
* instances which reference the same BufferData instance.
|
|
* New user data can be safely written only outside of the "dirty
|
|
* area" if the reference count is higher than 1 (that is, if
|
|
* more than one Buffer instance references the same BufferData).
|
|
*/
|
|
struct BufferData {
|
|
/* The reference count of an instance of this data structure.
|
|
* Each buffer which references an instance holds a count.
|
|
*/
|
|
uint32_t m_count;
|
|
/* the size of the m_data field below.
|
|
*/
|
|
uint32_t m_size;
|
|
/* offset from the start of the m_data field below to the
|
|
* start of the area in which user bytes were written.
|
|
*/
|
|
uint32_t m_dirtyStart;
|
|
/* offset from the start of the m_data field below to the
|
|
* end of the area in which user bytes were written.
|
|
*/
|
|
uint32_t m_dirtyEnd;
|
|
/* The real data buffer holds _at least_ one byte.
|
|
* Its real size is stored in the m_size field.
|
|
*/
|
|
uint8_t m_data[1];
|
|
};
|
|
typedef std::vector<struct BufferData*> BufferDataList;
|
|
|
|
static struct BufferData *BufferAllocate (uint32_t reqSize);
|
|
|
|
static void BufferDeallocate (struct BufferData *data);
|
|
|
|
|
|
} // namespace ns3
|
|
|
|
namespace ns3 {
|
|
|
|
#ifdef BUFFER_HEURISTICS
|
|
/* The following macros are pretty evil but they are needed to allow us to
|
|
* keep track of 3 possible states for the g_freeList variable:
|
|
* - uninitialized means that no one has created a buffer yet
|
|
* so no one has created the associated free list (it is created
|
|
* on-demand when the first buffer is created)
|
|
* - initialized means that the free list exists and is valid
|
|
* - destroyed means that the static destructors of this compilation unit
|
|
* have run so, the free list has been cleared from its content
|
|
* The key is that in destroyed state, we are careful not re-create it
|
|
* which is a typical weakness of lazy evaluation schemes which use
|
|
* '0' as a special value to indicate both un-initialized and destroyed.
|
|
* Note that it is important to use '0' as the marker for un-initialized state
|
|
* because the variable holding this state information is initialized to zero
|
|
* which the compiler assigns to zero-memory which is initialized to _zero_
|
|
* before the constructors run so this ensures perfect handling of crazy
|
|
* constructor orderings.
|
|
*/
|
|
#define MAGIC_DESTROYED (~(long) 0)
|
|
#define IS_UNINITIALIZED(x) (x == (BufferDataList*)0)
|
|
#define IS_DESTROYED(x) (x == (BufferDataList*)MAGIC_DESTROYED)
|
|
#define IS_INITIALIZED(x) (!IS_UNINITIALIZED(x) && !IS_DESTROYED(x))
|
|
#define DESTROYED ((BufferDataList*)MAGIC_DESTROYED)
|
|
#define UNINITIALIZED ((BufferDataList*)0)
|
|
static uint32_t g_recommendedStart = 0;
|
|
static uint64_t g_nAddNoRealloc = 0;
|
|
static uint64_t g_nAddRealloc = 0;
|
|
static BufferDataList *g_freeList = 0;
|
|
static uint32_t g_maxSize = 0;
|
|
static uint64_t g_nAllocs = 0;
|
|
static uint64_t g_nCreates = 0;
|
|
#endif /* BUFFER_HEURISTICS */
|
|
|
|
static struct LocalStaticDestructor {
|
|
~LocalStaticDestructor(void)
|
|
{
|
|
#ifdef PRINT_STATS
|
|
#ifdef BUFFER_HEURISTICS
|
|
double efficiency;
|
|
efficiency = g_nAllocs;
|
|
efficiency /= g_nCreates;
|
|
std::cout <<"buffer free list efficiency="<<efficiency<<" (lower is better)" << std::endl;
|
|
std::cout <<"buffer free list max size="<<g_maxSize<<std::endl;
|
|
std::cout <<"buffer free list recommended start="<<g_recommendedStart<<std::endl;
|
|
double addEfficiency;
|
|
addEfficiency = g_nAddRealloc;
|
|
addEfficiency /= g_nAddNoRealloc;
|
|
std::cout <<"buffer add efficiency=" << addEfficiency << " (lower is better)"<<std::endl;
|
|
//std::cout <<"n add reallocs="<< g_nAddRealloc << std::endl;
|
|
//std::cout <<"n add no reallocs="<< g_nAddNoRealloc << std::endl;
|
|
#endif /* BUFFER_HEURISTICS */
|
|
#endif /* PRINT_STATS */
|
|
if (IS_INITIALIZED(g_freeList))
|
|
{
|
|
for (BufferDataList::iterator i = g_freeList->begin ();
|
|
i != g_freeList->end (); i++)
|
|
{
|
|
BufferDeallocate (*i);
|
|
}
|
|
delete g_freeList;
|
|
g_freeList = DESTROYED;
|
|
}
|
|
}
|
|
} g_localStaticDestructor;
|
|
|
|
struct BufferData *
|
|
BufferAllocate (uint32_t reqSize)
|
|
{
|
|
if (reqSize == 0)
|
|
{
|
|
reqSize = 1;
|
|
}
|
|
NS_ASSERT (reqSize >= 1);
|
|
uint32_t size = reqSize - 1 + sizeof (struct BufferData);
|
|
uint8_t *b = new uint8_t [size];
|
|
struct BufferData *data = reinterpret_cast<struct BufferData*>(b);
|
|
data->m_size = reqSize;
|
|
data->m_count = 1;
|
|
return data;
|
|
}
|
|
|
|
void
|
|
BufferDeallocate (struct BufferData *data)
|
|
{
|
|
NS_ASSERT (data->m_count == 0);
|
|
uint8_t *buf = reinterpret_cast<uint8_t *> (data);
|
|
delete [] buf;
|
|
}
|
|
#ifdef BUFFER_HEURISTICS
|
|
void
|
|
Buffer::Recycle (struct BufferData *data)
|
|
{
|
|
NS_ASSERT (data->m_count == 0);
|
|
NS_ASSERT (!IS_UNINITIALIZED(g_freeList));
|
|
g_maxSize = std::max (g_maxSize, data->m_size);
|
|
/* feed into free list */
|
|
if (data->m_size < g_maxSize ||
|
|
IS_DESTROYED(g_freeList) ||
|
|
g_freeList->size () > 1000)
|
|
{
|
|
BufferDeallocate (data);
|
|
}
|
|
else
|
|
{
|
|
NS_ASSERT (IS_INITIALIZED(g_freeList));
|
|
g_freeList->push_back (data);
|
|
}
|
|
}
|
|
|
|
BufferData *
|
|
Buffer::Create (uint32_t dataSize)
|
|
{
|
|
/* try to find a buffer correctly sized. */
|
|
g_nCreates++;
|
|
if (IS_UNINITIALIZED(g_freeList))
|
|
{
|
|
g_freeList = new BufferDataList ();
|
|
}
|
|
else if (IS_INITIALIZED(g_freeList))
|
|
{
|
|
while (!g_freeList->empty ())
|
|
{
|
|
struct BufferData *data = g_freeList->back ();
|
|
g_freeList->pop_back ();
|
|
if (data->m_size >= dataSize)
|
|
{
|
|
data->m_count = 1;
|
|
return data;
|
|
}
|
|
BufferDeallocate (data);
|
|
}
|
|
}
|
|
g_nAllocs++;
|
|
struct BufferData *data = BufferAllocate (dataSize);
|
|
NS_ASSERT (data->m_count == 1);
|
|
return data;
|
|
}
|
|
#else
|
|
void
|
|
Buffer::Recycle (struct BufferData *data)
|
|
{
|
|
NS_ASSERT (data->m_count == 0);
|
|
BufferDeallocate (data);
|
|
}
|
|
|
|
BufferData *
|
|
Buffer::Create (uint32_t size)
|
|
{
|
|
return BufferAllocate (size);
|
|
}
|
|
#endif
|
|
|
|
Buffer::Buffer ()
|
|
{
|
|
NS_LOG_FUNCTION (this);
|
|
Initialize (0);
|
|
}
|
|
|
|
Buffer::Buffer (uint32_t dataSize)
|
|
{
|
|
NS_LOG_FUNCTION (this << dataSize);
|
|
Initialize (dataSize);
|
|
}
|
|
|
|
Buffer::Buffer (uint32_t dataSize, bool initialize)
|
|
{
|
|
NS_LOG_FUNCTION (this << dataSize << initialize);
|
|
if (initialize == true)
|
|
{
|
|
Initialize (dataSize);
|
|
}
|
|
}
|
|
|
|
bool
|
|
Buffer::CheckInternalState (void) const
|
|
{
|
|
#if 0
|
|
// If you want to modify any code in this file, enable this checking code.
|
|
// Otherwise, there is not much point is enabling it because the
|
|
// current implementation has been fairly seriously tested and the cost
|
|
// of this constant checking is pretty high, even for a debug build.
|
|
bool offsetsOk =
|
|
m_start <= m_zeroAreaStart &&
|
|
m_zeroAreaStart <= m_zeroAreaEnd &&
|
|
m_zeroAreaEnd <= m_end;
|
|
bool dirtyOk =
|
|
m_start >= m_data->m_dirtyStart &&
|
|
m_end <= m_data->m_dirtyEnd;
|
|
bool internalSizeOk = m_end - (m_zeroAreaEnd - m_zeroAreaStart) <= m_data->m_size &&
|
|
m_start <= m_data->m_size &&
|
|
m_zeroAreaStart <= m_data->m_size;
|
|
|
|
bool ok = m_data->m_count > 0 && offsetsOk && dirtyOk && internalSizeOk;
|
|
if (!ok)
|
|
{
|
|
LOG_INTERNAL_STATE ("check " << this <<
|
|
", " << (offsetsOk?"true":"false") <<
|
|
", " << (dirtyOk?"true":"false") <<
|
|
", " << (internalSizeOk?"true":"false") << " ");
|
|
}
|
|
return ok;
|
|
#else
|
|
return true;
|
|
#endif
|
|
}
|
|
|
|
void
|
|
Buffer::Initialize (uint32_t zeroSize)
|
|
{
|
|
NS_LOG_FUNCTION (this << zeroSize);
|
|
m_data = Buffer::Create (0);
|
|
#ifdef BUFFER_HEURISTICS
|
|
m_start = std::min (m_data->m_size, g_recommendedStart);
|
|
m_maxZeroAreaStart = m_start;
|
|
#else
|
|
m_start = 0;
|
|
#endif /* BUFFER_HEURISTICS */
|
|
m_zeroAreaStart = m_start;
|
|
m_zeroAreaEnd = m_zeroAreaStart + zeroSize;
|
|
m_end = m_zeroAreaEnd;
|
|
m_data->m_dirtyStart = m_start;
|
|
m_data->m_dirtyEnd = m_end;
|
|
NS_ASSERT (CheckInternalState ());
|
|
}
|
|
|
|
Buffer::Buffer (Buffer const&o)
|
|
: m_data (o.m_data),
|
|
#ifdef BUFFER_HEURISTICS
|
|
m_maxZeroAreaStart (o.m_zeroAreaStart),
|
|
#endif
|
|
m_zeroAreaStart (o.m_zeroAreaStart),
|
|
m_zeroAreaEnd (o.m_zeroAreaEnd),
|
|
m_start (o.m_start),
|
|
m_end (o.m_end)
|
|
{
|
|
NS_LOG_FUNCTION (this << &o);
|
|
m_data->m_count++;
|
|
NS_ASSERT (CheckInternalState ());
|
|
}
|
|
|
|
Buffer &
|
|
Buffer::operator = (Buffer const&o)
|
|
{
|
|
NS_LOG_FUNCTION (this << &o);
|
|
NS_ASSERT (CheckInternalState ());
|
|
if (m_data != o.m_data)
|
|
{
|
|
// not assignment to self.
|
|
m_data->m_count--;
|
|
if (m_data->m_count == 0)
|
|
{
|
|
Recycle (m_data);
|
|
}
|
|
m_data = o.m_data;
|
|
m_data->m_count++;
|
|
}
|
|
HEURISTICS (
|
|
g_recommendedStart = std::max (g_recommendedStart, m_maxZeroAreaStart);
|
|
m_maxZeroAreaStart = o.m_maxZeroAreaStart;
|
|
);
|
|
m_zeroAreaStart = o.m_zeroAreaStart;
|
|
m_zeroAreaEnd = o.m_zeroAreaEnd;
|
|
m_start = o.m_start;
|
|
m_end = o.m_end;
|
|
NS_ASSERT (CheckInternalState ());
|
|
return *this;
|
|
}
|
|
|
|
Buffer::~Buffer ()
|
|
{
|
|
NS_LOG_FUNCTION (this);
|
|
NS_ASSERT (CheckInternalState ());
|
|
HEURISTICS (g_recommendedStart = std::max (g_recommendedStart, m_maxZeroAreaStart));
|
|
m_data->m_count--;
|
|
if (m_data->m_count == 0)
|
|
{
|
|
Recycle (m_data);
|
|
}
|
|
}
|
|
|
|
Buffer::Iterator
|
|
Buffer::Begin (void) const
|
|
{
|
|
NS_ASSERT (CheckInternalState ());
|
|
return Buffer::Iterator (this);
|
|
}
|
|
Buffer::Iterator
|
|
Buffer::End (void) const
|
|
{
|
|
NS_ASSERT (CheckInternalState ());
|
|
return Buffer::Iterator (this, false);
|
|
}
|
|
|
|
uint32_t
|
|
Buffer::GetInternalSize (void) const
|
|
{
|
|
return m_zeroAreaStart - m_start + m_end - m_zeroAreaEnd;
|
|
}
|
|
uint32_t
|
|
Buffer::GetInternalEnd (void) const
|
|
{
|
|
return m_end - (m_zeroAreaEnd - m_zeroAreaStart);
|
|
}
|
|
|
|
bool
|
|
Buffer::AddAtStart (uint32_t start)
|
|
{
|
|
NS_LOG_FUNCTION (this << start);
|
|
bool dirty;
|
|
NS_ASSERT (CheckInternalState ());
|
|
bool isDirty = m_data->m_count > 1 && m_start > m_data->m_dirtyStart;
|
|
if (m_start >= start && !isDirty)
|
|
{
|
|
/* enough space in the buffer and not dirty.
|
|
* To add: |..|
|
|
* Before: |*****---------***|
|
|
* After: |***..---------***|
|
|
*/
|
|
NS_ASSERT (m_data->m_count == 1 || m_start == m_data->m_dirtyStart);
|
|
m_start -= start;
|
|
dirty = m_start > m_data->m_dirtyStart;
|
|
// update dirty area
|
|
m_data->m_dirtyStart = m_start;
|
|
HEURISTICS (g_nAddNoRealloc++);
|
|
}
|
|
else
|
|
{
|
|
uint32_t newSize = GetInternalSize () + start;
|
|
struct BufferData *newData = Buffer::Create (newSize);
|
|
memcpy (newData->m_data + start, m_data->m_data + m_start, GetInternalSize ());
|
|
m_data->m_count--;
|
|
if (m_data->m_count == 0)
|
|
{
|
|
Buffer::Recycle (m_data);
|
|
}
|
|
m_data = newData;
|
|
|
|
int32_t delta = start - m_start;
|
|
m_start += delta;
|
|
m_zeroAreaStart += delta;
|
|
m_zeroAreaEnd += delta;
|
|
m_end += delta;
|
|
m_start -= start;
|
|
|
|
// update dirty area
|
|
m_data->m_dirtyStart = m_start;
|
|
m_data->m_dirtyEnd = m_end;
|
|
|
|
dirty = true;
|
|
|
|
HEURISTICS (g_nAddRealloc++);
|
|
}
|
|
HEURISTICS (m_maxZeroAreaStart = std::max (m_maxZeroAreaStart, m_zeroAreaStart));
|
|
LOG_INTERNAL_STATE ("add start=" << start << ", ");
|
|
NS_ASSERT (CheckInternalState ());
|
|
return dirty;
|
|
}
|
|
bool
|
|
Buffer::AddAtEnd (uint32_t end)
|
|
{
|
|
NS_LOG_FUNCTION (this << end);
|
|
bool dirty;
|
|
NS_ASSERT (CheckInternalState ());
|
|
bool isDirty = m_data->m_count > 1 && m_end < m_data->m_dirtyEnd;
|
|
if (GetInternalEnd () + end <= m_data->m_size && !isDirty)
|
|
{
|
|
/* enough space in buffer and not dirty
|
|
* Add: |...|
|
|
* Before: |**----*****|
|
|
* After: |**----...**|
|
|
*/
|
|
NS_ASSERT (m_data->m_count == 1 || m_end == m_data->m_dirtyEnd);
|
|
m_end += end;
|
|
// update dirty area.
|
|
m_data->m_dirtyEnd = m_end;
|
|
|
|
dirty = m_end < m_data->m_dirtyEnd;
|
|
|
|
HEURISTICS (g_nAddNoRealloc++);
|
|
}
|
|
else
|
|
{
|
|
uint32_t newSize = GetInternalSize () + end;
|
|
struct BufferData *newData = Buffer::Create (newSize);
|
|
memcpy (newData->m_data, m_data->m_data + m_start, GetInternalSize ());
|
|
m_data->m_count--;
|
|
if (m_data->m_count == 0)
|
|
{
|
|
Buffer::Recycle (m_data);
|
|
}
|
|
m_data = newData;
|
|
|
|
int32_t delta = -m_start;
|
|
m_zeroAreaStart += delta;
|
|
m_zeroAreaEnd += delta;
|
|
m_end += delta;
|
|
m_start += delta;
|
|
m_end += end;
|
|
|
|
// update dirty area
|
|
m_data->m_dirtyStart = m_start;
|
|
m_data->m_dirtyEnd = m_end;
|
|
|
|
dirty = true;
|
|
|
|
HEURISTICS (g_nAddRealloc++);
|
|
}
|
|
HEURISTICS (m_maxZeroAreaStart = std::max (m_maxZeroAreaStart, m_zeroAreaStart));
|
|
LOG_INTERNAL_STATE ("add end=" << end << ", ");
|
|
NS_ASSERT (CheckInternalState ());
|
|
|
|
return dirty;
|
|
}
|
|
|
|
void
|
|
Buffer::AddAtEnd (const Buffer &o)
|
|
{
|
|
NS_LOG_FUNCTION (this << &o);
|
|
if (m_data->m_count == 1 &&
|
|
m_end == m_zeroAreaEnd &&
|
|
m_end == m_data->m_dirtyEnd &&
|
|
o.m_start == o.m_zeroAreaStart &&
|
|
o.m_zeroAreaEnd - o.m_zeroAreaStart > 0)
|
|
{
|
|
/**
|
|
* This is an optimization which kicks in when
|
|
* we attempt to aggregate two buffers which contain
|
|
* adjacent zero areas.
|
|
*/
|
|
uint32_t zeroSize = o.m_zeroAreaEnd - o.m_zeroAreaStart;
|
|
m_zeroAreaEnd += zeroSize;
|
|
m_end = m_zeroAreaEnd;
|
|
m_data->m_dirtyEnd = m_zeroAreaEnd;
|
|
uint32_t endData = o.m_end - o.m_zeroAreaEnd;
|
|
AddAtEnd (endData);
|
|
Buffer::Iterator dst = End ();
|
|
dst.Prev (endData);
|
|
Buffer::Iterator src = o.End ();
|
|
src.Prev (endData);
|
|
dst.Write (src, o.End ());
|
|
NS_ASSERT (CheckInternalState ());
|
|
return;
|
|
}
|
|
|
|
Buffer dst = CreateFullCopy ();
|
|
Buffer src = o.CreateFullCopy ();
|
|
|
|
dst.AddAtEnd (src.GetSize ());
|
|
Buffer::Iterator destStart = dst.End ();
|
|
destStart.Prev (src.GetSize ());
|
|
destStart.Write (src.Begin (), src.End ());
|
|
*this = dst;
|
|
NS_ASSERT (CheckInternalState ());
|
|
}
|
|
|
|
void
|
|
Buffer::RemoveAtStart (uint32_t start)
|
|
{
|
|
NS_LOG_FUNCTION (this << start);
|
|
NS_ASSERT (CheckInternalState ());
|
|
uint32_t newStart = m_start + start;
|
|
if (newStart <= m_zeroAreaStart)
|
|
{
|
|
/* only remove start of buffer
|
|
*/
|
|
m_start = newStart;
|
|
}
|
|
else if (newStart <= m_zeroAreaEnd)
|
|
{
|
|
/* remove start of buffer _and_ start of zero area
|
|
*/
|
|
uint32_t delta = newStart - m_zeroAreaStart;
|
|
m_start = m_zeroAreaStart;
|
|
m_zeroAreaEnd -= delta;
|
|
m_end -= delta;
|
|
}
|
|
else if (newStart <= m_end)
|
|
{
|
|
/* remove start of buffer, complete zero area, and part
|
|
* of end of buffer
|
|
*/
|
|
NS_ASSERT (m_end >= start);
|
|
uint32_t zeroSize = m_zeroAreaEnd - m_zeroAreaStart;
|
|
m_start = newStart - zeroSize;
|
|
m_end -= zeroSize;
|
|
m_zeroAreaStart = m_start;
|
|
m_zeroAreaEnd = m_start;
|
|
}
|
|
else
|
|
{
|
|
/* remove all buffer */
|
|
m_end -= m_zeroAreaEnd - m_zeroAreaStart;
|
|
m_start = m_end;
|
|
m_zeroAreaEnd = m_end;
|
|
m_zeroAreaStart = m_end;
|
|
}
|
|
HEURISTICS (m_maxZeroAreaStart = std::max (m_maxZeroAreaStart, m_zeroAreaStart));
|
|
LOG_INTERNAL_STATE ("rem start=" << start << ", ");
|
|
NS_ASSERT (CheckInternalState ());
|
|
}
|
|
void
|
|
Buffer::RemoveAtEnd (uint32_t end)
|
|
{
|
|
NS_LOG_FUNCTION (this << end);
|
|
NS_ASSERT (CheckInternalState ());
|
|
uint32_t newEnd = m_end - std::min (end, m_end - m_start);
|
|
if (newEnd > m_zeroAreaEnd)
|
|
{
|
|
/* remove part of end of buffer */
|
|
m_end = newEnd;
|
|
}
|
|
else if (newEnd > m_zeroAreaStart)
|
|
{
|
|
/* remove end of buffer, part of zero area */
|
|
m_end = newEnd;
|
|
m_zeroAreaEnd = newEnd;
|
|
}
|
|
else if (newEnd > m_start)
|
|
{
|
|
/* remove end of buffer, zero area, part of start of buffer */
|
|
m_end = newEnd;
|
|
m_zeroAreaEnd = newEnd;
|
|
m_zeroAreaStart = newEnd;
|
|
}
|
|
else
|
|
{
|
|
/* remove all buffer */
|
|
m_end = m_start;
|
|
m_zeroAreaEnd = m_start;
|
|
m_zeroAreaStart = m_start;
|
|
}
|
|
HEURISTICS (m_maxZeroAreaStart = std::max (m_maxZeroAreaStart, m_zeroAreaStart));
|
|
LOG_INTERNAL_STATE ("rem end=" << end << ", ");
|
|
NS_ASSERT (CheckInternalState ());
|
|
}
|
|
|
|
Buffer
|
|
Buffer::CreateFragment (uint32_t start, uint32_t length) const
|
|
{
|
|
NS_LOG_FUNCTION (this << start << length);
|
|
NS_ASSERT (CheckInternalState ());
|
|
Buffer tmp = *this;
|
|
tmp.RemoveAtStart (start);
|
|
tmp.RemoveAtEnd (GetSize () - (start + length));
|
|
NS_ASSERT (CheckInternalState ());
|
|
return tmp;
|
|
}
|
|
|
|
Buffer
|
|
Buffer::CreateFullCopy (void) const
|
|
{
|
|
NS_LOG_FUNCTION (this);
|
|
NS_ASSERT (CheckInternalState ());
|
|
if (m_zeroAreaEnd - m_zeroAreaStart != 0)
|
|
{
|
|
Buffer tmp;
|
|
tmp.AddAtStart (m_zeroAreaEnd - m_zeroAreaStart);
|
|
tmp.Begin ().WriteU8 (0, m_zeroAreaEnd - m_zeroAreaStart);
|
|
uint32_t dataStart = m_zeroAreaStart - m_start;
|
|
tmp.AddAtStart (dataStart);
|
|
tmp.Begin ().Write (m_data->m_data+m_start, dataStart);
|
|
uint32_t dataEnd = m_end - m_zeroAreaEnd;
|
|
tmp.AddAtEnd (dataEnd);
|
|
Buffer::Iterator i = tmp.End ();
|
|
i.Prev (dataEnd);
|
|
i.Write (m_data->m_data+m_zeroAreaStart,dataEnd);
|
|
NS_ASSERT (tmp.CheckInternalState ());
|
|
return tmp;
|
|
}
|
|
NS_ASSERT (CheckInternalState ());
|
|
return *this;
|
|
}
|
|
|
|
uint32_t
|
|
Buffer::GetSerializedSize (void) const
|
|
{
|
|
uint32_t dataStart = (m_zeroAreaStart - m_start + 3) & (~0x3);
|
|
uint32_t dataEnd = (m_end - m_zeroAreaEnd + 3) & (~0x3);
|
|
|
|
// total size 4-bytes for dataStart length
|
|
// + X number of bytes for dataStart
|
|
// + 4-bytes for dataEnd length
|
|
// + X number of bytes for dataEnd
|
|
uint32_t sz = sizeof (uint32_t)
|
|
+ sizeof (uint32_t)
|
|
+ dataStart
|
|
+ sizeof (uint32_t)
|
|
+ dataEnd;
|
|
|
|
return sz;
|
|
}
|
|
|
|
uint32_t
|
|
Buffer::Serialize (uint8_t* buffer, uint32_t maxSize) const
|
|
{
|
|
uint32_t* p = reinterpret_cast<uint32_t *> (buffer);
|
|
uint32_t size = 0;
|
|
|
|
NS_LOG_FUNCTION (this);
|
|
|
|
// Add the zero data length
|
|
if (size + 4 <= maxSize)
|
|
{
|
|
size += 4;
|
|
*p++ = m_zeroAreaEnd - m_zeroAreaStart;
|
|
}
|
|
else
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
// Add the length of actual start data
|
|
uint32_t dataStartLength = m_zeroAreaStart - m_start;
|
|
if (size + 4 <= maxSize)
|
|
{
|
|
size += 4;
|
|
*p++ = dataStartLength;
|
|
}
|
|
else
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
// Add the actual data
|
|
if (size + ((dataStartLength + 3) & (~3)) <= maxSize)
|
|
{
|
|
size += (dataStartLength + 3) & (~3);
|
|
memcpy(p, m_data->m_data + m_start, dataStartLength);
|
|
p += (((dataStartLength + 3) & (~3))/4); // Advance p, insuring 4 byte boundary
|
|
}
|
|
else
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
// Add the length of the actual end data
|
|
uint32_t dataEndLength = m_end - m_zeroAreaEnd;
|
|
if (size + 4 <= maxSize)
|
|
{
|
|
size += 4;
|
|
*p++ = dataEndLength;
|
|
}
|
|
else
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
// Add the actual data
|
|
if (size + ((dataEndLength + 3) & (~3)) <= maxSize)
|
|
{
|
|
size += (dataEndLength + 3) & (~3);
|
|
memcpy(p, m_data->m_data+m_zeroAreaStart,dataEndLength);
|
|
p += (((dataEndLength + 3) & (~3))/4); // Advance p, insuring 4 byte boundary
|
|
}
|
|
else
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
// Serialzed everything successfully
|
|
return 1;
|
|
}
|
|
|
|
uint32_t
|
|
Buffer::Deserialize (const uint8_t *buffer, uint32_t size)
|
|
{
|
|
const uint32_t* p = reinterpret_cast<const uint32_t *> (buffer);
|
|
uint32_t sizeCheck = size-4;
|
|
|
|
NS_ASSERT (sizeCheck >= 4);
|
|
uint32_t zeroDataLength = *p++;
|
|
sizeCheck -= 4;
|
|
|
|
// Create zero bytes
|
|
Initialize (zeroDataLength);
|
|
|
|
// Add start data
|
|
NS_ASSERT (sizeCheck >= 4);
|
|
uint32_t dataStartLength = *p++;
|
|
sizeCheck -= 4;
|
|
AddAtStart (dataStartLength);
|
|
|
|
NS_ASSERT (sizeCheck >= dataStartLength);
|
|
Begin ().Write (reinterpret_cast<uint8_t *> (const_cast<uint32_t *> (p)), dataStartLength);
|
|
p += (((dataStartLength+3)&(~3))/4); // Advance p, insuring 4 byte boundary
|
|
sizeCheck -= ((dataStartLength+3)&(~3));
|
|
|
|
// Add end data
|
|
NS_ASSERT (sizeCheck >= 4);
|
|
uint32_t dataEndLength = *p++;
|
|
sizeCheck -= 4;
|
|
AddAtEnd (dataEndLength);
|
|
|
|
NS_ASSERT (sizeCheck >= dataEndLength);
|
|
Buffer::Iterator tmp = End ();
|
|
tmp.Prev (dataEndLength);
|
|
tmp.Write (reinterpret_cast<uint8_t *> (const_cast<uint32_t *> (p)), dataEndLength);
|
|
p += (((dataEndLength+3)&(~3))/4); // Advance p, insuring 4 byte boundary
|
|
sizeCheck -= ((dataEndLength+3)&(~3));
|
|
|
|
NS_ASSERT (sizeCheck == 0);
|
|
// return zero if buffer did not
|
|
// contain a complete message
|
|
return (sizeCheck != 0) ? 0 : 1;
|
|
}
|
|
|
|
int32_t
|
|
Buffer::GetCurrentStartOffset (void) const
|
|
{
|
|
return m_start;
|
|
}
|
|
int32_t
|
|
Buffer::GetCurrentEndOffset (void) const
|
|
{
|
|
return m_end;
|
|
}
|
|
|
|
|
|
void
|
|
Buffer::TransformIntoRealBuffer (void) const
|
|
{
|
|
NS_ASSERT (CheckInternalState ());
|
|
Buffer tmp = CreateFullCopy ();
|
|
*const_cast<Buffer *> (this) = tmp;
|
|
NS_ASSERT (CheckInternalState ());
|
|
}
|
|
|
|
|
|
uint8_t const*
|
|
Buffer::PeekData (void) const
|
|
{
|
|
NS_ASSERT (CheckInternalState ());
|
|
TransformIntoRealBuffer ();
|
|
NS_ASSERT (CheckInternalState ());
|
|
return m_data->m_data + m_start;
|
|
}
|
|
|
|
void
|
|
Buffer::CopyData(std::ostream *os, uint32_t size) const
|
|
{
|
|
if (size > 0)
|
|
{
|
|
uint32_t tmpsize = std::min (m_zeroAreaStart-m_start, size);
|
|
os->write((const char*)(m_data->m_data + m_start), tmpsize);
|
|
if (size > tmpsize)
|
|
{
|
|
size -= m_zeroAreaStart-m_start;
|
|
tmpsize = std::min (m_zeroAreaEnd - m_zeroAreaStart, size);
|
|
uint32_t left = tmpsize;
|
|
while (left > 0)
|
|
{
|
|
uint32_t toWrite = std::min (left, g_zeroes.size);
|
|
os->write (g_zeroes.buffer, toWrite);
|
|
left -= toWrite;
|
|
}
|
|
if (size > tmpsize)
|
|
{
|
|
size -= tmpsize;
|
|
tmpsize = std::min (m_end - m_zeroAreaEnd, size);
|
|
os->write ((const char*)(m_data->m_data + m_zeroAreaStart), tmpsize);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
uint32_t
|
|
Buffer::CopyData (uint8_t *buffer, uint32_t size) const
|
|
{
|
|
uint32_t originalSize = size;
|
|
if (size > 0)
|
|
{
|
|
uint32_t tmpsize = std::min (m_zeroAreaStart-m_start, size);
|
|
memcpy (buffer, (const char*)(m_data->m_data + m_start), tmpsize);
|
|
buffer += tmpsize;
|
|
if (size > tmpsize)
|
|
{
|
|
size -= m_zeroAreaStart-m_start;
|
|
tmpsize = std::min (m_zeroAreaEnd - m_zeroAreaStart, size);
|
|
uint32_t left = tmpsize;
|
|
while (left > 0)
|
|
{
|
|
uint32_t toWrite = std::min (left, g_zeroes.size);
|
|
memcpy (buffer, g_zeroes.buffer, toWrite);
|
|
left -= toWrite;
|
|
buffer += toWrite;
|
|
}
|
|
if (size > tmpsize)
|
|
{
|
|
size -= tmpsize;
|
|
tmpsize = std::min (m_end - m_zeroAreaEnd, size);
|
|
memcpy (buffer, (const char*)(m_data->m_data + m_zeroAreaStart), tmpsize);
|
|
}
|
|
}
|
|
}
|
|
return originalSize - size;
|
|
}
|
|
|
|
/******************************************************
|
|
* The buffer iterator below.
|
|
******************************************************/
|
|
|
|
|
|
Buffer::Iterator::Iterator ()
|
|
: m_zeroStart (0),
|
|
m_zeroEnd (0),
|
|
m_dataStart (0),
|
|
m_dataEnd (0),
|
|
m_current (0),
|
|
m_data (0)
|
|
{}
|
|
Buffer::Iterator::Iterator (Buffer const*buffer)
|
|
{
|
|
Construct (buffer);
|
|
m_current = m_dataStart;
|
|
}
|
|
Buffer::Iterator::Iterator (Buffer const*buffer, bool dummy)
|
|
{
|
|
Construct (buffer);
|
|
m_current = m_dataEnd;
|
|
}
|
|
|
|
void
|
|
Buffer::Iterator::Construct (const Buffer *buffer)
|
|
{
|
|
m_zeroStart = buffer->m_zeroAreaStart;
|
|
m_zeroEnd = buffer->m_zeroAreaEnd;
|
|
m_dataStart = buffer->m_start;
|
|
m_dataEnd = buffer->m_end;
|
|
m_data = buffer->m_data->m_data;
|
|
}
|
|
|
|
void
|
|
Buffer::Iterator::Next (void)
|
|
{
|
|
NS_ASSERT (m_current + 1 <= m_dataEnd);
|
|
m_current++;
|
|
}
|
|
void
|
|
Buffer::Iterator::Prev (void)
|
|
{
|
|
NS_ASSERT (m_current >= 1);
|
|
m_current--;
|
|
}
|
|
void
|
|
Buffer::Iterator::Next (uint32_t delta)
|
|
{
|
|
NS_ASSERT (m_current + delta <= m_dataEnd);
|
|
m_current += delta;
|
|
}
|
|
void
|
|
Buffer::Iterator::Prev (uint32_t delta)
|
|
{
|
|
NS_ASSERT (m_current >= delta);
|
|
m_current -= delta;
|
|
}
|
|
uint32_t
|
|
Buffer::Iterator::GetDistanceFrom (Iterator const &o) const
|
|
{
|
|
NS_ASSERT (m_data == o.m_data);
|
|
int32_t diff = m_current - o.m_current;
|
|
if (diff < 0)
|
|
{
|
|
return -diff;
|
|
}
|
|
else
|
|
{
|
|
return diff;
|
|
}
|
|
}
|
|
|
|
bool
|
|
Buffer::Iterator::IsEnd (void) const
|
|
{
|
|
return m_current == m_dataEnd;
|
|
}
|
|
bool
|
|
Buffer::Iterator::IsStart (void) const
|
|
{
|
|
return m_current == m_dataStart;
|
|
}
|
|
|
|
bool
|
|
Buffer::Iterator::CheckNoZero (uint32_t start, uint32_t end) const
|
|
{
|
|
bool ok = true;
|
|
for (uint32_t i = start; i < end; i++)
|
|
{
|
|
if (!Check (i))
|
|
{
|
|
ok = false;
|
|
}
|
|
}
|
|
return ok;
|
|
}
|
|
bool
|
|
Buffer::Iterator::Check (uint32_t i) const
|
|
{
|
|
return i >= m_dataStart &&
|
|
!(i >= m_zeroStart && i < m_zeroEnd) &&
|
|
i <= m_dataEnd;
|
|
}
|
|
|
|
|
|
void
|
|
Buffer::Iterator::Write (Iterator start, Iterator end)
|
|
{
|
|
NS_ASSERT (start.m_data == end.m_data);
|
|
NS_ASSERT (start.m_current <= end.m_current);
|
|
NS_ASSERT (start.m_zeroStart == end.m_zeroStart);
|
|
NS_ASSERT (start.m_zeroEnd == end.m_zeroEnd);
|
|
NS_ASSERT (m_data != start.m_data);
|
|
uint32_t size = end.m_current - start.m_current;
|
|
Iterator cur = start;
|
|
NS_ASSERT (CheckNoZero (m_current, m_current + size));
|
|
if (start.m_current <= start.m_zeroStart)
|
|
{
|
|
uint32_t toCopy = std::min (size, start.m_zeroStart - start.m_current);
|
|
memcpy (&m_data[m_current], &start.m_data[start.m_current], toCopy);
|
|
start.m_current += toCopy;
|
|
m_current += toCopy;
|
|
size -= toCopy;
|
|
}
|
|
if (start.m_current <= start.m_zeroEnd)
|
|
{
|
|
uint32_t toCopy = std::min (size, start.m_zeroEnd - start.m_current);
|
|
memset (&m_data[m_current], 0, toCopy);
|
|
start.m_current += toCopy;
|
|
m_current += toCopy;
|
|
size -= toCopy;
|
|
}
|
|
uint32_t toCopy = std::min (size, start.m_dataEnd - start.m_current);
|
|
uint8_t *from = &start.m_data[start.m_current - (start.m_zeroEnd-start.m_zeroStart)];
|
|
uint8_t *to = &m_data[m_current];
|
|
memcpy (to, from, toCopy);
|
|
m_current += toCopy;
|
|
}
|
|
|
|
void
|
|
Buffer::Iterator::WriteU16 (uint16_t data)
|
|
{
|
|
WriteU8 (data & 0xff);
|
|
data >>= 8;
|
|
WriteU8 (data & 0xff);
|
|
}
|
|
void
|
|
Buffer::Iterator::WriteU32 (uint32_t data)
|
|
{
|
|
WriteU8 (data & 0xff);
|
|
data >>= 8;
|
|
WriteU8 (data & 0xff);
|
|
data >>= 8;
|
|
WriteU8 (data & 0xff);
|
|
data >>= 8;
|
|
WriteU8 (data & 0xff);
|
|
}
|
|
void
|
|
Buffer::Iterator::WriteU64 (uint64_t data)
|
|
{
|
|
WriteU8 (data & 0xff);
|
|
data >>= 8;
|
|
WriteU8 (data & 0xff);
|
|
data >>= 8;
|
|
WriteU8 (data & 0xff);
|
|
data >>= 8;
|
|
WriteU8 (data & 0xff);
|
|
data >>= 8;
|
|
WriteU8 (data & 0xff);
|
|
data >>= 8;
|
|
WriteU8 (data & 0xff);
|
|
data >>= 8;
|
|
WriteU8 (data & 0xff);
|
|
data >>= 8;
|
|
WriteU8 (data & 0xff);
|
|
}
|
|
void
|
|
Buffer::Iterator::WriteHtolsbU16 (uint16_t data)
|
|
{
|
|
WriteU8 ((data >> 0) & 0xff);
|
|
WriteU8 ((data >> 8) & 0xff);
|
|
}
|
|
void
|
|
Buffer::Iterator::WriteHtolsbU32 (uint32_t data)
|
|
{
|
|
WriteU8 ((data >> 0) & 0xff);
|
|
WriteU8 ((data >> 8) & 0xff);
|
|
WriteU8 ((data >> 16) & 0xff);
|
|
WriteU8 ((data >> 24) & 0xff);
|
|
}
|
|
void
|
|
Buffer::Iterator::WriteHtolsbU64 (uint64_t data)
|
|
{
|
|
WriteU8 ((data >> 0) & 0xff);
|
|
WriteU8 ((data >> 8) & 0xff);
|
|
WriteU8 ((data >> 16) & 0xff);
|
|
WriteU8 ((data >> 24) & 0xff);
|
|
WriteU8 ((data >> 32) & 0xff);
|
|
WriteU8 ((data >> 40) & 0xff);
|
|
WriteU8 ((data >> 48) & 0xff);
|
|
WriteU8 ((data >> 56) & 0xff);
|
|
}
|
|
|
|
void
|
|
Buffer::Iterator::WriteHtonU64 (uint64_t data)
|
|
{
|
|
WriteU8 ((data >> 56) & 0xff);
|
|
WriteU8 ((data >> 48) & 0xff);
|
|
WriteU8 ((data >> 40) & 0xff);
|
|
WriteU8 ((data >> 32) & 0xff);
|
|
WriteU8 ((data >> 24) & 0xff);
|
|
WriteU8 ((data >> 16) & 0xff);
|
|
WriteU8 ((data >> 8) & 0xff);
|
|
WriteU8 ((data >> 0) & 0xff);
|
|
}
|
|
void
|
|
Buffer::Iterator::Write (uint8_t const*buffer, uint32_t size)
|
|
{
|
|
NS_ASSERT (CheckNoZero (m_current, size));
|
|
uint8_t *to;
|
|
if (m_current <= m_zeroStart)
|
|
{
|
|
to = &m_data[m_current];
|
|
}
|
|
else
|
|
{
|
|
to = &m_data[m_current - (m_zeroEnd - m_zeroStart)];
|
|
}
|
|
memcpy (to, buffer, size);
|
|
m_current += size;
|
|
}
|
|
|
|
uint16_t
|
|
Buffer::Iterator::ReadU16 (void)
|
|
{
|
|
uint8_t byte0 = ReadU8 ();
|
|
uint8_t byte1 = ReadU8 ();
|
|
uint16_t data = byte1;
|
|
data <<= 8;
|
|
data |= byte0;
|
|
|
|
return data;
|
|
}
|
|
uint32_t
|
|
Buffer::Iterator::ReadU32 (void)
|
|
{
|
|
uint8_t byte0 = ReadU8 ();
|
|
uint8_t byte1 = ReadU8 ();
|
|
uint8_t byte2 = ReadU8 ();
|
|
uint8_t byte3 = ReadU8 ();
|
|
uint32_t data = byte3;
|
|
data <<= 8;
|
|
data |= byte2;
|
|
data <<= 8;
|
|
data |= byte1;
|
|
data <<= 8;
|
|
data |= byte0;
|
|
return data;
|
|
}
|
|
uint64_t
|
|
Buffer::Iterator::ReadU64 (void)
|
|
{
|
|
uint8_t byte0 = ReadU8 ();
|
|
uint8_t byte1 = ReadU8 ();
|
|
uint8_t byte2 = ReadU8 ();
|
|
uint8_t byte3 = ReadU8 ();
|
|
uint8_t byte4 = ReadU8 ();
|
|
uint8_t byte5 = ReadU8 ();
|
|
uint8_t byte6 = ReadU8 ();
|
|
uint8_t byte7 = ReadU8 ();
|
|
uint64_t data = byte7;
|
|
data <<= 8;
|
|
data |= byte6;
|
|
data <<= 8;
|
|
data |= byte5;
|
|
data <<= 8;
|
|
data |= byte4;
|
|
data <<= 8;
|
|
data |= byte3;
|
|
data <<= 8;
|
|
data |= byte2;
|
|
data <<= 8;
|
|
data |= byte1;
|
|
data <<= 8;
|
|
data |= byte0;
|
|
|
|
return data;
|
|
}
|
|
uint16_t
|
|
Buffer::Iterator::SlowReadNtohU16 (void)
|
|
{
|
|
uint16_t retval = 0;
|
|
retval |= ReadU8 ();
|
|
retval <<= 8;
|
|
retval |= ReadU8 ();
|
|
return retval;
|
|
}
|
|
uint32_t
|
|
Buffer::Iterator::SlowReadNtohU32 (void)
|
|
{
|
|
uint32_t retval = 0;
|
|
retval |= ReadU8 ();
|
|
retval <<= 8;
|
|
retval |= ReadU8 ();
|
|
retval <<= 8;
|
|
retval |= ReadU8 ();
|
|
retval <<= 8;
|
|
retval |= ReadU8 ();
|
|
return retval;
|
|
}
|
|
uint64_t
|
|
Buffer::Iterator::ReadNtohU64 (void)
|
|
{
|
|
uint64_t retval = 0;
|
|
retval |= ReadU8 ();
|
|
retval <<= 8;
|
|
retval |= ReadU8 ();
|
|
retval <<= 8;
|
|
retval |= ReadU8 ();
|
|
retval <<= 8;
|
|
retval |= ReadU8 ();
|
|
retval <<= 8;
|
|
retval |= ReadU8 ();
|
|
retval <<= 8;
|
|
retval |= ReadU8 ();
|
|
retval <<= 8;
|
|
retval |= ReadU8 ();
|
|
retval <<= 8;
|
|
retval |= ReadU8 ();
|
|
return retval;
|
|
}
|
|
uint16_t
|
|
Buffer::Iterator::ReadLsbtohU16 (void)
|
|
{
|
|
uint8_t byte0 = ReadU8 ();
|
|
uint8_t byte1 = ReadU8 ();
|
|
uint16_t data = byte1;
|
|
data <<= 8;
|
|
data |= byte0;
|
|
return data;
|
|
}
|
|
uint32_t
|
|
Buffer::Iterator::ReadLsbtohU32 (void)
|
|
{
|
|
uint8_t byte0 = ReadU8 ();
|
|
uint8_t byte1 = ReadU8 ();
|
|
uint8_t byte2 = ReadU8 ();
|
|
uint8_t byte3 = ReadU8 ();
|
|
uint32_t data = byte3;
|
|
data <<= 8;
|
|
data |= byte2;
|
|
data <<= 8;
|
|
data |= byte1;
|
|
data <<= 8;
|
|
data |= byte0;
|
|
return data;
|
|
}
|
|
uint64_t
|
|
Buffer::Iterator::ReadLsbtohU64 (void)
|
|
{
|
|
uint8_t byte0 = ReadU8 ();
|
|
uint8_t byte1 = ReadU8 ();
|
|
uint8_t byte2 = ReadU8 ();
|
|
uint8_t byte3 = ReadU8 ();
|
|
uint8_t byte4 = ReadU8 ();
|
|
uint8_t byte5 = ReadU8 ();
|
|
uint8_t byte6 = ReadU8 ();
|
|
uint8_t byte7 = ReadU8 ();
|
|
uint64_t data = byte7;
|
|
data <<= 8;
|
|
data |= byte6;
|
|
data <<= 8;
|
|
data |= byte5;
|
|
data <<= 8;
|
|
data |= byte4;
|
|
data <<= 8;
|
|
data |= byte3;
|
|
data <<= 8;
|
|
data |= byte2;
|
|
data <<= 8;
|
|
data |= byte1;
|
|
data <<= 8;
|
|
data |= byte0;
|
|
|
|
return data;
|
|
}
|
|
void
|
|
Buffer::Iterator::Read (uint8_t *buffer, uint32_t size)
|
|
{
|
|
for (uint32_t i = 0; i < size; i++)
|
|
{
|
|
buffer[i] = ReadU8 ();
|
|
}
|
|
}
|
|
|
|
uint16_t
|
|
Buffer::Iterator::CalculateIpChecksum(uint16_t size)
|
|
{
|
|
return CalculateIpChecksum(size, 0);
|
|
}
|
|
|
|
uint16_t
|
|
Buffer::Iterator::CalculateIpChecksum(uint16_t size, uint32_t initialChecksum)
|
|
{
|
|
/* see RFC 1071 to understand this code. */
|
|
uint32_t sum = initialChecksum;
|
|
|
|
for (int j = 0; j < size/2; j++)
|
|
sum += ReadU16 ();
|
|
|
|
if (size & 1)
|
|
sum += ReadU8 ();
|
|
|
|
while (sum >> 16)
|
|
sum = (sum & 0xffff) + (sum >> 16);
|
|
return ~sum;
|
|
}
|
|
|
|
uint32_t
|
|
Buffer::Iterator::GetSize (void) const
|
|
{
|
|
return m_dataEnd - m_dataStart;
|
|
}
|
|
|
|
//-----------------------------------------------------------------------------
|
|
// Unit tests
|
|
//-----------------------------------------------------------------------------
|
|
class BufferTest: public TestCase {
|
|
private:
|
|
bool EnsureWrittenBytes (Buffer b, uint32_t n, uint8_t array[]);
|
|
public:
|
|
virtual bool DoRun (void);
|
|
BufferTest ();
|
|
};
|
|
|
|
|
|
BufferTest::BufferTest ()
|
|
: TestCase ("Buffer") {}
|
|
|
|
bool
|
|
BufferTest::EnsureWrittenBytes (Buffer b, uint32_t n, uint8_t array[])
|
|
{
|
|
bool success = true;
|
|
uint8_t *expected = array;
|
|
uint8_t const*got;
|
|
got = b.PeekData ();
|
|
for (uint32_t j = 0; j < n; j++)
|
|
{
|
|
if (got[j] != expected[j])
|
|
{
|
|
success = false;
|
|
}
|
|
}
|
|
if (!success)
|
|
{
|
|
std::ostringstream failure;
|
|
failure << "Buffer -- ";
|
|
failure << "expected: n=";
|
|
failure << n << ", ";
|
|
failure.setf (std::ios::hex, std::ios::basefield);
|
|
for (uint32_t j = 0; j < n; j++)
|
|
{
|
|
failure << (uint16_t)expected[j] << " ";
|
|
}
|
|
failure.setf (std::ios::dec, std::ios::basefield);
|
|
failure << "got: ";
|
|
failure.setf (std::ios::hex, std::ios::basefield);
|
|
for (uint32_t j = 0; j < n; j++)
|
|
{
|
|
failure << (uint16_t)got[j] << " ";
|
|
}
|
|
failure << std::endl;
|
|
ReportTestFailure ("", "", "", failure.str(), __FILE__, __LINE__);
|
|
}
|
|
return success;
|
|
}
|
|
|
|
/* Note: works only when variadic macros are
|
|
* available which is the case for gcc.
|
|
* XXX
|
|
*/
|
|
#define ENSURE_WRITTEN_BYTES(buffer, n, ...) \
|
|
{ \
|
|
uint8_t bytes[] = {__VA_ARGS__}; \
|
|
if (!EnsureWrittenBytes (buffer, n , bytes)) \
|
|
{ \
|
|
SetErrorStatus (false); \
|
|
} \
|
|
}
|
|
|
|
bool
|
|
BufferTest::DoRun (void)
|
|
{
|
|
Buffer buffer;
|
|
Buffer::Iterator i;
|
|
buffer.AddAtStart (6);
|
|
i = buffer.Begin ();
|
|
i.WriteU8 (0x66);
|
|
ENSURE_WRITTEN_BYTES (buffer, 1, 0x66);
|
|
i = buffer.Begin ();
|
|
i.WriteU8 (0x67);
|
|
ENSURE_WRITTEN_BYTES (buffer, 1, 0x67);
|
|
i.WriteHtonU16 (0x6568);
|
|
i = buffer.Begin ();
|
|
ENSURE_WRITTEN_BYTES (buffer, 3, 0x67, 0x65, 0x68);
|
|
i.WriteHtonU16 (0x6369);
|
|
ENSURE_WRITTEN_BYTES (buffer, 3, 0x63, 0x69, 0x68);
|
|
i.WriteHtonU32 (0xdeadbeaf);
|
|
ENSURE_WRITTEN_BYTES (buffer, 6, 0x63, 0x69, 0xde, 0xad, 0xbe, 0xaf);
|
|
buffer.AddAtStart (2);
|
|
i = buffer.Begin ();
|
|
i.WriteU16 (0);
|
|
ENSURE_WRITTEN_BYTES (buffer, 8, 0, 0, 0x63, 0x69, 0xde, 0xad, 0xbe, 0xaf);
|
|
buffer.AddAtEnd (2);
|
|
i = buffer.Begin ();
|
|
i.Next (8);
|
|
i.WriteU16 (0);
|
|
ENSURE_WRITTEN_BYTES (buffer, 10, 0, 0, 0x63, 0x69, 0xde, 0xad, 0xbe, 0xaf, 0, 0);
|
|
buffer.RemoveAtStart (3);
|
|
i = buffer.Begin ();
|
|
ENSURE_WRITTEN_BYTES (buffer, 7, 0x69, 0xde, 0xad, 0xbe, 0xaf, 0, 0);
|
|
buffer.RemoveAtEnd (4);
|
|
i = buffer.Begin ();
|
|
ENSURE_WRITTEN_BYTES (buffer, 3, 0x69, 0xde, 0xad);
|
|
buffer.AddAtStart (1);
|
|
i = buffer.Begin ();
|
|
i.WriteU8 (0xff);
|
|
ENSURE_WRITTEN_BYTES (buffer, 4, 0xff, 0x69, 0xde, 0xad);
|
|
buffer.AddAtEnd (1);
|
|
i = buffer.Begin ();
|
|
i.Next (4);
|
|
i.WriteU8 (0xff);
|
|
i.Prev (2);
|
|
uint16_t saved = i.ReadU16 ();
|
|
i.Prev (2);
|
|
i.WriteHtonU16 (0xff00);
|
|
i.Prev (2);
|
|
if (i.ReadNtohU16 () != 0xff00)
|
|
{
|
|
SetErrorStatus (false);
|
|
}
|
|
i.Prev (2);
|
|
i.WriteU16 (saved);
|
|
ENSURE_WRITTEN_BYTES (buffer, 5, 0xff, 0x69, 0xde, 0xad, 0xff);
|
|
Buffer o = buffer;
|
|
ENSURE_WRITTEN_BYTES (o, 5, 0xff, 0x69, 0xde, 0xad, 0xff);
|
|
o.AddAtStart (1);
|
|
i = o.Begin ();
|
|
i.WriteU8 (0xfe);
|
|
ENSURE_WRITTEN_BYTES (o, 6, 0xfe, 0xff, 0x69, 0xde, 0xad, 0xff);
|
|
buffer.AddAtStart (2);
|
|
i = buffer.Begin ();
|
|
i.WriteU8 (0xfd);
|
|
i.WriteU8 (0xfd);
|
|
ENSURE_WRITTEN_BYTES (o, 6, 0xfe, 0xff, 0x69, 0xde, 0xad, 0xff);
|
|
ENSURE_WRITTEN_BYTES (buffer, 7, 0xfd, 0xfd, 0xff, 0x69, 0xde, 0xad, 0xff);
|
|
|
|
// test 64-bit read/write
|
|
Buffer buff64;
|
|
buff64.AddAtStart(8);
|
|
i = buff64.Begin();
|
|
i.WriteU64 (0x0123456789ABCDEFllu);
|
|
ENSURE_WRITTEN_BYTES (buff64, 8, 0xef, 0xcd, 0xab, 0x89, 0x67, 0x45, 0x23, 0x01);
|
|
i = buff64.Begin();
|
|
if (i.ReadLsbtohU64() != 0x0123456789abcdefllu)
|
|
{
|
|
SetErrorStatus (false);
|
|
}
|
|
i = buff64.Begin();
|
|
i.WriteHtolsbU64 (0x0123456789ABCDEFllu);
|
|
ENSURE_WRITTEN_BYTES (buff64, 8, 0xef, 0xcd, 0xab, 0x89, 0x67, 0x45, 0x23, 0x01);
|
|
i = buff64.Begin();
|
|
if (i.ReadLsbtohU64() != 0x0123456789abcdefllu)
|
|
{
|
|
SetErrorStatus (false);
|
|
}
|
|
i = buff64.Begin();
|
|
i.WriteHtonU64 (0x0123456789ABCDEFllu);
|
|
ENSURE_WRITTEN_BYTES (buff64, 8, 0x01, 0x23, 0x45, 0x67, 0x89, 0xab, 0xcd, 0xef);
|
|
i = buff64.Begin();
|
|
if (i.ReadNtohU64() != 0x0123456789abcdefllu)
|
|
{
|
|
SetErrorStatus (false);
|
|
}
|
|
|
|
// test self-assignment
|
|
{
|
|
Buffer a = o;
|
|
a = a;
|
|
}
|
|
|
|
// test Remove start.
|
|
buffer = Buffer (5);
|
|
ENSURE_WRITTEN_BYTES (buffer, 5, 0, 0, 0, 0, 0);
|
|
buffer.RemoveAtStart (1);
|
|
ENSURE_WRITTEN_BYTES (buffer, 4, 0, 0, 0, 0);
|
|
buffer.AddAtStart (1);
|
|
buffer.Begin ().WriteU8 (0xff);
|
|
ENSURE_WRITTEN_BYTES (buffer, 5, 0xff, 0, 0, 0, 0);
|
|
buffer.RemoveAtStart(3);
|
|
ENSURE_WRITTEN_BYTES (buffer, 2, 0, 0);
|
|
buffer.AddAtStart (4);
|
|
buffer.Begin ().WriteHtonU32 (0xdeadbeaf);
|
|
ENSURE_WRITTEN_BYTES (buffer, 6, 0xde, 0xad, 0xbe, 0xaf, 0, 0);
|
|
buffer.RemoveAtStart (2);
|
|
ENSURE_WRITTEN_BYTES (buffer, 4, 0xbe, 0xaf, 0, 0);
|
|
buffer.AddAtEnd (4);
|
|
i = buffer.Begin ();
|
|
i.Next (4);
|
|
i.WriteHtonU32 (0xdeadbeaf);
|
|
ENSURE_WRITTEN_BYTES (buffer, 8, 0xbe, 0xaf, 0, 0, 0xde, 0xad, 0xbe, 0xaf);
|
|
buffer.RemoveAtStart (5);
|
|
ENSURE_WRITTEN_BYTES (buffer, 3, 0xad, 0xbe, 0xaf);
|
|
// test Remove end
|
|
buffer = Buffer (5);
|
|
ENSURE_WRITTEN_BYTES (buffer, 5, 0, 0, 0, 0, 0);
|
|
buffer.RemoveAtEnd (1);
|
|
ENSURE_WRITTEN_BYTES (buffer, 4, 0, 0, 0, 0);
|
|
buffer.AddAtEnd (2);
|
|
i = buffer.Begin ();
|
|
i.Next (4);
|
|
i.WriteU8 (0xab);
|
|
i.WriteU8 (0xac);
|
|
ENSURE_WRITTEN_BYTES (buffer, 6, 0, 0, 0, 0, 0xab, 0xac);
|
|
buffer.RemoveAtEnd (1);
|
|
ENSURE_WRITTEN_BYTES (buffer, 5, 0, 0, 0, 0, 0xab);
|
|
buffer.RemoveAtEnd (3);
|
|
ENSURE_WRITTEN_BYTES (buffer, 2, 0, 0);
|
|
buffer.AddAtEnd (6);
|
|
i = buffer.Begin ();
|
|
i.Next (2);
|
|
i.WriteU8 (0xac);
|
|
i.WriteU8 (0xad);
|
|
i.WriteU8 (0xae);
|
|
i.WriteU8 (0xaf);
|
|
i.WriteU8 (0xba);
|
|
i.WriteU8 (0xbb);
|
|
ENSURE_WRITTEN_BYTES (buffer, 8, 0, 0, 0xac, 0xad, 0xae, 0xaf, 0xba, 0xbb);
|
|
buffer.AddAtStart (3);
|
|
i = buffer.Begin ();
|
|
i.WriteU8 (0x30);
|
|
i.WriteU8 (0x31);
|
|
i.WriteU8 (0x32);
|
|
ENSURE_WRITTEN_BYTES (buffer, 11, 0x30, 0x31, 0x32, 0, 0, 0xac, 0xad, 0xae, 0xaf, 0xba, 0xbb);
|
|
buffer.RemoveAtEnd (9);
|
|
ENSURE_WRITTEN_BYTES (buffer, 2, 0x30, 0x31);
|
|
buffer = Buffer (3);
|
|
buffer.AddAtEnd (2);
|
|
i = buffer.Begin ();
|
|
i.Next (3);
|
|
i.WriteHtonU16 (0xabcd);
|
|
buffer.AddAtStart (1);
|
|
buffer.Begin ().WriteU8 (0x21);
|
|
ENSURE_WRITTEN_BYTES (buffer, 6, 0x21, 0, 0, 0, 0xab, 0xcd);
|
|
buffer.RemoveAtEnd (8);
|
|
if (buffer.GetSize () != 0)
|
|
{
|
|
SetErrorStatus (false);
|
|
}
|
|
|
|
buffer = Buffer (6);
|
|
buffer.AddAtStart (9);
|
|
buffer.AddAtEnd (3);
|
|
i = buffer.End ();
|
|
i.Prev (1);
|
|
i.WriteU8 (1, 1);
|
|
|
|
buffer = Buffer (6);
|
|
buffer.AddAtStart (3);
|
|
buffer.RemoveAtEnd (8);
|
|
buffer.AddAtEnd (4);
|
|
i = buffer.End ();
|
|
i.Prev (4);
|
|
i.WriteU8 (1, 4);
|
|
|
|
buffer = Buffer (1);
|
|
buffer.AddAtEnd (100);
|
|
i = buffer.End ();
|
|
i.Prev (100);
|
|
i.WriteU8 (1, 100);
|
|
|
|
// Bug #54
|
|
{
|
|
const uint32_t actualSize = 72602;
|
|
const uint32_t chunkSize = 67624;
|
|
UniformVariable bytesRng (0, 256);
|
|
|
|
Buffer inputBuffer;
|
|
Buffer outputBuffer;
|
|
|
|
inputBuffer.AddAtEnd (actualSize);
|
|
{
|
|
Buffer::Iterator iter = inputBuffer.Begin ();
|
|
for (uint32_t i = 0; i < actualSize; i++)
|
|
iter.WriteU8 (static_cast<uint8_t> (bytesRng.GetValue ()));
|
|
}
|
|
|
|
outputBuffer.AddAtEnd (chunkSize);
|
|
Buffer::Iterator iter = outputBuffer.End ();
|
|
iter.Prev (chunkSize);
|
|
iter.Write (inputBuffer.PeekData (), chunkSize);
|
|
|
|
NS_TEST_EXPECT_MSG_EQ (memcmp (inputBuffer.PeekData (), outputBuffer.PeekData (), chunkSize), 0, "memcp works");
|
|
}
|
|
|
|
buffer = Buffer (5);
|
|
buffer.AddAtEnd (2);
|
|
i = buffer.End ();
|
|
i.Prev (2);
|
|
i.WriteU8 (0);
|
|
i.WriteU8 (0x66);
|
|
ENSURE_WRITTEN_BYTES (buffer, 7, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x66);
|
|
Buffer frag0 = buffer.CreateFragment (0, 2);
|
|
ENSURE_WRITTEN_BYTES (frag0, 2, 0x00, 0x00);
|
|
Buffer frag1 = buffer.CreateFragment (2, 5);
|
|
ENSURE_WRITTEN_BYTES (frag1, 5, 0x00, 0x00, 0x00, 0x00, 0x66);
|
|
frag0.AddAtEnd (frag1);
|
|
ENSURE_WRITTEN_BYTES (buffer, 7, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x66);
|
|
ENSURE_WRITTEN_BYTES (frag0, 7, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x66);
|
|
|
|
buffer = Buffer (5);
|
|
buffer.AddAtStart (2);
|
|
i = buffer.Begin ();
|
|
i.WriteU8 (0x1);
|
|
i.WriteU8 (0x2);
|
|
buffer.AddAtEnd (2);
|
|
i = buffer.End ();
|
|
i.Prev (2);
|
|
i.WriteU8 (0x3);
|
|
i.WriteU8 (0x4);
|
|
ENSURE_WRITTEN_BYTES (buffer, 9, 0x1, 0x2, 0x00, 0x00, 0x00, 0x00, 0x00, 0x3, 0x4);
|
|
Buffer other;
|
|
other.AddAtStart (9);
|
|
i = other.Begin ();
|
|
i.Write (buffer.Begin (), buffer.End ());
|
|
ENSURE_WRITTEN_BYTES (other, 9, 0x1, 0x2, 0x00, 0x00, 0x00, 0x00, 0x00, 0x3, 0x4);
|
|
|
|
return GetErrorStatus ();
|
|
}
|
|
//-----------------------------------------------------------------------------
|
|
class BufferTestSuite : public TestSuite
|
|
{
|
|
public:
|
|
BufferTestSuite ();
|
|
};
|
|
|
|
BufferTestSuite::BufferTestSuite ()
|
|
: TestSuite ("buffer", UNIT)
|
|
{
|
|
AddTestCase (new BufferTest);
|
|
}
|
|
|
|
BufferTestSuite g_bufferTestSuite;
|
|
|
|
} // namespace ns3
|
|
|
|
|