444839b842
Fix size of vertex_t on 64-bit architectures Fix __get_tls casts, the TLS area is a void** Cast through uintptr_t to store integers in a pointer Use %zu to print size_t Change-Id: I91079ec76025237e6d2081784c5348518b0ce122
174 lines
4.5 KiB
C++
174 lines
4.5 KiB
C++
/* libs/opengles/Tokenizer.cpp
|
|
**
|
|
** Copyright 2006, The Android Open Source Project
|
|
**
|
|
** Licensed under the Apache License, Version 2.0 (the "License");
|
|
** you may not use this file except in compliance with the License.
|
|
** You may obtain a copy of the License at
|
|
**
|
|
** http://www.apache.org/licenses/LICENSE-2.0
|
|
**
|
|
** Unless required by applicable law or agreed to in writing, software
|
|
** distributed under the License is distributed on an "AS IS" BASIS,
|
|
** WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
** See the License for the specific language governing permissions and
|
|
** limitations under the License.
|
|
*/
|
|
|
|
#include <stdio.h>
|
|
|
|
#include "Tokenizer.h"
|
|
|
|
// ----------------------------------------------------------------------------
|
|
|
|
namespace android {
|
|
|
|
ANDROID_BASIC_TYPES_TRAITS(Tokenizer::run_t)
|
|
|
|
Tokenizer::Tokenizer()
|
|
{
|
|
}
|
|
|
|
Tokenizer::Tokenizer(const Tokenizer& other)
|
|
: mRanges(other.mRanges)
|
|
{
|
|
}
|
|
|
|
Tokenizer::~Tokenizer()
|
|
{
|
|
}
|
|
|
|
uint32_t Tokenizer::acquire()
|
|
{
|
|
if (!mRanges.size() || mRanges[0].first) {
|
|
_insertTokenAt(0,0);
|
|
return 0;
|
|
}
|
|
|
|
// just extend the first run
|
|
const run_t& run = mRanges[0];
|
|
uint32_t token = run.first + run.length;
|
|
_insertTokenAt(token, 1);
|
|
return token;
|
|
}
|
|
|
|
bool Tokenizer::isAcquired(uint32_t token) const
|
|
{
|
|
return (_indexOrderOf(token) >= 0);
|
|
}
|
|
|
|
status_t Tokenizer::reserve(uint32_t token)
|
|
{
|
|
size_t o;
|
|
const ssize_t i = _indexOrderOf(token, &o);
|
|
if (i >= 0) {
|
|
return BAD_VALUE; // this token is already taken
|
|
}
|
|
ssize_t err = _insertTokenAt(token, o);
|
|
return (err<0) ? err : status_t(NO_ERROR);
|
|
}
|
|
|
|
status_t Tokenizer::release(uint32_t token)
|
|
{
|
|
const ssize_t i = _indexOrderOf(token);
|
|
if (i >= 0) {
|
|
const run_t& run = mRanges[i];
|
|
if ((token >= run.first) && (token < run.first+run.length)) {
|
|
// token in this range, we need to split
|
|
run_t& run = mRanges.editItemAt(i);
|
|
if ((token == run.first) || (token == run.first+run.length-1)) {
|
|
if (token == run.first) {
|
|
run.first += 1;
|
|
}
|
|
run.length -= 1;
|
|
if (run.length == 0) {
|
|
// XXX: should we systematically remove a run that's empty?
|
|
mRanges.removeItemsAt(i);
|
|
}
|
|
} else {
|
|
// split the run
|
|
run_t new_run;
|
|
new_run.first = token+1;
|
|
new_run.length = run.first+run.length - new_run.first;
|
|
run.length = token - run.first;
|
|
mRanges.insertAt(new_run, i+1);
|
|
}
|
|
return NO_ERROR;
|
|
}
|
|
}
|
|
return NAME_NOT_FOUND;
|
|
}
|
|
|
|
ssize_t Tokenizer::_indexOrderOf(uint32_t token, size_t* order) const
|
|
{
|
|
// binary search
|
|
ssize_t err = NAME_NOT_FOUND;
|
|
ssize_t l = 0;
|
|
ssize_t h = mRanges.size()-1;
|
|
ssize_t mid;
|
|
const run_t* a = mRanges.array();
|
|
while (l <= h) {
|
|
mid = l + (h - l)/2;
|
|
const run_t* const curr = a + mid;
|
|
int c = 0;
|
|
if (token < curr->first) c = 1;
|
|
else if (token >= curr->first+curr->length) c = -1;
|
|
if (c == 0) {
|
|
err = l = mid;
|
|
break;
|
|
} else if (c < 0) {
|
|
l = mid + 1;
|
|
} else {
|
|
h = mid - 1;
|
|
}
|
|
}
|
|
if (order) *order = l;
|
|
return err;
|
|
}
|
|
|
|
ssize_t Tokenizer::_insertTokenAt(uint32_t token, size_t index)
|
|
{
|
|
const size_t c = mRanges.size();
|
|
|
|
if (index >= 1) {
|
|
// do we need to merge with the previous run?
|
|
run_t& p = mRanges.editItemAt(index-1);
|
|
if (p.first+p.length == token) {
|
|
p.length += 1;
|
|
if (index < c) {
|
|
const run_t& n = mRanges[index];
|
|
if (token+1 == n.first) {
|
|
p.length += n.length;
|
|
mRanges.removeItemsAt(index);
|
|
}
|
|
}
|
|
return index;
|
|
}
|
|
}
|
|
|
|
if (index < c) {
|
|
// do we need to merge with the next run?
|
|
run_t& n = mRanges.editItemAt(index);
|
|
if (token+1 == n.first) {
|
|
n.first -= 1;
|
|
n.length += 1;
|
|
return index;
|
|
}
|
|
}
|
|
|
|
return mRanges.insertAt(run_t(token,1), index);
|
|
}
|
|
|
|
void Tokenizer::dump() const
|
|
{
|
|
const run_t* ranges = mRanges.array();
|
|
const size_t c = mRanges.size();
|
|
ALOGD("Tokenizer (%p, size = %zu)\n", this, c);
|
|
for (size_t i=0 ; i<c ; i++) {
|
|
ALOGD("%zu: (%u, %u)\n", i, ranges[i].first, ranges[i].length);
|
|
}
|
|
}
|
|
|
|
}; // namespace android
|
|
|