You've already forked godot
mirror of
https://github.com/godotengine/godot.git
synced 2025-11-27 15:57:02 +00:00
HarfBuzz: Update to version 8.0.0
This commit is contained in:
269
thirdparty/harfbuzz/src/hb-ot-cff-common.hh
vendored
269
thirdparty/harfbuzz/src/hb-ot-cff-common.hh
vendored
@@ -48,12 +48,24 @@ static inline const Type& StructAtOffsetOrNull (const void *P, unsigned int offs
|
||||
|
||||
struct code_pair_t
|
||||
{
|
||||
hb_codepoint_t code;
|
||||
unsigned code;
|
||||
hb_codepoint_t glyph;
|
||||
};
|
||||
|
||||
|
||||
using str_buff_t = hb_vector_t<unsigned char>;
|
||||
using str_buff_vec_t = hb_vector_t<str_buff_t>;
|
||||
using glyph_to_sid_map_t = hb_vector_t<code_pair_t>;
|
||||
|
||||
struct length_f_t
|
||||
{
|
||||
template <typename Iterable,
|
||||
hb_requires (hb_is_iterable (Iterable))>
|
||||
unsigned operator () (const Iterable &_) const { return hb_len (hb_iter (_)); }
|
||||
|
||||
unsigned operator () (unsigned _) const { return _; }
|
||||
}
|
||||
HB_FUNCOBJ (length_f);
|
||||
|
||||
/* CFF INDEX */
|
||||
template <typename COUNT>
|
||||
@@ -62,42 +74,52 @@ struct CFFIndex
|
||||
unsigned int offset_array_size () const
|
||||
{ return offSize * (count + 1); }
|
||||
|
||||
CFFIndex *copy (hb_serialize_context_t *c) const
|
||||
{
|
||||
TRACE_SERIALIZE (this);
|
||||
unsigned int size = get_size ();
|
||||
CFFIndex *out = c->allocate_size<CFFIndex> (size, false);
|
||||
if (likely (out))
|
||||
hb_memcpy (out, this, size);
|
||||
return_trace (out);
|
||||
}
|
||||
|
||||
template <typename Iterable,
|
||||
hb_requires (hb_is_iterable (Iterable))>
|
||||
bool serialize (hb_serialize_context_t *c,
|
||||
const Iterable &iterable)
|
||||
const Iterable &iterable,
|
||||
const unsigned *p_data_size = nullptr)
|
||||
{
|
||||
TRACE_SERIALIZE (this);
|
||||
unsigned data_size;
|
||||
if (p_data_size)
|
||||
data_size = *p_data_size;
|
||||
else
|
||||
total_size (iterable, &data_size);
|
||||
|
||||
auto it = hb_iter (iterable);
|
||||
serialize_header(c, + it | hb_map (hb_iter) | hb_map (hb_len));
|
||||
if (unlikely (!serialize_header (c, +it, data_size))) return_trace (false);
|
||||
unsigned char *ret = c->allocate_size<unsigned char> (data_size, false);
|
||||
if (unlikely (!ret)) return_trace (false);
|
||||
for (const auto &_ : +it)
|
||||
hb_iter (_).copy (c);
|
||||
{
|
||||
unsigned len = _.length;
|
||||
if (len <= 1)
|
||||
{
|
||||
if (!len)
|
||||
continue;
|
||||
*ret++ = *_.arrayZ;
|
||||
continue;
|
||||
}
|
||||
hb_memcpy (ret, _.arrayZ, len);
|
||||
ret += len;
|
||||
}
|
||||
return_trace (true);
|
||||
}
|
||||
|
||||
template <typename Iterator,
|
||||
hb_requires (hb_is_iterator (Iterator))>
|
||||
bool serialize_header (hb_serialize_context_t *c,
|
||||
Iterator it)
|
||||
Iterator it,
|
||||
unsigned data_size)
|
||||
{
|
||||
TRACE_SERIALIZE (this);
|
||||
|
||||
unsigned total = + it | hb_reduce (hb_add, 0);
|
||||
unsigned off_size = (hb_bit_storage (total + 1) + 7) / 8;
|
||||
unsigned off_size = (hb_bit_storage (data_size + 1) + 7) / 8;
|
||||
|
||||
/* serialize CFFIndex header */
|
||||
if (unlikely (!c->extend_min (this))) return_trace (false);
|
||||
this->count = it.len ();
|
||||
this->count = hb_len (it);
|
||||
if (!this->count) return_trace (true);
|
||||
if (unlikely (!c->extend (this->offSize))) return_trace (false);
|
||||
this->offSize = off_size;
|
||||
@@ -106,25 +128,88 @@ struct CFFIndex
|
||||
|
||||
/* serialize indices */
|
||||
unsigned int offset = 1;
|
||||
unsigned int i = 0;
|
||||
for (unsigned _ : +it)
|
||||
if (HB_OPTIMIZE_SIZE_VAL)
|
||||
{
|
||||
set_offset_at (i++, offset);
|
||||
offset += _;
|
||||
unsigned int i = 0;
|
||||
for (const auto &_ : +it)
|
||||
{
|
||||
set_offset_at (i++, offset);
|
||||
offset += length_f (_);
|
||||
}
|
||||
set_offset_at (i, offset);
|
||||
}
|
||||
set_offset_at (i, offset);
|
||||
else
|
||||
switch (off_size)
|
||||
{
|
||||
case 1:
|
||||
{
|
||||
HBUINT8 *p = (HBUINT8 *) offsets;
|
||||
for (const auto &_ : +it)
|
||||
{
|
||||
*p++ = offset;
|
||||
offset += length_f (_);
|
||||
}
|
||||
*p = offset;
|
||||
}
|
||||
break;
|
||||
case 2:
|
||||
{
|
||||
HBUINT16 *p = (HBUINT16 *) offsets;
|
||||
for (const auto &_ : +it)
|
||||
{
|
||||
*p++ = offset;
|
||||
offset += length_f (_);
|
||||
}
|
||||
*p = offset;
|
||||
}
|
||||
break;
|
||||
case 3:
|
||||
{
|
||||
HBUINT24 *p = (HBUINT24 *) offsets;
|
||||
for (const auto &_ : +it)
|
||||
{
|
||||
*p++ = offset;
|
||||
offset += length_f (_);
|
||||
}
|
||||
*p = offset;
|
||||
}
|
||||
break;
|
||||
case 4:
|
||||
{
|
||||
HBUINT32 *p = (HBUINT32 *) offsets;
|
||||
for (const auto &_ : +it)
|
||||
{
|
||||
*p++ = offset;
|
||||
offset += length_f (_);
|
||||
}
|
||||
*p = offset;
|
||||
}
|
||||
break;
|
||||
default:
|
||||
break;
|
||||
}
|
||||
|
||||
assert (offset == data_size + 1);
|
||||
return_trace (true);
|
||||
}
|
||||
|
||||
template <typename Iterable,
|
||||
hb_requires (hb_is_iterable (Iterable))>
|
||||
static unsigned total_size (const Iterable &iterable)
|
||||
static unsigned total_size (const Iterable &iterable, unsigned *data_size = nullptr)
|
||||
{
|
||||
auto it = + hb_iter (iterable) | hb_map (hb_iter) | hb_map (hb_len);
|
||||
if (!it) return 0;
|
||||
auto it = + hb_iter (iterable);
|
||||
if (!it)
|
||||
{
|
||||
if (data_size) *data_size = 0;
|
||||
return min_size;
|
||||
}
|
||||
|
||||
unsigned total = 0;
|
||||
for (const auto &_ : +it)
|
||||
total += length_f (_);
|
||||
|
||||
if (data_size) *data_size = total;
|
||||
|
||||
unsigned total = + it | hb_reduce (hb_add, 0);
|
||||
unsigned off_size = (hb_bit_storage (total + 1) + 7) / 8;
|
||||
|
||||
return min_size + HBUINT8::static_size + (hb_len (it) + 1) * off_size + total;
|
||||
@@ -133,13 +218,16 @@ struct CFFIndex
|
||||
void set_offset_at (unsigned int index, unsigned int offset)
|
||||
{
|
||||
assert (index <= count);
|
||||
HBUINT8 *p = offsets + offSize * index + offSize;
|
||||
|
||||
unsigned int size = offSize;
|
||||
for (; size; size--)
|
||||
const HBUINT8 *p = offsets;
|
||||
switch (size)
|
||||
{
|
||||
--p;
|
||||
*p = offset & 0xFF;
|
||||
offset >>= 8;
|
||||
case 1: ((HBUINT8 *) p)[index] = offset; break;
|
||||
case 2: ((HBUINT16 *) p)[index] = offset; break;
|
||||
case 3: ((HBUINT24 *) p)[index] = offset; break;
|
||||
case 4: ((HBUINT32 *) p)[index] = offset; break;
|
||||
default: return;
|
||||
}
|
||||
}
|
||||
|
||||
@@ -149,37 +237,30 @@ struct CFFIndex
|
||||
assert (index <= count);
|
||||
|
||||
unsigned int size = offSize;
|
||||
const HBUINT8 *p = offsets + size * index;
|
||||
const HBUINT8 *p = offsets;
|
||||
switch (size)
|
||||
{
|
||||
case 1: return * (HBUINT8 *) p;
|
||||
case 2: return * (HBUINT16 *) p;
|
||||
case 3: return * (HBUINT24 *) p;
|
||||
case 4: return * (HBUINT32 *) p;
|
||||
case 1: return ((HBUINT8 *) p)[index];
|
||||
case 2: return ((HBUINT16 *) p)[index];
|
||||
case 3: return ((HBUINT24 *) p)[index];
|
||||
case 4: return ((HBUINT32 *) p)[index];
|
||||
default: return 0;
|
||||
}
|
||||
}
|
||||
|
||||
unsigned int length_at (unsigned int index) const
|
||||
{
|
||||
unsigned offset0 = offset_at (index);
|
||||
unsigned offset1 = offset_at (index + 1);
|
||||
if (unlikely (offset1 < offset0 || offset1 > offset_at (count)))
|
||||
return 0;
|
||||
return offset1 - offset0;
|
||||
}
|
||||
|
||||
const unsigned char *data_base () const
|
||||
{ return (const unsigned char *) this + min_size + offSize.static_size + offset_array_size (); }
|
||||
{ return (const unsigned char *) this + min_size + offSize.static_size - 1 + offset_array_size (); }
|
||||
public:
|
||||
|
||||
hb_ubytes_t operator [] (unsigned int index) const
|
||||
{
|
||||
if (unlikely (index >= count)) return hb_ubytes_t ();
|
||||
_hb_compiler_memory_r_barrier ();
|
||||
unsigned length = length_at (index);
|
||||
if (unlikely (!length)) return hb_ubytes_t ();
|
||||
return hb_ubytes_t (data_base () + offset_at (index) - 1, length);
|
||||
unsigned offset0 = offset_at (index);
|
||||
unsigned offset1 = offset_at (index + 1);
|
||||
if (unlikely (offset1 < offset0 || offset1 > offset_at (count)))
|
||||
return hb_ubytes_t ();
|
||||
return hb_ubytes_t (data_base () + offset0, offset1 - offset0);
|
||||
}
|
||||
|
||||
unsigned int get_size () const
|
||||
@@ -197,7 +278,7 @@ struct CFFIndex
|
||||
(count < count + 1u &&
|
||||
c->check_struct (&offSize) && offSize >= 1 && offSize <= 4 &&
|
||||
c->check_array (offsets, offSize, count + 1u) &&
|
||||
c->check_array ((const HBUINT8*) data_base (), 1, offset_at (count) - 1)))));
|
||||
c->check_array ((const HBUINT8*) data_base (), 1, offset_at (count))))));
|
||||
}
|
||||
|
||||
public:
|
||||
@@ -211,47 +292,6 @@ struct CFFIndex
|
||||
DEFINE_SIZE_MIN (COUNT::static_size);
|
||||
};
|
||||
|
||||
template <typename COUNT, typename TYPE>
|
||||
struct CFFIndexOf : CFFIndex<COUNT>
|
||||
{
|
||||
template <typename DATA, typename PARAM1, typename PARAM2>
|
||||
bool serialize (hb_serialize_context_t *c,
|
||||
unsigned int offSize_,
|
||||
const DATA *dataArray,
|
||||
unsigned int dataArrayLen,
|
||||
const hb_vector_t<unsigned int> &dataSizeArray,
|
||||
const PARAM1 ¶m1,
|
||||
const PARAM2 ¶m2)
|
||||
{
|
||||
TRACE_SERIALIZE (this);
|
||||
/* serialize CFFIndex header */
|
||||
if (unlikely (!c->extend_min (this))) return_trace (false);
|
||||
this->count = dataArrayLen;
|
||||
this->offSize = offSize_;
|
||||
if (unlikely (!c->allocate_size<HBUINT8> (offSize_ * (dataArrayLen + 1), false)))
|
||||
return_trace (false);
|
||||
|
||||
/* serialize indices */
|
||||
unsigned int offset = 1;
|
||||
unsigned int i = 0;
|
||||
for (; i < dataArrayLen; i++)
|
||||
{
|
||||
this->set_offset_at (i, offset);
|
||||
offset += dataSizeArray[i];
|
||||
}
|
||||
this->set_offset_at (i, offset);
|
||||
|
||||
/* serialize data */
|
||||
for (unsigned int i = 0; i < dataArrayLen; i++)
|
||||
{
|
||||
TYPE *dest = c->start_embed<TYPE> ();
|
||||
if (unlikely (!dest || !dest->serialize (c, dataArray[i], param1, param2)))
|
||||
return_trace (false);
|
||||
}
|
||||
return_trace (true);
|
||||
}
|
||||
};
|
||||
|
||||
/* Top Dict, Font Dict, Private Dict */
|
||||
struct Dict : UnsizedByteStr
|
||||
{
|
||||
@@ -327,7 +367,7 @@ struct table_info_t
|
||||
};
|
||||
|
||||
template <typename COUNT>
|
||||
struct FDArray : CFFIndexOf<COUNT, FontDict>
|
||||
struct FDArray : CFFIndex<COUNT>
|
||||
{
|
||||
template <typename DICTVAL, typename INFO, typename Iterator, typename OP_SERIALIZER>
|
||||
bool serialize (hb_serialize_context_t *c,
|
||||
@@ -338,7 +378,11 @@ struct FDArray : CFFIndexOf<COUNT, FontDict>
|
||||
|
||||
/* serialize INDEX data */
|
||||
hb_vector_t<unsigned> sizes;
|
||||
if (it.is_random_access_iterator)
|
||||
sizes.alloc (hb_len (it));
|
||||
|
||||
c->push ();
|
||||
char *data_base = c->head;
|
||||
+ it
|
||||
| hb_map ([&] (const hb_pair_t<const DICTVAL&, const INFO&> &_)
|
||||
{
|
||||
@@ -348,10 +392,16 @@ struct FDArray : CFFIndexOf<COUNT, FontDict>
|
||||
})
|
||||
| hb_sink (sizes)
|
||||
;
|
||||
unsigned data_size = c->head - data_base;
|
||||
c->pop_pack (false);
|
||||
|
||||
if (unlikely (sizes.in_error ())) return_trace (false);
|
||||
|
||||
/* It just happens that the above is packed right after the header below.
|
||||
* Such a hack. */
|
||||
|
||||
/* serialize INDEX header */
|
||||
return_trace (CFFIndex<COUNT>::serialize_header (c, hb_iter (sizes)));
|
||||
return_trace (CFFIndex<COUNT>::serialize_header (c, hb_iter (sizes), data_size));
|
||||
}
|
||||
};
|
||||
|
||||
@@ -368,8 +418,11 @@ struct FDSelect0 {
|
||||
return_trace (true);
|
||||
}
|
||||
|
||||
hb_codepoint_t get_fd (hb_codepoint_t glyph) const
|
||||
{ return (hb_codepoint_t) fds[glyph]; }
|
||||
unsigned get_fd (hb_codepoint_t glyph) const
|
||||
{ return fds[glyph]; }
|
||||
|
||||
hb_pair_t<unsigned, hb_codepoint_t> get_fd_range (hb_codepoint_t glyph) const
|
||||
{ return {fds[glyph], glyph + 1}; }
|
||||
|
||||
unsigned int get_size (unsigned int num_glyphs) const
|
||||
{ return HBUINT8::static_size * num_glyphs; }
|
||||
@@ -427,12 +480,20 @@ struct FDSelect3_4
|
||||
return +1;
|
||||
}
|
||||
|
||||
hb_codepoint_t get_fd (hb_codepoint_t glyph) const
|
||||
unsigned get_fd (hb_codepoint_t glyph) const
|
||||
{
|
||||
auto *range = hb_bsearch (glyph, &ranges[0], nRanges () - 1, sizeof (ranges[0]), _cmp_range);
|
||||
return range ? range->fd : ranges[nRanges () - 1].fd;
|
||||
}
|
||||
|
||||
hb_pair_t<unsigned, hb_codepoint_t> get_fd_range (hb_codepoint_t glyph) const
|
||||
{
|
||||
auto *range = hb_bsearch (glyph, &ranges[0], nRanges () - 1, sizeof (ranges[0]), _cmp_range);
|
||||
unsigned fd = range ? range->fd : ranges[nRanges () - 1].fd;
|
||||
hb_codepoint_t end = range ? range[1].first : ranges[nRanges () - 1].first;
|
||||
return {fd, end};
|
||||
}
|
||||
|
||||
GID_TYPE &nRanges () { return ranges.len; }
|
||||
GID_TYPE nRanges () const { return ranges.len; }
|
||||
GID_TYPE &sentinel () { return StructAfter<GID_TYPE> (ranges[nRanges () - 1]); }
|
||||
@@ -469,7 +530,7 @@ struct FDSelect
|
||||
}
|
||||
}
|
||||
|
||||
hb_codepoint_t get_fd (hb_codepoint_t glyph) const
|
||||
unsigned get_fd (hb_codepoint_t glyph) const
|
||||
{
|
||||
if (this == &Null (FDSelect)) return 0;
|
||||
|
||||
@@ -480,6 +541,18 @@ struct FDSelect
|
||||
default:return 0;
|
||||
}
|
||||
}
|
||||
/* Returns pair of fd and one after last glyph in range. */
|
||||
hb_pair_t<unsigned, hb_codepoint_t> get_fd_range (hb_codepoint_t glyph) const
|
||||
{
|
||||
if (this == &Null (FDSelect)) return {0, 1};
|
||||
|
||||
switch (format)
|
||||
{
|
||||
case 0: return u.format0.get_fd_range (glyph);
|
||||
case 3: return u.format3.get_fd_range (glyph);
|
||||
default:return {0, 1};
|
||||
}
|
||||
}
|
||||
|
||||
bool sanitize (hb_sanitize_context_t *c, unsigned int fdcount) const
|
||||
{
|
||||
|
||||
Reference in New Issue
Block a user