blob: be3f57fcdb33d9a209ebc21faf52babbd549f768 [file] [log] [blame]
/*
* Copyright 2012 Google Inc.
*
* Use of this source code is governed by a BSD-style license that can be
* found in the LICENSE file.
*/
#include "Simplify.h"
#undef SkASSERT
#define SkASSERT(cond) while (!(cond)) { sk_throw(); }
// FIXME: remove once debugging is complete
#if 01 // set to 1 for no debugging whatsoever
//const bool gRunTestsInOneThread = false;
#define DEBUG_ACTIVE_LESS_THAN 0
#define DEBUG_ADD 0
#define DEBUG_ADD_BOTTOM_TS 0
#define DEBUG_ADD_INTERSECTING_TS 0
#define DEBUG_ADJUST_COINCIDENT 0
#define DEBUG_ASSEMBLE 0
#define DEBUG_BOTTOM 0
#define DEBUG_BRIDGE 0
#define DEBUG_DUMP 0
#define DEBUG_SORT_HORIZONTAL 0
#define DEBUG_OUT 0
#define DEBUG_OUT_LESS_THAN 0
#define DEBUG_SPLIT 0
#define DEBUG_STITCH_EDGE 0
#define DEBUG_TRIM_LINE 0
#else
//const bool gRunTestsInOneThread = true;
#define DEBUG_ACTIVE_LESS_THAN 0
#define DEBUG_ADD 01
#define DEBUG_ADD_BOTTOM_TS 0
#define DEBUG_ADD_INTERSECTING_TS 0
#define DEBUG_ADJUST_COINCIDENT 1
#define DEBUG_ASSEMBLE 1
#define DEBUG_BOTTOM 0
#define DEBUG_BRIDGE 1
#define DEBUG_DUMP 1
#define DEBUG_SORT_HORIZONTAL 01
#define DEBUG_OUT 01
#define DEBUG_OUT_LESS_THAN 0
#define DEBUG_SPLIT 1
#define DEBUG_STITCH_EDGE 1
#define DEBUG_TRIM_LINE 1
#endif
#if DEBUG_ASSEMBLE || DEBUG_BRIDGE
static const char* kLVerbStr[] = {"", "line", "quad", "cubic"};
#endif
#if DEBUG_STITCH_EDGE
static const char* kUVerbStr[] = {"", "Line", "Quad", "Cubic"};
#endif
static int LineIntersect(const SkPoint a[2], const SkPoint b[2],
Intersections& intersections) {
const _Line aLine = {{a[0].fX, a[0].fY}, {a[1].fX, a[1].fY}};
const _Line bLine = {{b[0].fX, b[0].fY}, {b[1].fX, b[1].fY}};
return intersect(aLine, bLine, intersections);
}
static int QuadLineIntersect(const SkPoint a[3], const SkPoint b[2],
Intersections& intersections) {
const Quadratic aQuad = {{a[0].fX, a[0].fY}, {a[1].fX, a[1].fY}, {a[2].fX, a[2].fY}};
const _Line bLine = {{b[0].fX, b[0].fY}, {b[1].fX, b[1].fY}};
intersect(aQuad, bLine, intersections);
return intersections.fUsed;
}
static int CubicLineIntersect(const SkPoint a[2], const SkPoint b[3],
Intersections& intersections) {
const Cubic aCubic = {{a[0].fX, a[0].fY}, {a[1].fX, a[1].fY}, {a[2].fX, a[2].fY},
{a[3].fX, a[3].fY}};
const _Line bLine = {{b[0].fX, b[0].fY}, {b[1].fX, b[1].fY}};
return intersect(aCubic, bLine, intersections);
}
static int QuadIntersect(const SkPoint a[3], const SkPoint b[3],
Intersections& intersections) {
const Quadratic aQuad = {{a[0].fX, a[0].fY}, {a[1].fX, a[1].fY}, {a[2].fX, a[2].fY}};
const Quadratic bQuad = {{b[0].fX, b[0].fY}, {b[1].fX, b[1].fY}, {b[2].fX, b[2].fY}};
intersect2(aQuad, bQuad, intersections);
return intersections.fUsed;
}
static int CubicIntersect(const SkPoint a[4], const SkPoint b[4],
Intersections& intersections) {
const Cubic aCubic = {{a[0].fX, a[0].fY}, {a[1].fX, a[1].fY}, {a[2].fX, a[2].fY},
{a[3].fX, a[3].fY}};
const Cubic bCubic = {{b[0].fX, b[0].fY}, {b[1].fX, b[1].fY}, {b[2].fX, b[2].fY},
{b[3].fX, b[3].fY}};
intersect(aCubic, bCubic, intersections);
return intersections.fUsed;
}
static int LineIntersect(const SkPoint a[2], SkScalar left, SkScalar right,
SkScalar y, double aRange[2]) {
const _Line aLine = {{a[0].fX, a[0].fY}, {a[1].fX, a[1].fY}};
return horizontalLineIntersect(aLine, left, right, y, aRange);
}
static int QuadIntersect(const SkPoint a[3], SkScalar left, SkScalar right,
SkScalar y, double aRange[3]) {
const Quadratic aQuad = {{a[0].fX, a[0].fY}, {a[1].fX, a[1].fY}, {a[2].fX, a[2].fY}};
return horizontalIntersect(aQuad, left, right, y, aRange);
}
static int CubicIntersect(const SkPoint a[4], SkScalar left, SkScalar right,
SkScalar y, double aRange[4]) {
const Cubic aCubic = {{a[0].fX, a[0].fY}, {a[1].fX, a[1].fY}, {a[2].fX, a[2].fY},
{a[3].fX, a[3].fY}};
return horizontalIntersect(aCubic, left, right, y, aRange);
}
static void LineXYAtT(const SkPoint a[2], double t, SkPoint* out) {
const _Line line = {{a[0].fX, a[0].fY}, {a[1].fX, a[1].fY}};
double x, y;
xy_at_t(line, t, x, y);
out->fX = SkDoubleToScalar(x);
out->fY = SkDoubleToScalar(y);
}
static void QuadXYAtT(const SkPoint a[3], double t, SkPoint* out) {
const Quadratic quad = {{a[0].fX, a[0].fY}, {a[1].fX, a[1].fY}, {a[2].fX, a[2].fY}};
double x, y;
xy_at_t(quad, t, x, y);
out->fX = SkDoubleToScalar(x);
out->fY = SkDoubleToScalar(y);
}
static void CubicXYAtT(const SkPoint a[4], double t, SkPoint* out) {
const Cubic cubic = {{a[0].fX, a[0].fY}, {a[1].fX, a[1].fY}, {a[2].fX, a[2].fY},
{a[3].fX, a[3].fY}};
double x, y;
xy_at_t(cubic, t, x, y);
out->fX = SkDoubleToScalar(x);
out->fY = SkDoubleToScalar(y);
}
static SkScalar LineYAtT(const SkPoint a[2], double t) {
const _Line aLine = {{a[0].fX, a[0].fY}, {a[1].fX, a[1].fY}};
double y;
xy_at_t(aLine, t, *(double*) 0, y);
return SkDoubleToScalar(y);
}
static SkScalar QuadYAtT(const SkPoint a[3], double t) {
const Quadratic quad = {{a[0].fX, a[0].fY}, {a[1].fX, a[1].fY}, {a[2].fX, a[2].fY}};
double y;
xy_at_t(quad, t, *(double*) 0, y);
return SkDoubleToScalar(y);
}
static SkScalar CubicYAtT(const SkPoint a[4], double t) {
const Cubic cubic = {{a[0].fX, a[0].fY}, {a[1].fX, a[1].fY}, {a[2].fX, a[2].fY},
{a[3].fX, a[3].fY}};
double y;
xy_at_t(cubic, t, *(double*) 0, y);
return SkDoubleToScalar(y);
}
static void LineSubDivide(const SkPoint a[2], double startT, double endT,
SkPoint sub[2]) {
const _Line aLine = {{a[0].fX, a[0].fY}, {a[1].fX, a[1].fY}};
_Line dst;
sub_divide(aLine, startT, endT, dst);
sub[0].fX = SkDoubleToScalar(dst[0].x);
sub[0].fY = SkDoubleToScalar(dst[0].y);
sub[1].fX = SkDoubleToScalar(dst[1].x);
sub[1].fY = SkDoubleToScalar(dst[1].y);
}
static void QuadSubDivide(const SkPoint a[3], double startT, double endT,
SkPoint sub[3]) {
const Quadratic aQuad = {{a[0].fX, a[0].fY}, {a[1].fX, a[1].fY},
{a[2].fX, a[2].fY}};
Quadratic dst;
sub_divide(aQuad, startT, endT, dst);
sub[0].fX = SkDoubleToScalar(dst[0].x);
sub[0].fY = SkDoubleToScalar(dst[0].y);
sub[1].fX = SkDoubleToScalar(dst[1].x);
sub[1].fY = SkDoubleToScalar(dst[1].y);
sub[2].fX = SkDoubleToScalar(dst[2].x);
sub[2].fY = SkDoubleToScalar(dst[2].y);
}
static void CubicSubDivide(const SkPoint a[4], double startT, double endT,
SkPoint sub[4]) {
const Cubic aCubic = {{a[0].fX, a[0].fY}, {a[1].fX, a[1].fY},
{a[2].fX, a[2].fY}, {a[3].fX, a[3].fY}};
Cubic dst;
sub_divide(aCubic, startT, endT, dst);
sub[0].fX = SkDoubleToScalar(dst[0].x);
sub[0].fY = SkDoubleToScalar(dst[0].y);
sub[1].fX = SkDoubleToScalar(dst[1].x);
sub[1].fY = SkDoubleToScalar(dst[1].y);
sub[2].fX = SkDoubleToScalar(dst[2].x);
sub[2].fY = SkDoubleToScalar(dst[2].y);
sub[3].fX = SkDoubleToScalar(dst[3].x);
sub[3].fY = SkDoubleToScalar(dst[3].y);
}
static void QuadSubBounds(const SkPoint a[3], double startT, double endT,
SkRect& bounds) {
SkPoint dst[3];
QuadSubDivide(a, startT, endT, dst);
bounds.fLeft = bounds.fRight = dst[0].fX;
bounds.fTop = bounds.fBottom = dst[0].fY;
for (int index = 1; index < 3; ++index) {
bounds.growToInclude(dst[index].fX, dst[index].fY);
}
}
static void CubicSubBounds(const SkPoint a[4], double startT, double endT,
SkRect& bounds) {
SkPoint dst[4];
CubicSubDivide(a, startT, endT, dst);
bounds.fLeft = bounds.fRight = dst[0].fX;
bounds.fTop = bounds.fBottom = dst[0].fY;
for (int index = 1; index < 4; ++index) {
bounds.growToInclude(dst[index].fX, dst[index].fY);
}
}
static SkPath::Verb QuadReduceOrder(SkPoint a[4]) {
const Quadratic aQuad = {{a[0].fX, a[0].fY}, {a[1].fX, a[1].fY},
{a[2].fX, a[2].fY}};
Quadratic dst;
int order = reduceOrder(aQuad, dst, kReduceOrder_TreatAsFill);
for (int index = 0; index < order; ++index) {
a[index].fX = SkDoubleToScalar(dst[index].x);
a[index].fY = SkDoubleToScalar(dst[index].y);
}
if (order == 1) { // FIXME: allow returning points, caller should discard
a[1] = a[0];
return (SkPath::Verb) order;
}
return (SkPath::Verb) (order - 1);
}
static SkPath::Verb CubicReduceOrder(SkPoint a[4]) {
const Cubic aCubic = {{a[0].fX, a[0].fY}, {a[1].fX, a[1].fY},
{a[2].fX, a[2].fY}, {a[3].fX, a[3].fY}};
Cubic dst;
int order = reduceOrder(aCubic, dst, kReduceOrder_QuadraticsAllowed, kReduceOrder_TreatAsFill);
for (int index = 0; index < order; ++index) {
a[index].fX = SkDoubleToScalar(dst[index].x);
a[index].fY = SkDoubleToScalar(dst[index].y);
}
if (order == 1) { // FIXME: allow returning points, caller should discard
a[1] = a[0];
return (SkPath::Verb) order;
}
return (SkPath::Verb) (order - 1);
}
static bool IsCoincident(const SkPoint a[2], const SkPoint& above,
const SkPoint& below) {
const _Line aLine = {{a[0].fX, a[0].fY}, {a[1].fX, a[1].fY}};
const _Line bLine = {{above.fX, above.fY}, {below.fX, below.fY}};
return implicit_matches_ulps(aLine, bLine, 32);
}
/*
list of edges
bounds for edge
sort
active T
if a contour's bounds is outside of the active area, no need to create edges
*/
/* given one or more paths,
find the bounds of each contour, select the active contours
for each active contour, compute a set of edges
each edge corresponds to one or more lines and curves
leave edges unbroken as long as possible
when breaking edges, compute the t at the break but leave the control points alone
*/
void contourBounds(const SkPath& path, SkTDArray<SkRect>& boundsArray) {
SkPath::Iter iter(path, false);
SkPoint pts[4];
SkPath::Verb verb;
SkRect bounds;
bounds.setEmpty();
int count = 0;
while ((verb = iter.next(pts)) != SkPath::kDone_Verb) {
switch (verb) {
case SkPath::kMove_Verb:
if (!bounds.isEmpty()) {
*boundsArray.append() = bounds;
}
bounds.set(pts[0].fX, pts[0].fY, pts[0].fX, pts[0].fY);
count = 0;
break;
case SkPath::kLine_Verb:
count = 1;
break;
case SkPath::kQuad_Verb:
count = 2;
break;
case SkPath::kCubic_Verb:
count = 3;
break;
case SkPath::kClose_Verb:
count = 0;
break;
default:
SkDEBUGFAIL("bad verb");
return;
}
for (int i = 1; i <= count; ++i) {
bounds.growToInclude(pts[i].fX, pts[i].fY);
}
}
}
static bool extendLine(const SkPoint line[2], const SkPoint& add) {
// FIXME: allow this to extend lines that have slopes that are nearly equal
SkScalar dx1 = line[1].fX - line[0].fX;
SkScalar dy1 = line[1].fY - line[0].fY;
SkScalar dx2 = add.fX - line[0].fX;
SkScalar dy2 = add.fY - line[0].fY;
return dx1 * dy2 == dx2 * dy1;
}
// OPTIMIZATION: this should point to a list of input data rather than duplicating
// the line data here. This would reduce the need to assemble the results.
struct OutEdge {
bool operator<(const OutEdge& rh) const {
const SkPoint& first = fPts[0];
const SkPoint& rhFirst = rh.fPts[0];
return first.fY == rhFirst.fY
? first.fX < rhFirst.fX
: first.fY < rhFirst.fY;
}
SkPoint fPts[4];
int fID; // id of edge generating data
uint8_t fVerb; // FIXME: not read from everywhere
bool fCloseCall; // edge is trimmable if not originally coincident
};
class OutEdgeBuilder {
public:
OutEdgeBuilder(bool fill)
: fFill(fill) {
}
void addCurve(const SkPoint line[4], SkPath::Verb verb, int id,
bool closeCall) {
OutEdge& newEdge = fEdges.push_back();
memcpy(newEdge.fPts, line, (verb + 1) * sizeof(SkPoint));
newEdge.fVerb = verb;
newEdge.fID = id;
newEdge.fCloseCall = closeCall;
}
bool trimLine(SkScalar y, int id) {
size_t count = fEdges.count();
while (count-- != 0) {
OutEdge& edge = fEdges[count];
if (edge.fID != id) {
continue;
}
if (edge.fCloseCall) {
return false;
}
SkASSERT(edge.fPts[0].fY <= y);
if (edge.fPts[1].fY <= y) {
continue;
}
edge.fPts[1].fX = edge.fPts[0].fX + (y - edge.fPts[0].fY)
* (edge.fPts[1].fX - edge.fPts[0].fX)
/ (edge.fPts[1].fY - edge.fPts[0].fY);
edge.fPts[1].fY = y;
#if DEBUG_TRIM_LINE
SkDebugf("%s edge=%d %1.9g,%1.9g\n", __FUNCTION__, id,
edge.fPts[1].fX, y);
#endif
return true;
}
return false;
}
void assemble(SkPath& simple) {
size_t listCount = fEdges.count();
if (listCount == 0) {
return;
}
do {
size_t listIndex = 0;
int advance = 1;
while (listIndex < listCount && fTops[listIndex] == 0) {
++listIndex;
}
if (listIndex >= listCount) {
break;
}
int closeEdgeIndex = -listIndex - 1;
// the curve is deferred and not added right away because the
// following edge may extend the first curve.
SkPoint firstPt, lastCurve[4];
uint8_t lastVerb;
#if DEBUG_ASSEMBLE
int firstIndex, lastIndex;
const int tab = 8;
#endif
bool doMove = true;
int edgeIndex;
do {
SkPoint* ptArray = fEdges[listIndex].fPts;
uint8_t verb = fEdges[listIndex].fVerb;
SkPoint* curve[4];
if (advance < 0) {
curve[0] = &ptArray[verb];
if (verb == SkPath::kCubic_Verb) {
curve[1] = &ptArray[2];
curve[2] = &ptArray[1];
}
curve[verb] = &ptArray[0];
} else {
curve[0] = &ptArray[0];
if (verb == SkPath::kCubic_Verb) {
curve[1] = &ptArray[1];
curve[2] = &ptArray[2];
}
curve[verb] = &ptArray[verb];
}
if (verb == SkPath::kQuad_Verb) {
curve[1] = &ptArray[1];
}
if (doMove) {
firstPt = *curve[0];
simple.moveTo(curve[0]->fX, curve[0]->fY);
#if DEBUG_ASSEMBLE
SkDebugf("%s %d moveTo (%g,%g)\n", __FUNCTION__,
listIndex + 1, curve[0]->fX, curve[0]->fY);
firstIndex = listIndex;
#endif
for (int index = 0; index <= verb; ++index) {
lastCurve[index] = *curve[index];
}
doMove = false;
} else {
bool gap = lastCurve[lastVerb] != *curve[0];
if (gap || lastVerb != SkPath::kLine_Verb) { // output the accumulated curve before the gap
// FIXME: see comment in bridge -- this probably
// conceals errors
SkASSERT(fFill && UlpsDiff(lastCurve[lastVerb].fY,
curve[0]->fY) <= 10);
switch (lastVerb) {
case SkPath::kLine_Verb:
simple.lineTo(lastCurve[1].fX, lastCurve[1].fY);
break;
case SkPath::kQuad_Verb:
simple.quadTo(lastCurve[1].fX, lastCurve[1].fY,
lastCurve[2].fX, lastCurve[2].fY);
break;
case SkPath::kCubic_Verb:
simple.cubicTo(lastCurve[1].fX, lastCurve[1].fY,
lastCurve[2].fX, lastCurve[2].fY,
lastCurve[3].fX, lastCurve[3].fY);
break;
}
#if DEBUG_ASSEMBLE
SkDebugf("%*s %d %sTo (%g,%g)\n", tab, "", lastIndex + 1,
kLVerbStr[lastVerb], lastCurve[lastVerb].fX,
lastCurve[lastVerb].fY);
#endif
}
int firstCopy = 1;
if (gap || (lastVerb == SkPath::kLine_Verb
&& (verb != SkPath::kLine_Verb
|| !extendLine(lastCurve, *curve[verb])))) {
// FIXME: see comment in bridge -- this probably
// conceals errors
SkASSERT(lastCurve[lastVerb] == *curve[0] ||
(fFill && UlpsDiff(lastCurve[lastVerb].fY,
curve[0]->fY) <= 10));
simple.lineTo(curve[0]->fX, curve[0]->fY);
#if DEBUG_ASSEMBLE
SkDebugf("%*s %d gap lineTo (%g,%g)\n", tab, "",
lastIndex + 1, curve[0]->fX, curve[0]->fY);
#endif
firstCopy = 0;
} else if (lastVerb != SkPath::kLine_Verb) {
firstCopy = 0;
}
for (int index = firstCopy; index <= verb; ++index) {
lastCurve[index] = *curve[index];
}
}
lastVerb = verb;
#if DEBUG_ASSEMBLE
lastIndex = listIndex;
#endif
if (advance < 0) {
edgeIndex = fTops[listIndex];
fTops[listIndex] = 0;
} else {
edgeIndex = fBottoms[listIndex];
fBottoms[listIndex] = 0;
}
if (edgeIndex) {
listIndex = abs(edgeIndex) - 1;
if (edgeIndex < 0) {
fTops[listIndex] = 0;
} else {
fBottoms[listIndex] = 0;
}
}
if (edgeIndex == closeEdgeIndex || edgeIndex == 0) {
switch (lastVerb) {
case SkPath::kLine_Verb:
simple.lineTo(lastCurve[1].fX, lastCurve[1].fY);
break;
case SkPath::kQuad_Verb:
simple.quadTo(lastCurve[1].fX, lastCurve[1].fY,
lastCurve[2].fX, lastCurve[2].fY);
break;
case SkPath::kCubic_Verb:
simple.cubicTo(lastCurve[1].fX, lastCurve[1].fY,
lastCurve[2].fX, lastCurve[2].fY,
lastCurve[3].fX, lastCurve[3].fY);
break;
}
#if DEBUG_ASSEMBLE
SkDebugf("%*s %d %sTo last (%g, %g)\n", tab, "",
lastIndex + 1, kLVerbStr[lastVerb],
lastCurve[lastVerb].fX, lastCurve[lastVerb].fY);
#endif
if (lastCurve[lastVerb] != firstPt) {
simple.lineTo(firstPt.fX, firstPt.fY);
#if DEBUG_ASSEMBLE
SkDebugf("%*s %d final line (%g, %g)\n", tab, "",
firstIndex + 1, firstPt.fX, firstPt.fY);
#endif
}
simple.close();
#if DEBUG_ASSEMBLE
SkDebugf("%*s close\n", tab, "");
#endif
break;
}
// if this and next edge go different directions
#if DEBUG_ASSEMBLE
SkDebugf("%*s advance=%d edgeIndex=%d flip=%s\n", tab, "",
advance, edgeIndex, advance > 0 ^ edgeIndex < 0 ?
"true" : "false");
#endif
if (advance > 0 ^ edgeIndex < 0) {
advance = -advance;
}
} while (edgeIndex);
} while (true);
}
// sort points by y, then x
// if x/y is identical, sort bottoms before tops
// if identical and both tops/bottoms, sort by angle
static bool lessThan(SkTArray<OutEdge>& edges, const int one,
const int two) {
const OutEdge& oneEdge = edges[abs(one) - 1];
int oneIndex = one < 0 ? 0 : oneEdge.fVerb;
const SkPoint& startPt1 = oneEdge.fPts[oneIndex];
const OutEdge& twoEdge = edges[abs(two) - 1];
int twoIndex = two < 0 ? 0 : twoEdge.fVerb;
const SkPoint& startPt2 = twoEdge.fPts[twoIndex];
if (startPt1.fY != startPt2.fY) {
#if DEBUG_OUT_LESS_THAN
SkDebugf("%s %d<%d (%g,%g) %s startPt1.fY < startPt2.fY\n", __FUNCTION__,
one, two, startPt1.fY, startPt2.fY,
startPt1.fY < startPt2.fY ? "true" : "false");
#endif
return startPt1.fY < startPt2.fY;
}
if (startPt1.fX != startPt2.fX) {
#if DEBUG_OUT_LESS_THAN
SkDebugf("%s %d<%d (%g,%g) %s startPt1.fX < startPt2.fX\n", __FUNCTION__,
one, two, startPt1.fX, startPt2.fX,
startPt1.fX < startPt2.fX ? "true" : "false");
#endif
return startPt1.fX < startPt2.fX;
}
const SkPoint& endPt1 = oneEdge.fPts[oneIndex ^ oneEdge.fVerb];
const SkPoint& endPt2 = twoEdge.fPts[twoIndex ^ twoEdge.fVerb];
SkScalar dy1 = startPt1.fY - endPt1.fY;
SkScalar dy2 = startPt2.fY - endPt2.fY;
SkScalar dy1y2 = dy1 * dy2;
if (dy1y2 < 0) { // different signs
#if DEBUG_OUT_LESS_THAN
SkDebugf("%s %d<%d %s dy1 > 0\n", __FUNCTION__, one, two,
dy1 > 0 ? "true" : "false");
#endif
return dy1 > 0; // one < two if one goes up and two goes down
}
if (dy1y2 == 0) {
#if DEBUG_OUT_LESS_THAN
SkDebugf("%s %d<%d %s endPt1.fX < endPt2.fX\n", __FUNCTION__,
one, two, endPt1.fX < endPt2.fX ? "true" : "false");
#endif
return endPt1.fX < endPt2.fX;
}
SkScalar dx1y2 = (startPt1.fX - endPt1.fX) * dy2;
SkScalar dx2y1 = (startPt2.fX - endPt2.fX) * dy1;
#if DEBUG_OUT_LESS_THAN
SkDebugf("%s %d<%d %s dy2 < 0 ^ dx1y2 < dx2y1\n", __FUNCTION__,
one, two, dy2 < 0 ^ dx1y2 < dx2y1 ? "true" : "false");
#endif
return dy2 > 0 ^ dx1y2 < dx2y1;
}
// Sort the indices of paired points and then create more indices so
// assemble() can find the next edge and connect the top or bottom
void bridge() {
size_t index;
size_t count = fEdges.count();
if (!count) {
return;
}
SkASSERT(!fFill || count > 1);
fTops.setCount(count);
sk_bzero(fTops.begin(), sizeof(fTops[0]) * count);
fBottoms.setCount(count);
sk_bzero(fBottoms.begin(), sizeof(fBottoms[0]) * count);
SkTDArray<int> order;
for (index = 1; index <= count; ++index) {
*order.append() = -index;
}
for (index = 1; index <= count; ++index) {
*order.append() = index;
}
QSort<SkTArray<OutEdge>, int>(fEdges, order.begin(), order.end() - 1, lessThan);
int* lastPtr = order.end() - 1;
int* leftPtr = order.begin();
while (leftPtr < lastPtr) {
int leftIndex = *leftPtr;
int leftOutIndex = abs(leftIndex) - 1;
const OutEdge& left = fEdges[leftOutIndex];
int* rightPtr = leftPtr + 1;
int rightIndex = *rightPtr;
int rightOutIndex = abs(rightIndex) - 1;
const OutEdge& right = fEdges[rightOutIndex];
bool pairUp = fFill;
if (!pairUp) {
const SkPoint& leftMatch =
left.fPts[leftIndex < 0 ? 0 : left.fVerb];
const SkPoint& rightMatch =
right.fPts[rightIndex < 0 ? 0 : right.fVerb];
pairUp = leftMatch == rightMatch;
} else {
#if DEBUG_OUT
// FIXME : not happy that error in low bit is allowed
// this probably conceals error elsewhere
if (UlpsDiff(left.fPts[leftIndex < 0 ? 0 : left.fVerb].fY,
right.fPts[rightIndex < 0 ? 0 : right.fVerb].fY) > 1) {
*fMismatches.append() = leftIndex;
if (rightPtr == lastPtr) {
*fMismatches.append() = rightIndex;
}
pairUp = false;
}
#else
SkASSERT(UlpsDiff(left.fPts[leftIndex < 0 ? 0 : left.fVerb].fY,
right.fPts[rightIndex < 0 ? 0 : right.fVerb].fY) <= 10);
#endif
}
if (pairUp) {
if (leftIndex < 0) {
fTops[leftOutIndex] = rightIndex;
} else {
fBottoms[leftOutIndex] = rightIndex;
}
if (rightIndex < 0) {
fTops[rightOutIndex] = leftIndex;
} else {
fBottoms[rightOutIndex] = leftIndex;
}
++rightPtr;
}
leftPtr = rightPtr;
}
#if DEBUG_OUT
int* mismatch = fMismatches.begin();
while (mismatch != fMismatches.end()) {
int leftIndex = *mismatch++;
int leftOutIndex = abs(leftIndex) - 1;
const OutEdge& left = fEdges[leftOutIndex];
const SkPoint& leftPt = left.fPts[leftIndex < 0 ? 0 : left.fVerb];
SkDebugf("%s left=%d %s (%1.9g,%1.9g)\n",
__FUNCTION__, left.fID, leftIndex < 0 ? "top" : "bot",
leftPt.fX, leftPt.fY);
}
SkASSERT(fMismatches.count() == 0);
#endif
#if DEBUG_BRIDGE
for (index = 0; index < count; ++index) {
const OutEdge& edge = fEdges[index];
uint8_t verb = edge.fVerb;
SkDebugf("%s %d edge=%d %s (%1.9g,%1.9g) (%1.9g,%1.9g)\n",
index == 0 ? __FUNCTION__ : " ",
index + 1, edge.fID, kLVerbStr[verb], edge.fPts[0].fX,
edge.fPts[0].fY, edge.fPts[verb].fX, edge.fPts[verb].fY);
}
for (index = 0; index < count; ++index) {
SkDebugf(" top of % 2d connects to %s of % 2d\n", index + 1,
fTops[index] < 0 ? "top " : "bottom", abs(fTops[index]));
SkDebugf(" bottom of % 2d connects to %s of % 2d\n", index + 1,
fBottoms[index] < 0 ? "top " : "bottom", abs(fBottoms[index]));
}
#endif
}
protected:
SkTArray<OutEdge> fEdges;
SkTDArray<int> fTops;
SkTDArray<int> fBottoms;
bool fFill;
#if DEBUG_OUT
SkTDArray<int> fMismatches;
#endif
};
// Bounds, unlike Rect, does not consider a vertical line to be empty.
struct Bounds : public SkRect {
static bool Intersects(const Bounds& a, const Bounds& b) {
return a.fLeft <= b.fRight && b.fLeft <= a.fRight &&
a.fTop <= b.fBottom && b.fTop <= a.fBottom;
}
bool isEmpty() {
return fLeft > fRight || fTop > fBottom
|| (fLeft == fRight && fTop == fBottom)
|| isnan(fLeft) || isnan(fRight)
|| isnan(fTop) || isnan(fBottom);
}
};
class Intercepts {
public:
Intercepts()
: fTopIntercepts(0)
, fBottomIntercepts(0)
, fExplicit(false) {
}
Intercepts& operator=(const Intercepts& src) {
fTs = src.fTs;
fTopIntercepts = src.fTopIntercepts;
fBottomIntercepts = src.fBottomIntercepts;
return *this;
}
// OPTIMIZATION: remove this function if it's never called
double t(int tIndex) const {
if (tIndex == 0) {
return 0;
}
if (tIndex > fTs.count()) {
return 1;
}
return fTs[tIndex - 1];
}
#if DEBUG_DUMP
void dump(const SkPoint* pts, SkPath::Verb verb) {
const char className[] = "Intercepts";
const int tab = 8;
for (int i = 0; i < fTs.count(); ++i) {
SkPoint out;
switch (verb) {
case SkPath::kLine_Verb:
LineXYAtT(pts, fTs[i], &out);
break;
case SkPath::kQuad_Verb:
QuadXYAtT(pts, fTs[i], &out);
break;
case SkPath::kCubic_Verb:
CubicXYAtT(pts, fTs[i], &out);
break;
default:
SkASSERT(0);
}
SkDebugf("%*s.fTs[%d]=%1.9g (%1.9g,%1.9g)\n", tab + sizeof(className),
className, i, fTs[i], out.fX, out.fY);
}
SkDebugf("%*s.fTopIntercepts=%u\n", tab + sizeof(className),
className, fTopIntercepts);
SkDebugf("%*s.fBottomIntercepts=%u\n", tab + sizeof(className),
className, fBottomIntercepts);
SkDebugf("%*s.fExplicit=%d\n", tab + sizeof(className),
className, fExplicit);
}
#endif
SkTDArray<double> fTs;
unsigned char fTopIntercepts; // 0=init state 1=1 edge >1=multiple edges
unsigned char fBottomIntercepts;
bool fExplicit; // if set, suppress 0 and 1
};
struct HorizontalEdge {
bool operator<(const HorizontalEdge& rh) const {
return fY == rh.fY ? fLeft == rh.fLeft ? fRight < rh.fRight
: fLeft < rh.fLeft : fY < rh.fY;
}
#if DEBUG_DUMP
void dump() {
const char className[] = "HorizontalEdge";
const int tab = 4;
SkDebugf("%*s.fLeft=%1.9g\n", tab + sizeof(className), className, fLeft);
SkDebugf("%*s.fRight=%1.9g\n", tab + sizeof(className), className, fRight);
SkDebugf("%*s.fY=%1.9g\n", tab + sizeof(className), className, fY);
}
#endif
SkScalar fLeft;
SkScalar fRight;
SkScalar fY;
};
struct InEdge {
bool operator<(const InEdge& rh) const {
return fBounds.fTop == rh.fBounds.fTop
? fBounds.fLeft < rh.fBounds.fLeft
: fBounds.fTop < rh.fBounds.fTop;
}
// Avoid collapsing t values that are close to the same since
// we walk ts to describe consecutive intersections. Since a pair of ts can
// be nearly equal, any problems caused by this should be taken care
// of later.
int add(double* ts, size_t count, ptrdiff_t verbIndex) {
// FIXME: in the pathological case where there is a ton of intercepts, binary search?
bool foundIntercept = false;
int insertedAt = -1;
Intercepts& intercepts = fIntercepts[verbIndex];
for (size_t index = 0; index < count; ++index) {
double t = ts[index];
if (t <= 0) {
intercepts.fTopIntercepts <<= 1;
fContainsIntercepts |= ++intercepts.fTopIntercepts > 1;
continue;
}
if (t >= 1) {
intercepts.fBottomIntercepts <<= 1;
fContainsIntercepts |= ++intercepts.fBottomIntercepts > 1;
continue;
}
fIntersected = true;
foundIntercept = true;
size_t tCount = intercepts.fTs.count();
double delta;
for (size_t idx2 = 0; idx2 < tCount; ++idx2) {
if (t <= intercepts.fTs[idx2]) {
// FIXME: ? if (t < intercepts.fTs[idx2]) // failed
delta = intercepts.fTs[idx2] - t;
if (delta > 0) {
insertedAt = idx2;
*intercepts.fTs.insert(idx2) = t;
}
goto nextPt;
}
}
if (tCount == 0 || (delta = t - intercepts.fTs[tCount - 1]) > 0) {
insertedAt = tCount;
*intercepts.fTs.append() = t;
}
nextPt:
;
}
fContainsIntercepts |= foundIntercept;
return insertedAt;
}
void addPartial(SkTArray<InEdge>& edges, int ptStart, int ptEnd,
int verbStart, int verbEnd) {
InEdge* edge = edges.push_back_n(1);
int verbCount = verbEnd - verbStart;
edge->fIntercepts.push_back_n(verbCount);
// uint8_t* verbs = &fVerbs[verbStart];
for (int ceptIdx = 0; ceptIdx < verbCount; ++ceptIdx) {
edge->fIntercepts[ceptIdx] = fIntercepts[verbStart + ceptIdx];
}
edge->fPts.append(ptEnd - ptStart, &fPts[ptStart]);
edge->fVerbs.append(verbCount, &fVerbs[verbStart]);
edge->setBounds();
edge->fWinding = fWinding;
edge->fContainsIntercepts = fContainsIntercepts; // FIXME: may not be correct -- but do we need to know?
}
void addSplit(SkTArray<InEdge>& edges, SkPoint* pts, uint8_t verb,
Intercepts& intercepts, int firstT, int lastT, bool flipped) {
InEdge* edge = edges.push_back_n(1);
edge->fIntercepts.push_back_n(1);
if (firstT == 0) {
*edge->fIntercepts[0].fTs.append() = 0;
} else {
*edge->fIntercepts[0].fTs.append() = intercepts.fTs[firstT - 1];
}
bool add1 = lastT == intercepts.fTs.count();
edge->fIntercepts[0].fTs.append(lastT - firstT, &intercepts.fTs[firstT]);
if (add1) {
*edge->fIntercepts[0].fTs.append() = 1;
}
edge->fIntercepts[0].fExplicit = true;
edge->fPts.append(verb + 1, pts);
edge->fVerbs.append(1, &verb);
// FIXME: bounds could be better for partial Ts
edge->setSubBounds();
edge->fContainsIntercepts = fContainsIntercepts; // FIXME: may not be correct -- but do we need to know?
if (flipped) {
edge->flipTs();
edge->fWinding = -fWinding;
} else {
edge->fWinding = fWinding;
}
}
bool cached(const InEdge* edge) {
// FIXME: in the pathological case where there is a ton of edges, binary search?
size_t count = fCached.count();
for (size_t index = 0; index < count; ++index) {
if (edge == fCached[index]) {
return true;
}
if (edge < fCached[index]) {
*fCached.insert(index) = edge;
return false;
}
}
*fCached.append() = edge;
return false;
}
void complete(signed char winding) {
setBounds();
fIntercepts.push_back_n(fVerbs.count());
if ((fWinding = winding) < 0) { // reverse verbs, pts, if bottom to top
flip();
}
fContainsIntercepts = fIntersected = false;
}
void flip() {
size_t index;
size_t last = fPts.count() - 1;
for (index = 0; index < last; ++index, --last) {
SkTSwap<SkPoint>(fPts[index], fPts[last]);
}
last = fVerbs.count() - 1;
for (index = 0; index < last; ++index, --last) {
SkTSwap<uint8_t>(fVerbs[index], fVerbs[last]);
}
}
void flipTs() {
SkASSERT(fIntercepts.count() == 1);
Intercepts& intercepts = fIntercepts[0];
SkASSERT(intercepts.fExplicit);
SkTDArray<double>& ts = intercepts.fTs;
size_t index;
size_t last = ts.count() - 1;
for (index = 0; index < last; ++index, --last) {
SkTSwap<double>(ts[index], ts[last]);
}
}
void reset() {
fCached.reset();
fIntercepts.reset();
fPts.reset();
fVerbs.reset();
fBounds.set(SK_ScalarMax, SK_ScalarMax, SK_ScalarMax, SK_ScalarMax);
fWinding = 0;
fContainsIntercepts = false;
fIntersected = false;
}
void setBounds() {
SkPoint* ptPtr = fPts.begin();
SkPoint* ptLast = fPts.end();
if (ptPtr == ptLast) {
SkDebugf("%s empty edge\n", __FUNCTION__);
SkASSERT(0);
// FIXME: delete empty edge?
return;
}
fBounds.set(ptPtr->fX, ptPtr->fY, ptPtr->fX, ptPtr->fY);
++ptPtr;
while (ptPtr != ptLast) {
fBounds.growToInclude(ptPtr->fX, ptPtr->fY);
++ptPtr;
}
}
// recompute bounds based on subrange of T values
void setSubBounds() {
SkASSERT(fIntercepts.count() == 1);
Intercepts& intercepts = fIntercepts[0];
SkASSERT(intercepts.fExplicit);
SkASSERT(fVerbs.count() == 1);
SkTDArray<double>& ts = intercepts.fTs;
if (fVerbs[0] == SkPath::kQuad_Verb) {
SkASSERT(fPts.count() == 3);
QuadSubBounds(fPts.begin(), ts[0], ts[ts.count() - 1], fBounds);
} else {
SkASSERT(fVerbs[0] == SkPath::kCubic_Verb);
SkASSERT(fPts.count() == 4);
CubicSubBounds(fPts.begin(), ts[0], ts[ts.count() - 1], fBounds);
}
}
void splitInflectionPts(SkTArray<InEdge>& edges) {
if (!fIntersected) {
return;
}
uint8_t* verbs = fVerbs.begin();
SkPoint* pts = fPts.begin();
int lastVerb = 0;
int lastPt = 0;
uint8_t verb;
bool edgeSplit = false;
for (int ceptIdx = 0; ceptIdx < fIntercepts.count(); ++ceptIdx, pts += verb) {
Intercepts& intercepts = fIntercepts[ceptIdx];
verb = *verbs++;
if (verb <= SkPath::kLine_Verb) {
continue;
}
size_t tCount = intercepts.fTs.count();
if (!tCount) {
continue;
}
size_t tIndex = (size_t) -1;
SkScalar y = pts[0].fY;
int lastSplit = 0;
int firstSplit = -1;
bool curveSplit = false;
while (++tIndex < tCount) {
double nextT = intercepts.fTs[tIndex];
SkScalar nextY = verb == SkPath::kQuad_Verb
? QuadYAtT(pts, nextT) : CubicYAtT(pts, nextT);
if (nextY < y) {
edgeSplit = curveSplit = true;
if (firstSplit < 0) {
firstSplit = tIndex;
int nextPt = pts - fPts.begin();
int nextVerb = verbs - 1 - fVerbs.begin();
if (lastVerb < nextVerb) {
addPartial(edges, lastPt, nextPt, lastVerb, nextVerb);
#if DEBUG_SPLIT
SkDebugf("%s addPartial 1\n", __FUNCTION__);
#endif
}
lastPt = nextPt;
lastVerb = nextVerb;
}
} else {
if (firstSplit >= 0) {
if (lastSplit < firstSplit) {
addSplit(edges, pts, verb, intercepts,
lastSplit, firstSplit, false);
#if DEBUG_SPLIT
SkDebugf("%s addSplit 1 tIndex=%d,%d\n",
__FUNCTION__, lastSplit, firstSplit);
#endif
}
addSplit(edges, pts, verb, intercepts,
firstSplit, tIndex, true);
#if DEBUG_SPLIT
SkDebugf("%s addSplit 2 tIndex=%d,%d flip\n",
__FUNCTION__, firstSplit, tIndex);
#endif
lastSplit = tIndex;
firstSplit = -1;
}
}
y = nextY;
}
if (curveSplit) {
if (firstSplit < 0) {
firstSplit = lastSplit;
} else {
addSplit(edges, pts, verb, intercepts, lastSplit,
firstSplit, false);
#if DEBUG_SPLIT
SkDebugf("%s addSplit 3 tIndex=%d,%d\n", __FUNCTION__,
lastSplit, firstSplit);
#endif
}
addSplit(edges, pts, verb, intercepts, firstSplit,
tIndex, pts[verb].fY < y);
#if DEBUG_SPLIT
SkDebugf("%s addSplit 4 tIndex=%d,%d %s\n", __FUNCTION__,
firstSplit, tIndex, pts[verb].fY < y ? "flip" : "");
#endif
}
}
// collapse remainder -- if there's nothing left, clear it somehow?
if (edgeSplit) {
int nextVerb = verbs - 1 - fVerbs.begin();
if (lastVerb < nextVerb) {
int nextPt = pts - fPts.begin();
addPartial(edges, lastPt, nextPt, lastVerb, nextVerb);
#if DEBUG_SPLIT
SkDebugf("%s addPartial 2\n", __FUNCTION__);
#endif
}
// OPTIMIZATION: reuse the edge instead of marking it empty
reset();
}
}
#if DEBUG_DUMP
void dump() {
int i;
const char className[] = "InEdge";
const int tab = 4;
SkDebugf("InEdge %p (edge=%d)\n", this, fID);
for (i = 0; i < fCached.count(); ++i) {
SkDebugf("%*s.fCached[%d]=0x%08x\n", tab + sizeof(className),
className, i, fCached[i]);
}
uint8_t* verbs = fVerbs.begin();
SkPoint* pts = fPts.begin();
for (i = 0; i < fIntercepts.count(); ++i) {
SkDebugf("%*s.fIntercepts[%d]:\n", tab + sizeof(className),
className, i);
fIntercepts[i].dump(pts, (SkPath::Verb) *verbs);
pts += *verbs++;
}
for (i = 0; i < fPts.count(); ++i) {
SkDebugf("%*s.fPts[%d]=(%1.9g,%1.9g)\n", tab + sizeof(className),
className, i, fPts[i].fX, fPts[i].fY);
}
for (i = 0; i < fVerbs.count(); ++i) {
SkDebugf("%*s.fVerbs[%d]=%d\n", tab + sizeof(className),
className, i, fVerbs[i]);
}
SkDebugf("%*s.fBounds=(%1.9g, %1.9g, %1.9g, %1.9g)\n", tab + sizeof(className),
className, fBounds.fLeft, fBounds.fTop,
fBounds.fRight, fBounds.fBottom);
SkDebugf("%*s.fWinding=%d\n", tab + sizeof(className), className,
fWinding);
SkDebugf("%*s.fContainsIntercepts=%d\n", tab + sizeof(className),
className, fContainsIntercepts);
SkDebugf("%*s.fIntersected=%d\n", tab + sizeof(className),
className, fIntersected);
}
#endif
// FIXME: temporary data : move this to a separate struct?
SkTDArray<const InEdge*> fCached; // list of edges already intercepted
SkTArray<Intercepts> fIntercepts; // one per verb
// persistent data
SkTDArray<SkPoint> fPts;
SkTDArray<uint8_t> fVerbs;
Bounds fBounds;
int fID;
signed char fWinding;
bool fContainsIntercepts;
bool fIntersected;
};
class InEdgeBuilder {
public:
InEdgeBuilder(const SkPath& path, bool ignoreHorizontal, SkTArray<InEdge>& edges,
SkTDArray<HorizontalEdge>& horizontalEdges)
: fPath(path)
, fCurrentEdge(NULL)
, fEdges(edges)
, fHorizontalEdges(horizontalEdges)
, fIgnoreHorizontal(ignoreHorizontal)
, fContainsCurves(false)
{
walk();
}
bool containsCurves() const {
return fContainsCurves;
}
protected:
void addEdge() {
SkASSERT(fCurrentEdge);
fCurrentEdge->fPts.append(fPtCount - fPtOffset, &fPts[fPtOffset]);
fPtOffset = 1;
*fCurrentEdge->fVerbs.append() = fVerb;
}
bool complete() {
if (fCurrentEdge && fCurrentEdge->fVerbs.count()) {
fCurrentEdge->complete(fWinding);
fCurrentEdge = NULL;
return true;
}
return false;
}
int direction(SkPath::Verb verb) {
fPtCount = verb + 1;
if (fIgnoreHorizontal && isHorizontal()) {
return 0;
}
return fPts[0].fY == fPts[verb].fY
? fPts[0].fX == fPts[verb].fX ? 0 : fPts[0].fX < fPts[verb].fX
? 1 : -1 : fPts[0].fY < fPts[verb].fY ? 1 : -1;
}
bool isHorizontal() {
SkScalar y = fPts[0].fY;
for (int i = 1; i < fPtCount; ++i) {
if (fPts[i].fY != y) {
return false;
}
}
return true;
}
void startEdge() {
if (!fCurrentEdge) {
fCurrentEdge = fEdges.push_back_n(1);
}
fWinding = 0;
fPtOffset = 0;
}
void walk() {
SkPath::Iter iter(fPath, true);
int winding = 0;
while ((fVerb = iter.next(fPts)) != SkPath::kDone_Verb) {
switch (fVerb) {
case SkPath::kMove_Verb:
startEdge();
continue;
case SkPath::kLine_Verb:
winding = direction(SkPath::kLine_Verb);
break;
case SkPath::kQuad_Verb:
fVerb = QuadReduceOrder(fPts);
winding = direction(fVerb);
fContainsCurves |= fVerb == SkPath::kQuad_Verb;
break;
case SkPath::kCubic_Verb:
fVerb = CubicReduceOrder(fPts);
winding = direction(fVerb);
fContainsCurves |= fVerb >= SkPath::kQuad_Verb;
break;
case SkPath::kClose_Verb:
SkASSERT(fCurrentEdge);
complete();
continue;
default:
SkDEBUGFAIL("bad verb");
return;
}
if (winding == 0) {
HorizontalEdge* horizontalEdge = fHorizontalEdges.append();
// FIXME: for degenerate quads and cubics, compute x extremes
horizontalEdge->fLeft = fPts[0].fX;
horizontalEdge->fRight = fPts[fVerb].fX;
horizontalEdge->fY = fPts[0].fY;
if (horizontalEdge->fLeft > horizontalEdge->fRight) {
SkTSwap<SkScalar>(horizontalEdge->fLeft, horizontalEdge->fRight);
}
if (complete()) {
startEdge();
}
continue;
}
if (fWinding + winding == 0) {
// FIXME: if prior verb or this verb is a horizontal line, reverse
// it instead of starting a new edge
SkASSERT(fCurrentEdge);
if (complete()) {
startEdge();
}
}
fWinding = winding;
addEdge();
}
if (!complete()) {
if (fCurrentEdge) {
fEdges.pop_back();
}
}
}
private:
const SkPath& fPath;
InEdge* fCurrentEdge;
SkTArray<InEdge>& fEdges;
SkTDArray<HorizontalEdge>& fHorizontalEdges;
SkPoint fPts[4];
SkPath::Verb fVerb;
int fPtCount;
int fPtOffset;
int8_t fWinding;
bool fIgnoreHorizontal;
bool fContainsCurves;
};
struct WorkEdge {
SkScalar bottom() const {
return fPts[verb()].fY;
}
void init(const InEdge* edge) {
fEdge = edge;
fPts = edge->fPts.begin();
fVerb = edge->fVerbs.begin();
}
bool advance() {
SkASSERT(fVerb < fEdge->fVerbs.end());
fPts += *fVerb++;
return fVerb != fEdge->fVerbs.end();
}
const SkPoint* lastPoints() const {
SkASSERT(fPts >= fEdge->fPts.begin() + lastVerb());
return &fPts[-lastVerb()];
}
SkPath::Verb lastVerb() const {
SkASSERT(fVerb > fEdge->fVerbs.begin());
return (SkPath::Verb) fVerb[-1];
}
const SkPoint* points() const {
return fPts;
}
SkPath::Verb verb() const {
return (SkPath::Verb) *fVerb;
}
ptrdiff_t verbIndex() const {
return fVerb - fEdge->fVerbs.begin();
}
int winding() const {
return fEdge->fWinding;
}
const InEdge* fEdge;
const SkPoint* fPts;
const uint8_t* fVerb;
};
// always constructed with SkTDArray because new edges are inserted
// this may be a inappropriate optimization, suggesting that a separate array of
// ActiveEdge* may be faster to insert and search
// OPTIMIZATION: Brian suggests that global sorting should be unnecessary, since
// as active edges are introduced, only local sorting should be required
class ActiveEdge {
public:
// this logic must be kept in sync with tooCloseToCall
// callers expect this to only read fAbove, fTangent
bool operator<(const ActiveEdge& rh) const {
if (fVerb == rh.fVerb) {
// FIXME: don't know what to do if verb is quad, cubic
return abCompare(fAbove, fBelow, rh.fAbove, rh.fBelow);
}
// figure out which is quad, line
// if cached data says line did not intersect quad, use top/bottom
if (fVerb != SkPath::kLine_Verb ? noIntersect(rh) : rh.noIntersect(*this)) {
return abCompare(fAbove, fBelow, rh.fAbove, rh.fBelow);
}
// use whichever of top/tangent tangent/bottom overlaps more
// with line top/bot
// assumes quad/cubic can already be upconverted to cubic/cubic
const SkPoint* line[2];
const SkPoint* curve[4];
if (fVerb != SkPath::kLine_Verb) {
line[0] = &rh.fAbove;
line[1] = &rh.fBelow;
curve[0] = &fAbove;
curve[1] = &fTangent;
curve[2] = &fBelow;
} else {
line[0] = &fAbove;
line[1] = &fBelow;
curve[0] = &rh.fAbove;
curve[1] = &rh.fTangent;
curve[2] = &rh.fBelow;
}
// FIXME: code has been abandoned, incomplete....
return false;
}
bool abCompare(const SkPoint& a1, const SkPoint& a2, const SkPoint& b1,
const SkPoint& b2) const {
double topD = a1.fX - b1.fX;
if (b1.fY < a1.fY) {
topD = (b2.fY - b1.fY) * topD - (a1.fY - b1.fY) * (b2.fX - b1.fX);
} else if (b1.fY > a1.fY) {
topD = (a2.fY - a1.fY) * topD + (b1.fY - a1.fY) * (a2.fX - a1.fX);
}
double botD = a2.fX - b2.fX;
if (b2.fY > a2.fY) {
botD = (b2.fY - b1.fY) * botD - (a2.fY - b2.fY) * (b2.fX - b1.fX);
} else if (b2.fY < a2.fY) {
botD = (a2.fY - a1.fY) * botD + (b2.fY - a2.fY) * (a2.fX - a1.fX);
}
// return sign of greater absolute value
return (fabs(topD) > fabs(botD) ? topD : botD) < 0;
}
// If a pair of edges are nearly coincident for some span, add a T in the
// edge so it can be shortened to match the other edge. Note that another
// approach is to trim the edge after it is added to the OutBuilder list --
// FIXME: since this has no effect if the edge is already done (i.e.,
// fYBottom >= y) maybe this can only be done by calling trimLine later.
void addTatYBelow(SkScalar y) {
if (fBelow.fY <= y || fYBottom >= y) {
return;
}
addTatYInner(y);
fFixBelow = true;
}
void addTatYAbove(SkScalar y) {
if (fBelow.fY <= y) {
return;
}
addTatYInner(y);
}
void addTatYInner(SkScalar y) {
if (fWorkEdge.fPts[0].fY > y) {
backup(y);
}
SkScalar left = fWorkEdge.fPts[0].fX;
SkScalar right = fWorkEdge.fPts[1].fX;
if (left > right) {
SkTSwap(left, right);
}
double ts[2];
SkASSERT(fWorkEdge.fVerb[0] == SkPath::kLine_Verb);
int pts = LineIntersect(fWorkEdge.fPts, left, right, y, ts);
SkASSERT(pts == 1);
// An ActiveEdge or WorkEdge has no need to modify the T values computed
// in the InEdge, except in the following case. If a pair of edges are
// nearly coincident, this may not be detected when the edges are
// intersected. Later, when sorted, and this near-coincidence is found,
// an additional t value must be added, requiring the cast below.
InEdge* writable = const_cast<InEdge*>(fWorkEdge.fEdge);
int insertedAt = writable->add(ts, pts, fWorkEdge.verbIndex());
#if DEBUG_ADJUST_COINCIDENT
SkDebugf("%s edge=%d y=%1.9g t=%1.9g\n", __FUNCTION__, ID(), y, ts[0]);
#endif
if (insertedAt >= 0) {
if (insertedAt + 1 < fTIndex) {
SkASSERT(insertedAt + 2 == fTIndex);
--fTIndex;
}
}
}
bool advanceT() {
SkASSERT(fTIndex <= fTs->count() - fExplicitTs);
return ++fTIndex <= fTs->count() - fExplicitTs;
}
bool advance() {
// FIXME: flip sense of next
bool result = fWorkEdge.advance();
fDone = !result;
initT();
return result;
}
void backup(SkScalar y) {
do {
SkASSERT(fWorkEdge.fEdge->fVerbs.begin() < fWorkEdge.fVerb);
fWorkEdge.fPts -= *--fWorkEdge.fVerb;
SkASSERT(fWorkEdge.fEdge->fPts.begin() <= fWorkEdge.fPts);
} while (fWorkEdge.fPts[0].fY >= y);
initT();
SkASSERT(!fExplicitTs);
fTIndex = fTs->count() + 1;
}
void calcAboveBelow(double tAbove, double tBelow) {
fVerb = fWorkEdge.verb();
switch (fVerb) {
case SkPath::kLine_Verb:
LineXYAtT(fWorkEdge.fPts, tAbove, &fAbove);
LineXYAtT(fWorkEdge.fPts, tBelow, &fTangent);
fBelow = fTangent;
break;
case SkPath::kQuad_Verb:
// FIXME: put array in struct to avoid copy?
SkPoint quad[3];
QuadSubDivide(fWorkEdge.fPts, tAbove, tBelow, quad);
fAbove = quad[0];
fTangent = quad[0] != quad[1] ? quad[1] : quad[2];
fBelow = quad[2];
break;
case SkPath::kCubic_Verb:
SkPoint cubic[3];
CubicSubDivide(fWorkEdge.fPts, tAbove, tBelow, cubic);
fAbove = cubic[0];
// FIXME: can't see how quad logic for how tangent is used
// extends to cubic
fTangent = cubic[0] != cubic[1] ? cubic[1]
: cubic[0] != cubic[2] ? cubic[2] : cubic[3];
fBelow = cubic[3];
break;
default:
SkASSERT(0);
}
}
void calcLeft(SkScalar y) {
// OPTIMIZE: put a kDone_Verb at the end of the verb list?
if (fDone || fBelow.fY > y) {
return; // nothing to do; use last
}
calcLeft();
if (fAbove.fY == fBelow.fY) {
SkDebugf("%s edge=%d fAbove.fY != fBelow.fY %1.9g\n", __FUNCTION__,
ID(), fAbove.fY);
}
}
void calcLeft() {
int add = (fTIndex <= fTs->count() - fExplicitTs) - 1;
double tAbove = t(fTIndex + add);
double tBelow = t(fTIndex - ~add);
// OPTIMIZATION: if fAbove, fBelow have already been computed
// for the fTIndex, don't do it again
calcAboveBelow(tAbove, tBelow);
// For identical x, this lets us know which edge is first.
// If both edges have T values < 1, check x at next T (fBelow).
SkASSERT(tAbove != tBelow);
// FIXME: this can loop forever
// need a break if we hit the end
// FIXME: in unit test, figure out how explicit Ts work as well
while (fAbove.fY == fBelow.fY) {
if (add < 0 || fTIndex == fTs->count()) {
add -= 1;
SkASSERT(fTIndex + add >= 0);
tAbove = t(fTIndex + add);
} else {
add += 1;
SkASSERT(fTIndex - ~add <= fTs->count() + 1);
tBelow = t(fTIndex - ~add);
}
calcAboveBelow(tAbove, tBelow);
}
fTAbove = tAbove;
fTBelow = tBelow;
}
bool done(SkScalar bottom) const {
return fDone || fYBottom >= bottom;
}
void fixBelow() {
if (fFixBelow) {
fTBelow = nextT();
calcAboveBelow(fTAbove, fTBelow);
fFixBelow = false;
}
}
void init(const InEdge* edge) {
fWorkEdge.init(edge);
fDone = false;
initT();
fBelow.fY = SK_ScalarMin;
fYBottom = SK_ScalarMin;
}
void initT() {
const Intercepts& intercepts = fWorkEdge.fEdge->fIntercepts.front();
SkASSERT(fWorkEdge.verbIndex() <= fWorkEdge.fEdge->fIntercepts.count());
const Intercepts* interceptPtr = &intercepts + fWorkEdge.verbIndex();
fTs = &interceptPtr->fTs;
fExplicitTs = interceptPtr->fExplicit;
// the above is conceptually the same as
// fTs = &fWorkEdge.fEdge->fIntercepts[fWorkEdge.verbIndex()].fTs;
// but templated arrays don't allow returning a pointer to the end() element
fTIndex = 0;
if (!fDone) {
fVerb = fWorkEdge.verb();
}
SkASSERT(fVerb > SkPath::kMove_Verb);
}
// OPTIMIZATION: record if two edges are coincident when the are intersected
// It's unclear how to do this -- seems more complicated than recording the
// t values, since the same t values could exist intersecting non-coincident
// edges.
bool isCoincidentWith(const ActiveEdge* edge) const {
if (fAbove != edge->fAbove || fBelow != edge->fBelow) {
return false;
}
if (fVerb != edge->fVerb) {
return false;
}
switch (fVerb) {
case SkPath::kLine_Verb:
return true;
default:
// FIXME: add support for quads, cubics
SkASSERT(0);
return false;
}
return false;
}
bool isUnordered(const ActiveEdge* edge) const {
return fAbove == edge->fAbove && fBelow == edge->fBelow;
}
// SkPath::Verb lastVerb() const {
// return fDone ? fWorkEdge.lastVerb() : fWorkEdge.verb();
// }
const SkPoint* lastPoints() const {
return fDone ? fWorkEdge.lastPoints() : fWorkEdge.points();
}
bool noIntersect(const ActiveEdge& ) const {
// incomplete
return false;
}
// The shortest close call edge should be moved into a position where
// it contributes if the winding is transitioning to or from zero.
bool swapClose(const ActiveEdge* next, int prev, int wind, int mask) const {
#if DEBUG_ADJUST_COINCIDENT
SkDebugf("%s edge=%d (%g) next=%d (%g) prev=%d wind=%d nextWind=%d\n",
__FUNCTION__, ID(), fBelow.fY, next->ID(), next->fBelow.fY,
prev, wind, wind + next->fWorkEdge.winding());
#endif
if ((prev & mask) == 0 || (wind & mask) == 0) {
return next->fBelow.fY < fBelow.fY;
}
int nextWinding = wind + next->fWorkEdge.winding();
if ((nextWinding & mask) == 0) {
return fBelow.fY < next->fBelow.fY;
}
return false;
}
bool swapCoincident(const ActiveEdge* edge, SkScalar bottom) const {
if (fBelow.fY >= bottom || fDone || edge->fDone) {
return false;
}
ActiveEdge thisWork = *this;
ActiveEdge edgeWork = *edge;
while ((thisWork.advanceT() || thisWork.advance())
&& (edgeWork.advanceT() || edgeWork.advance())) {
thisWork.calcLeft();
edgeWork.calcLeft();
if (thisWork < edgeWork) {
return false;
}
if (edgeWork < thisWork) {
return true;
}
}
return false;
}
bool swapUnordered(const ActiveEdge* edge, SkScalar /* bottom */) const {
SkASSERT(fVerb != SkPath::kLine_Verb
|| edge->fVerb != SkPath::kLine_Verb);
if (fDone || edge->fDone) {
return false;
}
ActiveEdge thisWork, edgeWork;
extractAboveBelow(thisWork);
edge->extractAboveBelow(edgeWork);
return edgeWork < thisWork;
}
bool tooCloseToCall(const ActiveEdge* edge) const {
int ulps;
double t1, t2, b1, b2;
// This logic must be kept in sync with operator <
if (edge->fAbove.fY < fAbove.fY) {
t1 = (edge->fTangent.fY - edge->fAbove.fY) * (fAbove.fX - edge->fAbove.fX);
t2 = (fAbove.fY - edge->fAbove.fY) * (edge->fTangent.fX - edge->fAbove.fX);
} else if (edge->fAbove.fY > fAbove.fY) {
t1 = (fTangent.fY - fAbove.fY) * (fAbove.fX - edge->fAbove.fX);
t2 = (fAbove.fY - edge->fAbove.fY) * (fTangent.fX - fAbove.fX);
} else {
t1 = fAbove.fX;
t2 = edge->fAbove.fX;
}
if (edge->fTangent.fY > fTangent.fY) {
b1 = (edge->fTangent.fY - edge->fAbove.fY) * (fTangent.fX - edge->fTangent.fX);
b2 = (fTangent.fY - edge->fTangent.fY) * (edge->fTangent.fX - edge->fAbove.fX);
} else if (edge->fTangent.fY < fTangent.fY) {
b1 = (fTangent.fY - fAbove.fY) * (fTangent.fX - edge->fTangent.fX);
b2 = (fTangent.fY - edge->fTangent.fY) * (fTangent.fX - fAbove.fX);
} else {
b1 = fTangent.fX;
b2 = edge->fTangent.fX;
}
if (fabs(t1 - t2) > fabs(b1 - b2)) {
ulps = UlpsDiff((float) t1, (float) t2);
} else {
ulps = UlpsDiff((float) b1, (float) b2);
}
#if DEBUG_ADJUST_COINCIDENT
SkDebugf("%s this=%d edge=%d ulps=%d\n", __FUNCTION__, ID(), edge->ID(),
ulps);
#endif
if (ulps < 0 || ulps > 32) {
return false;
}
if (fVerb == SkPath::kLine_Verb && edge->fVerb == SkPath::kLine_Verb) {
return true;
}
if (fVerb != SkPath::kLine_Verb && edge->fVerb != SkPath::kLine_Verb) {
return false;
}
double ts[2];
bool isLine = true;
bool curveQuad = true;
if (fVerb == SkPath::kCubic_Verb) {
ts[0] = (fTAbove * 2 + fTBelow) / 3;
ts[1] = (fTAbove + fTBelow * 2) / 3;
curveQuad = isLine = false;
} else if (edge->fVerb == SkPath::kCubic_Verb) {
ts[0] = (edge->fTAbove * 2 + edge->fTBelow) / 3;
ts[1] = (edge->fTAbove + edge->fTBelow * 2) / 3;
curveQuad = false;
} else if (fVerb == SkPath::kQuad_Verb) {
ts[0] = fTAbove;
ts[1] = (fTAbove + fTBelow) / 2;
isLine = false;
} else {
SkASSERT(edge->fVerb == SkPath::kQuad_Verb);
ts[0] = edge->fTAbove;
ts[1] = (edge->fTAbove + edge->fTBelow) / 2;
}
const SkPoint* curvePts = isLine ? edge->lastPoints() : lastPoints();
const ActiveEdge* lineEdge = isLine ? this : edge;
SkPoint curveSample[2];
for (int index = 0; index < 2; ++index) {
if (curveQuad) {
QuadXYAtT(curvePts, ts[index], &curveSample[index]);
} else {
CubicXYAtT(curvePts, ts[index], &curveSample[index]);
}
}
return IsCoincident(curveSample, lineEdge->fAbove, lineEdge->fBelow);
}
double nextT() const {
SkASSERT(fTIndex <= fTs->count() - fExplicitTs);
return t(fTIndex + 1);
}
double t() const {
return t(fTIndex);
}
double t(int tIndex) const {
if (fExplicitTs) {
SkASSERT(tIndex < fTs->count());
return (*fTs)[tIndex];
}
if (tIndex == 0) {
return 0;
}
if (tIndex > fTs->count()) {
return 1;
}
return (*fTs)[tIndex - 1];
}
// FIXME: debugging only
int ID() const {
return fWorkEdge.fEdge->fID;
}
private:
// utility used only by swapUnordered
void extractAboveBelow(ActiveEdge& extracted) const {
SkPoint curve[4];
switch (fVerb) {
case SkPath::kLine_Verb:
extracted.fAbove = fAbove;
extracted.fTangent = fTangent;
return;
case SkPath::kQuad_Verb:
QuadSubDivide(lastPoints(), fTAbove, fTBelow, curve);
break;
case SkPath::kCubic_Verb:
CubicSubDivide(lastPoints(), fTAbove, fTBelow, curve);
break;
default:
SkASSERT(0);
}
extracted.fAbove = curve[0];
extracted.fTangent = curve[1];
}
public:
WorkEdge fWorkEdge;
const SkTDArray<double>* fTs;
SkPoint fAbove;
SkPoint fTangent;
SkPoint fBelow;
double fTAbove; // OPTIMIZATION: only required if edge has quads or cubics
double fTBelow;
SkScalar fYBottom;
int fCoincident;
int fTIndex;
SkPath::Verb fVerb;
bool fSkip; // OPTIMIZATION: use bitfields?
bool fCloseCall;
bool fDone;
bool fFixBelow;
bool fExplicitTs;
};
static void addToActive(SkTDArray<ActiveEdge>& activeEdges, const InEdge* edge) {
size_t count = activeEdges.count();
for (size_t index = 0; index < count; ++index) {
if (edge == activeEdges[index].fWorkEdge.fEdge) {
return;
}
}
ActiveEdge* active = activeEdges.append();
active->init(edge);
}
// Find any intersections in the range of active edges. A pair of edges, on
// either side of another edge, may change the winding contribution for part of
// the edge.
// Keep horizontal edges just for
// the purpose of computing when edges change their winding contribution, since
// this is essentially computing the horizontal intersection.
static void addBottomT(InEdge** currentPtr, InEdge** lastPtr,
HorizontalEdge** horizontal) {
InEdge** testPtr = currentPtr - 1;
HorizontalEdge* horzEdge = *horizontal;
SkScalar left = horzEdge->fLeft;
SkScalar bottom = horzEdge->fY;
while (++testPtr != lastPtr) {
InEdge* test = *testPtr;
if (test->fBounds.fBottom <= bottom || test->fBounds.fRight <= left) {
continue;
}
WorkEdge wt;
wt.init(test);
do {
HorizontalEdge** sorted = horizontal;
horzEdge = *sorted;
do {
double wtTs[4];
int pts;
uint8_t verb = wt.verb();
switch (verb) {
case SkPath::kLine_Verb:
pts = LineIntersect(wt.fPts, horzEdge->fLeft,
horzEdge->fRight, horzEdge->fY, wtTs);
break;
case SkPath::kQuad_Verb:
pts = QuadIntersect(wt.fPts, horzEdge->fLeft,
horzEdge->fRight, horzEdge->fY, wtTs);
break;
case SkPath::kCubic_Verb:
pts = CubicIntersect(wt.fPts, horzEdge->fLeft,
horzEdge->fRight, horzEdge->fY, wtTs);
break;
}
if (pts) {
#if DEBUG_ADD_BOTTOM_TS
for (int x = 0; x < pts; ++x) {
SkDebugf("%s y=%g wtTs[0]=%g (%g,%g", __FUNCTION__,
horzEdge->fY, wtTs[x], wt.fPts[0].fX, wt.fPts[0].fY);
for (int y = 0; y < verb; ++y) {
SkDebugf(" %g,%g", wt.fPts[y + 1].fX, wt.fPts[y + 1].fY));
}
SkDebugf(")\n");
}
if (pts > verb) {
SkASSERT(0); // FIXME ? should this work?
SkDebugf("%s wtTs[1]=%g\n", __FUNCTION__, wtTs[1]);
}
#endif
test->add(wtTs, pts, wt.verbIndex());
}
horzEdge = *++sorted;
} while (horzEdge->fY == bottom
&& horzEdge->fLeft <= test->fBounds.fRight);
} while (wt.advance());
}
}
#if DEBUG_ADD_INTERSECTING_TS
static void debugShowLineIntersection(int pts, const WorkEdge& wt,
const WorkEdge& wn, const double wtTs[2], const double wnTs[2]) {
if (!pts) {
return;
}
SkPoint wtOutPt, wnOutPt;
LineXYAtT(wt.fPts, wtTs[0], &wtOutPt);
LineXYAtT(wn.fPts, wnTs[0], &wnOutPt);
SkDebugf("%s wtTs[0]=%g (%g,%g, %g,%g) (%g,%g)\n",
__FUNCTION__,
wtTs[0], wt.fPts[0].fX, wt.fPts[0].fY,
wt.fPts[1].fX, wt.fPts[1].fY, wtOutPt.fX, wtOutPt.fY);
if (pts == 2) {
SkDebugf("%s wtTs[1]=%g\n", __FUNCTION__, wtTs[1]);
}
SkDebugf("%s wnTs[0]=%g (%g,%g, %g,%g) (%g,%g)\n",
__FUNCTION__,
wnTs[0], wn.fPts[0].fX, wn.fPts[0].fY,
wn.fPts[1].fX, wn.fPts[1].fY, wnOutPt.fX, wnOutPt.fY);
if (pts == 2) {
SkDebugf("%s wnTs[1]=%g\n", __FUNCTION__, wnTs[1]);
}
}
#else
static void debugShowLineIntersection(int , const WorkEdge& ,
const WorkEdge& , const double [2], const double [2]) {
}
#endif
static void addIntersectingTs(InEdge** currentPtr, InEdge** lastPtr) {
InEdge** testPtr = currentPtr - 1;
// FIXME: lastPtr should be past the point of interest, so
// test below should be lastPtr - 2
// that breaks testSimplifyTriangle22, so further investigation is needed
while (++testPtr != lastPtr - 1) {
InEdge* test = *testPtr;
InEdge** nextPtr = testPtr;
do {
InEdge* next = *++nextPtr;
// FIXME: this compares against the sentinel sometimes
// OPTIMIZATION: this may never be needed since this gets called
// in two passes now. Verify that double hits are appropriate.
if (test->cached(next)) {
continue;
}
if (!Bounds::Intersects(test->fBounds, next->fBounds)) {
continue;
}
WorkEdge wt, wn;
wt.init(test);
wn.init(next);
do {
int pts;
Intersections ts;
bool swap = false;
switch (wt.verb()) {
case SkPath::kLine_Verb:
switch (wn.verb()) {
case SkPath::kLine_Verb: {
pts = LineIntersect(wt.fPts, wn.fPts, ts);
debugShowLineIntersection(pts, wt, wn,
ts.fT[0], ts.fT[1]);
break;
}
case SkPath::kQuad_Verb: {
swap = true;
pts = QuadLineIntersect(wn.fPts, wt.fPts, ts);
break;
}
case SkPath::kCubic_Verb: {
swap = true;
pts = CubicLineIntersect(wn.fPts, wt.fPts, ts);
break;
}
default:
SkASSERT(0);
}
break;
case SkPath::kQuad_Verb:
switch (wn.verb()) {
case SkPath::kLine_Verb: {
pts = QuadLineIntersect(wt.fPts, wn.fPts, ts);
break;
}
case SkPath::kQuad_Verb: {
pts = QuadIntersect(wt.fPts, wn.fPts, ts);
break;
}
case SkPath::kCubic_Verb: {
// FIXME: promote quad to cubic
pts = CubicIntersect(wt.fPts, wn.fPts, ts);
break;
}
default:
SkASSERT(0);
}
break;
case SkPath::kCubic_Verb:
switch (wn.verb()) {
case SkPath::kLine_Verb: {
pts = CubicLineIntersect(wt.fPts, wn.fPts, ts);
break;
}
case SkPath::kQuad_Verb: {
// FIXME: promote quad to cubic
pts = CubicIntersect(wt.fPts, wn.fPts, ts);
break;
}
case SkPath::kCubic_Verb: {
pts = CubicIntersect(wt.fPts, wn.fPts, ts);
break;
}
default:
SkASSERT(0);
}
break;
default:
SkASSERT(0);
}
test->add(ts.fT[swap], pts, wt.verbIndex());
next->add(ts.fT[!swap], pts, wn.verbIndex());
} while (wt.bottom() <= wn.bottom() ? wt.advance() : wn.advance());
} while (nextPtr != lastPtr);
}
}
static InEdge** advanceEdges(SkTDArray<ActiveEdge>* activeEdges,
InEdge** currentPtr, InEdge** lastPtr, SkScalar y) {
InEdge** testPtr = currentPtr - 1;
while (++testPtr != lastPtr) {
if ((*testPtr)->fBounds.fBottom > y) {
continue;
}
if (activeEdges) {
InEdge* test = *testPtr;
ActiveEdge* activePtr = activeEdges->begin() - 1;
ActiveEdge* lastActive = activeEdges->end();
while (++activePtr != lastActive) {
if (activePtr->fWorkEdge.fEdge == test) {
activeEdges->remove(activePtr - activeEdges->begin());
break;
}
}
}
if (testPtr == currentPtr) {
++currentPtr;
}
}
return currentPtr;
}
// OPTIMIZE: inline?
static HorizontalEdge** advanceHorizontal(HorizontalEdge** edge, SkScalar y) {
while ((*edge)->fY < y) {
++edge;
}
return edge;
}
// compute bottom taking into account any intersected edges
static SkScalar computeInterceptBottom(SkTDArray<ActiveEdge>& activeEdges,
SkScalar y, SkScalar bottom) {
ActiveEdge* activePtr = activeEdges.begin() - 1;
ActiveEdge* lastActive = activeEdges.end();
while (++activePtr != lastActive) {
const InEdge* test = activePtr->fWorkEdge.fEdge;
if (!test->fContainsIntercepts) {
continue;
}
WorkEdge wt;
wt.init(test);
do {
const Intercepts& intercepts = test->fIntercepts[wt.verbIndex()];
if (intercepts.fTopIntercepts > 1) {
SkScalar yTop = wt.fPts[0].fY;
if (yTop > y && bottom > yTop) {
bottom = yTop;
}
}
if (intercepts.fBottomIntercepts > 1) {
SkScalar yBottom = wt.fPts[wt.verb()].fY;
if (yBottom > y && bottom > yBottom) {
bottom = yBottom;
}
}
const SkTDArray<double>& fTs = intercepts.fTs;
size_t count = fTs.count();
for (size_t index = 0; index < count; ++index) {
SkScalar yIntercept;
switch (wt.verb()) {
case SkPath::kLine_Verb: {
yIntercept = LineYAtT(wt.fPts, fTs[index]);
break;
}
case SkPath::kQuad_Verb: {
yIntercept = QuadYAtT(wt.fPts, fTs[index]);
break;
}
case SkPath::kCubic_Verb: {
yIntercept = CubicYAtT(wt.fPts, fTs[index]);
break;
}
default:
SkASSERT(0); // should never get here
}
if (yIntercept > y && bottom > yIntercept) {
bottom = yIntercept;
}
}
} while (wt.advance());
}
#if DEBUG_BOTTOM
SkDebugf("%s bottom=%1.9g\n", __FUNCTION__, bottom);
#endif
return bottom;
}
static SkScalar findBottom(InEdge** currentPtr,
InEdge** edgeListEnd, SkTDArray<ActiveEdge>* activeEdges, SkScalar y,
bool /*asFill*/, InEdge**& testPtr) {
InEdge* current = *currentPtr;
SkScalar bottom = current->fBounds.fBottom;
// find the list of edges that cross y
InEdge* test = *testPtr;
while (testPtr != edgeListEnd) {
SkScalar testTop = test->fBounds.fTop;
if (bottom <= testTop) {
break;
}
SkScalar testBottom = test->fBounds.fBottom;
// OPTIMIZATION: Shortening the bottom is only interesting when filling
// and when the edge is to the left of a longer edge. If it's a framing
// edge, or part of the right, it won't effect the longer edges.
if (testTop > y) {
bottom = testTop;
break;
}
if (y < testBottom) {
if (bottom > testBottom) {
bottom = testBottom;
}
if (activeEdges) {
addToActive(*activeEdges, test);
}
}
test = *++testPtr;
}
#if DEBUG_BOTTOM
SkDebugf("%s %d bottom=%1.9g\n", __FUNCTION__, activeEdges ? 2 : 1, bottom);
#endif
return bottom;
}
static void makeEdgeList(SkTArray<InEdge>& edges, InEdge& edgeSentinel,
SkTDArray<InEdge*>& edgeList) {
size_t edgeCount = edges.count();
if (edgeCount == 0) {
return;
}
int id = 0;
for (size_t index = 0; index < edgeCount; ++index) {
InEdge& edge = edges[index];
if (!edge.fWinding) {
continue;
}
edge.fID = ++id;
*edgeList.append() = &edge;
}
*edgeList.append() = &edgeSentinel;
QSort<InEdge>(edgeList.begin(), edgeList.end() - 1);
}
static void makeHorizontalList(SkTDArray<HorizontalEdge>& edges,
HorizontalEdge& edgeSentinel, SkTDArray<HorizontalEdge*>& edgeList) {
size_t edgeCount = edges.count();
if (edgeCount == 0) {
return;
}
for (size_t index = 0; index < edgeCount; ++index) {
*edgeList.append() = &edges[index];
}
edgeSentinel.fLeft = edgeSentinel.fRight = edgeSentinel.fY = SK_ScalarMax;
*edgeList.append() = &edgeSentinel;
QSort<HorizontalEdge>(edgeList.begin(), edgeList.end() - 1);
}
static void skipCoincidence(int lastWinding, int winding, int windingMask,
ActiveEdge* activePtr, ActiveEdge* firstCoincident) {
if (((lastWinding & windingMask) == 0) ^ ((winding & windingMask) != 0)) {
return;
}
// FIXME: ? shouldn't this be if (lastWinding & windingMask) ?
if (lastWinding) {
#if DEBUG_ADJUST_COINCIDENT
SkDebugf("%s edge=%d 1 set skip=false\n", __FUNCTION__, activePtr->ID());
#endif
activePtr->fSkip = false;
} else {
#if DEBUG_ADJUST_COINCIDENT
SkDebugf("%s edge=%d 2 set skip=false\n", __FUNCTION__, firstCoincident->ID());
#endif
firstCoincident->fSkip = false;
}
}
static void sortHorizontal(SkTDArray<ActiveEdge>& activeEdges,
SkTDArray<ActiveEdge*>& edgeList, SkScalar y) {
size_t edgeCount = activeEdges.count();
if (edgeCount == 0) {
return;
}
#if DEBUG_SORT_HORIZONTAL
const int tab = 3; // FIXME: debugging only
SkDebugf("%s y=%1.9g\n", __FUNCTION__, y);
#endif
size_t index;
for (index = 0; index < edgeCount; ++index) {
ActiveEdge& activeEdge = activeEdges[index];
do {
activeEdge.calcLeft(y);
// skip segments that don't span y
if (activeEdge.fAbove != activeEdge.fBelow) {
break;
}
if (activeEdge.fDone) {
#if DEBUG_SORT_HORIZONTAL
SkDebugf("%*s edge=%d done\n", tab, "", activeEdge.ID());
#endif
goto nextEdge;
}
#if DEBUG_SORT_HORIZONTAL
SkDebugf("%*s edge=%d above==below\n", tab, "", activeEdge.ID());
#endif
} while (activeEdge.advanceT() || activeEdge.advance());
#if DEBUG_SORT_HORIZONTAL
SkDebugf("%*s edge=%d above=(%1.9g,%1.9g) (%1.9g) below=(%1.9g,%1.9g)"
" (%1.9g)\n", tab, "", activeEdge.ID(),
activeEdge.fAbove.fX, activeEdge.fAbove.fY, activeEdge.fTAbove,
activeEdge.fBelow.fX, activeEdge.fBelow.fY, activeEdge.fTBelow);
#endif
activeEdge.fSkip = activeEdge.fCloseCall = activeEdge.fFixBelow = false;
*edgeList.append() = &activeEdge;
nextEdge:
;
}
QSort<ActiveEdge>(edgeList.begin(), edgeList.end() - 1);
}
// remove coincident edges
// OPTIMIZE: remove edges? This is tricky because the current logic expects
// the winding count to be maintained while skipping coincident edges. In
// addition to removing the coincident edges, the remaining edges would need
// to have a different winding value, possibly different per intercept span.
static SkScalar adjustCoincident(SkTDArray<ActiveEdge*>& edgeList,
int windingMask, SkScalar y, SkScalar bottom, OutEdgeBuilder& outBuilder)
{
#if DEBUG_ADJUST_COINCIDENT
SkDebugf("%s y=%1.9g bottom=%1.9g\n", __FUNCTION__, y, bottom);
#endif
size_t edgeCount = edgeList.count();
if (edgeCount == 0) {
return bottom;
}
ActiveEdge* activePtr, * nextPtr = edgeList[0];
size_t index;
bool foundCoincident = false;
size_t firstIndex = 0;
for (index = 1; index < edgeCount; ++index) {
activePtr = nextPtr;
nextPtr = edgeList[index];
if (firstIndex != index - 1 && activePtr->fVerb > SkPath::kLine_Verb
&& nextPtr->fVerb == SkPath::kLine_Verb
&& activePtr->isUnordered(nextPtr)) {
// swap the line with the curve
// back up to the previous edge and retest
SkTSwap<ActiveEdge*>(edgeList[index - 1], edgeList[index]);
SkASSERT(index > 1);
index -= 2;
nextPtr = edgeList[index];
continue;
}
bool closeCall = false;
activePtr->fCoincident = firstIndex;
if (activePtr->isCoincidentWith(nextPtr)
|| (closeCall = activePtr->tooCloseToCall(nextPtr))) {
activePtr->fSkip = nextPtr->fSkip = foundCoincident = true;
activePtr->fCloseCall = nextPtr->fCloseCall = closeCall;
} else if (activePtr->isUnordered(nextPtr)) {
foundCoincident = true;
} else {
firstIndex = index;
}
}
nextPtr->fCoincident = firstIndex;
if (!foundCoincident) {
return bottom;
}
int winding = 0;
nextPtr = edgeList[0];
for (index = 1; index < edgeCount; ++index) {
int priorWinding = winding;
winding += activePtr->fWorkEdge.winding();
activePtr = nextPtr;
nextPtr = edgeList[index];
SkASSERT(activePtr == edgeList[index - 1]);
SkASSERT(nextPtr == edgeList[index]);
if (activePtr->fCoincident != nextPtr->fCoincident) {
continue;
}
// the coincident edges may not have been sorted above -- advance
// the edges and resort if needed
// OPTIMIZE: if sorting is done incrementally as new edges are added
// and not all at once as is done here, fold this test into the
// current less than test.
while ((!activePtr->fSkip || !nextPtr->fSkip)
&& activePtr->fCoincident == nextPtr->fCoincident) {
if (activePtr->swapUnordered(nextPtr, bottom)) {
winding -= activePtr->fWorkEdge.winding();
SkASSERT(activePtr == edgeList[index - 1]);
SkASSERT(nextPtr == edgeList[index]);
SkTSwap<ActiveEdge*>(edgeList[index - 1], edgeList[index]);
if (--index == 0) {
winding += activePtr->fWorkEdge.winding();
break;
}
// back up one
activePtr = edgeList[index - 1];
continue;
}
SkASSERT(activePtr == edgeList[index - 1]);
SkASSERT(nextPtr == edgeList[index]);
break;
}
if (activePtr->fSkip && nextPtr->fSkip) {
if (activePtr->fCloseCall ? activePtr->swapClose(nextPtr,
priorWinding, winding, windingMask)
: activePtr->swapCoincident(nextPtr, bottom)) {
winding -= activePtr->fWorkEdge.winding();
SkASSERT(activePtr == edgeList[index - 1]);
SkASSERT(nextPtr == edgeList[index]);
SkTSwap<ActiveEdge*>(edgeList[index - 1], edgeList[index]);
SkTSwap<ActiveEdge*>(activePtr, nextPtr);
winding += activePtr->fWorkEdge.winding();
SkASSERT(activePtr == edgeList[index - 1]);
SkASSERT(nextPtr == edgeList[index]);
}
}
}
int firstCoincidentWinding = 0;
ActiveEdge* firstCoincident = NULL;
winding = 0;
activePtr = edgeList[0];
for (index = 1; index < edgeCount; ++index) {
int priorWinding = winding;
winding += activePtr->fWorkEdge.winding();
nextPtr = edgeList[index];
if (activePtr->fSkip && nextPtr->fSkip
&& activePtr->fCoincident == nextPtr->fCoincident) {
if (!firstCoincident) {
firstCoincident = activePtr;
firstCoincidentWinding = priorWinding;
}
if (activePtr->fCloseCall) {
// If one of the edges has already been added to out as a non
// coincident edge, trim it back to the top of this span
if (outBuilder.trimLine(y, activePtr->ID())) {
activePtr->addTatYAbove(y);
#if DEBUG_ADJUST_COINCIDENT
SkDebugf("%s 1 edge=%d y=%1.9g (was fYBottom=%1.9g)\n",
__FUNCTION__, activePtr->ID(), y, activePtr->fYBottom);
#endif
activePtr->fYBottom = y;
}
if (outBuilder.trimLine(y, nextPtr->ID())) {
nextPtr->addTatYAbove(y);
#if DEBUG_ADJUST_COINCIDENT
SkDebugf("%s 2 edge=%d y=%1.9g (was fYBottom=%1.9g)\n",
__FUNCTION__, nextPtr->ID(), y, nextPtr->fYBottom);
#endif
nextPtr->fYBottom = y;
}
// add missing t values so edges can be the same length
SkScalar testY = activePtr->fBelow.fY;
nextPtr->addTatYBelow(testY);
if (bottom > testY && testY > y) {
#if DEBUG_ADJUST_COINCIDENT
SkDebugf("%s 3 edge=%d bottom=%1.9g (was bottom=%1.9g)\n",
__FUNCTION__, activePtr->ID(), testY, bottom);
#endif
bottom = testY;
}
testY = nextPtr->fBelow.fY;
activePtr->addTatYBelow(testY);
if (bottom > testY && testY > y) {
#if DEBUG_ADJUST_COINCIDENT
SkDebugf("%s 4 edge=%d bottom=%1.9g (was bottom=%1.9g)\n",
__FUNCTION__, nextPtr->ID(), testY, bottom);
#endif
bottom = testY;
}
}
} else if (firstCoincident) {
skipCoincidence(firstCoincidentWinding, winding, windingMask,
activePtr, firstCoincident);
firstCoincident = NULL;
}
activePtr = nextPtr;
}
if (firstCoincident) {
winding += activePtr->fWorkEdge.winding();
skipCoincidence(firstCoincidentWinding, winding, windingMask, activePtr,
firstCoincident);
}
// fix up the bottom for close call edges. OPTIMIZATION: maybe this could
// be in the loop above, but moved here since loop above reads fBelow and
// it felt unsafe to write it in that loop
for (index = 0; index < edgeCount; ++index) {
(edgeList[index])->fixBelow();
}
return bottom;
}
// stitch edge and t range that satisfies operation
static void stitchEdge(SkTDArray<ActiveEdge*>& edgeList, SkScalar
#if DEBUG_STITCH_EDGE
y
#endif
,
SkScalar bottom, int windingMask, bool fill, OutEdgeBuilder& outBuilder) {
int winding = 0;
ActiveEdge** activeHandle = edgeList.begin() - 1;
ActiveEdge** lastActive = edgeList.end();
#if DEBUG_STITCH_EDGE
const int tab = 7; // FIXME: debugging only
SkDebugf("%s y=%1.9g bottom=%1.9g\n", __FUNCTION__, y, bottom);
#endif
while (++activeHandle != lastActive) {
ActiveEdge* activePtr = *activeHandle;
const WorkEdge& wt = activePtr->fWorkEdge;
int lastWinding = winding;
winding += wt.winding();
#if DEBUG_STITCH_EDGE
SkDebugf("%*s edge=%d lastWinding=%d winding=%d skip=%d close=%d"
" above=%1.9g below=%1.9g\n",
tab-4, "", activePtr->ID(), lastWinding,
winding, activePtr->fSkip, activePtr->fCloseCall,
activePtr->fTAbove, activePtr->fTBelow);
#endif
if (activePtr->done(bottom)) {
#if DEBUG_STITCH_EDGE
SkDebugf("%*s fDone=%d || fYBottom=%1.9g >= bottom\n", tab, "",
activePtr->fDone, activePtr->fYBottom);
#endif
continue;
}
int opener = (lastWinding & windingMask) == 0;
bool closer = (winding & windingMask) == 0;
SkASSERT(!opener | !closer);
bool inWinding = opener | closer;
SkPoint clippedPts[4];
const SkPoint* clipped = NULL;
bool moreToDo, aboveBottom;
do {
double currentT = activePtr->t();
const SkPoint* points = wt.fPts;
double nextT;
SkPath::Verb verb = activePtr->fVerb;
do {
nextT = activePtr->nextT();
// FIXME: obtuse: want efficient way to say
// !currentT && currentT != 1 || !nextT && nextT != 1
if (currentT * nextT != 0 || currentT + nextT != 1) {
// OPTIMIZATION: if !inWinding, we only need
// clipped[1].fY
switch (verb) {
case SkPath::kLine_Verb:
LineSubDivide(points, currentT, nextT, clippedPts);
break;
case SkPath::kQuad_Verb:
QuadSubDivide(points, currentT, nextT, clippedPts);
break;
case SkPath::kCubic_Verb:
CubicSubDivide(points, currentT, nextT, clippedPts);
break;
default:
SkASSERT(0);
break;
}
clipped = clippedPts;
} else {
clipped = points;
}
if (inWinding && !activePtr->fSkip && (fill ? clipped[0].fY
!= clipped[verb].fY : clipped[0] != clipped[verb])) {
#if DEBUG_STITCH_EDGE
SkDebugf("%*s add%s %1.9g,%1.9g %1.9g,%1.9g edge=%d"
" v=%d t=(%1.9g,%1.9g)\n", tab, "",
kUVerbStr[verb], clipped[0].fX, clipped[0].fY,
clipped[verb].fX, clipped[verb].fY,
activePtr->ID(),
activePtr->fWorkEdge.fVerb
- activePtr->fWorkEdge.fEdge->fVerbs.begin(),
currentT, nextT);
#endif
outBuilder.addCurve(clipped, (SkPath::Verb) verb,
activePtr->fWorkEdge.fEdge->fID,
activePtr->fCloseCall);
} else {
#if DEBUG_STITCH_EDGE
SkDebugf("%*s skip%s %1.9g,%1.9g %1.9g,%1.9g"
" edge=%d v=%d t=(%1.9g,%1.9g)\n", tab, "",
kUVerbStr[verb], clipped[0].fX, clipped[0].fY,
clipped[verb].fX, clipped[verb].fY,
activePtr->ID(),
activePtr->fWorkEdge.fVerb
- activePtr->fWorkEdge.fEdge->fVerbs.begin(),
currentT, nextT);
#endif
}
// by advancing fAbove/fBelow, the next call to sortHorizontal
// will use these values if they're still valid instead of
// recomputing
if (clipped[verb].fY > activePtr->fBelow.fY
&& bottom >= activePtr->fBelow.fY
&& verb == SkPath::kLine_Verb) {
activePtr->fAbove = activePtr->fBelow;
activePtr->fBelow = activePtr->fTangent = clipped[verb];
activePtr->fTAbove = activePtr->fTBelow < 1
? activePtr->fTBelow : 0;
activePtr->fTBelow = nextT;
}
currentT = nextT;
moreToDo = activePtr->advanceT();
activePtr->fYBottom = clipped[verb].fY; // was activePtr->fCloseCall ? bottom :
// clearing the fSkip/fCloseCall bit here means that trailing edges
// fall out of sync, if one edge is long and another is a series of short pieces
// if fSkip/fCloseCall is set, need to recompute coincidence/too-close-to-call
// after advancing
// another approach would be to restrict bottom to smaller part of close call
// maybe this is already happening with coincidence when intersection is computed,
// and needs to be added to the close call computation as well
// this is hard to do because that the bottom is important is not known when
// the lines are intersected; only when the computation for edge sorting is done
// does the need for new bottoms become apparent.
// maybe this is good incentive to scrap the current sort and do an insertion
// sort that can take this into consideration when the x value is computed
// FIXME: initialized in sortHorizontal, cleared here as well so
// that next edge is not skipped -- but should skipped edges ever
// continue? (probably not)
aboveBottom = clipped[verb].fY < bottom;
if (clipped[0].fY != clipped[verb].fY) {
activePtr->fSkip = false;
activePtr->fCloseCall = false;
aboveBottom &= !activePtr->fCloseCall;
}
#if DEBUG_STITCH_EDGE
else {
if (activePtr->fSkip || activePtr->fCloseCall) {
SkDebugf("%s skip or close == %1.9g\n", __FUNCTION__,
clippedPts[0].fY);
}
}
#endif
} while (moreToDo & aboveBottom);
} while ((moreToDo || activePtr->advance()) & aboveBottom);
}
}
#if DEBUG_DUMP
static void dumpEdgeList(const SkTDArray<InEdge*>& edgeList,
const InEdge& edgeSentinel) {
InEdge** debugPtr = edgeList.begin();
do {
(*debugPtr++)->dump();
} while (*debugPtr != &edgeSentinel);
}
#else
static void dumpEdgeList(const SkTDArray<InEdge*>& ,
const InEdge& ) {
}
#endif
void simplify(const SkPath& path, bool asFill, SkPath& simple) {
// returns 1 for evenodd, -1 for winding, regardless of inverse-ness
int windingMask = (path.getFillType() & 1) ? 1 : -1;
simple.reset();
simple.setFillType(SkPath::kEvenOdd_FillType);
// turn path into list of edges increasing in y
// if an edge is a quad or a cubic with a y extrema, note it, but leave it
// unbroken. Once we have a list, sort it, then walk the list (walk edges
// twice that have y extrema's on top) and detect crossings -- look for raw
// bounds that cross over, then tight bounds that cross
SkTArray<InEdge> edges;
SkTDArray<HorizontalEdge> horizontalEdges;
InEdgeBuilder builder(path, asFill, edges, horizontalEdges);
SkTDArray<InEdge*> edgeList;
InEdge edgeSentinel;
edgeSentinel.reset();
makeEdgeList(edges, edgeSentinel, edgeList);
SkTDArray<HorizontalEdge*> horizontalList;
HorizontalEdge horizontalSentinel;
makeHorizontalList(horizontalEdges, horizontalSentinel, horizontalList);
InEdge** currentPtr = edgeList.begin();
if (!currentPtr) {
return;
}
// find all intersections between edges
// beyond looking for horizontal intercepts, we need to know if any active edges
// intersect edges below 'bottom', but above the active edge segment.
// maybe it makes more sense to compute all intercepts before doing anything
// else, since the intercept list is long-lived, at least in the current design.
SkScalar y = (*currentPtr)->fBounds.fTop;
HorizontalEdge** currentHorizontal = horizontalList.begin();
do {
InEdge** lastPtr = currentPtr; // find the edge below the bottom of the first set
SkScalar bottom = findBottom(currentPtr, edgeList.end(),
NULL, y, asFill, lastPtr);
if (lastPtr > currentPtr) {
if (currentHorizontal) {
if ((*currentHorizontal)->fY < SK_ScalarMax) {
addBottomT(currentPtr, lastPtr, currentHorizontal);
}
currentHorizontal = advanceHorizontal(currentHorizontal, bottom);
}
addIntersectingTs(currentPtr, lastPtr);
}
y = bottom;
currentPtr = advanceEdges(NULL, currentPtr, lastPtr, y);
} while (*currentPtr != &edgeSentinel);
// if a quadratic or cubic now has an intermediate T value, see if the Ts
// on either side cause the Y values to monotonically increase. If not, split
// the curve at the new T.
// try an alternate approach which does not split curves or stitch edges
// (may still need adjustCoincident, though)
// the idea is to output non-intersecting contours, then figure out their
// respective winding contribution
// each contour will need to know whether it is CW or CCW, and then whether
// a ray from that contour hits any a contour that contains it. The ray can
// move to the left and then arbitrarily move up or down (as long as it never
// moves to the right) to find a reference sibling contour or containing
// contour. If the contour is part of an intersection, the companion contour
// that is part of the intersection can determine the containership.
if (builder.containsCurves()) {
currentPtr = edgeList.begin();
SkTArray<InEdge> splits;
do {
(*currentPtr)->splitInflectionPts(splits);
} while (*++currentPtr != &edgeSentinel);
if (splits.count()) {
for (int index = 0; index < splits.count(); ++index) {
edges.push_back(splits[index]);
}
edgeList.reset();
makeEdgeList(edges, edgeSentinel, edgeList);
}
}
dumpEdgeList(edgeList, edgeSentinel);
// walk the sorted edges from top to bottom, computing accumulated winding
SkTDArray<ActiveEdge> activeEdges;
OutEdgeBuilder outBuilder(asFill);
currentPtr = edgeList.begin();
y = (*currentPtr)->fBounds.fTop;
do {
InEdge** lastPtr = currentPtr; // find the edge below the bottom of the first set
SkScalar bottom = findBottom(currentPtr, edgeList.end(),
&activeEdges, y, asFill, lastPtr);
if (lastPtr > currentPtr) {
bottom = computeInterceptBottom(activeEdges, y, bottom);
SkTDArray<ActiveEdge*> activeEdgeList;
sortHorizontal(activeEdges, activeEdgeList, y);
bottom = adjustCoincident(activeEdgeList, windingMask, y, bottom,
outBuilder);
stitchEdge(activeEdgeList, y, bottom, windingMask, asFill, outBuilder);
}
y = bottom;
// OPTIMIZATION: as edges expire, InEdge allocations could be released
currentPtr = advanceEdges(&activeEdges, currentPtr, lastPtr, y);
} while (*currentPtr != &edgeSentinel);
// assemble output path from string of pts, verbs
outBuilder.bridge();
outBuilder.assemble(simple);
}