gcc 4.7.2 in Ubuntu 12.10 detected an issue with assigning the result of getopt
to a char. When getopt returns -1 after exhausting the options, if it is
assigned to a char it sets all its bits to 1. When this result is then tested
for equality to the literal -1, the char was promoted to int, but the byte 0xff
turns into the signed integer 255. Aparently other compiling where assuming -1
is a char-type literal avoiding the conversions.
- primitive clipping is now more paranoid about roundoff errors
- significantly improved the kd-tree benchmark "kdbench"
- fixed a parallelization-related bug in the kd-tree builder
- fixed some remaining floating point issues in the min-max binning code
- completely moved TriMesh over to a more compact SoA representation
- the OpenGL renderer now sends smaller packets of geometry to the GPU, which
helps keep the operating system more responsive (NVidia drivers didn't like
the previous behavior much).
- Switched vectors classes to templates
- Moved most path handling over to boost_filesystem
- Nicer include file structure
- Better documentation for libcore
- Doxygen support
- Vector/Point/etc. data structures are now unitialized by default!