- Mesh objects split by material- many 3ds objects used more then 16 per mesh. and when a face looses its image texture its tedious to set again.
- Removed a lot of unneeded variable creation.
- append group: appends group + puts objects in scene
- link group: only links group, doesn't put objects in scene
- append particle system with group: appends group + objects in scene
- link particle system with group: only links group
+ Added note about using the config files. I repeat it here: a user NEVER
should edit config/(platform)-config.py directly. Instead, make a copy of
config/(platform)-config.py to user-config.py, and change that instead.
/Nathan
PS. now I can say "I told you", and be sure I will :P
If "export NAN_USE_FFMPEG_CONFIG=true" is added to user-def.mk,
the system executes the ffmeg-config program to set values
for NAN_FFMPEG (--prefix), NAN_FFMPEGLIBS (--libs avcodec avformat),
and NAN_FFMPEGCFLAGS (--cflags). Only one used so far is the
NAN_FFMPEGLIBS for linking on linux (if requested to do so).
Current default is not to do this.
always statically (you have to force it to build a dynamic library) the
resulting binary is redistributable.
The code is made ffmpeg-version independent using #ifdef's.
* Large sequencer rewrite to support:
- Audio-tracks, which are not completely loaded into memory (hdaudio) but
kept on disk instead.
- A dependency tree, that builds only the Imbufs, that are really needed
- Cleaner sequencer code
- Per instance data in sequencer plugins (without this, the Dynamic
Noise Reduction plugin would be impossible)
- A Luma Waveform display
- A U/V scatter plot display
- Memcache limiting in sequencer
- Buttons changed according to the boosted framecount limit
* Add ffmpeg-read support in anim.c and util.c
* Makes ImBufs refcountable. You can now increase an internal refcounter
in ImBufs (using IMB_refImBuf) which is decreased by freeImBuf.
This makes it possible to simply pass ImBuf pointers around in the
sequencer saving a few memcopies.
* Boosts the blender frame limit by changing the type of the frame number
from short to int everywhere. Without this, timelines longer than a few
minutes are impossible to handle.
* Adds several types for ffmpeg input/output, hdaudio-tracks in sequencer
integrating a mini-webserver (around 300 lines of code) into blender.
Using the VFAPI-plugin in contrib/windows it enables blender to
directly feed its output into TMPGEnc, a commercial high quality MPEG-Encoder.
Since it is a mini-webserver, you can probably easily use it for other
interfacing purposes.
- color-correction-hsv & color-correction-yuv
Do color correction in HSV or YUV-space, rather sophisticated but slow.
You can control setup, gain and gamma and saturation (seperated by
shadows, midtones and highlights).
- gamma
a simple RGB-Gamma plugin, but very fast.
- dnr
Dynamic Noise Reduction (plugin ported from VirtualDub).
This helps mpeg encoding alot, by ignoring noise /movement
below a given threshold between frames.
It is also a lot faster than the original VirtualDub-plugin preserving
it's quality.
- There's an option to enabled ABF in the UV Calculation panel.
- ABF works with seams, pinning, live transform etc. UI wise it's pretty
much the same as LSCM.
- ABF favors low angular distortion over area distortion, much like LSCM.
It does however not have the tendency to "collapse" under high stretch,
which basically means that it will be able to handle complex charts
without pinning.
- ABF does however have on serious drawback: it is sensitive to degenerate
geometry, especially very small angles. I'm working on this, but it seems
to be quite difficult to improve the situation. There's still a bunch of
ideas left to try, but I'll just commit this now as is. If ABF fails, it
will fall back to LSCM, and there will be a warning print in the console.
- This implementation has the matrix tricks from ABF++ also. The second part
of ABF++, simplification before unwrap, is being worked on. The current
- ABF is about 3-4 times slower than LSCM, which was to be expected. For
live transform however, once the initial computation is done, it's just as
snappy as the new LSCM code.
Results are quite nice though:
http://users.pandora.be/blendix/snurgle.pnghttp://users.pandora.be/blendix/snurgle_flattened.png
You may recognize the Snurgle model by Colin Levy there.