Age | Commit message (Collapse) | Author |
|
|
|
The officially supported compiled is now MSVC 2013, and MSVC 2008 will be faded out slowly over the coming weeks.
|
|
This also updates the configurations to build kernels for compute capability
5.0 cards, when using and older CUDA toolkit version this will be skipped.
Also includes tweaks to improve performance with this version:
* Increase max registers on sm_30, sm_35 and sm_50
* No longer use texture storage on sm_30
|
|
|
|
|
|
|
|
compilers that don't support it.
CMake still needs to updated to work the same for consistency, but this should
fix the OS X buildbot at least.
|
|
This should solve the missing -mavx option for Cycles, and also give better
performance since GCC 4.2 is now 5 years old already.
|
|
|
|
|
|
This is an attempt to make it so revision from git
SHA is being displayed, not from libs svn revision.
|
|
Consider this configuration final-1 for the migration,
Linux and OSX slaves seems to be fine, Windows slave
would need some changes from the slave side.
|
|
|
|
Need this so checkout of existing svn does not
interferes with new git checkout which would
need to be done.
|
|
|
|
use sw_vers -productVersion instead of uname -r, we must not redine it then from darwin version -> osx version
|
|
backend.
This means that if you have WITH_BF_QUICKTIME or WITH_CODEC_QUICKTIME enabled,
it will always use QTKit.
The old backend was only used on 32 bit OS X builds, now 32 and 64 bit builds will
give consistent input/output. On Windows or Linux quicktime isn't being used.
|
|
|
|
( always True )
|
|
|
|
|
|
|
|
|
|
|
|
using
a compiler older than CUDA 5.0 it will give a warning and skip this architecture.
|
|
Thanks Brecht and Thomas for reviewing ;)
|
|
|
|
|
|
|
|
|
|
Add numpy installation to blender player configuration,
this is so because player is building first and it installs
python, which prevented numpy installation from blender
configuration.
|
|
Numpy is compiled at build environment, but
importing it from blender somehow fails.
Will need help of Campbell for this.
|
|
|
|
Added new build option WITH_JACK_DYNLOAD for CMake and
WITH_BF_JACK_DYNLOAD for SCons, which means there'll be
no build-time linking against libjack and getting symbols
from libjack will happen runtime using dlopen and dlsym
tricks.
Alternative would be to use weak linking, but it'll require
having wrapper for preloading libjack.
This new options are disabled by default and they only
intended to be used on linux. Other platforms shall not
be using this and there shall be no functional changes
on non-linux platforms at all.
|
|
|
|
Didn't count indentation correct..
|
|
|
|
Make it so install directory is being nicely
cleaned before next build, which makes it
automatically removing all old files from
previous installations.
|
|
using methods from it
Otherwise linux buildbot fails dramatically.
That revision also leads to msvc runtime libs being removed, will fix this
in separate commit.
Take more respect for such creepy as linux and windows platforms, they're not so bad!
|
|
|
|
* Deprecate computing capability 1.3 (sm_13)
This commit disables auto build of sm_13 CUDA platform, which means that starting with Blender 2.67, we don't support sm_13 devices anymore. It has become difficult to support that and it was already feature incomplete (no render-passes, AO, Multi Closure etc).
It's still possible to manually enable sm_13 for own tests, but building might break in the future.
|
|
crosscompilation, but was caused by some bugs in toolkit itself.
Shall be all fine now (at least it is on my laptop now).
Let give it another try!
|
|
|
|
Crosscompiling of cubins doesn't work on linux with toolkit 4.2,
so use native toolkit for now. Disabled sm_13 for 32bit platform
for now.
Would keep cudakernels build target for a while. It doesn't hurt
being in the code and it could be helpful again once we'll switch
to toolkit 5.x where crosscompilation works fine.
Some further tweaks could probably be needed still, let's see how
building goes on buildbot now :)
|
|
- cycles ui used 'cscene' for scene.cycles and scene.cycles_curves
- style cleanup
|
|
|
|
Issue was caused by how CUDA devices availability done in Cycles.
Basically, if there's no WITH_CUDA_BINARIES buildtime, nvcc becomes
mandatory dependency.
Since kernels are building in separate target now, this logic broke
a bit.
Perhaps condition in util_cuda shall be changed to be a bit smarter,
but for now just work-around by enabling CUDA binaries when building
Cycles. Made it empty arch list to be sure no kernels will try to
re-compile after cudakernels target is done.
|
|
|
|
- BF_BITNESS should be passed as a command line argument
- Made it so CUDA binaries and OSL compiled scripts would
be installed regardless WITH_BF_PYTHON (which seems to
be quite obvious)
- Disable overwrite install, so CUDA kernels installed by
it's build target will be preserved when building blender
itself.
|
|
Solves memory limit issues when building sm_13 for 32bit platform.
Some further refinement of patch could be needed.
|