Welcome to mirror list, hosted at ThFree Co, Russian Federation.

github.com/torch/cutorch.git - Unnamed repository; edit this file 'description' to name the repository.
summaryrefslogtreecommitdiff
AgeCommit message (Collapse)Author
2016-12-17Revert "Bugfix of type in THCTensor macro."revert-639-patch-1Soumith Chintala
2016-12-16Merge pull request #639 from popol1991/patch-1Soumith Chintala
Bugfix of type in THCTensor macro.
2016-12-16Bugfix of type in THCTensor macro.Gao Yingkai
A fix for issue #632.
2016-12-16Merge pull request #637 from pavanky/test-fixesSoumith Chintala
Fixing various tests
2016-12-16Fixing various testsPavan Yalamanchili
- Increased the number of elements being used by distributions. - Fixed indexFill to generate a number that can be used by all types.
2016-12-15 fix wrong export directive for THCCachingHostAllocator (#633)Eric Cosatto
fix wrong export directive for THCCachingHostAllocator
2016-12-14Merge pull request #630 from apaszke/bernoulliSoumith Chintala
Implement bernoulli with element-wise probabilities for all types
2016-12-13Implement bernoulli with element-wise probabilities for all typesAdam Paszke
2016-12-12Merge pull request #628 from killeent/more-documentationSoumith Chintala
TensorInfo related code documentation
2016-12-12TensorInfo related code documentationTrevor Killeen
2016-12-02Merge pull request #619 from colesbury/cached_pinned_memory_fixSoumith Chintala
Process outstanding CUDA events in recordEvent
2016-12-02Process outstanding CUDA events in recordEventSam Gross
Without this, the cuda_events could continuously grow from calls to cudaMemcpyAsync, but would never be processed if there were no new pinned memory allocations. For example: t1 = cutorch.createCudaHostTensor(10) t2 = torch.CudaTensor(10) while true do t2:copyAsync(t1) end
2016-12-02Merge pull request #618 from colesbury/cached_pinned_memorySoumith Chintala
Add caching allocator for pinned (page-locked) memory
2016-12-02Add caching allocator for pinned (host) memorySam Gross
Adds a caching allocator for CUDA pinned (page-locked) memory. This avoid synchronization due to cudaFreeHost or cudaHostUnregister at the expense of potentially higher host memory usage. Correctness is preserved by recording CUDA events after each cudaMemcpyAsync involving the pinned memory. The pinned memory allocations are not reused until all events associated with it have completed.
2016-12-01Adds a CUDA "sleep" kernelSam Gross
Adds a CUDA "sleep" kernel which spins for the given number of iterations. This is useful for testing correct synchronization with streams.
2016-11-28Merge pull request #614 from BTNC/winSoumith Chintala
use local modified select_compute_arch.cmake for msvc
2016-11-28use local modified select_compute_arch.cmake for msvcRui Guo
2016-11-26Merge pull request #613 from colesbury/lazySoumith Chintala
Lazily initialize CUDA devices (take 2)
2016-11-26Lazily initialize CUDA devicesSam Gross
Previously, cutorch would initialize every CUDA device and enable P2P access between all pairs. This slows down start-up, especially with 8 devices. Now, THCudaInit does not initialize any devices and P2P access is enabled lazily. Setting the random number generator seed also does not initialize the device until random numbers are actually used.
2016-11-24Merge pull request #611 from torch/revert-610-lazySoumith Chintala
Revert "Lazily initialize CUDA devices"
2016-11-24Revert "Lazily initialize CUDA devices"revert-610-lazySoumith Chintala
2016-11-24remove spurious prints in testssoumith
2016-11-24Merge pull request #610 from colesbury/lazySoumith Chintala
Lazily initialize CUDA devices
2016-11-24Implemented cudaMemGetInfo for caching allocator (#600)Boris Fomitchev
* Implemented cudaMemGetInfo for caching allocator
2016-11-23Lazily initialize CUDA devicesSam Gross
Previously, cutorch would initialize every CUDA device and enable P2P access between all pairs. This slows down start-up, especially with 8 devices. Now, THCudaInit does not initialize any devices and P2P access is enabled lazily. Setting the random number generator seed also does not initialize the device until random numbers are actually used.
2016-11-18Merge pull request #607 from killeent/half-guardSoumith Chintala
guard random functions for half
2016-11-18guard random functions for halfTrevor Killeen
2016-11-18Merge pull request #605 from gchanan/halfAddrAddmvSoumith Chintala
Add half support for addmv and addr.
2016-11-18Add half support for addmv and addr.Gregory Chanan
2016-11-17Merge pull request #604 from killeent/memleakSoumith Chintala
fix memory leak in (equal)
2016-11-17fix memory leak in (equal)Trevor Killeen
2016-11-17Merge pull request #603 from killeent/remainderSoumith Chintala
Implement fmod, remainder, equal in Cutorch
2016-11-17add support for equal in cutorchTrevor Killeen
2016-11-17Merge pull request #602 from killeent/magmaSoumith Chintala
Magma functions to generic
2016-11-16add support for fmod in cutorchTrevor Killeen
2016-11-16add support for remainder in cutorchTrevor Killeen
2016-11-16[cutorch mag2gen] more cleanupTrevor Killeen
2016-11-16[cutorch mag2gen] some cleanupTrevor Killeen
2016-11-16[cutorch mag2gen] move qr to genericTrevor Killeen
2016-11-16[cutorch mag2gen] move potr* to genericTrevor Killeen
2016-11-16[cutorch mag2gen] move inverse to genericTrevor Killeen
2016-11-16[cutorch mag2gen] move svd to genericTrevor Killeen
2016-11-16[cutorch mag2gen] move eig to genericTrevor Killeen
2016-11-16[cutorch mag2gen] move symeig to genericTrevor Killeen
2016-11-16[cutorch mag2gen] move gels to genericTrevor Killeen
2016-11-16[cutorch mag2gen] code refactor to support generics; move gesv to genericTrevor Killeen
2016-11-16[cutorch mag2gen] generic MAGMA memory allocator functionTrevor Killeen
2016-11-16[cutorch potr*] API parity for potr* functions in cutorchTrevor Killeen
2016-11-15Merge pull request #601 from 1nadequacy/fix_baddbmmSoumith Chintala
[cutorch] remove syncing point from baddbmm
2016-11-15[cutorch] remove syncing point from baddbmmDenis Yarats
This change removes HtoD copies inside baddbmm. These copies introduce a syncing point which causes slow downs in a multi gpu training. Test plan: Run unittests for baddbmm.