Just a couple facts for setup:
- Thrust doesn't operate in-place for all of it's operations.
- You can supply custom allocators to
thrust::device_vectors
.
I've looked in thrust::system
and thrust::system::cuda
and haven't found anything that looks like a static system allocator. By that I mean, I can't see a way of replacing the allocator that thrust uses internally to allocate extra memory for the out-of-place algorithms.
I also find it hard to believe that the functions that are not in-place use the allocators for the given thrust::device_vectors
to allocator working memory.
Question: Does thrust
have a way of replacing the internal allocator with a user defined one?
Related questions:
Thrust's
custom_temporary_allocation
example demonstrates how to build your own custom allocator for the temporary storage used internally by Thrust algorithms. The example uses a caching scheme to perform allocation but in principle you could use any strategy you like.Basically, the idea is to build a custom backend derived from the CUDA backend specifically for the purpose of customizing allocation. Then, when you'd like to use an algorithm with your custom allocator, you point Thrust at your custom backend when you call the algorithm.
Note that this feature requires Thrust 1.6 or better.