alignment 32 ompi/mca/io/romio321/romio/adio/common/ad_aggregate_new.c static void align_fr (int fr_size, ADIO_Offset fr_off, int alignment, alignment 274 ompi/mca/io/romio321/romio/adio/common/ad_aggregate_new.c static void align_fr (int fr_size, ADIO_Offset fr_off, int alignment, alignment 276 ompi/mca/io/romio321/romio/adio/common/ad_aggregate_new.c *aligned_fr_off = fr_off - (fr_off % alignment); alignment 277 ompi/mca/io/romio321/romio/adio/common/ad_aggregate_new.c *aligned_fr_size = ((fr_off + fr_size) / alignment) * alignment - alignment 279 ompi/mca/io/romio321/romio/adio/common/ad_aggregate_new.c if ((fr_off + fr_size) % alignment) alignment 280 ompi/mca/io/romio321/romio/adio/common/ad_aggregate_new.c *aligned_fr_size += alignment; alignment 282 ompi/mca/io/romio321/romio/adio/common/ad_aggregate_new.c assert(!((*aligned_fr_off % alignment) || alignment 283 ompi/mca/io/romio321/romio/adio/common/ad_aggregate_new.c (*aligned_fr_size % alignment))); alignment 155 opal/mca/allocator/bucket/allocator_bucket_alloc.c size_t size, size_t alignment) alignment 170 opal/mca/allocator/bucket/allocator_bucket_alloc.c aligned_max_size = size + alignment + sizeof(mca_allocator_bucket_chunk_header_t) alignment 187 opal/mca/allocator/bucket/allocator_bucket_alloc.c alignment_off = ((size_t) aligned_memory) % alignment; alignment 188 opal/mca/allocator/bucket/allocator_bucket_alloc.c aligned_memory += (alignment - alignment_off); alignment 157 opal/mca/allocator/bucket/allocator_bucket_alloc.h size_t alignment); alignment 1139 oshmem/mca/memheap/ptmalloc/malloc.c void* mspace_memalign(mspace msp, size_t alignment, size_t bytes); alignment 4190 oshmem/mca/memheap/ptmalloc/malloc.c static void* internal_memalign(mstate m, size_t alignment, size_t bytes) { alignment 4191 oshmem/mca/memheap/ptmalloc/malloc.c if (alignment <= MALLOC_ALIGNMENT) /* Can just use malloc */ alignment 4193 oshmem/mca/memheap/ptmalloc/malloc.c if (alignment < MIN_CHUNK_SIZE) /* must be at least a minimum chunk size */ alignment 4194 oshmem/mca/memheap/ptmalloc/malloc.c alignment = MIN_CHUNK_SIZE; alignment 4195 oshmem/mca/memheap/ptmalloc/malloc.c if ((alignment & (alignment-SIZE_T_ONE)) != 0) {/* Ensure a power of 2 */ alignment 4197 oshmem/mca/memheap/ptmalloc/malloc.c while (a < alignment) a <<= 1; alignment 4198 oshmem/mca/memheap/ptmalloc/malloc.c alignment = a; alignment 4201 oshmem/mca/memheap/ptmalloc/malloc.c if (bytes >= MAX_REQUEST - alignment) { alignment 4208 oshmem/mca/memheap/ptmalloc/malloc.c size_t req = nb + alignment + MIN_CHUNK_SIZE - CHUNK_OVERHEAD; alignment 4216 oshmem/mca/memheap/ptmalloc/malloc.c if ((((size_t)(mem)) % alignment) != 0) { /* misaligned */ alignment 4226 oshmem/mca/memheap/ptmalloc/malloc.c alignment - alignment 4228 oshmem/mca/memheap/ptmalloc/malloc.c -alignment)); alignment 4230 oshmem/mca/memheap/ptmalloc/malloc.c br : br+alignment; alignment 4260 oshmem/mca/memheap/ptmalloc/malloc.c dl_assert((((size_t)(chunk2mem(p))) % alignment) == 0); alignment 4697 oshmem/mca/memheap/ptmalloc/malloc.c void* dlmemalign(size_t alignment, size_t bytes) { alignment 4698 oshmem/mca/memheap/ptmalloc/malloc.c return internal_memalign(gm, alignment, bytes); alignment 5114 oshmem/mca/memheap/ptmalloc/malloc.c void* mspace_memalign(mspace msp, size_t alignment, size_t bytes) { alignment 5120 oshmem/mca/memheap/ptmalloc/malloc.c return internal_memalign(ms, alignment, bytes);