diff --git a/cont.c b/cont.c index ba6120f9cd..75b259a3c2 100644 --- a/cont.c +++ b/cont.c @@ -313,7 +313,8 @@ fiber_pool_vacancy_reset(struct fiber_pool_vacancy * vacancy) } inline static struct fiber_pool_vacancy * -fiber_pool_vacancy_push(struct fiber_pool_vacancy * vacancy, struct fiber_pool_vacancy * head) { +fiber_pool_vacancy_push(struct fiber_pool_vacancy * vacancy, struct fiber_pool_vacancy * head) +{ vacancy->next = head; #ifdef FIBER_POOL_ALLOCATION_FREE @@ -327,36 +328,40 @@ fiber_pool_vacancy_push(struct fiber_pool_vacancy * vacancy, struct fiber_pool_v #ifdef FIBER_POOL_ALLOCATION_FREE static void -fiber_pool_vacancy_remove(struct fiber_pool_vacancy * vacancy) { +fiber_pool_vacancy_remove(struct fiber_pool_vacancy * vacancy) +{ if (vacancy->next) { vacancy->next->previous = vacancy->previous; } if (vacancy->previous) { vacancy->previous->next = vacancy->next; - } else { + } + else { // It's the head of the list: vacancy->stack.pool->vacancies = vacancy->next; } } inline static struct fiber_pool_vacancy * -fiber_pool_vacancy_pop(struct fiber_pool * pool) { +fiber_pool_vacancy_pop(struct fiber_pool * pool) +{ struct fiber_pool_vacancy * vacancy = pool->vacancies; if (vacancy) { - fiber_pool_vacancy_remove(vacancy); + fiber_pool_vacancy_remove(vacancy); } return vacancy; } #else inline static struct fiber_pool_vacancy * -fiber_pool_vacancy_pop(struct fiber_pool * pool) { +fiber_pool_vacancy_pop(struct fiber_pool * pool) +{ struct fiber_pool_vacancy * vacancy = pool->vacancies; if (vacancy) { - pool->vacancies = vacancy->next; + pool->vacancies = vacancy->next; } return vacancy; @@ -394,7 +399,8 @@ fiber_pool_allocate_memory(size_t * count, size_t stride) if (!base) { *count = (*count) >> 1; - } else { + } + else { return base; } #else @@ -403,7 +409,8 @@ fiber_pool_allocate_memory(size_t * count, size_t stride) if (base == MAP_FAILED) { *count = (*count) >> 1; - } else { + } + else { return base; } #endif @@ -466,9 +473,9 @@ fiber_pool_expand(struct fiber_pool * fiber_pool, size_t count) #endif vacancies = fiber_pool_vacancy_initialize( - fiber_pool, vacancies, - (char*)base + STACK_DIR_UPPER(0, RB_PAGE_SIZE), - size + fiber_pool, vacancies, + (char*)base + STACK_DIR_UPPER(0, RB_PAGE_SIZE), + size ); #ifdef FIBER_POOL_ALLOCATION_FREE @@ -536,14 +543,15 @@ fiber_pool_allocation_free(struct fiber_pool_allocation * allocation) } #ifdef _WIN32 - VirtualFree(allocation->base, 0, MEM_RELEASE); + VirtualFree(allocation->base, 0, MEM_RELEASE); #else - munmap(allocation->base, allocation->stride * allocation->count); + munmap(allocation->base, allocation->stride * allocation->count); #endif if (allocation->previous) { allocation->previous->next = allocation->next; - } else { + } + else { // We are the head of the list, so update the pool: allocation->pool->allocations = allocation->next; } @@ -598,7 +606,8 @@ fiber_pool_stack_acquire(struct fiber_pool * fiber_pool) { // We advise the operating system that the stack memory pages are no longer being used. // This introduce some performance overhead but allows system to relaim memory when there is pressure. static inline void -fiber_pool_stack_free(struct fiber_pool_stack * stack) { +fiber_pool_stack_free(struct fiber_pool_stack * stack) +{ void * base = fiber_pool_stack_base(stack); size_t size = stack->available; @@ -619,7 +628,8 @@ fiber_pool_stack_free(struct fiber_pool_stack * stack) { // Release and return a stack to the vacancy list. static void -fiber_pool_stack_release(struct fiber_pool_stack * stack) { +fiber_pool_stack_release(struct fiber_pool_stack * stack) +{ struct fiber_pool_vacancy * vacancy = fiber_pool_vacancy_pointer(stack->base, stack->size); if (DEBUG) fprintf(stderr, "fiber_pool_stack_release: %p used=%zu\n", stack->base, stack->pool->used); @@ -640,7 +650,8 @@ fiber_pool_stack_release(struct fiber_pool_stack * stack) { // Release address space and/or dirty memory: if (stack->allocation->used == 0) { fiber_pool_allocation_free(stack->allocation); - } else if (stack->pool->free_stacks) { + } + else if (stack->pool->free_stacks) { fiber_pool_stack_free(stack); } #else @@ -808,7 +819,7 @@ fiber_ptr(VALUE obj) NOINLINE(static VALUE cont_capture(volatile int *volatile stat)); #define THREAD_MUST_BE_RUNNING(th) do { \ - if (!(th)->ec->tag) rb_raise(rb_eThreadError, "not running thread"); \ + if (!(th)->ec->tag) rb_raise(rb_eThreadError, "not running thread"); \ } while (0) static VALUE @@ -886,7 +897,8 @@ cont_free(void *ptr) ruby_xfree(cont->saved_ec.vm_stack); ruby_xfree(cont->ensure_array); RUBY_FREE_UNLESS_NULL(cont->machine.stack); - } else { + } + else { rb_fiber_t *fiber = (rb_fiber_t*)cont; coroutine_destroy(&fiber->context); if (!fiber_is_root_p(fiber)) { @@ -1311,7 +1323,7 @@ cont_restore_1(rb_context_t *cont) if (cont->machine.stack_src) { FLUSH_REGISTER_WINDOWS; MEMCPY(cont->machine.stack_src, cont->machine.stack, - VALUE, cont->machine.stack_size); + VALUE, cont->machine.stack_size); } ruby_longjmp(cont->jmpbuf, 1); diff --git a/hash.c b/hash.c index 9c2da83fee..f458fd71aa 100644 --- a/hash.c +++ b/hash.c @@ -564,7 +564,8 @@ hash_ar_table_set(VALUE hash, ar_table *ar) #define RHASH_AR_TABLE_SIZE_INC(h) HASH_AR_TABLE_SIZE_ADD(h, 1) static inline void -RHASH_AR_TABLE_SIZE_DEC(VALUE h) { +RHASH_AR_TABLE_SIZE_DEC(VALUE h) +{ HASH_ASSERT(RHASH_AR_TABLE_P(h)); int new_size = RHASH_AR_TABLE_SIZE(h) - 1; diff --git a/node.h b/node.h index ebcc49c9be..a1f01ed088 100644 --- a/node.h +++ b/node.h @@ -138,7 +138,8 @@ typedef struct rb_code_location_struct { rb_code_position_t end_pos; } rb_code_location_t; -static inline rb_code_location_t code_loc_gen(rb_code_location_t *loc1, rb_code_location_t *loc2) +static inline rb_code_location_t +code_loc_gen(rb_code_location_t *loc1, rb_code_location_t *loc2) { rb_code_location_t loc; loc.beg_pos = loc1->beg_pos; diff --git a/numeric.c b/numeric.c index 9bf5e0a205..5b0ab8cbe8 100644 --- a/numeric.c +++ b/numeric.c @@ -4673,7 +4673,8 @@ compare_indexes(VALUE a, VALUE b) } static VALUE -generate_mask(VALUE len) { +generate_mask(VALUE len) +{ return rb_int_minus(rb_int_lshift(INT2FIX(1), len), INT2FIX(1)); }