mirror of
https://github.com/ruby/ruby.git
synced 2022-11-09 12:17:21 -05:00
internal/gc.h rework
Improved readability by reducing the use of macros. Also moved some part of internal/compilers.h into this file, because it seems to be the right place for them.
This commit is contained in:
parent
adc49f0f9a
commit
c27bcd7057
Notes:
git
2019-12-26 20:46:00 +09:00
2 changed files with 106 additions and 53 deletions
|
@ -236,22 +236,4 @@ rb_obj_builtin_type(VALUE obj)
|
|||
# define BITFIELD(type, name, size) unsigned int name : size
|
||||
#endif
|
||||
|
||||
#if defined(USE_UNALIGNED_MEMBER_ACCESS) && USE_UNALIGNED_MEMBER_ACCESS && \
|
||||
(defined(__clang__) || GCC_VERSION_SINCE(9, 0, 0))
|
||||
#include "warnings.h"
|
||||
# define UNALIGNED_MEMBER_ACCESS(expr) __extension__({ \
|
||||
COMPILER_WARNING_PUSH; \
|
||||
COMPILER_WARNING_IGNORED(-Waddress-of-packed-member); \
|
||||
typeof(expr) unaligned_member_access_result = (expr); \
|
||||
COMPILER_WARNING_POP; \
|
||||
unaligned_member_access_result; \
|
||||
})
|
||||
#else
|
||||
# define UNALIGNED_MEMBER_ACCESS(expr) expr
|
||||
#endif
|
||||
#define UNALIGNED_MEMBER_PTR(ptr, mem) UNALIGNED_MEMBER_ACCESS(&(ptr)->mem)
|
||||
|
||||
#undef RB_OBJ_WRITE
|
||||
#define RB_OBJ_WRITE(a, slot, b) UNALIGNED_MEMBER_ACCESS(rb_obj_write((VALUE)(a), (VALUE *)(slot), (VALUE)(b), __FILE__, __LINE__))
|
||||
|
||||
#endif /* INTERNAL_COMPILERS_H */
|
||||
|
|
141
internal/gc.h
141
internal/gc.h
|
@ -9,6 +9,48 @@
|
|||
* modify this file, provided that the conditions mentioned in the
|
||||
* file COPYING are met. Consult the file for details.
|
||||
*/
|
||||
#include "ruby/config.h"
|
||||
|
||||
#include <stddef.h> /* for size_t */
|
||||
|
||||
#include "internal/compilers.h" /* for __has_attribute */
|
||||
#include "ruby/ruby.h" /* for rb_event_flag_t */
|
||||
|
||||
struct rb_execution_context_struct; /* in vm_core.h */
|
||||
|
||||
#undef NEWOBJ_OF
|
||||
#undef RB_NEWOBJ_OF
|
||||
#undef RB_OBJ_WRITE
|
||||
|
||||
/* optimized version of NEWOBJ() */
|
||||
#define RB_NEWOBJ_OF(var, T, c, f) \
|
||||
T *(var) = (T *)(((f) & FL_WB_PROTECTED) ? \
|
||||
rb_wb_protected_newobj_of((c), (f) & ~FL_WB_PROTECTED) : \
|
||||
rb_wb_unprotected_newobj_of((c), (f)))
|
||||
#define NEWOBJ_OF(var, T, c, f) RB_NEWOBJ_OF((var), T, (c), (f))
|
||||
#define RB_OBJ_GC_FLAGS_MAX 6 /* used in ext/objspace */
|
||||
|
||||
#ifndef USE_UNALIGNED_MEMBER_ACCESS
|
||||
# define UNALIGNED_MEMBER_ACCESS(expr) (expr)
|
||||
#elif ! USE_UNALIGNED_MEMBER_ACCESS
|
||||
# define UNALIGNED_MEMBER_ACCESS(expr) (expr)
|
||||
#elif ! (__has_warning("-Waddress-of-packed-member") || GCC_VERSION_SINCE(9, 0, 0))
|
||||
# define UNALIGNED_MEMBER_ACCESS(expr) (expr)
|
||||
#else
|
||||
# include "internal/warnings.h"
|
||||
# define UNALIGNED_MEMBER_ACCESS(expr) __extension__({ \
|
||||
COMPILER_WARNING_PUSH; \
|
||||
COMPILER_WARNING_IGNORED(-Waddress-of-packed-member); \
|
||||
__typeof__(expr) unaligned_member_access_result = (expr); \
|
||||
COMPILER_WARNING_POP; \
|
||||
unaligned_member_access_result; \
|
||||
})
|
||||
#endif
|
||||
|
||||
#define UNALIGNED_MEMBER_PTR(ptr, mem) UNALIGNED_MEMBER_ACCESS(&(ptr)->mem)
|
||||
#define RB_OBJ_WRITE(a, slot, b) \
|
||||
UNALIGNED_MEMBER_ACCESS(\
|
||||
rb_obj_write((VALUE)(a), (VALUE *)(slot), (VALUE)(b), __FILE__, __LINE__))
|
||||
|
||||
/* gc.c */
|
||||
extern VALUE *ruby_initial_gc_stress_ptr;
|
||||
|
@ -16,62 +58,91 @@ extern int ruby_disable_gc;
|
|||
void *ruby_mimmalloc(size_t size) RUBY_ATTR_MALLOC;
|
||||
void ruby_mimfree(void *ptr);
|
||||
void rb_objspace_set_event_hook(const rb_event_flag_t event);
|
||||
#if USE_RGENGC
|
||||
void rb_gc_writebarrier_remember(VALUE obj);
|
||||
#else
|
||||
#define rb_gc_writebarrier_remember(obj) 0
|
||||
#endif
|
||||
void ruby_gc_set_params(void);
|
||||
void rb_copy_wb_protected_attribute(VALUE dest, VALUE obj);
|
||||
|
||||
#if defined(HAVE_MALLOC_USABLE_SIZE) || defined(HAVE_MALLOC_SIZE) || defined(_WIN32)
|
||||
#define ruby_sized_xrealloc(ptr, new_size, old_size) ruby_xrealloc(ptr, new_size)
|
||||
#define ruby_sized_xrealloc2(ptr, new_count, element_size, old_count) ruby_xrealloc2(ptr, new_count, element_size)
|
||||
#define ruby_sized_xfree(ptr, size) ruby_xfree(ptr)
|
||||
#define SIZED_REALLOC_N(var,type,n,old_n) REALLOC_N(var, type, n)
|
||||
#else
|
||||
RUBY_SYMBOL_EXPORT_BEGIN
|
||||
void *ruby_sized_xrealloc(void *ptr, size_t new_size, size_t old_size) RUBY_ATTR_RETURNS_NONNULL RUBY_ATTR_ALLOC_SIZE((2));
|
||||
void *ruby_sized_xrealloc2(void *ptr, size_t new_count, size_t element_size, size_t old_count) RUBY_ATTR_RETURNS_NONNULL RUBY_ATTR_ALLOC_SIZE((2, 3));
|
||||
void ruby_sized_xfree(void *x, size_t size);
|
||||
RUBY_SYMBOL_EXPORT_END
|
||||
#define SIZED_REALLOC_N(var,type,n,old_n) ((var)=(type*)ruby_sized_xrealloc2((void*)(var), (n), sizeof(type), (old_n)))
|
||||
#endif
|
||||
|
||||
/* optimized version of NEWOBJ() */
|
||||
#undef NEWOBJF_OF
|
||||
#undef RB_NEWOBJ_OF
|
||||
#define RB_NEWOBJ_OF(obj,type,klass,flags) \
|
||||
type *(obj) = (type*)(((flags) & FL_WB_PROTECTED) ? \
|
||||
rb_wb_protected_newobj_of(klass, (flags) & ~FL_WB_PROTECTED) : \
|
||||
rb_wb_unprotected_newobj_of(klass, flags))
|
||||
#define NEWOBJ_OF(obj,type,klass,flags) RB_NEWOBJ_OF(obj,type,klass,flags)
|
||||
|
||||
#if __has_attribute(alloc_align)
|
||||
__attribute__((__alloc_align__(1)))
|
||||
#endif
|
||||
void *rb_aligned_malloc(size_t, size_t) RUBY_ATTR_MALLOC RUBY_ATTR_ALLOC_SIZE((2));
|
||||
|
||||
size_t rb_size_mul_or_raise(size_t, size_t, VALUE); /* used in compile.c */
|
||||
size_t rb_size_mul_add_or_raise(size_t, size_t, size_t, VALUE); /* used in iseq.h */
|
||||
void *rb_xmalloc_mul_add(size_t, size_t, size_t) RUBY_ATTR_MALLOC;
|
||||
void *rb_xrealloc_mul_add(const void *, size_t, size_t, size_t);
|
||||
void *rb_xmalloc_mul_add_mul(size_t, size_t, size_t, size_t) RUBY_ATTR_MALLOC;
|
||||
void *rb_xcalloc_mul_add_mul(size_t, size_t, size_t, size_t) RUBY_ATTR_MALLOC;
|
||||
static inline void *ruby_sized_xrealloc_inlined(void *ptr, size_t new_size, size_t old_size) RUBY_ATTR_RETURNS_NONNULL RUBY_ATTR_ALLOC_SIZE((2));
|
||||
static inline void *ruby_sized_xrealloc2_inlined(void *ptr, size_t new_count, size_t elemsiz, size_t old_count) RUBY_ATTR_RETURNS_NONNULL RUBY_ATTR_ALLOC_SIZE((2, 3));
|
||||
static inline void ruby_sized_xfree_inlined(void *ptr, size_t size);
|
||||
|
||||
RUBY_SYMBOL_EXPORT_BEGIN
|
||||
const char *rb_objspace_data_type_name(VALUE obj);
|
||||
|
||||
/* gc.c (export) */
|
||||
const char *rb_objspace_data_type_name(VALUE obj);
|
||||
VALUE rb_wb_protected_newobj_of(VALUE, VALUE);
|
||||
VALUE rb_wb_unprotected_newobj_of(VALUE, VALUE);
|
||||
|
||||
size_t rb_obj_memsize_of(VALUE);
|
||||
void rb_gc_verify_internal_consistency(void);
|
||||
|
||||
#define RB_OBJ_GC_FLAGS_MAX 6
|
||||
size_t rb_obj_gc_flags(VALUE, ID[], size_t);
|
||||
void rb_gc_mark_values(long n, const VALUE *values);
|
||||
void rb_gc_mark_vm_stack_values(long n, const VALUE *values);
|
||||
void *ruby_sized_xrealloc(void *ptr, size_t new_size, size_t old_size) RUBY_ATTR_RETURNS_NONNULL RUBY_ATTR_ALLOC_SIZE((2));
|
||||
void *ruby_sized_xrealloc2(void *ptr, size_t new_count, size_t element_size, size_t old_count) RUBY_ATTR_RETURNS_NONNULL RUBY_ATTR_ALLOC_SIZE((2, 3));
|
||||
void ruby_sized_xfree(void *x, size_t size);
|
||||
RUBY_SYMBOL_EXPORT_END
|
||||
|
||||
MJIT_SYMBOL_EXPORT_BEGIN
|
||||
int rb_ec_stack_check(struct rb_execution_context_struct *ec);
|
||||
void rb_gc_writebarrier_remember(VALUE obj);
|
||||
const char *rb_obj_info(VALUE obj);
|
||||
MJIT_SYMBOL_EXPORT_END
|
||||
|
||||
#if defined(HAVE_MALLOC_USABLE_SIZE) || defined(HAVE_MALLOC_SIZE) || defined(_WIN32)
|
||||
|
||||
static inline void *
|
||||
ruby_sized_xrealloc_inlined(void *ptr, size_t new_size, size_t old_size)
|
||||
{
|
||||
return ruby_xrealloc(ptr, new_size);
|
||||
}
|
||||
|
||||
static inline void *
|
||||
ruby_sized_xrealloc2_inlined(void *ptr, size_t new_count, size_t elemsiz, size_t old_count)
|
||||
{
|
||||
return ruby_xrealloc2(ptr, new_count, elemsiz);
|
||||
}
|
||||
|
||||
static inline void
|
||||
ruby_sized_xfree_inlined(void *ptr, size_t size)
|
||||
{
|
||||
ruby_xfree(ptr);
|
||||
}
|
||||
|
||||
# define SIZED_REALLOC_N(x, y, z, w) REALLOC_N(x, y, z)
|
||||
|
||||
#else
|
||||
|
||||
static inline void *
|
||||
ruby_sized_xrealloc_inlined(void *ptr, size_t new_size, size_t old_size)
|
||||
{
|
||||
return ruby_sized_xrealloc(ptr, new_size, old_size);
|
||||
}
|
||||
|
||||
static inline void *
|
||||
ruby_sized_xrealloc2_inlined(void *ptr, size_t new_count, size_t elemsiz, size_t old_count)
|
||||
{
|
||||
return ruby_sized_xrealloc2(ptr, new_count, elemsiz, old_count);
|
||||
}
|
||||
|
||||
static inline void
|
||||
ruby_sized_xfree_inlined(void *ptr, size_t size)
|
||||
{
|
||||
ruby_sized_xfree(ptr, size);
|
||||
}
|
||||
|
||||
# define SIZED_REALLOC_N(v, T, m, n) \
|
||||
((v) = (T *)ruby_sized_xrealloc2((void *)(v), (n), sizeof(T), (m)))
|
||||
|
||||
#endif /* HAVE_MALLOC_USABLE_SIZE */
|
||||
|
||||
#define ruby_sized_xrealloc ruby_sized_xrealloc_inlined
|
||||
#define ruby_sized_xrealloc2 ruby_sized_xrealloc2_inlined
|
||||
#define ruby_sized_xfree ruby_sized_xfree_inlined
|
||||
#endif /* INTERNAL_GC_H */
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue