mirror of
https://github.com/ruby/ruby.git
synced 2022-11-09 12:17:21 -05:00
511b55bcef
* Enable unaligned accesses on arm64 64-bit Arm platforms support unaligned accesses. Running the string benchmarks this change improves performance by an average of 1.04x, min .96x, max 1.21x, median 1.01x * arm64 enable gc optimizations Similar to x86 and powerpc optimizations. | |compare-ruby|built-ruby| |:------|-----------:|---------:| |hash1 | 0.225| 0.237| | | -| 1.05x| |hash2 | 0.110| 0.110| | | 1.00x| -| * vm_exec.c: improve performance for arm64 | |compare-ruby|built-ruby| |:------------------------------|-----------:|---------:| |vm_array | 26.501M| 27.959M| | | -| 1.06x| |vm_attr_ivar | 21.606M| 31.429M| | | -| 1.45x| |vm_attr_ivar_set | 21.178M| 26.113M| | | -| 1.23x| |vm_backtrace | 6.621| 6.668| | | -| 1.01x| |vm_bigarray | 26.205M| 29.958M| | | -| 1.14x| |vm_bighash | 504.155k| 479.306k| | | 1.05x| -| |vm_block | 16.692M| 21.315M| | | -| 1.28x| |block_handler_type_iseq | 5.083| 7.004| | | -| 1.38x|
133 lines
4.1 KiB
C
133 lines
4.1 KiB
C
#ifndef RUBY_GC_H
|
|
#define RUBY_GC_H 1
|
|
#include "ruby/ruby.h"
|
|
|
|
#if defined(__x86_64__) && !defined(_ILP32) && defined(__GNUC__)
|
|
#define SET_MACHINE_STACK_END(p) __asm__ __volatile__ ("movq\t%%rsp, %0" : "=r" (*(p)))
|
|
#elif defined(__i386) && defined(__GNUC__)
|
|
#define SET_MACHINE_STACK_END(p) __asm__ __volatile__ ("movl\t%%esp, %0" : "=r" (*(p)))
|
|
#elif defined(__powerpc64__) && defined(__GNUC__)
|
|
#define SET_MACHINE_STACK_END(p) __asm__ __volatile__ ("mr\t%0, %%r1" : "=r" (*(p)))
|
|
#elif defined(__aarch64__) && defined(__GNUC__)
|
|
#define SET_MACHINE_STACK_END(p) __asm__ __volatile__ ("mov\t%0, sp" : "=r" (*(p)))
|
|
#else
|
|
NOINLINE(void rb_gc_set_stack_end(VALUE **stack_end_p));
|
|
#define SET_MACHINE_STACK_END(p) rb_gc_set_stack_end(p)
|
|
#define USE_CONSERVATIVE_STACK_END
|
|
#endif
|
|
|
|
/* for GC debug */
|
|
|
|
#ifndef RUBY_MARK_FREE_DEBUG
|
|
#define RUBY_MARK_FREE_DEBUG 0
|
|
#endif
|
|
|
|
#if RUBY_MARK_FREE_DEBUG
|
|
extern int ruby_gc_debug_indent;
|
|
|
|
static inline void
|
|
rb_gc_debug_indent(void)
|
|
{
|
|
printf("%*s", ruby_gc_debug_indent, "");
|
|
}
|
|
|
|
static inline void
|
|
rb_gc_debug_body(const char *mode, const char *msg, int st, void *ptr)
|
|
{
|
|
if (st == 0) {
|
|
ruby_gc_debug_indent--;
|
|
}
|
|
rb_gc_debug_indent();
|
|
printf("%s: %s %s (%p)\n", mode, st ? "->" : "<-", msg, ptr);
|
|
|
|
if (st) {
|
|
ruby_gc_debug_indent++;
|
|
}
|
|
|
|
fflush(stdout);
|
|
}
|
|
|
|
#define RUBY_MARK_ENTER(msg) rb_gc_debug_body("mark", (msg), 1, ptr)
|
|
#define RUBY_MARK_LEAVE(msg) rb_gc_debug_body("mark", (msg), 0, ptr)
|
|
#define RUBY_FREE_ENTER(msg) rb_gc_debug_body("free", (msg), 1, ptr)
|
|
#define RUBY_FREE_LEAVE(msg) rb_gc_debug_body("free", (msg), 0, ptr)
|
|
#define RUBY_GC_INFO rb_gc_debug_indent(); printf
|
|
|
|
#else
|
|
#define RUBY_MARK_ENTER(msg)
|
|
#define RUBY_MARK_LEAVE(msg)
|
|
#define RUBY_FREE_ENTER(msg)
|
|
#define RUBY_FREE_LEAVE(msg)
|
|
#define RUBY_GC_INFO if(0)printf
|
|
#endif
|
|
|
|
#define RUBY_MARK_MOVABLE_UNLESS_NULL(ptr) do { \
|
|
VALUE markobj = (ptr); \
|
|
if (RTEST(markobj)) {rb_gc_mark_movable(markobj);} \
|
|
} while (0)
|
|
#define RUBY_MARK_UNLESS_NULL(ptr) do { \
|
|
VALUE markobj = (ptr); \
|
|
if (RTEST(markobj)) {rb_gc_mark(markobj);} \
|
|
} while (0)
|
|
#define RUBY_FREE_UNLESS_NULL(ptr) if(ptr){ruby_xfree(ptr);(ptr)=NULL;}
|
|
|
|
#if STACK_GROW_DIRECTION > 0
|
|
# define STACK_UPPER(x, a, b) (a)
|
|
#elif STACK_GROW_DIRECTION < 0
|
|
# define STACK_UPPER(x, a, b) (b)
|
|
#else
|
|
RUBY_EXTERN int ruby_stack_grow_direction;
|
|
int ruby_get_stack_grow_direction(volatile VALUE *addr);
|
|
# define stack_growup_p(x) ( \
|
|
(ruby_stack_grow_direction ? \
|
|
ruby_stack_grow_direction : \
|
|
ruby_get_stack_grow_direction(x)) > 0)
|
|
# define STACK_UPPER(x, a, b) (stack_growup_p(x) ? (a) : (b))
|
|
#endif
|
|
|
|
/*
|
|
STACK_GROW_DIR_DETECTION is used with STACK_DIR_UPPER.
|
|
|
|
On most normal systems, stacks grow from high address to lower address. In
|
|
this case, STACK_DIR_UPPER(a, b) will return (b), but on exotic systems where
|
|
the stack grows UP (from low address to high address), it will return (a).
|
|
*/
|
|
|
|
#if STACK_GROW_DIRECTION
|
|
#define STACK_GROW_DIR_DETECTION
|
|
#define STACK_DIR_UPPER(a,b) STACK_UPPER(0, (a), (b))
|
|
#else
|
|
#define STACK_GROW_DIR_DETECTION VALUE stack_grow_dir_detection
|
|
#define STACK_DIR_UPPER(a,b) STACK_UPPER(&stack_grow_dir_detection, (a), (b))
|
|
#endif
|
|
#define IS_STACK_DIR_UPPER() STACK_DIR_UPPER(1,0)
|
|
|
|
const char *rb_obj_info(VALUE obj);
|
|
const char *rb_raw_obj_info(char *buff, const int buff_size, VALUE obj);
|
|
|
|
VALUE rb_gc_disable_no_rest(void);
|
|
|
|
struct rb_thread_struct;
|
|
|
|
RUBY_SYMBOL_EXPORT_BEGIN
|
|
|
|
/* exports for objspace module */
|
|
size_t rb_objspace_data_type_memsize(VALUE obj);
|
|
void rb_objspace_reachable_objects_from(VALUE obj, void (func)(VALUE, void *), void *data);
|
|
void rb_objspace_reachable_objects_from_root(void (func)(const char *category, VALUE, void *), void *data);
|
|
int rb_objspace_markable_object_p(VALUE obj);
|
|
int rb_objspace_internal_object_p(VALUE obj);
|
|
int rb_objspace_marked_object_p(VALUE obj);
|
|
int rb_objspace_garbage_object_p(VALUE obj);
|
|
|
|
void rb_objspace_each_objects(
|
|
int (*callback)(void *start, void *end, size_t stride, void *data),
|
|
void *data);
|
|
|
|
void rb_objspace_each_objects_without_setup(
|
|
int (*callback)(void *, void *, size_t, void *),
|
|
void *data);
|
|
|
|
RUBY_SYMBOL_EXPORT_END
|
|
|
|
#endif /* RUBY_GC_H */
|