mirror of
https://github.com/ruby/ruby.git
synced 2022-11-09 12:17:21 -05:00
mjit: get rid of rb_mjit_unit_node and use ccan/list
rb_mjit_unit can either exist in unit_queue or active_units, but not both. This will make state transitions for event-based MJIT process management easier. [ruby-core:89654] git-svn-id: svn+ssh://ci.ruby-lang.org/ruby/trunk@65468 b2dd03c8-39d4-4d8f-98ff-823fe69b080e
This commit is contained in:
parent
0c6edda08c
commit
c5177fa846
2 changed files with 66 additions and 130 deletions
67
mjit.c
67
mjit.c
|
@ -119,7 +119,7 @@ mjit_free_iseq(const rb_iseq_t *iseq)
|
|||
static void
|
||||
init_list(struct rb_mjit_unit_list *list)
|
||||
{
|
||||
list->head = NULL;
|
||||
list_head_init(&list->head);
|
||||
list->length = 0;
|
||||
}
|
||||
|
||||
|
@ -129,11 +129,11 @@ init_list(struct rb_mjit_unit_list *list)
|
|||
static void
|
||||
free_list(struct rb_mjit_unit_list *list)
|
||||
{
|
||||
struct rb_mjit_unit_node *node, *next;
|
||||
for (node = list->head; node != NULL; node = next) {
|
||||
next = node->next;
|
||||
free_unit(node->unit);
|
||||
xfree(node);
|
||||
struct rb_mjit_unit *unit = 0, *next;
|
||||
|
||||
list_for_each_safe(&list->head, unit, next, unode) {
|
||||
list_del(&unit->unode);
|
||||
free_unit(unit);
|
||||
}
|
||||
list->length = 0;
|
||||
}
|
||||
|
@ -248,24 +248,23 @@ unload_units(void)
|
|||
{
|
||||
rb_vm_t *vm = GET_THREAD()->vm;
|
||||
rb_thread_t *th = NULL;
|
||||
struct rb_mjit_unit_node *node, *next, *worst_node;
|
||||
struct rb_mjit_unit *unit = 0, *next, *worst;
|
||||
struct mjit_cont *cont;
|
||||
int delete_num, units_num = active_units.length;
|
||||
|
||||
/* For now, we don't unload units when ISeq is GCed. We should
|
||||
unload such ISeqs first here. */
|
||||
for (node = active_units.head; node != NULL; node = next) {
|
||||
next = node->next;
|
||||
if (node->unit->iseq == NULL) { /* ISeq is GCed. */
|
||||
free_unit(node->unit);
|
||||
remove_from_list(node, &active_units);
|
||||
list_for_each_safe(&active_units.head, unit, next, unode) {
|
||||
if (unit->iseq == NULL) { /* ISeq is GCed. */
|
||||
remove_from_list(unit, &active_units);
|
||||
free_unit(unit);
|
||||
}
|
||||
}
|
||||
|
||||
/* Detect units which are in use and can't be unloaded. */
|
||||
for (node = active_units.head; node != NULL; node = node->next) {
|
||||
assert(node->unit != NULL && node->unit->iseq != NULL && node->unit->handle != NULL);
|
||||
node->unit->used_code_p = FALSE;
|
||||
list_for_each_safe(&active_units.head, unit, next, unode) {
|
||||
assert(unit->iseq != NULL && unit->handle != NULL);
|
||||
unit->used_code_p = FALSE;
|
||||
}
|
||||
list_for_each(&vm->living_threads, th, vmlt_node) {
|
||||
mark_ec_units(th->ec);
|
||||
|
@ -280,23 +279,23 @@ unload_units(void)
|
|||
delete_num = active_units.length / 10;
|
||||
for (; active_units.length > mjit_opts.max_cache_size - delete_num;) {
|
||||
/* Find one unit that has the minimum total_calls. */
|
||||
worst_node = NULL;
|
||||
for (node = active_units.head; node != NULL; node = node->next) {
|
||||
if (node->unit->used_code_p) /* We can't unload code on stack. */
|
||||
worst = NULL;
|
||||
list_for_each_safe(&active_units.head, unit, next, unode) {
|
||||
if (unit->used_code_p) /* We can't unload code on stack. */
|
||||
continue;
|
||||
|
||||
if (worst_node == NULL || worst_node->unit->iseq->body->total_calls > node->unit->iseq->body->total_calls) {
|
||||
worst_node = node;
|
||||
if (worst == NULL || worst->iseq->body->total_calls > unit->iseq->body->total_calls) {
|
||||
worst = unit;
|
||||
}
|
||||
}
|
||||
if (worst_node == NULL)
|
||||
if (worst == NULL)
|
||||
break;
|
||||
|
||||
/* Unload the worst node. */
|
||||
verbose(2, "Unloading unit %d (calls=%lu)", worst_node->unit->id, worst_node->unit->iseq->body->total_calls);
|
||||
assert(worst_node->unit->handle != NULL);
|
||||
free_unit(worst_node->unit);
|
||||
remove_from_list(worst_node, &active_units);
|
||||
verbose(2, "Unloading unit %d (calls=%lu)", worst->id, worst->iseq->body->total_calls);
|
||||
assert(worst->handle != NULL);
|
||||
remove_from_list(worst, &active_units);
|
||||
free_unit(worst);
|
||||
}
|
||||
verbose(1, "Too many JIT code -- %d units unloaded", units_num - active_units.length);
|
||||
}
|
||||
|
@ -306,8 +305,6 @@ unload_units(void)
|
|||
void
|
||||
mjit_add_iseq_to_process(const rb_iseq_t *iseq)
|
||||
{
|
||||
struct rb_mjit_unit_node *node;
|
||||
|
||||
if (!mjit_enabled || pch_status == PCH_FAILED)
|
||||
return;
|
||||
|
||||
|
@ -317,14 +314,8 @@ mjit_add_iseq_to_process(const rb_iseq_t *iseq)
|
|||
/* Failure in creating the unit. */
|
||||
return;
|
||||
|
||||
node = create_list_node(iseq->body->jit_unit);
|
||||
if (node == NULL) {
|
||||
mjit_warning("failed to allocate a node to be added to unit_queue");
|
||||
return;
|
||||
}
|
||||
|
||||
CRITICAL_SECTION_START(3, "in add_iseq_to_process");
|
||||
add_to_list(node, &unit_queue);
|
||||
add_to_list(iseq->body->jit_unit, &unit_queue);
|
||||
if (active_units.length >= mjit_opts.max_cache_size) {
|
||||
unload_units();
|
||||
}
|
||||
|
@ -762,14 +753,14 @@ mjit_finish(void)
|
|||
void
|
||||
mjit_mark(void)
|
||||
{
|
||||
struct rb_mjit_unit_node *node;
|
||||
struct rb_mjit_unit *unit = 0;
|
||||
if (!mjit_enabled)
|
||||
return;
|
||||
RUBY_MARK_ENTER("mjit");
|
||||
CRITICAL_SECTION_START(4, "mjit_mark");
|
||||
for (node = unit_queue.head; node != NULL; node = node->next) {
|
||||
if (node->unit->iseq) { /* ISeq is still not GCed */
|
||||
VALUE iseq = (VALUE)node->unit->iseq;
|
||||
list_for_each(&unit_queue.head, unit, unode) {
|
||||
if (unit->iseq) { /* ISeq is still not GCed */
|
||||
VALUE iseq = (VALUE)unit->iseq;
|
||||
CRITICAL_SECTION_FINISH(4, "mjit_mark rb_gc_mark");
|
||||
|
||||
/* Don't wrap critical section with this. This may trigger GC,
|
||||
|
|
129
mjit_worker.c
129
mjit_worker.c
|
@ -139,18 +139,12 @@ struct rb_mjit_unit {
|
|||
#endif
|
||||
/* Only used by unload_units. Flag to check this unit is currently on stack or not. */
|
||||
char used_code_p;
|
||||
};
|
||||
|
||||
/* Node of linked list in struct rb_mjit_unit_list.
|
||||
TODO: use ccan/list for this */
|
||||
struct rb_mjit_unit_node {
|
||||
struct rb_mjit_unit *unit;
|
||||
struct rb_mjit_unit_node *next, *prev;
|
||||
struct list_node unode;
|
||||
};
|
||||
|
||||
/* Linked list of struct rb_mjit_unit. */
|
||||
struct rb_mjit_unit_list {
|
||||
struct rb_mjit_unit_node *head;
|
||||
struct list_head head;
|
||||
int length; /* the list length */
|
||||
};
|
||||
|
||||
|
@ -181,11 +175,11 @@ int mjit_call_p = FALSE;
|
|||
|
||||
/* Priority queue of iseqs waiting for JIT compilation.
|
||||
This variable is a pointer to head unit of the queue. */
|
||||
static struct rb_mjit_unit_list unit_queue;
|
||||
static struct rb_mjit_unit_list unit_queue = { LIST_HEAD_INIT(unit_queue.head) };
|
||||
/* List of units which are successfully compiled. */
|
||||
static struct rb_mjit_unit_list active_units;
|
||||
static struct rb_mjit_unit_list active_units = { LIST_HEAD_INIT(active_units.head) };
|
||||
/* List of compacted so files which will be deleted in `mjit_finish()`. */
|
||||
static struct rb_mjit_unit_list compact_units;
|
||||
static struct rb_mjit_unit_list compact_units = { LIST_HEAD_INIT(compact_units.head) };
|
||||
/* The number of so far processed ISEQs, used to generate unique id. */
|
||||
static int current_unit_num;
|
||||
/* A mutex for conitionals and critical sections. */
|
||||
|
@ -318,57 +312,20 @@ mjit_warning(const char *format, ...)
|
|||
}
|
||||
}
|
||||
|
||||
/* Allocate struct rb_mjit_unit_node and return it. This MUST NOT be
|
||||
called inside critical section because that causes deadlock. ZALLOC
|
||||
may fire GC and GC hooks mjit_gc_start_hook that starts critical section. */
|
||||
static struct rb_mjit_unit_node *
|
||||
create_list_node(struct rb_mjit_unit *unit)
|
||||
{
|
||||
struct rb_mjit_unit_node *node = calloc(1, sizeof(struct rb_mjit_unit_node)); /* To prevent GC, don't use ZALLOC */
|
||||
if (node == NULL) return NULL;
|
||||
node->unit = unit;
|
||||
return node;
|
||||
}
|
||||
|
||||
/* Add unit node to the tail of doubly linked LIST. It should be not in
|
||||
the list before. */
|
||||
static void
|
||||
add_to_list(struct rb_mjit_unit_node *node, struct rb_mjit_unit_list *list)
|
||||
add_to_list(struct rb_mjit_unit *unit, struct rb_mjit_unit_list *list)
|
||||
{
|
||||
/* Append iseq to list */
|
||||
if (list->head == NULL) {
|
||||
list->head = node;
|
||||
}
|
||||
else {
|
||||
struct rb_mjit_unit_node *tail = list->head;
|
||||
while (tail->next != NULL) {
|
||||
tail = tail->next;
|
||||
}
|
||||
tail->next = node;
|
||||
node->prev = tail;
|
||||
}
|
||||
list_add_tail(&list->head, &unit->unode);
|
||||
list->length++;
|
||||
}
|
||||
|
||||
static void
|
||||
remove_from_list(struct rb_mjit_unit_node *node, struct rb_mjit_unit_list *list)
|
||||
remove_from_list(struct rb_mjit_unit *unit, struct rb_mjit_unit_list *list)
|
||||
{
|
||||
if (node->prev && node->next) {
|
||||
node->prev->next = node->next;
|
||||
node->next->prev = node->prev;
|
||||
}
|
||||
else if (node->prev == NULL && node->next) {
|
||||
list->head = node->next;
|
||||
node->next->prev = NULL;
|
||||
}
|
||||
else if (node->prev && node->next == NULL) {
|
||||
node->prev->next = NULL;
|
||||
}
|
||||
else {
|
||||
list->head = NULL;
|
||||
}
|
||||
list_del(&unit->unode);
|
||||
list->length--;
|
||||
free(node);
|
||||
}
|
||||
|
||||
static void
|
||||
|
@ -497,28 +454,26 @@ mjit_valid_class_serial_p(rb_serial_t class_serial)
|
|||
/* Return the best unit from list. The best is the first
|
||||
high priority unit or the unit whose iseq has the biggest number
|
||||
of calls so far. */
|
||||
static struct rb_mjit_unit_node *
|
||||
static struct rb_mjit_unit *
|
||||
get_from_list(struct rb_mjit_unit_list *list)
|
||||
{
|
||||
struct rb_mjit_unit_node *node, *next, *best = NULL;
|
||||
|
||||
if (list->head == NULL)
|
||||
return NULL;
|
||||
struct rb_mjit_unit *unit = NULL, *next, *best = NULL;
|
||||
|
||||
/* Find iseq with max total_calls */
|
||||
for (node = list->head; node != NULL; node = next) {
|
||||
next = node->next;
|
||||
if (node->unit->iseq == NULL) { /* ISeq is GCed. */
|
||||
free_unit(node->unit);
|
||||
remove_from_list(node, list);
|
||||
list_for_each_safe(&list->head, unit, next, unode) {
|
||||
if (unit->iseq == NULL) { /* ISeq is GCed. */
|
||||
remove_from_list(unit, list);
|
||||
free_unit(unit);
|
||||
continue;
|
||||
}
|
||||
|
||||
if (best == NULL || best->unit->iseq->body->total_calls < node->unit->iseq->body->total_calls) {
|
||||
best = node;
|
||||
if (best == NULL || best->iseq->body->total_calls < unit->iseq->body->total_calls) {
|
||||
best = unit;
|
||||
}
|
||||
}
|
||||
|
||||
if (best) {
|
||||
remove_from_list(best, list);
|
||||
}
|
||||
return best;
|
||||
}
|
||||
|
||||
|
@ -881,8 +836,7 @@ static void
|
|||
compact_all_jit_code(void)
|
||||
{
|
||||
# ifndef _WIN32 /* This requires header transformation but we don't transform header on Windows for now */
|
||||
struct rb_mjit_unit *unit;
|
||||
struct rb_mjit_unit_node *node;
|
||||
struct rb_mjit_unit *unit, *cur = 0;
|
||||
double start_time, end_time;
|
||||
static const char so_ext[] = DLEXT;
|
||||
char so_file[MAXPATHLEN];
|
||||
|
@ -899,8 +853,8 @@ compact_all_jit_code(void)
|
|||
o_files = alloca(sizeof(char *) * (active_units.length + 1));
|
||||
o_files[active_units.length] = NULL;
|
||||
CRITICAL_SECTION_START(3, "in compact_all_jit_code to keep .o files");
|
||||
for (node = active_units.head; node != NULL; node = node->next) {
|
||||
o_files[i] = node->unit->o_file;
|
||||
list_for_each(&active_units.head, cur, unode) {
|
||||
o_files[i] = cur->o_file;
|
||||
i++;
|
||||
}
|
||||
|
||||
|
@ -924,27 +878,25 @@ compact_all_jit_code(void)
|
|||
unit->handle = handle;
|
||||
|
||||
/* lazily dlclose handle (and .so file for win32) on `mjit_finish()`. */
|
||||
node = calloc(1, sizeof(struct rb_mjit_unit_node)); /* To prevent GC, don't use ZALLOC */
|
||||
node->unit = unit;
|
||||
add_to_list(node, &compact_units);
|
||||
add_to_list(unit, &compact_units);
|
||||
|
||||
if (!mjit_opts.save_temps)
|
||||
remove_so_file(so_file, unit);
|
||||
|
||||
CRITICAL_SECTION_START(3, "in compact_all_jit_code to read list");
|
||||
for (node = active_units.head; node != NULL; node = node->next) {
|
||||
list_for_each(&active_units.head, cur, unode) {
|
||||
void *func;
|
||||
char funcname[35]; /* TODO: reconsider `35` */
|
||||
sprintf(funcname, "_mjit%d", node->unit->id);
|
||||
sprintf(funcname, "_mjit%d", cur->id);
|
||||
|
||||
if ((func = dlsym(handle, funcname)) == NULL) {
|
||||
mjit_warning("skipping to reload '%s' from '%s': %s", funcname, so_file, dlerror());
|
||||
continue;
|
||||
}
|
||||
|
||||
if (node->unit->iseq) { /* Check whether GCed or not */
|
||||
if (cur->iseq) { /* Check whether GCed or not */
|
||||
/* Usage of jit_code might be not in a critical section. */
|
||||
MJIT_ATOMIC_SET(node->unit->iseq->body->jit_func, (mjit_func_t)func);
|
||||
MJIT_ATOMIC_SET(cur->iseq->body->jit_func, (mjit_func_t)func);
|
||||
}
|
||||
}
|
||||
CRITICAL_SECTION_FINISH(3, "in compact_all_jit_code to read list");
|
||||
|
@ -1146,14 +1098,8 @@ convert_unit_to_func(struct rb_mjit_unit *unit, struct rb_call_cache *cc_entries
|
|||
remove_so_file(so_file, unit);
|
||||
|
||||
if ((uintptr_t)func > (uintptr_t)LAST_JIT_ISEQ_FUNC) {
|
||||
struct rb_mjit_unit_node *node = create_list_node(unit);
|
||||
if (node == NULL) {
|
||||
mjit_warning("failed to allocate a node to be added to active_units");
|
||||
return (mjit_func_t)NOT_COMPILED_JIT_ISEQ_FUNC;
|
||||
}
|
||||
|
||||
CRITICAL_SECTION_START(3, "end of jit");
|
||||
add_to_list(node, &active_units);
|
||||
add_to_list(unit, &active_units);
|
||||
if (unit->iseq)
|
||||
print_jit_result("success", unit, end_time - start_time, c_file);
|
||||
CRITICAL_SECTION_FINISH(3, "end of jit");
|
||||
|
@ -1216,22 +1162,22 @@ mjit_worker(void)
|
|||
|
||||
/* main worker loop */
|
||||
while (!stop_worker_p) {
|
||||
struct rb_mjit_unit_node *node;
|
||||
struct rb_mjit_unit *unit;
|
||||
|
||||
/* wait until unit is available */
|
||||
CRITICAL_SECTION_START(3, "in worker dequeue");
|
||||
while ((unit_queue.head == NULL || active_units.length >= mjit_opts.max_cache_size) && !stop_worker_p) {
|
||||
while ((list_empty(&unit_queue.head) || active_units.length >= mjit_opts.max_cache_size) && !stop_worker_p) {
|
||||
rb_native_cond_wait(&mjit_worker_wakeup, &mjit_engine_mutex);
|
||||
verbose(3, "Getting wakeup from client");
|
||||
}
|
||||
node = get_from_list(&unit_queue);
|
||||
unit = get_from_list(&unit_queue);
|
||||
CRITICAL_SECTION_FINISH(3, "in worker dequeue");
|
||||
|
||||
if (node) {
|
||||
if (unit) {
|
||||
mjit_func_t func;
|
||||
struct mjit_copy_job job;
|
||||
|
||||
job.body = node->unit->iseq->body;
|
||||
job.body = unit->iseq->body;
|
||||
job.cc_entries = NULL;
|
||||
if (job.body->ci_size > 0 || job.body->ci_kw_size > 0)
|
||||
job.cc_entries = alloca(sizeof(struct rb_call_cache) * (job.body->ci_size + job.body->ci_kw_size));
|
||||
|
@ -1250,14 +1196,13 @@ mjit_worker(void)
|
|||
}
|
||||
|
||||
/* JIT compile */
|
||||
func = convert_unit_to_func(node->unit, job.cc_entries, job.is_entries);
|
||||
func = convert_unit_to_func(unit, job.cc_entries, job.is_entries);
|
||||
|
||||
CRITICAL_SECTION_START(3, "in jit func replace");
|
||||
if (node->unit->iseq) { /* Check whether GCed or not */
|
||||
if (unit->iseq) { /* Check whether GCed or not */
|
||||
/* Usage of jit_code might be not in a critical section. */
|
||||
MJIT_ATOMIC_SET(node->unit->iseq->body->jit_func, func);
|
||||
MJIT_ATOMIC_SET(unit->iseq->body->jit_func, func);
|
||||
}
|
||||
remove_from_list(node, &unit_queue);
|
||||
CRITICAL_SECTION_FINISH(3, "in jit func replace");
|
||||
|
||||
#ifndef _MSC_VER
|
||||
|
|
Loading…
Reference in a new issue