mirror of
				https://github.com/ruby/ruby.git
				synced 2022-11-09 12:17:21 -05:00 
			
		
		
		
	mjit.c: copy call cache values to MJIT worker
same as r65275 but for call cache.
=== Optcarrot Benchmark ===
$ benchmark-driver benchmark.yml --rbenv 'before::before --disable-gems --jit;after::after --disable-gems --jit' -v --repeat-count 24
before: ruby 2.6.0dev (2018-10-21 trunk 65277) +JIT [x86_64-linux]
after: ruby 2.6.0dev (2018-10-21 trunk 65277) +JIT [x86_64-linux]
last_commit=mjit.c: copy call cache values to MJIT worker
Calculating -------------------------------------
                             before       after
Optcarrot Lan_Master.nes     85.372      85.359 fps
Comparison:
             Optcarrot Lan_Master.nes
                  before:        85.4 fps
                   after:        85.4 fps - 1.00x  slower
git-svn-id: svn+ssh://ci.ruby-lang.org/ruby/trunk@65279 b2dd03c8-39d4-4d8f-98ff-823fe69b080e
			
			
This commit is contained in:
		
							parent
							
								
									bca2cdcc0b
								
							
						
					
					
						commit
						13df05acfa
					
				
					 5 changed files with 48 additions and 27 deletions
				
			
		
							
								
								
									
										7
									
								
								mjit.c
									
										
									
									
									
								
							
							
						
						
									
										7
									
								
								mjit.c
									
										
									
									
									
								
							|  | @ -25,7 +25,12 @@ static void | ||||||
| mjit_copy_job_handler(void *data) | mjit_copy_job_handler(void *data) | ||||||
| { | { | ||||||
|     struct mjit_copy_job *job = (struct mjit_copy_job *)data; |     struct mjit_copy_job *job = (struct mjit_copy_job *)data; | ||||||
|     memcpy(job->is_entries, job->body->is_entries, sizeof(union iseq_inline_storage_entry) * job->body->is_size); |     if (job->cc_entries) { | ||||||
|  |         memcpy(job->cc_entries, job->body->cc_entries, sizeof(struct rb_call_cache) * (job->body->ci_size + job->body->ci_kw_size)); | ||||||
|  |     } | ||||||
|  |     if (job->is_entries) { | ||||||
|  |         memcpy(job->is_entries, job->body->is_entries, sizeof(union iseq_inline_storage_entry) * job->body->is_size); | ||||||
|  |     } | ||||||
| 
 | 
 | ||||||
|     CRITICAL_SECTION_START(3, "in MJIT copy job wait"); |     CRITICAL_SECTION_START(3, "in MJIT copy job wait"); | ||||||
|     job->finish_p = TRUE; |     job->finish_p = TRUE; | ||||||
|  |  | ||||||
							
								
								
									
										2
									
								
								mjit.h
									
										
									
									
									
								
							
							
						
						
									
										2
									
								
								mjit.h
									
										
									
									
									
								
							|  | @ -64,7 +64,7 @@ extern void mjit_add_iseq_to_process(const rb_iseq_t *iseq); | ||||||
| extern VALUE mjit_wait_call(rb_execution_context_t *ec, struct rb_iseq_constant_body *body); | extern VALUE mjit_wait_call(rb_execution_context_t *ec, struct rb_iseq_constant_body *body); | ||||||
| RUBY_SYMBOL_EXPORT_END | RUBY_SYMBOL_EXPORT_END | ||||||
| 
 | 
 | ||||||
| extern int mjit_compile(FILE *f, const struct rb_iseq_constant_body *body, const char *funcname, union iseq_inline_storage_entry *is_entries); | extern int mjit_compile(FILE *f, const struct rb_iseq_constant_body *body, const char *funcname, struct rb_call_cache *cc_entries, union iseq_inline_storage_entry *is_entries); | ||||||
| extern void mjit_init(struct mjit_options *opts); | extern void mjit_init(struct mjit_options *opts); | ||||||
| extern void mjit_finish(void); | extern void mjit_finish(void); | ||||||
| extern void mjit_gc_start_hook(void); | extern void mjit_gc_start_hook(void); | ||||||
|  |  | ||||||
|  | @ -34,8 +34,9 @@ struct compile_status { | ||||||
|     /* If TRUE, JIT-ed code will use local variables to store pushed values instead of
 |     /* If TRUE, JIT-ed code will use local variables to store pushed values instead of
 | ||||||
|        using VM's stack and moving stack pointer. */ |        using VM's stack and moving stack pointer. */ | ||||||
|     int local_stack_p; |     int local_stack_p; | ||||||
|     /* Safely-accessible is_entries copied from main thread. */ |     /* Safely-accessible cache entries copied from main thread. */ | ||||||
|     union iseq_inline_storage_entry *is_entries; |     union iseq_inline_storage_entry *is_entries; | ||||||
|  |     struct rb_call_cache *cc_entries; | ||||||
| }; | }; | ||||||
| 
 | 
 | ||||||
| /* Storage to keep data which is consistent in each conditional branch.
 | /* Storage to keep data which is consistent in each conditional branch.
 | ||||||
|  | @ -197,7 +198,7 @@ compile_cancel_handler(FILE *f, const struct rb_iseq_constant_body *body, struct | ||||||
| 
 | 
 | ||||||
| /* Compile ISeq to C code in F.  It returns 1 if it succeeds to compile. */ | /* Compile ISeq to C code in F.  It returns 1 if it succeeds to compile. */ | ||||||
| int | int | ||||||
| mjit_compile(FILE *f, const struct rb_iseq_constant_body *body, const char *funcname, union iseq_inline_storage_entry *is_entries) | mjit_compile(FILE *f, const struct rb_iseq_constant_body *body, const char *funcname, struct rb_call_cache *cc_entries, union iseq_inline_storage_entry *is_entries) | ||||||
| { | { | ||||||
|     struct compile_status status; |     struct compile_status status; | ||||||
|     status.success = TRUE; |     status.success = TRUE; | ||||||
|  | @ -206,6 +207,7 @@ mjit_compile(FILE *f, const struct rb_iseq_constant_body *body, const char *func | ||||||
|     if (status.stack_size_for_pos == NULL) |     if (status.stack_size_for_pos == NULL) | ||||||
|         return FALSE; |         return FALSE; | ||||||
|     memset(status.stack_size_for_pos, NOT_COMPILED_STACK_SIZE, sizeof(int) * body->iseq_size); |     memset(status.stack_size_for_pos, NOT_COMPILED_STACK_SIZE, sizeof(int) * body->iseq_size); | ||||||
|  |     status.cc_entries = cc_entries; | ||||||
|     status.is_entries = is_entries; |     status.is_entries = is_entries; | ||||||
| 
 | 
 | ||||||
|     /* For performance, we verify stack size only on compilation time (mjit_compile.inc.erb) without --jit-debug */ |     /* For performance, we verify stack size only on compilation time (mjit_compile.inc.erb) without --jit-debug */ | ||||||
|  |  | ||||||
|  | @ -1029,7 +1029,7 @@ compile_prelude(FILE *f) | ||||||
| /* Compile ISeq in UNIT and return function pointer of JIT-ed code.
 | /* Compile ISeq in UNIT and return function pointer of JIT-ed code.
 | ||||||
|    It may return NOT_COMPILED_JIT_ISEQ_FUNC if something went wrong. */ |    It may return NOT_COMPILED_JIT_ISEQ_FUNC if something went wrong. */ | ||||||
| static mjit_func_t | static mjit_func_t | ||||||
| convert_unit_to_func(struct rb_mjit_unit *unit, union iseq_inline_storage_entry *is_entries) | convert_unit_to_func(struct rb_mjit_unit *unit, struct rb_call_cache *cc_entries, union iseq_inline_storage_entry *is_entries) | ||||||
| { | { | ||||||
|     char c_file_buff[MAXPATHLEN], *c_file = c_file_buff, *so_file, funcname[35]; /* TODO: reconsider `35` */ |     char c_file_buff[MAXPATHLEN], *c_file = c_file_buff, *so_file, funcname[35]; /* TODO: reconsider `35` */ | ||||||
|     int success; |     int success; | ||||||
|  | @ -1097,7 +1097,7 @@ convert_unit_to_func(struct rb_mjit_unit *unit, union iseq_inline_storage_entry | ||||||
|         verbose(2, "start compilation: %s@%s:%d -> %s", label, path, lineno, c_file); |         verbose(2, "start compilation: %s@%s:%d -> %s", label, path, lineno, c_file); | ||||||
|         fprintf(f, "/* %s@%s:%d */\n\n", label, path, lineno); |         fprintf(f, "/* %s@%s:%d */\n\n", label, path, lineno); | ||||||
|     } |     } | ||||||
|     success = mjit_compile(f, unit->iseq->body, funcname, is_entries); |     success = mjit_compile(f, unit->iseq->body, funcname, cc_entries, is_entries); | ||||||
| 
 | 
 | ||||||
|     /* release blocking mjit_gc_start_hook */ |     /* release blocking mjit_gc_start_hook */ | ||||||
|     CRITICAL_SECTION_START(3, "after mjit_compile to wakeup client for GC"); |     CRITICAL_SECTION_START(3, "after mjit_compile to wakeup client for GC"); | ||||||
|  | @ -1163,12 +1163,29 @@ convert_unit_to_func(struct rb_mjit_unit *unit, union iseq_inline_storage_entry | ||||||
| 
 | 
 | ||||||
| struct mjit_copy_job { | struct mjit_copy_job { | ||||||
|     const struct rb_iseq_constant_body *body; |     const struct rb_iseq_constant_body *body; | ||||||
|  |     struct rb_call_cache *cc_entries; | ||||||
|     union iseq_inline_storage_entry *is_entries; |     union iseq_inline_storage_entry *is_entries; | ||||||
|     int finish_p; |     int finish_p; | ||||||
| }; | }; | ||||||
| 
 | 
 | ||||||
| static void mjit_copy_job_handler(void *data); | static void mjit_copy_job_handler(void *data); | ||||||
| 
 | 
 | ||||||
|  | /* We're lazily copying cache values from main thread because these cache values
 | ||||||
|  |    could be different between ones on enqueue timing and ones on dequeue timing. */ | ||||||
|  | static void | ||||||
|  | copy_cache_from_main_thread(struct mjit_copy_job *job) | ||||||
|  | { | ||||||
|  |     job->finish_p = FALSE; | ||||||
|  | 
 | ||||||
|  |     rb_postponed_job_register(0, mjit_copy_job_handler, (void *)job); | ||||||
|  |     CRITICAL_SECTION_START(3, "in MJIT copy job wait"); | ||||||
|  |     while (!job->finish_p) { | ||||||
|  |         rb_native_cond_wait(&mjit_worker_wakeup, &mjit_engine_mutex); | ||||||
|  |         verbose(3, "Getting wakeup from client"); | ||||||
|  |     } | ||||||
|  |     CRITICAL_SECTION_FINISH(3, "in MJIT copy job wait"); | ||||||
|  | } | ||||||
|  | 
 | ||||||
| /* The function implementing a worker. It is executed in a separate
 | /* The function implementing a worker. It is executed in a separate
 | ||||||
|    thread by rb_thread_create_mjit_thread. It compiles precompiled header |    thread by rb_thread_create_mjit_thread. It compiles precompiled header | ||||||
|    and then compiles requested ISeqs. */ |    and then compiles requested ISeqs. */ | ||||||
|  | @ -1207,24 +1224,21 @@ mjit_worker(void) | ||||||
|             mjit_func_t func; |             mjit_func_t func; | ||||||
|             struct mjit_copy_job job; |             struct mjit_copy_job job; | ||||||
| 
 | 
 | ||||||
|             /* Copy ISeq's inline caches from main thread. */ |  | ||||||
|             job.is_entries = NULL; |  | ||||||
|             job.body = node->unit->iseq->body; |             job.body = node->unit->iseq->body; | ||||||
|             if (job.body->is_size > 0) { |             job.cc_entries = NULL; | ||||||
|                 job.is_entries = malloc(sizeof(union iseq_inline_storage_entry) * job.body->is_size); |             if (job.body->ci_size > 0 || job.body->ci_kw_size > 0) | ||||||
|                 job.finish_p = FALSE; |                 job.cc_entries = alloca(sizeof(struct rb_call_cache) * (job.body->ci_size + job.body->ci_kw_size)); | ||||||
|  |             job.is_entries = NULL; | ||||||
|  |             if (job.body->is_size > 0) | ||||||
|  |                 job.is_entries = alloca(sizeof(union iseq_inline_storage_entry) * job.body->is_size); | ||||||
| 
 | 
 | ||||||
|                 rb_postponed_job_register(0, mjit_copy_job_handler, (void *)&job); |             /* Copy ISeq's inline caches values to avoid race condition. */ | ||||||
|                 CRITICAL_SECTION_START(3, "in MJIT copy job wait"); |             if (job.cc_entries != NULL || job.is_entries != NULL) { | ||||||
|                 while (!job.finish_p) { |                 copy_cache_from_main_thread(&job); | ||||||
|                     rb_native_cond_wait(&mjit_worker_wakeup, &mjit_engine_mutex); |  | ||||||
|                     verbose(3, "Getting wakeup from client"); |  | ||||||
|                 } |  | ||||||
|                 CRITICAL_SECTION_FINISH(3, "in MJIT copy job wait"); |  | ||||||
|             } |             } | ||||||
| 
 | 
 | ||||||
|             /* JIT compile */ |             /* JIT compile */ | ||||||
|             func = convert_unit_to_func(node->unit, job.is_entries); |             func = convert_unit_to_func(node->unit, job.cc_entries, job.is_entries); | ||||||
| 
 | 
 | ||||||
|             CRITICAL_SECTION_START(3, "in jit func replace"); |             CRITICAL_SECTION_START(3, "in jit func replace"); | ||||||
|             if (node->unit->iseq) { /* Check whether GCed or not */ |             if (node->unit->iseq) { /* Check whether GCed or not */ | ||||||
|  | @ -1241,9 +1255,6 @@ mjit_worker(void) | ||||||
|                 compact_all_jit_code(); |                 compact_all_jit_code(); | ||||||
|             } |             } | ||||||
| #endif | #endif | ||||||
|             if (job.is_entries != NULL) { |  | ||||||
|                 free(job.is_entries); |  | ||||||
|             } |  | ||||||
|         } |         } | ||||||
|     } |     } | ||||||
| 
 | 
 | ||||||
|  |  | ||||||
|  | @ -13,6 +13,9 @@ | ||||||
| % insn.opes.each_with_index do |ope, i| | % insn.opes.each_with_index do |ope, i| | ||||||
|     MAYBE_UNUSED(<%= ope.fetch(:decl) %>) = (<%= ope.fetch(:type) %>)operands[<%= i %>]; |     MAYBE_UNUSED(<%= ope.fetch(:decl) %>) = (<%= ope.fetch(:type) %>)operands[<%= i %>]; | ||||||
| % end | % end | ||||||
|  | % # compiler: Use copied cc to avoid race condition | ||||||
|  |     CALL_CACHE cc_copy = status->cc_entries + (cc - body->cc_entries); | ||||||
|  |     cc_copy = cc; | ||||||
| % | % | ||||||
|     if (has_valid_method_type(cc)) { |     if (has_valid_method_type(cc)) { | ||||||
|         const rb_iseq_t *iseq; |         const rb_iseq_t *iseq; | ||||||
|  | @ -21,7 +24,7 @@ | ||||||
|         argc += ((ci->flag & VM_CALL_ARGS_BLOCKARG) ? 1 : 0); |         argc += ((ci->flag & VM_CALL_ARGS_BLOCKARG) ? 1 : 0); | ||||||
| % end | % end | ||||||
| 
 | 
 | ||||||
|         if (cc->me->def->type == VM_METHOD_TYPE_ISEQ && inlinable_iseq_p(ci, cc, iseq = rb_iseq_check(cc->me->def->body.iseq.iseqptr))) { /* CC_SET_FASTPATH in vm_callee_setup_arg */ |         if (cc_copy->me->def->type == VM_METHOD_TYPE_ISEQ && inlinable_iseq_p(ci, cc, iseq = rb_iseq_check(cc_copy->me->def->body.iseq.iseqptr))) { /* CC_SET_FASTPATH in vm_callee_setup_arg */ | ||||||
|             int param_size = iseq->body->param.size; /* TODO: check calling->argc for argument_arity_error */ |             int param_size = iseq->body->param.size; /* TODO: check calling->argc for argument_arity_error */ | ||||||
| 
 | 
 | ||||||
|             fprintf(f, "{\n"); |             fprintf(f, "{\n"); | ||||||
|  | @ -31,8 +34,8 @@ | ||||||
|             } |             } | ||||||
| 
 | 
 | ||||||
| % # JIT: Invalidate call cache if it requires vm_search_method. This allows to inline some of following things. | % # JIT: Invalidate call cache if it requires vm_search_method. This allows to inline some of following things. | ||||||
|             fprintf(f, "    if (UNLIKELY(GET_GLOBAL_METHOD_STATE() != %"PRI_SERIALT_PREFIX"u ||\n", cc->method_state); |             fprintf(f, "    if (UNLIKELY(GET_GLOBAL_METHOD_STATE() != %"PRI_SERIALT_PREFIX"u ||\n", cc_copy->method_state); | ||||||
|             fprintf(f, "        RCLASS_SERIAL(CLASS_OF(stack[%d])) != %"PRI_SERIALT_PREFIX"u)) {\n", b->stack_size - 1 - argc, cc->class_serial); |             fprintf(f, "        RCLASS_SERIAL(CLASS_OF(stack[%d])) != %"PRI_SERIALT_PREFIX"u)) {\n", b->stack_size - 1 - argc, cc_copy->class_serial); | ||||||
|             fprintf(f, "        reg_cfp->pc = original_body_iseq + %d;\n", pos); |             fprintf(f, "        reg_cfp->pc = original_body_iseq + %d;\n", pos); | ||||||
|             fprintf(f, "        reg_cfp->sp = (VALUE *)reg_cfp->bp + %d;\n", b->stack_size + 1); |             fprintf(f, "        reg_cfp->sp = (VALUE *)reg_cfp->bp + %d;\n", b->stack_size + 1); | ||||||
|             fprintf(f, "        goto cancel;\n"); |             fprintf(f, "        goto cancel;\n"); | ||||||
|  | @ -52,11 +55,11 @@ | ||||||
|             fprintf(f, "        calling.argc = %d;\n", ci->orig_argc); |             fprintf(f, "        calling.argc = %d;\n", ci->orig_argc); | ||||||
|             fprintf(f, "        calling.recv = stack[%d];\n", b->stack_size - 1 - argc); |             fprintf(f, "        calling.recv = stack[%d];\n", b->stack_size - 1 - argc); | ||||||
| 
 | 
 | ||||||
| % # JIT: Special CALL_METHOD. Bypass cc->call and inline vm_call_iseq_setup_normal for vm_call_iseq_setup_func FASTPATH. | % # JIT: Special CALL_METHOD. Bypass cc_copy->call and inline vm_call_iseq_setup_normal for vm_call_iseq_setup_func FASTPATH. | ||||||
|             fprintf(f, "        {\n"); |             fprintf(f, "        {\n"); | ||||||
|             fprintf(f, "            VALUE v;\n"); |             fprintf(f, "            VALUE v;\n"); | ||||||
|             fprintf(f, "            vm_call_iseq_setup_normal(ec, reg_cfp, &calling, (const rb_callable_method_entry_t *)0x%"PRIxVALUE", 0, %d, %d);\n", |             fprintf(f, "            vm_call_iseq_setup_normal(ec, reg_cfp, &calling, (const rb_callable_method_entry_t *)0x%"PRIxVALUE", 0, %d, %d);\n", | ||||||
|                     (VALUE)cc->me, param_size, iseq->body->local_table_size); /* rb_simple_iseq_p checks rb_simple_iseq_p, which ensures has_opt == FALSE */ |                     (VALUE)cc_copy->me, param_size, iseq->body->local_table_size); /* rb_simple_iseq_p checks rb_simple_iseq_p, which ensures has_opt == FALSE */ | ||||||
|             if (iseq->body->catch_except_p) { |             if (iseq->body->catch_except_p) { | ||||||
|                 fprintf(f, "            VM_ENV_FLAGS_SET(ec->cfp->ep, VM_FRAME_FLAG_FINISH);\n"); |                 fprintf(f, "            VM_ENV_FLAGS_SET(ec->cfp->ep, VM_FRAME_FLAG_FINISH);\n"); | ||||||
|                 fprintf(f, "            v = vm_exec(ec, TRUE);\n"); |                 fprintf(f, "            v = vm_exec(ec, TRUE);\n"); | ||||||
|  |  | ||||||
		Loading…
	
	Add table
		Add a link
		
	
		Reference in a new issue
	
	 k0kubun
						k0kubun