mirror of
https://github.com/ruby/ruby.git
synced 2022-11-09 12:17:21 -05:00
Fix typos [ci skip]
This commit is contained in:
parent
d100c91445
commit
04f07713d1
7 changed files with 12 additions and 12 deletions
|
@ -52,7 +52,7 @@ rb_fiber_scheduler_io_result(ssize_t result, int error) {
|
|||
}
|
||||
|
||||
/**
|
||||
* Apply an io result to the local thread, returning the value of the orginal
|
||||
* Apply an io result to the local thread, returning the value of the original
|
||||
* system call that created it and updating `int errno`.
|
||||
*
|
||||
* You should not directly try to interpret the result value as it is considered
|
||||
|
|
|
@ -1843,7 +1843,7 @@ io_buffer_clear(int argc, VALUE *argv, VALUE self)
|
|||
|
||||
static
|
||||
size_t io_buffer_default_size(size_t page_size) {
|
||||
// Platform agnostic default size, based on emperical performance observation:
|
||||
// Platform agnostic default size, based on empirical performance observation:
|
||||
const size_t platform_agnostic_default_size = 64*1024;
|
||||
|
||||
// Allow user to specify custom default buffer size:
|
||||
|
|
|
@ -1456,7 +1456,7 @@ symname_equal(VALUE sym, const char *name, size_t nlen)
|
|||
}
|
||||
|
||||
#define BUILD_ASSERT_POSITIVE(n) \
|
||||
/* make 0 negative to workaround the "zero size array" GCC extention, */ \
|
||||
/* make 0 negative to workaround the "zero size array" GCC extension, */ \
|
||||
((sizeof(char [2*(ssize_t)(n)-1])+1)/2) /* assuming no overflow */
|
||||
#define symname_equal_lit(sym, sym_name) \
|
||||
symname_equal(sym, sym_name, BUILD_ASSERT_POSITIVE(rb_strlen_lit(sym_name)))
|
||||
|
|
2
struct.c
2
struct.c
|
@ -191,7 +191,7 @@ struct_member_pos(VALUE s, VALUE name)
|
|||
* call-seq:
|
||||
* StructClass::members -> array_of_symbols
|
||||
*
|
||||
* Returns the member names of the Struct decendant as an array:
|
||||
* Returns the member names of the Struct descendant as an array:
|
||||
*
|
||||
* Customer = Struct.new(:name, :address, :zip)
|
||||
* Customer.members # => [:name, :address, :zip]
|
||||
|
|
|
@ -1499,7 +1499,7 @@ rb_init_iv_list(VALUE obj)
|
|||
init_iv_list(obj, len, newsize, index_tbl);
|
||||
}
|
||||
|
||||
// Retreive or create the id-to-index mapping for a given object and an
|
||||
// Retrieve or create the id-to-index mapping for a given object and an
|
||||
// instance variable name.
|
||||
static struct ivar_update
|
||||
obj_ensure_iv_index_mapping(VALUE obj, ID id)
|
||||
|
|
|
@ -1344,7 +1344,7 @@ void mov(codeblock_t *cb, x86opnd_t dst, x86opnd_t src)
|
|||
cb_write_rm(cb, dst.num_bits == 16, dst.num_bits == 64, NO_OPND, dst, 0, 1, 0xC7);
|
||||
|
||||
const uint32_t output_num_bits = (dst.num_bits > 32u) ? 32u : dst.num_bits;
|
||||
// assert that we can write whole immediate without loss of infomation
|
||||
// assert that we can write whole immediate without loss of information
|
||||
assert (sig_imm_size(src.as.imm) <= output_num_bits);
|
||||
cb_write_int(cb, src.as.imm, output_num_bits);
|
||||
}
|
||||
|
|
|
@ -1731,7 +1731,7 @@ gen_get_ivar(jitstate_t *jit, ctx_t *ctx, const int max_chain_depth, VALUE compt
|
|||
mov(cb, out_opnd, REG1);
|
||||
}
|
||||
else {
|
||||
// Compile time value is *not* embeded.
|
||||
// Compile time value is *not* embedded.
|
||||
|
||||
// Guard that value is *not* embedded
|
||||
// See ROBJECT_IVPTR() from include/ruby/internal/core/robject.h
|
||||
|
@ -2274,7 +2274,7 @@ gen_opt_aref(jitstate_t *jit, ctx_t *ctx, codeblock_t *cb)
|
|||
|
||||
call_ptr(cb, REG0, (void *)rb_hash_aref);
|
||||
|
||||
// Pop the key and the reciever
|
||||
// Pop the key and the receiver
|
||||
(void)ctx_stack_pop(ctx, 2);
|
||||
|
||||
// Push the return value onto the stack
|
||||
|
@ -2682,7 +2682,7 @@ gen_opt_case_dispatch(jitstate_t *jit, ctx_t *ctx, codeblock_t *cb)
|
|||
// Normally this instruction would lookup the key in a hash and jump to an
|
||||
// offset based on that.
|
||||
// Instead we can take the fallback case and continue with the next
|
||||
// instruciton.
|
||||
// instruction.
|
||||
// We'd hope that our jitted code will be sufficiently fast without the
|
||||
// hash lookup, at least for small hashes, but it's worth revisiting this
|
||||
// assumption in the future.
|
||||
|
@ -2982,7 +2982,7 @@ jit_guard_known_klass(jitstate_t *jit, ctx_t *ctx, VALUE known_klass, insn_opnd_
|
|||
// Note that in case the sample instance has a singleton class that
|
||||
// doesn't attach to the sample instance, it means the sample instance
|
||||
// has an empty singleton class that hasn't been materialized yet. In
|
||||
// this case, comparing against the sample instance doesn't gurantee
|
||||
// this case, comparing against the sample instance doesn't guarantee
|
||||
// that its singleton class is empty, so we can't avoid the memory
|
||||
// access. As an example, `Object.new.singleton_class` is an object in
|
||||
// this situation.
|
||||
|
@ -4697,9 +4697,9 @@ gen_opt_getinlinecache(jitstate_t *jit, ctx_t *ctx, codeblock_t *cb)
|
|||
return YJIT_END_BLOCK;
|
||||
}
|
||||
|
||||
// Push the explict block parameter onto the temporary stack. Part of the
|
||||
// Push the explicit block parameter onto the temporary stack. Part of the
|
||||
// interpreter's scheme for avoiding Proc allocations when delegating
|
||||
// explict block parameters.
|
||||
// explicit block parameters.
|
||||
static codegen_status_t
|
||||
gen_getblockparamproxy(jitstate_t *jit, ctx_t *ctx, codeblock_t *cb)
|
||||
{
|
||||
|
|
Loading…
Add table
Reference in a new issue