12#include "ruby/internal/config.h"
22#include "internal/array.h"
23#include "internal/compile.h"
24#include "internal/complex.h"
25#include "internal/encoding.h"
26#include "internal/error.h"
27#include "internal/gc.h"
28#include "internal/hash.h"
29#include "internal/io.h"
30#include "internal/numeric.h"
31#include "internal/object.h"
32#include "internal/rational.h"
33#include "internal/re.h"
34#include "internal/ruby_parser.h"
35#include "internal/symbol.h"
36#include "internal/thread.h"
37#include "internal/variable.h"
43#include "vm_callinfo.h"
49#include "insns_info.inc"
51#define FIXNUM_INC(n, i) ((n)+(INT2FIX(i)&~FIXNUM_FLAG))
85 unsigned int rescued: 2;
86 unsigned int unremovable: 1;
91 enum ruby_vminsn_type insn_id;
121 const void *ensure_node;
126const ID rb_iseq_shared_exc_local_tbl[] = {idERROR_INFO};
146#define compile_debug CPDEBUG
148#define compile_debug ISEQ_COMPILE_DATA(iseq)->option->debug_level
153#define compile_debug_print_indent(level) \
154 ruby_debug_print_indent((level), compile_debug, gl_node_level * 2)
156#define debugp(header, value) (void) \
157 (compile_debug_print_indent(1) && \
158 ruby_debug_print_value(1, compile_debug, (header), (value)))
160#define debugi(header, id) (void) \
161 (compile_debug_print_indent(1) && \
162 ruby_debug_print_id(1, compile_debug, (header), (id)))
164#define debugp_param(header, value) (void) \
165 (compile_debug_print_indent(1) && \
166 ruby_debug_print_value(1, compile_debug, (header), (value)))
168#define debugp_verbose(header, value) (void) \
169 (compile_debug_print_indent(2) && \
170 ruby_debug_print_value(2, compile_debug, (header), (value)))
172#define debugp_verbose_node(header, value) (void) \
173 (compile_debug_print_indent(10) && \
174 ruby_debug_print_value(10, compile_debug, (header), (value)))
176#define debug_node_start(node) ((void) \
177 (compile_debug_print_indent(1) && \
178 (ruby_debug_print_node(1, CPDEBUG, "", (const NODE *)(node)), gl_node_level)), \
181#define debug_node_end() gl_node_level --
185#define debugi(header, id) ((void)0)
186#define debugp(header, value) ((void)0)
187#define debugp_verbose(header, value) ((void)0)
188#define debugp_verbose_node(header, value) ((void)0)
189#define debugp_param(header, value) ((void)0)
190#define debug_node_start(node) ((void)0)
191#define debug_node_end() ((void)0)
194#if CPDEBUG > 1 || CPDEBUG < 0
196#define printf ruby_debug_printf
197#define debugs if (compile_debug_print_indent(1)) ruby_debug_printf
198#define debug_compile(msg, v) ((void)(compile_debug_print_indent(1) && fputs((msg), stderr)), (v))
200#define debugs if(0)printf
201#define debug_compile(msg, v) (v)
204#define LVAR_ERRINFO (1)
207#define NEW_LABEL(l) new_label_body(iseq, (l))
208#define LABEL_FORMAT "<L%03d>"
210#define NEW_ISEQ(node, name, type, line_no) \
211 new_child_iseq(iseq, (node), rb_fstring(name), 0, (type), (line_no))
213#define NEW_CHILD_ISEQ(node, name, type, line_no) \
214 new_child_iseq(iseq, (node), rb_fstring(name), iseq, (type), (line_no))
216#define NEW_CHILD_ISEQ_WITH_CALLBACK(callback_func, name, type, line_no) \
217 new_child_iseq_with_callback(iseq, (callback_func), (name), iseq, (type), (line_no))
220#define ADD_SEQ(seq1, seq2) \
221 APPEND_LIST((seq1), (seq2))
224#define ADD_INSN(seq, line_node, insn) \
225 ADD_ELEM((seq), (LINK_ELEMENT *) new_insn_body(iseq, nd_line(line_node), nd_node_id(line_node), BIN(insn), 0))
228#define ADD_SYNTHETIC_INSN(seq, line_no, node_id, insn) \
229 ADD_ELEM((seq), (LINK_ELEMENT *) new_insn_body(iseq, (line_no), (node_id), BIN(insn), 0))
232#define INSERT_BEFORE_INSN(next, line_no, node_id, insn) \
233 ELEM_INSERT_PREV(&(next)->link, (LINK_ELEMENT *) new_insn_body(iseq, line_no, node_id, BIN(insn), 0))
236#define INSERT_AFTER_INSN(prev, line_no, node_id, insn) \
237 ELEM_INSERT_NEXT(&(prev)->link, (LINK_ELEMENT *) new_insn_body(iseq, line_no, node_id, BIN(insn), 0))
240#define ADD_INSN1(seq, line_node, insn, op1) \
241 ADD_ELEM((seq), (LINK_ELEMENT *) \
242 new_insn_body(iseq, nd_line(line_node), nd_node_id(line_node), BIN(insn), 1, (VALUE)(op1)))
245#define INSERT_BEFORE_INSN1(next, line_no, node_id, insn, op1) \
246 ELEM_INSERT_PREV(&(next)->link, (LINK_ELEMENT *) \
247 new_insn_body(iseq, line_no, node_id, BIN(insn), 1, (VALUE)(op1)))
250#define INSERT_AFTER_INSN1(prev, line_no, node_id, insn, op1) \
251 ELEM_INSERT_NEXT(&(prev)->link, (LINK_ELEMENT *) \
252 new_insn_body(iseq, line_no, node_id, BIN(insn), 1, (VALUE)(op1)))
254#define LABEL_REF(label) ((label)->refcnt++)
257#define ADD_INSNL(seq, line_node, insn, label) (ADD_INSN1(seq, line_node, insn, label), LABEL_REF(label))
259#define ADD_INSN2(seq, line_node, insn, op1, op2) \
260 ADD_ELEM((seq), (LINK_ELEMENT *) \
261 new_insn_body(iseq, nd_line(line_node), nd_node_id(line_node), BIN(insn), 2, (VALUE)(op1), (VALUE)(op2)))
263#define ADD_INSN3(seq, line_node, insn, op1, op2, op3) \
264 ADD_ELEM((seq), (LINK_ELEMENT *) \
265 new_insn_body(iseq, nd_line(line_node), nd_node_id(line_node), BIN(insn), 3, (VALUE)(op1), (VALUE)(op2), (VALUE)(op3)))
268#define ADD_SEND(seq, line_node, id, argc) \
269 ADD_SEND_R((seq), (line_node), (id), (argc), NULL, (VALUE)INT2FIX(0), NULL)
271#define ADD_SEND_WITH_FLAG(seq, line_node, id, argc, flag) \
272 ADD_SEND_R((seq), (line_node), (id), (argc), NULL, (VALUE)(flag), NULL)
274#define ADD_SEND_WITH_BLOCK(seq, line_node, id, argc, block) \
275 ADD_SEND_R((seq), (line_node), (id), (argc), (block), (VALUE)INT2FIX(0), NULL)
277#define ADD_CALL_RECEIVER(seq, line_node) \
278 ADD_INSN((seq), (line_node), putself)
280#define ADD_CALL(seq, line_node, id, argc) \
281 ADD_SEND_R((seq), (line_node), (id), (argc), NULL, (VALUE)INT2FIX(VM_CALL_FCALL), NULL)
283#define ADD_CALL_WITH_BLOCK(seq, line_node, id, argc, block) \
284 ADD_SEND_R((seq), (line_node), (id), (argc), (block), (VALUE)INT2FIX(VM_CALL_FCALL), NULL)
286#define ADD_SEND_R(seq, line_node, id, argc, block, flag, keywords) \
287 ADD_ELEM((seq), (LINK_ELEMENT *) new_insn_send(iseq, nd_line(line_node), nd_node_id(line_node), (id), (VALUE)(argc), (block), (VALUE)(flag), (keywords)))
289#define ADD_TRACE(seq, event) \
290 ADD_ELEM((seq), (LINK_ELEMENT *)new_trace_body(iseq, (event), 0))
291#define ADD_TRACE_WITH_DATA(seq, event, data) \
292 ADD_ELEM((seq), (LINK_ELEMENT *)new_trace_body(iseq, (event), (data)))
294static void iseq_add_getlocal(
rb_iseq_t *iseq,
LINK_ANCHOR *
const seq,
const NODE *
const line_node,
int idx,
int level);
295static void iseq_add_setlocal(
rb_iseq_t *iseq,
LINK_ANCHOR *
const seq,
const NODE *
const line_node,
int idx,
int level);
297#define ADD_GETLOCAL(seq, line_node, idx, level) iseq_add_getlocal(iseq, (seq), (line_node), (idx), (level))
298#define ADD_SETLOCAL(seq, line_node, idx, level) iseq_add_setlocal(iseq, (seq), (line_node), (idx), (level))
301#define ADD_LABEL(seq, label) \
302 ADD_ELEM((seq), (LINK_ELEMENT *) (label))
304#define APPEND_LABEL(seq, before, label) \
305 APPEND_ELEM((seq), (before), (LINK_ELEMENT *) (label))
307#define ADD_ADJUST(seq, line_node, label) \
308 ADD_ELEM((seq), (LINK_ELEMENT *) new_adjust_body(iseq, (label), nd_line(line_node)))
310#define ADD_ADJUST_RESTORE(seq, label) \
311 ADD_ELEM((seq), (LINK_ELEMENT *) new_adjust_body(iseq, (label), -1))
313#define LABEL_UNREMOVABLE(label) \
314 ((label) ? (LABEL_REF(label), (label)->unremovable=1) : 0)
315#define ADD_CATCH_ENTRY(type, ls, le, iseqv, lc) do { \
316 VALUE _e = rb_ary_new3(5, (type), \
317 (VALUE)(ls) | 1, (VALUE)(le) | 1, \
318 (VALUE)(iseqv), (VALUE)(lc) | 1); \
319 LABEL_UNREMOVABLE(ls); \
322 if (NIL_P(ISEQ_COMPILE_DATA(iseq)->catch_table_ary)) \
323 RB_OBJ_WRITE(iseq, &ISEQ_COMPILE_DATA(iseq)->catch_table_ary, rb_ary_hidden_new(3)); \
324 rb_ary_push(ISEQ_COMPILE_DATA(iseq)->catch_table_ary, freeze_hide_obj(_e)); \
328#define COMPILE(anchor, desc, node) \
329 (debug_compile("== " desc "\n", \
330 iseq_compile_each(iseq, (anchor), (node), 0)))
333#define COMPILE_POPPED(anchor, desc, node) \
334 (debug_compile("== " desc "\n", \
335 iseq_compile_each(iseq, (anchor), (node), 1)))
338#define COMPILE_(anchor, desc, node, popped) \
339 (debug_compile("== " desc "\n", \
340 iseq_compile_each(iseq, (anchor), (node), (popped))))
342#define COMPILE_RECV(anchor, desc, node, recv) \
343 (private_recv_p(node) ? \
344 (ADD_INSN(anchor, node, putself), VM_CALL_FCALL) : \
345 COMPILE(anchor, desc, recv) ? 0 : -1)
347#define OPERAND_AT(insn, idx) \
348 (((INSN*)(insn))->operands[(idx)])
350#define INSN_OF(insn) \
351 (((INSN*)(insn))->insn_id)
353#define IS_INSN(link) ((link)->type == ISEQ_ELEMENT_INSN)
354#define IS_LABEL(link) ((link)->type == ISEQ_ELEMENT_LABEL)
355#define IS_ADJUST(link) ((link)->type == ISEQ_ELEMENT_ADJUST)
356#define IS_TRACE(link) ((link)->type == ISEQ_ELEMENT_TRACE)
357#define IS_INSN_ID(iobj, insn) (INSN_OF(iobj) == BIN(insn))
358#define IS_NEXT_INSN_ID(link, insn) \
359 ((link)->next && IS_INSN((link)->next) && IS_INSN_ID((link)->next, insn))
367append_compile_error(const
rb_iseq_t *iseq,
int line, const
char *fmt, ...)
369 VALUE err_info = ISEQ_COMPILE_DATA(iseq)->err_info;
370 VALUE file = rb_iseq_path(iseq);
375 err = rb_syntax_error_append(err, file, line, -1, NULL, fmt, args);
377 if (
NIL_P(err_info)) {
378 RB_OBJ_WRITE(iseq, &ISEQ_COMPILE_DATA(iseq)->err_info, err);
381 else if (!err_info) {
392compile_bug(
rb_iseq_t *iseq,
int line,
const char *fmt, ...)
396 rb_report_bug_valist(rb_iseq_path(iseq), line, fmt, args);
402#define COMPILE_ERROR append_compile_error
404#define ERROR_ARGS_AT(n) iseq, nd_line(n),
405#define ERROR_ARGS ERROR_ARGS_AT(node)
407#define EXPECT_NODE(prefix, node, ndtype, errval) \
409 const NODE *error_node = (node); \
410 enum node_type error_type = nd_type(error_node); \
411 if (error_type != (ndtype)) { \
412 COMPILE_ERROR(ERROR_ARGS_AT(error_node) \
413 prefix ": " #ndtype " is expected, but %s", \
414 ruby_node_name(error_type)); \
419#define EXPECT_NODE_NONULL(prefix, parent, ndtype, errval) \
421 COMPILE_ERROR(ERROR_ARGS_AT(parent) \
422 prefix ": must be " #ndtype ", but 0"); \
426#define UNKNOWN_NODE(prefix, node, errval) \
428 const NODE *error_node = (node); \
429 COMPILE_ERROR(ERROR_ARGS_AT(error_node) prefix ": unknown node (%s)", \
430 ruby_node_name(nd_type(error_node))); \
437#define CHECK(sub) if (!(sub)) {BEFORE_RETURN;return COMPILE_NG;}
438#define NO_CHECK(sub) (void)(sub)
441#define DECL_ANCHOR(name) \
442 LINK_ANCHOR name[1] = {{{ISEQ_ELEMENT_ANCHOR,},&name[0].anchor}}
443#define INIT_ANCHOR(name) \
444 ((name->last = &name->anchor)->next = NULL)
447freeze_hide_obj(
VALUE obj)
450 RBASIC_CLEAR_CLASS(obj);
454#include "optinsn.inc"
455#if OPT_INSTRUCTIONS_UNIFICATION
456#include "optunifs.inc"
461#define ISEQ_ARG iseq,
462#define ISEQ_ARG_DECLARE rb_iseq_t *iseq,
465#define ISEQ_ARG_DECLARE
469#define gl_node_level ISEQ_COMPILE_DATA(iseq)->node_level
475static int insn_data_length(
INSN *iobj);
476static int calc_sp_depth(
int depth,
INSN *iobj);
478static INSN *new_insn_body(
rb_iseq_t *iseq,
int line_no,
int node_id,
enum ruby_vminsn_type insn_id,
int argc, ...);
491static int iseq_set_exception_local_table(
rb_iseq_t *iseq);
495static int iseq_set_exception_table(
rb_iseq_t *iseq);
496static int iseq_set_optargs_table(
rb_iseq_t *iseq);
497static int iseq_set_parameters_lvar_state(
const rb_iseq_t *iseq);
500static int compile_hash(
rb_iseq_t *iseq,
LINK_ANCHOR *
const ret,
const NODE *node,
int method_call_keywords,
int popped);
507verify_list(ISEQ_ARG_DECLARE
const char *info,
LINK_ANCHOR *
const anchor)
513 if (!compile_debug)
return;
515 list = anchor->anchor.next;
516 plist = &anchor->anchor;
518 if (plist != list->prev) {
525 if (anchor->last != plist && anchor->last != 0) {
530 rb_bug(
"list verify error: %08x (%s)", flag, info);
535#define verify_list(info, anchor) verify_list(iseq, (info), (anchor))
542 VALUE *original = rb_iseq_original_iseq(iseq);
544 while (i < ISEQ_BODY(iseq)->iseq_size) {
545 VALUE insn = original[i];
546 const char *types = insn_op_types(insn);
548 for (
int j=0; types[j]; j++) {
549 if (types[j] == TS_CALLDATA) {
553 if (cc != vm_cc_empty()) {
555 rb_bug(
"call cache is not initialized by vm_cc_empty()");
562 for (
unsigned int i=0; i<ISEQ_BODY(iseq)->ci_size; i++) {
563 struct rb_call_data *cd = &ISEQ_BODY(iseq)->call_data[i];
566 if (cc != NULL && cc != vm_cc_empty()) {
568 rb_bug(
"call cache is not initialized by vm_cc_empty()");
580 elem->prev = anchor->last;
581 anchor->last->next = elem;
583 verify_list(
"add", anchor);
593 elem->next = before->next;
594 elem->next->prev = elem;
596 if (before == anchor->last) anchor->last = elem;
597 verify_list(
"add", anchor);
600#define ADD_ELEM(anchor, elem) ADD_ELEM(iseq, (anchor), (elem))
601#define APPEND_ELEM(anchor, before, elem) APPEND_ELEM(iseq, (anchor), (before), (elem))
605branch_coverage_valid_p(
rb_iseq_t *iseq,
int first_line)
607 if (!ISEQ_COVERAGE(iseq))
return 0;
608 if (!ISEQ_BRANCH_COVERAGE(iseq))
return 0;
609 if (first_line <= 0)
return 0;
616 const int first_lineno = loc->beg_pos.lineno, first_column = loc->beg_pos.column;
617 const int last_lineno = loc->end_pos.lineno, last_column = loc->end_pos.column;
620 rb_hash_aset(structure, key, branch);
632 if (!branch_coverage_valid_p(iseq, loc->beg_pos.lineno))
return Qundef;
643 VALUE branch_base = rb_hash_aref(structure, key);
646 if (
NIL_P(branch_base)) {
647 branch_base = setup_branch(loc,
type, structure, key);
648 branches = rb_hash_new();
660generate_dummy_line_node(
int lineno,
int node_id)
663 nd_set_line(&dummy, lineno);
664 nd_set_node_id(&dummy, node_id);
671 if (!branch_coverage_valid_p(iseq, loc->beg_pos.lineno))
return;
682 VALUE branch = rb_hash_aref(branches, key);
686 branch = setup_branch(loc,
type, branches, key);
696 ADD_TRACE_WITH_DATA(seq, RUBY_EVENT_COVERAGE_BRANCH, counter_idx);
697 ADD_SYNTHETIC_INSN(seq, loc->end_pos.lineno, node_id, nop);
700#define ISEQ_LAST_LINE(iseq) (ISEQ_COMPILE_DATA(iseq)->last_line)
703validate_label(st_data_t name, st_data_t label, st_data_t arg)
707 if (!lobj->link.next) {
709 COMPILE_ERROR(iseq, lobj->position,
710 "%"PRIsVALUE
": undefined label",
720 st_foreach(labels_table, validate_label, (st_data_t)iseq);
721 st_free_table(labels_table);
725get_nd_recv(
const NODE *node)
727 switch (nd_type(node)) {
729 return RNODE_CALL(node)->nd_recv;
731 return RNODE_OPCALL(node)->nd_recv;
735 return RNODE_QCALL(node)->nd_recv;
739 return RNODE_ATTRASGN(node)->nd_recv;
741 return RNODE_OP_ASGN1(node)->nd_recv;
743 return RNODE_OP_ASGN2(node)->nd_recv;
745 rb_bug(
"unexpected node: %s", ruby_node_name(nd_type(node)));
750get_node_call_nd_mid(
const NODE *node)
752 switch (nd_type(node)) {
754 return RNODE_CALL(node)->nd_mid;
756 return RNODE_OPCALL(node)->nd_mid;
758 return RNODE_FCALL(node)->nd_mid;
760 return RNODE_QCALL(node)->nd_mid;
762 return RNODE_VCALL(node)->nd_mid;
764 return RNODE_ATTRASGN(node)->nd_mid;
766 rb_bug(
"unexpected node: %s", ruby_node_name(nd_type(node)));
771get_nd_args(
const NODE *node)
773 switch (nd_type(node)) {
775 return RNODE_CALL(node)->nd_args;
777 return RNODE_OPCALL(node)->nd_args;
779 return RNODE_FCALL(node)->nd_args;
781 return RNODE_QCALL(node)->nd_args;
785 return RNODE_ATTRASGN(node)->nd_args;
787 rb_bug(
"unexpected node: %s", ruby_node_name(nd_type(node)));
792get_node_colon_nd_mid(
const NODE *node)
794 switch (nd_type(node)) {
796 return RNODE_COLON2(node)->nd_mid;
798 return RNODE_COLON3(node)->nd_mid;
800 rb_bug(
"unexpected node: %s", ruby_node_name(nd_type(node)));
805get_nd_vid(
const NODE *node)
807 switch (nd_type(node)) {
809 return RNODE_LASGN(node)->nd_vid;
811 return RNODE_DASGN(node)->nd_vid;
813 return RNODE_IASGN(node)->nd_vid;
815 return RNODE_CVASGN(node)->nd_vid;
817 rb_bug(
"unexpected node: %s", ruby_node_name(nd_type(node)));
822get_nd_value(
const NODE *node)
824 switch (nd_type(node)) {
826 return RNODE_LASGN(node)->nd_value;
828 return RNODE_DASGN(node)->nd_value;
830 rb_bug(
"unexpected node: %s", ruby_node_name(nd_type(node)));
835get_string_value(
const NODE *node)
837 switch (nd_type(node)) {
839 return RB_OBJ_SET_SHAREABLE(rb_node_str_string_val(node));
841 return RB_OBJ_SET_SHAREABLE(rb_node_file_path_val(node));
843 rb_bug(
"unexpected node: %s", ruby_node_name(nd_type(node)));
853 (*ifunc->func)(iseq, ret, ifunc->data);
855 ADD_SYNTHETIC_INSN(ret, ISEQ_COMPILE_DATA(iseq)->last_line, -1, leave);
857 CHECK(iseq_setup_insn(iseq, ret));
858 return iseq_setup(iseq, ret);
861static bool drop_unreachable_return(
LINK_ANCHOR *ret);
870 NO_CHECK(COMPILE(ret,
"nil", node));
871 iseq_set_local_table(iseq, 0, 0);
874 else if (nd_type_p(node, NODE_SCOPE)) {
876 iseq_set_local_table(iseq, RNODE_SCOPE(node)->nd_tbl, (
NODE *)RNODE_SCOPE(node)->nd_args);
877 iseq_set_arguments(iseq, ret, (
NODE *)RNODE_SCOPE(node)->nd_args);
878 iseq_set_parameters_lvar_state(iseq);
880 switch (ISEQ_BODY(iseq)->
type) {
881 case ISEQ_TYPE_BLOCK:
883 LABEL *start = ISEQ_COMPILE_DATA(iseq)->start_label = NEW_LABEL(0);
884 LABEL *end = ISEQ_COMPILE_DATA(iseq)->end_label = NEW_LABEL(0);
886 start->rescued = LABEL_RESCUE_BEG;
887 end->rescued = LABEL_RESCUE_END;
890 ADD_SYNTHETIC_INSN(ret, ISEQ_BODY(iseq)->location.first_lineno, -1, nop);
891 ADD_LABEL(ret, start);
892 CHECK(COMPILE(ret,
"block body", RNODE_SCOPE(node)->nd_body));
895 ISEQ_COMPILE_DATA(iseq)->last_line = ISEQ_BODY(iseq)->location.code_location.end_pos.lineno;
898 ADD_CATCH_ENTRY(CATCH_TYPE_REDO, start, end, NULL, start);
899 ADD_CATCH_ENTRY(CATCH_TYPE_NEXT, start, end, NULL, end);
902 case ISEQ_TYPE_CLASS:
905 CHECK(COMPILE(ret,
"scoped node", RNODE_SCOPE(node)->nd_body));
907 ISEQ_COMPILE_DATA(iseq)->last_line = nd_line(node);
910 case ISEQ_TYPE_METHOD:
912 ISEQ_COMPILE_DATA(iseq)->root_node = RNODE_SCOPE(node)->nd_body;
914 CHECK(COMPILE(ret,
"scoped node", RNODE_SCOPE(node)->nd_body));
915 ISEQ_COMPILE_DATA(iseq)->root_node = RNODE_SCOPE(node)->nd_body;
917 ISEQ_COMPILE_DATA(iseq)->last_line = nd_line(node);
921 CHECK(COMPILE(ret,
"scoped node", RNODE_SCOPE(node)->nd_body));
928#define INVALID_ISEQ_TYPE(type) \
929 ISEQ_TYPE_##type: m = #type; goto invalid_iseq_type
930 switch (ISEQ_BODY(iseq)->
type) {
931 case INVALID_ISEQ_TYPE(
METHOD);
932 case INVALID_ISEQ_TYPE(CLASS);
933 case INVALID_ISEQ_TYPE(BLOCK);
934 case INVALID_ISEQ_TYPE(EVAL);
935 case INVALID_ISEQ_TYPE(MAIN);
936 case INVALID_ISEQ_TYPE(TOP);
937#undef INVALID_ISEQ_TYPE
938 case ISEQ_TYPE_RESCUE:
939 iseq_set_exception_local_table(iseq);
940 CHECK(COMPILE(ret,
"rescue", node));
942 case ISEQ_TYPE_ENSURE:
943 iseq_set_exception_local_table(iseq);
944 CHECK(COMPILE_POPPED(ret,
"ensure", node));
946 case ISEQ_TYPE_PLAIN:
947 CHECK(COMPILE(ret,
"ensure", node));
950 COMPILE_ERROR(ERROR_ARGS
"unknown scope: %d", ISEQ_BODY(iseq)->
type);
953 COMPILE_ERROR(ERROR_ARGS
"compile/ISEQ_TYPE_%s should not be reached", m);
958 if (ISEQ_BODY(iseq)->
type == ISEQ_TYPE_RESCUE || ISEQ_BODY(iseq)->
type == ISEQ_TYPE_ENSURE) {
959 NODE dummy_line_node = generate_dummy_line_node(0, -1);
960 ADD_GETLOCAL(ret, &dummy_line_node, LVAR_ERRINFO, 0);
961 ADD_INSN1(ret, &dummy_line_node,
throw,
INT2FIX(0) );
963 else if (!drop_unreachable_return(ret)) {
964 ADD_SYNTHETIC_INSN(ret, ISEQ_COMPILE_DATA(iseq)->last_line, -1, leave);
968 if (ISEQ_COMPILE_DATA(iseq)->labels_table) {
969 st_table *labels_table = ISEQ_COMPILE_DATA(iseq)->labels_table;
970 ISEQ_COMPILE_DATA(iseq)->labels_table = 0;
971 validate_labels(iseq, labels_table);
974 CHECK(iseq_setup_insn(iseq, ret));
975 return iseq_setup(iseq, ret);
979rb_iseq_translate_threaded_code(
rb_iseq_t *iseq)
981#if OPT_DIRECT_THREADED_CODE || OPT_CALL_THREADED_CODE
982 const void *
const *table = rb_vm_get_insns_address_table();
984 VALUE *encoded = (
VALUE *)ISEQ_BODY(iseq)->iseq_encoded;
986 for (i = 0; i < ISEQ_BODY(iseq)->iseq_size; ) {
987 int insn = (int)ISEQ_BODY(iseq)->iseq_encoded[i];
988 int len = insn_len(insn);
989 encoded[i] = (
VALUE)table[insn];
996 rb_yjit_live_iseq_count++;
997 rb_yjit_iseq_alloc_count++;
1004rb_iseq_original_iseq(
const rb_iseq_t *iseq)
1006 VALUE *original_code;
1008 if (ISEQ_ORIGINAL_ISEQ(iseq))
return ISEQ_ORIGINAL_ISEQ(iseq);
1009 original_code = ISEQ_ORIGINAL_ISEQ_ALLOC(iseq, ISEQ_BODY(iseq)->iseq_size);
1010 MEMCPY(original_code, ISEQ_BODY(iseq)->iseq_encoded,
VALUE, ISEQ_BODY(iseq)->iseq_size);
1012#if OPT_DIRECT_THREADED_CODE || OPT_CALL_THREADED_CODE
1016 for (i = 0; i < ISEQ_BODY(iseq)->iseq_size; ) {
1017 const void *addr = (
const void *)original_code[i];
1018 const int insn = rb_vm_insn_addr2insn(addr);
1020 original_code[i] = insn;
1021 i += insn_len(insn);
1025 return original_code;
1038#if defined(__sparc) && SIZEOF_VOIDP == 4 && defined(__GNUC__)
1039 #define STRICT_ALIGNMENT
1045#if defined(__OpenBSD__)
1046 #include <sys/endian.h>
1047 #ifdef __STRICT_ALIGNMENT
1048 #define STRICT_ALIGNMENT
1052#ifdef STRICT_ALIGNMENT
1053 #if defined(HAVE_TRUE_LONG_LONG) && SIZEOF_LONG_LONG > SIZEOF_VALUE
1054 #define ALIGNMENT_SIZE SIZEOF_LONG_LONG
1056 #define ALIGNMENT_SIZE SIZEOF_VALUE
1058 #define PADDING_SIZE_MAX ((size_t)((ALIGNMENT_SIZE) - 1))
1059 #define ALIGNMENT_SIZE_MASK PADDING_SIZE_MAX
1062 #define PADDING_SIZE_MAX 0
1065#ifdef STRICT_ALIGNMENT
1068calc_padding(
void *ptr,
size_t size)
1073 mis = (size_t)ptr & ALIGNMENT_SIZE_MASK;
1075 padding = ALIGNMENT_SIZE - mis;
1081#if ALIGNMENT_SIZE > SIZEOF_VALUE
1082 if (size ==
sizeof(
VALUE) && padding ==
sizeof(
VALUE)) {
1096#ifdef STRICT_ALIGNMENT
1097 size_t padding = calc_padding((
void *)&storage->buff[storage->pos], size);
1099 const size_t padding = 0;
1102 if (size >= INT_MAX - padding) rb_memerror();
1103 if (storage->pos + size + padding > storage->size) {
1104 unsigned int alloc_size = storage->size;
1106 while (alloc_size < size + PADDING_SIZE_MAX) {
1107 if (alloc_size >= INT_MAX / 2) rb_memerror();
1110 storage->next = (
void *)
ALLOC_N(
char, alloc_size +
1112 storage = *arena = storage->next;
1115 storage->size = alloc_size;
1116#ifdef STRICT_ALIGNMENT
1117 padding = calc_padding((
void *)&storage->buff[storage->pos], size);
1121#ifdef STRICT_ALIGNMENT
1122 storage->pos += (int)padding;
1125 ptr = (
void *)&storage->buff[storage->pos];
1126 storage->pos += (int)size;
1131compile_data_alloc(
rb_iseq_t *iseq,
size_t size)
1134 return compile_data_alloc_with_arena(arena, size);
1138compile_data_alloc2(
rb_iseq_t *iseq,
size_t x,
size_t y)
1141 return compile_data_alloc(iseq, size);
1145compile_data_calloc2(
rb_iseq_t *iseq,
size_t x,
size_t y)
1148 void *p = compile_data_alloc(iseq, size);
1157 return (
INSN *)compile_data_alloc_with_arena(arena,
sizeof(
INSN));
1161compile_data_alloc_label(
rb_iseq_t *iseq)
1163 return (
LABEL *)compile_data_alloc(iseq,
sizeof(
LABEL));
1167compile_data_alloc_adjust(
rb_iseq_t *iseq)
1169 return (
ADJUST *)compile_data_alloc(iseq,
sizeof(
ADJUST));
1173compile_data_alloc_trace(
rb_iseq_t *iseq)
1175 return (
TRACE *)compile_data_alloc(iseq,
sizeof(
TRACE));
1184 elem2->next = elem1->next;
1185 elem2->prev = elem1;
1186 elem1->next = elem2;
1188 elem2->next->prev = elem2;
1198 elem2->prev = elem1->prev;
1199 elem2->next = elem1;
1200 elem1->prev = elem2;
1202 elem2->prev->next = elem2;
1212 elem2->prev = elem1->prev;
1213 elem2->next = elem1->next;
1215 elem1->prev->next = elem2;
1218 elem1->next->prev = elem2;
1225 elem->prev->next = elem->next;
1227 elem->next->prev = elem->prev;
1234 return anchor->anchor.next;
1240 return anchor->last;
1247 switch (elem->type) {
1248 case ISEQ_ELEMENT_INSN:
1249 case ISEQ_ELEMENT_ADJUST:
1259LIST_INSN_SIZE_ONE(
const LINK_ANCHOR *
const anchor)
1261 LINK_ELEMENT *first_insn = ELEM_FIRST_INSN(FIRST_ELEMENT(anchor));
1262 if (first_insn != NULL &&
1263 ELEM_FIRST_INSN(first_insn->next) == NULL) {
1272LIST_INSN_SIZE_ZERO(
const LINK_ANCHOR *
const anchor)
1274 if (ELEM_FIRST_INSN(FIRST_ELEMENT(anchor)) == NULL) {
1292 if (anc2->anchor.next) {
1295 anc1->last->next = anc2->anchor.next;
1296 anc2->anchor.next->prev = anc1->last;
1297 anc1->last = anc2->last;
1302 verify_list(
"append", anc1);
1305#define APPEND_LIST(anc1, anc2) APPEND_LIST(iseq, (anc1), (anc2))
1314 printf(
"anch: %p, frst: %p, last: %p\n", (
void *)&anchor->anchor,
1315 (
void *)anchor->anchor.next, (
void *)anchor->last);
1317 printf(
"curr: %p, next: %p, prev: %p, type: %d\n", (
void *)list, (
void *)list->next,
1318 (
void *)list->prev, (
int)list->type);
1323 dump_disasm_list_with_cursor(anchor->anchor.next, cur, 0);
1324 verify_list(
"debug list", anchor);
1327#define debug_list(anc, cur) debug_list(iseq, (anc), (cur))
1330#define debug_list(anc, cur) ((void)0)
1336 TRACE *trace = compile_data_alloc_trace(iseq);
1338 trace->link.type = ISEQ_ELEMENT_TRACE;
1339 trace->link.next = NULL;
1340 trace->event = event;
1347new_label_body(
rb_iseq_t *iseq,
long line)
1349 LABEL *labelobj = compile_data_alloc_label(iseq);
1351 labelobj->link.type = ISEQ_ELEMENT_LABEL;
1352 labelobj->link.next = 0;
1354 labelobj->label_no = ISEQ_COMPILE_DATA(iseq)->label_no++;
1355 labelobj->sc_state = 0;
1357 labelobj->refcnt = 0;
1359 labelobj->rescued = LABEL_RESCUE_NONE;
1360 labelobj->unremovable = 0;
1361 labelobj->position = -1;
1368 ADJUST *adjust = compile_data_alloc_adjust(iseq);
1369 adjust->link.type = ISEQ_ELEMENT_ADJUST;
1370 adjust->link.next = 0;
1371 adjust->label = label;
1372 adjust->line_no = line;
1373 LABEL_UNREMOVABLE(label);
1380 const char *types = insn_op_types(insn->insn_id);
1381 for (
int j = 0; types[j]; j++) {
1382 char type = types[j];
1389 func(&OPERAND_AT(insn, j), data);
1398iseq_insn_each_object_write_barrier(
VALUE * obj,
VALUE iseq)
1407new_insn_core(
rb_iseq_t *iseq,
int line_no,
int node_id,
int insn_id,
int argc,
VALUE *argv)
1409 INSN *iobj = compile_data_alloc_insn(iseq);
1413 iobj->link.type = ISEQ_ELEMENT_INSN;
1414 iobj->link.next = 0;
1415 iobj->insn_id = insn_id;
1416 iobj->insn_info.line_no = line_no;
1417 iobj->insn_info.node_id = node_id;
1418 iobj->insn_info.events = 0;
1419 iobj->operands = argv;
1420 iobj->operand_size = argc;
1423 iseq_insn_each_markable_object(iobj, iseq_insn_each_object_write_barrier, (
VALUE)iseq);
1429new_insn_body(
rb_iseq_t *iseq,
int line_no,
int node_id,
enum ruby_vminsn_type insn_id,
int argc, ...)
1431 VALUE *operands = 0;
1435 va_start(argv, argc);
1436 operands = compile_data_alloc2(iseq,
sizeof(
VALUE), argc);
1437 for (i = 0; i < argc; i++) {
1443 return new_insn_core(iseq, line_no, node_id, insn_id, argc, operands);
1447insn_replace_with_operands(
rb_iseq_t *iseq,
INSN *iobj,
enum ruby_vminsn_type insn_id,
int argc, ...)
1449 VALUE *operands = 0;
1453 va_start(argv, argc);
1454 operands = compile_data_alloc2(iseq,
sizeof(
VALUE), argc);
1455 for (i = 0; i < argc; i++) {
1462 iobj->insn_id = insn_id;
1463 iobj->operand_size = argc;
1464 iobj->operands = operands;
1465 iseq_insn_each_markable_object(iobj, iseq_insn_each_object_write_barrier, (
VALUE)iseq);
1473 VM_ASSERT(argc >= 0);
1476 flag |= VM_CALL_KWARG;
1477 argc += kw_arg->keyword_len;
1480 if (!(flag & (VM_CALL_ARGS_SPLAT | VM_CALL_ARGS_BLOCKARG | VM_CALL_KWARG | VM_CALL_KW_SPLAT | VM_CALL_FORWARDING))
1481 && !has_blockiseq) {
1482 flag |= VM_CALL_ARGS_SIMPLE;
1485 ISEQ_BODY(iseq)->ci_size++;
1486 const struct rb_callinfo *ci = vm_ci_new(mid, flag, argc, kw_arg);
1494 VALUE *operands = compile_data_calloc2(iseq,
sizeof(
VALUE), 2);
1497 operands[1] = (
VALUE)blockiseq;
1504 if (vm_ci_flag((
struct rb_callinfo *)ci) & VM_CALL_FORWARDING) {
1505 insn = new_insn_core(iseq, line_no, node_id, BIN(sendforward), 2, operands);
1508 insn = new_insn_core(iseq, line_no, node_id, BIN(send), 2, operands);
1521 VALUE ast_value = rb_ruby_ast_new(node);
1523 debugs(
"[new_child_iseq]> ---------------------------------------\n");
1524 int isolated_depth = ISEQ_COMPILE_DATA(iseq)->isolated_depth;
1525 ret_iseq = rb_iseq_new_with_opt(ast_value, name,
1526 rb_iseq_path(iseq), rb_iseq_realpath(iseq),
1528 isolated_depth ? isolated_depth + 1 : 0,
1529 type, ISEQ_COMPILE_DATA(iseq)->option,
1530 ISEQ_BODY(iseq)->variable.script_lines);
1531 debugs(
"[new_child_iseq]< ---------------------------------------\n");
1541 debugs(
"[new_child_iseq_with_callback]> ---------------------------------------\n");
1542 ret_iseq = rb_iseq_new_with_callback(ifunc, name,
1543 rb_iseq_path(iseq), rb_iseq_realpath(iseq),
1544 line_no, parent,
type, ISEQ_COMPILE_DATA(iseq)->option);
1545 debugs(
"[new_child_iseq_with_callback]< ---------------------------------------\n");
1553 ISEQ_COMPILE_DATA(iseq)->catch_except_p =
true;
1554 if (ISEQ_BODY(iseq)->parent_iseq != NULL) {
1555 if (ISEQ_COMPILE_DATA(ISEQ_BODY(iseq)->parent_iseq)) {
1556 set_catch_except_p((
rb_iseq_t *) ISEQ_BODY(iseq)->parent_iseq);
1579 while (pos < body->iseq_size) {
1580 insn = rb_vm_insn_decode(body->iseq_encoded[pos]);
1581 if (insn == BIN(
throw)) {
1582 set_catch_except_p(iseq);
1585 pos += insn_len(insn);
1591 for (i = 0; i < ct->size; i++) {
1593 UNALIGNED_MEMBER_PTR(ct, entries[i]);
1594 if (entry->type != CATCH_TYPE_BREAK
1595 && entry->type != CATCH_TYPE_NEXT
1596 && entry->type != CATCH_TYPE_REDO) {
1598 ISEQ_COMPILE_DATA(iseq)->catch_except_p =
true;
1605iseq_insert_nop_between_end_and_cont(
rb_iseq_t *iseq)
1607 VALUE catch_table_ary = ISEQ_COMPILE_DATA(iseq)->catch_table_ary;
1608 if (
NIL_P(catch_table_ary))
return;
1609 unsigned int i, tlen = (
unsigned int)
RARRAY_LEN(catch_table_ary);
1611 for (i = 0; i < tlen; i++) {
1617 enum rb_catch_type ct = (
enum rb_catch_type)(ptr[0] & 0xffff);
1619 if (ct != CATCH_TYPE_BREAK
1620 && ct != CATCH_TYPE_NEXT
1621 && ct != CATCH_TYPE_REDO) {
1623 for (e = end; e && (IS_LABEL(e) || IS_TRACE(e)); e = e->next) {
1625 INSN *nop = new_insn_core(iseq, 0, -1, BIN(nop), 0, 0);
1626 ELEM_INSERT_NEXT(end, &nop->link);
1639 if (
RTEST(ISEQ_COMPILE_DATA(iseq)->err_info))
1644 if (compile_debug > 5)
1645 dump_disasm_list(FIRST_ELEMENT(anchor));
1647 debugs(
"[compile step 3.1 (iseq_optimize)]\n");
1648 iseq_optimize(iseq, anchor);
1650 if (compile_debug > 5)
1651 dump_disasm_list(FIRST_ELEMENT(anchor));
1653 if (ISEQ_COMPILE_DATA(iseq)->option->instructions_unification) {
1654 debugs(
"[compile step 3.2 (iseq_insns_unification)]\n");
1655 iseq_insns_unification(iseq, anchor);
1656 if (compile_debug > 5)
1657 dump_disasm_list(FIRST_ELEMENT(anchor));
1660 debugs(
"[compile step 3.4 (iseq_insert_nop_between_end_and_cont)]\n");
1661 iseq_insert_nop_between_end_and_cont(iseq);
1662 if (compile_debug > 5)
1663 dump_disasm_list(FIRST_ELEMENT(anchor));
1671 if (
RTEST(ISEQ_COMPILE_DATA(iseq)->err_info))
1674 debugs(
"[compile step 4.1 (iseq_set_sequence)]\n");
1675 if (!iseq_set_sequence(iseq, anchor))
return COMPILE_NG;
1676 if (compile_debug > 5)
1677 dump_disasm_list(FIRST_ELEMENT(anchor));
1679 debugs(
"[compile step 4.2 (iseq_set_exception_table)]\n");
1680 if (!iseq_set_exception_table(iseq))
return COMPILE_NG;
1682 debugs(
"[compile step 4.3 (set_optargs_table)] \n");
1683 if (!iseq_set_optargs_table(iseq))
return COMPILE_NG;
1685 debugs(
"[compile step 5 (iseq_translate_threaded_code)] \n");
1686 if (!rb_iseq_translate_threaded_code(iseq))
return COMPILE_NG;
1688 debugs(
"[compile step 6 (update_catch_except_flags)] \n");
1690 update_catch_except_flags(iseq, ISEQ_BODY(iseq));
1692 debugs(
"[compile step 6.1 (remove unused catch tables)] \n");
1694 if (!ISEQ_COMPILE_DATA(iseq)->catch_except_p && ISEQ_BODY(iseq)->catch_table) {
1695 ruby_sized_xfree(ISEQ_BODY(iseq)->catch_table, iseq_catch_table_bytes(ISEQ_BODY(iseq)->catch_table->size));
1696 ISEQ_BODY(iseq)->catch_table = NULL;
1699#if VM_INSN_INFO_TABLE_IMPL == 2
1700 if (ISEQ_BODY(iseq)->insns_info.succ_index_table == NULL) {
1701 debugs(
"[compile step 7 (rb_iseq_insns_info_encode_positions)] \n");
1702 rb_iseq_insns_info_encode_positions(iseq);
1706 if (compile_debug > 1) {
1707 VALUE str = rb_iseq_disasm(iseq);
1710 verify_call_cache(iseq);
1711 debugs(
"[compile step: finish]\n");
1717iseq_set_exception_local_table(
rb_iseq_t *iseq)
1719 ISEQ_BODY(iseq)->local_table_size = numberof(rb_iseq_shared_exc_local_tbl);
1720 ISEQ_BODY(iseq)->local_table = rb_iseq_shared_exc_local_tbl;
1721 ISEQ_BODY(iseq)->lvar_states = NULL;
1729 while (iseq != ISEQ_BODY(iseq)->local_iseq) {
1731 iseq = ISEQ_BODY(iseq)->parent_iseq;
1737get_dyna_var_idx_at_raw(
const rb_iseq_t *iseq,
ID id)
1741 for (i = 0; i < ISEQ_BODY(iseq)->local_table_size; i++) {
1742 if (ISEQ_BODY(iseq)->local_table[i] == id) {
1752 int idx = get_dyna_var_idx_at_raw(ISEQ_BODY(iseq)->local_iseq,
id);
1755 COMPILE_ERROR(iseq, ISEQ_LAST_LINE(iseq),
1756 "get_local_var_idx: %d", idx);
1763get_dyna_var_idx(
const rb_iseq_t *iseq,
ID id,
int *level,
int *ls)
1765 int lv = 0, idx = -1;
1766 const rb_iseq_t *
const topmost_iseq = iseq;
1769 idx = get_dyna_var_idx_at_raw(iseq,
id);
1773 iseq = ISEQ_BODY(iseq)->parent_iseq;
1778 COMPILE_ERROR(topmost_iseq, ISEQ_LAST_LINE(topmost_iseq),
1779 "get_dyna_var_idx: -1");
1783 *ls = ISEQ_BODY(iseq)->local_table_size;
1788iseq_local_block_param_p(
const rb_iseq_t *iseq,
unsigned int idx,
unsigned int level)
1792 iseq = ISEQ_BODY(iseq)->parent_iseq;
1795 body = ISEQ_BODY(iseq);
1796 if (body->local_iseq == iseq &&
1797 body->param.flags.has_block &&
1798 body->local_table_size - body->param.block_start == idx) {
1807iseq_block_param_id_p(
const rb_iseq_t *iseq,
ID id,
int *pidx,
int *plevel)
1810 int idx = get_dyna_var_idx(iseq,
id, &level, &ls);
1811 if (iseq_local_block_param_p(iseq, ls - idx, level)) {
1822access_outer_variables(
const rb_iseq_t *iseq,
int level,
ID id,
bool write)
1824 int isolated_depth = ISEQ_COMPILE_DATA(iseq)->isolated_depth;
1826 if (isolated_depth && level >= isolated_depth) {
1827 if (
id == rb_intern(
"yield")) {
1828 COMPILE_ERROR(iseq, ISEQ_LAST_LINE(iseq),
"can not yield from isolated Proc");
1831 COMPILE_ERROR(iseq, ISEQ_LAST_LINE(iseq),
"can not access variable '%s' from isolated Proc", rb_id2name(
id));
1835 for (
int i=0; i<level; i++) {
1837 struct rb_id_table *ovs = ISEQ_BODY(iseq)->outer_variables;
1840 ovs = ISEQ_BODY(iseq)->outer_variables = rb_id_table_create(8);
1843 if (rb_id_table_lookup(ISEQ_BODY(iseq)->outer_variables,
id, &val)) {
1844 if (write && !val) {
1845 rb_id_table_insert(ISEQ_BODY(iseq)->outer_variables,
id,
Qtrue);
1849 rb_id_table_insert(ISEQ_BODY(iseq)->outer_variables,
id, RBOOL(write));
1852 iseq = ISEQ_BODY(iseq)->parent_iseq;
1857iseq_lvar_id(
const rb_iseq_t *iseq,
int idx,
int level)
1859 for (
int i=0; i<level; i++) {
1860 iseq = ISEQ_BODY(iseq)->parent_iseq;
1863 ID id = ISEQ_BODY(iseq)->local_table[ISEQ_BODY(iseq)->local_table_size - idx];
1869update_lvar_state(
const rb_iseq_t *iseq,
int level,
int idx)
1871 for (
int i=0; i<level; i++) {
1872 iseq = ISEQ_BODY(iseq)->parent_iseq;
1875 enum lvar_state *states = ISEQ_BODY(iseq)->lvar_states;
1876 int table_idx = ISEQ_BODY(iseq)->local_table_size - idx;
1877 switch (states[table_idx]) {
1878 case lvar_uninitialized:
1879 states[table_idx] = lvar_initialized;
1881 case lvar_initialized:
1882 states[table_idx] = lvar_reassigned;
1884 case lvar_reassigned:
1888 rb_bug(
"unreachable");
1893iseq_set_parameters_lvar_state(
const rb_iseq_t *iseq)
1895 for (
unsigned int i=0; i<ISEQ_BODY(iseq)->param.size; i++) {
1896 ISEQ_BODY(iseq)->lvar_states[i] = lvar_initialized;
1899 int lead_num = ISEQ_BODY(iseq)->param.lead_num;
1900 int opt_num = ISEQ_BODY(iseq)->param.opt_num;
1901 for (
int i=0; i<opt_num; i++) {
1902 ISEQ_BODY(iseq)->lvar_states[lead_num + i] = lvar_uninitialized;
1911 if (iseq_local_block_param_p(iseq, idx, level)) {
1912 ADD_INSN2(seq, line_node, getblockparam,
INT2FIX((idx) + VM_ENV_DATA_SIZE - 1),
INT2FIX(level));
1915 ADD_INSN2(seq, line_node, getlocal,
INT2FIX((idx) + VM_ENV_DATA_SIZE - 1),
INT2FIX(level));
1917 if (level > 0) access_outer_variables(iseq, level, iseq_lvar_id(iseq, idx, level),
Qfalse);
1923 if (iseq_local_block_param_p(iseq, idx, level)) {
1924 ADD_INSN2(seq, line_node, setblockparam,
INT2FIX((idx) + VM_ENV_DATA_SIZE - 1),
INT2FIX(level));
1927 ADD_INSN2(seq, line_node, setlocal,
INT2FIX((idx) + VM_ENV_DATA_SIZE - 1),
INT2FIX(level));
1929 update_lvar_state(iseq, level, idx);
1930 if (level > 0) access_outer_variables(iseq, level, iseq_lvar_id(iseq, idx, level),
Qtrue);
1939 if (body->param.flags.has_opt ||
1940 body->param.flags.has_post ||
1941 body->param.flags.has_rest ||
1942 body->param.flags.has_block ||
1943 body->param.flags.has_kw ||
1944 body->param.flags.has_kwrest) {
1946 if (body->param.flags.has_block) {
1947 body->param.size = body->param.block_start + 1;
1949 else if (body->param.flags.has_kwrest) {
1950 body->param.size = body->param.keyword->rest_start + 1;
1952 else if (body->param.flags.has_kw) {
1953 body->param.size = body->param.keyword->bits_start + 1;
1955 else if (body->param.flags.has_post) {
1956 body->param.size = body->param.post_start + body->param.post_num;
1958 else if (body->param.flags.has_rest) {
1959 body->param.size = body->param.rest_start + 1;
1961 else if (body->param.flags.has_opt) {
1962 body->param.size = body->param.lead_num + body->param.opt_num;
1969 body->param.size = body->param.lead_num;
1979 struct rb_iseq_param_keyword *keyword;
1982 int kw = 0, rkw = 0, di = 0, i;
1984 body->param.flags.has_kw = TRUE;
1985 body->param.keyword = keyword =
ZALLOC_N(
struct rb_iseq_param_keyword, 1);
1989 node = node->nd_next;
1992 keyword->bits_start = arg_size++;
1994 node = args->kw_args;
1996 const NODE *val_node = get_nd_value(node->nd_body);
1999 if (val_node == NODE_SPECIAL_REQUIRED_KEYWORD) {
2003 switch (nd_type(val_node)) {
2005 dv = rb_node_sym_string_val(val_node);
2008 dv = rb_node_regx_string_val(val_node);
2011 dv = rb_node_line_lineno_val(val_node);
2014 dv = rb_node_integer_literal_val(val_node);
2017 dv = rb_node_float_literal_val(val_node);
2020 dv = rb_node_rational_literal_val(val_node);
2022 case NODE_IMAGINARY:
2023 dv = rb_node_imaginary_literal_val(val_node);
2026 dv = rb_node_encoding_val(val_node);
2038 NO_CHECK(COMPILE_POPPED(optargs,
"kwarg", RNODE(node)));
2042 keyword->num = ++di;
2046 node = node->nd_next;
2051 if (RNODE_DVAR(args->kw_rest_arg)->nd_vid != 0) {
2052 ID kw_id = ISEQ_BODY(iseq)->local_table[arg_size];
2053 keyword->rest_start = arg_size++;
2054 body->param.flags.has_kwrest = TRUE;
2056 if (kw_id == idPow) body->param.flags.anon_kwrest = TRUE;
2058 keyword->required_num = rkw;
2059 keyword->table = &body->local_table[keyword->bits_start - keyword->num];
2064 for (i = 0; i <
RARRAY_LEN(default_values); i++) {
2066 if (dv == complex_mark) dv =
Qundef;
2071 keyword->default_values = dvs;
2080 if (!body->param.flags.use_block) {
2081 body->param.flags.use_block = 1;
2086 st_data_t key = (st_data_t)rb_intern_str(body->location.label);
2087 set_insert(vm->unused_block_warning_table, key);
2095 debugs(
"iseq_set_arguments: %s\n", node_args ?
"" :
"0");
2099 const struct rb_args_info *
const args = &RNODE_ARGS(node_args)->nd_ainfo;
2105 EXPECT_NODE(
"iseq_set_arguments", node_args, NODE_ARGS, COMPILE_NG);
2107 body->param.lead_num = arg_size = (int)args->pre_args_num;
2108 if (body->param.lead_num > 0) body->param.flags.has_lead = TRUE;
2109 debugs(
" - argc: %d\n", body->param.lead_num);
2111 rest_id = args->rest_arg;
2112 if (rest_id == NODE_SPECIAL_EXCESSIVE_COMMA) {
2116 block_id = args->block_arg;
2118 bool optimized_forward = (args->forwarding && args->pre_args_num == 0 && !args->opt_args);
2120 if (optimized_forward) {
2125 if (args->opt_args) {
2133 label = NEW_LABEL(nd_line(RNODE(node)));
2135 ADD_LABEL(optargs, label);
2136 NO_CHECK(COMPILE_POPPED(optargs,
"optarg", node->nd_body));
2137 node = node->nd_next;
2142 label = NEW_LABEL(nd_line(node_args));
2144 ADD_LABEL(optargs, label);
2149 for (j = 0; j < i+1; j++) {
2154 body->param.flags.has_opt = TRUE;
2155 body->param.opt_num = i;
2156 body->param.opt_table = opt_table;
2161 body->param.rest_start = arg_size++;
2162 body->param.flags.has_rest = TRUE;
2163 if (rest_id ==
'*') body->param.flags.anon_rest = TRUE;
2167 if (args->first_post_arg) {
2168 body->param.post_start = arg_size;
2169 body->param.post_num = args->post_args_num;
2170 body->param.flags.has_post = TRUE;
2171 arg_size += args->post_args_num;
2173 if (body->param.flags.has_rest) {
2174 body->param.post_start = body->param.rest_start + 1;
2178 if (args->kw_args) {
2179 arg_size = iseq_set_arguments_keywords(iseq, optargs, args, arg_size);
2181 else if (args->kw_rest_arg && !optimized_forward) {
2182 ID kw_id = ISEQ_BODY(iseq)->local_table[arg_size];
2183 struct rb_iseq_param_keyword *keyword =
ZALLOC_N(
struct rb_iseq_param_keyword, 1);
2184 keyword->rest_start = arg_size++;
2185 body->param.keyword = keyword;
2186 body->param.flags.has_kwrest = TRUE;
2188 static ID anon_kwrest = 0;
2189 if (!anon_kwrest) anon_kwrest = rb_intern(
"**");
2190 if (kw_id == anon_kwrest) body->param.flags.anon_kwrest = TRUE;
2192 else if (args->no_kwarg) {
2193 body->param.flags.accepts_no_kwarg = TRUE;
2196 if (args->no_blockarg) {
2197 body->param.flags.accepts_no_block = TRUE;
2199 else if (block_id) {
2200 body->param.block_start = arg_size++;
2201 body->param.flags.has_block = TRUE;
2202 iseq_set_use_block(iseq);
2206 if (optimized_forward) {
2207 body->param.flags.use_block = 1;
2208 body->param.flags.forwardable = TRUE;
2212 iseq_calc_param_size(iseq);
2213 body->param.size = arg_size;
2215 if (args->pre_init) {
2216 NO_CHECK(COMPILE_POPPED(optargs,
"init arguments (m)", args->pre_init));
2218 if (args->post_init) {
2219 NO_CHECK(COMPILE_POPPED(optargs,
"init arguments (p)", args->post_init));
2222 if (body->type == ISEQ_TYPE_BLOCK) {
2223 if (body->param.flags.has_opt == FALSE &&
2224 body->param.flags.has_post == FALSE &&
2225 body->param.flags.has_rest == FALSE &&
2226 body->param.flags.has_kw == FALSE &&
2227 body->param.flags.has_kwrest == FALSE) {
2229 if (body->param.lead_num == 1 && last_comma == 0) {
2231 body->param.flags.ambiguous_param0 = TRUE;
2243 unsigned int size = tbl ? tbl->size : 0;
2244 unsigned int offset = 0;
2247 struct rb_args_info *args = &RNODE_ARGS(node_args)->nd_ainfo;
2252 if (args->forwarding && args->pre_args_num == 0 && !args->opt_args) {
2261 MEMCPY(ids, tbl->ids + offset,
ID, size);
2262 ISEQ_BODY(iseq)->local_table = ids;
2264 enum lvar_state *states =
ALLOC_N(
enum lvar_state, size);
2266 for (
unsigned int i=0; i<size; i++) {
2267 states[i] = lvar_uninitialized;
2270 ISEQ_BODY(iseq)->lvar_states = states;
2272 ISEQ_BODY(iseq)->local_table_size = size;
2274 debugs(
"iseq_set_local_table: %u\n", ISEQ_BODY(iseq)->local_table_size);
2286 else if ((tlit = OBJ_BUILTIN_TYPE(lit)) == -1) {
2289 else if ((tval = OBJ_BUILTIN_TYPE(val)) == -1) {
2292 else if (tlit != tval) {
2302 long x =
FIX2LONG(rb_big_cmp(lit, val));
2310 return rb_float_cmp(lit, val);
2313 const struct RRational *rat1 = RRATIONAL(val);
2314 const struct RRational *rat2 = RRATIONAL(lit);
2315 return rb_iseq_cdhash_cmp(rat1->num, rat2->num) || rb_iseq_cdhash_cmp(rat1->den, rat2->den);
2318 const struct RComplex *comp1 = RCOMPLEX(val);
2319 const struct RComplex *comp2 = RCOMPLEX(lit);
2320 return rb_iseq_cdhash_cmp(comp1->real, comp2->real) || rb_iseq_cdhash_cmp(comp1->imag, comp2->imag);
2323 return rb_reg_equal(val, lit) ? 0 : -1;
2331rb_iseq_cdhash_hash(
VALUE a)
2333 switch (OBJ_BUILTIN_TYPE(a)) {
2336 return (st_index_t)a;
2344 return rb_rational_hash(a);
2346 return rb_complex_hash(a);
2356 rb_iseq_cdhash_hash,
2370 rb_hash_aset(data->hash, key,
INT2FIX(lobj->position - (data->pos+data->len)));
2378 return INT2FIX(ISEQ_BODY(iseq)->ivc_size++);
2385 struct rb_id_table *tbl = ISEQ_COMPILE_DATA(iseq)->ivar_cache_table;
2387 if (rb_id_table_lookup(tbl,
id,&val)) {
2392 tbl = rb_id_table_create(1);
2393 ISEQ_COMPILE_DATA(iseq)->ivar_cache_table = tbl;
2395 val =
INT2FIX(ISEQ_BODY(iseq)->icvarc_size++);
2396 rb_id_table_insert(tbl,
id,val);
2400#define BADINSN_DUMP(anchor, list, dest) \
2401 dump_disasm_list_with_cursor(FIRST_ELEMENT(anchor), list, dest)
2403#define BADINSN_ERROR \
2404 (SIZED_FREE_N(generated_iseq, generated_iseq_size), \
2405 SIZED_FREE_N(insns_info, insns_info_size), \
2406 BADINSN_DUMP(anchor, list, NULL), \
2412 int stack_max = 0, sp = 0, line = 0;
2415 for (list = FIRST_ELEMENT(anchor); list; list = list->next) {
2416 if (IS_LABEL(list)) {
2422 for (list = FIRST_ELEMENT(anchor); list; list = list->next) {
2423 switch (list->type) {
2424 case ISEQ_ELEMENT_INSN:
2432 sp = calc_sp_depth(sp, iobj);
2434 BADINSN_DUMP(anchor, list, NULL);
2435 COMPILE_ERROR(iseq, iobj->insn_info.line_no,
2436 "argument stack underflow (%d)", sp);
2439 if (sp > stack_max) {
2443 line = iobj->insn_info.line_no;
2445 operands = iobj->operands;
2446 insn = iobj->insn_id;
2447 types = insn_op_types(insn);
2448 len = insn_len(insn);
2451 if (iobj->operand_size !=
len - 1) {
2453 BADINSN_DUMP(anchor, list, NULL);
2454 COMPILE_ERROR(iseq, iobj->insn_info.line_no,
2455 "operand size miss! (%d for %d)",
2456 iobj->operand_size,
len - 1);
2460 for (j = 0; types[j]; j++) {
2461 if (types[j] == TS_OFFSET) {
2465 BADINSN_DUMP(anchor, list, NULL);
2466 COMPILE_ERROR(iseq, iobj->insn_info.line_no,
2467 "unknown label: "LABEL_FORMAT, lobj->label_no);
2470 if (lobj->sp == -1) {
2473 else if (lobj->sp != sp) {
2474 debugs(
"%s:%d: sp inconsistency found but ignored (" LABEL_FORMAT
" sp: %d, calculated sp: %d)\n",
2475 RSTRING_PTR(rb_iseq_path(iseq)), line,
2476 lobj->label_no, lobj->sp, sp);
2482 case ISEQ_ELEMENT_LABEL:
2485 if (lobj->sp == -1) {
2489 if (lobj->sp != sp) {
2490 debugs(
"%s:%d: sp inconsistency found but ignored (" LABEL_FORMAT
" sp: %d, calculated sp: %d)\n",
2491 RSTRING_PTR(rb_iseq_path(iseq)), line,
2492 lobj->label_no, lobj->sp, sp);
2498 case ISEQ_ELEMENT_TRACE:
2503 case ISEQ_ELEMENT_ADJUST:
2508 sp = adjust->label ? adjust->label->sp : 0;
2509 if (adjust->line_no != -1 && orig_sp - sp < 0) {
2510 BADINSN_DUMP(anchor, list, NULL);
2511 COMPILE_ERROR(iseq, adjust->line_no,
2512 "iseq_set_sequence: adjust bug %d < %d",
2519 BADINSN_DUMP(anchor, list, NULL);
2520 COMPILE_ERROR(iseq, line,
"unknown list type: %d", list->type);
2529 int insns_info_index,
int code_index,
const INSN *iobj)
2531 if (insns_info_index == 0 ||
2532 insns_info[insns_info_index-1].line_no != iobj->insn_info.line_no ||
2533#ifdef USE_ISEQ_NODE_ID
2534 insns_info[insns_info_index-1].node_id != iobj->insn_info.node_id ||
2536 insns_info[insns_info_index-1].events != iobj->insn_info.events) {
2537 insns_info[insns_info_index].line_no = iobj->insn_info.line_no;
2538#ifdef USE_ISEQ_NODE_ID
2539 insns_info[insns_info_index].node_id = iobj->insn_info.node_id;
2541 insns_info[insns_info_index].events = iobj->insn_info.events;
2542 positions[insns_info_index] = code_index;
2550 int insns_info_index,
int code_index,
const ADJUST *adjust)
2552 insns_info[insns_info_index].line_no = adjust->line_no;
2553 insns_info[insns_info_index].node_id = -1;
2554 insns_info[insns_info_index].events = 0;
2555 positions[insns_info_index] = code_index;
2560array_to_idlist(
VALUE arr)
2565 for (
long i = 0; i < size; i++) {
2574idlist_to_array(
const ID *ids)
2591 unsigned int *positions;
2593 VALUE *generated_iseq;
2597 int insn_num, code_index, insns_info_index, sp = 0;
2598 int stack_max = fix_sp_depth(iseq, anchor);
2600 if (stack_max < 0)
return COMPILE_NG;
2603 insn_num = code_index = 0;
2604 for (list = FIRST_ELEMENT(anchor); list; list = list->next) {
2605 switch (list->type) {
2606 case ISEQ_ELEMENT_INSN:
2610 sp = calc_sp_depth(sp, iobj);
2612 events = iobj->insn_info.events |= events;
2613 if (ISEQ_COVERAGE(iseq)) {
2614 if (ISEQ_LINE_COVERAGE(iseq) && (events & RUBY_EVENT_COVERAGE_LINE) &&
2615 !(rb_get_coverage_mode() & COVERAGE_TARGET_ONESHOT_LINES)) {
2616 int line = iobj->insn_info.line_no - 1;
2617 if (line >= 0 && line <
RARRAY_LEN(ISEQ_LINE_COVERAGE(iseq))) {
2621 if (ISEQ_BRANCH_COVERAGE(iseq) && (events & RUBY_EVENT_COVERAGE_BRANCH)) {
2622 while (
RARRAY_LEN(ISEQ_PC2BRANCHINDEX(iseq)) <= code_index) {
2628 code_index += insn_data_length(iobj);
2633 case ISEQ_ELEMENT_LABEL:
2636 lobj->position = code_index;
2637 if (lobj->sp != sp) {
2638 debugs(
"%s: sp inconsistency found but ignored (" LABEL_FORMAT
" sp: %d, calculated sp: %d)\n",
2639 RSTRING_PTR(rb_iseq_path(iseq)),
2640 lobj->label_no, lobj->sp, sp);
2645 case ISEQ_ELEMENT_TRACE:
2648 events |= trace->event;
2649 if (trace->event & RUBY_EVENT_COVERAGE_BRANCH) data = trace->data;
2652 case ISEQ_ELEMENT_ADJUST:
2655 if (adjust->line_no != -1) {
2657 sp = adjust->label ? adjust->label->sp : 0;
2658 if (orig_sp - sp > 0) {
2659 if (orig_sp - sp > 1) code_index++;
2671 const int generated_iseq_size = code_index;
2674 const int insns_info_size = insn_num;
2677 const int positions_size = insn_num;
2678 positions =
ALLOC_N(
unsigned int, insn_num);
2679 if (ISEQ_IS_SIZE(body)) {
2683 body->is_entries = NULL;
2686 if (body->ci_size) {
2690 body->call_data = NULL;
2692 ISEQ_COMPILE_DATA(iseq)->ci_index = 0;
2699 iseq_bits_t * mark_offset_bits;
2700 int code_size = code_index;
2702 bool needs_bitmap =
false;
2704 const size_t mark_offset_bits_size = ISEQ_MBITS_BUFLEN(code_index);
2705 if (mark_offset_bits_size == 1) {
2706 mark_offset_bits = &ISEQ_COMPILE_DATA(iseq)->mark_bits.single;
2707 ISEQ_COMPILE_DATA(iseq)->is_single_mark_bit =
true;
2710 mark_offset_bits =
ZALLOC_N(iseq_bits_t, mark_offset_bits_size);
2711 ISEQ_COMPILE_DATA(iseq)->mark_bits.list = mark_offset_bits;
2712 ISEQ_COMPILE_DATA(iseq)->is_single_mark_bit =
false;
2715 ISEQ_COMPILE_DATA(iseq)->iseq_encoded = (
void *)generated_iseq;
2716 ISEQ_COMPILE_DATA(iseq)->iseq_size = code_index;
2718 list = FIRST_ELEMENT(anchor);
2719 insns_info_index = code_index = sp = 0;
2722 switch (list->type) {
2723 case ISEQ_ELEMENT_INSN:
2731 sp = calc_sp_depth(sp, iobj);
2733 operands = iobj->operands;
2734 insn = iobj->insn_id;
2735 generated_iseq[code_index] = insn;
2736 types = insn_op_types(insn);
2737 len = insn_len(insn);
2739 for (j = 0; types[j]; j++) {
2740 char type = types[j];
2748 generated_iseq[code_index + 1 + j] = lobj->position - (code_index +
len);
2753 VALUE map = operands[j];
2756 data.pos = code_index;
2760 rb_hash_rehash(map);
2761 freeze_hide_obj(map);
2763 generated_iseq[code_index + 1 + j] = map;
2764 ISEQ_MBITS_SET(mark_offset_bits, code_index + 1 + j);
2766 needs_bitmap =
true;
2771 generated_iseq[code_index + 1 + j] =
FIX2INT(operands[j]);
2776 VALUE v = operands[j];
2777 generated_iseq[code_index + 1 + j] = v;
2781 ISEQ_MBITS_SET(mark_offset_bits, code_index + 1 + j);
2782 needs_bitmap =
true;
2789 unsigned int ic_index = ISEQ_COMPILE_DATA(iseq)->ic_index++;
2790 IC ic = &ISEQ_IS_ENTRY_START(body,
type)[ic_index].ic_cache;
2791 if (UNLIKELY(ic_index >= body->ic_size)) {
2792 BADINSN_DUMP(anchor, &iobj->link, 0);
2793 COMPILE_ERROR(iseq, iobj->insn_info.line_no,
2794 "iseq_set_sequence: ic_index overflow: index: %d, size: %d",
2795 ic_index, ISEQ_IS_SIZE(body));
2798 ic->
segments = array_to_idlist(operands[j]);
2800 generated_iseq[code_index + 1 + j] = (
VALUE)ic;
2805 unsigned int ic_index =
FIX2UINT(operands[j]);
2807 IVC cache = ((
IVC)&body->is_entries[ic_index]);
2809 if (insn == BIN(setinstancevariable)) {
2810 cache->iv_set_name =
SYM2ID(operands[j - 1]);
2813 cache->iv_set_name = 0;
2816 vm_ic_attr_index_initialize(cache, INVALID_SHAPE_ID);
2821 unsigned int ic_index =
FIX2UINT(operands[j]);
2822 IC ic = &ISEQ_IS_ENTRY_START(body,
type)[ic_index].ic_cache;
2823 if (UNLIKELY(ic_index >= ISEQ_IS_SIZE(body))) {
2824 BADINSN_DUMP(anchor, &iobj->link, 0);
2825 COMPILE_ERROR(iseq, iobj->insn_info.line_no,
2826 "iseq_set_sequence: ic_index overflow: index: %d, size: %d",
2827 ic_index, ISEQ_IS_SIZE(body));
2829 generated_iseq[code_index + 1 + j] = (
VALUE)ic;
2836 RUBY_ASSERT(ISEQ_COMPILE_DATA(iseq)->ci_index <= body->ci_size);
2837 struct rb_call_data *cd = &body->call_data[ISEQ_COMPILE_DATA(iseq)->ci_index++];
2839 cd->cc = vm_cc_empty();
2840 generated_iseq[code_index + 1 + j] = (
VALUE)cd;
2844 generated_iseq[code_index + 1 + j] =
SYM2ID(operands[j]);
2847 generated_iseq[code_index + 1 + j] = operands[j];
2850 generated_iseq[code_index + 1 + j] = operands[j];
2853 BADINSN_ERROR(iseq, iobj->insn_info.line_no,
2854 "unknown operand type: %c",
type);
2858 if (add_insn_info(insns_info, positions, insns_info_index, code_index, iobj)) insns_info_index++;
2862 case ISEQ_ELEMENT_LABEL:
2865 if (lobj->sp != sp) {
2866 debugs(
"%s: sp inconsistency found but ignored (" LABEL_FORMAT
" sp: %d, calculated sp: %d)\n",
2867 RSTRING_PTR(rb_iseq_path(iseq)),
2868 lobj->label_no, lobj->sp, sp);
2873 case ISEQ_ELEMENT_ADJUST:
2878 if (adjust->label) {
2879 sp = adjust->label->sp;
2885 if (adjust->line_no != -1) {
2886 const int diff = orig_sp - sp;
2888 if (insns_info_index == 0) {
2889 COMPILE_ERROR(iseq, adjust->line_no,
2890 "iseq_set_sequence: adjust bug (ISEQ_ELEMENT_ADJUST must not be the first in iseq)");
2892 if (add_adjust_info(insns_info, positions, insns_info_index, code_index, adjust)) insns_info_index++;
2895 generated_iseq[code_index++] = BIN(adjuststack);
2896 generated_iseq[code_index++] = orig_sp - sp;
2898 else if (diff == 1) {
2899 generated_iseq[code_index++] = BIN(pop);
2901 else if (diff < 0) {
2902 int label_no = adjust->label ? adjust->label->label_no : -1;
2903 SIZED_FREE_N(generated_iseq, generated_iseq_size);
2904 SIZED_FREE_N(insns_info, insns_info_size);
2905 SIZED_FREE_N(positions, positions_size);
2906 if (ISEQ_MBITS_BUFLEN(code_size) > 1) {
2907 SIZED_FREE_N(mark_offset_bits, ISEQ_MBITS_BUFLEN(code_index));
2909 debug_list(anchor, list);
2910 COMPILE_ERROR(iseq, adjust->line_no,
2911 "iseq_set_sequence: adjust bug to %d %d < %d",
2912 label_no, orig_sp, sp);
2925 body->iseq_encoded = (
void *)generated_iseq;
2926 body->iseq_size = code_index;
2927 body->stack_max = stack_max;
2929 if (ISEQ_COMPILE_DATA(iseq)->is_single_mark_bit) {
2930 body->mark_bits.single = ISEQ_COMPILE_DATA(iseq)->mark_bits.single;
2934 body->mark_bits.list = mark_offset_bits;
2937 body->mark_bits.list = NULL;
2938 ISEQ_COMPILE_DATA(iseq)->mark_bits.list = NULL;
2939 SIZED_FREE_N(mark_offset_bits, mark_offset_bits_size);
2944 body->insns_info.body = insns_info;
2945 body->insns_info.positions = positions;
2948 body->insns_info.body = insns_info;
2949 SIZED_REALLOC_N(positions,
unsigned int, insns_info_index, positions_size);
2950 body->insns_info.positions = positions;
2951 body->insns_info.size = insns_info_index;
2957label_get_position(
LABEL *lobj)
2959 return lobj->position;
2963label_get_sp(
LABEL *lobj)
2969iseq_set_exception_table(
rb_iseq_t *iseq)
2971 const VALUE *tptr, *ptr;
2972 unsigned int tlen, i;
2975 ISEQ_BODY(iseq)->catch_table = NULL;
2977 VALUE catch_table_ary = ISEQ_COMPILE_DATA(iseq)->catch_table_ary;
2978 if (
NIL_P(catch_table_ary))
return COMPILE_OK;
2986 for (i = 0; i < table->size; i++) {
2989 entry = UNALIGNED_MEMBER_PTR(table, entries[i]);
2990 entry->type = (
enum rb_catch_type)(ptr[0] & 0xffff);
2991 pos = label_get_position((
LABEL *)(ptr[1] & ~1));
2993 entry->start = (
unsigned int)pos;
2994 pos = label_get_position((
LABEL *)(ptr[2] & ~1));
2996 entry->end = (
unsigned int)pos;
3003 entry->cont = label_get_position(lobj);
3004 entry->sp = label_get_sp(lobj);
3007 if (entry->type == CATCH_TYPE_RESCUE ||
3008 entry->type == CATCH_TYPE_BREAK ||
3009 entry->type == CATCH_TYPE_NEXT) {
3018 ISEQ_BODY(iseq)->catch_table = table;
3019 RB_OBJ_WRITE(iseq, &ISEQ_COMPILE_DATA(iseq)->catch_table_ary, 0);
3040 VALUE *opt_table = (
VALUE *)ISEQ_BODY(iseq)->param.opt_table;
3042 if (ISEQ_BODY(iseq)->param.flags.has_opt) {
3043 for (i = 0; i < ISEQ_BODY(iseq)->param.opt_num + 1; i++) {
3044 opt_table[i] = label_get_position((
LABEL *)opt_table[i]);
3051get_destination_insn(
INSN *iobj)
3057 list = lobj->link.next;
3059 switch (list->type) {
3060 case ISEQ_ELEMENT_INSN:
3061 case ISEQ_ELEMENT_ADJUST:
3063 case ISEQ_ELEMENT_LABEL:
3066 case ISEQ_ELEMENT_TRACE:
3069 events |= trace->event;
3077 if (list && IS_INSN(list)) {
3079 iobj->insn_info.events |= events;
3085get_next_insn(
INSN *iobj)
3090 if (IS_INSN(list) || IS_ADJUST(list)) {
3099get_prev_insn(
INSN *iobj)
3104 if (IS_INSN(list) || IS_ADJUST(list)) {
3113unref_destination(
INSN *iobj,
int pos)
3115 LABEL *lobj = (
LABEL *)OPERAND_AT(iobj, pos);
3117 if (!lobj->refcnt) ELEM_REMOVE(&lobj->link);
3121replace_destination(
INSN *dobj,
INSN *nobj)
3123 VALUE n = OPERAND_AT(nobj, 0);
3126 if (dl == nl)
return false;
3129 OPERAND_AT(dobj, 0) = n;
3130 if (!dl->refcnt) ELEM_REMOVE(&dl->link);
3135find_destination(
INSN *i)
3137 int pos,
len = insn_len(i->insn_id);
3138 for (pos = 0; pos <
len; ++pos) {
3139 if (insn_op_types(i->insn_id)[pos] == TS_OFFSET) {
3140 return (
LABEL *)OPERAND_AT(i, pos);
3150 int *unref_counts = 0, nlabels = ISEQ_COMPILE_DATA(iseq)->label_no;
3153 unref_counts =
ALLOCA_N(
int, nlabels);
3154 MEMZERO(unref_counts,
int, nlabels);
3159 if (IS_INSN_ID(i, leave)) {
3163 else if ((lab = find_destination((
INSN *)i)) != 0) {
3164 unref_counts[lab->label_no]++;
3167 else if (IS_LABEL(i)) {
3169 if (lab->unremovable)
return 0;
3170 if (lab->refcnt > unref_counts[lab->label_no]) {
3171 if (i == first)
return 0;
3176 else if (IS_TRACE(i)) {
3179 else if (IS_ADJUST(i)) {
3183 }
while ((i = i->next) != 0);
3188 VALUE insn = INSN_OF(i);
3189 int pos,
len = insn_len(insn);
3190 for (pos = 0; pos <
len; ++pos) {
3191 switch (insn_op_types(insn)[pos]) {
3193 unref_destination((
INSN *)i, pos);
3202 }
while ((i != end) && (i = i->next) != 0);
3209 switch (OPERAND_AT(iobj, 0)) {
3211 ELEM_REMOVE(&iobj->link);
3214 ELEM_REMOVE(&iobj->link);
3217 iobj->insn_id = BIN(adjuststack);
3223is_frozen_putstring(
INSN *insn,
VALUE *op)
3225 if (IS_INSN_ID(insn, putstring) || IS_INSN_ID(insn, putchilledstring)) {
3226 *op = OPERAND_AT(insn, 0);
3229 else if (IS_INSN_ID(insn, putobject)) {
3230 *op = OPERAND_AT(insn, 0);
3241 if (prev->type == ISEQ_ELEMENT_LABEL) {
3243 if (label->refcnt > 0) {
3247 else if (prev->type == ISEQ_ELEMENT_INSN) {
3280 INSN *niobj, *ciobj, *dup = 0;
3284 switch (INSN_OF(iobj)) {
3285 case BIN(putstring):
3286 case BIN(putchilledstring):
3292 case BIN(putobject):
3295 default:
return FALSE;
3298 ciobj = (
INSN *)get_next_insn(iobj);
3299 if (IS_INSN_ID(ciobj, jump)) {
3300 ciobj = (
INSN *)get_next_insn((
INSN*)OPERAND_AT(ciobj, 0));
3302 if (IS_INSN_ID(ciobj, dup)) {
3303 ciobj = (
INSN *)get_next_insn(dup = ciobj);
3305 if (!ciobj || !IS_INSN_ID(ciobj, checktype))
return FALSE;
3306 niobj = (
INSN *)get_next_insn(ciobj);
3311 switch (INSN_OF(niobj)) {
3313 if (OPERAND_AT(ciobj, 0) ==
type) {
3314 dest = (
LABEL *)OPERAND_AT(niobj, 0);
3317 case BIN(branchunless):
3318 if (OPERAND_AT(ciobj, 0) !=
type) {
3319 dest = (
LABEL *)OPERAND_AT(niobj, 0);
3325 line = ciobj->insn_info.line_no;
3326 node_id = ciobj->insn_info.node_id;
3328 if (niobj->link.next && IS_LABEL(niobj->link.next)) {
3329 dest = (
LABEL *)niobj->link.next;
3332 dest = NEW_LABEL(line);
3333 ELEM_INSERT_NEXT(&niobj->link, &dest->link);
3336 INSERT_AFTER_INSN1(iobj, line, node_id, jump, dest);
3338 if (!dup) INSERT_AFTER_INSN(iobj, line, node_id, pop);
3345 const struct rb_callinfo *nci = vm_ci_new(vm_ci_mid(ci),
3346 vm_ci_flag(ci) | add,
3356 const struct rb_callinfo *nci = vm_ci_new(vm_ci_mid(ci),
3364#define vm_ci_simple(ci) (vm_ci_flag(ci) & VM_CALL_ARGS_SIMPLE)
3372 optimize_checktype(iseq, iobj);
3374 if (IS_INSN_ID(iobj, jump)) {
3375 INSN *niobj, *diobj, *piobj;
3376 diobj = (
INSN *)get_destination_insn(iobj);
3377 niobj = (
INSN *)get_next_insn(iobj);
3379 if (diobj == niobj) {
3386 unref_destination(iobj, 0);
3387 ELEM_REMOVE(&iobj->link);
3390 else if (iobj != diobj && IS_INSN(&diobj->link) &&
3391 IS_INSN_ID(diobj, jump) &&
3392 OPERAND_AT(iobj, 0) != OPERAND_AT(diobj, 0) &&
3393 diobj->insn_info.events == 0) {
3404 if (replace_destination(iobj, diobj)) {
3405 remove_unreachable_chunk(iseq, iobj->link.next);
3409 else if (IS_INSN_ID(diobj, leave)) {
3422 unref_destination(iobj, 0);
3423 iobj->insn_id = BIN(leave);
3424 iobj->operand_size = 0;
3425 iobj->insn_info = diobj->insn_info;
3428 else if (IS_INSN(iobj->link.prev) &&
3429 (piobj = (
INSN *)iobj->link.prev) &&
3430 (IS_INSN_ID(piobj, branchif) ||
3431 IS_INSN_ID(piobj, branchunless))) {
3432 INSN *pdiobj = (
INSN *)get_destination_insn(piobj);
3433 if (niobj == pdiobj) {
3434 int refcnt = IS_LABEL(piobj->link.next) ?
3435 ((
LABEL *)piobj->link.next)->refcnt : 0;
3450 piobj->insn_id = (IS_INSN_ID(piobj, branchif))
3451 ? BIN(branchunless) : BIN(branchif);
3452 if (replace_destination(piobj, iobj) && refcnt <= 1) {
3453 ELEM_REMOVE(&iobj->link);
3460 else if (diobj == pdiobj) {
3474 INSN *popiobj = new_insn_core(iseq, iobj->insn_info.line_no, iobj->insn_info.node_id, BIN(pop), 0, 0);
3475 ELEM_REPLACE(&piobj->link, &popiobj->link);
3478 if (remove_unreachable_chunk(iseq, iobj->link.next)) {
3492 if (IS_INSN_ID(iobj, newrange)) {
3493 INSN *
const range = iobj;
3495 VALUE str_beg, str_end;
3497 if ((end = (
INSN *)get_prev_insn(range)) != 0 &&
3498 is_frozen_putstring(end, &str_end) &&
3499 (beg = (
INSN *)get_prev_insn(end)) != 0 &&
3500 is_frozen_putstring(beg, &str_beg) &&
3501 !(insn_has_label_before(&beg->link) || insn_has_label_before(&end->link))) {
3502 int excl =
FIX2INT(OPERAND_AT(range, 0));
3505 ELEM_REMOVE(&beg->link);
3506 ELEM_REMOVE(&end->link);
3507 range->insn_id = BIN(putobject);
3508 OPERAND_AT(range, 0) = lit_range;
3513 if (IS_INSN_ID(iobj, leave)) {
3514 remove_unreachable_chunk(iseq, iobj->link.next);
3526 if (IS_INSN_ID(iobj, duparray)) {
3528 if (IS_INSN(next) && (IS_INSN_ID(next, concatarray) || IS_INSN_ID(next, concattoarray))) {
3529 iobj->insn_id = BIN(putobject);
3539 if (IS_INSN_ID(iobj, duparray)) {
3541 if (IS_INSN(next) && (IS_INSN_ID(next, send))) {
3545 if (vm_ci_simple(ci) && vm_ci_argc(ci) == 0 && blockiseq == NULL && vm_ci_mid(ci) == idFreeze) {
3546 VALUE ary = iobj->operands[0];
3549 insn_replace_with_operands(iseq, iobj, BIN(opt_ary_freeze), 2, ary, (
VALUE)ci);
3561 if (IS_INSN_ID(iobj, duphash)) {
3563 if (IS_INSN(next) && (IS_INSN_ID(next, send))) {
3567 if (vm_ci_simple(ci) && vm_ci_argc(ci) == 0 && blockiseq == NULL && vm_ci_mid(ci) == idFreeze) {
3568 VALUE hash = iobj->operands[0];
3570 RB_OBJ_SET_SHAREABLE(hash);
3572 insn_replace_with_operands(iseq, iobj, BIN(opt_hash_freeze), 2, hash, (
VALUE)ci);
3584 if (IS_INSN_ID(iobj, newarray) && iobj->operands[0] ==
INT2FIX(0)) {
3586 if (IS_INSN(next) && (IS_INSN_ID(next, send))) {
3590 if (vm_ci_simple(ci) && vm_ci_argc(ci) == 0 && blockiseq == NULL && vm_ci_mid(ci) == idFreeze) {
3591 insn_replace_with_operands(iseq, iobj, BIN(opt_ary_freeze), 2, rb_cArray_empty_frozen, (
VALUE)ci);
3603 if (IS_INSN_ID(iobj, newhash) && iobj->operands[0] ==
INT2FIX(0)) {
3605 if (IS_INSN(next) && (IS_INSN_ID(next, send))) {
3609 if (vm_ci_simple(ci) && vm_ci_argc(ci) == 0 && blockiseq == NULL && vm_ci_mid(ci) == idFreeze) {
3610 insn_replace_with_operands(iseq, iobj, BIN(opt_hash_freeze), 2, rb_cHash_empty_frozen, (
VALUE)ci);
3616 if (IS_INSN_ID(iobj, branchif) ||
3617 IS_INSN_ID(iobj, branchnil) ||
3618 IS_INSN_ID(iobj, branchunless)) {
3627 INSN *nobj = (
INSN *)get_destination_insn(iobj);
3649 int stop_optimization =
3650 ISEQ_COVERAGE(iseq) && ISEQ_LINE_COVERAGE(iseq) &&
3651 nobj->link.type == ISEQ_ELEMENT_INSN &&
3652 nobj->insn_info.events;
3653 if (!stop_optimization) {
3654 INSN *pobj = (
INSN *)iobj->link.prev;
3657 if (!IS_INSN(&pobj->link))
3659 else if (IS_INSN_ID(pobj, dup))
3664 if (IS_INSN(&nobj->link) && IS_INSN_ID(nobj, jump)) {
3665 if (!replace_destination(iobj, nobj))
break;
3667 else if (prev_dup && IS_INSN_ID(nobj, dup) &&
3668 !!(nobj = (
INSN *)nobj->link.next) &&
3670 nobj->insn_id == iobj->insn_id) {
3686 if (!replace_destination(iobj, nobj))
break;
3714 if (prev_dup && IS_INSN(pobj->link.prev)) {
3715 pobj = (
INSN *)pobj->link.prev;
3717 if (IS_INSN_ID(pobj, putobject)) {
3718 cond = (IS_INSN_ID(iobj, branchif) ?
3719 OPERAND_AT(pobj, 0) !=
Qfalse :
3720 IS_INSN_ID(iobj, branchunless) ?
3721 OPERAND_AT(pobj, 0) ==
Qfalse :
3724 else if (IS_INSN_ID(pobj, putstring) ||
3725 IS_INSN_ID(pobj, duparray) ||
3726 IS_INSN_ID(pobj, newarray)) {
3727 cond = IS_INSN_ID(iobj, branchif);
3729 else if (IS_INSN_ID(pobj, putnil)) {
3730 cond = !IS_INSN_ID(iobj, branchif);
3733 if (prev_dup || !IS_INSN_ID(pobj, newarray)) {
3734 ELEM_REMOVE(iobj->link.prev);
3736 else if (!iseq_pop_newarray(iseq, pobj)) {
3737 pobj = new_insn_core(iseq, pobj->insn_info.line_no, pobj->insn_info.node_id, BIN(pop), 0, NULL);
3738 ELEM_INSERT_PREV(&iobj->link, &pobj->link);
3742 pobj = new_insn_core(iseq, pobj->insn_info.line_no, pobj->insn_info.node_id, BIN(putnil), 0, NULL);
3743 ELEM_INSERT_NEXT(&iobj->link, &pobj->link);
3745 iobj->insn_id = BIN(jump);
3749 unref_destination(iobj, 0);
3750 ELEM_REMOVE(&iobj->link);
3755 nobj = (
INSN *)get_destination_insn(nobj);
3760 if (IS_INSN_ID(iobj, pop)) {
3768 if (IS_INSN(prev)) {
3769 enum ruby_vminsn_type previ = ((
INSN *)prev)->insn_id;
3770 if (previ == BIN(putobject) || previ == BIN(putnil) ||
3771 previ == BIN(putself) || previ == BIN(putstring) ||
3772 previ == BIN(putchilledstring) ||
3773 previ == BIN(dup) ||
3774 previ == BIN(getlocal) ||
3775 previ == BIN(getblockparam) ||
3776 previ == BIN(getblockparamproxy) ||
3777 previ == BIN(getinstancevariable) ||
3778 previ == BIN(duparray)) {
3782 ELEM_REMOVE(&iobj->link);
3784 else if (previ == BIN(newarray) && iseq_pop_newarray(iseq, (
INSN*)prev)) {
3785 ELEM_REMOVE(&iobj->link);
3787 else if (previ == BIN(concatarray)) {
3789 INSERT_BEFORE_INSN1(piobj, piobj->insn_info.line_no, piobj->insn_info.node_id, splatarray,
Qfalse);
3790 INSN_OF(piobj) = BIN(pop);
3792 else if (previ == BIN(concatstrings)) {
3793 if (OPERAND_AT(prev, 0) ==
INT2FIX(1)) {
3797 ELEM_REMOVE(&iobj->link);
3798 INSN_OF(prev) = BIN(adjuststack);
3804 if (IS_INSN_ID(iobj, newarray) ||
3805 IS_INSN_ID(iobj, duparray) ||
3806 IS_INSN_ID(iobj, concatarray) ||
3807 IS_INSN_ID(iobj, splatarray) ||
3817 if (IS_INSN(next) && IS_INSN_ID(next, splatarray)) {
3823 if (IS_INSN_ID(iobj, newarray)) {
3825 if (IS_INSN(next) && IS_INSN_ID(next, expandarray) &&
3826 OPERAND_AT(next, 1) ==
INT2FIX(0)) {
3828 op1 = OPERAND_AT(iobj, 0);
3829 op2 = OPERAND_AT(next, 0);
3840 INSN_OF(iobj) = BIN(swap);
3841 iobj->operand_size = 0;
3850 INSN_OF(iobj) = BIN(opt_reverse);
3855 INSN_OF(iobj) = BIN(opt_reverse);
3856 OPERAND_AT(iobj, 0) = OPERAND_AT(next, 0);
3866 for (; diff > 0; diff--) {
3867 INSERT_BEFORE_INSN(iobj, iobj->insn_info.line_no, iobj->insn_info.node_id, pop);
3878 for (; diff < 0; diff++) {
3879 INSERT_BEFORE_INSN(iobj, iobj->insn_info.line_no, iobj->insn_info.node_id, putnil);
3886 if (IS_INSN_ID(iobj, duparray)) {
3895 if (IS_INSN(next) && IS_INSN_ID(next, expandarray)) {
3896 INSN_OF(iobj) = BIN(putobject);
3900 if (IS_INSN_ID(iobj, anytostring)) {
3908 if (IS_INSN(next) && IS_INSN_ID(next, concatstrings) &&
3909 OPERAND_AT(next, 0) ==
INT2FIX(1)) {
3914 if (IS_INSN_ID(iobj, putstring) || IS_INSN_ID(iobj, putchilledstring) ||
3922 if (IS_NEXT_INSN_ID(&iobj->link, concatstrings) &&
3923 RSTRING_LEN(OPERAND_AT(iobj, 0)) == 0) {
3924 INSN *next = (
INSN *)iobj->link.next;
3925 if ((OPERAND_AT(next, 0) = FIXNUM_INC(OPERAND_AT(next, 0), -1)) ==
INT2FIX(1)) {
3926 ELEM_REMOVE(&next->link);
3928 ELEM_REMOVE(&iobj->link);
3930 if (IS_NEXT_INSN_ID(&iobj->link, toregexp)) {
3931 INSN *next = (
INSN *)iobj->link.next;
3932 if (OPERAND_AT(next, 1) ==
INT2FIX(1)) {
3933 VALUE src = OPERAND_AT(iobj, 0);
3934 int opt = (int)
FIX2LONG(OPERAND_AT(next, 0));
3935 VALUE path = rb_iseq_path(iseq);
3936 int line = iobj->insn_info.line_no;
3937 VALUE errinfo = rb_errinfo();
3938 VALUE re = rb_reg_compile(src, opt, RSTRING_PTR(path), line);
3940 VALUE message = rb_attr_get(rb_errinfo(), idMesg);
3941 rb_set_errinfo(errinfo);
3942 COMPILE_ERROR(iseq, line,
"%" PRIsVALUE, message);
3945 RB_OBJ_SET_SHAREABLE(re);
3948 ELEM_REMOVE(iobj->link.next);
3953 if (IS_INSN_ID(iobj, concatstrings)) {
3962 if (IS_INSN(next) && IS_INSN_ID(next, jump))
3963 next = get_destination_insn(jump = (
INSN *)next);
3964 if (IS_INSN(next) && IS_INSN_ID(next, concatstrings)) {
3965 int n =
FIX2INT(OPERAND_AT(iobj, 0)) +
FIX2INT(OPERAND_AT(next, 0)) - 1;
3966 OPERAND_AT(iobj, 0) =
INT2FIX(n);
3968 LABEL *label = ((
LABEL *)OPERAND_AT(jump, 0));
3969 if (!--label->refcnt) {
3970 ELEM_REMOVE(&label->link);
3973 label = NEW_LABEL(0);
3974 OPERAND_AT(jump, 0) = (
VALUE)label;
3977 ELEM_INSERT_NEXT(next, &label->link);
3978 CHECK(iseq_peephole_optimize(iseq, get_next_insn(jump), do_tailcallopt));
3986 if (do_tailcallopt &&
3987 (IS_INSN_ID(iobj, send) ||
3988 IS_INSN_ID(iobj, invokesuper))) {
3997 if (iobj->link.next) {
4000 if (!IS_INSN(next)) {
4004 switch (INSN_OF(next)) {
4013 next = get_destination_insn((
INSN *)next);
4027 if (IS_INSN_ID(piobj, send) ||
4028 IS_INSN_ID(piobj, invokesuper)) {
4029 if (OPERAND_AT(piobj, 1) == 0) {
4030 ci = ci_flag_set(iseq, ci, VM_CALL_TAILCALL);
4031 OPERAND_AT(piobj, 0) = (
VALUE)ci;
4036 ci = ci_flag_set(iseq, ci, VM_CALL_TAILCALL);
4037 OPERAND_AT(piobj, 0) = (
VALUE)ci;
4043 if (IS_INSN_ID(iobj, dup)) {
4044 if (IS_NEXT_INSN_ID(&iobj->link, setlocal)) {
4055 if (IS_NEXT_INSN_ID(set1, setlocal)) {
4057 if (OPERAND_AT(set1, 0) == OPERAND_AT(set2, 0) &&
4058 OPERAND_AT(set1, 1) == OPERAND_AT(set2, 1)) {
4060 ELEM_REMOVE(&iobj->link);
4073 else if (IS_NEXT_INSN_ID(set1, dup) &&
4074 IS_NEXT_INSN_ID(set1->next, setlocal)) {
4075 set2 = set1->next->next;
4076 if (OPERAND_AT(set1, 0) == OPERAND_AT(set2, 0) &&
4077 OPERAND_AT(set1, 1) == OPERAND_AT(set2, 1)) {
4078 ELEM_REMOVE(set1->next);
4092 if (IS_INSN_ID(iobj, getlocal)) {
4094 if (IS_NEXT_INSN_ID(niobj, dup)) {
4095 niobj = niobj->next;
4097 if (IS_NEXT_INSN_ID(niobj, setlocal)) {
4099 if (OPERAND_AT(iobj, 0) == OPERAND_AT(set1, 0) &&
4100 OPERAND_AT(iobj, 1) == OPERAND_AT(set1, 1)) {
4116 if (IS_INSN_ID(iobj, opt_invokebuiltin_delegate)) {
4117 if (IS_TRACE(iobj->link.next)) {
4118 if (IS_NEXT_INSN_ID(iobj->link.next, leave)) {
4119 iobj->insn_id = BIN(opt_invokebuiltin_delegate_leave);
4121 if (iobj == (
INSN *)list && bf->argc == 0 && (ISEQ_BODY(iseq)->builtin_attrs & BUILTIN_ATTR_LEAF)) {
4122 ISEQ_BODY(iseq)->builtin_attrs |= BUILTIN_ATTR_SINGLE_NOARG_LEAF;
4135 if (IS_INSN_ID(iobj, getblockparam)) {
4136 if (IS_NEXT_INSN_ID(&iobj->link, branchif) || IS_NEXT_INSN_ID(&iobj->link, branchunless)) {
4137 iobj->insn_id = BIN(getblockparamproxy);
4141 if (IS_INSN_ID(iobj, splatarray) && OPERAND_AT(iobj, 0) ==
false) {
4143 if (IS_NEXT_INSN_ID(niobj, duphash)) {
4144 niobj = niobj->next;
4146 unsigned int set_flags = 0, unset_flags = 0;
4159 if (IS_NEXT_INSN_ID(niobj, send)) {
4160 siobj = niobj->next;
4161 set_flags = VM_CALL_ARGS_SPLAT|VM_CALL_KW_SPLAT|VM_CALL_KW_SPLAT_MUT;
4162 unset_flags = VM_CALL_ARGS_BLOCKARG;
4177 else if ((IS_NEXT_INSN_ID(niobj, getlocal) || IS_NEXT_INSN_ID(niobj, getinstancevariable) ||
4178 IS_NEXT_INSN_ID(niobj, getblockparamproxy)) && (IS_NEXT_INSN_ID(niobj->next, send))) {
4179 siobj = niobj->next->next;
4180 set_flags = VM_CALL_ARGS_SPLAT|VM_CALL_KW_SPLAT|VM_CALL_KW_SPLAT_MUT|VM_CALL_ARGS_BLOCKARG;
4185 unsigned int flags = vm_ci_flag(ci);
4186 if ((flags & set_flags) == set_flags && !(flags & unset_flags)) {
4187 ((
INSN*)niobj)->insn_id = BIN(putobject);
4188 RB_OBJ_WRITE(iseq, &OPERAND_AT(niobj, 0), RB_OBJ_SET_SHAREABLE(rb_hash_freeze(rb_hash_resurrect(OPERAND_AT(niobj, 0)))));
4190 const struct rb_callinfo *nci = vm_ci_new(vm_ci_mid(ci),
4191 flags & ~VM_CALL_KW_SPLAT_MUT, vm_ci_argc(ci), vm_ci_kwarg(ci));
4193 OPERAND_AT(siobj, 0) = (
VALUE)nci;
4203insn_set_specialized_instruction(
rb_iseq_t *iseq,
INSN *iobj,
int insn_id)
4205 if (insn_id == BIN(opt_neq)) {
4206 VALUE original_ci = iobj->operands[0];
4207 VALUE new_ci = (
VALUE)new_callinfo(iseq, idEq, 1, 0, NULL, FALSE);
4208 insn_replace_with_operands(iseq, iobj, insn_id, 2, new_ci, original_ci);
4211 iobj->insn_id = insn_id;
4212 iobj->operand_size = insn_len(insn_id) - 1;
4222 if (IS_INSN_ID(iobj, newarray) && iobj->link.next &&
4223 IS_INSN(iobj->link.next)) {
4227 INSN *niobj = (
INSN *)iobj->link.next;
4228 if (IS_INSN_ID(niobj, send)) {
4230 if (vm_ci_simple(ci) && vm_ci_argc(ci) == 0) {
4232 switch (vm_ci_mid(ci)) {
4234 method =
INT2FIX(VM_OPT_NEWARRAY_SEND_MAX);
4237 method =
INT2FIX(VM_OPT_NEWARRAY_SEND_MIN);
4240 method =
INT2FIX(VM_OPT_NEWARRAY_SEND_HASH);
4245 VALUE num = iobj->operands[0];
4246 insn_replace_with_operands(iseq, iobj, BIN(opt_newarray_send), 2, num, method);
4247 ELEM_REMOVE(&niobj->link);
4252 else if ((IS_INSN_ID(niobj, putstring) || IS_INSN_ID(niobj, putchilledstring) ||
4254 IS_NEXT_INSN_ID(&niobj->link, send)) {
4256 if (vm_ci_simple(ci) && vm_ci_argc(ci) == 1 && vm_ci_mid(ci) == idPack) {
4257 VALUE num = iobj->operands[0];
4258 insn_replace_with_operands(iseq, iobj, BIN(opt_newarray_send), 2, FIXNUM_INC(num, 1),
INT2FIX(VM_OPT_NEWARRAY_SEND_PACK));
4259 ELEM_REMOVE(&iobj->link);
4260 ELEM_REMOVE(niobj->link.next);
4261 ELEM_INSERT_NEXT(&niobj->link, &iobj->link);
4267 else if ((IS_INSN_ID(niobj, putstring) || IS_INSN_ID(niobj, putchilledstring) ||
4269 IS_NEXT_INSN_ID(&niobj->link, getlocal) &&
4270 (niobj->link.next && IS_NEXT_INSN_ID(niobj->link.next, send))) {
4273 if (vm_ci_mid(ci) == idPack && vm_ci_argc(ci) == 2 &&
4274 (kwarg && kwarg->keyword_len == 1 && kwarg->keywords[0] ==
rb_id2sym(idBuffer))) {
4275 VALUE num = iobj->operands[0];
4276 insn_replace_with_operands(iseq, iobj, BIN(opt_newarray_send), 2, FIXNUM_INC(num, 2),
INT2FIX(VM_OPT_NEWARRAY_SEND_PACK_BUFFER));
4278 ELEM_REMOVE((niobj->link.next)->next);
4280 ELEM_REMOVE(&iobj->link);
4282 ELEM_INSERT_NEXT(niobj->link.next, &iobj->link);
4290 if ((IS_INSN_ID(niobj, putstring) || IS_INSN_ID(niobj, putchilledstring) ||
4291 IS_INSN_ID(niobj, putobject) ||
4292 IS_INSN_ID(niobj, putself) ||
4293 IS_INSN_ID(niobj, getlocal) ||
4294 IS_INSN_ID(niobj, getinstancevariable)) &&
4295 IS_NEXT_INSN_ID(&niobj->link, send)) {
4302 sendobj = sendobj->next;
4303 ci = (
struct rb_callinfo *)OPERAND_AT(sendobj, 0);
4304 }
while (vm_ci_simple(ci) && vm_ci_argc(ci) == 0 && IS_NEXT_INSN_ID(sendobj, send));
4307 if (vm_ci_simple(ci) && vm_ci_argc(ci) == 1 && vm_ci_mid(ci) == idIncludeP) {
4308 VALUE num = iobj->operands[0];
4310 insn_replace_with_operands(iseq, sendins, BIN(opt_newarray_send), 2, FIXNUM_INC(num, 1),
INT2FIX(VM_OPT_NEWARRAY_SEND_INCLUDE_P));
4312 ELEM_REMOVE(&iobj->link);
4326 if (IS_INSN_ID(iobj, duparray) && iobj->link.next && IS_INSN(iobj->link.next)) {
4327 INSN *niobj = (
INSN *)iobj->link.next;
4328 if ((IS_INSN_ID(niobj, getlocal) ||
4329 IS_INSN_ID(niobj, getinstancevariable) ||
4330 IS_INSN_ID(niobj, putself)) &&
4331 IS_NEXT_INSN_ID(&niobj->link, send)) {
4338 sendobj = sendobj->next;
4339 ci = (
struct rb_callinfo *)OPERAND_AT(sendobj, 0);
4340 }
while (vm_ci_simple(ci) && vm_ci_argc(ci) == 0 && IS_NEXT_INSN_ID(sendobj, send));
4342 if (vm_ci_simple(ci) && vm_ci_argc(ci) == 1 && vm_ci_mid(ci) == idIncludeP) {
4344 VALUE ary = iobj->operands[0];
4348 insn_replace_with_operands(iseq, sendins, BIN(opt_duparray_send), 3, ary,
rb_id2sym(idIncludeP),
INT2FIX(1));
4351 ELEM_REMOVE(&iobj->link);
4358 if (IS_INSN_ID(iobj, send)) {
4362#define SP_INSN(opt) insn_set_specialized_instruction(iseq, iobj, BIN(opt_##opt))
4363 if (vm_ci_simple(ci)) {
4364 switch (vm_ci_argc(ci)) {
4366 switch (vm_ci_mid(ci)) {
4367 case idLength: SP_INSN(length);
return COMPILE_OK;
4368 case idSize: SP_INSN(size);
return COMPILE_OK;
4369 case idEmptyP: SP_INSN(empty_p);
return COMPILE_OK;
4370 case idNilP: SP_INSN(nil_p);
return COMPILE_OK;
4371 case idSucc: SP_INSN(succ);
return COMPILE_OK;
4372 case idNot: SP_INSN(not);
return COMPILE_OK;
4376 switch (vm_ci_mid(ci)) {
4377 case idPLUS: SP_INSN(plus);
return COMPILE_OK;
4378 case idMINUS: SP_INSN(minus);
return COMPILE_OK;
4379 case idMULT: SP_INSN(mult);
return COMPILE_OK;
4380 case idDIV: SP_INSN(div);
return COMPILE_OK;
4381 case idMOD: SP_INSN(mod);
return COMPILE_OK;
4382 case idEq: SP_INSN(eq);
return COMPILE_OK;
4383 case idNeq: SP_INSN(neq);
return COMPILE_OK;
4384 case idEqTilde:SP_INSN(regexpmatch2);
return COMPILE_OK;
4385 case idLT: SP_INSN(lt);
return COMPILE_OK;
4386 case idLE: SP_INSN(le);
return COMPILE_OK;
4387 case idGT: SP_INSN(gt);
return COMPILE_OK;
4388 case idGE: SP_INSN(ge);
return COMPILE_OK;
4389 case idLTLT: SP_INSN(ltlt);
return COMPILE_OK;
4390 case idAREF: SP_INSN(aref);
return COMPILE_OK;
4391 case idAnd: SP_INSN(and);
return COMPILE_OK;
4392 case idOr: SP_INSN(or);
return COMPILE_OK;
4396 switch (vm_ci_mid(ci)) {
4397 case idASET: SP_INSN(aset);
return COMPILE_OK;
4403 if ((vm_ci_flag(ci) & (VM_CALL_ARGS_BLOCKARG | VM_CALL_FORWARDING)) == 0 && blockiseq == NULL) {
4404 iobj->insn_id = BIN(opt_send_without_block);
4405 iobj->operand_size = insn_len(iobj->insn_id) - 1;
4416 switch (ISEQ_BODY(iseq)->
type) {
4418 case ISEQ_TYPE_EVAL:
4419 case ISEQ_TYPE_MAIN:
4421 case ISEQ_TYPE_RESCUE:
4422 case ISEQ_TYPE_ENSURE:
4434 const int do_peepholeopt = ISEQ_COMPILE_DATA(iseq)->option->peephole_optimization;
4435 const int do_tailcallopt = tailcallable_p(iseq) &&
4436 ISEQ_COMPILE_DATA(iseq)->option->tailcall_optimization;
4437 const int do_si = ISEQ_COMPILE_DATA(iseq)->option->specialized_instruction;
4438 const int do_ou = ISEQ_COMPILE_DATA(iseq)->option->operands_unification;
4439 const int do_without_ints = ISEQ_BODY(iseq)->builtin_attrs & BUILTIN_ATTR_WITHOUT_INTERRUPTS;
4440 int rescue_level = 0;
4441 int tailcallopt = do_tailcallopt;
4443 list = FIRST_ELEMENT(anchor);
4445 int do_block_optimization = 0;
4446 LABEL * block_loop_label = NULL;
4449 if (ISEQ_BODY(iseq)->
type == ISEQ_TYPE_BLOCK) {
4450 do_block_optimization = 1;
4455 if (IS_INSN(le) && IS_INSN_ID((
INSN *)le, nop) && IS_LABEL(le->next)) {
4456 block_loop_label = (
LABEL *)le->next;
4461 if (IS_INSN(list)) {
4462 if (do_peepholeopt) {
4463 iseq_peephole_optimize(iseq, list, tailcallopt);
4466 iseq_specialized_instruction(iseq, (
INSN *)list);
4469 insn_operands_unification((
INSN *)list);
4472 if (do_without_ints) {
4474 if (IS_INSN_ID(item, jump)) {
4475 item->insn_id = BIN(jump_without_ints);
4477 else if (IS_INSN_ID(item, branchif)) {
4478 item->insn_id = BIN(branchif_without_ints);
4480 else if (IS_INSN_ID(item, branchunless)) {
4481 item->insn_id = BIN(branchunless_without_ints);
4483 else if (IS_INSN_ID(item, branchnil)) {
4484 item->insn_id = BIN(branchnil_without_ints);
4488 if (do_block_optimization) {
4491 if (IS_INSN_ID(item,
throw)) {
4492 do_block_optimization = 0;
4497 const char *types = insn_op_types(item->insn_id);
4498 for (
int j = 0; types[j]; j++) {
4499 if (types[j] == TS_OFFSET) {
4504 LABEL * target = (
LABEL *)OPERAND_AT(item, j);
4505 if (target == block_loop_label) {
4506 do_block_optimization = 0;
4513 if (IS_LABEL(list)) {
4514 switch (((
LABEL *)list)->rescued) {
4515 case LABEL_RESCUE_BEG:
4517 tailcallopt = FALSE;
4519 case LABEL_RESCUE_END:
4520 if (!--rescue_level) tailcallopt = do_tailcallopt;
4527 if (do_block_optimization) {
4529 if (IS_INSN(le) && IS_INSN_ID((
INSN *)le, nop)) {
4536#if OPT_INSTRUCTIONS_UNIFICATION
4544 VALUE *operands = 0, *ptr = 0;
4548 for (i = 0; i < size; i++) {
4549 iobj = (
INSN *)list;
4550 argc += iobj->operand_size;
4555 ptr = operands = compile_data_alloc2(iseq,
sizeof(
VALUE), argc);
4560 for (i = 0; i < size; i++) {
4561 iobj = (
INSN *)list;
4562 MEMCPY(ptr, iobj->operands,
VALUE, iobj->operand_size);
4563 ptr += iobj->operand_size;
4567 return new_insn_core(iseq, iobj->insn_info.line_no, iobj->insn_info.node_id, insn_id, argc, operands);
4579#if OPT_INSTRUCTIONS_UNIFICATION
4585 list = FIRST_ELEMENT(anchor);
4587 if (IS_INSN(list)) {
4588 iobj = (
INSN *)list;
4590 if (unified_insns_data[
id] != 0) {
4591 const int *
const *entry = unified_insns_data[id];
4592 for (j = 1; j < (intptr_t)entry[0]; j++) {
4593 const int *unified = entry[j];
4595 for (k = 2; k < unified[1]; k++) {
4597 ((
INSN *)li)->insn_id != unified[k]) {
4604 new_unified_insn(iseq, unified[0], unified[1] - 1,
4609 niobj->link.next = li;
4628all_string_result_p(
const NODE *node)
4630 if (!node)
return FALSE;
4631 switch (nd_type(node)) {
4632 case NODE_STR:
case NODE_DSTR:
case NODE_FILE:
4634 case NODE_IF:
case NODE_UNLESS:
4635 if (!RNODE_IF(node)->nd_body || !RNODE_IF(node)->nd_else)
return FALSE;
4636 if (all_string_result_p(RNODE_IF(node)->nd_body))
4637 return all_string_result_p(RNODE_IF(node)->nd_else);
4639 case NODE_AND:
case NODE_OR:
4640 if (!RNODE_AND(node)->nd_2nd)
4641 return all_string_result_p(RNODE_AND(node)->nd_1st);
4642 if (!all_string_result_p(RNODE_AND(node)->nd_1st))
4644 return all_string_result_p(RNODE_AND(node)->nd_2nd);
4654 const NODE *lit_node;
4662 VALUE s = rb_str_new_mutable_parser_string(str);
4664 VALUE error = rb_reg_check_preprocess(s);
4665 if (!
NIL_P(error)) {
4666 COMPILE_ERROR(args->iseq, nd_line(node),
"%" PRIsVALUE, error);
4670 if (
NIL_P(args->lit)) {
4672 args->lit_node = node;
4681flush_dstr_fragment(
struct dstr_ctxt *args)
4683 if (!
NIL_P(args->lit)) {
4685 VALUE lit = args->lit;
4687 lit = rb_fstring(lit);
4688 ADD_INSN1(args->ret, args->lit_node, putobject, lit);
4695compile_dstr_fragments_0(
struct dstr_ctxt *args,
const NODE *
const node)
4697 const struct RNode_LIST *list = RNODE_DSTR(node)->nd_next;
4701 CHECK(append_dstr_fragment(args, node, str));
4705 const NODE *
const head = list->nd_head;
4706 if (nd_type_p(head, NODE_STR)) {
4707 CHECK(append_dstr_fragment(args, node, RNODE_STR(head)->
string));
4709 else if (nd_type_p(head, NODE_DSTR)) {
4710 CHECK(compile_dstr_fragments_0(args, head));
4713 flush_dstr_fragment(args);
4715 CHECK(COMPILE(args->ret,
"each string", head));
4727 .iseq = iseq, .ret = ret,
4728 .lit =
Qnil, .lit_node = NULL,
4729 .cnt = 0, .dregx = dregx,
4731 CHECK(compile_dstr_fragments_0(&args, node));
4732 flush_dstr_fragment(&args);
4742 while (node && nd_type_p(node, NODE_BLOCK)) {
4743 CHECK(COMPILE_(ret,
"BLOCK body", RNODE_BLOCK(node)->nd_head,
4744 (RNODE_BLOCK(node)->nd_next ? 1 : popped)));
4745 node = RNODE_BLOCK(node)->nd_next;
4748 CHECK(COMPILE_(ret,
"BLOCK next", RNODE_BLOCK(node)->nd_next, popped));
4757 if (!RNODE_DSTR(node)->nd_next) {
4758 VALUE lit = rb_node_dstr_string_val(node);
4759 ADD_INSN1(ret, node, putstring, lit);
4760 RB_OBJ_SET_SHAREABLE(lit);
4764 CHECK(compile_dstr_fragments(iseq, ret, node, &cnt, FALSE));
4765 ADD_INSN1(ret, node, concatstrings,
INT2FIX(cnt));
4774 int cflag = (int)RNODE_DREGX(node)->as.nd_cflag;
4776 if (!RNODE_DREGX(node)->nd_next) {
4778 VALUE src = rb_node_dregx_string_val(node);
4779 VALUE match = rb_reg_compile(src, cflag, NULL, 0);
4780 RB_OBJ_SET_SHAREABLE(match);
4781 ADD_INSN1(ret, node, putobject, match);
4787 CHECK(compile_dstr_fragments(iseq, ret, node, &cnt, TRUE));
4791 ADD_INSN(ret, node, pop);
4801 const int line = nd_line(node);
4802 LABEL *lend = NEW_LABEL(line);
4803 rb_num_t cnt = ISEQ_FLIP_CNT_INCREMENT(ISEQ_BODY(iseq)->local_iseq)
4804 + VM_SVAR_FLIPFLOP_START;
4807 ADD_INSN2(ret, node, getspecial, key,
INT2FIX(0));
4808 ADD_INSNL(ret, node, branchif, lend);
4811 CHECK(COMPILE(ret,
"flip2 beg", RNODE_FLIP2(node)->nd_beg));
4812 ADD_INSNL(ret, node, branchunless, else_label);
4813 ADD_INSN1(ret, node, putobject,
Qtrue);
4814 ADD_INSN1(ret, node, setspecial, key);
4816 ADD_INSNL(ret, node, jump, then_label);
4820 ADD_LABEL(ret, lend);
4821 CHECK(COMPILE(ret,
"flip2 end", RNODE_FLIP2(node)->nd_end));
4822 ADD_INSNL(ret, node, branchunless, then_label);
4823 ADD_INSN1(ret, node, putobject,
Qfalse);
4824 ADD_INSN1(ret, node, setspecial, key);
4825 ADD_INSNL(ret, node, jump, then_label);
4834#define COMPILE_SINGLE 2
4841 LABEL *label = NEW_LABEL(nd_line(cond));
4842 if (!then_label) then_label = label;
4843 else if (!else_label) else_label = label;
4845 CHECK(compile_branch_condition(iseq, seq, cond, then_label, else_label));
4847 if (LIST_INSN_SIZE_ONE(seq)) {
4848 INSN *insn = (
INSN *)ELEM_FIRST_INSN(FIRST_ELEMENT(seq));
4849 if (insn->insn_id == BIN(jump) && (
LABEL *)(insn->operands[0]) == label)
4852 if (!label->refcnt) {
4853 return COMPILE_SINGLE;
4855 ADD_LABEL(seq, label);
4865 DECL_ANCHOR(ignore);
4868 switch (nd_type(cond)) {
4870 CHECK(ok = compile_logical(iseq, ret, RNODE_AND(cond)->nd_1st, NULL, else_label));
4871 cond = RNODE_AND(cond)->nd_2nd;
4872 if (ok == COMPILE_SINGLE) {
4873 INIT_ANCHOR(ignore);
4875 then_label = NEW_LABEL(nd_line(cond));
4879 CHECK(ok = compile_logical(iseq, ret, RNODE_OR(cond)->nd_1st, then_label, NULL));
4880 cond = RNODE_OR(cond)->nd_2nd;
4881 if (ok == COMPILE_SINGLE) {
4882 INIT_ANCHOR(ignore);
4884 else_label = NEW_LABEL(nd_line(cond));
4894 case NODE_IMAGINARY:
4901 ADD_INSNL(ret, cond, jump, then_label);
4906 ADD_INSNL(ret, cond, jump, else_label);
4912 CHECK(COMPILE_POPPED(ret,
"branch condition", cond));
4913 ADD_INSNL(ret, cond, jump, then_label);
4916 CHECK(compile_flip_flop(iseq, ret, cond, TRUE, then_label, else_label));
4919 CHECK(compile_flip_flop(iseq, ret, cond, FALSE, then_label, else_label));
4922 CHECK(compile_defined_expr(iseq, ret, cond,
Qfalse, ret == ignore));
4926 DECL_ANCHOR(cond_seq);
4927 INIT_ANCHOR(cond_seq);
4929 CHECK(COMPILE(cond_seq,
"branch condition", cond));
4931 if (LIST_INSN_SIZE_ONE(cond_seq)) {
4932 INSN *insn = (
INSN *)ELEM_FIRST_INSN(FIRST_ELEMENT(cond_seq));
4933 if (insn->insn_id == BIN(putobject)) {
4934 if (
RTEST(insn->operands[0])) {
4935 ADD_INSNL(ret, cond, jump, then_label);
4940 ADD_INSNL(ret, cond, jump, else_label);
4945 ADD_SEQ(ret, cond_seq);
4950 ADD_INSNL(ret, cond, branchunless, else_label);
4951 ADD_INSNL(ret, cond, jump, then_label);
4958keyword_node_p(
const NODE *
const node)
4960 return nd_type_p(node, NODE_HASH) && (RNODE_HASH(node)->nd_brace & HASH_BRACE) != HASH_BRACE;
4966 switch (nd_type(node)) {
4968 return rb_node_sym_string_val(node);
4970 UNKNOWN_NODE(
"get_symbol_value", node,
Qnil);
4977 NODE *node = node_hash->nd_head;
4978 VALUE hash = rb_hash_new();
4981 for (
int i = 0; node != NULL; i++, node = RNODE_LIST(RNODE_LIST(node)->nd_next)->nd_next) {
4982 VALUE key = get_symbol_value(iseq, RNODE_LIST(node)->nd_head);
4983 VALUE idx = rb_hash_aref(hash, key);
4988 rb_hash_aset(hash, key,
INT2FIX(i));
4998 const NODE *
const root_node,
5006 if (RNODE_HASH(root_node)->nd_head && nd_type_p(RNODE_HASH(root_node)->nd_head, NODE_LIST)) {
5007 const NODE *node = RNODE_HASH(root_node)->nd_head;
5011 const NODE *key_node = RNODE_LIST(node)->nd_head;
5015 if (key_node && nd_type_p(key_node, NODE_SYM)) {
5020 *flag |= VM_CALL_KW_SPLAT;
5021 if (seen_nodes > 1 || RNODE_LIST(RNODE_LIST(node)->nd_next)->nd_next) {
5026 *flag |= VM_CALL_KW_SPLAT_MUT;
5031 node = RNODE_LIST(node)->nd_next;
5032 node = RNODE_LIST(node)->nd_next;
5036 node = RNODE_HASH(root_node)->nd_head;
5039 VALUE key_index = node_hash_unique_key_index(iseq, RNODE_HASH(root_node), &
len);
5042 VALUE *keywords = kw_arg->keywords;
5045 kw_arg->references = 0;
5046 kw_arg->keyword_len =
len;
5048 *kw_arg_ptr = kw_arg;
5050 for (i=0; node != NULL; i++, node = RNODE_LIST(RNODE_LIST(node)->nd_next)->nd_next) {
5051 const NODE *key_node = RNODE_LIST(node)->nd_head;
5052 const NODE *val_node = RNODE_LIST(RNODE_LIST(node)->nd_next)->nd_head;
5055 keywords[j] = get_symbol_value(iseq, key_node);
5059 NO_CHECK(COMPILE_(ret,
"keyword values", val_node, popped));
5073 for (; node;
len++, node = RNODE_LIST(node)->nd_next) {
5075 EXPECT_NODE(
"compile_args", node, NODE_LIST, -1);
5078 if (RNODE_LIST(node)->nd_next == NULL && keyword_node_p(RNODE_LIST(node)->nd_head)) {
5079 *kwnode_ptr = RNODE_LIST(node)->nd_head;
5082 RUBY_ASSERT(!keyword_node_p(RNODE_LIST(node)->nd_head));
5083 NO_CHECK(COMPILE_(ret,
"array element", RNODE_LIST(node)->nd_head, FALSE));
5091frozen_string_literal_p(
const rb_iseq_t *iseq)
5093 return ISEQ_COMPILE_DATA(iseq)->option->frozen_string_literal > 0;
5097static_literal_node_p(
const NODE *node,
const rb_iseq_t *iseq,
bool hash_key)
5099 switch (nd_type(node)) {
5107 case NODE_IMAGINARY:
5114 return hash_key || frozen_string_literal_p(iseq);
5123 switch (nd_type(node)) {
5126 VALUE lit = rb_node_integer_literal_val(node);
5132 VALUE lit = rb_node_float_literal_val(node);
5138 case NODE_IMAGINARY:
5147 return rb_node_sym_string_val(node);
5149 return RB_OBJ_SET_SHAREABLE(rb_node_regx_string_val(node));
5151 return rb_node_line_lineno_val(node);
5153 return rb_node_encoding_val(node);
5156 if (ISEQ_COMPILE_DATA(iseq)->option->debug_frozen_string_literal ||
RTEST(
ruby_debug)) {
5157 VALUE lit = get_string_value(node);
5158 VALUE str = rb_str_with_debug_created_info(lit, rb_iseq_path(iseq), (
int)nd_line(node));
5159 RB_OBJ_SET_SHAREABLE(str);
5163 return get_string_value(node);
5166 rb_bug(
"unexpected node: %s", ruby_node_name(nd_type(node)));
5173 const NODE *line_node = node;
5175 if (nd_type_p(node, NODE_ZLIST)) {
5177 ADD_INSN1(ret, line_node, newarray,
INT2FIX(0));
5182 EXPECT_NODE(
"compile_array", node, NODE_LIST, -1);
5185 for (; node; node = RNODE_LIST(node)->nd_next) {
5186 NO_CHECK(COMPILE_(ret,
"array element", RNODE_LIST(node)->nd_head, popped));
5228 const int max_stack_len = 0x100;
5229 const int min_tmp_ary_len = 0x40;
5233#define FLUSH_CHUNK \
5235 if (first_chunk) ADD_INSN1(ret, line_node, newarray, INT2FIX(stack_len)); \
5236 else ADD_INSN1(ret, line_node, pushtoarray, INT2FIX(stack_len)); \
5237 first_chunk = FALSE; \
5245 if (static_literal_node_p(RNODE_LIST(node)->nd_head, iseq,
false)) {
5247 const NODE *node_tmp = RNODE_LIST(node)->nd_next;
5248 for (; node_tmp && static_literal_node_p(RNODE_LIST(node_tmp)->nd_head, iseq,
false); node_tmp = RNODE_LIST(node_tmp)->nd_next)
5251 if ((first_chunk && stack_len == 0 && !node_tmp) || count >= min_tmp_ary_len) {
5256 for (; count; count--, node = RNODE_LIST(node)->nd_next)
5257 rb_ary_push(ary, static_literal_value(RNODE_LIST(node)->nd_head, iseq));
5258 RB_OBJ_SET_FROZEN_SHAREABLE(ary);
5263 ADD_INSN1(ret, line_node, duparray, ary);
5264 first_chunk = FALSE;
5267 ADD_INSN1(ret, line_node, putobject, ary);
5268 ADD_INSN(ret, line_node, concattoarray);
5270 RB_OBJ_SET_SHAREABLE(ary);
5276 for (; count; count--, node = RNODE_LIST(node)->nd_next) {
5278 EXPECT_NODE(
"compile_array", node, NODE_LIST, -1);
5281 if (!RNODE_LIST(node)->nd_next && keyword_node_p(RNODE_LIST(node)->nd_head)) {
5283 if (stack_len == 0 && first_chunk) {
5284 ADD_INSN1(ret, line_node, newarray,
INT2FIX(0));
5289 NO_CHECK(COMPILE_(ret,
"array element", RNODE_LIST(node)->nd_head, 0));
5290 ADD_INSN(ret, line_node, pushtoarraykwsplat);
5294 NO_CHECK(COMPILE_(ret,
"array element", RNODE_LIST(node)->nd_head, 0));
5299 if (stack_len >= max_stack_len) FLUSH_CHUNK;
5309static_literal_node_pair_p(
const NODE *node,
const rb_iseq_t *iseq)
5311 return RNODE_LIST(node)->nd_head && static_literal_node_p(RNODE_LIST(node)->nd_head, iseq,
true) && static_literal_node_p(RNODE_LIST(RNODE_LIST(node)->nd_next)->nd_head, iseq,
false);
5317 const NODE *line_node = node;
5319 node = RNODE_HASH(node)->nd_head;
5321 if (!node || nd_type_p(node, NODE_ZLIST)) {
5323 ADD_INSN1(ret, line_node, newhash,
INT2FIX(0));
5328 EXPECT_NODE(
"compile_hash", node, NODE_LIST, -1);
5331 for (; node; node = RNODE_LIST(node)->nd_next) {
5332 NO_CHECK(COMPILE_(ret,
"hash element", RNODE_LIST(node)->nd_head, popped));
5355 const int max_stack_len = 0x100;
5356 const int min_tmp_hash_len = 0x800;
5358 int first_chunk = 1;
5359 DECL_ANCHOR(anchor);
5360 INIT_ANCHOR(anchor);
5363#define FLUSH_CHUNK() \
5365 if (first_chunk) { \
5366 APPEND_LIST(ret, anchor); \
5367 ADD_INSN1(ret, line_node, newhash, INT2FIX(stack_len)); \
5370 ADD_INSN1(ret, line_node, putspecialobject, INT2FIX(VM_SPECIAL_OBJECT_VMCORE)); \
5371 ADD_INSN(ret, line_node, swap); \
5372 APPEND_LIST(ret, anchor); \
5373 ADD_SEND(ret, line_node, id_core_hash_merge_ptr, INT2FIX(stack_len + 1)); \
5375 INIT_ANCHOR(anchor); \
5376 first_chunk = stack_len = 0; \
5383 if (static_literal_node_pair_p(node, iseq)) {
5385 const NODE *node_tmp = RNODE_LIST(RNODE_LIST(node)->nd_next)->nd_next;
5386 for (; node_tmp && static_literal_node_pair_p(node_tmp, iseq); node_tmp = RNODE_LIST(RNODE_LIST(node_tmp)->nd_next)->nd_next)
5389 if ((first_chunk && stack_len == 0 && !node_tmp) || count >= min_tmp_hash_len) {
5394 for (; count; count--, node = RNODE_LIST(RNODE_LIST(node)->nd_next)->nd_next) {
5396 elem[0] = static_literal_value(RNODE_LIST(node)->nd_head, iseq);
5398 elem[1] = static_literal_value(RNODE_LIST(RNODE_LIST(node)->nd_next)->nd_head, iseq);
5404 hash = RB_OBJ_SET_FROZEN_SHAREABLE(
rb_obj_hide(hash));
5409 ADD_INSN1(ret, line_node, duphash, hash);
5413 ADD_INSN1(ret, line_node, putspecialobject,
INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
5414 ADD_INSN(ret, line_node, swap);
5416 ADD_INSN1(ret, line_node, putobject, hash);
5418 ADD_SEND(ret, line_node, id_core_hash_merge_kwd,
INT2FIX(2));
5425 for (; count; count--, node = RNODE_LIST(RNODE_LIST(node)->nd_next)->nd_next) {
5428 EXPECT_NODE(
"compile_hash", node, NODE_LIST, -1);
5431 if (RNODE_LIST(node)->nd_head) {
5433 NO_CHECK(COMPILE_(anchor,
"hash key element", RNODE_LIST(node)->nd_head, 0));
5434 NO_CHECK(COMPILE_(anchor,
"hash value element", RNODE_LIST(RNODE_LIST(node)->nd_next)->nd_head, 0));
5438 if (stack_len >= max_stack_len) FLUSH_CHUNK();
5444 const NODE *kw = RNODE_LIST(RNODE_LIST(node)->nd_next)->nd_head;
5445 int empty_kw = nd_type_p(kw, NODE_HASH) && (!RNODE_HASH(kw)->nd_head);
5446 int first_kw = first_chunk && stack_len == 0;
5447 int last_kw = !RNODE_LIST(RNODE_LIST(node)->nd_next)->nd_next;
5448 int only_kw = last_kw && first_kw;
5450 empty_kw = empty_kw || nd_type_p(kw, NODE_NIL);
5452 if (only_kw && method_call_keywords) {
5460 NO_CHECK(COMPILE(ret,
"keyword splat", kw));
5462 else if (first_kw) {
5466 ADD_INSN1(ret, line_node, newhash,
INT2FIX(0));
5473 if (only_kw && method_call_keywords) {
5479 NO_CHECK(COMPILE(ret,
"keyword splat", kw));
5486 ADD_INSN1(ret, line_node, putspecialobject,
INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
5487 if (first_kw) ADD_INSN1(ret, line_node, newhash,
INT2FIX(0));
5488 else ADD_INSN(ret, line_node, swap);
5490 NO_CHECK(COMPILE(ret,
"keyword splat", kw));
5492 ADD_SEND(ret, line_node, id_core_hash_merge_kwd,
INT2FIX(2));
5507rb_node_case_when_optimizable_literal(
const NODE *
const node)
5509 switch (nd_type(node)) {
5511 return rb_node_integer_literal_val(node);
5513 VALUE v = rb_node_float_literal_val(node);
5522 case NODE_IMAGINARY:
5531 return rb_node_sym_string_val(node);
5533 return rb_node_line_lineno_val(node);
5535 return rb_node_str_string_val(node);
5537 return rb_node_file_path_val(node);
5544 LABEL *l1,
int only_special_literals,
VALUE literals)
5547 const NODE *val = RNODE_LIST(vals)->nd_head;
5548 VALUE lit = rb_node_case_when_optimizable_literal(val);
5551 only_special_literals = 0;
5553 else if (
NIL_P(rb_hash_lookup(literals, lit))) {
5554 rb_hash_aset(literals, lit, (
VALUE)(l1) | 1);
5557 if (nd_type_p(val, NODE_STR) || nd_type_p(val, NODE_FILE)) {
5558 debugp_param(
"nd_lit", get_string_value(val));
5559 lit = get_string_value(val);
5560 ADD_INSN1(cond_seq, val, putobject, lit);
5564 if (!COMPILE(cond_seq,
"when cond", val))
return -1;
5568 ADD_INSN1(cond_seq, vals, topn,
INT2FIX(1));
5569 ADD_CALL(cond_seq, vals, idEqq,
INT2FIX(1));
5570 ADD_INSNL(cond_seq, val, branchif, l1);
5571 vals = RNODE_LIST(vals)->nd_next;
5573 return only_special_literals;
5578 LABEL *l1,
int only_special_literals,
VALUE literals)
5580 const NODE *line_node = vals;
5582 switch (nd_type(vals)) {
5584 if (when_vals(iseq, cond_seq, vals, l1, only_special_literals, literals) < 0)
5588 ADD_INSN (cond_seq, line_node, dup);
5589 CHECK(COMPILE(cond_seq,
"when splat", RNODE_SPLAT(vals)->nd_head));
5590 ADD_INSN1(cond_seq, line_node, splatarray,
Qfalse);
5591 ADD_INSN1(cond_seq, line_node, checkmatch,
INT2FIX(VM_CHECKMATCH_TYPE_CASE | VM_CHECKMATCH_ARRAY));
5592 ADD_INSNL(cond_seq, line_node, branchif, l1);
5595 CHECK(when_splat_vals(iseq, cond_seq, RNODE_ARGSCAT(vals)->nd_head, l1, only_special_literals, literals));
5596 CHECK(when_splat_vals(iseq, cond_seq, RNODE_ARGSCAT(vals)->nd_body, l1, only_special_literals, literals));
5599 CHECK(when_splat_vals(iseq, cond_seq, RNODE_ARGSPUSH(vals)->nd_head, l1, only_special_literals, literals));
5600 ADD_INSN (cond_seq, line_node, dup);
5601 CHECK(COMPILE(cond_seq,
"when argspush body", RNODE_ARGSPUSH(vals)->nd_body));
5602 ADD_INSN1(cond_seq, line_node, checkmatch,
INT2FIX(VM_CHECKMATCH_TYPE_CASE));
5603 ADD_INSNL(cond_seq, line_node, branchif, l1);
5606 ADD_INSN (cond_seq, line_node, dup);
5607 CHECK(COMPILE(cond_seq,
"when val", vals));
5608 ADD_INSN1(cond_seq, line_node, splatarray,
Qfalse);
5609 ADD_INSN1(cond_seq, line_node, checkmatch,
INT2FIX(VM_CHECKMATCH_TYPE_CASE | VM_CHECKMATCH_ARRAY));
5610 ADD_INSNL(cond_seq, line_node, branchif, l1);
5703 const NODE *line_node;
5718add_masgn_lhs_node(
struct masgn_state *state,
int lhs_pos,
const NODE *line_node,
int argc,
INSN *before_insn)
5721 rb_bug(
"no masgn_state");
5730 memo->before_insn = before_insn;
5731 memo->line_node = line_node;
5732 memo->argn = state->num_args + 1;
5733 memo->num_args = argc;
5734 state->num_args += argc;
5735 memo->lhs_pos = lhs_pos;
5737 if (!state->first_memo) {
5738 state->first_memo = memo;
5741 state->last_memo->next = memo;
5743 state->last_memo = memo;
5753 switch (nd_type(node)) {
5754 case NODE_ATTRASGN: {
5756 const NODE *line_node = node;
5758 CHECK(COMPILE_POPPED(pre,
"masgn lhs (NODE_ATTRASGN)", node));
5760 bool safenav_call =
false;
5762 iobj = (
INSN *)get_prev_insn((
INSN *)insn_element);
5764 ELEM_REMOVE(insn_element);
5765 if (!IS_INSN_ID(iobj, send)) {
5766 safenav_call =
true;
5767 iobj = (
INSN *)get_prev_insn(iobj);
5768 ELEM_INSERT_NEXT(&iobj->link, insn_element);
5770 (pre->last = iobj->link.prev)->next = 0;
5773 int argc = vm_ci_argc(ci) + 1;
5774 ci = ci_argc_set(iseq, ci, argc);
5775 OPERAND_AT(iobj, 0) = (
VALUE)ci;
5779 ADD_INSN(lhs, line_node, swap);
5782 ADD_INSN1(lhs, line_node, topn,
INT2FIX(argc));
5785 if (!add_masgn_lhs_node(state, lhs_pos, line_node, argc, (
INSN *)LAST_ELEMENT(lhs))) {
5789 iobj->link.prev = lhs->last;
5790 lhs->last->next = &iobj->link;
5791 for (lhs->last = &iobj->link; lhs->last->next; lhs->last = lhs->last->next);
5792 if (vm_ci_flag(ci) & VM_CALL_ARGS_SPLAT) {
5793 int argc = vm_ci_argc(ci);
5794 bool dupsplat =
false;
5795 ci = ci_argc_set(iseq, ci, argc - 1);
5796 if (!(vm_ci_flag(ci) & VM_CALL_ARGS_SPLAT_MUT)) {
5803 ci = ci_flag_set(iseq, ci, VM_CALL_ARGS_SPLAT_MUT);
5805 OPERAND_AT(iobj, 0) = (
VALUE)ci;
5814 int line_no = nd_line(line_node);
5815 int node_id = nd_node_id(line_node);
5818 INSERT_BEFORE_INSN(iobj, line_no, node_id, swap);
5819 INSERT_BEFORE_INSN1(iobj, line_no, node_id, splatarray,
Qtrue);
5820 INSERT_BEFORE_INSN(iobj, line_no, node_id, swap);
5822 INSERT_BEFORE_INSN1(iobj, line_no, node_id, pushtoarray,
INT2FIX(1));
5824 if (!safenav_call) {
5825 ADD_INSN(lhs, line_node, pop);
5827 ADD_INSN(lhs, line_node, pop);
5830 for (
int i=0; i < argc; i++) {
5831 ADD_INSN(post, line_node, pop);
5836 DECL_ANCHOR(nest_rhs);
5837 INIT_ANCHOR(nest_rhs);
5838 DECL_ANCHOR(nest_lhs);
5839 INIT_ANCHOR(nest_lhs);
5841 int prev_level = state->lhs_level;
5842 bool prev_nested = state->nested;
5844 state->lhs_level = lhs_pos - 1;
5845 CHECK(compile_massign0(iseq, pre, nest_rhs, nest_lhs, post, node, state, 1));
5846 state->lhs_level = prev_level;
5847 state->nested = prev_nested;
5849 ADD_SEQ(lhs, nest_rhs);
5850 ADD_SEQ(lhs, nest_lhs);
5854 if (!RNODE_CDECL(node)->nd_vid) {
5858 CHECK(COMPILE_POPPED(pre,
"masgn lhs (NODE_CDECL)", node));
5861 iobj = (
INSN *)insn_element;
5864 ELEM_REMOVE(insn_element);
5865 pre->last = iobj->link.prev;
5868 if (!add_masgn_lhs_node(state, lhs_pos, node, 1, (
INSN *)LAST_ELEMENT(lhs))) {
5872 ADD_INSN(post, node, pop);
5877 DECL_ANCHOR(anchor);
5878 INIT_ANCHOR(anchor);
5879 CHECK(COMPILE_POPPED(anchor,
"masgn lhs", node));
5880 ELEM_REMOVE(FIRST_ELEMENT(anchor));
5881 ADD_SEQ(lhs, anchor);
5892 CHECK(compile_massign_opt_lhs(iseq, ret, RNODE_LIST(lhsn)->nd_next));
5893 CHECK(compile_massign_lhs(iseq, ret, ret, ret, ret, RNODE_LIST(lhsn)->nd_head, NULL, 0));
5900 const NODE *rhsn,
const NODE *orig_lhsn)
5903 const int memsize = numberof(mem);
5905 int llen = 0, rlen = 0;
5907 const NODE *lhsn = orig_lhsn;
5909#define MEMORY(v) { \
5911 if (memindex == memsize) return 0; \
5912 for (i=0; i<memindex; i++) { \
5913 if (mem[i] == (v)) return 0; \
5915 mem[memindex++] = (v); \
5918 if (rhsn == 0 || !nd_type_p(rhsn, NODE_LIST)) {
5923 const NODE *ln = RNODE_LIST(lhsn)->nd_head;
5924 switch (nd_type(ln)) {
5929 MEMORY(get_nd_vid(ln));
5934 lhsn = RNODE_LIST(lhsn)->nd_next;
5940 NO_CHECK(COMPILE_POPPED(ret,
"masgn val (popped)", RNODE_LIST(rhsn)->nd_head));
5943 NO_CHECK(COMPILE(ret,
"masgn val", RNODE_LIST(rhsn)->nd_head));
5945 rhsn = RNODE_LIST(rhsn)->nd_next;
5950 for (i=0; i<llen-rlen; i++) {
5951 ADD_INSN(ret, orig_lhsn, putnil);
5955 compile_massign_opt_lhs(iseq, ret, orig_lhsn);
5962 const NODE *rhsn = RNODE_MASGN(node)->nd_value;
5963 const NODE *splatn = RNODE_MASGN(node)->nd_args;
5964 const NODE *lhsn = RNODE_MASGN(node)->nd_head;
5965 const NODE *lhsn_count = lhsn;
5966 int lhs_splat = (splatn && NODE_NAMED_REST_P(splatn)) ? 1 : 0;
5971 while (lhsn_count) {
5973 lhsn_count = RNODE_LIST(lhsn_count)->nd_next;
5976 CHECK(compile_massign_lhs(iseq, pre, rhs, lhs, post, RNODE_LIST(lhsn)->nd_head, state, (llen - lpos) + lhs_splat + state->lhs_level));
5978 lhsn = RNODE_LIST(lhsn)->nd_next;
5982 if (nd_type_p(splatn, NODE_POSTARG)) {
5984 const NODE *postn = RNODE_POSTARG(splatn)->nd_2nd;
5985 const NODE *restn = RNODE_POSTARG(splatn)->nd_1st;
5986 int plen = (int)RNODE_LIST(postn)->as.nd_alen;
5988 int flag = 0x02 | (NODE_NAMED_REST_P(restn) ? 0x01 : 0x00);
5990 ADD_INSN2(lhs, splatn, expandarray,
INT2FIX(plen),
INT2FIX(flag));
5992 if (NODE_NAMED_REST_P(restn)) {
5993 CHECK(compile_massign_lhs(iseq, pre, rhs, lhs, post, restn, state, 1 + plen + state->lhs_level));
5996 CHECK(compile_massign_lhs(iseq, pre, rhs, lhs, post, RNODE_LIST(postn)->nd_head, state, (plen - ppos) + state->lhs_level));
5998 postn = RNODE_LIST(postn)->nd_next;
6003 CHECK(compile_massign_lhs(iseq, pre, rhs, lhs, post, splatn, state, 1 + state->lhs_level));
6007 if (!state->nested) {
6008 NO_CHECK(COMPILE(rhs,
"normal masgn rhs", rhsn));
6012 ADD_INSN(rhs, node, dup);
6014 ADD_INSN2(rhs, node, expandarray,
INT2FIX(llen),
INT2FIX(lhs_splat));
6021 if (!popped || RNODE_MASGN(node)->nd_args || !compile_massign_opt(iseq, ret, RNODE_MASGN(node)->nd_value, RNODE_MASGN(node)->nd_head)) {
6023 state.lhs_level = popped ? 0 : 1;
6026 state.first_memo = NULL;
6027 state.last_memo = NULL;
6037 int ok = compile_massign0(iseq, pre, rhs, lhs, post, node, &state, popped);
6041 VALUE topn_arg =
INT2FIX((state.num_args - memo->argn) + memo->lhs_pos);
6042 for (
int i = 0; i < memo->num_args; i++) {
6043 INSERT_BEFORE_INSN1(memo->before_insn, nd_line(memo->line_node), nd_node_id(memo->line_node), topn, topn_arg);
6045 tmp_memo = memo->next;
6054 if (!popped && state.num_args >= 1) {
6056 ADD_INSN1(ret, node, setn,
INT2FIX(state.num_args));
6068 switch (nd_type(node)) {
6070 rb_ary_unshift(arr,
ID2SYM(RNODE_CONST(node)->nd_vid));
6071 RB_OBJ_SET_SHAREABLE(arr);
6074 rb_ary_unshift(arr,
ID2SYM(RNODE_COLON3(node)->nd_mid));
6075 rb_ary_unshift(arr,
ID2SYM(idNULL));
6076 RB_OBJ_SET_SHAREABLE(arr);
6079 rb_ary_unshift(arr,
ID2SYM(RNODE_COLON2(node)->nd_mid));
6080 node = RNODE_COLON2(node)->nd_head;
6089compile_const_prefix(
rb_iseq_t *iseq,
const NODE *
const node,
6092 switch (nd_type(node)) {
6094 debugi(
"compile_const_prefix - colon", RNODE_CONST(node)->nd_vid);
6095 ADD_INSN1(body, node, putobject,
Qtrue);
6096 ADD_INSN1(body, node, getconstant,
ID2SYM(RNODE_CONST(node)->nd_vid));
6099 debugi(
"compile_const_prefix - colon3", RNODE_COLON3(node)->nd_mid);
6100 ADD_INSN(body, node, pop);
6101 ADD_INSN1(body, node, putobject,
rb_cObject);
6102 ADD_INSN1(body, node, putobject,
Qtrue);
6103 ADD_INSN1(body, node, getconstant,
ID2SYM(RNODE_COLON3(node)->nd_mid));
6106 CHECK(compile_const_prefix(iseq, RNODE_COLON2(node)->nd_head, pref, body));
6107 debugi(
"compile_const_prefix - colon2", RNODE_COLON2(node)->nd_mid);
6108 ADD_INSN1(body, node, putobject,
Qfalse);
6109 ADD_INSN1(body, node, getconstant,
ID2SYM(RNODE_COLON2(node)->nd_mid));
6112 CHECK(COMPILE(pref,
"const colon2 prefix", node));
6121 if (nd_type_p(cpath, NODE_COLON3)) {
6123 ADD_INSN1(ret, cpath, putobject,
rb_cObject);
6124 return VM_DEFINECLASS_FLAG_SCOPED;
6126 else if (nd_type_p(cpath, NODE_COLON2) && RNODE_COLON2(cpath)->nd_head) {
6128 NO_CHECK(COMPILE(ret,
"nd_else->nd_head", RNODE_COLON2(cpath)->nd_head));
6129 return VM_DEFINECLASS_FLAG_SCOPED;
6133 ADD_INSN1(ret, cpath, putspecialobject,
6134 INT2FIX(VM_SPECIAL_OBJECT_CONST_BASE));
6140private_recv_p(
const NODE *node)
6142 NODE *recv = get_nd_recv(node);
6143 if (recv && nd_type_p(recv, NODE_SELF)) {
6144 return RNODE_SELF(recv)->nd_state != 0;
6151 const NODE *
const node,
LABEL **lfinish,
VALUE needstr,
bool ignore);
6154compile_call(
rb_iseq_t *iseq,
LINK_ANCHOR *
const ret,
const NODE *
const node,
const enum node_type
type,
const NODE *
const line_node,
int popped,
bool assume_receiver);
6161 enum defined_type expr_type = DEFINED_NOT_DEFINED;
6162 enum node_type
type;
6163 const int line = nd_line(node);
6164 const NODE *line_node = node;
6166 switch (
type = nd_type(node)) {
6170 expr_type = DEFINED_NIL;
6173 expr_type = DEFINED_SELF;
6176 expr_type = DEFINED_TRUE;
6179 expr_type = DEFINED_FALSE;
6184 const NODE *vals = (nd_type(node) == NODE_HASH) ? RNODE_HASH(node)->nd_head : node;
6188 if (RNODE_LIST(vals)->nd_head) {
6189 defined_expr0(iseq, ret, RNODE_LIST(vals)->nd_head, lfinish,
Qfalse,
false);
6192 lfinish[1] = NEW_LABEL(line);
6194 ADD_INSNL(ret, line_node, branchunless, lfinish[1]);
6196 }
while ((vals = RNODE_LIST(vals)->nd_next) != NULL);
6209 case NODE_IMAGINARY:
6214 expr_type = DEFINED_EXPR;
6218 defined_expr0(iseq, ret, RNODE_LIST(node)->nd_head, lfinish,
Qfalse,
false);
6220 lfinish[1] = NEW_LABEL(line);
6222 ADD_INSNL(ret, line_node, branchunless, lfinish[1]);
6223 expr_type = DEFINED_EXPR;
6229 expr_type = DEFINED_LVAR;
6232#define PUSH_VAL(type) (needstr == Qfalse ? Qtrue : rb_iseq_defined_string(type))
6234 ADD_INSN3(ret, line_node, definedivar,
6235 ID2SYM(RNODE_IVAR(node)->nd_vid), get_ivar_ic_value(iseq,RNODE_IVAR(node)->nd_vid), PUSH_VAL(DEFINED_IVAR));
6239 ADD_INSN(ret, line_node, putnil);
6240 ADD_INSN3(ret, line_node, defined,
INT2FIX(DEFINED_GVAR),
6241 ID2SYM(RNODE_GVAR(node)->nd_vid), PUSH_VAL(DEFINED_GVAR));
6245 ADD_INSN(ret, line_node, putnil);
6246 ADD_INSN3(ret, line_node, defined,
INT2FIX(DEFINED_CVAR),
6247 ID2SYM(RNODE_CVAR(node)->nd_vid), PUSH_VAL(DEFINED_CVAR));
6251 ADD_INSN(ret, line_node, putnil);
6252 ADD_INSN3(ret, line_node, defined,
INT2FIX(DEFINED_CONST),
6253 ID2SYM(RNODE_CONST(node)->nd_vid), PUSH_VAL(DEFINED_CONST));
6257 lfinish[1] = NEW_LABEL(line);
6259 defined_expr0(iseq, ret, RNODE_COLON2(node)->nd_head, lfinish,
Qfalse,
false);
6260 ADD_INSNL(ret, line_node, branchunless, lfinish[1]);
6261 NO_CHECK(COMPILE(ret,
"defined/colon2#nd_head", RNODE_COLON2(node)->nd_head));
6264 ADD_INSN3(ret, line_node, defined,
INT2FIX(DEFINED_CONST_FROM),
6265 ID2SYM(RNODE_COLON2(node)->nd_mid), PUSH_VAL(DEFINED_CONST));
6268 ADD_INSN3(ret, line_node, defined,
INT2FIX(DEFINED_METHOD),
6269 ID2SYM(RNODE_COLON2(node)->nd_mid), PUSH_VAL(DEFINED_METHOD));
6273 ADD_INSN1(ret, line_node, putobject,
rb_cObject);
6274 ADD_INSN3(ret, line_node, defined,
6275 INT2FIX(DEFINED_CONST_FROM),
ID2SYM(RNODE_COLON3(node)->nd_mid), PUSH_VAL(DEFINED_CONST));
6283 case NODE_ATTRASGN:{
6284 const int explicit_receiver =
6285 (
type == NODE_CALL ||
type == NODE_OPCALL ||
6286 (
type == NODE_ATTRASGN && !private_recv_p(node)));
6288 if (get_nd_args(node) || explicit_receiver) {
6290 lfinish[1] = NEW_LABEL(line);
6293 lfinish[2] = NEW_LABEL(line);
6296 if (get_nd_args(node)) {
6297 defined_expr0(iseq, ret, get_nd_args(node), lfinish,
Qfalse,
false);
6298 ADD_INSNL(ret, line_node, branchunless, lfinish[1]);
6300 if (explicit_receiver) {
6301 defined_expr0(iseq, ret, get_nd_recv(node), lfinish,
Qfalse,
true);
6302 switch (nd_type(get_nd_recv(node))) {
6308 ADD_INSNL(ret, line_node, branchunless, lfinish[2]);
6309 compile_call(iseq, ret, get_nd_recv(node), nd_type(get_nd_recv(node)), line_node, 0,
true);
6312 ADD_INSNL(ret, line_node, branchunless, lfinish[1]);
6313 NO_CHECK(COMPILE(ret,
"defined/recv", get_nd_recv(node)));
6317 ADD_INSN(ret, line_node, dup);
6319 ADD_INSN3(ret, line_node, defined,
INT2FIX(DEFINED_METHOD),
6320 ID2SYM(get_node_call_nd_mid(node)), PUSH_VAL(DEFINED_METHOD));
6323 ADD_INSN(ret, line_node, putself);
6325 ADD_INSN(ret, line_node, dup);
6327 ADD_INSN3(ret, line_node, defined,
INT2FIX(DEFINED_FUNC),
6328 ID2SYM(get_node_call_nd_mid(node)), PUSH_VAL(DEFINED_METHOD));
6334 ADD_INSN(ret, line_node, putnil);
6335 ADD_INSN3(ret, line_node, defined,
INT2FIX(DEFINED_YIELD), 0,
6336 PUSH_VAL(DEFINED_YIELD));
6337 iseq_set_use_block(ISEQ_BODY(iseq)->local_iseq);
6342 ADD_INSN(ret, line_node, putnil);
6343 ADD_INSN3(ret, line_node, defined,
INT2FIX(DEFINED_REF),
6344 INT2FIX((RNODE_BACK_REF(node)->nd_nth << 1) | (
type == NODE_BACK_REF)),
6345 PUSH_VAL(DEFINED_GVAR));
6350 ADD_INSN(ret, line_node, putnil);
6351 ADD_INSN3(ret, line_node, defined,
INT2FIX(DEFINED_ZSUPER), 0,
6352 PUSH_VAL(DEFINED_ZSUPER));
6358 case NODE_OP_ASGN_OR:
6359 case NODE_OP_ASGN_AND:
6368 expr_type = DEFINED_ASGN;
6375 VALUE str = rb_iseq_defined_string(expr_type);
6376 ADD_INSN1(ret, line_node, putobject, str);
6379 ADD_INSN1(ret, line_node, putobject,
Qtrue);
6386 ADD_SYNTHETIC_INSN(ret, 0, -1, putnil);
6387 iseq_set_exception_local_table(iseq);
6392 const NODE *
const node,
LABEL **lfinish,
VALUE needstr,
bool ignore)
6395 defined_expr0(iseq, ret, node, lfinish, needstr,
false);
6397 int line = nd_line(node);
6398 LABEL *lstart = NEW_LABEL(line);
6399 LABEL *lend = NEW_LABEL(line);
6402 rb_iseq_new_with_callback_new_callback(build_defined_rescue_iseq, NULL);
6403 rescue = NEW_CHILD_ISEQ_WITH_CALLBACK(ifunc,
6405 ISEQ_BODY(iseq)->location.label),
6406 ISEQ_TYPE_RESCUE, 0);
6407 lstart->rescued = LABEL_RESCUE_BEG;
6408 lend->rescued = LABEL_RESCUE_END;
6409 APPEND_LABEL(ret, lcur, lstart);
6410 ADD_LABEL(ret, lend);
6412 ADD_CATCH_ENTRY(CATCH_TYPE_RESCUE, lstart, lend, rescue, lfinish[1]);
6420 const int line = nd_line(node);
6421 const NODE *line_node = node;
6422 if (!RNODE_DEFINED(node)->nd_head) {
6423 VALUE str = rb_iseq_defined_string(DEFINED_NIL);
6424 ADD_INSN1(ret, line_node, putobject, str);
6429 lfinish[0] = NEW_LABEL(line);
6432 defined_expr(iseq, ret, RNODE_DEFINED(node)->nd_head, lfinish, needstr, ignore);
6434 ELEM_INSERT_NEXT(last, &new_insn_body(iseq, nd_line(line_node), nd_node_id(line_node), BIN(putnil), 0)->link);
6435 ADD_INSN(ret, line_node, swap);
6437 ADD_LABEL(ret, lfinish[2]);
6439 ADD_INSN(ret, line_node, pop);
6440 ADD_LABEL(ret, lfinish[1]);
6442 ADD_LABEL(ret, lfinish[0]);
6448make_name_for_block(
const rb_iseq_t *orig_iseq)
6453 if (ISEQ_BODY(orig_iseq)->parent_iseq != 0) {
6454 while (ISEQ_BODY(orig_iseq)->local_iseq != iseq) {
6455 if (ISEQ_BODY(iseq)->
type == ISEQ_TYPE_BLOCK) {
6458 iseq = ISEQ_BODY(iseq)->parent_iseq;
6463 return rb_sprintf(
"block in %"PRIsVALUE, ISEQ_BODY(iseq)->location.label);
6466 return rb_sprintf(
"block (%d levels) in %"PRIsVALUE, level, ISEQ_BODY(iseq)->location.label);
6475 enl->ensure_node = node;
6476 enl->prev = ISEQ_COMPILE_DATA(iseq)->ensure_node_stack;
6478 ISEQ_COMPILE_DATA(iseq)->ensure_node_stack = enl;
6488 while (erange->next != 0) {
6489 erange = erange->next;
6493 ne->end = erange->end;
6494 erange->end = lstart;
6500can_add_ensure_iseq(
const rb_iseq_t *iseq)
6503 if (ISEQ_COMPILE_DATA(iseq)->in_rescue && (e = ISEQ_COMPILE_DATA(iseq)->ensure_node_stack) != NULL) {
6505 if (e->ensure_node)
return false;
6518 ISEQ_COMPILE_DATA(iseq)->ensure_node_stack;
6520 DECL_ANCHOR(ensure);
6522 INIT_ANCHOR(ensure);
6524 if (enlp->erange != NULL) {
6525 DECL_ANCHOR(ensure_part);
6526 LABEL *lstart = NEW_LABEL(0);
6527 LABEL *lend = NEW_LABEL(0);
6528 INIT_ANCHOR(ensure_part);
6530 add_ensure_range(iseq, enlp->erange, lstart, lend);
6532 ISEQ_COMPILE_DATA(iseq)->ensure_node_stack = enlp->prev;
6533 ADD_LABEL(ensure_part, lstart);
6534 NO_CHECK(COMPILE_POPPED(ensure_part,
"ensure part", enlp->ensure_node));
6535 ADD_LABEL(ensure_part, lend);
6536 ADD_SEQ(ensure, ensure_part);
6545 ISEQ_COMPILE_DATA(iseq)->ensure_node_stack = prev_enlp;
6546 ADD_SEQ(ret, ensure);
6551check_keyword(
const NODE *node)
6555 if (nd_type_p(node, NODE_LIST)) {
6556 while (RNODE_LIST(node)->nd_next) {
6557 node = RNODE_LIST(node)->nd_next;
6559 node = RNODE_LIST(node)->nd_head;
6562 return keyword_node_p(node);
6567keyword_node_single_splat_p(
NODE *kwnode)
6571 NODE *node = RNODE_HASH(kwnode)->nd_head;
6572 return RNODE_LIST(node)->nd_head == NULL &&
6573 RNODE_LIST(RNODE_LIST(node)->nd_next)->nd_next == NULL;
6578 NODE *kwnode,
unsigned int *flag_ptr)
6580 *flag_ptr |= VM_CALL_KW_SPLAT_MUT;
6581 ADD_INSN1(args, argn, putspecialobject,
INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
6582 ADD_INSN1(args, argn, newhash,
INT2FIX(0));
6583 compile_hash(iseq, args, kwnode, TRUE, FALSE);
6584 ADD_SEND(args, argn, id_core_hash_merge_kwd,
INT2FIX(2));
6587#define SPLATARRAY_FALSE 0
6588#define SPLATARRAY_TRUE 1
6589#define DUP_SINGLE_KW_SPLAT 2
6593 unsigned int *dup_rest,
unsigned int *flag_ptr,
struct rb_callinfo_kwarg **kwarg_ptr)
6595 if (!argn)
return 0;
6597 NODE *kwnode = NULL;
6599 switch (nd_type(argn)) {
6602 int len = compile_args(iseq, args, argn, &kwnode);
6603 RUBY_ASSERT(flag_ptr == NULL || (*flag_ptr & VM_CALL_ARGS_SPLAT) == 0);
6606 if (compile_keyword_arg(iseq, args, kwnode, kwarg_ptr, flag_ptr)) {
6610 if (keyword_node_single_splat_p(kwnode) && (*dup_rest & DUP_SINGLE_KW_SPLAT)) {
6611 compile_single_keyword_splat_mutable(iseq, args, argn, kwnode, flag_ptr);
6614 compile_hash(iseq, args, kwnode, TRUE, FALSE);
6623 NO_CHECK(COMPILE(args,
"args (splat)", RNODE_SPLAT(argn)->nd_head));
6624 ADD_INSN1(args, argn, splatarray, RBOOL(*dup_rest & SPLATARRAY_TRUE));
6625 if (*dup_rest & SPLATARRAY_TRUE) *dup_rest &= ~SPLATARRAY_TRUE;
6626 if (flag_ptr) *flag_ptr |= VM_CALL_ARGS_SPLAT;
6627 RUBY_ASSERT(flag_ptr == NULL || (*flag_ptr & VM_CALL_KW_SPLAT) == 0);
6630 case NODE_ARGSCAT: {
6631 if (flag_ptr) *flag_ptr |= VM_CALL_ARGS_SPLAT;
6632 int argc = setup_args_core(iseq, args, RNODE_ARGSCAT(argn)->nd_head, dup_rest, NULL, NULL);
6633 bool args_pushed =
false;
6635 if (nd_type_p(RNODE_ARGSCAT(argn)->nd_body, NODE_LIST)) {
6636 int rest_len = compile_args(iseq, args, RNODE_ARGSCAT(argn)->nd_body, &kwnode);
6637 if (kwnode) rest_len--;
6638 ADD_INSN1(args, argn, pushtoarray,
INT2FIX(rest_len));
6642 RUBY_ASSERT(!check_keyword(RNODE_ARGSCAT(argn)->nd_body));
6643 NO_CHECK(COMPILE(args,
"args (cat: splat)", RNODE_ARGSCAT(argn)->nd_body));
6646 if (nd_type_p(RNODE_ARGSCAT(argn)->nd_head, NODE_LIST)) {
6647 ADD_INSN1(args, argn, splatarray, RBOOL(*dup_rest & SPLATARRAY_TRUE));
6648 if (*dup_rest & SPLATARRAY_TRUE) *dup_rest &= ~SPLATARRAY_TRUE;
6651 else if (!args_pushed) {
6652 ADD_INSN(args, argn, concattoarray);
6658 *flag_ptr |= VM_CALL_KW_SPLAT;
6659 compile_hash(iseq, args, kwnode, TRUE, FALSE);
6665 case NODE_ARGSPUSH: {
6666 if (flag_ptr) *flag_ptr |= VM_CALL_ARGS_SPLAT;
6667 int argc = setup_args_core(iseq, args, RNODE_ARGSPUSH(argn)->nd_head, dup_rest, NULL, NULL);
6669 if (nd_type_p(RNODE_ARGSPUSH(argn)->nd_body, NODE_LIST)) {
6670 int rest_len = compile_args(iseq, args, RNODE_ARGSPUSH(argn)->nd_body, &kwnode);
6671 if (kwnode) rest_len--;
6672 ADD_INSN1(args, argn, newarray,
INT2FIX(rest_len));
6673 ADD_INSN1(args, argn, pushtoarray,
INT2FIX(1));
6676 if (keyword_node_p(RNODE_ARGSPUSH(argn)->nd_body)) {
6677 kwnode = RNODE_ARGSPUSH(argn)->nd_body;
6680 NO_CHECK(COMPILE(args,
"args (cat: splat)", RNODE_ARGSPUSH(argn)->nd_body));
6681 ADD_INSN1(args, argn, pushtoarray,
INT2FIX(1));
6687 *flag_ptr |= VM_CALL_KW_SPLAT;
6688 if (!keyword_node_single_splat_p(kwnode)) {
6689 *flag_ptr |= VM_CALL_KW_SPLAT_MUT;
6690 compile_hash(iseq, args, kwnode, TRUE, FALSE);
6692 else if (*dup_rest & DUP_SINGLE_KW_SPLAT) {
6693 compile_single_keyword_splat_mutable(iseq, args, argn, kwnode, flag_ptr);
6696 compile_hash(iseq, args, kwnode, TRUE, FALSE);
6704 UNKNOWN_NODE(
"setup_arg", argn,
Qnil);
6710setup_args_splat_mut(
unsigned int *flag,
int dup_rest,
int initial_dup_rest)
6712 if ((*flag & VM_CALL_ARGS_SPLAT) && dup_rest != initial_dup_rest) {
6713 *flag |= VM_CALL_ARGS_SPLAT_MUT;
6718setup_args_dup_rest_p(
const NODE *argn)
6720 switch(nd_type(argn)) {
6731 case NODE_IMAGINARY:
6744 return setup_args_dup_rest_p(RNODE_COLON2(argn)->nd_head);
6747 if (setup_args_dup_rest_p(RNODE_LIST(argn)->nd_head)) {
6750 argn = RNODE_LIST(argn)->nd_next;
6763 unsigned int dup_rest = SPLATARRAY_TRUE, initial_dup_rest;
6766 const NODE *check_arg = nd_type_p(argn, NODE_BLOCK_PASS) ?
6767 RNODE_BLOCK_PASS(argn)->nd_head : argn;
6770 switch(nd_type(check_arg)) {
6773 dup_rest = SPLATARRAY_FALSE;
6777 dup_rest = !nd_type_p(RNODE_ARGSCAT(check_arg)->nd_head, NODE_LIST);
6779 case(NODE_ARGSPUSH):
6781 dup_rest = !((nd_type_p(RNODE_ARGSPUSH(check_arg)->nd_head, NODE_SPLAT) ||
6782 (nd_type_p(RNODE_ARGSPUSH(check_arg)->nd_head, NODE_ARGSCAT) &&
6783 nd_type_p(RNODE_ARGSCAT(RNODE_ARGSPUSH(check_arg)->nd_head)->nd_head, NODE_LIST))) &&
6784 nd_type_p(RNODE_ARGSPUSH(check_arg)->nd_body, NODE_HASH) &&
6785 !RNODE_HASH(RNODE_ARGSPUSH(check_arg)->nd_body)->nd_brace);
6787 if (dup_rest == SPLATARRAY_FALSE) {
6789 NODE *node = RNODE_HASH(RNODE_ARGSPUSH(check_arg)->nd_body)->nd_head;
6791 NODE *key_node = RNODE_LIST(node)->nd_head;
6792 if (key_node && setup_args_dup_rest_p(key_node)) {
6793 dup_rest = SPLATARRAY_TRUE;
6797 node = RNODE_LIST(node)->nd_next;
6798 NODE *value_node = RNODE_LIST(node)->nd_head;
6799 if (setup_args_dup_rest_p(value_node)) {
6800 dup_rest = SPLATARRAY_TRUE;
6804 node = RNODE_LIST(node)->nd_next;
6813 if (check_arg != argn && setup_args_dup_rest_p(RNODE_BLOCK_PASS(argn)->nd_body)) {
6815 dup_rest = SPLATARRAY_TRUE | DUP_SINGLE_KW_SPLAT;
6818 initial_dup_rest = dup_rest;
6820 if (argn && nd_type_p(argn, NODE_BLOCK_PASS)) {
6821 DECL_ANCHOR(arg_block);
6822 INIT_ANCHOR(arg_block);
6824 if (RNODE_BLOCK_PASS(argn)->forwarding && ISEQ_BODY(ISEQ_BODY(iseq)->local_iseq)->param.flags.forwardable) {
6825 int idx = ISEQ_BODY(ISEQ_BODY(iseq)->local_iseq)->local_table_size;
6827 RUBY_ASSERT(nd_type_p(RNODE_BLOCK_PASS(argn)->nd_head, NODE_ARGSPUSH));
6828 const NODE * arg_node =
6829 RNODE_ARGSPUSH(RNODE_BLOCK_PASS(argn)->nd_head)->nd_head;
6836 if (nd_type_p(arg_node, NODE_ARGSCAT)) {
6837 argc += setup_args_core(iseq, args, RNODE_ARGSCAT(arg_node)->nd_head, &dup_rest, flag, keywords);
6840 *flag |= VM_CALL_FORWARDING;
6842 ADD_GETLOCAL(args, argn, idx, get_lvar_level(iseq));
6843 setup_args_splat_mut(flag, dup_rest, initial_dup_rest);
6847 *flag |= VM_CALL_ARGS_BLOCKARG;
6849 NO_CHECK(COMPILE(arg_block,
"block", RNODE_BLOCK_PASS(argn)->nd_body));
6852 if (LIST_INSN_SIZE_ONE(arg_block)) {
6854 if (IS_INSN(elem)) {
6856 if (iobj->insn_id == BIN(getblockparam)) {
6857 iobj->insn_id = BIN(getblockparamproxy);
6861 ret =
INT2FIX(setup_args_core(iseq, args, RNODE_BLOCK_PASS(argn)->nd_head, &dup_rest, flag, keywords));
6862 ADD_SEQ(args, arg_block);
6865 ret =
INT2FIX(setup_args_core(iseq, args, argn, &dup_rest, flag, keywords));
6867 setup_args_splat_mut(flag, dup_rest, initial_dup_rest);
6874 const NODE *body = ptr;
6875 int line = nd_line(body);
6877 const rb_iseq_t *block = NEW_CHILD_ISEQ(body, make_name_for_block(ISEQ_BODY(iseq)->parent_iseq), ISEQ_TYPE_BLOCK, line);
6879 ADD_INSN1(ret, body, putspecialobject,
INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
6880 ADD_CALL_WITH_BLOCK(ret, body, id_core_set_postexe, argc, block);
6882 iseq_set_local_table(iseq, 0, 0);
6890 int line = nd_line(node);
6891 const NODE *line_node = node;
6892 LABEL *fail_label = NEW_LABEL(line), *end_label = NEW_LABEL(line);
6894#if !(defined(NAMED_CAPTURE_BY_SVAR) && NAMED_CAPTURE_BY_SVAR-0)
6895 ADD_INSN1(ret, line_node, getglobal,
ID2SYM(idBACKREF));
6899 ADD_INSN(ret, line_node, dup);
6900 ADD_INSNL(ret, line_node, branchunless, fail_label);
6902 for (vars = node; vars; vars = RNODE_BLOCK(vars)->nd_next) {
6904 if (RNODE_BLOCK(vars)->nd_next) {
6905 ADD_INSN(ret, line_node, dup);
6908 NO_CHECK(COMPILE_POPPED(ret,
"capture", RNODE_BLOCK(vars)->nd_head));
6910 cap = new_insn_send(iseq, nd_line(line_node), nd_node_id(line_node), idAREF,
INT2FIX(1),
6913#if !defined(NAMED_CAPTURE_SINGLE_OPT) || NAMED_CAPTURE_SINGLE_OPT-0
6914 if (!RNODE_BLOCK(vars)->nd_next && vars == node) {
6919 ADD_INSNL(nom, line_node, jump, end_label);
6920 ADD_LABEL(nom, fail_label);
6922 ADD_INSN(nom, line_node, pop);
6923 ADD_INSN(nom, line_node, putnil);
6925 ADD_LABEL(nom, end_label);
6926 (nom->last->next = cap->link.next)->prev = nom->last;
6927 (cap->link.next = nom->anchor.next)->prev = &cap->link;
6932 ADD_INSNL(ret, line_node, jump, end_label);
6933 ADD_LABEL(ret, fail_label);
6934 ADD_INSN(ret, line_node, pop);
6935 for (vars = node; vars; vars = RNODE_BLOCK(vars)->nd_next) {
6937 NO_CHECK(COMPILE_POPPED(ret,
"capture", RNODE_BLOCK(vars)->nd_head));
6939 ((
INSN*)last)->insn_id = BIN(putnil);
6940 ((
INSN*)last)->operand_size = 0;
6942 ADD_LABEL(ret, end_label);
6946optimizable_range_item_p(
const NODE *n)
6948 if (!n)
return FALSE;
6949 switch (nd_type(n)) {
6962optimized_range_item(
const NODE *n)
6964 switch (nd_type(n)) {
6966 return rb_node_line_lineno_val(n);
6968 return rb_node_integer_literal_val(n);
6970 return rb_node_float_literal_val(n);
6972 return rb_node_rational_literal_val(n);
6973 case NODE_IMAGINARY:
6974 return rb_node_imaginary_literal_val(n);
6978 rb_bug(
"unexpected node: %s", ruby_node_name(nd_type(n)));
6985 const NODE *
const node_body =
type == NODE_IF ? RNODE_IF(node)->nd_body : RNODE_UNLESS(node)->nd_else;
6986 const NODE *
const node_else =
type == NODE_IF ? RNODE_IF(node)->nd_else : RNODE_UNLESS(node)->nd_body;
6988 const int line = nd_line(node);
6989 const NODE *line_node = node;
6990 DECL_ANCHOR(cond_seq);
6991 LABEL *then_label, *else_label, *end_label;
6994 INIT_ANCHOR(cond_seq);
6995 then_label = NEW_LABEL(line);
6996 else_label = NEW_LABEL(line);
6999 NODE *cond = RNODE_IF(node)->nd_cond;
7000 if (nd_type(cond) == NODE_BLOCK) {
7001 cond = RNODE_BLOCK(cond)->nd_head;
7004 CHECK(compile_branch_condition(iseq, cond_seq, cond, then_label, else_label));
7005 ADD_SEQ(ret, cond_seq);
7007 if (then_label->refcnt && else_label->refcnt) {
7008 branches = decl_branch_base(iseq, PTR2NUM(node), nd_code_loc(node),
type == NODE_IF ?
"if" :
"unless");
7011 if (then_label->refcnt) {
7012 ADD_LABEL(ret, then_label);
7014 DECL_ANCHOR(then_seq);
7015 INIT_ANCHOR(then_seq);
7016 CHECK(COMPILE_(then_seq,
"then", node_body, popped));
7018 if (else_label->refcnt) {
7019 const NODE *
const coverage_node = node_body ? node_body : node;
7020 add_trace_branch_coverage(
7023 nd_code_loc(coverage_node),
7024 nd_node_id(coverage_node),
7026 type == NODE_IF ?
"then" :
"else",
7028 end_label = NEW_LABEL(line);
7029 ADD_INSNL(then_seq, line_node, jump, end_label);
7031 ADD_INSN(then_seq, line_node, pop);
7034 ADD_SEQ(ret, then_seq);
7037 if (else_label->refcnt) {
7038 ADD_LABEL(ret, else_label);
7040 DECL_ANCHOR(else_seq);
7041 INIT_ANCHOR(else_seq);
7042 CHECK(COMPILE_(else_seq,
"else", node_else, popped));
7044 if (then_label->refcnt) {
7045 const NODE *
const coverage_node = node_else ? node_else : node;
7046 add_trace_branch_coverage(
7049 nd_code_loc(coverage_node),
7050 nd_node_id(coverage_node),
7052 type == NODE_IF ?
"else" :
"then",
7055 ADD_SEQ(ret, else_seq);
7059 ADD_LABEL(ret, end_label);
7069 const NODE *node = orig_node;
7070 LABEL *endlabel, *elselabel;
7072 DECL_ANCHOR(body_seq);
7073 DECL_ANCHOR(cond_seq);
7074 int only_special_literals = 1;
7075 VALUE literals = rb_hash_new();
7077 enum node_type
type;
7078 const NODE *line_node;
7083 INIT_ANCHOR(body_seq);
7084 INIT_ANCHOR(cond_seq);
7086 RHASH_TBL_RAW(literals)->type = &cdhash_type;
7088 CHECK(COMPILE(head,
"case base", RNODE_CASE(node)->nd_head));
7090 branches = decl_branch_base(iseq, PTR2NUM(node), nd_code_loc(node),
"case");
7092 node = RNODE_CASE(node)->nd_body;
7093 EXPECT_NODE(
"NODE_CASE", node, NODE_WHEN, COMPILE_NG);
7094 type = nd_type(node);
7095 line = nd_line(node);
7098 endlabel = NEW_LABEL(line);
7099 elselabel = NEW_LABEL(line);
7103 while (
type == NODE_WHEN) {
7106 l1 = NEW_LABEL(line);
7107 ADD_LABEL(body_seq, l1);
7108 ADD_INSN(body_seq, line_node, pop);
7110 const NODE *
const coverage_node = RNODE_WHEN(node)->nd_body ? RNODE_WHEN(node)->nd_body : node;
7111 add_trace_branch_coverage(
7114 nd_code_loc(coverage_node),
7115 nd_node_id(coverage_node),
7120 CHECK(COMPILE_(body_seq,
"when body", RNODE_WHEN(node)->nd_body, popped));
7121 ADD_INSNL(body_seq, line_node, jump, endlabel);
7123 vals = RNODE_WHEN(node)->nd_head;
7125 switch (nd_type(vals)) {
7127 only_special_literals = when_vals(iseq, cond_seq, vals, l1, only_special_literals, literals);
7128 if (only_special_literals < 0)
return COMPILE_NG;
7133 only_special_literals = 0;
7134 CHECK(when_splat_vals(iseq, cond_seq, vals, l1, only_special_literals, literals));
7137 UNKNOWN_NODE(
"NODE_CASE", vals, COMPILE_NG);
7141 EXPECT_NODE_NONULL(
"NODE_CASE", node, NODE_LIST, COMPILE_NG);
7144 node = RNODE_WHEN(node)->nd_next;
7148 type = nd_type(node);
7149 line = nd_line(node);
7154 ADD_LABEL(cond_seq, elselabel);
7155 ADD_INSN(cond_seq, line_node, pop);
7156 add_trace_branch_coverage(iseq, cond_seq, nd_code_loc(node), nd_node_id(node), branch_id,
"else", branches);
7157 CHECK(COMPILE_(cond_seq,
"else", node, popped));
7158 ADD_INSNL(cond_seq, line_node, jump, endlabel);
7161 debugs(
"== else (implicit)\n");
7162 ADD_LABEL(cond_seq, elselabel);
7163 ADD_INSN(cond_seq, orig_node, pop);
7164 add_trace_branch_coverage(iseq, cond_seq, nd_code_loc(orig_node), nd_node_id(orig_node), branch_id,
"else", branches);
7166 ADD_INSN(cond_seq, orig_node, putnil);
7168 ADD_INSNL(cond_seq, orig_node, jump, endlabel);
7171 if (only_special_literals && ISEQ_COMPILE_DATA(iseq)->option->specialized_instruction) {
7172 ADD_INSN(ret, orig_node, dup);
7174 ADD_INSN2(ret, orig_node, opt_case_dispatch, literals, elselabel);
7176 LABEL_REF(elselabel);
7179 ADD_SEQ(ret, cond_seq);
7180 ADD_SEQ(ret, body_seq);
7181 ADD_LABEL(ret, endlabel);
7190 const NODE *node = RNODE_CASE2(orig_node)->nd_body;
7192 DECL_ANCHOR(body_seq);
7196 branches = decl_branch_base(iseq, PTR2NUM(orig_node), nd_code_loc(orig_node),
"case");
7198 INIT_ANCHOR(body_seq);
7199 endlabel = NEW_LABEL(nd_line(node));
7201 while (node && nd_type_p(node, NODE_WHEN)) {
7202 const int line = nd_line(node);
7203 LABEL *l1 = NEW_LABEL(line);
7204 ADD_LABEL(body_seq, l1);
7206 const NODE *
const coverage_node = RNODE_WHEN(node)->nd_body ? RNODE_WHEN(node)->nd_body : node;
7207 add_trace_branch_coverage(
7210 nd_code_loc(coverage_node),
7211 nd_node_id(coverage_node),
7216 CHECK(COMPILE_(body_seq,
"when", RNODE_WHEN(node)->nd_body, popped));
7217 ADD_INSNL(body_seq, node, jump, endlabel);
7219 vals = RNODE_WHEN(node)->nd_head;
7221 EXPECT_NODE_NONULL(
"NODE_WHEN", node, NODE_LIST, COMPILE_NG);
7223 switch (nd_type(vals)) {
7227 val = RNODE_LIST(vals)->nd_head;
7228 lnext = NEW_LABEL(nd_line(val));
7229 debug_compile(
"== when2\n", (
void)0);
7230 CHECK(compile_branch_condition(iseq, ret, val, l1, lnext));
7231 ADD_LABEL(ret, lnext);
7232 vals = RNODE_LIST(vals)->nd_next;
7238 ADD_INSN(ret, vals, putnil);
7239 CHECK(COMPILE(ret,
"when2/cond splat", vals));
7240 ADD_INSN1(ret, vals, checkmatch,
INT2FIX(VM_CHECKMATCH_TYPE_WHEN | VM_CHECKMATCH_ARRAY));
7241 ADD_INSNL(ret, vals, branchif, l1);
7244 UNKNOWN_NODE(
"NODE_WHEN", vals, COMPILE_NG);
7246 node = RNODE_WHEN(node)->nd_next;
7249 const NODE *
const coverage_node = node ? node : orig_node;
7250 add_trace_branch_coverage(
7253 nd_code_loc(coverage_node),
7254 nd_node_id(coverage_node),
7258 CHECK(COMPILE_(ret,
"else", node, popped));
7259 ADD_INSNL(ret, orig_node, jump, endlabel);
7261 ADD_SEQ(ret, body_seq);
7262 ADD_LABEL(ret, endlabel);
7266static int iseq_compile_pattern_match(
rb_iseq_t *iseq,
LINK_ANCHOR *
const ret,
const NODE *
const node,
LABEL *unmatched,
bool in_single_pattern,
bool in_alt_pattern,
int base_index,
bool use_deconstructed_cache);
7268static int iseq_compile_pattern_constant(
rb_iseq_t *iseq,
LINK_ANCHOR *
const ret,
const NODE *
const node,
LABEL *match_failed,
bool in_single_pattern,
int base_index);
7269static int iseq_compile_array_deconstruct(
rb_iseq_t *iseq,
LINK_ANCHOR *
const ret,
const NODE *
const node,
LABEL *deconstruct,
LABEL *deconstructed,
LABEL *match_failed,
LABEL *type_error,
bool in_single_pattern,
int base_index,
bool use_deconstructed_cache);
7270static int iseq_compile_pattern_set_general_errmsg(
rb_iseq_t *iseq,
LINK_ANCHOR *
const ret,
const NODE *
const node,
VALUE errmsg,
int base_index);
7271static int iseq_compile_pattern_set_length_errmsg(
rb_iseq_t *iseq,
LINK_ANCHOR *
const ret,
const NODE *
const node,
VALUE errmsg,
VALUE pattern_length,
int base_index);
7272static int iseq_compile_pattern_set_eqq_errmsg(
rb_iseq_t *iseq,
LINK_ANCHOR *
const ret,
const NODE *
const node,
int base_index);
7274#define CASE3_BI_OFFSET_DECONSTRUCTED_CACHE 0
7275#define CASE3_BI_OFFSET_ERROR_STRING 1
7276#define CASE3_BI_OFFSET_KEY_ERROR_P 2
7277#define CASE3_BI_OFFSET_KEY_ERROR_MATCHEE 3
7278#define CASE3_BI_OFFSET_KEY_ERROR_KEY 4
7281iseq_compile_pattern_each(
rb_iseq_t *iseq,
LINK_ANCHOR *
const ret,
const NODE *
const node,
LABEL *matched,
LABEL *unmatched,
bool in_single_pattern,
bool in_alt_pattern,
int base_index,
bool use_deconstructed_cache)
7283 const int line = nd_line(node);
7284 const NODE *line_node = node;
7286 switch (nd_type(node)) {
7340 const NODE *args = RNODE_ARYPTN(node)->pre_args;
7341 const int pre_args_num = RNODE_ARYPTN(node)->pre_args ?
rb_long2int(RNODE_LIST(RNODE_ARYPTN(node)->pre_args)->as.nd_alen) : 0;
7342 const int post_args_num = RNODE_ARYPTN(node)->post_args ?
rb_long2int(RNODE_LIST(RNODE_ARYPTN(node)->post_args)->as.nd_alen) : 0;
7344 const int min_argc = pre_args_num + post_args_num;
7345 const int use_rest_num = RNODE_ARYPTN(node)->rest_arg && (NODE_NAMED_REST_P(RNODE_ARYPTN(node)->rest_arg) ||
7346 (!NODE_NAMED_REST_P(RNODE_ARYPTN(node)->rest_arg) && post_args_num > 0));
7348 LABEL *match_failed, *type_error, *deconstruct, *deconstructed;
7350 match_failed = NEW_LABEL(line);
7351 type_error = NEW_LABEL(line);
7352 deconstruct = NEW_LABEL(line);
7353 deconstructed = NEW_LABEL(line);
7356 ADD_INSN1(ret, line_node, putobject,
INT2FIX(0));
7357 ADD_INSN(ret, line_node, swap);
7363 CHECK(iseq_compile_pattern_constant(iseq, ret, node, match_failed, in_single_pattern, base_index));
7365 CHECK(iseq_compile_array_deconstruct(iseq, ret, node, deconstruct, deconstructed, match_failed, type_error, in_single_pattern, base_index, use_deconstructed_cache));
7367 ADD_INSN(ret, line_node, dup);
7368 ADD_SEND(ret, line_node, idLength,
INT2FIX(0));
7369 ADD_INSN1(ret, line_node, putobject,
INT2FIX(min_argc));
7370 ADD_SEND(ret, line_node, RNODE_ARYPTN(node)->rest_arg ? idGE : idEq,
INT2FIX(1));
7371 if (in_single_pattern) {
7372 CHECK(iseq_compile_pattern_set_length_errmsg(iseq, ret, node,
7373 RNODE_ARYPTN(node)->rest_arg ? rb_fstring_lit(
"%p length mismatch (given %p, expected %p+)") :
7374 rb_fstring_lit(
"%p length mismatch (given %p, expected %p)"),
7375 INT2FIX(min_argc), base_index + 1 ));
7377 ADD_INSNL(ret, line_node, branchunless, match_failed);
7379 for (i = 0; i < pre_args_num; i++) {
7380 ADD_INSN(ret, line_node, dup);
7381 ADD_INSN1(ret, line_node, putobject,
INT2FIX(i));
7382 ADD_SEND(ret, line_node, idAREF,
INT2FIX(1));
7383 CHECK(iseq_compile_pattern_match(iseq, ret, RNODE_LIST(args)->nd_head, match_failed, in_single_pattern, in_alt_pattern, base_index + 1 ,
false));
7384 args = RNODE_LIST(args)->nd_next;
7387 if (RNODE_ARYPTN(node)->rest_arg) {
7388 if (NODE_NAMED_REST_P(RNODE_ARYPTN(node)->rest_arg)) {
7389 ADD_INSN(ret, line_node, dup);
7390 ADD_INSN1(ret, line_node, putobject,
INT2FIX(pre_args_num));
7391 ADD_INSN1(ret, line_node, topn,
INT2FIX(1));
7392 ADD_SEND(ret, line_node, idLength,
INT2FIX(0));
7393 ADD_INSN1(ret, line_node, putobject,
INT2FIX(min_argc));
7394 ADD_SEND(ret, line_node, idMINUS,
INT2FIX(1));
7395 ADD_INSN1(ret, line_node, setn,
INT2FIX(4));
7396 ADD_SEND(ret, line_node, idAREF,
INT2FIX(2));
7398 CHECK(iseq_compile_pattern_match(iseq, ret, RNODE_ARYPTN(node)->rest_arg, match_failed, in_single_pattern, in_alt_pattern, base_index + 1 ,
false));
7401 if (post_args_num > 0) {
7402 ADD_INSN(ret, line_node, dup);
7403 ADD_SEND(ret, line_node, idLength,
INT2FIX(0));
7404 ADD_INSN1(ret, line_node, putobject,
INT2FIX(min_argc));
7405 ADD_SEND(ret, line_node, idMINUS,
INT2FIX(1));
7406 ADD_INSN1(ret, line_node, setn,
INT2FIX(2));
7407 ADD_INSN(ret, line_node, pop);
7412 args = RNODE_ARYPTN(node)->post_args;
7413 for (i = 0; i < post_args_num; i++) {
7414 ADD_INSN(ret, line_node, dup);
7416 ADD_INSN1(ret, line_node, putobject,
INT2FIX(pre_args_num + i));
7417 ADD_INSN1(ret, line_node, topn,
INT2FIX(3));
7418 ADD_SEND(ret, line_node, idPLUS,
INT2FIX(1));
7420 ADD_SEND(ret, line_node, idAREF,
INT2FIX(1));
7421 CHECK(iseq_compile_pattern_match(iseq, ret, RNODE_LIST(args)->nd_head, match_failed, in_single_pattern, in_alt_pattern, base_index + 1 ,
false));
7422 args = RNODE_LIST(args)->nd_next;
7425 ADD_INSN(ret, line_node, pop);
7427 ADD_INSN(ret, line_node, pop);
7429 ADD_INSNL(ret, line_node, jump, matched);
7430 ADD_INSN(ret, line_node, putnil);
7432 ADD_INSN(ret, line_node, putnil);
7435 ADD_LABEL(ret, type_error);
7436 ADD_INSN1(ret, line_node, putspecialobject,
INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
7438 ADD_INSN1(ret, line_node, putobject, rb_fstring_lit(
"deconstruct must return Array"));
7439 ADD_SEND(ret, line_node, id_core_raise,
INT2FIX(2));
7440 ADD_INSN(ret, line_node, pop);
7442 ADD_LABEL(ret, match_failed);
7443 ADD_INSN(ret, line_node, pop);
7445 ADD_INSN(ret, line_node, pop);
7447 ADD_INSNL(ret, line_node, jump, unmatched);
7500 const NODE *args = RNODE_FNDPTN(node)->args;
7501 const int args_num = RNODE_FNDPTN(node)->args ?
rb_long2int(RNODE_LIST(RNODE_FNDPTN(node)->args)->as.nd_alen) : 0;
7503 LABEL *match_failed, *type_error, *deconstruct, *deconstructed;
7504 match_failed = NEW_LABEL(line);
7505 type_error = NEW_LABEL(line);
7506 deconstruct = NEW_LABEL(line);
7507 deconstructed = NEW_LABEL(line);
7509 CHECK(iseq_compile_pattern_constant(iseq, ret, node, match_failed, in_single_pattern, base_index));
7511 CHECK(iseq_compile_array_deconstruct(iseq, ret, node, deconstruct, deconstructed, match_failed, type_error, in_single_pattern, base_index, use_deconstructed_cache));
7513 ADD_INSN(ret, line_node, dup);
7514 ADD_SEND(ret, line_node, idLength,
INT2FIX(0));
7515 ADD_INSN1(ret, line_node, putobject,
INT2FIX(args_num));
7516 ADD_SEND(ret, line_node, idGE,
INT2FIX(1));
7517 if (in_single_pattern) {
7518 CHECK(iseq_compile_pattern_set_length_errmsg(iseq, ret, node, rb_fstring_lit(
"%p length mismatch (given %p, expected %p+)"),
INT2FIX(args_num), base_index + 1 ));
7520 ADD_INSNL(ret, line_node, branchunless, match_failed);
7523 LABEL *while_begin = NEW_LABEL(nd_line(node));
7524 LABEL *next_loop = NEW_LABEL(nd_line(node));
7525 LABEL *find_succeeded = NEW_LABEL(line);
7526 LABEL *find_failed = NEW_LABEL(nd_line(node));
7529 ADD_INSN(ret, line_node, dup);
7530 ADD_SEND(ret, line_node, idLength,
INT2FIX(0));
7532 ADD_INSN(ret, line_node, dup);
7533 ADD_INSN1(ret, line_node, putobject,
INT2FIX(args_num));
7534 ADD_SEND(ret, line_node, idMINUS,
INT2FIX(1));
7536 ADD_INSN1(ret, line_node, putobject,
INT2FIX(0));
7538 ADD_LABEL(ret, while_begin);
7540 ADD_INSN(ret, line_node, dup);
7541 ADD_INSN1(ret, line_node, topn,
INT2FIX(2));
7542 ADD_SEND(ret, line_node, idLE,
INT2FIX(1));
7543 ADD_INSNL(ret, line_node, branchunless, find_failed);
7545 for (j = 0; j < args_num; j++) {
7546 ADD_INSN1(ret, line_node, topn,
INT2FIX(3));
7547 ADD_INSN1(ret, line_node, topn,
INT2FIX(1));
7549 ADD_INSN1(ret, line_node, putobject,
INT2FIX(j));
7550 ADD_SEND(ret, line_node, idPLUS,
INT2FIX(1));
7552 ADD_SEND(ret, line_node, idAREF,
INT2FIX(1));
7554 CHECK(iseq_compile_pattern_match(iseq, ret, RNODE_LIST(args)->nd_head, next_loop, in_single_pattern, in_alt_pattern, base_index + 4 ,
false));
7555 args = RNODE_LIST(args)->nd_next;
7558 if (NODE_NAMED_REST_P(RNODE_FNDPTN(node)->pre_rest_arg)) {
7559 ADD_INSN1(ret, line_node, topn,
INT2FIX(3));
7560 ADD_INSN1(ret, line_node, putobject,
INT2FIX(0));
7561 ADD_INSN1(ret, line_node, topn,
INT2FIX(2));
7562 ADD_SEND(ret, line_node, idAREF,
INT2FIX(2));
7563 CHECK(iseq_compile_pattern_match(iseq, ret, RNODE_FNDPTN(node)->pre_rest_arg, find_failed, in_single_pattern, in_alt_pattern, base_index + 4 ,
false));
7565 if (NODE_NAMED_REST_P(RNODE_FNDPTN(node)->post_rest_arg)) {
7566 ADD_INSN1(ret, line_node, topn,
INT2FIX(3));
7567 ADD_INSN1(ret, line_node, topn,
INT2FIX(1));
7568 ADD_INSN1(ret, line_node, putobject,
INT2FIX(args_num));
7569 ADD_SEND(ret, line_node, idPLUS,
INT2FIX(1));
7570 ADD_INSN1(ret, line_node, topn,
INT2FIX(3));
7571 ADD_SEND(ret, line_node, idAREF,
INT2FIX(2));
7572 CHECK(iseq_compile_pattern_match(iseq, ret, RNODE_FNDPTN(node)->post_rest_arg, find_failed, in_single_pattern, in_alt_pattern, base_index + 4 ,
false));
7574 ADD_INSNL(ret, line_node, jump, find_succeeded);
7576 ADD_LABEL(ret, next_loop);
7577 ADD_INSN1(ret, line_node, putobject,
INT2FIX(1));
7578 ADD_SEND(ret, line_node, idPLUS,
INT2FIX(1));
7579 ADD_INSNL(ret, line_node, jump, while_begin);
7581 ADD_LABEL(ret, find_failed);
7582 ADD_INSN1(ret, line_node, adjuststack,
INT2FIX(3));
7583 if (in_single_pattern) {
7584 ADD_INSN1(ret, line_node, putspecialobject,
INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
7585 ADD_INSN1(ret, line_node, putobject, rb_fstring_lit(
"%p does not match to find pattern"));
7586 ADD_INSN1(ret, line_node, topn,
INT2FIX(2));
7587 ADD_SEND(ret, line_node, id_core_sprintf,
INT2FIX(2));
7588 ADD_INSN1(ret, line_node, setn,
INT2FIX(base_index + CASE3_BI_OFFSET_ERROR_STRING + 1 ));
7590 ADD_INSN1(ret, line_node, putobject,
Qfalse);
7591 ADD_INSN1(ret, line_node, setn,
INT2FIX(base_index + CASE3_BI_OFFSET_KEY_ERROR_P + 2 ));
7593 ADD_INSN(ret, line_node, pop);
7594 ADD_INSN(ret, line_node, pop);
7596 ADD_INSNL(ret, line_node, jump, match_failed);
7597 ADD_INSN1(ret, line_node, dupn,
INT2FIX(3));
7599 ADD_LABEL(ret, find_succeeded);
7600 ADD_INSN1(ret, line_node, adjuststack,
INT2FIX(3));
7603 ADD_INSN(ret, line_node, pop);
7604 ADD_INSNL(ret, line_node, jump, matched);
7605 ADD_INSN(ret, line_node, putnil);
7607 ADD_LABEL(ret, type_error);
7608 ADD_INSN1(ret, line_node, putspecialobject,
INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
7610 ADD_INSN1(ret, line_node, putobject, rb_fstring_lit(
"deconstruct must return Array"));
7611 ADD_SEND(ret, line_node, id_core_raise,
INT2FIX(2));
7612 ADD_INSN(ret, line_node, pop);
7614 ADD_LABEL(ret, match_failed);
7615 ADD_INSN(ret, line_node, pop);
7616 ADD_INSNL(ret, line_node, jump, unmatched);
7680 LABEL *match_failed, *type_error;
7683 match_failed = NEW_LABEL(line);
7684 type_error = NEW_LABEL(line);
7686 if (RNODE_HSHPTN(node)->nd_pkwargs && !RNODE_HSHPTN(node)->nd_pkwrestarg) {
7687 const NODE *kw_args = RNODE_HASH(RNODE_HSHPTN(node)->nd_pkwargs)->nd_head;
7688 keys =
rb_ary_new_capa(kw_args ? RNODE_LIST(kw_args)->as.nd_alen/2 : 0);
7690 rb_ary_push(keys, get_symbol_value(iseq, RNODE_LIST(kw_args)->nd_head));
7691 kw_args = RNODE_LIST(RNODE_LIST(kw_args)->nd_next)->nd_next;
7695 CHECK(iseq_compile_pattern_constant(iseq, ret, node, match_failed, in_single_pattern, base_index));
7697 ADD_INSN(ret, line_node, dup);
7698 ADD_INSN1(ret, line_node, putobject,
ID2SYM(rb_intern(
"deconstruct_keys")));
7699 ADD_SEND(ret, line_node, idRespond_to,
INT2FIX(1));
7700 if (in_single_pattern) {
7701 CHECK(iseq_compile_pattern_set_general_errmsg(iseq, ret, node, rb_fstring_lit(
"%p does not respond to #deconstruct_keys"), base_index + 1 ));
7703 ADD_INSNL(ret, line_node, branchunless, match_failed);
7706 ADD_INSN(ret, line_node, putnil);
7709 RB_OBJ_SET_FROZEN_SHAREABLE(keys);
7710 ADD_INSN1(ret, line_node, duparray, keys);
7713 ADD_SEND(ret, line_node, rb_intern(
"deconstruct_keys"),
INT2FIX(1));
7715 ADD_INSN(ret, line_node, dup);
7717 ADD_INSNL(ret, line_node, branchunless, type_error);
7719 if (RNODE_HSHPTN(node)->nd_pkwrestarg) {
7720 ADD_SEND(ret, line_node, rb_intern(
"dup"),
INT2FIX(0));
7723 if (RNODE_HSHPTN(node)->nd_pkwargs) {
7727 args = RNODE_HASH(RNODE_HSHPTN(node)->nd_pkwargs)->nd_head;
7729 DECL_ANCHOR(match_values);
7730 INIT_ANCHOR(match_values);
7731 keys_num =
rb_long2int(RNODE_LIST(args)->as.nd_alen) / 2;
7732 for (i = 0; i < keys_num; i++) {
7733 NODE *key_node = RNODE_LIST(args)->nd_head;
7734 NODE *value_node = RNODE_LIST(RNODE_LIST(args)->nd_next)->nd_head;
7735 VALUE key = get_symbol_value(iseq, key_node);
7737 ADD_INSN(ret, line_node, dup);
7738 ADD_INSN1(ret, line_node, putobject, key);
7739 ADD_SEND(ret, line_node, rb_intern(
"key?"),
INT2FIX(1));
7740 if (in_single_pattern) {
7741 LABEL *match_succeeded;
7742 match_succeeded = NEW_LABEL(line);
7744 ADD_INSN(ret, line_node, dup);
7745 ADD_INSNL(ret, line_node, branchif, match_succeeded);
7748 ADD_INSN1(ret, line_node, putobject, RB_OBJ_SET_SHAREABLE(str));
7749 ADD_INSN1(ret, line_node, setn,
INT2FIX(base_index + CASE3_BI_OFFSET_ERROR_STRING + 2 ));
7750 ADD_INSN1(ret, line_node, putobject,
Qtrue);
7751 ADD_INSN1(ret, line_node, setn,
INT2FIX(base_index + CASE3_BI_OFFSET_KEY_ERROR_P + 3 ));
7752 ADD_INSN1(ret, line_node, topn,
INT2FIX(3));
7753 ADD_INSN1(ret, line_node, setn,
INT2FIX(base_index + CASE3_BI_OFFSET_KEY_ERROR_MATCHEE + 4 ));
7754 ADD_INSN1(ret, line_node, putobject, key);
7755 ADD_INSN1(ret, line_node, setn,
INT2FIX(base_index + CASE3_BI_OFFSET_KEY_ERROR_KEY + 5 ));
7757 ADD_INSN1(ret, line_node, adjuststack,
INT2FIX(4));
7759 ADD_LABEL(ret, match_succeeded);
7761 ADD_INSNL(ret, line_node, branchunless, match_failed);
7763 ADD_INSN(match_values, line_node, dup);
7764 ADD_INSN1(match_values, line_node, putobject, key);
7765 ADD_SEND(match_values, line_node, RNODE_HSHPTN(node)->nd_pkwrestarg ? rb_intern(
"delete") : idAREF,
INT2FIX(1));
7766 CHECK(iseq_compile_pattern_match(iseq, match_values, value_node, match_failed, in_single_pattern, in_alt_pattern, base_index + 1 ,
false));
7767 args = RNODE_LIST(RNODE_LIST(args)->nd_next)->nd_next;
7769 ADD_SEQ(ret, match_values);
7773 ADD_INSN(ret, line_node, dup);
7774 ADD_SEND(ret, line_node, idEmptyP,
INT2FIX(0));
7775 if (in_single_pattern) {
7776 CHECK(iseq_compile_pattern_set_general_errmsg(iseq, ret, node, rb_fstring_lit(
"%p is not empty"), base_index + 1 ));
7778 ADD_INSNL(ret, line_node, branchunless, match_failed);
7781 if (RNODE_HSHPTN(node)->nd_pkwrestarg) {
7782 if (RNODE_HSHPTN(node)->nd_pkwrestarg == NODE_SPECIAL_NO_REST_KEYWORD) {
7783 ADD_INSN(ret, line_node, dup);
7784 ADD_SEND(ret, line_node, idEmptyP,
INT2FIX(0));
7785 if (in_single_pattern) {
7786 CHECK(iseq_compile_pattern_set_general_errmsg(iseq, ret, node, rb_fstring_lit(
"rest of %p is not empty"), base_index + 1 ));
7788 ADD_INSNL(ret, line_node, branchunless, match_failed);
7791 ADD_INSN(ret, line_node, dup);
7792 CHECK(iseq_compile_pattern_match(iseq, ret, RNODE_HSHPTN(node)->nd_pkwrestarg, match_failed, in_single_pattern, in_alt_pattern, base_index + 1 ,
false));
7796 ADD_INSN(ret, line_node, pop);
7797 ADD_INSNL(ret, line_node, jump, matched);
7798 ADD_INSN(ret, line_node, putnil);
7800 ADD_LABEL(ret, type_error);
7801 ADD_INSN1(ret, line_node, putspecialobject,
INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
7803 ADD_INSN1(ret, line_node, putobject, rb_fstring_lit(
"deconstruct_keys must return Hash"));
7804 ADD_SEND(ret, line_node, id_core_raise,
INT2FIX(2));
7805 ADD_INSN(ret, line_node, pop);
7807 ADD_LABEL(ret, match_failed);
7808 ADD_INSN(ret, line_node, pop);
7809 ADD_INSNL(ret, line_node, jump, unmatched);
7818 case NODE_IMAGINARY:
7846 CHECK(COMPILE(ret,
"case in literal", node));
7847 if (in_single_pattern) {
7848 ADD_INSN1(ret, line_node, dupn,
INT2FIX(2));
7850 ADD_INSN1(ret, line_node, checkmatch,
INT2FIX(VM_CHECKMATCH_TYPE_CASE));
7851 if (in_single_pattern) {
7852 CHECK(iseq_compile_pattern_set_eqq_errmsg(iseq, ret, node, base_index + 2 ));
7854 ADD_INSNL(ret, line_node, branchif, matched);
7855 ADD_INSNL(ret, line_node, jump, unmatched);
7859 ID id = RNODE_LASGN(node)->nd_vid;
7860 int idx = ISEQ_BODY(body->local_iseq)->local_table_size - get_local_var_idx(iseq,
id);
7862 if (in_alt_pattern) {
7863 const char *name = rb_id2name(
id);
7864 if (name && strlen(name) > 0 && name[0] !=
'_') {
7865 COMPILE_ERROR(ERROR_ARGS
"illegal variable in alternative pattern (%"PRIsVALUE
")",
7871 ADD_SETLOCAL(ret, line_node, idx, get_lvar_level(iseq));
7872 ADD_INSNL(ret, line_node, jump, matched);
7877 ID id = RNODE_DASGN(node)->nd_vid;
7879 idx = get_dyna_var_idx(iseq,
id, &lv, &ls);
7881 if (in_alt_pattern) {
7882 const char *name = rb_id2name(
id);
7883 if (name && strlen(name) > 0 && name[0] !=
'_') {
7884 COMPILE_ERROR(ERROR_ARGS
"illegal variable in alternative pattern (%"PRIsVALUE
")",
7891 COMPILE_ERROR(ERROR_ARGS
"NODE_DASGN: unknown id (%"PRIsVALUE
")",
7895 ADD_SETLOCAL(ret, line_node, ls - idx, lv);
7896 ADD_INSNL(ret, line_node, jump, matched);
7901 LABEL *match_failed;
7902 match_failed = unmatched;
7903 CHECK(iseq_compile_pattern_match(iseq, ret, RNODE_IF(node)->nd_body, unmatched, in_single_pattern, in_alt_pattern, base_index, use_deconstructed_cache));
7904 CHECK(COMPILE(ret,
"case in if", RNODE_IF(node)->nd_cond));
7905 if (in_single_pattern) {
7906 LABEL *match_succeeded;
7907 match_succeeded = NEW_LABEL(line);
7909 ADD_INSN(ret, line_node, dup);
7910 if (nd_type_p(node, NODE_IF)) {
7911 ADD_INSNL(ret, line_node, branchif, match_succeeded);
7914 ADD_INSNL(ret, line_node, branchunless, match_succeeded);
7917 ADD_INSN1(ret, line_node, putobject, rb_fstring_lit(
"guard clause does not return true"));
7918 ADD_INSN1(ret, line_node, setn,
INT2FIX(base_index + CASE3_BI_OFFSET_ERROR_STRING + 1 ));
7919 ADD_INSN1(ret, line_node, putobject,
Qfalse);
7920 ADD_INSN1(ret, line_node, setn,
INT2FIX(base_index + CASE3_BI_OFFSET_KEY_ERROR_P + 2 ));
7922 ADD_INSN(ret, line_node, pop);
7923 ADD_INSN(ret, line_node, pop);
7925 ADD_LABEL(ret, match_succeeded);
7927 if (nd_type_p(node, NODE_IF)) {
7928 ADD_INSNL(ret, line_node, branchunless, match_failed);
7931 ADD_INSNL(ret, line_node, branchif, match_failed);
7933 ADD_INSNL(ret, line_node, jump, matched);
7938 LABEL *match_failed;
7939 match_failed = NEW_LABEL(line);
7941 n = RNODE_HASH(node)->nd_head;
7942 if (! (nd_type_p(n, NODE_LIST) && RNODE_LIST(n)->as.nd_alen == 2)) {
7943 COMPILE_ERROR(ERROR_ARGS
"unexpected node");
7947 ADD_INSN(ret, line_node, dup);
7948 CHECK(iseq_compile_pattern_match(iseq, ret, RNODE_LIST(n)->nd_head, match_failed, in_single_pattern, in_alt_pattern, base_index + 1 , use_deconstructed_cache));
7949 CHECK(iseq_compile_pattern_each(iseq, ret, RNODE_LIST(RNODE_LIST(n)->nd_next)->nd_head, matched, match_failed, in_single_pattern, in_alt_pattern, base_index,
false));
7950 ADD_INSN(ret, line_node, putnil);
7952 ADD_LABEL(ret, match_failed);
7953 ADD_INSN(ret, line_node, pop);
7954 ADD_INSNL(ret, line_node, jump, unmatched);
7958 LABEL *match_succeeded, *fin;
7959 match_succeeded = NEW_LABEL(line);
7960 fin = NEW_LABEL(line);
7962 ADD_INSN(ret, line_node, dup);
7963 CHECK(iseq_compile_pattern_each(iseq, ret, RNODE_OR(node)->nd_1st, match_succeeded, fin, in_single_pattern,
true, base_index + 1 , use_deconstructed_cache));
7964 ADD_LABEL(ret, match_succeeded);
7965 ADD_INSN(ret, line_node, pop);
7966 ADD_INSNL(ret, line_node, jump, matched);
7967 ADD_INSN(ret, line_node, putnil);
7968 ADD_LABEL(ret, fin);
7969 CHECK(iseq_compile_pattern_each(iseq, ret, RNODE_OR(node)->nd_2nd, matched, unmatched, in_single_pattern,
true, base_index, use_deconstructed_cache));
7973 UNKNOWN_NODE(
"NODE_IN", node, COMPILE_NG);
7979iseq_compile_pattern_match(
rb_iseq_t *iseq,
LINK_ANCHOR *
const ret,
const NODE *
const node,
LABEL *unmatched,
bool in_single_pattern,
bool in_alt_pattern,
int base_index,
bool use_deconstructed_cache)
7981 LABEL *fin = NEW_LABEL(nd_line(node));
7982 CHECK(iseq_compile_pattern_each(iseq, ret, node, fin, unmatched, in_single_pattern, in_alt_pattern, base_index, use_deconstructed_cache));
7983 ADD_LABEL(ret, fin);
7988iseq_compile_pattern_constant(
rb_iseq_t *iseq,
LINK_ANCHOR *
const ret,
const NODE *
const node,
LABEL *match_failed,
bool in_single_pattern,
int base_index)
7990 const NODE *line_node = node;
7992 if (RNODE_ARYPTN(node)->nd_pconst) {
7993 ADD_INSN(ret, line_node, dup);
7994 CHECK(COMPILE(ret,
"constant", RNODE_ARYPTN(node)->nd_pconst));
7995 if (in_single_pattern) {
7996 ADD_INSN1(ret, line_node, dupn,
INT2FIX(2));
7998 ADD_INSN1(ret, line_node, checkmatch,
INT2FIX(VM_CHECKMATCH_TYPE_CASE));
7999 if (in_single_pattern) {
8000 CHECK(iseq_compile_pattern_set_eqq_errmsg(iseq, ret, node, base_index + 3 ));
8002 ADD_INSNL(ret, line_node, branchunless, match_failed);
8009iseq_compile_array_deconstruct(
rb_iseq_t *iseq,
LINK_ANCHOR *
const ret,
const NODE *
const node,
LABEL *deconstruct,
LABEL *deconstructed,
LABEL *match_failed,
LABEL *type_error,
bool in_single_pattern,
int base_index,
bool use_deconstructed_cache)
8011 const NODE *line_node = node;
8015 if (use_deconstructed_cache) {
8017 ADD_INSN1(ret, line_node, topn,
INT2FIX(base_index + CASE3_BI_OFFSET_DECONSTRUCTED_CACHE));
8018 ADD_INSNL(ret, line_node, branchnil, deconstruct);
8021 ADD_INSN1(ret, line_node, topn,
INT2FIX(base_index + CASE3_BI_OFFSET_DECONSTRUCTED_CACHE));
8022 ADD_INSNL(ret, line_node, branchunless, match_failed);
8025 ADD_INSN(ret, line_node, pop);
8026 ADD_INSN1(ret, line_node, topn,
INT2FIX(base_index + CASE3_BI_OFFSET_DECONSTRUCTED_CACHE - 1 ));
8027 ADD_INSNL(ret, line_node, jump, deconstructed);
8030 ADD_INSNL(ret, line_node, jump, deconstruct);
8033 ADD_LABEL(ret, deconstruct);
8034 ADD_INSN(ret, line_node, dup);
8035 ADD_INSN1(ret, line_node, putobject,
ID2SYM(rb_intern(
"deconstruct")));
8036 ADD_SEND(ret, line_node, idRespond_to,
INT2FIX(1));
8039 if (use_deconstructed_cache) {
8040 ADD_INSN1(ret, line_node, setn,
INT2FIX(base_index + CASE3_BI_OFFSET_DECONSTRUCTED_CACHE + 1 ));
8043 if (in_single_pattern) {
8044 CHECK(iseq_compile_pattern_set_general_errmsg(iseq, ret, node, rb_fstring_lit(
"%p does not respond to #deconstruct"), base_index + 1 ));
8047 ADD_INSNL(ret, line_node, branchunless, match_failed);
8049 ADD_SEND(ret, line_node, rb_intern(
"deconstruct"),
INT2FIX(0));
8052 if (use_deconstructed_cache) {
8053 ADD_INSN1(ret, line_node, setn,
INT2FIX(base_index + CASE3_BI_OFFSET_DECONSTRUCTED_CACHE));
8056 ADD_INSN(ret, line_node, dup);
8058 ADD_INSNL(ret, line_node, branchunless, type_error);
8060 ADD_LABEL(ret, deconstructed);
8076 const int line = nd_line(node);
8077 const NODE *line_node = node;
8078 LABEL *match_succeeded = NEW_LABEL(line);
8080 ADD_INSN(ret, line_node, dup);
8081 ADD_INSNL(ret, line_node, branchif, match_succeeded);
8083 ADD_INSN1(ret, line_node, putspecialobject,
INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
8084 ADD_INSN1(ret, line_node, putobject, errmsg);
8085 ADD_INSN1(ret, line_node, topn,
INT2FIX(3));
8086 ADD_SEND(ret, line_node, id_core_sprintf,
INT2FIX(2));
8087 ADD_INSN1(ret, line_node, setn,
INT2FIX(base_index + CASE3_BI_OFFSET_ERROR_STRING + 1 ));
8089 ADD_INSN1(ret, line_node, putobject,
Qfalse);
8090 ADD_INSN1(ret, line_node, setn,
INT2FIX(base_index + CASE3_BI_OFFSET_KEY_ERROR_P + 2 ));
8092 ADD_INSN(ret, line_node, pop);
8093 ADD_INSN(ret, line_node, pop);
8094 ADD_LABEL(ret, match_succeeded);
8110 const int line = nd_line(node);
8111 const NODE *line_node = node;
8112 LABEL *match_succeeded = NEW_LABEL(line);
8114 ADD_INSN(ret, line_node, dup);
8115 ADD_INSNL(ret, line_node, branchif, match_succeeded);
8117 ADD_INSN1(ret, line_node, putspecialobject,
INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
8118 ADD_INSN1(ret, line_node, putobject, errmsg);
8119 ADD_INSN1(ret, line_node, topn,
INT2FIX(3));
8120 ADD_INSN(ret, line_node, dup);
8121 ADD_SEND(ret, line_node, idLength,
INT2FIX(0));
8122 ADD_INSN1(ret, line_node, putobject, pattern_length);
8123 ADD_SEND(ret, line_node, id_core_sprintf,
INT2FIX(4));
8124 ADD_INSN1(ret, line_node, setn,
INT2FIX(base_index + CASE3_BI_OFFSET_ERROR_STRING + 1 ));
8126 ADD_INSN1(ret, line_node, putobject,
Qfalse);
8127 ADD_INSN1(ret, line_node, setn,
INT2FIX(base_index + CASE3_BI_OFFSET_KEY_ERROR_P + 2));
8129 ADD_INSN(ret, line_node, pop);
8130 ADD_INSN(ret, line_node, pop);
8131 ADD_LABEL(ret, match_succeeded);
8137iseq_compile_pattern_set_eqq_errmsg(
rb_iseq_t *iseq,
LINK_ANCHOR *
const ret,
const NODE *
const node,
int base_index)
8147 const int line = nd_line(node);
8148 const NODE *line_node = node;
8149 LABEL *match_succeeded = NEW_LABEL(line);
8151 ADD_INSN(ret, line_node, dup);
8152 ADD_INSNL(ret, line_node, branchif, match_succeeded);
8154 ADD_INSN1(ret, line_node, putspecialobject,
INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
8155 ADD_INSN1(ret, line_node, putobject, rb_fstring_lit(
"%p === %p does not return true"));
8156 ADD_INSN1(ret, line_node, topn,
INT2FIX(3));
8157 ADD_INSN1(ret, line_node, topn,
INT2FIX(5));
8158 ADD_SEND(ret, line_node, id_core_sprintf,
INT2FIX(3));
8159 ADD_INSN1(ret, line_node, setn,
INT2FIX(base_index + CASE3_BI_OFFSET_ERROR_STRING + 1 ));
8161 ADD_INSN1(ret, line_node, putobject,
Qfalse);
8162 ADD_INSN1(ret, line_node, setn,
INT2FIX(base_index + CASE3_BI_OFFSET_KEY_ERROR_P + 2 ));
8164 ADD_INSN(ret, line_node, pop);
8165 ADD_INSN(ret, line_node, pop);
8167 ADD_LABEL(ret, match_succeeded);
8168 ADD_INSN1(ret, line_node, setn,
INT2FIX(2));
8169 ADD_INSN(ret, line_node, pop);
8170 ADD_INSN(ret, line_node, pop);
8178 const NODE *pattern;
8179 const NODE *node = orig_node;
8180 LABEL *endlabel, *elselabel;
8182 DECL_ANCHOR(body_seq);
8183 DECL_ANCHOR(cond_seq);
8185 enum node_type
type;
8186 const NODE *line_node;
8189 bool single_pattern;
8192 INIT_ANCHOR(body_seq);
8193 INIT_ANCHOR(cond_seq);
8195 branches = decl_branch_base(iseq, PTR2NUM(node), nd_code_loc(node),
"case");
8197 node = RNODE_CASE3(node)->nd_body;
8198 EXPECT_NODE(
"NODE_CASE3", node, NODE_IN, COMPILE_NG);
8199 type = nd_type(node);
8200 line = nd_line(node);
8202 single_pattern = !RNODE_IN(node)->nd_next;
8204 endlabel = NEW_LABEL(line);
8205 elselabel = NEW_LABEL(line);
8207 if (single_pattern) {
8209 ADD_INSN(head, line_node, putnil);
8210 ADD_INSN(head, line_node, putnil);
8211 ADD_INSN1(head, line_node, putobject,
Qfalse);
8212 ADD_INSN(head, line_node, putnil);
8214 ADD_INSN(head, line_node, putnil);
8216 CHECK(COMPILE(head,
"case base", RNODE_CASE3(orig_node)->nd_head));
8220 while (
type == NODE_IN) {
8224 ADD_INSN(body_seq, line_node, putnil);
8226 l1 = NEW_LABEL(line);
8227 ADD_LABEL(body_seq, l1);
8228 ADD_INSN1(body_seq, line_node, adjuststack,
INT2FIX(single_pattern ? 6 : 2));
8230 const NODE *
const coverage_node = RNODE_IN(node)->nd_body ? RNODE_IN(node)->nd_body : node;
8231 add_trace_branch_coverage(
8234 nd_code_loc(coverage_node),
8235 nd_node_id(coverage_node),
8240 CHECK(COMPILE_(body_seq,
"in body", RNODE_IN(node)->nd_body, popped));
8241 ADD_INSNL(body_seq, line_node, jump, endlabel);
8243 pattern = RNODE_IN(node)->nd_head;
8245 int pat_line = nd_line(pattern);
8246 LABEL *next_pat = NEW_LABEL(pat_line);
8247 ADD_INSN (cond_seq, pattern, dup);
8249 CHECK(iseq_compile_pattern_each(iseq, cond_seq, pattern, l1, next_pat, single_pattern,
false, 2,
true));
8250 ADD_LABEL(cond_seq, next_pat);
8251 LABEL_UNREMOVABLE(next_pat);
8254 COMPILE_ERROR(ERROR_ARGS
"unexpected node");
8258 node = RNODE_IN(node)->nd_next;
8262 type = nd_type(node);
8263 line = nd_line(node);
8268 ADD_LABEL(cond_seq, elselabel);
8269 ADD_INSN(cond_seq, line_node, pop);
8270 ADD_INSN(cond_seq, line_node, pop);
8271 add_trace_branch_coverage(iseq, cond_seq, nd_code_loc(node), nd_node_id(node), branch_id,
"else", branches);
8272 CHECK(COMPILE_(cond_seq,
"else", node, popped));
8273 ADD_INSNL(cond_seq, line_node, jump, endlabel);
8274 ADD_INSN(cond_seq, line_node, putnil);
8276 ADD_INSN(cond_seq, line_node, putnil);
8280 debugs(
"== else (implicit)\n");
8281 ADD_LABEL(cond_seq, elselabel);
8282 add_trace_branch_coverage(iseq, cond_seq, nd_code_loc(orig_node), nd_node_id(orig_node), branch_id,
"else", branches);
8283 ADD_INSN1(cond_seq, orig_node, putspecialobject,
INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
8285 if (single_pattern) {
8293 LABEL *key_error, *fin;
8296 key_error = NEW_LABEL(line);
8297 fin = NEW_LABEL(line);
8300 kw_arg->references = 0;
8301 kw_arg->keyword_len = 2;
8302 kw_arg->keywords[0] =
ID2SYM(rb_intern(
"matchee"));
8303 kw_arg->keywords[1] =
ID2SYM(rb_intern(
"key"));
8305 ADD_INSN1(cond_seq, orig_node, topn,
INT2FIX(CASE3_BI_OFFSET_KEY_ERROR_P + 2));
8306 ADD_INSNL(cond_seq, orig_node, branchif, key_error);
8308 ADD_INSN1(cond_seq, orig_node, putspecialobject,
INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
8309 ADD_INSN1(cond_seq, orig_node, putobject, rb_fstring_lit(
"%p: %s"));
8310 ADD_INSN1(cond_seq, orig_node, topn,
INT2FIX(4));
8311 ADD_INSN1(cond_seq, orig_node, topn,
INT2FIX(CASE3_BI_OFFSET_ERROR_STRING + 6));
8312 ADD_SEND(cond_seq, orig_node, id_core_sprintf,
INT2FIX(3));
8313 ADD_SEND(cond_seq, orig_node, id_core_raise,
INT2FIX(2));
8314 ADD_INSNL(cond_seq, orig_node, jump, fin);
8316 ADD_LABEL(cond_seq, key_error);
8318 ADD_INSN1(cond_seq, orig_node, putspecialobject,
INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
8319 ADD_INSN1(cond_seq, orig_node, putobject, rb_fstring_lit(
"%p: %s"));
8320 ADD_INSN1(cond_seq, orig_node, topn,
INT2FIX(4));
8321 ADD_INSN1(cond_seq, orig_node, topn,
INT2FIX(CASE3_BI_OFFSET_ERROR_STRING + 6));
8322 ADD_SEND(cond_seq, orig_node, id_core_sprintf,
INT2FIX(3));
8323 ADD_INSN1(cond_seq, orig_node, topn,
INT2FIX(CASE3_BI_OFFSET_KEY_ERROR_MATCHEE + 4));
8324 ADD_INSN1(cond_seq, orig_node, topn,
INT2FIX(CASE3_BI_OFFSET_KEY_ERROR_KEY + 5));
8325 ADD_SEND_R(cond_seq, orig_node, rb_intern(
"new"),
INT2FIX(1), NULL,
INT2FIX(VM_CALL_KWARG), kw_arg);
8326 ADD_SEND(cond_seq, orig_node, id_core_raise,
INT2FIX(1));
8328 ADD_LABEL(cond_seq, fin);
8332 ADD_INSN1(cond_seq, orig_node, topn,
INT2FIX(2));
8333 ADD_SEND(cond_seq, orig_node, id_core_raise,
INT2FIX(2));
8335 ADD_INSN1(cond_seq, orig_node, adjuststack,
INT2FIX(single_pattern ? 7 : 3));
8337 ADD_INSN(cond_seq, orig_node, putnil);
8339 ADD_INSNL(cond_seq, orig_node, jump, endlabel);
8340 ADD_INSN1(cond_seq, orig_node, dupn,
INT2FIX(single_pattern ? 5 : 1));
8342 ADD_INSN(cond_seq, line_node, putnil);
8346 ADD_SEQ(ret, cond_seq);
8347 ADD_SEQ(ret, body_seq);
8348 ADD_LABEL(ret, endlabel);
8352#undef CASE3_BI_OFFSET_DECONSTRUCTED_CACHE
8353#undef CASE3_BI_OFFSET_ERROR_STRING
8354#undef CASE3_BI_OFFSET_KEY_ERROR_P
8355#undef CASE3_BI_OFFSET_KEY_ERROR_MATCHEE
8356#undef CASE3_BI_OFFSET_KEY_ERROR_KEY
8361 const int line = (int)nd_line(node);
8362 const NODE *line_node = node;
8364 LABEL *prev_start_label = ISEQ_COMPILE_DATA(iseq)->start_label;
8365 LABEL *prev_end_label = ISEQ_COMPILE_DATA(iseq)->end_label;
8366 LABEL *prev_redo_label = ISEQ_COMPILE_DATA(iseq)->redo_label;
8367 int prev_loopval_popped = ISEQ_COMPILE_DATA(iseq)->loopval_popped;
8372 LABEL *next_label = ISEQ_COMPILE_DATA(iseq)->start_label = NEW_LABEL(line);
8373 LABEL *redo_label = ISEQ_COMPILE_DATA(iseq)->redo_label = NEW_LABEL(line);
8374 LABEL *break_label = ISEQ_COMPILE_DATA(iseq)->end_label = NEW_LABEL(line);
8375 LABEL *end_label = NEW_LABEL(line);
8376 LABEL *adjust_label = NEW_LABEL(line);
8378 LABEL *next_catch_label = NEW_LABEL(line);
8379 LABEL *tmp_label = NULL;
8381 ISEQ_COMPILE_DATA(iseq)->loopval_popped = 0;
8382 push_ensure_entry(iseq, &enl, NULL, NULL);
8384 if (RNODE_WHILE(node)->nd_state == 1) {
8385 ADD_INSNL(ret, line_node, jump, next_label);
8388 tmp_label = NEW_LABEL(line);
8389 ADD_INSNL(ret, line_node, jump, tmp_label);
8391 ADD_LABEL(ret, adjust_label);
8392 ADD_INSN(ret, line_node, putnil);
8393 ADD_LABEL(ret, next_catch_label);
8394 ADD_INSN(ret, line_node, pop);
8395 ADD_INSNL(ret, line_node, jump, next_label);
8396 if (tmp_label) ADD_LABEL(ret, tmp_label);
8398 ADD_LABEL(ret, redo_label);
8399 branches = decl_branch_base(iseq, PTR2NUM(node), nd_code_loc(node),
type == NODE_WHILE ?
"while" :
"until");
8401 const NODE *
const coverage_node = RNODE_WHILE(node)->nd_body ? RNODE_WHILE(node)->nd_body : node;
8402 add_trace_branch_coverage(
8405 nd_code_loc(coverage_node),
8406 nd_node_id(coverage_node),
8411 CHECK(COMPILE_POPPED(ret,
"while body", RNODE_WHILE(node)->nd_body));
8412 ADD_LABEL(ret, next_label);
8414 if (
type == NODE_WHILE) {
8415 CHECK(compile_branch_condition(iseq, ret, RNODE_WHILE(node)->nd_cond,
8416 redo_label, end_label));
8420 CHECK(compile_branch_condition(iseq, ret, RNODE_WHILE(node)->nd_cond,
8421 end_label, redo_label));
8424 ADD_LABEL(ret, end_label);
8425 ADD_ADJUST_RESTORE(ret, adjust_label);
8427 if (UNDEF_P(RNODE_WHILE(node)->nd_state)) {
8429 COMPILE_ERROR(ERROR_ARGS
"unsupported: putundef");
8433 ADD_INSN(ret, line_node, putnil);
8436 ADD_LABEL(ret, break_label);
8439 ADD_INSN(ret, line_node, pop);
8442 ADD_CATCH_ENTRY(CATCH_TYPE_BREAK, redo_label, break_label, NULL,
8444 ADD_CATCH_ENTRY(CATCH_TYPE_NEXT, redo_label, break_label, NULL,
8446 ADD_CATCH_ENTRY(CATCH_TYPE_REDO, redo_label, break_label, NULL,
8447 ISEQ_COMPILE_DATA(iseq)->redo_label);
8449 ISEQ_COMPILE_DATA(iseq)->start_label = prev_start_label;
8450 ISEQ_COMPILE_DATA(iseq)->end_label = prev_end_label;
8451 ISEQ_COMPILE_DATA(iseq)->redo_label = prev_redo_label;
8452 ISEQ_COMPILE_DATA(iseq)->loopval_popped = prev_loopval_popped;
8453 ISEQ_COMPILE_DATA(iseq)->ensure_node_stack = ISEQ_COMPILE_DATA(iseq)->ensure_node_stack->prev;
8460 const int line = nd_line(node);
8461 const NODE *line_node = node;
8462 const rb_iseq_t *prevblock = ISEQ_COMPILE_DATA(iseq)->current_block;
8463 LABEL *retry_label = NEW_LABEL(line);
8464 LABEL *retry_end_l = NEW_LABEL(line);
8467 ADD_LABEL(ret, retry_label);
8468 if (nd_type_p(node, NODE_FOR)) {
8469 CHECK(COMPILE(ret,
"iter caller (for)", RNODE_FOR(node)->nd_iter));
8471 ISEQ_COMPILE_DATA(iseq)->current_block = child_iseq =
8472 NEW_CHILD_ISEQ(RNODE_FOR(node)->nd_body, make_name_for_block(iseq),
8473 ISEQ_TYPE_BLOCK, line);
8474 ADD_SEND_WITH_BLOCK(ret, line_node, idEach,
INT2FIX(0), child_iseq);
8477 ISEQ_COMPILE_DATA(iseq)->current_block = child_iseq =
8478 NEW_CHILD_ISEQ(RNODE_ITER(node)->nd_body, make_name_for_block(iseq),
8479 ISEQ_TYPE_BLOCK, line);
8480 CHECK(COMPILE(ret,
"iter caller", RNODE_ITER(node)->nd_iter));
8494 iobj = IS_INSN(last_elem) ? (
INSN*) last_elem : (
INSN*) get_prev_insn((
INSN*) last_elem);
8495 while (!IS_INSN_ID(iobj, send) && !IS_INSN_ID(iobj, invokesuper) && !IS_INSN_ID(iobj, sendforward) && !IS_INSN_ID(iobj, invokesuperforward)) {
8496 iobj = (
INSN*) get_prev_insn(iobj);
8498 ELEM_INSERT_NEXT(&iobj->link, (
LINK_ELEMENT*) retry_end_l);
8502 if (&iobj->link == LAST_ELEMENT(ret)) {
8508 ADD_INSN(ret, line_node, pop);
8511 ISEQ_COMPILE_DATA(iseq)->current_block = prevblock;
8513 ADD_CATCH_ENTRY(CATCH_TYPE_BREAK, retry_label, retry_end_l, child_iseq, retry_end_l);
8523 const NODE *line_node = node;
8524 const NODE *var = RNODE_FOR_MASGN(node)->nd_var;
8525 LABEL *not_single = NEW_LABEL(nd_line(var));
8526 LABEL *not_ary = NEW_LABEL(nd_line(var));
8527 CHECK(COMPILE(ret,
"for var", var));
8528 ADD_INSN(ret, line_node, dup);
8529 ADD_CALL(ret, line_node, idLength,
INT2FIX(0));
8530 ADD_INSN1(ret, line_node, putobject,
INT2FIX(1));
8531 ADD_CALL(ret, line_node, idEq,
INT2FIX(1));
8532 ADD_INSNL(ret, line_node, branchunless, not_single);
8533 ADD_INSN(ret, line_node, dup);
8534 ADD_INSN1(ret, line_node, putobject,
INT2FIX(0));
8535 ADD_CALL(ret, line_node, idAREF,
INT2FIX(1));
8536 ADD_INSN1(ret, line_node, putobject,
rb_cArray);
8537 ADD_INSN(ret, line_node, swap);
8538 ADD_CALL(ret, line_node, rb_intern(
"try_convert"),
INT2FIX(1));
8539 ADD_INSN(ret, line_node, dup);
8540 ADD_INSNL(ret, line_node, branchunless, not_ary);
8541 ADD_INSN(ret, line_node, swap);
8542 ADD_LABEL(ret, not_ary);
8543 ADD_INSN(ret, line_node, pop);
8544 ADD_LABEL(ret, not_single);
8551 const NODE *line_node = node;
8552 unsigned long throw_flag = 0;
8554 if (ISEQ_COMPILE_DATA(iseq)->redo_label != 0 && can_add_ensure_iseq(iseq)) {
8556 LABEL *splabel = NEW_LABEL(0);
8557 ADD_LABEL(ret, splabel);
8558 ADD_ADJUST(ret, line_node, ISEQ_COMPILE_DATA(iseq)->redo_label);
8559 CHECK(COMPILE_(ret,
"break val (while/until)", RNODE_BREAK(node)->nd_stts,
8560 ISEQ_COMPILE_DATA(iseq)->loopval_popped));
8561 add_ensure_iseq(ret, iseq, 0);
8562 ADD_INSNL(ret, line_node, jump, ISEQ_COMPILE_DATA(iseq)->end_label);
8563 ADD_ADJUST_RESTORE(ret, splabel);
8566 ADD_INSN(ret, line_node, putnil);
8573 if (!ISEQ_COMPILE_DATA(ip)) {
8578 if (ISEQ_COMPILE_DATA(ip)->redo_label != 0) {
8579 throw_flag = VM_THROW_NO_ESCAPE_FLAG;
8581 else if (ISEQ_BODY(ip)->
type == ISEQ_TYPE_BLOCK) {
8584 else if (ISEQ_BODY(ip)->
type == ISEQ_TYPE_EVAL) {
8585 COMPILE_ERROR(ERROR_ARGS
"Can't escape from eval with break");
8589 ip = ISEQ_BODY(ip)->parent_iseq;
8594 CHECK(COMPILE(ret,
"break val (block)", RNODE_BREAK(node)->nd_stts));
8595 ADD_INSN1(ret, line_node,
throw,
INT2FIX(throw_flag | TAG_BREAK));
8597 ADD_INSN(ret, line_node, pop);
8601 COMPILE_ERROR(ERROR_ARGS
"Invalid break");
8610 const NODE *line_node = node;
8611 unsigned long throw_flag = 0;
8613 if (ISEQ_COMPILE_DATA(iseq)->redo_label != 0 && can_add_ensure_iseq(iseq)) {
8614 LABEL *splabel = NEW_LABEL(0);
8615 debugs(
"next in while loop\n");
8616 ADD_LABEL(ret, splabel);
8617 CHECK(COMPILE(ret,
"next val/valid syntax?", RNODE_NEXT(node)->nd_stts));
8618 add_ensure_iseq(ret, iseq, 0);
8619 ADD_ADJUST(ret, line_node, ISEQ_COMPILE_DATA(iseq)->redo_label);
8620 ADD_INSNL(ret, line_node, jump, ISEQ_COMPILE_DATA(iseq)->start_label);
8621 ADD_ADJUST_RESTORE(ret, splabel);
8623 ADD_INSN(ret, line_node, putnil);
8626 else if (ISEQ_COMPILE_DATA(iseq)->end_label && can_add_ensure_iseq(iseq)) {
8627 LABEL *splabel = NEW_LABEL(0);
8628 debugs(
"next in block\n");
8629 ADD_LABEL(ret, splabel);
8630 ADD_ADJUST(ret, line_node, ISEQ_COMPILE_DATA(iseq)->start_label);
8631 CHECK(COMPILE(ret,
"next val", RNODE_NEXT(node)->nd_stts));
8632 add_ensure_iseq(ret, iseq, 0);
8633 ADD_INSNL(ret, line_node, jump, ISEQ_COMPILE_DATA(iseq)->end_label);
8634 ADD_ADJUST_RESTORE(ret, splabel);
8637 ADD_INSN(ret, line_node, putnil);
8644 if (!ISEQ_COMPILE_DATA(ip)) {
8649 throw_flag = VM_THROW_NO_ESCAPE_FLAG;
8650 if (ISEQ_COMPILE_DATA(ip)->redo_label != 0) {
8654 else if (ISEQ_BODY(ip)->
type == ISEQ_TYPE_BLOCK) {
8657 else if (ISEQ_BODY(ip)->
type == ISEQ_TYPE_EVAL) {
8658 COMPILE_ERROR(ERROR_ARGS
"Can't escape from eval with next");
8662 ip = ISEQ_BODY(ip)->parent_iseq;
8665 CHECK(COMPILE(ret,
"next val", RNODE_NEXT(node)->nd_stts));
8666 ADD_INSN1(ret, line_node,
throw,
INT2FIX(throw_flag | TAG_NEXT));
8669 ADD_INSN(ret, line_node, pop);
8673 COMPILE_ERROR(ERROR_ARGS
"Invalid next");
8683 const NODE *line_node = node;
8685 if (ISEQ_COMPILE_DATA(iseq)->redo_label && can_add_ensure_iseq(iseq)) {
8686 LABEL *splabel = NEW_LABEL(0);
8687 debugs(
"redo in while");
8688 ADD_LABEL(ret, splabel);
8689 ADD_ADJUST(ret, line_node, ISEQ_COMPILE_DATA(iseq)->redo_label);
8690 add_ensure_iseq(ret, iseq, 0);
8691 ADD_INSNL(ret, line_node, jump, ISEQ_COMPILE_DATA(iseq)->redo_label);
8692 ADD_ADJUST_RESTORE(ret, splabel);
8694 ADD_INSN(ret, line_node, putnil);
8697 else if (ISEQ_BODY(iseq)->
type != ISEQ_TYPE_EVAL && ISEQ_COMPILE_DATA(iseq)->start_label && can_add_ensure_iseq(iseq)) {
8698 LABEL *splabel = NEW_LABEL(0);
8700 debugs(
"redo in block");
8701 ADD_LABEL(ret, splabel);
8702 add_ensure_iseq(ret, iseq, 0);
8703 ADD_ADJUST(ret, line_node, ISEQ_COMPILE_DATA(iseq)->start_label);
8704 ADD_INSNL(ret, line_node, jump, ISEQ_COMPILE_DATA(iseq)->start_label);
8705 ADD_ADJUST_RESTORE(ret, splabel);
8708 ADD_INSN(ret, line_node, putnil);
8715 if (!ISEQ_COMPILE_DATA(ip)) {
8720 if (ISEQ_COMPILE_DATA(ip)->redo_label != 0) {
8723 else if (ISEQ_BODY(ip)->
type == ISEQ_TYPE_BLOCK) {
8726 else if (ISEQ_BODY(ip)->
type == ISEQ_TYPE_EVAL) {
8727 COMPILE_ERROR(ERROR_ARGS
"Can't escape from eval with redo");
8731 ip = ISEQ_BODY(ip)->parent_iseq;
8734 ADD_INSN(ret, line_node, putnil);
8735 ADD_INSN1(ret, line_node,
throw,
INT2FIX(VM_THROW_NO_ESCAPE_FLAG | TAG_REDO));
8738 ADD_INSN(ret, line_node, pop);
8742 COMPILE_ERROR(ERROR_ARGS
"Invalid redo");
8752 const NODE *line_node = node;
8754 if (ISEQ_BODY(iseq)->
type == ISEQ_TYPE_RESCUE) {
8755 ADD_INSN(ret, line_node, putnil);
8756 ADD_INSN1(ret, line_node,
throw,
INT2FIX(TAG_RETRY));
8759 ADD_INSN(ret, line_node, pop);
8763 COMPILE_ERROR(ERROR_ARGS
"Invalid retry");
8772 const int line = nd_line(node);
8773 const NODE *line_node = node;
8774 LABEL *lstart = NEW_LABEL(line);
8775 LABEL *lend = NEW_LABEL(line);
8776 LABEL *lcont = NEW_LABEL(line);
8777 const rb_iseq_t *rescue = NEW_CHILD_ISEQ(RNODE_RESCUE(node)->nd_resq,
8779 ISEQ_BODY(iseq)->location.label),
8780 ISEQ_TYPE_RESCUE, line);
8782 lstart->rescued = LABEL_RESCUE_BEG;
8783 lend->rescued = LABEL_RESCUE_END;
8784 ADD_LABEL(ret, lstart);
8786 bool prev_in_rescue = ISEQ_COMPILE_DATA(iseq)->in_rescue;
8787 ISEQ_COMPILE_DATA(iseq)->in_rescue =
true;
8789 CHECK(COMPILE(ret,
"rescue head", RNODE_RESCUE(node)->nd_head));
8791 ISEQ_COMPILE_DATA(iseq)->in_rescue = prev_in_rescue;
8793 ADD_LABEL(ret, lend);
8794 if (RNODE_RESCUE(node)->nd_else) {
8795 ADD_INSN(ret, line_node, pop);
8796 CHECK(COMPILE(ret,
"rescue else", RNODE_RESCUE(node)->nd_else));
8798 ADD_INSN(ret, line_node, nop);
8799 ADD_LABEL(ret, lcont);
8802 ADD_INSN(ret, line_node, pop);
8806 ADD_CATCH_ENTRY(CATCH_TYPE_RESCUE, lstart, lend, rescue, lcont);
8807 ADD_CATCH_ENTRY(CATCH_TYPE_RETRY, lend, lcont, NULL, lstart);
8814 const int line = nd_line(node);
8815 const NODE *line_node = node;
8816 const NODE *resq = node;
8818 LABEL *label_miss, *label_hit;
8821 label_miss = NEW_LABEL(line);
8822 label_hit = NEW_LABEL(line);
8824 narg = RNODE_RESBODY(resq)->nd_args;
8826 switch (nd_type(narg)) {
8829 ADD_GETLOCAL(ret, line_node, LVAR_ERRINFO, 0);
8830 CHECK(COMPILE(ret,
"rescue arg", RNODE_LIST(narg)->nd_head));
8831 ADD_INSN1(ret, line_node, checkmatch,
INT2FIX(VM_CHECKMATCH_TYPE_RESCUE));
8832 ADD_INSNL(ret, line_node, branchif, label_hit);
8833 narg = RNODE_LIST(narg)->nd_next;
8839 ADD_GETLOCAL(ret, line_node, LVAR_ERRINFO, 0);
8840 CHECK(COMPILE(ret,
"rescue/cond splat", narg));
8841 ADD_INSN1(ret, line_node, checkmatch,
INT2FIX(VM_CHECKMATCH_TYPE_RESCUE | VM_CHECKMATCH_ARRAY));
8842 ADD_INSNL(ret, line_node, branchif, label_hit);
8845 UNKNOWN_NODE(
"NODE_RESBODY", narg, COMPILE_NG);
8849 ADD_GETLOCAL(ret, line_node, LVAR_ERRINFO, 0);
8851 ADD_INSN1(ret, line_node, checkmatch,
INT2FIX(VM_CHECKMATCH_TYPE_RESCUE));
8852 ADD_INSNL(ret, line_node, branchif, label_hit);
8854 ADD_INSNL(ret, line_node, jump, label_miss);
8855 ADD_LABEL(ret, label_hit);
8858 if (RNODE_RESBODY(resq)->nd_exc_var) {
8859 CHECK(COMPILE_POPPED(ret,
"resbody exc_var", RNODE_RESBODY(resq)->nd_exc_var));
8862 if (nd_type(RNODE_RESBODY(resq)->nd_body) == NODE_BEGIN && RNODE_BEGIN(RNODE_RESBODY(resq)->nd_body)->nd_body == NULL && !RNODE_RESBODY(resq)->nd_exc_var) {
8864 ADD_SYNTHETIC_INSN(ret, nd_line(RNODE_RESBODY(resq)->nd_body), -1, putnil);
8867 CHECK(COMPILE(ret,
"resbody body", RNODE_RESBODY(resq)->nd_body));
8870 if (ISEQ_COMPILE_DATA(iseq)->option->tailcall_optimization) {
8871 ADD_INSN(ret, line_node, nop);
8873 ADD_INSN(ret, line_node, leave);
8874 ADD_LABEL(ret, label_miss);
8875 resq = RNODE_RESBODY(resq)->nd_next;
8883 const int line = nd_line(RNODE_ENSURE(node)->nd_ensr);
8884 const NODE *line_node = node;
8886 const rb_iseq_t *ensure = NEW_CHILD_ISEQ(RNODE_ENSURE(node)->nd_ensr,
8888 ISEQ_TYPE_ENSURE, line);
8889 LABEL *lstart = NEW_LABEL(line);
8890 LABEL *lend = NEW_LABEL(line);
8891 LABEL *lcont = NEW_LABEL(line);
8899 CHECK(COMPILE_POPPED(ensr,
"ensure ensr", RNODE_ENSURE(node)->nd_ensr));
8901 last_leave = last && IS_INSN(last) && IS_INSN_ID(last, leave);
8906 push_ensure_entry(iseq, &enl, &er, RNODE_ENSURE(node)->nd_ensr);
8908 ADD_LABEL(ret, lstart);
8909 CHECK(COMPILE_(ret,
"ensure head", RNODE_ENSURE(node)->nd_head, (popped | last_leave)));
8910 ADD_LABEL(ret, lend);
8912 if (!popped && last_leave) ADD_INSN(ret, line_node, putnil);
8913 ADD_LABEL(ret, lcont);
8914 if (last_leave) ADD_INSN(ret, line_node, pop);
8916 erange = ISEQ_COMPILE_DATA(iseq)->ensure_node_stack->erange;
8917 if (lstart->link.next != &lend->link) {
8919 ADD_CATCH_ENTRY(CATCH_TYPE_ENSURE, erange->begin, erange->end,
8921 erange = erange->next;
8925 ISEQ_COMPILE_DATA(iseq)->ensure_node_stack = enl.prev;
8932 const NODE *line_node = node;
8935 enum rb_iseq_type
type = ISEQ_BODY(iseq)->type;
8937 enum rb_iseq_type t =
type;
8938 const NODE *retval = RNODE_RETURN(node)->nd_stts;
8941 while (t == ISEQ_TYPE_RESCUE || t == ISEQ_TYPE_ENSURE) {
8942 if (!(is = ISEQ_BODY(is)->parent_iseq))
break;
8943 t = ISEQ_BODY(is)->type;
8947 case ISEQ_TYPE_MAIN:
8949 rb_warn(
"argument of top-level return is ignored");
8953 type = ISEQ_TYPE_METHOD;
8960 if (
type == ISEQ_TYPE_METHOD) {
8961 splabel = NEW_LABEL(0);
8962 ADD_LABEL(ret, splabel);
8963 ADD_ADJUST(ret, line_node, 0);
8966 CHECK(COMPILE(ret,
"return nd_stts (return val)", retval));
8968 if (
type == ISEQ_TYPE_METHOD && can_add_ensure_iseq(iseq)) {
8969 add_ensure_iseq(ret, iseq, 1);
8971 ADD_INSN(ret, line_node, leave);
8972 ADD_ADJUST_RESTORE(ret, splabel);
8975 ADD_INSN(ret, line_node, putnil);
8979 ADD_INSN1(ret, line_node,
throw,
INT2FIX(TAG_RETURN));
8981 ADD_INSN(ret, line_node, pop);
8992 if (!i)
return false;
8993 if (IS_TRACE(i)) i = i->prev;
8994 if (!IS_INSN(i) || !IS_INSN_ID(i, putnil))
return false;
8996 if (IS_ADJUST(i)) i = i->prev;
8997 if (!IS_INSN(i))
return false;
8998 switch (INSN_OF(i)) {
9005 (ret->last = last->prev)->next = NULL;
9012 CHECK(COMPILE_(ret,
"nd_body", node, popped));
9014 if (!popped && !all_string_result_p(node)) {
9015 const NODE *line_node = node;
9016 const unsigned int flag = VM_CALL_FCALL;
9020 ADD_INSN(ret, line_node, dup);
9021 ADD_INSN1(ret, line_node, objtostring, new_callinfo(iseq, idTo_s, 0, flag, NULL, FALSE));
9022 ADD_INSN(ret, line_node, anytostring);
9030 int idx = ISEQ_BODY(ISEQ_BODY(iseq)->local_iseq)->local_table_size - get_local_var_idx(iseq,
id);
9032 debugs(
"id: %s idx: %d\n", rb_id2name(
id), idx);
9033 ADD_GETLOCAL(ret, line_node, idx, get_lvar_level(iseq));
9039 LABEL *else_label = NEW_LABEL(nd_line(line_node));
9042 br = decl_branch_base(iseq, PTR2NUM(node), nd_code_loc(node),
"&.");
9044 ADD_INSN(recv, line_node, dup);
9045 ADD_INSNL(recv, line_node, branchnil, else_label);
9046 add_trace_branch_coverage(iseq, recv, nd_code_loc(node), nd_node_id(node), 0,
"then", br);
9054 if (!else_label)
return;
9055 end_label = NEW_LABEL(nd_line(line_node));
9056 ADD_INSNL(ret, line_node, jump, end_label);
9057 ADD_LABEL(ret, else_label);
9058 add_trace_branch_coverage(iseq, ret, nd_code_loc(node), nd_node_id(node), 1,
"else", branches);
9059 ADD_LABEL(ret, end_label);
9068 if (get_nd_recv(node) &&
9069 (nd_type_p(get_nd_recv(node), NODE_STR) || nd_type_p(get_nd_recv(node), NODE_FILE)) &&
9070 (get_node_call_nd_mid(node) == idFreeze || get_node_call_nd_mid(node) == idUMinus) &&
9071 get_nd_args(node) == NULL &&
9072 ISEQ_COMPILE_DATA(iseq)->current_block == NULL &&
9073 ISEQ_COMPILE_DATA(iseq)->option->specialized_instruction) {
9074 VALUE str = get_string_value(get_nd_recv(node));
9075 if (get_node_call_nd_mid(node) == idUMinus) {
9076 ADD_INSN2(ret, line_node, opt_str_uminus, str,
9077 new_callinfo(iseq, idUMinus, 0, 0, NULL, FALSE));
9080 ADD_INSN2(ret, line_node, opt_str_freeze, str,
9081 new_callinfo(iseq, idFreeze, 0, 0, NULL, FALSE));
9085 ADD_INSN(ret, line_node, pop);
9093iseq_has_builtin_function_table(
const rb_iseq_t *iseq)
9095 return ISEQ_COMPILE_DATA(iseq)->builtin_function_table != NULL;
9099iseq_builtin_function_lookup(
const rb_iseq_t *iseq,
const char *name)
9102 const struct rb_builtin_function *table = ISEQ_COMPILE_DATA(iseq)->builtin_function_table;
9103 for (i=0; table[i].index != -1; i++) {
9104 if (strcmp(table[i].name, name) == 0) {
9112iseq_builtin_function_name(
const enum node_type
type,
const NODE *recv,
ID mid)
9114 const char *name = rb_id2name(mid);
9115 static const char prefix[] =
"__builtin_";
9116 const size_t prefix_len =
sizeof(prefix) - 1;
9121 switch (nd_type(recv)) {
9123 if (RNODE_VCALL(recv)->nd_mid == rb_intern(
"__builtin")) {
9128 if (RNODE_CONST(recv)->nd_vid == rb_intern(
"Primitive")) {
9138 if (UNLIKELY(strncmp(prefix, name, prefix_len) == 0)) {
9139 return &name[prefix_len];
9148delegate_call_p(
const rb_iseq_t *iseq,
unsigned int argc,
const LINK_ANCHOR *args,
unsigned int *pstart_index)
9155 else if (argc <= ISEQ_BODY(iseq)->local_table_size) {
9156 unsigned int start=0;
9161 argc + start <= ISEQ_BODY(iseq)->local_table_size;
9165 for (
unsigned int i=start; i-start<argc; i++) {
9166 if (IS_INSN(elem) &&
9167 INSN_OF(elem) == BIN(getlocal)) {
9168 int local_index =
FIX2INT(OPERAND_AT(elem, 0));
9169 int local_level =
FIX2INT(OPERAND_AT(elem, 1));
9171 if (local_level == 0) {
9172 unsigned int index = ISEQ_BODY(iseq)->local_table_size - (local_index - VM_ENV_DATA_SIZE + 1);
9174 fprintf(stderr,
"lvar:%s (%d), id:%s (%d) local_index:%d, local_size:%d\n",
9175 rb_id2name(ISEQ_BODY(iseq)->local_table[i]), i,
9176 rb_id2name(ISEQ_BODY(iseq)->local_table[index]), index,
9177 local_index, (
int)ISEQ_BODY(iseq)->local_table_size);
9201 *pstart_index = start;
9215 if (!node)
goto no_arg;
9217 if (!nd_type_p(node, NODE_LIST))
goto bad_arg;
9218 const NODE *next = RNODE_LIST(node)->nd_next;
9220 node = RNODE_LIST(node)->nd_head;
9221 if (!node)
goto no_arg;
9222 switch (nd_type(node)) {
9224 symbol = rb_node_sym_string_val(node);
9230 if (!
SYMBOL_P(symbol))
goto non_symbol_arg;
9233 if (strcmp(RSTRING_PTR(
string),
"leaf") == 0) {
9234 ISEQ_BODY(iseq)->builtin_attrs |= BUILTIN_ATTR_LEAF;
9236 else if (strcmp(RSTRING_PTR(
string),
"inline_block") == 0) {
9237 ISEQ_BODY(iseq)->builtin_attrs |= BUILTIN_ATTR_INLINE_BLOCK;
9239 else if (strcmp(RSTRING_PTR(
string),
"use_block") == 0) {
9240 iseq_set_use_block(iseq);
9242 else if (strcmp(RSTRING_PTR(
string),
"c_trace") == 0) {
9244 ISEQ_BODY(iseq)->builtin_attrs |= BUILTIN_ATTR_C_TRACE;
9246 else if (strcmp(RSTRING_PTR(
string),
"without_interrupts") == 0) {
9247 ISEQ_BODY(iseq)->builtin_attrs |= BUILTIN_ATTR_WITHOUT_INTERRUPTS;
9256 COMPILE_ERROR(ERROR_ARGS
"attr!: no argument");
9259 COMPILE_ERROR(ERROR_ARGS
"non symbol argument to attr!: %s", rb_builtin_class_name(symbol));
9262 COMPILE_ERROR(ERROR_ARGS
"unknown argument to attr!: %s", RSTRING_PTR(
string));
9265 UNKNOWN_NODE(
"attr!", node, COMPILE_NG);
9273 if (!node)
goto no_arg;
9274 if (!nd_type_p(node, NODE_LIST))
goto bad_arg;
9275 if (RNODE_LIST(node)->nd_next)
goto too_many_arg;
9276 node = RNODE_LIST(node)->nd_head;
9277 if (!node)
goto no_arg;
9278 switch (nd_type(node)) {
9280 name = rb_node_sym_string_val(node);
9285 if (!
SYMBOL_P(name))
goto non_symbol_arg;
9287 compile_lvar(iseq, ret, line_node,
SYM2ID(name));
9291 COMPILE_ERROR(ERROR_ARGS
"arg!: no argument");
9294 COMPILE_ERROR(ERROR_ARGS
"arg!: too many argument");
9297 COMPILE_ERROR(ERROR_ARGS
"non symbol argument to arg!: %s",
9298 rb_builtin_class_name(name));
9301 UNKNOWN_NODE(
"arg!", node, COMPILE_NG);
9307 const NODE *node = ISEQ_COMPILE_DATA(iseq)->root_node;
9308 if (nd_type(node) == NODE_IF && RNODE_IF(node)->nd_cond == cond_node) {
9309 return RNODE_IF(node)->nd_body;
9312 rb_bug(
"mandatory_node: can't find mandatory node");
9317compile_builtin_mandatory_only_method(
rb_iseq_t *iseq,
const NODE *node,
const NODE *line_node)
9321 .pre_args_num = ISEQ_BODY(iseq)->param.lead_num,
9324 rb_node_init(RNODE(&args_node), NODE_ARGS);
9325 args_node.nd_ainfo = args;
9328 const int skip_local_size = ISEQ_BODY(iseq)->param.size - ISEQ_BODY(iseq)->param.lead_num;
9329 const int table_size = ISEQ_BODY(iseq)->local_table_size - skip_local_size;
9333 tbl->size = table_size;
9338 for (i=0; i<ISEQ_BODY(iseq)->param.lead_num; i++) {
9339 tbl->ids[i] = ISEQ_BODY(iseq)->local_table[i];
9342 for (; i<table_size; i++) {
9343 tbl->ids[i] = ISEQ_BODY(iseq)->local_table[i + skip_local_size];
9347 rb_node_init(RNODE(&scope_node), NODE_SCOPE);
9348 scope_node.nd_tbl = tbl;
9349 scope_node.nd_body = mandatory_node(iseq, node);
9350 scope_node.nd_parent = NULL;
9351 scope_node.nd_args = &args_node;
9353 VALUE ast_value = rb_ruby_ast_new(RNODE(&scope_node));
9356 rb_iseq_new_with_opt(ast_value, rb_iseq_base_label(iseq),
9357 rb_iseq_path(iseq), rb_iseq_realpath(iseq),
9358 nd_line(line_node), NULL, 0,
9359 ISEQ_TYPE_METHOD, ISEQ_COMPILE_DATA(iseq)->option,
9360 ISEQ_BODY(iseq)->variable.script_lines);
9361 RB_OBJ_WRITE(iseq, &ISEQ_BODY(iseq)->mandatory_only_iseq, (
VALUE)mandatory_only_iseq);
9371 NODE *args_node = get_nd_args(node);
9373 if (parent_block != NULL) {
9374 COMPILE_ERROR(ERROR_ARGS_AT(line_node)
"should not call builtins here.");
9378# define BUILTIN_INLINE_PREFIX "_bi"
9379 char inline_func[
sizeof(BUILTIN_INLINE_PREFIX) +
DECIMAL_SIZE_OF(
int)];
9380 bool cconst =
false;
9385 if (strcmp(
"cstmt!", builtin_func) == 0 ||
9386 strcmp(
"cexpr!", builtin_func) == 0) {
9389 else if (strcmp(
"cconst!", builtin_func) == 0) {
9392 else if (strcmp(
"cinit!", builtin_func) == 0) {
9396 else if (strcmp(
"attr!", builtin_func) == 0) {
9397 return compile_builtin_attr(iseq, args_node);
9399 else if (strcmp(
"arg!", builtin_func) == 0) {
9400 return compile_builtin_arg(iseq, ret, args_node, line_node, popped);
9402 else if (strcmp(
"mandatory_only?", builtin_func) == 0) {
9404 rb_bug(
"mandatory_only? should be in if condition");
9406 else if (!LIST_INSN_SIZE_ZERO(ret)) {
9407 rb_bug(
"mandatory_only? should be put on top");
9410 ADD_INSN1(ret, line_node, putobject,
Qfalse);
9411 return compile_builtin_mandatory_only_method(iseq, node, line_node);
9414 rb_bug(
"can't find builtin function:%s", builtin_func);
9417 COMPILE_ERROR(ERROR_ARGS
"can't find builtin function:%s", builtin_func);
9421 int inline_index = nd_line(node);
9422 snprintf(inline_func,
sizeof(inline_func), BUILTIN_INLINE_PREFIX
"%d", inline_index);
9423 builtin_func = inline_func;
9429 typedef VALUE(*builtin_func0)(
void *,
VALUE);
9430 VALUE const_val = (*(builtin_func0)(uintptr_t)bf->func_ptr)(NULL,
Qnil);
9431 ADD_INSN1(ret, line_node, putobject, const_val);
9437 unsigned int flag = 0;
9439 VALUE argc = setup_args(iseq, args, args_node, &flag, &keywords);
9441 if (
FIX2INT(argc) != bf->argc) {
9442 COMPILE_ERROR(ERROR_ARGS
"argc is not match for builtin function:%s (expect %d but %d)",
9443 builtin_func, bf->argc,
FIX2INT(argc));
9447 unsigned int start_index;
9448 if (delegate_call_p(iseq,
FIX2INT(argc), args, &start_index)) {
9449 ADD_INSN2(ret, line_node, opt_invokebuiltin_delegate, bf,
INT2FIX(start_index));
9453 ADD_INSN1(ret, line_node, invokebuiltin, bf);
9456 if (popped) ADD_INSN(ret, line_node, pop);
9462compile_call(
rb_iseq_t *iseq,
LINK_ANCHOR *
const ret,
const NODE *
const node,
const enum node_type
type,
const NODE *
const line_node,
int popped,
bool assume_receiver)
9470 ID mid = get_node_call_nd_mid(node);
9472 unsigned int flag = 0;
9474 const rb_iseq_t *parent_block = ISEQ_COMPILE_DATA(iseq)->current_block;
9475 LABEL *else_label = NULL;
9478 ISEQ_COMPILE_DATA(iseq)->current_block = NULL;
9484 if (nd_type_p(node, NODE_VCALL)) {
9489 CONST_ID(id_answer,
"the_answer_to_life_the_universe_and_everything");
9491 if (mid == id_bitblt) {
9492 ADD_INSN(ret, line_node, bitblt);
9495 else if (mid == id_answer) {
9496 ADD_INSN(ret, line_node, answer);
9508 if (nd_type_p(node, NODE_FCALL) &&
9509 (mid == goto_id || mid == label_id)) {
9512 st_table *labels_table = ISEQ_COMPILE_DATA(iseq)->labels_table;
9515 if (!labels_table) {
9516 labels_table = st_init_numtable();
9517 ISEQ_COMPILE_DATA(iseq)->labels_table = labels_table;
9520 COMPILE_ERROR(ERROR_ARGS
"invalid goto/label format");
9524 if (mid == goto_id) {
9525 ADD_INSNL(ret, line_node, jump, label);
9528 ADD_LABEL(ret, label);
9535 const char *builtin_func;
9536 if (UNLIKELY(iseq_has_builtin_function_table(iseq)) &&
9537 (builtin_func = iseq_builtin_function_name(
type, get_nd_recv(node), mid)) != NULL) {
9538 return compile_builtin_function_call(iseq, ret, node, line_node, popped, parent_block, args, builtin_func);
9542 if (!assume_receiver) {
9543 if (
type == NODE_CALL ||
type == NODE_OPCALL ||
type == NODE_QCALL) {
9546 if (mid == idCall &&
9547 nd_type_p(get_nd_recv(node), NODE_LVAR) &&
9548 iseq_block_param_id_p(iseq, RNODE_LVAR(get_nd_recv(node))->nd_vid, &idx, &level)) {
9549 ADD_INSN2(recv, get_nd_recv(node), getblockparamproxy,
INT2FIX(idx + VM_ENV_DATA_SIZE - 1),
INT2FIX(level));
9551 else if (private_recv_p(node)) {
9552 ADD_INSN(recv, node, putself);
9553 flag |= VM_CALL_FCALL;
9556 CHECK(COMPILE(recv,
"recv", get_nd_recv(node)));
9559 if (
type == NODE_QCALL) {
9560 else_label = qcall_branch_start(iseq, recv, &branches, node, line_node);
9563 else if (
type == NODE_FCALL ||
type == NODE_VCALL) {
9564 ADD_CALL_RECEIVER(recv, line_node);
9569 if (
type != NODE_VCALL) {
9570 argc = setup_args(iseq, args, get_nd_args(node), &flag, &keywords);
9571 CHECK(!
NIL_P(argc));
9579 bool inline_new = ISEQ_COMPILE_DATA(iseq)->option->specialized_instruction &&
9580 mid == rb_intern(
"new") &&
9581 parent_block == NULL &&
9582 !(flag & VM_CALL_ARGS_BLOCKARG);
9585 ADD_INSN(ret, node, putnil);
9586 ADD_INSN(ret, node, swap);
9591 debugp_param(
"call args argc", argc);
9592 debugp_param(
"call method",
ID2SYM(mid));
9594 switch ((
int)
type) {
9596 flag |= VM_CALL_VCALL;
9599 flag |= VM_CALL_FCALL;
9602 if ((flag & VM_CALL_ARGS_BLOCKARG) && (flag & VM_CALL_KW_SPLAT) && !(flag & VM_CALL_KW_SPLAT_MUT)) {
9603 ADD_INSN(ret, line_node, splatkw);
9606 LABEL *not_basic_new = NEW_LABEL(nd_line(node));
9607 LABEL *not_basic_new_finish = NEW_LABEL(nd_line(node));
9612 if (flag & VM_CALL_FORWARDING) {
9613 ci = (
VALUE)new_callinfo(iseq, mid,
NUM2INT(argc) + 1, flag, keywords, 0);
9616 ci = (
VALUE)new_callinfo(iseq, mid,
NUM2INT(argc), flag, keywords, 0);
9618 ADD_INSN2(ret, node, opt_new, ci, not_basic_new);
9619 LABEL_REF(not_basic_new);
9622 ADD_SEND_R(ret, line_node, rb_intern(
"initialize"), argc, parent_block,
INT2FIX(flag | VM_CALL_FCALL), keywords);
9623 ADD_INSNL(ret, line_node, jump, not_basic_new_finish);
9625 ADD_LABEL(ret, not_basic_new);
9627 ADD_SEND_R(ret, line_node, mid, argc, parent_block,
INT2FIX(flag), keywords);
9628 ADD_INSN(ret, line_node, swap);
9630 ADD_LABEL(ret, not_basic_new_finish);
9631 ADD_INSN(ret, line_node, pop);
9634 ADD_SEND_R(ret, line_node, mid, argc, parent_block,
INT2FIX(flag), keywords);
9637 qcall_branch_end(iseq, ret, else_label, branches, node, line_node);
9639 ADD_INSN(ret, line_node, pop);
9647 const int line = nd_line(node);
9649 unsigned int flag = 0;
9651 ID id = RNODE_OP_ASGN1(node)->nd_mid;
9677 ADD_INSN(ret, node, putnil);
9679 asgnflag = COMPILE_RECV(ret,
"NODE_OP_ASGN1 recv", node, RNODE_OP_ASGN1(node)->nd_recv);
9680 CHECK(asgnflag != -1);
9681 switch (nd_type(RNODE_OP_ASGN1(node)->nd_index)) {
9686 argc = setup_args(iseq, ret, RNODE_OP_ASGN1(node)->nd_index, &flag, NULL);
9687 CHECK(!
NIL_P(argc));
9689 int dup_argn =
FIX2INT(argc) + 1;
9690 ADD_INSN1(ret, node, dupn,
INT2FIX(dup_argn));
9692 ADD_SEND_R(ret, node, idAREF, argc, NULL,
INT2FIX(flag & ~VM_CALL_ARGS_SPLAT_MUT), NULL);
9694 if (
id == idOROP ||
id == idANDOP) {
9703 LABEL *label = NEW_LABEL(line);
9704 LABEL *lfin = NEW_LABEL(line);
9706 ADD_INSN(ret, node, dup);
9708 ADD_INSNL(ret, node, branchif, label);
9711 ADD_INSNL(ret, node, branchunless, label);
9713 ADD_INSN(ret, node, pop);
9715 CHECK(COMPILE(ret,
"NODE_OP_ASGN1 nd_rvalue: ", RNODE_OP_ASGN1(node)->nd_rvalue));
9717 ADD_INSN1(ret, node, setn,
INT2FIX(dup_argn+1));
9719 if (flag & VM_CALL_ARGS_SPLAT) {
9720 if (!(flag & VM_CALL_ARGS_SPLAT_MUT)) {
9721 ADD_INSN(ret, node, swap);
9722 ADD_INSN1(ret, node, splatarray,
Qtrue);
9723 ADD_INSN(ret, node, swap);
9724 flag |= VM_CALL_ARGS_SPLAT_MUT;
9726 ADD_INSN1(ret, node, pushtoarray,
INT2FIX(1));
9727 ADD_SEND_R(ret, node, idASET, argc, NULL,
INT2FIX(flag), NULL);
9730 ADD_SEND_R(ret, node, idASET, FIXNUM_INC(argc, 1), NULL,
INT2FIX(flag), NULL);
9732 ADD_INSN(ret, node, pop);
9733 ADD_INSNL(ret, node, jump, lfin);
9734 ADD_LABEL(ret, label);
9736 ADD_INSN1(ret, node, setn,
INT2FIX(dup_argn+1));
9738 ADD_INSN1(ret, node, adjuststack,
INT2FIX(dup_argn+1));
9739 ADD_LABEL(ret, lfin);
9742 CHECK(COMPILE(ret,
"NODE_OP_ASGN1 nd_rvalue: ", RNODE_OP_ASGN1(node)->nd_rvalue));
9743 ADD_SEND(ret, node,
id,
INT2FIX(1));
9745 ADD_INSN1(ret, node, setn,
INT2FIX(dup_argn+1));
9747 if (flag & VM_CALL_ARGS_SPLAT) {
9748 if (flag & VM_CALL_KW_SPLAT) {
9749 ADD_INSN1(ret, node, topn,
INT2FIX(2));
9750 if (!(flag & VM_CALL_ARGS_SPLAT_MUT)) {
9751 ADD_INSN1(ret, node, splatarray,
Qtrue);
9752 flag |= VM_CALL_ARGS_SPLAT_MUT;
9754 ADD_INSN(ret, node, swap);
9755 ADD_INSN1(ret, node, pushtoarray,
INT2FIX(1));
9756 ADD_INSN1(ret, node, setn,
INT2FIX(2));
9757 ADD_INSN(ret, node, pop);
9760 if (!(flag & VM_CALL_ARGS_SPLAT_MUT)) {
9761 ADD_INSN(ret, node, swap);
9762 ADD_INSN1(ret, node, splatarray,
Qtrue);
9763 ADD_INSN(ret, node, swap);
9764 flag |= VM_CALL_ARGS_SPLAT_MUT;
9766 ADD_INSN1(ret, node, pushtoarray,
INT2FIX(1));
9768 ADD_SEND_R(ret, node, idASET, argc, NULL,
INT2FIX(flag), NULL);
9771 ADD_SEND_R(ret, node, idASET, FIXNUM_INC(argc, 1), NULL,
INT2FIX(flag), NULL);
9773 ADD_INSN(ret, node, pop);
9781 const int line = nd_line(node);
9782 ID atype = RNODE_OP_ASGN2(node)->nd_mid;
9783 ID vid = RNODE_OP_ASGN2(node)->nd_vid, aid = rb_id_attrset(vid);
9785 LABEL *lfin = NEW_LABEL(line);
9786 LABEL *lcfin = NEW_LABEL(line);
9841 asgnflag = COMPILE_RECV(ret,
"NODE_OP_ASGN2#recv", node, RNODE_OP_ASGN2(node)->nd_recv);
9842 CHECK(asgnflag != -1);
9843 if (RNODE_OP_ASGN2(node)->nd_aid) {
9844 lskip = NEW_LABEL(line);
9845 ADD_INSN(ret, node, dup);
9846 ADD_INSNL(ret, node, branchnil, lskip);
9848 ADD_INSN(ret, node, dup);
9849 ADD_SEND_WITH_FLAG(ret, node, vid,
INT2FIX(0),
INT2FIX(asgnflag));
9851 if (atype == idOROP || atype == idANDOP) {
9853 ADD_INSN(ret, node, dup);
9855 if (atype == idOROP) {
9856 ADD_INSNL(ret, node, branchif, lcfin);
9859 ADD_INSNL(ret, node, branchunless, lcfin);
9862 ADD_INSN(ret, node, pop);
9864 CHECK(COMPILE(ret,
"NODE_OP_ASGN2 val", RNODE_OP_ASGN2(node)->nd_value));
9866 ADD_INSN(ret, node, swap);
9867 ADD_INSN1(ret, node, topn,
INT2FIX(1));
9869 ADD_SEND_WITH_FLAG(ret, node, aid,
INT2FIX(1),
INT2FIX(asgnflag));
9870 ADD_INSNL(ret, node, jump, lfin);
9872 ADD_LABEL(ret, lcfin);
9874 ADD_INSN(ret, node, swap);
9877 ADD_LABEL(ret, lfin);
9880 CHECK(COMPILE(ret,
"NODE_OP_ASGN2 val", RNODE_OP_ASGN2(node)->nd_value));
9881 ADD_SEND(ret, node, atype,
INT2FIX(1));
9883 ADD_INSN(ret, node, swap);
9884 ADD_INSN1(ret, node, topn,
INT2FIX(1));
9886 ADD_SEND_WITH_FLAG(ret, node, aid,
INT2FIX(1),
INT2FIX(asgnflag));
9888 if (lskip && popped) {
9889 ADD_LABEL(ret, lskip);
9891 ADD_INSN(ret, node, pop);
9892 if (lskip && !popped) {
9893 ADD_LABEL(ret, lskip);
9898static int compile_shareable_constant_value(
rb_iseq_t *iseq,
LINK_ANCHOR *ret,
enum rb_parser_shareability shareable,
const NODE *lhs,
const NODE *value);
9903 const int line = nd_line(node);
9908 switch (nd_type(RNODE_OP_CDECL(node)->nd_head)) {
9913 CHECK(COMPILE(ret,
"NODE_OP_CDECL/colon2#nd_head", RNODE_COLON2(RNODE_OP_CDECL(node)->nd_head)->nd_head));
9916 COMPILE_ERROR(ERROR_ARGS
"%s: invalid node in NODE_OP_CDECL",
9917 ruby_node_name(nd_type(RNODE_OP_CDECL(node)->nd_head)));
9920 mid = get_node_colon_nd_mid(RNODE_OP_CDECL(node)->nd_head);
9922 if (RNODE_OP_CDECL(node)->nd_aid == idOROP) {
9923 lassign = NEW_LABEL(line);
9924 ADD_INSN(ret, node, dup);
9925 ADD_INSN3(ret, node, defined,
INT2FIX(DEFINED_CONST_FROM),
9927 ADD_INSNL(ret, node, branchunless, lassign);
9929 ADD_INSN(ret, node, dup);
9930 ADD_INSN1(ret, node, putobject,
Qtrue);
9931 ADD_INSN1(ret, node, getconstant,
ID2SYM(mid));
9933 if (RNODE_OP_CDECL(node)->nd_aid == idOROP || RNODE_OP_CDECL(node)->nd_aid == idANDOP) {
9934 lfin = NEW_LABEL(line);
9935 if (!popped) ADD_INSN(ret, node, dup);
9936 if (RNODE_OP_CDECL(node)->nd_aid == idOROP)
9937 ADD_INSNL(ret, node, branchif, lfin);
9939 ADD_INSNL(ret, node, branchunless, lfin);
9941 if (!popped) ADD_INSN(ret, node, pop);
9942 if (lassign) ADD_LABEL(ret, lassign);
9943 CHECK(compile_shareable_constant_value(iseq, ret, RNODE_OP_CDECL(node)->shareability, RNODE_OP_CDECL(node)->nd_head, RNODE_OP_CDECL(node)->nd_value));
9946 ADD_INSN1(ret, node, topn,
INT2FIX(1));
9948 ADD_INSN1(ret, node, dupn,
INT2FIX(2));
9949 ADD_INSN(ret, node, swap);
9951 ADD_INSN1(ret, node, setconstant,
ID2SYM(mid));
9952 ADD_LABEL(ret, lfin);
9953 if (!popped) ADD_INSN(ret, node, swap);
9954 ADD_INSN(ret, node, pop);
9957 CHECK(compile_shareable_constant_value(iseq, ret, RNODE_OP_CDECL(node)->shareability, RNODE_OP_CDECL(node)->nd_head, RNODE_OP_CDECL(node)->nd_value));
9959 ADD_CALL(ret, node, RNODE_OP_CDECL(node)->nd_aid,
INT2FIX(1));
9961 ADD_INSN(ret, node, swap);
9963 ADD_INSN1(ret, node, topn,
INT2FIX(1));
9964 ADD_INSN(ret, node, swap);
9966 ADD_INSN1(ret, node, setconstant,
ID2SYM(mid));
9974 const int line = nd_line(node);
9975 LABEL *lfin = NEW_LABEL(line);
9978 if (
type == NODE_OP_ASGN_OR && !nd_type_p(RNODE_OP_ASGN_OR(node)->nd_head, NODE_IVAR)) {
9982 defined_expr(iseq, ret, RNODE_OP_ASGN_OR(node)->nd_head, lfinish,
Qfalse,
false);
9983 lassign = lfinish[1];
9985 lassign = NEW_LABEL(line);
9987 ADD_INSNL(ret, node, branchunless, lassign);
9990 lassign = NEW_LABEL(line);
9993 CHECK(COMPILE(ret,
"NODE_OP_ASGN_AND/OR#nd_head", RNODE_OP_ASGN_OR(node)->nd_head));
9996 ADD_INSN(ret, node, dup);
9999 if (
type == NODE_OP_ASGN_AND) {
10000 ADD_INSNL(ret, node, branchunless, lfin);
10003 ADD_INSNL(ret, node, branchif, lfin);
10007 ADD_INSN(ret, node, pop);
10010 ADD_LABEL(ret, lassign);
10011 CHECK(COMPILE_(ret,
"NODE_OP_ASGN_AND/OR#nd_value", RNODE_OP_ASGN_OR(node)->nd_value, popped));
10012 ADD_LABEL(ret, lfin);
10022 unsigned int flag = 0;
10024 const rb_iseq_t *parent_block = ISEQ_COMPILE_DATA(iseq)->current_block;
10028 ISEQ_COMPILE_DATA(iseq)->current_block = NULL;
10030 if (
type == NODE_SUPER) {
10031 VALUE vargc = setup_args(iseq, args, RNODE_SUPER(node)->nd_args, &flag, &keywords);
10032 CHECK(!
NIL_P(vargc));
10034 if ((flag & VM_CALL_ARGS_BLOCKARG) && (flag & VM_CALL_KW_SPLAT) && !(flag & VM_CALL_KW_SPLAT_MUT)) {
10035 ADD_INSN(args, node, splatkw);
10038 if (flag & VM_CALL_ARGS_BLOCKARG) {
10045 const rb_iseq_t *liseq = body->local_iseq;
10047 const struct rb_iseq_param_keyword *
const local_kwd = local_body->param.keyword;
10048 int lvar_level = get_lvar_level(iseq);
10050 argc = local_body->param.lead_num;
10053 for (i = 0; i < local_body->param.lead_num; i++) {
10054 int idx = local_body->local_table_size - i;
10055 ADD_GETLOCAL(args, node, idx, lvar_level);
10059 if (local_body->param.flags.forwardable) {
10060 flag |= VM_CALL_FORWARDING;
10061 int idx = local_body->local_table_size - get_local_var_idx(liseq, idDot3);
10062 ADD_GETLOCAL(args, node, idx, lvar_level);
10065 if (local_body->param.flags.has_opt) {
10068 for (j = 0; j < local_body->param.opt_num; j++) {
10069 int idx = local_body->local_table_size - (i + j);
10070 ADD_GETLOCAL(args, node, idx, lvar_level);
10075 if (local_body->param.flags.has_rest) {
10077 int idx = local_body->local_table_size - local_body->param.rest_start;
10078 ADD_GETLOCAL(args, node, idx, lvar_level);
10079 ADD_INSN1(args, node, splatarray, RBOOL(local_body->param.flags.has_post));
10081 argc = local_body->param.rest_start + 1;
10082 flag |= VM_CALL_ARGS_SPLAT;
10084 if (local_body->param.flags.has_post) {
10086 int post_len = local_body->param.post_num;
10087 int post_start = local_body->param.post_start;
10089 if (local_body->param.flags.has_rest) {
10091 for (j=0; j<post_len; j++) {
10092 int idx = local_body->local_table_size - (post_start + j);
10093 ADD_GETLOCAL(args, node, idx, lvar_level);
10095 ADD_INSN1(args, node, pushtoarray,
INT2FIX(j));
10096 flag |= VM_CALL_ARGS_SPLAT_MUT;
10101 for (j=0; j<post_len; j++) {
10102 int idx = local_body->local_table_size - (post_start + j);
10103 ADD_GETLOCAL(args, node, idx, lvar_level);
10105 argc = post_len + post_start;
10109 if (local_body->param.flags.has_kw) {
10110 int local_size = local_body->local_table_size;
10113 ADD_INSN1(args, node, putspecialobject,
INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
10115 if (local_body->param.flags.has_kwrest) {
10116 int idx = local_body->local_table_size - local_kwd->rest_start;
10117 ADD_GETLOCAL(args, node, idx, lvar_level);
10119 ADD_SEND (args, node, rb_intern(
"dup"),
INT2FIX(0));
10122 ADD_INSN1(args, node, newhash,
INT2FIX(0));
10124 for (i = 0; i < local_kwd->num; ++i) {
10125 ID id = local_kwd->table[i];
10126 int idx = local_size - get_local_var_idx(liseq,
id);
10127 ADD_INSN1(args, node, putobject,
ID2SYM(
id));
10128 ADD_GETLOCAL(args, node, idx, lvar_level);
10130 ADD_SEND(args, node, id_core_hash_merge_ptr,
INT2FIX(i * 2 + 1));
10131 flag |= VM_CALL_KW_SPLAT| VM_CALL_KW_SPLAT_MUT;
10133 else if (local_body->param.flags.has_kwrest) {
10134 int idx = local_body->local_table_size - local_kwd->rest_start;
10135 ADD_GETLOCAL(args, node, idx, lvar_level);
10137 flag |= VM_CALL_KW_SPLAT;
10141 if (use_block && parent_block == NULL) {
10142 iseq_set_use_block(ISEQ_BODY(iseq)->local_iseq);
10145 flag |= VM_CALL_SUPER | VM_CALL_FCALL;
10146 if (
type == NODE_ZSUPER) flag |= VM_CALL_ZSUPER;
10147 ADD_INSN(ret, node, putself);
10148 ADD_SEQ(ret, args);
10150 const struct rb_callinfo * ci = new_callinfo(iseq, 0, argc, flag, keywords, parent_block != NULL);
10152 if (vm_ci_flag(ci) & VM_CALL_FORWARDING) {
10153 ADD_INSN2(ret, node, invokesuperforward, ci, parent_block);
10156 ADD_INSN2(ret, node, invokesuper, ci, parent_block);
10160 ADD_INSN(ret, node, pop);
10170 unsigned int flag = 0;
10175 switch (ISEQ_BODY(ISEQ_BODY(iseq)->local_iseq)->
type) {
10176 case ISEQ_TYPE_TOP:
10177 case ISEQ_TYPE_MAIN:
10178 case ISEQ_TYPE_CLASS:
10179 COMPILE_ERROR(ERROR_ARGS
"Invalid yield");
10184 if (RNODE_YIELD(node)->nd_head) {
10185 argc = setup_args(iseq, args, RNODE_YIELD(node)->nd_head, &flag, &keywords);
10186 CHECK(!
NIL_P(argc));
10192 ADD_SEQ(ret, args);
10193 ADD_INSN1(ret, node, invokeblock, new_callinfo(iseq, 0,
FIX2INT(argc), flag, keywords, FALSE));
10194 iseq_set_use_block(ISEQ_BODY(iseq)->local_iseq);
10197 ADD_INSN(ret, node, pop);
10202 for (; tmp_iseq != ISEQ_BODY(iseq)->local_iseq; level++ ) {
10203 tmp_iseq = ISEQ_BODY(tmp_iseq)->parent_iseq;
10205 if (level > 0) access_outer_variables(iseq, level, rb_intern(
"yield"),
true);
10218 switch ((
int)
type) {
10221 VALUE re = rb_node_regx_string_val(node);
10222 RB_OBJ_SET_FROZEN_SHAREABLE(re);
10223 ADD_INSN1(recv, node, putobject, re);
10224 ADD_INSN2(val, node, getspecial,
INT2FIX(0),
10229 CHECK(COMPILE(recv,
"receiver", RNODE_MATCH2(node)->nd_recv));
10230 CHECK(COMPILE(val,
"value", RNODE_MATCH2(node)->nd_value));
10233 CHECK(COMPILE(recv,
"receiver", RNODE_MATCH3(node)->nd_value));
10234 CHECK(COMPILE(val,
"value", RNODE_MATCH3(node)->nd_recv));
10238 ADD_SEQ(ret, recv);
10240 ADD_SEND(ret, node, idEqTilde,
INT2FIX(1));
10242 if (nd_type_p(node, NODE_MATCH2) && RNODE_MATCH2(node)->nd_args) {
10243 compile_named_capture_assign(iseq, ret, RNODE_MATCH2(node)->nd_args);
10247 ADD_INSN(ret, node, pop);
10258 if (ISEQ_COMPILE_DATA(iseq)->option->inline_const_cache &&
10259 (segments = collect_const_segments(iseq, node))) {
10260 ISEQ_BODY(iseq)->ic_size++;
10261 ADD_INSN1(ret, node, opt_getconstant_path, segments);
10271 CHECK(compile_const_prefix(iseq, node, pref, body));
10272 if (LIST_INSN_SIZE_ZERO(pref)) {
10273 ADD_INSN(ret, node, putnil);
10274 ADD_SEQ(ret, body);
10277 ADD_SEQ(ret, pref);
10278 ADD_SEQ(ret, body);
10284 ADD_CALL_RECEIVER(ret, node);
10285 CHECK(COMPILE(ret,
"colon2#nd_head", RNODE_COLON2(node)->nd_head));
10286 ADD_CALL(ret, node, RNODE_COLON2(node)->nd_mid,
INT2FIX(1));
10289 ADD_INSN(ret, node, pop);
10297 debugi(
"colon3#nd_mid", RNODE_COLON3(node)->nd_mid);
10300 if (ISEQ_COMPILE_DATA(iseq)->option->inline_const_cache) {
10301 ISEQ_BODY(iseq)->ic_size++;
10302 VALUE segments = rb_ary_new_from_args(2,
ID2SYM(idNULL),
ID2SYM(RNODE_COLON3(node)->nd_mid));
10303 RB_OBJ_SET_FROZEN_SHAREABLE(segments);
10304 ADD_INSN1(ret, node, opt_getconstant_path, segments);
10308 ADD_INSN1(ret, node, putobject,
rb_cObject);
10309 ADD_INSN1(ret, node, putobject,
Qtrue);
10310 ADD_INSN1(ret, node, getconstant,
ID2SYM(RNODE_COLON3(node)->nd_mid));
10314 ADD_INSN(ret, node, pop);
10323 const NODE *b = RNODE_DOT2(node)->nd_beg;
10324 const NODE *e = RNODE_DOT2(node)->nd_end;
10326 if (optimizable_range_item_p(b) && optimizable_range_item_p(e)) {
10328 VALUE bv = optimized_range_item(b);
10329 VALUE ev = optimized_range_item(e);
10332 ADD_INSN1(ret, node, putobject, val);
10337 CHECK(COMPILE_(ret,
"min", b, popped));
10338 CHECK(COMPILE_(ret,
"max", e, popped));
10340 ADD_INSN1(ret, node, newrange, flag);
10350 if (ISEQ_BODY(iseq)->
type == ISEQ_TYPE_RESCUE) {
10351 ADD_GETLOCAL(ret, node, LVAR_ERRINFO, 0);
10357 if (ISEQ_BODY(ip)->
type == ISEQ_TYPE_RESCUE) {
10360 ip = ISEQ_BODY(ip)->parent_iseq;
10364 ADD_GETLOCAL(ret, node, LVAR_ERRINFO, level);
10367 ADD_INSN(ret, node, putnil);
10378 LABEL *end_label = NEW_LABEL(nd_line(node));
10379 const NODE *default_value = get_nd_value(RNODE_KW_ARG(node)->nd_body);
10381 if (default_value == NODE_SPECIAL_REQUIRED_KEYWORD) {
10383 COMPILE_ERROR(ERROR_ARGS
"unreachable");
10386 else if (nd_type_p(default_value, NODE_SYM) ||
10387 nd_type_p(default_value, NODE_REGX) ||
10388 nd_type_p(default_value, NODE_LINE) ||
10389 nd_type_p(default_value, NODE_INTEGER) ||
10390 nd_type_p(default_value, NODE_FLOAT) ||
10391 nd_type_p(default_value, NODE_RATIONAL) ||
10392 nd_type_p(default_value, NODE_IMAGINARY) ||
10393 nd_type_p(default_value, NODE_NIL) ||
10394 nd_type_p(default_value, NODE_TRUE) ||
10395 nd_type_p(default_value, NODE_FALSE)) {
10396 COMPILE_ERROR(ERROR_ARGS
"unreachable");
10404 int kw_bits_idx = body->local_table_size - body->param.keyword->bits_start;
10405 int keyword_idx = body->param.keyword->num;
10407 ADD_INSN2(ret, node, checkkeyword,
INT2FIX(kw_bits_idx + VM_ENV_DATA_SIZE - 1),
INT2FIX(keyword_idx));
10408 ADD_INSNL(ret, node, branchif, end_label);
10409 CHECK(COMPILE_POPPED(ret,
"keyword default argument", RNODE_KW_ARG(node)->nd_body));
10410 ADD_LABEL(ret, end_label);
10420 unsigned int flag = 0;
10421 ID mid = RNODE_ATTRASGN(node)->nd_mid;
10423 LABEL *else_label = NULL;
10428 argc = setup_args(iseq, args, RNODE_ATTRASGN(node)->nd_args, &flag, NULL);
10429 CHECK(!
NIL_P(argc));
10431 int asgnflag = COMPILE_RECV(recv,
"recv", node, RNODE_ATTRASGN(node)->nd_recv);
10432 CHECK(asgnflag != -1);
10433 flag |= (
unsigned int)asgnflag;
10435 debugp_param(
"argc", argc);
10436 debugp_param(
"nd_mid",
ID2SYM(mid));
10440 mid = rb_id_attrset(mid);
10441 else_label = qcall_branch_start(iseq, recv, &branches, node, node);
10444 ADD_INSN(ret, node, putnil);
10445 ADD_SEQ(ret, recv);
10446 ADD_SEQ(ret, args);
10448 if (flag & VM_CALL_ARGS_SPLAT) {
10449 ADD_INSN(ret, node, dup);
10450 ADD_INSN1(ret, node, putobject,
INT2FIX(-1));
10451 ADD_SEND_WITH_FLAG(ret, node, idAREF,
INT2FIX(1),
INT2FIX(asgnflag));
10452 ADD_INSN1(ret, node, setn, FIXNUM_INC(argc, 2));
10453 ADD_INSN (ret, node, pop);
10456 ADD_INSN1(ret, node, setn, FIXNUM_INC(argc, 1));
10460 ADD_SEQ(ret, recv);
10461 ADD_SEQ(ret, args);
10463 ADD_SEND_WITH_FLAG(ret, node, mid, argc,
INT2FIX(flag));
10464 qcall_branch_end(iseq, ret, else_label, branches, node, node);
10465 ADD_INSN(ret, node, pop);
10472 ADD_INSN1(ret, value, putspecialobject,
INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
10480 ADD_SEND_WITH_FLAG(ret, value, rb_intern(
"make_shareable_copy"),
INT2FIX(1),
INT2FIX(VM_CALL_ARGS_SIMPLE));
10487 ADD_SEND_WITH_FLAG(ret, value, rb_intern(
"make_shareable"),
INT2FIX(1),
INT2FIX(VM_CALL_ARGS_SIMPLE));
10494node_const_decl_val(
const NODE *node)
10497 switch (nd_type(node)) {
10499 if (RNODE_CDECL(node)->nd_vid) {
10500 path = rb_id2str(RNODE_CDECL(node)->nd_vid);
10504 node = RNODE_CDECL(node)->nd_else;
10512 rb_str_append(path, rb_id2str(RNODE_COLON3(node)->nd_mid));
10515 rb_bug(
"unexpected node: %s", ruby_node_name(nd_type(node)));
10521 for (; node && nd_type_p(node, NODE_COLON2); node = RNODE_COLON2(node)->nd_head) {
10522 rb_ary_push(path, rb_id2str(RNODE_COLON2(node)->nd_mid));
10524 if (node && nd_type_p(node, NODE_CONST)) {
10526 rb_ary_push(path, rb_id2str(RNODE_CONST(node)->nd_vid));
10528 else if (node && nd_type_p(node, NODE_COLON3)) {
10530 rb_ary_push(path, rb_id2str(RNODE_COLON3(node)->nd_mid));
10540 path = rb_fstring(path);
10545const_decl_path(
NODE *dest)
10548 if (!nd_type_p(dest, NODE_CALL)) {
10549 path = node_const_decl_val(dest);
10560 VALUE path = const_decl_path(dest);
10561 ADD_INSN1(ret, value, putspecialobject,
INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
10562 CHECK(COMPILE(ret,
"compile_ensure_shareable_node", value));
10563 ADD_INSN1(ret, value, putobject, path);
10565 ADD_SEND_WITH_FLAG(ret, value, rb_intern(
"ensure_shareable"),
INT2FIX(2),
INT2FIX(VM_CALL_ARGS_SIMPLE));
10570#ifndef SHAREABLE_BARE_EXPRESSION
10571#define SHAREABLE_BARE_EXPRESSION 1
10575compile_shareable_literal_constant(
rb_iseq_t *iseq,
LINK_ANCHOR *ret,
enum rb_parser_shareability shareable,
NODE *dest,
const NODE *node,
size_t level,
VALUE *value_p,
int *shareable_literal_p)
10577# define compile_shareable_literal_constant_next(node, anchor, value_p, shareable_literal_p) \
10578 compile_shareable_literal_constant(iseq, anchor, shareable, dest, node, level+1, value_p, shareable_literal_p)
10580 DECL_ANCHOR(anchor);
10582 enum node_type
type = node ? nd_type(node) : NODE_NIL;
10594 *value_p = rb_node_sym_string_val(node);
10597 *value_p = rb_node_regx_string_val(node);
10600 *value_p = rb_node_line_lineno_val(node);
10603 *value_p = rb_node_integer_literal_val(node);
10606 *value_p = rb_node_float_literal_val(node);
10608 case NODE_RATIONAL:
10609 *value_p = rb_node_rational_literal_val(node);
10611 case NODE_IMAGINARY:
10612 *value_p = rb_node_imaginary_literal_val(node);
10614 case NODE_ENCODING:
10615 *value_p = rb_node_encoding_val(node);
10618 CHECK(COMPILE(ret,
"shareable_literal_constant", node));
10619 *shareable_literal_p = 1;
10623 CHECK(COMPILE(ret,
"shareable_literal_constant", node));
10624 if (shareable == rb_parser_shareable_literal) {
10630 ADD_SEND_WITH_FLAG(ret, node, idUMinus,
INT2FIX(0),
INT2FIX(VM_CALL_ARGS_SIMPLE));
10633 *shareable_literal_p = 1;
10637 VALUE lit = rb_node_str_string_val(node);
10638 ADD_INSN1(ret, node, putobject, lit);
10641 *shareable_literal_p = 1;
10647 VALUE lit = rb_node_file_path_val(node);
10648 ADD_INSN1(ret, node, putobject, lit);
10651 *shareable_literal_p = 1;
10659 ADD_INSN1(ret, node, putobject, lit);
10662 *shareable_literal_p = 1;
10668 INIT_ANCHOR(anchor);
10670 for (
NODE *n = (
NODE *)node; n; n = RNODE_LIST(n)->nd_next) {
10672 int shareable_literal_p2;
10673 NODE *elt = RNODE_LIST(n)->nd_head;
10675 CHECK(compile_shareable_literal_constant_next(elt, anchor, &val, &shareable_literal_p2));
10676 if (shareable_literal_p2) {
10679 else if (
RTEST(lit)) {
10685 if (!UNDEF_P(val)) {
10697 if (!RNODE_HASH(node)->nd_brace) {
10699 *shareable_literal_p = 0;
10702 for (
NODE *n = RNODE_HASH(node)->nd_head; n; n = RNODE_LIST(RNODE_LIST(n)->nd_next)->nd_next) {
10703 if (!RNODE_LIST(n)->nd_head) {
10705 goto compile_shareable;
10709 INIT_ANCHOR(anchor);
10710 lit = rb_hash_new();
10711 for (
NODE *n = RNODE_HASH(node)->nd_head; n; n = RNODE_LIST(RNODE_LIST(n)->nd_next)->nd_next) {
10713 VALUE value_val = 0;
10714 int shareable_literal_p2;
10715 NODE *key = RNODE_LIST(n)->nd_head;
10716 NODE *val = RNODE_LIST(RNODE_LIST(n)->nd_next)->nd_head;
10717 CHECK(compile_shareable_literal_constant_next(key, anchor, &key_val, &shareable_literal_p2));
10718 if (shareable_literal_p2) {
10721 else if (
RTEST(lit)) {
10722 rb_hash_clear(lit);
10725 CHECK(compile_shareable_literal_constant_next(val, anchor, &value_val, &shareable_literal_p2));
10726 if (shareable_literal_p2) {
10729 else if (
RTEST(lit)) {
10730 rb_hash_clear(lit);
10734 if (!UNDEF_P(key_val) && !UNDEF_P(value_val)) {
10735 rb_hash_aset(lit, key_val, value_val);
10738 rb_hash_clear(lit);
10749 if (shareable == rb_parser_shareable_literal &&
10750 (SHAREABLE_BARE_EXPRESSION || level > 0)) {
10751 CHECK(compile_ensure_shareable_node(iseq, ret, dest, node));
10753 *shareable_literal_p = 1;
10756 CHECK(COMPILE(ret,
"shareable_literal_constant", node));
10758 *shareable_literal_p = 0;
10764 if (nd_type(node) == NODE_LIST) {
10765 ADD_INSN1(anchor, node, newarray,
INT2FIX(RNODE_LIST(node)->as.nd_alen));
10767 else if (nd_type(node) == NODE_HASH) {
10768 int len = (int)RNODE_LIST(RNODE_HASH(node)->nd_head)->as.nd_alen;
10769 ADD_INSN1(anchor, node, newhash,
INT2FIX(
len));
10772 *shareable_literal_p = 0;
10773 ADD_SEQ(ret, anchor);
10779 if (nd_type(node) == NODE_LIST) {
10780 ADD_INSN1(anchor, node, newarray,
INT2FIX(RNODE_LIST(node)->as.nd_alen));
10782 else if (nd_type(node) == NODE_HASH) {
10783 int len = (int)RNODE_LIST(RNODE_HASH(node)->nd_head)->as.nd_alen;
10784 ADD_INSN1(anchor, node, newhash,
INT2FIX(
len));
10786 CHECK(compile_make_shareable_node(iseq, ret, anchor, node,
false));
10788 *shareable_literal_p = 1;
10792 ADD_INSN1(ret, node, putobject, val);
10795 *shareable_literal_p = 1;
10802compile_shareable_constant_value(
rb_iseq_t *iseq,
LINK_ANCHOR *ret,
enum rb_parser_shareability shareable,
const NODE *lhs,
const NODE *value)
10806 DECL_ANCHOR(anchor);
10807 INIT_ANCHOR(anchor);
10809 switch (shareable) {
10810 case rb_parser_shareable_none:
10811 CHECK(COMPILE(ret,
"compile_shareable_constant_value", value));
10814 case rb_parser_shareable_literal:
10815 CHECK(compile_shareable_literal_constant(iseq, anchor, shareable, (
NODE *)lhs, value, 0, &val, &literal_p));
10816 ADD_SEQ(ret, anchor);
10819 case rb_parser_shareable_copy:
10820 case rb_parser_shareable_everything:
10821 CHECK(compile_shareable_literal_constant(iseq, anchor, shareable, (
NODE *)lhs, value, 0, &val, &literal_p));
10823 CHECK(compile_make_shareable_node(iseq, ret, anchor, value, shareable == rb_parser_shareable_copy));
10826 ADD_SEQ(ret, anchor);
10830 rb_bug(
"unexpected rb_parser_shareability: %d", shareable);
10847 int lineno = ISEQ_COMPILE_DATA(iseq)->last_line;
10848 if (lineno == 0) lineno =
FIX2INT(rb_iseq_first_lineno(iseq));
10849 debugs(
"node: NODE_NIL(implicit)\n");
10850 ADD_SYNTHETIC_INSN(ret, lineno, -1, putnil);
10854 return iseq_compile_each0(iseq, ret, node, popped);
10860 const int line = (int)nd_line(node);
10861 const enum node_type
type = nd_type(node);
10864 if (ISEQ_COMPILE_DATA(iseq)->last_line == line) {
10868 if (nd_fl_newline(node)) {
10870 ISEQ_COMPILE_DATA(iseq)->last_line = line;
10871 if (line > 0 && ISEQ_COVERAGE(iseq) && ISEQ_LINE_COVERAGE(iseq)) {
10872 event |= RUBY_EVENT_COVERAGE_LINE;
10874 ADD_TRACE(ret, event);
10878 debug_node_start(node);
10879#undef BEFORE_RETURN
10880#define BEFORE_RETURN debug_node_end()
10884 CHECK(compile_block(iseq, ret, node, popped));
10888 CHECK(compile_if(iseq, ret, node, popped,
type));
10891 CHECK(compile_case(iseq, ret, node, popped));
10894 CHECK(compile_case2(iseq, ret, node, popped));
10897 CHECK(compile_case3(iseq, ret, node, popped));
10901 CHECK(compile_loop(iseq, ret, node, popped,
type));
10905 CHECK(compile_iter(iseq, ret, node, popped));
10907 case NODE_FOR_MASGN:
10908 CHECK(compile_for_masgn(iseq, ret, node, popped));
10911 CHECK(compile_break(iseq, ret, node, popped));
10914 CHECK(compile_next(iseq, ret, node, popped));
10917 CHECK(compile_redo(iseq, ret, node, popped));
10920 CHECK(compile_retry(iseq, ret, node, popped));
10923 CHECK(COMPILE_(ret,
"NODE_BEGIN", RNODE_BEGIN(node)->nd_body, popped));
10927 CHECK(compile_rescue(iseq, ret, node, popped));
10930 CHECK(compile_resbody(iseq, ret, node, popped));
10933 CHECK(compile_ensure(iseq, ret, node, popped));
10938 LABEL *end_label = NEW_LABEL(line);
10939 CHECK(COMPILE(ret,
"nd_1st", RNODE_OR(node)->nd_1st));
10941 ADD_INSN(ret, node, dup);
10943 if (
type == NODE_AND) {
10944 ADD_INSNL(ret, node, branchunless, end_label);
10947 ADD_INSNL(ret, node, branchif, end_label);
10950 ADD_INSN(ret, node, pop);
10952 CHECK(COMPILE_(ret,
"nd_2nd", RNODE_OR(node)->nd_2nd, popped));
10953 ADD_LABEL(ret, end_label);
10958 compile_massign(iseq, ret, node, popped);
10963 ID id = RNODE_LASGN(node)->nd_vid;
10964 int idx = ISEQ_BODY(body->local_iseq)->local_table_size - get_local_var_idx(iseq,
id);
10966 debugs(
"lvar: %s idx: %d\n", rb_id2name(
id), idx);
10967 CHECK(COMPILE(ret,
"rvalue", RNODE_LASGN(node)->nd_value));
10970 ADD_INSN(ret, node, dup);
10972 ADD_SETLOCAL(ret, node, idx, get_lvar_level(iseq));
10977 ID id = RNODE_DASGN(node)->nd_vid;
10978 CHECK(COMPILE(ret,
"dvalue", RNODE_DASGN(node)->nd_value));
10979 debugi(
"dassn id", rb_id2str(
id) ?
id :
'*');
10982 ADD_INSN(ret, node, dup);
10985 idx = get_dyna_var_idx(iseq,
id, &lv, &ls);
10988 COMPILE_ERROR(ERROR_ARGS
"NODE_DASGN: unknown id (%"PRIsVALUE
")",
10992 ADD_SETLOCAL(ret, node, ls - idx, lv);
10996 CHECK(COMPILE(ret,
"lvalue", RNODE_GASGN(node)->nd_value));
10999 ADD_INSN(ret, node, dup);
11001 ADD_INSN1(ret, node, setglobal,
ID2SYM(RNODE_GASGN(node)->nd_vid));
11005 CHECK(COMPILE(ret,
"lvalue", RNODE_IASGN(node)->nd_value));
11007 ADD_INSN(ret, node, dup);
11009 ADD_INSN2(ret, node, setinstancevariable,
11010 ID2SYM(RNODE_IASGN(node)->nd_vid),
11011 get_ivar_ic_value(iseq,RNODE_IASGN(node)->nd_vid));
11015 if (RNODE_CDECL(node)->nd_vid) {
11016 CHECK(compile_shareable_constant_value(iseq, ret, RNODE_CDECL(node)->shareability, node, RNODE_CDECL(node)->nd_value));
11019 ADD_INSN(ret, node, dup);
11022 ADD_INSN1(ret, node, putspecialobject,
11023 INT2FIX(VM_SPECIAL_OBJECT_CONST_BASE));
11024 ADD_INSN1(ret, node, setconstant,
ID2SYM(RNODE_CDECL(node)->nd_vid));
11027 compile_cpath(ret, iseq, RNODE_CDECL(node)->nd_else);
11028 CHECK(compile_shareable_constant_value(iseq, ret, RNODE_CDECL(node)->shareability, node, RNODE_CDECL(node)->nd_value));
11029 ADD_INSN(ret, node, swap);
11032 ADD_INSN1(ret, node, topn,
INT2FIX(1));
11033 ADD_INSN(ret, node, swap);
11036 ADD_INSN1(ret, node, setconstant,
ID2SYM(get_node_colon_nd_mid(RNODE_CDECL(node)->nd_else)));
11041 CHECK(COMPILE(ret,
"cvasgn val", RNODE_CVASGN(node)->nd_value));
11043 ADD_INSN(ret, node, dup);
11045 ADD_INSN2(ret, node, setclassvariable,
11046 ID2SYM(RNODE_CVASGN(node)->nd_vid),
11047 get_cvar_ic_value(iseq, RNODE_CVASGN(node)->nd_vid));
11050 case NODE_OP_ASGN1:
11051 CHECK(compile_op_asgn1(iseq, ret, node, popped));
11053 case NODE_OP_ASGN2:
11054 CHECK(compile_op_asgn2(iseq, ret, node, popped));
11056 case NODE_OP_CDECL:
11057 CHECK(compile_op_cdecl(iseq, ret, node, popped));
11059 case NODE_OP_ASGN_AND:
11060 case NODE_OP_ASGN_OR:
11061 CHECK(compile_op_log(iseq, ret, node, popped,
type));
11065 if (compile_call_precheck_freeze(iseq, ret, node, node, popped) == TRUE) {
11071 if (compile_call(iseq, ret, node,
type, node, popped,
false) == COMPILE_NG) {
11077 CHECK(compile_super(iseq, ret, node, popped,
type));
11080 CHECK(compile_array(iseq, ret, node, popped, TRUE) >= 0);
11085 ADD_INSN1(ret, node, newarray,
INT2FIX(0));
11090 CHECK(compile_hash(iseq, ret, node, FALSE, popped) >= 0);
11093 CHECK(compile_return(iseq, ret, node, popped));
11096 CHECK(compile_yield(iseq, ret, node, popped));
11100 compile_lvar(iseq, ret, node, RNODE_LVAR(node)->nd_vid);
11106 debugi(
"nd_vid", RNODE_DVAR(node)->nd_vid);
11108 idx = get_dyna_var_idx(iseq, RNODE_DVAR(node)->nd_vid, &lv, &ls);
11110 COMPILE_ERROR(ERROR_ARGS
"unknown dvar (%"PRIsVALUE
")",
11111 rb_id2str(RNODE_DVAR(node)->nd_vid));
11114 ADD_GETLOCAL(ret, node, ls - idx, lv);
11119 ADD_INSN1(ret, node, getglobal,
ID2SYM(RNODE_GVAR(node)->nd_vid));
11121 ADD_INSN(ret, node, pop);
11126 debugi(
"nd_vid", RNODE_IVAR(node)->nd_vid);
11128 ADD_INSN2(ret, node, getinstancevariable,
11129 ID2SYM(RNODE_IVAR(node)->nd_vid),
11130 get_ivar_ic_value(iseq, RNODE_IVAR(node)->nd_vid));
11135 debugi(
"nd_vid", RNODE_CONST(node)->nd_vid);
11137 if (ISEQ_COMPILE_DATA(iseq)->option->inline_const_cache) {
11139 VALUE segments = rb_ary_new_from_args(1,
ID2SYM(RNODE_CONST(node)->nd_vid));
11140 RB_OBJ_SET_FROZEN_SHAREABLE(segments);
11141 ADD_INSN1(ret, node, opt_getconstant_path, segments);
11145 ADD_INSN(ret, node, putnil);
11146 ADD_INSN1(ret, node, putobject,
Qtrue);
11147 ADD_INSN1(ret, node, getconstant,
ID2SYM(RNODE_CONST(node)->nd_vid));
11151 ADD_INSN(ret, node, pop);
11157 ADD_INSN2(ret, node, getclassvariable,
11158 ID2SYM(RNODE_CVAR(node)->nd_vid),
11159 get_cvar_ic_value(iseq, RNODE_CVAR(node)->nd_vid));
11163 case NODE_NTH_REF:{
11165 if (!RNODE_NTH_REF(node)->nd_nth) {
11166 ADD_INSN(ret, node, putnil);
11169 ADD_INSN2(ret, node, getspecial,
INT2FIX(1) ,
11170 INT2FIX(RNODE_NTH_REF(node)->nd_nth << 1));
11174 case NODE_BACK_REF:{
11176 ADD_INSN2(ret, node, getspecial,
INT2FIX(1) ,
11177 INT2FIX(0x01 | (RNODE_BACK_REF(node)->nd_nth << 1)));
11184 CHECK(compile_match(iseq, ret, node, popped,
type));
11188 ADD_INSN1(ret, node, putobject, rb_node_sym_string_val(node));
11194 ADD_INSN1(ret, node, putobject, rb_node_line_lineno_val(node));
11198 case NODE_ENCODING:{
11200 ADD_INSN1(ret, node, putobject, rb_node_encoding_val(node));
11204 case NODE_INTEGER:{
11205 VALUE lit = rb_node_integer_literal_val(node);
11207 debugp_param(
"integer", lit);
11209 ADD_INSN1(ret, node, putobject, lit);
11215 VALUE lit = rb_node_float_literal_val(node);
11217 debugp_param(
"float", lit);
11219 ADD_INSN1(ret, node, putobject, lit);
11224 case NODE_RATIONAL:{
11225 VALUE lit = rb_node_rational_literal_val(node);
11227 debugp_param(
"rational", lit);
11229 ADD_INSN1(ret, node, putobject, lit);
11234 case NODE_IMAGINARY:{
11235 VALUE lit = rb_node_imaginary_literal_val(node);
11237 debugp_param(
"imaginary", lit);
11239 ADD_INSN1(ret, node, putobject, lit);
11246 debugp_param(
"nd_lit", get_string_value(node));
11248 VALUE lit = get_string_value(node);
11251 option->frozen_string_literal != ISEQ_FROZEN_STRING_LITERAL_DISABLED) {
11252 lit = rb_str_with_debug_created_info(lit, rb_iseq_path(iseq), line);
11253 RB_OBJ_SET_SHAREABLE(lit);
11255 switch (option->frozen_string_literal) {
11256 case ISEQ_FROZEN_STRING_LITERAL_UNSET:
11257 ADD_INSN1(ret, node, putchilledstring, lit);
11259 case ISEQ_FROZEN_STRING_LITERAL_DISABLED:
11260 ADD_INSN1(ret, node, putstring, lit);
11262 case ISEQ_FROZEN_STRING_LITERAL_ENABLED:
11263 ADD_INSN1(ret, node, putobject, lit);
11266 rb_bug(
"invalid frozen_string_literal");
11273 compile_dstr(iseq, ret, node);
11276 ADD_INSN(ret, node, pop);
11281 ADD_CALL_RECEIVER(ret, node);
11282 VALUE str = rb_node_str_string_val(node);
11283 ADD_INSN1(ret, node, putobject, str);
11285 ADD_CALL(ret, node, idBackquote,
INT2FIX(1));
11288 ADD_INSN(ret, node, pop);
11293 ADD_CALL_RECEIVER(ret, node);
11294 compile_dstr(iseq, ret, node);
11295 ADD_CALL(ret, node, idBackquote,
INT2FIX(1));
11298 ADD_INSN(ret, node, pop);
11303 CHECK(compile_evstr(iseq, ret, RNODE_EVSTR(node)->nd_body, popped));
11307 VALUE lit = rb_node_regx_string_val(node);
11308 RB_OBJ_SET_SHAREABLE(lit);
11309 ADD_INSN1(ret, node, putobject, lit);
11315 compile_dregx(iseq, ret, node, popped);
11318 int ic_index = body->ise_size++;
11320 block_iseq = NEW_CHILD_ISEQ(RNODE_ONCE(node)->nd_body, make_name_for_block(iseq), ISEQ_TYPE_PLAIN, line);
11322 ADD_INSN2(ret, node, once, block_iseq,
INT2FIX(ic_index));
11326 ADD_INSN(ret, node, pop);
11330 case NODE_ARGSCAT:{
11332 CHECK(COMPILE(ret,
"argscat head", RNODE_ARGSCAT(node)->nd_head));
11333 ADD_INSN1(ret, node, splatarray,
Qfalse);
11334 ADD_INSN(ret, node, pop);
11335 CHECK(COMPILE(ret,
"argscat body", RNODE_ARGSCAT(node)->nd_body));
11336 ADD_INSN1(ret, node, splatarray,
Qfalse);
11337 ADD_INSN(ret, node, pop);
11340 CHECK(COMPILE(ret,
"argscat head", RNODE_ARGSCAT(node)->nd_head));
11341 const NODE *body_node = RNODE_ARGSCAT(node)->nd_body;
11342 if (nd_type_p(body_node, NODE_LIST)) {
11343 CHECK(compile_array(iseq, ret, body_node, popped, FALSE) >= 0);
11346 CHECK(COMPILE(ret,
"argscat body", body_node));
11347 ADD_INSN(ret, node, concattoarray);
11352 case NODE_ARGSPUSH:{
11354 CHECK(COMPILE(ret,
"argspush head", RNODE_ARGSPUSH(node)->nd_head));
11355 ADD_INSN1(ret, node, splatarray,
Qfalse);
11356 ADD_INSN(ret, node, pop);
11357 CHECK(COMPILE_(ret,
"argspush body", RNODE_ARGSPUSH(node)->nd_body, popped));
11360 CHECK(COMPILE(ret,
"argspush head", RNODE_ARGSPUSH(node)->nd_head));
11361 const NODE *body_node = RNODE_ARGSPUSH(node)->nd_body;
11362 if (keyword_node_p(body_node)) {
11363 CHECK(COMPILE_(ret,
"array element", body_node, FALSE));
11364 ADD_INSN(ret, node, pushtoarraykwsplat);
11366 else if (static_literal_node_p(body_node, iseq,
false)) {
11367 ADD_INSN1(ret, body_node, putobject, static_literal_value(body_node, iseq));
11368 ADD_INSN1(ret, node, pushtoarray,
INT2FIX(1));
11371 CHECK(COMPILE_(ret,
"array element", body_node, FALSE));
11372 ADD_INSN1(ret, node, pushtoarray,
INT2FIX(1));
11378 CHECK(COMPILE(ret,
"splat", RNODE_SPLAT(node)->nd_head));
11379 ADD_INSN1(ret, node, splatarray,
Qtrue);
11382 ADD_INSN(ret, node, pop);
11387 ID mid = RNODE_DEFN(node)->nd_mid;
11388 const rb_iseq_t *method_iseq = NEW_ISEQ(RNODE_DEFN(node)->nd_defn,
11390 ISEQ_TYPE_METHOD, line);
11392 debugp_param(
"defn/iseq", rb_iseqw_new(method_iseq));
11393 ADD_INSN2(ret, node, definemethod,
ID2SYM(mid), method_iseq);
11397 ADD_INSN1(ret, node, putobject,
ID2SYM(mid));
11403 ID mid = RNODE_DEFS(node)->nd_mid;
11404 const rb_iseq_t * singleton_method_iseq = NEW_ISEQ(RNODE_DEFS(node)->nd_defn,
11406 ISEQ_TYPE_METHOD, line);
11408 debugp_param(
"defs/iseq", rb_iseqw_new(singleton_method_iseq));
11409 CHECK(COMPILE(ret,
"defs: recv", RNODE_DEFS(node)->nd_recv));
11410 ADD_INSN2(ret, node, definesmethod,
ID2SYM(mid), singleton_method_iseq);
11414 ADD_INSN1(ret, node, putobject,
ID2SYM(mid));
11419 ADD_INSN1(ret, node, putspecialobject,
INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
11420 ADD_INSN1(ret, node, putspecialobject,
INT2FIX(VM_SPECIAL_OBJECT_CBASE));
11421 CHECK(COMPILE(ret,
"alias arg1", RNODE_ALIAS(node)->nd_1st));
11422 CHECK(COMPILE(ret,
"alias arg2", RNODE_ALIAS(node)->nd_2nd));
11423 ADD_SEND(ret, node, id_core_set_method_alias,
INT2FIX(3));
11426 ADD_INSN(ret, node, pop);
11431 ADD_INSN1(ret, node, putspecialobject,
INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
11432 ADD_INSN1(ret, node, putobject,
ID2SYM(RNODE_VALIAS(node)->nd_alias));
11433 ADD_INSN1(ret, node, putobject,
ID2SYM(RNODE_VALIAS(node)->nd_orig));
11434 ADD_SEND(ret, node, id_core_set_variable_alias,
INT2FIX(2));
11437 ADD_INSN(ret, node, pop);
11444 for (
long i = 0; i < ary->len; i++) {
11445 ADD_INSN1(ret, node, putspecialobject,
INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
11446 ADD_INSN1(ret, node, putspecialobject,
INT2FIX(VM_SPECIAL_OBJECT_CBASE));
11447 CHECK(COMPILE(ret,
"undef arg", ary->data[i]));
11448 ADD_SEND(ret, node, id_core_undef_method,
INT2FIX(2));
11450 if (i < ary->
len - 1) {
11451 ADD_INSN(ret, node, pop);
11456 ADD_INSN(ret, node, pop);
11461 const rb_iseq_t *class_iseq = NEW_CHILD_ISEQ(RNODE_CLASS(node)->nd_body,
11462 rb_str_freeze(rb_sprintf(
"<class:%"PRIsVALUE
">", rb_id2str(get_node_colon_nd_mid(RNODE_CLASS(node)->nd_cpath)))),
11463 ISEQ_TYPE_CLASS, line);
11464 const int flags = VM_DEFINECLASS_TYPE_CLASS |
11465 (RNODE_CLASS(node)->nd_super ? VM_DEFINECLASS_FLAG_HAS_SUPERCLASS : 0) |
11466 compile_cpath(ret, iseq, RNODE_CLASS(node)->nd_cpath);
11468 CHECK(COMPILE(ret,
"super", RNODE_CLASS(node)->nd_super));
11469 ADD_INSN3(ret, node, defineclass,
ID2SYM(get_node_colon_nd_mid(RNODE_CLASS(node)->nd_cpath)), class_iseq,
INT2FIX(flags));
11473 ADD_INSN(ret, node, pop);
11478 const rb_iseq_t *module_iseq = NEW_CHILD_ISEQ(RNODE_MODULE(node)->nd_body,
11479 rb_str_freeze(rb_sprintf(
"<module:%"PRIsVALUE
">", rb_id2str(get_node_colon_nd_mid(RNODE_MODULE(node)->nd_cpath)))),
11480 ISEQ_TYPE_CLASS, line);
11481 const int flags = VM_DEFINECLASS_TYPE_MODULE |
11482 compile_cpath(ret, iseq, RNODE_MODULE(node)->nd_cpath);
11484 ADD_INSN (ret, node, putnil);
11485 ADD_INSN3(ret, node, defineclass,
ID2SYM(get_node_colon_nd_mid(RNODE_MODULE(node)->nd_cpath)), module_iseq,
INT2FIX(flags));
11489 ADD_INSN(ret, node, pop);
11495 const rb_iseq_t *singleton_class = NEW_ISEQ(RNODE_SCLASS(node)->nd_body, rb_fstring_lit(
"singleton class"),
11496 ISEQ_TYPE_CLASS, line);
11498 CHECK(COMPILE(ret,
"sclass#recv", RNODE_SCLASS(node)->nd_recv));
11499 ADD_INSN (ret, node, putnil);
11500 CONST_ID(singletonclass,
"singletonclass");
11501 ADD_INSN3(ret, node, defineclass,
11502 ID2SYM(singletonclass), singleton_class,
11503 INT2FIX(VM_DEFINECLASS_TYPE_SINGLETON_CLASS));
11507 ADD_INSN(ret, node, pop);
11512 CHECK(compile_colon2(iseq, ret, node, popped));
11515 CHECK(compile_colon3(iseq, ret, node, popped));
11518 CHECK(compile_dots(iseq, ret, node, popped, FALSE));
11521 CHECK(compile_dots(iseq, ret, node, popped, TRUE));
11525 LABEL *lend = NEW_LABEL(line);
11526 LABEL *ltrue = NEW_LABEL(line);
11527 LABEL *lfalse = NEW_LABEL(line);
11528 CHECK(compile_flip_flop(iseq, ret, node,
type == NODE_FLIP2,
11530 ADD_LABEL(ret, ltrue);
11531 ADD_INSN1(ret, node, putobject,
Qtrue);
11532 ADD_INSNL(ret, node, jump, lend);
11533 ADD_LABEL(ret, lfalse);
11534 ADD_INSN1(ret, node, putobject,
Qfalse);
11535 ADD_LABEL(ret, lend);
11540 ADD_INSN(ret, node, putself);
11546 ADD_INSN(ret, node, putnil);
11552 ADD_INSN1(ret, node, putobject,
Qtrue);
11558 ADD_INSN1(ret, node, putobject,
Qfalse);
11563 CHECK(compile_errinfo(iseq, ret, node, popped));
11567 CHECK(compile_defined_expr(iseq, ret, node,
Qtrue,
false));
11570 case NODE_POSTEXE:{
11574 int is_index = body->ise_size++;
11576 rb_iseq_new_with_callback_new_callback(build_postexe_iseq, RNODE_POSTEXE(node)->nd_body);
11578 NEW_CHILD_ISEQ_WITH_CALLBACK(ifunc, rb_fstring(make_name_for_block(iseq)), ISEQ_TYPE_BLOCK, line);
11580 ADD_INSN2(ret, node, once, once_iseq,
INT2FIX(is_index));
11584 ADD_INSN(ret, node, pop);
11589 CHECK(compile_kw_arg(iseq, ret, node, popped));
11592 compile_dstr(iseq, ret, node);
11594 ADD_INSN(ret, node, intern);
11597 ADD_INSN(ret, node, pop);
11601 case NODE_ATTRASGN:
11602 CHECK(compile_attrasgn(iseq, ret, node, popped));
11606 const rb_iseq_t *block = NEW_CHILD_ISEQ(RNODE_LAMBDA(node)->nd_body, make_name_for_block(iseq), ISEQ_TYPE_BLOCK, line);
11609 ADD_INSN1(ret, node, putspecialobject,
INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
11610 ADD_CALL_WITH_BLOCK(ret, node, idLambda, argc, block);
11614 ADD_INSN(ret, node, pop);
11619 UNKNOWN_NODE(
"iseq_compile_each", node, COMPILE_NG);
11634insn_data_length(
INSN *iobj)
11636 return insn_len(iobj->insn_id);
11640calc_sp_depth(
int depth,
INSN *insn)
11642 return comptime_insn_stack_increase(depth, insn->insn_id, insn->operands);
11646opobj_inspect(
VALUE obj)
11666insn_data_to_s_detail(
INSN *iobj)
11668 VALUE str = rb_sprintf(
"%-20s ", insn_name(iobj->insn_id));
11670 if (iobj->operands) {
11671 const char *types = insn_op_types(iobj->insn_id);
11674 for (j = 0; types[j]; j++) {
11675 char type = types[j];
11681 rb_str_catf(str, LABEL_FORMAT, lobj->label_no);
11699 VALUE v = OPERAND_AT(iobj, j);
11714 rb_str_catf(str,
"<ivc:%d>",
FIX2INT(OPERAND_AT(iobj, j)));
11717 rb_str_catf(str,
"<icvarc:%d>",
FIX2INT(OPERAND_AT(iobj, j)));
11720 rb_str_catf(str,
"<ise:%d>",
FIX2INT(OPERAND_AT(iobj, j)));
11726 if (vm_ci_mid(ci)) rb_str_catf(str,
"%"PRIsVALUE, rb_id2str(vm_ci_mid(ci)));
11727 rb_str_catf(str,
", %d>", vm_ci_argc(ci));
11735 void *func = (
void *)OPERAND_AT(iobj, j);
11738 if (dladdr(func, &info) && info.dli_sname) {
11743 rb_str_catf(str,
"<%p>", func);
11753 if (types[j + 1]) {
11764 dump_disasm_list_with_cursor(link, NULL, NULL);
11775 printf(
"-- raw disasm--------\n");
11778 if (curr) printf(curr == link ?
"*" :
" ");
11779 switch (link->type) {
11780 case ISEQ_ELEMENT_INSN:
11782 iobj = (
INSN *)link;
11783 str = insn_data_to_s_detail(iobj);
11784 printf(
" %04d %-65s(%4u)\n", pos,
StringValueCStr(str), iobj->insn_info.line_no);
11785 pos += insn_data_length(iobj);
11788 case ISEQ_ELEMENT_LABEL:
11790 lobj = (
LABEL *)link;
11791 printf(LABEL_FORMAT
" [sp: %d, unremovable: %d, refcnt: %d]%s\n", lobj->label_no, lobj->sp, lobj->unremovable, lobj->refcnt,
11792 dest == lobj ?
" <---" :
"");
11795 case ISEQ_ELEMENT_TRACE:
11798 printf(
" trace: %0x\n", trace->event);
11801 case ISEQ_ELEMENT_ADJUST:
11804 printf(
" adjust: [label: %d]\n", adjust->label ? adjust->label->label_no : -1);
11809 rb_raise(
rb_eSyntaxError,
"dump_disasm_list error: %d\n", (
int)link->type);
11813 printf(
"---------------------\n");
11818rb_insn_len(
VALUE insn)
11820 return insn_len(insn);
11824rb_insns_name(
int i)
11826 return insn_name(i);
11830rb_insns_name_array(
void)
11834 for (i = 0; i < VM_INSTRUCTION_SIZE; i++) {
11845 obj = rb_to_symbol_type(obj);
11847 if (st_lookup(labels_table, obj, &tmp) == 0) {
11848 label = NEW_LABEL(0);
11849 st_insert(labels_table, obj, (st_data_t)label);
11852 label = (
LABEL *)tmp;
11859get_exception_sym2type(
VALUE sym)
11861 static VALUE symRescue, symEnsure, symRetry;
11862 static VALUE symBreak, symRedo, symNext;
11864 if (symRescue == 0) {
11873 if (sym == symRescue)
return CATCH_TYPE_RESCUE;
11874 if (sym == symEnsure)
return CATCH_TYPE_ENSURE;
11875 if (sym == symRetry)
return CATCH_TYPE_RETRY;
11876 if (sym == symBreak)
return CATCH_TYPE_BREAK;
11877 if (sym == symRedo)
return CATCH_TYPE_REDO;
11878 if (sym == symNext)
return CATCH_TYPE_NEXT;
11879 rb_raise(
rb_eSyntaxError,
"invalid exception symbol: %+"PRIsVALUE, sym);
11892 LABEL *lstart, *lend, *lcont;
11907 lstart = register_label(iseq, labels_table,
RARRAY_AREF(v, 2));
11908 lend = register_label(iseq, labels_table,
RARRAY_AREF(v, 3));
11909 lcont = register_label(iseq, labels_table,
RARRAY_AREF(v, 4));
11913 if (
type == CATCH_TYPE_RESCUE ||
11914 type == CATCH_TYPE_BREAK ||
11915 type == CATCH_TYPE_NEXT) {
11921 ADD_CATCH_ENTRY(
type, lstart, lend, eiseq, lcont);
11929insn_make_insn_table(
void)
11933 table = st_init_numtable_with_size(VM_INSTRUCTION_SIZE);
11935 for (i=0; i<VM_INSTRUCTION_SIZE; i++) {
11949 iseqw = rb_iseq_load(op, (
VALUE)iseq,
Qnil);
11951 else if (
CLASS_OF(op) == rb_cISeq) {
11958 loaded_iseq = rb_iseqw_to_iseq(iseqw);
11959 return loaded_iseq;
11967 unsigned int flag = 0;
11978 if (!
NIL_P(vorig_argc)) orig_argc =
FIX2INT(vorig_argc);
11980 if (!
NIL_P(vkw_arg)) {
11983 size_t n = rb_callinfo_kwarg_bytes(
len);
11986 kw_arg->references = 0;
11987 kw_arg->keyword_len =
len;
11988 for (i = 0; i <
len; i++) {
11991 kw_arg->keywords[i] = kw;
11996 const struct rb_callinfo *ci = new_callinfo(iseq, mid, orig_argc, flag, kw_arg, (flag & VM_CALL_ARGS_SIMPLE) == 0);
12002event_name_to_flag(
VALUE sym)
12004#define CHECK_EVENT(ev) if (sym == ID2SYM(rb_intern_const(#ev))) return ev;
12025 int line_no = 0, node_id = -1, insn_idx = 0;
12026 int ret = COMPILE_OK;
12031 static struct st_table *insn_table;
12033 if (insn_table == 0) {
12034 insn_table = insn_make_insn_table();
12037 for (i=0; i<
len; i++) {
12043 ADD_TRACE(anchor, event);
12046 LABEL *label = register_label(iseq, labels_table, obj);
12047 ADD_LABEL(anchor, label);
12064 if (st_lookup(insn_table, (st_data_t)insn, &insn_id) == 0) {
12066 COMPILE_ERROR(iseq, line_no,
12067 "unknown instruction: %+"PRIsVALUE, insn);
12072 if (argc != insn_len((
VALUE)insn_id)-1) {
12073 COMPILE_ERROR(iseq, line_no,
12074 "operand size mismatch");
12080 argv = compile_data_calloc2(iseq,
sizeof(
VALUE), argc);
12085 (
enum ruby_vminsn_type)insn_id, argc, argv));
12087 for (j=0; j<argc; j++) {
12089 switch (insn_op_type((
VALUE)insn_id, j)) {
12091 LABEL *label = register_label(iseq, labels_table, op);
12092 argv[j] = (
VALUE)label;
12107 VALUE v = (
VALUE)iseq_build_load_iseq(iseq, op);
12118 if (
NUM2UINT(op) >= ISEQ_BODY(iseq)->ise_size) {
12119 ISEQ_BODY(iseq)->ise_size =
NUM2INT(op) + 1;
12125 op = rb_to_array_type(op);
12129 sym = rb_to_symbol_type(sym);
12134 argv[j] = segments;
12136 ISEQ_BODY(iseq)->ic_size++;
12141 if (
NUM2UINT(op) >= ISEQ_BODY(iseq)->ivc_size) {
12142 ISEQ_BODY(iseq)->ivc_size =
NUM2INT(op) + 1;
12147 if (
NUM2UINT(op) >= ISEQ_BODY(iseq)->icvarc_size) {
12148 ISEQ_BODY(iseq)->icvarc_size =
NUM2INT(op) + 1;
12152 argv[j] = iseq_build_callinfo_from_hash(iseq, op);
12155 argv[j] = rb_to_symbol_type(op);
12162 RHASH_TBL_RAW(map)->type = &cdhash_type;
12163 op = rb_to_array_type(op);
12168 register_label(iseq, labels_table, sym);
12169 rb_hash_aset(map, key, (
VALUE)label | 1);
12179#if SIZEOF_VALUE <= SIZEOF_LONG
12184 argv[j] = (
VALUE)funcptr;
12195 (
enum ruby_vminsn_type)insn_id, argc, NULL));
12199 rb_raise(
rb_eTypeError,
"unexpected object for instruction");
12204 validate_labels(iseq, labels_table);
12205 if (!ret)
return ret;
12206 return iseq_setup(iseq, anchor);
12209#define CHECK_ARRAY(v) rb_to_array_type(v)
12210#define CHECK_SYMBOL(v) rb_to_symbol_type(v)
12215 VALUE val = rb_hash_aref(param, sym);
12220 else if (!
NIL_P(val)) {
12221 rb_raise(
rb_eTypeError,
"invalid %+"PRIsVALUE
" Fixnum: %+"PRIsVALUE,
12227static const struct rb_iseq_param_keyword *
12233 VALUE key, sym, default_val;
12236 struct rb_iseq_param_keyword *keyword =
ZALLOC(
struct rb_iseq_param_keyword);
12238 ISEQ_BODY(iseq)->param.flags.has_kw = TRUE;
12240 keyword->num =
len;
12241#define SYM(s) ID2SYM(rb_intern_const(#s))
12242 (void)int_param(&keyword->bits_start, params, SYM(kwbits));
12243 i = keyword->bits_start - keyword->num;
12244 ids = (
ID *)&ISEQ_BODY(iseq)->local_table[i];
12248 for (i = 0; i <
len; i++) {
12252 goto default_values;
12255 keyword->required_num++;
12259 default_len =
len - i;
12260 if (default_len == 0) {
12261 keyword->table = ids;
12264 else if (default_len < 0) {
12270 for (j = 0; i <
len; i++, j++) {
12284 rb_raise(
rb_eTypeError,
"keyword default has unsupported len %+"PRIsVALUE, key);
12290 keyword->table = ids;
12291 keyword->default_values = dvs;
12297iseq_insn_each_object_mark_and_move(
VALUE * obj,
VALUE _)
12299 rb_gc_mark_and_move(obj);
12306 size_t size =
sizeof(
INSN);
12307 unsigned int pos = 0;
12310#ifdef STRICT_ALIGNMENT
12311 size_t padding = calc_padding((
void *)&storage->buff[pos], size);
12313 const size_t padding = 0;
12315 size_t offset = pos + size + padding;
12316 if (offset > storage->size || offset > storage->pos) {
12318 storage = storage->next;
12321#ifdef STRICT_ALIGNMENT
12322 pos += (int)padding;
12325 iobj = (
INSN *)&storage->buff[pos];
12327 if (iobj->operands) {
12328 iseq_insn_each_markable_object(iobj, iseq_insn_each_object_mark_and_move, (
VALUE)0);
12348#define SYM(s) ID2SYM(rb_intern_const(#s))
12350 unsigned int arg_size, local_size, stack_max;
12352 struct st_table *labels_table = st_init_numtable();
12354 VALUE arg_opt_labels = rb_hash_aref(params, SYM(opt));
12355 VALUE keywords = rb_hash_aref(params, SYM(keyword));
12357 DECL_ANCHOR(anchor);
12358 INIT_ANCHOR(anchor);
12361 ISEQ_BODY(iseq)->local_table_size =
len;
12362 ISEQ_BODY(iseq)->local_table = tbl =
len > 0 ? (
ID *)
ALLOC_N(
ID, ISEQ_BODY(iseq)->local_table_size) : NULL;
12364 for (i = 0; i <
len; i++) {
12367 if (sym_arg_rest == lv) {
12375#define INT_PARAM(F) int_param(&ISEQ_BODY(iseq)->param.F, params, SYM(F))
12376 if (INT_PARAM(lead_num)) {
12377 ISEQ_BODY(iseq)->param.flags.has_lead = TRUE;
12379 if (INT_PARAM(post_num)) ISEQ_BODY(iseq)->param.flags.has_post = TRUE;
12380 if (INT_PARAM(post_start)) ISEQ_BODY(iseq)->param.flags.has_post = TRUE;
12381 if (INT_PARAM(rest_start)) ISEQ_BODY(iseq)->param.flags.has_rest = TRUE;
12382 if (INT_PARAM(block_start)) ISEQ_BODY(iseq)->param.flags.has_block = TRUE;
12385#define INT_PARAM(F) F = (int_param(&x, misc, SYM(F)) ? (unsigned int)x : 0)
12387 INT_PARAM(arg_size);
12388 INT_PARAM(local_size);
12389 INT_PARAM(stack_max);
12394#ifdef USE_ISEQ_NODE_ID
12395 node_ids = rb_hash_aref(misc,
ID2SYM(rb_intern(
"node_ids")));
12403 ISEQ_BODY(iseq)->param.flags.has_opt = !!(
len - 1 >= 0);
12405 if (ISEQ_BODY(iseq)->param.flags.has_opt) {
12408 for (i = 0; i <
len; i++) {
12410 LABEL *label = register_label(iseq, labels_table, ent);
12411 opt_table[i] = (
VALUE)label;
12414 ISEQ_BODY(iseq)->param.opt_num =
len - 1;
12415 ISEQ_BODY(iseq)->param.opt_table = opt_table;
12418 else if (!
NIL_P(arg_opt_labels)) {
12419 rb_raise(
rb_eTypeError,
":opt param is not an array: %+"PRIsVALUE,
12424 ISEQ_BODY(iseq)->param.keyword = iseq_build_kw(iseq, params, keywords);
12426 else if (!
NIL_P(keywords)) {
12427 rb_raise(
rb_eTypeError,
":keywords param is not an array: %+"PRIsVALUE,
12431 if (
Qtrue == rb_hash_aref(params, SYM(ambiguous_param0))) {
12432 ISEQ_BODY(iseq)->param.flags.ambiguous_param0 = TRUE;
12435 if (
Qtrue == rb_hash_aref(params, SYM(use_block))) {
12436 ISEQ_BODY(iseq)->param.flags.use_block = TRUE;
12439 if (int_param(&i, params, SYM(kwrest))) {
12440 struct rb_iseq_param_keyword *keyword = (
struct rb_iseq_param_keyword *)ISEQ_BODY(iseq)->param.keyword;
12441 if (keyword == NULL) {
12442 ISEQ_BODY(iseq)->param.keyword = keyword =
ZALLOC(
struct rb_iseq_param_keyword);
12444 keyword->rest_start = i;
12445 ISEQ_BODY(iseq)->param.flags.has_kwrest = TRUE;
12448 iseq_calc_param_size(iseq);
12451 iseq_build_from_ary_exception(iseq, labels_table, exception);
12454 iseq_build_from_ary_body(iseq, anchor, body, node_ids, labels_wrapper);
12456 ISEQ_BODY(iseq)->param.size = arg_size;
12457 ISEQ_BODY(iseq)->local_table_size = local_size;
12458 ISEQ_BODY(iseq)->stack_max = stack_max;
12468 while (body->type == ISEQ_TYPE_BLOCK ||
12469 body->type == ISEQ_TYPE_RESCUE ||
12470 body->type == ISEQ_TYPE_ENSURE ||
12471 body->type == ISEQ_TYPE_EVAL ||
12472 body->type == ISEQ_TYPE_MAIN
12476 for (i = 0; i < body->local_table_size; i++) {
12477 if (body->local_table[i] ==
id) {
12481 iseq = body->parent_iseq;
12482 body = ISEQ_BODY(iseq);
12495 for (i=0; i<body->local_table_size; i++) {
12496 if (body->local_table[i] ==
id) {
12506#ifndef IBF_ISEQ_DEBUG
12507#define IBF_ISEQ_DEBUG 0
12510#ifndef IBF_ISEQ_ENABLE_LOCAL_BUFFER
12511#define IBF_ISEQ_ENABLE_LOCAL_BUFFER 0
12514typedef uint32_t ibf_offset_t;
12515#define IBF_OFFSET(ptr) ((ibf_offset_t)(VALUE)(ptr))
12517#define IBF_MAJOR_VERSION ISEQ_MAJOR_VERSION
12519#define IBF_DEVEL_VERSION 5
12520#define IBF_MINOR_VERSION (ISEQ_MINOR_VERSION * 10000 + IBF_DEVEL_VERSION)
12522#define IBF_MINOR_VERSION ISEQ_MINOR_VERSION
12525static const char IBF_ENDIAN_MARK =
12526#ifdef WORDS_BIGENDIAN
12535 uint32_t major_version;
12536 uint32_t minor_version;
12538 uint32_t extra_size;
12540 uint32_t iseq_list_size;
12541 uint32_t global_object_list_size;
12542 ibf_offset_t iseq_list_offset;
12543 ibf_offset_t global_object_list_offset;
12564 unsigned int obj_list_size;
12565 ibf_offset_t obj_list_offset;
12584pinned_list_mark(
void *ptr)
12588 for (i = 0; i < list->size; i++) {
12589 if (list->buffer[i]) {
12590 rb_gc_mark(list->buffer[i]);
12606pinned_list_fetch(
VALUE list,
long offset)
12612 if (offset >= ptr->size) {
12613 rb_raise(
rb_eIndexError,
"object index out of range: %ld", offset);
12616 return ptr->buffer[offset];
12620pinned_list_store(
VALUE list,
long offset,
VALUE object)
12626 if (offset >= ptr->size) {
12627 rb_raise(
rb_eIndexError,
"object index out of range: %ld", offset);
12634pinned_list_new(
long size)
12636 size_t memsize = offsetof(
struct pinned_list, buffer) + size *
sizeof(
VALUE);
12637 VALUE obj_list = rb_data_typed_object_zalloc(0, memsize, &pinned_list_type);
12638 struct pinned_list * ptr = RTYPEDDATA_GET_DATA(obj_list);
12644ibf_dump_pos(
struct ibf_dump *dump)
12646 long pos = RSTRING_LEN(dump->current_buffer->str);
12647#if SIZEOF_LONG > SIZEOF_INT
12648 if (pos >= UINT_MAX) {
12652 return (
unsigned int)pos;
12656ibf_dump_align(
struct ibf_dump *dump,
size_t align)
12658 ibf_offset_t pos = ibf_dump_pos(dump);
12660 static const char padding[
sizeof(
VALUE)];
12661 size_t size = align - ((size_t)pos % align);
12662#if SIZEOF_LONG > SIZEOF_INT
12663 if (pos + size >= UINT_MAX) {
12667 for (; size >
sizeof(padding); size -=
sizeof(padding)) {
12668 rb_str_cat(dump->current_buffer->str, padding,
sizeof(padding));
12670 rb_str_cat(dump->current_buffer->str, padding, size);
12675ibf_dump_write(
struct ibf_dump *dump,
const void *buff,
unsigned long size)
12677 ibf_offset_t pos = ibf_dump_pos(dump);
12678#if SIZEOF_LONG > SIZEOF_INT
12680 if (size >= UINT_MAX || pos + size >= UINT_MAX) {
12684 rb_str_cat(dump->current_buffer->str, (
const char *)buff, size);
12689ibf_dump_write_byte(
struct ibf_dump *dump,
unsigned char byte)
12691 return ibf_dump_write(dump, &
byte,
sizeof(
unsigned char));
12695ibf_dump_overwrite(
struct ibf_dump *dump,
void *buff,
unsigned int size,
long offset)
12697 VALUE str = dump->current_buffer->str;
12698 char *ptr = RSTRING_PTR(str);
12699 if ((
unsigned long)(size + offset) > (
unsigned long)RSTRING_LEN(str))
12700 rb_bug(
"ibf_dump_overwrite: overflow");
12701 memcpy(ptr + offset, buff, size);
12705ibf_load_ptr(
const struct ibf_load *load, ibf_offset_t *offset,
int size)
12707 ibf_offset_t beg = *offset;
12709 return load->current_buffer->buff + beg;
12713ibf_load_alloc(
const struct ibf_load *load, ibf_offset_t offset,
size_t x,
size_t y)
12715 void *buff = ruby_xmalloc2(x, y);
12716 size_t size = x * y;
12717 memcpy(buff, load->current_buffer->buff + offset, size);
12721#define IBF_W_ALIGN(type) (RUBY_ALIGNOF(type) > 1 ? ibf_dump_align(dump, RUBY_ALIGNOF(type)) : (void)0)
12723#define IBF_W(b, type, n) (IBF_W_ALIGN(type), (type *)(VALUE)IBF_WP(b, type, n))
12724#define IBF_WV(variable) ibf_dump_write(dump, &(variable), sizeof(variable))
12725#define IBF_WP(b, type, n) ibf_dump_write(dump, (b), sizeof(type) * (n))
12726#define IBF_R(val, type, n) (type *)ibf_load_alloc(load, IBF_OFFSET(val), sizeof(type), (n))
12727#define IBF_ZERO(variable) memset(&(variable), 0, sizeof(variable))
12730ibf_table_lookup(
struct st_table *table, st_data_t key)
12734 if (st_lookup(table, key, &val)) {
12743ibf_table_find_or_insert(
struct st_table *table, st_data_t key)
12745 int index = ibf_table_lookup(table, key);
12748 index = (int)table->num_entries;
12749 st_insert(table, key, (st_data_t)index);
12757static void ibf_dump_object_list(
struct ibf_dump *dump, ibf_offset_t *obj_list_offset,
unsigned int *obj_list_size);
12763ibf_dump_object_table_new(
void)
12765 st_table *obj_table = st_init_numtable();
12766 st_insert(obj_table, (st_data_t)
Qnil, (st_data_t)0);
12774 return ibf_table_find_or_insert(dump->current_buffer->obj_table, (st_data_t)obj);
12780 if (
id == 0 || rb_id2name(
id) == NULL) {
12783 return ibf_dump_object(dump,
rb_id2sym(
id));
12787ibf_load_id(
const struct ibf_load *load,
const ID id_index)
12789 if (id_index == 0) {
12792 VALUE sym = ibf_load_object(load, id_index);
12802static ibf_offset_t ibf_dump_iseq_each(
struct ibf_dump *dump,
const rb_iseq_t *iseq);
12807 if (iseq == NULL) {
12811 return ibf_table_find_or_insert(dump->iseq_table, (st_data_t)iseq);
12815static unsigned char
12816ibf_load_byte(
const struct ibf_load *load, ibf_offset_t *offset)
12818 if (*offset >= load->current_buffer->size) { rb_raise(
rb_eRuntimeError,
"invalid bytecode"); }
12819 return (
unsigned char)load->current_buffer->buff[(*offset)++];
12835 if (
sizeof(
VALUE) > 8 || CHAR_BIT != 8) {
12836 ibf_dump_write(dump, &x,
sizeof(
VALUE));
12840 enum { max_byte_length =
sizeof(
VALUE) + 1 };
12842 unsigned char bytes[max_byte_length];
12845 for (n = 0; n <
sizeof(
VALUE) && (x >> (7 - n)); n++, x >>= 8) {
12846 bytes[max_byte_length - 1 - n] = (
unsigned char)x;
12852 bytes[max_byte_length - 1 - n] = (
unsigned char)x;
12855 ibf_dump_write(dump, bytes + max_byte_length - n, n);
12859ibf_load_small_value(
const struct ibf_load *load, ibf_offset_t *offset)
12861 if (
sizeof(
VALUE) > 8 || CHAR_BIT != 8) {
12862 union {
char s[
sizeof(
VALUE)];
VALUE v; } x;
12864 memcpy(x.s, load->current_buffer->buff + *offset,
sizeof(
VALUE));
12865 *offset +=
sizeof(
VALUE);
12870 enum { max_byte_length =
sizeof(
VALUE) + 1 };
12872 const unsigned char *buffer = (
const unsigned char *)load->current_buffer->buff;
12873 const unsigned char c = buffer[*offset];
12877 c == 0 ? 9 : ntz_int32(c) + 1;
12880 if (*offset + n > load->current_buffer->size) {
12885 for (i = 1; i < n; i++) {
12887 x |= (
VALUE)buffer[*offset + i];
12901 ibf_dump_write_small_value(dump, (
VALUE)bf->index);
12903 size_t len = strlen(bf->name);
12904 ibf_dump_write_small_value(dump, (
VALUE)
len);
12905 ibf_dump_write(dump, bf->name,
len);
12909ibf_load_builtin(
const struct ibf_load *load, ibf_offset_t *offset)
12911 int i = (int)ibf_load_small_value(load, offset);
12912 int len = (int)ibf_load_small_value(load, offset);
12913 const char *name = (
char *)ibf_load_ptr(load, offset,
len);
12916 fprintf(stderr,
"%.*s!!\n",
len, name);
12920 if (table == NULL) rb_raise(rb_eArgError,
"builtin function table is not provided");
12921 if (strncmp(table[i].name, name,
len) != 0) {
12922 rb_raise(rb_eArgError,
"builtin function index (%d) mismatch (expect %s but %s)", i, name, table[i].name);
12933 const int iseq_size = body->iseq_size;
12935 const VALUE *orig_code = rb_iseq_original_iseq(iseq);
12937 ibf_offset_t offset = ibf_dump_pos(dump);
12939 for (code_index=0; code_index<iseq_size;) {
12940 const VALUE insn = orig_code[code_index++];
12941 const char *types = insn_op_types(insn);
12946 ibf_dump_write_small_value(dump, insn);
12949 for (op_index=0; types[op_index]; op_index++, code_index++) {
12950 VALUE op = orig_code[code_index];
12953 switch (types[op_index]) {
12956 wv = ibf_dump_object(dump, op);
12965 wv = ibf_dump_object(dump, arr);
12973 wv = is - ISEQ_IS_ENTRY_START(body, types[op_index]);
12981 wv = ibf_dump_id(dump, (
ID)op);
12993 ibf_dump_write_small_value(dump, wv);
13003ibf_load_code(
const struct ibf_load *load,
rb_iseq_t *iseq, ibf_offset_t bytecode_offset, ibf_offset_t bytecode_size,
unsigned int iseq_size)
13006 unsigned int code_index;
13007 ibf_offset_t reading_pos = bytecode_offset;
13011 struct rb_call_data *cd_entries = load_body->call_data;
13014 load_body->iseq_encoded = code;
13015 load_body->iseq_size = 0;
13017 iseq_bits_t * mark_offset_bits;
13019 iseq_bits_t tmp[1] = {0};
13021 if (ISEQ_MBITS_BUFLEN(iseq_size) == 1) {
13022 mark_offset_bits = tmp;
13025 mark_offset_bits =
ZALLOC_N(iseq_bits_t, ISEQ_MBITS_BUFLEN(iseq_size));
13027 bool needs_bitmap =
false;
13029 for (code_index=0; code_index<iseq_size;) {
13031 const VALUE insn = code[code_index] = ibf_load_small_value(load, &reading_pos);
13032 const char *types = insn_op_types(insn);
13038 for (op_index=0; types[op_index]; op_index++, code_index++) {
13039 const char operand_type = types[op_index];
13040 switch (operand_type) {
13043 VALUE op = ibf_load_small_value(load, &reading_pos);
13044 VALUE v = ibf_load_object(load, op);
13045 code[code_index] = v;
13048 ISEQ_MBITS_SET(mark_offset_bits, code_index);
13049 needs_bitmap =
true;
13055 VALUE op = ibf_load_small_value(load, &reading_pos);
13056 VALUE v = ibf_load_object(load, op);
13057 v = rb_hash_dup(v);
13058 RHASH_TBL_RAW(v)->type = &cdhash_type;
13060 RB_OBJ_SET_SHAREABLE(freeze_hide_obj(v));
13065 pinned_list_store(load->current_buffer->obj_list, (
long)op, v);
13067 code[code_index] = v;
13068 ISEQ_MBITS_SET(mark_offset_bits, code_index);
13070 needs_bitmap =
true;
13075 VALUE op = (
VALUE)ibf_load_small_value(load, &reading_pos);
13077 code[code_index] = v;
13080 ISEQ_MBITS_SET(mark_offset_bits, code_index);
13081 needs_bitmap =
true;
13087 VALUE op = ibf_load_small_value(load, &reading_pos);
13088 VALUE arr = ibf_load_object(load, op);
13090 IC ic = &ISEQ_IS_IC_ENTRY(load_body, ic_index++);
13091 ic->
segments = array_to_idlist(arr);
13093 code[code_index] = (
VALUE)ic;
13100 unsigned int op = (
unsigned int)ibf_load_small_value(load, &reading_pos);
13102 ISE ic = ISEQ_IS_ENTRY_START(load_body, operand_type) + op;
13103 code[code_index] = (
VALUE)ic;
13105 if (operand_type == TS_IVC) {
13108 if (insn == BIN(setinstancevariable)) {
13109 ID iv_name = (
ID)code[code_index - 1];
13110 cache->iv_set_name = iv_name;
13113 cache->iv_set_name = 0;
13116 vm_ic_attr_index_initialize(cache, INVALID_SHAPE_ID);
13123 code[code_index] = (
VALUE)cd_entries++;
13128 VALUE op = ibf_load_small_value(load, &reading_pos);
13129 code[code_index] = ibf_load_id(load, (
ID)(
VALUE)op);
13136 code[code_index] = (
VALUE)ibf_load_builtin(load, &reading_pos);
13139 code[code_index] = ibf_load_small_value(load, &reading_pos);
13143 if (insn_len(insn) != op_index+1) {
13148 load_body->iseq_size = code_index;
13150 if (ISEQ_MBITS_BUFLEN(load_body->iseq_size) == 1) {
13151 load_body->mark_bits.single = mark_offset_bits[0];
13154 if (needs_bitmap) {
13155 load_body->mark_bits.list = mark_offset_bits;
13158 load_body->mark_bits.list = 0;
13159 SIZED_FREE_N(mark_offset_bits, ISEQ_MBITS_BUFLEN(iseq_size));
13164 RUBY_ASSERT(reading_pos == bytecode_offset + bytecode_size);
13171 int opt_num = ISEQ_BODY(iseq)->param.opt_num;
13174 IBF_W_ALIGN(
VALUE);
13175 return ibf_dump_write(dump, ISEQ_BODY(iseq)->param.opt_table,
sizeof(
VALUE) * (opt_num + 1));
13178 return ibf_dump_pos(dump);
13183ibf_load_param_opt_table(
const struct ibf_load *load, ibf_offset_t opt_table_offset,
int opt_num)
13187 MEMCPY(table, load->current_buffer->buff + opt_table_offset,
VALUE, opt_num+1);
13198 const struct rb_iseq_param_keyword *kw = ISEQ_BODY(iseq)->param.keyword;
13201 struct rb_iseq_param_keyword dump_kw = *kw;
13202 int dv_num = kw->num - kw->required_num;
13207 for (i=0; i<kw->num; i++) ids[i] = (
ID)ibf_dump_id(dump, kw->table[i]);
13208 for (i=0; i<dv_num; i++) dvs[i] = (
VALUE)ibf_dump_object(dump, kw->default_values[i]);
13210 dump_kw.table = IBF_W(ids,
ID, kw->num);
13211 dump_kw.default_values = IBF_W(dvs,
VALUE, dv_num);
13212 IBF_W_ALIGN(
struct rb_iseq_param_keyword);
13213 return ibf_dump_write(dump, &dump_kw,
sizeof(
struct rb_iseq_param_keyword) * 1);
13220static const struct rb_iseq_param_keyword *
13221ibf_load_param_keyword(
const struct ibf_load *load, ibf_offset_t param_keyword_offset)
13223 if (param_keyword_offset) {
13224 struct rb_iseq_param_keyword *kw = IBF_R(param_keyword_offset,
struct rb_iseq_param_keyword, 1);
13225 int dv_num = kw->num - kw->required_num;
13226 VALUE *dvs = dv_num ? IBF_R(kw->default_values,
VALUE, dv_num) : NULL;
13229 for (i=0; i<dv_num; i++) {
13230 dvs[i] = ibf_load_object(load, dvs[i]);
13236 kw->default_values = dvs;
13247 ibf_offset_t offset = ibf_dump_pos(dump);
13251 for (i = 0; i < ISEQ_BODY(iseq)->insns_info.size; i++) {
13252 ibf_dump_write_small_value(dump, entries[i].line_no);
13253#ifdef USE_ISEQ_NODE_ID
13254 ibf_dump_write_small_value(dump, entries[i].node_id);
13256 ibf_dump_write_small_value(dump, entries[i].events);
13263ibf_load_insns_info_body(
const struct ibf_load *load, ibf_offset_t body_offset,
unsigned int size)
13265 ibf_offset_t reading_pos = body_offset;
13269 for (i = 0; i < size; i++) {
13270 entries[i].line_no = (int)ibf_load_small_value(load, &reading_pos);
13271#ifdef USE_ISEQ_NODE_ID
13272 entries[i].node_id = (int)ibf_load_small_value(load, &reading_pos);
13274 entries[i].events = (
rb_event_flag_t)ibf_load_small_value(load, &reading_pos);
13281ibf_dump_insns_info_positions(
struct ibf_dump *dump,
const unsigned int *positions,
unsigned int size)
13283 ibf_offset_t offset = ibf_dump_pos(dump);
13285 unsigned int last = 0;
13287 for (i = 0; i < size; i++) {
13288 ibf_dump_write_small_value(dump, positions[i] - last);
13289 last = positions[i];
13295static unsigned int *
13296ibf_load_insns_info_positions(
const struct ibf_load *load, ibf_offset_t positions_offset,
unsigned int size)
13298 ibf_offset_t reading_pos = positions_offset;
13299 unsigned int *positions =
ALLOC_N(
unsigned int, size);
13301 unsigned int last = 0;
13303 for (i = 0; i < size; i++) {
13304 positions[i] = last + (
unsigned int)ibf_load_small_value(load, &reading_pos);
13305 last = positions[i];
13315 const int size = body->local_table_size;
13319 for (i=0; i<size; i++) {
13320 VALUE v = ibf_dump_id(dump, body->local_table[i]);
13323 v = ibf_dump_object(dump,
ULONG2NUM(body->local_table[i]));
13329 return ibf_dump_write(dump, table,
sizeof(
ID) * size);
13333ibf_load_local_table(
const struct ibf_load *load, ibf_offset_t local_table_offset,
int size)
13336 ID *table = IBF_R(local_table_offset,
ID, size);
13339 for (i=0; i<size; i++) {
13340 table[i] = ibf_load_id(load, table[i]);
13343 if (size == 1 && table[0] == idERROR_INFO) {
13344 ruby_sized_xfree(table,
sizeof(
ID) * size);
13345 return rb_iseq_shared_exc_local_tbl;
13360 const int size = body->local_table_size;
13361 IBF_W_ALIGN(
enum lvar_state);
13362 return ibf_dump_write(dump, body->lvar_states,
sizeof(
enum lvar_state) * (body->lvar_states ? size : 0));
13365static enum lvar_state *
13366ibf_load_lvar_states(
const struct ibf_load *load, ibf_offset_t lvar_states_offset,
int size,
const ID *local_table)
13368 if (local_table == rb_iseq_shared_exc_local_tbl ||
13373 enum lvar_state *states = IBF_R(lvar_states_offset,
enum lvar_state, size);
13384 int *iseq_indices =
ALLOCA_N(
int, table->size);
13387 for (i=0; i<table->size; i++) {
13388 iseq_indices[i] = ibf_dump_iseq(dump, table->entries[i].iseq);
13391 const ibf_offset_t offset = ibf_dump_pos(dump);
13393 for (i=0; i<table->size; i++) {
13394 ibf_dump_write_small_value(dump, iseq_indices[i]);
13395 ibf_dump_write_small_value(dump, table->entries[i].type);
13396 ibf_dump_write_small_value(dump, table->entries[i].start);
13397 ibf_dump_write_small_value(dump, table->entries[i].end);
13398 ibf_dump_write_small_value(dump, table->entries[i].cont);
13399 ibf_dump_write_small_value(dump, table->entries[i].sp);
13404 return ibf_dump_pos(dump);
13409ibf_load_catch_table(
const struct ibf_load *load, ibf_offset_t catch_table_offset,
unsigned int size,
const rb_iseq_t *parent_iseq)
13412 struct iseq_catch_table *table = ruby_xcalloc(1, iseq_catch_table_bytes(size));
13413 table->size = size;
13414 ISEQ_BODY(parent_iseq)->catch_table = table;
13416 ibf_offset_t reading_pos = catch_table_offset;
13419 for (i=0; i<table->size; i++) {
13420 int iseq_index = (int)ibf_load_small_value(load, &reading_pos);
13421 table->entries[i].type = (
enum rb_catch_type)ibf_load_small_value(load, &reading_pos);
13422 table->entries[i].start = (
unsigned int)ibf_load_small_value(load, &reading_pos);
13423 table->entries[i].end = (
unsigned int)ibf_load_small_value(load, &reading_pos);
13424 table->entries[i].cont = (
unsigned int)ibf_load_small_value(load, &reading_pos);
13425 table->entries[i].sp = (
unsigned int)ibf_load_small_value(load, &reading_pos);
13428 RB_OBJ_WRITE(parent_iseq, UNALIGNED_MEMBER_PTR(&table->entries[i], iseq), catch_iseq);
13432 ISEQ_BODY(parent_iseq)->catch_table = NULL;
13440 const unsigned int ci_size = body->ci_size;
13443 ibf_offset_t offset = ibf_dump_pos(dump);
13447 for (i = 0; i < ci_size; i++) {
13450 ibf_dump_write_small_value(dump, ibf_dump_id(dump, vm_ci_mid(ci)));
13451 ibf_dump_write_small_value(dump, vm_ci_flag(ci));
13452 ibf_dump_write_small_value(dump, vm_ci_argc(ci));
13456 int len = kwarg->keyword_len;
13457 ibf_dump_write_small_value(dump,
len);
13458 for (
int j=0; j<
len; j++) {
13459 VALUE keyword = ibf_dump_object(dump, kwarg->keywords[j]);
13460 ibf_dump_write_small_value(dump, keyword);
13464 ibf_dump_write_small_value(dump, 0);
13469 ibf_dump_write_small_value(dump, (
VALUE)-1);
13487static enum rb_id_table_iterator_result
13488store_outer_variable(
ID id,
VALUE val,
void *dump)
13493 pair->name = rb_id2str(
id);
13495 return ID_TABLE_CONTINUE;
13499outer_variable_cmp(
const void *a,
const void *b,
void *arg)
13507 else if (!bp->name) {
13517 struct rb_id_table * ovs = ISEQ_BODY(iseq)->outer_variables;
13519 ibf_offset_t offset = ibf_dump_pos(dump);
13521 size_t size = ovs ? rb_id_table_size(ovs) : 0;
13522 ibf_dump_write_small_value(dump, (
VALUE)size);
13531 rb_id_table_foreach(ovs, store_outer_variable, ovlist);
13533 for (
size_t i = 0; i < size; ++i) {
13534 ID id = ovlist->pairs[i].id;
13535 ID val = ovlist->pairs[i].val;
13536 ibf_dump_write_small_value(dump, ibf_dump_id(dump,
id));
13537 ibf_dump_write_small_value(dump, val);
13546ibf_load_ci_entries(
const struct ibf_load *load,
13547 ibf_offset_t ci_entries_offset,
13548 unsigned int ci_size,
13556 ibf_offset_t reading_pos = ci_entries_offset;
13563 for (i = 0; i < ci_size; i++) {
13564 VALUE mid_index = ibf_load_small_value(load, &reading_pos);
13565 if (mid_index != (
VALUE)-1) {
13566 ID mid = ibf_load_id(load, mid_index);
13567 unsigned int flag = (
unsigned int)ibf_load_small_value(load, &reading_pos);
13568 unsigned int argc = (
unsigned int)ibf_load_small_value(load, &reading_pos);
13571 int kwlen = (int)ibf_load_small_value(load, &reading_pos);
13574 kwarg->references = 0;
13575 kwarg->keyword_len = kwlen;
13576 for (
int j=0; j<kwlen; j++) {
13577 VALUE keyword = ibf_load_small_value(load, &reading_pos);
13578 kwarg->keywords[j] = ibf_load_object(load, keyword);
13582 cds[i].ci = vm_ci_new(mid, flag, argc, kwarg);
13584 cds[i].cc = vm_cc_empty();
13595ibf_load_outer_variables(
const struct ibf_load * load, ibf_offset_t outer_variables_offset)
13597 ibf_offset_t reading_pos = outer_variables_offset;
13601 size_t table_size = (size_t)ibf_load_small_value(load, &reading_pos);
13603 if (table_size > 0) {
13604 tbl = rb_id_table_create(table_size);
13607 for (
size_t i = 0; i < table_size; i++) {
13608 ID key = ibf_load_id(load, (
ID)ibf_load_small_value(load, &reading_pos));
13609 VALUE value = ibf_load_small_value(load, &reading_pos);
13610 if (!key) key = rb_make_temporary_id(i);
13611 rb_id_table_insert(tbl, key, value);
13620 RUBY_ASSERT(dump->current_buffer == &dump->global_buffer);
13622 unsigned int *positions;
13626 const VALUE location_pathobj_index = ibf_dump_object(dump, body->location.pathobj);
13627 const VALUE location_base_label_index = ibf_dump_object(dump, body->location.base_label);
13628 const VALUE location_label_index = ibf_dump_object(dump, body->location.label);
13630#if IBF_ISEQ_ENABLE_LOCAL_BUFFER
13631 ibf_offset_t iseq_start = ibf_dump_pos(dump);
13636 buffer.obj_table = ibf_dump_object_table_new();
13637 dump->current_buffer = &buffer;
13640 const ibf_offset_t bytecode_offset = ibf_dump_code(dump, iseq);
13641 const ibf_offset_t bytecode_size = ibf_dump_pos(dump) - bytecode_offset;
13642 const ibf_offset_t param_opt_table_offset = ibf_dump_param_opt_table(dump, iseq);
13643 const ibf_offset_t param_keyword_offset = ibf_dump_param_keyword(dump, iseq);
13644 const ibf_offset_t insns_info_body_offset = ibf_dump_insns_info_body(dump, iseq);
13646 positions = rb_iseq_insns_info_decode_positions(ISEQ_BODY(iseq));
13647 const ibf_offset_t insns_info_positions_offset = ibf_dump_insns_info_positions(dump, positions, body->insns_info.size);
13648 SIZED_FREE_N(positions, ISEQ_BODY(iseq)->insns_info.size);
13650 const ibf_offset_t local_table_offset = ibf_dump_local_table(dump, iseq);
13651 const ibf_offset_t lvar_states_offset = ibf_dump_lvar_states(dump, iseq);
13652 const unsigned int catch_table_size = body->catch_table ? body->catch_table->size : 0;
13653 const ibf_offset_t catch_table_offset = ibf_dump_catch_table(dump, iseq);
13654 const int parent_iseq_index = ibf_dump_iseq(dump, ISEQ_BODY(iseq)->parent_iseq);
13655 const int local_iseq_index = ibf_dump_iseq(dump, ISEQ_BODY(iseq)->local_iseq);
13656 const int mandatory_only_iseq_index = ibf_dump_iseq(dump, ISEQ_BODY(iseq)->mandatory_only_iseq);
13657 const ibf_offset_t ci_entries_offset = ibf_dump_ci_entries(dump, iseq);
13658 const ibf_offset_t outer_variables_offset = ibf_dump_outer_variables(dump, iseq);
13660#if IBF_ISEQ_ENABLE_LOCAL_BUFFER
13661 ibf_offset_t local_obj_list_offset;
13662 unsigned int local_obj_list_size;
13664 ibf_dump_object_list(dump, &local_obj_list_offset, &local_obj_list_size);
13667 ibf_offset_t body_offset = ibf_dump_pos(dump);
13670 unsigned int param_flags =
13671 (body->param.flags.has_lead << 0) |
13672 (body->param.flags.has_opt << 1) |
13673 (body->param.flags.has_rest << 2) |
13674 (body->param.flags.has_post << 3) |
13675 (body->param.flags.has_kw << 4) |
13676 (body->param.flags.has_kwrest << 5) |
13677 (body->param.flags.has_block << 6) |
13678 (body->param.flags.ambiguous_param0 << 7) |
13679 (body->param.flags.accepts_no_kwarg << 8) |
13680 (body->param.flags.ruby2_keywords << 9) |
13681 (body->param.flags.anon_rest << 10) |
13682 (body->param.flags.anon_kwrest << 11) |
13683 (body->param.flags.use_block << 12) |
13684 (body->param.flags.forwardable << 13) |
13685 (body->param.flags.accepts_no_block << 14);
13687#if IBF_ISEQ_ENABLE_LOCAL_BUFFER
13688# define IBF_BODY_OFFSET(x) (x)
13690# define IBF_BODY_OFFSET(x) (body_offset - (x))
13693 ibf_dump_write_small_value(dump, body->type);
13694 ibf_dump_write_small_value(dump, body->iseq_size);
13695 ibf_dump_write_small_value(dump, IBF_BODY_OFFSET(bytecode_offset));
13696 ibf_dump_write_small_value(dump, bytecode_size);
13697 ibf_dump_write_small_value(dump, param_flags);
13698 ibf_dump_write_small_value(dump, body->param.size);
13699 ibf_dump_write_small_value(dump, body->param.lead_num);
13700 ibf_dump_write_small_value(dump, body->param.opt_num);
13701 ibf_dump_write_small_value(dump, body->param.rest_start);
13702 ibf_dump_write_small_value(dump, body->param.post_start);
13703 ibf_dump_write_small_value(dump, body->param.post_num);
13704 ibf_dump_write_small_value(dump, body->param.block_start);
13705 ibf_dump_write_small_value(dump, IBF_BODY_OFFSET(param_opt_table_offset));
13706 ibf_dump_write_small_value(dump, param_keyword_offset);
13707 ibf_dump_write_small_value(dump, location_pathobj_index);
13708 ibf_dump_write_small_value(dump, location_base_label_index);
13709 ibf_dump_write_small_value(dump, location_label_index);
13710 ibf_dump_write_small_value(dump, body->location.first_lineno);
13711 ibf_dump_write_small_value(dump, body->location.node_id);
13712 ibf_dump_write_small_value(dump, body->location.code_location.beg_pos.lineno);
13713 ibf_dump_write_small_value(dump, body->location.code_location.beg_pos.column);
13714 ibf_dump_write_small_value(dump, body->location.code_location.end_pos.lineno);
13715 ibf_dump_write_small_value(dump, body->location.code_location.end_pos.column);
13716 ibf_dump_write_small_value(dump, IBF_BODY_OFFSET(insns_info_body_offset));
13717 ibf_dump_write_small_value(dump, IBF_BODY_OFFSET(insns_info_positions_offset));
13718 ibf_dump_write_small_value(dump, body->insns_info.size);
13719 ibf_dump_write_small_value(dump, IBF_BODY_OFFSET(local_table_offset));
13720 ibf_dump_write_small_value(dump, IBF_BODY_OFFSET(lvar_states_offset));
13721 ibf_dump_write_small_value(dump, catch_table_size);
13722 ibf_dump_write_small_value(dump, IBF_BODY_OFFSET(catch_table_offset));
13723 ibf_dump_write_small_value(dump, parent_iseq_index);
13724 ibf_dump_write_small_value(dump, local_iseq_index);
13725 ibf_dump_write_small_value(dump, mandatory_only_iseq_index);
13726 ibf_dump_write_small_value(dump, IBF_BODY_OFFSET(ci_entries_offset));
13727 ibf_dump_write_small_value(dump, IBF_BODY_OFFSET(outer_variables_offset));
13728 ibf_dump_write_small_value(dump, body->variable.flip_count);
13729 ibf_dump_write_small_value(dump, body->local_table_size);
13730 ibf_dump_write_small_value(dump, body->ivc_size);
13731 ibf_dump_write_small_value(dump, body->icvarc_size);
13732 ibf_dump_write_small_value(dump, body->ise_size);
13733 ibf_dump_write_small_value(dump, body->ic_size);
13734 ibf_dump_write_small_value(dump, body->ci_size);
13735 ibf_dump_write_small_value(dump, body->stack_max);
13736 ibf_dump_write_small_value(dump, body->builtin_attrs);
13737 ibf_dump_write_small_value(dump, body->prism ? 1 : 0);
13739#undef IBF_BODY_OFFSET
13741#if IBF_ISEQ_ENABLE_LOCAL_BUFFER
13742 ibf_offset_t iseq_length_bytes = ibf_dump_pos(dump);
13744 dump->current_buffer = saved_buffer;
13745 ibf_dump_write(dump, RSTRING_PTR(buffer.str), iseq_length_bytes);
13747 ibf_offset_t offset = ibf_dump_pos(dump);
13748 ibf_dump_write_small_value(dump, iseq_start);
13749 ibf_dump_write_small_value(dump, iseq_length_bytes);
13750 ibf_dump_write_small_value(dump, body_offset);
13752 ibf_dump_write_small_value(dump, local_obj_list_offset);
13753 ibf_dump_write_small_value(dump, local_obj_list_size);
13755 st_free_table(buffer.obj_table);
13759 return body_offset;
13764ibf_load_location_str(
const struct ibf_load *load,
VALUE str_index)
13766 VALUE str = ibf_load_object(load, str_index);
13768 str = rb_fstring(str);
13778 ibf_offset_t reading_pos = offset;
13780#if IBF_ISEQ_ENABLE_LOCAL_BUFFER
13782 load->current_buffer = &load->global_buffer;
13784 const ibf_offset_t iseq_start = (ibf_offset_t)ibf_load_small_value(load, &reading_pos);
13785 const ibf_offset_t iseq_length_bytes = (ibf_offset_t)ibf_load_small_value(load, &reading_pos);
13786 const ibf_offset_t body_offset = (ibf_offset_t)ibf_load_small_value(load, &reading_pos);
13789 buffer.buff = load->global_buffer.buff + iseq_start;
13790 buffer.size = iseq_length_bytes;
13791 buffer.obj_list_offset = (ibf_offset_t)ibf_load_small_value(load, &reading_pos);
13792 buffer.obj_list_size = (ibf_offset_t)ibf_load_small_value(load, &reading_pos);
13793 buffer.obj_list = pinned_list_new(buffer.obj_list_size);
13795 load->current_buffer = &buffer;
13796 reading_pos = body_offset;
13799#if IBF_ISEQ_ENABLE_LOCAL_BUFFER
13800# define IBF_BODY_OFFSET(x) (x)
13802# define IBF_BODY_OFFSET(x) (offset - (x))
13805 const unsigned int type = (
unsigned int)ibf_load_small_value(load, &reading_pos);
13806 const unsigned int iseq_size = (
unsigned int)ibf_load_small_value(load, &reading_pos);
13807 const ibf_offset_t bytecode_offset = (ibf_offset_t)IBF_BODY_OFFSET(ibf_load_small_value(load, &reading_pos));
13808 const ibf_offset_t bytecode_size = (ibf_offset_t)ibf_load_small_value(load, &reading_pos);
13809 const unsigned int param_flags = (
unsigned int)ibf_load_small_value(load, &reading_pos);
13810 const unsigned int param_size = (
unsigned int)ibf_load_small_value(load, &reading_pos);
13811 const int param_lead_num = (int)ibf_load_small_value(load, &reading_pos);
13812 const int param_opt_num = (int)ibf_load_small_value(load, &reading_pos);
13813 const int param_rest_start = (int)ibf_load_small_value(load, &reading_pos);
13814 const int param_post_start = (int)ibf_load_small_value(load, &reading_pos);
13815 const int param_post_num = (int)ibf_load_small_value(load, &reading_pos);
13816 const int param_block_start = (int)ibf_load_small_value(load, &reading_pos);
13817 const ibf_offset_t param_opt_table_offset = (ibf_offset_t)IBF_BODY_OFFSET(ibf_load_small_value(load, &reading_pos));
13818 const ibf_offset_t param_keyword_offset = (ibf_offset_t)ibf_load_small_value(load, &reading_pos);
13819 const VALUE location_pathobj_index = ibf_load_small_value(load, &reading_pos);
13820 const VALUE location_base_label_index = ibf_load_small_value(load, &reading_pos);
13821 const VALUE location_label_index = ibf_load_small_value(load, &reading_pos);
13822 const int location_first_lineno = (int)ibf_load_small_value(load, &reading_pos);
13823 const int location_node_id = (int)ibf_load_small_value(load, &reading_pos);
13824 const int location_code_location_beg_pos_lineno = (int)ibf_load_small_value(load, &reading_pos);
13825 const int location_code_location_beg_pos_column = (int)ibf_load_small_value(load, &reading_pos);
13826 const int location_code_location_end_pos_lineno = (int)ibf_load_small_value(load, &reading_pos);
13827 const int location_code_location_end_pos_column = (int)ibf_load_small_value(load, &reading_pos);
13828 const ibf_offset_t insns_info_body_offset = (ibf_offset_t)IBF_BODY_OFFSET(ibf_load_small_value(load, &reading_pos));
13829 const ibf_offset_t insns_info_positions_offset = (ibf_offset_t)IBF_BODY_OFFSET(ibf_load_small_value(load, &reading_pos));
13830 const unsigned int insns_info_size = (
unsigned int)ibf_load_small_value(load, &reading_pos);
13831 const ibf_offset_t local_table_offset = (ibf_offset_t)IBF_BODY_OFFSET(ibf_load_small_value(load, &reading_pos));
13832 const ibf_offset_t lvar_states_offset = (ibf_offset_t)IBF_BODY_OFFSET(ibf_load_small_value(load, &reading_pos));
13833 const unsigned int catch_table_size = (
unsigned int)ibf_load_small_value(load, &reading_pos);
13834 const ibf_offset_t catch_table_offset = (ibf_offset_t)IBF_BODY_OFFSET(ibf_load_small_value(load, &reading_pos));
13835 const int parent_iseq_index = (int)ibf_load_small_value(load, &reading_pos);
13836 const int local_iseq_index = (int)ibf_load_small_value(load, &reading_pos);
13837 const int mandatory_only_iseq_index = (int)ibf_load_small_value(load, &reading_pos);
13838 const ibf_offset_t ci_entries_offset = (ibf_offset_t)IBF_BODY_OFFSET(ibf_load_small_value(load, &reading_pos));
13839 const ibf_offset_t outer_variables_offset = (ibf_offset_t)IBF_BODY_OFFSET(ibf_load_small_value(load, &reading_pos));
13840 const rb_snum_t variable_flip_count = (rb_snum_t)ibf_load_small_value(load, &reading_pos);
13841 const unsigned int local_table_size = (
unsigned int)ibf_load_small_value(load, &reading_pos);
13843 const unsigned int ivc_size = (
unsigned int)ibf_load_small_value(load, &reading_pos);
13844 const unsigned int icvarc_size = (
unsigned int)ibf_load_small_value(load, &reading_pos);
13845 const unsigned int ise_size = (
unsigned int)ibf_load_small_value(load, &reading_pos);
13846 const unsigned int ic_size = (
unsigned int)ibf_load_small_value(load, &reading_pos);
13848 const unsigned int ci_size = (
unsigned int)ibf_load_small_value(load, &reading_pos);
13849 const unsigned int stack_max = (
unsigned int)ibf_load_small_value(load, &reading_pos);
13850 const unsigned int builtin_attrs = (
unsigned int)ibf_load_small_value(load, &reading_pos);
13851 const bool prism = (bool)ibf_load_small_value(load, &reading_pos);
13854 VALUE path = ibf_load_object(load, location_pathobj_index);
13859 realpath = path = rb_fstring(path);
13862 VALUE pathobj = path;
13868 if (!
NIL_P(realpath)) {
13870 rb_raise(rb_eArgError,
"unexpected realpath %"PRIxVALUE
13871 "(%x), path=%+"PRIsVALUE,
13872 realpath,
TYPE(realpath), path);
13874 realpath = rb_fstring(realpath);
13880 rb_iseq_pathobj_set(iseq, path, realpath);
13885 VALUE dummy_frame = rb_vm_push_frame_fname(ec, path);
13887#undef IBF_BODY_OFFSET
13889 load_body->type =
type;
13890 load_body->stack_max = stack_max;
13891 load_body->param.flags.has_lead = (param_flags >> 0) & 1;
13892 load_body->param.flags.has_opt = (param_flags >> 1) & 1;
13893 load_body->param.flags.has_rest = (param_flags >> 2) & 1;
13894 load_body->param.flags.has_post = (param_flags >> 3) & 1;
13895 load_body->param.flags.has_kw = FALSE;
13896 load_body->param.flags.has_kwrest = (param_flags >> 5) & 1;
13897 load_body->param.flags.has_block = (param_flags >> 6) & 1;
13898 load_body->param.flags.ambiguous_param0 = (param_flags >> 7) & 1;
13899 load_body->param.flags.accepts_no_kwarg = (param_flags >> 8) & 1;
13900 load_body->param.flags.ruby2_keywords = (param_flags >> 9) & 1;
13901 load_body->param.flags.anon_rest = (param_flags >> 10) & 1;
13902 load_body->param.flags.anon_kwrest = (param_flags >> 11) & 1;
13903 load_body->param.flags.use_block = (param_flags >> 12) & 1;
13904 load_body->param.flags.forwardable = (param_flags >> 13) & 1;
13905 load_body->param.flags.accepts_no_block = (param_flags >> 14) & 1;
13906 load_body->param.size = param_size;
13907 load_body->param.lead_num = param_lead_num;
13908 load_body->param.opt_num = param_opt_num;
13909 load_body->param.rest_start = param_rest_start;
13910 load_body->param.post_start = param_post_start;
13911 load_body->param.post_num = param_post_num;
13912 load_body->param.block_start = param_block_start;
13913 load_body->local_table_size = local_table_size;
13914 load_body->ci_size = ci_size;
13915 load_body->insns_info.size = insns_info_size;
13917 ISEQ_COVERAGE_SET(iseq,
Qnil);
13918 ISEQ_ORIGINAL_ISEQ_CLEAR(iseq);
13919 load_body->variable.flip_count = variable_flip_count;
13920 load_body->variable.script_lines =
Qnil;
13922 load_body->location.first_lineno = location_first_lineno;
13923 load_body->location.node_id = location_node_id;
13924 load_body->location.code_location.beg_pos.lineno = location_code_location_beg_pos_lineno;
13925 load_body->location.code_location.beg_pos.column = location_code_location_beg_pos_column;
13926 load_body->location.code_location.end_pos.lineno = location_code_location_end_pos_lineno;
13927 load_body->location.code_location.end_pos.column = location_code_location_end_pos_column;
13928 load_body->builtin_attrs = builtin_attrs;
13929 load_body->prism = prism;
13931 load_body->ivc_size = ivc_size;
13932 load_body->icvarc_size = icvarc_size;
13933 load_body->ise_size = ise_size;
13934 load_body->ic_size = ic_size;
13936 if (ISEQ_IS_SIZE(load_body)) {
13940 load_body->is_entries = NULL;
13942 ibf_load_ci_entries(load, ci_entries_offset, ci_size, &load_body->call_data);
13943 load_body->outer_variables = ibf_load_outer_variables(load, outer_variables_offset);
13944 load_body->param.opt_table = ibf_load_param_opt_table(load, param_opt_table_offset, param_opt_num);
13945 load_body->param.keyword = ibf_load_param_keyword(load, param_keyword_offset);
13946 load_body->param.flags.has_kw = (param_flags >> 4) & 1;
13947 load_body->insns_info.body = ibf_load_insns_info_body(load, insns_info_body_offset, insns_info_size);
13948 load_body->insns_info.positions = ibf_load_insns_info_positions(load, insns_info_positions_offset, insns_info_size);
13949 load_body->local_table = ibf_load_local_table(load, local_table_offset, local_table_size);
13950 load_body->lvar_states = ibf_load_lvar_states(load, lvar_states_offset, local_table_size, load_body->local_table);
13951 ibf_load_catch_table(load, catch_table_offset, catch_table_size, iseq);
13955 const rb_iseq_t *mandatory_only_iseq = ibf_load_iseq(load, (
const rb_iseq_t *)(
VALUE)mandatory_only_iseq_index);
13957 RB_OBJ_WRITE(iseq, &load_body->parent_iseq, parent_iseq);
13958 RB_OBJ_WRITE(iseq, &load_body->local_iseq, local_iseq);
13959 RB_OBJ_WRITE(iseq, &load_body->mandatory_only_iseq, mandatory_only_iseq);
13962 if (load_body->param.keyword != NULL) {
13964 struct rb_iseq_param_keyword *keyword = (
struct rb_iseq_param_keyword *) load_body->param.keyword;
13965 keyword->table = &load_body->local_table[keyword->bits_start - keyword->num];
13968 ibf_load_code(load, iseq, bytecode_offset, bytecode_size, iseq_size);
13969#if VM_INSN_INFO_TABLE_IMPL == 2
13970 rb_iseq_insns_info_encode_positions(iseq);
13973 rb_iseq_translate_threaded_code(iseq);
13975#if IBF_ISEQ_ENABLE_LOCAL_BUFFER
13976 load->current_buffer = &load->global_buffer;
13979 RB_OBJ_WRITE(iseq, &load_body->location.base_label, ibf_load_location_str(load, location_base_label_index));
13980 RB_OBJ_WRITE(iseq, &load_body->location.label, ibf_load_location_str(load, location_label_index));
13982#if IBF_ISEQ_ENABLE_LOCAL_BUFFER
13983 load->current_buffer = saved_buffer;
13985 verify_call_cache(iseq);
13988 rb_vm_pop_frame_no_int(ec);
13998ibf_dump_iseq_list_i(st_data_t key, st_data_t val, st_data_t ptr)
14003 ibf_offset_t offset = ibf_dump_iseq_each(args->dump, iseq);
14006 return ST_CONTINUE;
14016 args.offset_list = offset_list;
14018 st_foreach(dump->iseq_table, ibf_dump_iseq_list_i, (st_data_t)&args);
14021 st_index_t size = dump->iseq_table->num_entries;
14022 ibf_offset_t *offsets =
ALLOCA_N(ibf_offset_t, size);
14024 for (i = 0; i < size; i++) {
14028 ibf_dump_align(dump,
sizeof(ibf_offset_t));
14029 header->iseq_list_offset = ibf_dump_write(dump, offsets,
sizeof(ibf_offset_t) * size);
14030 header->iseq_list_size = (
unsigned int)size;
14040 unsigned int type: 5;
14041 unsigned int special_const: 1;
14042 unsigned int frozen: 1;
14043 unsigned int internal: 1;
14046enum ibf_object_class_index {
14047 IBF_OBJECT_CLASS_OBJECT,
14048 IBF_OBJECT_CLASS_ARRAY,
14049 IBF_OBJECT_CLASS_STANDARD_ERROR,
14050 IBF_OBJECT_CLASS_NO_MATCHING_PATTERN_ERROR,
14051 IBF_OBJECT_CLASS_TYPE_ERROR,
14052 IBF_OBJECT_CLASS_NO_MATCHING_PATTERN_KEY_ERROR,
14062 long keyval[FLEX_ARY_LEN];
14075 BDIGIT digits[FLEX_ARY_LEN];
14078enum ibf_object_data_type {
14079 IBF_OBJECT_DATA_ENCODING,
14090#define IBF_ALIGNED_OFFSET(align, offset) \
14091 ((((offset) - 1) / (align) + 1) * (align))
14096#define IBF_OBJBODY(type, offset) \
14097 ibf_load_check_offset(load, IBF_ALIGNED_OFFSET(RUBY_ALIGNOF(type), offset))
14100ibf_load_check_offset(
const struct ibf_load *load,
size_t offset)
14102 if (offset >= load->current_buffer->size) {
14103 rb_raise(
rb_eIndexError,
"object offset out of range: %"PRIdSIZE, offset);
14105 return load->current_buffer->buff + offset;
14108NORETURN(
static void ibf_dump_object_unsupported(
struct ibf_dump *dump,
VALUE obj));
14111ibf_dump_object_unsupported(
struct ibf_dump *dump,
VALUE obj)
14114 rb_raw_obj_info(buff,
sizeof(buff), obj);
14123 rb_raise(rb_eArgError,
"unsupported");
14130 enum ibf_object_class_index cindex;
14132 cindex = IBF_OBJECT_CLASS_OBJECT;
14135 cindex = IBF_OBJECT_CLASS_ARRAY;
14138 cindex = IBF_OBJECT_CLASS_STANDARD_ERROR;
14141 cindex = IBF_OBJECT_CLASS_NO_MATCHING_PATTERN_ERROR;
14144 cindex = IBF_OBJECT_CLASS_TYPE_ERROR;
14147 cindex = IBF_OBJECT_CLASS_NO_MATCHING_PATTERN_KEY_ERROR;
14150 rb_obj_info_dump(obj);
14152 rb_bug(
"unsupported class");
14154 ibf_dump_write_small_value(dump, (
VALUE)cindex);
14160 enum ibf_object_class_index cindex = (
enum ibf_object_class_index)ibf_load_small_value(load, &offset);
14163 case IBF_OBJECT_CLASS_OBJECT:
14165 case IBF_OBJECT_CLASS_ARRAY:
14167 case IBF_OBJECT_CLASS_STANDARD_ERROR:
14169 case IBF_OBJECT_CLASS_NO_MATCHING_PATTERN_ERROR:
14171 case IBF_OBJECT_CLASS_TYPE_ERROR:
14173 case IBF_OBJECT_CLASS_NO_MATCHING_PATTERN_KEY_ERROR:
14177 rb_raise(rb_eArgError,
"ibf_load_object_class: unknown class (%d)", (
int)cindex);
14185 (void)IBF_W(&dbl,
double, 1);
14193 memcpy(&d, IBF_OBJBODY(
double, offset),
sizeof(d));
14195 if (!
FLONUM_P(f)) RB_OBJ_SET_SHAREABLE(f);
14202 long encindex = (long)rb_enc_get_index(obj);
14203 long len = RSTRING_LEN(obj);
14204 const char *ptr = RSTRING_PTR(obj);
14206 if (encindex > RUBY_ENCINDEX_BUILTIN_MAX) {
14207 rb_encoding *enc = rb_enc_from_index((
int)encindex);
14208 const char *enc_name = rb_enc_name(enc);
14209 encindex = RUBY_ENCINDEX_BUILTIN_MAX + ibf_dump_object(dump,
rb_str_new2(enc_name));
14212 ibf_dump_write_small_value(dump, encindex);
14213 ibf_dump_write_small_value(dump,
len);
14214 IBF_WP(ptr,
char,
len);
14220 ibf_offset_t reading_pos = offset;
14222 int encindex = (int)ibf_load_small_value(load, &reading_pos);
14223 const long len = (long)ibf_load_small_value(load, &reading_pos);
14224 const char *ptr = load->current_buffer->buff + reading_pos;
14226 if (encindex > RUBY_ENCINDEX_BUILTIN_MAX) {
14227 VALUE enc_name_str = ibf_load_object(load, encindex - RUBY_ENCINDEX_BUILTIN_MAX);
14228 encindex = rb_enc_find_index(RSTRING_PTR(enc_name_str));
14232 if (header->frozen && !header->internal) {
14233 str = rb_enc_literal_str(ptr,
len, rb_enc_from_index(encindex));
14236 str = rb_enc_str_new(ptr,
len, rb_enc_from_index(encindex));
14239 if (header->frozen) str = rb_fstring(str);
14250 regexp.srcstr = (long)ibf_dump_object(dump, srcstr);
14252 ibf_dump_write_byte(dump, (
unsigned char)regexp.option);
14253 ibf_dump_write_small_value(dump, regexp.srcstr);
14260 regexp.option = ibf_load_byte(load, &offset);
14261 regexp.srcstr = ibf_load_small_value(load, &offset);
14263 VALUE srcstr = ibf_load_object(load, regexp.srcstr);
14264 VALUE reg = rb_reg_compile(srcstr, (
int)regexp.option, NULL, 0);
14267 if (header->frozen) RB_OBJ_SET_SHAREABLE(
rb_obj_freeze(reg));
14276 ibf_dump_write_small_value(dump,
len);
14277 for (i=0; i<
len; i++) {
14278 long index = (long)ibf_dump_object(dump,
RARRAY_AREF(obj, i));
14279 ibf_dump_write_small_value(dump, index);
14286 ibf_offset_t reading_pos = offset;
14288 const long len = (long)ibf_load_small_value(load, &reading_pos);
14293 for (i=0; i<
len; i++) {
14294 const VALUE index = ibf_load_small_value(load, &reading_pos);
14298 if (header->frozen) {
14307ibf_dump_object_hash_i(st_data_t key, st_data_t val, st_data_t ptr)
14311 VALUE key_index = ibf_dump_object(dump, (
VALUE)key);
14312 VALUE val_index = ibf_dump_object(dump, (
VALUE)val);
14314 ibf_dump_write_small_value(dump, key_index);
14315 ibf_dump_write_small_value(dump, val_index);
14316 return ST_CONTINUE;
14323 ibf_dump_write_small_value(dump, (
VALUE)
len);
14331 long len = (long)ibf_load_small_value(load, &offset);
14332 VALUE obj = rb_hash_new_with_size(
len);
14335 for (i = 0; i <
len; i++) {
14336 VALUE key_index = ibf_load_small_value(load, &offset);
14337 VALUE val_index = ibf_load_small_value(load, &offset);
14339 VALUE key = ibf_load_object(load, key_index);
14340 VALUE val = ibf_load_object(load, val_index);
14341 rb_hash_aset(obj, key, val);
14343 rb_hash_rehash(obj);
14346 if (header->frozen) {
14347 RB_OBJ_SET_FROZEN_SHAREABLE(obj);
14361 range.class_index = 0;
14364 range.beg = (long)ibf_dump_object(dump, beg);
14365 range.end = (long)ibf_dump_object(dump, end);
14371 rb_raise(
rb_eNotImpError,
"ibf_dump_object_struct: unsupported class %"PRIsVALUE,
14380 VALUE beg = ibf_load_object(load, range->beg);
14381 VALUE end = ibf_load_object(load, range->end);
14384 if (header->frozen) RB_OBJ_SET_FROZEN_SHAREABLE(obj);
14391 ssize_t
len = BIGNUM_LEN(obj);
14392 ssize_t slen = BIGNUM_SIGN(obj) > 0 ?
len :
len * -1;
14393 BDIGIT *d = BIGNUM_DIGITS(obj);
14395 (void)IBF_W(&slen, ssize_t, 1);
14396 IBF_WP(d, BDIGIT,
len);
14403 int sign = bignum->slen > 0;
14404 ssize_t
len = sign > 0 ? bignum->slen : -1 * bignum->slen;
14405 const int big_unpack_flags =
14408 VALUE obj = rb_integer_unpack(bignum->digits,
len,
sizeof(BDIGIT), 0,
14412 if (header->frozen) RB_OBJ_SET_FROZEN_SHAREABLE(obj);
14419 if (rb_data_is_encoding(obj)) {
14421 const char *name = rb_enc_name(enc);
14422 long len = strlen(name) + 1;
14424 data[0] = IBF_OBJECT_DATA_ENCODING;
14426 (void)IBF_W(data,
long, 2);
14427 IBF_WP(name,
char,
len);
14430 ibf_dump_object_unsupported(dump, obj);
14437 const long *body = IBF_OBJBODY(
long, offset);
14438 const enum ibf_object_data_type
type = (
enum ibf_object_data_type)body[0];
14440 const char *data = (
const char *)&body[2];
14443 case IBF_OBJECT_DATA_ENCODING:
14445 VALUE encobj = rb_enc_from_encoding(rb_enc_find(data));
14450 return ibf_load_object_unsupported(load, header, offset);
14454ibf_dump_object_complex_rational(
struct ibf_dump *dump,
VALUE obj)
14457 data[0] = (long)ibf_dump_object(dump, RCOMPLEX(obj)->real);
14458 data[1] = (long)ibf_dump_object(dump, RCOMPLEX(obj)->imag);
14460 (void)IBF_W(data,
long, 2);
14464ibf_load_object_complex_rational(
const struct ibf_load *load,
const struct ibf_object_header *header, ibf_offset_t offset)
14467 VALUE a = ibf_load_object(load, nums->a);
14468 VALUE b = ibf_load_object(load, nums->b);
14480 ibf_dump_object_string(dump,
rb_sym2str(obj));
14486 ibf_offset_t reading_pos = offset;
14488 int encindex = (int)ibf_load_small_value(load, &reading_pos);
14489 const long len = (long)ibf_load_small_value(load, &reading_pos);
14490 const char *ptr = load->current_buffer->buff + reading_pos;
14492 if (encindex > RUBY_ENCINDEX_BUILTIN_MAX) {
14493 VALUE enc_name_str = ibf_load_object(load, encindex - RUBY_ENCINDEX_BUILTIN_MAX);
14494 encindex = rb_enc_find_index(RSTRING_PTR(enc_name_str));
14497 ID id = rb_intern3(ptr,
len, rb_enc_from_index(encindex));
14501typedef void (*ibf_dump_object_function)(
struct ibf_dump *dump,
VALUE obj);
14502static const ibf_dump_object_function dump_object_functions[
RUBY_T_MASK+1] = {
14503 ibf_dump_object_unsupported,
14504 ibf_dump_object_unsupported,
14505 ibf_dump_object_class,
14506 ibf_dump_object_unsupported,
14507 ibf_dump_object_float,
14508 ibf_dump_object_string,
14509 ibf_dump_object_regexp,
14510 ibf_dump_object_array,
14511 ibf_dump_object_hash,
14512 ibf_dump_object_struct,
14513 ibf_dump_object_bignum,
14514 ibf_dump_object_unsupported,
14515 ibf_dump_object_data,
14516 ibf_dump_object_unsupported,
14517 ibf_dump_object_complex_rational,
14518 ibf_dump_object_complex_rational,
14519 ibf_dump_object_unsupported,
14520 ibf_dump_object_unsupported,
14521 ibf_dump_object_unsupported,
14522 ibf_dump_object_unsupported,
14523 ibf_dump_object_symbol,
14524 ibf_dump_object_unsupported,
14525 ibf_dump_object_unsupported,
14526 ibf_dump_object_unsupported,
14527 ibf_dump_object_unsupported,
14528 ibf_dump_object_unsupported,
14529 ibf_dump_object_unsupported,
14530 ibf_dump_object_unsupported,
14531 ibf_dump_object_unsupported,
14532 ibf_dump_object_unsupported,
14533 ibf_dump_object_unsupported,
14534 ibf_dump_object_unsupported,
14540 unsigned char byte =
14541 (header.type << 0) |
14542 (header.special_const << 5) |
14543 (header.frozen << 6) |
14544 (header.internal << 7);
14550ibf_load_object_object_header(const struct
ibf_load *load, ibf_offset_t *offset)
14552 unsigned char byte = ibf_load_byte(load, offset);
14555 header.type = (
byte >> 0) & 0x1f;
14556 header.special_const = (
byte >> 5) & 0x01;
14557 header.frozen = (
byte >> 6) & 0x01;
14558 header.internal = (
byte >> 7) & 0x01;
14567 ibf_offset_t current_offset;
14568 IBF_ZERO(obj_header);
14569 obj_header.type =
TYPE(obj);
14571 IBF_W_ALIGN(ibf_offset_t);
14572 current_offset = ibf_dump_pos(dump);
14577 obj_header.special_const = TRUE;
14578 obj_header.frozen = TRUE;
14579 obj_header.internal = TRUE;
14580 ibf_dump_object_object_header(dump, obj_header);
14581 ibf_dump_write_small_value(dump, obj);
14585 obj_header.special_const = FALSE;
14586 obj_header.frozen =
OBJ_FROZEN(obj) ? TRUE : FALSE;
14587 ibf_dump_object_object_header(dump, obj_header);
14588 (*dump_object_functions[obj_header.type])(dump, obj);
14591 return current_offset;
14595static const ibf_load_object_function load_object_functions[
RUBY_T_MASK+1] = {
14596 ibf_load_object_unsupported,
14597 ibf_load_object_unsupported,
14598 ibf_load_object_class,
14599 ibf_load_object_unsupported,
14600 ibf_load_object_float,
14601 ibf_load_object_string,
14602 ibf_load_object_regexp,
14603 ibf_load_object_array,
14604 ibf_load_object_hash,
14605 ibf_load_object_struct,
14606 ibf_load_object_bignum,
14607 ibf_load_object_unsupported,
14608 ibf_load_object_data,
14609 ibf_load_object_unsupported,
14610 ibf_load_object_complex_rational,
14611 ibf_load_object_complex_rational,
14612 ibf_load_object_unsupported,
14613 ibf_load_object_unsupported,
14614 ibf_load_object_unsupported,
14615 ibf_load_object_unsupported,
14616 ibf_load_object_symbol,
14617 ibf_load_object_unsupported,
14618 ibf_load_object_unsupported,
14619 ibf_load_object_unsupported,
14620 ibf_load_object_unsupported,
14621 ibf_load_object_unsupported,
14622 ibf_load_object_unsupported,
14623 ibf_load_object_unsupported,
14624 ibf_load_object_unsupported,
14625 ibf_load_object_unsupported,
14626 ibf_load_object_unsupported,
14627 ibf_load_object_unsupported,
14631ibf_load_object(
const struct ibf_load *load,
VALUE object_index)
14633 if (object_index == 0) {
14637 VALUE obj = pinned_list_fetch(load->current_buffer->obj_list, (
long)object_index);
14639 ibf_offset_t *offsets = (ibf_offset_t *)(load->current_buffer->obj_list_offset + load->current_buffer->buff);
14640 ibf_offset_t offset = offsets[object_index];
14641 const struct ibf_object_header header = ibf_load_object_object_header(load, &offset);
14644 fprintf(stderr,
"ibf_load_object: list=%#x offsets=%p offset=%#x\n",
14645 load->current_buffer->obj_list_offset, (
void *)offsets, offset);
14646 fprintf(stderr,
"ibf_load_object: type=%#x special=%d frozen=%d internal=%d\n",
14647 header.type, header.special_const, header.frozen, header.internal);
14649 if (offset >= load->current_buffer->size) {
14650 rb_raise(
rb_eIndexError,
"object offset out of range: %u", offset);
14653 if (header.special_const) {
14654 ibf_offset_t reading_pos = offset;
14656 obj = ibf_load_small_value(load, &reading_pos);
14659 obj = (*load_object_functions[header.type])(load, &header, offset);
14662 pinned_list_store(load->current_buffer->obj_list, (
long)object_index, obj);
14665 fprintf(stderr,
"ibf_load_object: index=%#"PRIxVALUE
" obj=%#"PRIxVALUE
"\n",
14666 object_index, obj);
14679ibf_dump_object_list_i(st_data_t key, st_data_t val, st_data_t ptr)
14684 ibf_offset_t offset = ibf_dump_object_object(args->dump, obj);
14687 return ST_CONTINUE;
14691ibf_dump_object_list(
struct ibf_dump *dump, ibf_offset_t *obj_list_offset,
unsigned int *obj_list_size)
14693 st_table *obj_table = dump->current_buffer->obj_table;
14698 args.offset_list = offset_list;
14700 st_foreach(obj_table, ibf_dump_object_list_i, (st_data_t)&args);
14702 IBF_W_ALIGN(ibf_offset_t);
14703 *obj_list_offset = ibf_dump_pos(dump);
14705 st_index_t size = obj_table->num_entries;
14708 for (i=0; i<size; i++) {
14713 *obj_list_size = (
unsigned int)size;
14717ibf_dump_mark(
void *ptr)
14720 rb_gc_mark(dump->global_buffer.str);
14722 rb_mark_set(dump->global_buffer.obj_table);
14723 rb_mark_set(dump->iseq_table);
14727ibf_dump_free(
void *ptr)
14730 if (dump->global_buffer.obj_table) {
14731 st_free_table(dump->global_buffer.obj_table);
14732 dump->global_buffer.obj_table = 0;
14734 if (dump->iseq_table) {
14735 st_free_table(dump->iseq_table);
14736 dump->iseq_table = 0;
14741ibf_dump_memsize(
const void *ptr)
14745 if (dump->iseq_table) size += st_memsize(dump->iseq_table);
14746 if (dump->global_buffer.obj_table) size += st_memsize(dump->global_buffer.obj_table);
14752 {ibf_dump_mark, ibf_dump_free, ibf_dump_memsize,},
14759 dump->global_buffer.obj_table = NULL;
14760 dump->iseq_table = NULL;
14763 dump->global_buffer.obj_table = ibf_dump_object_table_new();
14764 dump->iseq_table = st_init_numtable();
14766 dump->current_buffer = &dump->global_buffer;
14777 if (ISEQ_BODY(iseq)->parent_iseq != NULL ||
14778 ISEQ_BODY(iseq)->local_iseq != iseq) {
14781 if (
RTEST(ISEQ_COVERAGE(iseq))) {
14786 ibf_dump_setup(dump, dump_obj);
14788 ibf_dump_write(dump, &header,
sizeof(header));
14789 ibf_dump_iseq(dump, iseq);
14791 header.magic[0] =
'Y';
14792 header.magic[1] =
'A';
14793 header.magic[2] =
'R';
14794 header.magic[3] =
'B';
14795 header.major_version = IBF_MAJOR_VERSION;
14796 header.minor_version = IBF_MINOR_VERSION;
14797 header.endian = IBF_ENDIAN_MARK;
14799 ibf_dump_iseq_list(dump, &header);
14800 ibf_dump_object_list(dump, &header.global_object_list_offset, &header.global_object_list_size);
14801 header.size = ibf_dump_pos(dump);
14804 VALUE opt_str = opt;
14807 ibf_dump_write(dump, ptr, header.extra_size);
14810 header.extra_size = 0;
14813 ibf_dump_overwrite(dump, &header,
sizeof(header), 0);
14815 str = dump->global_buffer.str;
14820static const ibf_offset_t *
14821ibf_iseq_list(
const struct ibf_load *load)
14823 return (
const ibf_offset_t *)(load->global_buffer.buff + load->header->iseq_list_offset);
14827rb_ibf_load_iseq_complete(
rb_iseq_t *iseq)
14831 ibf_offset_t offset = ibf_iseq_list(load)[iseq->aux.loader.index];
14834 fprintf(stderr,
"rb_ibf_load_iseq_complete: index=%#x offset=%#x size=%#x\n",
14835 iseq->aux.loader.index, offset,
14836 load->header->size);
14838 ibf_load_iseq_each(load, iseq, offset);
14839 ISEQ_COMPILE_DATA_CLEAR(iseq);
14841 rb_iseq_init_trace(iseq);
14842 load->iseq = prev_src_iseq;
14849 rb_ibf_load_iseq_complete((
rb_iseq_t *)iseq);
14857 int iseq_index = (int)(
VALUE)index_iseq;
14860 fprintf(stderr,
"ibf_load_iseq: index_iseq=%p iseq_list=%p\n",
14861 (
void *)index_iseq, (
void *)load->iseq_list);
14863 if (iseq_index == -1) {
14867 VALUE iseqv = pinned_list_fetch(load->iseq_list, iseq_index);
14870 fprintf(stderr,
"ibf_load_iseq: iseqv=%p\n", (
void *)iseqv);
14878 fprintf(stderr,
"ibf_load_iseq: new iseq=%p\n", (
void *)iseq);
14881 iseq->aux.loader.obj = load->loader_obj;
14882 iseq->aux.loader.index = iseq_index;
14884 fprintf(stderr,
"ibf_load_iseq: iseq=%p loader_obj=%p index=%d\n",
14885 (
void *)iseq, (
void *)load->loader_obj, iseq_index);
14887 pinned_list_store(load->iseq_list, iseq_index, (
VALUE)iseq);
14889 if (!USE_LAZY_LOAD || GET_VM()->builtin_function_table) {
14891 fprintf(stderr,
"ibf_load_iseq: loading iseq=%p\n", (
void *)iseq);
14893 rb_ibf_load_iseq_complete(iseq);
14897 fprintf(stderr,
"ibf_load_iseq: iseq=%p loaded %p\n",
14898 (
void *)iseq, (
void *)load->iseq);
14906ibf_load_setup_bytes(
struct ibf_load *load,
VALUE loader_obj,
const char *bytes,
size_t size)
14909 load->loader_obj = loader_obj;
14910 load->global_buffer.buff = bytes;
14911 load->header = header;
14912 load->global_buffer.size = header->size;
14913 load->global_buffer.obj_list_offset = header->global_object_list_offset;
14914 load->global_buffer.obj_list_size = header->global_object_list_size;
14915 RB_OBJ_WRITE(loader_obj, &load->iseq_list, pinned_list_new(header->iseq_list_size));
14916 RB_OBJ_WRITE(loader_obj, &load->global_buffer.obj_list, pinned_list_new(load->global_buffer.obj_list_size));
14919 load->current_buffer = &load->global_buffer;
14921 if (size < header->size) {
14924 if (strncmp(header->magic,
"YARB", 4) != 0) {
14927 if (header->major_version != IBF_MAJOR_VERSION ||
14928 header->minor_version != IBF_MINOR_VERSION) {
14930 header->major_version, header->minor_version, IBF_MAJOR_VERSION, IBF_MINOR_VERSION);
14932 if (header->endian != IBF_ENDIAN_MARK) {
14938 if (header->iseq_list_offset %
RUBY_ALIGNOF(ibf_offset_t)) {
14939 rb_raise(rb_eArgError,
"unaligned iseq list offset: %u",
14940 header->iseq_list_offset);
14942 if (load->global_buffer.obj_list_offset %
RUBY_ALIGNOF(ibf_offset_t)) {
14943 rb_raise(rb_eArgError,
"unaligned object list offset: %u",
14944 load->global_buffer.obj_list_offset);
14957 if (USE_LAZY_LOAD) {
14958 str =
rb_str_new(RSTRING_PTR(str), RSTRING_LEN(str));
14961 ibf_load_setup_bytes(load, loader_obj, RSTRING_PTR(str), RSTRING_LEN(str));
14966ibf_loader_mark(
void *ptr)
14969 rb_gc_mark(load->str);
14970 rb_gc_mark(load->iseq_list);
14971 rb_gc_mark(load->global_buffer.obj_list);
14975ibf_loader_free(
void *ptr)
14982ibf_loader_memsize(
const void *ptr)
14989 {ibf_loader_mark, ibf_loader_free, ibf_loader_memsize,},
14994rb_iseq_ibf_load(
VALUE str)
15000 ibf_load_setup(load, loader_obj, str);
15001 iseq = ibf_load_iseq(load, 0);
15008rb_iseq_ibf_load_bytes(
const char *bytes,
size_t size)
15014 ibf_load_setup_bytes(load, loader_obj, bytes, size);
15015 iseq = ibf_load_iseq(load, 0);
15022rb_iseq_ibf_load_extra_data(
VALUE str)
15028 ibf_load_setup(load, loader_obj, str);
15029 extra_str =
rb_str_new(load->global_buffer.buff + load->header->size, load->header->extra_size);
15034#include "prism_compile.c"
#define RUBY_ASSERT(...)
Asserts that the given expression is truthy if and only if RUBY_DEBUG is truthy.
#define RUBY_ALIGNOF
Wraps (or simulates) alignof.
#define RUBY_EVENT_END
Encountered an end of a class clause.
#define RUBY_EVENT_C_CALL
A method, written in C, is called.
#define RUBY_EVENT_B_RETURN
Encountered a next statement.
#define RUBY_EVENT_CLASS
Encountered a new class.
#define RUBY_EVENT_NONE
No events.
#define RUBY_EVENT_LINE
Encountered a new line.
#define RUBY_EVENT_RETURN
Encountered a return statement.
#define RUBY_EVENT_C_RETURN
Return from a method, written in C.
#define RUBY_EVENT_B_CALL
Encountered an yield statement.
uint32_t rb_event_flag_t
Represents event(s).
#define RUBY_EVENT_CALL
A method, written in Ruby, is called.
#define RUBY_EVENT_RESCUE
Encountered a rescue statement.
#define rb_str_new2
Old name of rb_str_new_cstr.
#define T_COMPLEX
Old name of RUBY_T_COMPLEX.
#define TYPE(_)
Old name of rb_type.
#define NUM2ULONG
Old name of RB_NUM2ULONG.
#define NUM2LL
Old name of RB_NUM2LL.
#define ALLOCV
Old name of RB_ALLOCV.
#define RFLOAT_VALUE
Old name of rb_float_value.
#define T_STRING
Old name of RUBY_T_STRING.
#define Qundef
Old name of RUBY_Qundef.
#define INT2FIX
Old name of RB_INT2FIX.
#define OBJ_FROZEN
Old name of RB_OBJ_FROZEN.
#define rb_str_cat2
Old name of rb_str_cat_cstr.
#define T_NIL
Old name of RUBY_T_NIL.
#define UNREACHABLE
Old name of RBIMPL_UNREACHABLE.
#define T_FLOAT
Old name of RUBY_T_FLOAT.
#define ID2SYM
Old name of RB_ID2SYM.
#define T_BIGNUM
Old name of RUBY_T_BIGNUM.
#define SPECIAL_CONST_P
Old name of RB_SPECIAL_CONST_P.
#define OBJ_FREEZE
Old name of RB_OBJ_FREEZE.
#define ULONG2NUM
Old name of RB_ULONG2NUM.
#define UNREACHABLE_RETURN
Old name of RBIMPL_UNREACHABLE_RETURN.
#define SYM2ID
Old name of RB_SYM2ID.
#define FIX2UINT
Old name of RB_FIX2UINT.
#define ZALLOC
Old name of RB_ZALLOC.
#define CLASS_OF
Old name of rb_class_of.
#define FIXABLE
Old name of RB_FIXABLE.
#define xmalloc
Old name of ruby_xmalloc.
#define LONG2FIX
Old name of RB_INT2FIX.
#define FIX2INT
Old name of RB_FIX2INT.
#define NUM2UINT
Old name of RB_NUM2UINT.
#define ZALLOC_N
Old name of RB_ZALLOC_N.
#define ASSUME
Old name of RBIMPL_ASSUME.
#define T_RATIONAL
Old name of RUBY_T_RATIONAL.
#define T_HASH
Old name of RUBY_T_HASH.
#define ALLOC_N
Old name of RB_ALLOC_N.
#define FL_SET
Old name of RB_FL_SET.
#define FLONUM_P
Old name of RB_FLONUM_P.
#define Qtrue
Old name of RUBY_Qtrue.
#define NUM2INT
Old name of RB_NUM2INT.
#define Qnil
Old name of RUBY_Qnil.
#define Qfalse
Old name of RUBY_Qfalse.
#define FIX2LONG
Old name of RB_FIX2LONG.
#define T_ARRAY
Old name of RUBY_T_ARRAY.
#define NIL_P
Old name of RB_NIL_P.
#define T_SYMBOL
Old name of RUBY_T_SYMBOL.
#define DBL2NUM
Old name of rb_float_new.
#define BUILTIN_TYPE
Old name of RB_BUILTIN_TYPE.
#define NUM2LONG
Old name of RB_NUM2LONG.
#define FL_UNSET
Old name of RB_FL_UNSET.
#define UINT2NUM
Old name of RB_UINT2NUM.
#define FIXNUM_P
Old name of RB_FIXNUM_P.
#define CONST_ID
Old name of RUBY_CONST_ID.
#define ALLOCV_END
Old name of RB_ALLOCV_END.
#define SYMBOL_P
Old name of RB_SYMBOL_P.
#define T_REGEXP
Old name of RUBY_T_REGEXP.
#define ruby_debug
This variable controls whether the interpreter is in debug mode.
VALUE rb_eNotImpError
NotImplementedError exception.
VALUE rb_eStandardError
StandardError exception.
VALUE rb_eTypeError
TypeError exception.
VALUE rb_eNoMatchingPatternError
NoMatchingPatternError exception.
void rb_exc_fatal(VALUE mesg)
Raises a fatal error in the current thread.
VALUE rb_eRuntimeError
RuntimeError exception.
void rb_warn(const char *fmt,...)
Identical to rb_warning(), except it reports unless $VERBOSE is nil.
VALUE rb_eNoMatchingPatternKeyError
NoMatchingPatternKeyError exception.
VALUE rb_eIndexError
IndexError exception.
VALUE rb_eSyntaxError
SyntaxError exception.
@ RB_WARN_CATEGORY_STRICT_UNUSED_BLOCK
Warning is for checking unused block strictly.
VALUE rb_obj_reveal(VALUE obj, VALUE klass)
Make a hidden object visible again.
VALUE rb_cArray
Array class.
VALUE rb_cObject
Object class.
VALUE rb_obj_hide(VALUE obj)
Make the object invisible from Ruby code.
VALUE rb_cHash
Hash class.
VALUE rb_inspect(VALUE obj)
Generates a human-readable textual representation of the given object.
VALUE rb_cRange
Range class.
VALUE rb_obj_is_kind_of(VALUE obj, VALUE klass)
Queries if the given object is an instance (of possibly descendants) of the given class.
VALUE rb_obj_freeze(VALUE obj)
Just calls rb_obj_freeze_inline() inside.
#define RB_OBJ_WRITTEN(old, oldv, young)
Identical to RB_OBJ_WRITE(), except it doesn't write any values, but only a WB declaration.
#define RB_OBJ_WRITE(old, slot, young)
Declaration of a "back" pointer.
Defines RBIMPL_HAS_BUILTIN.
VALUE rb_ary_reverse(VALUE ary)
Destructively reverses the passed array in-place.
VALUE rb_ary_dup(VALUE ary)
Duplicates an array.
VALUE rb_ary_cat(VALUE ary, const VALUE *train, long len)
Destructively appends multiple elements at the end of the array.
VALUE rb_ary_new(void)
Allocates a new, empty array.
VALUE rb_ary_new_capa(long capa)
Identical to rb_ary_new(), except it additionally specifies how many rooms of objects it should alloc...
VALUE rb_ary_hidden_new(long capa)
Allocates a hidden (no class) empty array.
VALUE rb_ary_clear(VALUE ary)
Destructively removes everything form an array.
VALUE rb_ary_push(VALUE ary, VALUE elem)
Special case of rb_ary_cat() that it adds only one element.
VALUE rb_ary_freeze(VALUE obj)
Freeze an array, preventing further modifications.
VALUE rb_ary_entry(VALUE ary, long off)
Queries an element of an array.
VALUE rb_ary_join(VALUE ary, VALUE sep)
Recursively stringises the elements of the passed array, flattens that result, then joins the sequenc...
void rb_ary_store(VALUE ary, long key, VALUE val)
Destructively stores the passed value to the passed array's passed index.
#define INTEGER_PACK_NATIVE_BYTE_ORDER
Means either INTEGER_PACK_MSBYTE_FIRST or INTEGER_PACK_LSBYTE_FIRST, depending on the host processor'...
#define INTEGER_PACK_NEGATIVE
Interprets the input as a signed negative number (unpack only).
#define INTEGER_PACK_LSWORD_FIRST
Stores/interprets the least significant word as the first word.
int rb_is_const_id(ID id)
Classifies the given ID, then sees if it is a constant.
int rb_is_attrset_id(ID id)
Classifies the given ID, then sees if it is an attribute writer.
int rb_range_values(VALUE range, VALUE *begp, VALUE *endp, int *exclp)
Deconstructs a range into its components.
VALUE rb_range_new(VALUE beg, VALUE end, int excl)
Creates a new Range.
VALUE rb_rational_new(VALUE num, VALUE den)
Constructs a Rational, with reduction.
int rb_reg_options(VALUE re)
Queries the options of the passed regular expression.
VALUE rb_str_append(VALUE dst, VALUE src)
Identical to rb_str_buf_append(), except it converts the right hand side before concatenating.
VALUE rb_str_tmp_new(long len)
Allocates a "temporary" string.
int rb_str_hash_cmp(VALUE str1, VALUE str2)
Compares two strings.
#define rb_str_new(str, len)
Allocates an instance of rb_cString.
st_index_t rb_str_hash(VALUE str)
Calculates a hash value of a string.
VALUE rb_str_cat(VALUE dst, const char *src, long srclen)
Destructively appends the passed contents to the string.
VALUE rb_str_buf_append(VALUE dst, VALUE src)
Identical to rb_str_cat_cstr(), except it takes Ruby's string instead of C's.
int rb_str_cmp(VALUE lhs, VALUE rhs)
Compares two strings, as in strcmp(3).
VALUE rb_str_concat(VALUE dst, VALUE src)
Identical to rb_str_append(), except it also accepts an integer as a codepoint.
VALUE rb_str_freeze(VALUE str)
This is the implementation of String#freeze.
#define rb_str_new_cstr(str)
Identical to rb_str_new, except it assumes the passed pointer is a pointer to a C string.
VALUE rb_class_name(VALUE obj)
Queries the name of the given object's class.
static ID rb_intern_const(const char *str)
This is a "tiny optimisation" over rb_intern().
VALUE rb_id2sym(ID id)
Allocates an instance of rb_cSymbol that has the given id.
VALUE rb_sym2str(VALUE symbol)
Obtain a frozen string representation of a symbol (not including the leading colon).
ID rb_sym2id(VALUE obj)
Converts an instance of rb_cSymbol into an ID.
int len
Length of the buffer.
#define RB_OBJ_SHAREABLE_P(obj)
Queries if the passed object has previously classified as shareable or not.
VALUE rb_ractor_make_shareable(VALUE obj)
Destructively transforms the passed object so that multiple Ractors can share it.
#define DECIMAL_SIZE_OF(expr)
An approximation of decimal representation size.
void ruby_qsort(void *, const size_t, const size_t, int(*)(const void *, const void *, void *), void *)
Reentrant implementation of quick sort.
#define rb_long2int
Just another name of rb_long2int_inline.
#define MEMCPY(p1, p2, type, n)
Handy macro to call memcpy.
#define ALLOCA_N(type, n)
#define MEMZERO(p, type, n)
Handy macro to erase a region of memory.
#define RB_GC_GUARD(v)
Prevents premature destruction of local objects.
#define RB_ALLOCV(v, n)
Identical to RB_ALLOCV_N(), except that it allocates a number of bytes and returns a void* .
VALUE type(ANYARGS)
ANYARGS-ed function type.
void rb_hash_foreach(VALUE q, int_type *w, VALUE e)
Iteration over the given hash.
#define RBIMPL_ATTR_NORETURN()
Wraps (or simulates) [[noreturn]]
#define RARRAY_LEN
Just another name of rb_array_len.
static int RARRAY_LENINT(VALUE ary)
Identical to rb_array_len(), except it differs for the return type.
static void RARRAY_ASET(VALUE ary, long i, VALUE v)
Assigns an object in an array.
#define RARRAY_AREF(a, i)
#define RARRAY_CONST_PTR
Just another name of rb_array_const_ptr.
static VALUE RBASIC_CLASS(VALUE obj)
Queries the class of an object.
#define RUBY_DEFAULT_FREE
This is a value you can set to RData::dfree.
void(* RUBY_DATA_FUNC)(void *)
This is the type of callbacks registered to RData.
#define RHASH_SIZE(h)
Queries the size of the hash.
static VALUE RREGEXP_SRC(VALUE rexp)
Convenient getter function.
#define StringValue(v)
Ensures that the parameter object is a String.
#define StringValuePtr(v)
Identical to StringValue, except it returns a char*.
static int RSTRING_LENINT(VALUE str)
Identical to RSTRING_LEN(), except it differs for the return type.
#define StringValueCStr(v)
Identical to StringValuePtr, except it additionally checks for the contents for viability as a C stri...
#define RTYPEDDATA_DATA(v)
Convenient getter macro.
#define RUBY_TYPED_FREE_IMMEDIATELY
Macros to see if each corresponding flag is defined.
#define TypedData_Get_Struct(obj, type, data_type, sval)
Obtains a C struct from inside of a wrapper Ruby object.
#define TypedData_Wrap_Struct(klass, data_type, sval)
Converts sval, a pointer to your struct, into a Ruby object.
#define TypedData_Make_Struct(klass, type, data_type, sval)
Identical to TypedData_Wrap_Struct, except it allocates a new data region internally instead of takin...
void rb_p(VALUE obj)
Inspects an object.
static bool RB_SPECIAL_CONST_P(VALUE obj)
Checks if the given object is of enum ruby_special_consts.
#define RTEST
This is an old name of RB_TEST.
#define _(args)
This was a transition path from K&R to ANSI.
Internal header for Complex.
Internal header for Rational.
const ID * segments
A null-terminated list of ids, used to represent a constant's path idNULL is used to represent the ::...
This is the struct that holds necessary info for a struct.
const char * wrap_struct_name
Name of structs of this kind.
uintptr_t ID
Type that represents a Ruby identifier such as a variable name.
#define SIZEOF_VALUE
Identical to sizeof(VALUE), except it is a macro that can also be used inside of preprocessor directi...
uintptr_t VALUE
Type that represents a Ruby object.
static bool RB_FLOAT_TYPE_P(VALUE obj)
Queries if the object is an instance of rb_cFloat.
static bool rb_integer_type_p(VALUE obj)
Queries if the object is an instance of rb_cInteger.
static bool RB_TYPE_P(VALUE obj, enum ruby_value_type t)
Queries if the given object is of given type.
@ RUBY_T_MASK
Bitmask of ruby_value_type.