Ruby 3.5.0dev (2025-04-03 revision 1dddc6c78b5f6dc6ae18ee04ebe44abfce3b0433)
compile.c (1dddc6c78b5f6dc6ae18ee04ebe44abfce3b0433)
1/**********************************************************************
2
3 compile.c - ruby node tree -> VM instruction sequence
4
5 $Author$
6 created at: 04/01/01 03:42:15 JST
7
8 Copyright (C) 2004-2007 Koichi Sasada
9
10**********************************************************************/
11
12#include "ruby/internal/config.h"
13#include <math.h>
14
15#ifdef HAVE_DLADDR
16# include <dlfcn.h>
17#endif
18
19#include "encindex.h"
20#include "id_table.h"
21#include "internal.h"
22#include "internal/array.h"
23#include "internal/compile.h"
24#include "internal/complex.h"
25#include "internal/encoding.h"
26#include "internal/error.h"
27#include "internal/gc.h"
28#include "internal/hash.h"
29#include "internal/io.h"
30#include "internal/numeric.h"
31#include "internal/object.h"
32#include "internal/rational.h"
33#include "internal/re.h"
34#include "internal/ruby_parser.h"
35#include "internal/symbol.h"
36#include "internal/thread.h"
37#include "internal/variable.h"
38#include "iseq.h"
39#include "ruby/ractor.h"
40#include "ruby/re.h"
41#include "ruby/util.h"
42#include "vm_core.h"
43#include "vm_callinfo.h"
44#include "vm_debug.h"
45#include "yjit.h"
46
47#include "builtin.h"
48#include "insns.inc"
49#include "insns_info.inc"
50
51#define FIXNUM_INC(n, i) ((n)+(INT2FIX(i)&~FIXNUM_FLAG))
52
53typedef struct iseq_link_element {
54 enum {
55 ISEQ_ELEMENT_ANCHOR,
56 ISEQ_ELEMENT_LABEL,
57 ISEQ_ELEMENT_INSN,
58 ISEQ_ELEMENT_ADJUST,
59 ISEQ_ELEMENT_TRACE,
60 } type;
61 struct iseq_link_element *next;
62 struct iseq_link_element *prev;
64
65typedef struct iseq_link_anchor {
66 LINK_ELEMENT anchor;
67 LINK_ELEMENT *last;
69
70typedef enum {
71 LABEL_RESCUE_NONE,
72 LABEL_RESCUE_BEG,
73 LABEL_RESCUE_END,
74 LABEL_RESCUE_TYPE_MAX
75} LABEL_RESCUE_TYPE;
76
77typedef struct iseq_label_data {
78 LINK_ELEMENT link;
79 int label_no;
80 int position;
81 int sc_state;
82 int sp;
83 int refcnt;
84 unsigned int set: 1;
85 unsigned int rescued: 2;
86 unsigned int unremovable: 1;
87} LABEL;
88
89typedef struct iseq_insn_data {
90 LINK_ELEMENT link;
91 enum ruby_vminsn_type insn_id;
92 int operand_size;
93 int sc_state;
94 VALUE *operands;
95 struct {
96 int line_no;
97 int node_id;
98 rb_event_flag_t events;
99 } insn_info;
100} INSN;
101
102typedef struct iseq_adjust_data {
103 LINK_ELEMENT link;
104 LABEL *label;
105 int line_no;
106} ADJUST;
107
108typedef struct iseq_trace_data {
109 LINK_ELEMENT link;
110 rb_event_flag_t event;
111 long data;
112} TRACE;
113
115 LABEL *begin;
116 LABEL *end;
117 struct ensure_range *next;
118};
119
121 const void *ensure_node;
123 struct ensure_range *erange;
124};
125
126const ID rb_iseq_shared_exc_local_tbl[] = {idERROR_INFO};
127
141#ifndef CPDEBUG
142#define CPDEBUG 0
143#endif
144
145#if CPDEBUG >= 0
146#define compile_debug CPDEBUG
147#else
148#define compile_debug ISEQ_COMPILE_DATA(iseq)->option->debug_level
149#endif
150
151#if CPDEBUG
152
153#define compile_debug_print_indent(level) \
154 ruby_debug_print_indent((level), compile_debug, gl_node_level * 2)
155
156#define debugp(header, value) (void) \
157 (compile_debug_print_indent(1) && \
158 ruby_debug_print_value(1, compile_debug, (header), (value)))
159
160#define debugi(header, id) (void) \
161 (compile_debug_print_indent(1) && \
162 ruby_debug_print_id(1, compile_debug, (header), (id)))
163
164#define debugp_param(header, value) (void) \
165 (compile_debug_print_indent(1) && \
166 ruby_debug_print_value(1, compile_debug, (header), (value)))
167
168#define debugp_verbose(header, value) (void) \
169 (compile_debug_print_indent(2) && \
170 ruby_debug_print_value(2, compile_debug, (header), (value)))
171
172#define debugp_verbose_node(header, value) (void) \
173 (compile_debug_print_indent(10) && \
174 ruby_debug_print_value(10, compile_debug, (header), (value)))
175
176#define debug_node_start(node) ((void) \
177 (compile_debug_print_indent(1) && \
178 (ruby_debug_print_node(1, CPDEBUG, "", (const NODE *)(node)), gl_node_level)), \
179 gl_node_level++)
180
181#define debug_node_end() gl_node_level --
182
183#else
184
185#define debugi(header, id) ((void)0)
186#define debugp(header, value) ((void)0)
187#define debugp_verbose(header, value) ((void)0)
188#define debugp_verbose_node(header, value) ((void)0)
189#define debugp_param(header, value) ((void)0)
190#define debug_node_start(node) ((void)0)
191#define debug_node_end() ((void)0)
192#endif
193
194#if CPDEBUG > 1 || CPDEBUG < 0
195#undef printf
196#define printf ruby_debug_printf
197#define debugs if (compile_debug_print_indent(1)) ruby_debug_printf
198#define debug_compile(msg, v) ((void)(compile_debug_print_indent(1) && fputs((msg), stderr)), (v))
199#else
200#define debugs if(0)printf
201#define debug_compile(msg, v) (v)
202#endif
203
204#define LVAR_ERRINFO (1)
205
206/* create new label */
207#define NEW_LABEL(l) new_label_body(iseq, (l))
208#define LABEL_FORMAT "<L%03d>"
209
210#define NEW_ISEQ(node, name, type, line_no) \
211 new_child_iseq(iseq, (node), rb_fstring(name), 0, (type), (line_no))
212
213#define NEW_CHILD_ISEQ(node, name, type, line_no) \
214 new_child_iseq(iseq, (node), rb_fstring(name), iseq, (type), (line_no))
215
216#define NEW_CHILD_ISEQ_WITH_CALLBACK(callback_func, name, type, line_no) \
217 new_child_iseq_with_callback(iseq, (callback_func), (name), iseq, (type), (line_no))
218
219/* add instructions */
220#define ADD_SEQ(seq1, seq2) \
221 APPEND_LIST((seq1), (seq2))
222
223/* add an instruction */
224#define ADD_INSN(seq, line_node, insn) \
225 ADD_ELEM((seq), (LINK_ELEMENT *) new_insn_body(iseq, nd_line(line_node), nd_node_id(line_node), BIN(insn), 0))
226
227/* add an instruction with the given line number and node id */
228#define ADD_SYNTHETIC_INSN(seq, line_no, node_id, insn) \
229 ADD_ELEM((seq), (LINK_ELEMENT *) new_insn_body(iseq, (line_no), (node_id), BIN(insn), 0))
230
231/* insert an instruction before next */
232#define INSERT_BEFORE_INSN(next, line_no, node_id, insn) \
233 ELEM_INSERT_PREV(&(next)->link, (LINK_ELEMENT *) new_insn_body(iseq, line_no, node_id, BIN(insn), 0))
234
235/* insert an instruction after prev */
236#define INSERT_AFTER_INSN(prev, line_no, node_id, insn) \
237 ELEM_INSERT_NEXT(&(prev)->link, (LINK_ELEMENT *) new_insn_body(iseq, line_no, node_id, BIN(insn), 0))
238
239/* add an instruction with some operands (1, 2, 3, 5) */
240#define ADD_INSN1(seq, line_node, insn, op1) \
241 ADD_ELEM((seq), (LINK_ELEMENT *) \
242 new_insn_body(iseq, nd_line(line_node), nd_node_id(line_node), BIN(insn), 1, (VALUE)(op1)))
243
244/* insert an instruction with some operands (1, 2, 3, 5) before next */
245#define INSERT_BEFORE_INSN1(next, line_no, node_id, insn, op1) \
246 ELEM_INSERT_PREV(&(next)->link, (LINK_ELEMENT *) \
247 new_insn_body(iseq, line_no, node_id, BIN(insn), 1, (VALUE)(op1)))
248
249/* insert an instruction with some operands (1, 2, 3, 5) after prev */
250#define INSERT_AFTER_INSN1(prev, line_no, node_id, insn, op1) \
251 ELEM_INSERT_NEXT(&(prev)->link, (LINK_ELEMENT *) \
252 new_insn_body(iseq, line_no, node_id, BIN(insn), 1, (VALUE)(op1)))
253
254#define LABEL_REF(label) ((label)->refcnt++)
255
256/* add an instruction with label operand (alias of ADD_INSN1) */
257#define ADD_INSNL(seq, line_node, insn, label) (ADD_INSN1(seq, line_node, insn, label), LABEL_REF(label))
258
259#define ADD_INSN2(seq, line_node, insn, op1, op2) \
260 ADD_ELEM((seq), (LINK_ELEMENT *) \
261 new_insn_body(iseq, nd_line(line_node), nd_node_id(line_node), BIN(insn), 2, (VALUE)(op1), (VALUE)(op2)))
262
263#define ADD_INSN3(seq, line_node, insn, op1, op2, op3) \
264 ADD_ELEM((seq), (LINK_ELEMENT *) \
265 new_insn_body(iseq, nd_line(line_node), nd_node_id(line_node), BIN(insn), 3, (VALUE)(op1), (VALUE)(op2), (VALUE)(op3)))
266
267/* Specific Insn factory */
268#define ADD_SEND(seq, line_node, id, argc) \
269 ADD_SEND_R((seq), (line_node), (id), (argc), NULL, (VALUE)INT2FIX(0), NULL)
270
271#define ADD_SEND_WITH_FLAG(seq, line_node, id, argc, flag) \
272 ADD_SEND_R((seq), (line_node), (id), (argc), NULL, (VALUE)(flag), NULL)
273
274#define ADD_SEND_WITH_BLOCK(seq, line_node, id, argc, block) \
275 ADD_SEND_R((seq), (line_node), (id), (argc), (block), (VALUE)INT2FIX(0), NULL)
276
277#define ADD_CALL_RECEIVER(seq, line_node) \
278 ADD_INSN((seq), (line_node), putself)
279
280#define ADD_CALL(seq, line_node, id, argc) \
281 ADD_SEND_R((seq), (line_node), (id), (argc), NULL, (VALUE)INT2FIX(VM_CALL_FCALL), NULL)
282
283#define ADD_CALL_WITH_BLOCK(seq, line_node, id, argc, block) \
284 ADD_SEND_R((seq), (line_node), (id), (argc), (block), (VALUE)INT2FIX(VM_CALL_FCALL), NULL)
285
286#define ADD_SEND_R(seq, line_node, id, argc, block, flag, keywords) \
287 ADD_ELEM((seq), (LINK_ELEMENT *) new_insn_send(iseq, nd_line(line_node), nd_node_id(line_node), (id), (VALUE)(argc), (block), (VALUE)(flag), (keywords)))
288
289#define ADD_TRACE(seq, event) \
290 ADD_ELEM((seq), (LINK_ELEMENT *)new_trace_body(iseq, (event), 0))
291#define ADD_TRACE_WITH_DATA(seq, event, data) \
292 ADD_ELEM((seq), (LINK_ELEMENT *)new_trace_body(iseq, (event), (data)))
293
294static void iseq_add_getlocal(rb_iseq_t *iseq, LINK_ANCHOR *const seq, const NODE *const line_node, int idx, int level);
295static void iseq_add_setlocal(rb_iseq_t *iseq, LINK_ANCHOR *const seq, const NODE *const line_node, int idx, int level);
296
297#define ADD_GETLOCAL(seq, line_node, idx, level) iseq_add_getlocal(iseq, (seq), (line_node), (idx), (level))
298#define ADD_SETLOCAL(seq, line_node, idx, level) iseq_add_setlocal(iseq, (seq), (line_node), (idx), (level))
299
300/* add label */
301#define ADD_LABEL(seq, label) \
302 ADD_ELEM((seq), (LINK_ELEMENT *) (label))
303
304#define APPEND_LABEL(seq, before, label) \
305 APPEND_ELEM((seq), (before), (LINK_ELEMENT *) (label))
306
307#define ADD_ADJUST(seq, line_node, label) \
308 ADD_ELEM((seq), (LINK_ELEMENT *) new_adjust_body(iseq, (label), nd_line(line_node)))
309
310#define ADD_ADJUST_RESTORE(seq, label) \
311 ADD_ELEM((seq), (LINK_ELEMENT *) new_adjust_body(iseq, (label), -1))
312
313#define LABEL_UNREMOVABLE(label) \
314 ((label) ? (LABEL_REF(label), (label)->unremovable=1) : 0)
315#define ADD_CATCH_ENTRY(type, ls, le, iseqv, lc) do { \
316 VALUE _e = rb_ary_new3(5, (type), \
317 (VALUE)(ls) | 1, (VALUE)(le) | 1, \
318 (VALUE)(iseqv), (VALUE)(lc) | 1); \
319 LABEL_UNREMOVABLE(ls); \
320 LABEL_REF(le); \
321 LABEL_REF(lc); \
322 if (NIL_P(ISEQ_COMPILE_DATA(iseq)->catch_table_ary)) \
323 RB_OBJ_WRITE(iseq, &ISEQ_COMPILE_DATA(iseq)->catch_table_ary, rb_ary_hidden_new(3)); \
324 rb_ary_push(ISEQ_COMPILE_DATA(iseq)->catch_table_ary, freeze_hide_obj(_e)); \
325} while (0)
326
327/* compile node */
328#define COMPILE(anchor, desc, node) \
329 (debug_compile("== " desc "\n", \
330 iseq_compile_each(iseq, (anchor), (node), 0)))
331
332/* compile node, this node's value will be popped */
333#define COMPILE_POPPED(anchor, desc, node) \
334 (debug_compile("== " desc "\n", \
335 iseq_compile_each(iseq, (anchor), (node), 1)))
336
337/* compile node, which is popped when 'popped' is true */
338#define COMPILE_(anchor, desc, node, popped) \
339 (debug_compile("== " desc "\n", \
340 iseq_compile_each(iseq, (anchor), (node), (popped))))
341
342#define COMPILE_RECV(anchor, desc, node, recv) \
343 (private_recv_p(node) ? \
344 (ADD_INSN(anchor, node, putself), VM_CALL_FCALL) : \
345 COMPILE(anchor, desc, recv) ? 0 : -1)
346
347#define OPERAND_AT(insn, idx) \
348 (((INSN*)(insn))->operands[(idx)])
349
350#define INSN_OF(insn) \
351 (((INSN*)(insn))->insn_id)
352
353#define IS_INSN(link) ((link)->type == ISEQ_ELEMENT_INSN)
354#define IS_LABEL(link) ((link)->type == ISEQ_ELEMENT_LABEL)
355#define IS_ADJUST(link) ((link)->type == ISEQ_ELEMENT_ADJUST)
356#define IS_TRACE(link) ((link)->type == ISEQ_ELEMENT_TRACE)
357#define IS_INSN_ID(iobj, insn) (INSN_OF(iobj) == BIN(insn))
358#define IS_NEXT_INSN_ID(link, insn) \
359 ((link)->next && IS_INSN((link)->next) && IS_INSN_ID((link)->next, insn))
360
361/* error */
362#if CPDEBUG > 0
364#endif
365RBIMPL_ATTR_FORMAT(RBIMPL_PRINTF_FORMAT, 3, 4)
366static void
367append_compile_error(const rb_iseq_t *iseq, int line, const char *fmt, ...)
368{
369 VALUE err_info = ISEQ_COMPILE_DATA(iseq)->err_info;
370 VALUE file = rb_iseq_path(iseq);
371 VALUE err = err_info == Qtrue ? Qfalse : err_info;
372 va_list args;
373
374 va_start(args, fmt);
375 err = rb_syntax_error_append(err, file, line, -1, NULL, fmt, args);
376 va_end(args);
377 if (NIL_P(err_info)) {
378 RB_OBJ_WRITE(iseq, &ISEQ_COMPILE_DATA(iseq)->err_info, err);
379 rb_set_errinfo(err);
380 }
381 else if (!err_info) {
382 RB_OBJ_WRITE(iseq, &ISEQ_COMPILE_DATA(iseq)->err_info, Qtrue);
383 }
384 if (compile_debug) {
385 if (SPECIAL_CONST_P(err)) err = rb_eSyntaxError;
386 rb_exc_fatal(err);
387 }
388}
389
390#if 0
391static void
392compile_bug(rb_iseq_t *iseq, int line, const char *fmt, ...)
393{
394 va_list args;
395 va_start(args, fmt);
396 rb_report_bug_valist(rb_iseq_path(iseq), line, fmt, args);
397 va_end(args);
398 abort();
399}
400#endif
401
402#define COMPILE_ERROR append_compile_error
403
404#define ERROR_ARGS_AT(n) iseq, nd_line(n),
405#define ERROR_ARGS ERROR_ARGS_AT(node)
406
407#define EXPECT_NODE(prefix, node, ndtype, errval) \
408do { \
409 const NODE *error_node = (node); \
410 enum node_type error_type = nd_type(error_node); \
411 if (error_type != (ndtype)) { \
412 COMPILE_ERROR(ERROR_ARGS_AT(error_node) \
413 prefix ": " #ndtype " is expected, but %s", \
414 ruby_node_name(error_type)); \
415 return errval; \
416 } \
417} while (0)
418
419#define EXPECT_NODE_NONULL(prefix, parent, ndtype, errval) \
420do { \
421 COMPILE_ERROR(ERROR_ARGS_AT(parent) \
422 prefix ": must be " #ndtype ", but 0"); \
423 return errval; \
424} while (0)
425
426#define UNKNOWN_NODE(prefix, node, errval) \
427do { \
428 const NODE *error_node = (node); \
429 COMPILE_ERROR(ERROR_ARGS_AT(error_node) prefix ": unknown node (%s)", \
430 ruby_node_name(nd_type(error_node))); \
431 return errval; \
432} while (0)
433
434#define COMPILE_OK 1
435#define COMPILE_NG 0
436
437#define CHECK(sub) if (!(sub)) {BEFORE_RETURN;return COMPILE_NG;}
438#define NO_CHECK(sub) (void)(sub)
439#define BEFORE_RETURN
440
441#define DECL_ANCHOR(name) \
442 LINK_ANCHOR name[1] = {{{ISEQ_ELEMENT_ANCHOR,},&name[0].anchor}}
443#define INIT_ANCHOR(name) \
444 ((name->last = &name->anchor)->next = NULL) /* re-initialize */
445
446static inline VALUE
447freeze_hide_obj(VALUE obj)
448{
449 OBJ_FREEZE(obj);
450 RBASIC_CLEAR_CLASS(obj);
451 return obj;
452}
453
454#include "optinsn.inc"
455#if OPT_INSTRUCTIONS_UNIFICATION
456#include "optunifs.inc"
457#endif
458
459/* for debug */
460#if CPDEBUG < 0
461#define ISEQ_ARG iseq,
462#define ISEQ_ARG_DECLARE rb_iseq_t *iseq,
463#else
464#define ISEQ_ARG
465#define ISEQ_ARG_DECLARE
466#endif
467
468#if CPDEBUG
469#define gl_node_level ISEQ_COMPILE_DATA(iseq)->node_level
470#endif
471
472static void dump_disasm_list_with_cursor(const LINK_ELEMENT *link, const LINK_ELEMENT *curr, const LABEL *dest);
473static void dump_disasm_list(const LINK_ELEMENT *elem);
474
475static int insn_data_length(INSN *iobj);
476static int calc_sp_depth(int depth, INSN *iobj);
477
478static INSN *new_insn_body(rb_iseq_t *iseq, int line_no, int node_id, enum ruby_vminsn_type insn_id, int argc, ...);
479static LABEL *new_label_body(rb_iseq_t *iseq, long line);
480static ADJUST *new_adjust_body(rb_iseq_t *iseq, LABEL *label, int line);
481static TRACE *new_trace_body(rb_iseq_t *iseq, rb_event_flag_t event, long data);
482
483
484static int iseq_compile_each(rb_iseq_t *iseq, LINK_ANCHOR *anchor, const NODE *n, int);
485static int iseq_setup(rb_iseq_t *iseq, LINK_ANCHOR *const anchor);
486static int iseq_setup_insn(rb_iseq_t *iseq, LINK_ANCHOR *const anchor);
487static int iseq_optimize(rb_iseq_t *iseq, LINK_ANCHOR *const anchor);
488static int iseq_insns_unification(rb_iseq_t *iseq, LINK_ANCHOR *const anchor);
489
490static int iseq_set_local_table(rb_iseq_t *iseq, const rb_ast_id_table_t *tbl, const NODE *const node_args);
491static int iseq_set_exception_local_table(rb_iseq_t *iseq);
492static int iseq_set_arguments(rb_iseq_t *iseq, LINK_ANCHOR *const anchor, const NODE *const node);
493
494static int iseq_set_sequence(rb_iseq_t *iseq, LINK_ANCHOR *const anchor);
495static int iseq_set_exception_table(rb_iseq_t *iseq);
496static int iseq_set_optargs_table(rb_iseq_t *iseq);
497
498static int compile_defined_expr(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, VALUE needstr, bool ignore);
499static int compile_hash(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *node, int method_call_keywords, int popped);
500
501/*
502 * To make Array to LinkedList, use link_anchor
503 */
504
505static void
506verify_list(ISEQ_ARG_DECLARE const char *info, LINK_ANCHOR *const anchor)
507{
508#if CPDEBUG
509 int flag = 0;
510 LINK_ELEMENT *list, *plist;
511
512 if (!compile_debug) return;
513
514 list = anchor->anchor.next;
515 plist = &anchor->anchor;
516 while (list) {
517 if (plist != list->prev) {
518 flag += 1;
519 }
520 plist = list;
521 list = list->next;
522 }
523
524 if (anchor->last != plist && anchor->last != 0) {
525 flag |= 0x70000;
526 }
527
528 if (flag != 0) {
529 rb_bug("list verify error: %08x (%s)", flag, info);
530 }
531#endif
532}
533#if CPDEBUG < 0
534#define verify_list(info, anchor) verify_list(iseq, (info), (anchor))
535#endif
536
537static void
538verify_call_cache(rb_iseq_t *iseq)
539{
540#if CPDEBUG
541 VALUE *original = rb_iseq_original_iseq(iseq);
542 size_t i = 0;
543 while (i < ISEQ_BODY(iseq)->iseq_size) {
544 VALUE insn = original[i];
545 const char *types = insn_op_types(insn);
546
547 for (int j=0; types[j]; j++) {
548 if (types[j] == TS_CALLDATA) {
549 struct rb_call_data *cd = (struct rb_call_data *)original[i+j+1];
550 const struct rb_callinfo *ci = cd->ci;
551 const struct rb_callcache *cc = cd->cc;
552 if (cc != vm_cc_empty()) {
553 vm_ci_dump(ci);
554 rb_bug("call cache is not initialized by vm_cc_empty()");
555 }
556 }
557 }
558 i += insn_len(insn);
559 }
560
561 for (unsigned int i=0; i<ISEQ_BODY(iseq)->ci_size; i++) {
562 struct rb_call_data *cd = &ISEQ_BODY(iseq)->call_data[i];
563 const struct rb_callinfo *ci = cd->ci;
564 const struct rb_callcache *cc = cd->cc;
565 if (cc != NULL && cc != vm_cc_empty()) {
566 vm_ci_dump(ci);
567 rb_bug("call cache is not initialized by vm_cc_empty()");
568 }
569 }
570#endif
571}
572
573/*
574 * elem1, elem2 => elem1, elem2, elem
575 */
576static void
577ADD_ELEM(ISEQ_ARG_DECLARE LINK_ANCHOR *const anchor, LINK_ELEMENT *elem)
578{
579 elem->prev = anchor->last;
580 anchor->last->next = elem;
581 anchor->last = elem;
582 verify_list("add", anchor);
583}
584
585/*
586 * elem1, before, elem2 => elem1, before, elem, elem2
587 */
588static void
589APPEND_ELEM(ISEQ_ARG_DECLARE LINK_ANCHOR *const anchor, LINK_ELEMENT *before, LINK_ELEMENT *elem)
590{
591 elem->prev = before;
592 elem->next = before->next;
593 elem->next->prev = elem;
594 before->next = elem;
595 if (before == anchor->last) anchor->last = elem;
596 verify_list("add", anchor);
597}
598#if CPDEBUG < 0
599#define ADD_ELEM(anchor, elem) ADD_ELEM(iseq, (anchor), (elem))
600#define APPEND_ELEM(anchor, before, elem) APPEND_ELEM(iseq, (anchor), (before), (elem))
601#endif
602
603static int
604branch_coverage_valid_p(rb_iseq_t *iseq, int first_line)
605{
606 if (!ISEQ_COVERAGE(iseq)) return 0;
607 if (!ISEQ_BRANCH_COVERAGE(iseq)) return 0;
608 if (first_line <= 0) return 0;
609 return 1;
610}
611
612#define PTR2NUM(x) (rb_int2inum((intptr_t)(void *)(x)))
613
614static VALUE
615setup_branch(const rb_code_location_t *loc, const char *type, VALUE structure, VALUE key)
616{
617 const int first_lineno = loc->beg_pos.lineno, first_column = loc->beg_pos.column;
618 const int last_lineno = loc->end_pos.lineno, last_column = loc->end_pos.column;
619 VALUE branch = rb_ary_hidden_new(6);
620
621 rb_hash_aset(structure, key, branch);
622 rb_ary_push(branch, ID2SYM(rb_intern(type)));
623 rb_ary_push(branch, INT2FIX(first_lineno));
624 rb_ary_push(branch, INT2FIX(first_column));
625 rb_ary_push(branch, INT2FIX(last_lineno));
626 rb_ary_push(branch, INT2FIX(last_column));
627 return branch;
628}
629
630static VALUE
631decl_branch_base(rb_iseq_t *iseq, VALUE key, const rb_code_location_t *loc, const char *type)
632{
633 if (!branch_coverage_valid_p(iseq, loc->beg_pos.lineno)) return Qundef;
634
635 /*
636 * if !structure[node]
637 * structure[node] = [type, first_lineno, first_column, last_lineno, last_column, branches = {}]
638 * else
639 * branches = structure[node][5]
640 * end
641 */
642
643 VALUE structure = RARRAY_AREF(ISEQ_BRANCH_COVERAGE(iseq), 0);
644 VALUE branch_base = rb_hash_aref(structure, key);
645 VALUE branches;
646
647 if (NIL_P(branch_base)) {
648 branch_base = setup_branch(loc, type, structure, key);
649 branches = rb_hash_new();
650 rb_obj_hide(branches);
651 rb_ary_push(branch_base, branches);
652 }
653 else {
654 branches = RARRAY_AREF(branch_base, 5);
655 }
656
657 return branches;
658}
659
660static NODE
661generate_dummy_line_node(int lineno, int node_id)
662{
663 NODE dummy = { 0 };
664 nd_set_line(&dummy, lineno);
665 nd_set_node_id(&dummy, node_id);
666 return dummy;
667}
668
669static void
670add_trace_branch_coverage(rb_iseq_t *iseq, LINK_ANCHOR *const seq, const rb_code_location_t *loc, int node_id, int branch_id, const char *type, VALUE branches)
671{
672 if (!branch_coverage_valid_p(iseq, loc->beg_pos.lineno)) return;
673
674 /*
675 * if !branches[branch_id]
676 * branches[branch_id] = [type, first_lineno, first_column, last_lineno, last_column, counter_idx]
677 * else
678 * counter_idx= branches[branch_id][5]
679 * end
680 */
681
682 VALUE key = INT2FIX(branch_id);
683 VALUE branch = rb_hash_aref(branches, key);
684 long counter_idx;
685
686 if (NIL_P(branch)) {
687 branch = setup_branch(loc, type, branches, key);
688 VALUE counters = RARRAY_AREF(ISEQ_BRANCH_COVERAGE(iseq), 1);
689 counter_idx = RARRAY_LEN(counters);
690 rb_ary_push(branch, LONG2FIX(counter_idx));
691 rb_ary_push(counters, INT2FIX(0));
692 }
693 else {
694 counter_idx = FIX2LONG(RARRAY_AREF(branch, 5));
695 }
696
697 ADD_TRACE_WITH_DATA(seq, RUBY_EVENT_COVERAGE_BRANCH, counter_idx);
698 ADD_SYNTHETIC_INSN(seq, loc->end_pos.lineno, node_id, nop);
699}
700
701#define ISEQ_LAST_LINE(iseq) (ISEQ_COMPILE_DATA(iseq)->last_line)
702
703static int
704validate_label(st_data_t name, st_data_t label, st_data_t arg)
705{
706 rb_iseq_t *iseq = (rb_iseq_t *)arg;
707 LABEL *lobj = (LABEL *)label;
708 if (!lobj->link.next) {
709 do {
710 COMPILE_ERROR(iseq, lobj->position,
711 "%"PRIsVALUE": undefined label",
712 rb_sym2str((VALUE)name));
713 } while (0);
714 }
715 return ST_CONTINUE;
716}
717
718static void
719validate_labels(rb_iseq_t *iseq, st_table *labels_table)
720{
721 st_foreach(labels_table, validate_label, (st_data_t)iseq);
722 st_free_table(labels_table);
723}
724
725static NODE *
726get_nd_recv(const NODE *node)
727{
728 switch (nd_type(node)) {
729 case NODE_CALL:
730 return RNODE_CALL(node)->nd_recv;
731 case NODE_OPCALL:
732 return RNODE_OPCALL(node)->nd_recv;
733 case NODE_FCALL:
734 return 0;
735 case NODE_QCALL:
736 return RNODE_QCALL(node)->nd_recv;
737 case NODE_VCALL:
738 return 0;
739 case NODE_ATTRASGN:
740 return RNODE_ATTRASGN(node)->nd_recv;
741 case NODE_OP_ASGN1:
742 return RNODE_OP_ASGN1(node)->nd_recv;
743 case NODE_OP_ASGN2:
744 return RNODE_OP_ASGN2(node)->nd_recv;
745 default:
746 rb_bug("unexpected node: %s", ruby_node_name(nd_type(node)));
747 }
748}
749
750static ID
751get_node_call_nd_mid(const NODE *node)
752{
753 switch (nd_type(node)) {
754 case NODE_CALL:
755 return RNODE_CALL(node)->nd_mid;
756 case NODE_OPCALL:
757 return RNODE_OPCALL(node)->nd_mid;
758 case NODE_FCALL:
759 return RNODE_FCALL(node)->nd_mid;
760 case NODE_QCALL:
761 return RNODE_QCALL(node)->nd_mid;
762 case NODE_VCALL:
763 return RNODE_VCALL(node)->nd_mid;
764 case NODE_ATTRASGN:
765 return RNODE_ATTRASGN(node)->nd_mid;
766 default:
767 rb_bug("unexpected node: %s", ruby_node_name(nd_type(node)));
768 }
769}
770
771static NODE *
772get_nd_args(const NODE *node)
773{
774 switch (nd_type(node)) {
775 case NODE_CALL:
776 return RNODE_CALL(node)->nd_args;
777 case NODE_OPCALL:
778 return RNODE_OPCALL(node)->nd_args;
779 case NODE_FCALL:
780 return RNODE_FCALL(node)->nd_args;
781 case NODE_QCALL:
782 return RNODE_QCALL(node)->nd_args;
783 case NODE_VCALL:
784 return 0;
785 case NODE_ATTRASGN:
786 return RNODE_ATTRASGN(node)->nd_args;
787 default:
788 rb_bug("unexpected node: %s", ruby_node_name(nd_type(node)));
789 }
790}
791
792static ID
793get_node_colon_nd_mid(const NODE *node)
794{
795 switch (nd_type(node)) {
796 case NODE_COLON2:
797 return RNODE_COLON2(node)->nd_mid;
798 case NODE_COLON3:
799 return RNODE_COLON3(node)->nd_mid;
800 default:
801 rb_bug("unexpected node: %s", ruby_node_name(nd_type(node)));
802 }
803}
804
805static ID
806get_nd_vid(const NODE *node)
807{
808 switch (nd_type(node)) {
809 case NODE_LASGN:
810 return RNODE_LASGN(node)->nd_vid;
811 case NODE_DASGN:
812 return RNODE_DASGN(node)->nd_vid;
813 case NODE_IASGN:
814 return RNODE_IASGN(node)->nd_vid;
815 case NODE_CVASGN:
816 return RNODE_CVASGN(node)->nd_vid;
817 default:
818 rb_bug("unexpected node: %s", ruby_node_name(nd_type(node)));
819 }
820}
821
822static NODE *
823get_nd_value(const NODE *node)
824{
825 switch (nd_type(node)) {
826 case NODE_LASGN:
827 return RNODE_LASGN(node)->nd_value;
828 case NODE_DASGN:
829 return RNODE_DASGN(node)->nd_value;
830 default:
831 rb_bug("unexpected node: %s", ruby_node_name(nd_type(node)));
832 }
833}
834
835static VALUE
836get_string_value(const NODE *node)
837{
838 switch (nd_type(node)) {
839 case NODE_STR:
840 return rb_node_str_string_val(node);
841 case NODE_FILE:
842 return rb_node_file_path_val(node);
843 default:
844 rb_bug("unexpected node: %s", ruby_node_name(nd_type(node)));
845 }
846}
847
848VALUE
849rb_iseq_compile_callback(rb_iseq_t *iseq, const struct rb_iseq_new_with_callback_callback_func * ifunc)
850{
851 DECL_ANCHOR(ret);
852 INIT_ANCHOR(ret);
853
854 (*ifunc->func)(iseq, ret, ifunc->data);
855
856 ADD_SYNTHETIC_INSN(ret, ISEQ_COMPILE_DATA(iseq)->last_line, -1, leave);
857
858 CHECK(iseq_setup_insn(iseq, ret));
859 return iseq_setup(iseq, ret);
860}
861
862static bool drop_unreachable_return(LINK_ANCHOR *ret);
863
864VALUE
865rb_iseq_compile_node(rb_iseq_t *iseq, const NODE *node)
866{
867 DECL_ANCHOR(ret);
868 INIT_ANCHOR(ret);
869
870 if (node == 0) {
871 NO_CHECK(COMPILE(ret, "nil", node));
872 iseq_set_local_table(iseq, 0, 0);
873 }
874 /* assume node is T_NODE */
875 else if (nd_type_p(node, NODE_SCOPE)) {
876 /* iseq type of top, method, class, block */
877 iseq_set_local_table(iseq, RNODE_SCOPE(node)->nd_tbl, (NODE *)RNODE_SCOPE(node)->nd_args);
878 iseq_set_arguments(iseq, ret, (NODE *)RNODE_SCOPE(node)->nd_args);
879
880 switch (ISEQ_BODY(iseq)->type) {
881 case ISEQ_TYPE_BLOCK:
882 {
883 LABEL *start = ISEQ_COMPILE_DATA(iseq)->start_label = NEW_LABEL(0);
884 LABEL *end = ISEQ_COMPILE_DATA(iseq)->end_label = NEW_LABEL(0);
885
886 start->rescued = LABEL_RESCUE_BEG;
887 end->rescued = LABEL_RESCUE_END;
888
889 ADD_TRACE(ret, RUBY_EVENT_B_CALL);
890 ADD_SYNTHETIC_INSN(ret, ISEQ_BODY(iseq)->location.first_lineno, -1, nop);
891 ADD_LABEL(ret, start);
892 CHECK(COMPILE(ret, "block body", RNODE_SCOPE(node)->nd_body));
893 ADD_LABEL(ret, end);
894 ADD_TRACE(ret, RUBY_EVENT_B_RETURN);
895 ISEQ_COMPILE_DATA(iseq)->last_line = ISEQ_BODY(iseq)->location.code_location.end_pos.lineno;
896
897 /* wide range catch handler must put at last */
898 ADD_CATCH_ENTRY(CATCH_TYPE_REDO, start, end, NULL, start);
899 ADD_CATCH_ENTRY(CATCH_TYPE_NEXT, start, end, NULL, end);
900 break;
901 }
902 case ISEQ_TYPE_CLASS:
903 {
904 ADD_TRACE(ret, RUBY_EVENT_CLASS);
905 CHECK(COMPILE(ret, "scoped node", RNODE_SCOPE(node)->nd_body));
906 ADD_TRACE(ret, RUBY_EVENT_END);
907 ISEQ_COMPILE_DATA(iseq)->last_line = nd_line(node);
908 break;
909 }
910 case ISEQ_TYPE_METHOD:
911 {
912 ISEQ_COMPILE_DATA(iseq)->root_node = RNODE_SCOPE(node)->nd_body;
913 ADD_TRACE(ret, RUBY_EVENT_CALL);
914 CHECK(COMPILE(ret, "scoped node", RNODE_SCOPE(node)->nd_body));
915 ISEQ_COMPILE_DATA(iseq)->root_node = RNODE_SCOPE(node)->nd_body;
916 ADD_TRACE(ret, RUBY_EVENT_RETURN);
917 ISEQ_COMPILE_DATA(iseq)->last_line = nd_line(node);
918 break;
919 }
920 default: {
921 CHECK(COMPILE(ret, "scoped node", RNODE_SCOPE(node)->nd_body));
922 break;
923 }
924 }
925 }
926 else {
927 const char *m;
928#define INVALID_ISEQ_TYPE(type) \
929 ISEQ_TYPE_##type: m = #type; goto invalid_iseq_type
930 switch (ISEQ_BODY(iseq)->type) {
931 case INVALID_ISEQ_TYPE(METHOD);
932 case INVALID_ISEQ_TYPE(CLASS);
933 case INVALID_ISEQ_TYPE(BLOCK);
934 case INVALID_ISEQ_TYPE(EVAL);
935 case INVALID_ISEQ_TYPE(MAIN);
936 case INVALID_ISEQ_TYPE(TOP);
937#undef INVALID_ISEQ_TYPE /* invalid iseq types end */
938 case ISEQ_TYPE_RESCUE:
939 iseq_set_exception_local_table(iseq);
940 CHECK(COMPILE(ret, "rescue", node));
941 break;
942 case ISEQ_TYPE_ENSURE:
943 iseq_set_exception_local_table(iseq);
944 CHECK(COMPILE_POPPED(ret, "ensure", node));
945 break;
946 case ISEQ_TYPE_PLAIN:
947 CHECK(COMPILE(ret, "ensure", node));
948 break;
949 default:
950 COMPILE_ERROR(ERROR_ARGS "unknown scope: %d", ISEQ_BODY(iseq)->type);
951 return COMPILE_NG;
952 invalid_iseq_type:
953 COMPILE_ERROR(ERROR_ARGS "compile/ISEQ_TYPE_%s should not be reached", m);
954 return COMPILE_NG;
955 }
956 }
957
958 if (ISEQ_BODY(iseq)->type == ISEQ_TYPE_RESCUE || ISEQ_BODY(iseq)->type == ISEQ_TYPE_ENSURE) {
959 NODE dummy_line_node = generate_dummy_line_node(0, -1);
960 ADD_GETLOCAL(ret, &dummy_line_node, LVAR_ERRINFO, 0);
961 ADD_INSN1(ret, &dummy_line_node, throw, INT2FIX(0) /* continue throw */ );
962 }
963 else if (!drop_unreachable_return(ret)) {
964 ADD_SYNTHETIC_INSN(ret, ISEQ_COMPILE_DATA(iseq)->last_line, -1, leave);
965 }
966
967#if OPT_SUPPORT_JOKE
968 if (ISEQ_COMPILE_DATA(iseq)->labels_table) {
969 st_table *labels_table = ISEQ_COMPILE_DATA(iseq)->labels_table;
970 ISEQ_COMPILE_DATA(iseq)->labels_table = 0;
971 validate_labels(iseq, labels_table);
972 }
973#endif
974 CHECK(iseq_setup_insn(iseq, ret));
975 return iseq_setup(iseq, ret);
976}
977
978static int
979rb_iseq_translate_threaded_code(rb_iseq_t *iseq)
980{
981#if OPT_DIRECT_THREADED_CODE || OPT_CALL_THREADED_CODE
982 const void * const *table = rb_vm_get_insns_address_table();
983 unsigned int i;
984 VALUE *encoded = (VALUE *)ISEQ_BODY(iseq)->iseq_encoded;
985
986 for (i = 0; i < ISEQ_BODY(iseq)->iseq_size; /* */ ) {
987 int insn = (int)ISEQ_BODY(iseq)->iseq_encoded[i];
988 int len = insn_len(insn);
989 encoded[i] = (VALUE)table[insn];
990 i += len;
991 }
992 FL_SET((VALUE)iseq, ISEQ_TRANSLATED);
993#endif
994
995#if USE_YJIT
996 rb_yjit_live_iseq_count++;
997 rb_yjit_iseq_alloc_count++;
998#endif
999
1000 return COMPILE_OK;
1001}
1002
1003VALUE *
1004rb_iseq_original_iseq(const rb_iseq_t *iseq) /* cold path */
1005{
1006 VALUE *original_code;
1007
1008 if (ISEQ_ORIGINAL_ISEQ(iseq)) return ISEQ_ORIGINAL_ISEQ(iseq);
1009 original_code = ISEQ_ORIGINAL_ISEQ_ALLOC(iseq, ISEQ_BODY(iseq)->iseq_size);
1010 MEMCPY(original_code, ISEQ_BODY(iseq)->iseq_encoded, VALUE, ISEQ_BODY(iseq)->iseq_size);
1011
1012#if OPT_DIRECT_THREADED_CODE || OPT_CALL_THREADED_CODE
1013 {
1014 unsigned int i;
1015
1016 for (i = 0; i < ISEQ_BODY(iseq)->iseq_size; /* */ ) {
1017 const void *addr = (const void *)original_code[i];
1018 const int insn = rb_vm_insn_addr2insn(addr);
1019
1020 original_code[i] = insn;
1021 i += insn_len(insn);
1022 }
1023 }
1024#endif
1025 return original_code;
1026}
1027
1028/*********************************************/
1029/* definition of data structure for compiler */
1030/*********************************************/
1031
1032/*
1033 * On 32-bit SPARC, GCC by default generates SPARC V7 code that may require
1034 * 8-byte word alignment. On the other hand, Oracle Solaris Studio seems to
1035 * generate SPARCV8PLUS code with unaligned memory access instructions.
1036 * That is why the STRICT_ALIGNMENT is defined only with GCC.
1037 */
1038#if defined(__sparc) && SIZEOF_VOIDP == 4 && defined(__GNUC__)
1039 #define STRICT_ALIGNMENT
1040#endif
1041
1042/*
1043 * Some OpenBSD platforms (including sparc64) require strict alignment.
1044 */
1045#if defined(__OpenBSD__)
1046 #include <sys/endian.h>
1047 #ifdef __STRICT_ALIGNMENT
1048 #define STRICT_ALIGNMENT
1049 #endif
1050#endif
1051
1052#ifdef STRICT_ALIGNMENT
1053 #if defined(HAVE_TRUE_LONG_LONG) && SIZEOF_LONG_LONG > SIZEOF_VALUE
1054 #define ALIGNMENT_SIZE SIZEOF_LONG_LONG
1055 #else
1056 #define ALIGNMENT_SIZE SIZEOF_VALUE
1057 #endif
1058 #define PADDING_SIZE_MAX ((size_t)((ALIGNMENT_SIZE) - 1))
1059 #define ALIGNMENT_SIZE_MASK PADDING_SIZE_MAX
1060 /* Note: ALIGNMENT_SIZE == (2 ** N) is expected. */
1061#else
1062 #define PADDING_SIZE_MAX 0
1063#endif /* STRICT_ALIGNMENT */
1064
1065#ifdef STRICT_ALIGNMENT
1066/* calculate padding size for aligned memory access */
1067static size_t
1068calc_padding(void *ptr, size_t size)
1069{
1070 size_t mis;
1071 size_t padding = 0;
1072
1073 mis = (size_t)ptr & ALIGNMENT_SIZE_MASK;
1074 if (mis > 0) {
1075 padding = ALIGNMENT_SIZE - mis;
1076 }
1077/*
1078 * On 32-bit sparc or equivalents, when a single VALUE is requested
1079 * and padding == sizeof(VALUE), it is clear that no padding is needed.
1080 */
1081#if ALIGNMENT_SIZE > SIZEOF_VALUE
1082 if (size == sizeof(VALUE) && padding == sizeof(VALUE)) {
1083 padding = 0;
1084 }
1085#endif
1086
1087 return padding;
1088}
1089#endif /* STRICT_ALIGNMENT */
1090
1091static void *
1092compile_data_alloc_with_arena(struct iseq_compile_data_storage **arena, size_t size)
1093{
1094 void *ptr = 0;
1095 struct iseq_compile_data_storage *storage = *arena;
1096#ifdef STRICT_ALIGNMENT
1097 size_t padding = calc_padding((void *)&storage->buff[storage->pos], size);
1098#else
1099 const size_t padding = 0; /* expected to be optimized by compiler */
1100#endif /* STRICT_ALIGNMENT */
1101
1102 if (size >= INT_MAX - padding) rb_memerror();
1103 if (storage->pos + size + padding > storage->size) {
1104 unsigned int alloc_size = storage->size;
1105
1106 while (alloc_size < size + PADDING_SIZE_MAX) {
1107 if (alloc_size >= INT_MAX / 2) rb_memerror();
1108 alloc_size *= 2;
1109 }
1110 storage->next = (void *)ALLOC_N(char, alloc_size +
1111 offsetof(struct iseq_compile_data_storage, buff));
1112 storage = *arena = storage->next;
1113 storage->next = 0;
1114 storage->pos = 0;
1115 storage->size = alloc_size;
1116#ifdef STRICT_ALIGNMENT
1117 padding = calc_padding((void *)&storage->buff[storage->pos], size);
1118#endif /* STRICT_ALIGNMENT */
1119 }
1120
1121#ifdef STRICT_ALIGNMENT
1122 storage->pos += (int)padding;
1123#endif /* STRICT_ALIGNMENT */
1124
1125 ptr = (void *)&storage->buff[storage->pos];
1126 storage->pos += (int)size;
1127 return ptr;
1128}
1129
1130static void *
1131compile_data_alloc(rb_iseq_t *iseq, size_t size)
1132{
1133 struct iseq_compile_data_storage ** arena = &ISEQ_COMPILE_DATA(iseq)->node.storage_current;
1134 return compile_data_alloc_with_arena(arena, size);
1135}
1136
1137static inline void *
1138compile_data_alloc2(rb_iseq_t *iseq, size_t x, size_t y)
1139{
1140 size_t size = rb_size_mul_or_raise(x, y, rb_eRuntimeError);
1141 return compile_data_alloc(iseq, size);
1142}
1143
1144static inline void *
1145compile_data_calloc2(rb_iseq_t *iseq, size_t x, size_t y)
1146{
1147 size_t size = rb_size_mul_or_raise(x, y, rb_eRuntimeError);
1148 void *p = compile_data_alloc(iseq, size);
1149 memset(p, 0, size);
1150 return p;
1151}
1152
1153static INSN *
1154compile_data_alloc_insn(rb_iseq_t *iseq)
1155{
1156 struct iseq_compile_data_storage ** arena = &ISEQ_COMPILE_DATA(iseq)->insn.storage_current;
1157 return (INSN *)compile_data_alloc_with_arena(arena, sizeof(INSN));
1158}
1159
1160static LABEL *
1161compile_data_alloc_label(rb_iseq_t *iseq)
1162{
1163 return (LABEL *)compile_data_alloc(iseq, sizeof(LABEL));
1164}
1165
1166static ADJUST *
1167compile_data_alloc_adjust(rb_iseq_t *iseq)
1168{
1169 return (ADJUST *)compile_data_alloc(iseq, sizeof(ADJUST));
1170}
1171
1172static TRACE *
1173compile_data_alloc_trace(rb_iseq_t *iseq)
1174{
1175 return (TRACE *)compile_data_alloc(iseq, sizeof(TRACE));
1176}
1177
1178/*
1179 * elem1, elemX => elem1, elem2, elemX
1180 */
1181static void
1182ELEM_INSERT_NEXT(LINK_ELEMENT *elem1, LINK_ELEMENT *elem2)
1183{
1184 elem2->next = elem1->next;
1185 elem2->prev = elem1;
1186 elem1->next = elem2;
1187 if (elem2->next) {
1188 elem2->next->prev = elem2;
1189 }
1190}
1191
1192/*
1193 * elem1, elemX => elemX, elem2, elem1
1194 */
1195static void
1196ELEM_INSERT_PREV(LINK_ELEMENT *elem1, LINK_ELEMENT *elem2)
1197{
1198 elem2->prev = elem1->prev;
1199 elem2->next = elem1;
1200 elem1->prev = elem2;
1201 if (elem2->prev) {
1202 elem2->prev->next = elem2;
1203 }
1204}
1205
1206/*
1207 * elemX, elem1, elemY => elemX, elem2, elemY
1208 */
1209static void
1210ELEM_REPLACE(LINK_ELEMENT *elem1, LINK_ELEMENT *elem2)
1211{
1212 elem2->prev = elem1->prev;
1213 elem2->next = elem1->next;
1214 if (elem1->prev) {
1215 elem1->prev->next = elem2;
1216 }
1217 if (elem1->next) {
1218 elem1->next->prev = elem2;
1219 }
1220}
1221
1222static void
1223ELEM_REMOVE(LINK_ELEMENT *elem)
1224{
1225 elem->prev->next = elem->next;
1226 if (elem->next) {
1227 elem->next->prev = elem->prev;
1228 }
1229}
1230
1231static LINK_ELEMENT *
1232FIRST_ELEMENT(const LINK_ANCHOR *const anchor)
1233{
1234 return anchor->anchor.next;
1235}
1236
1237static LINK_ELEMENT *
1238LAST_ELEMENT(LINK_ANCHOR *const anchor)
1239{
1240 return anchor->last;
1241}
1242
1243static LINK_ELEMENT *
1244ELEM_FIRST_INSN(LINK_ELEMENT *elem)
1245{
1246 while (elem) {
1247 switch (elem->type) {
1248 case ISEQ_ELEMENT_INSN:
1249 case ISEQ_ELEMENT_ADJUST:
1250 return elem;
1251 default:
1252 elem = elem->next;
1253 }
1254 }
1255 return NULL;
1256}
1257
1258static int
1259LIST_INSN_SIZE_ONE(const LINK_ANCHOR *const anchor)
1260{
1261 LINK_ELEMENT *first_insn = ELEM_FIRST_INSN(FIRST_ELEMENT(anchor));
1262 if (first_insn != NULL &&
1263 ELEM_FIRST_INSN(first_insn->next) == NULL) {
1264 return TRUE;
1265 }
1266 else {
1267 return FALSE;
1268 }
1269}
1270
1271static int
1272LIST_INSN_SIZE_ZERO(const LINK_ANCHOR *const anchor)
1273{
1274 if (ELEM_FIRST_INSN(FIRST_ELEMENT(anchor)) == NULL) {
1275 return TRUE;
1276 }
1277 else {
1278 return FALSE;
1279 }
1280}
1281
1282/*
1283 * anc1: e1, e2, e3
1284 * anc2: e4, e5
1285 *#=>
1286 * anc1: e1, e2, e3, e4, e5
1287 * anc2: e4, e5 (broken)
1288 */
1289static void
1290APPEND_LIST(ISEQ_ARG_DECLARE LINK_ANCHOR *const anc1, LINK_ANCHOR *const anc2)
1291{
1292 if (anc2->anchor.next) {
1293 /* LINK_ANCHOR must not loop */
1294 RUBY_ASSERT(anc2->last != &anc2->anchor);
1295 anc1->last->next = anc2->anchor.next;
1296 anc2->anchor.next->prev = anc1->last;
1297 anc1->last = anc2->last;
1298 }
1299 else {
1300 RUBY_ASSERT(anc2->last == &anc2->anchor);
1301 }
1302 verify_list("append", anc1);
1303}
1304#if CPDEBUG < 0
1305#define APPEND_LIST(anc1, anc2) APPEND_LIST(iseq, (anc1), (anc2))
1306#endif
1307
1308#if CPDEBUG && 0
1309static void
1310debug_list(ISEQ_ARG_DECLARE LINK_ANCHOR *const anchor, LINK_ELEMENT *cur)
1311{
1312 LINK_ELEMENT *list = FIRST_ELEMENT(anchor);
1313 printf("----\n");
1314 printf("anch: %p, frst: %p, last: %p\n", (void *)&anchor->anchor,
1315 (void *)anchor->anchor.next, (void *)anchor->last);
1316 while (list) {
1317 printf("curr: %p, next: %p, prev: %p, type: %d\n", (void *)list, (void *)list->next,
1318 (void *)list->prev, (int)list->type);
1319 list = list->next;
1320 }
1321 printf("----\n");
1322
1323 dump_disasm_list_with_cursor(anchor->anchor.next, cur, 0);
1324 verify_list("debug list", anchor);
1325}
1326#if CPDEBUG < 0
1327#define debug_list(anc, cur) debug_list(iseq, (anc), (cur))
1328#endif
1329#else
1330#define debug_list(anc, cur) ((void)0)
1331#endif
1332
1333static TRACE *
1334new_trace_body(rb_iseq_t *iseq, rb_event_flag_t event, long data)
1335{
1336 TRACE *trace = compile_data_alloc_trace(iseq);
1337
1338 trace->link.type = ISEQ_ELEMENT_TRACE;
1339 trace->link.next = NULL;
1340 trace->event = event;
1341 trace->data = data;
1342
1343 return trace;
1344}
1345
1346static LABEL *
1347new_label_body(rb_iseq_t *iseq, long line)
1348{
1349 LABEL *labelobj = compile_data_alloc_label(iseq);
1350
1351 labelobj->link.type = ISEQ_ELEMENT_LABEL;
1352 labelobj->link.next = 0;
1353
1354 labelobj->label_no = ISEQ_COMPILE_DATA(iseq)->label_no++;
1355 labelobj->sc_state = 0;
1356 labelobj->sp = -1;
1357 labelobj->refcnt = 0;
1358 labelobj->set = 0;
1359 labelobj->rescued = LABEL_RESCUE_NONE;
1360 labelobj->unremovable = 0;
1361 labelobj->position = -1;
1362 return labelobj;
1363}
1364
1365static ADJUST *
1366new_adjust_body(rb_iseq_t *iseq, LABEL *label, int line)
1367{
1368 ADJUST *adjust = compile_data_alloc_adjust(iseq);
1369 adjust->link.type = ISEQ_ELEMENT_ADJUST;
1370 adjust->link.next = 0;
1371 adjust->label = label;
1372 adjust->line_no = line;
1373 LABEL_UNREMOVABLE(label);
1374 return adjust;
1375}
1376
1377static void
1378iseq_insn_each_markable_object(INSN *insn, void (*func)(VALUE *, VALUE), VALUE data)
1379{
1380 const char *types = insn_op_types(insn->insn_id);
1381 for (int j = 0; types[j]; j++) {
1382 char type = types[j];
1383 switch (type) {
1384 case TS_CDHASH:
1385 case TS_ISEQ:
1386 case TS_VALUE:
1387 case TS_IC: // constant path array
1388 case TS_CALLDATA: // ci is stored.
1389 func(&OPERAND_AT(insn, j), data);
1390 break;
1391 default:
1392 break;
1393 }
1394 }
1395}
1396
1397static void
1398iseq_insn_each_object_write_barrier(VALUE * obj, VALUE iseq)
1399{
1400 RB_OBJ_WRITTEN(iseq, Qundef, *obj);
1401}
1402
1403static INSN *
1404new_insn_core(rb_iseq_t *iseq, int line_no, int node_id, int insn_id, int argc, VALUE *argv)
1405{
1406 INSN *iobj = compile_data_alloc_insn(iseq);
1407
1408 /* printf("insn_id: %d, line: %d\n", insn_id, nd_line(line_node)); */
1409
1410 iobj->link.type = ISEQ_ELEMENT_INSN;
1411 iobj->link.next = 0;
1412 iobj->insn_id = insn_id;
1413 iobj->insn_info.line_no = line_no;
1414 iobj->insn_info.node_id = node_id;
1415 iobj->insn_info.events = 0;
1416 iobj->operands = argv;
1417 iobj->operand_size = argc;
1418 iobj->sc_state = 0;
1419
1420 iseq_insn_each_markable_object(iobj, iseq_insn_each_object_write_barrier, (VALUE)iseq);
1421
1422 return iobj;
1423}
1424
1425static INSN *
1426new_insn_body(rb_iseq_t *iseq, int line_no, int node_id, enum ruby_vminsn_type insn_id, int argc, ...)
1427{
1428 VALUE *operands = 0;
1429 va_list argv;
1430 if (argc > 0) {
1431 int i;
1432 va_start(argv, argc);
1433 operands = compile_data_alloc2(iseq, sizeof(VALUE), argc);
1434 for (i = 0; i < argc; i++) {
1435 VALUE v = va_arg(argv, VALUE);
1436 operands[i] = v;
1437 }
1438 va_end(argv);
1439 }
1440 return new_insn_core(iseq, line_no, node_id, insn_id, argc, operands);
1441}
1442
1443static const struct rb_callinfo *
1444new_callinfo(rb_iseq_t *iseq, ID mid, int argc, unsigned int flag, struct rb_callinfo_kwarg *kw_arg, int has_blockiseq)
1445{
1446 VM_ASSERT(argc >= 0);
1447
1448 if (kw_arg) {
1449 flag |= VM_CALL_KWARG;
1450 argc += kw_arg->keyword_len;
1451 }
1452
1453 if (!(flag & (VM_CALL_ARGS_SPLAT | VM_CALL_ARGS_BLOCKARG | VM_CALL_KWARG | VM_CALL_KW_SPLAT | VM_CALL_FORWARDING))
1454 && !has_blockiseq) {
1455 flag |= VM_CALL_ARGS_SIMPLE;
1456 }
1457
1458 ISEQ_BODY(iseq)->ci_size++;
1459 const struct rb_callinfo *ci = vm_ci_new(mid, flag, argc, kw_arg);
1460 RB_OBJ_WRITTEN(iseq, Qundef, ci);
1461 return ci;
1462}
1463
1464static INSN *
1465new_insn_send(rb_iseq_t *iseq, int line_no, int node_id, ID id, VALUE argc, const rb_iseq_t *blockiseq, VALUE flag, struct rb_callinfo_kwarg *keywords)
1466{
1467 VALUE *operands = compile_data_calloc2(iseq, sizeof(VALUE), 2);
1468 VALUE ci = (VALUE)new_callinfo(iseq, id, FIX2INT(argc), FIX2INT(flag), keywords, blockiseq != NULL);
1469 operands[0] = ci;
1470 operands[1] = (VALUE)blockiseq;
1471 if (blockiseq) {
1472 RB_OBJ_WRITTEN(iseq, Qundef, blockiseq);
1473 }
1474
1475 INSN *insn;
1476
1477 if (vm_ci_flag((struct rb_callinfo *)ci) & VM_CALL_FORWARDING) {
1478 insn = new_insn_core(iseq, line_no, node_id, BIN(sendforward), 2, operands);
1479 }
1480 else {
1481 insn = new_insn_core(iseq, line_no, node_id, BIN(send), 2, operands);
1482 }
1483
1484 RB_OBJ_WRITTEN(iseq, Qundef, ci);
1485 RB_GC_GUARD(ci);
1486 return insn;
1487}
1488
1489static rb_iseq_t *
1490new_child_iseq(rb_iseq_t *iseq, const NODE *const node,
1491 VALUE name, const rb_iseq_t *parent, enum rb_iseq_type type, int line_no)
1492{
1493 rb_iseq_t *ret_iseq;
1494 VALUE ast_value = rb_ruby_ast_new(node);
1495
1496 debugs("[new_child_iseq]> ---------------------------------------\n");
1497 int isolated_depth = ISEQ_COMPILE_DATA(iseq)->isolated_depth;
1498 ret_iseq = rb_iseq_new_with_opt(ast_value, name,
1499 rb_iseq_path(iseq), rb_iseq_realpath(iseq),
1500 line_no, parent,
1501 isolated_depth ? isolated_depth + 1 : 0,
1502 type, ISEQ_COMPILE_DATA(iseq)->option,
1503 ISEQ_BODY(iseq)->variable.script_lines);
1504 debugs("[new_child_iseq]< ---------------------------------------\n");
1505 return ret_iseq;
1506}
1507
1508static rb_iseq_t *
1509new_child_iseq_with_callback(rb_iseq_t *iseq, const struct rb_iseq_new_with_callback_callback_func *ifunc,
1510 VALUE name, const rb_iseq_t *parent, enum rb_iseq_type type, int line_no)
1511{
1512 rb_iseq_t *ret_iseq;
1513
1514 debugs("[new_child_iseq_with_callback]> ---------------------------------------\n");
1515 ret_iseq = rb_iseq_new_with_callback(ifunc, name,
1516 rb_iseq_path(iseq), rb_iseq_realpath(iseq),
1517 line_no, parent, type, ISEQ_COMPILE_DATA(iseq)->option);
1518 debugs("[new_child_iseq_with_callback]< ---------------------------------------\n");
1519 return ret_iseq;
1520}
1521
1522static void
1523set_catch_except_p(rb_iseq_t *iseq)
1524{
1525 RUBY_ASSERT(ISEQ_COMPILE_DATA(iseq));
1526 ISEQ_COMPILE_DATA(iseq)->catch_except_p = true;
1527 if (ISEQ_BODY(iseq)->parent_iseq != NULL) {
1528 if (ISEQ_COMPILE_DATA(ISEQ_BODY(iseq)->parent_iseq)) {
1529 set_catch_except_p((rb_iseq_t *) ISEQ_BODY(iseq)->parent_iseq);
1530 }
1531 }
1532}
1533
1534/* Set body->catch_except_p to true if the ISeq may catch an exception. If it is false,
1535 JIT-ed code may be optimized. If we are extremely conservative, we should set true
1536 if catch table exists. But we want to optimize while loop, which always has catch
1537 table entries for break/next/redo.
1538
1539 So this function sets true for limited ISeqs with break/next/redo catch table entries
1540 whose child ISeq would really raise an exception. */
1541static void
1542update_catch_except_flags(rb_iseq_t *iseq, struct rb_iseq_constant_body *body)
1543{
1544 unsigned int pos;
1545 size_t i;
1546 int insn;
1547 const struct iseq_catch_table *ct = body->catch_table;
1548
1549 /* This assumes that a block has parent_iseq which may catch an exception from the block, and that
1550 BREAK/NEXT/REDO catch table entries are used only when `throw` insn is used in the block. */
1551 pos = 0;
1552 while (pos < body->iseq_size) {
1553 insn = rb_vm_insn_decode(body->iseq_encoded[pos]);
1554 if (insn == BIN(throw)) {
1555 set_catch_except_p(iseq);
1556 break;
1557 }
1558 pos += insn_len(insn);
1559 }
1560
1561 if (ct == NULL)
1562 return;
1563
1564 for (i = 0; i < ct->size; i++) {
1565 const struct iseq_catch_table_entry *entry =
1566 UNALIGNED_MEMBER_PTR(ct, entries[i]);
1567 if (entry->type != CATCH_TYPE_BREAK
1568 && entry->type != CATCH_TYPE_NEXT
1569 && entry->type != CATCH_TYPE_REDO) {
1570 RUBY_ASSERT(ISEQ_COMPILE_DATA(iseq));
1571 ISEQ_COMPILE_DATA(iseq)->catch_except_p = true;
1572 break;
1573 }
1574 }
1575}
1576
1577static void
1578iseq_insert_nop_between_end_and_cont(rb_iseq_t *iseq)
1579{
1580 VALUE catch_table_ary = ISEQ_COMPILE_DATA(iseq)->catch_table_ary;
1581 if (NIL_P(catch_table_ary)) return;
1582 unsigned int i, tlen = (unsigned int)RARRAY_LEN(catch_table_ary);
1583 const VALUE *tptr = RARRAY_CONST_PTR(catch_table_ary);
1584 for (i = 0; i < tlen; i++) {
1585 const VALUE *ptr = RARRAY_CONST_PTR(tptr[i]);
1586 LINK_ELEMENT *end = (LINK_ELEMENT *)(ptr[2] & ~1);
1587 LINK_ELEMENT *cont = (LINK_ELEMENT *)(ptr[4] & ~1);
1588 LINK_ELEMENT *e;
1589
1590 enum rb_catch_type ct = (enum rb_catch_type)(ptr[0] & 0xffff);
1591
1592 if (ct != CATCH_TYPE_BREAK
1593 && ct != CATCH_TYPE_NEXT
1594 && ct != CATCH_TYPE_REDO) {
1595
1596 for (e = end; e && (IS_LABEL(e) || IS_TRACE(e)); e = e->next) {
1597 if (e == cont) {
1598 INSN *nop = new_insn_core(iseq, 0, -1, BIN(nop), 0, 0);
1599 ELEM_INSERT_NEXT(end, &nop->link);
1600 break;
1601 }
1602 }
1603 }
1604 }
1605
1606 RB_GC_GUARD(catch_table_ary);
1607}
1608
1609static int
1610iseq_setup_insn(rb_iseq_t *iseq, LINK_ANCHOR *const anchor)
1611{
1612 if (RTEST(ISEQ_COMPILE_DATA(iseq)->err_info))
1613 return COMPILE_NG;
1614
1615 /* debugs("[compile step 2] (iseq_array_to_linkedlist)\n"); */
1616
1617 if (compile_debug > 5)
1618 dump_disasm_list(FIRST_ELEMENT(anchor));
1619
1620 debugs("[compile step 3.1 (iseq_optimize)]\n");
1621 iseq_optimize(iseq, anchor);
1622
1623 if (compile_debug > 5)
1624 dump_disasm_list(FIRST_ELEMENT(anchor));
1625
1626 if (ISEQ_COMPILE_DATA(iseq)->option->instructions_unification) {
1627 debugs("[compile step 3.2 (iseq_insns_unification)]\n");
1628 iseq_insns_unification(iseq, anchor);
1629 if (compile_debug > 5)
1630 dump_disasm_list(FIRST_ELEMENT(anchor));
1631 }
1632
1633 debugs("[compile step 3.4 (iseq_insert_nop_between_end_and_cont)]\n");
1634 iseq_insert_nop_between_end_and_cont(iseq);
1635 if (compile_debug > 5)
1636 dump_disasm_list(FIRST_ELEMENT(anchor));
1637
1638 return COMPILE_OK;
1639}
1640
1641static int
1642iseq_setup(rb_iseq_t *iseq, LINK_ANCHOR *const anchor)
1643{
1644 if (RTEST(ISEQ_COMPILE_DATA(iseq)->err_info))
1645 return COMPILE_NG;
1646
1647 debugs("[compile step 4.1 (iseq_set_sequence)]\n");
1648 if (!iseq_set_sequence(iseq, anchor)) return COMPILE_NG;
1649 if (compile_debug > 5)
1650 dump_disasm_list(FIRST_ELEMENT(anchor));
1651
1652 debugs("[compile step 4.2 (iseq_set_exception_table)]\n");
1653 if (!iseq_set_exception_table(iseq)) return COMPILE_NG;
1654
1655 debugs("[compile step 4.3 (set_optargs_table)] \n");
1656 if (!iseq_set_optargs_table(iseq)) return COMPILE_NG;
1657
1658 debugs("[compile step 5 (iseq_translate_threaded_code)] \n");
1659 if (!rb_iseq_translate_threaded_code(iseq)) return COMPILE_NG;
1660
1661 debugs("[compile step 6 (update_catch_except_flags)] \n");
1662 RUBY_ASSERT(ISEQ_COMPILE_DATA(iseq));
1663 update_catch_except_flags(iseq, ISEQ_BODY(iseq));
1664
1665 debugs("[compile step 6.1 (remove unused catch tables)] \n");
1666 RUBY_ASSERT(ISEQ_COMPILE_DATA(iseq));
1667 if (!ISEQ_COMPILE_DATA(iseq)->catch_except_p && ISEQ_BODY(iseq)->catch_table) {
1668 xfree(ISEQ_BODY(iseq)->catch_table);
1669 ISEQ_BODY(iseq)->catch_table = NULL;
1670 }
1671
1672#if VM_INSN_INFO_TABLE_IMPL == 2
1673 if (ISEQ_BODY(iseq)->insns_info.succ_index_table == NULL) {
1674 debugs("[compile step 7 (rb_iseq_insns_info_encode_positions)] \n");
1675 rb_iseq_insns_info_encode_positions(iseq);
1676 }
1677#endif
1678
1679 if (compile_debug > 1) {
1680 VALUE str = rb_iseq_disasm(iseq);
1681 printf("%s\n", StringValueCStr(str));
1682 }
1683 verify_call_cache(iseq);
1684 debugs("[compile step: finish]\n");
1685
1686 return COMPILE_OK;
1687}
1688
1689static int
1690iseq_set_exception_local_table(rb_iseq_t *iseq)
1691{
1692 ISEQ_BODY(iseq)->local_table_size = numberof(rb_iseq_shared_exc_local_tbl);
1693 ISEQ_BODY(iseq)->local_table = rb_iseq_shared_exc_local_tbl;
1694 return COMPILE_OK;
1695}
1696
1697static int
1698get_lvar_level(const rb_iseq_t *iseq)
1699{
1700 int lev = 0;
1701 while (iseq != ISEQ_BODY(iseq)->local_iseq) {
1702 lev++;
1703 iseq = ISEQ_BODY(iseq)->parent_iseq;
1704 }
1705 return lev;
1706}
1707
1708static int
1709get_dyna_var_idx_at_raw(const rb_iseq_t *iseq, ID id)
1710{
1711 unsigned int i;
1712
1713 for (i = 0; i < ISEQ_BODY(iseq)->local_table_size; i++) {
1714 if (ISEQ_BODY(iseq)->local_table[i] == id) {
1715 return (int)i;
1716 }
1717 }
1718 return -1;
1719}
1720
1721static int
1722get_local_var_idx(const rb_iseq_t *iseq, ID id)
1723{
1724 int idx = get_dyna_var_idx_at_raw(ISEQ_BODY(iseq)->local_iseq, id);
1725
1726 if (idx < 0) {
1727 COMPILE_ERROR(iseq, ISEQ_LAST_LINE(iseq),
1728 "get_local_var_idx: %d", idx);
1729 }
1730
1731 return idx;
1732}
1733
1734static int
1735get_dyna_var_idx(const rb_iseq_t *iseq, ID id, int *level, int *ls)
1736{
1737 int lv = 0, idx = -1;
1738 const rb_iseq_t *const topmost_iseq = iseq;
1739
1740 while (iseq) {
1741 idx = get_dyna_var_idx_at_raw(iseq, id);
1742 if (idx >= 0) {
1743 break;
1744 }
1745 iseq = ISEQ_BODY(iseq)->parent_iseq;
1746 lv++;
1747 }
1748
1749 if (idx < 0) {
1750 COMPILE_ERROR(topmost_iseq, ISEQ_LAST_LINE(topmost_iseq),
1751 "get_dyna_var_idx: -1");
1752 }
1753
1754 *level = lv;
1755 *ls = ISEQ_BODY(iseq)->local_table_size;
1756 return idx;
1757}
1758
1759static int
1760iseq_local_block_param_p(const rb_iseq_t *iseq, unsigned int idx, unsigned int level)
1761{
1762 const struct rb_iseq_constant_body *body;
1763 while (level > 0) {
1764 iseq = ISEQ_BODY(iseq)->parent_iseq;
1765 level--;
1766 }
1767 body = ISEQ_BODY(iseq);
1768 if (body->local_iseq == iseq && /* local variables */
1769 body->param.flags.has_block &&
1770 body->local_table_size - body->param.block_start == idx) {
1771 return TRUE;
1772 }
1773 else {
1774 return FALSE;
1775 }
1776}
1777
1778static int
1779iseq_block_param_id_p(const rb_iseq_t *iseq, ID id, int *pidx, int *plevel)
1780{
1781 int level, ls;
1782 int idx = get_dyna_var_idx(iseq, id, &level, &ls);
1783 if (iseq_local_block_param_p(iseq, ls - idx, level)) {
1784 *pidx = ls - idx;
1785 *plevel = level;
1786 return TRUE;
1787 }
1788 else {
1789 return FALSE;
1790 }
1791}
1792
1793static void
1794access_outer_variables(const rb_iseq_t *iseq, int level, ID id, bool write)
1795{
1796 int isolated_depth = ISEQ_COMPILE_DATA(iseq)->isolated_depth;
1797
1798 if (isolated_depth && level >= isolated_depth) {
1799 if (id == rb_intern("yield")) {
1800 COMPILE_ERROR(iseq, ISEQ_LAST_LINE(iseq), "can not yield from isolated Proc");
1801 }
1802 else {
1803 COMPILE_ERROR(iseq, ISEQ_LAST_LINE(iseq), "can not access variable '%s' from isolated Proc", rb_id2name(id));
1804 }
1805 }
1806
1807 for (int i=0; i<level; i++) {
1808 VALUE val;
1809 struct rb_id_table *ovs = ISEQ_BODY(iseq)->outer_variables;
1810
1811 if (!ovs) {
1812 ovs = ISEQ_BODY(iseq)->outer_variables = rb_id_table_create(8);
1813 }
1814
1815 if (rb_id_table_lookup(ISEQ_BODY(iseq)->outer_variables, id, &val)) {
1816 if (write && !val) {
1817 rb_id_table_insert(ISEQ_BODY(iseq)->outer_variables, id, Qtrue);
1818 }
1819 }
1820 else {
1821 rb_id_table_insert(ISEQ_BODY(iseq)->outer_variables, id, RBOOL(write));
1822 }
1823
1824 iseq = ISEQ_BODY(iseq)->parent_iseq;
1825 }
1826}
1827
1828static ID
1829iseq_lvar_id(const rb_iseq_t *iseq, int idx, int level)
1830{
1831 for (int i=0; i<level; i++) {
1832 iseq = ISEQ_BODY(iseq)->parent_iseq;
1833 }
1834
1835 ID id = ISEQ_BODY(iseq)->local_table[ISEQ_BODY(iseq)->local_table_size - idx];
1836 // fprintf(stderr, "idx:%d level:%d ID:%s\n", idx, level, rb_id2name(id));
1837 return id;
1838}
1839
1840static void
1841iseq_add_getlocal(rb_iseq_t *iseq, LINK_ANCHOR *const seq, const NODE *const line_node, int idx, int level)
1842{
1843 if (iseq_local_block_param_p(iseq, idx, level)) {
1844 ADD_INSN2(seq, line_node, getblockparam, INT2FIX((idx) + VM_ENV_DATA_SIZE - 1), INT2FIX(level));
1845 }
1846 else {
1847 ADD_INSN2(seq, line_node, getlocal, INT2FIX((idx) + VM_ENV_DATA_SIZE - 1), INT2FIX(level));
1848 }
1849 if (level > 0) access_outer_variables(iseq, level, iseq_lvar_id(iseq, idx, level), Qfalse);
1850}
1851
1852static void
1853iseq_add_setlocal(rb_iseq_t *iseq, LINK_ANCHOR *const seq, const NODE *const line_node, int idx, int level)
1854{
1855 if (iseq_local_block_param_p(iseq, idx, level)) {
1856 ADD_INSN2(seq, line_node, setblockparam, INT2FIX((idx) + VM_ENV_DATA_SIZE - 1), INT2FIX(level));
1857 }
1858 else {
1859 ADD_INSN2(seq, line_node, setlocal, INT2FIX((idx) + VM_ENV_DATA_SIZE - 1), INT2FIX(level));
1860 }
1861 if (level > 0) access_outer_variables(iseq, level, iseq_lvar_id(iseq, idx, level), Qtrue);
1862}
1863
1864
1865
1866static void
1867iseq_calc_param_size(rb_iseq_t *iseq)
1868{
1869 struct rb_iseq_constant_body *const body = ISEQ_BODY(iseq);
1870 if (body->param.flags.has_opt ||
1871 body->param.flags.has_post ||
1872 body->param.flags.has_rest ||
1873 body->param.flags.has_block ||
1874 body->param.flags.has_kw ||
1875 body->param.flags.has_kwrest) {
1876
1877 if (body->param.flags.has_block) {
1878 body->param.size = body->param.block_start + 1;
1879 }
1880 else if (body->param.flags.has_kwrest) {
1881 body->param.size = body->param.keyword->rest_start + 1;
1882 }
1883 else if (body->param.flags.has_kw) {
1884 body->param.size = body->param.keyword->bits_start + 1;
1885 }
1886 else if (body->param.flags.has_post) {
1887 body->param.size = body->param.post_start + body->param.post_num;
1888 }
1889 else if (body->param.flags.has_rest) {
1890 body->param.size = body->param.rest_start + 1;
1891 }
1892 else if (body->param.flags.has_opt) {
1893 body->param.size = body->param.lead_num + body->param.opt_num;
1894 }
1895 else {
1897 }
1898 }
1899 else {
1900 body->param.size = body->param.lead_num;
1901 }
1902}
1903
1904static int
1905iseq_set_arguments_keywords(rb_iseq_t *iseq, LINK_ANCHOR *const optargs,
1906 const struct rb_args_info *args, int arg_size)
1907{
1908 const rb_node_kw_arg_t *node = args->kw_args;
1909 struct rb_iseq_constant_body *const body = ISEQ_BODY(iseq);
1910 struct rb_iseq_param_keyword *keyword;
1911 const VALUE default_values = rb_ary_hidden_new(1);
1912 const VALUE complex_mark = rb_str_tmp_new(0);
1913 int kw = 0, rkw = 0, di = 0, i;
1914
1915 body->param.flags.has_kw = TRUE;
1916 body->param.keyword = keyword = ZALLOC_N(struct rb_iseq_param_keyword, 1);
1917
1918 while (node) {
1919 kw++;
1920 node = node->nd_next;
1921 }
1922 arg_size += kw;
1923 keyword->bits_start = arg_size++;
1924
1925 node = args->kw_args;
1926 while (node) {
1927 const NODE *val_node = get_nd_value(node->nd_body);
1928 VALUE dv;
1929
1930 if (val_node == NODE_SPECIAL_REQUIRED_KEYWORD) {
1931 ++rkw;
1932 }
1933 else {
1934 switch (nd_type(val_node)) {
1935 case NODE_SYM:
1936 dv = rb_node_sym_string_val(val_node);
1937 break;
1938 case NODE_REGX:
1939 dv = rb_node_regx_string_val(val_node);
1940 break;
1941 case NODE_LINE:
1942 dv = rb_node_line_lineno_val(val_node);
1943 break;
1944 case NODE_INTEGER:
1945 dv = rb_node_integer_literal_val(val_node);
1946 break;
1947 case NODE_FLOAT:
1948 dv = rb_node_float_literal_val(val_node);
1949 break;
1950 case NODE_RATIONAL:
1951 dv = rb_node_rational_literal_val(val_node);
1952 break;
1953 case NODE_IMAGINARY:
1954 dv = rb_node_imaginary_literal_val(val_node);
1955 break;
1956 case NODE_ENCODING:
1957 dv = rb_node_encoding_val(val_node);
1958 break;
1959 case NODE_NIL:
1960 dv = Qnil;
1961 break;
1962 case NODE_TRUE:
1963 dv = Qtrue;
1964 break;
1965 case NODE_FALSE:
1966 dv = Qfalse;
1967 break;
1968 default:
1969 NO_CHECK(COMPILE_POPPED(optargs, "kwarg", RNODE(node))); /* nd_type_p(node, NODE_KW_ARG) */
1970 dv = complex_mark;
1971 }
1972
1973 keyword->num = ++di;
1974 rb_ary_push(default_values, dv);
1975 }
1976
1977 node = node->nd_next;
1978 }
1979
1980 keyword->num = kw;
1981
1982 if (RNODE_DVAR(args->kw_rest_arg)->nd_vid != 0) {
1983 ID kw_id = ISEQ_BODY(iseq)->local_table[arg_size];
1984 keyword->rest_start = arg_size++;
1985 body->param.flags.has_kwrest = TRUE;
1986
1987 if (kw_id == idPow) body->param.flags.anon_kwrest = TRUE;
1988 }
1989 keyword->required_num = rkw;
1990 keyword->table = &body->local_table[keyword->bits_start - keyword->num];
1991
1992 if (RARRAY_LEN(default_values)) {
1993 VALUE *dvs = ALLOC_N(VALUE, RARRAY_LEN(default_values));
1994
1995 for (i = 0; i < RARRAY_LEN(default_values); i++) {
1996 VALUE dv = RARRAY_AREF(default_values, i);
1997 if (dv == complex_mark) dv = Qundef;
1998 RB_OBJ_WRITE(iseq, &dvs[i], dv);
1999 }
2000
2001 keyword->default_values = dvs;
2002 }
2003 return arg_size;
2004}
2005
2006static void
2007iseq_set_use_block(rb_iseq_t *iseq)
2008{
2009 struct rb_iseq_constant_body *const body = ISEQ_BODY(iseq);
2010 if (!body->param.flags.use_block) {
2011 body->param.flags.use_block = 1;
2012
2013 rb_vm_t *vm = GET_VM();
2014
2015 if (!rb_warning_category_enabled_p(RB_WARN_CATEGORY_STRICT_UNUSED_BLOCK)) {
2016 st_data_t key = (st_data_t)rb_intern_str(body->location.label); // String -> ID
2017 st_insert(vm->unused_block_warning_table, key, 1);
2018 }
2019 }
2020}
2021
2022static int
2023iseq_set_arguments(rb_iseq_t *iseq, LINK_ANCHOR *const optargs, const NODE *const node_args)
2024{
2025 debugs("iseq_set_arguments: %s\n", node_args ? "" : "0");
2026
2027 if (node_args) {
2028 struct rb_iseq_constant_body *const body = ISEQ_BODY(iseq);
2029 struct rb_args_info *args = &RNODE_ARGS(node_args)->nd_ainfo;
2030 ID rest_id = 0;
2031 int last_comma = 0;
2032 ID block_id = 0;
2033 int arg_size;
2034
2035 EXPECT_NODE("iseq_set_arguments", node_args, NODE_ARGS, COMPILE_NG);
2036
2037 body->param.flags.ruby2_keywords = args->ruby2_keywords;
2038 body->param.lead_num = arg_size = (int)args->pre_args_num;
2039 if (body->param.lead_num > 0) body->param.flags.has_lead = TRUE;
2040 debugs(" - argc: %d\n", body->param.lead_num);
2041
2042 rest_id = args->rest_arg;
2043 if (rest_id == NODE_SPECIAL_EXCESSIVE_COMMA) {
2044 last_comma = 1;
2045 rest_id = 0;
2046 }
2047 block_id = args->block_arg;
2048
2049 bool optimized_forward = (args->forwarding && args->pre_args_num == 0 && !args->opt_args);
2050
2051 if (optimized_forward) {
2052 rest_id = 0;
2053 block_id = 0;
2054 }
2055
2056 if (args->opt_args) {
2057 const rb_node_opt_arg_t *node = args->opt_args;
2058 LABEL *label;
2059 VALUE labels = rb_ary_hidden_new(1);
2060 VALUE *opt_table;
2061 int i = 0, j;
2062
2063 while (node) {
2064 label = NEW_LABEL(nd_line(RNODE(node)));
2065 rb_ary_push(labels, (VALUE)label | 1);
2066 ADD_LABEL(optargs, label);
2067 NO_CHECK(COMPILE_POPPED(optargs, "optarg", node->nd_body));
2068 node = node->nd_next;
2069 i += 1;
2070 }
2071
2072 /* last label */
2073 label = NEW_LABEL(nd_line(node_args));
2074 rb_ary_push(labels, (VALUE)label | 1);
2075 ADD_LABEL(optargs, label);
2076
2077 opt_table = ALLOC_N(VALUE, i+1);
2078
2079 MEMCPY(opt_table, RARRAY_CONST_PTR(labels), VALUE, i+1);
2080 for (j = 0; j < i+1; j++) {
2081 opt_table[j] &= ~1;
2082 }
2083 rb_ary_clear(labels);
2084
2085 body->param.flags.has_opt = TRUE;
2086 body->param.opt_num = i;
2087 body->param.opt_table = opt_table;
2088 arg_size += i;
2089 }
2090
2091 if (rest_id) {
2092 body->param.rest_start = arg_size++;
2093 body->param.flags.has_rest = TRUE;
2094 if (rest_id == '*') body->param.flags.anon_rest = TRUE;
2095 RUBY_ASSERT(body->param.rest_start != -1);
2096 }
2097
2098 if (args->first_post_arg) {
2099 body->param.post_start = arg_size;
2100 body->param.post_num = args->post_args_num;
2101 body->param.flags.has_post = TRUE;
2102 arg_size += args->post_args_num;
2103
2104 if (body->param.flags.has_rest) { /* TODO: why that? */
2105 body->param.post_start = body->param.rest_start + 1;
2106 }
2107 }
2108
2109 if (args->kw_args) {
2110 arg_size = iseq_set_arguments_keywords(iseq, optargs, args, arg_size);
2111 }
2112 else if (args->kw_rest_arg && !optimized_forward) {
2113 ID kw_id = ISEQ_BODY(iseq)->local_table[arg_size];
2114 struct rb_iseq_param_keyword *keyword = ZALLOC_N(struct rb_iseq_param_keyword, 1);
2115 keyword->rest_start = arg_size++;
2116 body->param.keyword = keyword;
2117 body->param.flags.has_kwrest = TRUE;
2118
2119 static ID anon_kwrest = 0;
2120 if (!anon_kwrest) anon_kwrest = rb_intern("**");
2121 if (kw_id == anon_kwrest) body->param.flags.anon_kwrest = TRUE;
2122 }
2123 else if (args->no_kwarg) {
2124 body->param.flags.accepts_no_kwarg = TRUE;
2125 }
2126
2127 if (block_id) {
2128 body->param.block_start = arg_size++;
2129 body->param.flags.has_block = TRUE;
2130 iseq_set_use_block(iseq);
2131 }
2132
2133 // Only optimize specifically methods like this: `foo(...)`
2134 if (optimized_forward) {
2135 body->param.flags.use_block = 1;
2136 body->param.flags.forwardable = TRUE;
2137 arg_size = 1;
2138 }
2139
2140 iseq_calc_param_size(iseq);
2141 body->param.size = arg_size;
2142
2143 if (args->pre_init) { /* m_init */
2144 NO_CHECK(COMPILE_POPPED(optargs, "init arguments (m)", args->pre_init));
2145 }
2146 if (args->post_init) { /* p_init */
2147 NO_CHECK(COMPILE_POPPED(optargs, "init arguments (p)", args->post_init));
2148 }
2149
2150 if (body->type == ISEQ_TYPE_BLOCK) {
2151 if (body->param.flags.has_opt == FALSE &&
2152 body->param.flags.has_post == FALSE &&
2153 body->param.flags.has_rest == FALSE &&
2154 body->param.flags.has_kw == FALSE &&
2155 body->param.flags.has_kwrest == FALSE) {
2156
2157 if (body->param.lead_num == 1 && last_comma == 0) {
2158 /* {|a|} */
2159 body->param.flags.ambiguous_param0 = TRUE;
2160 }
2161 }
2162 }
2163 }
2164
2165 return COMPILE_OK;
2166}
2167
2168static int
2169iseq_set_local_table(rb_iseq_t *iseq, const rb_ast_id_table_t *tbl, const NODE *const node_args)
2170{
2171 unsigned int size = tbl ? tbl->size : 0;
2172 unsigned int offset = 0;
2173
2174 if (node_args) {
2175 struct rb_args_info *args = &RNODE_ARGS(node_args)->nd_ainfo;
2176
2177 // If we have a function that only has `...` as the parameter,
2178 // then its local table should only be `...`
2179 // FIXME: I think this should be fixed in the AST rather than special case here.
2180 if (args->forwarding && args->pre_args_num == 0 && !args->opt_args) {
2181 size -= 3;
2182 offset += 3;
2183 }
2184 }
2185
2186 if (size > 0) {
2187 ID *ids = (ID *)ALLOC_N(ID, size);
2188 MEMCPY(ids, tbl->ids + offset, ID, size);
2189 ISEQ_BODY(iseq)->local_table = ids;
2190 }
2191 ISEQ_BODY(iseq)->local_table_size = size;
2192
2193 debugs("iseq_set_local_table: %u\n", ISEQ_BODY(iseq)->local_table_size);
2194 return COMPILE_OK;
2195}
2196
2197int
2198rb_iseq_cdhash_cmp(VALUE val, VALUE lit)
2199{
2200 int tval, tlit;
2201
2202 if (val == lit) {
2203 return 0;
2204 }
2205 else if ((tlit = OBJ_BUILTIN_TYPE(lit)) == -1) {
2206 return val != lit;
2207 }
2208 else if ((tval = OBJ_BUILTIN_TYPE(val)) == -1) {
2209 return -1;
2210 }
2211 else if (tlit != tval) {
2212 return -1;
2213 }
2214 else if (tlit == T_SYMBOL) {
2215 return val != lit;
2216 }
2217 else if (tlit == T_STRING) {
2218 return rb_str_hash_cmp(lit, val);
2219 }
2220 else if (tlit == T_BIGNUM) {
2221 long x = FIX2LONG(rb_big_cmp(lit, val));
2222
2223 /* Given lit and val are both Bignum, x must be -1, 0, 1.
2224 * There is no need to call rb_fix2int here. */
2225 RUBY_ASSERT((x == 1) || (x == 0) || (x == -1));
2226 return (int)x;
2227 }
2228 else if (tlit == T_FLOAT) {
2229 return rb_float_cmp(lit, val);
2230 }
2231 else if (tlit == T_RATIONAL) {
2232 const struct RRational *rat1 = RRATIONAL(val);
2233 const struct RRational *rat2 = RRATIONAL(lit);
2234 return rb_iseq_cdhash_cmp(rat1->num, rat2->num) || rb_iseq_cdhash_cmp(rat1->den, rat2->den);
2235 }
2236 else if (tlit == T_COMPLEX) {
2237 const struct RComplex *comp1 = RCOMPLEX(val);
2238 const struct RComplex *comp2 = RCOMPLEX(lit);
2239 return rb_iseq_cdhash_cmp(comp1->real, comp2->real) || rb_iseq_cdhash_cmp(comp1->imag, comp2->imag);
2240 }
2241 else if (tlit == T_REGEXP) {
2242 return rb_reg_equal(val, lit) ? 0 : -1;
2243 }
2244 else {
2246 }
2247}
2248
2249st_index_t
2250rb_iseq_cdhash_hash(VALUE a)
2251{
2252 switch (OBJ_BUILTIN_TYPE(a)) {
2253 case -1:
2254 case T_SYMBOL:
2255 return (st_index_t)a;
2256 case T_STRING:
2257 return rb_str_hash(a);
2258 case T_BIGNUM:
2259 return FIX2LONG(rb_big_hash(a));
2260 case T_FLOAT:
2261 return rb_dbl_long_hash(RFLOAT_VALUE(a));
2262 case T_RATIONAL:
2263 return rb_rational_hash(a);
2264 case T_COMPLEX:
2265 return rb_complex_hash(a);
2266 case T_REGEXP:
2267 return NUM2LONG(rb_reg_hash(a));
2268 default:
2270 }
2271}
2272
2273static const struct st_hash_type cdhash_type = {
2274 rb_iseq_cdhash_cmp,
2275 rb_iseq_cdhash_hash,
2276};
2277
2279 VALUE hash;
2280 int pos;
2281 int len;
2282};
2283
2284static int
2285cdhash_set_label_i(VALUE key, VALUE val, VALUE ptr)
2286{
2287 struct cdhash_set_label_struct *data = (struct cdhash_set_label_struct *)ptr;
2288 LABEL *lobj = (LABEL *)(val & ~1);
2289 rb_hash_aset(data->hash, key, INT2FIX(lobj->position - (data->pos+data->len)));
2290 return ST_CONTINUE;
2291}
2292
2293
2294static inline VALUE
2295get_ivar_ic_value(rb_iseq_t *iseq,ID id)
2296{
2297 return INT2FIX(ISEQ_BODY(iseq)->ivc_size++);
2298}
2299
2300static inline VALUE
2301get_cvar_ic_value(rb_iseq_t *iseq,ID id)
2302{
2303 VALUE val;
2304 struct rb_id_table *tbl = ISEQ_COMPILE_DATA(iseq)->ivar_cache_table;
2305 if (tbl) {
2306 if (rb_id_table_lookup(tbl,id,&val)) {
2307 return val;
2308 }
2309 }
2310 else {
2311 tbl = rb_id_table_create(1);
2312 ISEQ_COMPILE_DATA(iseq)->ivar_cache_table = tbl;
2313 }
2314 val = INT2FIX(ISEQ_BODY(iseq)->icvarc_size++);
2315 rb_id_table_insert(tbl,id,val);
2316 return val;
2317}
2318
2319#define BADINSN_DUMP(anchor, list, dest) \
2320 dump_disasm_list_with_cursor(FIRST_ELEMENT(anchor), list, dest)
2321
2322#define BADINSN_ERROR \
2323 (xfree(generated_iseq), \
2324 xfree(insns_info), \
2325 BADINSN_DUMP(anchor, list, NULL), \
2326 COMPILE_ERROR)
2327
2328static int
2329fix_sp_depth(rb_iseq_t *iseq, LINK_ANCHOR *const anchor)
2330{
2331 int stack_max = 0, sp = 0, line = 0;
2332 LINK_ELEMENT *list;
2333
2334 for (list = FIRST_ELEMENT(anchor); list; list = list->next) {
2335 if (IS_LABEL(list)) {
2336 LABEL *lobj = (LABEL *)list;
2337 lobj->set = TRUE;
2338 }
2339 }
2340
2341 for (list = FIRST_ELEMENT(anchor); list; list = list->next) {
2342 switch (list->type) {
2343 case ISEQ_ELEMENT_INSN:
2344 {
2345 int j, len, insn;
2346 const char *types;
2347 VALUE *operands;
2348 INSN *iobj = (INSN *)list;
2349
2350 /* update sp */
2351 sp = calc_sp_depth(sp, iobj);
2352 if (sp < 0) {
2353 BADINSN_DUMP(anchor, list, NULL);
2354 COMPILE_ERROR(iseq, iobj->insn_info.line_no,
2355 "argument stack underflow (%d)", sp);
2356 return -1;
2357 }
2358 if (sp > stack_max) {
2359 stack_max = sp;
2360 }
2361
2362 line = iobj->insn_info.line_no;
2363 /* fprintf(stderr, "insn: %-16s, sp: %d\n", insn_name(iobj->insn_id), sp); */
2364 operands = iobj->operands;
2365 insn = iobj->insn_id;
2366 types = insn_op_types(insn);
2367 len = insn_len(insn);
2368
2369 /* operand check */
2370 if (iobj->operand_size != len - 1) {
2371 /* printf("operand size miss! (%d, %d)\n", iobj->operand_size, len); */
2372 BADINSN_DUMP(anchor, list, NULL);
2373 COMPILE_ERROR(iseq, iobj->insn_info.line_no,
2374 "operand size miss! (%d for %d)",
2375 iobj->operand_size, len - 1);
2376 return -1;
2377 }
2378
2379 for (j = 0; types[j]; j++) {
2380 if (types[j] == TS_OFFSET) {
2381 /* label(destination position) */
2382 LABEL *lobj = (LABEL *)operands[j];
2383 if (!lobj->set) {
2384 BADINSN_DUMP(anchor, list, NULL);
2385 COMPILE_ERROR(iseq, iobj->insn_info.line_no,
2386 "unknown label: "LABEL_FORMAT, lobj->label_no);
2387 return -1;
2388 }
2389 if (lobj->sp == -1) {
2390 lobj->sp = sp;
2391 }
2392 else if (lobj->sp != sp) {
2393 debugs("%s:%d: sp inconsistency found but ignored (" LABEL_FORMAT " sp: %d, calculated sp: %d)\n",
2394 RSTRING_PTR(rb_iseq_path(iseq)), line,
2395 lobj->label_no, lobj->sp, sp);
2396 }
2397 }
2398 }
2399 break;
2400 }
2401 case ISEQ_ELEMENT_LABEL:
2402 {
2403 LABEL *lobj = (LABEL *)list;
2404 if (lobj->sp == -1) {
2405 lobj->sp = sp;
2406 }
2407 else {
2408 if (lobj->sp != sp) {
2409 debugs("%s:%d: sp inconsistency found but ignored (" LABEL_FORMAT " sp: %d, calculated sp: %d)\n",
2410 RSTRING_PTR(rb_iseq_path(iseq)), line,
2411 lobj->label_no, lobj->sp, sp);
2412 }
2413 sp = lobj->sp;
2414 }
2415 break;
2416 }
2417 case ISEQ_ELEMENT_TRACE:
2418 {
2419 /* ignore */
2420 break;
2421 }
2422 case ISEQ_ELEMENT_ADJUST:
2423 {
2424 ADJUST *adjust = (ADJUST *)list;
2425 int orig_sp = sp;
2426
2427 sp = adjust->label ? adjust->label->sp : 0;
2428 if (adjust->line_no != -1 && orig_sp - sp < 0) {
2429 BADINSN_DUMP(anchor, list, NULL);
2430 COMPILE_ERROR(iseq, adjust->line_no,
2431 "iseq_set_sequence: adjust bug %d < %d",
2432 orig_sp, sp);
2433 return -1;
2434 }
2435 break;
2436 }
2437 default:
2438 BADINSN_DUMP(anchor, list, NULL);
2439 COMPILE_ERROR(iseq, line, "unknown list type: %d", list->type);
2440 return -1;
2441 }
2442 }
2443 return stack_max;
2444}
2445
2446static int
2447add_insn_info(struct iseq_insn_info_entry *insns_info, unsigned int *positions,
2448 int insns_info_index, int code_index, const INSN *iobj)
2449{
2450 if (insns_info_index == 0 ||
2451 insns_info[insns_info_index-1].line_no != iobj->insn_info.line_no ||
2452#ifdef USE_ISEQ_NODE_ID
2453 insns_info[insns_info_index-1].node_id != iobj->insn_info.node_id ||
2454#endif
2455 insns_info[insns_info_index-1].events != iobj->insn_info.events) {
2456 insns_info[insns_info_index].line_no = iobj->insn_info.line_no;
2457#ifdef USE_ISEQ_NODE_ID
2458 insns_info[insns_info_index].node_id = iobj->insn_info.node_id;
2459#endif
2460 insns_info[insns_info_index].events = iobj->insn_info.events;
2461 positions[insns_info_index] = code_index;
2462 return TRUE;
2463 }
2464 return FALSE;
2465}
2466
2467static int
2468add_adjust_info(struct iseq_insn_info_entry *insns_info, unsigned int *positions,
2469 int insns_info_index, int code_index, const ADJUST *adjust)
2470{
2471 insns_info[insns_info_index].line_no = adjust->line_no;
2472 insns_info[insns_info_index].node_id = -1;
2473 insns_info[insns_info_index].events = 0;
2474 positions[insns_info_index] = code_index;
2475 return TRUE;
2476}
2477
2478static ID *
2479array_to_idlist(VALUE arr)
2480{
2482 long size = RARRAY_LEN(arr);
2483 ID *ids = (ID *)ALLOC_N(ID, size + 1);
2484 for (int i = 0; i < size; i++) {
2485 VALUE sym = RARRAY_AREF(arr, i);
2486 ids[i] = SYM2ID(sym);
2487 }
2488 ids[size] = 0;
2489 return ids;
2490}
2491
2492static VALUE
2493idlist_to_array(const ID *ids)
2494{
2495 VALUE arr = rb_ary_new();
2496 while (*ids) {
2497 rb_ary_push(arr, ID2SYM(*ids++));
2498 }
2499 return arr;
2500}
2501
2505static int
2506iseq_set_sequence(rb_iseq_t *iseq, LINK_ANCHOR *const anchor)
2507{
2508 struct iseq_insn_info_entry *insns_info;
2509 struct rb_iseq_constant_body *const body = ISEQ_BODY(iseq);
2510 unsigned int *positions;
2511 LINK_ELEMENT *list;
2512 VALUE *generated_iseq;
2513 rb_event_flag_t events = 0;
2514 long data = 0;
2515
2516 int insn_num, code_index, insns_info_index, sp = 0;
2517 int stack_max = fix_sp_depth(iseq, anchor);
2518
2519 if (stack_max < 0) return COMPILE_NG;
2520
2521 /* fix label position */
2522 insn_num = code_index = 0;
2523 for (list = FIRST_ELEMENT(anchor); list; list = list->next) {
2524 switch (list->type) {
2525 case ISEQ_ELEMENT_INSN:
2526 {
2527 INSN *iobj = (INSN *)list;
2528 /* update sp */
2529 sp = calc_sp_depth(sp, iobj);
2530 insn_num++;
2531 events = iobj->insn_info.events |= events;
2532 if (ISEQ_COVERAGE(iseq)) {
2533 if (ISEQ_LINE_COVERAGE(iseq) && (events & RUBY_EVENT_COVERAGE_LINE) &&
2534 !(rb_get_coverage_mode() & COVERAGE_TARGET_ONESHOT_LINES)) {
2535 int line = iobj->insn_info.line_no - 1;
2536 if (line >= 0 && line < RARRAY_LEN(ISEQ_LINE_COVERAGE(iseq))) {
2537 RARRAY_ASET(ISEQ_LINE_COVERAGE(iseq), line, INT2FIX(0));
2538 }
2539 }
2540 if (ISEQ_BRANCH_COVERAGE(iseq) && (events & RUBY_EVENT_COVERAGE_BRANCH)) {
2541 while (RARRAY_LEN(ISEQ_PC2BRANCHINDEX(iseq)) <= code_index) {
2542 rb_ary_push(ISEQ_PC2BRANCHINDEX(iseq), Qnil);
2543 }
2544 RARRAY_ASET(ISEQ_PC2BRANCHINDEX(iseq), code_index, INT2FIX(data));
2545 }
2546 }
2547 code_index += insn_data_length(iobj);
2548 events = 0;
2549 data = 0;
2550 break;
2551 }
2552 case ISEQ_ELEMENT_LABEL:
2553 {
2554 LABEL *lobj = (LABEL *)list;
2555 lobj->position = code_index;
2556 if (lobj->sp != sp) {
2557 debugs("%s: sp inconsistency found but ignored (" LABEL_FORMAT " sp: %d, calculated sp: %d)\n",
2558 RSTRING_PTR(rb_iseq_path(iseq)),
2559 lobj->label_no, lobj->sp, sp);
2560 }
2561 sp = lobj->sp;
2562 break;
2563 }
2564 case ISEQ_ELEMENT_TRACE:
2565 {
2566 TRACE *trace = (TRACE *)list;
2567 events |= trace->event;
2568 if (trace->event & RUBY_EVENT_COVERAGE_BRANCH) data = trace->data;
2569 break;
2570 }
2571 case ISEQ_ELEMENT_ADJUST:
2572 {
2573 ADJUST *adjust = (ADJUST *)list;
2574 if (adjust->line_no != -1) {
2575 int orig_sp = sp;
2576 sp = adjust->label ? adjust->label->sp : 0;
2577 if (orig_sp - sp > 0) {
2578 if (orig_sp - sp > 1) code_index++; /* 1 operand */
2579 code_index++; /* insn */
2580 insn_num++;
2581 }
2582 }
2583 break;
2584 }
2585 default: break;
2586 }
2587 }
2588
2589 /* make instruction sequence */
2590 generated_iseq = ALLOC_N(VALUE, code_index);
2591 insns_info = ALLOC_N(struct iseq_insn_info_entry, insn_num);
2592 positions = ALLOC_N(unsigned int, insn_num);
2593 if (ISEQ_IS_SIZE(body)) {
2594 body->is_entries = ZALLOC_N(union iseq_inline_storage_entry, ISEQ_IS_SIZE(body));
2595 }
2596 else {
2597 body->is_entries = NULL;
2598 }
2599 body->call_data = ZALLOC_N(struct rb_call_data, body->ci_size);
2600 ISEQ_COMPILE_DATA(iseq)->ci_index = 0;
2601
2602 // Calculate the bitmask buffer size.
2603 // Round the generated_iseq size up to the nearest multiple
2604 // of the number of bits in an unsigned long.
2605
2606 // Allocate enough room for the bitmask list
2607 iseq_bits_t * mark_offset_bits;
2608 int code_size = code_index;
2609
2610 bool needs_bitmap = false;
2611
2612 if (ISEQ_MBITS_BUFLEN(code_index) == 1) {
2613 mark_offset_bits = &ISEQ_COMPILE_DATA(iseq)->mark_bits.single;
2614 ISEQ_COMPILE_DATA(iseq)->is_single_mark_bit = true;
2615 }
2616 else {
2617 mark_offset_bits = ZALLOC_N(iseq_bits_t, ISEQ_MBITS_BUFLEN(code_index));
2618 ISEQ_COMPILE_DATA(iseq)->mark_bits.list = mark_offset_bits;
2619 ISEQ_COMPILE_DATA(iseq)->is_single_mark_bit = false;
2620 }
2621
2622 ISEQ_COMPILE_DATA(iseq)->iseq_encoded = (void *)generated_iseq;
2623 ISEQ_COMPILE_DATA(iseq)->iseq_size = code_index;
2624
2625 list = FIRST_ELEMENT(anchor);
2626 insns_info_index = code_index = sp = 0;
2627
2628 while (list) {
2629 switch (list->type) {
2630 case ISEQ_ELEMENT_INSN:
2631 {
2632 int j, len, insn;
2633 const char *types;
2634 VALUE *operands;
2635 INSN *iobj = (INSN *)list;
2636
2637 /* update sp */
2638 sp = calc_sp_depth(sp, iobj);
2639 /* fprintf(stderr, "insn: %-16s, sp: %d\n", insn_name(iobj->insn_id), sp); */
2640 operands = iobj->operands;
2641 insn = iobj->insn_id;
2642 generated_iseq[code_index] = insn;
2643 types = insn_op_types(insn);
2644 len = insn_len(insn);
2645
2646 for (j = 0; types[j]; j++) {
2647 char type = types[j];
2648
2649 /* printf("--> [%c - (%d-%d)]\n", type, k, j); */
2650 switch (type) {
2651 case TS_OFFSET:
2652 {
2653 /* label(destination position) */
2654 LABEL *lobj = (LABEL *)operands[j];
2655 generated_iseq[code_index + 1 + j] = lobj->position - (code_index + len);
2656 break;
2657 }
2658 case TS_CDHASH:
2659 {
2660 VALUE map = operands[j];
2661 struct cdhash_set_label_struct data;
2662 data.hash = map;
2663 data.pos = code_index;
2664 data.len = len;
2665 rb_hash_foreach(map, cdhash_set_label_i, (VALUE)&data);
2666
2667 rb_hash_rehash(map);
2668 freeze_hide_obj(map);
2669 generated_iseq[code_index + 1 + j] = map;
2670 ISEQ_MBITS_SET(mark_offset_bits, code_index + 1 + j);
2671 RB_OBJ_WRITTEN(iseq, Qundef, map);
2672 needs_bitmap = true;
2673 break;
2674 }
2675 case TS_LINDEX:
2676 case TS_NUM: /* ulong */
2677 generated_iseq[code_index + 1 + j] = FIX2INT(operands[j]);
2678 break;
2679 case TS_ISEQ: /* iseq */
2680 case TS_VALUE: /* VALUE */
2681 {
2682 VALUE v = operands[j];
2683 generated_iseq[code_index + 1 + j] = v;
2684 /* to mark ruby object */
2685 if (!SPECIAL_CONST_P(v)) {
2686 RB_OBJ_WRITTEN(iseq, Qundef, v);
2687 ISEQ_MBITS_SET(mark_offset_bits, code_index + 1 + j);
2688 needs_bitmap = true;
2689 }
2690 break;
2691 }
2692 /* [ TS_IVC | TS_ICVARC | TS_ISE | TS_IC ] */
2693 case TS_IC: /* inline cache: constants */
2694 {
2695 unsigned int ic_index = ISEQ_COMPILE_DATA(iseq)->ic_index++;
2696 IC ic = &ISEQ_IS_ENTRY_START(body, type)[ic_index].ic_cache;
2697 if (UNLIKELY(ic_index >= body->ic_size)) {
2698 BADINSN_DUMP(anchor, &iobj->link, 0);
2699 COMPILE_ERROR(iseq, iobj->insn_info.line_no,
2700 "iseq_set_sequence: ic_index overflow: index: %d, size: %d",
2701 ic_index, ISEQ_IS_SIZE(body));
2702 }
2703
2704 ic->segments = array_to_idlist(operands[j]);
2705
2706 generated_iseq[code_index + 1 + j] = (VALUE)ic;
2707 }
2708 break;
2709 case TS_IVC: /* inline ivar cache */
2710 {
2711 unsigned int ic_index = FIX2UINT(operands[j]);
2712
2713 IVC cache = ((IVC)&body->is_entries[ic_index]);
2714
2715 if (insn == BIN(setinstancevariable)) {
2716 cache->iv_set_name = SYM2ID(operands[j - 1]);
2717 }
2718 else {
2719 cache->iv_set_name = 0;
2720 }
2721
2722 vm_ic_attr_index_initialize(cache, INVALID_SHAPE_ID);
2723 }
2724 case TS_ISE: /* inline storage entry: `once` insn */
2725 case TS_ICVARC: /* inline cvar cache */
2726 {
2727 unsigned int ic_index = FIX2UINT(operands[j]);
2728 IC ic = &ISEQ_IS_ENTRY_START(body, type)[ic_index].ic_cache;
2729 if (UNLIKELY(ic_index >= ISEQ_IS_SIZE(body))) {
2730 BADINSN_DUMP(anchor, &iobj->link, 0);
2731 COMPILE_ERROR(iseq, iobj->insn_info.line_no,
2732 "iseq_set_sequence: ic_index overflow: index: %d, size: %d",
2733 ic_index, ISEQ_IS_SIZE(body));
2734 }
2735 generated_iseq[code_index + 1 + j] = (VALUE)ic;
2736
2737 break;
2738 }
2739 case TS_CALLDATA:
2740 {
2741 const struct rb_callinfo *source_ci = (const struct rb_callinfo *)operands[j];
2742 RUBY_ASSERT(ISEQ_COMPILE_DATA(iseq)->ci_index <= body->ci_size);
2743 struct rb_call_data *cd = &body->call_data[ISEQ_COMPILE_DATA(iseq)->ci_index++];
2744 cd->ci = source_ci;
2745 cd->cc = vm_cc_empty();
2746 generated_iseq[code_index + 1 + j] = (VALUE)cd;
2747 break;
2748 }
2749 case TS_ID: /* ID */
2750 generated_iseq[code_index + 1 + j] = SYM2ID(operands[j]);
2751 break;
2752 case TS_FUNCPTR:
2753 generated_iseq[code_index + 1 + j] = operands[j];
2754 break;
2755 case TS_BUILTIN:
2756 generated_iseq[code_index + 1 + j] = operands[j];
2757 break;
2758 default:
2759 BADINSN_ERROR(iseq, iobj->insn_info.line_no,
2760 "unknown operand type: %c", type);
2761 return COMPILE_NG;
2762 }
2763 }
2764 if (add_insn_info(insns_info, positions, insns_info_index, code_index, iobj)) insns_info_index++;
2765 code_index += len;
2766 break;
2767 }
2768 case ISEQ_ELEMENT_LABEL:
2769 {
2770 LABEL *lobj = (LABEL *)list;
2771 if (lobj->sp != sp) {
2772 debugs("%s: sp inconsistency found but ignored (" LABEL_FORMAT " sp: %d, calculated sp: %d)\n",
2773 RSTRING_PTR(rb_iseq_path(iseq)),
2774 lobj->label_no, lobj->sp, sp);
2775 }
2776 sp = lobj->sp;
2777 break;
2778 }
2779 case ISEQ_ELEMENT_ADJUST:
2780 {
2781 ADJUST *adjust = (ADJUST *)list;
2782 int orig_sp = sp;
2783
2784 if (adjust->label) {
2785 sp = adjust->label->sp;
2786 }
2787 else {
2788 sp = 0;
2789 }
2790
2791 if (adjust->line_no != -1) {
2792 const int diff = orig_sp - sp;
2793 if (diff > 0) {
2794 if (insns_info_index == 0) {
2795 COMPILE_ERROR(iseq, adjust->line_no,
2796 "iseq_set_sequence: adjust bug (ISEQ_ELEMENT_ADJUST must not be the first in iseq)");
2797 }
2798 if (add_adjust_info(insns_info, positions, insns_info_index, code_index, adjust)) insns_info_index++;
2799 }
2800 if (diff > 1) {
2801 generated_iseq[code_index++] = BIN(adjuststack);
2802 generated_iseq[code_index++] = orig_sp - sp;
2803 }
2804 else if (diff == 1) {
2805 generated_iseq[code_index++] = BIN(pop);
2806 }
2807 else if (diff < 0) {
2808 int label_no = adjust->label ? adjust->label->label_no : -1;
2809 xfree(generated_iseq);
2810 xfree(insns_info);
2811 xfree(positions);
2812 if (ISEQ_MBITS_BUFLEN(code_size) > 1) {
2813 xfree(mark_offset_bits);
2814 }
2815 debug_list(anchor, list);
2816 COMPILE_ERROR(iseq, adjust->line_no,
2817 "iseq_set_sequence: adjust bug to %d %d < %d",
2818 label_no, orig_sp, sp);
2819 return COMPILE_NG;
2820 }
2821 }
2822 break;
2823 }
2824 default:
2825 /* ignore */
2826 break;
2827 }
2828 list = list->next;
2829 }
2830
2831 body->iseq_encoded = (void *)generated_iseq;
2832 body->iseq_size = code_index;
2833 body->stack_max = stack_max;
2834
2835 if (ISEQ_COMPILE_DATA(iseq)->is_single_mark_bit) {
2836 body->mark_bits.single = ISEQ_COMPILE_DATA(iseq)->mark_bits.single;
2837 }
2838 else {
2839 if (needs_bitmap) {
2840 body->mark_bits.list = mark_offset_bits;
2841 }
2842 else {
2843 body->mark_bits.list = NULL;
2844 ISEQ_COMPILE_DATA(iseq)->mark_bits.list = NULL;
2845 ruby_xfree(mark_offset_bits);
2846 }
2847 }
2848
2849 /* get rid of memory leak when REALLOC failed */
2850 body->insns_info.body = insns_info;
2851 body->insns_info.positions = positions;
2852
2853 REALLOC_N(insns_info, struct iseq_insn_info_entry, insns_info_index);
2854 body->insns_info.body = insns_info;
2855 REALLOC_N(positions, unsigned int, insns_info_index);
2856 body->insns_info.positions = positions;
2857 body->insns_info.size = insns_info_index;
2858
2859 return COMPILE_OK;
2860}
2861
2862static int
2863label_get_position(LABEL *lobj)
2864{
2865 return lobj->position;
2866}
2867
2868static int
2869label_get_sp(LABEL *lobj)
2870{
2871 return lobj->sp;
2872}
2873
2874static int
2875iseq_set_exception_table(rb_iseq_t *iseq)
2876{
2877 const VALUE *tptr, *ptr;
2878 unsigned int tlen, i;
2879 struct iseq_catch_table_entry *entry;
2880
2881 ISEQ_BODY(iseq)->catch_table = NULL;
2882
2883 VALUE catch_table_ary = ISEQ_COMPILE_DATA(iseq)->catch_table_ary;
2884 if (NIL_P(catch_table_ary)) return COMPILE_OK;
2885 tlen = (int)RARRAY_LEN(catch_table_ary);
2886 tptr = RARRAY_CONST_PTR(catch_table_ary);
2887
2888 if (tlen > 0) {
2889 struct iseq_catch_table *table = xmalloc(iseq_catch_table_bytes(tlen));
2890 table->size = tlen;
2891
2892 for (i = 0; i < table->size; i++) {
2893 int pos;
2894 ptr = RARRAY_CONST_PTR(tptr[i]);
2895 entry = UNALIGNED_MEMBER_PTR(table, entries[i]);
2896 entry->type = (enum rb_catch_type)(ptr[0] & 0xffff);
2897 pos = label_get_position((LABEL *)(ptr[1] & ~1));
2898 RUBY_ASSERT(pos >= 0);
2899 entry->start = (unsigned int)pos;
2900 pos = label_get_position((LABEL *)(ptr[2] & ~1));
2901 RUBY_ASSERT(pos >= 0);
2902 entry->end = (unsigned int)pos;
2903 entry->iseq = (rb_iseq_t *)ptr[3];
2904 RB_OBJ_WRITTEN(iseq, Qundef, entry->iseq);
2905
2906 /* stack depth */
2907 if (ptr[4]) {
2908 LABEL *lobj = (LABEL *)(ptr[4] & ~1);
2909 entry->cont = label_get_position(lobj);
2910 entry->sp = label_get_sp(lobj);
2911
2912 /* TODO: Dirty Hack! Fix me */
2913 if (entry->type == CATCH_TYPE_RESCUE ||
2914 entry->type == CATCH_TYPE_BREAK ||
2915 entry->type == CATCH_TYPE_NEXT) {
2916 RUBY_ASSERT(entry->sp > 0);
2917 entry->sp--;
2918 }
2919 }
2920 else {
2921 entry->cont = 0;
2922 }
2923 }
2924 ISEQ_BODY(iseq)->catch_table = table;
2925 RB_OBJ_WRITE(iseq, &ISEQ_COMPILE_DATA(iseq)->catch_table_ary, 0); /* free */
2926 }
2927
2928 RB_GC_GUARD(catch_table_ary);
2929
2930 return COMPILE_OK;
2931}
2932
2933/*
2934 * set optional argument table
2935 * def foo(a, b=expr1, c=expr2)
2936 * =>
2937 * b:
2938 * expr1
2939 * c:
2940 * expr2
2941 */
2942static int
2943iseq_set_optargs_table(rb_iseq_t *iseq)
2944{
2945 int i;
2946 VALUE *opt_table = (VALUE *)ISEQ_BODY(iseq)->param.opt_table;
2947
2948 if (ISEQ_BODY(iseq)->param.flags.has_opt) {
2949 for (i = 0; i < ISEQ_BODY(iseq)->param.opt_num + 1; i++) {
2950 opt_table[i] = label_get_position((LABEL *)opt_table[i]);
2951 }
2952 }
2953 return COMPILE_OK;
2954}
2955
2956static LINK_ELEMENT *
2957get_destination_insn(INSN *iobj)
2958{
2959 LABEL *lobj = (LABEL *)OPERAND_AT(iobj, 0);
2960 LINK_ELEMENT *list;
2961 rb_event_flag_t events = 0;
2962
2963 list = lobj->link.next;
2964 while (list) {
2965 switch (list->type) {
2966 case ISEQ_ELEMENT_INSN:
2967 case ISEQ_ELEMENT_ADJUST:
2968 goto found;
2969 case ISEQ_ELEMENT_LABEL:
2970 /* ignore */
2971 break;
2972 case ISEQ_ELEMENT_TRACE:
2973 {
2974 TRACE *trace = (TRACE *)list;
2975 events |= trace->event;
2976 }
2977 break;
2978 default: break;
2979 }
2980 list = list->next;
2981 }
2982 found:
2983 if (list && IS_INSN(list)) {
2984 INSN *iobj = (INSN *)list;
2985 iobj->insn_info.events |= events;
2986 }
2987 return list;
2988}
2989
2990static LINK_ELEMENT *
2991get_next_insn(INSN *iobj)
2992{
2993 LINK_ELEMENT *list = iobj->link.next;
2994
2995 while (list) {
2996 if (IS_INSN(list) || IS_ADJUST(list)) {
2997 return list;
2998 }
2999 list = list->next;
3000 }
3001 return 0;
3002}
3003
3004static LINK_ELEMENT *
3005get_prev_insn(INSN *iobj)
3006{
3007 LINK_ELEMENT *list = iobj->link.prev;
3008
3009 while (list) {
3010 if (IS_INSN(list) || IS_ADJUST(list)) {
3011 return list;
3012 }
3013 list = list->prev;
3014 }
3015 return 0;
3016}
3017
3018static void
3019unref_destination(INSN *iobj, int pos)
3020{
3021 LABEL *lobj = (LABEL *)OPERAND_AT(iobj, pos);
3022 --lobj->refcnt;
3023 if (!lobj->refcnt) ELEM_REMOVE(&lobj->link);
3024}
3025
3026static bool
3027replace_destination(INSN *dobj, INSN *nobj)
3028{
3029 VALUE n = OPERAND_AT(nobj, 0);
3030 LABEL *dl = (LABEL *)OPERAND_AT(dobj, 0);
3031 LABEL *nl = (LABEL *)n;
3032 if (dl == nl) return false;
3033 --dl->refcnt;
3034 ++nl->refcnt;
3035 OPERAND_AT(dobj, 0) = n;
3036 if (!dl->refcnt) ELEM_REMOVE(&dl->link);
3037 return true;
3038}
3039
3040static LABEL*
3041find_destination(INSN *i)
3042{
3043 int pos, len = insn_len(i->insn_id);
3044 for (pos = 0; pos < len; ++pos) {
3045 if (insn_op_types(i->insn_id)[pos] == TS_OFFSET) {
3046 return (LABEL *)OPERAND_AT(i, pos);
3047 }
3048 }
3049 return 0;
3050}
3051
3052static int
3053remove_unreachable_chunk(rb_iseq_t *iseq, LINK_ELEMENT *i)
3054{
3055 LINK_ELEMENT *first = i, *end;
3056 int *unref_counts = 0, nlabels = ISEQ_COMPILE_DATA(iseq)->label_no;
3057
3058 if (!i) return 0;
3059 unref_counts = ALLOCA_N(int, nlabels);
3060 MEMZERO(unref_counts, int, nlabels);
3061 end = i;
3062 do {
3063 LABEL *lab;
3064 if (IS_INSN(i)) {
3065 if (IS_INSN_ID(i, leave)) {
3066 end = i;
3067 break;
3068 }
3069 else if ((lab = find_destination((INSN *)i)) != 0) {
3070 unref_counts[lab->label_no]++;
3071 }
3072 }
3073 else if (IS_LABEL(i)) {
3074 lab = (LABEL *)i;
3075 if (lab->unremovable) return 0;
3076 if (lab->refcnt > unref_counts[lab->label_no]) {
3077 if (i == first) return 0;
3078 break;
3079 }
3080 continue;
3081 }
3082 else if (IS_TRACE(i)) {
3083 /* do nothing */
3084 }
3085 else if (IS_ADJUST(i)) {
3086 return 0;
3087 }
3088 end = i;
3089 } while ((i = i->next) != 0);
3090 i = first;
3091 do {
3092 if (IS_INSN(i)) {
3093 struct rb_iseq_constant_body *body = ISEQ_BODY(iseq);
3094 VALUE insn = INSN_OF(i);
3095 int pos, len = insn_len(insn);
3096 for (pos = 0; pos < len; ++pos) {
3097 switch (insn_op_types(insn)[pos]) {
3098 case TS_OFFSET:
3099 unref_destination((INSN *)i, pos);
3100 break;
3101 case TS_CALLDATA:
3102 --(body->ci_size);
3103 break;
3104 }
3105 }
3106 }
3107 ELEM_REMOVE(i);
3108 } while ((i != end) && (i = i->next) != 0);
3109 return 1;
3110}
3111
3112static int
3113iseq_pop_newarray(rb_iseq_t *iseq, INSN *iobj)
3114{
3115 switch (OPERAND_AT(iobj, 0)) {
3116 case INT2FIX(0): /* empty array */
3117 ELEM_REMOVE(&iobj->link);
3118 return TRUE;
3119 case INT2FIX(1): /* single element array */
3120 ELEM_REMOVE(&iobj->link);
3121 return FALSE;
3122 default:
3123 iobj->insn_id = BIN(adjuststack);
3124 return TRUE;
3125 }
3126}
3127
3128static int
3129is_frozen_putstring(INSN *insn, VALUE *op)
3130{
3131 if (IS_INSN_ID(insn, putstring) || IS_INSN_ID(insn, putchilledstring)) {
3132 *op = OPERAND_AT(insn, 0);
3133 return 1;
3134 }
3135 else if (IS_INSN_ID(insn, putobject)) { /* frozen_string_literal */
3136 *op = OPERAND_AT(insn, 0);
3137 return RB_TYPE_P(*op, T_STRING);
3138 }
3139 return 0;
3140}
3141
3142static int
3143optimize_checktype(rb_iseq_t *iseq, INSN *iobj)
3144{
3145 /*
3146 * putobject obj
3147 * dup
3148 * checktype T_XXX
3149 * branchif l1
3150 * l2:
3151 * ...
3152 * l1:
3153 *
3154 * => obj is a T_XXX
3155 *
3156 * putobject obj (T_XXX)
3157 * jump L1
3158 * L1:
3159 *
3160 * => obj is not a T_XXX
3161 *
3162 * putobject obj (T_XXX)
3163 * jump L2
3164 * L2:
3165 */
3166 int line, node_id;
3167 INSN *niobj, *ciobj, *dup = 0;
3168 LABEL *dest = 0;
3169 VALUE type;
3170
3171 switch (INSN_OF(iobj)) {
3172 case BIN(putstring):
3173 case BIN(putchilledstring):
3175 break;
3176 case BIN(putnil):
3177 type = INT2FIX(T_NIL);
3178 break;
3179 case BIN(putobject):
3180 type = INT2FIX(TYPE(OPERAND_AT(iobj, 0)));
3181 break;
3182 default: return FALSE;
3183 }
3184
3185 ciobj = (INSN *)get_next_insn(iobj);
3186 if (IS_INSN_ID(ciobj, jump)) {
3187 ciobj = (INSN *)get_next_insn((INSN*)OPERAND_AT(ciobj, 0));
3188 }
3189 if (IS_INSN_ID(ciobj, dup)) {
3190 ciobj = (INSN *)get_next_insn(dup = ciobj);
3191 }
3192 if (!ciobj || !IS_INSN_ID(ciobj, checktype)) return FALSE;
3193 niobj = (INSN *)get_next_insn(ciobj);
3194 if (!niobj) {
3195 /* TODO: putobject true/false */
3196 return FALSE;
3197 }
3198 switch (INSN_OF(niobj)) {
3199 case BIN(branchif):
3200 if (OPERAND_AT(ciobj, 0) == type) {
3201 dest = (LABEL *)OPERAND_AT(niobj, 0);
3202 }
3203 break;
3204 case BIN(branchunless):
3205 if (OPERAND_AT(ciobj, 0) != type) {
3206 dest = (LABEL *)OPERAND_AT(niobj, 0);
3207 }
3208 break;
3209 default:
3210 return FALSE;
3211 }
3212 line = ciobj->insn_info.line_no;
3213 node_id = ciobj->insn_info.node_id;
3214 if (!dest) {
3215 if (niobj->link.next && IS_LABEL(niobj->link.next)) {
3216 dest = (LABEL *)niobj->link.next; /* reuse label */
3217 }
3218 else {
3219 dest = NEW_LABEL(line);
3220 ELEM_INSERT_NEXT(&niobj->link, &dest->link);
3221 }
3222 }
3223 INSERT_AFTER_INSN1(iobj, line, node_id, jump, dest);
3224 LABEL_REF(dest);
3225 if (!dup) INSERT_AFTER_INSN(iobj, line, node_id, pop);
3226 return TRUE;
3227}
3228
3229static const struct rb_callinfo *
3230ci_flag_set(const rb_iseq_t *iseq, const struct rb_callinfo *ci, unsigned int add)
3231{
3232 const struct rb_callinfo *nci = vm_ci_new(vm_ci_mid(ci),
3233 vm_ci_flag(ci) | add,
3234 vm_ci_argc(ci),
3235 vm_ci_kwarg(ci));
3236 RB_OBJ_WRITTEN(iseq, ci, nci);
3237 return nci;
3238}
3239
3240static const struct rb_callinfo *
3241ci_argc_set(const rb_iseq_t *iseq, const struct rb_callinfo *ci, int argc)
3242{
3243 const struct rb_callinfo *nci = vm_ci_new(vm_ci_mid(ci),
3244 vm_ci_flag(ci),
3245 argc,
3246 vm_ci_kwarg(ci));
3247 RB_OBJ_WRITTEN(iseq, ci, nci);
3248 return nci;
3249}
3250
3251#define vm_ci_simple(ci) (vm_ci_flag(ci) & VM_CALL_ARGS_SIMPLE)
3252
3253static int
3254iseq_peephole_optimize(rb_iseq_t *iseq, LINK_ELEMENT *list, const int do_tailcallopt)
3255{
3256 INSN *const iobj = (INSN *)list;
3257
3258 again:
3259 optimize_checktype(iseq, iobj);
3260
3261 if (IS_INSN_ID(iobj, jump)) {
3262 INSN *niobj, *diobj, *piobj;
3263 diobj = (INSN *)get_destination_insn(iobj);
3264 niobj = (INSN *)get_next_insn(iobj);
3265
3266 if (diobj == niobj) {
3267 /*
3268 * jump LABEL
3269 * LABEL:
3270 * =>
3271 * LABEL:
3272 */
3273 unref_destination(iobj, 0);
3274 ELEM_REMOVE(&iobj->link);
3275 return COMPILE_OK;
3276 }
3277 else if (iobj != diobj && IS_INSN(&diobj->link) &&
3278 IS_INSN_ID(diobj, jump) &&
3279 OPERAND_AT(iobj, 0) != OPERAND_AT(diobj, 0) &&
3280 diobj->insn_info.events == 0) {
3281 /*
3282 * useless jump elimination:
3283 * jump LABEL1
3284 * ...
3285 * LABEL1:
3286 * jump LABEL2
3287 *
3288 * => in this case, first jump instruction should jump to
3289 * LABEL2 directly
3290 */
3291 if (replace_destination(iobj, diobj)) {
3292 remove_unreachable_chunk(iseq, iobj->link.next);
3293 goto again;
3294 }
3295 }
3296 else if (IS_INSN_ID(diobj, leave)) {
3297 /*
3298 * jump LABEL
3299 * ...
3300 * LABEL:
3301 * leave
3302 * =>
3303 * leave
3304 * ...
3305 * LABEL:
3306 * leave
3307 */
3308 /* replace */
3309 unref_destination(iobj, 0);
3310 iobj->insn_id = BIN(leave);
3311 iobj->operand_size = 0;
3312 iobj->insn_info = diobj->insn_info;
3313 goto again;
3314 }
3315 else if (IS_INSN(iobj->link.prev) &&
3316 (piobj = (INSN *)iobj->link.prev) &&
3317 (IS_INSN_ID(piobj, branchif) ||
3318 IS_INSN_ID(piobj, branchunless))) {
3319 INSN *pdiobj = (INSN *)get_destination_insn(piobj);
3320 if (niobj == pdiobj) {
3321 int refcnt = IS_LABEL(piobj->link.next) ?
3322 ((LABEL *)piobj->link.next)->refcnt : 0;
3323 /*
3324 * useless jump elimination (if/unless destination):
3325 * if L1
3326 * jump L2
3327 * L1:
3328 * ...
3329 * L2:
3330 *
3331 * ==>
3332 * unless L2
3333 * L1:
3334 * ...
3335 * L2:
3336 */
3337 piobj->insn_id = (IS_INSN_ID(piobj, branchif))
3338 ? BIN(branchunless) : BIN(branchif);
3339 if (replace_destination(piobj, iobj) && refcnt <= 1) {
3340 ELEM_REMOVE(&iobj->link);
3341 }
3342 else {
3343 /* TODO: replace other branch destinations too */
3344 }
3345 return COMPILE_OK;
3346 }
3347 else if (diobj == pdiobj) {
3348 /*
3349 * useless jump elimination (if/unless before jump):
3350 * L1:
3351 * ...
3352 * if L1
3353 * jump L1
3354 *
3355 * ==>
3356 * L1:
3357 * ...
3358 * pop
3359 * jump L1
3360 */
3361 INSN *popiobj = new_insn_core(iseq, iobj->insn_info.line_no, iobj->insn_info.node_id, BIN(pop), 0, 0);
3362 ELEM_REPLACE(&piobj->link, &popiobj->link);
3363 }
3364 }
3365 if (remove_unreachable_chunk(iseq, iobj->link.next)) {
3366 goto again;
3367 }
3368 }
3369
3370 /*
3371 * putstring "beg"
3372 * putstring "end"
3373 * newrange excl
3374 *
3375 * ==>
3376 *
3377 * putobject "beg".."end"
3378 */
3379 if (IS_INSN_ID(iobj, newrange)) {
3380 INSN *const range = iobj;
3381 INSN *beg, *end;
3382 VALUE str_beg, str_end;
3383
3384 if ((end = (INSN *)get_prev_insn(range)) != 0 &&
3385 is_frozen_putstring(end, &str_end) &&
3386 (beg = (INSN *)get_prev_insn(end)) != 0 &&
3387 is_frozen_putstring(beg, &str_beg)) {
3388 int excl = FIX2INT(OPERAND_AT(range, 0));
3389 VALUE lit_range = rb_range_new(str_beg, str_end, excl);
3390
3391 ELEM_REMOVE(&beg->link);
3392 ELEM_REMOVE(&end->link);
3393 range->insn_id = BIN(putobject);
3394 OPERAND_AT(range, 0) = lit_range;
3395 RB_OBJ_WRITTEN(iseq, Qundef, lit_range);
3396 }
3397 }
3398
3399 if (IS_INSN_ID(iobj, leave)) {
3400 remove_unreachable_chunk(iseq, iobj->link.next);
3401 }
3402
3403 /*
3404 * ...
3405 * duparray [...]
3406 * concatarray | concattoarray
3407 * =>
3408 * ...
3409 * putobject [...]
3410 * concatarray | concattoarray
3411 */
3412 if (IS_INSN_ID(iobj, duparray)) {
3413 LINK_ELEMENT *next = iobj->link.next;
3414 if (IS_INSN(next) && (IS_INSN_ID(next, concatarray) || IS_INSN_ID(next, concattoarray))) {
3415 iobj->insn_id = BIN(putobject);
3416 }
3417 }
3418
3419 /*
3420 * duparray [...]
3421 * send <calldata!mid:freeze, argc:0, ARGS_SIMPLE>, nil
3422 * =>
3423 * opt_ary_freeze [...], <calldata!mid:freeze, argc:0, ARGS_SIMPLE>
3424 */
3425 if (IS_INSN_ID(iobj, duparray)) {
3426 LINK_ELEMENT *next = iobj->link.next;
3427 if (IS_INSN(next) && (IS_INSN_ID(next, send))) {
3428 const struct rb_callinfo *ci = (struct rb_callinfo *)OPERAND_AT(next, 0);
3429 const rb_iseq_t *blockiseq = (rb_iseq_t *)OPERAND_AT(next, 1);
3430
3431 if (vm_ci_simple(ci) && vm_ci_argc(ci) == 0 && blockiseq == NULL && vm_ci_mid(ci) == idFreeze) {
3432 VALUE ary = iobj->operands[0];
3434
3435 iobj->insn_id = BIN(opt_ary_freeze);
3436 iobj->operand_size = 2;
3437 iobj->operands = compile_data_calloc2(iseq, iobj->operand_size, sizeof(VALUE));
3438 iobj->operands[0] = ary;
3439 iobj->operands[1] = (VALUE)ci;
3440 ELEM_REMOVE(next);
3441 }
3442 }
3443 }
3444
3445 /*
3446 * duphash {...}
3447 * send <calldata!mid:freeze, argc:0, ARGS_SIMPLE>, nil
3448 * =>
3449 * opt_hash_freeze {...}, <calldata!mid:freeze, argc:0, ARGS_SIMPLE>
3450 */
3451 if (IS_INSN_ID(iobj, duphash)) {
3452 LINK_ELEMENT *next = iobj->link.next;
3453 if (IS_INSN(next) && (IS_INSN_ID(next, send))) {
3454 const struct rb_callinfo *ci = (struct rb_callinfo *)OPERAND_AT(next, 0);
3455 const rb_iseq_t *blockiseq = (rb_iseq_t *)OPERAND_AT(next, 1);
3456
3457 if (vm_ci_simple(ci) && vm_ci_argc(ci) == 0 && blockiseq == NULL && vm_ci_mid(ci) == idFreeze) {
3458 VALUE hash = iobj->operands[0];
3459 rb_obj_reveal(hash, rb_cHash);
3460
3461 iobj->insn_id = BIN(opt_hash_freeze);
3462 iobj->operand_size = 2;
3463 iobj->operands = compile_data_calloc2(iseq, iobj->operand_size, sizeof(VALUE));
3464 iobj->operands[0] = hash;
3465 iobj->operands[1] = (VALUE)ci;
3466 ELEM_REMOVE(next);
3467 }
3468 }
3469 }
3470
3471 /*
3472 * newarray 0
3473 * send <calldata!mid:freeze, argc:0, ARGS_SIMPLE>, nil
3474 * =>
3475 * opt_ary_freeze [], <calldata!mid:freeze, argc:0, ARGS_SIMPLE>
3476 */
3477 if (IS_INSN_ID(iobj, newarray) && iobj->operands[0] == INT2FIX(0)) {
3478 LINK_ELEMENT *next = iobj->link.next;
3479 if (IS_INSN(next) && (IS_INSN_ID(next, send))) {
3480 const struct rb_callinfo *ci = (struct rb_callinfo *)OPERAND_AT(next, 0);
3481 const rb_iseq_t *blockiseq = (rb_iseq_t *)OPERAND_AT(next, 1);
3482
3483 if (vm_ci_simple(ci) && vm_ci_argc(ci) == 0 && blockiseq == NULL && vm_ci_mid(ci) == idFreeze) {
3484 iobj->insn_id = BIN(opt_ary_freeze);
3485 iobj->operand_size = 2;
3486 iobj->operands = compile_data_calloc2(iseq, iobj->operand_size, sizeof(VALUE));
3487 iobj->operands[0] = rb_cArray_empty_frozen;
3488 iobj->operands[1] = (VALUE)ci;
3489 ELEM_REMOVE(next);
3490 }
3491 }
3492 }
3493
3494 /*
3495 * newhash 0
3496 * send <calldata!mid:freeze, argc:0, ARGS_SIMPLE>, nil
3497 * =>
3498 * opt_hash_freeze {}, <calldata!mid:freeze, argc:0, ARGS_SIMPLE>
3499 */
3500 if (IS_INSN_ID(iobj, newhash) && iobj->operands[0] == INT2FIX(0)) {
3501 LINK_ELEMENT *next = iobj->link.next;
3502 if (IS_INSN(next) && (IS_INSN_ID(next, send))) {
3503 const struct rb_callinfo *ci = (struct rb_callinfo *)OPERAND_AT(next, 0);
3504 const rb_iseq_t *blockiseq = (rb_iseq_t *)OPERAND_AT(next, 1);
3505
3506 if (vm_ci_simple(ci) && vm_ci_argc(ci) == 0 && blockiseq == NULL && vm_ci_mid(ci) == idFreeze) {
3507 iobj->insn_id = BIN(opt_hash_freeze);
3508 iobj->operand_size = 2;
3509 iobj->operands = compile_data_calloc2(iseq, iobj->operand_size, sizeof(VALUE));
3510 iobj->operands[0] = rb_cHash_empty_frozen;
3511 iobj->operands[1] = (VALUE)ci;
3512 ELEM_REMOVE(next);
3513 }
3514 }
3515 }
3516
3517 if (IS_INSN_ID(iobj, branchif) ||
3518 IS_INSN_ID(iobj, branchnil) ||
3519 IS_INSN_ID(iobj, branchunless)) {
3520 /*
3521 * if L1
3522 * ...
3523 * L1:
3524 * jump L2
3525 * =>
3526 * if L2
3527 */
3528 INSN *nobj = (INSN *)get_destination_insn(iobj);
3529
3530 /* This is super nasty hack!!!
3531 *
3532 * This jump-jump optimization may ignore event flags of the jump
3533 * instruction being skipped. Actually, Line 2 TracePoint event
3534 * is never fired in the following code:
3535 *
3536 * 1: raise if 1 == 2
3537 * 2: while true
3538 * 3: break
3539 * 4: end
3540 *
3541 * This is critical for coverage measurement. [Bug #15980]
3542 *
3543 * This is a stopgap measure: stop the jump-jump optimization if
3544 * coverage measurement is enabled and if the skipped instruction
3545 * has any event flag.
3546 *
3547 * Note that, still, TracePoint Line event does not occur on Line 2.
3548 * This should be fixed in future.
3549 */
3550 int stop_optimization =
3551 ISEQ_COVERAGE(iseq) && ISEQ_LINE_COVERAGE(iseq) &&
3552 nobj->link.type == ISEQ_ELEMENT_INSN &&
3553 nobj->insn_info.events;
3554 if (!stop_optimization) {
3555 INSN *pobj = (INSN *)iobj->link.prev;
3556 int prev_dup = 0;
3557 if (pobj) {
3558 if (!IS_INSN(&pobj->link))
3559 pobj = 0;
3560 else if (IS_INSN_ID(pobj, dup))
3561 prev_dup = 1;
3562 }
3563
3564 for (;;) {
3565 if (IS_INSN(&nobj->link) && IS_INSN_ID(nobj, jump)) {
3566 if (!replace_destination(iobj, nobj)) break;
3567 }
3568 else if (prev_dup && IS_INSN_ID(nobj, dup) &&
3569 !!(nobj = (INSN *)nobj->link.next) &&
3570 /* basic blocks, with no labels in the middle */
3571 nobj->insn_id == iobj->insn_id) {
3572 /*
3573 * dup
3574 * if L1
3575 * ...
3576 * L1:
3577 * dup
3578 * if L2
3579 * =>
3580 * dup
3581 * if L2
3582 * ...
3583 * L1:
3584 * dup
3585 * if L2
3586 */
3587 if (!replace_destination(iobj, nobj)) break;
3588 }
3589 else if (pobj) {
3590 /*
3591 * putnil
3592 * if L1
3593 * =>
3594 * # nothing
3595 *
3596 * putobject true
3597 * if L1
3598 * =>
3599 * jump L1
3600 *
3601 * putstring ".."
3602 * if L1
3603 * =>
3604 * jump L1
3605 *
3606 * putstring ".."
3607 * dup
3608 * if L1
3609 * =>
3610 * putstring ".."
3611 * jump L1
3612 *
3613 */
3614 int cond;
3615 if (prev_dup && IS_INSN(pobj->link.prev)) {
3616 pobj = (INSN *)pobj->link.prev;
3617 }
3618 if (IS_INSN_ID(pobj, putobject)) {
3619 cond = (IS_INSN_ID(iobj, branchif) ?
3620 OPERAND_AT(pobj, 0) != Qfalse :
3621 IS_INSN_ID(iobj, branchunless) ?
3622 OPERAND_AT(pobj, 0) == Qfalse :
3623 FALSE);
3624 }
3625 else if (IS_INSN_ID(pobj, putstring) ||
3626 IS_INSN_ID(pobj, duparray) ||
3627 IS_INSN_ID(pobj, newarray)) {
3628 cond = IS_INSN_ID(iobj, branchif);
3629 }
3630 else if (IS_INSN_ID(pobj, putnil)) {
3631 cond = !IS_INSN_ID(iobj, branchif);
3632 }
3633 else break;
3634 if (prev_dup || !IS_INSN_ID(pobj, newarray)) {
3635 ELEM_REMOVE(iobj->link.prev);
3636 }
3637 else if (!iseq_pop_newarray(iseq, pobj)) {
3638 pobj = new_insn_core(iseq, pobj->insn_info.line_no, pobj->insn_info.node_id, BIN(pop), 0, NULL);
3639 ELEM_INSERT_PREV(&iobj->link, &pobj->link);
3640 }
3641 if (cond) {
3642 if (prev_dup) {
3643 pobj = new_insn_core(iseq, pobj->insn_info.line_no, pobj->insn_info.node_id, BIN(putnil), 0, NULL);
3644 ELEM_INSERT_NEXT(&iobj->link, &pobj->link);
3645 }
3646 iobj->insn_id = BIN(jump);
3647 goto again;
3648 }
3649 else {
3650 unref_destination(iobj, 0);
3651 ELEM_REMOVE(&iobj->link);
3652 }
3653 break;
3654 }
3655 else break;
3656 nobj = (INSN *)get_destination_insn(nobj);
3657 }
3658 }
3659 }
3660
3661 if (IS_INSN_ID(iobj, pop)) {
3662 /*
3663 * putself / putnil / putobject obj / putstring "..."
3664 * pop
3665 * =>
3666 * # do nothing
3667 */
3668 LINK_ELEMENT *prev = iobj->link.prev;
3669 if (IS_INSN(prev)) {
3670 enum ruby_vminsn_type previ = ((INSN *)prev)->insn_id;
3671 if (previ == BIN(putobject) || previ == BIN(putnil) ||
3672 previ == BIN(putself) || previ == BIN(putstring) ||
3673 previ == BIN(putchilledstring) ||
3674 previ == BIN(dup) ||
3675 previ == BIN(getlocal) ||
3676 previ == BIN(getblockparam) ||
3677 previ == BIN(getblockparamproxy) ||
3678 previ == BIN(getinstancevariable) ||
3679 previ == BIN(duparray)) {
3680 /* just push operand or static value and pop soon, no
3681 * side effects */
3682 ELEM_REMOVE(prev);
3683 ELEM_REMOVE(&iobj->link);
3684 }
3685 else if (previ == BIN(newarray) && iseq_pop_newarray(iseq, (INSN*)prev)) {
3686 ELEM_REMOVE(&iobj->link);
3687 }
3688 else if (previ == BIN(concatarray)) {
3689 INSN *piobj = (INSN *)prev;
3690 INSERT_BEFORE_INSN1(piobj, piobj->insn_info.line_no, piobj->insn_info.node_id, splatarray, Qfalse);
3691 INSN_OF(piobj) = BIN(pop);
3692 }
3693 else if (previ == BIN(concatstrings)) {
3694 if (OPERAND_AT(prev, 0) == INT2FIX(1)) {
3695 ELEM_REMOVE(prev);
3696 }
3697 else {
3698 ELEM_REMOVE(&iobj->link);
3699 INSN_OF(prev) = BIN(adjuststack);
3700 }
3701 }
3702 }
3703 }
3704
3705 if (IS_INSN_ID(iobj, newarray) ||
3706 IS_INSN_ID(iobj, duparray) ||
3707 IS_INSN_ID(iobj, concatarray) ||
3708 IS_INSN_ID(iobj, splatarray) ||
3709 0) {
3710 /*
3711 * newarray N
3712 * splatarray
3713 * =>
3714 * newarray N
3715 * newarray always puts an array
3716 */
3717 LINK_ELEMENT *next = iobj->link.next;
3718 if (IS_INSN(next) && IS_INSN_ID(next, splatarray)) {
3719 /* remove splatarray following always-array insn */
3720 ELEM_REMOVE(next);
3721 }
3722 }
3723
3724 if (IS_INSN_ID(iobj, newarray)) {
3725 LINK_ELEMENT *next = iobj->link.next;
3726 if (IS_INSN(next) && IS_INSN_ID(next, expandarray) &&
3727 OPERAND_AT(next, 1) == INT2FIX(0)) {
3728 VALUE op1, op2;
3729 op1 = OPERAND_AT(iobj, 0);
3730 op2 = OPERAND_AT(next, 0);
3731 ELEM_REMOVE(next);
3732
3733 if (op1 == op2) {
3734 /*
3735 * newarray 2
3736 * expandarray 2, 0
3737 * =>
3738 * swap
3739 */
3740 if (op1 == INT2FIX(2)) {
3741 INSN_OF(iobj) = BIN(swap);
3742 iobj->operand_size = 0;
3743 }
3744 /*
3745 * newarray X
3746 * expandarray X, 0
3747 * =>
3748 * opt_reverse X
3749 */
3750 else {
3751 INSN_OF(iobj) = BIN(opt_reverse);
3752 }
3753 }
3754 else {
3755 long diff = FIX2LONG(op1) - FIX2LONG(op2);
3756 INSN_OF(iobj) = BIN(opt_reverse);
3757 OPERAND_AT(iobj, 0) = OPERAND_AT(next, 0);
3758
3759 if (op1 > op2) {
3760 /* X > Y
3761 * newarray X
3762 * expandarray Y, 0
3763 * =>
3764 * pop * (Y-X)
3765 * opt_reverse Y
3766 */
3767 for (; diff > 0; diff--) {
3768 INSERT_BEFORE_INSN(iobj, iobj->insn_info.line_no, iobj->insn_info.node_id, pop);
3769 }
3770 }
3771 else { /* (op1 < op2) */
3772 /* X < Y
3773 * newarray X
3774 * expandarray Y, 0
3775 * =>
3776 * putnil * (Y-X)
3777 * opt_reverse Y
3778 */
3779 for (; diff < 0; diff++) {
3780 INSERT_BEFORE_INSN(iobj, iobj->insn_info.line_no, iobj->insn_info.node_id, putnil);
3781 }
3782 }
3783 }
3784 }
3785 }
3786
3787 if (IS_INSN_ID(iobj, duparray)) {
3788 LINK_ELEMENT *next = iobj->link.next;
3789 /*
3790 * duparray obj
3791 * expandarray X, 0
3792 * =>
3793 * putobject obj
3794 * expandarray X, 0
3795 */
3796 if (IS_INSN(next) && IS_INSN_ID(next, expandarray)) {
3797 INSN_OF(iobj) = BIN(putobject);
3798 }
3799 }
3800
3801 if (IS_INSN_ID(iobj, anytostring)) {
3802 LINK_ELEMENT *next = iobj->link.next;
3803 /*
3804 * anytostring
3805 * concatstrings 1
3806 * =>
3807 * anytostring
3808 */
3809 if (IS_INSN(next) && IS_INSN_ID(next, concatstrings) &&
3810 OPERAND_AT(next, 0) == INT2FIX(1)) {
3811 ELEM_REMOVE(next);
3812 }
3813 }
3814
3815 if (IS_INSN_ID(iobj, putstring) || IS_INSN_ID(iobj, putchilledstring) ||
3816 (IS_INSN_ID(iobj, putobject) && RB_TYPE_P(OPERAND_AT(iobj, 0), T_STRING))) {
3817 /*
3818 * putstring ""
3819 * concatstrings N
3820 * =>
3821 * concatstrings N-1
3822 */
3823 if (IS_NEXT_INSN_ID(&iobj->link, concatstrings) &&
3824 RSTRING_LEN(OPERAND_AT(iobj, 0)) == 0) {
3825 INSN *next = (INSN *)iobj->link.next;
3826 if ((OPERAND_AT(next, 0) = FIXNUM_INC(OPERAND_AT(next, 0), -1)) == INT2FIX(1)) {
3827 ELEM_REMOVE(&next->link);
3828 }
3829 ELEM_REMOVE(&iobj->link);
3830 }
3831 if (IS_NEXT_INSN_ID(&iobj->link, toregexp)) {
3832 INSN *next = (INSN *)iobj->link.next;
3833 if (OPERAND_AT(next, 1) == INT2FIX(1)) {
3834 VALUE src = OPERAND_AT(iobj, 0);
3835 int opt = (int)FIX2LONG(OPERAND_AT(next, 0));
3836 VALUE path = rb_iseq_path(iseq);
3837 int line = iobj->insn_info.line_no;
3838 VALUE errinfo = rb_errinfo();
3839 VALUE re = rb_reg_compile(src, opt, RSTRING_PTR(path), line);
3840 if (NIL_P(re)) {
3841 VALUE message = rb_attr_get(rb_errinfo(), idMesg);
3842 rb_set_errinfo(errinfo);
3843 COMPILE_ERROR(iseq, line, "%" PRIsVALUE, message);
3844 }
3845 RB_OBJ_WRITE(iseq, &OPERAND_AT(iobj, 0), re);
3846 ELEM_REMOVE(iobj->link.next);
3847 }
3848 }
3849 }
3850
3851 if (IS_INSN_ID(iobj, concatstrings)) {
3852 /*
3853 * concatstrings N
3854 * concatstrings M
3855 * =>
3856 * concatstrings N+M-1
3857 */
3858 LINK_ELEMENT *next = iobj->link.next;
3859 INSN *jump = 0;
3860 if (IS_INSN(next) && IS_INSN_ID(next, jump))
3861 next = get_destination_insn(jump = (INSN *)next);
3862 if (IS_INSN(next) && IS_INSN_ID(next, concatstrings)) {
3863 int n = FIX2INT(OPERAND_AT(iobj, 0)) + FIX2INT(OPERAND_AT(next, 0)) - 1;
3864 OPERAND_AT(iobj, 0) = INT2FIX(n);
3865 if (jump) {
3866 LABEL *label = ((LABEL *)OPERAND_AT(jump, 0));
3867 if (!--label->refcnt) {
3868 ELEM_REMOVE(&label->link);
3869 }
3870 else {
3871 label = NEW_LABEL(0);
3872 OPERAND_AT(jump, 0) = (VALUE)label;
3873 }
3874 label->refcnt++;
3875 ELEM_INSERT_NEXT(next, &label->link);
3876 CHECK(iseq_peephole_optimize(iseq, get_next_insn(jump), do_tailcallopt));
3877 }
3878 else {
3879 ELEM_REMOVE(next);
3880 }
3881 }
3882 }
3883
3884 if (do_tailcallopt &&
3885 (IS_INSN_ID(iobj, send) ||
3886 IS_INSN_ID(iobj, opt_aref_with) ||
3887 IS_INSN_ID(iobj, opt_aset_with) ||
3888 IS_INSN_ID(iobj, invokesuper))) {
3889 /*
3890 * send ...
3891 * leave
3892 * =>
3893 * send ..., ... | VM_CALL_TAILCALL, ...
3894 * leave # unreachable
3895 */
3896 INSN *piobj = NULL;
3897 if (iobj->link.next) {
3898 LINK_ELEMENT *next = iobj->link.next;
3899 do {
3900 if (!IS_INSN(next)) {
3901 next = next->next;
3902 continue;
3903 }
3904 switch (INSN_OF(next)) {
3905 case BIN(nop):
3906 next = next->next;
3907 break;
3908 case BIN(jump):
3909 /* if cond
3910 * return tailcall
3911 * end
3912 */
3913 next = get_destination_insn((INSN *)next);
3914 break;
3915 case BIN(leave):
3916 piobj = iobj;
3917 /* fall through */
3918 default:
3919 next = NULL;
3920 break;
3921 }
3922 } while (next);
3923 }
3924
3925 if (piobj) {
3926 const struct rb_callinfo *ci = (struct rb_callinfo *)OPERAND_AT(piobj, 0);
3927 if (IS_INSN_ID(piobj, send) ||
3928 IS_INSN_ID(piobj, invokesuper)) {
3929 if (OPERAND_AT(piobj, 1) == 0) { /* no blockiseq */
3930 ci = ci_flag_set(iseq, ci, VM_CALL_TAILCALL);
3931 OPERAND_AT(piobj, 0) = (VALUE)ci;
3932 RB_OBJ_WRITTEN(iseq, Qundef, ci);
3933 }
3934 }
3935 else {
3936 ci = ci_flag_set(iseq, ci, VM_CALL_TAILCALL);
3937 OPERAND_AT(piobj, 0) = (VALUE)ci;
3938 RB_OBJ_WRITTEN(iseq, Qundef, ci);
3939 }
3940 }
3941 }
3942
3943 if (IS_INSN_ID(iobj, dup)) {
3944 if (IS_NEXT_INSN_ID(&iobj->link, setlocal)) {
3945 LINK_ELEMENT *set1 = iobj->link.next, *set2 = NULL;
3946
3947 /*
3948 * dup
3949 * setlocal x, y
3950 * setlocal x, y
3951 * =>
3952 * dup
3953 * setlocal x, y
3954 */
3955 if (IS_NEXT_INSN_ID(set1, setlocal)) {
3956 set2 = set1->next;
3957 if (OPERAND_AT(set1, 0) == OPERAND_AT(set2, 0) &&
3958 OPERAND_AT(set1, 1) == OPERAND_AT(set2, 1)) {
3959 ELEM_REMOVE(set1);
3960 ELEM_REMOVE(&iobj->link);
3961 }
3962 }
3963
3964 /*
3965 * dup
3966 * setlocal x, y
3967 * dup
3968 * setlocal x, y
3969 * =>
3970 * dup
3971 * setlocal x, y
3972 */
3973 else if (IS_NEXT_INSN_ID(set1, dup) &&
3974 IS_NEXT_INSN_ID(set1->next, setlocal)) {
3975 set2 = set1->next->next;
3976 if (OPERAND_AT(set1, 0) == OPERAND_AT(set2, 0) &&
3977 OPERAND_AT(set1, 1) == OPERAND_AT(set2, 1)) {
3978 ELEM_REMOVE(set1->next);
3979 ELEM_REMOVE(set2);
3980 }
3981 }
3982 }
3983 }
3984
3985 /*
3986 * getlocal x, y
3987 * dup
3988 * setlocal x, y
3989 * =>
3990 * dup
3991 */
3992 if (IS_INSN_ID(iobj, getlocal)) {
3993 LINK_ELEMENT *niobj = &iobj->link;
3994 if (IS_NEXT_INSN_ID(niobj, dup)) {
3995 niobj = niobj->next;
3996 }
3997 if (IS_NEXT_INSN_ID(niobj, setlocal)) {
3998 LINK_ELEMENT *set1 = niobj->next;
3999 if (OPERAND_AT(iobj, 0) == OPERAND_AT(set1, 0) &&
4000 OPERAND_AT(iobj, 1) == OPERAND_AT(set1, 1)) {
4001 ELEM_REMOVE(set1);
4002 ELEM_REMOVE(niobj);
4003 }
4004 }
4005 }
4006
4007 /*
4008 * opt_invokebuiltin_delegate
4009 * trace
4010 * leave
4011 * =>
4012 * opt_invokebuiltin_delegate_leave
4013 * trace
4014 * leave
4015 */
4016 if (IS_INSN_ID(iobj, opt_invokebuiltin_delegate)) {
4017 if (IS_TRACE(iobj->link.next)) {
4018 if (IS_NEXT_INSN_ID(iobj->link.next, leave)) {
4019 iobj->insn_id = BIN(opt_invokebuiltin_delegate_leave);
4020 const struct rb_builtin_function *bf = (const struct rb_builtin_function *)iobj->operands[0];
4021 if (iobj == (INSN *)list && bf->argc == 0 && (ISEQ_BODY(iseq)->builtin_attrs & BUILTIN_ATTR_LEAF)) {
4022 ISEQ_BODY(iseq)->builtin_attrs |= BUILTIN_ATTR_SINGLE_NOARG_LEAF;
4023 }
4024 }
4025 }
4026 }
4027
4028 /*
4029 * getblockparam
4030 * branchif / branchunless
4031 * =>
4032 * getblockparamproxy
4033 * branchif / branchunless
4034 */
4035 if (IS_INSN_ID(iobj, getblockparam)) {
4036 if (IS_NEXT_INSN_ID(&iobj->link, branchif) || IS_NEXT_INSN_ID(&iobj->link, branchunless)) {
4037 iobj->insn_id = BIN(getblockparamproxy);
4038 }
4039 }
4040
4041 if (IS_INSN_ID(iobj, splatarray) && OPERAND_AT(iobj, 0) == false) {
4042 LINK_ELEMENT *niobj = &iobj->link;
4043 if (IS_NEXT_INSN_ID(niobj, duphash)) {
4044 niobj = niobj->next;
4045 LINK_ELEMENT *siobj;
4046 unsigned int set_flags = 0, unset_flags = 0;
4047
4048 /*
4049 * Eliminate hash allocation for f(*a, kw: 1)
4050 *
4051 * splatarray false
4052 * duphash
4053 * send ARGS_SPLAT|KW_SPLAT|KW_SPLAT_MUT and not ARGS_BLOCKARG
4054 * =>
4055 * splatarray false
4056 * putobject
4057 * send ARGS_SPLAT|KW_SPLAT
4058 */
4059 if (IS_NEXT_INSN_ID(niobj, send)) {
4060 siobj = niobj->next;
4061 set_flags = VM_CALL_ARGS_SPLAT|VM_CALL_KW_SPLAT|VM_CALL_KW_SPLAT_MUT;
4062 unset_flags = VM_CALL_ARGS_BLOCKARG;
4063 }
4064 /*
4065 * Eliminate hash allocation for f(*a, kw: 1, &{arg,lvar,@iv})
4066 *
4067 * splatarray false
4068 * duphash
4069 * getlocal / getinstancevariable / getblockparamproxy
4070 * send ARGS_SPLAT|KW_SPLAT|KW_SPLAT_MUT|ARGS_BLOCKARG
4071 * =>
4072 * splatarray false
4073 * putobject
4074 * getlocal / getinstancevariable / getblockparamproxy
4075 * send ARGS_SPLAT|KW_SPLAT|ARGS_BLOCKARG
4076 */
4077 else if ((IS_NEXT_INSN_ID(niobj, getlocal) || IS_NEXT_INSN_ID(niobj, getinstancevariable) ||
4078 IS_NEXT_INSN_ID(niobj, getblockparamproxy)) && (IS_NEXT_INSN_ID(niobj->next, send))) {
4079 siobj = niobj->next->next;
4080 set_flags = VM_CALL_ARGS_SPLAT|VM_CALL_KW_SPLAT|VM_CALL_KW_SPLAT_MUT|VM_CALL_ARGS_BLOCKARG;
4081 }
4082
4083 if (set_flags) {
4084 const struct rb_callinfo *ci = (const struct rb_callinfo *)OPERAND_AT(siobj, 0);
4085 unsigned int flags = vm_ci_flag(ci);
4086 if ((flags & set_flags) == set_flags && !(flags & unset_flags)) {
4087 ((INSN*)niobj)->insn_id = BIN(putobject);
4088 OPERAND_AT(niobj, 0) = rb_hash_freeze(rb_hash_resurrect(OPERAND_AT(niobj, 0)));
4089
4090 const struct rb_callinfo *nci = vm_ci_new(vm_ci_mid(ci),
4091 flags & ~VM_CALL_KW_SPLAT_MUT, vm_ci_argc(ci), vm_ci_kwarg(ci));
4092 RB_OBJ_WRITTEN(iseq, ci, nci);
4093 OPERAND_AT(siobj, 0) = (VALUE)nci;
4094 }
4095 }
4096 }
4097 }
4098
4099 return COMPILE_OK;
4100}
4101
4102static int
4103insn_set_specialized_instruction(rb_iseq_t *iseq, INSN *iobj, int insn_id)
4104{
4105 iobj->insn_id = insn_id;
4106 iobj->operand_size = insn_len(insn_id) - 1;
4107 iobj->insn_info.events |= RUBY_EVENT_C_CALL | RUBY_EVENT_C_RETURN;
4108
4109 if (insn_id == BIN(opt_neq)) {
4110 VALUE original_ci = iobj->operands[0];
4111 iobj->operand_size = 2;
4112 iobj->operands = compile_data_calloc2(iseq, iobj->operand_size, sizeof(VALUE));
4113 iobj->operands[0] = (VALUE)new_callinfo(iseq, idEq, 1, 0, NULL, FALSE);
4114 iobj->operands[1] = original_ci;
4115 }
4116
4117 return COMPILE_OK;
4118}
4119
4120static int
4121iseq_specialized_instruction(rb_iseq_t *iseq, INSN *iobj)
4122{
4123 if (IS_INSN_ID(iobj, newarray) && iobj->link.next &&
4124 IS_INSN(iobj->link.next)) {
4125 /*
4126 * [a, b, ...].max/min -> a, b, c, opt_newarray_send max/min
4127 */
4128 INSN *niobj = (INSN *)iobj->link.next;
4129 if (IS_INSN_ID(niobj, send)) {
4130 const struct rb_callinfo *ci = (struct rb_callinfo *)OPERAND_AT(niobj, 0);
4131 if (vm_ci_simple(ci) && vm_ci_argc(ci) == 0) {
4132 VALUE method = INT2FIX(0);
4133 switch (vm_ci_mid(ci)) {
4134 case idMax:
4135 method = INT2FIX(VM_OPT_NEWARRAY_SEND_MAX);
4136 break;
4137 case idMin:
4138 method = INT2FIX(VM_OPT_NEWARRAY_SEND_MIN);
4139 break;
4140 case idHash:
4141 method = INT2FIX(VM_OPT_NEWARRAY_SEND_HASH);
4142 break;
4143 }
4144
4145 if (method != INT2FIX(0)) {
4146 VALUE num = iobj->operands[0];
4147 int operand_len = insn_len(BIN(opt_newarray_send)) - 1;
4148 iobj->insn_id = BIN(opt_newarray_send);
4149 iobj->operands = compile_data_calloc2(iseq, operand_len, sizeof(VALUE));
4150 iobj->operands[0] = num;
4151 iobj->operands[1] = method;
4152 iobj->operand_size = operand_len;
4153 ELEM_REMOVE(&niobj->link);
4154 return COMPILE_OK;
4155 }
4156 }
4157 }
4158 else if ((IS_INSN_ID(niobj, putstring) || IS_INSN_ID(niobj, putchilledstring) ||
4159 (IS_INSN_ID(niobj, putobject) && RB_TYPE_P(OPERAND_AT(niobj, 0), T_STRING))) &&
4160 IS_NEXT_INSN_ID(&niobj->link, send)) {
4161 const struct rb_callinfo *ci = (struct rb_callinfo *)OPERAND_AT((INSN *)niobj->link.next, 0);
4162 if (vm_ci_simple(ci) && vm_ci_argc(ci) == 1 && vm_ci_mid(ci) == idPack) {
4163 VALUE num = iobj->operands[0];
4164 int operand_len = insn_len(BIN(opt_newarray_send)) - 1;
4165 iobj->insn_id = BIN(opt_newarray_send);
4166 iobj->operands = compile_data_calloc2(iseq, operand_len, sizeof(VALUE));
4167 iobj->operands[0] = FIXNUM_INC(num, 1);
4168 iobj->operands[1] = INT2FIX(VM_OPT_NEWARRAY_SEND_PACK);
4169 iobj->operand_size = operand_len;
4170 ELEM_REMOVE(&iobj->link);
4171 ELEM_REMOVE(niobj->link.next);
4172 ELEM_INSERT_NEXT(&niobj->link, &iobj->link);
4173 return COMPILE_OK;
4174 }
4175 }
4176 // newarray n, putchilledstring "E", getlocal b, send :pack with {buffer: b}
4177 // -> putchilledstring "E", getlocal b, opt_newarray_send n+2, :pack, :buffer
4178 else if ((IS_INSN_ID(niobj, putstring) || IS_INSN_ID(niobj, putchilledstring) ||
4179 (IS_INSN_ID(niobj, putobject) && RB_TYPE_P(OPERAND_AT(niobj, 0), T_STRING))) &&
4180 IS_NEXT_INSN_ID(&niobj->link, getlocal) &&
4181 (niobj->link.next && IS_NEXT_INSN_ID(niobj->link.next, send))) {
4182 const struct rb_callinfo *ci = (struct rb_callinfo *)OPERAND_AT((INSN *)(niobj->link.next)->next, 0);
4183 const struct rb_callinfo_kwarg *kwarg = vm_ci_kwarg(ci);
4184 if (vm_ci_mid(ci) == idPack && vm_ci_argc(ci) == 2 &&
4185 (kwarg && kwarg->keyword_len == 1 && kwarg->keywords[0] == rb_id2sym(idBuffer))) {
4186 VALUE num = iobj->operands[0];
4187 int operand_len = insn_len(BIN(opt_newarray_send)) - 1;
4188 iobj->insn_id = BIN(opt_newarray_send);
4189 iobj->operands = compile_data_calloc2(iseq, operand_len, sizeof(VALUE));
4190 iobj->operands[0] = FIXNUM_INC(num, 2);
4191 iobj->operands[1] = INT2FIX(VM_OPT_NEWARRAY_SEND_PACK_BUFFER);
4192 iobj->operand_size = operand_len;
4193 // Remove the "send" insn.
4194 ELEM_REMOVE((niobj->link.next)->next);
4195 // Remove the modified insn from its original "newarray" position...
4196 ELEM_REMOVE(&iobj->link);
4197 // and insert it after the buffer insn.
4198 ELEM_INSERT_NEXT(niobj->link.next, &iobj->link);
4199 return COMPILE_OK;
4200 }
4201 }
4202
4203 // Break the "else if" chain since some prior checks abort after sub-ifs.
4204 // We already found "newarray". To match `[...].include?(arg)` we look for
4205 // the instruction(s) representing the argument followed by a "send".
4206 if ((IS_INSN_ID(niobj, putstring) || IS_INSN_ID(niobj, putchilledstring) ||
4207 IS_INSN_ID(niobj, putobject) ||
4208 IS_INSN_ID(niobj, putself) ||
4209 IS_INSN_ID(niobj, getlocal) ||
4210 IS_INSN_ID(niobj, getinstancevariable)) &&
4211 IS_NEXT_INSN_ID(&niobj->link, send)) {
4212
4213 LINK_ELEMENT *sendobj = &(niobj->link); // Below we call ->next;
4214 const struct rb_callinfo *ci;
4215 // Allow any number (0 or more) of simple method calls on the argument
4216 // (as in `[...].include?(arg.method1.method2)`.
4217 do {
4218 sendobj = sendobj->next;
4219 ci = (struct rb_callinfo *)OPERAND_AT(sendobj, 0);
4220 } while (vm_ci_simple(ci) && vm_ci_argc(ci) == 0 && IS_NEXT_INSN_ID(sendobj, send));
4221
4222 // If this send is for .include? with one arg we can do our opt.
4223 if (vm_ci_simple(ci) && vm_ci_argc(ci) == 1 && vm_ci_mid(ci) == idIncludeP) {
4224 VALUE num = iobj->operands[0];
4225 INSN *sendins = (INSN *)sendobj;
4226 sendins->insn_id = BIN(opt_newarray_send);
4227 sendins->operand_size = insn_len(sendins->insn_id) - 1;
4228 sendins->operands = compile_data_calloc2(iseq, sendins->operand_size, sizeof(VALUE));
4229 sendins->operands[0] = FIXNUM_INC(num, 1);
4230 sendins->operands[1] = INT2FIX(VM_OPT_NEWARRAY_SEND_INCLUDE_P);
4231 // Remove the original "newarray" insn.
4232 ELEM_REMOVE(&iobj->link);
4233 return COMPILE_OK;
4234 }
4235 }
4236 }
4237
4238 /*
4239 * duparray [...]
4240 * some insn for the arg...
4241 * send <calldata!mid:include?, argc:1, ARGS_SIMPLE>, nil
4242 * =>
4243 * arg insn...
4244 * opt_duparray_send [...], :include?, 1
4245 */
4246 if (IS_INSN_ID(iobj, duparray) && iobj->link.next && IS_INSN(iobj->link.next)) {
4247 INSN *niobj = (INSN *)iobj->link.next;
4248 if ((IS_INSN_ID(niobj, getlocal) ||
4249 IS_INSN_ID(niobj, getinstancevariable) ||
4250 IS_INSN_ID(niobj, putself)) &&
4251 IS_NEXT_INSN_ID(&niobj->link, send)) {
4252
4253 LINK_ELEMENT *sendobj = &(niobj->link); // Below we call ->next;
4254 const struct rb_callinfo *ci;
4255 // Allow any number (0 or more) of simple method calls on the argument
4256 // (as in `[...].include?(arg.method1.method2)`.
4257 do {
4258 sendobj = sendobj->next;
4259 ci = (struct rb_callinfo *)OPERAND_AT(sendobj, 0);
4260 } while (vm_ci_simple(ci) && vm_ci_argc(ci) == 0 && IS_NEXT_INSN_ID(sendobj, send));
4261
4262 if (vm_ci_simple(ci) && vm_ci_argc(ci) == 1 && vm_ci_mid(ci) == idIncludeP) {
4263 // Move the array arg from duparray to opt_duparray_send.
4264 VALUE ary = iobj->operands[0];
4266
4267 INSN *sendins = (INSN *)sendobj;
4268 sendins->insn_id = BIN(opt_duparray_send);
4269 sendins->operand_size = insn_len(sendins->insn_id) - 1;;
4270 sendins->operands = compile_data_calloc2(iseq, sendins->operand_size, sizeof(VALUE));
4271 sendins->operands[0] = ary;
4272 sendins->operands[1] = rb_id2sym(idIncludeP);
4273 sendins->operands[2] = INT2FIX(1);
4274
4275 // Remove the duparray insn.
4276 ELEM_REMOVE(&iobj->link);
4277 return COMPILE_OK;
4278 }
4279 }
4280 }
4281
4282
4283 if (IS_INSN_ID(iobj, send)) {
4284 const struct rb_callinfo *ci = (struct rb_callinfo *)OPERAND_AT(iobj, 0);
4285 const rb_iseq_t *blockiseq = (rb_iseq_t *)OPERAND_AT(iobj, 1);
4286
4287#define SP_INSN(opt) insn_set_specialized_instruction(iseq, iobj, BIN(opt_##opt))
4288 if (vm_ci_simple(ci)) {
4289 switch (vm_ci_argc(ci)) {
4290 case 0:
4291 switch (vm_ci_mid(ci)) {
4292 case idLength: SP_INSN(length); return COMPILE_OK;
4293 case idSize: SP_INSN(size); return COMPILE_OK;
4294 case idEmptyP: SP_INSN(empty_p);return COMPILE_OK;
4295 case idNilP: SP_INSN(nil_p); return COMPILE_OK;
4296 case idSucc: SP_INSN(succ); return COMPILE_OK;
4297 case idNot: SP_INSN(not); return COMPILE_OK;
4298 }
4299 break;
4300 case 1:
4301 switch (vm_ci_mid(ci)) {
4302 case idPLUS: SP_INSN(plus); return COMPILE_OK;
4303 case idMINUS: SP_INSN(minus); return COMPILE_OK;
4304 case idMULT: SP_INSN(mult); return COMPILE_OK;
4305 case idDIV: SP_INSN(div); return COMPILE_OK;
4306 case idMOD: SP_INSN(mod); return COMPILE_OK;
4307 case idEq: SP_INSN(eq); return COMPILE_OK;
4308 case idNeq: SP_INSN(neq); return COMPILE_OK;
4309 case idEqTilde:SP_INSN(regexpmatch2);return COMPILE_OK;
4310 case idLT: SP_INSN(lt); return COMPILE_OK;
4311 case idLE: SP_INSN(le); return COMPILE_OK;
4312 case idGT: SP_INSN(gt); return COMPILE_OK;
4313 case idGE: SP_INSN(ge); return COMPILE_OK;
4314 case idLTLT: SP_INSN(ltlt); return COMPILE_OK;
4315 case idAREF: SP_INSN(aref); return COMPILE_OK;
4316 case idAnd: SP_INSN(and); return COMPILE_OK;
4317 case idOr: SP_INSN(or); return COMPILE_OK;
4318 }
4319 break;
4320 case 2:
4321 switch (vm_ci_mid(ci)) {
4322 case idASET: SP_INSN(aset); return COMPILE_OK;
4323 }
4324 break;
4325 }
4326 }
4327
4328 if ((vm_ci_flag(ci) & (VM_CALL_ARGS_BLOCKARG | VM_CALL_FORWARDING)) == 0 && blockiseq == NULL) {
4329 iobj->insn_id = BIN(opt_send_without_block);
4330 iobj->operand_size = insn_len(iobj->insn_id) - 1;
4331 }
4332 }
4333#undef SP_INSN
4334
4335 return COMPILE_OK;
4336}
4337
4338static inline int
4339tailcallable_p(rb_iseq_t *iseq)
4340{
4341 switch (ISEQ_BODY(iseq)->type) {
4342 case ISEQ_TYPE_TOP:
4343 case ISEQ_TYPE_EVAL:
4344 case ISEQ_TYPE_MAIN:
4345 /* not tail callable because cfp will be over popped */
4346 case ISEQ_TYPE_RESCUE:
4347 case ISEQ_TYPE_ENSURE:
4348 /* rescue block can't tail call because of errinfo */
4349 return FALSE;
4350 default:
4351 return TRUE;
4352 }
4353}
4354
4355static int
4356iseq_optimize(rb_iseq_t *iseq, LINK_ANCHOR *const anchor)
4357{
4358 LINK_ELEMENT *list;
4359 const int do_peepholeopt = ISEQ_COMPILE_DATA(iseq)->option->peephole_optimization;
4360 const int do_tailcallopt = tailcallable_p(iseq) &&
4361 ISEQ_COMPILE_DATA(iseq)->option->tailcall_optimization;
4362 const int do_si = ISEQ_COMPILE_DATA(iseq)->option->specialized_instruction;
4363 const int do_ou = ISEQ_COMPILE_DATA(iseq)->option->operands_unification;
4364 int rescue_level = 0;
4365 int tailcallopt = do_tailcallopt;
4366
4367 list = FIRST_ELEMENT(anchor);
4368
4369 int do_block_optimization = 0;
4370 LABEL * block_loop_label = NULL;
4371
4372 // If we're optimizing a block
4373 if (ISEQ_BODY(iseq)->type == ISEQ_TYPE_BLOCK) {
4374 do_block_optimization = 1;
4375
4376 // If the block starts with a nop and a label,
4377 // record the label so we can detect if it's a jump target
4378 LINK_ELEMENT * le = FIRST_ELEMENT(anchor)->next;
4379 if (IS_INSN(le) && IS_INSN_ID((INSN *)le, nop) && IS_LABEL(le->next)) {
4380 block_loop_label = (LABEL *)le->next;
4381 }
4382 }
4383
4384 while (list) {
4385 if (IS_INSN(list)) {
4386 if (do_peepholeopt) {
4387 iseq_peephole_optimize(iseq, list, tailcallopt);
4388 }
4389 if (do_si) {
4390 iseq_specialized_instruction(iseq, (INSN *)list);
4391 }
4392 if (do_ou) {
4393 insn_operands_unification((INSN *)list);
4394 }
4395
4396 if (do_block_optimization) {
4397 INSN * item = (INSN *)list;
4398 // Give up if there is a throw
4399 if (IS_INSN_ID(item, throw)) {
4400 do_block_optimization = 0;
4401 }
4402 else {
4403 // If the instruction has a jump target, check if the
4404 // jump target is the block loop label
4405 const char *types = insn_op_types(item->insn_id);
4406 for (int j = 0; types[j]; j++) {
4407 if (types[j] == TS_OFFSET) {
4408 // If the jump target is equal to the block loop
4409 // label, then we can't do the optimization because
4410 // the leading `nop` instruction fires the block
4411 // entry tracepoint
4412 LABEL * target = (LABEL *)OPERAND_AT(item, j);
4413 if (target == block_loop_label) {
4414 do_block_optimization = 0;
4415 }
4416 }
4417 }
4418 }
4419 }
4420 }
4421 if (IS_LABEL(list)) {
4422 switch (((LABEL *)list)->rescued) {
4423 case LABEL_RESCUE_BEG:
4424 rescue_level++;
4425 tailcallopt = FALSE;
4426 break;
4427 case LABEL_RESCUE_END:
4428 if (!--rescue_level) tailcallopt = do_tailcallopt;
4429 break;
4430 }
4431 }
4432 list = list->next;
4433 }
4434
4435 if (do_block_optimization) {
4436 LINK_ELEMENT * le = FIRST_ELEMENT(anchor)->next;
4437 if (IS_INSN(le) && IS_INSN_ID((INSN *)le, nop)) {
4438 ELEM_REMOVE(le);
4439 }
4440 }
4441 return COMPILE_OK;
4442}
4443
4444#if OPT_INSTRUCTIONS_UNIFICATION
4445static INSN *
4446new_unified_insn(rb_iseq_t *iseq,
4447 int insn_id, int size, LINK_ELEMENT *seq_list)
4448{
4449 INSN *iobj = 0;
4450 LINK_ELEMENT *list = seq_list;
4451 int i, argc = 0;
4452 VALUE *operands = 0, *ptr = 0;
4453
4454
4455 /* count argc */
4456 for (i = 0; i < size; i++) {
4457 iobj = (INSN *)list;
4458 argc += iobj->operand_size;
4459 list = list->next;
4460 }
4461
4462 if (argc > 0) {
4463 ptr = operands = compile_data_alloc2(iseq, sizeof(VALUE), argc);
4464 }
4465
4466 /* copy operands */
4467 list = seq_list;
4468 for (i = 0; i < size; i++) {
4469 iobj = (INSN *)list;
4470 MEMCPY(ptr, iobj->operands, VALUE, iobj->operand_size);
4471 ptr += iobj->operand_size;
4472 list = list->next;
4473 }
4474
4475 return new_insn_core(iseq, iobj->insn_info.line_no, iobj->insn_info.node_id, insn_id, argc, operands);
4476}
4477#endif
4478
4479/*
4480 * This scheme can get more performance if do this optimize with
4481 * label address resolving.
4482 * It's future work (if compile time was bottle neck).
4483 */
4484static int
4485iseq_insns_unification(rb_iseq_t *iseq, LINK_ANCHOR *const anchor)
4486{
4487#if OPT_INSTRUCTIONS_UNIFICATION
4488 LINK_ELEMENT *list;
4489 INSN *iobj, *niobj;
4490 int id, k;
4491 intptr_t j;
4492
4493 list = FIRST_ELEMENT(anchor);
4494 while (list) {
4495 if (IS_INSN(list)) {
4496 iobj = (INSN *)list;
4497 id = iobj->insn_id;
4498 if (unified_insns_data[id] != 0) {
4499 const int *const *entry = unified_insns_data[id];
4500 for (j = 1; j < (intptr_t)entry[0]; j++) {
4501 const int *unified = entry[j];
4502 LINK_ELEMENT *li = list->next;
4503 for (k = 2; k < unified[1]; k++) {
4504 if (!IS_INSN(li) ||
4505 ((INSN *)li)->insn_id != unified[k]) {
4506 goto miss;
4507 }
4508 li = li->next;
4509 }
4510 /* matched */
4511 niobj =
4512 new_unified_insn(iseq, unified[0], unified[1] - 1,
4513 list);
4514
4515 /* insert to list */
4516 niobj->link.prev = (LINK_ELEMENT *)iobj->link.prev;
4517 niobj->link.next = li;
4518 if (li) {
4519 li->prev = (LINK_ELEMENT *)niobj;
4520 }
4521
4522 list->prev->next = (LINK_ELEMENT *)niobj;
4523 list = (LINK_ELEMENT *)niobj;
4524 break;
4525 miss:;
4526 }
4527 }
4528 }
4529 list = list->next;
4530 }
4531#endif
4532 return COMPILE_OK;
4533}
4534
4535static int
4536all_string_result_p(const NODE *node)
4537{
4538 if (!node) return FALSE;
4539 switch (nd_type(node)) {
4540 case NODE_STR: case NODE_DSTR: case NODE_FILE:
4541 return TRUE;
4542 case NODE_IF: case NODE_UNLESS:
4543 if (!RNODE_IF(node)->nd_body || !RNODE_IF(node)->nd_else) return FALSE;
4544 if (all_string_result_p(RNODE_IF(node)->nd_body))
4545 return all_string_result_p(RNODE_IF(node)->nd_else);
4546 return FALSE;
4547 case NODE_AND: case NODE_OR:
4548 if (!RNODE_AND(node)->nd_2nd)
4549 return all_string_result_p(RNODE_AND(node)->nd_1st);
4550 if (!all_string_result_p(RNODE_AND(node)->nd_1st))
4551 return FALSE;
4552 return all_string_result_p(RNODE_AND(node)->nd_2nd);
4553 default:
4554 return FALSE;
4555 }
4556}
4557
4559 rb_iseq_t *const iseq;
4560 LINK_ANCHOR *const ret;
4561 VALUE lit;
4562 const NODE *lit_node;
4563 int cnt;
4564 int dregx;
4565};
4566
4567static int
4568append_dstr_fragment(struct dstr_ctxt *args, const NODE *const node, rb_parser_string_t *str)
4569{
4570 VALUE s = rb_str_new_mutable_parser_string(str);
4571 if (args->dregx) {
4572 VALUE error = rb_reg_check_preprocess(s);
4573 if (!NIL_P(error)) {
4574 COMPILE_ERROR(args->iseq, nd_line(node), "%" PRIsVALUE, error);
4575 return COMPILE_NG;
4576 }
4577 }
4578 if (NIL_P(args->lit)) {
4579 args->lit = s;
4580 args->lit_node = node;
4581 }
4582 else {
4583 rb_str_buf_append(args->lit, s);
4584 }
4585 return COMPILE_OK;
4586}
4587
4588static void
4589flush_dstr_fragment(struct dstr_ctxt *args)
4590{
4591 if (!NIL_P(args->lit)) {
4592 rb_iseq_t *iseq = args->iseq;
4593 VALUE lit = args->lit;
4594 args->lit = Qnil;
4595 lit = rb_fstring(lit);
4596 ADD_INSN1(args->ret, args->lit_node, putobject, lit);
4597 RB_OBJ_WRITTEN(args->iseq, Qundef, lit);
4598 args->cnt++;
4599 }
4600}
4601
4602static int
4603compile_dstr_fragments_0(struct dstr_ctxt *args, const NODE *const node)
4604{
4605 const struct RNode_LIST *list = RNODE_DSTR(node)->nd_next;
4606 rb_parser_string_t *str = RNODE_DSTR(node)->string;
4607
4608 if (str) {
4609 CHECK(append_dstr_fragment(args, node, str));
4610 }
4611
4612 while (list) {
4613 const NODE *const head = list->nd_head;
4614 if (nd_type_p(head, NODE_STR)) {
4615 CHECK(append_dstr_fragment(args, node, RNODE_STR(head)->string));
4616 }
4617 else if (nd_type_p(head, NODE_DSTR)) {
4618 CHECK(compile_dstr_fragments_0(args, head));
4619 }
4620 else {
4621 flush_dstr_fragment(args);
4622 rb_iseq_t *iseq = args->iseq;
4623 CHECK(COMPILE(args->ret, "each string", head));
4624 args->cnt++;
4625 }
4626 list = (struct RNode_LIST *)list->nd_next;
4627 }
4628 return COMPILE_OK;
4629}
4630
4631static int
4632compile_dstr_fragments(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, int *cntp, int dregx)
4633{
4634 struct dstr_ctxt args = {
4635 .iseq = iseq, .ret = ret,
4636 .lit = Qnil, .lit_node = NULL,
4637 .cnt = 0, .dregx = dregx,
4638 };
4639 CHECK(compile_dstr_fragments_0(&args, node));
4640 flush_dstr_fragment(&args);
4641
4642 *cntp = args.cnt;
4643
4644 return COMPILE_OK;
4645}
4646
4647static int
4648compile_block(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *node, int popped)
4649{
4650 while (node && nd_type_p(node, NODE_BLOCK)) {
4651 CHECK(COMPILE_(ret, "BLOCK body", RNODE_BLOCK(node)->nd_head,
4652 (RNODE_BLOCK(node)->nd_next ? 1 : popped)));
4653 node = RNODE_BLOCK(node)->nd_next;
4654 }
4655 if (node) {
4656 CHECK(COMPILE_(ret, "BLOCK next", RNODE_BLOCK(node)->nd_next, popped));
4657 }
4658 return COMPILE_OK;
4659}
4660
4661static int
4662compile_dstr(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node)
4663{
4664 int cnt;
4665 if (!RNODE_DSTR(node)->nd_next) {
4666 VALUE lit = rb_node_dstr_string_val(node);
4667 ADD_INSN1(ret, node, putstring, lit);
4668 RB_OBJ_WRITTEN(iseq, Qundef, lit);
4669 }
4670 else {
4671 CHECK(compile_dstr_fragments(iseq, ret, node, &cnt, FALSE));
4672 ADD_INSN1(ret, node, concatstrings, INT2FIX(cnt));
4673 }
4674 return COMPILE_OK;
4675}
4676
4677static int
4678compile_dregx(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, int popped)
4679{
4680 int cnt;
4681 int cflag = (int)RNODE_DREGX(node)->as.nd_cflag;
4682
4683 if (!RNODE_DREGX(node)->nd_next) {
4684 if (!popped) {
4685 VALUE src = rb_node_dregx_string_val(node);
4686 VALUE match = rb_reg_compile(src, cflag, NULL, 0);
4687 ADD_INSN1(ret, node, putobject, match);
4688 RB_OBJ_WRITTEN(iseq, Qundef, match);
4689 }
4690 return COMPILE_OK;
4691 }
4692
4693 CHECK(compile_dstr_fragments(iseq, ret, node, &cnt, TRUE));
4694 ADD_INSN2(ret, node, toregexp, INT2FIX(cflag), INT2FIX(cnt));
4695
4696 if (popped) {
4697 ADD_INSN(ret, node, pop);
4698 }
4699
4700 return COMPILE_OK;
4701}
4702
4703static int
4704compile_flip_flop(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, int again,
4705 LABEL *then_label, LABEL *else_label)
4706{
4707 const int line = nd_line(node);
4708 LABEL *lend = NEW_LABEL(line);
4709 rb_num_t cnt = ISEQ_FLIP_CNT_INCREMENT(ISEQ_BODY(iseq)->local_iseq)
4710 + VM_SVAR_FLIPFLOP_START;
4711 VALUE key = INT2FIX(cnt);
4712
4713 ADD_INSN2(ret, node, getspecial, key, INT2FIX(0));
4714 ADD_INSNL(ret, node, branchif, lend);
4715
4716 /* *flip == 0 */
4717 CHECK(COMPILE(ret, "flip2 beg", RNODE_FLIP2(node)->nd_beg));
4718 ADD_INSNL(ret, node, branchunless, else_label);
4719 ADD_INSN1(ret, node, putobject, Qtrue);
4720 ADD_INSN1(ret, node, setspecial, key);
4721 if (!again) {
4722 ADD_INSNL(ret, node, jump, then_label);
4723 }
4724
4725 /* *flip == 1 */
4726 ADD_LABEL(ret, lend);
4727 CHECK(COMPILE(ret, "flip2 end", RNODE_FLIP2(node)->nd_end));
4728 ADD_INSNL(ret, node, branchunless, then_label);
4729 ADD_INSN1(ret, node, putobject, Qfalse);
4730 ADD_INSN1(ret, node, setspecial, key);
4731 ADD_INSNL(ret, node, jump, then_label);
4732
4733 return COMPILE_OK;
4734}
4735
4736static int
4737compile_branch_condition(rb_iseq_t *iseq, LINK_ANCHOR *ret, const NODE *cond,
4738 LABEL *then_label, LABEL *else_label);
4739
4740#define COMPILE_SINGLE 2
4741static int
4742compile_logical(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *cond,
4743 LABEL *then_label, LABEL *else_label)
4744{
4745 DECL_ANCHOR(seq);
4746 INIT_ANCHOR(seq);
4747 LABEL *label = NEW_LABEL(nd_line(cond));
4748 if (!then_label) then_label = label;
4749 else if (!else_label) else_label = label;
4750
4751 CHECK(compile_branch_condition(iseq, seq, cond, then_label, else_label));
4752
4753 if (LIST_INSN_SIZE_ONE(seq)) {
4754 INSN *insn = (INSN *)ELEM_FIRST_INSN(FIRST_ELEMENT(seq));
4755 if (insn->insn_id == BIN(jump) && (LABEL *)(insn->operands[0]) == label)
4756 return COMPILE_OK;
4757 }
4758 if (!label->refcnt) {
4759 return COMPILE_SINGLE;
4760 }
4761 ADD_LABEL(seq, label);
4762 ADD_SEQ(ret, seq);
4763 return COMPILE_OK;
4764}
4765
4766static int
4767compile_branch_condition(rb_iseq_t *iseq, LINK_ANCHOR *ret, const NODE *cond,
4768 LABEL *then_label, LABEL *else_label)
4769{
4770 int ok;
4771 DECL_ANCHOR(ignore);
4772
4773 again:
4774 switch (nd_type(cond)) {
4775 case NODE_AND:
4776 CHECK(ok = compile_logical(iseq, ret, RNODE_AND(cond)->nd_1st, NULL, else_label));
4777 cond = RNODE_AND(cond)->nd_2nd;
4778 if (ok == COMPILE_SINGLE) {
4779 INIT_ANCHOR(ignore);
4780 ret = ignore;
4781 then_label = NEW_LABEL(nd_line(cond));
4782 }
4783 goto again;
4784 case NODE_OR:
4785 CHECK(ok = compile_logical(iseq, ret, RNODE_OR(cond)->nd_1st, then_label, NULL));
4786 cond = RNODE_OR(cond)->nd_2nd;
4787 if (ok == COMPILE_SINGLE) {
4788 INIT_ANCHOR(ignore);
4789 ret = ignore;
4790 else_label = NEW_LABEL(nd_line(cond));
4791 }
4792 goto again;
4793 case NODE_SYM:
4794 case NODE_LINE:
4795 case NODE_FILE:
4796 case NODE_ENCODING:
4797 case NODE_INTEGER: /* NODE_INTEGER is always true */
4798 case NODE_FLOAT: /* NODE_FLOAT is always true */
4799 case NODE_RATIONAL: /* NODE_RATIONAL is always true */
4800 case NODE_IMAGINARY: /* NODE_IMAGINARY is always true */
4801 case NODE_TRUE:
4802 case NODE_STR:
4803 case NODE_REGX:
4804 case NODE_ZLIST:
4805 case NODE_LAMBDA:
4806 /* printf("useless condition eliminate (%s)\n", ruby_node_name(nd_type(cond))); */
4807 ADD_INSNL(ret, cond, jump, then_label);
4808 return COMPILE_OK;
4809 case NODE_FALSE:
4810 case NODE_NIL:
4811 /* printf("useless condition eliminate (%s)\n", ruby_node_name(nd_type(cond))); */
4812 ADD_INSNL(ret, cond, jump, else_label);
4813 return COMPILE_OK;
4814 case NODE_LIST:
4815 case NODE_ARGSCAT:
4816 case NODE_DREGX:
4817 case NODE_DSTR:
4818 CHECK(COMPILE_POPPED(ret, "branch condition", cond));
4819 ADD_INSNL(ret, cond, jump, then_label);
4820 return COMPILE_OK;
4821 case NODE_FLIP2:
4822 CHECK(compile_flip_flop(iseq, ret, cond, TRUE, then_label, else_label));
4823 return COMPILE_OK;
4824 case NODE_FLIP3:
4825 CHECK(compile_flip_flop(iseq, ret, cond, FALSE, then_label, else_label));
4826 return COMPILE_OK;
4827 case NODE_DEFINED:
4828 CHECK(compile_defined_expr(iseq, ret, cond, Qfalse, ret == ignore));
4829 break;
4830 default:
4831 {
4832 DECL_ANCHOR(cond_seq);
4833 INIT_ANCHOR(cond_seq);
4834
4835 CHECK(COMPILE(cond_seq, "branch condition", cond));
4836
4837 if (LIST_INSN_SIZE_ONE(cond_seq)) {
4838 INSN *insn = (INSN *)ELEM_FIRST_INSN(FIRST_ELEMENT(cond_seq));
4839 if (insn->insn_id == BIN(putobject)) {
4840 if (RTEST(insn->operands[0])) {
4841 ADD_INSNL(ret, cond, jump, then_label);
4842 // maybe unreachable
4843 return COMPILE_OK;
4844 }
4845 else {
4846 ADD_INSNL(ret, cond, jump, else_label);
4847 return COMPILE_OK;
4848 }
4849 }
4850 }
4851 ADD_SEQ(ret, cond_seq);
4852 }
4853 break;
4854 }
4855
4856 ADD_INSNL(ret, cond, branchunless, else_label);
4857 ADD_INSNL(ret, cond, jump, then_label);
4858 return COMPILE_OK;
4859}
4860
4861#define HASH_BRACE 1
4862
4863static int
4864keyword_node_p(const NODE *const node)
4865{
4866 return nd_type_p(node, NODE_HASH) && (RNODE_HASH(node)->nd_brace & HASH_BRACE) != HASH_BRACE;
4867}
4868
4869static VALUE
4870get_symbol_value(rb_iseq_t *iseq, const NODE *node)
4871{
4872 switch (nd_type(node)) {
4873 case NODE_SYM:
4874 return rb_node_sym_string_val(node);
4875 default:
4876 UNKNOWN_NODE("get_symbol_value", node, Qnil);
4877 }
4878}
4879
4880static VALUE
4881node_hash_unique_key_index(rb_iseq_t *iseq, rb_node_hash_t *node_hash, int *count_ptr)
4882{
4883 NODE *node = node_hash->nd_head;
4884 VALUE hash = rb_hash_new();
4885 VALUE ary = rb_ary_new();
4886
4887 for (int i = 0; node != NULL; i++, node = RNODE_LIST(RNODE_LIST(node)->nd_next)->nd_next) {
4888 VALUE key = get_symbol_value(iseq, RNODE_LIST(node)->nd_head);
4889 VALUE idx = rb_hash_aref(hash, key);
4890 if (!NIL_P(idx)) {
4891 rb_ary_store(ary, FIX2INT(idx), Qfalse);
4892 (*count_ptr)--;
4893 }
4894 rb_hash_aset(hash, key, INT2FIX(i));
4895 rb_ary_store(ary, i, Qtrue);
4896 (*count_ptr)++;
4897 }
4898
4899 return ary;
4900}
4901
4902static int
4903compile_keyword_arg(rb_iseq_t *iseq, LINK_ANCHOR *const ret,
4904 const NODE *const root_node,
4905 struct rb_callinfo_kwarg **const kw_arg_ptr,
4906 unsigned int *flag)
4907{
4908 RUBY_ASSERT(nd_type_p(root_node, NODE_HASH));
4909 RUBY_ASSERT(kw_arg_ptr != NULL);
4910 RUBY_ASSERT(flag != NULL);
4911
4912 if (RNODE_HASH(root_node)->nd_head && nd_type_p(RNODE_HASH(root_node)->nd_head, NODE_LIST)) {
4913 const NODE *node = RNODE_HASH(root_node)->nd_head;
4914 int seen_nodes = 0;
4915
4916 while (node) {
4917 const NODE *key_node = RNODE_LIST(node)->nd_head;
4918 seen_nodes++;
4919
4920 RUBY_ASSERT(nd_type_p(node, NODE_LIST));
4921 if (key_node && nd_type_p(key_node, NODE_SYM)) {
4922 /* can be keywords */
4923 }
4924 else {
4925 if (flag) {
4926 *flag |= VM_CALL_KW_SPLAT;
4927 if (seen_nodes > 1 || RNODE_LIST(RNODE_LIST(node)->nd_next)->nd_next) {
4928 /* A new hash will be created for the keyword arguments
4929 * in this case, so mark the method as passing mutable
4930 * keyword splat.
4931 */
4932 *flag |= VM_CALL_KW_SPLAT_MUT;
4933 }
4934 }
4935 return FALSE;
4936 }
4937 node = RNODE_LIST(node)->nd_next; /* skip value node */
4938 node = RNODE_LIST(node)->nd_next;
4939 }
4940
4941 /* may be keywords */
4942 node = RNODE_HASH(root_node)->nd_head;
4943 {
4944 int len = 0;
4945 VALUE key_index = node_hash_unique_key_index(iseq, RNODE_HASH(root_node), &len);
4946 struct rb_callinfo_kwarg *kw_arg =
4947 rb_xmalloc_mul_add(len, sizeof(VALUE), sizeof(struct rb_callinfo_kwarg));
4948 VALUE *keywords = kw_arg->keywords;
4949 int i = 0;
4950 int j = 0;
4951 kw_arg->references = 0;
4952 kw_arg->keyword_len = len;
4953
4954 *kw_arg_ptr = kw_arg;
4955
4956 for (i=0; node != NULL; i++, node = RNODE_LIST(RNODE_LIST(node)->nd_next)->nd_next) {
4957 const NODE *key_node = RNODE_LIST(node)->nd_head;
4958 const NODE *val_node = RNODE_LIST(RNODE_LIST(node)->nd_next)->nd_head;
4959 int popped = TRUE;
4960 if (rb_ary_entry(key_index, i)) {
4961 keywords[j] = get_symbol_value(iseq, key_node);
4962 j++;
4963 popped = FALSE;
4964 }
4965 NO_CHECK(COMPILE_(ret, "keyword values", val_node, popped));
4966 }
4967 RUBY_ASSERT(j == len);
4968 return TRUE;
4969 }
4970 }
4971 return FALSE;
4972}
4973
4974static int
4975compile_args(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *node, NODE **kwnode_ptr)
4976{
4977 int len = 0;
4978
4979 for (; node; len++, node = RNODE_LIST(node)->nd_next) {
4980 if (CPDEBUG > 0) {
4981 EXPECT_NODE("compile_args", node, NODE_LIST, -1);
4982 }
4983
4984 if (RNODE_LIST(node)->nd_next == NULL && keyword_node_p(RNODE_LIST(node)->nd_head)) { /* last node is kwnode */
4985 *kwnode_ptr = RNODE_LIST(node)->nd_head;
4986 }
4987 else {
4988 RUBY_ASSERT(!keyword_node_p(RNODE_LIST(node)->nd_head));
4989 NO_CHECK(COMPILE_(ret, "array element", RNODE_LIST(node)->nd_head, FALSE));
4990 }
4991 }
4992
4993 return len;
4994}
4995
4996static inline bool
4997frozen_string_literal_p(const rb_iseq_t *iseq)
4998{
4999 return ISEQ_COMPILE_DATA(iseq)->option->frozen_string_literal > 0;
5000}
5001
5002static inline bool
5003static_literal_node_p(const NODE *node, const rb_iseq_t *iseq, bool hash_key)
5004{
5005 switch (nd_type(node)) {
5006 case NODE_SYM:
5007 case NODE_REGX:
5008 case NODE_LINE:
5009 case NODE_ENCODING:
5010 case NODE_INTEGER:
5011 case NODE_FLOAT:
5012 case NODE_RATIONAL:
5013 case NODE_IMAGINARY:
5014 case NODE_NIL:
5015 case NODE_TRUE:
5016 case NODE_FALSE:
5017 return TRUE;
5018 case NODE_STR:
5019 case NODE_FILE:
5020 return hash_key || frozen_string_literal_p(iseq);
5021 default:
5022 return FALSE;
5023 }
5024}
5025
5026static inline VALUE
5027static_literal_value(const NODE *node, rb_iseq_t *iseq)
5028{
5029 switch (nd_type(node)) {
5030 case NODE_INTEGER:
5031 return rb_node_integer_literal_val(node);
5032 case NODE_FLOAT:
5033 return rb_node_float_literal_val(node);
5034 case NODE_RATIONAL:
5035 return rb_node_rational_literal_val(node);
5036 case NODE_IMAGINARY:
5037 return rb_node_imaginary_literal_val(node);
5038 case NODE_NIL:
5039 return Qnil;
5040 case NODE_TRUE:
5041 return Qtrue;
5042 case NODE_FALSE:
5043 return Qfalse;
5044 case NODE_SYM:
5045 return rb_node_sym_string_val(node);
5046 case NODE_REGX:
5047 return rb_node_regx_string_val(node);
5048 case NODE_LINE:
5049 return rb_node_line_lineno_val(node);
5050 case NODE_ENCODING:
5051 return rb_node_encoding_val(node);
5052 case NODE_FILE:
5053 case NODE_STR:
5054 if (ISEQ_COMPILE_DATA(iseq)->option->debug_frozen_string_literal || RTEST(ruby_debug)) {
5055 VALUE lit = get_string_value(node);
5056 return rb_str_with_debug_created_info(lit, rb_iseq_path(iseq), (int)nd_line(node));
5057 }
5058 else {
5059 return get_string_value(node);
5060 }
5061 default:
5062 rb_bug("unexpected node: %s", ruby_node_name(nd_type(node)));
5063 }
5064}
5065
5066static int
5067compile_array(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *node, int popped, bool first_chunk)
5068{
5069 const NODE *line_node = node;
5070
5071 if (nd_type_p(node, NODE_ZLIST)) {
5072 if (!popped) {
5073 ADD_INSN1(ret, line_node, newarray, INT2FIX(0));
5074 }
5075 return 0;
5076 }
5077
5078 EXPECT_NODE("compile_array", node, NODE_LIST, -1);
5079
5080 if (popped) {
5081 for (; node; node = RNODE_LIST(node)->nd_next) {
5082 NO_CHECK(COMPILE_(ret, "array element", RNODE_LIST(node)->nd_head, popped));
5083 }
5084 return 1;
5085 }
5086
5087 /* Compilation of an array literal.
5088 * The following code is essentially the same as:
5089 *
5090 * for (int count = 0; node; count++; node->nd_next) {
5091 * compile(node->nd_head);
5092 * }
5093 * ADD_INSN(newarray, count);
5094 *
5095 * However, there are three points.
5096 *
5097 * - The code above causes stack overflow for a big string literal.
5098 * The following limits the stack length up to max_stack_len.
5099 *
5100 * [x1,x2,...,x10000] =>
5101 * push x1 ; push x2 ; ...; push x256; newarray 256;
5102 * push x257; push x258; ...; push x512; pushtoarray 256;
5103 * push x513; push x514; ...; push x768; pushtoarray 256;
5104 * ...
5105 *
5106 * - Long subarray can be optimized by pre-allocating a hidden array.
5107 *
5108 * [1,2,3,...,100] =>
5109 * duparray [1,2,3,...,100]
5110 *
5111 * [x, 1,2,3,...,100, z] =>
5112 * push x; newarray 1;
5113 * putobject [1,2,3,...,100] (<- hidden array); concattoarray;
5114 * push z; pushtoarray 1;
5115 *
5116 * - If the last element is a keyword, pushtoarraykwsplat should be emitted
5117 * to only push it onto the array if it is not empty
5118 * (Note: a keyword is NODE_HASH which is not static_literal_node_p.)
5119 *
5120 * [1,2,3,**kw] =>
5121 * putobject 1; putobject 2; putobject 3; newarray 3; ...; pushtoarraykwsplat kw
5122 */
5123
5124 const int max_stack_len = 0x100;
5125 const int min_tmp_ary_len = 0x40;
5126 int stack_len = 0;
5127
5128 /* Either create a new array, or push to the existing array */
5129#define FLUSH_CHUNK \
5130 if (stack_len) { \
5131 if (first_chunk) ADD_INSN1(ret, line_node, newarray, INT2FIX(stack_len)); \
5132 else ADD_INSN1(ret, line_node, pushtoarray, INT2FIX(stack_len)); \
5133 first_chunk = FALSE; \
5134 stack_len = 0; \
5135 }
5136
5137 while (node) {
5138 int count = 1;
5139
5140 /* pre-allocation check (this branch can be omittable) */
5141 if (static_literal_node_p(RNODE_LIST(node)->nd_head, iseq, false)) {
5142 /* count the elements that are optimizable */
5143 const NODE *node_tmp = RNODE_LIST(node)->nd_next;
5144 for (; node_tmp && static_literal_node_p(RNODE_LIST(node_tmp)->nd_head, iseq, false); node_tmp = RNODE_LIST(node_tmp)->nd_next)
5145 count++;
5146
5147 if ((first_chunk && stack_len == 0 && !node_tmp) || count >= min_tmp_ary_len) {
5148 /* The literal contains only optimizable elements, or the subarray is long enough */
5149 VALUE ary = rb_ary_hidden_new(count);
5150
5151 /* Create a hidden array */
5152 for (; count; count--, node = RNODE_LIST(node)->nd_next)
5153 rb_ary_push(ary, static_literal_value(RNODE_LIST(node)->nd_head, iseq));
5154 OBJ_FREEZE(ary);
5155
5156 /* Emit optimized code */
5157 FLUSH_CHUNK;
5158 if (first_chunk) {
5159 ADD_INSN1(ret, line_node, duparray, ary);
5160 first_chunk = FALSE;
5161 }
5162 else {
5163 ADD_INSN1(ret, line_node, putobject, ary);
5164 ADD_INSN(ret, line_node, concattoarray);
5165 }
5166 RB_OBJ_WRITTEN(iseq, Qundef, ary);
5167 }
5168 }
5169
5170 /* Base case: Compile "count" elements */
5171 for (; count; count--, node = RNODE_LIST(node)->nd_next) {
5172 if (CPDEBUG > 0) {
5173 EXPECT_NODE("compile_array", node, NODE_LIST, -1);
5174 }
5175
5176 if (!RNODE_LIST(node)->nd_next && keyword_node_p(RNODE_LIST(node)->nd_head)) {
5177 /* Create array or push existing non-keyword elements onto array */
5178 if (stack_len == 0 && first_chunk) {
5179 ADD_INSN1(ret, line_node, newarray, INT2FIX(0));
5180 }
5181 else {
5182 FLUSH_CHUNK;
5183 }
5184 NO_CHECK(COMPILE_(ret, "array element", RNODE_LIST(node)->nd_head, 0));
5185 ADD_INSN(ret, line_node, pushtoarraykwsplat);
5186 return 1;
5187 }
5188 else {
5189 NO_CHECK(COMPILE_(ret, "array element", RNODE_LIST(node)->nd_head, 0));
5190 stack_len++;
5191 }
5192
5193 /* If there are many pushed elements, flush them to avoid stack overflow */
5194 if (stack_len >= max_stack_len) FLUSH_CHUNK;
5195 }
5196 }
5197
5198 FLUSH_CHUNK;
5199#undef FLUSH_CHUNK
5200 return 1;
5201}
5202
5203static inline int
5204static_literal_node_pair_p(const NODE *node, const rb_iseq_t *iseq)
5205{
5206 return RNODE_LIST(node)->nd_head && static_literal_node_p(RNODE_LIST(node)->nd_head, iseq, true) && static_literal_node_p(RNODE_LIST(RNODE_LIST(node)->nd_next)->nd_head, iseq, false);
5207}
5208
5209static int
5210compile_hash(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *node, int method_call_keywords, int popped)
5211{
5212 const NODE *line_node = node;
5213
5214 node = RNODE_HASH(node)->nd_head;
5215
5216 if (!node || nd_type_p(node, NODE_ZLIST)) {
5217 if (!popped) {
5218 ADD_INSN1(ret, line_node, newhash, INT2FIX(0));
5219 }
5220 return 0;
5221 }
5222
5223 EXPECT_NODE("compile_hash", node, NODE_LIST, -1);
5224
5225 if (popped) {
5226 for (; node; node = RNODE_LIST(node)->nd_next) {
5227 NO_CHECK(COMPILE_(ret, "hash element", RNODE_LIST(node)->nd_head, popped));
5228 }
5229 return 1;
5230 }
5231
5232 /* Compilation of a hash literal (or keyword arguments).
5233 * This is very similar to compile_array, but there are some differences:
5234 *
5235 * - It contains key-value pairs. So we need to take every two elements.
5236 * We can assume that the length is always even.
5237 *
5238 * - Merging is done by a method call (id_core_hash_merge_ptr).
5239 * Sometimes we need to insert the receiver, so "anchor" is needed.
5240 * In addition, a method call is much slower than concatarray.
5241 * So it pays only when the subsequence is really long.
5242 * (min_tmp_hash_len must be much larger than min_tmp_ary_len.)
5243 *
5244 * - We need to handle keyword splat: **kw.
5245 * For **kw, the key part (node->nd_head) is NULL, and the value part
5246 * (node->nd_next->nd_head) is "kw".
5247 * The code is a bit difficult to avoid hash allocation for **{}.
5248 */
5249
5250 const int max_stack_len = 0x100;
5251 const int min_tmp_hash_len = 0x800;
5252 int stack_len = 0;
5253 int first_chunk = 1;
5254 DECL_ANCHOR(anchor);
5255 INIT_ANCHOR(anchor);
5256
5257 /* Convert pushed elements to a hash, and merge if needed */
5258#define FLUSH_CHUNK() \
5259 if (stack_len) { \
5260 if (first_chunk) { \
5261 APPEND_LIST(ret, anchor); \
5262 ADD_INSN1(ret, line_node, newhash, INT2FIX(stack_len)); \
5263 } \
5264 else { \
5265 ADD_INSN1(ret, line_node, putspecialobject, INT2FIX(VM_SPECIAL_OBJECT_VMCORE)); \
5266 ADD_INSN(ret, line_node, swap); \
5267 APPEND_LIST(ret, anchor); \
5268 ADD_SEND(ret, line_node, id_core_hash_merge_ptr, INT2FIX(stack_len + 1)); \
5269 } \
5270 INIT_ANCHOR(anchor); \
5271 first_chunk = stack_len = 0; \
5272 }
5273
5274 while (node) {
5275 int count = 1;
5276
5277 /* pre-allocation check (this branch can be omittable) */
5278 if (static_literal_node_pair_p(node, iseq)) {
5279 /* count the elements that are optimizable */
5280 const NODE *node_tmp = RNODE_LIST(RNODE_LIST(node)->nd_next)->nd_next;
5281 for (; node_tmp && static_literal_node_pair_p(node_tmp, iseq); node_tmp = RNODE_LIST(RNODE_LIST(node_tmp)->nd_next)->nd_next)
5282 count++;
5283
5284 if ((first_chunk && stack_len == 0 && !node_tmp) || count >= min_tmp_hash_len) {
5285 /* The literal contains only optimizable elements, or the subsequence is long enough */
5286 VALUE ary = rb_ary_hidden_new(count);
5287
5288 /* Create a hidden hash */
5289 for (; count; count--, node = RNODE_LIST(RNODE_LIST(node)->nd_next)->nd_next) {
5290 VALUE elem[2];
5291 elem[0] = static_literal_value(RNODE_LIST(node)->nd_head, iseq);
5292 elem[1] = static_literal_value(RNODE_LIST(RNODE_LIST(node)->nd_next)->nd_head, iseq);
5293 rb_ary_cat(ary, elem, 2);
5294 }
5295 VALUE hash = rb_hash_new_with_size(RARRAY_LEN(ary) / 2);
5296 rb_hash_bulk_insert(RARRAY_LEN(ary), RARRAY_CONST_PTR(ary), hash);
5297 hash = rb_obj_hide(hash);
5298 OBJ_FREEZE(hash);
5299
5300 /* Emit optimized code */
5301 FLUSH_CHUNK();
5302 if (first_chunk) {
5303 ADD_INSN1(ret, line_node, duphash, hash);
5304 first_chunk = 0;
5305 }
5306 else {
5307 ADD_INSN1(ret, line_node, putspecialobject, INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
5308 ADD_INSN(ret, line_node, swap);
5309
5310 ADD_INSN1(ret, line_node, putobject, hash);
5311
5312 ADD_SEND(ret, line_node, id_core_hash_merge_kwd, INT2FIX(2));
5313 }
5314 RB_OBJ_WRITTEN(iseq, Qundef, hash);
5315 }
5316 }
5317
5318 /* Base case: Compile "count" elements */
5319 for (; count; count--, node = RNODE_LIST(RNODE_LIST(node)->nd_next)->nd_next) {
5320
5321 if (CPDEBUG > 0) {
5322 EXPECT_NODE("compile_hash", node, NODE_LIST, -1);
5323 }
5324
5325 if (RNODE_LIST(node)->nd_head) {
5326 /* Normal key-value pair */
5327 NO_CHECK(COMPILE_(anchor, "hash key element", RNODE_LIST(node)->nd_head, 0));
5328 NO_CHECK(COMPILE_(anchor, "hash value element", RNODE_LIST(RNODE_LIST(node)->nd_next)->nd_head, 0));
5329 stack_len += 2;
5330
5331 /* If there are many pushed elements, flush them to avoid stack overflow */
5332 if (stack_len >= max_stack_len) FLUSH_CHUNK();
5333 }
5334 else {
5335 /* kwsplat case: foo(..., **kw, ...) */
5336 FLUSH_CHUNK();
5337
5338 const NODE *kw = RNODE_LIST(RNODE_LIST(node)->nd_next)->nd_head;
5339 int empty_kw = nd_type_p(kw, NODE_HASH) && (!RNODE_HASH(kw)->nd_head); /* foo( ..., **{}, ...) */
5340 int first_kw = first_chunk && stack_len == 0; /* foo(1,2,3, **kw, ...) */
5341 int last_kw = !RNODE_LIST(RNODE_LIST(node)->nd_next)->nd_next; /* foo( ..., **kw) */
5342 int only_kw = last_kw && first_kw; /* foo(1,2,3, **kw) */
5343
5344 empty_kw = empty_kw || nd_type_p(kw, NODE_NIL); /* foo( ..., **nil, ...) */
5345 if (empty_kw) {
5346 if (only_kw && method_call_keywords) {
5347 /* **{} appears at the only keyword argument in method call,
5348 * so it won't be modified.
5349 * kw is a special NODE_LIT that contains a special empty hash,
5350 * so this emits: putobject {}.
5351 * This is only done for method calls and not for literal hashes,
5352 * because literal hashes should always result in a new hash.
5353 */
5354 NO_CHECK(COMPILE(ret, "keyword splat", kw));
5355 }
5356 else if (first_kw) {
5357 /* **{} appears as the first keyword argument, so it may be modified.
5358 * We need to create a fresh hash object.
5359 */
5360 ADD_INSN1(ret, line_node, newhash, INT2FIX(0));
5361 }
5362 /* Any empty keyword splats that are not the first can be ignored.
5363 * since merging an empty hash into the existing hash is the same
5364 * as not merging it. */
5365 }
5366 else {
5367 if (only_kw && method_call_keywords) {
5368 /* **kw is only keyword argument in method call.
5369 * Use directly. This will be not be flagged as mutable.
5370 * This is only done for method calls and not for literal hashes,
5371 * because literal hashes should always result in a new hash.
5372 */
5373 NO_CHECK(COMPILE(ret, "keyword splat", kw));
5374 }
5375 else {
5376 /* There is more than one keyword argument, or this is not a method
5377 * call. In that case, we need to add an empty hash (if first keyword),
5378 * or merge the hash to the accumulated hash (if not the first keyword).
5379 */
5380 ADD_INSN1(ret, line_node, putspecialobject, INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
5381 if (first_kw) ADD_INSN1(ret, line_node, newhash, INT2FIX(0));
5382 else ADD_INSN(ret, line_node, swap);
5383
5384 NO_CHECK(COMPILE(ret, "keyword splat", kw));
5385
5386 ADD_SEND(ret, line_node, id_core_hash_merge_kwd, INT2FIX(2));
5387 }
5388 }
5389
5390 first_chunk = 0;
5391 }
5392 }
5393 }
5394
5395 FLUSH_CHUNK();
5396#undef FLUSH_CHUNK
5397 return 1;
5398}
5399
5400VALUE
5401rb_node_case_when_optimizable_literal(const NODE *const node)
5402{
5403 switch (nd_type(node)) {
5404 case NODE_INTEGER:
5405 return rb_node_integer_literal_val(node);
5406 case NODE_FLOAT: {
5407 VALUE v = rb_node_float_literal_val(node);
5408 double ival;
5409
5410 if (modf(RFLOAT_VALUE(v), &ival) == 0.0) {
5411 return FIXABLE(ival) ? LONG2FIX((long)ival) : rb_dbl2big(ival);
5412 }
5413 return v;
5414 }
5415 case NODE_RATIONAL:
5416 case NODE_IMAGINARY:
5417 return Qundef;
5418 case NODE_NIL:
5419 return Qnil;
5420 case NODE_TRUE:
5421 return Qtrue;
5422 case NODE_FALSE:
5423 return Qfalse;
5424 case NODE_SYM:
5425 return rb_node_sym_string_val(node);
5426 case NODE_LINE:
5427 return rb_node_line_lineno_val(node);
5428 case NODE_STR:
5429 return rb_node_str_string_val(node);
5430 case NODE_FILE:
5431 return rb_node_file_path_val(node);
5432 }
5433 return Qundef;
5434}
5435
5436static int
5437when_vals(rb_iseq_t *iseq, LINK_ANCHOR *const cond_seq, const NODE *vals,
5438 LABEL *l1, int only_special_literals, VALUE literals)
5439{
5440 while (vals) {
5441 const NODE *val = RNODE_LIST(vals)->nd_head;
5442 VALUE lit = rb_node_case_when_optimizable_literal(val);
5443
5444 if (UNDEF_P(lit)) {
5445 only_special_literals = 0;
5446 }
5447 else if (NIL_P(rb_hash_lookup(literals, lit))) {
5448 rb_hash_aset(literals, lit, (VALUE)(l1) | 1);
5449 }
5450
5451 if (nd_type_p(val, NODE_STR) || nd_type_p(val, NODE_FILE)) {
5452 debugp_param("nd_lit", get_string_value(val));
5453 lit = get_string_value(val);
5454 ADD_INSN1(cond_seq, val, putobject, lit);
5455 RB_OBJ_WRITTEN(iseq, Qundef, lit);
5456 }
5457 else {
5458 if (!COMPILE(cond_seq, "when cond", val)) return -1;
5459 }
5460
5461 // Emit pattern === target
5462 ADD_INSN1(cond_seq, vals, topn, INT2FIX(1));
5463 ADD_CALL(cond_seq, vals, idEqq, INT2FIX(1));
5464 ADD_INSNL(cond_seq, val, branchif, l1);
5465 vals = RNODE_LIST(vals)->nd_next;
5466 }
5467 return only_special_literals;
5468}
5469
5470static int
5471when_splat_vals(rb_iseq_t *iseq, LINK_ANCHOR *const cond_seq, const NODE *vals,
5472 LABEL *l1, int only_special_literals, VALUE literals)
5473{
5474 const NODE *line_node = vals;
5475
5476 switch (nd_type(vals)) {
5477 case NODE_LIST:
5478 if (when_vals(iseq, cond_seq, vals, l1, only_special_literals, literals) < 0)
5479 return COMPILE_NG;
5480 break;
5481 case NODE_SPLAT:
5482 ADD_INSN (cond_seq, line_node, dup);
5483 CHECK(COMPILE(cond_seq, "when splat", RNODE_SPLAT(vals)->nd_head));
5484 ADD_INSN1(cond_seq, line_node, splatarray, Qfalse);
5485 ADD_INSN1(cond_seq, line_node, checkmatch, INT2FIX(VM_CHECKMATCH_TYPE_CASE | VM_CHECKMATCH_ARRAY));
5486 ADD_INSNL(cond_seq, line_node, branchif, l1);
5487 break;
5488 case NODE_ARGSCAT:
5489 CHECK(when_splat_vals(iseq, cond_seq, RNODE_ARGSCAT(vals)->nd_head, l1, only_special_literals, literals));
5490 CHECK(when_splat_vals(iseq, cond_seq, RNODE_ARGSCAT(vals)->nd_body, l1, only_special_literals, literals));
5491 break;
5492 case NODE_ARGSPUSH:
5493 CHECK(when_splat_vals(iseq, cond_seq, RNODE_ARGSPUSH(vals)->nd_head, l1, only_special_literals, literals));
5494 ADD_INSN (cond_seq, line_node, dup);
5495 CHECK(COMPILE(cond_seq, "when argspush body", RNODE_ARGSPUSH(vals)->nd_body));
5496 ADD_INSN1(cond_seq, line_node, checkmatch, INT2FIX(VM_CHECKMATCH_TYPE_CASE));
5497 ADD_INSNL(cond_seq, line_node, branchif, l1);
5498 break;
5499 default:
5500 ADD_INSN (cond_seq, line_node, dup);
5501 CHECK(COMPILE(cond_seq, "when val", vals));
5502 ADD_INSN1(cond_seq, line_node, splatarray, Qfalse);
5503 ADD_INSN1(cond_seq, line_node, checkmatch, INT2FIX(VM_CHECKMATCH_TYPE_CASE | VM_CHECKMATCH_ARRAY));
5504 ADD_INSNL(cond_seq, line_node, branchif, l1);
5505 break;
5506 }
5507 return COMPILE_OK;
5508}
5509
5510/* Multiple Assignment Handling
5511 *
5512 * In order to handle evaluation of multiple assignment such that the left hand side
5513 * is evaluated before the right hand side, we need to process the left hand side
5514 * and see if there are any attributes that need to be assigned, or constants set
5515 * on explicit objects. If so, we add instructions to evaluate the receiver of
5516 * any assigned attributes or constants before we process the right hand side.
5517 *
5518 * For a multiple assignment such as:
5519 *
5520 * l1.m1, l2[0] = r3, r4
5521 *
5522 * We start off evaluating l1 and l2, then we evaluate r3 and r4, then we
5523 * assign the result of r3 to l1.m1, and then the result of r4 to l2.m2.
5524 * On the VM stack, this looks like:
5525 *
5526 * self # putself
5527 * l1 # send
5528 * l1, self # putself
5529 * l1, l2 # send
5530 * l1, l2, 0 # putobject 0
5531 * l1, l2, 0, [r3, r4] # after evaluation of RHS
5532 * l1, l2, 0, [r3, r4], r4, r3 # expandarray
5533 * l1, l2, 0, [r3, r4], r4, r3, l1 # topn 5
5534 * l1, l2, 0, [r3, r4], r4, l1, r3 # swap
5535 * l1, l2, 0, [r3, r4], r4, m1= # send
5536 * l1, l2, 0, [r3, r4], r4 # pop
5537 * l1, l2, 0, [r3, r4], r4, l2 # topn 3
5538 * l1, l2, 0, [r3, r4], r4, l2, 0 # topn 3
5539 * l1, l2, 0, [r3, r4], r4, l2, 0, r4 # topn 2
5540 * l1, l2, 0, [r3, r4], r4, []= # send
5541 * l1, l2, 0, [r3, r4], r4 # pop
5542 * l1, l2, 0, [r3, r4] # pop
5543 * [r3, r4], l2, 0, [r3, r4] # setn 3
5544 * [r3, r4], l2, 0 # pop
5545 * [r3, r4], l2 # pop
5546 * [r3, r4] # pop
5547 *
5548 * This is made more complex when you have to handle splats, post args,
5549 * and arbitrary levels of nesting. You need to keep track of the total
5550 * number of attributes to set, and for each attribute, how many entries
5551 * are on the stack before the final attribute, in order to correctly
5552 * calculate the topn value to use to get the receiver of the attribute
5553 * setter method.
5554 *
5555 * A brief description of the VM stack for simple multiple assignment
5556 * with no splat (rhs_array will not be present if the return value of
5557 * the multiple assignment is not needed):
5558 *
5559 * lhs_attr1, lhs_attr2, ..., rhs_array, ..., rhs_arg2, rhs_arg1
5560 *
5561 * For multiple assignment with splats, while processing the part before
5562 * the splat (splat+post here is an array of the splat and the post arguments):
5563 *
5564 * lhs_attr1, lhs_attr2, ..., rhs_array, splat+post, ..., rhs_arg2, rhs_arg1
5565 *
5566 * When processing the splat and post arguments:
5567 *
5568 * lhs_attr1, lhs_attr2, ..., rhs_array, ..., post_arg2, post_arg1, splat
5569 *
5570 * When processing nested multiple assignment, existing values on the stack
5571 * are kept. So for:
5572 *
5573 * (l1.m1, l2.m2), l3.m3, l4* = [r1, r2], r3, r4
5574 *
5575 * The stack layout would be the following before processing the nested
5576 * multiple assignment:
5577 *
5578 * l1, l2, [[r1, r2], r3, r4], [r4], r3, [r1, r2]
5579 *
5580 * In order to handle this correctly, we need to keep track of the nesting
5581 * level for each attribute assignment, as well as the attribute number
5582 * (left hand side attributes are processed left to right) and number of
5583 * arguments to pass to the setter method. struct masgn_lhs_node tracks
5584 * this information.
5585 *
5586 * We also need to track information for the entire multiple assignment, such
5587 * as the total number of arguments, and the current nesting level, to
5588 * handle both nested multiple assignment as well as cases where the
5589 * rhs is not needed. We also need to keep track of all attribute
5590 * assignments in this, which we do using a linked listed. struct masgn_state
5591 * tracks this information.
5592 */
5593
5595 INSN *before_insn;
5596 struct masgn_lhs_node *next;
5597 const NODE *line_node;
5598 int argn;
5599 int num_args;
5600 int lhs_pos;
5601};
5602
5604 struct masgn_lhs_node *first_memo;
5605 struct masgn_lhs_node *last_memo;
5606 int lhs_level;
5607 int num_args;
5608 bool nested;
5609};
5610
5611static int
5612add_masgn_lhs_node(struct masgn_state *state, int lhs_pos, const NODE *line_node, int argc, INSN *before_insn)
5613{
5614 if (!state) {
5615 rb_bug("no masgn_state");
5616 }
5617
5618 struct masgn_lhs_node *memo;
5619 memo = malloc(sizeof(struct masgn_lhs_node));
5620 if (!memo) {
5621 return COMPILE_NG;
5622 }
5623
5624 memo->before_insn = before_insn;
5625 memo->line_node = line_node;
5626 memo->argn = state->num_args + 1;
5627 memo->num_args = argc;
5628 state->num_args += argc;
5629 memo->lhs_pos = lhs_pos;
5630 memo->next = NULL;
5631 if (!state->first_memo) {
5632 state->first_memo = memo;
5633 }
5634 else {
5635 state->last_memo->next = memo;
5636 }
5637 state->last_memo = memo;
5638
5639 return COMPILE_OK;
5640}
5641
5642static int compile_massign0(rb_iseq_t *iseq, LINK_ANCHOR *const pre, LINK_ANCHOR *const rhs, LINK_ANCHOR *const lhs, LINK_ANCHOR *const post, const NODE *const node, struct masgn_state *state, int popped);
5643
5644static int
5645compile_massign_lhs(rb_iseq_t *iseq, LINK_ANCHOR *const pre, LINK_ANCHOR *const rhs, LINK_ANCHOR *const lhs, LINK_ANCHOR *const post, const NODE *const node, struct masgn_state *state, int lhs_pos)
5646{
5647 switch (nd_type(node)) {
5648 case NODE_ATTRASGN: {
5649 INSN *iobj;
5650 const NODE *line_node = node;
5651
5652 CHECK(COMPILE_POPPED(pre, "masgn lhs (NODE_ATTRASGN)", node));
5653
5654 bool safenav_call = false;
5655 LINK_ELEMENT *insn_element = LAST_ELEMENT(pre);
5656 iobj = (INSN *)get_prev_insn((INSN *)insn_element); /* send insn */
5657 ASSUME(iobj);
5658 ELEM_REMOVE(insn_element);
5659 if (!IS_INSN_ID(iobj, send)) {
5660 safenav_call = true;
5661 iobj = (INSN *)get_prev_insn(iobj);
5662 ELEM_INSERT_NEXT(&iobj->link, insn_element);
5663 }
5664 (pre->last = iobj->link.prev)->next = 0;
5665
5666 const struct rb_callinfo *ci = (struct rb_callinfo *)OPERAND_AT(iobj, 0);
5667 int argc = vm_ci_argc(ci) + 1;
5668 ci = ci_argc_set(iseq, ci, argc);
5669 OPERAND_AT(iobj, 0) = (VALUE)ci;
5670 RB_OBJ_WRITTEN(iseq, Qundef, ci);
5671
5672 if (argc == 1) {
5673 ADD_INSN(lhs, line_node, swap);
5674 }
5675 else {
5676 ADD_INSN1(lhs, line_node, topn, INT2FIX(argc));
5677 }
5678
5679 if (!add_masgn_lhs_node(state, lhs_pos, line_node, argc, (INSN *)LAST_ELEMENT(lhs))) {
5680 return COMPILE_NG;
5681 }
5682
5683 iobj->link.prev = lhs->last;
5684 lhs->last->next = &iobj->link;
5685 for (lhs->last = &iobj->link; lhs->last->next; lhs->last = lhs->last->next);
5686 if (vm_ci_flag(ci) & VM_CALL_ARGS_SPLAT) {
5687 int argc = vm_ci_argc(ci);
5688 bool dupsplat = false;
5689 ci = ci_argc_set(iseq, ci, argc - 1);
5690 if (!(vm_ci_flag(ci) & VM_CALL_ARGS_SPLAT_MUT)) {
5691 /* Given h[*a], _ = ary
5692 * setup_args sets VM_CALL_ARGS_SPLAT and not VM_CALL_ARGS_SPLAT_MUT
5693 * `a` must be dupped, because it will be appended with ary[0]
5694 * Since you are dupping `a`, you can set VM_CALL_ARGS_SPLAT_MUT
5695 */
5696 dupsplat = true;
5697 ci = ci_flag_set(iseq, ci, VM_CALL_ARGS_SPLAT_MUT);
5698 }
5699 OPERAND_AT(iobj, 0) = (VALUE)ci;
5700 RB_OBJ_WRITTEN(iseq, Qundef, iobj);
5701
5702 /* Given: h[*a], h[*b, 1] = ary
5703 * h[*a] uses splatarray false and does not set VM_CALL_ARGS_SPLAT_MUT,
5704 * so this uses splatarray true on a to dup it before using pushtoarray
5705 * h[*b, 1] uses splatarray true and sets VM_CALL_ARGS_SPLAT_MUT,
5706 * so you can use pushtoarray directly
5707 */
5708 int line_no = nd_line(line_node);
5709 int node_id = nd_node_id(line_node);
5710
5711 if (dupsplat) {
5712 INSERT_BEFORE_INSN(iobj, line_no, node_id, swap);
5713 INSERT_BEFORE_INSN1(iobj, line_no, node_id, splatarray, Qtrue);
5714 INSERT_BEFORE_INSN(iobj, line_no, node_id, swap);
5715 }
5716 INSERT_BEFORE_INSN1(iobj, line_no, node_id, pushtoarray, INT2FIX(1));
5717 }
5718 if (!safenav_call) {
5719 ADD_INSN(lhs, line_node, pop);
5720 if (argc != 1) {
5721 ADD_INSN(lhs, line_node, pop);
5722 }
5723 }
5724 for (int i=0; i < argc; i++) {
5725 ADD_INSN(post, line_node, pop);
5726 }
5727 break;
5728 }
5729 case NODE_MASGN: {
5730 DECL_ANCHOR(nest_rhs);
5731 INIT_ANCHOR(nest_rhs);
5732 DECL_ANCHOR(nest_lhs);
5733 INIT_ANCHOR(nest_lhs);
5734
5735 int prev_level = state->lhs_level;
5736 bool prev_nested = state->nested;
5737 state->nested = 1;
5738 state->lhs_level = lhs_pos - 1;
5739 CHECK(compile_massign0(iseq, pre, nest_rhs, nest_lhs, post, node, state, 1));
5740 state->lhs_level = prev_level;
5741 state->nested = prev_nested;
5742
5743 ADD_SEQ(lhs, nest_rhs);
5744 ADD_SEQ(lhs, nest_lhs);
5745 break;
5746 }
5747 case NODE_CDECL:
5748 if (!RNODE_CDECL(node)->nd_vid) {
5749 /* Special handling only needed for expr::C, not for C */
5750 INSN *iobj;
5751
5752 CHECK(COMPILE_POPPED(pre, "masgn lhs (NODE_CDECL)", node));
5753
5754 LINK_ELEMENT *insn_element = LAST_ELEMENT(pre);
5755 iobj = (INSN *)insn_element; /* setconstant insn */
5756 ELEM_REMOVE((LINK_ELEMENT *)get_prev_insn((INSN *)get_prev_insn(iobj)));
5757 ELEM_REMOVE((LINK_ELEMENT *)get_prev_insn(iobj));
5758 ELEM_REMOVE(insn_element);
5759 pre->last = iobj->link.prev;
5760 ADD_ELEM(lhs, (LINK_ELEMENT *)iobj);
5761
5762 if (!add_masgn_lhs_node(state, lhs_pos, node, 1, (INSN *)LAST_ELEMENT(lhs))) {
5763 return COMPILE_NG;
5764 }
5765
5766 ADD_INSN(post, node, pop);
5767 break;
5768 }
5769 /* Fallthrough */
5770 default: {
5771 DECL_ANCHOR(anchor);
5772 INIT_ANCHOR(anchor);
5773 CHECK(COMPILE_POPPED(anchor, "masgn lhs", node));
5774 ELEM_REMOVE(FIRST_ELEMENT(anchor));
5775 ADD_SEQ(lhs, anchor);
5776 }
5777 }
5778
5779 return COMPILE_OK;
5780}
5781
5782static int
5783compile_massign_opt_lhs(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *lhsn)
5784{
5785 if (lhsn) {
5786 CHECK(compile_massign_opt_lhs(iseq, ret, RNODE_LIST(lhsn)->nd_next));
5787 CHECK(compile_massign_lhs(iseq, ret, ret, ret, ret, RNODE_LIST(lhsn)->nd_head, NULL, 0));
5788 }
5789 return COMPILE_OK;
5790}
5791
5792static int
5793compile_massign_opt(rb_iseq_t *iseq, LINK_ANCHOR *const ret,
5794 const NODE *rhsn, const NODE *orig_lhsn)
5795{
5796 VALUE mem[64];
5797 const int memsize = numberof(mem);
5798 int memindex = 0;
5799 int llen = 0, rlen = 0;
5800 int i;
5801 const NODE *lhsn = orig_lhsn;
5802
5803#define MEMORY(v) { \
5804 int i; \
5805 if (memindex == memsize) return 0; \
5806 for (i=0; i<memindex; i++) { \
5807 if (mem[i] == (v)) return 0; \
5808 } \
5809 mem[memindex++] = (v); \
5810}
5811
5812 if (rhsn == 0 || !nd_type_p(rhsn, NODE_LIST)) {
5813 return 0;
5814 }
5815
5816 while (lhsn) {
5817 const NODE *ln = RNODE_LIST(lhsn)->nd_head;
5818 switch (nd_type(ln)) {
5819 case NODE_LASGN:
5820 case NODE_DASGN:
5821 case NODE_IASGN:
5822 case NODE_CVASGN:
5823 MEMORY(get_nd_vid(ln));
5824 break;
5825 default:
5826 return 0;
5827 }
5828 lhsn = RNODE_LIST(lhsn)->nd_next;
5829 llen++;
5830 }
5831
5832 while (rhsn) {
5833 if (llen <= rlen) {
5834 NO_CHECK(COMPILE_POPPED(ret, "masgn val (popped)", RNODE_LIST(rhsn)->nd_head));
5835 }
5836 else {
5837 NO_CHECK(COMPILE(ret, "masgn val", RNODE_LIST(rhsn)->nd_head));
5838 }
5839 rhsn = RNODE_LIST(rhsn)->nd_next;
5840 rlen++;
5841 }
5842
5843 if (llen > rlen) {
5844 for (i=0; i<llen-rlen; i++) {
5845 ADD_INSN(ret, orig_lhsn, putnil);
5846 }
5847 }
5848
5849 compile_massign_opt_lhs(iseq, ret, orig_lhsn);
5850 return 1;
5851}
5852
5853static int
5854compile_massign0(rb_iseq_t *iseq, LINK_ANCHOR *const pre, LINK_ANCHOR *const rhs, LINK_ANCHOR *const lhs, LINK_ANCHOR *const post, const NODE *const node, struct masgn_state *state, int popped)
5855{
5856 const NODE *rhsn = RNODE_MASGN(node)->nd_value;
5857 const NODE *splatn = RNODE_MASGN(node)->nd_args;
5858 const NODE *lhsn = RNODE_MASGN(node)->nd_head;
5859 const NODE *lhsn_count = lhsn;
5860 int lhs_splat = (splatn && NODE_NAMED_REST_P(splatn)) ? 1 : 0;
5861
5862 int llen = 0;
5863 int lpos = 0;
5864
5865 while (lhsn_count) {
5866 llen++;
5867 lhsn_count = RNODE_LIST(lhsn_count)->nd_next;
5868 }
5869 while (lhsn) {
5870 CHECK(compile_massign_lhs(iseq, pre, rhs, lhs, post, RNODE_LIST(lhsn)->nd_head, state, (llen - lpos) + lhs_splat + state->lhs_level));
5871 lpos++;
5872 lhsn = RNODE_LIST(lhsn)->nd_next;
5873 }
5874
5875 if (lhs_splat) {
5876 if (nd_type_p(splatn, NODE_POSTARG)) {
5877 /*a, b, *r, p1, p2 */
5878 const NODE *postn = RNODE_POSTARG(splatn)->nd_2nd;
5879 const NODE *restn = RNODE_POSTARG(splatn)->nd_1st;
5880 int plen = (int)RNODE_LIST(postn)->as.nd_alen;
5881 int ppos = 0;
5882 int flag = 0x02 | (NODE_NAMED_REST_P(restn) ? 0x01 : 0x00);
5883
5884 ADD_INSN2(lhs, splatn, expandarray, INT2FIX(plen), INT2FIX(flag));
5885
5886 if (NODE_NAMED_REST_P(restn)) {
5887 CHECK(compile_massign_lhs(iseq, pre, rhs, lhs, post, restn, state, 1 + plen + state->lhs_level));
5888 }
5889 while (postn) {
5890 CHECK(compile_massign_lhs(iseq, pre, rhs, lhs, post, RNODE_LIST(postn)->nd_head, state, (plen - ppos) + state->lhs_level));
5891 ppos++;
5892 postn = RNODE_LIST(postn)->nd_next;
5893 }
5894 }
5895 else {
5896 /* a, b, *r */
5897 CHECK(compile_massign_lhs(iseq, pre, rhs, lhs, post, splatn, state, 1 + state->lhs_level));
5898 }
5899 }
5900
5901 if (!state->nested) {
5902 NO_CHECK(COMPILE(rhs, "normal masgn rhs", rhsn));
5903 }
5904
5905 if (!popped) {
5906 ADD_INSN(rhs, node, dup);
5907 }
5908 ADD_INSN2(rhs, node, expandarray, INT2FIX(llen), INT2FIX(lhs_splat));
5909 return COMPILE_OK;
5910}
5911
5912static int
5913compile_massign(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, int popped)
5914{
5915 if (!popped || RNODE_MASGN(node)->nd_args || !compile_massign_opt(iseq, ret, RNODE_MASGN(node)->nd_value, RNODE_MASGN(node)->nd_head)) {
5916 struct masgn_state state;
5917 state.lhs_level = popped ? 0 : 1;
5918 state.nested = 0;
5919 state.num_args = 0;
5920 state.first_memo = NULL;
5921 state.last_memo = NULL;
5922
5923 DECL_ANCHOR(pre);
5924 INIT_ANCHOR(pre);
5925 DECL_ANCHOR(rhs);
5926 INIT_ANCHOR(rhs);
5927 DECL_ANCHOR(lhs);
5928 INIT_ANCHOR(lhs);
5929 DECL_ANCHOR(post);
5930 INIT_ANCHOR(post);
5931 int ok = compile_massign0(iseq, pre, rhs, lhs, post, node, &state, popped);
5932
5933 struct masgn_lhs_node *memo = state.first_memo, *tmp_memo;
5934 while (memo) {
5935 VALUE topn_arg = INT2FIX((state.num_args - memo->argn) + memo->lhs_pos);
5936 for (int i = 0; i < memo->num_args; i++) {
5937 INSERT_BEFORE_INSN1(memo->before_insn, nd_line(memo->line_node), nd_node_id(memo->line_node), topn, topn_arg);
5938 }
5939 tmp_memo = memo->next;
5940 free(memo);
5941 memo = tmp_memo;
5942 }
5943 CHECK(ok);
5944
5945 ADD_SEQ(ret, pre);
5946 ADD_SEQ(ret, rhs);
5947 ADD_SEQ(ret, lhs);
5948 if (!popped && state.num_args >= 1) {
5949 /* make sure rhs array is returned before popping */
5950 ADD_INSN1(ret, node, setn, INT2FIX(state.num_args));
5951 }
5952 ADD_SEQ(ret, post);
5953 }
5954 return COMPILE_OK;
5955}
5956
5957static VALUE
5958collect_const_segments(rb_iseq_t *iseq, const NODE *node)
5959{
5960 VALUE arr = rb_ary_new();
5961 for (;;) {
5962 switch (nd_type(node)) {
5963 case NODE_CONST:
5964 rb_ary_unshift(arr, ID2SYM(RNODE_CONST(node)->nd_vid));
5965 return arr;
5966 case NODE_COLON3:
5967 rb_ary_unshift(arr, ID2SYM(RNODE_COLON3(node)->nd_mid));
5968 rb_ary_unshift(arr, ID2SYM(idNULL));
5969 return arr;
5970 case NODE_COLON2:
5971 rb_ary_unshift(arr, ID2SYM(RNODE_COLON2(node)->nd_mid));
5972 node = RNODE_COLON2(node)->nd_head;
5973 break;
5974 default:
5975 return Qfalse;
5976 }
5977 }
5978}
5979
5980static int
5981compile_const_prefix(rb_iseq_t *iseq, const NODE *const node,
5982 LINK_ANCHOR *const pref, LINK_ANCHOR *const body)
5983{
5984 switch (nd_type(node)) {
5985 case NODE_CONST:
5986 debugi("compile_const_prefix - colon", RNODE_CONST(node)->nd_vid);
5987 ADD_INSN1(body, node, putobject, Qtrue);
5988 ADD_INSN1(body, node, getconstant, ID2SYM(RNODE_CONST(node)->nd_vid));
5989 break;
5990 case NODE_COLON3:
5991 debugi("compile_const_prefix - colon3", RNODE_COLON3(node)->nd_mid);
5992 ADD_INSN(body, node, pop);
5993 ADD_INSN1(body, node, putobject, rb_cObject);
5994 ADD_INSN1(body, node, putobject, Qtrue);
5995 ADD_INSN1(body, node, getconstant, ID2SYM(RNODE_COLON3(node)->nd_mid));
5996 break;
5997 case NODE_COLON2:
5998 CHECK(compile_const_prefix(iseq, RNODE_COLON2(node)->nd_head, pref, body));
5999 debugi("compile_const_prefix - colon2", RNODE_COLON2(node)->nd_mid);
6000 ADD_INSN1(body, node, putobject, Qfalse);
6001 ADD_INSN1(body, node, getconstant, ID2SYM(RNODE_COLON2(node)->nd_mid));
6002 break;
6003 default:
6004 CHECK(COMPILE(pref, "const colon2 prefix", node));
6005 break;
6006 }
6007 return COMPILE_OK;
6008}
6009
6010static int
6011compile_cpath(LINK_ANCHOR *const ret, rb_iseq_t *iseq, const NODE *cpath)
6012{
6013 if (nd_type_p(cpath, NODE_COLON3)) {
6014 /* toplevel class ::Foo */
6015 ADD_INSN1(ret, cpath, putobject, rb_cObject);
6016 return VM_DEFINECLASS_FLAG_SCOPED;
6017 }
6018 else if (nd_type_p(cpath, NODE_COLON2) && RNODE_COLON2(cpath)->nd_head) {
6019 /* Bar::Foo */
6020 NO_CHECK(COMPILE(ret, "nd_else->nd_head", RNODE_COLON2(cpath)->nd_head));
6021 return VM_DEFINECLASS_FLAG_SCOPED;
6022 }
6023 else {
6024 /* class at cbase Foo */
6025 ADD_INSN1(ret, cpath, putspecialobject,
6026 INT2FIX(VM_SPECIAL_OBJECT_CONST_BASE));
6027 return 0;
6028 }
6029}
6030
6031static inline int
6032private_recv_p(const NODE *node)
6033{
6034 NODE *recv = get_nd_recv(node);
6035 if (recv && nd_type_p(recv, NODE_SELF)) {
6036 return RNODE_SELF(recv)->nd_state != 0;
6037 }
6038 return 0;
6039}
6040
6041static void
6042defined_expr(rb_iseq_t *iseq, LINK_ANCHOR *const ret,
6043 const NODE *const node, LABEL **lfinish, VALUE needstr, bool ignore);
6044
6045static int
6046compile_call(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, const enum node_type type, const NODE *const line_node, int popped, bool assume_receiver);
6047
6048static void
6049defined_expr0(rb_iseq_t *iseq, LINK_ANCHOR *const ret,
6050 const NODE *const node, LABEL **lfinish, VALUE needstr,
6051 bool keep_result)
6052{
6053 enum defined_type expr_type = DEFINED_NOT_DEFINED;
6054 enum node_type type;
6055 const int line = nd_line(node);
6056 const NODE *line_node = node;
6057
6058 switch (type = nd_type(node)) {
6059
6060 /* easy literals */
6061 case NODE_NIL:
6062 expr_type = DEFINED_NIL;
6063 break;
6064 case NODE_SELF:
6065 expr_type = DEFINED_SELF;
6066 break;
6067 case NODE_TRUE:
6068 expr_type = DEFINED_TRUE;
6069 break;
6070 case NODE_FALSE:
6071 expr_type = DEFINED_FALSE;
6072 break;
6073
6074 case NODE_HASH:
6075 case NODE_LIST:{
6076 const NODE *vals = (nd_type(node) == NODE_HASH) ? RNODE_HASH(node)->nd_head : node;
6077
6078 if (vals) {
6079 do {
6080 if (RNODE_LIST(vals)->nd_head) {
6081 defined_expr0(iseq, ret, RNODE_LIST(vals)->nd_head, lfinish, Qfalse, false);
6082
6083 if (!lfinish[1]) {
6084 lfinish[1] = NEW_LABEL(line);
6085 }
6086 ADD_INSNL(ret, line_node, branchunless, lfinish[1]);
6087 }
6088 } while ((vals = RNODE_LIST(vals)->nd_next) != NULL);
6089 }
6090 }
6091 /* fall through */
6092 case NODE_STR:
6093 case NODE_SYM:
6094 case NODE_REGX:
6095 case NODE_LINE:
6096 case NODE_FILE:
6097 case NODE_ENCODING:
6098 case NODE_INTEGER:
6099 case NODE_FLOAT:
6100 case NODE_RATIONAL:
6101 case NODE_IMAGINARY:
6102 case NODE_ZLIST:
6103 case NODE_AND:
6104 case NODE_OR:
6105 default:
6106 expr_type = DEFINED_EXPR;
6107 break;
6108
6109 case NODE_SPLAT:
6110 defined_expr0(iseq, ret, RNODE_LIST(node)->nd_head, lfinish, Qfalse, false);
6111 if (!lfinish[1]) {
6112 lfinish[1] = NEW_LABEL(line);
6113 }
6114 ADD_INSNL(ret, line_node, branchunless, lfinish[1]);
6115 expr_type = DEFINED_EXPR;
6116 break;
6117
6118 /* variables */
6119 case NODE_LVAR:
6120 case NODE_DVAR:
6121 expr_type = DEFINED_LVAR;
6122 break;
6123
6124#define PUSH_VAL(type) (needstr == Qfalse ? Qtrue : rb_iseq_defined_string(type))
6125 case NODE_IVAR:
6126 ADD_INSN3(ret, line_node, definedivar,
6127 ID2SYM(RNODE_IVAR(node)->nd_vid), get_ivar_ic_value(iseq,RNODE_IVAR(node)->nd_vid), PUSH_VAL(DEFINED_IVAR));
6128 return;
6129
6130 case NODE_GVAR:
6131 ADD_INSN(ret, line_node, putnil);
6132 ADD_INSN3(ret, line_node, defined, INT2FIX(DEFINED_GVAR),
6133 ID2SYM(RNODE_GVAR(node)->nd_vid), PUSH_VAL(DEFINED_GVAR));
6134 return;
6135
6136 case NODE_CVAR:
6137 ADD_INSN(ret, line_node, putnil);
6138 ADD_INSN3(ret, line_node, defined, INT2FIX(DEFINED_CVAR),
6139 ID2SYM(RNODE_CVAR(node)->nd_vid), PUSH_VAL(DEFINED_CVAR));
6140 return;
6141
6142 case NODE_CONST:
6143 ADD_INSN(ret, line_node, putnil);
6144 ADD_INSN3(ret, line_node, defined, INT2FIX(DEFINED_CONST),
6145 ID2SYM(RNODE_CONST(node)->nd_vid), PUSH_VAL(DEFINED_CONST));
6146 return;
6147 case NODE_COLON2:
6148 if (!lfinish[1]) {
6149 lfinish[1] = NEW_LABEL(line);
6150 }
6151 defined_expr0(iseq, ret, RNODE_COLON2(node)->nd_head, lfinish, Qfalse, false);
6152 ADD_INSNL(ret, line_node, branchunless, lfinish[1]);
6153 NO_CHECK(COMPILE(ret, "defined/colon2#nd_head", RNODE_COLON2(node)->nd_head));
6154
6155 if (rb_is_const_id(RNODE_COLON2(node)->nd_mid)) {
6156 ADD_INSN3(ret, line_node, defined, INT2FIX(DEFINED_CONST_FROM),
6157 ID2SYM(RNODE_COLON2(node)->nd_mid), PUSH_VAL(DEFINED_CONST));
6158 }
6159 else {
6160 ADD_INSN3(ret, line_node, defined, INT2FIX(DEFINED_METHOD),
6161 ID2SYM(RNODE_COLON2(node)->nd_mid), PUSH_VAL(DEFINED_METHOD));
6162 }
6163 return;
6164 case NODE_COLON3:
6165 ADD_INSN1(ret, line_node, putobject, rb_cObject);
6166 ADD_INSN3(ret, line_node, defined,
6167 INT2FIX(DEFINED_CONST_FROM), ID2SYM(RNODE_COLON3(node)->nd_mid), PUSH_VAL(DEFINED_CONST));
6168 return;
6169
6170 /* method dispatch */
6171 case NODE_CALL:
6172 case NODE_OPCALL:
6173 case NODE_VCALL:
6174 case NODE_FCALL:
6175 case NODE_ATTRASGN:{
6176 const int explicit_receiver =
6177 (type == NODE_CALL || type == NODE_OPCALL ||
6178 (type == NODE_ATTRASGN && !private_recv_p(node)));
6179
6180 if (get_nd_args(node) || explicit_receiver) {
6181 if (!lfinish[1]) {
6182 lfinish[1] = NEW_LABEL(line);
6183 }
6184 if (!lfinish[2]) {
6185 lfinish[2] = NEW_LABEL(line);
6186 }
6187 }
6188 if (get_nd_args(node)) {
6189 defined_expr0(iseq, ret, get_nd_args(node), lfinish, Qfalse, false);
6190 ADD_INSNL(ret, line_node, branchunless, lfinish[1]);
6191 }
6192 if (explicit_receiver) {
6193 defined_expr0(iseq, ret, get_nd_recv(node), lfinish, Qfalse, true);
6194 switch (nd_type(get_nd_recv(node))) {
6195 case NODE_CALL:
6196 case NODE_OPCALL:
6197 case NODE_VCALL:
6198 case NODE_FCALL:
6199 case NODE_ATTRASGN:
6200 ADD_INSNL(ret, line_node, branchunless, lfinish[2]);
6201 compile_call(iseq, ret, get_nd_recv(node), nd_type(get_nd_recv(node)), line_node, 0, true);
6202 break;
6203 default:
6204 ADD_INSNL(ret, line_node, branchunless, lfinish[1]);
6205 NO_CHECK(COMPILE(ret, "defined/recv", get_nd_recv(node)));
6206 break;
6207 }
6208 if (keep_result) {
6209 ADD_INSN(ret, line_node, dup);
6210 }
6211 ADD_INSN3(ret, line_node, defined, INT2FIX(DEFINED_METHOD),
6212 ID2SYM(get_node_call_nd_mid(node)), PUSH_VAL(DEFINED_METHOD));
6213 }
6214 else {
6215 ADD_INSN(ret, line_node, putself);
6216 if (keep_result) {
6217 ADD_INSN(ret, line_node, dup);
6218 }
6219 ADD_INSN3(ret, line_node, defined, INT2FIX(DEFINED_FUNC),
6220 ID2SYM(get_node_call_nd_mid(node)), PUSH_VAL(DEFINED_METHOD));
6221 }
6222 return;
6223 }
6224
6225 case NODE_YIELD:
6226 ADD_INSN(ret, line_node, putnil);
6227 ADD_INSN3(ret, line_node, defined, INT2FIX(DEFINED_YIELD), 0,
6228 PUSH_VAL(DEFINED_YIELD));
6229 iseq_set_use_block(ISEQ_BODY(iseq)->local_iseq);
6230 return;
6231
6232 case NODE_BACK_REF:
6233 case NODE_NTH_REF:
6234 ADD_INSN(ret, line_node, putnil);
6235 ADD_INSN3(ret, line_node, defined, INT2FIX(DEFINED_REF),
6236 INT2FIX((RNODE_BACK_REF(node)->nd_nth << 1) | (type == NODE_BACK_REF)),
6237 PUSH_VAL(DEFINED_GVAR));
6238 return;
6239
6240 case NODE_SUPER:
6241 case NODE_ZSUPER:
6242 ADD_INSN(ret, line_node, putnil);
6243 ADD_INSN3(ret, line_node, defined, INT2FIX(DEFINED_ZSUPER), 0,
6244 PUSH_VAL(DEFINED_ZSUPER));
6245 return;
6246
6247#undef PUSH_VAL
6248 case NODE_OP_ASGN1:
6249 case NODE_OP_ASGN2:
6250 case NODE_OP_ASGN_OR:
6251 case NODE_OP_ASGN_AND:
6252 case NODE_MASGN:
6253 case NODE_LASGN:
6254 case NODE_DASGN:
6255 case NODE_GASGN:
6256 case NODE_IASGN:
6257 case NODE_CDECL:
6258 case NODE_CVASGN:
6259 case NODE_OP_CDECL:
6260 expr_type = DEFINED_ASGN;
6261 break;
6262 }
6263
6264 RUBY_ASSERT(expr_type != DEFINED_NOT_DEFINED);
6265
6266 if (needstr != Qfalse) {
6267 VALUE str = rb_iseq_defined_string(expr_type);
6268 ADD_INSN1(ret, line_node, putobject, str);
6269 }
6270 else {
6271 ADD_INSN1(ret, line_node, putobject, Qtrue);
6272 }
6273}
6274
6275static void
6276build_defined_rescue_iseq(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const void *unused)
6277{
6278 ADD_SYNTHETIC_INSN(ret, 0, -1, putnil);
6279 iseq_set_exception_local_table(iseq);
6280}
6281
6282static void
6283defined_expr(rb_iseq_t *iseq, LINK_ANCHOR *const ret,
6284 const NODE *const node, LABEL **lfinish, VALUE needstr, bool ignore)
6285{
6286 LINK_ELEMENT *lcur = ret->last;
6287 defined_expr0(iseq, ret, node, lfinish, needstr, false);
6288 if (lfinish[1]) {
6289 int line = nd_line(node);
6290 LABEL *lstart = NEW_LABEL(line);
6291 LABEL *lend = NEW_LABEL(line);
6292 const rb_iseq_t *rescue;
6294 rb_iseq_new_with_callback_new_callback(build_defined_rescue_iseq, NULL);
6295 rescue = NEW_CHILD_ISEQ_WITH_CALLBACK(ifunc,
6296 rb_str_concat(rb_str_new2("defined guard in "),
6297 ISEQ_BODY(iseq)->location.label),
6298 ISEQ_TYPE_RESCUE, 0);
6299 lstart->rescued = LABEL_RESCUE_BEG;
6300 lend->rescued = LABEL_RESCUE_END;
6301 APPEND_LABEL(ret, lcur, lstart);
6302 ADD_LABEL(ret, lend);
6303 if (!ignore) {
6304 ADD_CATCH_ENTRY(CATCH_TYPE_RESCUE, lstart, lend, rescue, lfinish[1]);
6305 }
6306 }
6307}
6308
6309static int
6310compile_defined_expr(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, VALUE needstr, bool ignore)
6311{
6312 const int line = nd_line(node);
6313 const NODE *line_node = node;
6314 if (!RNODE_DEFINED(node)->nd_head) {
6315 VALUE str = rb_iseq_defined_string(DEFINED_NIL);
6316 ADD_INSN1(ret, line_node, putobject, str);
6317 }
6318 else {
6319 LABEL *lfinish[3];
6320 LINK_ELEMENT *last = ret->last;
6321 lfinish[0] = NEW_LABEL(line);
6322 lfinish[1] = 0;
6323 lfinish[2] = 0;
6324 defined_expr(iseq, ret, RNODE_DEFINED(node)->nd_head, lfinish, needstr, ignore);
6325 if (lfinish[1]) {
6326 ELEM_INSERT_NEXT(last, &new_insn_body(iseq, nd_line(line_node), nd_node_id(line_node), BIN(putnil), 0)->link);
6327 ADD_INSN(ret, line_node, swap);
6328 if (lfinish[2]) {
6329 ADD_LABEL(ret, lfinish[2]);
6330 }
6331 ADD_INSN(ret, line_node, pop);
6332 ADD_LABEL(ret, lfinish[1]);
6333 }
6334 ADD_LABEL(ret, lfinish[0]);
6335 }
6336 return COMPILE_OK;
6337}
6338
6339static VALUE
6340make_name_for_block(const rb_iseq_t *orig_iseq)
6341{
6342 int level = 1;
6343 const rb_iseq_t *iseq = orig_iseq;
6344
6345 if (ISEQ_BODY(orig_iseq)->parent_iseq != 0) {
6346 while (ISEQ_BODY(orig_iseq)->local_iseq != iseq) {
6347 if (ISEQ_BODY(iseq)->type == ISEQ_TYPE_BLOCK) {
6348 level++;
6349 }
6350 iseq = ISEQ_BODY(iseq)->parent_iseq;
6351 }
6352 }
6353
6354 if (level == 1) {
6355 return rb_sprintf("block in %"PRIsVALUE, ISEQ_BODY(iseq)->location.label);
6356 }
6357 else {
6358 return rb_sprintf("block (%d levels) in %"PRIsVALUE, level, ISEQ_BODY(iseq)->location.label);
6359 }
6360}
6361
6362static void
6363push_ensure_entry(rb_iseq_t *iseq,
6365 struct ensure_range *er, const void *const node)
6366{
6367 enl->ensure_node = node;
6368 enl->prev = ISEQ_COMPILE_DATA(iseq)->ensure_node_stack; /* prev */
6369 enl->erange = er;
6370 ISEQ_COMPILE_DATA(iseq)->ensure_node_stack = enl;
6371}
6372
6373static void
6374add_ensure_range(rb_iseq_t *iseq, struct ensure_range *erange,
6375 LABEL *lstart, LABEL *lend)
6376{
6377 struct ensure_range *ne =
6378 compile_data_alloc(iseq, sizeof(struct ensure_range));
6379
6380 while (erange->next != 0) {
6381 erange = erange->next;
6382 }
6383 ne->next = 0;
6384 ne->begin = lend;
6385 ne->end = erange->end;
6386 erange->end = lstart;
6387
6388 erange->next = ne;
6389}
6390
6391static bool
6392can_add_ensure_iseq(const rb_iseq_t *iseq)
6393{
6395 if (ISEQ_COMPILE_DATA(iseq)->in_rescue && (e = ISEQ_COMPILE_DATA(iseq)->ensure_node_stack) != NULL) {
6396 while (e) {
6397 if (e->ensure_node) return false;
6398 e = e->prev;
6399 }
6400 }
6401 return true;
6402}
6403
6404static void
6405add_ensure_iseq(LINK_ANCHOR *const ret, rb_iseq_t *iseq, int is_return)
6406{
6407 RUBY_ASSERT(can_add_ensure_iseq(iseq));
6408
6410 ISEQ_COMPILE_DATA(iseq)->ensure_node_stack;
6411 struct iseq_compile_data_ensure_node_stack *prev_enlp = enlp;
6412 DECL_ANCHOR(ensure);
6413
6414 INIT_ANCHOR(ensure);
6415 while (enlp) {
6416 if (enlp->erange != NULL) {
6417 DECL_ANCHOR(ensure_part);
6418 LABEL *lstart = NEW_LABEL(0);
6419 LABEL *lend = NEW_LABEL(0);
6420 INIT_ANCHOR(ensure_part);
6421
6422 add_ensure_range(iseq, enlp->erange, lstart, lend);
6423
6424 ISEQ_COMPILE_DATA(iseq)->ensure_node_stack = enlp->prev;
6425 ADD_LABEL(ensure_part, lstart);
6426 NO_CHECK(COMPILE_POPPED(ensure_part, "ensure part", enlp->ensure_node));
6427 ADD_LABEL(ensure_part, lend);
6428 ADD_SEQ(ensure, ensure_part);
6429 }
6430 else {
6431 if (!is_return) {
6432 break;
6433 }
6434 }
6435 enlp = enlp->prev;
6436 }
6437 ISEQ_COMPILE_DATA(iseq)->ensure_node_stack = prev_enlp;
6438 ADD_SEQ(ret, ensure);
6439}
6440
6441#if RUBY_DEBUG
6442static int
6443check_keyword(const NODE *node)
6444{
6445 /* This check is essentially a code clone of compile_keyword_arg. */
6446
6447 if (nd_type_p(node, NODE_LIST)) {
6448 while (RNODE_LIST(node)->nd_next) {
6449 node = RNODE_LIST(node)->nd_next;
6450 }
6451 node = RNODE_LIST(node)->nd_head;
6452 }
6453
6454 return keyword_node_p(node);
6455}
6456#endif
6457
6458static bool
6459keyword_node_single_splat_p(NODE *kwnode)
6460{
6461 RUBY_ASSERT(keyword_node_p(kwnode));
6462
6463 NODE *node = RNODE_HASH(kwnode)->nd_head;
6464 return RNODE_LIST(node)->nd_head == NULL &&
6465 RNODE_LIST(RNODE_LIST(node)->nd_next)->nd_next == NULL;
6466}
6467
6468static void
6469compile_single_keyword_splat_mutable(rb_iseq_t *iseq, LINK_ANCHOR *const args, const NODE *argn,
6470 NODE *kwnode, unsigned int *flag_ptr)
6471{
6472 *flag_ptr |= VM_CALL_KW_SPLAT_MUT;
6473 ADD_INSN1(args, argn, putspecialobject, INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
6474 ADD_INSN1(args, argn, newhash, INT2FIX(0));
6475 compile_hash(iseq, args, kwnode, TRUE, FALSE);
6476 ADD_SEND(args, argn, id_core_hash_merge_kwd, INT2FIX(2));
6477}
6478
6479#define SPLATARRAY_FALSE 0
6480#define SPLATARRAY_TRUE 1
6481#define DUP_SINGLE_KW_SPLAT 2
6482
6483static int
6484setup_args_core(rb_iseq_t *iseq, LINK_ANCHOR *const args, const NODE *argn,
6485 unsigned int *dup_rest, unsigned int *flag_ptr, struct rb_callinfo_kwarg **kwarg_ptr)
6486{
6487 if (!argn) return 0;
6488
6489 NODE *kwnode = NULL;
6490
6491 switch (nd_type(argn)) {
6492 case NODE_LIST: {
6493 // f(x, y, z)
6494 int len = compile_args(iseq, args, argn, &kwnode);
6495 RUBY_ASSERT(flag_ptr == NULL || (*flag_ptr & VM_CALL_ARGS_SPLAT) == 0);
6496
6497 if (kwnode) {
6498 if (compile_keyword_arg(iseq, args, kwnode, kwarg_ptr, flag_ptr)) {
6499 len -= 1;
6500 }
6501 else {
6502 if (keyword_node_single_splat_p(kwnode) && (*dup_rest & DUP_SINGLE_KW_SPLAT)) {
6503 compile_single_keyword_splat_mutable(iseq, args, argn, kwnode, flag_ptr);
6504 }
6505 else {
6506 compile_hash(iseq, args, kwnode, TRUE, FALSE);
6507 }
6508 }
6509 }
6510
6511 return len;
6512 }
6513 case NODE_SPLAT: {
6514 // f(*a)
6515 NO_CHECK(COMPILE(args, "args (splat)", RNODE_SPLAT(argn)->nd_head));
6516 ADD_INSN1(args, argn, splatarray, RBOOL(*dup_rest & SPLATARRAY_TRUE));
6517 if (*dup_rest & SPLATARRAY_TRUE) *dup_rest &= ~SPLATARRAY_TRUE;
6518 if (flag_ptr) *flag_ptr |= VM_CALL_ARGS_SPLAT;
6519 RUBY_ASSERT(flag_ptr == NULL || (*flag_ptr & VM_CALL_KW_SPLAT) == 0);
6520 return 1;
6521 }
6522 case NODE_ARGSCAT: {
6523 if (flag_ptr) *flag_ptr |= VM_CALL_ARGS_SPLAT;
6524 int argc = setup_args_core(iseq, args, RNODE_ARGSCAT(argn)->nd_head, dup_rest, NULL, NULL);
6525 bool args_pushed = false;
6526
6527 if (nd_type_p(RNODE_ARGSCAT(argn)->nd_body, NODE_LIST)) {
6528 int rest_len = compile_args(iseq, args, RNODE_ARGSCAT(argn)->nd_body, &kwnode);
6529 if (kwnode) rest_len--;
6530 ADD_INSN1(args, argn, pushtoarray, INT2FIX(rest_len));
6531 args_pushed = true;
6532 }
6533 else {
6534 RUBY_ASSERT(!check_keyword(RNODE_ARGSCAT(argn)->nd_body));
6535 NO_CHECK(COMPILE(args, "args (cat: splat)", RNODE_ARGSCAT(argn)->nd_body));
6536 }
6537
6538 if (nd_type_p(RNODE_ARGSCAT(argn)->nd_head, NODE_LIST)) {
6539 ADD_INSN1(args, argn, splatarray, RBOOL(*dup_rest & SPLATARRAY_TRUE));
6540 if (*dup_rest & SPLATARRAY_TRUE) *dup_rest &= ~SPLATARRAY_TRUE;
6541 argc += 1;
6542 }
6543 else if (!args_pushed) {
6544 ADD_INSN(args, argn, concattoarray);
6545 }
6546
6547 // f(..., *a, ..., k1:1, ...) #=> f(..., *[*a, ...], **{k1:1, ...})
6548 if (kwnode) {
6549 // kwsplat
6550 *flag_ptr |= VM_CALL_KW_SPLAT;
6551 compile_hash(iseq, args, kwnode, TRUE, FALSE);
6552 argc += 1;
6553 }
6554
6555 return argc;
6556 }
6557 case NODE_ARGSPUSH: {
6558 if (flag_ptr) *flag_ptr |= VM_CALL_ARGS_SPLAT;
6559 int argc = setup_args_core(iseq, args, RNODE_ARGSPUSH(argn)->nd_head, dup_rest, NULL, NULL);
6560
6561 if (nd_type_p(RNODE_ARGSPUSH(argn)->nd_body, NODE_LIST)) {
6562 int rest_len = compile_args(iseq, args, RNODE_ARGSPUSH(argn)->nd_body, &kwnode);
6563 if (kwnode) rest_len--;
6564 ADD_INSN1(args, argn, newarray, INT2FIX(rest_len));
6565 ADD_INSN1(args, argn, pushtoarray, INT2FIX(1));
6566 }
6567 else {
6568 if (keyword_node_p(RNODE_ARGSPUSH(argn)->nd_body)) {
6569 kwnode = RNODE_ARGSPUSH(argn)->nd_body;
6570 }
6571 else {
6572 NO_CHECK(COMPILE(args, "args (cat: splat)", RNODE_ARGSPUSH(argn)->nd_body));
6573 ADD_INSN1(args, argn, pushtoarray, INT2FIX(1));
6574 }
6575 }
6576
6577 if (kwnode) {
6578 // f(*a, k:1)
6579 *flag_ptr |= VM_CALL_KW_SPLAT;
6580 if (!keyword_node_single_splat_p(kwnode)) {
6581 *flag_ptr |= VM_CALL_KW_SPLAT_MUT;
6582 compile_hash(iseq, args, kwnode, TRUE, FALSE);
6583 }
6584 else if (*dup_rest & DUP_SINGLE_KW_SPLAT) {
6585 compile_single_keyword_splat_mutable(iseq, args, argn, kwnode, flag_ptr);
6586 }
6587 else {
6588 compile_hash(iseq, args, kwnode, TRUE, FALSE);
6589 }
6590 argc += 1;
6591 }
6592
6593 return argc;
6594 }
6595 default: {
6596 UNKNOWN_NODE("setup_arg", argn, Qnil);
6597 }
6598 }
6599}
6600
6601static void
6602setup_args_splat_mut(unsigned int *flag, int dup_rest, int initial_dup_rest)
6603{
6604 if ((*flag & VM_CALL_ARGS_SPLAT) && dup_rest != initial_dup_rest) {
6605 *flag |= VM_CALL_ARGS_SPLAT_MUT;
6606 }
6607}
6608
6609static bool
6610setup_args_dup_rest_p(const NODE *argn)
6611{
6612 switch(nd_type(argn)) {
6613 case NODE_LVAR:
6614 case NODE_DVAR:
6615 case NODE_GVAR:
6616 case NODE_IVAR:
6617 case NODE_CVAR:
6618 case NODE_CONST:
6619 case NODE_COLON3:
6620 case NODE_INTEGER:
6621 case NODE_FLOAT:
6622 case NODE_RATIONAL:
6623 case NODE_IMAGINARY:
6624 case NODE_STR:
6625 case NODE_SYM:
6626 case NODE_REGX:
6627 case NODE_SELF:
6628 case NODE_NIL:
6629 case NODE_TRUE:
6630 case NODE_FALSE:
6631 case NODE_LAMBDA:
6632 case NODE_NTH_REF:
6633 case NODE_BACK_REF:
6634 return false;
6635 case NODE_COLON2:
6636 return setup_args_dup_rest_p(RNODE_COLON2(argn)->nd_head);
6637 default:
6638 return true;
6639 }
6640}
6641
6642static VALUE
6643setup_args(rb_iseq_t *iseq, LINK_ANCHOR *const args, const NODE *argn,
6644 unsigned int *flag, struct rb_callinfo_kwarg **keywords)
6645{
6646 VALUE ret;
6647 unsigned int dup_rest = SPLATARRAY_TRUE, initial_dup_rest;
6648
6649 if (argn) {
6650 const NODE *check_arg = nd_type_p(argn, NODE_BLOCK_PASS) ?
6651 RNODE_BLOCK_PASS(argn)->nd_head : argn;
6652
6653 if (check_arg) {
6654 switch(nd_type(check_arg)) {
6655 case(NODE_SPLAT):
6656 // avoid caller side array allocation for f(*arg)
6657 dup_rest = SPLATARRAY_FALSE;
6658 break;
6659 case(NODE_ARGSCAT):
6660 // avoid caller side array allocation for f(1, *arg)
6661 dup_rest = !nd_type_p(RNODE_ARGSCAT(check_arg)->nd_head, NODE_LIST);
6662 break;
6663 case(NODE_ARGSPUSH):
6664 // avoid caller side array allocation for f(*arg, **hash) and f(1, *arg, **hash)
6665 dup_rest = !((nd_type_p(RNODE_ARGSPUSH(check_arg)->nd_head, NODE_SPLAT) ||
6666 (nd_type_p(RNODE_ARGSPUSH(check_arg)->nd_head, NODE_ARGSCAT) &&
6667 nd_type_p(RNODE_ARGSCAT(RNODE_ARGSPUSH(check_arg)->nd_head)->nd_head, NODE_LIST))) &&
6668 nd_type_p(RNODE_ARGSPUSH(check_arg)->nd_body, NODE_HASH) &&
6669 !RNODE_HASH(RNODE_ARGSPUSH(check_arg)->nd_body)->nd_brace);
6670
6671 if (dup_rest == SPLATARRAY_FALSE) {
6672 // require allocation for keyword key/value/splat that may modify splatted argument
6673 NODE *node = RNODE_HASH(RNODE_ARGSPUSH(check_arg)->nd_body)->nd_head;
6674 while (node) {
6675 NODE *key_node = RNODE_LIST(node)->nd_head;
6676 if (key_node && setup_args_dup_rest_p(key_node)) {
6677 dup_rest = SPLATARRAY_TRUE;
6678 break;
6679 }
6680
6681 node = RNODE_LIST(node)->nd_next;
6682 NODE *value_node = RNODE_LIST(node)->nd_head;
6683 if (setup_args_dup_rest_p(value_node)) {
6684 dup_rest = SPLATARRAY_TRUE;
6685 break;
6686 }
6687
6688 node = RNODE_LIST(node)->nd_next;
6689 }
6690 }
6691 break;
6692 default:
6693 break;
6694 }
6695 }
6696
6697 if (check_arg != argn && setup_args_dup_rest_p(RNODE_BLOCK_PASS(argn)->nd_body)) {
6698 // for block pass that may modify splatted argument, dup rest and kwrest if given
6699 dup_rest = SPLATARRAY_TRUE | DUP_SINGLE_KW_SPLAT;
6700 }
6701 }
6702 initial_dup_rest = dup_rest;
6703
6704 if (argn && nd_type_p(argn, NODE_BLOCK_PASS)) {
6705 DECL_ANCHOR(arg_block);
6706 INIT_ANCHOR(arg_block);
6707
6708 if (RNODE_BLOCK_PASS(argn)->forwarding && ISEQ_BODY(ISEQ_BODY(iseq)->local_iseq)->param.flags.forwardable) {
6709 int idx = ISEQ_BODY(ISEQ_BODY(iseq)->local_iseq)->local_table_size;// - get_local_var_idx(iseq, idDot3);
6710
6711 RUBY_ASSERT(nd_type_p(RNODE_BLOCK_PASS(argn)->nd_head, NODE_ARGSPUSH));
6712 const NODE * arg_node =
6713 RNODE_ARGSPUSH(RNODE_BLOCK_PASS(argn)->nd_head)->nd_head;
6714
6715 int argc = 0;
6716
6717 // Only compile leading args:
6718 // foo(x, y, ...)
6719 // ^^^^
6720 if (nd_type_p(arg_node, NODE_ARGSCAT)) {
6721 argc += setup_args_core(iseq, args, RNODE_ARGSCAT(arg_node)->nd_head, &dup_rest, flag, keywords);
6722 }
6723
6724 *flag |= VM_CALL_FORWARDING;
6725
6726 ADD_GETLOCAL(args, argn, idx, get_lvar_level(iseq));
6727 setup_args_splat_mut(flag, dup_rest, initial_dup_rest);
6728 return INT2FIX(argc);
6729 }
6730 else {
6731 *flag |= VM_CALL_ARGS_BLOCKARG;
6732
6733 NO_CHECK(COMPILE(arg_block, "block", RNODE_BLOCK_PASS(argn)->nd_body));
6734 }
6735
6736 if (LIST_INSN_SIZE_ONE(arg_block)) {
6737 LINK_ELEMENT *elem = FIRST_ELEMENT(arg_block);
6738 if (IS_INSN(elem)) {
6739 INSN *iobj = (INSN *)elem;
6740 if (iobj->insn_id == BIN(getblockparam)) {
6741 iobj->insn_id = BIN(getblockparamproxy);
6742 }
6743 }
6744 }
6745 ret = INT2FIX(setup_args_core(iseq, args, RNODE_BLOCK_PASS(argn)->nd_head, &dup_rest, flag, keywords));
6746 ADD_SEQ(args, arg_block);
6747 }
6748 else {
6749 ret = INT2FIX(setup_args_core(iseq, args, argn, &dup_rest, flag, keywords));
6750 }
6751 setup_args_splat_mut(flag, dup_rest, initial_dup_rest);
6752 return ret;
6753}
6754
6755static void
6756build_postexe_iseq(rb_iseq_t *iseq, LINK_ANCHOR *ret, const void *ptr)
6757{
6758 const NODE *body = ptr;
6759 int line = nd_line(body);
6760 VALUE argc = INT2FIX(0);
6761 const rb_iseq_t *block = NEW_CHILD_ISEQ(body, make_name_for_block(ISEQ_BODY(iseq)->parent_iseq), ISEQ_TYPE_BLOCK, line);
6762
6763 ADD_INSN1(ret, body, putspecialobject, INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
6764 ADD_CALL_WITH_BLOCK(ret, body, id_core_set_postexe, argc, block);
6765 RB_OBJ_WRITTEN(iseq, Qundef, (VALUE)block);
6766 iseq_set_local_table(iseq, 0, 0);
6767}
6768
6769static void
6770compile_named_capture_assign(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node)
6771{
6772 const NODE *vars;
6773 LINK_ELEMENT *last;
6774 int line = nd_line(node);
6775 const NODE *line_node = node;
6776 LABEL *fail_label = NEW_LABEL(line), *end_label = NEW_LABEL(line);
6777
6778#if !(defined(NAMED_CAPTURE_BY_SVAR) && NAMED_CAPTURE_BY_SVAR-0)
6779 ADD_INSN1(ret, line_node, getglobal, ID2SYM(idBACKREF));
6780#else
6781 ADD_INSN2(ret, line_node, getspecial, INT2FIX(1) /* '~' */, INT2FIX(0));
6782#endif
6783 ADD_INSN(ret, line_node, dup);
6784 ADD_INSNL(ret, line_node, branchunless, fail_label);
6785
6786 for (vars = node; vars; vars = RNODE_BLOCK(vars)->nd_next) {
6787 INSN *cap;
6788 if (RNODE_BLOCK(vars)->nd_next) {
6789 ADD_INSN(ret, line_node, dup);
6790 }
6791 last = ret->last;
6792 NO_CHECK(COMPILE_POPPED(ret, "capture", RNODE_BLOCK(vars)->nd_head));
6793 last = last->next; /* putobject :var */
6794 cap = new_insn_send(iseq, nd_line(line_node), nd_node_id(line_node), idAREF, INT2FIX(1),
6795 NULL, INT2FIX(0), NULL);
6796 ELEM_INSERT_PREV(last->next, (LINK_ELEMENT *)cap);
6797#if !defined(NAMED_CAPTURE_SINGLE_OPT) || NAMED_CAPTURE_SINGLE_OPT-0
6798 if (!RNODE_BLOCK(vars)->nd_next && vars == node) {
6799 /* only one name */
6800 DECL_ANCHOR(nom);
6801
6802 INIT_ANCHOR(nom);
6803 ADD_INSNL(nom, line_node, jump, end_label);
6804 ADD_LABEL(nom, fail_label);
6805# if 0 /* $~ must be MatchData or nil */
6806 ADD_INSN(nom, line_node, pop);
6807 ADD_INSN(nom, line_node, putnil);
6808# endif
6809 ADD_LABEL(nom, end_label);
6810 (nom->last->next = cap->link.next)->prev = nom->last;
6811 (cap->link.next = nom->anchor.next)->prev = &cap->link;
6812 return;
6813 }
6814#endif
6815 }
6816 ADD_INSNL(ret, line_node, jump, end_label);
6817 ADD_LABEL(ret, fail_label);
6818 ADD_INSN(ret, line_node, pop);
6819 for (vars = node; vars; vars = RNODE_BLOCK(vars)->nd_next) {
6820 last = ret->last;
6821 NO_CHECK(COMPILE_POPPED(ret, "capture", RNODE_BLOCK(vars)->nd_head));
6822 last = last->next; /* putobject :var */
6823 ((INSN*)last)->insn_id = BIN(putnil);
6824 ((INSN*)last)->operand_size = 0;
6825 }
6826 ADD_LABEL(ret, end_label);
6827}
6828
6829static int
6830optimizable_range_item_p(const NODE *n)
6831{
6832 if (!n) return FALSE;
6833 switch (nd_type(n)) {
6834 case NODE_LINE:
6835 return TRUE;
6836 case NODE_INTEGER:
6837 return TRUE;
6838 case NODE_NIL:
6839 return TRUE;
6840 default:
6841 return FALSE;
6842 }
6843}
6844
6845static VALUE
6846optimized_range_item(const NODE *n)
6847{
6848 switch (nd_type(n)) {
6849 case NODE_LINE:
6850 return rb_node_line_lineno_val(n);
6851 case NODE_INTEGER:
6852 return rb_node_integer_literal_val(n);
6853 case NODE_FLOAT:
6854 return rb_node_float_literal_val(n);
6855 case NODE_RATIONAL:
6856 return rb_node_rational_literal_val(n);
6857 case NODE_IMAGINARY:
6858 return rb_node_imaginary_literal_val(n);
6859 case NODE_NIL:
6860 return Qnil;
6861 default:
6862 rb_bug("unexpected node: %s", ruby_node_name(nd_type(n)));
6863 }
6864}
6865
6866static int
6867compile_if(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, int popped, const enum node_type type)
6868{
6869 const NODE *const node_body = type == NODE_IF ? RNODE_IF(node)->nd_body : RNODE_UNLESS(node)->nd_else;
6870 const NODE *const node_else = type == NODE_IF ? RNODE_IF(node)->nd_else : RNODE_UNLESS(node)->nd_body;
6871
6872 const int line = nd_line(node);
6873 const NODE *line_node = node;
6874 DECL_ANCHOR(cond_seq);
6875 LABEL *then_label, *else_label, *end_label;
6876 VALUE branches = Qfalse;
6877
6878 INIT_ANCHOR(cond_seq);
6879 then_label = NEW_LABEL(line);
6880 else_label = NEW_LABEL(line);
6881 end_label = 0;
6882
6883 NODE *cond = RNODE_IF(node)->nd_cond;
6884 if (nd_type(cond) == NODE_BLOCK) {
6885 cond = RNODE_BLOCK(cond)->nd_head;
6886 }
6887
6888 CHECK(compile_branch_condition(iseq, cond_seq, cond, then_label, else_label));
6889 ADD_SEQ(ret, cond_seq);
6890
6891 if (then_label->refcnt && else_label->refcnt) {
6892 branches = decl_branch_base(iseq, PTR2NUM(node), nd_code_loc(node), type == NODE_IF ? "if" : "unless");
6893 }
6894
6895 if (then_label->refcnt) {
6896 ADD_LABEL(ret, then_label);
6897
6898 DECL_ANCHOR(then_seq);
6899 INIT_ANCHOR(then_seq);
6900 CHECK(COMPILE_(then_seq, "then", node_body, popped));
6901
6902 if (else_label->refcnt) {
6903 const NODE *const coverage_node = node_body ? node_body : node;
6904 add_trace_branch_coverage(
6905 iseq,
6906 ret,
6907 nd_code_loc(coverage_node),
6908 nd_node_id(coverage_node),
6909 0,
6910 type == NODE_IF ? "then" : "else",
6911 branches);
6912 end_label = NEW_LABEL(line);
6913 ADD_INSNL(then_seq, line_node, jump, end_label);
6914 if (!popped) {
6915 ADD_INSN(then_seq, line_node, pop);
6916 }
6917 }
6918 ADD_SEQ(ret, then_seq);
6919 }
6920
6921 if (else_label->refcnt) {
6922 ADD_LABEL(ret, else_label);
6923
6924 DECL_ANCHOR(else_seq);
6925 INIT_ANCHOR(else_seq);
6926 CHECK(COMPILE_(else_seq, "else", node_else, popped));
6927
6928 if (then_label->refcnt) {
6929 const NODE *const coverage_node = node_else ? node_else : node;
6930 add_trace_branch_coverage(
6931 iseq,
6932 ret,
6933 nd_code_loc(coverage_node),
6934 nd_node_id(coverage_node),
6935 1,
6936 type == NODE_IF ? "else" : "then",
6937 branches);
6938 }
6939 ADD_SEQ(ret, else_seq);
6940 }
6941
6942 if (end_label) {
6943 ADD_LABEL(ret, end_label);
6944 }
6945
6946 return COMPILE_OK;
6947}
6948
6949static int
6950compile_case(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const orig_node, int popped)
6951{
6952 const NODE *vals;
6953 const NODE *node = orig_node;
6954 LABEL *endlabel, *elselabel;
6955 DECL_ANCHOR(head);
6956 DECL_ANCHOR(body_seq);
6957 DECL_ANCHOR(cond_seq);
6958 int only_special_literals = 1;
6959 VALUE literals = rb_hash_new();
6960 int line;
6961 enum node_type type;
6962 const NODE *line_node;
6963 VALUE branches = Qfalse;
6964 int branch_id = 0;
6965
6966 INIT_ANCHOR(head);
6967 INIT_ANCHOR(body_seq);
6968 INIT_ANCHOR(cond_seq);
6969
6970 RHASH_TBL_RAW(literals)->type = &cdhash_type;
6971
6972 CHECK(COMPILE(head, "case base", RNODE_CASE(node)->nd_head));
6973
6974 branches = decl_branch_base(iseq, PTR2NUM(node), nd_code_loc(node), "case");
6975
6976 node = RNODE_CASE(node)->nd_body;
6977 EXPECT_NODE("NODE_CASE", node, NODE_WHEN, COMPILE_NG);
6978 type = nd_type(node);
6979 line = nd_line(node);
6980 line_node = node;
6981
6982 endlabel = NEW_LABEL(line);
6983 elselabel = NEW_LABEL(line);
6984
6985 ADD_SEQ(ret, head); /* case VAL */
6986
6987 while (type == NODE_WHEN) {
6988 LABEL *l1;
6989
6990 l1 = NEW_LABEL(line);
6991 ADD_LABEL(body_seq, l1);
6992 ADD_INSN(body_seq, line_node, pop);
6993
6994 const NODE *const coverage_node = RNODE_WHEN(node)->nd_body ? RNODE_WHEN(node)->nd_body : node;
6995 add_trace_branch_coverage(
6996 iseq,
6997 body_seq,
6998 nd_code_loc(coverage_node),
6999 nd_node_id(coverage_node),
7000 branch_id++,
7001 "when",
7002 branches);
7003
7004 CHECK(COMPILE_(body_seq, "when body", RNODE_WHEN(node)->nd_body, popped));
7005 ADD_INSNL(body_seq, line_node, jump, endlabel);
7006
7007 vals = RNODE_WHEN(node)->nd_head;
7008 if (vals) {
7009 switch (nd_type(vals)) {
7010 case NODE_LIST:
7011 only_special_literals = when_vals(iseq, cond_seq, vals, l1, only_special_literals, literals);
7012 if (only_special_literals < 0) return COMPILE_NG;
7013 break;
7014 case NODE_SPLAT:
7015 case NODE_ARGSCAT:
7016 case NODE_ARGSPUSH:
7017 only_special_literals = 0;
7018 CHECK(when_splat_vals(iseq, cond_seq, vals, l1, only_special_literals, literals));
7019 break;
7020 default:
7021 UNKNOWN_NODE("NODE_CASE", vals, COMPILE_NG);
7022 }
7023 }
7024 else {
7025 EXPECT_NODE_NONULL("NODE_CASE", node, NODE_LIST, COMPILE_NG);
7026 }
7027
7028 node = RNODE_WHEN(node)->nd_next;
7029 if (!node) {
7030 break;
7031 }
7032 type = nd_type(node);
7033 line = nd_line(node);
7034 line_node = node;
7035 }
7036 /* else */
7037 if (node) {
7038 ADD_LABEL(cond_seq, elselabel);
7039 ADD_INSN(cond_seq, line_node, pop);
7040 add_trace_branch_coverage(iseq, cond_seq, nd_code_loc(node), nd_node_id(node), branch_id, "else", branches);
7041 CHECK(COMPILE_(cond_seq, "else", node, popped));
7042 ADD_INSNL(cond_seq, line_node, jump, endlabel);
7043 }
7044 else {
7045 debugs("== else (implicit)\n");
7046 ADD_LABEL(cond_seq, elselabel);
7047 ADD_INSN(cond_seq, orig_node, pop);
7048 add_trace_branch_coverage(iseq, cond_seq, nd_code_loc(orig_node), nd_node_id(orig_node), branch_id, "else", branches);
7049 if (!popped) {
7050 ADD_INSN(cond_seq, orig_node, putnil);
7051 }
7052 ADD_INSNL(cond_seq, orig_node, jump, endlabel);
7053 }
7054
7055 if (only_special_literals && ISEQ_COMPILE_DATA(iseq)->option->specialized_instruction) {
7056 ADD_INSN(ret, orig_node, dup);
7057 ADD_INSN2(ret, orig_node, opt_case_dispatch, literals, elselabel);
7058 RB_OBJ_WRITTEN(iseq, Qundef, literals);
7059 LABEL_REF(elselabel);
7060 }
7061
7062 ADD_SEQ(ret, cond_seq);
7063 ADD_SEQ(ret, body_seq);
7064 ADD_LABEL(ret, endlabel);
7065 return COMPILE_OK;
7066}
7067
7068static int
7069compile_case2(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const orig_node, int popped)
7070{
7071 const NODE *vals;
7072 const NODE *val;
7073 const NODE *node = RNODE_CASE2(orig_node)->nd_body;
7074 LABEL *endlabel;
7075 DECL_ANCHOR(body_seq);
7076 VALUE branches = Qfalse;
7077 int branch_id = 0;
7078
7079 branches = decl_branch_base(iseq, PTR2NUM(orig_node), nd_code_loc(orig_node), "case");
7080
7081 INIT_ANCHOR(body_seq);
7082 endlabel = NEW_LABEL(nd_line(node));
7083
7084 while (node && nd_type_p(node, NODE_WHEN)) {
7085 const int line = nd_line(node);
7086 LABEL *l1 = NEW_LABEL(line);
7087 ADD_LABEL(body_seq, l1);
7088
7089 const NODE *const coverage_node = RNODE_WHEN(node)->nd_body ? RNODE_WHEN(node)->nd_body : node;
7090 add_trace_branch_coverage(
7091 iseq,
7092 body_seq,
7093 nd_code_loc(coverage_node),
7094 nd_node_id(coverage_node),
7095 branch_id++,
7096 "when",
7097 branches);
7098
7099 CHECK(COMPILE_(body_seq, "when", RNODE_WHEN(node)->nd_body, popped));
7100 ADD_INSNL(body_seq, node, jump, endlabel);
7101
7102 vals = RNODE_WHEN(node)->nd_head;
7103 if (!vals) {
7104 EXPECT_NODE_NONULL("NODE_WHEN", node, NODE_LIST, COMPILE_NG);
7105 }
7106 switch (nd_type(vals)) {
7107 case NODE_LIST:
7108 while (vals) {
7109 LABEL *lnext;
7110 val = RNODE_LIST(vals)->nd_head;
7111 lnext = NEW_LABEL(nd_line(val));
7112 debug_compile("== when2\n", (void)0);
7113 CHECK(compile_branch_condition(iseq, ret, val, l1, lnext));
7114 ADD_LABEL(ret, lnext);
7115 vals = RNODE_LIST(vals)->nd_next;
7116 }
7117 break;
7118 case NODE_SPLAT:
7119 case NODE_ARGSCAT:
7120 case NODE_ARGSPUSH:
7121 ADD_INSN(ret, vals, putnil);
7122 CHECK(COMPILE(ret, "when2/cond splat", vals));
7123 ADD_INSN1(ret, vals, checkmatch, INT2FIX(VM_CHECKMATCH_TYPE_WHEN | VM_CHECKMATCH_ARRAY));
7124 ADD_INSNL(ret, vals, branchif, l1);
7125 break;
7126 default:
7127 UNKNOWN_NODE("NODE_WHEN", vals, COMPILE_NG);
7128 }
7129 node = RNODE_WHEN(node)->nd_next;
7130 }
7131 /* else */
7132 const NODE *const coverage_node = node ? node : orig_node;
7133 add_trace_branch_coverage(
7134 iseq,
7135 ret,
7136 nd_code_loc(coverage_node),
7137 nd_node_id(coverage_node),
7138 branch_id,
7139 "else",
7140 branches);
7141 CHECK(COMPILE_(ret, "else", node, popped));
7142 ADD_INSNL(ret, orig_node, jump, endlabel);
7143
7144 ADD_SEQ(ret, body_seq);
7145 ADD_LABEL(ret, endlabel);
7146 return COMPILE_OK;
7147}
7148
7149static int iseq_compile_pattern_match(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, LABEL *unmatched, bool in_single_pattern, bool in_alt_pattern, int base_index, bool use_deconstructed_cache);
7150
7151static int iseq_compile_pattern_constant(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, LABEL *match_failed, bool in_single_pattern, int base_index);
7152static int iseq_compile_array_deconstruct(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, LABEL *deconstruct, LABEL *deconstructed, LABEL *match_failed, LABEL *type_error, bool in_single_pattern, int base_index, bool use_deconstructed_cache);
7153static int iseq_compile_pattern_set_general_errmsg(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, VALUE errmsg, int base_index);
7154static int iseq_compile_pattern_set_length_errmsg(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, VALUE errmsg, VALUE pattern_length, int base_index);
7155static int iseq_compile_pattern_set_eqq_errmsg(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, int base_index);
7156
7157#define CASE3_BI_OFFSET_DECONSTRUCTED_CACHE 0
7158#define CASE3_BI_OFFSET_ERROR_STRING 1
7159#define CASE3_BI_OFFSET_KEY_ERROR_P 2
7160#define CASE3_BI_OFFSET_KEY_ERROR_MATCHEE 3
7161#define CASE3_BI_OFFSET_KEY_ERROR_KEY 4
7162
7163static int
7164iseq_compile_pattern_each(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, LABEL *matched, LABEL *unmatched, bool in_single_pattern, bool in_alt_pattern, int base_index, bool use_deconstructed_cache)
7165{
7166 const int line = nd_line(node);
7167 const NODE *line_node = node;
7168
7169 switch (nd_type(node)) {
7170 case NODE_ARYPTN: {
7171 /*
7172 * if pattern.use_rest_num?
7173 * rest_num = 0
7174 * end
7175 * if pattern.has_constant_node?
7176 * unless pattern.constant === obj
7177 * goto match_failed
7178 * end
7179 * end
7180 * unless obj.respond_to?(:deconstruct)
7181 * goto match_failed
7182 * end
7183 * d = obj.deconstruct
7184 * unless Array === d
7185 * goto type_error
7186 * end
7187 * min_argc = pattern.pre_args_num + pattern.post_args_num
7188 * if pattern.has_rest_arg?
7189 * unless d.length >= min_argc
7190 * goto match_failed
7191 * end
7192 * else
7193 * unless d.length == min_argc
7194 * goto match_failed
7195 * end
7196 * end
7197 * pattern.pre_args_num.each do |i|
7198 * unless pattern.pre_args[i].match?(d[i])
7199 * goto match_failed
7200 * end
7201 * end
7202 * if pattern.use_rest_num?
7203 * rest_num = d.length - min_argc
7204 * if pattern.has_rest_arg? && pattern.has_rest_arg_id # not `*`, but `*rest`
7205 * unless pattern.rest_arg.match?(d[pattern.pre_args_num, rest_num])
7206 * goto match_failed
7207 * end
7208 * end
7209 * end
7210 * pattern.post_args_num.each do |i|
7211 * j = pattern.pre_args_num + i
7212 * j += rest_num
7213 * unless pattern.post_args[i].match?(d[j])
7214 * goto match_failed
7215 * end
7216 * end
7217 * goto matched
7218 * type_error:
7219 * FrozenCore.raise TypeError
7220 * match_failed:
7221 * goto unmatched
7222 */
7223 const NODE *args = RNODE_ARYPTN(node)->pre_args;
7224 const int pre_args_num = RNODE_ARYPTN(node)->pre_args ? rb_long2int(RNODE_LIST(RNODE_ARYPTN(node)->pre_args)->as.nd_alen) : 0;
7225 const int post_args_num = RNODE_ARYPTN(node)->post_args ? rb_long2int(RNODE_LIST(RNODE_ARYPTN(node)->post_args)->as.nd_alen) : 0;
7226
7227 const int min_argc = pre_args_num + post_args_num;
7228 const int use_rest_num = RNODE_ARYPTN(node)->rest_arg && (NODE_NAMED_REST_P(RNODE_ARYPTN(node)->rest_arg) ||
7229 (!NODE_NAMED_REST_P(RNODE_ARYPTN(node)->rest_arg) && post_args_num > 0));
7230
7231 LABEL *match_failed, *type_error, *deconstruct, *deconstructed;
7232 int i;
7233 match_failed = NEW_LABEL(line);
7234 type_error = NEW_LABEL(line);
7235 deconstruct = NEW_LABEL(line);
7236 deconstructed = NEW_LABEL(line);
7237
7238 if (use_rest_num) {
7239 ADD_INSN1(ret, line_node, putobject, INT2FIX(0)); /* allocate stack for rest_num */
7240 ADD_INSN(ret, line_node, swap);
7241 if (base_index) {
7242 base_index++;
7243 }
7244 }
7245
7246 CHECK(iseq_compile_pattern_constant(iseq, ret, node, match_failed, in_single_pattern, base_index));
7247
7248 CHECK(iseq_compile_array_deconstruct(iseq, ret, node, deconstruct, deconstructed, match_failed, type_error, in_single_pattern, base_index, use_deconstructed_cache));
7249
7250 ADD_INSN(ret, line_node, dup);
7251 ADD_SEND(ret, line_node, idLength, INT2FIX(0));
7252 ADD_INSN1(ret, line_node, putobject, INT2FIX(min_argc));
7253 ADD_SEND(ret, line_node, RNODE_ARYPTN(node)->rest_arg ? idGE : idEq, INT2FIX(1)); // (1)
7254 if (in_single_pattern) {
7255 CHECK(iseq_compile_pattern_set_length_errmsg(iseq, ret, node,
7256 RNODE_ARYPTN(node)->rest_arg ? rb_fstring_lit("%p length mismatch (given %p, expected %p+)") :
7257 rb_fstring_lit("%p length mismatch (given %p, expected %p)"),
7258 INT2FIX(min_argc), base_index + 1 /* (1) */));
7259 }
7260 ADD_INSNL(ret, line_node, branchunless, match_failed);
7261
7262 for (i = 0; i < pre_args_num; i++) {
7263 ADD_INSN(ret, line_node, dup);
7264 ADD_INSN1(ret, line_node, putobject, INT2FIX(i));
7265 ADD_SEND(ret, line_node, idAREF, INT2FIX(1)); // (2)
7266 CHECK(iseq_compile_pattern_match(iseq, ret, RNODE_LIST(args)->nd_head, match_failed, in_single_pattern, in_alt_pattern, base_index + 1 /* (2) */, false));
7267 args = RNODE_LIST(args)->nd_next;
7268 }
7269
7270 if (RNODE_ARYPTN(node)->rest_arg) {
7271 if (NODE_NAMED_REST_P(RNODE_ARYPTN(node)->rest_arg)) {
7272 ADD_INSN(ret, line_node, dup);
7273 ADD_INSN1(ret, line_node, putobject, INT2FIX(pre_args_num));
7274 ADD_INSN1(ret, line_node, topn, INT2FIX(1));
7275 ADD_SEND(ret, line_node, idLength, INT2FIX(0));
7276 ADD_INSN1(ret, line_node, putobject, INT2FIX(min_argc));
7277 ADD_SEND(ret, line_node, idMINUS, INT2FIX(1));
7278 ADD_INSN1(ret, line_node, setn, INT2FIX(4));
7279 ADD_SEND(ret, line_node, idAREF, INT2FIX(2)); // (3)
7280
7281 CHECK(iseq_compile_pattern_match(iseq, ret, RNODE_ARYPTN(node)->rest_arg, match_failed, in_single_pattern, in_alt_pattern, base_index + 1 /* (3) */, false));
7282 }
7283 else {
7284 if (post_args_num > 0) {
7285 ADD_INSN(ret, line_node, dup);
7286 ADD_SEND(ret, line_node, idLength, INT2FIX(0));
7287 ADD_INSN1(ret, line_node, putobject, INT2FIX(min_argc));
7288 ADD_SEND(ret, line_node, idMINUS, INT2FIX(1));
7289 ADD_INSN1(ret, line_node, setn, INT2FIX(2));
7290 ADD_INSN(ret, line_node, pop);
7291 }
7292 }
7293 }
7294
7295 args = RNODE_ARYPTN(node)->post_args;
7296 for (i = 0; i < post_args_num; i++) {
7297 ADD_INSN(ret, line_node, dup);
7298
7299 ADD_INSN1(ret, line_node, putobject, INT2FIX(pre_args_num + i));
7300 ADD_INSN1(ret, line_node, topn, INT2FIX(3));
7301 ADD_SEND(ret, line_node, idPLUS, INT2FIX(1));
7302
7303 ADD_SEND(ret, line_node, idAREF, INT2FIX(1)); // (4)
7304 CHECK(iseq_compile_pattern_match(iseq, ret, RNODE_LIST(args)->nd_head, match_failed, in_single_pattern, in_alt_pattern, base_index + 1 /* (4) */, false));
7305 args = RNODE_LIST(args)->nd_next;
7306 }
7307
7308 ADD_INSN(ret, line_node, pop);
7309 if (use_rest_num) {
7310 ADD_INSN(ret, line_node, pop);
7311 }
7312 ADD_INSNL(ret, line_node, jump, matched);
7313 ADD_INSN(ret, line_node, putnil);
7314 if (use_rest_num) {
7315 ADD_INSN(ret, line_node, putnil);
7316 }
7317
7318 ADD_LABEL(ret, type_error);
7319 ADD_INSN1(ret, line_node, putspecialobject, INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
7320 ADD_INSN1(ret, line_node, putobject, rb_eTypeError);
7321 ADD_INSN1(ret, line_node, putobject, rb_fstring_lit("deconstruct must return Array"));
7322 ADD_SEND(ret, line_node, id_core_raise, INT2FIX(2));
7323 ADD_INSN(ret, line_node, pop);
7324
7325 ADD_LABEL(ret, match_failed);
7326 ADD_INSN(ret, line_node, pop);
7327 if (use_rest_num) {
7328 ADD_INSN(ret, line_node, pop);
7329 }
7330 ADD_INSNL(ret, line_node, jump, unmatched);
7331
7332 break;
7333 }
7334 case NODE_FNDPTN: {
7335 /*
7336 * if pattern.has_constant_node?
7337 * unless pattern.constant === obj
7338 * goto match_failed
7339 * end
7340 * end
7341 * unless obj.respond_to?(:deconstruct)
7342 * goto match_failed
7343 * end
7344 * d = obj.deconstruct
7345 * unless Array === d
7346 * goto type_error
7347 * end
7348 * unless d.length >= pattern.args_num
7349 * goto match_failed
7350 * end
7351 *
7352 * begin
7353 * len = d.length
7354 * limit = d.length - pattern.args_num
7355 * i = 0
7356 * while i <= limit
7357 * if pattern.args_num.times.all? {|j| pattern.args[j].match?(d[i+j]) }
7358 * if pattern.has_pre_rest_arg_id
7359 * unless pattern.pre_rest_arg.match?(d[0, i])
7360 * goto find_failed
7361 * end
7362 * end
7363 * if pattern.has_post_rest_arg_id
7364 * unless pattern.post_rest_arg.match?(d[i+pattern.args_num, len])
7365 * goto find_failed
7366 * end
7367 * end
7368 * goto find_succeeded
7369 * end
7370 * i+=1
7371 * end
7372 * find_failed:
7373 * goto match_failed
7374 * find_succeeded:
7375 * end
7376 *
7377 * goto matched
7378 * type_error:
7379 * FrozenCore.raise TypeError
7380 * match_failed:
7381 * goto unmatched
7382 */
7383 const NODE *args = RNODE_FNDPTN(node)->args;
7384 const int args_num = RNODE_FNDPTN(node)->args ? rb_long2int(RNODE_LIST(RNODE_FNDPTN(node)->args)->as.nd_alen) : 0;
7385
7386 LABEL *match_failed, *type_error, *deconstruct, *deconstructed;
7387 match_failed = NEW_LABEL(line);
7388 type_error = NEW_LABEL(line);
7389 deconstruct = NEW_LABEL(line);
7390 deconstructed = NEW_LABEL(line);
7391
7392 CHECK(iseq_compile_pattern_constant(iseq, ret, node, match_failed, in_single_pattern, base_index));
7393
7394 CHECK(iseq_compile_array_deconstruct(iseq, ret, node, deconstruct, deconstructed, match_failed, type_error, in_single_pattern, base_index, use_deconstructed_cache));
7395
7396 ADD_INSN(ret, line_node, dup);
7397 ADD_SEND(ret, line_node, idLength, INT2FIX(0));
7398 ADD_INSN1(ret, line_node, putobject, INT2FIX(args_num));
7399 ADD_SEND(ret, line_node, idGE, INT2FIX(1)); // (1)
7400 if (in_single_pattern) {
7401 CHECK(iseq_compile_pattern_set_length_errmsg(iseq, ret, node, rb_fstring_lit("%p length mismatch (given %p, expected %p+)"), INT2FIX(args_num), base_index + 1 /* (1) */));
7402 }
7403 ADD_INSNL(ret, line_node, branchunless, match_failed);
7404
7405 {
7406 LABEL *while_begin = NEW_LABEL(nd_line(node));
7407 LABEL *next_loop = NEW_LABEL(nd_line(node));
7408 LABEL *find_succeeded = NEW_LABEL(line);
7409 LABEL *find_failed = NEW_LABEL(nd_line(node));
7410 int j;
7411
7412 ADD_INSN(ret, line_node, dup); /* allocate stack for len */
7413 ADD_SEND(ret, line_node, idLength, INT2FIX(0)); // (2)
7414
7415 ADD_INSN(ret, line_node, dup); /* allocate stack for limit */
7416 ADD_INSN1(ret, line_node, putobject, INT2FIX(args_num));
7417 ADD_SEND(ret, line_node, idMINUS, INT2FIX(1)); // (3)
7418
7419 ADD_INSN1(ret, line_node, putobject, INT2FIX(0)); /* allocate stack for i */ // (4)
7420
7421 ADD_LABEL(ret, while_begin);
7422
7423 ADD_INSN(ret, line_node, dup);
7424 ADD_INSN1(ret, line_node, topn, INT2FIX(2));
7425 ADD_SEND(ret, line_node, idLE, INT2FIX(1));
7426 ADD_INSNL(ret, line_node, branchunless, find_failed);
7427
7428 for (j = 0; j < args_num; j++) {
7429 ADD_INSN1(ret, line_node, topn, INT2FIX(3));
7430 ADD_INSN1(ret, line_node, topn, INT2FIX(1));
7431 if (j != 0) {
7432 ADD_INSN1(ret, line_node, putobject, INT2FIX(j));
7433 ADD_SEND(ret, line_node, idPLUS, INT2FIX(1));
7434 }
7435 ADD_SEND(ret, line_node, idAREF, INT2FIX(1)); // (5)
7436
7437 CHECK(iseq_compile_pattern_match(iseq, ret, RNODE_LIST(args)->nd_head, next_loop, in_single_pattern, in_alt_pattern, base_index + 4 /* (2), (3), (4), (5) */, false));
7438 args = RNODE_LIST(args)->nd_next;
7439 }
7440
7441 if (NODE_NAMED_REST_P(RNODE_FNDPTN(node)->pre_rest_arg)) {
7442 ADD_INSN1(ret, line_node, topn, INT2FIX(3));
7443 ADD_INSN1(ret, line_node, putobject, INT2FIX(0));
7444 ADD_INSN1(ret, line_node, topn, INT2FIX(2));
7445 ADD_SEND(ret, line_node, idAREF, INT2FIX(2)); // (6)
7446 CHECK(iseq_compile_pattern_match(iseq, ret, RNODE_FNDPTN(node)->pre_rest_arg, find_failed, in_single_pattern, in_alt_pattern, base_index + 4 /* (2), (3), (4), (6) */, false));
7447 }
7448 if (NODE_NAMED_REST_P(RNODE_FNDPTN(node)->post_rest_arg)) {
7449 ADD_INSN1(ret, line_node, topn, INT2FIX(3));
7450 ADD_INSN1(ret, line_node, topn, INT2FIX(1));
7451 ADD_INSN1(ret, line_node, putobject, INT2FIX(args_num));
7452 ADD_SEND(ret, line_node, idPLUS, INT2FIX(1));
7453 ADD_INSN1(ret, line_node, topn, INT2FIX(3));
7454 ADD_SEND(ret, line_node, idAREF, INT2FIX(2)); // (7)
7455 CHECK(iseq_compile_pattern_match(iseq, ret, RNODE_FNDPTN(node)->post_rest_arg, find_failed, in_single_pattern, in_alt_pattern, base_index + 4 /* (2), (3),(4), (7) */, false));
7456 }
7457 ADD_INSNL(ret, line_node, jump, find_succeeded);
7458
7459 ADD_LABEL(ret, next_loop);
7460 ADD_INSN1(ret, line_node, putobject, INT2FIX(1));
7461 ADD_SEND(ret, line_node, idPLUS, INT2FIX(1));
7462 ADD_INSNL(ret, line_node, jump, while_begin);
7463
7464 ADD_LABEL(ret, find_failed);
7465 ADD_INSN1(ret, line_node, adjuststack, INT2FIX(3));
7466 if (in_single_pattern) {
7467 ADD_INSN1(ret, line_node, putspecialobject, INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
7468 ADD_INSN1(ret, line_node, putobject, rb_fstring_lit("%p does not match to find pattern"));
7469 ADD_INSN1(ret, line_node, topn, INT2FIX(2));
7470 ADD_SEND(ret, line_node, id_core_sprintf, INT2FIX(2)); // (8)
7471 ADD_INSN1(ret, line_node, setn, INT2FIX(base_index + CASE3_BI_OFFSET_ERROR_STRING + 1 /* (8) */)); // (9)
7472
7473 ADD_INSN1(ret, line_node, putobject, Qfalse);
7474 ADD_INSN1(ret, line_node, setn, INT2FIX(base_index + CASE3_BI_OFFSET_KEY_ERROR_P + 2 /* (8), (9) */));
7475
7476 ADD_INSN(ret, line_node, pop);
7477 ADD_INSN(ret, line_node, pop);
7478 }
7479 ADD_INSNL(ret, line_node, jump, match_failed);
7480 ADD_INSN1(ret, line_node, dupn, INT2FIX(3));
7481
7482 ADD_LABEL(ret, find_succeeded);
7483 ADD_INSN1(ret, line_node, adjuststack, INT2FIX(3));
7484 }
7485
7486 ADD_INSN(ret, line_node, pop);
7487 ADD_INSNL(ret, line_node, jump, matched);
7488 ADD_INSN(ret, line_node, putnil);
7489
7490 ADD_LABEL(ret, type_error);
7491 ADD_INSN1(ret, line_node, putspecialobject, INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
7492 ADD_INSN1(ret, line_node, putobject, rb_eTypeError);
7493 ADD_INSN1(ret, line_node, putobject, rb_fstring_lit("deconstruct must return Array"));
7494 ADD_SEND(ret, line_node, id_core_raise, INT2FIX(2));
7495 ADD_INSN(ret, line_node, pop);
7496
7497 ADD_LABEL(ret, match_failed);
7498 ADD_INSN(ret, line_node, pop);
7499 ADD_INSNL(ret, line_node, jump, unmatched);
7500
7501 break;
7502 }
7503 case NODE_HSHPTN: {
7504 /*
7505 * keys = nil
7506 * if pattern.has_kw_args_node? && !pattern.has_kw_rest_arg_node?
7507 * keys = pattern.kw_args_node.keys
7508 * end
7509 * if pattern.has_constant_node?
7510 * unless pattern.constant === obj
7511 * goto match_failed
7512 * end
7513 * end
7514 * unless obj.respond_to?(:deconstruct_keys)
7515 * goto match_failed
7516 * end
7517 * d = obj.deconstruct_keys(keys)
7518 * unless Hash === d
7519 * goto type_error
7520 * end
7521 * if pattern.has_kw_rest_arg_node?
7522 * d = d.dup
7523 * end
7524 * if pattern.has_kw_args_node?
7525 * pattern.kw_args_node.each |k,|
7526 * unless d.key?(k)
7527 * goto match_failed
7528 * end
7529 * end
7530 * pattern.kw_args_node.each |k, pat|
7531 * if pattern.has_kw_rest_arg_node?
7532 * unless pat.match?(d.delete(k))
7533 * goto match_failed
7534 * end
7535 * else
7536 * unless pat.match?(d[k])
7537 * goto match_failed
7538 * end
7539 * end
7540 * end
7541 * else
7542 * unless d.empty?
7543 * goto match_failed
7544 * end
7545 * end
7546 * if pattern.has_kw_rest_arg_node?
7547 * if pattern.no_rest_keyword?
7548 * unless d.empty?
7549 * goto match_failed
7550 * end
7551 * else
7552 * unless pattern.kw_rest_arg_node.match?(d)
7553 * goto match_failed
7554 * end
7555 * end
7556 * end
7557 * goto matched
7558 * type_error:
7559 * FrozenCore.raise TypeError
7560 * match_failed:
7561 * goto unmatched
7562 */
7563 LABEL *match_failed, *type_error;
7564 VALUE keys = Qnil;
7565
7566 match_failed = NEW_LABEL(line);
7567 type_error = NEW_LABEL(line);
7568
7569 if (RNODE_HSHPTN(node)->nd_pkwargs && !RNODE_HSHPTN(node)->nd_pkwrestarg) {
7570 const NODE *kw_args = RNODE_HASH(RNODE_HSHPTN(node)->nd_pkwargs)->nd_head;
7571 keys = rb_ary_new_capa(kw_args ? RNODE_LIST(kw_args)->as.nd_alen/2 : 0);
7572 while (kw_args) {
7573 rb_ary_push(keys, get_symbol_value(iseq, RNODE_LIST(kw_args)->nd_head));
7574 kw_args = RNODE_LIST(RNODE_LIST(kw_args)->nd_next)->nd_next;
7575 }
7576 }
7577
7578 CHECK(iseq_compile_pattern_constant(iseq, ret, node, match_failed, in_single_pattern, base_index));
7579
7580 ADD_INSN(ret, line_node, dup);
7581 ADD_INSN1(ret, line_node, putobject, ID2SYM(rb_intern("deconstruct_keys")));
7582 ADD_SEND(ret, line_node, idRespond_to, INT2FIX(1)); // (1)
7583 if (in_single_pattern) {
7584 CHECK(iseq_compile_pattern_set_general_errmsg(iseq, ret, node, rb_fstring_lit("%p does not respond to #deconstruct_keys"), base_index + 1 /* (1) */));
7585 }
7586 ADD_INSNL(ret, line_node, branchunless, match_failed);
7587
7588 if (NIL_P(keys)) {
7589 ADD_INSN(ret, line_node, putnil);
7590 }
7591 else {
7592 ADD_INSN1(ret, line_node, duparray, keys);
7593 RB_OBJ_WRITTEN(iseq, Qundef, rb_obj_hide(keys));
7594 }
7595 ADD_SEND(ret, line_node, rb_intern("deconstruct_keys"), INT2FIX(1)); // (2)
7596
7597 ADD_INSN(ret, line_node, dup);
7598 ADD_INSN1(ret, line_node, checktype, INT2FIX(T_HASH));
7599 ADD_INSNL(ret, line_node, branchunless, type_error);
7600
7601 if (RNODE_HSHPTN(node)->nd_pkwrestarg) {
7602 ADD_SEND(ret, line_node, rb_intern("dup"), INT2FIX(0));
7603 }
7604
7605 if (RNODE_HSHPTN(node)->nd_pkwargs) {
7606 int i;
7607 int keys_num;
7608 const NODE *args;
7609 args = RNODE_HASH(RNODE_HSHPTN(node)->nd_pkwargs)->nd_head;
7610 if (args) {
7611 DECL_ANCHOR(match_values);
7612 INIT_ANCHOR(match_values);
7613 keys_num = rb_long2int(RNODE_LIST(args)->as.nd_alen) / 2;
7614 for (i = 0; i < keys_num; i++) {
7615 NODE *key_node = RNODE_LIST(args)->nd_head;
7616 NODE *value_node = RNODE_LIST(RNODE_LIST(args)->nd_next)->nd_head;
7617 VALUE key = get_symbol_value(iseq, key_node);
7618
7619 ADD_INSN(ret, line_node, dup);
7620 ADD_INSN1(ret, line_node, putobject, key);
7621 ADD_SEND(ret, line_node, rb_intern("key?"), INT2FIX(1)); // (3)
7622 if (in_single_pattern) {
7623 LABEL *match_succeeded;
7624 match_succeeded = NEW_LABEL(line);
7625
7626 ADD_INSN(ret, line_node, dup);
7627 ADD_INSNL(ret, line_node, branchif, match_succeeded);
7628
7629 ADD_INSN1(ret, line_node, putobject, rb_str_freeze(rb_sprintf("key not found: %+"PRIsVALUE, key))); // (4)
7630 ADD_INSN1(ret, line_node, setn, INT2FIX(base_index + CASE3_BI_OFFSET_ERROR_STRING + 2 /* (3), (4) */));
7631 ADD_INSN1(ret, line_node, putobject, Qtrue); // (5)
7632 ADD_INSN1(ret, line_node, setn, INT2FIX(base_index + CASE3_BI_OFFSET_KEY_ERROR_P + 3 /* (3), (4), (5) */));
7633 ADD_INSN1(ret, line_node, topn, INT2FIX(3)); // (6)
7634 ADD_INSN1(ret, line_node, setn, INT2FIX(base_index + CASE3_BI_OFFSET_KEY_ERROR_MATCHEE + 4 /* (3), (4), (5), (6) */));
7635 ADD_INSN1(ret, line_node, putobject, key); // (7)
7636 ADD_INSN1(ret, line_node, setn, INT2FIX(base_index + CASE3_BI_OFFSET_KEY_ERROR_KEY + 5 /* (3), (4), (5), (6), (7) */));
7637
7638 ADD_INSN1(ret, line_node, adjuststack, INT2FIX(4));
7639
7640 ADD_LABEL(ret, match_succeeded);
7641 }
7642 ADD_INSNL(ret, line_node, branchunless, match_failed);
7643
7644 ADD_INSN(match_values, line_node, dup);
7645 ADD_INSN1(match_values, line_node, putobject, key);
7646 ADD_SEND(match_values, line_node, RNODE_HSHPTN(node)->nd_pkwrestarg ? rb_intern("delete") : idAREF, INT2FIX(1)); // (8)
7647 CHECK(iseq_compile_pattern_match(iseq, match_values, value_node, match_failed, in_single_pattern, in_alt_pattern, base_index + 1 /* (8) */, false));
7648 args = RNODE_LIST(RNODE_LIST(args)->nd_next)->nd_next;
7649 }
7650 ADD_SEQ(ret, match_values);
7651 }
7652 }
7653 else {
7654 ADD_INSN(ret, line_node, dup);
7655 ADD_SEND(ret, line_node, idEmptyP, INT2FIX(0)); // (9)
7656 if (in_single_pattern) {
7657 CHECK(iseq_compile_pattern_set_general_errmsg(iseq, ret, node, rb_fstring_lit("%p is not empty"), base_index + 1 /* (9) */));
7658 }
7659 ADD_INSNL(ret, line_node, branchunless, match_failed);
7660 }
7661
7662 if (RNODE_HSHPTN(node)->nd_pkwrestarg) {
7663 if (RNODE_HSHPTN(node)->nd_pkwrestarg == NODE_SPECIAL_NO_REST_KEYWORD) {
7664 ADD_INSN(ret, line_node, dup);
7665 ADD_SEND(ret, line_node, idEmptyP, INT2FIX(0)); // (10)
7666 if (in_single_pattern) {
7667 CHECK(iseq_compile_pattern_set_general_errmsg(iseq, ret, node, rb_fstring_lit("rest of %p is not empty"), base_index + 1 /* (10) */));
7668 }
7669 ADD_INSNL(ret, line_node, branchunless, match_failed);
7670 }
7671 else {
7672 ADD_INSN(ret, line_node, dup); // (11)
7673 CHECK(iseq_compile_pattern_match(iseq, ret, RNODE_HSHPTN(node)->nd_pkwrestarg, match_failed, in_single_pattern, in_alt_pattern, base_index + 1 /* (11) */, false));
7674 }
7675 }
7676
7677 ADD_INSN(ret, line_node, pop);
7678 ADD_INSNL(ret, line_node, jump, matched);
7679 ADD_INSN(ret, line_node, putnil);
7680
7681 ADD_LABEL(ret, type_error);
7682 ADD_INSN1(ret, line_node, putspecialobject, INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
7683 ADD_INSN1(ret, line_node, putobject, rb_eTypeError);
7684 ADD_INSN1(ret, line_node, putobject, rb_fstring_lit("deconstruct_keys must return Hash"));
7685 ADD_SEND(ret, line_node, id_core_raise, INT2FIX(2));
7686 ADD_INSN(ret, line_node, pop);
7687
7688 ADD_LABEL(ret, match_failed);
7689 ADD_INSN(ret, line_node, pop);
7690 ADD_INSNL(ret, line_node, jump, unmatched);
7691 break;
7692 }
7693 case NODE_SYM:
7694 case NODE_REGX:
7695 case NODE_LINE:
7696 case NODE_INTEGER:
7697 case NODE_FLOAT:
7698 case NODE_RATIONAL:
7699 case NODE_IMAGINARY:
7700 case NODE_FILE:
7701 case NODE_ENCODING:
7702 case NODE_STR:
7703 case NODE_XSTR:
7704 case NODE_DSTR:
7705 case NODE_DSYM:
7706 case NODE_DREGX:
7707 case NODE_LIST:
7708 case NODE_ZLIST:
7709 case NODE_LAMBDA:
7710 case NODE_DOT2:
7711 case NODE_DOT3:
7712 case NODE_CONST:
7713 case NODE_LVAR:
7714 case NODE_DVAR:
7715 case NODE_IVAR:
7716 case NODE_CVAR:
7717 case NODE_GVAR:
7718 case NODE_TRUE:
7719 case NODE_FALSE:
7720 case NODE_SELF:
7721 case NODE_NIL:
7722 case NODE_COLON2:
7723 case NODE_COLON3:
7724 case NODE_BEGIN:
7725 case NODE_BLOCK:
7726 case NODE_ONCE:
7727 CHECK(COMPILE(ret, "case in literal", node)); // (1)
7728 if (in_single_pattern) {
7729 ADD_INSN1(ret, line_node, dupn, INT2FIX(2));
7730 }
7731 ADD_INSN1(ret, line_node, checkmatch, INT2FIX(VM_CHECKMATCH_TYPE_CASE)); // (2)
7732 if (in_single_pattern) {
7733 CHECK(iseq_compile_pattern_set_eqq_errmsg(iseq, ret, node, base_index + 2 /* (1), (2) */));
7734 }
7735 ADD_INSNL(ret, line_node, branchif, matched);
7736 ADD_INSNL(ret, line_node, jump, unmatched);
7737 break;
7738 case NODE_LASGN: {
7739 struct rb_iseq_constant_body *const body = ISEQ_BODY(iseq);
7740 ID id = RNODE_LASGN(node)->nd_vid;
7741 int idx = ISEQ_BODY(body->local_iseq)->local_table_size - get_local_var_idx(iseq, id);
7742
7743 if (in_alt_pattern) {
7744 const char *name = rb_id2name(id);
7745 if (name && strlen(name) > 0 && name[0] != '_') {
7746 COMPILE_ERROR(ERROR_ARGS "illegal variable in alternative pattern (%"PRIsVALUE")",
7747 rb_id2str(id));
7748 return COMPILE_NG;
7749 }
7750 }
7751
7752 ADD_SETLOCAL(ret, line_node, idx, get_lvar_level(iseq));
7753 ADD_INSNL(ret, line_node, jump, matched);
7754 break;
7755 }
7756 case NODE_DASGN: {
7757 int idx, lv, ls;
7758 ID id = RNODE_DASGN(node)->nd_vid;
7759
7760 idx = get_dyna_var_idx(iseq, id, &lv, &ls);
7761
7762 if (in_alt_pattern) {
7763 const char *name = rb_id2name(id);
7764 if (name && strlen(name) > 0 && name[0] != '_') {
7765 COMPILE_ERROR(ERROR_ARGS "illegal variable in alternative pattern (%"PRIsVALUE")",
7766 rb_id2str(id));
7767 return COMPILE_NG;
7768 }
7769 }
7770
7771 if (idx < 0) {
7772 COMPILE_ERROR(ERROR_ARGS "NODE_DASGN: unknown id (%"PRIsVALUE")",
7773 rb_id2str(id));
7774 return COMPILE_NG;
7775 }
7776 ADD_SETLOCAL(ret, line_node, ls - idx, lv);
7777 ADD_INSNL(ret, line_node, jump, matched);
7778 break;
7779 }
7780 case NODE_IF:
7781 case NODE_UNLESS: {
7782 LABEL *match_failed;
7783 match_failed = unmatched;
7784 CHECK(iseq_compile_pattern_match(iseq, ret, RNODE_IF(node)->nd_body, unmatched, in_single_pattern, in_alt_pattern, base_index, use_deconstructed_cache));
7785 CHECK(COMPILE(ret, "case in if", RNODE_IF(node)->nd_cond));
7786 if (in_single_pattern) {
7787 LABEL *match_succeeded;
7788 match_succeeded = NEW_LABEL(line);
7789
7790 ADD_INSN(ret, line_node, dup);
7791 if (nd_type_p(node, NODE_IF)) {
7792 ADD_INSNL(ret, line_node, branchif, match_succeeded);
7793 }
7794 else {
7795 ADD_INSNL(ret, line_node, branchunless, match_succeeded);
7796 }
7797
7798 ADD_INSN1(ret, line_node, putobject, rb_fstring_lit("guard clause does not return true")); // (1)
7799 ADD_INSN1(ret, line_node, setn, INT2FIX(base_index + CASE3_BI_OFFSET_ERROR_STRING + 1 /* (1) */)); // (2)
7800 ADD_INSN1(ret, line_node, putobject, Qfalse);
7801 ADD_INSN1(ret, line_node, setn, INT2FIX(base_index + CASE3_BI_OFFSET_KEY_ERROR_P + 2 /* (1), (2) */));
7802
7803 ADD_INSN(ret, line_node, pop);
7804 ADD_INSN(ret, line_node, pop);
7805
7806 ADD_LABEL(ret, match_succeeded);
7807 }
7808 if (nd_type_p(node, NODE_IF)) {
7809 ADD_INSNL(ret, line_node, branchunless, match_failed);
7810 }
7811 else {
7812 ADD_INSNL(ret, line_node, branchif, match_failed);
7813 }
7814 ADD_INSNL(ret, line_node, jump, matched);
7815 break;
7816 }
7817 case NODE_HASH: {
7818 NODE *n;
7819 LABEL *match_failed;
7820 match_failed = NEW_LABEL(line);
7821
7822 n = RNODE_HASH(node)->nd_head;
7823 if (! (nd_type_p(n, NODE_LIST) && RNODE_LIST(n)->as.nd_alen == 2)) {
7824 COMPILE_ERROR(ERROR_ARGS "unexpected node");
7825 return COMPILE_NG;
7826 }
7827
7828 ADD_INSN(ret, line_node, dup); // (1)
7829 CHECK(iseq_compile_pattern_match(iseq, ret, RNODE_LIST(n)->nd_head, match_failed, in_single_pattern, in_alt_pattern, base_index + 1 /* (1) */, use_deconstructed_cache));
7830 CHECK(iseq_compile_pattern_each(iseq, ret, RNODE_LIST(RNODE_LIST(n)->nd_next)->nd_head, matched, match_failed, in_single_pattern, in_alt_pattern, base_index, false));
7831 ADD_INSN(ret, line_node, putnil);
7832
7833 ADD_LABEL(ret, match_failed);
7834 ADD_INSN(ret, line_node, pop);
7835 ADD_INSNL(ret, line_node, jump, unmatched);
7836 break;
7837 }
7838 case NODE_OR: {
7839 LABEL *match_succeeded, *fin;
7840 match_succeeded = NEW_LABEL(line);
7841 fin = NEW_LABEL(line);
7842
7843 ADD_INSN(ret, line_node, dup); // (1)
7844 CHECK(iseq_compile_pattern_each(iseq, ret, RNODE_OR(node)->nd_1st, match_succeeded, fin, in_single_pattern, true, base_index + 1 /* (1) */, use_deconstructed_cache));
7845 ADD_LABEL(ret, match_succeeded);
7846 ADD_INSN(ret, line_node, pop);
7847 ADD_INSNL(ret, line_node, jump, matched);
7848 ADD_INSN(ret, line_node, putnil);
7849 ADD_LABEL(ret, fin);
7850 CHECK(iseq_compile_pattern_each(iseq, ret, RNODE_OR(node)->nd_2nd, matched, unmatched, in_single_pattern, true, base_index, use_deconstructed_cache));
7851 break;
7852 }
7853 default:
7854 UNKNOWN_NODE("NODE_IN", node, COMPILE_NG);
7855 }
7856 return COMPILE_OK;
7857}
7858
7859static int
7860iseq_compile_pattern_match(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, LABEL *unmatched, bool in_single_pattern, bool in_alt_pattern, int base_index, bool use_deconstructed_cache)
7861{
7862 LABEL *fin = NEW_LABEL(nd_line(node));
7863 CHECK(iseq_compile_pattern_each(iseq, ret, node, fin, unmatched, in_single_pattern, in_alt_pattern, base_index, use_deconstructed_cache));
7864 ADD_LABEL(ret, fin);
7865 return COMPILE_OK;
7866}
7867
7868static int
7869iseq_compile_pattern_constant(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, LABEL *match_failed, bool in_single_pattern, int base_index)
7870{
7871 const NODE *line_node = node;
7872
7873 if (RNODE_ARYPTN(node)->nd_pconst) {
7874 ADD_INSN(ret, line_node, dup); // (1)
7875 CHECK(COMPILE(ret, "constant", RNODE_ARYPTN(node)->nd_pconst)); // (2)
7876 if (in_single_pattern) {
7877 ADD_INSN1(ret, line_node, dupn, INT2FIX(2));
7878 }
7879 ADD_INSN1(ret, line_node, checkmatch, INT2FIX(VM_CHECKMATCH_TYPE_CASE)); // (3)
7880 if (in_single_pattern) {
7881 CHECK(iseq_compile_pattern_set_eqq_errmsg(iseq, ret, node, base_index + 3 /* (1), (2), (3) */));
7882 }
7883 ADD_INSNL(ret, line_node, branchunless, match_failed);
7884 }
7885 return COMPILE_OK;
7886}
7887
7888
7889static int
7890iseq_compile_array_deconstruct(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, LABEL *deconstruct, LABEL *deconstructed, LABEL *match_failed, LABEL *type_error, bool in_single_pattern, int base_index, bool use_deconstructed_cache)
7891{
7892 const NODE *line_node = node;
7893
7894 // NOTE: this optimization allows us to re-use the #deconstruct value
7895 // (or its absence).
7896 if (use_deconstructed_cache) {
7897 // If value is nil then we haven't tried to deconstruct
7898 ADD_INSN1(ret, line_node, topn, INT2FIX(base_index + CASE3_BI_OFFSET_DECONSTRUCTED_CACHE));
7899 ADD_INSNL(ret, line_node, branchnil, deconstruct);
7900
7901 // If false then the value is not deconstructable
7902 ADD_INSN1(ret, line_node, topn, INT2FIX(base_index + CASE3_BI_OFFSET_DECONSTRUCTED_CACHE));
7903 ADD_INSNL(ret, line_node, branchunless, match_failed);
7904
7905 // Drop value, add deconstructed to the stack and jump
7906 ADD_INSN(ret, line_node, pop); // (1)
7907 ADD_INSN1(ret, line_node, topn, INT2FIX(base_index + CASE3_BI_OFFSET_DECONSTRUCTED_CACHE - 1 /* (1) */));
7908 ADD_INSNL(ret, line_node, jump, deconstructed);
7909 }
7910 else {
7911 ADD_INSNL(ret, line_node, jump, deconstruct);
7912 }
7913
7914 ADD_LABEL(ret, deconstruct);
7915 ADD_INSN(ret, line_node, dup);
7916 ADD_INSN1(ret, line_node, putobject, ID2SYM(rb_intern("deconstruct")));
7917 ADD_SEND(ret, line_node, idRespond_to, INT2FIX(1)); // (2)
7918
7919 // Cache the result of respond_to? (in case it's false is stays there, if true - it's overwritten after #deconstruct)
7920 if (use_deconstructed_cache) {
7921 ADD_INSN1(ret, line_node, setn, INT2FIX(base_index + CASE3_BI_OFFSET_DECONSTRUCTED_CACHE + 1 /* (2) */));
7922 }
7923
7924 if (in_single_pattern) {
7925 CHECK(iseq_compile_pattern_set_general_errmsg(iseq, ret, node, rb_fstring_lit("%p does not respond to #deconstruct"), base_index + 1 /* (2) */));
7926 }
7927
7928 ADD_INSNL(ret, line_node, branchunless, match_failed);
7929
7930 ADD_SEND(ret, line_node, rb_intern("deconstruct"), INT2FIX(0));
7931
7932 // Cache the result (if it's cacheable - currently, only top-level array patterns)
7933 if (use_deconstructed_cache) {
7934 ADD_INSN1(ret, line_node, setn, INT2FIX(base_index + CASE3_BI_OFFSET_DECONSTRUCTED_CACHE));
7935 }
7936
7937 ADD_INSN(ret, line_node, dup);
7938 ADD_INSN1(ret, line_node, checktype, INT2FIX(T_ARRAY));
7939 ADD_INSNL(ret, line_node, branchunless, type_error);
7940
7941 ADD_LABEL(ret, deconstructed);
7942
7943 return COMPILE_OK;
7944}
7945
7946static int
7947iseq_compile_pattern_set_general_errmsg(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, VALUE errmsg, int base_index)
7948{
7949 /*
7950 * if match_succeeded?
7951 * goto match_succeeded
7952 * end
7953 * error_string = FrozenCore.sprintf(errmsg, matchee)
7954 * key_error_p = false
7955 * match_succeeded:
7956 */
7957 const int line = nd_line(node);
7958 const NODE *line_node = node;
7959 LABEL *match_succeeded = NEW_LABEL(line);
7960
7961 ADD_INSN(ret, line_node, dup);
7962 ADD_INSNL(ret, line_node, branchif, match_succeeded);
7963
7964 ADD_INSN1(ret, line_node, putspecialobject, INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
7965 ADD_INSN1(ret, line_node, putobject, errmsg);
7966 ADD_INSN1(ret, line_node, topn, INT2FIX(3));
7967 ADD_SEND(ret, line_node, id_core_sprintf, INT2FIX(2)); // (1)
7968 ADD_INSN1(ret, line_node, setn, INT2FIX(base_index + CASE3_BI_OFFSET_ERROR_STRING + 1 /* (1) */)); // (2)
7969
7970 ADD_INSN1(ret, line_node, putobject, Qfalse);
7971 ADD_INSN1(ret, line_node, setn, INT2FIX(base_index + CASE3_BI_OFFSET_KEY_ERROR_P + 2 /* (1), (2) */));
7972
7973 ADD_INSN(ret, line_node, pop);
7974 ADD_INSN(ret, line_node, pop);
7975 ADD_LABEL(ret, match_succeeded);
7976
7977 return COMPILE_OK;
7978}
7979
7980static int
7981iseq_compile_pattern_set_length_errmsg(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, VALUE errmsg, VALUE pattern_length, int base_index)
7982{
7983 /*
7984 * if match_succeeded?
7985 * goto match_succeeded
7986 * end
7987 * error_string = FrozenCore.sprintf(errmsg, matchee, matchee.length, pat.length)
7988 * key_error_p = false
7989 * match_succeeded:
7990 */
7991 const int line = nd_line(node);
7992 const NODE *line_node = node;
7993 LABEL *match_succeeded = NEW_LABEL(line);
7994
7995 ADD_INSN(ret, line_node, dup);
7996 ADD_INSNL(ret, line_node, branchif, match_succeeded);
7997
7998 ADD_INSN1(ret, line_node, putspecialobject, INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
7999 ADD_INSN1(ret, line_node, putobject, errmsg);
8000 ADD_INSN1(ret, line_node, topn, INT2FIX(3));
8001 ADD_INSN(ret, line_node, dup);
8002 ADD_SEND(ret, line_node, idLength, INT2FIX(0));
8003 ADD_INSN1(ret, line_node, putobject, pattern_length);
8004 ADD_SEND(ret, line_node, id_core_sprintf, INT2FIX(4)); // (1)
8005 ADD_INSN1(ret, line_node, setn, INT2FIX(base_index + CASE3_BI_OFFSET_ERROR_STRING + 1 /* (1) */)); // (2)
8006
8007 ADD_INSN1(ret, line_node, putobject, Qfalse);
8008 ADD_INSN1(ret, line_node, setn, INT2FIX(base_index + CASE3_BI_OFFSET_KEY_ERROR_P + 2/* (1), (2) */));
8009
8010 ADD_INSN(ret, line_node, pop);
8011 ADD_INSN(ret, line_node, pop);
8012 ADD_LABEL(ret, match_succeeded);
8013
8014 return COMPILE_OK;
8015}
8016
8017static int
8018iseq_compile_pattern_set_eqq_errmsg(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, int base_index)
8019{
8020 /*
8021 * if match_succeeded?
8022 * goto match_succeeded
8023 * end
8024 * error_string = FrozenCore.sprintf("%p === %p does not return true", pat, matchee)
8025 * key_error_p = false
8026 * match_succeeded:
8027 */
8028 const int line = nd_line(node);
8029 const NODE *line_node = node;
8030 LABEL *match_succeeded = NEW_LABEL(line);
8031
8032 ADD_INSN(ret, line_node, dup);
8033 ADD_INSNL(ret, line_node, branchif, match_succeeded);
8034
8035 ADD_INSN1(ret, line_node, putspecialobject, INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
8036 ADD_INSN1(ret, line_node, putobject, rb_fstring_lit("%p === %p does not return true"));
8037 ADD_INSN1(ret, line_node, topn, INT2FIX(3));
8038 ADD_INSN1(ret, line_node, topn, INT2FIX(5));
8039 ADD_SEND(ret, line_node, id_core_sprintf, INT2FIX(3)); // (1)
8040 ADD_INSN1(ret, line_node, setn, INT2FIX(base_index + CASE3_BI_OFFSET_ERROR_STRING + 1 /* (1) */)); // (2)
8041
8042 ADD_INSN1(ret, line_node, putobject, Qfalse);
8043 ADD_INSN1(ret, line_node, setn, INT2FIX(base_index + CASE3_BI_OFFSET_KEY_ERROR_P + 2 /* (1), (2) */));
8044
8045 ADD_INSN(ret, line_node, pop);
8046 ADD_INSN(ret, line_node, pop);
8047
8048 ADD_LABEL(ret, match_succeeded);
8049 ADD_INSN1(ret, line_node, setn, INT2FIX(2));
8050 ADD_INSN(ret, line_node, pop);
8051 ADD_INSN(ret, line_node, pop);
8052
8053 return COMPILE_OK;
8054}
8055
8056static int
8057compile_case3(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const orig_node, int popped)
8058{
8059 const NODE *pattern;
8060 const NODE *node = orig_node;
8061 LABEL *endlabel, *elselabel;
8062 DECL_ANCHOR(head);
8063 DECL_ANCHOR(body_seq);
8064 DECL_ANCHOR(cond_seq);
8065 int line;
8066 enum node_type type;
8067 const NODE *line_node;
8068 VALUE branches = 0;
8069 int branch_id = 0;
8070 bool single_pattern;
8071
8072 INIT_ANCHOR(head);
8073 INIT_ANCHOR(body_seq);
8074 INIT_ANCHOR(cond_seq);
8075
8076 branches = decl_branch_base(iseq, PTR2NUM(node), nd_code_loc(node), "case");
8077
8078 node = RNODE_CASE3(node)->nd_body;
8079 EXPECT_NODE("NODE_CASE3", node, NODE_IN, COMPILE_NG);
8080 type = nd_type(node);
8081 line = nd_line(node);
8082 line_node = node;
8083 single_pattern = !RNODE_IN(node)->nd_next;
8084
8085 endlabel = NEW_LABEL(line);
8086 elselabel = NEW_LABEL(line);
8087
8088 if (single_pattern) {
8089 /* allocate stack for ... */
8090 ADD_INSN(head, line_node, putnil); /* key_error_key */
8091 ADD_INSN(head, line_node, putnil); /* key_error_matchee */
8092 ADD_INSN1(head, line_node, putobject, Qfalse); /* key_error_p */
8093 ADD_INSN(head, line_node, putnil); /* error_string */
8094 }
8095 ADD_INSN(head, line_node, putnil); /* allocate stack for cached #deconstruct value */
8096
8097 CHECK(COMPILE(head, "case base", RNODE_CASE3(orig_node)->nd_head));
8098
8099 ADD_SEQ(ret, head); /* case VAL */
8100
8101 while (type == NODE_IN) {
8102 LABEL *l1;
8103
8104 if (branch_id) {
8105 ADD_INSN(body_seq, line_node, putnil);
8106 }
8107 l1 = NEW_LABEL(line);
8108 ADD_LABEL(body_seq, l1);
8109 ADD_INSN1(body_seq, line_node, adjuststack, INT2FIX(single_pattern ? 6 : 2));
8110
8111 const NODE *const coverage_node = RNODE_IN(node)->nd_body ? RNODE_IN(node)->nd_body : node;
8112 add_trace_branch_coverage(
8113 iseq,
8114 body_seq,
8115 nd_code_loc(coverage_node),
8116 nd_node_id(coverage_node),
8117 branch_id++,
8118 "in",
8119 branches);
8120
8121 CHECK(COMPILE_(body_seq, "in body", RNODE_IN(node)->nd_body, popped));
8122 ADD_INSNL(body_seq, line_node, jump, endlabel);
8123
8124 pattern = RNODE_IN(node)->nd_head;
8125 if (pattern) {
8126 int pat_line = nd_line(pattern);
8127 LABEL *next_pat = NEW_LABEL(pat_line);
8128 ADD_INSN (cond_seq, pattern, dup); /* dup case VAL */
8129 // NOTE: set base_index (it's "under" the matchee value, so it's position is 2)
8130 CHECK(iseq_compile_pattern_each(iseq, cond_seq, pattern, l1, next_pat, single_pattern, false, 2, true));
8131 ADD_LABEL(cond_seq, next_pat);
8132 LABEL_UNREMOVABLE(next_pat);
8133 }
8134 else {
8135 COMPILE_ERROR(ERROR_ARGS "unexpected node");
8136 return COMPILE_NG;
8137 }
8138
8139 node = RNODE_IN(node)->nd_next;
8140 if (!node) {
8141 break;
8142 }
8143 type = nd_type(node);
8144 line = nd_line(node);
8145 line_node = node;
8146 }
8147 /* else */
8148 if (node) {
8149 ADD_LABEL(cond_seq, elselabel);
8150 ADD_INSN(cond_seq, line_node, pop);
8151 ADD_INSN(cond_seq, line_node, pop); /* discard cached #deconstruct value */
8152 add_trace_branch_coverage(iseq, cond_seq, nd_code_loc(node), nd_node_id(node), branch_id, "else", branches);
8153 CHECK(COMPILE_(cond_seq, "else", node, popped));
8154 ADD_INSNL(cond_seq, line_node, jump, endlabel);
8155 ADD_INSN(cond_seq, line_node, putnil);
8156 if (popped) {
8157 ADD_INSN(cond_seq, line_node, putnil);
8158 }
8159 }
8160 else {
8161 debugs("== else (implicit)\n");
8162 ADD_LABEL(cond_seq, elselabel);
8163 add_trace_branch_coverage(iseq, cond_seq, nd_code_loc(orig_node), nd_node_id(orig_node), branch_id, "else", branches);
8164 ADD_INSN1(cond_seq, orig_node, putspecialobject, INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
8165
8166 if (single_pattern) {
8167 /*
8168 * if key_error_p
8169 * FrozenCore.raise NoMatchingPatternKeyError.new(FrozenCore.sprintf("%p: %s", case_val, error_string), matchee: key_error_matchee, key: key_error_key)
8170 * else
8171 * FrozenCore.raise NoMatchingPatternError, FrozenCore.sprintf("%p: %s", case_val, error_string)
8172 * end
8173 */
8174 LABEL *key_error, *fin;
8175 struct rb_callinfo_kwarg *kw_arg;
8176
8177 key_error = NEW_LABEL(line);
8178 fin = NEW_LABEL(line);
8179
8180 kw_arg = rb_xmalloc_mul_add(2, sizeof(VALUE), sizeof(struct rb_callinfo_kwarg));
8181 kw_arg->references = 0;
8182 kw_arg->keyword_len = 2;
8183 kw_arg->keywords[0] = ID2SYM(rb_intern("matchee"));
8184 kw_arg->keywords[1] = ID2SYM(rb_intern("key"));
8185
8186 ADD_INSN1(cond_seq, orig_node, topn, INT2FIX(CASE3_BI_OFFSET_KEY_ERROR_P + 2));
8187 ADD_INSNL(cond_seq, orig_node, branchif, key_error);
8188 ADD_INSN1(cond_seq, orig_node, putobject, rb_eNoMatchingPatternError);
8189 ADD_INSN1(cond_seq, orig_node, putspecialobject, INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
8190 ADD_INSN1(cond_seq, orig_node, putobject, rb_fstring_lit("%p: %s"));
8191 ADD_INSN1(cond_seq, orig_node, topn, INT2FIX(4)); /* case VAL */
8192 ADD_INSN1(cond_seq, orig_node, topn, INT2FIX(CASE3_BI_OFFSET_ERROR_STRING + 6));
8193 ADD_SEND(cond_seq, orig_node, id_core_sprintf, INT2FIX(3));
8194 ADD_SEND(cond_seq, orig_node, id_core_raise, INT2FIX(2));
8195 ADD_INSNL(cond_seq, orig_node, jump, fin);
8196
8197 ADD_LABEL(cond_seq, key_error);
8198 ADD_INSN1(cond_seq, orig_node, putobject, rb_eNoMatchingPatternKeyError);
8199 ADD_INSN1(cond_seq, orig_node, putspecialobject, INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
8200 ADD_INSN1(cond_seq, orig_node, putobject, rb_fstring_lit("%p: %s"));
8201 ADD_INSN1(cond_seq, orig_node, topn, INT2FIX(4)); /* case VAL */
8202 ADD_INSN1(cond_seq, orig_node, topn, INT2FIX(CASE3_BI_OFFSET_ERROR_STRING + 6));
8203 ADD_SEND(cond_seq, orig_node, id_core_sprintf, INT2FIX(3));
8204 ADD_INSN1(cond_seq, orig_node, topn, INT2FIX(CASE3_BI_OFFSET_KEY_ERROR_MATCHEE + 4));
8205 ADD_INSN1(cond_seq, orig_node, topn, INT2FIX(CASE3_BI_OFFSET_KEY_ERROR_KEY + 5));
8206 ADD_SEND_R(cond_seq, orig_node, rb_intern("new"), INT2FIX(1), NULL, INT2FIX(VM_CALL_KWARG), kw_arg);
8207 ADD_SEND(cond_seq, orig_node, id_core_raise, INT2FIX(1));
8208
8209 ADD_LABEL(cond_seq, fin);
8210 }
8211 else {
8212 ADD_INSN1(cond_seq, orig_node, putobject, rb_eNoMatchingPatternError);
8213 ADD_INSN1(cond_seq, orig_node, topn, INT2FIX(2));
8214 ADD_SEND(cond_seq, orig_node, id_core_raise, INT2FIX(2));
8215 }
8216 ADD_INSN1(cond_seq, orig_node, adjuststack, INT2FIX(single_pattern ? 7 : 3));
8217 if (!popped) {
8218 ADD_INSN(cond_seq, orig_node, putnil);
8219 }
8220 ADD_INSNL(cond_seq, orig_node, jump, endlabel);
8221 ADD_INSN1(cond_seq, orig_node, dupn, INT2FIX(single_pattern ? 5 : 1));
8222 if (popped) {
8223 ADD_INSN(cond_seq, line_node, putnil);
8224 }
8225 }
8226
8227 ADD_SEQ(ret, cond_seq);
8228 ADD_SEQ(ret, body_seq);
8229 ADD_LABEL(ret, endlabel);
8230 return COMPILE_OK;
8231}
8232
8233#undef CASE3_BI_OFFSET_DECONSTRUCTED_CACHE
8234#undef CASE3_BI_OFFSET_ERROR_STRING
8235#undef CASE3_BI_OFFSET_KEY_ERROR_P
8236#undef CASE3_BI_OFFSET_KEY_ERROR_MATCHEE
8237#undef CASE3_BI_OFFSET_KEY_ERROR_KEY
8238
8239static int
8240compile_loop(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, int popped, const enum node_type type)
8241{
8242 const int line = (int)nd_line(node);
8243 const NODE *line_node = node;
8244
8245 LABEL *prev_start_label = ISEQ_COMPILE_DATA(iseq)->start_label;
8246 LABEL *prev_end_label = ISEQ_COMPILE_DATA(iseq)->end_label;
8247 LABEL *prev_redo_label = ISEQ_COMPILE_DATA(iseq)->redo_label;
8248 int prev_loopval_popped = ISEQ_COMPILE_DATA(iseq)->loopval_popped;
8249 VALUE branches = Qfalse;
8250
8252
8253 LABEL *next_label = ISEQ_COMPILE_DATA(iseq)->start_label = NEW_LABEL(line); /* next */
8254 LABEL *redo_label = ISEQ_COMPILE_DATA(iseq)->redo_label = NEW_LABEL(line); /* redo */
8255 LABEL *break_label = ISEQ_COMPILE_DATA(iseq)->end_label = NEW_LABEL(line); /* break */
8256 LABEL *end_label = NEW_LABEL(line);
8257 LABEL *adjust_label = NEW_LABEL(line);
8258
8259 LABEL *next_catch_label = NEW_LABEL(line);
8260 LABEL *tmp_label = NULL;
8261
8262 ISEQ_COMPILE_DATA(iseq)->loopval_popped = 0;
8263 push_ensure_entry(iseq, &enl, NULL, NULL);
8264
8265 if (RNODE_WHILE(node)->nd_state == 1) {
8266 ADD_INSNL(ret, line_node, jump, next_label);
8267 }
8268 else {
8269 tmp_label = NEW_LABEL(line);
8270 ADD_INSNL(ret, line_node, jump, tmp_label);
8271 }
8272 ADD_LABEL(ret, adjust_label);
8273 ADD_INSN(ret, line_node, putnil);
8274 ADD_LABEL(ret, next_catch_label);
8275 ADD_INSN(ret, line_node, pop);
8276 ADD_INSNL(ret, line_node, jump, next_label);
8277 if (tmp_label) ADD_LABEL(ret, tmp_label);
8278
8279 ADD_LABEL(ret, redo_label);
8280 branches = decl_branch_base(iseq, PTR2NUM(node), nd_code_loc(node), type == NODE_WHILE ? "while" : "until");
8281
8282 const NODE *const coverage_node = RNODE_WHILE(node)->nd_body ? RNODE_WHILE(node)->nd_body : node;
8283 add_trace_branch_coverage(
8284 iseq,
8285 ret,
8286 nd_code_loc(coverage_node),
8287 nd_node_id(coverage_node),
8288 0,
8289 "body",
8290 branches);
8291
8292 CHECK(COMPILE_POPPED(ret, "while body", RNODE_WHILE(node)->nd_body));
8293 ADD_LABEL(ret, next_label); /* next */
8294
8295 if (type == NODE_WHILE) {
8296 CHECK(compile_branch_condition(iseq, ret, RNODE_WHILE(node)->nd_cond,
8297 redo_label, end_label));
8298 }
8299 else {
8300 /* until */
8301 CHECK(compile_branch_condition(iseq, ret, RNODE_WHILE(node)->nd_cond,
8302 end_label, redo_label));
8303 }
8304
8305 ADD_LABEL(ret, end_label);
8306 ADD_ADJUST_RESTORE(ret, adjust_label);
8307
8308 if (UNDEF_P(RNODE_WHILE(node)->nd_state)) {
8309 /* ADD_INSN(ret, line_node, putundef); */
8310 COMPILE_ERROR(ERROR_ARGS "unsupported: putundef");
8311 return COMPILE_NG;
8312 }
8313 else {
8314 ADD_INSN(ret, line_node, putnil);
8315 }
8316
8317 ADD_LABEL(ret, break_label); /* break */
8318
8319 if (popped) {
8320 ADD_INSN(ret, line_node, pop);
8321 }
8322
8323 ADD_CATCH_ENTRY(CATCH_TYPE_BREAK, redo_label, break_label, NULL,
8324 break_label);
8325 ADD_CATCH_ENTRY(CATCH_TYPE_NEXT, redo_label, break_label, NULL,
8326 next_catch_label);
8327 ADD_CATCH_ENTRY(CATCH_TYPE_REDO, redo_label, break_label, NULL,
8328 ISEQ_COMPILE_DATA(iseq)->redo_label);
8329
8330 ISEQ_COMPILE_DATA(iseq)->start_label = prev_start_label;
8331 ISEQ_COMPILE_DATA(iseq)->end_label = prev_end_label;
8332 ISEQ_COMPILE_DATA(iseq)->redo_label = prev_redo_label;
8333 ISEQ_COMPILE_DATA(iseq)->loopval_popped = prev_loopval_popped;
8334 ISEQ_COMPILE_DATA(iseq)->ensure_node_stack = ISEQ_COMPILE_DATA(iseq)->ensure_node_stack->prev;
8335 return COMPILE_OK;
8336}
8337
8338static int
8339compile_iter(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, int popped)
8340{
8341 const int line = nd_line(node);
8342 const NODE *line_node = node;
8343 const rb_iseq_t *prevblock = ISEQ_COMPILE_DATA(iseq)->current_block;
8344 LABEL *retry_label = NEW_LABEL(line);
8345 LABEL *retry_end_l = NEW_LABEL(line);
8346 const rb_iseq_t *child_iseq;
8347
8348 ADD_LABEL(ret, retry_label);
8349 if (nd_type_p(node, NODE_FOR)) {
8350 CHECK(COMPILE(ret, "iter caller (for)", RNODE_FOR(node)->nd_iter));
8351
8352 ISEQ_COMPILE_DATA(iseq)->current_block = child_iseq =
8353 NEW_CHILD_ISEQ(RNODE_FOR(node)->nd_body, make_name_for_block(iseq),
8354 ISEQ_TYPE_BLOCK, line);
8355 ADD_SEND_WITH_BLOCK(ret, line_node, idEach, INT2FIX(0), child_iseq);
8356 }
8357 else {
8358 ISEQ_COMPILE_DATA(iseq)->current_block = child_iseq =
8359 NEW_CHILD_ISEQ(RNODE_ITER(node)->nd_body, make_name_for_block(iseq),
8360 ISEQ_TYPE_BLOCK, line);
8361 CHECK(COMPILE(ret, "iter caller", RNODE_ITER(node)->nd_iter));
8362 }
8363
8364 {
8365 // We need to put the label "retry_end_l" immediately after the last "send" instruction.
8366 // This because vm_throw checks if the break cont is equal to the index of next insn of the "send".
8367 // (Otherwise, it is considered "break from proc-closure". See "TAG_BREAK" handling in "vm_throw_start".)
8368 //
8369 // Normally, "send" instruction is at the last.
8370 // However, qcall under branch coverage measurement adds some instructions after the "send".
8371 //
8372 // Note that "invokesuper", "invokesuperforward" appears instead of "send".
8373 INSN *iobj;
8374 LINK_ELEMENT *last_elem = LAST_ELEMENT(ret);
8375 iobj = IS_INSN(last_elem) ? (INSN*) last_elem : (INSN*) get_prev_insn((INSN*) last_elem);
8376 while (!IS_INSN_ID(iobj, send) && !IS_INSN_ID(iobj, invokesuper) && !IS_INSN_ID(iobj, sendforward) && !IS_INSN_ID(iobj, invokesuperforward)) {
8377 iobj = (INSN*) get_prev_insn(iobj);
8378 }
8379 ELEM_INSERT_NEXT(&iobj->link, (LINK_ELEMENT*) retry_end_l);
8380
8381 // LINK_ANCHOR has a pointer to the last element, but ELEM_INSERT_NEXT does not update it
8382 // even if we add an insn to the last of LINK_ANCHOR. So this updates it manually.
8383 if (&iobj->link == LAST_ELEMENT(ret)) {
8384 ret->last = (LINK_ELEMENT*) retry_end_l;
8385 }
8386 }
8387
8388 if (popped) {
8389 ADD_INSN(ret, line_node, pop);
8390 }
8391
8392 ISEQ_COMPILE_DATA(iseq)->current_block = prevblock;
8393
8394 ADD_CATCH_ENTRY(CATCH_TYPE_BREAK, retry_label, retry_end_l, child_iseq, retry_end_l);
8395 return COMPILE_OK;
8396}
8397
8398static int
8399compile_for_masgn(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, int popped)
8400{
8401 /* massign to var in "for"
8402 * (args.length == 1 && Array.try_convert(args[0])) || args
8403 */
8404 const NODE *line_node = node;
8405 const NODE *var = RNODE_FOR_MASGN(node)->nd_var;
8406 LABEL *not_single = NEW_LABEL(nd_line(var));
8407 LABEL *not_ary = NEW_LABEL(nd_line(var));
8408 CHECK(COMPILE(ret, "for var", var));
8409 ADD_INSN(ret, line_node, dup);
8410 ADD_CALL(ret, line_node, idLength, INT2FIX(0));
8411 ADD_INSN1(ret, line_node, putobject, INT2FIX(1));
8412 ADD_CALL(ret, line_node, idEq, INT2FIX(1));
8413 ADD_INSNL(ret, line_node, branchunless, not_single);
8414 ADD_INSN(ret, line_node, dup);
8415 ADD_INSN1(ret, line_node, putobject, INT2FIX(0));
8416 ADD_CALL(ret, line_node, idAREF, INT2FIX(1));
8417 ADD_INSN1(ret, line_node, putobject, rb_cArray);
8418 ADD_INSN(ret, line_node, swap);
8419 ADD_CALL(ret, line_node, rb_intern("try_convert"), INT2FIX(1));
8420 ADD_INSN(ret, line_node, dup);
8421 ADD_INSNL(ret, line_node, branchunless, not_ary);
8422 ADD_INSN(ret, line_node, swap);
8423 ADD_LABEL(ret, not_ary);
8424 ADD_INSN(ret, line_node, pop);
8425 ADD_LABEL(ret, not_single);
8426 return COMPILE_OK;
8427}
8428
8429static int
8430compile_break(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, int popped)
8431{
8432 const NODE *line_node = node;
8433 unsigned long throw_flag = 0;
8434
8435 if (ISEQ_COMPILE_DATA(iseq)->redo_label != 0 && can_add_ensure_iseq(iseq)) {
8436 /* while/until */
8437 LABEL *splabel = NEW_LABEL(0);
8438 ADD_LABEL(ret, splabel);
8439 ADD_ADJUST(ret, line_node, ISEQ_COMPILE_DATA(iseq)->redo_label);
8440 CHECK(COMPILE_(ret, "break val (while/until)", RNODE_BREAK(node)->nd_stts,
8441 ISEQ_COMPILE_DATA(iseq)->loopval_popped));
8442 add_ensure_iseq(ret, iseq, 0);
8443 ADD_INSNL(ret, line_node, jump, ISEQ_COMPILE_DATA(iseq)->end_label);
8444 ADD_ADJUST_RESTORE(ret, splabel);
8445
8446 if (!popped) {
8447 ADD_INSN(ret, line_node, putnil);
8448 }
8449 }
8450 else {
8451 const rb_iseq_t *ip = iseq;
8452
8453 while (ip) {
8454 if (!ISEQ_COMPILE_DATA(ip)) {
8455 ip = 0;
8456 break;
8457 }
8458
8459 if (ISEQ_COMPILE_DATA(ip)->redo_label != 0) {
8460 throw_flag = VM_THROW_NO_ESCAPE_FLAG;
8461 }
8462 else if (ISEQ_BODY(ip)->type == ISEQ_TYPE_BLOCK) {
8463 throw_flag = 0;
8464 }
8465 else if (ISEQ_BODY(ip)->type == ISEQ_TYPE_EVAL) {
8466 COMPILE_ERROR(ERROR_ARGS "Can't escape from eval with break");
8467 return COMPILE_NG;
8468 }
8469 else {
8470 ip = ISEQ_BODY(ip)->parent_iseq;
8471 continue;
8472 }
8473
8474 /* escape from block */
8475 CHECK(COMPILE(ret, "break val (block)", RNODE_BREAK(node)->nd_stts));
8476 ADD_INSN1(ret, line_node, throw, INT2FIX(throw_flag | TAG_BREAK));
8477 if (popped) {
8478 ADD_INSN(ret, line_node, pop);
8479 }
8480 return COMPILE_OK;
8481 }
8482 COMPILE_ERROR(ERROR_ARGS "Invalid break");
8483 return COMPILE_NG;
8484 }
8485 return COMPILE_OK;
8486}
8487
8488static int
8489compile_next(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, int popped)
8490{
8491 const NODE *line_node = node;
8492 unsigned long throw_flag = 0;
8493
8494 if (ISEQ_COMPILE_DATA(iseq)->redo_label != 0 && can_add_ensure_iseq(iseq)) {
8495 LABEL *splabel = NEW_LABEL(0);
8496 debugs("next in while loop\n");
8497 ADD_LABEL(ret, splabel);
8498 CHECK(COMPILE(ret, "next val/valid syntax?", RNODE_NEXT(node)->nd_stts));
8499 add_ensure_iseq(ret, iseq, 0);
8500 ADD_ADJUST(ret, line_node, ISEQ_COMPILE_DATA(iseq)->redo_label);
8501 ADD_INSNL(ret, line_node, jump, ISEQ_COMPILE_DATA(iseq)->start_label);
8502 ADD_ADJUST_RESTORE(ret, splabel);
8503 if (!popped) {
8504 ADD_INSN(ret, line_node, putnil);
8505 }
8506 }
8507 else if (ISEQ_COMPILE_DATA(iseq)->end_label && can_add_ensure_iseq(iseq)) {
8508 LABEL *splabel = NEW_LABEL(0);
8509 debugs("next in block\n");
8510 ADD_LABEL(ret, splabel);
8511 ADD_ADJUST(ret, line_node, ISEQ_COMPILE_DATA(iseq)->start_label);
8512 CHECK(COMPILE(ret, "next val", RNODE_NEXT(node)->nd_stts));
8513 add_ensure_iseq(ret, iseq, 0);
8514 ADD_INSNL(ret, line_node, jump, ISEQ_COMPILE_DATA(iseq)->end_label);
8515 ADD_ADJUST_RESTORE(ret, splabel);
8516
8517 if (!popped) {
8518 ADD_INSN(ret, line_node, putnil);
8519 }
8520 }
8521 else {
8522 const rb_iseq_t *ip = iseq;
8523
8524 while (ip) {
8525 if (!ISEQ_COMPILE_DATA(ip)) {
8526 ip = 0;
8527 break;
8528 }
8529
8530 throw_flag = VM_THROW_NO_ESCAPE_FLAG;
8531 if (ISEQ_COMPILE_DATA(ip)->redo_label != 0) {
8532 /* while loop */
8533 break;
8534 }
8535 else if (ISEQ_BODY(ip)->type == ISEQ_TYPE_BLOCK) {
8536 break;
8537 }
8538 else if (ISEQ_BODY(ip)->type == ISEQ_TYPE_EVAL) {
8539 COMPILE_ERROR(ERROR_ARGS "Can't escape from eval with next");
8540 return COMPILE_NG;
8541 }
8542
8543 ip = ISEQ_BODY(ip)->parent_iseq;
8544 }
8545 if (ip != 0) {
8546 CHECK(COMPILE(ret, "next val", RNODE_NEXT(node)->nd_stts));
8547 ADD_INSN1(ret, line_node, throw, INT2FIX(throw_flag | TAG_NEXT));
8548
8549 if (popped) {
8550 ADD_INSN(ret, line_node, pop);
8551 }
8552 }
8553 else {
8554 COMPILE_ERROR(ERROR_ARGS "Invalid next");
8555 return COMPILE_NG;
8556 }
8557 }
8558 return COMPILE_OK;
8559}
8560
8561static int
8562compile_redo(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, int popped)
8563{
8564 const NODE *line_node = node;
8565
8566 if (ISEQ_COMPILE_DATA(iseq)->redo_label && can_add_ensure_iseq(iseq)) {
8567 LABEL *splabel = NEW_LABEL(0);
8568 debugs("redo in while");
8569 ADD_LABEL(ret, splabel);
8570 ADD_ADJUST(ret, line_node, ISEQ_COMPILE_DATA(iseq)->redo_label);
8571 add_ensure_iseq(ret, iseq, 0);
8572 ADD_INSNL(ret, line_node, jump, ISEQ_COMPILE_DATA(iseq)->redo_label);
8573 ADD_ADJUST_RESTORE(ret, splabel);
8574 if (!popped) {
8575 ADD_INSN(ret, line_node, putnil);
8576 }
8577 }
8578 else if (ISEQ_BODY(iseq)->type != ISEQ_TYPE_EVAL && ISEQ_COMPILE_DATA(iseq)->start_label && can_add_ensure_iseq(iseq)) {
8579 LABEL *splabel = NEW_LABEL(0);
8580
8581 debugs("redo in block");
8582 ADD_LABEL(ret, splabel);
8583 add_ensure_iseq(ret, iseq, 0);
8584 ADD_ADJUST(ret, line_node, ISEQ_COMPILE_DATA(iseq)->start_label);
8585 ADD_INSNL(ret, line_node, jump, ISEQ_COMPILE_DATA(iseq)->start_label);
8586 ADD_ADJUST_RESTORE(ret, splabel);
8587
8588 if (!popped) {
8589 ADD_INSN(ret, line_node, putnil);
8590 }
8591 }
8592 else {
8593 const rb_iseq_t *ip = iseq;
8594
8595 while (ip) {
8596 if (!ISEQ_COMPILE_DATA(ip)) {
8597 ip = 0;
8598 break;
8599 }
8600
8601 if (ISEQ_COMPILE_DATA(ip)->redo_label != 0) {
8602 break;
8603 }
8604 else if (ISEQ_BODY(ip)->type == ISEQ_TYPE_BLOCK) {
8605 break;
8606 }
8607 else if (ISEQ_BODY(ip)->type == ISEQ_TYPE_EVAL) {
8608 COMPILE_ERROR(ERROR_ARGS "Can't escape from eval with redo");
8609 return COMPILE_NG;
8610 }
8611
8612 ip = ISEQ_BODY(ip)->parent_iseq;
8613 }
8614 if (ip != 0) {
8615 ADD_INSN(ret, line_node, putnil);
8616 ADD_INSN1(ret, line_node, throw, INT2FIX(VM_THROW_NO_ESCAPE_FLAG | TAG_REDO));
8617
8618 if (popped) {
8619 ADD_INSN(ret, line_node, pop);
8620 }
8621 }
8622 else {
8623 COMPILE_ERROR(ERROR_ARGS "Invalid redo");
8624 return COMPILE_NG;
8625 }
8626 }
8627 return COMPILE_OK;
8628}
8629
8630static int
8631compile_retry(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, int popped)
8632{
8633 const NODE *line_node = node;
8634
8635 if (ISEQ_BODY(iseq)->type == ISEQ_TYPE_RESCUE) {
8636 ADD_INSN(ret, line_node, putnil);
8637 ADD_INSN1(ret, line_node, throw, INT2FIX(TAG_RETRY));
8638
8639 if (popped) {
8640 ADD_INSN(ret, line_node, pop);
8641 }
8642 }
8643 else {
8644 COMPILE_ERROR(ERROR_ARGS "Invalid retry");
8645 return COMPILE_NG;
8646 }
8647 return COMPILE_OK;
8648}
8649
8650static int
8651compile_rescue(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, int popped)
8652{
8653 const int line = nd_line(node);
8654 const NODE *line_node = node;
8655 LABEL *lstart = NEW_LABEL(line);
8656 LABEL *lend = NEW_LABEL(line);
8657 LABEL *lcont = NEW_LABEL(line);
8658 const rb_iseq_t *rescue = NEW_CHILD_ISEQ(RNODE_RESCUE(node)->nd_resq,
8659 rb_str_concat(rb_str_new2("rescue in "),
8660 ISEQ_BODY(iseq)->location.label),
8661 ISEQ_TYPE_RESCUE, line);
8662
8663 lstart->rescued = LABEL_RESCUE_BEG;
8664 lend->rescued = LABEL_RESCUE_END;
8665 ADD_LABEL(ret, lstart);
8666
8667 bool prev_in_rescue = ISEQ_COMPILE_DATA(iseq)->in_rescue;
8668 ISEQ_COMPILE_DATA(iseq)->in_rescue = true;
8669 {
8670 CHECK(COMPILE(ret, "rescue head", RNODE_RESCUE(node)->nd_head));
8671 }
8672 ISEQ_COMPILE_DATA(iseq)->in_rescue = prev_in_rescue;
8673
8674 ADD_LABEL(ret, lend);
8675 if (RNODE_RESCUE(node)->nd_else) {
8676 ADD_INSN(ret, line_node, pop);
8677 CHECK(COMPILE(ret, "rescue else", RNODE_RESCUE(node)->nd_else));
8678 }
8679 ADD_INSN(ret, line_node, nop);
8680 ADD_LABEL(ret, lcont);
8681
8682 if (popped) {
8683 ADD_INSN(ret, line_node, pop);
8684 }
8685
8686 /* register catch entry */
8687 ADD_CATCH_ENTRY(CATCH_TYPE_RESCUE, lstart, lend, rescue, lcont);
8688 ADD_CATCH_ENTRY(CATCH_TYPE_RETRY, lend, lcont, NULL, lstart);
8689 return COMPILE_OK;
8690}
8691
8692static int
8693compile_resbody(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, int popped)
8694{
8695 const int line = nd_line(node);
8696 const NODE *line_node = node;
8697 const NODE *resq = node;
8698 const NODE *narg;
8699 LABEL *label_miss, *label_hit;
8700
8701 while (resq) {
8702 label_miss = NEW_LABEL(line);
8703 label_hit = NEW_LABEL(line);
8704
8705 narg = RNODE_RESBODY(resq)->nd_args;
8706 if (narg) {
8707 switch (nd_type(narg)) {
8708 case NODE_LIST:
8709 while (narg) {
8710 ADD_GETLOCAL(ret, line_node, LVAR_ERRINFO, 0);
8711 CHECK(COMPILE(ret, "rescue arg", RNODE_LIST(narg)->nd_head));
8712 ADD_INSN1(ret, line_node, checkmatch, INT2FIX(VM_CHECKMATCH_TYPE_RESCUE));
8713 ADD_INSNL(ret, line_node, branchif, label_hit);
8714 narg = RNODE_LIST(narg)->nd_next;
8715 }
8716 break;
8717 case NODE_SPLAT:
8718 case NODE_ARGSCAT:
8719 case NODE_ARGSPUSH:
8720 ADD_GETLOCAL(ret, line_node, LVAR_ERRINFO, 0);
8721 CHECK(COMPILE(ret, "rescue/cond splat", narg));
8722 ADD_INSN1(ret, line_node, checkmatch, INT2FIX(VM_CHECKMATCH_TYPE_RESCUE | VM_CHECKMATCH_ARRAY));
8723 ADD_INSNL(ret, line_node, branchif, label_hit);
8724 break;
8725 default:
8726 UNKNOWN_NODE("NODE_RESBODY", narg, COMPILE_NG);
8727 }
8728 }
8729 else {
8730 ADD_GETLOCAL(ret, line_node, LVAR_ERRINFO, 0);
8731 ADD_INSN1(ret, line_node, putobject, rb_eStandardError);
8732 ADD_INSN1(ret, line_node, checkmatch, INT2FIX(VM_CHECKMATCH_TYPE_RESCUE));
8733 ADD_INSNL(ret, line_node, branchif, label_hit);
8734 }
8735 ADD_INSNL(ret, line_node, jump, label_miss);
8736 ADD_LABEL(ret, label_hit);
8737 ADD_TRACE(ret, RUBY_EVENT_RESCUE);
8738
8739 if (RNODE_RESBODY(resq)->nd_exc_var) {
8740 CHECK(COMPILE_POPPED(ret, "resbody exc_var", RNODE_RESBODY(resq)->nd_exc_var));
8741 }
8742
8743 if (nd_type(RNODE_RESBODY(resq)->nd_body) == NODE_BEGIN && RNODE_BEGIN(RNODE_RESBODY(resq)->nd_body)->nd_body == NULL && !RNODE_RESBODY(resq)->nd_exc_var) {
8744 // empty body
8745 ADD_SYNTHETIC_INSN(ret, nd_line(RNODE_RESBODY(resq)->nd_body), -1, putnil);
8746 }
8747 else {
8748 CHECK(COMPILE(ret, "resbody body", RNODE_RESBODY(resq)->nd_body));
8749 }
8750
8751 if (ISEQ_COMPILE_DATA(iseq)->option->tailcall_optimization) {
8752 ADD_INSN(ret, line_node, nop);
8753 }
8754 ADD_INSN(ret, line_node, leave);
8755 ADD_LABEL(ret, label_miss);
8756 resq = RNODE_RESBODY(resq)->nd_next;
8757 }
8758 return COMPILE_OK;
8759}
8760
8761static int
8762compile_ensure(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, int popped)
8763{
8764 const int line = nd_line(RNODE_ENSURE(node)->nd_ensr);
8765 const NODE *line_node = node;
8766 DECL_ANCHOR(ensr);
8767 const rb_iseq_t *ensure = NEW_CHILD_ISEQ(RNODE_ENSURE(node)->nd_ensr,
8768 rb_str_concat(rb_str_new2 ("ensure in "), ISEQ_BODY(iseq)->location.label),
8769 ISEQ_TYPE_ENSURE, line);
8770 LABEL *lstart = NEW_LABEL(line);
8771 LABEL *lend = NEW_LABEL(line);
8772 LABEL *lcont = NEW_LABEL(line);
8773 LINK_ELEMENT *last;
8774 int last_leave = 0;
8775 struct ensure_range er;
8777 struct ensure_range *erange;
8778
8779 INIT_ANCHOR(ensr);
8780 CHECK(COMPILE_POPPED(ensr, "ensure ensr", RNODE_ENSURE(node)->nd_ensr));
8781 last = ensr->last;
8782 last_leave = last && IS_INSN(last) && IS_INSN_ID(last, leave);
8783
8784 er.begin = lstart;
8785 er.end = lend;
8786 er.next = 0;
8787 push_ensure_entry(iseq, &enl, &er, RNODE_ENSURE(node)->nd_ensr);
8788
8789 ADD_LABEL(ret, lstart);
8790 CHECK(COMPILE_(ret, "ensure head", RNODE_ENSURE(node)->nd_head, (popped | last_leave)));
8791 ADD_LABEL(ret, lend);
8792 ADD_SEQ(ret, ensr);
8793 if (!popped && last_leave) ADD_INSN(ret, line_node, putnil);
8794 ADD_LABEL(ret, lcont);
8795 if (last_leave) ADD_INSN(ret, line_node, pop);
8796
8797 erange = ISEQ_COMPILE_DATA(iseq)->ensure_node_stack->erange;
8798 if (lstart->link.next != &lend->link) {
8799 while (erange) {
8800 ADD_CATCH_ENTRY(CATCH_TYPE_ENSURE, erange->begin, erange->end,
8801 ensure, lcont);
8802 erange = erange->next;
8803 }
8804 }
8805
8806 ISEQ_COMPILE_DATA(iseq)->ensure_node_stack = enl.prev;
8807 return COMPILE_OK;
8808}
8809
8810static int
8811compile_return(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, int popped)
8812{
8813 const NODE *line_node = node;
8814
8815 if (iseq) {
8816 enum rb_iseq_type type = ISEQ_BODY(iseq)->type;
8817 const rb_iseq_t *is = iseq;
8818 enum rb_iseq_type t = type;
8819 const NODE *retval = RNODE_RETURN(node)->nd_stts;
8820 LABEL *splabel = 0;
8821
8822 while (t == ISEQ_TYPE_RESCUE || t == ISEQ_TYPE_ENSURE) {
8823 if (!(is = ISEQ_BODY(is)->parent_iseq)) break;
8824 t = ISEQ_BODY(is)->type;
8825 }
8826 switch (t) {
8827 case ISEQ_TYPE_TOP:
8828 case ISEQ_TYPE_MAIN:
8829 if (retval) {
8830 rb_warn("argument of top-level return is ignored");
8831 }
8832 if (is == iseq) {
8833 /* plain top-level, leave directly */
8834 type = ISEQ_TYPE_METHOD;
8835 }
8836 break;
8837 default:
8838 break;
8839 }
8840
8841 if (type == ISEQ_TYPE_METHOD) {
8842 splabel = NEW_LABEL(0);
8843 ADD_LABEL(ret, splabel);
8844 ADD_ADJUST(ret, line_node, 0);
8845 }
8846
8847 CHECK(COMPILE(ret, "return nd_stts (return val)", retval));
8848
8849 if (type == ISEQ_TYPE_METHOD && can_add_ensure_iseq(iseq)) {
8850 add_ensure_iseq(ret, iseq, 1);
8851 ADD_TRACE(ret, RUBY_EVENT_RETURN);
8852 ADD_INSN(ret, line_node, leave);
8853 ADD_ADJUST_RESTORE(ret, splabel);
8854
8855 if (!popped) {
8856 ADD_INSN(ret, line_node, putnil);
8857 }
8858 }
8859 else {
8860 ADD_INSN1(ret, line_node, throw, INT2FIX(TAG_RETURN));
8861 if (popped) {
8862 ADD_INSN(ret, line_node, pop);
8863 }
8864 }
8865 }
8866 return COMPILE_OK;
8867}
8868
8869static bool
8870drop_unreachable_return(LINK_ANCHOR *ret)
8871{
8872 LINK_ELEMENT *i = ret->last, *last;
8873 if (!i) return false;
8874 if (IS_TRACE(i)) i = i->prev;
8875 if (!IS_INSN(i) || !IS_INSN_ID(i, putnil)) return false;
8876 last = i = i->prev;
8877 if (IS_ADJUST(i)) i = i->prev;
8878 if (!IS_INSN(i)) return false;
8879 switch (INSN_OF(i)) {
8880 case BIN(leave):
8881 case BIN(jump):
8882 break;
8883 default:
8884 return false;
8885 }
8886 (ret->last = last->prev)->next = NULL;
8887 return true;
8888}
8889
8890static int
8891compile_evstr(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, int popped)
8892{
8893 CHECK(COMPILE_(ret, "nd_body", node, popped));
8894
8895 if (!popped && !all_string_result_p(node)) {
8896 const NODE *line_node = node;
8897 const unsigned int flag = VM_CALL_FCALL;
8898
8899 // Note, this dup could be removed if we are willing to change anytostring. It pops
8900 // two VALUEs off the stack when it could work by replacing the top most VALUE.
8901 ADD_INSN(ret, line_node, dup);
8902 ADD_INSN1(ret, line_node, objtostring, new_callinfo(iseq, idTo_s, 0, flag, NULL, FALSE));
8903 ADD_INSN(ret, line_node, anytostring);
8904 }
8905 return COMPILE_OK;
8906}
8907
8908static void
8909compile_lvar(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *line_node, ID id)
8910{
8911 int idx = ISEQ_BODY(ISEQ_BODY(iseq)->local_iseq)->local_table_size - get_local_var_idx(iseq, id);
8912
8913 debugs("id: %s idx: %d\n", rb_id2name(id), idx);
8914 ADD_GETLOCAL(ret, line_node, idx, get_lvar_level(iseq));
8915}
8916
8917static LABEL *
8918qcall_branch_start(rb_iseq_t *iseq, LINK_ANCHOR *const recv, VALUE *branches, const NODE *node, const NODE *line_node)
8919{
8920 LABEL *else_label = NEW_LABEL(nd_line(line_node));
8921 VALUE br = 0;
8922
8923 br = decl_branch_base(iseq, PTR2NUM(node), nd_code_loc(node), "&.");
8924 *branches = br;
8925 ADD_INSN(recv, line_node, dup);
8926 ADD_INSNL(recv, line_node, branchnil, else_label);
8927 add_trace_branch_coverage(iseq, recv, nd_code_loc(node), nd_node_id(node), 0, "then", br);
8928 return else_label;
8929}
8930
8931static void
8932qcall_branch_end(rb_iseq_t *iseq, LINK_ANCHOR *const ret, LABEL *else_label, VALUE branches, const NODE *node, const NODE *line_node)
8933{
8934 LABEL *end_label;
8935 if (!else_label) return;
8936 end_label = NEW_LABEL(nd_line(line_node));
8937 ADD_INSNL(ret, line_node, jump, end_label);
8938 ADD_LABEL(ret, else_label);
8939 add_trace_branch_coverage(iseq, ret, nd_code_loc(node), nd_node_id(node), 1, "else", branches);
8940 ADD_LABEL(ret, end_label);
8941}
8942
8943static int
8944compile_call_precheck_freeze(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, const NODE *line_node, int popped)
8945{
8946 /* optimization shortcut
8947 * "literal".freeze -> opt_str_freeze("literal")
8948 */
8949 if (get_nd_recv(node) &&
8950 (nd_type_p(get_nd_recv(node), NODE_STR) || nd_type_p(get_nd_recv(node), NODE_FILE)) &&
8951 (get_node_call_nd_mid(node) == idFreeze || get_node_call_nd_mid(node) == idUMinus) &&
8952 get_nd_args(node) == NULL &&
8953 ISEQ_COMPILE_DATA(iseq)->current_block == NULL &&
8954 ISEQ_COMPILE_DATA(iseq)->option->specialized_instruction) {
8955 VALUE str = get_string_value(get_nd_recv(node));
8956 if (get_node_call_nd_mid(node) == idUMinus) {
8957 ADD_INSN2(ret, line_node, opt_str_uminus, str,
8958 new_callinfo(iseq, idUMinus, 0, 0, NULL, FALSE));
8959 }
8960 else {
8961 ADD_INSN2(ret, line_node, opt_str_freeze, str,
8962 new_callinfo(iseq, idFreeze, 0, 0, NULL, FALSE));
8963 }
8964 RB_OBJ_WRITTEN(iseq, Qundef, str);
8965 if (popped) {
8966 ADD_INSN(ret, line_node, pop);
8967 }
8968 return TRUE;
8969 }
8970 /* optimization shortcut
8971 * obj["literal"] -> opt_aref_with(obj, "literal")
8972 */
8973 if (get_node_call_nd_mid(node) == idAREF && !private_recv_p(node) && get_nd_args(node) &&
8974 nd_type_p(get_nd_args(node), NODE_LIST) && RNODE_LIST(get_nd_args(node))->as.nd_alen == 1 &&
8975 (nd_type_p(RNODE_LIST(get_nd_args(node))->nd_head, NODE_STR) || nd_type_p(RNODE_LIST(get_nd_args(node))->nd_head, NODE_FILE)) &&
8976 ISEQ_COMPILE_DATA(iseq)->current_block == NULL &&
8977 !frozen_string_literal_p(iseq) &&
8978 ISEQ_COMPILE_DATA(iseq)->option->specialized_instruction) {
8979 VALUE str = get_string_value(RNODE_LIST(get_nd_args(node))->nd_head);
8980 CHECK(COMPILE(ret, "recv", get_nd_recv(node)));
8981 ADD_INSN2(ret, line_node, opt_aref_with, str,
8982 new_callinfo(iseq, idAREF, 1, 0, NULL, FALSE));
8983 RB_OBJ_WRITTEN(iseq, Qundef, str);
8984 if (popped) {
8985 ADD_INSN(ret, line_node, pop);
8986 }
8987 return TRUE;
8988 }
8989 return FALSE;
8990}
8991
8992static int
8993iseq_has_builtin_function_table(const rb_iseq_t *iseq)
8994{
8995 return ISEQ_COMPILE_DATA(iseq)->builtin_function_table != NULL;
8996}
8997
8998static const struct rb_builtin_function *
8999iseq_builtin_function_lookup(const rb_iseq_t *iseq, const char *name)
9000{
9001 int i;
9002 const struct rb_builtin_function *table = ISEQ_COMPILE_DATA(iseq)->builtin_function_table;
9003 for (i=0; table[i].index != -1; i++) {
9004 if (strcmp(table[i].name, name) == 0) {
9005 return &table[i];
9006 }
9007 }
9008 return NULL;
9009}
9010
9011static const char *
9012iseq_builtin_function_name(const enum node_type type, const NODE *recv, ID mid)
9013{
9014 const char *name = rb_id2name(mid);
9015 static const char prefix[] = "__builtin_";
9016 const size_t prefix_len = sizeof(prefix) - 1;
9017
9018 switch (type) {
9019 case NODE_CALL:
9020 if (recv) {
9021 switch (nd_type(recv)) {
9022 case NODE_VCALL:
9023 if (RNODE_VCALL(recv)->nd_mid == rb_intern("__builtin")) {
9024 return name;
9025 }
9026 break;
9027 case NODE_CONST:
9028 if (RNODE_CONST(recv)->nd_vid == rb_intern("Primitive")) {
9029 return name;
9030 }
9031 break;
9032 default: break;
9033 }
9034 }
9035 break;
9036 case NODE_VCALL:
9037 case NODE_FCALL:
9038 if (UNLIKELY(strncmp(prefix, name, prefix_len) == 0)) {
9039 return &name[prefix_len];
9040 }
9041 break;
9042 default: break;
9043 }
9044 return NULL;
9045}
9046
9047static int
9048delegate_call_p(const rb_iseq_t *iseq, unsigned int argc, const LINK_ANCHOR *args, unsigned int *pstart_index)
9049{
9050
9051 if (argc == 0) {
9052 *pstart_index = 0;
9053 return TRUE;
9054 }
9055 else if (argc <= ISEQ_BODY(iseq)->local_table_size) {
9056 unsigned int start=0;
9057
9058 // local_table: [p1, p2, p3, l1, l2, l3]
9059 // arguments: [p3, l1, l2] -> 2
9060 for (start = 0;
9061 argc + start <= ISEQ_BODY(iseq)->local_table_size;
9062 start++) {
9063 const LINK_ELEMENT *elem = FIRST_ELEMENT(args);
9064
9065 for (unsigned int i=start; i-start<argc; i++) {
9066 if (IS_INSN(elem) &&
9067 INSN_OF(elem) == BIN(getlocal)) {
9068 int local_index = FIX2INT(OPERAND_AT(elem, 0));
9069 int local_level = FIX2INT(OPERAND_AT(elem, 1));
9070
9071 if (local_level == 0) {
9072 unsigned int index = ISEQ_BODY(iseq)->local_table_size - (local_index - VM_ENV_DATA_SIZE + 1);
9073 if (0) { // for debug
9074 fprintf(stderr, "lvar:%s (%d), id:%s (%d) local_index:%d, local_size:%d\n",
9075 rb_id2name(ISEQ_BODY(iseq)->local_table[i]), i,
9076 rb_id2name(ISEQ_BODY(iseq)->local_table[index]), index,
9077 local_index, (int)ISEQ_BODY(iseq)->local_table_size);
9078 }
9079 if (i == index) {
9080 elem = elem->next;
9081 continue; /* for */
9082 }
9083 else {
9084 goto next;
9085 }
9086 }
9087 else {
9088 goto fail; // level != 0 is unsupported
9089 }
9090 }
9091 else {
9092 goto fail; // insn is not a getlocal
9093 }
9094 }
9095 goto success;
9096 next:;
9097 }
9098 fail:
9099 return FALSE;
9100 success:
9101 *pstart_index = start;
9102 return TRUE;
9103 }
9104 else {
9105 return FALSE;
9106 }
9107}
9108
9109// Compile Primitive.attr! :leaf, ...
9110static int
9111compile_builtin_attr(rb_iseq_t *iseq, const NODE *node)
9112{
9113 VALUE symbol;
9114 VALUE string;
9115 if (!node) goto no_arg;
9116 while (node) {
9117 if (!nd_type_p(node, NODE_LIST)) goto bad_arg;
9118 const NODE *next = RNODE_LIST(node)->nd_next;
9119
9120 node = RNODE_LIST(node)->nd_head;
9121 if (!node) goto no_arg;
9122 switch (nd_type(node)) {
9123 case NODE_SYM:
9124 symbol = rb_node_sym_string_val(node);
9125 break;
9126 default:
9127 goto bad_arg;
9128 }
9129
9130 if (!SYMBOL_P(symbol)) goto non_symbol_arg;
9131
9132 string = rb_sym2str(symbol);
9133 if (strcmp(RSTRING_PTR(string), "leaf") == 0) {
9134 ISEQ_BODY(iseq)->builtin_attrs |= BUILTIN_ATTR_LEAF;
9135 }
9136 else if (strcmp(RSTRING_PTR(string), "inline_block") == 0) {
9137 ISEQ_BODY(iseq)->builtin_attrs |= BUILTIN_ATTR_INLINE_BLOCK;
9138 }
9139 else if (strcmp(RSTRING_PTR(string), "use_block") == 0) {
9140 iseq_set_use_block(iseq);
9141 }
9142 else if (strcmp(RSTRING_PTR(string), "c_trace") == 0) {
9143 // Let the iseq act like a C method in backtraces
9144 ISEQ_BODY(iseq)->builtin_attrs |= BUILTIN_ATTR_C_TRACE;
9145 }
9146 else {
9147 goto unknown_arg;
9148 }
9149 node = next;
9150 }
9151 return COMPILE_OK;
9152 no_arg:
9153 COMPILE_ERROR(ERROR_ARGS "attr!: no argument");
9154 return COMPILE_NG;
9155 non_symbol_arg:
9156 COMPILE_ERROR(ERROR_ARGS "non symbol argument to attr!: %s", rb_builtin_class_name(symbol));
9157 return COMPILE_NG;
9158 unknown_arg:
9159 COMPILE_ERROR(ERROR_ARGS "unknown argument to attr!: %s", RSTRING_PTR(string));
9160 return COMPILE_NG;
9161 bad_arg:
9162 UNKNOWN_NODE("attr!", node, COMPILE_NG);
9163}
9164
9165static int
9166compile_builtin_arg(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *node, const NODE *line_node, int popped)
9167{
9168 VALUE name;
9169
9170 if (!node) goto no_arg;
9171 if (!nd_type_p(node, NODE_LIST)) goto bad_arg;
9172 if (RNODE_LIST(node)->nd_next) goto too_many_arg;
9173 node = RNODE_LIST(node)->nd_head;
9174 if (!node) goto no_arg;
9175 switch (nd_type(node)) {
9176 case NODE_SYM:
9177 name = rb_node_sym_string_val(node);
9178 break;
9179 default:
9180 goto bad_arg;
9181 }
9182 if (!SYMBOL_P(name)) goto non_symbol_arg;
9183 if (!popped) {
9184 compile_lvar(iseq, ret, line_node, SYM2ID(name));
9185 }
9186 return COMPILE_OK;
9187 no_arg:
9188 COMPILE_ERROR(ERROR_ARGS "arg!: no argument");
9189 return COMPILE_NG;
9190 too_many_arg:
9191 COMPILE_ERROR(ERROR_ARGS "arg!: too many argument");
9192 return COMPILE_NG;
9193 non_symbol_arg:
9194 COMPILE_ERROR(ERROR_ARGS "non symbol argument to arg!: %s",
9195 rb_builtin_class_name(name));
9196 return COMPILE_NG;
9197 bad_arg:
9198 UNKNOWN_NODE("arg!", node, COMPILE_NG);
9199}
9200
9201static NODE *
9202mandatory_node(const rb_iseq_t *iseq, const NODE *cond_node)
9203{
9204 const NODE *node = ISEQ_COMPILE_DATA(iseq)->root_node;
9205 if (nd_type(node) == NODE_IF && RNODE_IF(node)->nd_cond == cond_node) {
9206 return RNODE_IF(node)->nd_body;
9207 }
9208 else {
9209 rb_bug("mandatory_node: can't find mandatory node");
9210 }
9211}
9212
9213static int
9214compile_builtin_mandatory_only_method(rb_iseq_t *iseq, const NODE *node, const NODE *line_node)
9215{
9216 // arguments
9217 struct rb_args_info args = {
9218 .pre_args_num = ISEQ_BODY(iseq)->param.lead_num,
9219 };
9220 rb_node_args_t args_node;
9221 rb_node_init(RNODE(&args_node), NODE_ARGS);
9222 args_node.nd_ainfo = args;
9223
9224 // local table without non-mandatory parameters
9225 const int skip_local_size = ISEQ_BODY(iseq)->param.size - ISEQ_BODY(iseq)->param.lead_num;
9226 const int table_size = ISEQ_BODY(iseq)->local_table_size - skip_local_size;
9227
9228 VALUE idtmp = 0;
9229 rb_ast_id_table_t *tbl = ALLOCV(idtmp, sizeof(rb_ast_id_table_t) + table_size * sizeof(ID));
9230 tbl->size = table_size;
9231
9232 int i;
9233
9234 // lead parameters
9235 for (i=0; i<ISEQ_BODY(iseq)->param.lead_num; i++) {
9236 tbl->ids[i] = ISEQ_BODY(iseq)->local_table[i];
9237 }
9238 // local variables
9239 for (; i<table_size; i++) {
9240 tbl->ids[i] = ISEQ_BODY(iseq)->local_table[i + skip_local_size];
9241 }
9242
9243 rb_node_scope_t scope_node;
9244 rb_node_init(RNODE(&scope_node), NODE_SCOPE);
9245 scope_node.nd_tbl = tbl;
9246 scope_node.nd_body = mandatory_node(iseq, node);
9247 scope_node.nd_args = &args_node;
9248
9249 VALUE ast_value = rb_ruby_ast_new(RNODE(&scope_node));
9250
9251 ISEQ_BODY(iseq)->mandatory_only_iseq =
9252 rb_iseq_new_with_opt(ast_value, rb_iseq_base_label(iseq),
9253 rb_iseq_path(iseq), rb_iseq_realpath(iseq),
9254 nd_line(line_node), NULL, 0,
9255 ISEQ_TYPE_METHOD, ISEQ_COMPILE_DATA(iseq)->option,
9256 ISEQ_BODY(iseq)->variable.script_lines);
9257
9258 ALLOCV_END(idtmp);
9259 return COMPILE_OK;
9260}
9261
9262static int
9263compile_builtin_function_call(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, const NODE *line_node, int popped,
9264 const rb_iseq_t *parent_block, LINK_ANCHOR *args, const char *builtin_func)
9265{
9266 NODE *args_node = get_nd_args(node);
9267
9268 if (parent_block != NULL) {
9269 COMPILE_ERROR(ERROR_ARGS_AT(line_node) "should not call builtins here.");
9270 return COMPILE_NG;
9271 }
9272 else {
9273# define BUILTIN_INLINE_PREFIX "_bi"
9274 char inline_func[sizeof(BUILTIN_INLINE_PREFIX) + DECIMAL_SIZE_OF(int)];
9275 bool cconst = false;
9276 retry:;
9277 const struct rb_builtin_function *bf = iseq_builtin_function_lookup(iseq, builtin_func);
9278
9279 if (bf == NULL) {
9280 if (strcmp("cstmt!", builtin_func) == 0 ||
9281 strcmp("cexpr!", builtin_func) == 0) {
9282 // ok
9283 }
9284 else if (strcmp("cconst!", builtin_func) == 0) {
9285 cconst = true;
9286 }
9287 else if (strcmp("cinit!", builtin_func) == 0) {
9288 // ignore
9289 return COMPILE_OK;
9290 }
9291 else if (strcmp("attr!", builtin_func) == 0) {
9292 return compile_builtin_attr(iseq, args_node);
9293 }
9294 else if (strcmp("arg!", builtin_func) == 0) {
9295 return compile_builtin_arg(iseq, ret, args_node, line_node, popped);
9296 }
9297 else if (strcmp("mandatory_only?", builtin_func) == 0) {
9298 if (popped) {
9299 rb_bug("mandatory_only? should be in if condition");
9300 }
9301 else if (!LIST_INSN_SIZE_ZERO(ret)) {
9302 rb_bug("mandatory_only? should be put on top");
9303 }
9304
9305 ADD_INSN1(ret, line_node, putobject, Qfalse);
9306 return compile_builtin_mandatory_only_method(iseq, node, line_node);
9307 }
9308 else if (1) {
9309 rb_bug("can't find builtin function:%s", builtin_func);
9310 }
9311 else {
9312 COMPILE_ERROR(ERROR_ARGS "can't find builtin function:%s", builtin_func);
9313 return COMPILE_NG;
9314 }
9315
9316 int inline_index = nd_line(node);
9317 snprintf(inline_func, sizeof(inline_func), BUILTIN_INLINE_PREFIX "%d", inline_index);
9318 builtin_func = inline_func;
9319 args_node = NULL;
9320 goto retry;
9321 }
9322
9323 if (cconst) {
9324 typedef VALUE(*builtin_func0)(void *, VALUE);
9325 VALUE const_val = (*(builtin_func0)(uintptr_t)bf->func_ptr)(NULL, Qnil);
9326 ADD_INSN1(ret, line_node, putobject, const_val);
9327 return COMPILE_OK;
9328 }
9329
9330 // fprintf(stderr, "func_name:%s -> %p\n", builtin_func, bf->func_ptr);
9331
9332 unsigned int flag = 0;
9333 struct rb_callinfo_kwarg *keywords = NULL;
9334 VALUE argc = setup_args(iseq, args, args_node, &flag, &keywords);
9335
9336 if (FIX2INT(argc) != bf->argc) {
9337 COMPILE_ERROR(ERROR_ARGS "argc is not match for builtin function:%s (expect %d but %d)",
9338 builtin_func, bf->argc, FIX2INT(argc));
9339 return COMPILE_NG;
9340 }
9341
9342 unsigned int start_index;
9343 if (delegate_call_p(iseq, FIX2INT(argc), args, &start_index)) {
9344 ADD_INSN2(ret, line_node, opt_invokebuiltin_delegate, bf, INT2FIX(start_index));
9345 }
9346 else {
9347 ADD_SEQ(ret, args);
9348 ADD_INSN1(ret, line_node, invokebuiltin, bf);
9349 }
9350
9351 if (popped) ADD_INSN(ret, line_node, pop);
9352 return COMPILE_OK;
9353 }
9354}
9355
9356static int
9357compile_call(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, const enum node_type type, const NODE *const line_node, int popped, bool assume_receiver)
9358{
9359 /* call: obj.method(...)
9360 * fcall: func(...)
9361 * vcall: func
9362 */
9363 DECL_ANCHOR(recv);
9364 DECL_ANCHOR(args);
9365 ID mid = get_node_call_nd_mid(node);
9366 VALUE argc;
9367 unsigned int flag = 0;
9368 struct rb_callinfo_kwarg *keywords = NULL;
9369 const rb_iseq_t *parent_block = ISEQ_COMPILE_DATA(iseq)->current_block;
9370 LABEL *else_label = NULL;
9371 VALUE branches = Qfalse;
9372
9373 ISEQ_COMPILE_DATA(iseq)->current_block = NULL;
9374
9375 INIT_ANCHOR(recv);
9376 INIT_ANCHOR(args);
9377#if OPT_SUPPORT_JOKE
9378 if (nd_type_p(node, NODE_VCALL)) {
9379 ID id_bitblt;
9380 ID id_answer;
9381
9382 CONST_ID(id_bitblt, "bitblt");
9383 CONST_ID(id_answer, "the_answer_to_life_the_universe_and_everything");
9384
9385 if (mid == id_bitblt) {
9386 ADD_INSN(ret, line_node, bitblt);
9387 return COMPILE_OK;
9388 }
9389 else if (mid == id_answer) {
9390 ADD_INSN(ret, line_node, answer);
9391 return COMPILE_OK;
9392 }
9393 }
9394 /* only joke */
9395 {
9396 ID goto_id;
9397 ID label_id;
9398
9399 CONST_ID(goto_id, "__goto__");
9400 CONST_ID(label_id, "__label__");
9401
9402 if (nd_type_p(node, NODE_FCALL) &&
9403 (mid == goto_id || mid == label_id)) {
9404 LABEL *label;
9405 st_data_t data;
9406 st_table *labels_table = ISEQ_COMPILE_DATA(iseq)->labels_table;
9407 VALUE label_name;
9408
9409 if (!labels_table) {
9410 labels_table = st_init_numtable();
9411 ISEQ_COMPILE_DATA(iseq)->labels_table = labels_table;
9412 }
9413 {
9414 COMPILE_ERROR(ERROR_ARGS "invalid goto/label format");
9415 return COMPILE_NG;
9416 }
9417
9418 if (mid == goto_id) {
9419 ADD_INSNL(ret, line_node, jump, label);
9420 }
9421 else {
9422 ADD_LABEL(ret, label);
9423 }
9424 return COMPILE_OK;
9425 }
9426 }
9427#endif
9428
9429 const char *builtin_func;
9430 if (UNLIKELY(iseq_has_builtin_function_table(iseq)) &&
9431 (builtin_func = iseq_builtin_function_name(type, get_nd_recv(node), mid)) != NULL) {
9432 return compile_builtin_function_call(iseq, ret, node, line_node, popped, parent_block, args, builtin_func);
9433 }
9434
9435 /* receiver */
9436 if (!assume_receiver) {
9437 if (type == NODE_CALL || type == NODE_OPCALL || type == NODE_QCALL) {
9438 int idx, level;
9439
9440 if (mid == idCall &&
9441 nd_type_p(get_nd_recv(node), NODE_LVAR) &&
9442 iseq_block_param_id_p(iseq, RNODE_LVAR(get_nd_recv(node))->nd_vid, &idx, &level)) {
9443 ADD_INSN2(recv, get_nd_recv(node), getblockparamproxy, INT2FIX(idx + VM_ENV_DATA_SIZE - 1), INT2FIX(level));
9444 }
9445 else if (private_recv_p(node)) {
9446 ADD_INSN(recv, node, putself);
9447 flag |= VM_CALL_FCALL;
9448 }
9449 else {
9450 CHECK(COMPILE(recv, "recv", get_nd_recv(node)));
9451 }
9452
9453 if (type == NODE_QCALL) {
9454 else_label = qcall_branch_start(iseq, recv, &branches, node, line_node);
9455 }
9456 }
9457 else if (type == NODE_FCALL || type == NODE_VCALL) {
9458 ADD_CALL_RECEIVER(recv, line_node);
9459 }
9460 }
9461
9462 /* args */
9463 if (type != NODE_VCALL) {
9464 argc = setup_args(iseq, args, get_nd_args(node), &flag, &keywords);
9465 CHECK(!NIL_P(argc));
9466 }
9467 else {
9468 argc = INT2FIX(0);
9469 }
9470
9471 ADD_SEQ(ret, recv);
9472 ADD_SEQ(ret, args);
9473
9474 debugp_param("call args argc", argc);
9475 debugp_param("call method", ID2SYM(mid));
9476
9477 switch ((int)type) {
9478 case NODE_VCALL:
9479 flag |= VM_CALL_VCALL;
9480 /* VCALL is funcall, so fall through */
9481 case NODE_FCALL:
9482 flag |= VM_CALL_FCALL;
9483 }
9484
9485 if ((flag & VM_CALL_ARGS_BLOCKARG) && (flag & VM_CALL_KW_SPLAT) && !(flag & VM_CALL_KW_SPLAT_MUT)) {
9486 ADD_INSN(ret, line_node, splatkw);
9487 }
9488 ADD_SEND_R(ret, line_node, mid, argc, parent_block, INT2FIX(flag), keywords);
9489
9490 qcall_branch_end(iseq, ret, else_label, branches, node, line_node);
9491 if (popped) {
9492 ADD_INSN(ret, line_node, pop);
9493 }
9494 return COMPILE_OK;
9495}
9496
9497static int
9498compile_op_asgn1(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, int popped)
9499{
9500 const int line = nd_line(node);
9501 VALUE argc;
9502 unsigned int flag = 0;
9503 int asgnflag = 0;
9504 ID id = RNODE_OP_ASGN1(node)->nd_mid;
9505
9506 /*
9507 * a[x] (op)= y
9508 *
9509 * nil # nil
9510 * eval a # nil a
9511 * eval x # nil a x
9512 * dupn 2 # nil a x a x
9513 * send :[] # nil a x a[x]
9514 * eval y # nil a x a[x] y
9515 * send op # nil a x ret
9516 * setn 3 # ret a x ret
9517 * send []= # ret ?
9518 * pop # ret
9519 */
9520
9521 /*
9522 * nd_recv[nd_args->nd_body] (nd_mid)= nd_args->nd_head;
9523 * NODE_OP_ASGN nd_recv
9524 * nd_args->nd_head
9525 * nd_args->nd_body
9526 * nd_mid
9527 */
9528
9529 if (!popped) {
9530 ADD_INSN(ret, node, putnil);
9531 }
9532 asgnflag = COMPILE_RECV(ret, "NODE_OP_ASGN1 recv", node, RNODE_OP_ASGN1(node)->nd_recv);
9533 CHECK(asgnflag != -1);
9534 switch (nd_type(RNODE_OP_ASGN1(node)->nd_index)) {
9535 case NODE_ZLIST:
9536 argc = INT2FIX(0);
9537 break;
9538 default:
9539 argc = setup_args(iseq, ret, RNODE_OP_ASGN1(node)->nd_index, &flag, NULL);
9540 CHECK(!NIL_P(argc));
9541 }
9542 int dup_argn = FIX2INT(argc) + 1;
9543 ADD_INSN1(ret, node, dupn, INT2FIX(dup_argn));
9544 flag |= asgnflag;
9545 ADD_SEND_R(ret, node, idAREF, argc, NULL, INT2FIX(flag & ~VM_CALL_ARGS_SPLAT_MUT), NULL);
9546
9547 if (id == idOROP || id == idANDOP) {
9548 /* a[x] ||= y or a[x] &&= y
9549
9550 unless/if a[x]
9551 a[x]= y
9552 else
9553 nil
9554 end
9555 */
9556 LABEL *label = NEW_LABEL(line);
9557 LABEL *lfin = NEW_LABEL(line);
9558
9559 ADD_INSN(ret, node, dup);
9560 if (id == idOROP) {
9561 ADD_INSNL(ret, node, branchif, label);
9562 }
9563 else { /* idANDOP */
9564 ADD_INSNL(ret, node, branchunless, label);
9565 }
9566 ADD_INSN(ret, node, pop);
9567
9568 CHECK(COMPILE(ret, "NODE_OP_ASGN1 nd_rvalue: ", RNODE_OP_ASGN1(node)->nd_rvalue));
9569 if (!popped) {
9570 ADD_INSN1(ret, node, setn, INT2FIX(dup_argn+1));
9571 }
9572 if (flag & VM_CALL_ARGS_SPLAT) {
9573 if (!(flag & VM_CALL_ARGS_SPLAT_MUT)) {
9574 ADD_INSN(ret, node, swap);
9575 ADD_INSN1(ret, node, splatarray, Qtrue);
9576 ADD_INSN(ret, node, swap);
9577 flag |= VM_CALL_ARGS_SPLAT_MUT;
9578 }
9579 ADD_INSN1(ret, node, pushtoarray, INT2FIX(1));
9580 ADD_SEND_R(ret, node, idASET, argc, NULL, INT2FIX(flag), NULL);
9581 }
9582 else {
9583 ADD_SEND_R(ret, node, idASET, FIXNUM_INC(argc, 1), NULL, INT2FIX(flag), NULL);
9584 }
9585 ADD_INSN(ret, node, pop);
9586 ADD_INSNL(ret, node, jump, lfin);
9587 ADD_LABEL(ret, label);
9588 if (!popped) {
9589 ADD_INSN1(ret, node, setn, INT2FIX(dup_argn+1));
9590 }
9591 ADD_INSN1(ret, node, adjuststack, INT2FIX(dup_argn+1));
9592 ADD_LABEL(ret, lfin);
9593 }
9594 else {
9595 CHECK(COMPILE(ret, "NODE_OP_ASGN1 nd_rvalue: ", RNODE_OP_ASGN1(node)->nd_rvalue));
9596 ADD_SEND(ret, node, id, INT2FIX(1));
9597 if (!popped) {
9598 ADD_INSN1(ret, node, setn, INT2FIX(dup_argn+1));
9599 }
9600 if (flag & VM_CALL_ARGS_SPLAT) {
9601 if (flag & VM_CALL_KW_SPLAT) {
9602 ADD_INSN1(ret, node, topn, INT2FIX(2));
9603 if (!(flag & VM_CALL_ARGS_SPLAT_MUT)) {
9604 ADD_INSN1(ret, node, splatarray, Qtrue);
9605 flag |= VM_CALL_ARGS_SPLAT_MUT;
9606 }
9607 ADD_INSN(ret, node, swap);
9608 ADD_INSN1(ret, node, pushtoarray, INT2FIX(1));
9609 ADD_INSN1(ret, node, setn, INT2FIX(2));
9610 ADD_INSN(ret, node, pop);
9611 }
9612 else {
9613 if (!(flag & VM_CALL_ARGS_SPLAT_MUT)) {
9614 ADD_INSN(ret, node, swap);
9615 ADD_INSN1(ret, node, splatarray, Qtrue);
9616 ADD_INSN(ret, node, swap);
9617 flag |= VM_CALL_ARGS_SPLAT_MUT;
9618 }
9619 ADD_INSN1(ret, node, pushtoarray, INT2FIX(1));
9620 }
9621 ADD_SEND_R(ret, node, idASET, argc, NULL, INT2FIX(flag), NULL);
9622 }
9623 else {
9624 ADD_SEND_R(ret, node, idASET, FIXNUM_INC(argc, 1), NULL, INT2FIX(flag), NULL);
9625 }
9626 ADD_INSN(ret, node, pop);
9627 }
9628 return COMPILE_OK;
9629}
9630
9631static int
9632compile_op_asgn2(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, int popped)
9633{
9634 const int line = nd_line(node);
9635 ID atype = RNODE_OP_ASGN2(node)->nd_mid;
9636 ID vid = RNODE_OP_ASGN2(node)->nd_vid, aid = rb_id_attrset(vid);
9637 int asgnflag;
9638 LABEL *lfin = NEW_LABEL(line);
9639 LABEL *lcfin = NEW_LABEL(line);
9640 LABEL *lskip = 0;
9641 /*
9642 class C; attr_accessor :c; end
9643 r = C.new
9644 r.a &&= v # asgn2
9645
9646 eval r # r
9647 dup # r r
9648 eval r.a # r o
9649
9650 # or
9651 dup # r o o
9652 if lcfin # r o
9653 pop # r
9654 eval v # r v
9655 swap # v r
9656 topn 1 # v r v
9657 send a= # v ?
9658 jump lfin # v ?
9659
9660 lcfin: # r o
9661 swap # o r
9662
9663 lfin: # o ?
9664 pop # o
9665
9666 # or (popped)
9667 if lcfin # r
9668 eval v # r v
9669 send a= # ?
9670 jump lfin # ?
9671
9672 lcfin: # r
9673
9674 lfin: # ?
9675 pop #
9676
9677 # and
9678 dup # r o o
9679 unless lcfin
9680 pop # r
9681 eval v # r v
9682 swap # v r
9683 topn 1 # v r v
9684 send a= # v ?
9685 jump lfin # v ?
9686
9687 # others
9688 eval v # r o v
9689 send ?? # r w
9690 send a= # w
9691
9692 */
9693
9694 asgnflag = COMPILE_RECV(ret, "NODE_OP_ASGN2#recv", node, RNODE_OP_ASGN2(node)->nd_recv);
9695 CHECK(asgnflag != -1);
9696 if (RNODE_OP_ASGN2(node)->nd_aid) {
9697 lskip = NEW_LABEL(line);
9698 ADD_INSN(ret, node, dup);
9699 ADD_INSNL(ret, node, branchnil, lskip);
9700 }
9701 ADD_INSN(ret, node, dup);
9702 ADD_SEND_WITH_FLAG(ret, node, vid, INT2FIX(0), INT2FIX(asgnflag));
9703
9704 if (atype == idOROP || atype == idANDOP) {
9705 if (!popped) {
9706 ADD_INSN(ret, node, dup);
9707 }
9708 if (atype == idOROP) {
9709 ADD_INSNL(ret, node, branchif, lcfin);
9710 }
9711 else { /* idANDOP */
9712 ADD_INSNL(ret, node, branchunless, lcfin);
9713 }
9714 if (!popped) {
9715 ADD_INSN(ret, node, pop);
9716 }
9717 CHECK(COMPILE(ret, "NODE_OP_ASGN2 val", RNODE_OP_ASGN2(node)->nd_value));
9718 if (!popped) {
9719 ADD_INSN(ret, node, swap);
9720 ADD_INSN1(ret, node, topn, INT2FIX(1));
9721 }
9722 ADD_SEND_WITH_FLAG(ret, node, aid, INT2FIX(1), INT2FIX(asgnflag));
9723 ADD_INSNL(ret, node, jump, lfin);
9724
9725 ADD_LABEL(ret, lcfin);
9726 if (!popped) {
9727 ADD_INSN(ret, node, swap);
9728 }
9729
9730 ADD_LABEL(ret, lfin);
9731 }
9732 else {
9733 CHECK(COMPILE(ret, "NODE_OP_ASGN2 val", RNODE_OP_ASGN2(node)->nd_value));
9734 ADD_SEND(ret, node, atype, INT2FIX(1));
9735 if (!popped) {
9736 ADD_INSN(ret, node, swap);
9737 ADD_INSN1(ret, node, topn, INT2FIX(1));
9738 }
9739 ADD_SEND_WITH_FLAG(ret, node, aid, INT2FIX(1), INT2FIX(asgnflag));
9740 }
9741 if (lskip && popped) {
9742 ADD_LABEL(ret, lskip);
9743 }
9744 ADD_INSN(ret, node, pop);
9745 if (lskip && !popped) {
9746 ADD_LABEL(ret, lskip);
9747 }
9748 return COMPILE_OK;
9749}
9750
9751static int compile_shareable_constant_value(rb_iseq_t *iseq, LINK_ANCHOR *ret, enum rb_parser_shareability shareable, const NODE *lhs, const NODE *value);
9752
9753static int
9754compile_op_cdecl(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, int popped)
9755{
9756 const int line = nd_line(node);
9757 LABEL *lfin = 0;
9758 LABEL *lassign = 0;
9759 ID mid;
9760
9761 switch (nd_type(RNODE_OP_CDECL(node)->nd_head)) {
9762 case NODE_COLON3:
9763 ADD_INSN1(ret, node, putobject, rb_cObject);
9764 break;
9765 case NODE_COLON2:
9766 CHECK(COMPILE(ret, "NODE_OP_CDECL/colon2#nd_head", RNODE_COLON2(RNODE_OP_CDECL(node)->nd_head)->nd_head));
9767 break;
9768 default:
9769 COMPILE_ERROR(ERROR_ARGS "%s: invalid node in NODE_OP_CDECL",
9770 ruby_node_name(nd_type(RNODE_OP_CDECL(node)->nd_head)));
9771 return COMPILE_NG;
9772 }
9773 mid = get_node_colon_nd_mid(RNODE_OP_CDECL(node)->nd_head);
9774 /* cref */
9775 if (RNODE_OP_CDECL(node)->nd_aid == idOROP) {
9776 lassign = NEW_LABEL(line);
9777 ADD_INSN(ret, node, dup); /* cref cref */
9778 ADD_INSN3(ret, node, defined, INT2FIX(DEFINED_CONST_FROM),
9779 ID2SYM(mid), Qtrue); /* cref bool */
9780 ADD_INSNL(ret, node, branchunless, lassign); /* cref */
9781 }
9782 ADD_INSN(ret, node, dup); /* cref cref */
9783 ADD_INSN1(ret, node, putobject, Qtrue);
9784 ADD_INSN1(ret, node, getconstant, ID2SYM(mid)); /* cref obj */
9785
9786 if (RNODE_OP_CDECL(node)->nd_aid == idOROP || RNODE_OP_CDECL(node)->nd_aid == idANDOP) {
9787 lfin = NEW_LABEL(line);
9788 if (!popped) ADD_INSN(ret, node, dup); /* cref [obj] obj */
9789 if (RNODE_OP_CDECL(node)->nd_aid == idOROP)
9790 ADD_INSNL(ret, node, branchif, lfin);
9791 else /* idANDOP */
9792 ADD_INSNL(ret, node, branchunless, lfin);
9793 /* cref [obj] */
9794 if (!popped) ADD_INSN(ret, node, pop); /* cref */
9795 if (lassign) ADD_LABEL(ret, lassign);
9796 CHECK(compile_shareable_constant_value(iseq, ret, RNODE_OP_CDECL(node)->shareability, RNODE_OP_CDECL(node)->nd_head, RNODE_OP_CDECL(node)->nd_value));
9797 /* cref value */
9798 if (popped)
9799 ADD_INSN1(ret, node, topn, INT2FIX(1)); /* cref value cref */
9800 else {
9801 ADD_INSN1(ret, node, dupn, INT2FIX(2)); /* cref value cref value */
9802 ADD_INSN(ret, node, swap); /* cref value value cref */
9803 }
9804 ADD_INSN1(ret, node, setconstant, ID2SYM(mid)); /* cref [value] */
9805 ADD_LABEL(ret, lfin); /* cref [value] */
9806 if (!popped) ADD_INSN(ret, node, swap); /* [value] cref */
9807 ADD_INSN(ret, node, pop); /* [value] */
9808 }
9809 else {
9810 CHECK(compile_shareable_constant_value(iseq, ret, RNODE_OP_CDECL(node)->shareability, RNODE_OP_CDECL(node)->nd_head, RNODE_OP_CDECL(node)->nd_value));
9811 /* cref obj value */
9812 ADD_CALL(ret, node, RNODE_OP_CDECL(node)->nd_aid, INT2FIX(1));
9813 /* cref value */
9814 ADD_INSN(ret, node, swap); /* value cref */
9815 if (!popped) {
9816 ADD_INSN1(ret, node, topn, INT2FIX(1)); /* value cref value */
9817 ADD_INSN(ret, node, swap); /* value value cref */
9818 }
9819 ADD_INSN1(ret, node, setconstant, ID2SYM(mid));
9820 }
9821 return COMPILE_OK;
9822}
9823
9824static int
9825compile_op_log(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, int popped, const enum node_type type)
9826{
9827 const int line = nd_line(node);
9828 LABEL *lfin = NEW_LABEL(line);
9829 LABEL *lassign;
9830
9831 if (type == NODE_OP_ASGN_OR && !nd_type_p(RNODE_OP_ASGN_OR(node)->nd_head, NODE_IVAR)) {
9832 LABEL *lfinish[2];
9833 lfinish[0] = lfin;
9834 lfinish[1] = 0;
9835 defined_expr(iseq, ret, RNODE_OP_ASGN_OR(node)->nd_head, lfinish, Qfalse, false);
9836 lassign = lfinish[1];
9837 if (!lassign) {
9838 lassign = NEW_LABEL(line);
9839 }
9840 ADD_INSNL(ret, node, branchunless, lassign);
9841 }
9842 else {
9843 lassign = NEW_LABEL(line);
9844 }
9845
9846 CHECK(COMPILE(ret, "NODE_OP_ASGN_AND/OR#nd_head", RNODE_OP_ASGN_OR(node)->nd_head));
9847
9848 if (!popped) {
9849 ADD_INSN(ret, node, dup);
9850 }
9851
9852 if (type == NODE_OP_ASGN_AND) {
9853 ADD_INSNL(ret, node, branchunless, lfin);
9854 }
9855 else {
9856 ADD_INSNL(ret, node, branchif, lfin);
9857 }
9858
9859 if (!popped) {
9860 ADD_INSN(ret, node, pop);
9861 }
9862
9863 ADD_LABEL(ret, lassign);
9864 CHECK(COMPILE_(ret, "NODE_OP_ASGN_AND/OR#nd_value", RNODE_OP_ASGN_OR(node)->nd_value, popped));
9865 ADD_LABEL(ret, lfin);
9866 return COMPILE_OK;
9867}
9868
9869static int
9870compile_super(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, int popped, const enum node_type type)
9871{
9872 struct rb_iseq_constant_body *const body = ISEQ_BODY(iseq);
9873 DECL_ANCHOR(args);
9874 int argc;
9875 unsigned int flag = 0;
9876 struct rb_callinfo_kwarg *keywords = NULL;
9877 const rb_iseq_t *parent_block = ISEQ_COMPILE_DATA(iseq)->current_block;
9878 int use_block = 1;
9879
9880 INIT_ANCHOR(args);
9881 ISEQ_COMPILE_DATA(iseq)->current_block = NULL;
9882
9883 if (type == NODE_SUPER) {
9884 VALUE vargc = setup_args(iseq, args, RNODE_SUPER(node)->nd_args, &flag, &keywords);
9885 CHECK(!NIL_P(vargc));
9886 argc = FIX2INT(vargc);
9887 if ((flag & VM_CALL_ARGS_BLOCKARG) && (flag & VM_CALL_KW_SPLAT) && !(flag & VM_CALL_KW_SPLAT_MUT)) {
9888 ADD_INSN(args, node, splatkw);
9889 }
9890
9891 if (flag & VM_CALL_ARGS_BLOCKARG) {
9892 use_block = 0;
9893 }
9894 }
9895 else {
9896 /* NODE_ZSUPER */
9897 int i;
9898 const rb_iseq_t *liseq = body->local_iseq;
9899 const struct rb_iseq_constant_body *const local_body = ISEQ_BODY(liseq);
9900 const struct rb_iseq_param_keyword *const local_kwd = local_body->param.keyword;
9901 int lvar_level = get_lvar_level(iseq);
9902
9903 argc = local_body->param.lead_num;
9904
9905 /* normal arguments */
9906 for (i = 0; i < local_body->param.lead_num; i++) {
9907 int idx = local_body->local_table_size - i;
9908 ADD_GETLOCAL(args, node, idx, lvar_level);
9909 }
9910
9911 /* forward ... */
9912 if (local_body->param.flags.forwardable) {
9913 flag |= VM_CALL_FORWARDING;
9914 int idx = local_body->local_table_size - get_local_var_idx(liseq, idDot3);
9915 ADD_GETLOCAL(args, node, idx, lvar_level);
9916 }
9917
9918 if (local_body->param.flags.has_opt) {
9919 /* optional arguments */
9920 int j;
9921 for (j = 0; j < local_body->param.opt_num; j++) {
9922 int idx = local_body->local_table_size - (i + j);
9923 ADD_GETLOCAL(args, node, idx, lvar_level);
9924 }
9925 i += j;
9926 argc = i;
9927 }
9928 if (local_body->param.flags.has_rest) {
9929 /* rest argument */
9930 int idx = local_body->local_table_size - local_body->param.rest_start;
9931 ADD_GETLOCAL(args, node, idx, lvar_level);
9932 ADD_INSN1(args, node, splatarray, RBOOL(local_body->param.flags.has_post));
9933
9934 argc = local_body->param.rest_start + 1;
9935 flag |= VM_CALL_ARGS_SPLAT;
9936 }
9937 if (local_body->param.flags.has_post) {
9938 /* post arguments */
9939 int post_len = local_body->param.post_num;
9940 int post_start = local_body->param.post_start;
9941
9942 if (local_body->param.flags.has_rest) {
9943 int j;
9944 for (j=0; j<post_len; j++) {
9945 int idx = local_body->local_table_size - (post_start + j);
9946 ADD_GETLOCAL(args, node, idx, lvar_level);
9947 }
9948 ADD_INSN1(args, node, pushtoarray, INT2FIX(j));
9949 flag |= VM_CALL_ARGS_SPLAT_MUT;
9950 /* argc is settled at above */
9951 }
9952 else {
9953 int j;
9954 for (j=0; j<post_len; j++) {
9955 int idx = local_body->local_table_size - (post_start + j);
9956 ADD_GETLOCAL(args, node, idx, lvar_level);
9957 }
9958 argc = post_len + post_start;
9959 }
9960 }
9961
9962 if (local_body->param.flags.has_kw) { /* TODO: support keywords */
9963 int local_size = local_body->local_table_size;
9964 argc++;
9965
9966 ADD_INSN1(args, node, putspecialobject, INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
9967
9968 if (local_body->param.flags.has_kwrest) {
9969 int idx = local_body->local_table_size - local_kwd->rest_start;
9970 ADD_GETLOCAL(args, node, idx, lvar_level);
9971 RUBY_ASSERT(local_kwd->num > 0);
9972 ADD_SEND (args, node, rb_intern("dup"), INT2FIX(0));
9973 }
9974 else {
9975 ADD_INSN1(args, node, newhash, INT2FIX(0));
9976 }
9977 for (i = 0; i < local_kwd->num; ++i) {
9978 ID id = local_kwd->table[i];
9979 int idx = local_size - get_local_var_idx(liseq, id);
9980 ADD_INSN1(args, node, putobject, ID2SYM(id));
9981 ADD_GETLOCAL(args, node, idx, lvar_level);
9982 }
9983 ADD_SEND(args, node, id_core_hash_merge_ptr, INT2FIX(i * 2 + 1));
9984 flag |= VM_CALL_KW_SPLAT| VM_CALL_KW_SPLAT_MUT;
9985 }
9986 else if (local_body->param.flags.has_kwrest) {
9987 int idx = local_body->local_table_size - local_kwd->rest_start;
9988 ADD_GETLOCAL(args, node, idx, lvar_level);
9989 argc++;
9990 flag |= VM_CALL_KW_SPLAT;
9991 }
9992 }
9993
9994 if (use_block && parent_block == NULL) {
9995 iseq_set_use_block(ISEQ_BODY(iseq)->local_iseq);
9996 }
9997
9998 flag |= VM_CALL_SUPER | VM_CALL_FCALL;
9999 if (type == NODE_ZSUPER) flag |= VM_CALL_ZSUPER;
10000 ADD_INSN(ret, node, putself);
10001 ADD_SEQ(ret, args);
10002
10003 const struct rb_callinfo * ci = new_callinfo(iseq, 0, argc, flag, keywords, parent_block != NULL);
10004
10005 if (vm_ci_flag(ci) & VM_CALL_FORWARDING) {
10006 ADD_INSN2(ret, node, invokesuperforward, ci, parent_block);
10007 }
10008 else {
10009 ADD_INSN2(ret, node, invokesuper, ci, parent_block);
10010 }
10011
10012 if (popped) {
10013 ADD_INSN(ret, node, pop);
10014 }
10015 return COMPILE_OK;
10016}
10017
10018static int
10019compile_yield(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, int popped)
10020{
10021 DECL_ANCHOR(args);
10022 VALUE argc;
10023 unsigned int flag = 0;
10024 struct rb_callinfo_kwarg *keywords = NULL;
10025
10026 INIT_ANCHOR(args);
10027
10028 switch (ISEQ_BODY(ISEQ_BODY(iseq)->local_iseq)->type) {
10029 case ISEQ_TYPE_TOP:
10030 case ISEQ_TYPE_MAIN:
10031 case ISEQ_TYPE_CLASS:
10032 COMPILE_ERROR(ERROR_ARGS "Invalid yield");
10033 return COMPILE_NG;
10034 default: /* valid */;
10035 }
10036
10037 if (RNODE_YIELD(node)->nd_head) {
10038 argc = setup_args(iseq, args, RNODE_YIELD(node)->nd_head, &flag, &keywords);
10039 CHECK(!NIL_P(argc));
10040 }
10041 else {
10042 argc = INT2FIX(0);
10043 }
10044
10045 ADD_SEQ(ret, args);
10046 ADD_INSN1(ret, node, invokeblock, new_callinfo(iseq, 0, FIX2INT(argc), flag, keywords, FALSE));
10047 iseq_set_use_block(ISEQ_BODY(iseq)->local_iseq);
10048
10049 if (popped) {
10050 ADD_INSN(ret, node, pop);
10051 }
10052
10053 int level = 0;
10054 const rb_iseq_t *tmp_iseq = iseq;
10055 for (; tmp_iseq != ISEQ_BODY(iseq)->local_iseq; level++ ) {
10056 tmp_iseq = ISEQ_BODY(tmp_iseq)->parent_iseq;
10057 }
10058 if (level > 0) access_outer_variables(iseq, level, rb_intern("yield"), true);
10059
10060 return COMPILE_OK;
10061}
10062
10063static int
10064compile_match(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, int popped, const enum node_type type)
10065{
10066 DECL_ANCHOR(recv);
10067 DECL_ANCHOR(val);
10068
10069 INIT_ANCHOR(recv);
10070 INIT_ANCHOR(val);
10071 switch ((int)type) {
10072 case NODE_MATCH:
10073 ADD_INSN1(recv, node, putobject, rb_node_regx_string_val(node));
10074 ADD_INSN2(val, node, getspecial, INT2FIX(0),
10075 INT2FIX(0));
10076 break;
10077 case NODE_MATCH2:
10078 CHECK(COMPILE(recv, "receiver", RNODE_MATCH2(node)->nd_recv));
10079 CHECK(COMPILE(val, "value", RNODE_MATCH2(node)->nd_value));
10080 break;
10081 case NODE_MATCH3:
10082 CHECK(COMPILE(recv, "receiver", RNODE_MATCH3(node)->nd_value));
10083 CHECK(COMPILE(val, "value", RNODE_MATCH3(node)->nd_recv));
10084 break;
10085 }
10086
10087 ADD_SEQ(ret, recv);
10088 ADD_SEQ(ret, val);
10089 ADD_SEND(ret, node, idEqTilde, INT2FIX(1));
10090
10091 if (nd_type_p(node, NODE_MATCH2) && RNODE_MATCH2(node)->nd_args) {
10092 compile_named_capture_assign(iseq, ret, RNODE_MATCH2(node)->nd_args);
10093 }
10094
10095 if (popped) {
10096 ADD_INSN(ret, node, pop);
10097 }
10098 return COMPILE_OK;
10099}
10100
10101static int
10102compile_colon2(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, int popped)
10103{
10104 if (rb_is_const_id(RNODE_COLON2(node)->nd_mid)) {
10105 /* constant */
10106 VALUE segments;
10107 if (ISEQ_COMPILE_DATA(iseq)->option->inline_const_cache &&
10108 (segments = collect_const_segments(iseq, node))) {
10109 ISEQ_BODY(iseq)->ic_size++;
10110 ADD_INSN1(ret, node, opt_getconstant_path, segments);
10111 RB_OBJ_WRITTEN(iseq, Qundef, segments);
10112 }
10113 else {
10114 /* constant */
10115 DECL_ANCHOR(pref);
10116 DECL_ANCHOR(body);
10117
10118 INIT_ANCHOR(pref);
10119 INIT_ANCHOR(body);
10120 CHECK(compile_const_prefix(iseq, node, pref, body));
10121 if (LIST_INSN_SIZE_ZERO(pref)) {
10122 ADD_INSN(ret, node, putnil);
10123 ADD_SEQ(ret, body);
10124 }
10125 else {
10126 ADD_SEQ(ret, pref);
10127 ADD_SEQ(ret, body);
10128 }
10129 }
10130 }
10131 else {
10132 /* function call */
10133 ADD_CALL_RECEIVER(ret, node);
10134 CHECK(COMPILE(ret, "colon2#nd_head", RNODE_COLON2(node)->nd_head));
10135 ADD_CALL(ret, node, RNODE_COLON2(node)->nd_mid, INT2FIX(1));
10136 }
10137 if (popped) {
10138 ADD_INSN(ret, node, pop);
10139 }
10140 return COMPILE_OK;
10141}
10142
10143static int
10144compile_colon3(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, int popped)
10145{
10146 debugi("colon3#nd_mid", RNODE_COLON3(node)->nd_mid);
10147
10148 /* add cache insn */
10149 if (ISEQ_COMPILE_DATA(iseq)->option->inline_const_cache) {
10150 ISEQ_BODY(iseq)->ic_size++;
10151 VALUE segments = rb_ary_new_from_args(2, ID2SYM(idNULL), ID2SYM(RNODE_COLON3(node)->nd_mid));
10152 ADD_INSN1(ret, node, opt_getconstant_path, segments);
10153 RB_OBJ_WRITTEN(iseq, Qundef, segments);
10154 }
10155 else {
10156 ADD_INSN1(ret, node, putobject, rb_cObject);
10157 ADD_INSN1(ret, node, putobject, Qtrue);
10158 ADD_INSN1(ret, node, getconstant, ID2SYM(RNODE_COLON3(node)->nd_mid));
10159 }
10160
10161 if (popped) {
10162 ADD_INSN(ret, node, pop);
10163 }
10164 return COMPILE_OK;
10165}
10166
10167static int
10168compile_dots(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, int popped, const int excl)
10169{
10170 VALUE flag = INT2FIX(excl);
10171 const NODE *b = RNODE_DOT2(node)->nd_beg;
10172 const NODE *e = RNODE_DOT2(node)->nd_end;
10173
10174 if (optimizable_range_item_p(b) && optimizable_range_item_p(e)) {
10175 if (!popped) {
10176 VALUE bv = optimized_range_item(b);
10177 VALUE ev = optimized_range_item(e);
10178 VALUE val = rb_range_new(bv, ev, excl);
10179 ADD_INSN1(ret, node, putobject, val);
10180 RB_OBJ_WRITTEN(iseq, Qundef, val);
10181 }
10182 }
10183 else {
10184 CHECK(COMPILE_(ret, "min", b, popped));
10185 CHECK(COMPILE_(ret, "max", e, popped));
10186 if (!popped) {
10187 ADD_INSN1(ret, node, newrange, flag);
10188 }
10189 }
10190 return COMPILE_OK;
10191}
10192
10193static int
10194compile_errinfo(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, int popped)
10195{
10196 if (!popped) {
10197 if (ISEQ_BODY(iseq)->type == ISEQ_TYPE_RESCUE) {
10198 ADD_GETLOCAL(ret, node, LVAR_ERRINFO, 0);
10199 }
10200 else {
10201 const rb_iseq_t *ip = iseq;
10202 int level = 0;
10203 while (ip) {
10204 if (ISEQ_BODY(ip)->type == ISEQ_TYPE_RESCUE) {
10205 break;
10206 }
10207 ip = ISEQ_BODY(ip)->parent_iseq;
10208 level++;
10209 }
10210 if (ip) {
10211 ADD_GETLOCAL(ret, node, LVAR_ERRINFO, level);
10212 }
10213 else {
10214 ADD_INSN(ret, node, putnil);
10215 }
10216 }
10217 }
10218 return COMPILE_OK;
10219}
10220
10221static int
10222compile_kw_arg(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, int popped)
10223{
10224 struct rb_iseq_constant_body *const body = ISEQ_BODY(iseq);
10225 LABEL *end_label = NEW_LABEL(nd_line(node));
10226 const NODE *default_value = get_nd_value(RNODE_KW_ARG(node)->nd_body);
10227
10228 if (default_value == NODE_SPECIAL_REQUIRED_KEYWORD) {
10229 /* required argument. do nothing */
10230 COMPILE_ERROR(ERROR_ARGS "unreachable");
10231 return COMPILE_NG;
10232 }
10233 else if (nd_type_p(default_value, NODE_SYM) ||
10234 nd_type_p(default_value, NODE_REGX) ||
10235 nd_type_p(default_value, NODE_LINE) ||
10236 nd_type_p(default_value, NODE_INTEGER) ||
10237 nd_type_p(default_value, NODE_FLOAT) ||
10238 nd_type_p(default_value, NODE_RATIONAL) ||
10239 nd_type_p(default_value, NODE_IMAGINARY) ||
10240 nd_type_p(default_value, NODE_NIL) ||
10241 nd_type_p(default_value, NODE_TRUE) ||
10242 nd_type_p(default_value, NODE_FALSE)) {
10243 COMPILE_ERROR(ERROR_ARGS "unreachable");
10244 return COMPILE_NG;
10245 }
10246 else {
10247 /* if keywordcheck(_kw_bits, nth_keyword)
10248 * kw = default_value
10249 * end
10250 */
10251 int kw_bits_idx = body->local_table_size - body->param.keyword->bits_start;
10252 int keyword_idx = body->param.keyword->num;
10253
10254 ADD_INSN2(ret, node, checkkeyword, INT2FIX(kw_bits_idx + VM_ENV_DATA_SIZE - 1), INT2FIX(keyword_idx));
10255 ADD_INSNL(ret, node, branchif, end_label);
10256 CHECK(COMPILE_POPPED(ret, "keyword default argument", RNODE_KW_ARG(node)->nd_body));
10257 ADD_LABEL(ret, end_label);
10258 }
10259 return COMPILE_OK;
10260}
10261
10262static int
10263compile_attrasgn(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, int popped)
10264{
10265 DECL_ANCHOR(recv);
10266 DECL_ANCHOR(args);
10267 unsigned int flag = 0;
10268 ID mid = RNODE_ATTRASGN(node)->nd_mid;
10269 VALUE argc;
10270 LABEL *else_label = NULL;
10271 VALUE branches = Qfalse;
10272
10273 /* optimization shortcut
10274 * obj["literal"] = value -> opt_aset_with(obj, "literal", value)
10275 */
10276 if (!ISEQ_COMPILE_DATA(iseq)->in_masgn &&
10277 mid == idASET && !private_recv_p(node) && RNODE_ATTRASGN(node)->nd_args &&
10278 nd_type_p(RNODE_ATTRASGN(node)->nd_args, NODE_LIST) && RNODE_LIST(RNODE_ATTRASGN(node)->nd_args)->as.nd_alen == 2 &&
10279 (nd_type_p(RNODE_LIST(RNODE_ATTRASGN(node)->nd_args)->nd_head, NODE_STR) || nd_type_p(RNODE_LIST(RNODE_ATTRASGN(node)->nd_args)->nd_head, NODE_FILE)) &&
10280 ISEQ_COMPILE_DATA(iseq)->current_block == NULL &&
10281 !frozen_string_literal_p(iseq) &&
10282 ISEQ_COMPILE_DATA(iseq)->option->specialized_instruction)
10283 {
10284 VALUE str = get_string_value(RNODE_LIST(RNODE_ATTRASGN(node)->nd_args)->nd_head);
10285 CHECK(COMPILE(ret, "recv", RNODE_ATTRASGN(node)->nd_recv));
10286 CHECK(COMPILE(ret, "value", RNODE_LIST(RNODE_LIST(RNODE_ATTRASGN(node)->nd_args)->nd_next)->nd_head));
10287 if (!popped) {
10288 ADD_INSN(ret, node, swap);
10289 ADD_INSN1(ret, node, topn, INT2FIX(1));
10290 }
10291 ADD_INSN2(ret, node, opt_aset_with, str,
10292 new_callinfo(iseq, idASET, 2, 0, NULL, FALSE));
10293 RB_OBJ_WRITTEN(iseq, Qundef, str);
10294 ADD_INSN(ret, node, pop);
10295 return COMPILE_OK;
10296 }
10297
10298 INIT_ANCHOR(recv);
10299 INIT_ANCHOR(args);
10300 argc = setup_args(iseq, args, RNODE_ATTRASGN(node)->nd_args, &flag, NULL);
10301 CHECK(!NIL_P(argc));
10302
10303 int asgnflag = COMPILE_RECV(recv, "recv", node, RNODE_ATTRASGN(node)->nd_recv);
10304 CHECK(asgnflag != -1);
10305 flag |= (unsigned int)asgnflag;
10306
10307 debugp_param("argc", argc);
10308 debugp_param("nd_mid", ID2SYM(mid));
10309
10310 if (!rb_is_attrset_id(mid)) {
10311 /* safe nav attr */
10312 mid = rb_id_attrset(mid);
10313 else_label = qcall_branch_start(iseq, recv, &branches, node, node);
10314 }
10315 if (!popped) {
10316 ADD_INSN(ret, node, putnil);
10317 ADD_SEQ(ret, recv);
10318 ADD_SEQ(ret, args);
10319
10320 if (flag & VM_CALL_ARGS_SPLAT) {
10321 ADD_INSN(ret, node, dup);
10322 ADD_INSN1(ret, node, putobject, INT2FIX(-1));
10323 ADD_SEND_WITH_FLAG(ret, node, idAREF, INT2FIX(1), INT2FIX(asgnflag));
10324 ADD_INSN1(ret, node, setn, FIXNUM_INC(argc, 2));
10325 ADD_INSN (ret, node, pop);
10326 }
10327 else {
10328 ADD_INSN1(ret, node, setn, FIXNUM_INC(argc, 1));
10329 }
10330 }
10331 else {
10332 ADD_SEQ(ret, recv);
10333 ADD_SEQ(ret, args);
10334 }
10335 ADD_SEND_WITH_FLAG(ret, node, mid, argc, INT2FIX(flag));
10336 qcall_branch_end(iseq, ret, else_label, branches, node, node);
10337 ADD_INSN(ret, node, pop);
10338 return COMPILE_OK;
10339}
10340
10341static int
10342compile_make_shareable_node(rb_iseq_t *iseq, LINK_ANCHOR *ret, LINK_ANCHOR *sub, const NODE *value, bool copy)
10343{
10344 ADD_INSN1(ret, value, putspecialobject, INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
10345 ADD_SEQ(ret, sub);
10346
10347 if (copy) {
10348 /*
10349 * NEW_CALL(fcore, rb_intern("make_shareable_copy"),
10350 * NEW_LIST(value, loc), loc);
10351 */
10352 ADD_SEND_WITH_FLAG(ret, value, rb_intern("make_shareable_copy"), INT2FIX(1), INT2FIX(VM_CALL_ARGS_SIMPLE));
10353 }
10354 else {
10355 /*
10356 * NEW_CALL(fcore, rb_intern("make_shareable"),
10357 * NEW_LIST(value, loc), loc);
10358 */
10359 ADD_SEND_WITH_FLAG(ret, value, rb_intern("make_shareable"), INT2FIX(1), INT2FIX(VM_CALL_ARGS_SIMPLE));
10360 }
10361
10362 return COMPILE_OK;
10363}
10364
10365static VALUE
10366node_const_decl_val(const NODE *node)
10367{
10368 VALUE path;
10369 switch (nd_type(node)) {
10370 case NODE_CDECL:
10371 if (RNODE_CDECL(node)->nd_vid) {
10372 path = rb_id2str(RNODE_CDECL(node)->nd_vid);
10373 goto end;
10374 }
10375 else {
10376 node = RNODE_CDECL(node)->nd_else;
10377 }
10378 break;
10379 case NODE_COLON2:
10380 break;
10381 case NODE_COLON3:
10382 // ::Const
10383 path = rb_str_new_cstr("::");
10384 rb_str_append(path, rb_id2str(RNODE_COLON3(node)->nd_mid));
10385 goto end;
10386 default:
10387 rb_bug("unexpected node: %s", ruby_node_name(nd_type(node)));
10389 }
10390
10391 path = rb_ary_new();
10392 if (node) {
10393 for (; node && nd_type_p(node, NODE_COLON2); node = RNODE_COLON2(node)->nd_head) {
10394 rb_ary_push(path, rb_id2str(RNODE_COLON2(node)->nd_mid));
10395 }
10396 if (node && nd_type_p(node, NODE_CONST)) {
10397 // Const::Name
10398 rb_ary_push(path, rb_id2str(RNODE_CONST(node)->nd_vid));
10399 }
10400 else if (node && nd_type_p(node, NODE_COLON3)) {
10401 // ::Const::Name
10402 rb_ary_push(path, rb_id2str(RNODE_COLON3(node)->nd_mid));
10403 rb_ary_push(path, rb_str_new(0, 0));
10404 }
10405 else {
10406 // expression::Name
10407 rb_ary_push(path, rb_str_new_cstr("..."));
10408 }
10409 path = rb_ary_join(rb_ary_reverse(path), rb_str_new_cstr("::"));
10410 }
10411 end:
10412 path = rb_fstring(path);
10413 return path;
10414}
10415
10416static VALUE
10417const_decl_path(NODE *dest)
10418{
10419 VALUE path = Qnil;
10420 if (!nd_type_p(dest, NODE_CALL)) {
10421 path = node_const_decl_val(dest);
10422 }
10423 return path;
10424}
10425
10426static int
10427compile_ensure_shareable_node(rb_iseq_t *iseq, LINK_ANCHOR *ret, NODE *dest, const NODE *value)
10428{
10429 /*
10430 *. RubyVM::FrozenCore.ensure_shareable(value, const_decl_path(dest))
10431 */
10432 VALUE path = const_decl_path(dest);
10433 ADD_INSN1(ret, value, putspecialobject, INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
10434 CHECK(COMPILE(ret, "compile_ensure_shareable_node", value));
10435 ADD_INSN1(ret, value, putobject, path);
10436 RB_OBJ_WRITTEN(iseq, Qundef, path);
10437 ADD_SEND_WITH_FLAG(ret, value, rb_intern("ensure_shareable"), INT2FIX(2), INT2FIX(VM_CALL_ARGS_SIMPLE));
10438
10439 return COMPILE_OK;
10440}
10441
10442#ifndef SHAREABLE_BARE_EXPRESSION
10443#define SHAREABLE_BARE_EXPRESSION 1
10444#endif
10445
10446static int
10447compile_shareable_literal_constant(rb_iseq_t *iseq, LINK_ANCHOR *ret, enum rb_parser_shareability shareable, NODE *dest, const NODE *node, size_t level, VALUE *value_p, int *shareable_literal_p)
10448{
10449# define compile_shareable_literal_constant_next(node, anchor, value_p, shareable_literal_p) \
10450 compile_shareable_literal_constant(iseq, anchor, shareable, dest, node, level+1, value_p, shareable_literal_p)
10451 VALUE lit = Qnil;
10452 DECL_ANCHOR(anchor);
10453
10454 enum node_type type = node ? nd_type(node) : NODE_NIL;
10455 switch (type) {
10456 case NODE_TRUE:
10457 *value_p = Qtrue;
10458 goto compile;
10459 case NODE_FALSE:
10460 *value_p = Qfalse;
10461 goto compile;
10462 case NODE_NIL:
10463 *value_p = Qnil;
10464 goto compile;
10465 case NODE_SYM:
10466 *value_p = rb_node_sym_string_val(node);
10467 goto compile;
10468 case NODE_REGX:
10469 *value_p = rb_node_regx_string_val(node);
10470 goto compile;
10471 case NODE_LINE:
10472 *value_p = rb_node_line_lineno_val(node);
10473 goto compile;
10474 case NODE_INTEGER:
10475 *value_p = rb_node_integer_literal_val(node);
10476 goto compile;
10477 case NODE_FLOAT:
10478 *value_p = rb_node_float_literal_val(node);
10479 goto compile;
10480 case NODE_RATIONAL:
10481 *value_p = rb_node_rational_literal_val(node);
10482 goto compile;
10483 case NODE_IMAGINARY:
10484 *value_p = rb_node_imaginary_literal_val(node);
10485 goto compile;
10486 case NODE_ENCODING:
10487 *value_p = rb_node_encoding_val(node);
10488
10489 compile:
10490 CHECK(COMPILE(ret, "shareable_literal_constant", node));
10491 *shareable_literal_p = 1;
10492 return COMPILE_OK;
10493
10494 case NODE_DSTR:
10495 CHECK(COMPILE(ret, "shareable_literal_constant", node));
10496 if (shareable == rb_parser_shareable_literal) {
10497 /*
10498 * NEW_CALL(node, idUMinus, 0, loc);
10499 *
10500 * -"#{var}"
10501 */
10502 ADD_SEND_WITH_FLAG(ret, node, idUMinus, INT2FIX(0), INT2FIX(VM_CALL_ARGS_SIMPLE));
10503 }
10504 *value_p = Qundef;
10505 *shareable_literal_p = 1;
10506 return COMPILE_OK;
10507
10508 case NODE_STR:{
10509 VALUE lit = rb_node_str_string_val(node);
10510 ADD_INSN1(ret, node, putobject, lit);
10511 RB_OBJ_WRITTEN(iseq, Qundef, lit);
10512 *value_p = lit;
10513 *shareable_literal_p = 1;
10514
10515 return COMPILE_OK;
10516 }
10517
10518 case NODE_FILE:{
10519 VALUE lit = rb_node_file_path_val(node);
10520 ADD_INSN1(ret, node, putobject, lit);
10521 RB_OBJ_WRITTEN(iseq, Qundef, lit);
10522 *value_p = lit;
10523 *shareable_literal_p = 1;
10524
10525 return COMPILE_OK;
10526 }
10527
10528 case NODE_ZLIST:{
10529 VALUE lit = rb_ary_new();
10530 OBJ_FREEZE(lit);
10531 ADD_INSN1(ret, node, putobject, lit);
10532 RB_OBJ_WRITTEN(iseq, Qundef, lit);
10533 *value_p = lit;
10534 *shareable_literal_p = 1;
10535
10536 return COMPILE_OK;
10537 }
10538
10539 case NODE_LIST:{
10540 INIT_ANCHOR(anchor);
10541 lit = rb_ary_new();
10542 for (NODE *n = (NODE *)node; n; n = RNODE_LIST(n)->nd_next) {
10543 VALUE val;
10544 int shareable_literal_p2;
10545 NODE *elt = RNODE_LIST(n)->nd_head;
10546 if (elt) {
10547 CHECK(compile_shareable_literal_constant_next(elt, anchor, &val, &shareable_literal_p2));
10548 if (shareable_literal_p2) {
10549 /* noop */
10550 }
10551 else if (RTEST(lit)) {
10552 rb_ary_clear(lit);
10553 lit = Qfalse;
10554 }
10555 }
10556 if (RTEST(lit)) {
10557 if (!UNDEF_P(val)) {
10558 rb_ary_push(lit, val);
10559 }
10560 else {
10561 rb_ary_clear(lit);
10562 lit = Qnil; /* make shareable at runtime */
10563 }
10564 }
10565 }
10566 break;
10567 }
10568 case NODE_HASH:{
10569 if (!RNODE_HASH(node)->nd_brace) {
10570 *value_p = Qundef;
10571 *shareable_literal_p = 0;
10572 return COMPILE_OK;
10573 }
10574 for (NODE *n = RNODE_HASH(node)->nd_head; n; n = RNODE_LIST(RNODE_LIST(n)->nd_next)->nd_next) {
10575 if (!RNODE_LIST(n)->nd_head) {
10576 // If the hash node have a keyword splat, fall back to the default case.
10577 goto compile_shareable;
10578 }
10579 }
10580
10581 INIT_ANCHOR(anchor);
10582 lit = rb_hash_new();
10583 for (NODE *n = RNODE_HASH(node)->nd_head; n; n = RNODE_LIST(RNODE_LIST(n)->nd_next)->nd_next) {
10584 VALUE key_val = 0;
10585 VALUE value_val = 0;
10586 int shareable_literal_p2;
10587 NODE *key = RNODE_LIST(n)->nd_head;
10588 NODE *val = RNODE_LIST(RNODE_LIST(n)->nd_next)->nd_head;
10589 CHECK(compile_shareable_literal_constant_next(key, anchor, &key_val, &shareable_literal_p2));
10590 if (shareable_literal_p2) {
10591 /* noop */
10592 }
10593 else if (RTEST(lit)) {
10594 rb_hash_clear(lit);
10595 lit = Qfalse;
10596 }
10597 CHECK(compile_shareable_literal_constant_next(val, anchor, &value_val, &shareable_literal_p2));
10598 if (shareable_literal_p2) {
10599 /* noop */
10600 }
10601 else if (RTEST(lit)) {
10602 rb_hash_clear(lit);
10603 lit = Qfalse;
10604 }
10605 if (RTEST(lit)) {
10606 if (!UNDEF_P(key_val) && !UNDEF_P(value_val)) {
10607 rb_hash_aset(lit, key_val, value_val);
10608 }
10609 else {
10610 rb_hash_clear(lit);
10611 lit = Qnil; /* make shareable at runtime */
10612 }
10613 }
10614 }
10615 break;
10616 }
10617
10618 default:
10619
10620 compile_shareable:
10621 if (shareable == rb_parser_shareable_literal &&
10622 (SHAREABLE_BARE_EXPRESSION || level > 0)) {
10623 CHECK(compile_ensure_shareable_node(iseq, ret, dest, node));
10624 *value_p = Qundef;
10625 *shareable_literal_p = 1;
10626 return COMPILE_OK;
10627 }
10628 CHECK(COMPILE(ret, "shareable_literal_constant", node));
10629 *value_p = Qundef;
10630 *shareable_literal_p = 0;
10631 return COMPILE_OK;
10632 }
10633
10634 /* Array or Hash that does not have keyword splat */
10635 if (!lit) {
10636 if (nd_type(node) == NODE_LIST) {
10637 ADD_INSN1(anchor, node, newarray, INT2FIX(RNODE_LIST(node)->as.nd_alen));
10638 }
10639 else if (nd_type(node) == NODE_HASH) {
10640 int len = (int)RNODE_LIST(RNODE_HASH(node)->nd_head)->as.nd_alen;
10641 ADD_INSN1(anchor, node, newhash, INT2FIX(len));
10642 }
10643 *value_p = Qundef;
10644 *shareable_literal_p = 0;
10645 ADD_SEQ(ret, anchor);
10646 return COMPILE_OK;
10647 }
10648 if (NIL_P(lit)) {
10649 // if shareable_literal, all elements should have been ensured
10650 // as shareable
10651 if (nd_type(node) == NODE_LIST) {
10652 ADD_INSN1(anchor, node, newarray, INT2FIX(RNODE_LIST(node)->as.nd_alen));
10653 }
10654 else if (nd_type(node) == NODE_HASH) {
10655 int len = (int)RNODE_LIST(RNODE_HASH(node)->nd_head)->as.nd_alen;
10656 ADD_INSN1(anchor, node, newhash, INT2FIX(len));
10657 }
10658 CHECK(compile_make_shareable_node(iseq, ret, anchor, node, false));
10659 *value_p = Qundef;
10660 *shareable_literal_p = 1;
10661 }
10662 else {
10664 ADD_INSN1(ret, node, putobject, val);
10665 RB_OBJ_WRITTEN(iseq, Qundef, val);
10666 *value_p = val;
10667 *shareable_literal_p = 1;
10668 }
10669
10670 return COMPILE_OK;
10671}
10672
10673static int
10674compile_shareable_constant_value(rb_iseq_t *iseq, LINK_ANCHOR *ret, enum rb_parser_shareability shareable, const NODE *lhs, const NODE *value)
10675{
10676 int literal_p = 0;
10677 VALUE val;
10678 DECL_ANCHOR(anchor);
10679 INIT_ANCHOR(anchor);
10680
10681 switch (shareable) {
10682 case rb_parser_shareable_none:
10683 CHECK(COMPILE(ret, "compile_shareable_constant_value", value));
10684 return COMPILE_OK;
10685
10686 case rb_parser_shareable_literal:
10687 CHECK(compile_shareable_literal_constant(iseq, anchor, shareable, (NODE *)lhs, value, 0, &val, &literal_p));
10688 ADD_SEQ(ret, anchor);
10689 return COMPILE_OK;
10690
10691 case rb_parser_shareable_copy:
10692 case rb_parser_shareable_everything:
10693 CHECK(compile_shareable_literal_constant(iseq, anchor, shareable, (NODE *)lhs, value, 0, &val, &literal_p));
10694 if (!literal_p) {
10695 CHECK(compile_make_shareable_node(iseq, ret, anchor, value, shareable == rb_parser_shareable_copy));
10696 }
10697 else {
10698 ADD_SEQ(ret, anchor);
10699 }
10700 return COMPILE_OK;
10701 default:
10702 rb_bug("unexpected rb_parser_shareability: %d", shareable);
10703 }
10704}
10705
10706static int iseq_compile_each0(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, int popped);
10714static int
10715iseq_compile_each(rb_iseq_t *iseq, LINK_ANCHOR *ret, const NODE *node, int popped)
10716{
10717 if (node == 0) {
10718 if (!popped) {
10719 int lineno = ISEQ_COMPILE_DATA(iseq)->last_line;
10720 if (lineno == 0) lineno = FIX2INT(rb_iseq_first_lineno(iseq));
10721 debugs("node: NODE_NIL(implicit)\n");
10722 ADD_SYNTHETIC_INSN(ret, lineno, -1, putnil);
10723 }
10724 return COMPILE_OK;
10725 }
10726 return iseq_compile_each0(iseq, ret, node, popped);
10727}
10728
10729static int
10730iseq_compile_each0(rb_iseq_t *iseq, LINK_ANCHOR *const ret, const NODE *const node, int popped)
10731{
10732 const int line = (int)nd_line(node);
10733 const enum node_type type = nd_type(node);
10734 struct rb_iseq_constant_body *const body = ISEQ_BODY(iseq);
10735
10736 if (ISEQ_COMPILE_DATA(iseq)->last_line == line) {
10737 /* ignore */
10738 }
10739 else {
10740 if (nd_fl_newline(node)) {
10741 int event = RUBY_EVENT_LINE;
10742 ISEQ_COMPILE_DATA(iseq)->last_line = line;
10743 if (ISEQ_COVERAGE(iseq) && ISEQ_LINE_COVERAGE(iseq)) {
10744 event |= RUBY_EVENT_COVERAGE_LINE;
10745 }
10746 ADD_TRACE(ret, event);
10747 }
10748 }
10749
10750 debug_node_start(node);
10751#undef BEFORE_RETURN
10752#define BEFORE_RETURN debug_node_end()
10753
10754 switch (type) {
10755 case NODE_BLOCK:
10756 CHECK(compile_block(iseq, ret, node, popped));
10757 break;
10758 case NODE_IF:
10759 case NODE_UNLESS:
10760 CHECK(compile_if(iseq, ret, node, popped, type));
10761 break;
10762 case NODE_CASE:
10763 CHECK(compile_case(iseq, ret, node, popped));
10764 break;
10765 case NODE_CASE2:
10766 CHECK(compile_case2(iseq, ret, node, popped));
10767 break;
10768 case NODE_CASE3:
10769 CHECK(compile_case3(iseq, ret, node, popped));
10770 break;
10771 case NODE_WHILE:
10772 case NODE_UNTIL:
10773 CHECK(compile_loop(iseq, ret, node, popped, type));
10774 break;
10775 case NODE_FOR:
10776 case NODE_ITER:
10777 CHECK(compile_iter(iseq, ret, node, popped));
10778 break;
10779 case NODE_FOR_MASGN:
10780 CHECK(compile_for_masgn(iseq, ret, node, popped));
10781 break;
10782 case NODE_BREAK:
10783 CHECK(compile_break(iseq, ret, node, popped));
10784 break;
10785 case NODE_NEXT:
10786 CHECK(compile_next(iseq, ret, node, popped));
10787 break;
10788 case NODE_REDO:
10789 CHECK(compile_redo(iseq, ret, node, popped));
10790 break;
10791 case NODE_RETRY:
10792 CHECK(compile_retry(iseq, ret, node, popped));
10793 break;
10794 case NODE_BEGIN:{
10795 CHECK(COMPILE_(ret, "NODE_BEGIN", RNODE_BEGIN(node)->nd_body, popped));
10796 break;
10797 }
10798 case NODE_RESCUE:
10799 CHECK(compile_rescue(iseq, ret, node, popped));
10800 break;
10801 case NODE_RESBODY:
10802 CHECK(compile_resbody(iseq, ret, node, popped));
10803 break;
10804 case NODE_ENSURE:
10805 CHECK(compile_ensure(iseq, ret, node, popped));
10806 break;
10807
10808 case NODE_AND:
10809 case NODE_OR:{
10810 LABEL *end_label = NEW_LABEL(line);
10811 CHECK(COMPILE(ret, "nd_1st", RNODE_OR(node)->nd_1st));
10812 if (!popped) {
10813 ADD_INSN(ret, node, dup);
10814 }
10815 if (type == NODE_AND) {
10816 ADD_INSNL(ret, node, branchunless, end_label);
10817 }
10818 else {
10819 ADD_INSNL(ret, node, branchif, end_label);
10820 }
10821 if (!popped) {
10822 ADD_INSN(ret, node, pop);
10823 }
10824 CHECK(COMPILE_(ret, "nd_2nd", RNODE_OR(node)->nd_2nd, popped));
10825 ADD_LABEL(ret, end_label);
10826 break;
10827 }
10828
10829 case NODE_MASGN:{
10830 bool prev_in_masgn = ISEQ_COMPILE_DATA(iseq)->in_masgn;
10831 ISEQ_COMPILE_DATA(iseq)->in_masgn = true;
10832 compile_massign(iseq, ret, node, popped);
10833 ISEQ_COMPILE_DATA(iseq)->in_masgn = prev_in_masgn;
10834 break;
10835 }
10836
10837 case NODE_LASGN:{
10838 ID id = RNODE_LASGN(node)->nd_vid;
10839 int idx = ISEQ_BODY(body->local_iseq)->local_table_size - get_local_var_idx(iseq, id);
10840
10841 debugs("lvar: %s idx: %d\n", rb_id2name(id), idx);
10842 CHECK(COMPILE(ret, "rvalue", RNODE_LASGN(node)->nd_value));
10843
10844 if (!popped) {
10845 ADD_INSN(ret, node, dup);
10846 }
10847 ADD_SETLOCAL(ret, node, idx, get_lvar_level(iseq));
10848 break;
10849 }
10850 case NODE_DASGN: {
10851 int idx, lv, ls;
10852 ID id = RNODE_DASGN(node)->nd_vid;
10853 CHECK(COMPILE(ret, "dvalue", RNODE_DASGN(node)->nd_value));
10854 debugi("dassn id", rb_id2str(id) ? id : '*');
10855
10856 if (!popped) {
10857 ADD_INSN(ret, node, dup);
10858 }
10859
10860 idx = get_dyna_var_idx(iseq, id, &lv, &ls);
10861
10862 if (idx < 0) {
10863 COMPILE_ERROR(ERROR_ARGS "NODE_DASGN: unknown id (%"PRIsVALUE")",
10864 rb_id2str(id));
10865 goto ng;
10866 }
10867 ADD_SETLOCAL(ret, node, ls - idx, lv);
10868 break;
10869 }
10870 case NODE_GASGN:{
10871 CHECK(COMPILE(ret, "lvalue", RNODE_GASGN(node)->nd_value));
10872
10873 if (!popped) {
10874 ADD_INSN(ret, node, dup);
10875 }
10876 ADD_INSN1(ret, node, setglobal, ID2SYM(RNODE_GASGN(node)->nd_vid));
10877 break;
10878 }
10879 case NODE_IASGN:{
10880 CHECK(COMPILE(ret, "lvalue", RNODE_IASGN(node)->nd_value));
10881 if (!popped) {
10882 ADD_INSN(ret, node, dup);
10883 }
10884 ADD_INSN2(ret, node, setinstancevariable,
10885 ID2SYM(RNODE_IASGN(node)->nd_vid),
10886 get_ivar_ic_value(iseq,RNODE_IASGN(node)->nd_vid));
10887 break;
10888 }
10889 case NODE_CDECL:{
10890 if (RNODE_CDECL(node)->nd_vid) {
10891 CHECK(compile_shareable_constant_value(iseq, ret, RNODE_CDECL(node)->shareability, node, RNODE_CDECL(node)->nd_value));
10892
10893 if (!popped) {
10894 ADD_INSN(ret, node, dup);
10895 }
10896
10897 ADD_INSN1(ret, node, putspecialobject,
10898 INT2FIX(VM_SPECIAL_OBJECT_CONST_BASE));
10899 ADD_INSN1(ret, node, setconstant, ID2SYM(RNODE_CDECL(node)->nd_vid));
10900 }
10901 else {
10902 compile_cpath(ret, iseq, RNODE_CDECL(node)->nd_else);
10903 CHECK(compile_shareable_constant_value(iseq, ret, RNODE_CDECL(node)->shareability, node, RNODE_CDECL(node)->nd_value));
10904 ADD_INSN(ret, node, swap);
10905
10906 if (!popped) {
10907 ADD_INSN1(ret, node, topn, INT2FIX(1));
10908 ADD_INSN(ret, node, swap);
10909 }
10910
10911 ADD_INSN1(ret, node, setconstant, ID2SYM(get_node_colon_nd_mid(RNODE_CDECL(node)->nd_else)));
10912 }
10913 break;
10914 }
10915 case NODE_CVASGN:{
10916 CHECK(COMPILE(ret, "cvasgn val", RNODE_CVASGN(node)->nd_value));
10917 if (!popped) {
10918 ADD_INSN(ret, node, dup);
10919 }
10920 ADD_INSN2(ret, node, setclassvariable,
10921 ID2SYM(RNODE_CVASGN(node)->nd_vid),
10922 get_cvar_ic_value(iseq, RNODE_CVASGN(node)->nd_vid));
10923 break;
10924 }
10925 case NODE_OP_ASGN1:
10926 CHECK(compile_op_asgn1(iseq, ret, node, popped));
10927 break;
10928 case NODE_OP_ASGN2:
10929 CHECK(compile_op_asgn2(iseq, ret, node, popped));
10930 break;
10931 case NODE_OP_CDECL:
10932 CHECK(compile_op_cdecl(iseq, ret, node, popped));
10933 break;
10934 case NODE_OP_ASGN_AND:
10935 case NODE_OP_ASGN_OR:
10936 CHECK(compile_op_log(iseq, ret, node, popped, type));
10937 break;
10938 case NODE_CALL: /* obj.foo */
10939 case NODE_OPCALL: /* foo[] */
10940 if (compile_call_precheck_freeze(iseq, ret, node, node, popped) == TRUE) {
10941 break;
10942 }
10943 case NODE_QCALL: /* obj&.foo */
10944 case NODE_FCALL: /* foo() */
10945 case NODE_VCALL: /* foo (variable or call) */
10946 if (compile_call(iseq, ret, node, type, node, popped, false) == COMPILE_NG) {
10947 goto ng;
10948 }
10949 break;
10950 case NODE_SUPER:
10951 case NODE_ZSUPER:
10952 CHECK(compile_super(iseq, ret, node, popped, type));
10953 break;
10954 case NODE_LIST:{
10955 CHECK(compile_array(iseq, ret, node, popped, TRUE) >= 0);
10956 break;
10957 }
10958 case NODE_ZLIST:{
10959 if (!popped) {
10960 ADD_INSN1(ret, node, newarray, INT2FIX(0));
10961 }
10962 break;
10963 }
10964 case NODE_HASH:
10965 CHECK(compile_hash(iseq, ret, node, FALSE, popped) >= 0);
10966 break;
10967 case NODE_RETURN:
10968 CHECK(compile_return(iseq, ret, node, popped));
10969 break;
10970 case NODE_YIELD:
10971 CHECK(compile_yield(iseq, ret, node, popped));
10972 break;
10973 case NODE_LVAR:{
10974 if (!popped) {
10975 compile_lvar(iseq, ret, node, RNODE_LVAR(node)->nd_vid);
10976 }
10977 break;
10978 }
10979 case NODE_DVAR:{
10980 int lv, idx, ls;
10981 debugi("nd_vid", RNODE_DVAR(node)->nd_vid);
10982 if (!popped) {
10983 idx = get_dyna_var_idx(iseq, RNODE_DVAR(node)->nd_vid, &lv, &ls);
10984 if (idx < 0) {
10985 COMPILE_ERROR(ERROR_ARGS "unknown dvar (%"PRIsVALUE")",
10986 rb_id2str(RNODE_DVAR(node)->nd_vid));
10987 goto ng;
10988 }
10989 ADD_GETLOCAL(ret, node, ls - idx, lv);
10990 }
10991 break;
10992 }
10993 case NODE_GVAR:{
10994 ADD_INSN1(ret, node, getglobal, ID2SYM(RNODE_GVAR(node)->nd_vid));
10995 if (popped) {
10996 ADD_INSN(ret, node, pop);
10997 }
10998 break;
10999 }
11000 case NODE_IVAR:{
11001 debugi("nd_vid", RNODE_IVAR(node)->nd_vid);
11002 if (!popped) {
11003 ADD_INSN2(ret, node, getinstancevariable,
11004 ID2SYM(RNODE_IVAR(node)->nd_vid),
11005 get_ivar_ic_value(iseq, RNODE_IVAR(node)->nd_vid));
11006 }
11007 break;
11008 }
11009 case NODE_CONST:{
11010 debugi("nd_vid", RNODE_CONST(node)->nd_vid);
11011
11012 if (ISEQ_COMPILE_DATA(iseq)->option->inline_const_cache) {
11013 body->ic_size++;
11014 VALUE segments = rb_ary_new_from_args(1, ID2SYM(RNODE_CONST(node)->nd_vid));
11015 ADD_INSN1(ret, node, opt_getconstant_path, segments);
11016 RB_OBJ_WRITTEN(iseq, Qundef, segments);
11017 }
11018 else {
11019 ADD_INSN(ret, node, putnil);
11020 ADD_INSN1(ret, node, putobject, Qtrue);
11021 ADD_INSN1(ret, node, getconstant, ID2SYM(RNODE_CONST(node)->nd_vid));
11022 }
11023
11024 if (popped) {
11025 ADD_INSN(ret, node, pop);
11026 }
11027 break;
11028 }
11029 case NODE_CVAR:{
11030 if (!popped) {
11031 ADD_INSN2(ret, node, getclassvariable,
11032 ID2SYM(RNODE_CVAR(node)->nd_vid),
11033 get_cvar_ic_value(iseq, RNODE_CVAR(node)->nd_vid));
11034 }
11035 break;
11036 }
11037 case NODE_NTH_REF:{
11038 if (!popped) {
11039 if (!RNODE_NTH_REF(node)->nd_nth) {
11040 ADD_INSN(ret, node, putnil);
11041 break;
11042 }
11043 ADD_INSN2(ret, node, getspecial, INT2FIX(1) /* '~' */,
11044 INT2FIX(RNODE_NTH_REF(node)->nd_nth << 1));
11045 }
11046 break;
11047 }
11048 case NODE_BACK_REF:{
11049 if (!popped) {
11050 ADD_INSN2(ret, node, getspecial, INT2FIX(1) /* '~' */,
11051 INT2FIX(0x01 | (RNODE_BACK_REF(node)->nd_nth << 1)));
11052 }
11053 break;
11054 }
11055 case NODE_MATCH:
11056 case NODE_MATCH2:
11057 case NODE_MATCH3:
11058 CHECK(compile_match(iseq, ret, node, popped, type));
11059 break;
11060 case NODE_SYM:{
11061 if (!popped) {
11062 ADD_INSN1(ret, node, putobject, rb_node_sym_string_val(node));
11063 }
11064 break;
11065 }
11066 case NODE_LINE:{
11067 if (!popped) {
11068 ADD_INSN1(ret, node, putobject, rb_node_line_lineno_val(node));
11069 }
11070 break;
11071 }
11072 case NODE_ENCODING:{
11073 if (!popped) {
11074 ADD_INSN1(ret, node, putobject, rb_node_encoding_val(node));
11075 }
11076 break;
11077 }
11078 case NODE_INTEGER:{
11079 VALUE lit = rb_node_integer_literal_val(node);
11080 debugp_param("integer", lit);
11081 if (!popped) {
11082 ADD_INSN1(ret, node, putobject, lit);
11083 RB_OBJ_WRITTEN(iseq, Qundef, lit);
11084 }
11085 break;
11086 }
11087 case NODE_FLOAT:{
11088 VALUE lit = rb_node_float_literal_val(node);
11089 debugp_param("float", lit);
11090 if (!popped) {
11091 ADD_INSN1(ret, node, putobject, lit);
11092 RB_OBJ_WRITTEN(iseq, Qundef, lit);
11093 }
11094 break;
11095 }
11096 case NODE_RATIONAL:{
11097 VALUE lit = rb_node_rational_literal_val(node);
11098 debugp_param("rational", lit);
11099 if (!popped) {
11100 ADD_INSN1(ret, node, putobject, lit);
11101 RB_OBJ_WRITTEN(iseq, Qundef, lit);
11102 }
11103 break;
11104 }
11105 case NODE_IMAGINARY:{
11106 VALUE lit = rb_node_imaginary_literal_val(node);
11107 debugp_param("imaginary", lit);
11108 if (!popped) {
11109 ADD_INSN1(ret, node, putobject, lit);
11110 RB_OBJ_WRITTEN(iseq, Qundef, lit);
11111 }
11112 break;
11113 }
11114 case NODE_FILE:
11115 case NODE_STR:{
11116 debugp_param("nd_lit", get_string_value(node));
11117 if (!popped) {
11118 VALUE lit = get_string_value(node);
11119 const rb_compile_option_t *option = ISEQ_COMPILE_DATA(iseq)->option;
11120 if ((option->debug_frozen_string_literal || RTEST(ruby_debug)) &&
11121 option->frozen_string_literal != ISEQ_FROZEN_STRING_LITERAL_DISABLED) {
11122 lit = rb_str_with_debug_created_info(lit, rb_iseq_path(iseq), line);
11123 }
11124 switch (option->frozen_string_literal) {
11125 case ISEQ_FROZEN_STRING_LITERAL_UNSET:
11126 ADD_INSN1(ret, node, putchilledstring, lit);
11127 break;
11128 case ISEQ_FROZEN_STRING_LITERAL_DISABLED:
11129 ADD_INSN1(ret, node, putstring, lit);
11130 break;
11131 case ISEQ_FROZEN_STRING_LITERAL_ENABLED:
11132 ADD_INSN1(ret, node, putobject, lit);
11133 break;
11134 default:
11135 rb_bug("invalid frozen_string_literal");
11136 }
11137 RB_OBJ_WRITTEN(iseq, Qundef, lit);
11138 }
11139 break;
11140 }
11141 case NODE_DSTR:{
11142 compile_dstr(iseq, ret, node);
11143
11144 if (popped) {
11145 ADD_INSN(ret, node, pop);
11146 }
11147 break;
11148 }
11149 case NODE_XSTR:{
11150 ADD_CALL_RECEIVER(ret, node);
11151 VALUE str = rb_node_str_string_val(node);
11152 ADD_INSN1(ret, node, putobject, str);
11153 RB_OBJ_WRITTEN(iseq, Qundef, str);
11154 ADD_CALL(ret, node, idBackquote, INT2FIX(1));
11155
11156 if (popped) {
11157 ADD_INSN(ret, node, pop);
11158 }
11159 break;
11160 }
11161 case NODE_DXSTR:{
11162 ADD_CALL_RECEIVER(ret, node);
11163 compile_dstr(iseq, ret, node);
11164 ADD_CALL(ret, node, idBackquote, INT2FIX(1));
11165
11166 if (popped) {
11167 ADD_INSN(ret, node, pop);
11168 }
11169 break;
11170 }
11171 case NODE_EVSTR:
11172 CHECK(compile_evstr(iseq, ret, RNODE_EVSTR(node)->nd_body, popped));
11173 break;
11174 case NODE_REGX:{
11175 if (!popped) {
11176 VALUE lit = rb_node_regx_string_val(node);
11177 ADD_INSN1(ret, node, putobject, lit);
11178 RB_OBJ_WRITTEN(iseq, Qundef, lit);
11179 }
11180 break;
11181 }
11182 case NODE_DREGX:
11183 compile_dregx(iseq, ret, node, popped);
11184 break;
11185 case NODE_ONCE:{
11186 int ic_index = body->ise_size++;
11187 const rb_iseq_t *block_iseq;
11188 block_iseq = NEW_CHILD_ISEQ(RNODE_ONCE(node)->nd_body, make_name_for_block(iseq), ISEQ_TYPE_PLAIN, line);
11189
11190 ADD_INSN2(ret, node, once, block_iseq, INT2FIX(ic_index));
11191 RB_OBJ_WRITTEN(iseq, Qundef, (VALUE)block_iseq);
11192
11193 if (popped) {
11194 ADD_INSN(ret, node, pop);
11195 }
11196 break;
11197 }
11198 case NODE_ARGSCAT:{
11199 if (popped) {
11200 CHECK(COMPILE(ret, "argscat head", RNODE_ARGSCAT(node)->nd_head));
11201 ADD_INSN1(ret, node, splatarray, Qfalse);
11202 ADD_INSN(ret, node, pop);
11203 CHECK(COMPILE(ret, "argscat body", RNODE_ARGSCAT(node)->nd_body));
11204 ADD_INSN1(ret, node, splatarray, Qfalse);
11205 ADD_INSN(ret, node, pop);
11206 }
11207 else {
11208 CHECK(COMPILE(ret, "argscat head", RNODE_ARGSCAT(node)->nd_head));
11209 const NODE *body_node = RNODE_ARGSCAT(node)->nd_body;
11210 if (nd_type_p(body_node, NODE_LIST)) {
11211 CHECK(compile_array(iseq, ret, body_node, popped, FALSE) >= 0);
11212 }
11213 else {
11214 CHECK(COMPILE(ret, "argscat body", body_node));
11215 ADD_INSN(ret, node, concattoarray);
11216 }
11217 }
11218 break;
11219 }
11220 case NODE_ARGSPUSH:{
11221 if (popped) {
11222 CHECK(COMPILE(ret, "argspush head", RNODE_ARGSPUSH(node)->nd_head));
11223 ADD_INSN1(ret, node, splatarray, Qfalse);
11224 ADD_INSN(ret, node, pop);
11225 CHECK(COMPILE_(ret, "argspush body", RNODE_ARGSPUSH(node)->nd_body, popped));
11226 }
11227 else {
11228 CHECK(COMPILE(ret, "argspush head", RNODE_ARGSPUSH(node)->nd_head));
11229 const NODE *body_node = RNODE_ARGSPUSH(node)->nd_body;
11230 if (keyword_node_p(body_node)) {
11231 CHECK(COMPILE_(ret, "array element", body_node, FALSE));
11232 ADD_INSN(ret, node, pushtoarraykwsplat);
11233 }
11234 else if (static_literal_node_p(body_node, iseq, false)) {
11235 ADD_INSN1(ret, body_node, putobject, static_literal_value(body_node, iseq));
11236 ADD_INSN1(ret, node, pushtoarray, INT2FIX(1));
11237 }
11238 else {
11239 CHECK(COMPILE_(ret, "array element", body_node, FALSE));
11240 ADD_INSN1(ret, node, pushtoarray, INT2FIX(1));
11241 }
11242 }
11243 break;
11244 }
11245 case NODE_SPLAT:{
11246 CHECK(COMPILE(ret, "splat", RNODE_SPLAT(node)->nd_head));
11247 ADD_INSN1(ret, node, splatarray, Qtrue);
11248
11249 if (popped) {
11250 ADD_INSN(ret, node, pop);
11251 }
11252 break;
11253 }
11254 case NODE_DEFN:{
11255 ID mid = RNODE_DEFN(node)->nd_mid;
11256 const rb_iseq_t *method_iseq = NEW_ISEQ(RNODE_DEFN(node)->nd_defn,
11257 rb_id2str(mid),
11258 ISEQ_TYPE_METHOD, line);
11259
11260 debugp_param("defn/iseq", rb_iseqw_new(method_iseq));
11261 ADD_INSN2(ret, node, definemethod, ID2SYM(mid), method_iseq);
11262 RB_OBJ_WRITTEN(iseq, Qundef, (VALUE)method_iseq);
11263
11264 if (!popped) {
11265 ADD_INSN1(ret, node, putobject, ID2SYM(mid));
11266 }
11267
11268 break;
11269 }
11270 case NODE_DEFS:{
11271 ID mid = RNODE_DEFS(node)->nd_mid;
11272 const rb_iseq_t * singleton_method_iseq = NEW_ISEQ(RNODE_DEFS(node)->nd_defn,
11273 rb_id2str(mid),
11274 ISEQ_TYPE_METHOD, line);
11275
11276 debugp_param("defs/iseq", rb_iseqw_new(singleton_method_iseq));
11277 CHECK(COMPILE(ret, "defs: recv", RNODE_DEFS(node)->nd_recv));
11278 ADD_INSN2(ret, node, definesmethod, ID2SYM(mid), singleton_method_iseq);
11279 RB_OBJ_WRITTEN(iseq, Qundef, (VALUE)singleton_method_iseq);
11280
11281 if (!popped) {
11282 ADD_INSN1(ret, node, putobject, ID2SYM(mid));
11283 }
11284 break;
11285 }
11286 case NODE_ALIAS:{
11287 ADD_INSN1(ret, node, putspecialobject, INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
11288 ADD_INSN1(ret, node, putspecialobject, INT2FIX(VM_SPECIAL_OBJECT_CBASE));
11289 CHECK(COMPILE(ret, "alias arg1", RNODE_ALIAS(node)->nd_1st));
11290 CHECK(COMPILE(ret, "alias arg2", RNODE_ALIAS(node)->nd_2nd));
11291 ADD_SEND(ret, node, id_core_set_method_alias, INT2FIX(3));
11292
11293 if (popped) {
11294 ADD_INSN(ret, node, pop);
11295 }
11296 break;
11297 }
11298 case NODE_VALIAS:{
11299 ADD_INSN1(ret, node, putspecialobject, INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
11300 ADD_INSN1(ret, node, putobject, ID2SYM(RNODE_VALIAS(node)->nd_alias));
11301 ADD_INSN1(ret, node, putobject, ID2SYM(RNODE_VALIAS(node)->nd_orig));
11302 ADD_SEND(ret, node, id_core_set_variable_alias, INT2FIX(2));
11303
11304 if (popped) {
11305 ADD_INSN(ret, node, pop);
11306 }
11307 break;
11308 }
11309 case NODE_UNDEF:{
11310 const rb_parser_ary_t *ary = RNODE_UNDEF(node)->nd_undefs;
11311
11312 for (long i = 0; i < ary->len; i++) {
11313 ADD_INSN1(ret, node, putspecialobject, INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
11314 ADD_INSN1(ret, node, putspecialobject, INT2FIX(VM_SPECIAL_OBJECT_CBASE));
11315 CHECK(COMPILE(ret, "undef arg", ary->data[i]));
11316 ADD_SEND(ret, node, id_core_undef_method, INT2FIX(2));
11317
11318 if (i < ary->len - 1) {
11319 ADD_INSN(ret, node, pop);
11320 }
11321 }
11322
11323 if (popped) {
11324 ADD_INSN(ret, node, pop);
11325 }
11326 break;
11327 }
11328 case NODE_CLASS:{
11329 const rb_iseq_t *class_iseq = NEW_CHILD_ISEQ(RNODE_CLASS(node)->nd_body,
11330 rb_str_freeze(rb_sprintf("<class:%"PRIsVALUE">", rb_id2str(get_node_colon_nd_mid(RNODE_CLASS(node)->nd_cpath)))),
11331 ISEQ_TYPE_CLASS, line);
11332 const int flags = VM_DEFINECLASS_TYPE_CLASS |
11333 (RNODE_CLASS(node)->nd_super ? VM_DEFINECLASS_FLAG_HAS_SUPERCLASS : 0) |
11334 compile_cpath(ret, iseq, RNODE_CLASS(node)->nd_cpath);
11335
11336 CHECK(COMPILE(ret, "super", RNODE_CLASS(node)->nd_super));
11337 ADD_INSN3(ret, node, defineclass, ID2SYM(get_node_colon_nd_mid(RNODE_CLASS(node)->nd_cpath)), class_iseq, INT2FIX(flags));
11338 RB_OBJ_WRITTEN(iseq, Qundef, (VALUE)class_iseq);
11339
11340 if (popped) {
11341 ADD_INSN(ret, node, pop);
11342 }
11343 break;
11344 }
11345 case NODE_MODULE:{
11346 const rb_iseq_t *module_iseq = NEW_CHILD_ISEQ(RNODE_MODULE(node)->nd_body,
11347 rb_str_freeze(rb_sprintf("<module:%"PRIsVALUE">", rb_id2str(get_node_colon_nd_mid(RNODE_MODULE(node)->nd_cpath)))),
11348 ISEQ_TYPE_CLASS, line);
11349 const int flags = VM_DEFINECLASS_TYPE_MODULE |
11350 compile_cpath(ret, iseq, RNODE_MODULE(node)->nd_cpath);
11351
11352 ADD_INSN (ret, node, putnil); /* dummy */
11353 ADD_INSN3(ret, node, defineclass, ID2SYM(get_node_colon_nd_mid(RNODE_MODULE(node)->nd_cpath)), module_iseq, INT2FIX(flags));
11354 RB_OBJ_WRITTEN(iseq, Qundef, (VALUE)module_iseq);
11355
11356 if (popped) {
11357 ADD_INSN(ret, node, pop);
11358 }
11359 break;
11360 }
11361 case NODE_SCLASS:{
11362 ID singletonclass;
11363 const rb_iseq_t *singleton_class = NEW_ISEQ(RNODE_SCLASS(node)->nd_body, rb_fstring_lit("singleton class"),
11364 ISEQ_TYPE_CLASS, line);
11365
11366 CHECK(COMPILE(ret, "sclass#recv", RNODE_SCLASS(node)->nd_recv));
11367 ADD_INSN (ret, node, putnil);
11368 CONST_ID(singletonclass, "singletonclass");
11369 ADD_INSN3(ret, node, defineclass,
11370 ID2SYM(singletonclass), singleton_class,
11371 INT2FIX(VM_DEFINECLASS_TYPE_SINGLETON_CLASS));
11372 RB_OBJ_WRITTEN(iseq, Qundef, (VALUE)singleton_class);
11373
11374 if (popped) {
11375 ADD_INSN(ret, node, pop);
11376 }
11377 break;
11378 }
11379 case NODE_COLON2:
11380 CHECK(compile_colon2(iseq, ret, node, popped));
11381 break;
11382 case NODE_COLON3:
11383 CHECK(compile_colon3(iseq, ret, node, popped));
11384 break;
11385 case NODE_DOT2:
11386 CHECK(compile_dots(iseq, ret, node, popped, FALSE));
11387 break;
11388 case NODE_DOT3:
11389 CHECK(compile_dots(iseq, ret, node, popped, TRUE));
11390 break;
11391 case NODE_FLIP2:
11392 case NODE_FLIP3:{
11393 LABEL *lend = NEW_LABEL(line);
11394 LABEL *ltrue = NEW_LABEL(line);
11395 LABEL *lfalse = NEW_LABEL(line);
11396 CHECK(compile_flip_flop(iseq, ret, node, type == NODE_FLIP2,
11397 ltrue, lfalse));
11398 ADD_LABEL(ret, ltrue);
11399 ADD_INSN1(ret, node, putobject, Qtrue);
11400 ADD_INSNL(ret, node, jump, lend);
11401 ADD_LABEL(ret, lfalse);
11402 ADD_INSN1(ret, node, putobject, Qfalse);
11403 ADD_LABEL(ret, lend);
11404 break;
11405 }
11406 case NODE_SELF:{
11407 if (!popped) {
11408 ADD_INSN(ret, node, putself);
11409 }
11410 break;
11411 }
11412 case NODE_NIL:{
11413 if (!popped) {
11414 ADD_INSN(ret, node, putnil);
11415 }
11416 break;
11417 }
11418 case NODE_TRUE:{
11419 if (!popped) {
11420 ADD_INSN1(ret, node, putobject, Qtrue);
11421 }
11422 break;
11423 }
11424 case NODE_FALSE:{
11425 if (!popped) {
11426 ADD_INSN1(ret, node, putobject, Qfalse);
11427 }
11428 break;
11429 }
11430 case NODE_ERRINFO:
11431 CHECK(compile_errinfo(iseq, ret, node, popped));
11432 break;
11433 case NODE_DEFINED:
11434 if (!popped) {
11435 CHECK(compile_defined_expr(iseq, ret, node, Qtrue, false));
11436 }
11437 break;
11438 case NODE_POSTEXE:{
11439 /* compiled to:
11440 * ONCE{ rb_mRubyVMFrozenCore::core#set_postexe{ ... } }
11441 */
11442 int is_index = body->ise_size++;
11444 rb_iseq_new_with_callback_new_callback(build_postexe_iseq, RNODE_POSTEXE(node)->nd_body);
11445 const rb_iseq_t *once_iseq =
11446 NEW_CHILD_ISEQ_WITH_CALLBACK(ifunc, rb_fstring(make_name_for_block(iseq)), ISEQ_TYPE_BLOCK, line);
11447
11448 ADD_INSN2(ret, node, once, once_iseq, INT2FIX(is_index));
11449 RB_OBJ_WRITTEN(iseq, Qundef, (VALUE)once_iseq);
11450
11451 if (popped) {
11452 ADD_INSN(ret, node, pop);
11453 }
11454 break;
11455 }
11456 case NODE_KW_ARG:
11457 CHECK(compile_kw_arg(iseq, ret, node, popped));
11458 break;
11459 case NODE_DSYM:{
11460 compile_dstr(iseq, ret, node);
11461 if (!popped) {
11462 ADD_INSN(ret, node, intern);
11463 }
11464 else {
11465 ADD_INSN(ret, node, pop);
11466 }
11467 break;
11468 }
11469 case NODE_ATTRASGN:
11470 CHECK(compile_attrasgn(iseq, ret, node, popped));
11471 break;
11472 case NODE_LAMBDA:{
11473 /* compile same as lambda{...} */
11474 const rb_iseq_t *block = NEW_CHILD_ISEQ(RNODE_LAMBDA(node)->nd_body, make_name_for_block(iseq), ISEQ_TYPE_BLOCK, line);
11475 VALUE argc = INT2FIX(0);
11476
11477 ADD_INSN1(ret, node, putspecialobject, INT2FIX(VM_SPECIAL_OBJECT_VMCORE));
11478 ADD_CALL_WITH_BLOCK(ret, node, idLambda, argc, block);
11479 RB_OBJ_WRITTEN(iseq, Qundef, (VALUE)block);
11480
11481 if (popped) {
11482 ADD_INSN(ret, node, pop);
11483 }
11484 break;
11485 }
11486 default:
11487 UNKNOWN_NODE("iseq_compile_each", node, COMPILE_NG);
11488 ng:
11489 debug_node_end();
11490 return COMPILE_NG;
11491 }
11492
11493 debug_node_end();
11494 return COMPILE_OK;
11495}
11496
11497/***************************/
11498/* instruction information */
11499/***************************/
11500
11501static int
11502insn_data_length(INSN *iobj)
11503{
11504 return insn_len(iobj->insn_id);
11505}
11506
11507static int
11508calc_sp_depth(int depth, INSN *insn)
11509{
11510 return comptime_insn_stack_increase(depth, insn->insn_id, insn->operands);
11511}
11512
11513static VALUE
11514opobj_inspect(VALUE obj)
11515{
11516 if (!SPECIAL_CONST_P(obj) && !RBASIC_CLASS(obj)) {
11517 switch (BUILTIN_TYPE(obj)) {
11518 case T_STRING:
11519 obj = rb_str_new_cstr(RSTRING_PTR(obj));
11520 break;
11521 case T_ARRAY:
11522 obj = rb_ary_dup(obj);
11523 break;
11524 default:
11525 break;
11526 }
11527 }
11528 return rb_inspect(obj);
11529}
11530
11531
11532
11533static VALUE
11534insn_data_to_s_detail(INSN *iobj)
11535{
11536 VALUE str = rb_sprintf("%-20s ", insn_name(iobj->insn_id));
11537
11538 if (iobj->operands) {
11539 const char *types = insn_op_types(iobj->insn_id);
11540 int j;
11541
11542 for (j = 0; types[j]; j++) {
11543 char type = types[j];
11544
11545 switch (type) {
11546 case TS_OFFSET: /* label(destination position) */
11547 {
11548 LABEL *lobj = (LABEL *)OPERAND_AT(iobj, j);
11549 rb_str_catf(str, LABEL_FORMAT, lobj->label_no);
11550 break;
11551 }
11552 break;
11553 case TS_ISEQ: /* iseq */
11554 {
11555 rb_iseq_t *iseq = (rb_iseq_t *)OPERAND_AT(iobj, j);
11556 VALUE val = Qnil;
11557 if (0 && iseq) { /* TODO: invalidate now */
11558 val = (VALUE)iseq;
11559 }
11560 rb_str_concat(str, opobj_inspect(val));
11561 }
11562 break;
11563 case TS_LINDEX:
11564 case TS_NUM: /* ulong */
11565 case TS_VALUE: /* VALUE */
11566 {
11567 VALUE v = OPERAND_AT(iobj, j);
11568 if (!CLASS_OF(v))
11569 rb_str_cat2(str, "<hidden>");
11570 else {
11571 rb_str_concat(str, opobj_inspect(v));
11572 }
11573 break;
11574 }
11575 case TS_ID: /* ID */
11576 rb_str_concat(str, opobj_inspect(OPERAND_AT(iobj, j)));
11577 break;
11578 case TS_IC: /* inline cache */
11579 rb_str_concat(str, opobj_inspect(OPERAND_AT(iobj, j)));
11580 break;
11581 case TS_IVC: /* inline ivar cache */
11582 rb_str_catf(str, "<ivc:%d>", FIX2INT(OPERAND_AT(iobj, j)));
11583 break;
11584 case TS_ICVARC: /* inline cvar cache */
11585 rb_str_catf(str, "<icvarc:%d>", FIX2INT(OPERAND_AT(iobj, j)));
11586 break;
11587 case TS_ISE: /* inline storage entry */
11588 rb_str_catf(str, "<ise:%d>", FIX2INT(OPERAND_AT(iobj, j)));
11589 break;
11590 case TS_CALLDATA: /* we store these as call infos at compile time */
11591 {
11592 const struct rb_callinfo *ci = (struct rb_callinfo *)OPERAND_AT(iobj, j);
11593 rb_str_cat2(str, "<calldata:");
11594 if (vm_ci_mid(ci)) rb_str_catf(str, "%"PRIsVALUE, rb_id2str(vm_ci_mid(ci)));
11595 rb_str_catf(str, ", %d>", vm_ci_argc(ci));
11596 break;
11597 }
11598 case TS_CDHASH: /* case/when condition cache */
11599 rb_str_cat2(str, "<ch>");
11600 break;
11601 case TS_FUNCPTR:
11602 {
11603 void *func = (void *)OPERAND_AT(iobj, j);
11604#ifdef HAVE_DLADDR
11605 Dl_info info;
11606 if (dladdr(func, &info) && info.dli_sname) {
11607 rb_str_cat2(str, info.dli_sname);
11608 break;
11609 }
11610#endif
11611 rb_str_catf(str, "<%p>", func);
11612 }
11613 break;
11614 case TS_BUILTIN:
11615 rb_str_cat2(str, "<TS_BUILTIN>");
11616 break;
11617 default:{
11618 rb_raise(rb_eSyntaxError, "unknown operand type: %c", type);
11619 }
11620 }
11621 if (types[j + 1]) {
11622 rb_str_cat2(str, ", ");
11623 }
11624 }
11625 }
11626 return str;
11627}
11628
11629static void
11630dump_disasm_list(const LINK_ELEMENT *link)
11631{
11632 dump_disasm_list_with_cursor(link, NULL, NULL);
11633}
11634
11635static void
11636dump_disasm_list_with_cursor(const LINK_ELEMENT *link, const LINK_ELEMENT *curr, const LABEL *dest)
11637{
11638 int pos = 0;
11639 INSN *iobj;
11640 LABEL *lobj;
11641 VALUE str;
11642
11643 printf("-- raw disasm--------\n");
11644
11645 while (link) {
11646 if (curr) printf(curr == link ? "*" : " ");
11647 switch (link->type) {
11648 case ISEQ_ELEMENT_INSN:
11649 {
11650 iobj = (INSN *)link;
11651 str = insn_data_to_s_detail(iobj);
11652 printf(" %04d %-65s(%4u)\n", pos, StringValueCStr(str), iobj->insn_info.line_no);
11653 pos += insn_data_length(iobj);
11654 break;
11655 }
11656 case ISEQ_ELEMENT_LABEL:
11657 {
11658 lobj = (LABEL *)link;
11659 printf(LABEL_FORMAT" [sp: %d, unremovable: %d, refcnt: %d]%s\n", lobj->label_no, lobj->sp, lobj->unremovable, lobj->refcnt,
11660 dest == lobj ? " <---" : "");
11661 break;
11662 }
11663 case ISEQ_ELEMENT_TRACE:
11664 {
11665 TRACE *trace = (TRACE *)link;
11666 printf(" trace: %0x\n", trace->event);
11667 break;
11668 }
11669 case ISEQ_ELEMENT_ADJUST:
11670 {
11671 ADJUST *adjust = (ADJUST *)link;
11672 printf(" adjust: [label: %d]\n", adjust->label ? adjust->label->label_no : -1);
11673 break;
11674 }
11675 default:
11676 /* ignore */
11677 rb_raise(rb_eSyntaxError, "dump_disasm_list error: %d\n", (int)link->type);
11678 }
11679 link = link->next;
11680 }
11681 printf("---------------------\n");
11682 fflush(stdout);
11683}
11684
11685int
11686rb_insn_len(VALUE insn)
11687{
11688 return insn_len(insn);
11689}
11690
11691const char *
11692rb_insns_name(int i)
11693{
11694 return insn_name(i);
11695}
11696
11697VALUE
11698rb_insns_name_array(void)
11699{
11700 VALUE ary = rb_ary_new_capa(VM_INSTRUCTION_SIZE);
11701 int i;
11702 for (i = 0; i < VM_INSTRUCTION_SIZE; i++) {
11703 rb_ary_push(ary, rb_fstring_cstr(insn_name(i)));
11704 }
11705 return rb_ary_freeze(ary);
11706}
11707
11708static LABEL *
11709register_label(rb_iseq_t *iseq, struct st_table *labels_table, VALUE obj)
11710{
11711 LABEL *label = 0;
11712 st_data_t tmp;
11713 obj = rb_to_symbol_type(obj);
11714
11715 if (st_lookup(labels_table, obj, &tmp) == 0) {
11716 label = NEW_LABEL(0);
11717 st_insert(labels_table, obj, (st_data_t)label);
11718 }
11719 else {
11720 label = (LABEL *)tmp;
11721 }
11722 LABEL_REF(label);
11723 return label;
11724}
11725
11726static VALUE
11727get_exception_sym2type(VALUE sym)
11728{
11729 static VALUE symRescue, symEnsure, symRetry;
11730 static VALUE symBreak, symRedo, symNext;
11731
11732 if (symRescue == 0) {
11733 symRescue = ID2SYM(rb_intern_const("rescue"));
11734 symEnsure = ID2SYM(rb_intern_const("ensure"));
11735 symRetry = ID2SYM(rb_intern_const("retry"));
11736 symBreak = ID2SYM(rb_intern_const("break"));
11737 symRedo = ID2SYM(rb_intern_const("redo"));
11738 symNext = ID2SYM(rb_intern_const("next"));
11739 }
11740
11741 if (sym == symRescue) return CATCH_TYPE_RESCUE;
11742 if (sym == symEnsure) return CATCH_TYPE_ENSURE;
11743 if (sym == symRetry) return CATCH_TYPE_RETRY;
11744 if (sym == symBreak) return CATCH_TYPE_BREAK;
11745 if (sym == symRedo) return CATCH_TYPE_REDO;
11746 if (sym == symNext) return CATCH_TYPE_NEXT;
11747 rb_raise(rb_eSyntaxError, "invalid exception symbol: %+"PRIsVALUE, sym);
11748 return 0;
11749}
11750
11751static int
11752iseq_build_from_ary_exception(rb_iseq_t *iseq, struct st_table *labels_table,
11753 VALUE exception)
11754{
11755 int i;
11756
11757 for (i=0; i<RARRAY_LEN(exception); i++) {
11758 const rb_iseq_t *eiseq;
11759 VALUE v, type;
11760 LABEL *lstart, *lend, *lcont;
11761 unsigned int sp;
11762
11763 v = rb_to_array_type(RARRAY_AREF(exception, i));
11764 if (RARRAY_LEN(v) != 6) {
11765 rb_raise(rb_eSyntaxError, "wrong exception entry");
11766 }
11767 type = get_exception_sym2type(RARRAY_AREF(v, 0));
11768 if (NIL_P(RARRAY_AREF(v, 1))) {
11769 eiseq = NULL;
11770 }
11771 else {
11772 eiseq = rb_iseqw_to_iseq(rb_iseq_load(RARRAY_AREF(v, 1), (VALUE)iseq, Qnil));
11773 }
11774
11775 lstart = register_label(iseq, labels_table, RARRAY_AREF(v, 2));
11776 lend = register_label(iseq, labels_table, RARRAY_AREF(v, 3));
11777 lcont = register_label(iseq, labels_table, RARRAY_AREF(v, 4));
11778 sp = NUM2UINT(RARRAY_AREF(v, 5));
11779
11780 /* TODO: Dirty Hack! Fix me */
11781 if (type == CATCH_TYPE_RESCUE ||
11782 type == CATCH_TYPE_BREAK ||
11783 type == CATCH_TYPE_NEXT) {
11784 ++sp;
11785 }
11786
11787 lcont->sp = sp;
11788
11789 ADD_CATCH_ENTRY(type, lstart, lend, eiseq, lcont);
11790
11791 RB_GC_GUARD(v);
11792 }
11793 return COMPILE_OK;
11794}
11795
11796static struct st_table *
11797insn_make_insn_table(void)
11798{
11799 struct st_table *table;
11800 int i;
11801 table = st_init_numtable_with_size(VM_INSTRUCTION_SIZE);
11802
11803 for (i=0; i<VM_INSTRUCTION_SIZE; i++) {
11804 st_insert(table, ID2SYM(rb_intern_const(insn_name(i))), i);
11805 }
11806
11807 return table;
11808}
11809
11810static const rb_iseq_t *
11811iseq_build_load_iseq(const rb_iseq_t *iseq, VALUE op)
11812{
11813 VALUE iseqw;
11814 const rb_iseq_t *loaded_iseq;
11815
11816 if (RB_TYPE_P(op, T_ARRAY)) {
11817 iseqw = rb_iseq_load(op, (VALUE)iseq, Qnil);
11818 }
11819 else if (CLASS_OF(op) == rb_cISeq) {
11820 iseqw = op;
11821 }
11822 else {
11823 rb_raise(rb_eSyntaxError, "ISEQ is required");
11824 }
11825
11826 loaded_iseq = rb_iseqw_to_iseq(iseqw);
11827 return loaded_iseq;
11828}
11829
11830static VALUE
11831iseq_build_callinfo_from_hash(rb_iseq_t *iseq, VALUE op)
11832{
11833 ID mid = 0;
11834 int orig_argc = 0;
11835 unsigned int flag = 0;
11836 struct rb_callinfo_kwarg *kw_arg = 0;
11837
11838 if (!NIL_P(op)) {
11839 VALUE vmid = rb_hash_aref(op, ID2SYM(rb_intern_const("mid")));
11840 VALUE vflag = rb_hash_aref(op, ID2SYM(rb_intern_const("flag")));
11841 VALUE vorig_argc = rb_hash_aref(op, ID2SYM(rb_intern_const("orig_argc")));
11842 VALUE vkw_arg = rb_hash_aref(op, ID2SYM(rb_intern_const("kw_arg")));
11843
11844 if (!NIL_P(vmid)) mid = SYM2ID(vmid);
11845 if (!NIL_P(vflag)) flag = NUM2UINT(vflag);
11846 if (!NIL_P(vorig_argc)) orig_argc = FIX2INT(vorig_argc);
11847
11848 if (!NIL_P(vkw_arg)) {
11849 int i;
11850 int len = RARRAY_LENINT(vkw_arg);
11851 size_t n = rb_callinfo_kwarg_bytes(len);
11852
11853 kw_arg = xmalloc(n);
11854 kw_arg->references = 0;
11855 kw_arg->keyword_len = len;
11856 for (i = 0; i < len; i++) {
11857 VALUE kw = RARRAY_AREF(vkw_arg, i);
11858 SYM2ID(kw); /* make immortal */
11859 kw_arg->keywords[i] = kw;
11860 }
11861 }
11862 }
11863
11864 const struct rb_callinfo *ci = new_callinfo(iseq, mid, orig_argc, flag, kw_arg, (flag & VM_CALL_ARGS_SIMPLE) == 0);
11865 RB_OBJ_WRITTEN(iseq, Qundef, ci);
11866 return (VALUE)ci;
11867}
11868
11869static rb_event_flag_t
11870event_name_to_flag(VALUE sym)
11871{
11872#define CHECK_EVENT(ev) if (sym == ID2SYM(rb_intern_const(#ev))) return ev;
11873 CHECK_EVENT(RUBY_EVENT_LINE);
11874 CHECK_EVENT(RUBY_EVENT_CLASS);
11875 CHECK_EVENT(RUBY_EVENT_END);
11876 CHECK_EVENT(RUBY_EVENT_CALL);
11877 CHECK_EVENT(RUBY_EVENT_RETURN);
11878 CHECK_EVENT(RUBY_EVENT_B_CALL);
11879 CHECK_EVENT(RUBY_EVENT_B_RETURN);
11880 CHECK_EVENT(RUBY_EVENT_RESCUE);
11881#undef CHECK_EVENT
11882 return RUBY_EVENT_NONE;
11883}
11884
11885static int
11886iseq_build_from_ary_body(rb_iseq_t *iseq, LINK_ANCHOR *const anchor,
11887 VALUE body, VALUE node_ids, VALUE labels_wrapper)
11888{
11889 /* TODO: body should be frozen */
11890 long i, len = RARRAY_LEN(body);
11891 struct st_table *labels_table = RTYPEDDATA_DATA(labels_wrapper);
11892 int j;
11893 int line_no = 0, node_id = -1, insn_idx = 0;
11894 int ret = COMPILE_OK;
11895
11896 /*
11897 * index -> LABEL *label
11898 */
11899 static struct st_table *insn_table;
11900
11901 if (insn_table == 0) {
11902 insn_table = insn_make_insn_table();
11903 }
11904
11905 for (i=0; i<len; i++) {
11906 VALUE obj = RARRAY_AREF(body, i);
11907
11908 if (SYMBOL_P(obj)) {
11909 rb_event_flag_t event;
11910 if ((event = event_name_to_flag(obj)) != RUBY_EVENT_NONE) {
11911 ADD_TRACE(anchor, event);
11912 }
11913 else {
11914 LABEL *label = register_label(iseq, labels_table, obj);
11915 ADD_LABEL(anchor, label);
11916 }
11917 }
11918 else if (FIXNUM_P(obj)) {
11919 line_no = NUM2INT(obj);
11920 }
11921 else if (RB_TYPE_P(obj, T_ARRAY)) {
11922 VALUE *argv = 0;
11923 int argc = RARRAY_LENINT(obj) - 1;
11924 st_data_t insn_id;
11925 VALUE insn;
11926
11927 if (node_ids) {
11928 node_id = NUM2INT(rb_ary_entry(node_ids, insn_idx++));
11929 }
11930
11931 insn = (argc < 0) ? Qnil : RARRAY_AREF(obj, 0);
11932 if (st_lookup(insn_table, (st_data_t)insn, &insn_id) == 0) {
11933 /* TODO: exception */
11934 COMPILE_ERROR(iseq, line_no,
11935 "unknown instruction: %+"PRIsVALUE, insn);
11936 ret = COMPILE_NG;
11937 break;
11938 }
11939
11940 if (argc != insn_len((VALUE)insn_id)-1) {
11941 COMPILE_ERROR(iseq, line_no,
11942 "operand size mismatch");
11943 ret = COMPILE_NG;
11944 break;
11945 }
11946
11947 if (argc > 0) {
11948 argv = compile_data_calloc2(iseq, sizeof(VALUE), argc);
11949
11950 // add element before operand setup to make GC root
11951 ADD_ELEM(anchor,
11952 (LINK_ELEMENT*)new_insn_core(iseq, line_no, node_id,
11953 (enum ruby_vminsn_type)insn_id, argc, argv));
11954
11955 for (j=0; j<argc; j++) {
11956 VALUE op = rb_ary_entry(obj, j+1);
11957 switch (insn_op_type((VALUE)insn_id, j)) {
11958 case TS_OFFSET: {
11959 LABEL *label = register_label(iseq, labels_table, op);
11960 argv[j] = (VALUE)label;
11961 break;
11962 }
11963 case TS_LINDEX:
11964 case TS_NUM:
11965 (void)NUM2INT(op);
11966 argv[j] = op;
11967 break;
11968 case TS_VALUE:
11969 argv[j] = op;
11970 RB_OBJ_WRITTEN(iseq, Qundef, op);
11971 break;
11972 case TS_ISEQ:
11973 {
11974 if (op != Qnil) {
11975 VALUE v = (VALUE)iseq_build_load_iseq(iseq, op);
11976 argv[j] = v;
11977 RB_OBJ_WRITTEN(iseq, Qundef, v);
11978 }
11979 else {
11980 argv[j] = 0;
11981 }
11982 }
11983 break;
11984 case TS_ISE:
11985 argv[j] = op;
11986 if (NUM2UINT(op) >= ISEQ_BODY(iseq)->ise_size) {
11987 ISEQ_BODY(iseq)->ise_size = NUM2INT(op) + 1;
11988 }
11989 break;
11990 case TS_IC:
11991 {
11992 VALUE segments = rb_ary_new();
11993 op = rb_to_array_type(op);
11994
11995 for (int i = 0; i < RARRAY_LEN(op); i++) {
11996 VALUE sym = RARRAY_AREF(op, i);
11997 sym = rb_to_symbol_type(sym);
11998 rb_ary_push(segments, sym);
11999 }
12000
12001 RB_GC_GUARD(op);
12002 argv[j] = segments;
12003 RB_OBJ_WRITTEN(iseq, Qundef, segments);
12004 ISEQ_BODY(iseq)->ic_size++;
12005 }
12006 break;
12007 case TS_IVC: /* inline ivar cache */
12008 argv[j] = op;
12009 if (NUM2UINT(op) >= ISEQ_BODY(iseq)->ivc_size) {
12010 ISEQ_BODY(iseq)->ivc_size = NUM2INT(op) + 1;
12011 }
12012 break;
12013 case TS_ICVARC: /* inline cvar cache */
12014 argv[j] = op;
12015 if (NUM2UINT(op) >= ISEQ_BODY(iseq)->icvarc_size) {
12016 ISEQ_BODY(iseq)->icvarc_size = NUM2INT(op) + 1;
12017 }
12018 break;
12019 case TS_CALLDATA:
12020 argv[j] = iseq_build_callinfo_from_hash(iseq, op);
12021 break;
12022 case TS_ID:
12023 argv[j] = rb_to_symbol_type(op);
12024 break;
12025 case TS_CDHASH:
12026 {
12027 int i;
12028 VALUE map = rb_hash_new_with_size(RARRAY_LEN(op)/2);
12029
12030 RHASH_TBL_RAW(map)->type = &cdhash_type;
12031 op = rb_to_array_type(op);
12032 for (i=0; i<RARRAY_LEN(op); i+=2) {
12033 VALUE key = RARRAY_AREF(op, i);
12034 VALUE sym = RARRAY_AREF(op, i+1);
12035 LABEL *label =
12036 register_label(iseq, labels_table, sym);
12037 rb_hash_aset(map, key, (VALUE)label | 1);
12038 }
12039 RB_GC_GUARD(op);
12040 argv[j] = map;
12041 RB_OBJ_WRITTEN(iseq, Qundef, map);
12042 }
12043 break;
12044 case TS_FUNCPTR:
12045 {
12046#if SIZEOF_VALUE <= SIZEOF_LONG
12047 long funcptr = NUM2LONG(op);
12048#else
12049 LONG_LONG funcptr = NUM2LL(op);
12050#endif
12051 argv[j] = (VALUE)funcptr;
12052 }
12053 break;
12054 default:
12055 rb_raise(rb_eSyntaxError, "unknown operand: %c", insn_op_type((VALUE)insn_id, j));
12056 }
12057 }
12058 }
12059 else {
12060 ADD_ELEM(anchor,
12061 (LINK_ELEMENT*)new_insn_core(iseq, line_no, node_id,
12062 (enum ruby_vminsn_type)insn_id, argc, NULL));
12063 }
12064 }
12065 else {
12066 rb_raise(rb_eTypeError, "unexpected object for instruction");
12067 }
12068 }
12069 RTYPEDDATA_DATA(labels_wrapper) = 0;
12070 RB_GC_GUARD(labels_wrapper);
12071 validate_labels(iseq, labels_table);
12072 if (!ret) return ret;
12073 return iseq_setup(iseq, anchor);
12074}
12075
12076#define CHECK_ARRAY(v) rb_to_array_type(v)
12077#define CHECK_SYMBOL(v) rb_to_symbol_type(v)
12078
12079static int
12080int_param(int *dst, VALUE param, VALUE sym)
12081{
12082 VALUE val = rb_hash_aref(param, sym);
12083 if (FIXNUM_P(val)) {
12084 *dst = FIX2INT(val);
12085 return TRUE;
12086 }
12087 else if (!NIL_P(val)) {
12088 rb_raise(rb_eTypeError, "invalid %+"PRIsVALUE" Fixnum: %+"PRIsVALUE,
12089 sym, val);
12090 }
12091 return FALSE;
12092}
12093
12094static const struct rb_iseq_param_keyword *
12095iseq_build_kw(rb_iseq_t *iseq, VALUE params, VALUE keywords)
12096{
12097 int i, j;
12098 int len = RARRAY_LENINT(keywords);
12099 int default_len;
12100 VALUE key, sym, default_val;
12101 VALUE *dvs;
12102 ID *ids;
12103 struct rb_iseq_param_keyword *keyword = ZALLOC(struct rb_iseq_param_keyword);
12104
12105 ISEQ_BODY(iseq)->param.flags.has_kw = TRUE;
12106
12107 keyword->num = len;
12108#define SYM(s) ID2SYM(rb_intern_const(#s))
12109 (void)int_param(&keyword->bits_start, params, SYM(kwbits));
12110 i = keyword->bits_start - keyword->num;
12111 ids = (ID *)&ISEQ_BODY(iseq)->local_table[i];
12112#undef SYM
12113
12114 /* required args */
12115 for (i = 0; i < len; i++) {
12116 VALUE val = RARRAY_AREF(keywords, i);
12117
12118 if (!SYMBOL_P(val)) {
12119 goto default_values;
12120 }
12121 ids[i] = SYM2ID(val);
12122 keyword->required_num++;
12123 }
12124
12125 default_values: /* note: we intentionally preserve `i' from previous loop */
12126 default_len = len - i;
12127 if (default_len == 0) {
12128 keyword->table = ids;
12129 return keyword;
12130 }
12131 else if (default_len < 0) {
12133 }
12134
12135 dvs = ALLOC_N(VALUE, (unsigned int)default_len);
12136
12137 for (j = 0; i < len; i++, j++) {
12138 key = RARRAY_AREF(keywords, i);
12139 CHECK_ARRAY(key);
12140
12141 switch (RARRAY_LEN(key)) {
12142 case 1:
12143 sym = RARRAY_AREF(key, 0);
12144 default_val = Qundef;
12145 break;
12146 case 2:
12147 sym = RARRAY_AREF(key, 0);
12148 default_val = RARRAY_AREF(key, 1);
12149 break;
12150 default:
12151 rb_raise(rb_eTypeError, "keyword default has unsupported len %+"PRIsVALUE, key);
12152 }
12153 ids[i] = SYM2ID(sym);
12154 RB_OBJ_WRITE(iseq, &dvs[j], default_val);
12155 }
12156
12157 keyword->table = ids;
12158 keyword->default_values = dvs;
12159
12160 return keyword;
12161}
12162
12163static void
12164iseq_insn_each_object_mark_and_move(VALUE * obj, VALUE _)
12165{
12166 rb_gc_mark_and_move(obj);
12167}
12168
12169void
12170rb_iseq_mark_and_move_insn_storage(struct iseq_compile_data_storage *storage)
12171{
12172 INSN *iobj = 0;
12173 size_t size = sizeof(INSN);
12174 unsigned int pos = 0;
12175
12176 while (storage) {
12177#ifdef STRICT_ALIGNMENT
12178 size_t padding = calc_padding((void *)&storage->buff[pos], size);
12179#else
12180 const size_t padding = 0; /* expected to be optimized by compiler */
12181#endif /* STRICT_ALIGNMENT */
12182 size_t offset = pos + size + padding;
12183 if (offset > storage->size || offset > storage->pos) {
12184 pos = 0;
12185 storage = storage->next;
12186 }
12187 else {
12188#ifdef STRICT_ALIGNMENT
12189 pos += (int)padding;
12190#endif /* STRICT_ALIGNMENT */
12191
12192 iobj = (INSN *)&storage->buff[pos];
12193
12194 if (iobj->operands) {
12195 iseq_insn_each_markable_object(iobj, iseq_insn_each_object_mark_and_move, (VALUE)0);
12196 }
12197 pos += (int)size;
12198 }
12199 }
12200}
12201
12202static const rb_data_type_t labels_wrapper_type = {
12203 .wrap_struct_name = "compiler/labels_wrapper",
12204 .function = {
12205 .dmark = (RUBY_DATA_FUNC)rb_mark_set,
12206 .dfree = (RUBY_DATA_FUNC)st_free_table,
12207 },
12208 .flags = RUBY_TYPED_FREE_IMMEDIATELY | RUBY_TYPED_WB_PROTECTED,
12209};
12210
12211void
12212rb_iseq_build_from_ary(rb_iseq_t *iseq, VALUE misc, VALUE locals, VALUE params,
12213 VALUE exception, VALUE body)
12214{
12215#define SYM(s) ID2SYM(rb_intern_const(#s))
12216 int i, len;
12217 unsigned int arg_size, local_size, stack_max;
12218 ID *tbl;
12219 struct st_table *labels_table = st_init_numtable();
12220 VALUE labels_wrapper = TypedData_Wrap_Struct(0, &labels_wrapper_type, labels_table);
12221 VALUE arg_opt_labels = rb_hash_aref(params, SYM(opt));
12222 VALUE keywords = rb_hash_aref(params, SYM(keyword));
12223 VALUE sym_arg_rest = ID2SYM(rb_intern_const("#arg_rest"));
12224 DECL_ANCHOR(anchor);
12225 INIT_ANCHOR(anchor);
12226
12227 len = RARRAY_LENINT(locals);
12228 ISEQ_BODY(iseq)->local_table_size = len;
12229 ISEQ_BODY(iseq)->local_table = tbl = len > 0 ? (ID *)ALLOC_N(ID, ISEQ_BODY(iseq)->local_table_size) : NULL;
12230
12231 for (i = 0; i < len; i++) {
12232 VALUE lv = RARRAY_AREF(locals, i);
12233
12234 if (sym_arg_rest == lv) {
12235 tbl[i] = 0;
12236 }
12237 else {
12238 tbl[i] = FIXNUM_P(lv) ? (ID)FIX2LONG(lv) : SYM2ID(CHECK_SYMBOL(lv));
12239 }
12240 }
12241
12242#define INT_PARAM(F) int_param(&ISEQ_BODY(iseq)->param.F, params, SYM(F))
12243 if (INT_PARAM(lead_num)) {
12244 ISEQ_BODY(iseq)->param.flags.has_lead = TRUE;
12245 }
12246 if (INT_PARAM(post_num)) ISEQ_BODY(iseq)->param.flags.has_post = TRUE;
12247 if (INT_PARAM(post_start)) ISEQ_BODY(iseq)->param.flags.has_post = TRUE;
12248 if (INT_PARAM(rest_start)) ISEQ_BODY(iseq)->param.flags.has_rest = TRUE;
12249 if (INT_PARAM(block_start)) ISEQ_BODY(iseq)->param.flags.has_block = TRUE;
12250#undef INT_PARAM
12251 {
12252#define INT_PARAM(F) F = (int_param(&x, misc, SYM(F)) ? (unsigned int)x : 0)
12253 int x;
12254 INT_PARAM(arg_size);
12255 INT_PARAM(local_size);
12256 INT_PARAM(stack_max);
12257#undef INT_PARAM
12258 }
12259
12260 VALUE node_ids = Qfalse;
12261#ifdef USE_ISEQ_NODE_ID
12262 node_ids = rb_hash_aref(misc, ID2SYM(rb_intern("node_ids")));
12263 if (!RB_TYPE_P(node_ids, T_ARRAY)) {
12264 rb_raise(rb_eTypeError, "node_ids is not an array");
12265 }
12266#endif
12267
12268 if (RB_TYPE_P(arg_opt_labels, T_ARRAY)) {
12269 len = RARRAY_LENINT(arg_opt_labels);
12270 ISEQ_BODY(iseq)->param.flags.has_opt = !!(len - 1 >= 0);
12271
12272 if (ISEQ_BODY(iseq)->param.flags.has_opt) {
12273 VALUE *opt_table = ALLOC_N(VALUE, len);
12274
12275 for (i = 0; i < len; i++) {
12276 VALUE ent = RARRAY_AREF(arg_opt_labels, i);
12277 LABEL *label = register_label(iseq, labels_table, ent);
12278 opt_table[i] = (VALUE)label;
12279 }
12280
12281 ISEQ_BODY(iseq)->param.opt_num = len - 1;
12282 ISEQ_BODY(iseq)->param.opt_table = opt_table;
12283 }
12284 }
12285 else if (!NIL_P(arg_opt_labels)) {
12286 rb_raise(rb_eTypeError, ":opt param is not an array: %+"PRIsVALUE,
12287 arg_opt_labels);
12288 }
12289
12290 if (RB_TYPE_P(keywords, T_ARRAY)) {
12291 ISEQ_BODY(iseq)->param.keyword = iseq_build_kw(iseq, params, keywords);
12292 }
12293 else if (!NIL_P(keywords)) {
12294 rb_raise(rb_eTypeError, ":keywords param is not an array: %+"PRIsVALUE,
12295 keywords);
12296 }
12297
12298 if (Qtrue == rb_hash_aref(params, SYM(ambiguous_param0))) {
12299 ISEQ_BODY(iseq)->param.flags.ambiguous_param0 = TRUE;
12300 }
12301
12302 if (Qtrue == rb_hash_aref(params, SYM(use_block))) {
12303 ISEQ_BODY(iseq)->param.flags.use_block = TRUE;
12304 }
12305
12306 if (int_param(&i, params, SYM(kwrest))) {
12307 struct rb_iseq_param_keyword *keyword = (struct rb_iseq_param_keyword *)ISEQ_BODY(iseq)->param.keyword;
12308 if (keyword == NULL) {
12309 ISEQ_BODY(iseq)->param.keyword = keyword = ZALLOC(struct rb_iseq_param_keyword);
12310 }
12311 keyword->rest_start = i;
12312 ISEQ_BODY(iseq)->param.flags.has_kwrest = TRUE;
12313 }
12314#undef SYM
12315 iseq_calc_param_size(iseq);
12316
12317 /* exception */
12318 iseq_build_from_ary_exception(iseq, labels_table, exception);
12319
12320 /* body */
12321 iseq_build_from_ary_body(iseq, anchor, body, node_ids, labels_wrapper);
12322
12323 ISEQ_BODY(iseq)->param.size = arg_size;
12324 ISEQ_BODY(iseq)->local_table_size = local_size;
12325 ISEQ_BODY(iseq)->stack_max = stack_max;
12326}
12327
12328/* for parser */
12329
12330int
12331rb_dvar_defined(ID id, const rb_iseq_t *iseq)
12332{
12333 if (iseq) {
12334 const struct rb_iseq_constant_body *body = ISEQ_BODY(iseq);
12335 while (body->type == ISEQ_TYPE_BLOCK ||
12336 body->type == ISEQ_TYPE_RESCUE ||
12337 body->type == ISEQ_TYPE_ENSURE ||
12338 body->type == ISEQ_TYPE_EVAL ||
12339 body->type == ISEQ_TYPE_MAIN
12340 ) {
12341 unsigned int i;
12342
12343 for (i = 0; i < body->local_table_size; i++) {
12344 if (body->local_table[i] == id) {
12345 return 1;
12346 }
12347 }
12348 iseq = body->parent_iseq;
12349 body = ISEQ_BODY(iseq);
12350 }
12351 }
12352 return 0;
12353}
12354
12355int
12356rb_local_defined(ID id, const rb_iseq_t *iseq)
12357{
12358 if (iseq) {
12359 unsigned int i;
12360 const struct rb_iseq_constant_body *const body = ISEQ_BODY(ISEQ_BODY(iseq)->local_iseq);
12361
12362 for (i=0; i<body->local_table_size; i++) {
12363 if (body->local_table[i] == id) {
12364 return 1;
12365 }
12366 }
12367 }
12368 return 0;
12369}
12370
12371/* ISeq binary format */
12372
12373#ifndef IBF_ISEQ_DEBUG
12374#define IBF_ISEQ_DEBUG 0
12375#endif
12376
12377#ifndef IBF_ISEQ_ENABLE_LOCAL_BUFFER
12378#define IBF_ISEQ_ENABLE_LOCAL_BUFFER 0
12379#endif
12380
12381typedef uint32_t ibf_offset_t;
12382#define IBF_OFFSET(ptr) ((ibf_offset_t)(VALUE)(ptr))
12383
12384#define IBF_MAJOR_VERSION ISEQ_MAJOR_VERSION
12385#ifdef RUBY_DEVEL
12386#define IBF_DEVEL_VERSION 4
12387#define IBF_MINOR_VERSION (ISEQ_MINOR_VERSION * 10000 + IBF_DEVEL_VERSION)
12388#else
12389#define IBF_MINOR_VERSION ISEQ_MINOR_VERSION
12390#endif
12391
12392static const char IBF_ENDIAN_MARK =
12393#ifdef WORDS_BIGENDIAN
12394 'b'
12395#else
12396 'l'
12397#endif
12398 ;
12399
12401 char magic[4]; /* YARB */
12402 uint32_t major_version;
12403 uint32_t minor_version;
12404 uint32_t size;
12405 uint32_t extra_size;
12406
12407 uint32_t iseq_list_size;
12408 uint32_t global_object_list_size;
12409 ibf_offset_t iseq_list_offset;
12410 ibf_offset_t global_object_list_offset;
12411 uint8_t endian;
12412 uint8_t wordsize; /* assume no 2048-bit CPU */
12413};
12414
12416 VALUE str;
12417 st_table *obj_table; /* obj -> obj number */
12418};
12419
12420struct ibf_dump {
12421 st_table *iseq_table; /* iseq -> iseq number */
12422 struct ibf_dump_buffer global_buffer;
12423 struct ibf_dump_buffer *current_buffer;
12424};
12425
12427 const char *buff;
12428 ibf_offset_t size;
12429
12430 VALUE obj_list; /* [obj0, ...] */
12431 unsigned int obj_list_size;
12432 ibf_offset_t obj_list_offset;
12433};
12434
12435struct ibf_load {
12436 const struct ibf_header *header;
12437 VALUE iseq_list; /* [iseq0, ...] */
12438 struct ibf_load_buffer global_buffer;
12439 VALUE loader_obj;
12440 rb_iseq_t *iseq;
12441 VALUE str;
12442 struct ibf_load_buffer *current_buffer;
12443};
12444
12446 long size;
12447 VALUE buffer[1];
12448};
12449
12450static void
12451pinned_list_mark(void *ptr)
12452{
12453 long i;
12454 struct pinned_list *list = (struct pinned_list *)ptr;
12455 for (i = 0; i < list->size; i++) {
12456 if (list->buffer[i]) {
12457 rb_gc_mark(list->buffer[i]);
12458 }
12459 }
12460}
12461
12462static const rb_data_type_t pinned_list_type = {
12463 "pinned_list",
12464 {
12465 pinned_list_mark,
12467 NULL, // No external memory to report,
12468 },
12469 0, 0, RUBY_TYPED_WB_PROTECTED | RUBY_TYPED_FREE_IMMEDIATELY | RUBY_TYPED_EMBEDDABLE
12470};
12471
12472static VALUE
12473pinned_list_fetch(VALUE list, long offset)
12474{
12475 struct pinned_list * ptr;
12476
12477 TypedData_Get_Struct(list, struct pinned_list, &pinned_list_type, ptr);
12478
12479 if (offset >= ptr->size) {
12480 rb_raise(rb_eIndexError, "object index out of range: %ld", offset);
12481 }
12482
12483 return ptr->buffer[offset];
12484}
12485
12486static void
12487pinned_list_store(VALUE list, long offset, VALUE object)
12488{
12489 struct pinned_list * ptr;
12490
12491 TypedData_Get_Struct(list, struct pinned_list, &pinned_list_type, ptr);
12492
12493 if (offset >= ptr->size) {
12494 rb_raise(rb_eIndexError, "object index out of range: %ld", offset);
12495 }
12496
12497 RB_OBJ_WRITE(list, &ptr->buffer[offset], object);
12498}
12499
12500static VALUE
12501pinned_list_new(long size)
12502{
12503 size_t memsize = offsetof(struct pinned_list, buffer) + size * sizeof(VALUE);
12504 VALUE obj_list = rb_data_typed_object_zalloc(0, memsize, &pinned_list_type);
12505 struct pinned_list * ptr = RTYPEDDATA_GET_DATA(obj_list);
12506 ptr->size = size;
12507 return obj_list;
12508}
12509
12510static ibf_offset_t
12511ibf_dump_pos(struct ibf_dump *dump)
12512{
12513 long pos = RSTRING_LEN(dump->current_buffer->str);
12514#if SIZEOF_LONG > SIZEOF_INT
12515 if (pos >= UINT_MAX) {
12516 rb_raise(rb_eRuntimeError, "dump size exceeds");
12517 }
12518#endif
12519 return (unsigned int)pos;
12520}
12521
12522static void
12523ibf_dump_align(struct ibf_dump *dump, size_t align)
12524{
12525 ibf_offset_t pos = ibf_dump_pos(dump);
12526 if (pos % align) {
12527 static const char padding[sizeof(VALUE)];
12528 size_t size = align - ((size_t)pos % align);
12529#if SIZEOF_LONG > SIZEOF_INT
12530 if (pos + size >= UINT_MAX) {
12531 rb_raise(rb_eRuntimeError, "dump size exceeds");
12532 }
12533#endif
12534 for (; size > sizeof(padding); size -= sizeof(padding)) {
12535 rb_str_cat(dump->current_buffer->str, padding, sizeof(padding));
12536 }
12537 rb_str_cat(dump->current_buffer->str, padding, size);
12538 }
12539}
12540
12541static ibf_offset_t
12542ibf_dump_write(struct ibf_dump *dump, const void *buff, unsigned long size)
12543{
12544 ibf_offset_t pos = ibf_dump_pos(dump);
12545 rb_str_cat(dump->current_buffer->str, (const char *)buff, size);
12546 /* TODO: overflow check */
12547 return pos;
12548}
12549
12550static ibf_offset_t
12551ibf_dump_write_byte(struct ibf_dump *dump, unsigned char byte)
12552{
12553 return ibf_dump_write(dump, &byte, sizeof(unsigned char));
12554}
12555
12556static void
12557ibf_dump_overwrite(struct ibf_dump *dump, void *buff, unsigned int size, long offset)
12558{
12559 VALUE str = dump->current_buffer->str;
12560 char *ptr = RSTRING_PTR(str);
12561 if ((unsigned long)(size + offset) > (unsigned long)RSTRING_LEN(str))
12562 rb_bug("ibf_dump_overwrite: overflow");
12563 memcpy(ptr + offset, buff, size);
12564}
12565
12566static const void *
12567ibf_load_ptr(const struct ibf_load *load, ibf_offset_t *offset, int size)
12568{
12569 ibf_offset_t beg = *offset;
12570 *offset += size;
12571 return load->current_buffer->buff + beg;
12572}
12573
12574static void *
12575ibf_load_alloc(const struct ibf_load *load, ibf_offset_t offset, size_t x, size_t y)
12576{
12577 void *buff = ruby_xmalloc2(x, y);
12578 size_t size = x * y;
12579 memcpy(buff, load->current_buffer->buff + offset, size);
12580 return buff;
12581}
12582
12583#define IBF_W_ALIGN(type) (RUBY_ALIGNOF(type) > 1 ? ibf_dump_align(dump, RUBY_ALIGNOF(type)) : (void)0)
12584
12585#define IBF_W(b, type, n) (IBF_W_ALIGN(type), (type *)(VALUE)IBF_WP(b, type, n))
12586#define IBF_WV(variable) ibf_dump_write(dump, &(variable), sizeof(variable))
12587#define IBF_WP(b, type, n) ibf_dump_write(dump, (b), sizeof(type) * (n))
12588#define IBF_R(val, type, n) (type *)ibf_load_alloc(load, IBF_OFFSET(val), sizeof(type), (n))
12589#define IBF_ZERO(variable) memset(&(variable), 0, sizeof(variable))
12590
12591static int
12592ibf_table_lookup(struct st_table *table, st_data_t key)
12593{
12594 st_data_t val;
12595
12596 if (st_lookup(table, key, &val)) {
12597 return (int)val;
12598 }
12599 else {
12600 return -1;
12601 }
12602}
12603
12604static int
12605ibf_table_find_or_insert(struct st_table *table, st_data_t key)
12606{
12607 int index = ibf_table_lookup(table, key);
12608
12609 if (index < 0) { /* not found */
12610 index = (int)table->num_entries;
12611 st_insert(table, key, (st_data_t)index);
12612 }
12613
12614 return index;
12615}
12616
12617/* dump/load generic */
12618
12619static void ibf_dump_object_list(struct ibf_dump *dump, ibf_offset_t *obj_list_offset, unsigned int *obj_list_size);
12620
12621static VALUE ibf_load_object(const struct ibf_load *load, VALUE object_index);
12622static rb_iseq_t *ibf_load_iseq(const struct ibf_load *load, const rb_iseq_t *index_iseq);
12623
12624static st_table *
12625ibf_dump_object_table_new(void)
12626{
12627 st_table *obj_table = st_init_numtable(); /* need free */
12628 st_insert(obj_table, (st_data_t)Qnil, (st_data_t)0); /* 0th is nil */
12629
12630 return obj_table;
12631}
12632
12633static VALUE
12634ibf_dump_object(struct ibf_dump *dump, VALUE obj)
12635{
12636 return ibf_table_find_or_insert(dump->current_buffer->obj_table, (st_data_t)obj);
12637}
12638
12639static VALUE
12640ibf_dump_id(struct ibf_dump *dump, ID id)
12641{
12642 if (id == 0 || rb_id2name(id) == NULL) {
12643 return 0;
12644 }
12645 return ibf_dump_object(dump, rb_id2sym(id));
12646}
12647
12648static ID
12649ibf_load_id(const struct ibf_load *load, const ID id_index)
12650{
12651 if (id_index == 0) {
12652 return 0;
12653 }
12654 VALUE sym = ibf_load_object(load, id_index);
12655 if (rb_integer_type_p(sym)) {
12656 /* Load hidden local variables as indexes */
12657 return NUM2ULONG(sym);
12658 }
12659 return rb_sym2id(sym);
12660}
12661
12662/* dump/load: code */
12663
12664static ibf_offset_t ibf_dump_iseq_each(struct ibf_dump *dump, const rb_iseq_t *iseq);
12665
12666static int
12667ibf_dump_iseq(struct ibf_dump *dump, const rb_iseq_t *iseq)
12668{
12669 if (iseq == NULL) {
12670 return -1;
12671 }
12672 else {
12673 return ibf_table_find_or_insert(dump->iseq_table, (st_data_t)iseq);
12674 }
12675}
12676
12677static unsigned char
12678ibf_load_byte(const struct ibf_load *load, ibf_offset_t *offset)
12679{
12680 if (*offset >= load->current_buffer->size) { rb_raise(rb_eRuntimeError, "invalid bytecode"); }
12681 return (unsigned char)load->current_buffer->buff[(*offset)++];
12682}
12683
12684/*
12685 * Small uint serialization
12686 * 0x00000000_00000000 - 0x00000000_0000007f: 1byte | XXXX XXX1 |
12687 * 0x00000000_00000080 - 0x00000000_00003fff: 2byte | XXXX XX10 | XXXX XXXX |
12688 * 0x00000000_00004000 - 0x00000000_001fffff: 3byte | XXXX X100 | XXXX XXXX | XXXX XXXX |
12689 * 0x00000000_00020000 - 0x00000000_0fffffff: 4byte | XXXX 1000 | XXXX XXXX | XXXX XXXX | XXXX XXXX |
12690 * ...
12691 * 0x00010000_00000000 - 0x00ffffff_ffffffff: 8byte | 1000 0000 | XXXX XXXX | XXXX XXXX | XXXX XXXX | XXXX XXXX | XXXX XXXX | XXXX XXXX | XXXX XXXX |
12692 * 0x01000000_00000000 - 0xffffffff_ffffffff: 9byte | 0000 0000 | XXXX XXXX | XXXX XXXX | XXXX XXXX | XXXX XXXX | XXXX XXXX | XXXX XXXX | XXXX XXXX | XXXX XXXX |
12693 */
12694static void
12695ibf_dump_write_small_value(struct ibf_dump *dump, VALUE x)
12696{
12697 if (sizeof(VALUE) > 8 || CHAR_BIT != 8) {
12698 ibf_dump_write(dump, &x, sizeof(VALUE));
12699 return;
12700 }
12701
12702 enum { max_byte_length = sizeof(VALUE) + 1 };
12703
12704 unsigned char bytes[max_byte_length];
12705 ibf_offset_t n;
12706
12707 for (n = 0; n < sizeof(VALUE) && (x >> (7 - n)); n++, x >>= 8) {
12708 bytes[max_byte_length - 1 - n] = (unsigned char)x;
12709 }
12710
12711 x <<= 1;
12712 x |= 1;
12713 x <<= n;
12714 bytes[max_byte_length - 1 - n] = (unsigned char)x;
12715 n++;
12716
12717 ibf_dump_write(dump, bytes + max_byte_length - n, n);
12718}
12719
12720static VALUE
12721ibf_load_small_value(const struct ibf_load *load, ibf_offset_t *offset)
12722{
12723 if (sizeof(VALUE) > 8 || CHAR_BIT != 8) {
12724 union { char s[sizeof(VALUE)]; VALUE v; } x;
12725
12726 memcpy(x.s, load->current_buffer->buff + *offset, sizeof(VALUE));
12727 *offset += sizeof(VALUE);
12728
12729 return x.v;
12730 }
12731
12732 enum { max_byte_length = sizeof(VALUE) + 1 };
12733
12734 const unsigned char *buffer = (const unsigned char *)load->current_buffer->buff;
12735 const unsigned char c = buffer[*offset];
12736
12737 ibf_offset_t n =
12738 c & 1 ? 1 :
12739 c == 0 ? 9 : ntz_int32(c) + 1;
12740 VALUE x = (VALUE)c >> n;
12741
12742 if (*offset + n > load->current_buffer->size) {
12743 rb_raise(rb_eRuntimeError, "invalid byte sequence");
12744 }
12745
12746 ibf_offset_t i;
12747 for (i = 1; i < n; i++) {
12748 x <<= 8;
12749 x |= (VALUE)buffer[*offset + i];
12750 }
12751
12752 *offset += n;
12753 return x;
12754}
12755
12756static void
12757ibf_dump_builtin(struct ibf_dump *dump, const struct rb_builtin_function *bf)
12758{
12759 // short: index
12760 // short: name.length
12761 // bytes: name
12762 // // omit argc (only verify with name)
12763 ibf_dump_write_small_value(dump, (VALUE)bf->index);
12764
12765 size_t len = strlen(bf->name);
12766 ibf_dump_write_small_value(dump, (VALUE)len);
12767 ibf_dump_write(dump, bf->name, len);
12768}
12769
12770static const struct rb_builtin_function *
12771ibf_load_builtin(const struct ibf_load *load, ibf_offset_t *offset)
12772{
12773 int i = (int)ibf_load_small_value(load, offset);
12774 int len = (int)ibf_load_small_value(load, offset);
12775 const char *name = (char *)ibf_load_ptr(load, offset, len);
12776
12777 if (0) {
12778 fprintf(stderr, "%.*s!!\n", len, name);
12779 }
12780
12781 const struct rb_builtin_function *table = GET_VM()->builtin_function_table;
12782 if (table == NULL) rb_raise(rb_eArgError, "builtin function table is not provided");
12783 if (strncmp(table[i].name, name, len) != 0) {
12784 rb_raise(rb_eArgError, "builtin function index (%d) mismatch (expect %s but %s)", i, name, table[i].name);
12785 }
12786 // fprintf(stderr, "load-builtin: name:%s(%d)\n", table[i].name, table[i].argc);
12787
12788 return &table[i];
12789}
12790
12791static ibf_offset_t
12792ibf_dump_code(struct ibf_dump *dump, const rb_iseq_t *iseq)
12793{
12794 const struct rb_iseq_constant_body *const body = ISEQ_BODY(iseq);
12795 const int iseq_size = body->iseq_size;
12796 int code_index;
12797 const VALUE *orig_code = rb_iseq_original_iseq(iseq);
12798
12799 ibf_offset_t offset = ibf_dump_pos(dump);
12800
12801 for (code_index=0; code_index<iseq_size;) {
12802 const VALUE insn = orig_code[code_index++];
12803 const char *types = insn_op_types(insn);
12804 int op_index;
12805
12806 /* opcode */
12807 if (insn >= 0x100) { rb_raise(rb_eRuntimeError, "invalid instruction"); }
12808 ibf_dump_write_small_value(dump, insn);
12809
12810 /* operands */
12811 for (op_index=0; types[op_index]; op_index++, code_index++) {
12812 VALUE op = orig_code[code_index];
12813 VALUE wv;
12814
12815 switch (types[op_index]) {
12816 case TS_CDHASH:
12817 case TS_VALUE:
12818 wv = ibf_dump_object(dump, op);
12819 break;
12820 case TS_ISEQ:
12821 wv = (VALUE)ibf_dump_iseq(dump, (const rb_iseq_t *)op);
12822 break;
12823 case TS_IC:
12824 {
12825 IC ic = (IC)op;
12826 VALUE arr = idlist_to_array(ic->segments);
12827 wv = ibf_dump_object(dump, arr);
12828 }
12829 break;
12830 case TS_ISE:
12831 case TS_IVC:
12832 case TS_ICVARC:
12833 {
12835 wv = is - ISEQ_IS_ENTRY_START(body, types[op_index]);
12836 }
12837 break;
12838 case TS_CALLDATA:
12839 {
12840 goto skip_wv;
12841 }
12842 case TS_ID:
12843 wv = ibf_dump_id(dump, (ID)op);
12844 break;
12845 case TS_FUNCPTR:
12846 rb_raise(rb_eRuntimeError, "TS_FUNCPTR is not supported");
12847 goto skip_wv;
12848 case TS_BUILTIN:
12849 ibf_dump_builtin(dump, (const struct rb_builtin_function *)op);
12850 goto skip_wv;
12851 default:
12852 wv = op;
12853 break;
12854 }
12855 ibf_dump_write_small_value(dump, wv);
12856 skip_wv:;
12857 }
12858 RUBY_ASSERT(insn_len(insn) == op_index+1);
12859 }
12860
12861 return offset;
12862}
12863
12864static VALUE *
12865ibf_load_code(const struct ibf_load *load, rb_iseq_t *iseq, ibf_offset_t bytecode_offset, ibf_offset_t bytecode_size, unsigned int iseq_size)
12866{
12867 VALUE iseqv = (VALUE)iseq;
12868 unsigned int code_index;
12869 ibf_offset_t reading_pos = bytecode_offset;
12870 VALUE *code = ALLOC_N(VALUE, iseq_size);
12871
12872 struct rb_iseq_constant_body *load_body = ISEQ_BODY(iseq);
12873 struct rb_call_data *cd_entries = load_body->call_data;
12874 int ic_index = 0;
12875
12876 iseq_bits_t * mark_offset_bits;
12877
12878 iseq_bits_t tmp[1] = {0};
12879
12880 if (ISEQ_MBITS_BUFLEN(iseq_size) == 1) {
12881 mark_offset_bits = tmp;
12882 }
12883 else {
12884 mark_offset_bits = ZALLOC_N(iseq_bits_t, ISEQ_MBITS_BUFLEN(iseq_size));
12885 }
12886 bool needs_bitmap = false;
12887
12888 for (code_index=0; code_index<iseq_size;) {
12889 /* opcode */
12890 const VALUE insn = code[code_index] = ibf_load_small_value(load, &reading_pos);
12891 const char *types = insn_op_types(insn);
12892 int op_index;
12893
12894 code_index++;
12895
12896 /* operands */
12897 for (op_index=0; types[op_index]; op_index++, code_index++) {
12898 const char operand_type = types[op_index];
12899 switch (operand_type) {
12900 case TS_VALUE:
12901 {
12902 VALUE op = ibf_load_small_value(load, &reading_pos);
12903 VALUE v = ibf_load_object(load, op);
12904 code[code_index] = v;
12905 if (!SPECIAL_CONST_P(v)) {
12906 RB_OBJ_WRITTEN(iseqv, Qundef, v);
12907 ISEQ_MBITS_SET(mark_offset_bits, code_index);
12908 needs_bitmap = true;
12909 }
12910 break;
12911 }
12912 case TS_CDHASH:
12913 {
12914 VALUE op = ibf_load_small_value(load, &reading_pos);
12915 VALUE v = ibf_load_object(load, op);
12916 v = rb_hash_dup(v); // hash dumped as frozen
12917 RHASH_TBL_RAW(v)->type = &cdhash_type;
12918 rb_hash_rehash(v); // hash function changed
12919 freeze_hide_obj(v);
12920
12921 // Overwrite the existing hash in the object list. This
12922 // is to keep the object alive during load time.
12923 // [Bug #17984] [ruby-core:104259]
12924 pinned_list_store(load->current_buffer->obj_list, (long)op, v);
12925
12926 code[code_index] = v;
12927 ISEQ_MBITS_SET(mark_offset_bits, code_index);
12928 RB_OBJ_WRITTEN(iseqv, Qundef, v);
12929 needs_bitmap = true;
12930 break;
12931 }
12932 case TS_ISEQ:
12933 {
12934 VALUE op = (VALUE)ibf_load_small_value(load, &reading_pos);
12935 VALUE v = (VALUE)ibf_load_iseq(load, (const rb_iseq_t *)op);
12936 code[code_index] = v;
12937 if (!SPECIAL_CONST_P(v)) {
12938 RB_OBJ_WRITTEN(iseqv, Qundef, v);
12939 ISEQ_MBITS_SET(mark_offset_bits, code_index);
12940 needs_bitmap = true;
12941 }
12942 break;
12943 }
12944 case TS_IC:
12945 {
12946 VALUE op = ibf_load_small_value(load, &reading_pos);
12947 VALUE arr = ibf_load_object(load, op);
12948
12949 IC ic = &ISEQ_IS_IC_ENTRY(load_body, ic_index++);
12950 ic->segments = array_to_idlist(arr);
12951
12952 code[code_index] = (VALUE)ic;
12953 }
12954 break;
12955 case TS_ISE:
12956 case TS_ICVARC:
12957 case TS_IVC:
12958 {
12959 unsigned int op = (unsigned int)ibf_load_small_value(load, &reading_pos);
12960
12961 ISE ic = ISEQ_IS_ENTRY_START(load_body, operand_type) + op;
12962 code[code_index] = (VALUE)ic;
12963
12964 if (operand_type == TS_IVC) {
12965 IVC cache = (IVC)ic;
12966
12967 if (insn == BIN(setinstancevariable)) {
12968 ID iv_name = (ID)code[code_index - 1];
12969 cache->iv_set_name = iv_name;
12970 }
12971 else {
12972 cache->iv_set_name = 0;
12973 }
12974
12975 vm_ic_attr_index_initialize(cache, INVALID_SHAPE_ID);
12976 }
12977
12978 }
12979 break;
12980 case TS_CALLDATA:
12981 {
12982 code[code_index] = (VALUE)cd_entries++;
12983 }
12984 break;
12985 case TS_ID:
12986 {
12987 VALUE op = ibf_load_small_value(load, &reading_pos);
12988 code[code_index] = ibf_load_id(load, (ID)(VALUE)op);
12989 }
12990 break;
12991 case TS_FUNCPTR:
12992 rb_raise(rb_eRuntimeError, "TS_FUNCPTR is not supported");
12993 break;
12994 case TS_BUILTIN:
12995 code[code_index] = (VALUE)ibf_load_builtin(load, &reading_pos);
12996 break;
12997 default:
12998 code[code_index] = ibf_load_small_value(load, &reading_pos);
12999 continue;
13000 }
13001 }
13002 if (insn_len(insn) != op_index+1) {
13003 rb_raise(rb_eRuntimeError, "operand size mismatch");
13004 }
13005 }
13006
13007 load_body->iseq_encoded = code;
13008 load_body->iseq_size = code_index;
13009
13010 if (ISEQ_MBITS_BUFLEN(load_body->iseq_size) == 1) {
13011 load_body->mark_bits.single = mark_offset_bits[0];
13012 }
13013 else {
13014 if (needs_bitmap) {
13015 load_body->mark_bits.list = mark_offset_bits;
13016 }
13017 else {
13018 load_body->mark_bits.list = 0;
13019 ruby_xfree(mark_offset_bits);
13020 }
13021 }
13022
13023 RUBY_ASSERT(code_index == iseq_size);
13024 RUBY_ASSERT(reading_pos == bytecode_offset + bytecode_size);
13025 return code;
13026}
13027
13028static ibf_offset_t
13029ibf_dump_param_opt_table(struct ibf_dump *dump, const rb_iseq_t *iseq)
13030{
13031 int opt_num = ISEQ_BODY(iseq)->param.opt_num;
13032
13033 if (opt_num > 0) {
13034 IBF_W_ALIGN(VALUE);
13035 return ibf_dump_write(dump, ISEQ_BODY(iseq)->param.opt_table, sizeof(VALUE) * (opt_num + 1));
13036 }
13037 else {
13038 return ibf_dump_pos(dump);
13039 }
13040}
13041
13042static VALUE *
13043ibf_load_param_opt_table(const struct ibf_load *load, ibf_offset_t opt_table_offset, int opt_num)
13044{
13045 if (opt_num > 0) {
13046 VALUE *table = ALLOC_N(VALUE, opt_num+1);
13047 MEMCPY(table, load->current_buffer->buff + opt_table_offset, VALUE, opt_num+1);
13048 return table;
13049 }
13050 else {
13051 return NULL;
13052 }
13053}
13054
13055static ibf_offset_t
13056ibf_dump_param_keyword(struct ibf_dump *dump, const rb_iseq_t *iseq)
13057{
13058 const struct rb_iseq_param_keyword *kw = ISEQ_BODY(iseq)->param.keyword;
13059
13060 if (kw) {
13061 struct rb_iseq_param_keyword dump_kw = *kw;
13062 int dv_num = kw->num - kw->required_num;
13063 ID *ids = kw->num > 0 ? ALLOCA_N(ID, kw->num) : NULL;
13064 VALUE *dvs = dv_num > 0 ? ALLOCA_N(VALUE, dv_num) : NULL;
13065 int i;
13066
13067 for (i=0; i<kw->num; i++) ids[i] = (ID)ibf_dump_id(dump, kw->table[i]);
13068 for (i=0; i<dv_num; i++) dvs[i] = (VALUE)ibf_dump_object(dump, kw->default_values[i]);
13069
13070 dump_kw.table = IBF_W(ids, ID, kw->num);
13071 dump_kw.default_values = IBF_W(dvs, VALUE, dv_num);
13072 IBF_W_ALIGN(struct rb_iseq_param_keyword);
13073 return ibf_dump_write(dump, &dump_kw, sizeof(struct rb_iseq_param_keyword) * 1);
13074 }
13075 else {
13076 return 0;
13077 }
13078}
13079
13080static const struct rb_iseq_param_keyword *
13081ibf_load_param_keyword(const struct ibf_load *load, ibf_offset_t param_keyword_offset)
13082{
13083 if (param_keyword_offset) {
13084 struct rb_iseq_param_keyword *kw = IBF_R(param_keyword_offset, struct rb_iseq_param_keyword, 1);
13085 int dv_num = kw->num - kw->required_num;
13086 VALUE *dvs = dv_num ? IBF_R(kw->default_values, VALUE, dv_num) : NULL;
13087
13088 int i;
13089 for (i=0; i<dv_num; i++) {
13090 dvs[i] = ibf_load_object(load, dvs[i]);
13091 }
13092
13093 // Will be set once the local table is loaded.
13094 kw->table = NULL;
13095
13096 kw->default_values = dvs;
13097 return kw;
13098 }
13099 else {
13100 return NULL;
13101 }
13102}
13103
13104static ibf_offset_t
13105ibf_dump_insns_info_body(struct ibf_dump *dump, const rb_iseq_t *iseq)
13106{
13107 ibf_offset_t offset = ibf_dump_pos(dump);
13108 const struct iseq_insn_info_entry *entries = ISEQ_BODY(iseq)->insns_info.body;
13109
13110 unsigned int i;
13111 for (i = 0; i < ISEQ_BODY(iseq)->insns_info.size; i++) {
13112 ibf_dump_write_small_value(dump, entries[i].line_no);
13113#ifdef USE_ISEQ_NODE_ID
13114 ibf_dump_write_small_value(dump, entries[i].node_id);
13115#endif
13116 ibf_dump_write_small_value(dump, entries[i].events);
13117 }
13118
13119 return offset;
13120}
13121
13122static struct iseq_insn_info_entry *
13123ibf_load_insns_info_body(const struct ibf_load *load, ibf_offset_t body_offset, unsigned int size)
13124{
13125 ibf_offset_t reading_pos = body_offset;
13126 struct iseq_insn_info_entry *entries = ALLOC_N(struct iseq_insn_info_entry, size);
13127
13128 unsigned int i;
13129 for (i = 0; i < size; i++) {
13130 entries[i].line_no = (int)ibf_load_small_value(load, &reading_pos);
13131#ifdef USE_ISEQ_NODE_ID
13132 entries[i].node_id = (int)ibf_load_small_value(load, &reading_pos);
13133#endif
13134 entries[i].events = (rb_event_flag_t)ibf_load_small_value(load, &reading_pos);
13135 }
13136
13137 return entries;
13138}
13139
13140static ibf_offset_t
13141ibf_dump_insns_info_positions(struct ibf_dump *dump, const unsigned int *positions, unsigned int size)
13142{
13143 ibf_offset_t offset = ibf_dump_pos(dump);
13144
13145 unsigned int last = 0;
13146 unsigned int i;
13147 for (i = 0; i < size; i++) {
13148 ibf_dump_write_small_value(dump, positions[i] - last);
13149 last = positions[i];
13150 }
13151
13152 return offset;
13153}
13154
13155static unsigned int *
13156ibf_load_insns_info_positions(const struct ibf_load *load, ibf_offset_t positions_offset, unsigned int size)
13157{
13158 ibf_offset_t reading_pos = positions_offset;
13159 unsigned int *positions = ALLOC_N(unsigned int, size);
13160
13161 unsigned int last = 0;
13162 unsigned int i;
13163 for (i = 0; i < size; i++) {
13164 positions[i] = last + (unsigned int)ibf_load_small_value(load, &reading_pos);
13165 last = positions[i];
13166 }
13167
13168 return positions;
13169}
13170
13171static ibf_offset_t
13172ibf_dump_local_table(struct ibf_dump *dump, const rb_iseq_t *iseq)
13173{
13174 const struct rb_iseq_constant_body *const body = ISEQ_BODY(iseq);
13175 const int size = body->local_table_size;
13176 ID *table = ALLOCA_N(ID, size);
13177 int i;
13178
13179 for (i=0; i<size; i++) {
13180 VALUE v = ibf_dump_id(dump, body->local_table[i]);
13181 if (v == 0) {
13182 /* Dump hidden local variables as indexes, so load_from_binary will work with them */
13183 v = ibf_dump_object(dump, ULONG2NUM(body->local_table[i]));
13184 }
13185 table[i] = v;
13186 }
13187
13188 IBF_W_ALIGN(ID);
13189 return ibf_dump_write(dump, table, sizeof(ID) * size);
13190}
13191
13192static ID *
13193ibf_load_local_table(const struct ibf_load *load, ibf_offset_t local_table_offset, int size)
13194{
13195 if (size > 0) {
13196 ID *table = IBF_R(local_table_offset, ID, size);
13197 int i;
13198
13199 for (i=0; i<size; i++) {
13200 table[i] = ibf_load_id(load, table[i]);
13201 }
13202 return table;
13203 }
13204 else {
13205 return NULL;
13206 }
13207}
13208
13209static ibf_offset_t
13210ibf_dump_catch_table(struct ibf_dump *dump, const rb_iseq_t *iseq)
13211{
13212 const struct iseq_catch_table *table = ISEQ_BODY(iseq)->catch_table;
13213
13214 if (table) {
13215 int *iseq_indices = ALLOCA_N(int, table->size);
13216 unsigned int i;
13217
13218 for (i=0; i<table->size; i++) {
13219 iseq_indices[i] = ibf_dump_iseq(dump, table->entries[i].iseq);
13220 }
13221
13222 const ibf_offset_t offset = ibf_dump_pos(dump);
13223
13224 for (i=0; i<table->size; i++) {
13225 ibf_dump_write_small_value(dump, iseq_indices[i]);
13226 ibf_dump_write_small_value(dump, table->entries[i].type);
13227 ibf_dump_write_small_value(dump, table->entries[i].start);
13228 ibf_dump_write_small_value(dump, table->entries[i].end);
13229 ibf_dump_write_small_value(dump, table->entries[i].cont);
13230 ibf_dump_write_small_value(dump, table->entries[i].sp);
13231 }
13232 return offset;
13233 }
13234 else {
13235 return ibf_dump_pos(dump);
13236 }
13237}
13238
13239static struct iseq_catch_table *
13240ibf_load_catch_table(const struct ibf_load *load, ibf_offset_t catch_table_offset, unsigned int size)
13241{
13242 if (size) {
13243 struct iseq_catch_table *table = ruby_xmalloc(iseq_catch_table_bytes(size));
13244 table->size = size;
13245
13246 ibf_offset_t reading_pos = catch_table_offset;
13247
13248 unsigned int i;
13249 for (i=0; i<table->size; i++) {
13250 int iseq_index = (int)ibf_load_small_value(load, &reading_pos);
13251 table->entries[i].type = (enum rb_catch_type)ibf_load_small_value(load, &reading_pos);
13252 table->entries[i].start = (unsigned int)ibf_load_small_value(load, &reading_pos);
13253 table->entries[i].end = (unsigned int)ibf_load_small_value(load, &reading_pos);
13254 table->entries[i].cont = (unsigned int)ibf_load_small_value(load, &reading_pos);
13255 table->entries[i].sp = (unsigned int)ibf_load_small_value(load, &reading_pos);
13256
13257 table->entries[i].iseq = ibf_load_iseq(load, (const rb_iseq_t *)(VALUE)iseq_index);
13258 }
13259 return table;
13260 }
13261 else {
13262 return NULL;
13263 }
13264}
13265
13266static ibf_offset_t
13267ibf_dump_ci_entries(struct ibf_dump *dump, const rb_iseq_t *iseq)
13268{
13269 const struct rb_iseq_constant_body *const body = ISEQ_BODY(iseq);
13270 const unsigned int ci_size = body->ci_size;
13271 const struct rb_call_data *cds = body->call_data;
13272
13273 ibf_offset_t offset = ibf_dump_pos(dump);
13274
13275 unsigned int i;
13276
13277 for (i = 0; i < ci_size; i++) {
13278 const struct rb_callinfo *ci = cds[i].ci;
13279 if (ci != NULL) {
13280 ibf_dump_write_small_value(dump, ibf_dump_id(dump, vm_ci_mid(ci)));
13281 ibf_dump_write_small_value(dump, vm_ci_flag(ci));
13282 ibf_dump_write_small_value(dump, vm_ci_argc(ci));
13283
13284 const struct rb_callinfo_kwarg *kwarg = vm_ci_kwarg(ci);
13285 if (kwarg) {
13286 int len = kwarg->keyword_len;
13287 ibf_dump_write_small_value(dump, len);
13288 for (int j=0; j<len; j++) {
13289 VALUE keyword = ibf_dump_object(dump, kwarg->keywords[j]);
13290 ibf_dump_write_small_value(dump, keyword);
13291 }
13292 }
13293 else {
13294 ibf_dump_write_small_value(dump, 0);
13295 }
13296 }
13297 else {
13298 // TODO: truncate NULL ci from call_data.
13299 ibf_dump_write_small_value(dump, (VALUE)-1);
13300 }
13301 }
13302
13303 return offset;
13304}
13305
13307 ID id;
13308 VALUE name;
13309 VALUE val;
13310};
13311
13313 size_t num;
13314 struct outer_variable_pair pairs[1];
13315};
13316
13317static enum rb_id_table_iterator_result
13318store_outer_variable(ID id, VALUE val, void *dump)
13319{
13320 struct outer_variable_list *ovlist = dump;
13321 struct outer_variable_pair *pair = &ovlist->pairs[ovlist->num++];
13322 pair->id = id;
13323 pair->name = rb_id2str(id);
13324 pair->val = val;
13325 return ID_TABLE_CONTINUE;
13326}
13327
13328static int
13329outer_variable_cmp(const void *a, const void *b, void *arg)
13330{
13331 const struct outer_variable_pair *ap = (const struct outer_variable_pair *)a;
13332 const struct outer_variable_pair *bp = (const struct outer_variable_pair *)b;
13333 return rb_str_cmp(ap->name, bp->name);
13334}
13335
13336static ibf_offset_t
13337ibf_dump_outer_variables(struct ibf_dump *dump, const rb_iseq_t *iseq)
13338{
13339 struct rb_id_table * ovs = ISEQ_BODY(iseq)->outer_variables;
13340
13341 ibf_offset_t offset = ibf_dump_pos(dump);
13342
13343 size_t size = ovs ? rb_id_table_size(ovs) : 0;
13344 ibf_dump_write_small_value(dump, (VALUE)size);
13345 if (size > 0) {
13346 VALUE buff;
13347 size_t buffsize =
13348 rb_size_mul_add_or_raise(sizeof(struct outer_variable_pair), size,
13349 offsetof(struct outer_variable_list, pairs),
13350 rb_eArgError);
13351 struct outer_variable_list *ovlist = RB_ALLOCV(buff, buffsize);
13352 ovlist->num = 0;
13353 rb_id_table_foreach(ovs, store_outer_variable, ovlist);
13354 ruby_qsort(ovlist->pairs, size, sizeof(struct outer_variable_pair), outer_variable_cmp, NULL);
13355 for (size_t i = 0; i < size; ++i) {
13356 ID id = ovlist->pairs[i].id;
13357 ID val = ovlist->pairs[i].val;
13358 ibf_dump_write_small_value(dump, ibf_dump_id(dump, id));
13359 ibf_dump_write_small_value(dump, val);
13360 }
13361 }
13362
13363 return offset;
13364}
13365
13366/* note that we dump out rb_call_info but load back rb_call_data */
13367static void
13368ibf_load_ci_entries(const struct ibf_load *load,
13369 ibf_offset_t ci_entries_offset,
13370 unsigned int ci_size,
13371 struct rb_call_data **cd_ptr)
13372{
13373 ibf_offset_t reading_pos = ci_entries_offset;
13374
13375 unsigned int i;
13376
13377 struct rb_call_data *cds = ZALLOC_N(struct rb_call_data, ci_size);
13378 *cd_ptr = cds;
13379
13380 for (i = 0; i < ci_size; i++) {
13381 VALUE mid_index = ibf_load_small_value(load, &reading_pos);
13382 if (mid_index != (VALUE)-1) {
13383 ID mid = ibf_load_id(load, mid_index);
13384 unsigned int flag = (unsigned int)ibf_load_small_value(load, &reading_pos);
13385 unsigned int argc = (unsigned int)ibf_load_small_value(load, &reading_pos);
13386
13387 struct rb_callinfo_kwarg *kwarg = NULL;
13388 int kwlen = (int)ibf_load_small_value(load, &reading_pos);
13389 if (kwlen > 0) {
13390 kwarg = rb_xmalloc_mul_add(kwlen, sizeof(VALUE), sizeof(struct rb_callinfo_kwarg));
13391 kwarg->references = 0;
13392 kwarg->keyword_len = kwlen;
13393 for (int j=0; j<kwlen; j++) {
13394 VALUE keyword = ibf_load_small_value(load, &reading_pos);
13395 kwarg->keywords[j] = ibf_load_object(load, keyword);
13396 }
13397 }
13398
13399 cds[i].ci = vm_ci_new(mid, flag, argc, kwarg);
13400 RB_OBJ_WRITTEN(load->iseq, Qundef, cds[i].ci);
13401 cds[i].cc = vm_cc_empty();
13402 }
13403 else {
13404 // NULL ci
13405 cds[i].ci = NULL;
13406 cds[i].cc = NULL;
13407 }
13408 }
13409}
13410
13411static struct rb_id_table *
13412ibf_load_outer_variables(const struct ibf_load * load, ibf_offset_t outer_variables_offset)
13413{
13414 ibf_offset_t reading_pos = outer_variables_offset;
13415
13416 struct rb_id_table *tbl = NULL;
13417
13418 size_t table_size = (size_t)ibf_load_small_value(load, &reading_pos);
13419
13420 if (table_size > 0) {
13421 tbl = rb_id_table_create(table_size);
13422 }
13423
13424 for (size_t i = 0; i < table_size; i++) {
13425 ID key = ibf_load_id(load, (ID)ibf_load_small_value(load, &reading_pos));
13426 VALUE value = ibf_load_small_value(load, &reading_pos);
13427 if (!key) key = rb_make_temporary_id(i);
13428 rb_id_table_insert(tbl, key, value);
13429 }
13430
13431 return tbl;
13432}
13433
13434static ibf_offset_t
13435ibf_dump_iseq_each(struct ibf_dump *dump, const rb_iseq_t *iseq)
13436{
13437 RUBY_ASSERT(dump->current_buffer == &dump->global_buffer);
13438
13439 unsigned int *positions;
13440
13441 const struct rb_iseq_constant_body *body = ISEQ_BODY(iseq);
13442
13443 const VALUE location_pathobj_index = ibf_dump_object(dump, body->location.pathobj); /* TODO: freeze */
13444 const VALUE location_base_label_index = ibf_dump_object(dump, body->location.base_label);
13445 const VALUE location_label_index = ibf_dump_object(dump, body->location.label);
13446
13447#if IBF_ISEQ_ENABLE_LOCAL_BUFFER
13448 ibf_offset_t iseq_start = ibf_dump_pos(dump);
13449
13450 struct ibf_dump_buffer *saved_buffer = dump->current_buffer;
13451 struct ibf_dump_buffer buffer;
13452 buffer.str = rb_str_new(0, 0);
13453 buffer.obj_table = ibf_dump_object_table_new();
13454 dump->current_buffer = &buffer;
13455#endif
13456
13457 const ibf_offset_t bytecode_offset = ibf_dump_code(dump, iseq);
13458 const ibf_offset_t bytecode_size = ibf_dump_pos(dump) - bytecode_offset;
13459 const ibf_offset_t param_opt_table_offset = ibf_dump_param_opt_table(dump, iseq);
13460 const ibf_offset_t param_keyword_offset = ibf_dump_param_keyword(dump, iseq);
13461 const ibf_offset_t insns_info_body_offset = ibf_dump_insns_info_body(dump, iseq);
13462
13463 positions = rb_iseq_insns_info_decode_positions(ISEQ_BODY(iseq));
13464 const ibf_offset_t insns_info_positions_offset = ibf_dump_insns_info_positions(dump, positions, body->insns_info.size);
13465 ruby_xfree(positions);
13466
13467 const ibf_offset_t local_table_offset = ibf_dump_local_table(dump, iseq);
13468 const unsigned int catch_table_size = body->catch_table ? body->catch_table->size : 0;
13469 const ibf_offset_t catch_table_offset = ibf_dump_catch_table(dump, iseq);
13470 const int parent_iseq_index = ibf_dump_iseq(dump, ISEQ_BODY(iseq)->parent_iseq);
13471 const int local_iseq_index = ibf_dump_iseq(dump, ISEQ_BODY(iseq)->local_iseq);
13472 const int mandatory_only_iseq_index = ibf_dump_iseq(dump, ISEQ_BODY(iseq)->mandatory_only_iseq);
13473 const ibf_offset_t ci_entries_offset = ibf_dump_ci_entries(dump, iseq);
13474 const ibf_offset_t outer_variables_offset = ibf_dump_outer_variables(dump, iseq);
13475
13476#if IBF_ISEQ_ENABLE_LOCAL_BUFFER
13477 ibf_offset_t local_obj_list_offset;
13478 unsigned int local_obj_list_size;
13479
13480 ibf_dump_object_list(dump, &local_obj_list_offset, &local_obj_list_size);
13481#endif
13482
13483 ibf_offset_t body_offset = ibf_dump_pos(dump);
13484
13485 /* dump the constant body */
13486 unsigned int param_flags =
13487 (body->param.flags.has_lead << 0) |
13488 (body->param.flags.has_opt << 1) |
13489 (body->param.flags.has_rest << 2) |
13490 (body->param.flags.has_post << 3) |
13491 (body->param.flags.has_kw << 4) |
13492 (body->param.flags.has_kwrest << 5) |
13493 (body->param.flags.has_block << 6) |
13494 (body->param.flags.ambiguous_param0 << 7) |
13495 (body->param.flags.accepts_no_kwarg << 8) |
13496 (body->param.flags.ruby2_keywords << 9) |
13497 (body->param.flags.anon_rest << 10) |
13498 (body->param.flags.anon_kwrest << 11) |
13499 (body->param.flags.use_block << 12) |
13500 (body->param.flags.forwardable << 13) ;
13501
13502#if IBF_ISEQ_ENABLE_LOCAL_BUFFER
13503# define IBF_BODY_OFFSET(x) (x)
13504#else
13505# define IBF_BODY_OFFSET(x) (body_offset - (x))
13506#endif
13507
13508 ibf_dump_write_small_value(dump, body->type);
13509 ibf_dump_write_small_value(dump, body->iseq_size);
13510 ibf_dump_write_small_value(dump, IBF_BODY_OFFSET(bytecode_offset));
13511 ibf_dump_write_small_value(dump, bytecode_size);
13512 ibf_dump_write_small_value(dump, param_flags);
13513 ibf_dump_write_small_value(dump, body->param.size);
13514 ibf_dump_write_small_value(dump, body->param.lead_num);
13515 ibf_dump_write_small_value(dump, body->param.opt_num);
13516 ibf_dump_write_small_value(dump, body->param.rest_start);
13517 ibf_dump_write_small_value(dump, body->param.post_start);
13518 ibf_dump_write_small_value(dump, body->param.post_num);
13519 ibf_dump_write_small_value(dump, body->param.block_start);
13520 ibf_dump_write_small_value(dump, IBF_BODY_OFFSET(param_opt_table_offset));
13521 ibf_dump_write_small_value(dump, param_keyword_offset);
13522 ibf_dump_write_small_value(dump, location_pathobj_index);
13523 ibf_dump_write_small_value(dump, location_base_label_index);
13524 ibf_dump_write_small_value(dump, location_label_index);
13525 ibf_dump_write_small_value(dump, body->location.first_lineno);
13526 ibf_dump_write_small_value(dump, body->location.node_id);
13527 ibf_dump_write_small_value(dump, body->location.code_location.beg_pos.lineno);
13528 ibf_dump_write_small_value(dump, body->location.code_location.beg_pos.column);
13529 ibf_dump_write_small_value(dump, body->location.code_location.end_pos.lineno);
13530 ibf_dump_write_small_value(dump, body->location.code_location.end_pos.column);
13531 ibf_dump_write_small_value(dump, IBF_BODY_OFFSET(insns_info_body_offset));
13532 ibf_dump_write_small_value(dump, IBF_BODY_OFFSET(insns_info_positions_offset));
13533 ibf_dump_write_small_value(dump, body->insns_info.size);
13534 ibf_dump_write_small_value(dump, IBF_BODY_OFFSET(local_table_offset));
13535 ibf_dump_write_small_value(dump, catch_table_size);
13536 ibf_dump_write_small_value(dump, IBF_BODY_OFFSET(catch_table_offset));
13537 ibf_dump_write_small_value(dump, parent_iseq_index);
13538 ibf_dump_write_small_value(dump, local_iseq_index);
13539 ibf_dump_write_small_value(dump, mandatory_only_iseq_index);
13540 ibf_dump_write_small_value(dump, IBF_BODY_OFFSET(ci_entries_offset));
13541 ibf_dump_write_small_value(dump, IBF_BODY_OFFSET(outer_variables_offset));
13542 ibf_dump_write_small_value(dump, body->variable.flip_count);
13543 ibf_dump_write_small_value(dump, body->local_table_size);
13544 ibf_dump_write_small_value(dump, body->ivc_size);
13545 ibf_dump_write_small_value(dump, body->icvarc_size);
13546 ibf_dump_write_small_value(dump, body->ise_size);
13547 ibf_dump_write_small_value(dump, body->ic_size);
13548 ibf_dump_write_small_value(dump, body->ci_size);
13549 ibf_dump_write_small_value(dump, body->stack_max);
13550 ibf_dump_write_small_value(dump, body->builtin_attrs);
13551 ibf_dump_write_small_value(dump, body->prism ? 1 : 0);
13552
13553#undef IBF_BODY_OFFSET
13554
13555#if IBF_ISEQ_ENABLE_LOCAL_BUFFER
13556 ibf_offset_t iseq_length_bytes = ibf_dump_pos(dump);
13557
13558 dump->current_buffer = saved_buffer;
13559 ibf_dump_write(dump, RSTRING_PTR(buffer.str), iseq_length_bytes);
13560
13561 ibf_offset_t offset = ibf_dump_pos(dump);
13562 ibf_dump_write_small_value(dump, iseq_start);
13563 ibf_dump_write_small_value(dump, iseq_length_bytes);
13564 ibf_dump_write_small_value(dump, body_offset);
13565
13566 ibf_dump_write_small_value(dump, local_obj_list_offset);
13567 ibf_dump_write_small_value(dump, local_obj_list_size);
13568
13569 st_free_table(buffer.obj_table); // TODO: this leaks in case of exception
13570
13571 return offset;
13572#else
13573 return body_offset;
13574#endif
13575}
13576
13577static VALUE
13578ibf_load_location_str(const struct ibf_load *load, VALUE str_index)
13579{
13580 VALUE str = ibf_load_object(load, str_index);
13581 if (str != Qnil) {
13582 str = rb_fstring(str);
13583 }
13584 return str;
13585}
13586
13587static void
13588ibf_load_iseq_each(struct ibf_load *load, rb_iseq_t *iseq, ibf_offset_t offset)
13589{
13590 struct rb_iseq_constant_body *load_body = ISEQ_BODY(iseq) = rb_iseq_constant_body_alloc();
13591
13592 ibf_offset_t reading_pos = offset;
13593
13594#if IBF_ISEQ_ENABLE_LOCAL_BUFFER
13595 struct ibf_load_buffer *saved_buffer = load->current_buffer;
13596 load->current_buffer = &load->global_buffer;
13597
13598 const ibf_offset_t iseq_start = (ibf_offset_t)ibf_load_small_value(load, &reading_pos);
13599 const ibf_offset_t iseq_length_bytes = (ibf_offset_t)ibf_load_small_value(load, &reading_pos);
13600 const ibf_offset_t body_offset = (ibf_offset_t)ibf_load_small_value(load, &reading_pos);
13601
13602 struct ibf_load_buffer buffer;
13603 buffer.buff = load->global_buffer.buff + iseq_start;
13604 buffer.size = iseq_length_bytes;
13605 buffer.obj_list_offset = (ibf_offset_t)ibf_load_small_value(load, &reading_pos);
13606 buffer.obj_list_size = (ibf_offset_t)ibf_load_small_value(load, &reading_pos);
13607 buffer.obj_list = pinned_list_new(buffer.obj_list_size);
13608
13609 load->current_buffer = &buffer;
13610 reading_pos = body_offset;
13611#endif
13612
13613#if IBF_ISEQ_ENABLE_LOCAL_BUFFER
13614# define IBF_BODY_OFFSET(x) (x)
13615#else
13616# define IBF_BODY_OFFSET(x) (offset - (x))
13617#endif
13618
13619 const unsigned int type = (unsigned int)ibf_load_small_value(load, &reading_pos);
13620 const unsigned int iseq_size = (unsigned int)ibf_load_small_value(load, &reading_pos);
13621 const ibf_offset_t bytecode_offset = (ibf_offset_t)IBF_BODY_OFFSET(ibf_load_small_value(load, &reading_pos));
13622 const ibf_offset_t bytecode_size = (ibf_offset_t)ibf_load_small_value(load, &reading_pos);
13623 const unsigned int param_flags = (unsigned int)ibf_load_small_value(load, &reading_pos);
13624 const unsigned int param_size = (unsigned int)ibf_load_small_value(load, &reading_pos);
13625 const int param_lead_num = (int)ibf_load_small_value(load, &reading_pos);
13626 const int param_opt_num = (int)ibf_load_small_value(load, &reading_pos);
13627 const int param_rest_start = (int)ibf_load_small_value(load, &reading_pos);
13628 const int param_post_start = (int)ibf_load_small_value(load, &reading_pos);
13629 const int param_post_num = (int)ibf_load_small_value(load, &reading_pos);
13630 const int param_block_start = (int)ibf_load_small_value(load, &reading_pos);
13631 const ibf_offset_t param_opt_table_offset = (ibf_offset_t)IBF_BODY_OFFSET(ibf_load_small_value(load, &reading_pos));
13632 const ibf_offset_t param_keyword_offset = (ibf_offset_t)ibf_load_small_value(load, &reading_pos);
13633 const VALUE location_pathobj_index = ibf_load_small_value(load, &reading_pos);
13634 const VALUE location_base_label_index = ibf_load_small_value(load, &reading_pos);
13635 const VALUE location_label_index = ibf_load_small_value(load, &reading_pos);
13636 const int location_first_lineno = (int)ibf_load_small_value(load, &reading_pos);
13637 const int location_node_id = (int)ibf_load_small_value(load, &reading_pos);
13638 const int location_code_location_beg_pos_lineno = (int)ibf_load_small_value(load, &reading_pos);
13639 const int location_code_location_beg_pos_column = (int)ibf_load_small_value(load, &reading_pos);
13640 const int location_code_location_end_pos_lineno = (int)ibf_load_small_value(load, &reading_pos);
13641 const int location_code_location_end_pos_column = (int)ibf_load_small_value(load, &reading_pos);
13642 const ibf_offset_t insns_info_body_offset = (ibf_offset_t)IBF_BODY_OFFSET(ibf_load_small_value(load, &reading_pos));
13643 const ibf_offset_t insns_info_positions_offset = (ibf_offset_t)IBF_BODY_OFFSET(ibf_load_small_value(load, &reading_pos));
13644 const unsigned int insns_info_size = (unsigned int)ibf_load_small_value(load, &reading_pos);
13645 const ibf_offset_t local_table_offset = (ibf_offset_t)IBF_BODY_OFFSET(ibf_load_small_value(load, &reading_pos));
13646 const unsigned int catch_table_size = (unsigned int)ibf_load_small_value(load, &reading_pos);
13647 const ibf_offset_t catch_table_offset = (ibf_offset_t)IBF_BODY_OFFSET(ibf_load_small_value(load, &reading_pos));
13648 const int parent_iseq_index = (int)ibf_load_small_value(load, &reading_pos);
13649 const int local_iseq_index = (int)ibf_load_small_value(load, &reading_pos);
13650 const int mandatory_only_iseq_index = (int)ibf_load_small_value(load, &reading_pos);
13651 const ibf_offset_t ci_entries_offset = (ibf_offset_t)IBF_BODY_OFFSET(ibf_load_small_value(load, &reading_pos));
13652 const ibf_offset_t outer_variables_offset = (ibf_offset_t)IBF_BODY_OFFSET(ibf_load_small_value(load, &reading_pos));
13653 const rb_snum_t variable_flip_count = (rb_snum_t)ibf_load_small_value(load, &reading_pos);
13654 const unsigned int local_table_size = (unsigned int)ibf_load_small_value(load, &reading_pos);
13655
13656 const unsigned int ivc_size = (unsigned int)ibf_load_small_value(load, &reading_pos);
13657 const unsigned int icvarc_size = (unsigned int)ibf_load_small_value(load, &reading_pos);
13658 const unsigned int ise_size = (unsigned int)ibf_load_small_value(load, &reading_pos);
13659 const unsigned int ic_size = (unsigned int)ibf_load_small_value(load, &reading_pos);
13660
13661 const unsigned int ci_size = (unsigned int)ibf_load_small_value(load, &reading_pos);
13662 const unsigned int stack_max = (unsigned int)ibf_load_small_value(load, &reading_pos);
13663 const unsigned int builtin_attrs = (unsigned int)ibf_load_small_value(load, &reading_pos);
13664 const bool prism = (bool)ibf_load_small_value(load, &reading_pos);
13665
13666 // setup fname and dummy frame
13667 VALUE path = ibf_load_object(load, location_pathobj_index);
13668 {
13669 VALUE realpath = Qnil;
13670
13671 if (RB_TYPE_P(path, T_STRING)) {
13672 realpath = path = rb_fstring(path);
13673 }
13674 else if (RB_TYPE_P(path, T_ARRAY)) {
13675 VALUE pathobj = path;
13676 if (RARRAY_LEN(pathobj) != 2) {
13677 rb_raise(rb_eRuntimeError, "path object size mismatch");
13678 }
13679 path = rb_fstring(RARRAY_AREF(pathobj, 0));
13680 realpath = RARRAY_AREF(pathobj, 1);
13681 if (!NIL_P(realpath)) {
13682 if (!RB_TYPE_P(realpath, T_STRING)) {
13683 rb_raise(rb_eArgError, "unexpected realpath %"PRIxVALUE
13684 "(%x), path=%+"PRIsVALUE,
13685 realpath, TYPE(realpath), path);
13686 }
13687 realpath = rb_fstring(realpath);
13688 }
13689 }
13690 else {
13691 rb_raise(rb_eRuntimeError, "unexpected path object");
13692 }
13693 rb_iseq_pathobj_set(iseq, path, realpath);
13694 }
13695
13696 // push dummy frame
13697 rb_execution_context_t *ec = GET_EC();
13698 VALUE dummy_frame = rb_vm_push_frame_fname(ec, path);
13699
13700#undef IBF_BODY_OFFSET
13701
13702 load_body->type = type;
13703 load_body->stack_max = stack_max;
13704 load_body->param.flags.has_lead = (param_flags >> 0) & 1;
13705 load_body->param.flags.has_opt = (param_flags >> 1) & 1;
13706 load_body->param.flags.has_rest = (param_flags >> 2) & 1;
13707 load_body->param.flags.has_post = (param_flags >> 3) & 1;
13708 load_body->param.flags.has_kw = FALSE;
13709 load_body->param.flags.has_kwrest = (param_flags >> 5) & 1;
13710 load_body->param.flags.has_block = (param_flags >> 6) & 1;
13711 load_body->param.flags.ambiguous_param0 = (param_flags >> 7) & 1;
13712 load_body->param.flags.accepts_no_kwarg = (param_flags >> 8) & 1;
13713 load_body->param.flags.ruby2_keywords = (param_flags >> 9) & 1;
13714 load_body->param.flags.anon_rest = (param_flags >> 10) & 1;
13715 load_body->param.flags.anon_kwrest = (param_flags >> 11) & 1;
13716 load_body->param.flags.use_block = (param_flags >> 12) & 1;
13717 load_body->param.flags.forwardable = (param_flags >> 13) & 1;
13718 load_body->param.size = param_size;
13719 load_body->param.lead_num = param_lead_num;
13720 load_body->param.opt_num = param_opt_num;
13721 load_body->param.rest_start = param_rest_start;
13722 load_body->param.post_start = param_post_start;
13723 load_body->param.post_num = param_post_num;
13724 load_body->param.block_start = param_block_start;
13725 load_body->local_table_size = local_table_size;
13726 load_body->ci_size = ci_size;
13727 load_body->insns_info.size = insns_info_size;
13728
13729 ISEQ_COVERAGE_SET(iseq, Qnil);
13730 ISEQ_ORIGINAL_ISEQ_CLEAR(iseq);
13731 load_body->variable.flip_count = variable_flip_count;
13732 load_body->variable.script_lines = Qnil;
13733
13734 load_body->location.first_lineno = location_first_lineno;
13735 load_body->location.node_id = location_node_id;
13736 load_body->location.code_location.beg_pos.lineno = location_code_location_beg_pos_lineno;
13737 load_body->location.code_location.beg_pos.column = location_code_location_beg_pos_column;
13738 load_body->location.code_location.end_pos.lineno = location_code_location_end_pos_lineno;
13739 load_body->location.code_location.end_pos.column = location_code_location_end_pos_column;
13740 load_body->builtin_attrs = builtin_attrs;
13741 load_body->prism = prism;
13742
13743 load_body->ivc_size = ivc_size;
13744 load_body->icvarc_size = icvarc_size;
13745 load_body->ise_size = ise_size;
13746 load_body->ic_size = ic_size;
13747
13748 if (ISEQ_IS_SIZE(load_body)) {
13749 load_body->is_entries = ZALLOC_N(union iseq_inline_storage_entry, ISEQ_IS_SIZE(load_body));
13750 }
13751 else {
13752 load_body->is_entries = NULL;
13753 }
13754 ibf_load_ci_entries(load, ci_entries_offset, ci_size, &load_body->call_data);
13755 load_body->outer_variables = ibf_load_outer_variables(load, outer_variables_offset);
13756 load_body->param.opt_table = ibf_load_param_opt_table(load, param_opt_table_offset, param_opt_num);
13757 load_body->param.keyword = ibf_load_param_keyword(load, param_keyword_offset);
13758 load_body->param.flags.has_kw = (param_flags >> 4) & 1;
13759 load_body->insns_info.body = ibf_load_insns_info_body(load, insns_info_body_offset, insns_info_size);
13760 load_body->insns_info.positions = ibf_load_insns_info_positions(load, insns_info_positions_offset, insns_info_size);
13761 load_body->local_table = ibf_load_local_table(load, local_table_offset, local_table_size);
13762 load_body->catch_table = ibf_load_catch_table(load, catch_table_offset, catch_table_size);
13763 load_body->parent_iseq = ibf_load_iseq(load, (const rb_iseq_t *)(VALUE)parent_iseq_index);
13764 load_body->local_iseq = ibf_load_iseq(load, (const rb_iseq_t *)(VALUE)local_iseq_index);
13765 load_body->mandatory_only_iseq = ibf_load_iseq(load, (const rb_iseq_t *)(VALUE)mandatory_only_iseq_index);
13766
13767 // This must be done after the local table is loaded.
13768 if (load_body->param.keyword != NULL) {
13769 RUBY_ASSERT(load_body->local_table);
13770 struct rb_iseq_param_keyword *keyword = (struct rb_iseq_param_keyword *) load_body->param.keyword;
13771 keyword->table = &load_body->local_table[keyword->bits_start - keyword->num];
13772 }
13773
13774 ibf_load_code(load, iseq, bytecode_offset, bytecode_size, iseq_size);
13775#if VM_INSN_INFO_TABLE_IMPL == 2
13776 rb_iseq_insns_info_encode_positions(iseq);
13777#endif
13778
13779 rb_iseq_translate_threaded_code(iseq);
13780
13781#if IBF_ISEQ_ENABLE_LOCAL_BUFFER
13782 load->current_buffer = &load->global_buffer;
13783#endif
13784
13785 RB_OBJ_WRITE(iseq, &load_body->location.base_label, ibf_load_location_str(load, location_base_label_index));
13786 RB_OBJ_WRITE(iseq, &load_body->location.label, ibf_load_location_str(load, location_label_index));
13787
13788#if IBF_ISEQ_ENABLE_LOCAL_BUFFER
13789 load->current_buffer = saved_buffer;
13790#endif
13791 verify_call_cache(iseq);
13792
13793 RB_GC_GUARD(dummy_frame);
13794 rb_vm_pop_frame_no_int(ec);
13795}
13796
13798{
13799 struct ibf_dump *dump;
13800 VALUE offset_list;
13801};
13802
13803static int
13804ibf_dump_iseq_list_i(st_data_t key, st_data_t val, st_data_t ptr)
13805{
13806 const rb_iseq_t *iseq = (const rb_iseq_t *)key;
13807 struct ibf_dump_iseq_list_arg *args = (struct ibf_dump_iseq_list_arg *)ptr;
13808
13809 ibf_offset_t offset = ibf_dump_iseq_each(args->dump, iseq);
13810 rb_ary_push(args->offset_list, UINT2NUM(offset));
13811
13812 return ST_CONTINUE;
13813}
13814
13815static void
13816ibf_dump_iseq_list(struct ibf_dump *dump, struct ibf_header *header)
13817{
13818 VALUE offset_list = rb_ary_hidden_new(dump->iseq_table->num_entries);
13819
13820 struct ibf_dump_iseq_list_arg args;
13821 args.dump = dump;
13822 args.offset_list = offset_list;
13823
13824 st_foreach(dump->iseq_table, ibf_dump_iseq_list_i, (st_data_t)&args);
13825
13826 st_index_t i;
13827 st_index_t size = dump->iseq_table->num_entries;
13828 ibf_offset_t *offsets = ALLOCA_N(ibf_offset_t, size);
13829
13830 for (i = 0; i < size; i++) {
13831 offsets[i] = NUM2UINT(RARRAY_AREF(offset_list, i));
13832 }
13833
13834 ibf_dump_align(dump, sizeof(ibf_offset_t));
13835 header->iseq_list_offset = ibf_dump_write(dump, offsets, sizeof(ibf_offset_t) * size);
13836 header->iseq_list_size = (unsigned int)size;
13837}
13838
13839#define IBF_OBJECT_INTERNAL FL_PROMOTED0
13840
13841/*
13842 * Binary format
13843 * - ibf_object_header
13844 * - ibf_object_xxx (xxx is type)
13845 */
13846
13848 unsigned int type: 5;
13849 unsigned int special_const: 1;
13850 unsigned int frozen: 1;
13851 unsigned int internal: 1;
13852};
13853
13854enum ibf_object_class_index {
13855 IBF_OBJECT_CLASS_OBJECT,
13856 IBF_OBJECT_CLASS_ARRAY,
13857 IBF_OBJECT_CLASS_STANDARD_ERROR,
13858 IBF_OBJECT_CLASS_NO_MATCHING_PATTERN_ERROR,
13859 IBF_OBJECT_CLASS_TYPE_ERROR,
13860 IBF_OBJECT_CLASS_NO_MATCHING_PATTERN_KEY_ERROR,
13861};
13862
13864 long srcstr;
13865 char option;
13866};
13867
13869 long len;
13870 long keyval[FLEX_ARY_LEN];
13871};
13872
13874 long class_index;
13875 long len;
13876 long beg;
13877 long end;
13878 int excl;
13879};
13880
13882 ssize_t slen;
13883 BDIGIT digits[FLEX_ARY_LEN];
13884};
13885
13886enum ibf_object_data_type {
13887 IBF_OBJECT_DATA_ENCODING,
13888};
13889
13891 long a, b;
13892};
13893
13895 long str;
13896};
13897
13898#define IBF_ALIGNED_OFFSET(align, offset) /* offset > 0 */ \
13899 ((((offset) - 1) / (align) + 1) * (align))
13900#define IBF_OBJBODY(type, offset) (const type *)\
13901 ibf_load_check_offset(load, IBF_ALIGNED_OFFSET(RUBY_ALIGNOF(type), offset))
13902
13903static const void *
13904ibf_load_check_offset(const struct ibf_load *load, size_t offset)
13905{
13906 if (offset >= load->current_buffer->size) {
13907 rb_raise(rb_eIndexError, "object offset out of range: %"PRIdSIZE, offset);
13908 }
13909 return load->current_buffer->buff + offset;
13910}
13911
13912NORETURN(static void ibf_dump_object_unsupported(struct ibf_dump *dump, VALUE obj));
13913
13914static void
13915ibf_dump_object_unsupported(struct ibf_dump *dump, VALUE obj)
13916{
13917 char buff[0x100];
13918 rb_raw_obj_info(buff, sizeof(buff), obj);
13919 rb_raise(rb_eNotImpError, "ibf_dump_object_unsupported: %s", buff);
13920}
13921
13922NORETURN(static VALUE ibf_load_object_unsupported(const struct ibf_load *load, const struct ibf_object_header *header, ibf_offset_t offset));
13923
13924static VALUE
13925ibf_load_object_unsupported(const struct ibf_load *load, const struct ibf_object_header *header, ibf_offset_t offset)
13926{
13927 rb_raise(rb_eArgError, "unsupported");
13929}
13930
13931static void
13932ibf_dump_object_class(struct ibf_dump *dump, VALUE obj)
13933{
13934 enum ibf_object_class_index cindex;
13935 if (obj == rb_cObject) {
13936 cindex = IBF_OBJECT_CLASS_OBJECT;
13937 }
13938 else if (obj == rb_cArray) {
13939 cindex = IBF_OBJECT_CLASS_ARRAY;
13940 }
13941 else if (obj == rb_eStandardError) {
13942 cindex = IBF_OBJECT_CLASS_STANDARD_ERROR;
13943 }
13944 else if (obj == rb_eNoMatchingPatternError) {
13945 cindex = IBF_OBJECT_CLASS_NO_MATCHING_PATTERN_ERROR;
13946 }
13947 else if (obj == rb_eTypeError) {
13948 cindex = IBF_OBJECT_CLASS_TYPE_ERROR;
13949 }
13950 else if (obj == rb_eNoMatchingPatternKeyError) {
13951 cindex = IBF_OBJECT_CLASS_NO_MATCHING_PATTERN_KEY_ERROR;
13952 }
13953 else {
13954 rb_obj_info_dump(obj);
13955 rb_p(obj);
13956 rb_bug("unsupported class");
13957 }
13958 ibf_dump_write_small_value(dump, (VALUE)cindex);
13959}
13960
13961static VALUE
13962ibf_load_object_class(const struct ibf_load *load, const struct ibf_object_header *header, ibf_offset_t offset)
13963{
13964 enum ibf_object_class_index cindex = (enum ibf_object_class_index)ibf_load_small_value(load, &offset);
13965
13966 switch (cindex) {
13967 case IBF_OBJECT_CLASS_OBJECT:
13968 return rb_cObject;
13969 case IBF_OBJECT_CLASS_ARRAY:
13970 return rb_cArray;
13971 case IBF_OBJECT_CLASS_STANDARD_ERROR:
13972 return rb_eStandardError;
13973 case IBF_OBJECT_CLASS_NO_MATCHING_PATTERN_ERROR:
13975 case IBF_OBJECT_CLASS_TYPE_ERROR:
13976 return rb_eTypeError;
13977 case IBF_OBJECT_CLASS_NO_MATCHING_PATTERN_KEY_ERROR:
13979 }
13980
13981 rb_raise(rb_eArgError, "ibf_load_object_class: unknown class (%d)", (int)cindex);
13982}
13983
13984
13985static void
13986ibf_dump_object_float(struct ibf_dump *dump, VALUE obj)
13987{
13988 double dbl = RFLOAT_VALUE(obj);
13989 (void)IBF_W(&dbl, double, 1);
13990}
13991
13992static VALUE
13993ibf_load_object_float(const struct ibf_load *load, const struct ibf_object_header *header, ibf_offset_t offset)
13994{
13995 const double *dblp = IBF_OBJBODY(double, offset);
13996 return DBL2NUM(*dblp);
13997}
13998
13999static void
14000ibf_dump_object_string(struct ibf_dump *dump, VALUE obj)
14001{
14002 long encindex = (long)rb_enc_get_index(obj);
14003 long len = RSTRING_LEN(obj);
14004 const char *ptr = RSTRING_PTR(obj);
14005
14006 if (encindex > RUBY_ENCINDEX_BUILTIN_MAX) {
14007 rb_encoding *enc = rb_enc_from_index((int)encindex);
14008 const char *enc_name = rb_enc_name(enc);
14009 encindex = RUBY_ENCINDEX_BUILTIN_MAX + ibf_dump_object(dump, rb_str_new2(enc_name));
14010 }
14011
14012 ibf_dump_write_small_value(dump, encindex);
14013 ibf_dump_write_small_value(dump, len);
14014 IBF_WP(ptr, char, len);
14015}
14016
14017static VALUE
14018ibf_load_object_string(const struct ibf_load *load, const struct ibf_object_header *header, ibf_offset_t offset)
14019{
14020 ibf_offset_t reading_pos = offset;
14021
14022 int encindex = (int)ibf_load_small_value(load, &reading_pos);
14023 const long len = (long)ibf_load_small_value(load, &reading_pos);
14024 const char *ptr = load->current_buffer->buff + reading_pos;
14025
14026 if (encindex > RUBY_ENCINDEX_BUILTIN_MAX) {
14027 VALUE enc_name_str = ibf_load_object(load, encindex - RUBY_ENCINDEX_BUILTIN_MAX);
14028 encindex = rb_enc_find_index(RSTRING_PTR(enc_name_str));
14029 }
14030
14031 VALUE str;
14032 if (header->frozen && !header->internal) {
14033 str = rb_enc_literal_str(ptr, len, rb_enc_from_index(encindex));
14034 }
14035 else {
14036 str = rb_enc_str_new(ptr, len, rb_enc_from_index(encindex));
14037
14038 if (header->internal) rb_obj_hide(str);
14039 if (header->frozen) str = rb_fstring(str);
14040 }
14041 return str;
14042}
14043
14044static void
14045ibf_dump_object_regexp(struct ibf_dump *dump, VALUE obj)
14046{
14047 VALUE srcstr = RREGEXP_SRC(obj);
14048 struct ibf_object_regexp regexp;
14049 regexp.option = (char)rb_reg_options(obj);
14050 regexp.srcstr = (long)ibf_dump_object(dump, srcstr);
14051
14052 ibf_dump_write_byte(dump, (unsigned char)regexp.option);
14053 ibf_dump_write_small_value(dump, regexp.srcstr);
14054}
14055
14056static VALUE
14057ibf_load_object_regexp(const struct ibf_load *load, const struct ibf_object_header *header, ibf_offset_t offset)
14058{
14059 struct ibf_object_regexp regexp;
14060 regexp.option = ibf_load_byte(load, &offset);
14061 regexp.srcstr = ibf_load_small_value(load, &offset);
14062
14063 VALUE srcstr = ibf_load_object(load, regexp.srcstr);
14064 VALUE reg = rb_reg_compile(srcstr, (int)regexp.option, NULL, 0);
14065
14066 if (header->internal) rb_obj_hide(reg);
14067 if (header->frozen) rb_obj_freeze(reg);
14068
14069 return reg;
14070}
14071
14072static void
14073ibf_dump_object_array(struct ibf_dump *dump, VALUE obj)
14074{
14075 long i, len = RARRAY_LEN(obj);
14076 ibf_dump_write_small_value(dump, len);
14077 for (i=0; i<len; i++) {
14078 long index = (long)ibf_dump_object(dump, RARRAY_AREF(obj, i));
14079 ibf_dump_write_small_value(dump, index);
14080 }
14081}
14082
14083static VALUE
14084ibf_load_object_array(const struct ibf_load *load, const struct ibf_object_header *header, ibf_offset_t offset)
14085{
14086 ibf_offset_t reading_pos = offset;
14087
14088 const long len = (long)ibf_load_small_value(load, &reading_pos);
14089
14090 VALUE ary = header->internal ? rb_ary_hidden_new(len) : rb_ary_new_capa(len);
14091 int i;
14092
14093 for (i=0; i<len; i++) {
14094 const VALUE index = ibf_load_small_value(load, &reading_pos);
14095 rb_ary_push(ary, ibf_load_object(load, index));
14096 }
14097
14098 if (header->frozen) rb_ary_freeze(ary);
14099
14100 return ary;
14101}
14102
14103static int
14104ibf_dump_object_hash_i(st_data_t key, st_data_t val, st_data_t ptr)
14105{
14106 struct ibf_dump *dump = (struct ibf_dump *)ptr;
14107
14108 VALUE key_index = ibf_dump_object(dump, (VALUE)key);
14109 VALUE val_index = ibf_dump_object(dump, (VALUE)val);
14110
14111 ibf_dump_write_small_value(dump, key_index);
14112 ibf_dump_write_small_value(dump, val_index);
14113 return ST_CONTINUE;
14114}
14115
14116static void
14117ibf_dump_object_hash(struct ibf_dump *dump, VALUE obj)
14118{
14119 long len = RHASH_SIZE(obj);
14120 ibf_dump_write_small_value(dump, (VALUE)len);
14121
14122 if (len > 0) rb_hash_foreach(obj, ibf_dump_object_hash_i, (VALUE)dump);
14123}
14124
14125static VALUE
14126ibf_load_object_hash(const struct ibf_load *load, const struct ibf_object_header *header, ibf_offset_t offset)
14127{
14128 long len = (long)ibf_load_small_value(load, &offset);
14129 VALUE obj = rb_hash_new_with_size(len);
14130 int i;
14131
14132 for (i = 0; i < len; i++) {
14133 VALUE key_index = ibf_load_small_value(load, &offset);
14134 VALUE val_index = ibf_load_small_value(load, &offset);
14135
14136 VALUE key = ibf_load_object(load, key_index);
14137 VALUE val = ibf_load_object(load, val_index);
14138 rb_hash_aset(obj, key, val);
14139 }
14140 rb_hash_rehash(obj);
14141
14142 if (header->internal) rb_obj_hide(obj);
14143 if (header->frozen) rb_obj_freeze(obj);
14144
14145 return obj;
14146}
14147
14148static void
14149ibf_dump_object_struct(struct ibf_dump *dump, VALUE obj)
14150{
14151 if (rb_obj_is_kind_of(obj, rb_cRange)) {
14152 struct ibf_object_struct_range range;
14153 VALUE beg, end;
14154 IBF_ZERO(range);
14155 range.len = 3;
14156 range.class_index = 0;
14157
14158 rb_range_values(obj, &beg, &end, &range.excl);
14159 range.beg = (long)ibf_dump_object(dump, beg);
14160 range.end = (long)ibf_dump_object(dump, end);
14161
14162 IBF_W_ALIGN(struct ibf_object_struct_range);
14163 IBF_WV(range);
14164 }
14165 else {
14166 rb_raise(rb_eNotImpError, "ibf_dump_object_struct: unsupported class %"PRIsVALUE,
14167 rb_class_name(CLASS_OF(obj)));
14168 }
14169}
14170
14171static VALUE
14172ibf_load_object_struct(const struct ibf_load *load, const struct ibf_object_header *header, ibf_offset_t offset)
14173{
14174 const struct ibf_object_struct_range *range = IBF_OBJBODY(struct ibf_object_struct_range, offset);
14175 VALUE beg = ibf_load_object(load, range->beg);
14176 VALUE end = ibf_load_object(load, range->end);
14177 VALUE obj = rb_range_new(beg, end, range->excl);
14178 if (header->internal) rb_obj_hide(obj);
14179 if (header->frozen) rb_obj_freeze(obj);
14180 return obj;
14181}
14182
14183static void
14184ibf_dump_object_bignum(struct ibf_dump *dump, VALUE obj)
14185{
14186 ssize_t len = BIGNUM_LEN(obj);
14187 ssize_t slen = BIGNUM_SIGN(obj) > 0 ? len : len * -1;
14188 BDIGIT *d = BIGNUM_DIGITS(obj);
14189
14190 (void)IBF_W(&slen, ssize_t, 1);
14191 IBF_WP(d, BDIGIT, len);
14192}
14193
14194static VALUE
14195ibf_load_object_bignum(const struct ibf_load *load, const struct ibf_object_header *header, ibf_offset_t offset)
14196{
14197 const struct ibf_object_bignum *bignum = IBF_OBJBODY(struct ibf_object_bignum, offset);
14198 int sign = bignum->slen > 0;
14199 ssize_t len = sign > 0 ? bignum->slen : -1 * bignum->slen;
14200 const int big_unpack_flags = /* c.f. rb_big_unpack() */
14203 VALUE obj = rb_integer_unpack(bignum->digits, len, sizeof(BDIGIT), 0,
14204 big_unpack_flags |
14205 (sign == 0 ? INTEGER_PACK_NEGATIVE : 0));
14206 if (header->internal) rb_obj_hide(obj);
14207 if (header->frozen) rb_obj_freeze(obj);
14208 return obj;
14209}
14210
14211static void
14212ibf_dump_object_data(struct ibf_dump *dump, VALUE obj)
14213{
14214 if (rb_data_is_encoding(obj)) {
14215 rb_encoding *enc = rb_to_encoding(obj);
14216 const char *name = rb_enc_name(enc);
14217 long len = strlen(name) + 1;
14218 long data[2];
14219 data[0] = IBF_OBJECT_DATA_ENCODING;
14220 data[1] = len;
14221 (void)IBF_W(data, long, 2);
14222 IBF_WP(name, char, len);
14223 }
14224 else {
14225 ibf_dump_object_unsupported(dump, obj);
14226 }
14227}
14228
14229static VALUE
14230ibf_load_object_data(const struct ibf_load *load, const struct ibf_object_header *header, ibf_offset_t offset)
14231{
14232 const long *body = IBF_OBJBODY(long, offset);
14233 const enum ibf_object_data_type type = (enum ibf_object_data_type)body[0];
14234 /* const long len = body[1]; */
14235 const char *data = (const char *)&body[2];
14236
14237 switch (type) {
14238 case IBF_OBJECT_DATA_ENCODING:
14239 {
14240 VALUE encobj = rb_enc_from_encoding(rb_enc_find(data));
14241 return encobj;
14242 }
14243 }
14244
14245 return ibf_load_object_unsupported(load, header, offset);
14246}
14247
14248static void
14249ibf_dump_object_complex_rational(struct ibf_dump *dump, VALUE obj)
14250{
14251 long data[2];
14252 data[0] = (long)ibf_dump_object(dump, RCOMPLEX(obj)->real);
14253 data[1] = (long)ibf_dump_object(dump, RCOMPLEX(obj)->imag);
14254
14255 (void)IBF_W(data, long, 2);
14256}
14257
14258static VALUE
14259ibf_load_object_complex_rational(const struct ibf_load *load, const struct ibf_object_header *header, ibf_offset_t offset)
14260{
14261 const struct ibf_object_complex_rational *nums = IBF_OBJBODY(struct ibf_object_complex_rational, offset);
14262 VALUE a = ibf_load_object(load, nums->a);
14263 VALUE b = ibf_load_object(load, nums->b);
14264 VALUE obj = header->type == T_COMPLEX ?
14265 rb_complex_new(a, b) : rb_rational_new(a, b);
14266
14267 if (header->internal) rb_obj_hide(obj);
14268 if (header->frozen) rb_obj_freeze(obj);
14269 return obj;
14270}
14271
14272static void
14273ibf_dump_object_symbol(struct ibf_dump *dump, VALUE obj)
14274{
14275 ibf_dump_object_string(dump, rb_sym2str(obj));
14276}
14277
14278static VALUE
14279ibf_load_object_symbol(const struct ibf_load *load, const struct ibf_object_header *header, ibf_offset_t offset)
14280{
14281 ibf_offset_t reading_pos = offset;
14282
14283 int encindex = (int)ibf_load_small_value(load, &reading_pos);
14284 const long len = (long)ibf_load_small_value(load, &reading_pos);
14285 const char *ptr = load->current_buffer->buff + reading_pos;
14286
14287 if (encindex > RUBY_ENCINDEX_BUILTIN_MAX) {
14288 VALUE enc_name_str = ibf_load_object(load, encindex - RUBY_ENCINDEX_BUILTIN_MAX);
14289 encindex = rb_enc_find_index(RSTRING_PTR(enc_name_str));
14290 }
14291
14292 ID id = rb_intern3(ptr, len, rb_enc_from_index(encindex));
14293 return ID2SYM(id);
14294}
14295
14296typedef void (*ibf_dump_object_function)(struct ibf_dump *dump, VALUE obj);
14297static const ibf_dump_object_function dump_object_functions[RUBY_T_MASK+1] = {
14298 ibf_dump_object_unsupported, /* T_NONE */
14299 ibf_dump_object_unsupported, /* T_OBJECT */
14300 ibf_dump_object_class, /* T_CLASS */
14301 ibf_dump_object_unsupported, /* T_MODULE */
14302 ibf_dump_object_float, /* T_FLOAT */
14303 ibf_dump_object_string, /* T_STRING */
14304 ibf_dump_object_regexp, /* T_REGEXP */
14305 ibf_dump_object_array, /* T_ARRAY */
14306 ibf_dump_object_hash, /* T_HASH */
14307 ibf_dump_object_struct, /* T_STRUCT */
14308 ibf_dump_object_bignum, /* T_BIGNUM */
14309 ibf_dump_object_unsupported, /* T_FILE */
14310 ibf_dump_object_data, /* T_DATA */
14311 ibf_dump_object_unsupported, /* T_MATCH */
14312 ibf_dump_object_complex_rational, /* T_COMPLEX */
14313 ibf_dump_object_complex_rational, /* T_RATIONAL */
14314 ibf_dump_object_unsupported, /* 0x10 */
14315 ibf_dump_object_unsupported, /* 0x11 T_NIL */
14316 ibf_dump_object_unsupported, /* 0x12 T_TRUE */
14317 ibf_dump_object_unsupported, /* 0x13 T_FALSE */
14318 ibf_dump_object_symbol, /* 0x14 T_SYMBOL */
14319 ibf_dump_object_unsupported, /* T_FIXNUM */
14320 ibf_dump_object_unsupported, /* T_UNDEF */
14321 ibf_dump_object_unsupported, /* 0x17 */
14322 ibf_dump_object_unsupported, /* 0x18 */
14323 ibf_dump_object_unsupported, /* 0x19 */
14324 ibf_dump_object_unsupported, /* T_IMEMO 0x1a */
14325 ibf_dump_object_unsupported, /* T_NODE 0x1b */
14326 ibf_dump_object_unsupported, /* T_ICLASS 0x1c */
14327 ibf_dump_object_unsupported, /* T_ZOMBIE 0x1d */
14328 ibf_dump_object_unsupported, /* 0x1e */
14329 ibf_dump_object_unsupported, /* 0x1f */
14330};
14331
14332static void
14333ibf_dump_object_object_header(struct ibf_dump *dump, const struct ibf_object_header header)
14334{
14335 unsigned char byte =
14336 (header.type << 0) |
14337 (header.special_const << 5) |
14338 (header.frozen << 6) |
14339 (header.internal << 7);
14340
14341 IBF_WV(byte);
14342}
14343
14344static struct ibf_object_header
14345ibf_load_object_object_header(const struct ibf_load *load, ibf_offset_t *offset)
14346{
14347 unsigned char byte = ibf_load_byte(load, offset);
14348
14349 struct ibf_object_header header;
14350 header.type = (byte >> 0) & 0x1f;
14351 header.special_const = (byte >> 5) & 0x01;
14352 header.frozen = (byte >> 6) & 0x01;
14353 header.internal = (byte >> 7) & 0x01;
14354
14355 return header;
14356}
14357
14358static ibf_offset_t
14359ibf_dump_object_object(struct ibf_dump *dump, VALUE obj)
14360{
14361 struct ibf_object_header obj_header;
14362 ibf_offset_t current_offset;
14363 IBF_ZERO(obj_header);
14364 obj_header.type = TYPE(obj);
14365
14366 IBF_W_ALIGN(ibf_offset_t);
14367 current_offset = ibf_dump_pos(dump);
14368
14369 if (SPECIAL_CONST_P(obj) &&
14370 ! (SYMBOL_P(obj) ||
14371 RB_FLOAT_TYPE_P(obj))) {
14372 obj_header.special_const = TRUE;
14373 obj_header.frozen = TRUE;
14374 obj_header.internal = TRUE;
14375 ibf_dump_object_object_header(dump, obj_header);
14376 ibf_dump_write_small_value(dump, obj);
14377 }
14378 else {
14379 obj_header.internal = SPECIAL_CONST_P(obj) ? FALSE : (RBASIC_CLASS(obj) == 0) ? TRUE : FALSE;
14380 obj_header.special_const = FALSE;
14381 obj_header.frozen = FL_TEST(obj, FL_FREEZE) ? TRUE : FALSE;
14382 ibf_dump_object_object_header(dump, obj_header);
14383 (*dump_object_functions[obj_header.type])(dump, obj);
14384 }
14385
14386 return current_offset;
14387}
14388
14389typedef VALUE (*ibf_load_object_function)(const struct ibf_load *load, const struct ibf_object_header *header, ibf_offset_t offset);
14390static const ibf_load_object_function load_object_functions[RUBY_T_MASK+1] = {
14391 ibf_load_object_unsupported, /* T_NONE */
14392 ibf_load_object_unsupported, /* T_OBJECT */
14393 ibf_load_object_class, /* T_CLASS */
14394 ibf_load_object_unsupported, /* T_MODULE */
14395 ibf_load_object_float, /* T_FLOAT */
14396 ibf_load_object_string, /* T_STRING */
14397 ibf_load_object_regexp, /* T_REGEXP */
14398 ibf_load_object_array, /* T_ARRAY */
14399 ibf_load_object_hash, /* T_HASH */
14400 ibf_load_object_struct, /* T_STRUCT */
14401 ibf_load_object_bignum, /* T_BIGNUM */
14402 ibf_load_object_unsupported, /* T_FILE */
14403 ibf_load_object_data, /* T_DATA */
14404 ibf_load_object_unsupported, /* T_MATCH */
14405 ibf_load_object_complex_rational, /* T_COMPLEX */
14406 ibf_load_object_complex_rational, /* T_RATIONAL */
14407 ibf_load_object_unsupported, /* 0x10 */
14408 ibf_load_object_unsupported, /* T_NIL */
14409 ibf_load_object_unsupported, /* T_TRUE */
14410 ibf_load_object_unsupported, /* T_FALSE */
14411 ibf_load_object_symbol,
14412 ibf_load_object_unsupported, /* T_FIXNUM */
14413 ibf_load_object_unsupported, /* T_UNDEF */
14414 ibf_load_object_unsupported, /* 0x17 */
14415 ibf_load_object_unsupported, /* 0x18 */
14416 ibf_load_object_unsupported, /* 0x19 */
14417 ibf_load_object_unsupported, /* T_IMEMO 0x1a */
14418 ibf_load_object_unsupported, /* T_NODE 0x1b */
14419 ibf_load_object_unsupported, /* T_ICLASS 0x1c */
14420 ibf_load_object_unsupported, /* T_ZOMBIE 0x1d */
14421 ibf_load_object_unsupported, /* 0x1e */
14422 ibf_load_object_unsupported, /* 0x1f */
14423};
14424
14425static VALUE
14426ibf_load_object(const struct ibf_load *load, VALUE object_index)
14427{
14428 if (object_index == 0) {
14429 return Qnil;
14430 }
14431 else {
14432 VALUE obj = pinned_list_fetch(load->current_buffer->obj_list, (long)object_index);
14433 if (!obj) {
14434 ibf_offset_t *offsets = (ibf_offset_t *)(load->current_buffer->obj_list_offset + load->current_buffer->buff);
14435 ibf_offset_t offset = offsets[object_index];
14436 const struct ibf_object_header header = ibf_load_object_object_header(load, &offset);
14437
14438#if IBF_ISEQ_DEBUG
14439 fprintf(stderr, "ibf_load_object: list=%#x offsets=%p offset=%#x\n",
14440 load->current_buffer->obj_list_offset, (void *)offsets, offset);
14441 fprintf(stderr, "ibf_load_object: type=%#x special=%d frozen=%d internal=%d\n",
14442 header.type, header.special_const, header.frozen, header.internal);
14443#endif
14444 if (offset >= load->current_buffer->size) {
14445 rb_raise(rb_eIndexError, "object offset out of range: %u", offset);
14446 }
14447
14448 if (header.special_const) {
14449 ibf_offset_t reading_pos = offset;
14450
14451 obj = ibf_load_small_value(load, &reading_pos);
14452 }
14453 else {
14454 obj = (*load_object_functions[header.type])(load, &header, offset);
14455 }
14456
14457 pinned_list_store(load->current_buffer->obj_list, (long)object_index, obj);
14458 }
14459#if IBF_ISEQ_DEBUG
14460 fprintf(stderr, "ibf_load_object: index=%#"PRIxVALUE" obj=%#"PRIxVALUE"\n",
14461 object_index, obj);
14462#endif
14463 return obj;
14464 }
14465}
14466
14468{
14469 struct ibf_dump *dump;
14470 VALUE offset_list;
14471};
14472
14473static int
14474ibf_dump_object_list_i(st_data_t key, st_data_t val, st_data_t ptr)
14475{
14476 VALUE obj = (VALUE)key;
14477 struct ibf_dump_object_list_arg *args = (struct ibf_dump_object_list_arg *)ptr;
14478
14479 ibf_offset_t offset = ibf_dump_object_object(args->dump, obj);
14480 rb_ary_push(args->offset_list, UINT2NUM(offset));
14481
14482 return ST_CONTINUE;
14483}
14484
14485static void
14486ibf_dump_object_list(struct ibf_dump *dump, ibf_offset_t *obj_list_offset, unsigned int *obj_list_size)
14487{
14488 st_table *obj_table = dump->current_buffer->obj_table;
14489 VALUE offset_list = rb_ary_hidden_new(obj_table->num_entries);
14490
14491 struct ibf_dump_object_list_arg args;
14492 args.dump = dump;
14493 args.offset_list = offset_list;
14494
14495 st_foreach(obj_table, ibf_dump_object_list_i, (st_data_t)&args);
14496
14497 IBF_W_ALIGN(ibf_offset_t);
14498 *obj_list_offset = ibf_dump_pos(dump);
14499
14500 st_index_t size = obj_table->num_entries;
14501 st_index_t i;
14502
14503 for (i=0; i<size; i++) {
14504 ibf_offset_t offset = NUM2UINT(RARRAY_AREF(offset_list, i));
14505 IBF_WV(offset);
14506 }
14507
14508 *obj_list_size = (unsigned int)size;
14509}
14510
14511static void
14512ibf_dump_mark(void *ptr)
14513{
14514 struct ibf_dump *dump = (struct ibf_dump *)ptr;
14515 rb_gc_mark(dump->global_buffer.str);
14516
14517 rb_mark_set(dump->global_buffer.obj_table);
14518 rb_mark_set(dump->iseq_table);
14519}
14520
14521static void
14522ibf_dump_free(void *ptr)
14523{
14524 struct ibf_dump *dump = (struct ibf_dump *)ptr;
14525 if (dump->global_buffer.obj_table) {
14526 st_free_table(dump->global_buffer.obj_table);
14527 dump->global_buffer.obj_table = 0;
14528 }
14529 if (dump->iseq_table) {
14530 st_free_table(dump->iseq_table);
14531 dump->iseq_table = 0;
14532 }
14533}
14534
14535static size_t
14536ibf_dump_memsize(const void *ptr)
14537{
14538 struct ibf_dump *dump = (struct ibf_dump *)ptr;
14539 size_t size = 0;
14540 if (dump->iseq_table) size += st_memsize(dump->iseq_table);
14541 if (dump->global_buffer.obj_table) size += st_memsize(dump->global_buffer.obj_table);
14542 return size;
14543}
14544
14545static const rb_data_type_t ibf_dump_type = {
14546 "ibf_dump",
14547 {ibf_dump_mark, ibf_dump_free, ibf_dump_memsize,},
14548 0, 0, RUBY_TYPED_FREE_IMMEDIATELY | RUBY_TYPED_EMBEDDABLE
14549};
14550
14551static void
14552ibf_dump_setup(struct ibf_dump *dump, VALUE dumper_obj)
14553{
14554 dump->global_buffer.obj_table = NULL; // GC may run before a value is assigned
14555 dump->iseq_table = NULL;
14556
14557 RB_OBJ_WRITE(dumper_obj, &dump->global_buffer.str, rb_str_new(0, 0));
14558 dump->global_buffer.obj_table = ibf_dump_object_table_new();
14559 dump->iseq_table = st_init_numtable(); /* need free */
14560
14561 dump->current_buffer = &dump->global_buffer;
14562}
14563
14564VALUE
14565rb_iseq_ibf_dump(const rb_iseq_t *iseq, VALUE opt)
14566{
14567 struct ibf_dump *dump;
14568 struct ibf_header header = {{0}};
14569 VALUE dump_obj;
14570 VALUE str;
14571
14572 if (ISEQ_BODY(iseq)->parent_iseq != NULL ||
14573 ISEQ_BODY(iseq)->local_iseq != iseq) {
14574 rb_raise(rb_eRuntimeError, "should be top of iseq");
14575 }
14576 if (RTEST(ISEQ_COVERAGE(iseq))) {
14577 rb_raise(rb_eRuntimeError, "should not compile with coverage");
14578 }
14579
14580 dump_obj = TypedData_Make_Struct(0, struct ibf_dump, &ibf_dump_type, dump);
14581 ibf_dump_setup(dump, dump_obj);
14582
14583 ibf_dump_write(dump, &header, sizeof(header));
14584 ibf_dump_iseq(dump, iseq);
14585
14586 header.magic[0] = 'Y'; /* YARB */
14587 header.magic[1] = 'A';
14588 header.magic[2] = 'R';
14589 header.magic[3] = 'B';
14590 header.major_version = IBF_MAJOR_VERSION;
14591 header.minor_version = IBF_MINOR_VERSION;
14592 header.endian = IBF_ENDIAN_MARK;
14593 header.wordsize = (uint8_t)SIZEOF_VALUE;
14594 ibf_dump_iseq_list(dump, &header);
14595 ibf_dump_object_list(dump, &header.global_object_list_offset, &header.global_object_list_size);
14596 header.size = ibf_dump_pos(dump);
14597
14598 if (RTEST(opt)) {
14599 VALUE opt_str = opt;
14600 const char *ptr = StringValuePtr(opt_str);
14601 header.extra_size = RSTRING_LENINT(opt_str);
14602 ibf_dump_write(dump, ptr, header.extra_size);
14603 }
14604 else {
14605 header.extra_size = 0;
14606 }
14607
14608 ibf_dump_overwrite(dump, &header, sizeof(header), 0);
14609
14610 str = dump->global_buffer.str;
14611 RB_GC_GUARD(dump_obj);
14612 return str;
14613}
14614
14615static const ibf_offset_t *
14616ibf_iseq_list(const struct ibf_load *load)
14617{
14618 return (const ibf_offset_t *)(load->global_buffer.buff + load->header->iseq_list_offset);
14619}
14620
14621void
14622rb_ibf_load_iseq_complete(rb_iseq_t *iseq)
14623{
14624 struct ibf_load *load = RTYPEDDATA_DATA(iseq->aux.loader.obj);
14625 rb_iseq_t *prev_src_iseq = load->iseq;
14626 ibf_offset_t offset = ibf_iseq_list(load)[iseq->aux.loader.index];
14627 load->iseq = iseq;
14628#if IBF_ISEQ_DEBUG
14629 fprintf(stderr, "rb_ibf_load_iseq_complete: index=%#x offset=%#x size=%#x\n",
14630 iseq->aux.loader.index, offset,
14631 load->header->size);
14632#endif
14633 ibf_load_iseq_each(load, iseq, offset);
14634 ISEQ_COMPILE_DATA_CLEAR(iseq);
14635 FL_UNSET((VALUE)iseq, ISEQ_NOT_LOADED_YET);
14636 rb_iseq_init_trace(iseq);
14637 load->iseq = prev_src_iseq;
14638}
14639
14640#if USE_LAZY_LOAD
14641const rb_iseq_t *
14642rb_iseq_complete(const rb_iseq_t *iseq)
14643{
14644 rb_ibf_load_iseq_complete((rb_iseq_t *)iseq);
14645 return iseq;
14646}
14647#endif
14648
14649static rb_iseq_t *
14650ibf_load_iseq(const struct ibf_load *load, const rb_iseq_t *index_iseq)
14651{
14652 int iseq_index = (int)(VALUE)index_iseq;
14653
14654#if IBF_ISEQ_DEBUG
14655 fprintf(stderr, "ibf_load_iseq: index_iseq=%p iseq_list=%p\n",
14656 (void *)index_iseq, (void *)load->iseq_list);
14657#endif
14658 if (iseq_index == -1) {
14659 return NULL;
14660 }
14661 else {
14662 VALUE iseqv = pinned_list_fetch(load->iseq_list, iseq_index);
14663
14664#if IBF_ISEQ_DEBUG
14665 fprintf(stderr, "ibf_load_iseq: iseqv=%p\n", (void *)iseqv);
14666#endif
14667 if (iseqv) {
14668 return (rb_iseq_t *)iseqv;
14669 }
14670 else {
14671 rb_iseq_t *iseq = iseq_imemo_alloc();
14672#if IBF_ISEQ_DEBUG
14673 fprintf(stderr, "ibf_load_iseq: new iseq=%p\n", (void *)iseq);
14674#endif
14675 FL_SET((VALUE)iseq, ISEQ_NOT_LOADED_YET);
14676 iseq->aux.loader.obj = load->loader_obj;
14677 iseq->aux.loader.index = iseq_index;
14678#if IBF_ISEQ_DEBUG
14679 fprintf(stderr, "ibf_load_iseq: iseq=%p loader_obj=%p index=%d\n",
14680 (void *)iseq, (void *)load->loader_obj, iseq_index);
14681#endif
14682 pinned_list_store(load->iseq_list, iseq_index, (VALUE)iseq);
14683
14684 if (!USE_LAZY_LOAD || GET_VM()->builtin_function_table) {
14685#if IBF_ISEQ_DEBUG
14686 fprintf(stderr, "ibf_load_iseq: loading iseq=%p\n", (void *)iseq);
14687#endif
14688 rb_ibf_load_iseq_complete(iseq);
14689 }
14690
14691#if IBF_ISEQ_DEBUG
14692 fprintf(stderr, "ibf_load_iseq: iseq=%p loaded %p\n",
14693 (void *)iseq, (void *)load->iseq);
14694#endif
14695 return iseq;
14696 }
14697 }
14698}
14699
14700static void
14701ibf_load_setup_bytes(struct ibf_load *load, VALUE loader_obj, const char *bytes, size_t size)
14702{
14703 struct ibf_header *header = (struct ibf_header *)bytes;
14704 load->loader_obj = loader_obj;
14705 load->global_buffer.buff = bytes;
14706 load->header = header;
14707 load->global_buffer.size = header->size;
14708 load->global_buffer.obj_list_offset = header->global_object_list_offset;
14709 load->global_buffer.obj_list_size = header->global_object_list_size;
14710 RB_OBJ_WRITE(loader_obj, &load->iseq_list, pinned_list_new(header->iseq_list_size));
14711 RB_OBJ_WRITE(loader_obj, &load->global_buffer.obj_list, pinned_list_new(load->global_buffer.obj_list_size));
14712 load->iseq = NULL;
14713
14714 load->current_buffer = &load->global_buffer;
14715
14716 if (size < header->size) {
14717 rb_raise(rb_eRuntimeError, "broken binary format");
14718 }
14719 if (strncmp(header->magic, "YARB", 4) != 0) {
14720 rb_raise(rb_eRuntimeError, "unknown binary format");
14721 }
14722 if (header->major_version != IBF_MAJOR_VERSION ||
14723 header->minor_version != IBF_MINOR_VERSION) {
14724 rb_raise(rb_eRuntimeError, "unmatched version file (%u.%u for %u.%u)",
14725 header->major_version, header->minor_version, IBF_MAJOR_VERSION, IBF_MINOR_VERSION);
14726 }
14727 if (header->endian != IBF_ENDIAN_MARK) {
14728 rb_raise(rb_eRuntimeError, "unmatched endian: %c", header->endian);
14729 }
14730 if (header->wordsize != SIZEOF_VALUE) {
14731 rb_raise(rb_eRuntimeError, "unmatched word size: %d", header->wordsize);
14732 }
14733 if (header->iseq_list_offset % RUBY_ALIGNOF(ibf_offset_t)) {
14734 rb_raise(rb_eArgError, "unaligned iseq list offset: %u",
14735 header->iseq_list_offset);
14736 }
14737 if (load->global_buffer.obj_list_offset % RUBY_ALIGNOF(ibf_offset_t)) {
14738 rb_raise(rb_eArgError, "unaligned object list offset: %u",
14739 load->global_buffer.obj_list_offset);
14740 }
14741}
14742
14743static void
14744ibf_load_setup(struct ibf_load *load, VALUE loader_obj, VALUE str)
14745{
14746 StringValue(str);
14747
14748 if (RSTRING_LENINT(str) < (int)sizeof(struct ibf_header)) {
14749 rb_raise(rb_eRuntimeError, "broken binary format");
14750 }
14751
14752 if (USE_LAZY_LOAD) {
14753 str = rb_str_new(RSTRING_PTR(str), RSTRING_LEN(str));
14754 }
14755
14756 ibf_load_setup_bytes(load, loader_obj, RSTRING_PTR(str), RSTRING_LEN(str));
14757 RB_OBJ_WRITE(loader_obj, &load->str, str);
14758}
14759
14760static void
14761ibf_loader_mark(void *ptr)
14762{
14763 struct ibf_load *load = (struct ibf_load *)ptr;
14764 rb_gc_mark(load->str);
14765 rb_gc_mark(load->iseq_list);
14766 rb_gc_mark(load->global_buffer.obj_list);
14767}
14768
14769static void
14770ibf_loader_free(void *ptr)
14771{
14772 struct ibf_load *load = (struct ibf_load *)ptr;
14773 ruby_xfree(load);
14774}
14775
14776static size_t
14777ibf_loader_memsize(const void *ptr)
14778{
14779 return sizeof(struct ibf_load);
14780}
14781
14782static const rb_data_type_t ibf_load_type = {
14783 "ibf_loader",
14784 {ibf_loader_mark, ibf_loader_free, ibf_loader_memsize,},
14785 0, 0, RUBY_TYPED_WB_PROTECTED | RUBY_TYPED_FREE_IMMEDIATELY
14786};
14787
14788const rb_iseq_t *
14789rb_iseq_ibf_load(VALUE str)
14790{
14791 struct ibf_load *load;
14792 rb_iseq_t *iseq;
14793 VALUE loader_obj = TypedData_Make_Struct(0, struct ibf_load, &ibf_load_type, load);
14794
14795 ibf_load_setup(load, loader_obj, str);
14796 iseq = ibf_load_iseq(load, 0);
14797
14798 RB_GC_GUARD(loader_obj);
14799 return iseq;
14800}
14801
14802const rb_iseq_t *
14803rb_iseq_ibf_load_bytes(const char *bytes, size_t size)
14804{
14805 struct ibf_load *load;
14806 rb_iseq_t *iseq;
14807 VALUE loader_obj = TypedData_Make_Struct(0, struct ibf_load, &ibf_load_type, load);
14808
14809 ibf_load_setup_bytes(load, loader_obj, bytes, size);
14810 iseq = ibf_load_iseq(load, 0);
14811
14812 RB_GC_GUARD(loader_obj);
14813 return iseq;
14814}
14815
14816VALUE
14817rb_iseq_ibf_load_extra_data(VALUE str)
14818{
14819 struct ibf_load *load;
14820 VALUE loader_obj = TypedData_Make_Struct(0, struct ibf_load, &ibf_load_type, load);
14821 VALUE extra_str;
14822
14823 ibf_load_setup(load, loader_obj, str);
14824 extra_str = rb_str_new(load->global_buffer.buff + load->header->size, load->header->extra_size);
14825 RB_GC_GUARD(loader_obj);
14826 return extra_str;
14827}
14828
14829#include "prism_compile.c"
#define RUBY_ASSERT(...)
Asserts that the given expression is truthy if and only if RUBY_DEBUG is truthy.
Definition assert.h:219
#define LONG_LONG
Definition long_long.h:38
#define RUBY_ALIGNOF
Wraps (or simulates) alignof.
Definition stdalign.h:28
#define RUBY_EVENT_END
Encountered an end of a class clause.
Definition event.h:40
#define RUBY_EVENT_C_CALL
A method, written in C, is called.
Definition event.h:43
#define RUBY_EVENT_B_RETURN
Encountered a next statement.
Definition event.h:56
#define RUBY_EVENT_CLASS
Encountered a new class.
Definition event.h:39
#define RUBY_EVENT_NONE
No events.
Definition event.h:37
#define RUBY_EVENT_LINE
Encountered a new line.
Definition event.h:38
#define RUBY_EVENT_RETURN
Encountered a return statement.
Definition event.h:42
#define RUBY_EVENT_C_RETURN
Return from a method, written in C.
Definition event.h:44
#define RUBY_EVENT_B_CALL
Encountered an yield statement.
Definition event.h:55
uint32_t rb_event_flag_t
Represents event(s).
Definition event.h:108
#define RUBY_EVENT_CALL
A method, written in Ruby, is called.
Definition event.h:41
#define RUBY_EVENT_RESCUE
Encountered a rescue statement.
Definition event.h:61
#define RBIMPL_ATTR_FORMAT(x, y, z)
Wraps (or simulates) __attribute__((format))
Definition format.h:29
#define rb_str_new2
Old name of rb_str_new_cstr.
Definition string.h:1675
#define T_COMPLEX
Old name of RUBY_T_COMPLEX.
Definition value_type.h:59
#define TYPE(_)
Old name of rb_type.
Definition value_type.h:108
#define NUM2ULONG
Old name of RB_NUM2ULONG.
Definition long.h:52
#define NUM2LL
Old name of RB_NUM2LL.
Definition long_long.h:34
#define REALLOC_N
Old name of RB_REALLOC_N.
Definition memory.h:403
#define ALLOCV
Old name of RB_ALLOCV.
Definition memory.h:404
#define RFLOAT_VALUE
Old name of rb_float_value.
Definition double.h:28
#define T_STRING
Old name of RUBY_T_STRING.
Definition value_type.h:78
#define xfree
Old name of ruby_xfree.
Definition xmalloc.h:58
#define Qundef
Old name of RUBY_Qundef.
#define INT2FIX
Old name of RB_INT2FIX.
Definition long.h:48
#define rb_str_cat2
Old name of rb_str_cat_cstr.
Definition string.h:1683
#define T_NIL
Old name of RUBY_T_NIL.
Definition value_type.h:72
#define UNREACHABLE
Old name of RBIMPL_UNREACHABLE.
Definition assume.h:28
#define T_FLOAT
Old name of RUBY_T_FLOAT.
Definition value_type.h:64
#define ID2SYM
Old name of RB_ID2SYM.
Definition symbol.h:44
#define T_BIGNUM
Old name of RUBY_T_BIGNUM.
Definition value_type.h:57
#define SPECIAL_CONST_P
Old name of RB_SPECIAL_CONST_P.
#define OBJ_FREEZE
Old name of RB_OBJ_FREEZE.
Definition fl_type.h:135
#define ULONG2NUM
Old name of RB_ULONG2NUM.
Definition long.h:60
#define UNREACHABLE_RETURN
Old name of RBIMPL_UNREACHABLE_RETURN.
Definition assume.h:29
#define SYM2ID
Old name of RB_SYM2ID.
Definition symbol.h:45
#define FIX2UINT
Old name of RB_FIX2UINT.
Definition int.h:42
#define ZALLOC
Old name of RB_ZALLOC.
Definition memory.h:402
#define CLASS_OF
Old name of rb_class_of.
Definition globals.h:203
#define FIXABLE
Old name of RB_FIXABLE.
Definition fixnum.h:25
#define xmalloc
Old name of ruby_xmalloc.
Definition xmalloc.h:53
#define LONG2FIX
Old name of RB_INT2FIX.
Definition long.h:49
#define FIX2INT
Old name of RB_FIX2INT.
Definition int.h:41
#define NUM2UINT
Old name of RB_NUM2UINT.
Definition int.h:45
#define ZALLOC_N
Old name of RB_ZALLOC_N.
Definition memory.h:401
#define ASSUME
Old name of RBIMPL_ASSUME.
Definition assume.h:27
#define T_RATIONAL
Old name of RUBY_T_RATIONAL.
Definition value_type.h:76
#define T_HASH
Old name of RUBY_T_HASH.
Definition value_type.h:65
#define ALLOC_N
Old name of RB_ALLOC_N.
Definition memory.h:399
#define FL_SET
Old name of RB_FL_SET.
Definition fl_type.h:129
#define Qtrue
Old name of RUBY_Qtrue.
#define NUM2INT
Old name of RB_NUM2INT.
Definition int.h:44
#define Qnil
Old name of RUBY_Qnil.
#define Qfalse
Old name of RUBY_Qfalse.
#define FIX2LONG
Old name of RB_FIX2LONG.
Definition long.h:46
#define T_ARRAY
Old name of RUBY_T_ARRAY.
Definition value_type.h:56
#define NIL_P
Old name of RB_NIL_P.
#define T_SYMBOL
Old name of RUBY_T_SYMBOL.
Definition value_type.h:80
#define DBL2NUM
Old name of rb_float_new.
Definition double.h:29
#define BUILTIN_TYPE
Old name of RB_BUILTIN_TYPE.
Definition value_type.h:85
#define FL_TEST
Old name of RB_FL_TEST.
Definition fl_type.h:131
#define FL_FREEZE
Old name of RUBY_FL_FREEZE.
Definition fl_type.h:67
#define NUM2LONG
Old name of RB_NUM2LONG.
Definition long.h:51
#define FL_UNSET
Old name of RB_FL_UNSET.
Definition fl_type.h:133
#define UINT2NUM
Old name of RB_UINT2NUM.
Definition int.h:46
#define FIXNUM_P
Old name of RB_FIXNUM_P.
#define CONST_ID
Old name of RUBY_CONST_ID.
Definition symbol.h:47
#define ALLOCV_END
Old name of RB_ALLOCV_END.
Definition memory.h:406
#define SYMBOL_P
Old name of RB_SYMBOL_P.
Definition value_type.h:88
#define T_REGEXP
Old name of RUBY_T_REGEXP.
Definition value_type.h:77
#define ruby_debug
This variable controls whether the interpreter is in debug mode.
Definition error.h:486
VALUE rb_eNotImpError
NotImplementedError exception.
Definition error.c:1440
VALUE rb_eStandardError
StandardError exception.
Definition error.c:1427
VALUE rb_eTypeError
TypeError exception.
Definition error.c:1430
VALUE rb_eNoMatchingPatternError
NoMatchingPatternError exception.
Definition error.c:1443
void rb_exc_fatal(VALUE mesg)
Raises a fatal error in the current thread.
Definition eval.c:688
VALUE rb_eRuntimeError
RuntimeError exception.
Definition error.c:1428
void rb_warn(const char *fmt,...)
Identical to rb_warning(), except it reports unless $VERBOSE is nil.
Definition error.c:466
VALUE rb_eNoMatchingPatternKeyError
NoMatchingPatternKeyError exception.
Definition error.c:1444
VALUE rb_eIndexError
IndexError exception.
Definition error.c:1432
VALUE rb_eSyntaxError
SyntaxError exception.
Definition error.c:1447
@ RB_WARN_CATEGORY_STRICT_UNUSED_BLOCK
Warning is for checking unused block strictly.
Definition error.h:57
VALUE rb_obj_reveal(VALUE obj, VALUE klass)
Make a hidden object visible again.
Definition object.c:113
VALUE rb_cArray
Array class.
Definition array.c:40
VALUE rb_obj_hide(VALUE obj)
Make the object invisible from Ruby code.
Definition object.c:104
VALUE rb_cHash
Hash class.
Definition hash.c:113
VALUE rb_inspect(VALUE obj)
Generates a human-readable textual representation of the given object.
Definition object.c:680
VALUE rb_cRange
Range class.
Definition range.c:31
VALUE rb_obj_is_kind_of(VALUE obj, VALUE klass)
Queries if the given object is an instance (of possibly descendants) of the given class.
Definition object.c:865
VALUE rb_obj_freeze(VALUE obj)
Just calls rb_obj_freeze_inline() inside.
Definition object.c:1260
#define RB_OBJ_WRITTEN(old, oldv, young)
Identical to RB_OBJ_WRITE(), except it doesn't write any values, but only a WB declaration.
Definition gc.h:615
#define RB_OBJ_WRITE(old, slot, young)
Declaration of a "back" pointer.
Definition gc.h:603
#define INTEGER_PACK_NATIVE_BYTE_ORDER
Means either INTEGER_PACK_MSBYTE_FIRST or INTEGER_PACK_LSBYTE_FIRST, depending on the host processor'...
Definition bignum.h:546
#define INTEGER_PACK_NEGATIVE
Interprets the input as a signed negative number (unpack only).
Definition bignum.h:564
#define INTEGER_PACK_LSWORD_FIRST
Stores/interprets the least significant word as the first word.
Definition bignum.h:528
int rb_is_const_id(ID id)
Classifies the given ID, then sees if it is a constant.
Definition symbol.c:1079
int rb_is_attrset_id(ID id)
Classifies the given ID, then sees if it is an attribute writer.
Definition symbol.c:1103
int rb_range_values(VALUE range, VALUE *begp, VALUE *endp, int *exclp)
Deconstructs a range into its components.
Definition range.c:1838
VALUE rb_range_new(VALUE beg, VALUE end, int excl)
Creates a new Range.
Definition range.c:68
VALUE rb_rational_new(VALUE num, VALUE den)
Constructs a Rational, with reduction.
Definition rational.c:1974
int rb_reg_options(VALUE re)
Queries the options of the passed regular expression.
Definition re.c:4219
VALUE rb_str_append(VALUE dst, VALUE src)
Identical to rb_str_buf_append(), except it converts the right hand side before concatenating.
Definition string.c:3695
VALUE rb_str_tmp_new(long len)
Allocates a "temporary" string.
Definition string.c:1688
int rb_str_hash_cmp(VALUE str1, VALUE str2)
Compares two strings.
Definition string.c:4064
#define rb_str_new(str, len)
Allocates an instance of rb_cString.
Definition string.h:1498
st_index_t rb_str_hash(VALUE str)
Calculates a hash value of a string.
Definition string.c:4050
VALUE rb_str_cat(VALUE dst, const char *src, long srclen)
Destructively appends the passed contents to the string.
Definition string.c:3463
VALUE rb_str_buf_append(VALUE dst, VALUE src)
Identical to rb_str_cat_cstr(), except it takes Ruby's string instead of C's.
Definition string.c:3661
int rb_str_cmp(VALUE lhs, VALUE rhs)
Compares two strings, as in strcmp(3).
Definition string.c:4120
VALUE rb_str_concat(VALUE dst, VALUE src)
Identical to rb_str_append(), except it also accepts an integer as a codepoint.
Definition string.c:3937
VALUE rb_str_freeze(VALUE str)
This is the implementation of String#freeze.
Definition string.c:3195
#define rb_str_new_cstr(str)
Identical to rb_str_new, except it assumes the passed pointer is a pointer to a C string.
Definition string.h:1514
VALUE rb_class_name(VALUE obj)
Queries the name of the given object's class.
Definition variable.c:492
static ID rb_intern_const(const char *str)
This is a "tiny optimisation" over rb_intern().
Definition symbol.h:284
VALUE rb_id2sym(ID id)
Allocates an instance of rb_cSymbol that has the given id.
Definition symbol.c:967
VALUE rb_sym2str(VALUE symbol)
Obtain a frozen string representation of a symbol (not including the leading colon).
Definition symbol.c:986
ID rb_sym2id(VALUE obj)
Converts an instance of rb_cSymbol into an ID.
Definition symbol.c:933
int len
Length of the buffer.
Definition io.h:8
VALUE rb_ractor_make_shareable(VALUE obj)
Destructively transforms the passed object so that multiple Ractors can share it.
Definition ractor.c:3109
#define DECIMAL_SIZE_OF(expr)
An approximation of decimal representation size.
Definition util.h:48
void ruby_qsort(void *, const size_t, const size_t, int(*)(const void *, const void *, void *), void *)
Reentrant implementation of quick sort.
#define rb_long2int
Just another name of rb_long2int_inline.
Definition long.h:62
#define MEMCPY(p1, p2, type, n)
Handy macro to call memcpy.
Definition memory.h:372
#define ALLOCA_N(type, n)
Definition memory.h:292
#define MEMZERO(p, type, n)
Handy macro to erase a region of memory.
Definition memory.h:360
#define RB_GC_GUARD(v)
Prevents premature destruction of local objects.
Definition memory.h:167
#define RB_ALLOCV(v, n)
Identical to RB_ALLOCV_N(), except that it allocates a number of bytes and returns a void* .
Definition memory.h:304
VALUE type(ANYARGS)
ANYARGS-ed function type.
void rb_hash_foreach(VALUE q, int_type *w, VALUE e)
Iteration over the given hash.
#define RBIMPL_ATTR_NORETURN()
Wraps (or simulates) [[noreturn]]
Definition noreturn.h:38
#define RARRAY_LEN
Just another name of rb_array_len.
Definition rarray.h:51
static int RARRAY_LENINT(VALUE ary)
Identical to rb_array_len(), except it differs for the return type.
Definition rarray.h:281
static void RARRAY_ASET(VALUE ary, long i, VALUE v)
Assigns an object in an array.
Definition rarray.h:386
#define RARRAY_AREF(a, i)
Definition rarray.h:403
#define RARRAY_CONST_PTR
Just another name of rb_array_const_ptr.
Definition rarray.h:52
static VALUE RBASIC_CLASS(VALUE obj)
Queries the class of an object.
Definition rbasic.h:150
#define RUBY_DEFAULT_FREE
This is a value you can set to RData::dfree.
Definition rdata.h:78
void(* RUBY_DATA_FUNC)(void *)
This is the type of callbacks registered to RData.
Definition rdata.h:104
#define RHASH_SIZE(h)
Queries the size of the hash.
Definition rhash.h:69
static VALUE RREGEXP_SRC(VALUE rexp)
Convenient getter function.
Definition rregexp.h:103
#define StringValue(v)
Ensures that the parameter object is a String.
Definition rstring.h:66
#define StringValuePtr(v)
Identical to StringValue, except it returns a char*.
Definition rstring.h:76
static int RSTRING_LENINT(VALUE str)
Identical to RSTRING_LEN(), except it differs for the return type.
Definition rstring.h:468
#define StringValueCStr(v)
Identical to StringValuePtr, except it additionally checks for the contents for viability as a C stri...
Definition rstring.h:89
#define RTYPEDDATA_DATA(v)
Convenient getter macro.
Definition rtypeddata.h:102
#define TypedData_Get_Struct(obj, type, data_type, sval)
Obtains a C struct from inside of a wrapper Ruby object.
Definition rtypeddata.h:515
#define TypedData_Wrap_Struct(klass, data_type, sval)
Converts sval, a pointer to your struct, into a Ruby object.
Definition rtypeddata.h:449
#define TypedData_Make_Struct(klass, type, data_type, sval)
Identical to TypedData_Wrap_Struct, except it allocates a new data region internally instead of takin...
Definition rtypeddata.h:497
void rb_p(VALUE obj)
Inspects an object.
Definition io.c:9060
#define RTEST
This is an old name of RB_TEST.
#define _(args)
This was a transition path from K&R to ANSI.
Definition stdarg.h:35
Definition proc.c:29
Internal header for Complex.
Definition complex.h:13
Internal header for Rational.
Definition rational.h:16
Definition iseq.h:280
const ID * segments
A null-terminated list of ids, used to represent a constant's path idNULL is used to represent the ::...
Definition vm_core.h:285
Definition vm_core.h:288
Definition iseq.h:251
This is the struct that holds necessary info for a struct.
Definition rtypeddata.h:200
const char * wrap_struct_name
Name of structs of this kind.
Definition rtypeddata.h:207
struct rb_iseq_constant_body::@155 param
parameter information
Definition st.h:79
Definition vm_core.h:297
uintptr_t ID
Type that represents a Ruby identifier such as a variable name.
Definition value.h:52
#define SIZEOF_VALUE
Identical to sizeof(VALUE), except it is a macro that can also be used inside of preprocessor directi...
Definition value.h:69
uintptr_t VALUE
Type that represents a Ruby object.
Definition value.h:40
static bool RB_FLOAT_TYPE_P(VALUE obj)
Queries if the object is an instance of rb_cFloat.
Definition value_type.h:264
static bool rb_integer_type_p(VALUE obj)
Queries if the object is an instance of rb_cInteger.
Definition value_type.h:204
static bool RB_TYPE_P(VALUE obj, enum ruby_value_type t)
Queries if the given object is of given type.
Definition value_type.h:376
@ RUBY_T_MASK
Bitmask of ruby_value_type.
Definition value_type.h:145