1 /* Gimple walk support.
3 Copyright (C) 2007-2014 Free Software Foundation, Inc.
4 Contributed by Aldy Hernandez <aldyh@redhat.com>
6 This file is part of GCC.
8 GCC is free software; you can redistribute it and/or modify it under
9 the terms of the GNU General Public License as published by the Free
10 Software Foundation; either version 3, or (at your option) any later
13 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
14 WARRANTY; without even the implied warranty of MERCHANTABILITY or
15 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
18 You should have received a copy of the GNU General Public License
19 along with GCC; see the file COPYING3. If not see
20 <http://www.gnu.org/licenses/>. */
24 #include "coretypes.h"
33 #include "hard-reg-set.h"
36 #include "basic-block.h"
37 #include "tree-ssa-alias.h"
38 #include "internal-fn.h"
39 #include "gimple-expr.h"
42 #include "gimple-iterator.h"
43 #include "gimple-walk.h"
44 #include "gimple-walk.h"
47 /* Walk all the statements in the sequence *PSEQ calling walk_gimple_stmt
48 on each one. WI is as in walk_gimple_stmt.
50 If walk_gimple_stmt returns non-NULL, the walk is stopped, and the
51 value is stored in WI->CALLBACK_RESULT. Also, the statement that
52 produced the value is returned if this statement has not been
53 removed by a callback (wi->removed_stmt). If the statement has
54 been removed, NULL is returned.
56 Otherwise, all the statements are walked and NULL returned. */
59 walk_gimple_seq_mod (gimple_seq *pseq, walk_stmt_fn callback_stmt,
60 walk_tree_fn callback_op, struct walk_stmt_info *wi)
62 gimple_stmt_iterator gsi;
64 for (gsi = gsi_start (*pseq); !gsi_end_p (gsi); )
66 tree ret = walk_gimple_stmt (&gsi, callback_stmt, callback_op, wi);
69 /* If CALLBACK_STMT or CALLBACK_OP return a value, WI must exist
72 wi->callback_result = ret;
74 return wi->removed_stmt ? NULL : gsi_stmt (gsi);
77 if (!wi->removed_stmt)
82 wi->callback_result = NULL_TREE;
88 /* Like walk_gimple_seq_mod, but ensure that the head of SEQ isn't
89 changed by the callbacks. */
92 walk_gimple_seq (gimple_seq seq, walk_stmt_fn callback_stmt,
93 walk_tree_fn callback_op, struct walk_stmt_info *wi)
95 gimple_seq seq2 = seq;
96 gimple ret = walk_gimple_seq_mod (&seq2, callback_stmt, callback_op, wi);
97 gcc_assert (seq2 == seq);
102 /* Helper function for walk_gimple_stmt. Walk operands of a GIMPLE_ASM. */
105 walk_gimple_asm (gasm *stmt, walk_tree_fn callback_op,
106 struct walk_stmt_info *wi)
110 const char **oconstraints;
112 const char *constraint;
113 bool allows_mem, allows_reg, is_inout;
115 noutputs = gimple_asm_noutputs (stmt);
116 oconstraints = (const char **) alloca ((noutputs) * sizeof (const char *));
121 for (i = 0; i < noutputs; i++)
123 op = gimple_asm_output_op (stmt, i);
124 constraint = TREE_STRING_POINTER (TREE_VALUE (TREE_PURPOSE (op)));
125 oconstraints[i] = constraint;
126 parse_output_constraint (&constraint, i, 0, 0, &allows_mem, &allows_reg,
129 wi->val_only = (allows_reg || !allows_mem);
130 ret = walk_tree (&TREE_VALUE (op), callback_op, wi, NULL);
135 n = gimple_asm_ninputs (stmt);
136 for (i = 0; i < n; i++)
138 op = gimple_asm_input_op (stmt, i);
139 constraint = TREE_STRING_POINTER (TREE_VALUE (TREE_PURPOSE (op)));
140 parse_input_constraint (&constraint, 0, 0, noutputs, 0,
141 oconstraints, &allows_mem, &allows_reg);
144 wi->val_only = (allows_reg || !allows_mem);
145 /* Although input "m" is not really a LHS, we need a lvalue. */
146 wi->is_lhs = !wi->val_only;
148 ret = walk_tree (&TREE_VALUE (op), callback_op, wi, NULL);
159 n = gimple_asm_nlabels (stmt);
160 for (i = 0; i < n; i++)
162 op = gimple_asm_label_op (stmt, i);
163 ret = walk_tree (&TREE_VALUE (op), callback_op, wi, NULL);
172 /* Helper function of WALK_GIMPLE_STMT. Walk every tree operand in
173 STMT. CALLBACK_OP and WI are as in WALK_GIMPLE_STMT.
175 CALLBACK_OP is called on each operand of STMT via walk_tree.
176 Additional parameters to walk_tree must be stored in WI. For each operand
177 OP, walk_tree is called as:
179 walk_tree (&OP, CALLBACK_OP, WI, WI->PSET)
181 If CALLBACK_OP returns non-NULL for an operand, the remaining
182 operands are not scanned.
184 The return value is that returned by the last call to walk_tree, or
185 NULL_TREE if no CALLBACK_OP is specified. */
188 walk_gimple_op (gimple stmt, walk_tree_fn callback_op,
189 struct walk_stmt_info *wi)
191 hash_set<tree> *pset = (wi) ? wi->pset : NULL;
193 tree ret = NULL_TREE;
195 switch (gimple_code (stmt))
198 /* Walk the RHS operands. If the LHS is of a non-renamable type or
199 is a register variable, we may use a COMPONENT_REF on the RHS. */
202 tree lhs = gimple_assign_lhs (stmt);
204 = (is_gimple_reg_type (TREE_TYPE (lhs)) && !is_gimple_reg (lhs))
205 || gimple_assign_rhs_class (stmt) != GIMPLE_SINGLE_RHS;
208 for (i = 1; i < gimple_num_ops (stmt); i++)
210 ret = walk_tree (gimple_op_ptr (stmt, i), callback_op, wi,
216 /* Walk the LHS. If the RHS is appropriate for a memory, we
217 may use a COMPONENT_REF on the LHS. */
220 /* If the RHS is of a non-renamable type or is a register variable,
221 we may use a COMPONENT_REF on the LHS. */
222 tree rhs1 = gimple_assign_rhs1 (stmt);
224 = (is_gimple_reg_type (TREE_TYPE (rhs1)) && !is_gimple_reg (rhs1))
225 || gimple_assign_rhs_class (stmt) != GIMPLE_SINGLE_RHS;
229 ret = walk_tree (gimple_op_ptr (stmt, 0), callback_op, wi, pset);
247 ret = walk_tree (gimple_call_chain_ptr (as_a <gcall *> (stmt)),
248 callback_op, wi, pset);
252 ret = walk_tree (gimple_call_fn_ptr (stmt), callback_op, wi, pset);
256 for (i = 0; i < gimple_call_num_args (stmt); i++)
260 = is_gimple_reg_type (TREE_TYPE (gimple_call_arg (stmt, i)));
261 ret = walk_tree (gimple_call_arg_ptr (stmt, i), callback_op, wi,
267 if (gimple_call_lhs (stmt))
273 = is_gimple_reg_type (TREE_TYPE (gimple_call_lhs (stmt)));
276 ret = walk_tree (gimple_call_lhs_ptr (stmt), callback_op, wi, pset);
289 ret = walk_tree (gimple_catch_types_ptr (as_a <gcatch *> (stmt)),
290 callback_op, wi, pset);
295 case GIMPLE_EH_FILTER:
296 ret = walk_tree (gimple_eh_filter_types_ptr (stmt), callback_op, wi,
303 ret = walk_gimple_asm (as_a <gasm *> (stmt), callback_op, wi);
308 case GIMPLE_OMP_CONTINUE:
310 gomp_continue *cont_stmt = as_a <gomp_continue *> (stmt);
311 ret = walk_tree (gimple_omp_continue_control_def_ptr (cont_stmt),
312 callback_op, wi, pset);
316 ret = walk_tree (gimple_omp_continue_control_use_ptr (cont_stmt),
317 callback_op, wi, pset);
323 case GIMPLE_OMP_CRITICAL:
324 ret = walk_tree (gimple_omp_critical_name_ptr (
325 as_a <gomp_critical *> (stmt)),
326 callback_op, wi, pset);
332 ret = walk_tree (gimple_omp_for_clauses_ptr (stmt), callback_op, wi,
336 for (i = 0; i < gimple_omp_for_collapse (stmt); i++)
338 ret = walk_tree (gimple_omp_for_index_ptr (stmt, i), callback_op,
342 ret = walk_tree (gimple_omp_for_initial_ptr (stmt, i), callback_op,
346 ret = walk_tree (gimple_omp_for_final_ptr (stmt, i), callback_op,
350 ret = walk_tree (gimple_omp_for_incr_ptr (stmt, i), callback_op,
357 case GIMPLE_OMP_PARALLEL:
359 gomp_parallel *omp_par_stmt = as_a <gomp_parallel *> (stmt);
360 ret = walk_tree (gimple_omp_parallel_clauses_ptr (omp_par_stmt),
361 callback_op, wi, pset);
364 ret = walk_tree (gimple_omp_parallel_child_fn_ptr (omp_par_stmt),
365 callback_op, wi, pset);
368 ret = walk_tree (gimple_omp_parallel_data_arg_ptr (omp_par_stmt),
369 callback_op, wi, pset);
375 case GIMPLE_OMP_TASK:
376 ret = walk_tree (gimple_omp_task_clauses_ptr (stmt), callback_op,
380 ret = walk_tree (gimple_omp_task_child_fn_ptr (stmt), callback_op,
384 ret = walk_tree (gimple_omp_task_data_arg_ptr (stmt), callback_op,
388 ret = walk_tree (gimple_omp_task_copy_fn_ptr (stmt), callback_op,
392 ret = walk_tree (gimple_omp_task_arg_size_ptr (stmt), callback_op,
396 ret = walk_tree (gimple_omp_task_arg_align_ptr (stmt), callback_op,
402 case GIMPLE_OMP_SECTIONS:
403 ret = walk_tree (gimple_omp_sections_clauses_ptr (stmt), callback_op,
408 ret = walk_tree (gimple_omp_sections_control_ptr (stmt), callback_op,
415 case GIMPLE_OMP_SINGLE:
416 ret = walk_tree (gimple_omp_single_clauses_ptr (stmt), callback_op, wi,
422 case GIMPLE_OMP_TARGET:
423 ret = walk_tree (gimple_omp_target_clauses_ptr (stmt), callback_op, wi,
429 case GIMPLE_OMP_TEAMS:
430 ret = walk_tree (gimple_omp_teams_clauses_ptr (stmt), callback_op, wi,
436 case GIMPLE_OMP_ATOMIC_LOAD:
438 gomp_atomic_load *omp_stmt = as_a <gomp_atomic_load *> (stmt);
439 ret = walk_tree (gimple_omp_atomic_load_lhs_ptr (omp_stmt),
440 callback_op, wi, pset);
444 ret = walk_tree (gimple_omp_atomic_load_rhs_ptr (omp_stmt),
445 callback_op, wi, pset);
451 case GIMPLE_OMP_ATOMIC_STORE:
452 ret = walk_tree (gimple_omp_atomic_store_val_ptr (
453 as_a <gomp_atomic_store *> (stmt)),
454 callback_op, wi, pset);
459 case GIMPLE_TRANSACTION:
460 ret = walk_tree (gimple_transaction_label_ptr (
461 as_a <gtransaction *> (stmt)),
462 callback_op, wi, pset);
467 case GIMPLE_OMP_RETURN:
468 ret = walk_tree (gimple_omp_return_lhs_ptr (stmt), callback_op, wi,
474 /* Tuples that do not have operands. */
482 enum gimple_statement_structure_enum gss;
483 gss = gimple_statement_structure (stmt);
484 if (gss == GSS_WITH_OPS || gss == GSS_WITH_MEM_OPS)
485 for (i = 0; i < gimple_num_ops (stmt); i++)
487 ret = walk_tree (gimple_op_ptr (stmt, i), callback_op, wi, pset);
499 /* Walk the current statement in GSI (optionally using traversal state
500 stored in WI). If WI is NULL, no state is kept during traversal.
501 The callback CALLBACK_STMT is called. If CALLBACK_STMT indicates
502 that it has handled all the operands of the statement, its return
503 value is returned. Otherwise, the return value from CALLBACK_STMT
504 is discarded and its operands are scanned.
506 If CALLBACK_STMT is NULL or it didn't handle the operands,
507 CALLBACK_OP is called on each operand of the statement via
508 walk_gimple_op. If walk_gimple_op returns non-NULL for any
509 operand, the remaining operands are not scanned. In this case, the
510 return value from CALLBACK_OP is returned.
512 In any other case, NULL_TREE is returned. */
515 walk_gimple_stmt (gimple_stmt_iterator *gsi, walk_stmt_fn callback_stmt,
516 walk_tree_fn callback_op, struct walk_stmt_info *wi)
520 gimple stmt = gsi_stmt (*gsi);
525 wi->removed_stmt = false;
527 if (wi->want_locations && gimple_has_location (stmt))
528 input_location = gimple_location (stmt);
533 /* Invoke the statement callback. Return if the callback handled
534 all of STMT operands by itself. */
537 bool handled_ops = false;
538 tree_ret = callback_stmt (gsi, &handled_ops, wi);
542 /* If CALLBACK_STMT did not handle operands, it should not have
543 a value to return. */
544 gcc_assert (tree_ret == NULL);
546 if (wi && wi->removed_stmt)
549 /* Re-read stmt in case the callback changed it. */
550 stmt = gsi_stmt (*gsi);
553 /* If CALLBACK_OP is defined, invoke it on every operand of STMT. */
556 tree_ret = walk_gimple_op (stmt, callback_op, wi);
561 /* If STMT can have statements inside (e.g. GIMPLE_BIND), walk them. */
562 switch (gimple_code (stmt))
565 ret = walk_gimple_seq_mod (gimple_bind_body_ptr (as_a <gbind *> (stmt)),
566 callback_stmt, callback_op, wi);
568 return wi->callback_result;
572 ret = walk_gimple_seq_mod (gimple_catch_handler_ptr (
573 as_a <gcatch *> (stmt)),
574 callback_stmt, callback_op, wi);
576 return wi->callback_result;
579 case GIMPLE_EH_FILTER:
580 ret = walk_gimple_seq_mod (gimple_eh_filter_failure_ptr (stmt), callback_stmt,
583 return wi->callback_result;
588 geh_else *eh_else_stmt = as_a <geh_else *> (stmt);
589 ret = walk_gimple_seq_mod (gimple_eh_else_n_body_ptr (eh_else_stmt),
590 callback_stmt, callback_op, wi);
592 return wi->callback_result;
593 ret = walk_gimple_seq_mod (gimple_eh_else_e_body_ptr (eh_else_stmt),
594 callback_stmt, callback_op, wi);
596 return wi->callback_result;
601 ret = walk_gimple_seq_mod (gimple_try_eval_ptr (stmt), callback_stmt, callback_op,
604 return wi->callback_result;
606 ret = walk_gimple_seq_mod (gimple_try_cleanup_ptr (stmt), callback_stmt,
609 return wi->callback_result;
613 ret = walk_gimple_seq_mod (gimple_omp_for_pre_body_ptr (stmt), callback_stmt,
616 return wi->callback_result;
619 case GIMPLE_OMP_CRITICAL:
620 case GIMPLE_OMP_MASTER:
621 case GIMPLE_OMP_TASKGROUP:
622 case GIMPLE_OMP_ORDERED:
623 case GIMPLE_OMP_SECTION:
624 case GIMPLE_OMP_PARALLEL:
625 case GIMPLE_OMP_TASK:
626 case GIMPLE_OMP_SECTIONS:
627 case GIMPLE_OMP_SINGLE:
628 case GIMPLE_OMP_TARGET:
629 case GIMPLE_OMP_TEAMS:
630 ret = walk_gimple_seq_mod (gimple_omp_body_ptr (stmt), callback_stmt,
633 return wi->callback_result;
636 case GIMPLE_WITH_CLEANUP_EXPR:
637 ret = walk_gimple_seq_mod (gimple_wce_cleanup_ptr (stmt), callback_stmt,
640 return wi->callback_result;
643 case GIMPLE_TRANSACTION:
644 ret = walk_gimple_seq_mod (gimple_transaction_body_ptr (
645 as_a <gtransaction *> (stmt)),
646 callback_stmt, callback_op, wi);
648 return wi->callback_result;
652 gcc_assert (!gimple_has_substatements (stmt));
659 /* From a tree operand OP return the base of a load or store operation
660 or NULL_TREE if OP is not a load or a store. */
663 get_base_loadstore (tree op)
665 while (handled_component_p (op))
666 op = TREE_OPERAND (op, 0);
668 || INDIRECT_REF_P (op)
669 || TREE_CODE (op) == MEM_REF
670 || TREE_CODE (op) == TARGET_MEM_REF)
676 /* For the statement STMT call the callbacks VISIT_LOAD, VISIT_STORE and
677 VISIT_ADDR if non-NULL on loads, store and address-taken operands
678 passing the STMT, the base of the operand, the operand itself containing
679 the base and DATA to it. The base will be either a decl, an indirect
680 reference (including TARGET_MEM_REF) or the argument of an address
682 Returns the results of these callbacks or'ed. */
685 walk_stmt_load_store_addr_ops (gimple stmt, void *data,
686 walk_stmt_load_store_addr_fn visit_load,
687 walk_stmt_load_store_addr_fn visit_store,
688 walk_stmt_load_store_addr_fn visit_addr)
692 if (gimple_assign_single_p (stmt))
697 arg = gimple_assign_lhs (stmt);
698 lhs = get_base_loadstore (arg);
700 ret |= visit_store (stmt, lhs, arg, data);
702 arg = gimple_assign_rhs1 (stmt);
704 while (handled_component_p (rhs))
705 rhs = TREE_OPERAND (rhs, 0);
708 if (TREE_CODE (rhs) == ADDR_EXPR)
709 ret |= visit_addr (stmt, TREE_OPERAND (rhs, 0), arg, data);
710 else if (TREE_CODE (rhs) == TARGET_MEM_REF
711 && TREE_CODE (TMR_BASE (rhs)) == ADDR_EXPR)
712 ret |= visit_addr (stmt, TREE_OPERAND (TMR_BASE (rhs), 0), arg,
714 else if (TREE_CODE (rhs) == OBJ_TYPE_REF
715 && TREE_CODE (OBJ_TYPE_REF_OBJECT (rhs)) == ADDR_EXPR)
716 ret |= visit_addr (stmt, TREE_OPERAND (OBJ_TYPE_REF_OBJECT (rhs),
718 else if (TREE_CODE (rhs) == CONSTRUCTOR)
723 FOR_EACH_CONSTRUCTOR_VALUE (CONSTRUCTOR_ELTS (rhs), ix, val)
724 if (TREE_CODE (val) == ADDR_EXPR)
725 ret |= visit_addr (stmt, TREE_OPERAND (val, 0), arg, data);
726 else if (TREE_CODE (val) == OBJ_TYPE_REF
727 && TREE_CODE (OBJ_TYPE_REF_OBJECT (val)) == ADDR_EXPR)
728 ret |= visit_addr (stmt,
729 TREE_OPERAND (OBJ_TYPE_REF_OBJECT (val),
732 lhs = gimple_assign_lhs (stmt);
733 if (TREE_CODE (lhs) == TARGET_MEM_REF
734 && TREE_CODE (TMR_BASE (lhs)) == ADDR_EXPR)
735 ret |= visit_addr (stmt, TREE_OPERAND (TMR_BASE (lhs), 0), lhs, data);
739 rhs = get_base_loadstore (rhs);
741 ret |= visit_load (stmt, rhs, arg, data);
745 && (is_gimple_assign (stmt)
746 || gimple_code (stmt) == GIMPLE_COND))
748 for (i = 0; i < gimple_num_ops (stmt); ++i)
750 tree op = gimple_op (stmt, i);
753 else if (TREE_CODE (op) == ADDR_EXPR)
754 ret |= visit_addr (stmt, TREE_OPERAND (op, 0), op, data);
755 /* COND_EXPR and VCOND_EXPR rhs1 argument is a comparison
756 tree with two operands. */
757 else if (i == 1 && COMPARISON_CLASS_P (op))
759 if (TREE_CODE (TREE_OPERAND (op, 0)) == ADDR_EXPR)
760 ret |= visit_addr (stmt, TREE_OPERAND (TREE_OPERAND (op, 0),
762 if (TREE_CODE (TREE_OPERAND (op, 1)) == ADDR_EXPR)
763 ret |= visit_addr (stmt, TREE_OPERAND (TREE_OPERAND (op, 1),
768 else if (gcall *call_stmt = dyn_cast <gcall *> (stmt))
772 tree arg = gimple_call_lhs (call_stmt);
775 tree lhs = get_base_loadstore (arg);
777 ret |= visit_store (stmt, lhs, arg, data);
780 if (visit_load || visit_addr)
781 for (i = 0; i < gimple_call_num_args (call_stmt); ++i)
783 tree arg = gimple_call_arg (call_stmt, i);
785 && TREE_CODE (arg) == ADDR_EXPR)
786 ret |= visit_addr (stmt, TREE_OPERAND (arg, 0), arg, data);
789 tree rhs = get_base_loadstore (arg);
791 ret |= visit_load (stmt, rhs, arg, data);
795 && gimple_call_chain (call_stmt)
796 && TREE_CODE (gimple_call_chain (call_stmt)) == ADDR_EXPR)
797 ret |= visit_addr (stmt, TREE_OPERAND (gimple_call_chain (call_stmt), 0),
798 gimple_call_chain (call_stmt), data);
800 && gimple_call_return_slot_opt_p (call_stmt)
801 && gimple_call_lhs (call_stmt) != NULL_TREE
802 && TREE_ADDRESSABLE (TREE_TYPE (gimple_call_lhs (call_stmt))))
803 ret |= visit_addr (stmt, gimple_call_lhs (call_stmt),
804 gimple_call_lhs (call_stmt), data);
806 else if (gasm *asm_stmt = dyn_cast <gasm *> (stmt))
809 const char *constraint;
810 const char **oconstraints;
811 bool allows_mem, allows_reg, is_inout;
812 noutputs = gimple_asm_noutputs (asm_stmt);
813 oconstraints = XALLOCAVEC (const char *, noutputs);
814 if (visit_store || visit_addr)
815 for (i = 0; i < gimple_asm_noutputs (asm_stmt); ++i)
817 tree link = gimple_asm_output_op (asm_stmt, i);
818 tree op = get_base_loadstore (TREE_VALUE (link));
819 if (op && visit_store)
820 ret |= visit_store (stmt, op, TREE_VALUE (link), data);
823 constraint = TREE_STRING_POINTER
824 (TREE_VALUE (TREE_PURPOSE (link)));
825 oconstraints[i] = constraint;
826 parse_output_constraint (&constraint, i, 0, 0, &allows_mem,
827 &allows_reg, &is_inout);
828 if (op && !allows_reg && allows_mem)
829 ret |= visit_addr (stmt, op, TREE_VALUE (link), data);
832 if (visit_load || visit_addr)
833 for (i = 0; i < gimple_asm_ninputs (asm_stmt); ++i)
835 tree link = gimple_asm_input_op (asm_stmt, i);
836 tree op = TREE_VALUE (link);
838 && TREE_CODE (op) == ADDR_EXPR)
839 ret |= visit_addr (stmt, TREE_OPERAND (op, 0), op, data);
840 else if (visit_load || visit_addr)
842 op = get_base_loadstore (op);
846 ret |= visit_load (stmt, op, TREE_VALUE (link), data);
849 constraint = TREE_STRING_POINTER
850 (TREE_VALUE (TREE_PURPOSE (link)));
851 parse_input_constraint (&constraint, 0, 0, noutputs,
853 &allows_mem, &allows_reg);
854 if (!allows_reg && allows_mem)
855 ret |= visit_addr (stmt, op, TREE_VALUE (link),
862 else if (greturn *return_stmt = dyn_cast <greturn *> (stmt))
864 tree op = gimple_return_retval (return_stmt);
868 && TREE_CODE (op) == ADDR_EXPR)
869 ret |= visit_addr (stmt, TREE_OPERAND (op, 0), op, data);
872 tree base = get_base_loadstore (op);
874 ret |= visit_load (stmt, base, op, data);
879 && gimple_code (stmt) == GIMPLE_PHI)
881 for (i = 0; i < gimple_phi_num_args (stmt); ++i)
883 tree op = gimple_phi_arg_def (stmt, i);
884 if (TREE_CODE (op) == ADDR_EXPR)
885 ret |= visit_addr (stmt, TREE_OPERAND (op, 0), op, data);
889 && gimple_code (stmt) == GIMPLE_GOTO)
891 tree op = gimple_goto_dest (stmt);
892 if (TREE_CODE (op) == ADDR_EXPR)
893 ret |= visit_addr (stmt, TREE_OPERAND (op, 0), op, data);
899 /* Like walk_stmt_load_store_addr_ops but with NULL visit_addr. IPA-CP
900 should make a faster clone for this case. */
903 walk_stmt_load_store_ops (gimple stmt, void *data,
904 walk_stmt_load_store_addr_fn visit_load,
905 walk_stmt_load_store_addr_fn visit_store)
907 return walk_stmt_load_store_addr_ops (stmt, data,
908 visit_load, visit_store, NULL);