Lines Matching refs:worklist

62 #define WHILE_WORKLIST(worklist, len, i) do { \  argument
66 ZEND_BITSET_FOREACH(worklist, len, i) { \
67 zend_bitset_excl(worklist, i); \
97 zend_bitset_incl(worklist, _var); \
105 zend_bitset_incl(worklist, _var); \
272 zend_bitset worklist; in zend_ssa_find_false_dependencies() local
281 worklist = do_alloca(sizeof(zend_ulong) * zend_bitset_len(ssa_vars_count), use_heap); in zend_ssa_find_false_dependencies()
282 memset(worklist, 0, sizeof(zend_ulong) * zend_bitset_len(ssa_vars_count)); in zend_ssa_find_false_dependencies()
290 zend_bitset_incl(worklist, i); in zend_ssa_find_false_dependencies()
297 WHILE_WORKLIST(worklist, zend_bitset_len(ssa_vars_count), i) { in zend_ssa_find_false_dependencies()
304 zend_bitset_incl(worklist, p->sources[0]); in zend_ssa_find_false_dependencies()
310 zend_bitset_incl(worklist, p->sources[j]); in zend_ssa_find_false_dependencies()
317 free_alloca(worklist, use_heap); in zend_ssa_find_false_dependencies()
1554 zend_check_inner_cycles(op_array, ssa, worklist, visited, var2)) { \
1559 …_inner_cycles(const zend_op_array *op_array, zend_ssa *ssa, zend_bitset worklist, zend_bitset visi… in zend_check_inner_cycles() argument
1561 if (zend_bitset_in(worklist, var)) { in zend_check_inner_cycles()
1564 zend_bitset_incl(worklist, var); in zend_check_inner_cycles()
1577 zend_bitset worklist = do_alloca(sizeof(zend_ulong) * worklist_len * 2, use_heap); in zend_infer_ranges_warmup() local
1578 zend_bitset visited = worklist + worklist_len; in zend_infer_ranges_warmup()
1582 memset(worklist, 0, sizeof(zend_ulong) * worklist_len); in zend_infer_ranges_warmup()
1587 zend_check_inner_cycles(op_array, ssa, worklist, visited, j)) { in zend_infer_ranges_warmup()
1595 memset(worklist, 0, sizeof(zend_ulong) * worklist_len); in zend_infer_ranges_warmup()
1601 zend_bitset_incl(worklist, j); in zend_infer_ranges_warmup()
1608 WHILE_WORKLIST(worklist, worklist_len, j) { in zend_infer_ranges_warmup()
1665 free_alloca(worklist, use_heap); in zend_infer_ranges_warmup()
1671 zend_bitset worklist; in zend_infer_ranges() local
1679 worklist = do_alloca( in zend_infer_ranges()
1683 next_scc_var = (int*)((char*)worklist + ZEND_MM_ALIGNED_SIZE(sizeof(zend_ulong) * worklist_len)); in zend_infer_ranges()
1708 memset(worklist, 0, sizeof(zend_ulong) * worklist_len); in zend_infer_ranges()
1711 zend_bitset_incl(worklist, j); in zend_infer_ranges()
1720 zend_bitset_incl(worklist, j); in zend_infer_ranges()
1726 WHILE_WORKLIST(worklist, worklist_len, j) { in zend_infer_ranges()
1741 zend_bitset_incl(worklist, j); in zend_infer_ranges()
1745 WHILE_WORKLIST(worklist, worklist_len, j) { in zend_infer_ranges()
1761 free_alloca(worklist, use_heap); in zend_infer_ranges()
1786 handle_type_narrowing(op_array, ssa, worklist, \
1791 add_usages(op_array, ssa, worklist, __var); \
1804 add_usages(op_array, ssa, worklist, var); \
1820 static void add_usages(const zend_op_array *op_array, zend_ssa *ssa, zend_bitset worklist, int var) in add_usages() argument
1825 zend_bitset_incl(worklist, p->ssa_var); in add_usages()
1836 zend_bitset_incl(worklist, op->result_def); in add_usages()
1839 zend_bitset_incl(worklist, op->op1_def); in add_usages()
1842 zend_bitset_incl(worklist, op->op2_def); in add_usages()
1847 zend_bitset_incl(worklist, op->result_def); in add_usages()
1850 zend_bitset_incl(worklist, op->op1_def); in add_usages()
1853 zend_bitset_incl(worklist, op->op2_def); in add_usages()
1861 static void reset_dependent_vars(const zend_op_array *op_array, zend_ssa *ssa, zend_bitset worklist in reset_dependent_vars() argument
1873 zend_bitset_incl(worklist, p->ssa_var); in reset_dependent_vars()
1874 reset_dependent_vars(op_array, ssa, worklist, p->ssa_var); in reset_dependent_vars()
1882 zend_bitset_incl(worklist, ssa_ops[use].op1_def); in reset_dependent_vars()
1883 reset_dependent_vars(op_array, ssa, worklist, ssa_ops[use].op1_def); in reset_dependent_vars()
1887 zend_bitset_incl(worklist, ssa_ops[use].op2_def); in reset_dependent_vars()
1888 reset_dependent_vars(op_array, ssa, worklist, ssa_ops[use].op2_def); in reset_dependent_vars()
1892 zend_bitset_incl(worklist, ssa_ops[use].result_def); in reset_dependent_vars()
1893 reset_dependent_vars(op_array, ssa, worklist, ssa_ops[use].result_def); in reset_dependent_vars()
1898 zend_bitset_incl(worklist, ssa_ops[use+1].op1_def); in reset_dependent_vars()
1899 reset_dependent_vars(op_array, ssa, worklist, ssa_ops[use+1].op1_def); in reset_dependent_vars()
1903 zend_bitset_incl(worklist, ssa_ops[use+1].op2_def); in reset_dependent_vars()
1904 reset_dependent_vars(op_array, ssa, worklist, ssa_ops[use+1].op2_def); in reset_dependent_vars()
1908 zend_bitset_incl(worklist, ssa_ops[use+1].result_def); in reset_dependent_vars()
1909 reset_dependent_vars(op_array, ssa, worklist, ssa_ops[use+1].result_def); in reset_dependent_vars()
1919 zend_bitset_incl(worklist, p->ssa_var); in reset_dependent_vars()
1920 reset_dependent_vars(op_array, ssa, worklist, p->ssa_var); in reset_dependent_vars()
1926 …ype_narrowing(const zend_op_array *op_array, zend_ssa *ssa, zend_bitset worklist, int var, uint32_… in handle_type_narrowing() argument
1940 reset_dependent_vars(op_array, ssa, worklist, var); in handle_type_narrowing()
2168 zend_bitset worklist, in zend_update_type_info() argument
3333 …s_ex(const zend_op_array *op_array, const zend_script *script, zend_ssa *ssa, zend_bitset worklist) in zend_infer_types_ex() argument
3342 while (!zend_bitset_empty(worklist, worklist_len)) { in zend_infer_types_ex()
3343 j = zend_bitset_first(worklist, worklist_len); in zend_infer_types_ex()
3344 zend_bitset_excl(worklist, j); in zend_infer_types_ex()
3399 if (zend_update_type_info(op_array, ssa, script, worklist, i) == FAILURE) { in zend_infer_types_ex()
3574 zend_bitset visited, worklist; in zend_type_narrowing() local
3581 worklist = visited + bitset_len; in zend_type_narrowing()
3583 zend_bitset_clear(worklist, bitset_len); in zend_type_narrowing()
3605 zend_bitset_union(worklist, visited, bitset_len); in zend_type_narrowing()
3615 if (zend_infer_types_ex(op_array, script, ssa, worklist) != SUCCESS) { in zend_type_narrowing()
3868 zend_bitset worklist; in zend_infer_types() local
3871 worklist = do_alloca(sizeof(zend_ulong) * zend_bitset_len(ssa_vars_count), use_heap); in zend_infer_types()
3872 memset(worklist, 0, sizeof(zend_ulong) * zend_bitset_len(ssa_vars_count)); in zend_infer_types()
3876 zend_bitset_incl(worklist, j); in zend_infer_types()
3880 if (zend_infer_types_ex(op_array, script, ssa, worklist) != SUCCESS) { in zend_infer_types()
3881 free_alloca(worklist, use_heap); in zend_infer_types()
3911 free_alloca(worklist, use_heap); in zend_infer_types()
3957 zend_bitset worklist; in zend_inference_check_recursive_dependencies() local
3965 worklist = do_alloca(sizeof(zend_ulong) * worklist_len, use_heap); in zend_inference_check_recursive_dependencies()
3966 memset(worklist, 0, sizeof(zend_ulong) * worklist_len); in zend_inference_check_recursive_dependencies()
3971 …zend_bitset_incl(worklist, info->ssa.ops[call_info->caller_call_opline - op_array->opcodes].result… in zend_inference_check_recursive_dependencies()
3975 WHILE_WORKLIST(worklist, worklist_len, i) { in zend_inference_check_recursive_dependencies()
3978 add_usages(op_array, &info->ssa, worklist, i); in zend_inference_check_recursive_dependencies()
3981 free_alloca(worklist, use_heap); in zend_inference_check_recursive_dependencies()