|
1 | 1 | skip_dict = {
|
2 |
| - "../../../../test/distributed/fsdp/test_fsdp_checkpoint.py": ( |
3 |
| - "test_checkpoint_fsdp_wrapping_cpu_offload0_offload_activations_False_use_orig_params_False", |
4 |
| - "test_checkpoint_fsdp_wrapping_cpu_offload1_offload_activations_False_use_orig_params_False", |
5 |
| - "test_checkpoint_fsdp_wrapping_cpu_offload1_offload_activations_True_use_orig_params_False", |
6 |
| - "test_checkpoint_submodule_use_reentrant_False_xpu", |
7 |
| - ), |
| 2 | + "../../../../test/distributed/fsdp/test_fsdp_checkpoint.py": None, |
8 | 3 | "../../../../test/distributed/fsdp/test_fsdp_apply.py": None,
|
9 | 4 | "../../../../test/distributed/fsdp/test_fsdp_clip_grad_norm.py": (
|
10 | 5 | "test_ddp_parity_xpu",
|
11 | 6 | ),
|
12 | 7 | "../../../../test/distributed/fsdp/test_fsdp_comm.py": None,
|
13 |
| - "../../../../test/distributed/fsdp/test_fsdp_core.py": ( |
14 |
| - "test_delayed_optim_step_offload_false_no_shard_xpu", |
15 |
| - "test_delayed_optim_step_offload_false_none_xpu", |
16 |
| - "test_delayed_optim_step_offload_false_shard_grad_op_xpu", |
17 |
| - "test_delayed_optim_step_offload_true_none_xpu", |
18 |
| - "test_delayed_optim_step_offload_true_shard_grad_op_xpu", |
19 |
| - "test_delayed_reduce_scatter_offload_false_no_shard_xpu", |
20 |
| - "test_delayed_reduce_scatter_offload_false_none_xpu", |
21 |
| - "test_delayed_reduce_scatter_offload_false_shard_grad_op_xpu", |
22 |
| - "test_delayed_reduce_scatter_offload_true_none_xpu", |
23 |
| - "test_delayed_reduce_scatter_offload_true_shard_grad_op_xpu", |
24 |
| - "test_mixture_of_experts_offload_false_no_shard_xpu", |
25 |
| - "test_mixture_of_experts_offload_false_none_xpu", |
26 |
| - "test_mixture_of_experts_offload_false_shard_grad_op_xpu", |
27 |
| - "test_mixture_of_experts_offload_true_no_shard_xpu", |
28 |
| - "test_mixture_of_experts_offload_true_none_xpu", |
29 |
| - "test_mixture_of_experts_offload_true_shard_grad_op_xpu", |
30 |
| - "test_mixture_of_experts_with_delay_before_free_offload_false_no_shard_xpu", |
31 |
| - "test_mixture_of_experts_with_delay_before_free_offload_false_none_xpu", |
32 |
| - "test_mixture_of_experts_with_delay_before_free_offload_false_shard_grad_op_xpu", |
33 |
| - "test_mixture_of_experts_with_delay_before_free_offload_true_no_shard_xpu", |
34 |
| - "test_mixture_of_experts_with_delay_before_free_offload_true_none_xpu", |
35 |
| - "test_mixture_of_experts_with_delay_before_free_offload_true_shard_grad_op_xpu", |
36 |
| - "test_nested_always_wrap_model_offload_false_no_shard_xpu", |
37 |
| - "test_nested_always_wrap_model_offload_false_none_xpu", |
38 |
| - "test_nested_always_wrap_model_offload_false_shard_grad_op_xpu", |
39 |
| - "test_nested_always_wrap_model_offload_true_none_xpu", |
40 |
| - "test_nested_always_wrap_model_offload_true_shard_grad_op_xpu", |
41 |
| - "test_nested_wrapped_model_offload_false_no_shard_xpu", |
42 |
| - "test_nested_wrapped_model_offload_false_none_xpu", |
43 |
| - "test_nested_wrapped_model_offload_false_shard_grad_op_xpu", |
44 |
| - "test_nested_wrapped_model_offload_true_none_xpu", |
45 |
| - "test_nested_wrapped_model_offload_true_shard_grad_op_xpu", |
46 |
| - "test_transformer_offload_false_no_shard_xpu", |
47 |
| - "test_transformer_offload_false_none_xpu", |
48 |
| - "test_transformer_offload_false_shard_grad_op_xpu", |
49 |
| - "test_transformer_offload_true_none_xpu", |
50 |
| - "test_transformer_offload_true_shard_grad_op_xpu", |
51 |
| - # https://github.com/intel/torch-xpu-ops/issues/1475 |
52 |
| - "test_transformer_no_grad_mixed_precision_True_xpu", |
53 |
| - "test_transformer_no_grad_mixed_precision_False_xpu", |
54 |
| - ), |
55 |
| - # Will add them back after debugging |
56 |
| - # "../../../../test/distributed/fsdp/test_fsdp_dtensor_state_dict.py": ( |
57 |
| - # "test_dtensor_sharded_model_load_state_dict_offload_to_cpu_False_is_even_sharded_model_False_xpu", |
58 |
| - # "test_dtensor_sharded_model_load_state_dict_offload_to_cpu_False_is_even_sharded_model_True_xpu", |
59 |
| - # "test_dtensor_sharded_model_load_state_dict_offload_to_cpu_True_is_even_sharded_model_False_xpu", |
60 |
| - # "test_dtensor_sharded_model_load_state_dict_offload_to_cpu_True_is_even_sharded_model_True_xpu", |
61 |
| - # "test_dtensor_sharded_optim_load_state_dict_offload_to_cpu_False_is_even_sharded_model_False_xpu", |
62 |
| - # "test_dtensor_sharded_optim_load_state_dict_offload_to_cpu_False_is_even_sharded_model_True_xpu", |
63 |
| - # "test_dtensor_sharded_optim_load_state_dict_offload_to_cpu_True_is_even_sharded_model_False_xpu", |
64 |
| - # "test_dtensor_sharded_optim_load_state_dict_offload_to_cpu_True_is_even_sharded_model_True_xpu", |
65 |
| - # "test_dtensor_sharded_tensor_state_dict_identical_offload_to_cpu_False_is_even_sharded_model_False_xpu", |
66 |
| - # "test_dtensor_sharded_tensor_state_dict_identical_offload_to_cpu_False_is_even_sharded_model_True_xpu", |
67 |
| - # "test_dtensor_sharded_tensor_state_dict_identical_offload_to_cpu_True_is_even_sharded_model_False_xpu", |
68 |
| - # "test_dtensor_sharded_tensor_state_dict_identical_offload_to_cpu_True_is_even_sharded_model_True_xpu", |
69 |
| - # "test_fsdp_init_with_device_mesh_is_even_sharded_model_False_xpu", |
70 |
| - # "test_fsdp_init_with_device_mesh_is_even_sharded_model_True_xpu", |
71 |
| - # "test_raises_warning_or_errors_xpu", |
72 |
| - # ), |
73 |
| - "../../../../test/distributed/fsdp/test_fsdp_exec_order.py": ( |
74 |
| - "test_invalid_first_iter_order_sharding_strategy1_xpu", |
75 |
| - "test_train_eval_sharding_strategy1_xpu", |
76 |
| - ), |
| 8 | + "../../../../test/distributed/fsdp/test_fsdp_comm_hooks.py": None, |
| 9 | + "../../../../test/distributed/fsdp/test_fsdp_core.py": None, |
| 10 | + "../../../../test/distributed/fsdp/test_fsdp_exec_order.py": None, |
77 | 11 | "../../../../test/distributed/fsdp/test_fsdp_fine_tune.py": (
|
78 | 12 | "test_parity_with_non_frozen_fsdp_xpu",
|
79 | 13 | "test_parity_with_ddp_xpu",
|
80 | 14 | ),
|
81 | 15 | "../../../../test/distributed/fsdp/test_fsdp_fx.py": None,
|
82 |
| - # will bring back after oneccl upgrade to 2021.16.1 |
83 |
| - # "../../../../test/distributed/fsdp/test_fsdp_input.py": None, |
| 16 | + "../../../../test/distributed/fsdp/test_fsdp_input.py": None, |
84 | 17 | "../../../../test/distributed/fsdp/test_fsdp_multiple_forward.py": None,
|
85 | 18 | "../../../../test/distributed/fsdp/test_fsdp_multiple_wrapping.py": (
|
86 | 19 | "test_transformer_no_grad_mixed_precision_True_xpu",
|
87 | 20 | ),
|
88 | 21 | "../../../../test/distributed/fsdp/test_fsdp_uneven.py": None,
|
89 |
| - # Will add them back after debugging |
90 |
| - # "../../../../test/distributed/fsdp/test_hsdp_dtensor_state_dict.py": ( |
91 |
| - # "test_dtensor_sharded_model_load_state_dict_offload_to_cpu_False_xpu", |
92 |
| - # "test_dtensor_sharded_model_load_state_dict_offload_to_cpu_True_xpu", |
93 |
| - # "test_dtensor_sharded_optim_load_state_dict_offload_to_cpu_False_xpu", |
94 |
| - # "test_dtensor_sharded_optim_load_state_dict_offload_to_cpu_True_xpu", |
95 |
| - # "test_dtensor_sharded_tensor_state_dict_identical_offload_to_cpu_False_xpu", |
96 |
| - # "test_dtensor_sharded_tensor_state_dict_identical_offload_to_cpu_True_xpu", |
97 |
| - # "test_hsdp_init_with_device_mesh_xpu", |
98 |
| - # "test_root_module_is_not_FSDP_xpu", |
99 |
| - # ), |
100 | 22 | "../../../../test/distributed/fsdp/test_utils.py": None,
|
101 | 23 | "distributed/test_c10d_xccl.py": (
|
102 |
| - # will bring back after oneccl upgrade to 2021.16.1 |
103 |
| - "test_xccl_barrier", |
| 24 | + # https://github.com/intel/torch-xpu-ops/issues/2046 |
| 25 | + "test_unwaited", |
| 26 | + ), |
| 27 | + "distributed/test_c10d_ops_xccl.py": None, |
| 28 | + "../../../../test/distributed/fsdp/test_fsdp_misc.py": None, |
| 29 | + "../../../../test/distributed/test_functional_api.py": ( |
| 30 | + # depends on https://github.com/pytorch/pytorch/pull/159473 |
| 31 | + "test_tracing_with_fakepg_xpu", |
104 | 32 | ),
|
| 33 | + "../../../../test/distributed/_tools/test_fsdp2_mem_tracker.py": None, |
| 34 | + "../../../../test/distributed/_tools/test_mem_tracker.py": None, |
| 35 | + "../../../../test/distributed/_tools/test_memory_tracker.py": None, |
105 | 36 | }
|
0 commit comments