|
1 | 1 | skip_dict = {
|
2 |
| - "../../../../test/distributed/fsdp/test_fsdp_checkpoint.py": ( |
3 |
| - "test_checkpoint_fsdp_wrapping_cpu_offload0_offload_activations_False_use_orig_params_False", |
4 |
| - "test_checkpoint_fsdp_wrapping_cpu_offload1_offload_activations_False_use_orig_params_False", |
5 |
| - "test_checkpoint_fsdp_wrapping_cpu_offload1_offload_activations_True_use_orig_params_False", |
6 |
| - "test_checkpoint_submodule_use_reentrant_False_xpu", |
7 |
| - ), |
| 2 | + "../../../../test/distributed/fsdp/test_fsdp_checkpoint.py": None, |
8 | 3 | "../../../../test/distributed/fsdp/test_fsdp_apply.py": None,
|
9 | 4 | "../../../../test/distributed/fsdp/test_fsdp_clip_grad_norm.py": (
|
10 | 5 | "test_ddp_parity_xpu",
|
11 | 6 | ),
|
12 | 7 | "../../../../test/distributed/fsdp/test_fsdp_comm.py": None,
|
13 |
| - "../../../../test/distributed/fsdp/test_fsdp_core.py": ( |
14 |
| - "test_delayed_optim_step_offload_false_no_shard_xpu", |
15 |
| - "test_delayed_optim_step_offload_false_none_xpu", |
16 |
| - "test_delayed_optim_step_offload_false_shard_grad_op_xpu", |
17 |
| - "test_delayed_optim_step_offload_true_none_xpu", |
18 |
| - "test_delayed_optim_step_offload_true_shard_grad_op_xpu", |
19 |
| - "test_delayed_reduce_scatter_offload_false_no_shard_xpu", |
20 |
| - "test_delayed_reduce_scatter_offload_false_none_xpu", |
21 |
| - "test_delayed_reduce_scatter_offload_false_shard_grad_op_xpu", |
22 |
| - "test_delayed_reduce_scatter_offload_true_none_xpu", |
23 |
| - "test_delayed_reduce_scatter_offload_true_shard_grad_op_xpu", |
24 |
| - "test_mixture_of_experts_offload_false_no_shard_xpu", |
25 |
| - "test_mixture_of_experts_offload_false_none_xpu", |
26 |
| - "test_mixture_of_experts_offload_false_shard_grad_op_xpu", |
27 |
| - "test_mixture_of_experts_offload_true_no_shard_xpu", |
28 |
| - "test_mixture_of_experts_offload_true_none_xpu", |
29 |
| - "test_mixture_of_experts_offload_true_shard_grad_op_xpu", |
30 |
| - "test_mixture_of_experts_with_delay_before_free_offload_false_no_shard_xpu", |
31 |
| - "test_mixture_of_experts_with_delay_before_free_offload_false_none_xpu", |
32 |
| - "test_mixture_of_experts_with_delay_before_free_offload_false_shard_grad_op_xpu", |
33 |
| - "test_mixture_of_experts_with_delay_before_free_offload_true_no_shard_xpu", |
34 |
| - "test_mixture_of_experts_with_delay_before_free_offload_true_none_xpu", |
35 |
| - "test_mixture_of_experts_with_delay_before_free_offload_true_shard_grad_op_xpu", |
36 |
| - "test_nested_always_wrap_model_offload_false_no_shard_xpu", |
37 |
| - "test_nested_always_wrap_model_offload_false_none_xpu", |
38 |
| - "test_nested_always_wrap_model_offload_false_shard_grad_op_xpu", |
39 |
| - "test_nested_always_wrap_model_offload_true_none_xpu", |
40 |
| - "test_nested_always_wrap_model_offload_true_shard_grad_op_xpu", |
41 |
| - "test_nested_wrapped_model_offload_false_no_shard_xpu", |
42 |
| - "test_nested_wrapped_model_offload_false_none_xpu", |
43 |
| - "test_nested_wrapped_model_offload_false_shard_grad_op_xpu", |
44 |
| - "test_nested_wrapped_model_offload_true_none_xpu", |
45 |
| - "test_nested_wrapped_model_offload_true_shard_grad_op_xpu", |
46 |
| - "test_transformer_offload_false_no_shard_xpu", |
47 |
| - "test_transformer_offload_false_none_xpu", |
48 |
| - "test_transformer_offload_false_shard_grad_op_xpu", |
49 |
| - "test_transformer_offload_true_none_xpu", |
50 |
| - "test_transformer_offload_true_shard_grad_op_xpu", |
51 |
| - # https://github.com/intel/torch-xpu-ops/issues/1475 |
52 |
| - "test_transformer_no_grad_mixed_precision_True_xpu", |
53 |
| - "test_transformer_no_grad_mixed_precision_False_xpu", |
54 |
| - ), |
| 8 | + "../../../../test/distributed/fsdp/test_fsdp_comm_hooks.py": None, |
| 9 | + "../../../../test/distributed/fsdp/test_fsdp_core.py": None, |
55 | 10 | # Will add them back after debugging
|
56 | 11 | # "../../../../test/distributed/fsdp/test_fsdp_dtensor_state_dict.py": (
|
57 | 12 | # "test_dtensor_sharded_model_load_state_dict_offload_to_cpu_False_is_even_sharded_model_False_xpu",
|
|
70 | 25 | # "test_fsdp_init_with_device_mesh_is_even_sharded_model_True_xpu",
|
71 | 26 | # "test_raises_warning_or_errors_xpu",
|
72 | 27 | # ),
|
73 |
| - "../../../../test/distributed/fsdp/test_fsdp_exec_order.py": ( |
74 |
| - "test_invalid_first_iter_order_sharding_strategy1_xpu", |
75 |
| - "test_train_eval_sharding_strategy1_xpu", |
76 |
| - ), |
| 28 | + "../../../../test/distributed/fsdp/test_fsdp_exec_order.py": None, |
77 | 29 | "../../../../test/distributed/fsdp/test_fsdp_fine_tune.py": (
|
78 | 30 | "test_parity_with_non_frozen_fsdp_xpu",
|
79 | 31 | "test_parity_with_ddp_xpu",
|
80 | 32 | ),
|
81 | 33 | "../../../../test/distributed/fsdp/test_fsdp_fx.py": None,
|
82 |
| - # will bring back after oneccl upgrade to 2021.16.1 |
83 |
| - # "../../../../test/distributed/fsdp/test_fsdp_input.py": None, |
| 34 | + "../../../../test/distributed/fsdp/test_fsdp_input.py": None, |
84 | 35 | "../../../../test/distributed/fsdp/test_fsdp_multiple_forward.py": None,
|
85 | 36 | "../../../../test/distributed/fsdp/test_fsdp_multiple_wrapping.py": (
|
86 | 37 | "test_transformer_no_grad_mixed_precision_True_xpu",
|
|
98 | 49 | # "test_root_module_is_not_FSDP_xpu",
|
99 | 50 | # ),
|
100 | 51 | "../../../../test/distributed/fsdp/test_utils.py": None,
|
101 |
| - "distributed/test_c10d_xccl.py": ( |
102 |
| - # will bring back after oneccl upgrade to 2021.16.1 |
103 |
| - "test_xccl_barrier", |
| 52 | + "distributed/test_c10d_xccl.py": None, |
| 53 | + "../../../../test/distributed/fsdp/test_fsdp_misc.py": None, |
| 54 | + "../../../../test/distributed/test_functional_api.py": ( |
| 55 | + # depends on https://github.com/pytorch/pytorch/pull/159473 |
| 56 | + "test_tracing_with_fakepg_xpu", |
104 | 57 | ),
|
| 58 | + "../../../../test/distributed/_tools/test_fsdp2_mem_tracker.py": None, |
| 59 | + "../../../../test/distributed/_tools/test_mem_tracker.py": None, |
| 60 | + "../../../../test/distributed/_tools/test_memory_tracker.py": None, |
105 | 61 | }
|
0 commit comments