Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
1 change: 1 addition & 0 deletions .github/workflows/_linux_build.yml
Original file line number Diff line number Diff line change
Expand Up @@ -94,6 +94,7 @@ jobs:
- name: Build Pytorch on ${{ needs.runner.outputs.hostname }}
run: |
export USE_XCCL=1
export IS_XPU_CI=1
# only build pvc for CI
if [ "${{ github.event_name }}" == "pull_request" ];then
export TORCH_XPU_ARCH_LIST='pvc'
Expand Down
3 changes: 3 additions & 0 deletions CMakeLists.txt
Original file line number Diff line number Diff line change
Expand Up @@ -25,6 +25,9 @@ set(PROJECT_VERSION "2.3.0")
# Avoid SYCL compiler error
if(NOT WIN32)
string(APPEND CMAKE_CXX_FLAGS " -Wno-error")
if("$ENV{IS_XPU_CI}" STREQUAL "1")
string(APPEND CMAKE_CXX_FLAGS " -Werror=unused-variable")
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Should we always turn on this build option?

endif()
endif()

cmake_policy(SET CMP0048 NEW)
Expand Down
3 changes: 0 additions & 3 deletions src/ATen/native/xpu/sycl/LayerNormKernels.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -1103,9 +1103,6 @@ void _layer_norm_backward_kernel(
// affecting performance and behavior.
const scalar_t* dY_data = dY.const_data_ptr<scalar_t>();
const scalar_t* X_data = X.const_data_ptr<scalar_t>();
weight_t* dg_data =
dgamma.defined() ? dgamma.data_ptr<weight_t>() : nullptr;
weight_t* db_data = dbeta.defined() ? dbeta.data_ptr<weight_t>() : nullptr;
Tensor dgamma_blocks;
Tensor dbeta_blocks;
weight_t* dgamma_blocks_ptr = nullptr;
Expand Down
1 change: 0 additions & 1 deletion src/ATen/native/xpu/sycl/RoiAlignKernels.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -440,7 +440,6 @@ Tensor roi_align_kernel(

at::Tensor output = at::zeros(
{num_rois, channels, pooled_height, pooled_width}, input.options());
auto output_size = num_rois * pooled_height * pooled_width * channels;

if (output.numel() == 0) {
return output;
Expand Down
Loading