Skip to content

Conversation

moqingyan
Copy link

Description:

This PR adds support for running the connected components CUDA extension on multiple GPUs in parallel.
Motivation:
Currently, the extension only supports single-GPU execution. Enabling multi-GPU allows for larger batch processing and improved performance in distributed training setups.

Changes:

Refactored kernel launch code to select devices dynamically.
Updated get_connected_componnets to accept a device argument and process batches in parallel across available GPUs.

@facebook-github-bot
Copy link

Hi @moqingyan!

Thank you for your pull request and welcome to our community.

Action Required

In order to merge any pull request (code, docs, etc.), we require contributors to sign our Contributor License Agreement, and we don't seem to have one on file for you.

Process

In order for us to review and merge your suggested changes, please sign at https://code.facebook.com/cla. If you are contributing on behalf of someone else (eg your employer), the individual CLA may not be sufficient and your employer may need to sign the corporate CLA.

Once the CLA is signed, our tooling will perform checks and validations. Afterwards, the pull request will be tagged with CLA signed. The tagging process may take up to 1 hour after signing. Please give it that time before contacting us about it.

If you have received this in error or have any questions, please contact us at [email protected]. Thanks!

@facebook-github-bot
Copy link

Thank you for signing our Contributor License Agreement. We can now accept your code for this (and any) Meta Open Source project. Thanks!

@hjj-lmx
Copy link

hjj-lmx commented Aug 4, 2025

描述:

此 PR 增加了在多个 GPU 上并行运行 Connected Components CUDA 扩展的支持。 动机: 目前,该扩展仅支持单 GPU 执行。启用多 GPU 可以实现更大的批量处理,并提升分布式训练设置的性能。

变化:

重构了内核启动代码,使其能够动态选择设备。 更新了 get_connected_componnets,使其能够接受设备参数,并在可用的 GPU 上并行处理批次。

这是针对训练还是推理的多GPU

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants