Skip to content

Commit

Permalink
fix(gpu): fix multi-gpu error in division
Browse files Browse the repository at this point in the history
  • Loading branch information
agnesLeroy committed Aug 2, 2024
1 parent 5547d92 commit 16c19ea
Show file tree
Hide file tree
Showing 3 changed files with 5 additions and 5 deletions.
4 changes: 2 additions & 2 deletions .github/workflows/integer_multi_bit_multi_gpu_benchmark.yml
Original file line number Diff line number Diff line change
Expand Up @@ -49,7 +49,7 @@ jobs:
slab-url: ${{ secrets.SLAB_BASE_URL }}
job-secret: ${{ secrets.JOB_SECRET }}
backend: hyperstack
profile: multi-h100
profile: multi-h100-nvlink

cuda-integer-multi-bit-multi-gpu-benchmarks:
name: Execute multi GPU integer multi-bit benchmarks
Expand Down Expand Up @@ -153,7 +153,7 @@ jobs:
run: |
python3 ./ci/benchmark_parser.py target/criterion ${{ env.RESULTS_FILENAME }} \
--database tfhe_rs \
--hardware "n3-H100x8" \
--hardware "n3-H100x8-NVLink" \
--backend gpu \
--project-version "${{ env.COMMIT_HASH }}" \
--branch ${{ github.ref_name }} \
Expand Down
4 changes: 2 additions & 2 deletions .github/workflows/integer_multi_gpu_full_benchmark.yml
Original file line number Diff line number Diff line change
Expand Up @@ -36,7 +36,7 @@ jobs:
slab-url: ${{ secrets.SLAB_BASE_URL }}
job-secret: ${{ secrets.JOB_SECRET }}
backend: hyperstack
profile: multi-h100
profile: multi-h100-nvlink

cuda-integer-full-multi-gpu-benchmarks:
name: Execute multi GPU integer benchmarks for all operations flavor
Expand Down Expand Up @@ -133,7 +133,7 @@ jobs:
run: |
python3 ./ci/benchmark_parser.py target/criterion ${{ env.RESULTS_FILENAME }} \
--database tfhe_rs \
--hardware "n3-H100x8" \
--hardware "n3-H100x8-NVLink" \
--backend gpu \
--project-version "${{ env.COMMIT_HASH }}" \
--branch ${{ github.ref_name }} \
Expand Down
2 changes: 1 addition & 1 deletion backends/tfhe-cuda-backend/cuda/include/integer.h
Original file line number Diff line number Diff line change
Expand Up @@ -2796,7 +2796,7 @@ template <typename Torus> struct int_div_rem_memory {
int_div_rem_memory(cudaStream_t *streams, uint32_t *gpu_indexes,
uint32_t gpu_count, int_radix_params params,
uint32_t num_blocks, bool allocate_gpu_memory) {
active_gpu_count = get_active_gpu_count(num_blocks, gpu_count);
active_gpu_count = get_active_gpu_count(2 * num_blocks, gpu_count);

this->params = params;
shift_mem_1 = new int_logical_scalar_shift_buffer<Torus>(
Expand Down

0 comments on commit 16c19ea

Please sign in to comment.