Skip to content
This repository has been archived by the owner on Nov 4, 2024. It is now read-only.

Commit

Permalink
fix: fix the typos
Browse files Browse the repository at this point in the history
  • Loading branch information
avik-pal committed Jul 4, 2024
1 parent c2e1b18 commit 1f58ea7
Show file tree
Hide file tree
Showing 4 changed files with 8 additions and 3 deletions.
5 changes: 5 additions & 0 deletions .typos.toml
Original file line number Diff line number Diff line change
@@ -0,0 +1,5 @@
[default.extend-words]
numer = "numer"
nd = "nd"
Ba = "Ba"

2 changes: 1 addition & 1 deletion ext/LuxLibCUDAExt/cublaslt.jl
Original file line number Diff line number Diff line change
Expand Up @@ -125,7 +125,7 @@ function LuxLib._cublaslt_matmul_fused!(
lthandle = Ref{CUBLAS.cublasLtHandle_t}()
CUBLAS.cublasLtCreate(lthandle)

# Seach for the best algorithm
# Search for the best algorithm
heuristic = Ref{CUBLAS.cublasLtMatmulHeuristicResult_t}()
returnedResults = Ref{Cint}(0)
CUBLAS.cublasLtMatmulAlgoGetHeuristic(
Expand Down
2 changes: 1 addition & 1 deletion ext/LuxLibForwardDiffExt.jl
Original file line number Diff line number Diff line change
Expand Up @@ -12,7 +12,7 @@ LuxLib.__has_dual(::AbstractArray{<:ForwardDiff.Dual}) = true
return ForwardDiff.valtype(eltype(x))
end

# Convolutions: We might want to capture these furthur down in `conv!`
# Convolutions: We might want to capture these further down in `conv!`
# NOTE: In principle we can concatenate all of the partials along the batch dimension
# and cut down substantially on the time to compute jacobians.
# Here we should be broadcasting with `Tag` for safety but that breaks GPU compilation.
Expand Down
2 changes: 1 addition & 1 deletion src/utils.jl
Original file line number Diff line number Diff line change
Expand Up @@ -27,7 +27,7 @@ EnzymeRules.inactive_noinl(::typeof(_copy_autodiff_barrier), ::Any...) = nothing
__is_tracked(x) = x == :TrackedArray || x == :TrackedVector
__is_tracked(args...) = any(__is_tracked, args)

# Droping ForwardDiff Gradients
# Dropping ForwardDiff Gradients
function _drop_forwarddiff_partials end

_drop_forwarddiff_partials(x::AbstractArray) = x
Expand Down

0 comments on commit 1f58ea7

Please sign in to comment.