Skip to content

Commit

Permalink
[Doc] Add Nemotron to supported model docs (vllm-project#6843)
Browse files Browse the repository at this point in the history
  • Loading branch information
mgoin authored Jul 26, 2024
1 parent 3bbb493 commit 281977b
Show file tree
Hide file tree
Showing 2 changed files with 5 additions and 3 deletions.
4 changes: 4 additions & 0 deletions docs/source/models/supported_models.rst
Original file line number Diff line number Diff line change
Expand Up @@ -113,6 +113,10 @@ Decoder-only Language Models
- MPT, MPT-Instruct, MPT-Chat, MPT-StoryWriter
- :code:`mosaicml/mpt-7b`, :code:`mosaicml/mpt-7b-storywriter`, :code:`mosaicml/mpt-30b`, etc.
-
* - :code:`NemotronForCausalLM`
- Nemotron-3, Nemotron-4, Minitron
- :code:`nvidia/Minitron-8B-Base`, :code:`mgoin/Nemotron-4-340B-Base-hf-FP8`, etc.
- ✅︎
* - :code:`OLMoForCausalLM`
- OLMo
- :code:`allenai/OLMo-1B-hf`, :code:`allenai/OLMo-7B-hf`, etc.
Expand Down
4 changes: 1 addition & 3 deletions vllm/model_executor/layers/activation.py
Original file line number Diff line number Diff line change
Expand Up @@ -166,9 +166,7 @@ class ReLUSquaredActivation(CustomOp):

def forward_native(self, x: torch.Tensor) -> torch.Tensor:
"""PyTorch-native implementation equivalent to forward()."""
relu_applied = nn.functional.relu(x)
squared = torch.square(relu_applied)
return squared
return torch.square(F.relu(x))

def forward_cuda(self, x: torch.Tensor) -> torch.Tensor:
return self.forward_native(x)
Expand Down

0 comments on commit 281977b

Please sign in to comment.