[Doc] Add Nemotron to supported model docs (#6843)
This commit is contained in:
parent
3bbb4936dc
commit
281977bd6e
@ -113,6 +113,10 @@ Decoder-only Language Models
|
|||||||
- MPT, MPT-Instruct, MPT-Chat, MPT-StoryWriter
|
- MPT, MPT-Instruct, MPT-Chat, MPT-StoryWriter
|
||||||
- :code:`mosaicml/mpt-7b`, :code:`mosaicml/mpt-7b-storywriter`, :code:`mosaicml/mpt-30b`, etc.
|
- :code:`mosaicml/mpt-7b`, :code:`mosaicml/mpt-7b-storywriter`, :code:`mosaicml/mpt-30b`, etc.
|
||||||
-
|
-
|
||||||
|
* - :code:`NemotronForCausalLM`
|
||||||
|
- Nemotron-3, Nemotron-4, Minitron
|
||||||
|
- :code:`nvidia/Minitron-8B-Base`, :code:`mgoin/Nemotron-4-340B-Base-hf-FP8`, etc.
|
||||||
|
- ✅︎
|
||||||
* - :code:`OLMoForCausalLM`
|
* - :code:`OLMoForCausalLM`
|
||||||
- OLMo
|
- OLMo
|
||||||
- :code:`allenai/OLMo-1B-hf`, :code:`allenai/OLMo-7B-hf`, etc.
|
- :code:`allenai/OLMo-1B-hf`, :code:`allenai/OLMo-7B-hf`, etc.
|
||||||
|
@ -166,9 +166,7 @@ class ReLUSquaredActivation(CustomOp):
|
|||||||
|
|
||||||
def forward_native(self, x: torch.Tensor) -> torch.Tensor:
|
def forward_native(self, x: torch.Tensor) -> torch.Tensor:
|
||||||
"""PyTorch-native implementation equivalent to forward()."""
|
"""PyTorch-native implementation equivalent to forward()."""
|
||||||
relu_applied = nn.functional.relu(x)
|
return torch.square(F.relu(x))
|
||||||
squared = torch.square(relu_applied)
|
|
||||||
return squared
|
|
||||||
|
|
||||||
def forward_cuda(self, x: torch.Tensor) -> torch.Tensor:
|
def forward_cuda(self, x: torch.Tensor) -> torch.Tensor:
|
||||||
return self.forward_native(x)
|
return self.forward_native(x)
|
||||||
|
Loading…
x
Reference in New Issue
Block a user