Skip to content

Commit

Permalink
remove some comments
Browse files Browse the repository at this point in the history
  • Loading branch information
xffxff committed Nov 21, 2024
1 parent eba7955 commit e40ac1a
Showing 1 changed file with 1 addition and 14 deletions.
15 changes: 1 addition & 14 deletions aria/vllm/aria.py
Original file line number Diff line number Diff line change
Expand Up @@ -637,9 +637,6 @@ class MoELayer(nn.Module):
This layer implements the MoE mechanism, which routes input tokens to different experts
based on a routing algorithm, processes them through the experts, and then combines
the outputs.
Args:
config (AriaMoELMConfig): Configuration object for the MoE layer.
"""

def __init__(
Expand Down Expand Up @@ -680,10 +677,6 @@ class MoEDecoderLayer(LlamaDecoderLayer):
"""
Custom Decoder Layer for the AriaMoE model which modifies the standard `LlamaDecoderLayer` by
replacing the traditional MLP with a Mixture of Experts (MoE) Layer.
Args:
config (LlamaConfig): Configuration object for the layer.
layer_idx (int): Index of the current layer in the model.
"""

def __init__(
Expand Down Expand Up @@ -736,12 +729,6 @@ class AriaMoELMModel(LlamaModel):
"""
Custom LlamaModel for the AriaMoE model which modifies the standard LlamaModel by
replacing the `LlamaDecoderLayer` with `MoEDecoderLayer`.
This model implements a Mixture of Experts (MoE) approach, where each layer contains
multiple expert networks that specialize in different aspects of the input.
Args:
config (LlamaConfig): Configuration object for the model.
"""

def __init__(self, *, vllm_config: VllmConfig, prefix: str = ""):
Expand All @@ -751,7 +738,7 @@ def __init__(self, *, vllm_config: VllmConfig, prefix: str = ""):
cache_config = vllm_config.cache_config
quant_config = vllm_config.quant_config

# FIXME(zhoufan): this is a hack to avoid the error: AttributeError: 'AriaMoELMModel' object has no attribute 'do_not_compile'.
# FIXME: this is a hack to disable the compilation of the model
self.do_not_compile = True

self.layers = None
Expand Down

0 comments on commit e40ac1a

Please sign in to comment.