Change the name to vLLM (#150)
This commit is contained in:
@ -1,7 +1,7 @@
|
||||
import torch
|
||||
import torch.nn.functional as F
|
||||
|
||||
from cacheflow import activation_ops
|
||||
from vllm import activation_ops
|
||||
|
||||
|
||||
def ref_silu_and_mul(x: torch.Tensor) -> torch.Tensor:
|
||||
|
||||
@ -5,7 +5,7 @@ import torch
|
||||
from xformers import ops as xops
|
||||
from xformers.ops.fmha.attn_bias import BlockDiagonalCausalMask
|
||||
|
||||
from cacheflow import attention_ops
|
||||
from vllm import attention_ops
|
||||
|
||||
MAX_SEQ_LEN = 4096
|
||||
TEST_SEED = 0
|
||||
|
||||
@ -2,7 +2,7 @@ import random
|
||||
|
||||
import torch
|
||||
|
||||
from cacheflow import cache_ops
|
||||
from vllm import cache_ops
|
||||
|
||||
|
||||
@torch.inference_mode()
|
||||
|
||||
@ -1,7 +1,7 @@
|
||||
import torch
|
||||
import torch.nn as nn
|
||||
|
||||
from cacheflow import layernorm_ops
|
||||
from vllm import layernorm_ops
|
||||
|
||||
|
||||
class RefRMSNorm(nn.Module):
|
||||
|
||||
@ -4,7 +4,7 @@ import torch
|
||||
import torch.nn as nn
|
||||
import torch.nn.functional as F
|
||||
|
||||
from cacheflow import pos_encoding_ops
|
||||
from vllm import pos_encoding_ops
|
||||
|
||||
|
||||
def rotate_half(x: torch.Tensor) -> torch.Tensor:
|
||||
|
||||
Reference in New Issue
Block a user