Skip to content

Commit

Permalink
improve docs
Browse files Browse the repository at this point in the history
  • Loading branch information
merrymercy committed Oct 9, 2023
1 parent 85833d4 commit 0adf31c
Show file tree
Hide file tree
Showing 5 changed files with 10 additions and 9 deletions.
6 changes: 3 additions & 3 deletions fastchat/model/model_adapter.py
Original file line number Diff line number Diff line change
Expand Up @@ -27,9 +27,6 @@
)

from fastchat.constants import CPU_ISA
from fastchat.modules.gptq import GptqConfig, load_gptq_quantized
from fastchat.modules.awq import AWQConfig, load_awq_quantized
from fastchat.modules.exllama import ExllamaConfig, load_exllama_model
from fastchat.conversation import Conversation, get_conv_template
from fastchat.model.compression import load_compress_model
from fastchat.model.llama_condense_monkey_patch import replace_llama_with_condense
Expand All @@ -40,6 +37,9 @@
from fastchat.model.monkey_patch_non_inplace import (
replace_llama_attn_with_non_inplace_operations,
)
from fastchat.modules.awq import AWQConfig, load_awq_quantized
from fastchat.modules.exllama import ExllamaConfig, load_exllama_model
from fastchat.modules.gptq import GptqConfig, load_gptq_quantized
from fastchat.utils import get_gpu_memory

# Check an environment variable to check if we should be sharing Peft model
Expand Down
5 changes: 3 additions & 2 deletions fastchat/model/model_exllama.py
Original file line number Diff line number Diff line change
@@ -1,8 +1,9 @@
import sys
import torch
import gc
import sys
from typing import Dict

import torch


def generate_stream_exllama(
model,
Expand Down
2 changes: 1 addition & 1 deletion fastchat/serve/cli.py
Original file line number Diff line number Diff line change
Expand Up @@ -29,9 +29,9 @@
import torch

from fastchat.model.model_adapter import add_model_args
from fastchat.modules.gptq import GptqConfig
from fastchat.modules.awq import AWQConfig
from fastchat.modules.exllama import ExllamaConfig
from fastchat.modules.gptq import GptqConfig
from fastchat.serve.inference import ChatIO, chat_loop
from fastchat.utils import str_to_torch_dtype

Expand Down
2 changes: 1 addition & 1 deletion fastchat/serve/inference.py
Original file line number Diff line number Diff line change
Expand Up @@ -35,8 +35,8 @@
get_conversation_template,
get_generate_stream_function,
)
from fastchat.modules.gptq import GptqConfig
from fastchat.modules.awq import AWQConfig
from fastchat.modules.gptq import GptqConfig
from fastchat.modules.exllama import ExllamaConfig
from fastchat.utils import is_partial_stop, is_sentence_complete, get_context_length

Expand Down
4 changes: 2 additions & 2 deletions fastchat/serve/model_worker.py
Original file line number Diff line number Diff line change
Expand Up @@ -45,15 +45,15 @@
get_conversation_template,
get_generate_stream_function,
)
from fastchat.modules.gptq import GptqConfig
from fastchat.modules.awq import AWQConfig
from fastchat.modules.exllama import ExllamaConfig
from fastchat.modules.gptq import GptqConfig
from fastchat.utils import (
build_logger,
pretty_print_semaphore,
get_context_length,
str_to_torch_dtype,
)
from fastchat.modules.exllama import ExllamaConfig
from fastchat.utils import build_logger, pretty_print_semaphore, get_context_length


Expand Down

0 comments on commit 0adf31c

Please sign in to comment.