|
31 | 31 | from typing import Any, Optional
|
32 | 32 |
|
33 | 33 | import numpy as np
|
34 |
| -from backend_request_func import (ASYNC_REQUEST_FUNCS, |
35 |
| - OPENAI_COMPATIBLE_BACKENDS, RequestFuncInput, |
36 |
| - RequestFuncOutput, get_tokenizer) |
37 |
| -from benchmark_dataset import (AIMODataset, BurstGPTDataset, |
38 |
| - ConversationDataset, HuggingFaceDataset, |
39 |
| - InstructCoderDataset, RandomDataset, |
40 |
| - SampleRequest, ShareGPTDataset, SonnetDataset, |
41 |
| - VisionArenaDataset) |
42 |
| -from benchmark_utils import convert_to_pytorch_benchmark_format, write_to_json |
43 | 34 | from tqdm.asyncio import tqdm
|
44 | 35 | from transformers import PreTrainedTokenizerBase
|
45 | 36 |
|
| 37 | +from .backend_request_func import (ASYNC_REQUEST_FUNCS, |
| 38 | + OPENAI_COMPATIBLE_BACKENDS, RequestFuncInput, |
| 39 | + RequestFuncOutput, get_tokenizer) |
| 40 | +from .benchmark_dataset import (AIMODataset, BurstGPTDataset, |
| 41 | + ConversationDataset, HuggingFaceDataset, |
| 42 | + InstructCoderDataset, RandomDataset, |
| 43 | + SampleRequest, ShareGPTDataset, SonnetDataset, |
| 44 | + VisionArenaDataset) |
| 45 | +from .benchmark_utils import convert_to_pytorch_benchmark_format, write_to_json |
| 46 | + |
46 | 47 | MILLISECONDS_TO_SECONDS_CONVERSION = 1000
|
47 | 48 |
|
48 | 49 |
|
|
0 commit comments