|
110 | 110 | "metadata": {},
|
111 | 111 | "outputs": [],
|
112 | 112 | "source": [
|
113 |
| - "from fate_llm.algo.inferdpt.inference.api import APICompletionInference\n", |
| 113 | + "from fate_llm.inference.api import APICompletionInference\n", |
114 | 114 | "from fate_llm.algo.inferdpt import inferdpt\n",
|
115 | 115 | "from fate_llm.algo.inferdpt.utils import InferDPTKit\n",
|
116 |
| - "from fate_llm.algo.inferdpt.inferdpt import InferDPTClient, InferDPTServer\n", |
117 |
| - "from jinja2 import Template\n", |
118 |
| - "from fate.arch import Context\n", |
119 | 116 | "import sys\n",
|
120 | 117 | "\n",
|
121 | 118 | "arbiter = (\"arbiter\", 10000)\n",
|
|
225 | 222 | "metadata": {},
|
226 | 223 | "outputs": [],
|
227 | 224 | "source": [
|
228 |
| - "from fate_llm.algo.inferdpt.utils import InferDPTKit\n", |
229 |
| - "from fate_llm.algo.inferdpt.inferdpt import InferDPTClient, InferDPTServer\n", |
230 |
| - "from jinja2 import Template\n", |
231 |
| - "from fate.arch import Context\n", |
| 225 | + "from fate_llm.algo.inferdpt.inferdpt import InferDPTServer\n", |
232 | 226 | "import sys\n",
|
233 |
| - "from fate_llm.algo.inferdpt.inference.api import APICompletionInference\n", |
| 227 | + "from fate_llm.inference.api import APICompletionInference\n", |
234 | 228 | "\n",
|
235 | 229 | "arbiter = (\"arbiter\", 10000)\n",
|
236 | 230 | "guest = (\"guest\", 10000)\n",
|
|
297 | 291 | },
|
298 | 292 | "outputs": [],
|
299 | 293 | "source": [
|
300 |
| - "from fate_llm.algo.inferdpt.inference.api import APICompletionInference\n", |
| 294 | + "from fate_llm.inference.api import APICompletionInference\n", |
301 | 295 | "from fate_llm.algo.pdss.encoder_decoder.slm_encoder_decoder import SLMEncoderDecoderClient\n",
|
302 | 296 | "\n",
|
303 | 297 | "arbiter = (\"arbiter\", 10000)\n",
|
|
407 | 401 | "metadata": {},
|
408 | 402 | "outputs": [],
|
409 | 403 | "source": [
|
410 |
| - "from fate_llm.algo.inferdpt.inference.api import APICompletionInference\n", |
| 404 | + "from fate_llm.inference.api import APICompletionInference\n", |
411 | 405 | "from fate_llm.algo.pdss.encoder_decoder.slm_encoder_decoder import SLMEncoderDecoderServer\n",
|
412 | 406 | "\n",
|
413 | 407 | "arbiter = (\"arbiter\", 10000)\n",
|
|
833 | 827 | "source": [
|
834 | 828 | "from fate_llm.algo.inferdpt.inferdpt import InferDPTServer\n",
|
835 | 829 | "from fate_llm.algo.pdss.pdss_trainer import PDSSTraineServer\n",
|
836 |
| - "from jinja2 import Template\n", |
837 |
| - "from fate.arch import Context\n", |
838 | 830 | "import sys\n",
|
839 | 831 | "\n",
|
840 | 832 | "\n",
|
|
954 | 946 | "metadata": {},
|
955 | 947 | "outputs": [],
|
956 | 948 | "source": [
|
957 |
| - "from fate_llm.runner.pdss_runner import PDSSRunner\n", |
958 |
| - "from fate.components.components.nn.nn_runner import loader_load_from_conf\n", |
959 |
| - "from fate.components.components.nn.loader import Loader\n", |
960 |
| - "from fate_llm.dataset.pdss_dataset import PrefixDataset\n", |
961 |
| - "from fate_client.pipeline.components.fate.nn.loader import ModelLoader, DatasetLoader, CustFuncLoader, Loader\n", |
962 |
| - "from transformers import (\n", |
963 |
| - " AutoConfig,\n", |
964 |
| - " AutoModel,\n", |
965 |
| - " AutoTokenizer,\n", |
966 |
| - " DataCollatorForSeq2Seq,\n", |
967 |
| - " HfArgumentParser,\n", |
968 |
| - " Seq2SeqTrainingArguments,\n", |
969 |
| - " set_seed,\n", |
970 |
| - " Trainer\n", |
971 |
| - ")\n", |
| 949 | + "from fate_client.pipeline.components.fate.nn.loader import Loader\n", |
972 | 950 | "import argparse\n",
|
973 | 951 | "from fate_client.pipeline.utils import test_utils\n",
|
974 |
| - "from fate_client.pipeline.components.fate.evaluation import Evaluation\n", |
975 | 952 | "from fate_client.pipeline.components.fate.reader import Reader\n",
|
976 | 953 | "from fate_client.pipeline import FateFlowPipeline\n",
|
977 |
| - "from fate_client.pipeline.components.fate.nn.torch import nn, optim\n", |
978 |
| - "from fate_client.pipeline.components.fate.nn.torch.base import Sequential\n", |
979 |
| - "from fate_client.pipeline.components.fate.homo_nn import HomoNN, get_config_of_default_runner\n", |
980 |
| - "from fate_client.pipeline.components.fate.nn.algo_params import TrainingArguments, FedAVGArguments\n", |
| 954 | + "\n", |
981 | 955 | "\n",
|
982 | 956 | "def main(config=\"../../config.yaml\", namespace=\"\"):\n",
|
983 | 957 | " # obtain config\n",
|
|
0 commit comments