Skip to content

Commit b4278bb

Browse files
committed
💄 run prettier.
1 parent edb66fa commit b4278bb

File tree

2 files changed

+36
-36
lines changed

2 files changed

+36
-36
lines changed

packages/inference/src/providers/featherless-ai.ts

+35-35
Original file line numberDiff line numberDiff line change
@@ -3,49 +3,49 @@ import type { ChatCompletionOutput, TextGenerationOutputFinishReason } from "@hu
33
import { InferenceOutputError } from "../lib/InferenceOutputError";
44

55
interface FeatherlessAITextCompletionOutput extends Omit<ChatCompletionOutput, "choices"> {
6-
choices: Array<{
7-
text: string;
8-
finish_reason: TextGenerationOutputFinishReason;
9-
seed: number;
10-
logprobs: unknown;
11-
index: number;
12-
}>;
6+
choices: Array<{
7+
text: string;
8+
finish_reason: TextGenerationOutputFinishReason;
9+
seed: number;
10+
logprobs: unknown;
11+
index: number;
12+
}>;
1313
}
1414

1515
const FEATHERLESS_API_BASE_URL = "https://api.featherless.ai";
1616

1717
export class FeatherlessAIConversationalTask extends BaseConversationalTask {
18-
constructor() {
19-
super("featherless-ai", FEATHERLESS_API_BASE_URL);
20-
}
18+
constructor() {
19+
super("featherless-ai", FEATHERLESS_API_BASE_URL);
20+
}
2121
}
2222

2323
export class FeatherlessAITextGenerationTask extends BaseTextGenerationTask {
24-
constructor() {
25-
super("featherless-ai", FEATHERLESS_API_BASE_URL);
26-
}
24+
constructor() {
25+
super("featherless-ai", FEATHERLESS_API_BASE_URL);
26+
}
2727

28-
override preparePayload(params: BodyParams): Record<string, unknown> {
29-
return {
30-
model: params.model,
31-
...params.args,
32-
...params.args.parameters,
33-
prompt: params.args.inputs,
34-
};
35-
}
28+
override preparePayload(params: BodyParams): Record<string, unknown> {
29+
return {
30+
model: params.model,
31+
...params.args,
32+
...params.args.parameters,
33+
prompt: params.args.inputs,
34+
};
35+
}
3636

37-
override async getResponse(response: FeatherlessAITextCompletionOutput): Promise<TextGenerationOutput> {
38-
if (
39-
typeof response === "object" &&
40-
"choices" in response &&
41-
Array.isArray(response?.choices) &&
42-
typeof response?.model === "string"
43-
) {
44-
const completion = response.choices[0];
45-
return {
46-
generated_text: completion.text,
47-
};
48-
}
49-
throw new InferenceOutputError("Expected Together text generation response format");
50-
}
37+
override async getResponse(response: FeatherlessAITextCompletionOutput): Promise<TextGenerationOutput> {
38+
if (
39+
typeof response === "object" &&
40+
"choices" in response &&
41+
Array.isArray(response?.choices) &&
42+
typeof response?.model === "string"
43+
) {
44+
const completion = response.choices[0];
45+
return {
46+
generated_text: completion.text,
47+
};
48+
}
49+
throw new InferenceOutputError("Expected Together text generation response format");
50+
}
5151
}

packages/inference/test/InferenceClient.spec.ts

+1-1
Original file line numberDiff line numberDiff line change
@@ -829,7 +829,7 @@ describe.concurrent("InferenceClient", () => {
829829
describe.concurrent(
830830
"Featherless",
831831
() => {
832-
HARDCODED_MODEL_ID_MAPPING['featherless-ai'] = {
832+
HARDCODED_MODEL_ID_MAPPING["featherless-ai"] = {
833833
"meta-llama/Llama-3.1-8B": "meta-llama/Meta-Llama-3.1-8B",
834834
"meta-llama/Llama-3.1-8B-Instruct": "meta-llama/Meta-Llama-3.1-8B-Instruct",
835835
};

0 commit comments

Comments
 (0)