huggi / src /lib /server /modelEndpoint.ts
Colin Dablain
feat: factor out HF_API_ROOT to allow different inference endpoints (#332)
3baa389 unverified
raw
history blame
782 Bytes
import { HF_ACCESS_TOKEN, HF_API_ROOT } from "$env/static/private";
import { sum } from "$lib/utils/sum";
import type { BackendModel } from "./models";
/**
* Find a random load-balanced endpoint
*/
export function modelEndpoint(model: BackendModel): {
url: string;
authorization: string;
weight: number;
} {
if (!model.endpoints) {
return {
url: `${HF_API_ROOT}/${model.name}`,
authorization: `Bearer ${HF_ACCESS_TOKEN}`,
weight: 1,
};
}
const endpoints = model.endpoints;
const totalWeight = sum(endpoints.map((e) => e.weight));
let random = Math.random() * totalWeight;
for (const endpoint of endpoints) {
if (random < endpoint.weight) {
return endpoint;
}
random -= endpoint.weight;
}
throw new Error("Invalid config, no endpoint found");
}