Spaces:
Sleeping
Sleeping
from fastapi import FastAPI | |
from fastapi.responses import RedirectResponse | |
from langserve import add_routes | |
from langchain.prompts import ChatPromptTemplate | |
from langchain.chat_models import ChatOpenAI | |
from dotenv import load_dotenv | |
from langchain_google_genai import ChatGoogleGenerativeAI | |
load_dotenv() | |
app = FastAPI( | |
title="LangChain Server", | |
version="1.0", | |
description="A simple api server using Langchain's Runnable interfaces", | |
) | |
async def root(): | |
# return json response | |
return {"message": "LangChain Server is running"} | |
add_routes( | |
path = "/openai", | |
app = app, | |
runnable= ChatOpenAI(model="gpt-4-1106-preview"), | |
) | |
add_routes( | |
path = "/google", | |
app = app, | |
runnable= ChatGoogleGenerativeAI(model="gemini-pro"), | |
) | |
if __name__ == "__main__": | |
import uvicorn | |
uvicorn.run(app, host="localhost", port=7860) |