LLMs are an extremely general-purpose tool to have in your toolkit. GeniePy lets you seamlessly integrate LLMs in your application to solve a variety of tasks.
GeniePy uses langchain
to interact with LLMs. The LangChain library provides a
standardized interface to work with a variety of language models, which is especially
useful in case you need to switch between models.
As a default, GeniePy uses the language models provided by OpenAI, in particular
gpt-3.5-turbo
. This makes it easy for you to get started with, and you don't have to
worry about deploying and maintaining your own models.
Configuration
Set the OPENAI_API_KEY
environment variable to your own API key (you can generate one
here if you don't have one already).
The gpt-3.5-turbo
model is the default. If you'd like to use a different model, update
app.constants.OPENAI_MODEL
to the model name.
Usage
All the code around LLM is organized in the app.services.llm
module.
As an example, the app.services.llm.random_product_features
function is provided that
generates 5 random features for a fictitious product with a given name. You can take
this function and write your own function based on the problem at hand.