Skip to content

By default, forks are named the same as their upstream repository. You can customize the name to distinguish it further.

License

Notifications You must be signed in to change notification settings

NeonGeckoCom/neon-llm-gemini

 
 

Repository files navigation

NeonAI LLM Gemini

Proxies API calls to Google Gemini.

Request Format

API requests should include history, a list of tuples of strings, and the current query

Example Request:

{
 "history": [["user", "hello"], ["llm", "hi"]],
 "query": "how are you?"
}

Response Format

Responses will be returned as dictionaries. Responses should contain the following:

  • response - String LLM response to the query

Docker Configuration

When running this as a docker container, the XDG_CONFIG_HOME envvar is set to /config. A configuration file at /config/neon/diana.yaml is required and should look like:

MQ:
  port: <MQ Port>
  server: <MQ Hostname or IP>
  users:
    neon_llm_gemini:
      password: <neon_gemini user's password>
      user: neon_gemini
LLM_GEMINI:
  model: "gemini-pro"
  key_path: ""
  role: "You are trying to give a short answer in less than 40 words."
  context_depth: 3
  max_tokens: 100
  num_parallel_processes: 2

For example, if your configuration resides in ~/.config:

export CONFIG_PATH="/home/${USER}/.config"
docker run -v ${CONFIG_PATH}:/config neon_llm_gemini

Note: If connecting to a local MQ server, you may need to specify --network host

About

By default, forks are named the same as their upstream repository. You can customize the name to distinguish it further.

Resources

License

Stars

Watchers

Forks

Packages

 
 
 

Languages

  • Python 98.6%
  • Dockerfile 1.4%