mirror of
https://github.com/Direct-Dev-Ru/go-lcg.git
synced 2025-11-16 01:29:55 +00:00
225 lines
5.8 KiB
Go
225 lines
5.8 KiB
Go
// package main
|
|
|
|
// import (
|
|
// _ "embed"
|
|
// "fmt"
|
|
// "math"
|
|
// "os"
|
|
// "os/user"
|
|
// "path"
|
|
// "strings"
|
|
// "time"
|
|
|
|
// "github.com/atotto/clipboard"
|
|
// "github.com/direct-dev-ru/linux-command-gpt/gpt"
|
|
// "github.com/direct-dev-ru/linux-command-gpt/reader"
|
|
// )
|
|
|
|
// //go:embed VERSION.txt
|
|
// var Version string
|
|
|
|
// var cwd, _ = os.Getwd()
|
|
|
|
// var (
|
|
// HOST = getEnv("LCG_HOST", "http://192.168.87.108:11434/")
|
|
// COMPLETIONS = getEnv("LCG_COMPLETIONS_PATH", "api/chat") // relative part of endpoint
|
|
// MODEL = getEnv("LCG_MODEL", "codegeex4")
|
|
// PROMPT = getEnv("LCG_PROMPT", "Reply with linux command and nothing else. Output with plain response - no need formatting. No need explanation. No need code blocks. No need ` symbols.")
|
|
// API_KEY_FILE = getEnv("LCG_API_KEY_FILE", ".openai_api_key")
|
|
// RESULT_FOLDER = getEnv("LCG_RESULT_FOLDER", path.Join(cwd, "gpt_results"))
|
|
|
|
// // HOST = "https://api.openai.com/v1/"
|
|
// // COMPLETIONS = "chat/completions"
|
|
|
|
// // MODEL = "gpt-4o-mini"
|
|
// // MODEL = "codellama:13b"
|
|
|
|
// // This file is created in the user's home directory
|
|
// // Example: /home/username/.openai_api_key
|
|
// // API_KEY_FILE = ".openai_api_key"
|
|
|
|
// HELP = `
|
|
|
|
// Usage: lcg [options]
|
|
|
|
// --help -h output usage information
|
|
// --version -v output the version number
|
|
// --file -f read part of command from file or bash feature $(...)
|
|
// --update-key -u update the API key
|
|
// --delete-key -d delete the API key
|
|
|
|
// Example Usage: lcg I want to extract linux-command-gpt.tar.gz file
|
|
// Example Usage: lcg --file /path/to/file.json I want to print object questions with jq
|
|
|
|
// Env Vars:
|
|
// LCG_HOST - defaults to "http://192.168.87.108:11434/" - endpoint for Ollama or other LLM API
|
|
// LCG_COMPLETIONS_PATH -defaults to "api/chat" - relative part of endpoint
|
|
// LCG_MODEL - defaults to "codegeex4"
|
|
// LCG_PROMPT - defaults to Reply with linux command and nothing else. Output with plain response - no need formatting. No need explanation. No need code blocks.
|
|
// LCG_API_KEY_FILE - defaults to ${HOME}/.openai_api_key - file with API key
|
|
// LCG_RESULT_FOLDER - defaults to $(pwd)/gpt_results - folder to save results
|
|
// `
|
|
|
|
// VERSION = Version
|
|
// CMD_HELP = 100
|
|
// CMD_VERSION = 101
|
|
// CMD_UPDATE = 102
|
|
// CMD_DELETE = 103
|
|
// CMD_COMPLETION = 110
|
|
// )
|
|
|
|
// // getEnv retrieves the value of the environment variable `key` or returns `defaultValue` if not set.
|
|
// func getEnv(key, defaultValue string) string {
|
|
// if value, exists := os.LookupEnv(key); exists {
|
|
// return value
|
|
// }
|
|
// return defaultValue
|
|
// }
|
|
|
|
// func handleCommand(cmd string) int {
|
|
// if cmd == "" || cmd == "--help" || cmd == "-h" {
|
|
// return CMD_HELP
|
|
// }
|
|
// if cmd == "--version" || cmd == "-v" {
|
|
// return CMD_VERSION
|
|
// }
|
|
// if cmd == "--update-key" || cmd == "-u" {
|
|
// return CMD_UPDATE
|
|
// }
|
|
// if cmd == "--delete-key" || cmd == "-d" {
|
|
// return CMD_DELETE
|
|
// }
|
|
// return CMD_COMPLETION
|
|
// }
|
|
|
|
// func getCommand(gpt3 gpt.Gpt3, cmd string) (string, float64) {
|
|
// gpt3.InitKey()
|
|
// s := time.Now()
|
|
// done := make(chan bool)
|
|
// go func() {
|
|
// loadingChars := []rune{'-', '\\', '|', '/'}
|
|
// i := 0
|
|
// for {
|
|
// select {
|
|
// case <-done:
|
|
// fmt.Printf("\r")
|
|
// return
|
|
// default:
|
|
// fmt.Printf("\rLoading %c", loadingChars[i])
|
|
// i = (i + 1) % len(loadingChars)
|
|
// time.Sleep(30 * time.Millisecond)
|
|
// }
|
|
// }
|
|
// }()
|
|
|
|
// r := gpt3.Completions(cmd)
|
|
// done <- true
|
|
// elapsed := time.Since(s).Seconds()
|
|
// elapsed = math.Round(elapsed*100) / 100
|
|
|
|
// if r == "" {
|
|
// return "", elapsed
|
|
// }
|
|
// return r, elapsed
|
|
// }
|
|
|
|
// func main() {
|
|
// currentUser, err := user.Current()
|
|
// if err != nil {
|
|
// panic(err)
|
|
// }
|
|
|
|
// args := os.Args
|
|
// cmd := ""
|
|
// file := ""
|
|
// if len(args) > 1 {
|
|
// start := 1
|
|
// if args[1] == "--file" || args[1] == "-f" {
|
|
// file = args[2]
|
|
// start = 3
|
|
// }
|
|
// cmd = strings.Join(args[start:], " ")
|
|
// }
|
|
|
|
// if file != "" {
|
|
// err := reader.FileToPrompt(&cmd, file)
|
|
// if err != nil {
|
|
// fmt.Println(err)
|
|
// return
|
|
// }
|
|
// }
|
|
|
|
// if _, err := os.Stat(RESULT_FOLDER); os.IsNotExist(err) {
|
|
// os.MkdirAll(RESULT_FOLDER, 0755)
|
|
// }
|
|
|
|
// h := handleCommand(cmd)
|
|
|
|
// if h == CMD_HELP {
|
|
// fmt.Println(HELP)
|
|
// return
|
|
// }
|
|
|
|
// if h == CMD_VERSION {
|
|
// fmt.Println(VERSION)
|
|
// return
|
|
// }
|
|
|
|
// gpt3 := gpt.Gpt3{
|
|
// CompletionUrl: HOST + COMPLETIONS,
|
|
// Model: MODEL,
|
|
// Prompt: PROMPT,
|
|
// HomeDir: currentUser.HomeDir,
|
|
// ApiKeyFile: API_KEY_FILE,
|
|
// Temperature: 0.01,
|
|
// }
|
|
|
|
// if h == CMD_UPDATE {
|
|
// gpt3.UpdateKey()
|
|
// return
|
|
// }
|
|
|
|
// if h == CMD_DELETE {
|
|
// gpt3.DeleteKey()
|
|
// return
|
|
// }
|
|
|
|
// c := "R"
|
|
// r := ""
|
|
// elapsed := 0.0
|
|
// for c == "R" || c == "r" {
|
|
// r, elapsed = getCommand(gpt3, cmd)
|
|
// c = "N"
|
|
// fmt.Printf("Completed in %v seconds\n\n", elapsed)
|
|
// fmt.Println(r)
|
|
// fmt.Print("\nDo you want to (c)opy, (s)ave to file, (r)egenerate, or take (N)o action on the command? (c/r/N): ")
|
|
// fmt.Scanln(&c)
|
|
|
|
// // no action
|
|
// if c == "N" || c == "n" {
|
|
// return
|
|
// }
|
|
// }
|
|
|
|
// if r == "" {
|
|
// return
|
|
// }
|
|
|
|
// // Copy to clipboard
|
|
// if c == "C" || c == "c" {
|
|
// clipboard.WriteAll(r)
|
|
// fmt.Println("\033[33mCopied to clipboard")
|
|
// return
|
|
// }
|
|
|
|
// if c == "S" || c == "s" {
|
|
// timestamp := time.Now().Format("2006-01-02_15-04-05") // Format: YYYY-MM-DD_HH-MM-SS
|
|
// filename := fmt.Sprintf("gpt_request_%s(%s).md", timestamp, gpt3.Model)
|
|
// filePath := path.Join(RESULT_FOLDER, filename)
|
|
// resultString := fmt.Sprintf("## Prompt:\n\n%s\n\n------------------\n\n## Response:\n\n%s\n\n", cmd+". "+gpt3.Prompt, r)
|
|
// os.WriteFile(filePath, []byte(resultString), 0644)
|
|
// fmt.Println("\033[33mSaved to file")
|
|
// return
|
|
// }
|
|
// }
|