Migrated repository
Go to file
Michael Yang acfad42c4e
!10 支持 opensearch 向量存储
Merge pull request !10 from 读钓/main
2024-08-11 10:53:22 +00:00
agents-flex-bom fix: update bom module 2024-07-23 12:04:54 +08:00
agents-flex-chain build: v1.0.0-beta.8 release (^.^)YYa!! 2024-07-14 17:40:33 +08:00
agents-flex-core 支持 opensearch 向量存储 2024-08-11 18:41:56 +08:00
agents-flex-document-parser build: v1.0.0-beta.8 release (^.^)YYa!! 2024-07-14 17:40:33 +08:00
agents-flex-image build: v1.0.0-beta.8 release (^.^)YYa!! 2024-07-14 17:40:33 +08:00
agents-flex-llm feat: close https://github.com/agents-flex/agents-flex/issues/5 2024-08-05 09:45:32 +08:00
agents-flex-samples build: v1.0.0-beta.8 release (^.^)YYa!! 2024-07-14 17:40:33 +08:00
agents-flex-spring-boot-starter 支持 SpringBoot3;支持 idea 配置提示;修复通义千问模型config为空 2024-08-07 19:38:12 +08:00
agents-flex-store 支持 opensearch 向量存储 2024-08-11 18:41:56 +08:00
agents-flex-test build: v1.0.0-beta.8 release (^.^)YYa!! 2024-07-14 17:40:33 +08:00
docs Update what-is-agentsflex.md 2024-07-15 20:14:30 +08:00
testresource test: add .pdf and .doc parse test 2024-07-03 12:37:23 +08:00
.editorconfig init 2024-01-12 16:24:14 +08:00
.gitignore chore: update .gitignore 2024-05-08 20:15:39 +08:00
changes.md build: v1.0.0-beta.8 release (^.^)YYa!! 2024-07-14 17:40:33 +08:00
LICENSE init 2024-01-12 16:24:14 +08:00
pom.xml build: v1.0.0-beta.8 release (^.^)YYa!! 2024-07-14 17:40:33 +08:00
readme_ja.md doc: update readme 2024-07-15 09:13:03 +08:00
readme_zh.md docs: add Japanese README 2024-07-08 01:40:13 +09:00
readme.md doc: update readme 2024-07-15 09:13:03 +08:00

English | 简体中文 | 日本語

Agents-Flex is a LLM Application Framework like LangChain base on Java.


Features

  • LLM Visit
  • Prompt、Prompt Template
  • Function Calling Definer, Invoker、Running
  • Memory
  • Embedding
  • Vector Store
  • Resource Loaders
  • Document
    • Splitter
    • Loader
    • Parser
      • PoiParser
      • PdfBoxParser
  • Agent
    • LLM Agent
  • Chain
    • SequentialChain
    • ParallelChain
    • LoopChain
    • ChainNode
      • AgentNode
      • EndNode
      • RouterNode
        • GroovyRouterNode
        • QLExpressRouterNode
        • LLMRouterNode

Simple Chat

use OpenAi LLM:

 @Test
public void testChat() {
    OpenAiLlmConfig config = new OpenAiLlmConfig();
    config.setApiKey("sk-rts5NF6n*******");

    Llm llm = new OpenAiLlm(config);
    String response = llm.chat("what is your name?");

    System.out.println(response);
}

use Qwen LLM:

 @Test
public void testChat() {
    QwenLlmConfig config = new QwenLlmConfig();
    config.setApiKey("sk-28a6be3236****");
    config.setModel("qwen-turbo");

    Llm llm = new QwenLlm(config);
    String response = llm.chat("what is your name?");

    System.out.println(response);
}

use SparkAi LLM:

 @Test
public void testChat() {
    SparkLlmConfig config = new SparkLlmConfig();
    config.setAppId("****");
    config.setApiKey("****");
    config.setApiSecret("****");

    Llm llm = new SparkLlm(config);
    String response = llm.chat("what is your name?");

    System.out.println(response);
}

Chat With Histories

public static void main(String[] args) {
    SparkLlmConfig config = new SparkLlmConfig();
    config.setAppId("****");
    config.setApiKey("****");
    config.setApiSecret("****");

    Llm llm = new SparkLlm(config);

    HistoriesPrompt prompt = new HistoriesPrompt();

    System.out.println("ask for something...");
    Scanner scanner = new Scanner(System.in);
    String userInput = scanner.nextLine();

    while (userInput != null) {

        prompt.addMessage(new HumanMessage(userInput));

        llm.chatStream(prompt, (context, response) -> {
            System.out.println(">>>> " + response.getMessage().getContent());
        });

        userInput = scanner.nextLine();
    }
}

Function Calling

  • step 1: define the function native
public class WeatherUtil {

    @FunctionDef(name = "get_the_weather_info", description = "get the weather info")
    public static String getWeatherInfo(
        @FunctionParam(name = "city", description = "the city name") String name
    ) {
        //we should invoke the third part api for weather info here
        return "Today it will be dull and overcast in " + name;
    }
}

  • step 2: invoke the function from LLM
 public static void main(String[] args) {
    OpenAiLlmConfig config = new OpenAiLlmConfig();
    config.setApiKey("sk-rts5NF6n*******");

    OpenAiLlm llm = new OpenAiLlm(config);

    FunctionPrompt prompt = new FunctionPrompt("How is the weather in Beijing today?", WeatherUtil.class);
    FunctionResultResponse response = llm.chat(prompt);

    Object result = response.getFunctionResult();

    System.out.println(result);
    //Today it will be dull and overcast in Beijing
}

Communication

Modules