Sign in

DagKnows Architecture Overview

There was a problem that the LLM was not able to address. Please rephrase your prompt and try again.

DagKnows is an Agentic AI platform for automation. With DagKnows, users can effortlessly automate various tasks by simply chatting with an AI agent in natural language. However, the AI Agent part is optional. If users prefer to build and manage automation manually, DagKnows makes that easy too. Let's discuss the AI Agent first.




As this figure shows, DagKnows integrates with all your favorite tools as well as your infrastructure (Cloud as well as on-prem). It provides a natural language layer to interact with all the tools and infrastructure. A user can just "chat" with DagKnows and ask it to perform various actions on the connected tools and infra and DagKnows takes care of it. Underneath, DagKnows makes LLM calls to turn user requests into code and executes the code on your infrastructure or tools via API calls, CLI commands etc.


Here's a high level architecture of DagKnows. As the following figure shows, DagKnows has two main components.

  1. DagKnows webapp: This is the main controller with a UI. It manages and orchestrates users, automation tasks etc.
  2. DagKnows proxy: Think of it like a worker node. It actually executes the scripts and commands in the runbooks or workflows. While the figure shows only one proxy, in reality you can have multiple proxies. It is the proxy that connects to your environment and executes scripts/commands.




Here's a typical flow:


  1. User logs into the webapp and starts a conversation with AI agent. They can ask any question or specify actions etc. e.g. "List my ec2 instances with CPU utilization less than 10%"
  2. User prompt undergoes prompt engineering + RAG context attachment in our system
  3. The modified prompt is then sent to LLM. This can be any LLM you configure. We integrate with Amazon Bedrock, Azure AI, OpenAI, Anthropic and more. It can also be your own LLM on your premises.
  4. LLM responds with a code to execute or tool to use
  5. DagKnows web app puts together a script with the code or tool and sends it to the selected proxy to execute it on
  6. Proxy executes the script. In the script there may be calls to connect to tools or servers etc. Proxy will make the necessary calls and finally respond with the execution results.
  7. Execution results come to DagKnows webapp and it will send those to the LLM to ensure user's goal is achieved.
  8. If LLM thinks the goal is achieved, the results are displayed to the user in the webapp. If not achieved, LLM may respond with modified code or tool to retry.
  9. When the conversation with AI is over, user can optionally turn all the scripts generated in the process into a workflow or runbook by clicking a button.
  10. The workflow can then be executed as many times as you want, without any AI involvement. The workflows execution can be triggered via alerts or in a periodic schedule.


Thus, the beauty of DagKnows AI platform is that it enables users to build reusable workflows. AI interaction is needed only once -- while creating workflow. Once the workflow is ready, it can be deployed in multiple environments and there's no need to call LLMs.