* Update README.md for docker instruction issues some fixes for docker build instructions Signed-off-by: Tsai, Louie <louie.tsai@intel.com> * Update README.md for new frontend Signed-off-by: Tsai, Louie <louie.tsai@intel.com> * Update README.md for output example Signed-off-by: Tsai, Louie <louie.tsai@intel.com> * pre-commit change Signed-off-by: Tsai, Louie <louie.tsai@intel.com> * fix for pre-commit Signed-off-by: Tsai, Louie <louie.tsai@intel.com> --------- Signed-off-by: Tsai, Louie <louie.tsai@intel.com>
Document Summarization Application
In a world where data, information, and legal complexities are prevalent, the volume of legal documents is growing rapidly. Law firms, legal professionals, and businesses are dealing with an ever-increasing number of legal texts, including contracts, court rulings, statutes, and regulations. These documents contain important insights, but understanding them can be overwhelming. This is where the demand for legal document summarization comes in.
Large Language Models (LLMs) have revolutionized the way we interact with text. These models can be used to create summaries of news articles, research papers, technical documents, and other types of text. Suppose you have a set of documents (PDFs, Notion pages, customer questions, etc.) and you want to summarize the content. In this example use case, we utilize LangChain to implement summarization strategies and facilitate LLM inference using Text Generation Inference on Intel Xeon and Gaudi2 processors.
The architecture for document summarization will be illustrated/described below:
Deploy Document Summarization Service
The Document Summarization service can be effortlessly deployed on either Intel Gaudi2 or Intel XEON Scalable Processors.
Deploy Document Summarization on Gaudi
Refer to the Gaudi Guide for instructions on deploying Document Summarization on Gaudi.
Deploy Document Summarization on Xeon
Refer to the Xeon Guide for instructions on deploying Document Summarization on Xeon.

