From 79bb7a689305eabd342c9cec9bf8ac79191c0063 Mon Sep 17 00:00:00 2001 From: Yi Yao Date: Tue, 19 Nov 2024 16:28:33 +0800 Subject: [PATCH 01/10] Initial commit of v1.1 release notes --- release_notes/v1.1.md | 140 ++++++++++++++++++++++++++++++++++++++++++ 1 file changed, 140 insertions(+) create mode 100644 release_notes/v1.1.md diff --git a/release_notes/v1.1.md b/release_notes/v1.1.md new file mode 100644 index 00000000..a0012f01 --- /dev/null +++ b/release_notes/v1.1.md @@ -0,0 +1,140 @@ +# OPEA Release Notes v1.1 +We are pleased to announce the release of OPEA version 1.1, which includes significant contributions from the open-source community. This release addresses over 450 pull requests from xxx, enhancing xxxx. + +More information about how to get started with OPEA v1.1 can be found at [Getting Started](https://opea-project.github.io/latest/index.html) page. All project source code is maintained in the [repository](https://github.com/opea-project). To pull Docker images, please access the [hub](https://hub.docker.com/u/opea). + +## What's New in OPEA v1.1 +This release introduces more scenarios with general availability, including the [Image-to-Video](https://github.com/opea-project/GenAIComps/pull/465), [Test-to-Image](https://github.com/opea-project/GenAIComps/pull/729), [Text-to-SQL](https://github.com/opea-project/GenAIExamples/tree/main/DBQnA) and [Avatar Animation](Animation microservice) capabilities; introduces [Prediction Guard](https://github.com/opea-project/GenAIComps/pull/677) to enhance AI security by Factuality Check, PII Detection, Prompt Injection Detection and Toxicity Detection; introduces [Generative AI Studio](https://github.com/opea-project/GenAIStudio) that streamlines the creation of enterprise Generative AI applications, offering an alternative to manual processes; expands the hardware support portfolio, like [Intel® Arc™ GPUs](https://github.com/opea-project/GenAIComps/pull/641) and [AMD® GPUs](https://github.com/opea-project/GenAIExamples/issues/1153); introduces enhanced monitoring system, providing real-time insights into performance and system resource utilization for [CPU and Intel® Gaudi® AI Accelerator](https://github.com/opea-project/GenAIEval/issues/195), as well as [Horizontal Pod Autoscaler(HPA)](https://github.com/opea-project/GenAIInfra/pull/531); integrates benchmark tools for long-context language models (LCLMs), like [LongBench](https://github.com/opea-project/GenAIEval/pull/179) and [HELMET](https://github.com/opea-project/GenAIEval/pull/182). + +### Highlights +#### New GenAI Examples +- [AvatarChatbot](https://github.com/opea-project/GenAIExamples/tree/main/AvatarChatbot): a chatbot that combines a virtual "avatar" on either Intel Gaudi 2 AI Acclerator or Intel Xeon Scalable Processors. +- [DBQnA](https://github.com/opea-project/GenAIExamples/tree/main/DBQnA): it seamlessly translates natural language queries into SQL and delivers real-time database results. +- [EdgeCraftRAG](https://github.com/opea-project/GenAIExamples/tree/main/EdgeCraftRAG): a customizable and tunable RAG example for edge solutions on Intel® Arc™ GPUs. +- [GraphRAG](https://github.com/opea-project/GenAIExamples/tree/main/GraphRAG): a use case presents a Graph RAG approach to query-focused summarization. +- [Text2Image](https://github.com/opea-project/GenAIExamples/tree/main/Text2Image): an application that generates image conditioning on the provided texts. +- [WorkflowExecAgent](https://github.com/opea-project/GenAIExamples/tree/main/WorkflowExecAgent): a workflow executor example to handle data/AI workflow operations via LangChain agents to execute custom-defined workflow-based tools. + +#### New GenAI Components +- Text-to-Image: add [Stable Diffusion microservice](https://github.com/opea-project/GenAIComps/pull/729) +- Image-to-Video: add [Stable Video Diffusion microservice](https://github.com/opea-project/GenAIComps/pull/465) +- Text-to-SQL: add [Text-to-SQL microservice](https://github.com/opea-project/GenAIComps/pull/736) +- Text-to-Speech: add [GPT-SoVITS microservice](https://github.com/opea-project/GenAIComps/pull/784) +- Avatar Animation: add [Animation microservice](https://github.com/opea-project/GenAIComps/pull/775) +- RAG: add [GraphRAG with llama-index microservice](https://github.com/opea-project/GenAIComps/pull/793) + +#### GenAIStudio +[GenAI Studio](https://github.com/opea-project/GenAIStudio), a new project of OPEA, streamlines the creation of enterprise Generative AI applications by providing alternate to manual processes with a seamless, end-to-end solution. From GenAI app development and evaluation to performance benchmarking and deployment, GenAI Studio empowers developers to effortlessly build, test, optimize their LLM solutions and create the deployment package. Its intuitive no-code/low-code interface accelerates innovation, enabling rapid development and deployment of cutting-edge AI applications with unparalleled efficiency and precision. + +#### Prediction Guard +[Prediction Guard](https://docs.predictionguard.com) allows you to utilize hosted open access LLMs, LVMs, and embedding functionality with seamlessly integrated safeguards. In addition to providing a scalable access to open models, Prediction Guard allows you to configure factual consistency checks, toxicity filters, PII filters, and prompt injection blocking. + +#### Newly Supported Models +- llama-3.2 (1B/3B/11B/90B) +- glm-4-9b-chat +- Qwen2/2.5 (7B/32B/72B) + +#### Newly Supported Hardware +- Intel® Arc™ GPU: [vLLM powered by OpenVINO](https://github.com/opea-project/GenAIComps/pull/729) can perform optimal model serving on Intel® Arc™ GPU. +- AMD® GPU: deploy GenAI examples on AMD® GPU using AMD® ROCm™: [CodeTrans](https://github.com/opea-project/GenAIExamples/pull/1138), [CodeGen](https://github.com/opea-project/GenAIExamples/pull/1130), [FaqGen](https://github.com/opea-project/GenAIExamples/pull/1126), [DocSum](https://github.com/opea-project/GenAIExamples/pull/1125), [ChatQnA](https://github.com/opea-project/GenAIExamples/pull/1122). + +### Notable Changes + +
GenAIExamples +- x +
+ +
GenAIComps + +- Functionalities + - New microservices + - Add stable diffusion microservice (#729) + - Add image2video microservice (Stable Video Diffusion) (#465) + - Text to SQL microservice (#736) + - Add GPT-SoVITS microservice (#784) + - Add image2image microservice (#794) + - Initiate "animation" component (#775) + - GraphRAG with llama-index (#793) + - Enhanced microservices + - Support Chinese for Docsum (#799) + - Support file upload summary for DocSum microservice (#823) + - MultimodalQnA Image and Audio Support Phase 1 (#852) + - add dynamic batching embedding/reranking (#774) + - refine codetrans prompt, support parameter input (#822) + - Update RAGAgentLlama and ReActLlama (#843) + - Set a higher default value(1.2) about repetition_penalty for codegen example to reduce repetition (#820) + - Add E2E Promeheus metrics to applications (#845) + - [Agent] support custom prompt (#798) + - support faqgen upload file in UI (#866) + - Removed microservices + - Remove useless vllm ray (#859) + - Enhanced Security + - Prediction Guard Guardrails components (#677) + - Add WildGuard Guardrail Microservice (#710) + - upgrade setuptools version to fix CVE-2024-6345 (#806) + - Remote TGI/TGI services with OAuth Client Credentials authentication (#836) + - Async support for microservices + - Support async for embedding micorservice (#742) + - TEI rerank microservice async support (#746) + - Async support for some microservices (#763) + - Add RAG agent and ReAct agent implemention for llama3.1 served by TGI-gaudi (#722) + - Support Llama3.2 vision and vision guard model (#753) + - Enable vllm for Agent (#752) + - Add Intel/toxic-prompt-roberta to toxicity detection microservice (#749) + - Refactor milvus dataprep and retriever (#728) +- Performance + - Fix vllm microservice performance issue. (#731) + - [Dataprep] Reduce Upload File Time Consumption (#744) +- New Hardware Support + - Add vLLM ARC support with OpenVINO backend (#641) + + + + + +- Enable bash scr to to be path-independent using $0 to address ERROR: failed to solve: failed to read dockerfile: open Dockerfile.intel_hpu: no such file or director when following README (#808) +- Add DPO support in finetuning microservice (#857) +- Combine CI/CD docker compose. (#861) +- vLLM support for Codegen (#886) +- agent short & long term memory with langgraph. (#851) +- Fix missing end of file chars (#874) +- quick fix (#894) +- Replace HTTP "inprogress" gauge with megaservice "request_pending" one (#864) +- Add support for Audio and Video summarization to Docsum (#865) +- vLLM support for FAQGen (#884) +- vLLM support for DocSum (#885) +- Block links that require real person verification (#896) +- Multiple models support for LLM TGI (#835) +- Block links that require real person verification (#897) +- Multiple models and remote service support for langchain vLLM text-generation (#887) +- Minor simplication to ServiceOrchestrator code (#889) +- Embedding compatible with OpenAI API (#892) +- Standardize the naming format of images (#898) +- fix history content from agent memory. (#899) +- Fix LLM special token issue (#895) +- trim input to TGI, moved clustering and summarization to dataprep and store in DB (#893) +- Bugfix for follow-up query with a .png image (#900) +- Docsum Gateway Fix (#902) +- vllm hpu fix version for bug fix (#903) +- add zero-shot vc readme (#904) +- Fix units of incorrect caption timestamps (#907) +- Add env for pass down model id in ChatQnA gateway (#906) +- Add "--no-verbose" flag to wget download commands in entrypoint (#909) +- Add empty list check (#914) +
+ +
GenAIEvals +- x +
+ +
GenAIInfra +- x +
+ +
GenAIStudio +- x +
+ +
docs +- x +
\ No newline at end of file From c39a066444c4d905e613f15e34311413e232f770 Mon Sep 17 00:00:00 2001 From: Yi Yao Date: Wed, 20 Nov 2024 17:41:05 +0800 Subject: [PATCH 02/10] Add notable changess for v1.1 release notes --- release_notes/v1.1.md | 336 +++++++++++++++++++++++++++++++----------- 1 file changed, 254 insertions(+), 82 deletions(-) diff --git a/release_notes/v1.1.md b/release_notes/v1.1.md index a0012f01..667a29c2 100644 --- a/release_notes/v1.1.md +++ b/release_notes/v1.1.md @@ -1,10 +1,10 @@ # OPEA Release Notes v1.1 -We are pleased to announce the release of OPEA version 1.1, which includes significant contributions from the open-source community. This release addresses over 450 pull requests from xxx, enhancing xxxx. +We are pleased to announce the release of OPEA version 1.1, which includes significant contributions from the open-source community. This release addresses over 450 pull requests. More information about how to get started with OPEA v1.1 can be found at [Getting Started](https://opea-project.github.io/latest/index.html) page. All project source code is maintained in the [repository](https://github.com/opea-project). To pull Docker images, please access the [hub](https://hub.docker.com/u/opea). ## What's New in OPEA v1.1 -This release introduces more scenarios with general availability, including the [Image-to-Video](https://github.com/opea-project/GenAIComps/pull/465), [Test-to-Image](https://github.com/opea-project/GenAIComps/pull/729), [Text-to-SQL](https://github.com/opea-project/GenAIExamples/tree/main/DBQnA) and [Avatar Animation](Animation microservice) capabilities; introduces [Prediction Guard](https://github.com/opea-project/GenAIComps/pull/677) to enhance AI security by Factuality Check, PII Detection, Prompt Injection Detection and Toxicity Detection; introduces [Generative AI Studio](https://github.com/opea-project/GenAIStudio) that streamlines the creation of enterprise Generative AI applications, offering an alternative to manual processes; expands the hardware support portfolio, like [Intel® Arc™ GPUs](https://github.com/opea-project/GenAIComps/pull/641) and [AMD® GPUs](https://github.com/opea-project/GenAIExamples/issues/1153); introduces enhanced monitoring system, providing real-time insights into performance and system resource utilization for [CPU and Intel® Gaudi® AI Accelerator](https://github.com/opea-project/GenAIEval/issues/195), as well as [Horizontal Pod Autoscaler(HPA)](https://github.com/opea-project/GenAIInfra/pull/531); integrates benchmark tools for long-context language models (LCLMs), like [LongBench](https://github.com/opea-project/GenAIEval/pull/179) and [HELMET](https://github.com/opea-project/GenAIEval/pull/182). +This release introduces more scenarios with general availability, including the [Image-to-Video](https://github.com/opea-project/GenAIComps/pull/465), [Test-to-Image](https://github.com/opea-project/GenAIComps/pull/729), [Text-to-SQL](https://github.com/opea-project/GenAIExamples/tree/main/DBQnA) and [Avatar Animation](https://github.com/opea-project/GenAIComps/pull/775) capabilities; introduces [Prediction Guard](https://github.com/opea-project/GenAIComps/pull/677) to enhance AI security by Factuality Check, PII Detection, Prompt Injection Detection and Toxicity Detection; introduces [Generative AI Studio](https://github.com/opea-project/GenAIStudio) that streamlines the creation of enterprise Generative AI applications, offering an alternative to manual processes; expands the hardware support portfolio, like [Intel® Arc™ GPUs](https://github.com/opea-project/GenAIComps/pull/641) and [AMD® GPUs](https://github.com/opea-project/GenAIExamples/issues/1153); introduces enhanced monitoring system, providing real-time insights into performance and system resource utilization for [CPU and Intel® Gaudi® AI Accelerator](https://github.com/opea-project/GenAIEval/issues/195), as well as [Horizontal Pod Autoscaler(HPA)](https://github.com/opea-project/GenAIInfra/pull/531); integrates benchmark tools for long-context language models (LCLMs), like [LongBench](https://github.com/opea-project/GenAIEval/pull/179) and [HELMET](https://github.com/opea-project/GenAIEval/pull/182). ### Highlights #### New GenAI Examples @@ -15,6 +15,10 @@ This release introduces more scenarios with general availability, including the - [Text2Image](https://github.com/opea-project/GenAIExamples/tree/main/Text2Image): an application that generates image conditioning on the provided texts. - [WorkflowExecAgent](https://github.com/opea-project/GenAIExamples/tree/main/WorkflowExecAgent): a workflow executor example to handle data/AI workflow operations via LangChain agents to execute custom-defined workflow-based tools. +#### Enhanced GenAI Examples +- Multi-media support: [DocSum](https://github.com/opea-project/GenAIExamples/commit/eb91d1f), [MultimodalQnA](https://github.com/opea-project/GenAIExamples/commit/bbc95bb) +- Multi-language support: [AudioQnA](https://github.com/opea-project/GenAIExamples/commit/658867f), [DocSum](https://github.com/opea-project/GenAIExamples/commit/b0f7c9c) + #### New GenAI Components - Text-to-Image: add [Stable Diffusion microservice](https://github.com/opea-project/GenAIComps/pull/729) - Image-to-Video: add [Stable Video Diffusion microservice](https://github.com/opea-project/GenAIComps/pull/465) @@ -23,12 +27,26 @@ This release introduces more scenarios with general availability, including the - Avatar Animation: add [Animation microservice](https://github.com/opea-project/GenAIComps/pull/775) - RAG: add [GraphRAG with llama-index microservice](https://github.com/opea-project/GenAIComps/pull/793) +#### Enhanced GenAI Components +- Asynchronous support for microservices ([28672956](https://github.com/opea-project/GenAIComps/commit/28672956), [9df4b3c0](https://github.com/opea-project/GenAIComps/commit/9df4b3c0), [f3746dc8](https://github.com/opea-project/GenAIComps/commit/f3746dc8)) +- Add vLLM backends for [summarization](https://github.com/opea-project/GenAIComps/commit/550325d8), [FAQ generation](https://github.com/opea-project/GenAIComps/commit/f5c60f10), [code generation](https://github.com/opea-project/GenAIComps/commit/24b9f03f), and [Agents](https://github.com/opea-project/GenAIComps/commit/4638c1d4) +- Multimedia support ([29ef6426](https://github.com/opea-project/GenAIComps/commit/29ef6426), [baafa402](https://github.com/opea-project/GenAIComps/commit/baafa402)) + #### GenAIStudio [GenAI Studio](https://github.com/opea-project/GenAIStudio), a new project of OPEA, streamlines the creation of enterprise Generative AI applications by providing alternate to manual processes with a seamless, end-to-end solution. From GenAI app development and evaluation to performance benchmarking and deployment, GenAI Studio empowers developers to effortlessly build, test, optimize their LLM solutions and create the deployment package. Its intuitive no-code/low-code interface accelerates innovation, enabling rapid development and deployment of cutting-edge AI applications with unparalleled efficiency and precision. #### Prediction Guard [Prediction Guard](https://docs.predictionguard.com) allows you to utilize hosted open access LLMs, LVMs, and embedding functionality with seamlessly integrated safeguards. In addition to providing a scalable access to open models, Prediction Guard allows you to configure factual consistency checks, toxicity filters, PII filters, and prompt injection blocking. +#### Enhanced Observability +Observability offers real-time insights into component performance and system resource utilization. We enhanced this capability by monitoring [key system metrics](https://github.com/opea-project/GenAIInfra/pull/541), including [CPU](https://github.com/opea-project/GenAIInfra/pull/541), host memory, storage, network, and accelerators (such as [Intel Gaudi](https://github.com/opea-project/GenAIInfra/pull/541)), as well as tracking [OPEA application scaling](https://github.com/opea-project/GenAIInfra/pull/541). + +#### Long-context Benchmark Support +We added the following two benchmark kit to response to the community's requirements of long-context language models. + +- HELMET: a comprehensive benchmark for long-context language models covering seven diverse categories of tasks. The datasets are application-centric and are designed to evaluate models at different lengths and levels of complexity. +- [LongBench](https://github.com/THUDM/LongBench): a benchmark tool for bilingual, multitask, and comprehensive assessment of long context understanding capabilities of large language models. + #### Newly Supported Models - llama-3.2 (1B/3B/11B/90B) - glm-4-9b-chat @@ -41,100 +59,254 @@ This release introduces more scenarios with general availability, including the ### Notable Changes
GenAIExamples -- x + +- Functionalities + - New GenAI Examples + - [AvatarChatbot] Initiate "AvatarChatbot" (audio) example ([cfffb4c](https://github.com/opea-project/GenAIExamples/commit/cfffb4c), [960805a](https://github.com/opea-project/GenAIExamples/commit/960805a)) + - [DBQnA] Adding DBQnA example in GenAIExamples ([c0643b7](https://github.com/opea-project/GenAIExamples/commit/c0643b7), [6b9a27d](https://github.com/opea-project/GenAIExamples/commit/6b9a27d)) + - [EdgeCraftRag] Add EdgeCraftRag as a GenAIExample ([c9088eb](https://github.com/opea-project/GenAIExamples/commit/c9088eb), [7949045](https://github.com/opea-project/GenAIExamples/commit/7949045), [096a37a](https://github.com/opea-project/GenAIExamples/commit/096a37a)) + - [GraphRAG] Add GraphRAG example [a65640b](https://github.com/opea-project/GenAIExamples/commit/a65640b) + - [Text2Image]: Add example for text2image [085d859](https://github.com/opea-project/GenAIExamples/commit/085d859) + - [WorkflowExecAgent] Add Workflow Executor Example [bf5c391](https://github.com/opea-project/GenAIExamples/commit/bf5c391) + - Enhanced GenAI Examples + - [AudioQnA] Add multi-language AudioQnA on Xeon [658867f](https://github.com/opea-project/GenAIExamples/commit/658867f) + - [AgentQnA] Update AgentQnA example for v1.1 release [5eb3d28](https://github.com/opea-project/GenAIExamples/commit/5eb3d28) + - [ChatQnA] Enable vLLM Profiling for ChatQnA ([00d9bb6](https://github.com/opea-project/GenAIExamples/commit/00d9bb6), [7adbba6](https://github.com/opea-project/GenAIExamples/commit/7adbba6)) + - [ChatQnA] Add Terraform and Ansible Modules information [7c9ed04](https://github.com/opea-project/GenAIExamples/commit/7c9ed04) + - [ChatQnA] Add chatqna wrapper for multiple model selection [fb514bb](https://github.com/opea-project/GenAIExamples/commit/fb514bb) + - [DocSum] Supported multimedia and added new GUI powered by gradio ([eb91d1f](https://github.com/opea-project/GenAIExamples/commit/eb91d1f), [0cdeb94](https://github.com/opea-project/GenAIExamples/commit/0cdeb94)) + - [DocSum] Support Chinese for Docsum [b0f7c9c](https://github.com/opea-project/GenAIExamples/commit/b0f7c9c) + - [DocIndexRetriever] Update DocIndexRetriever Example to allow user passing in retriever/reranker params [62e06a0](https://github.com/opea-project/GenAIExamples/commit/62e06a0) + - [MultimodalQnA] Image and Audio Support Phase 1 [bbc95bb](https://github.com/opea-project/GenAIExamples/commit/bbc95bb) + - [Text2Image] Add Text2Image UI, UI tests, Readme, and Docker support [c6fc92d](https://github.com/opea-project/GenAIExamples/commit/c6fc92d) + - update examples accuracy [088ab98](https://github.com/opea-project/GenAIExamples/commit/088ab98) + - Add one-button benchmark launcher ([5720cd4](https://github.com/opea-project/GenAIExamples/commit/5720cd4), [ced68e1](https://github.com/opea-project/GenAIExamples/commit/ced68e1)) + - Removed GenAI Pipelines + - [ChatQnA] remove ChatQnA vllm-on-ray [40386d9](https://github.com/opea-project/GenAIExamples/commit/40386d9) + - Changed Defaults + - [ChatQnA] Set no wrapper ChatQnA as default [619d941](https://github.com/opea-project/GenAIExamples/commit/619d941) + - [Codegen] Replace codegen default Model to Qwen/Qwen2.5-Coder-7B-Instruct. [2332d22](https://github.com/opea-project/GenAIExamples/commit/2332d22) + - [CodeTrans] update codetrans default model to Mistral-7B-Instruct-v0.3 [a2afce1](https://github.com/opea-project/GenAIExamples/commit/a2afce1) + +- Enhanced Security + - upgrade setuptools version to fix CVE-2024-6345 [2b2c7ee](https://github.com/opea-project/GenAIExamples/commit/2b2c7ee) + +- New Hardware Support + - [ChatQnA] Add compose example for ChatQnA AMD ROCm deployment [6d3a017](https://github.com/opea-project/GenAIExamples/commit/6d3a017) + - [CodeGen] Adding files to deploy CodeGen application on AMD GPU [83172e9](https://github.com/opea-project/GenAIExamples/commit/83172e9) + - [CodeTrans] Adding files to deploy CodeTrans application on AMD GPU [7e62175](https://github.com/opea-project/GenAIExamples/commit/7e62175) + - [DocSum] Add compose example for DocSum amd rocm deployment [b1bb6db](https://github.com/opea-project/GenAIExamples/commit/b1bb6db) + - [FaqGen] Add compose example for FaqGen AMD ROCm [5648839](https://github.com/opea-project/GenAIExamples/commit/5648839) + +- Dependency Versioning + - [gradio] Bump gradio from 4.44.0 to 5.0.0 in /MultimodalQnA/ui/gradio [f2f6c09](https://github.com/opea-project/GenAIExamples/commit/f2f6c09) + - [TGI-CPU] Update TGI CPU image to latest official release 2.4.0-intel-cpu [0306c62](https://github.com/opea-project/GenAIExamples/commit/0306c62) + - [TGI-Gaudi] Remove redundant code and update tgi version to 2.0.5 [33b9d4e](https://github.com/opea-project/GenAIExamples/commit/33b9d4e) + - [TEI-Gaudi] Use fixed version(1.5.0) of TEI Gaudi for stability [9ff7df9](https://github.com/opea-project/GenAIExamples/commit/9ff7df9) + - [vLLM-Gaudi] align vllm hpu version to latest vllm-fork [e9b1645](https://github.com/opea-project/GenAIExamples/commit/e9b1645) + +- Deployment + - [ChatQnA] Add instructions of modifying reranking docker image for NVGPU [2587179](https://github.com/opea-project/GenAIExamples/commit/2587179) + - [ChatQnA] setup ollama service in aipc docker compose [def39cf](https://github.com/opea-project/GenAIExamples/commit/def39cf) + - [ChatQnA] Make rerank run on gaudi for hpu docker compose [3c164f3](https://github.com/opea-project/GenAIExamples/commit/3c164f3) + - [ChatQnA] Added the K8s yaml for vLLM support [e2f9037](https://github.com/opea-project/GenAIExamples/commit/e2f9037) + - [ChatQnA] manage your own ChatQnA pipelines. [d16c80e](https://github.com/opea-project/GenAIExamples/commit/d16c80e) + - [ChatQnA] docker install instruction for csp [75df2c9](https://github.com/opea-project/GenAIExamples/commit/75df2c9) + - [ChatQnA] ChatQnA with Remote Inference Endpoints (Kubernetes) [56f770c](https://github.com/opea-project/GenAIExamples/commit/56f770c) + - [ProductivitySuite] Simplify the deployment ProductivitySuite on kubernetes [afc39fa](https://github.com/opea-project/GenAIExamples/commit/afc39fa) + +- Fixed Issues + - [AvatarChatbot] Fix left issue of tgi version update [393367e](https://github.com/opea-project/GenAIExamples/commit/393367e) + - [ChatQnA] Fix the service connection issue on GPU and modify the emb backend [944ae47](https://github.com/opea-project/GenAIExamples/commit/944ae47) + - [ChatQnA] Fix AIPC docker container network issue [95b58b5](https://github.com/opea-project/GenAIExamples/commit/95b58b5) + - [ChatQnA] Fix top_n rerank docs [4a265ab](https://github.com/opea-project/GenAIExamples/commit/4a265ab) + - [ChatQnA] fix chatqna accuracy issue with incorrect penalty [b0487fe](https://github.com/opea-project/GenAIExamples/commit/b0487fe) + - [ChatQnA] Fix AIPC retriever and UI error [773c32b](https://github.com/opea-project/GenAIExamples/commit/773c32b) + - [DocSum] Fix docSum ui error in accessing parsed files [3744bb8](https://github.com/opea-project/GenAIExamples/commit/3744bb8) + - image build bug fix [82801d0](https://github.com/opea-project/GenAIExamples/commit/82801d0) + +- Documentation + - [AudioQnA] Update AudioQnA README.md for its workflow [63bad29](https://github.com/opea-project/GenAIExamples/commit/63bad29) + - [AudioQnA] Update AudioQnA README to add a couple usage details [184e9a4](https://github.com/opea-project/GenAIExamples/commit/184e9a4) + - [AgentQnA] Update Agent README.md for workflow [23b820e](https://github.com/opea-project/GenAIExamples/commit/23b820e) + - [AgentQnA] Update README.md for usage experience [a8f4245](https://github.com/opea-project/GenAIExamples/commit/a8f4245) + - [ChatQnA] Add steps to deploy opea services using minikube [6263b51](https://github.com/opea-project/GenAIExamples/commit/6263b51) + - [ChatQnA] Update ChatQnA Readme for LLM Endpoint [aa314f6](https://github.com/opea-project/GenAIExamples/commit/aa314f6) + - [ChatQnA] Update ChatQnA AIPC README [b056ce6](https://github.com/opea-project/GenAIExamples/commit/b056ce6) + - [CodeGen] Update CodeGen README for its workflow [12469c9](https://github.com/opea-project/GenAIExamples/commit/12469c9) + - [DocSum] Update DocSum README.md for its workflow [fbde15b](https://github.com/opea-project/GenAIExamples/commit/fbde15b) + - [FaqGen] Update FaqGen README.md for its workflow [0c6b044](https://github.com/opea-project/GenAIExamples/commit/0c6b044) + - [InstructionTuning] instruction finetune README improvement [644c3a6](https://github.com/opea-project/GenAIExamples/commit/644c3a6) + - [MultiModalQnA] Update MultiModal README.md for workflow [40800b0](https://github.com/opea-project/GenAIExamples/commit/40800b0) + - [ProductivitySuite] Update Productivity README.md for workflow [0edff26](https://github.com/opea-project/GenAIExamples/commit/0edff26) + - [DocIndexRetriever] Update DocIndexRetriever README.md for workflow [a3f9811](https://github.com/opea-project/GenAIExamples/commit/a3f9811) + - [SearchQnA] Update SearchQnA README.md for its workflow [bf28c7f](https://github.com/opea-project/GenAIExamples/commit/bf28c7f) + - [Translation] Update Translation README.md for workflow [35a4fef](https://github.com/opea-project/GenAIExamples/commit/35a4fef) + - [VideoQnA] Update VideoQnA README.md for workflow [1929dfd](https://github.com/opea-project/GenAIExamples/commit/1929dfd) + +- CI/CD/UT + - Add nightly image build and publish action [78331ee](https://github.com/opea-project/GenAIExamples/commit/78331ee) + - optimize hardware list for test [3b1a9fe](https://github.com/opea-project/GenAIExamples/commit/3b1a9fe) + - open manifest test in CI when dockerfile changed [620ef76](https://github.com/opea-project/GenAIExamples/commit/620ef76) + - Optimize path and link validity check. [7dec001](https://github.com/opea-project/GenAIExamples/commit/7dec001)
GenAIComps - Functionalities - - New microservices - - Add stable diffusion microservice (#729) - - Add image2video microservice (Stable Video Diffusion) (#465) - - Text to SQL microservice (#736) - - Add GPT-SoVITS microservice (#784) - - Add image2image microservice (#794) - - Initiate "animation" component (#775) - - GraphRAG with llama-index (#793) - - Enhanced microservices - - Support Chinese for Docsum (#799) - - Support file upload summary for DocSum microservice (#823) - - MultimodalQnA Image and Audio Support Phase 1 (#852) - - add dynamic batching embedding/reranking (#774) - - refine codetrans prompt, support parameter input (#822) - - Update RAGAgentLlama and ReActLlama (#843) - - Set a higher default value(1.2) about repetition_penalty for codegen example to reduce repetition (#820) - - Add E2E Promeheus metrics to applications (#845) - - [Agent] support custom prompt (#798) - - support faqgen upload file in UI (#866) + - New microservices: + - Add stable diffusion microservice [5d0c4367](https://github.com/opea-project/GenAIComps/commit/5d0c4367) + - Add image2video microservice (Stable Video Diffusion) [a03e7a55](https://github.com/opea-project/GenAIComps/commit/a03e7a55) + - Text to SQL microservice [827e3d40](https://github.com/opea-project/GenAIComps/commit/827e3d40) + - Add GPT-SoVITS microservice [6da7db9e](https://github.com/opea-project/GenAIComps/commit/6da7db9e) + - Add image2image microservice [52c1826f](https://github.com/opea-project/GenAIComps/commit/52c1826f) + - Initiate "animation" component [c26d37e7](https://github.com/opea-project/GenAIComps/commit/c26d37e7) + - GraphRAG with llama-index [19330ea2](https://github.com/opea-project/GenAIComps/commit/19330ea2) + - Enhanced microservices: + - Add DPO support in finetuning microservice [37f35140](https://github.com/opea-project/GenAIComps/commit/37f35140) + - Support Chinese for Docsum [9a00a3ea](https://github.com/opea-project/GenAIComps/commit/9a00a3ea) + - Support file upload summary for DocSum microservice [fa2ea642](https://github.com/opea-project/GenAIComps/commit/fa2ea642) + - Add support for Audio and Video summarization to Docsum [baafa402](https://github.com/opea-project/GenAIComps/commit/baafa402) + - vLLM support for FAQGen [f5c60f10](https://github.com/opea-project/GenAIComps/commit/f5c60f10) + - vLLM support for DocSum [550325d8](https://github.com/opea-project/GenAIComps/commit/550325d8) + - vLLM support for Codegen [24b9f03f](https://github.com/opea-project/GenAIComps/commit/24b9f03f) + - Enable vllm for Agent [4638c1d4](https://github.com/opea-project/GenAIComps/commit/4638c1d4) + - Multiple models and remote service support for langchain vLLM text-generation [e3812a74](https://github.com/opea-project/GenAIComps/commit/e3812a74) + - Set a higher default value(1.2) about repetition_penalty for codegen example to reduce repetition [5ed428f4](https://github.com/opea-project/GenAIComps/commit/5ed428f4) + - MultimodalQnA Image and Audio Support Phase 1 [29ef6426](https://github.com/opea-project/GenAIComps/commit/29ef6426) + - refine codetrans prompt, support parameter input [0bb019f8](https://github.com/opea-project/GenAIComps/commit/0bb019f8) + - add dynamic batching embedding/reranking [518cdfb6](https://github.com/opea-project/GenAIComps/commit/518cdfb6) + - Embedding compatible with OpenAI API [7bf1953c](https://github.com/opea-project/GenAIComps/commit/7bf1953c) + - Update RAGAgentLlama and ReActLlama [c8e36390](https://github.com/opea-project/GenAIComps/commit/c8e36390) + - [Agent] support custom prompt [3473bfb3](https://github.com/opea-project/GenAIComps/commit/3473bfb3) + - agent short & long term memory with langgraph. [e39b08f3](https://github.com/opea-project/GenAIComps/commit/e39b08f3) + - support faqgen upload file in UI [453ff726](https://github.com/opea-project/GenAIComps/commit/453ff726) + - Add E2E Promeheus metrics to applications [a6998a1d](https://github.com/opea-project/GenAIComps/commit/a6998a1d) + - Multiple models support for LLM TGI [e879366c](https://github.com/opea-project/GenAIComps/commit/e879366c) + - Add RAG agent and ReAct agent implemention for llama3.1 served by TGI-gaudi [e7fdf537](https://github.com/opea-project/GenAIComps/commit/e7fdf537) + - Support Llama3.2 vision and vision guard model [534c227a](https://github.com/opea-project/GenAIComps/commit/534c227a) + - Add Intel/toxic-prompt-roberta to toxicity detection microservice [f6f620a2](https://github.com/opea-project/GenAIComps/commit/f6f620a2) + - Refactor milvus dataprep and retriever [84374a57](https://github.com/opea-project/GenAIComps/commit/84374a57) - Removed microservices - - Remove useless vllm ray (#859) - - Enhanced Security - - Prediction Guard Guardrails components (#677) - - Add WildGuard Guardrail Microservice (#710) - - upgrade setuptools version to fix CVE-2024-6345 (#806) - - Remote TGI/TGI services with OAuth Client Credentials authentication (#836) + - Remove vllm ray [617e119f](https://github.com/opea-project/GenAIComps/commit/617e119f) - Async support for microservices - - Support async for embedding micorservice (#742) - - TEI rerank microservice async support (#746) - - Async support for some microservices (#763) - - Add RAG agent and ReAct agent implemention for llama3.1 served by TGI-gaudi (#722) - - Support Llama3.2 vision and vision guard model (#753) - - Enable vllm for Agent (#752) - - Add Intel/toxic-prompt-roberta to toxicity detection microservice (#749) - - Refactor milvus dataprep and retriever (#728) + - Support async for embedding micorservice [28672956](https://github.com/opea-project/GenAIComps/commit/28672956) + - TEI rerank microservice async support [9df4b3c0](https://github.com/opea-project/GenAIComps/commit/9df4b3c0) + - Async support for some microservices [f3746dc8](https://github.com/opea-project/GenAIComps/commit/f3746dc8) - Performance - - Fix vllm microservice performance issue. (#731) - - [Dataprep] Reduce Upload File Time Consumption (#744) + - Fix vllm microservice performance issue. [2159f9ad](https://github.com/opea-project/GenAIComps/commit/2159f9ad) + - [Dataprep] Reduce Upload File Time Consumption [71348998](https://github.com/opea-project/GenAIComps/commit/71348998) - New Hardware Support - - Add vLLM ARC support with OpenVINO backend (#641) - - - - - -- Enable bash scr to to be path-independent using $0 to address ERROR: failed to solve: failed to read dockerfile: open Dockerfile.intel_hpu: no such file or director when following README (#808) -- Add DPO support in finetuning microservice (#857) -- Combine CI/CD docker compose. (#861) -- vLLM support for Codegen (#886) -- agent short & long term memory with langgraph. (#851) -- Fix missing end of file chars (#874) -- quick fix (#894) -- Replace HTTP "inprogress" gauge with megaservice "request_pending" one (#864) -- Add support for Audio and Video summarization to Docsum (#865) -- vLLM support for FAQGen (#884) -- vLLM support for DocSum (#885) -- Block links that require real person verification (#896) -- Multiple models support for LLM TGI (#835) -- Block links that require real person verification (#897) -- Multiple models and remote service support for langchain vLLM text-generation (#887) -- Minor simplication to ServiceOrchestrator code (#889) -- Embedding compatible with OpenAI API (#892) -- Standardize the naming format of images (#898) -- fix history content from agent memory. (#899) -- Fix LLM special token issue (#895) -- trim input to TGI, moved clustering and summarization to dataprep and store in DB (#893) -- Bugfix for follow-up query with a .png image (#900) -- Docsum Gateway Fix (#902) -- vllm hpu fix version for bug fix (#903) -- add zero-shot vc readme (#904) -- Fix units of incorrect caption timestamps (#907) -- Add env for pass down model id in ChatQnA gateway (#906) -- Add "--no-verbose" flag to wget download commands in entrypoint (#909) -- Add empty list check (#914) + - Add vLLM ARC support with OpenVINO backend [a2b9d95f](https://github.com/opea-project/GenAIComps/commit/a2b9d95f) +- Enhanced Security + - Prediction Guard Guardrails components [4bbc7a2f](https://github.com/opea-project/GenAIComps/commit/4bbc7a2f) + - Add WildGuard Guardrail Microservice [5bb4046b](https://github.com/opea-project/GenAIComps/commit/5bb4046b)F + - upgrade setuptools version to fix CVE-2024-6345 [6518c0f0](https://github.com/opea-project/GenAIComps/commit/6518c0f0) + - Remote TGI/TGI services with OAuth Client Credentials authentication [74df6bb7](https://github.com/opea-project/GenAIComps/commit/74df6bb7) +- Validation + - Combine CI/CD docker compose. [23c99c11](https://github.com/opea-project/GenAIComps/commit/23c99c11704cfcd0b13b3b463819f633ba02b42d)
-
GenAIEvals -- x +
GenAIEvals + +- New Benchmark + - Support HELMET [4c8f048](https://github.com/opea-project/GenAIEval/commit/4c8f048) + - Support Longbench [021193f](https://github.com/opea-project/GenAIEval/commit/021193f) +- Performance + - Add new constant loader & Fix poisson loader issue [e11588c](https://github.com/opea-project/GenAIEval/commit/e11588c) + - Support Poisson distributed requests for benchmark [7305ea3](https://github.com/opea-project/GenAIEval/commit/7305ea3) + - Support customized prompts and max new tokens in chatqna e2e test [79a4ad3](https://github.com/opea-project/GenAIEval/commit/79a4ad3) + - Add namespace support for k8s performance test [70697d1](https://github.com/opea-project/GenAIEval/commit/70697d1) + - Support sharegpt dataset in chatqna e2e test [028bf63](https://github.com/opea-project/GenAIEval/commit/028bf63) + - [Benchmark] Get benchmark reports. [946c439](https://github.com/opea-project/GenAIEval/commit/946c439) +- Accuracy + - Control the concurrent number of requests in codegen acc test. [84e077e](https://github.com/opea-project/GenAIEval/commit/84e077e) + - integrate deepeval metric with remote endpoint, like tgi server. [ffa65dc](https://github.com/opea-project/GenAIEval/commit/ffa65dc) + - Ragaaf - adding new metric 'context recall' [cc7cebd](https://github.com/opea-project/GenAIEval/commit/cc7cebd) + - Ragaaf - adding new metric 'context relevance' [f995c9c](https://github.com/opea-project/GenAIEval/commit/f995c9c) + - Ragaaf (RAG assessment annotation free) [2413e70](https://github.com/opea-project/GenAIEval/commit/2413e70) + - Adding new metrics to ragas offering [d1c1337](https://github.com/opea-project/GenAIEval/commit/d1c1337) + - add crud ragas evaluation. [f2bff45](https://github.com/opea-project/GenAIEval/commit/f2bff45) + - Minimize requirements for user data for OPEA ragas [f1593ea](https://github.com/opea-project/GenAIEval/commit/f1593ea) +- Monitoring + - Add node metrics Grafana dashboard [a19f42e](https://github.com/opea-project/GenAIEval/commit/a19f42e) + - Add CPU Grafana dashboard [38e69eb](https://github.com/opea-project/GenAIEval/commit/38e69eb) + - add the grafana dashboard json file for Gaudi metrics [6c9ae91](https://github.com/opea-project/GenAIEval/commit/6c9ae91) + - Enhance the Grafana JSON file [8653efb](https://github.com/opea-project/GenAIEval/commit/8653efb) +- Fixed Issues + - [ChatQnA Benchmark] Fixed the output token in chatqnafixed.py [2c8ca26](https://github.com/opea-project/GenAIEval/commit/2c8ca26) + - Fix test duration time inaccurate issue [9d76832](https://github.com/opea-project/GenAIEval/commit/9d76832) + - Fix llm output token length issue [99ef325](https://github.com/opea-project/GenAIEval/commit/99ef325) + - Fix llm serving benchmark issue [d6bafbd](https://github.com/opea-project/GenAIEval/commit/d6bafbd) + - Fix input token size(1024) [30adcbe](https://github.com/opea-project/GenAIEval/commit/30adcbe) + - Ragas fix for use of metrics argument [0cf3631](https://github.com/opea-project/GenAIEval/commit/0cf3631) + - fixed the number of ouput token & fixed the top_k=1 [4af0a62](https://github.com/opea-project/GenAIEval/commit/4af0a62) + - Fix JSON Return Format in getReqData Function [a4be366](https://github.com/opea-project/GenAIEval/commit/a4be366) +- Documentation + - Add setup guide of gaudi prometheus exporter [e9b8637](https://github.com/opea-project/GenAIEval/commit/e9b8637) + - Add README for running OPEA ragas using HF endpoint on Gaudi [0dff0d3](https://github.com/opea-project/GenAIEval/commit/0dff0d3)
GenAIInfra -- x -
-
GenAIStudio -- x +- GMC + - Add manifests for new components [e51fd62](https://github.com/opea-project/GenAIInfra/commit/e51fd62) + +- HelmChart + - [AgentQnA] Helm Chart for AgentQnA [66de41c](https://github.com/opea-project/GenAIInfra/commit/66de41c) + - [AudioQnA] helm: Add audioQnA e2e helm chart [9efacee](https://github.com/opea-project/GenAIInfra/commit/9efacee) + - [AudioQnA] helm-charts: Add gpt-sovits support [1f55e1a](https://github.com/opea-project/GenAIInfra/commit/1f55e1a) + - [ChatQnA] Implement the nowrapper version chatqna [71c81d0](https://github.com/opea-project/GenAIInfra/commit/71c81d0) + - [FaqGen] Add FaqGen helm chart [f847e05](https://github.com/opea-project/GenAIInfra/commit/f847e05) + - [FaqGen] helm: Add llm-faqgen-tgi support [325126e](https://github.com/opea-project/GenAIInfra/commit/325126e) + - [HPA] helm/manifest: Sync HPA related K8S probe settings [c399578](https://github.com/opea-project/GenAIInfra/commit/c399578) + - [VisualQnA] Add helm chart for VisualQnA example [b077d44](https://github.com/opea-project/GenAIInfra/commit/b077d44) + - [UI] support variants for multiple examples [96af2ad](https://github.com/opea-project/GenAIInfra/commit/96af2ad) + - [Nginx] helm-chart: Make nginx service type configurable [a5c96ab](https://github.com/opea-project/GenAIInfra/commit/a5c96ab) + - [Milvus] Add milvus support for data-prep and retriever-usvc [d289b4e](https://github.com/opea-project/GenAIInfra/commit/d289b4e) + - Add helm chart for 3 components [881e2b5](https://github.com/opea-project/GenAIInfra/commit/881e2b5) + - accelerate also teirerank with Gaudi [620963f](https://github.com/opea-project/GenAIInfra/commit/620963f) + +- CSP + - terraform: add AWS/EKS deployment for ChatQnA [bdb9af9](https://github.com/opea-project/GenAIInfra/commit/bdb9af9) + +- Monitoring + - Add Grafana dashboard for monitoring OPEA application scaling in k8s [691bbc5](https://github.com/opea-project/GenAIInfra/commit/691bbc5) + - Add ServiceMonitors for rest of OPEA applications [fc6235a](https://github.com/opea-project/GenAIInfra/commit/fc6235a) + - Add monitoring option to (ChatQnA) Helm charts [dbd607e](https://github.com/opea-project/GenAIInfra/commit/dbd607e) + - Support alternative metrics on accelerated TGI / TEI instances [cdd3585](https://github.com/opea-project/GenAIInfra/commit/cdd3585) + - Expose options such as collector.interval of memory bandwidth exporter in k8s manifests and docker for user configuration. [2517e79](https://github.com/opea-project/GenAIInfra/commit/2517e79) + +- Dependency Versioning + - [TEI-Gaudi] Upgrade tei-gaudi version to 1.5.0 [c6a9c90](https://github.com/opea-project/GenAIInfra/commit/c6a9c90) + - [TGI-CPU] Update tgi cpu image version to 2.4.0-intel-cpu [f6c180e](https://github.com/opea-project/GenAIInfra/commit/f6c180e) + - [TGI-Gaudi] Upgrade tgi-gaudi to version 2.0.6 [915baa0](https://github.com/opea-project/GenAIInfra/commit/915baa0) + - Update the image version for ChatQnA examples [593458c](https://github.com/opea-project/GenAIInfra/commit/593458c) + +- Changed Defaults + - Change default model of codegen and codetrans [74476b7](https://github.com/opea-project/GenAIInfra/commit/74476b7) + +- Documentation + - Update observability README + fix typos [1d77b81](https://github.com/opea-project/GenAIInfra/commit/1d77b81) + - Monitoring, Observability and HPA doc improvements [14198fe](https://github.com/opea-project/GenAIInfra/commit/14198fe) + - Update GMC manifest changes and misc fixes [87dc673](https://github.com/opea-project/GenAIInfra/commit/87dc673) + - Improve Helm charts README [7b8c510](https://github.com/opea-project/GenAIInfra/commit/7b8c510) + - Create troubleshooting.md [d55ded4](https://github.com/opea-project/GenAIInfra/commit/d55ded4) + - Enhance helm chart repo usage in README [0de5535](https://github.com/opea-project/GenAIInfra/commit/0de5535) + +- CI/CD/UT + - Refactor CI scripts to support more components [e09270a](https://github.com/opea-project/GenAIInfra/commit/e09270a) + - Add github workflows to release helm chart [3910e3b](https://github.com/opea-project/GenAIInfra/commit/3910e3b) + - Fix link check failure (#481) (5 weeks ago) [fc87ef3](https://github.com/opea-project/GenAIInfra/commit/fc87ef3) + - Fix CI failures (#477) (5 weeks ago) [7e7b8ab](https://github.com/opea-project/GenAIInfra/commit/7e7b8ab) + - Optimize path and link validity check. [91bd163](https://github.com/opea-project/GenAIInfra/commit/91bd163) + - enable image build process for memory-bandwidth-exporter [ddeac46](https://github.com/opea-project/GenAIInfra/commit/ddeac46) + - Add hyperlinks and paths validation. [d8cd3a1](https://github.com/opea-project/GenAIInfra/commit/d8cd3a1)
-
docs -- x -
\ No newline at end of file +### Full Changelog +- GenAIExamples: [v1.0...v1.1](https://github.com/opea-project/GenAIExamples/compare/v1.0...v1.1) +- GenAIComps: [v1.0...v1.1](https://github.com/opea-project/GenAIComps/compare/v1.0...v1.1) +- GenAIInfra: [v1.0...v1.1](https://github.com/opea-project/GenAIInfra/compare/v1.0...v1.1) +- GenAIEval: [v1.0...v1.1](https://github.com/opea-project/GenAIEval/compare/v1.0...v1.1) +- GenAIStudio: [...v1.1](https://github.com/opea-project/GenAIStudio/compare/v1.1...v1.1) \ No newline at end of file From 963b5f6444de95ab5f34e5b7817bd5b048af2b0f Mon Sep 17 00:00:00 2001 From: Yi Yao Date: Wed, 20 Nov 2024 21:17:12 +0800 Subject: [PATCH 03/10] Add Helm Charts Support in highlights of v1.1 release notes --- release_notes/v1.1.md | 13 ++++++++----- 1 file changed, 8 insertions(+), 5 deletions(-) diff --git a/release_notes/v1.1.md b/release_notes/v1.1.md index 667a29c2..b5bb5061 100644 --- a/release_notes/v1.1.md +++ b/release_notes/v1.1.md @@ -1,10 +1,10 @@ # OPEA Release Notes v1.1 -We are pleased to announce the release of OPEA version 1.1, which includes significant contributions from the open-source community. This release addresses over 450 pull requests. +We are pleased to announce the release of OPEA version 1.1, which includes significant contributions from the open-source community. This release addresses over 460 pull requests. More information about how to get started with OPEA v1.1 can be found at [Getting Started](https://opea-project.github.io/latest/index.html) page. All project source code is maintained in the [repository](https://github.com/opea-project). To pull Docker images, please access the [hub](https://hub.docker.com/u/opea). ## What's New in OPEA v1.1 -This release introduces more scenarios with general availability, including the [Image-to-Video](https://github.com/opea-project/GenAIComps/pull/465), [Test-to-Image](https://github.com/opea-project/GenAIComps/pull/729), [Text-to-SQL](https://github.com/opea-project/GenAIExamples/tree/main/DBQnA) and [Avatar Animation](https://github.com/opea-project/GenAIComps/pull/775) capabilities; introduces [Prediction Guard](https://github.com/opea-project/GenAIComps/pull/677) to enhance AI security by Factuality Check, PII Detection, Prompt Injection Detection and Toxicity Detection; introduces [Generative AI Studio](https://github.com/opea-project/GenAIStudio) that streamlines the creation of enterprise Generative AI applications, offering an alternative to manual processes; expands the hardware support portfolio, like [Intel® Arc™ GPUs](https://github.com/opea-project/GenAIComps/pull/641) and [AMD® GPUs](https://github.com/opea-project/GenAIExamples/issues/1153); introduces enhanced monitoring system, providing real-time insights into performance and system resource utilization for [CPU and Intel® Gaudi® AI Accelerator](https://github.com/opea-project/GenAIEval/issues/195), as well as [Horizontal Pod Autoscaler(HPA)](https://github.com/opea-project/GenAIInfra/pull/531); integrates benchmark tools for long-context language models (LCLMs), like [LongBench](https://github.com/opea-project/GenAIEval/pull/179) and [HELMET](https://github.com/opea-project/GenAIEval/pull/182). +This release introduces more scenarios with general availability, including the [Image-to-Video](https://github.com/opea-project/GenAIComps/pull/465), [Text-to-Image](https://github.com/opea-project/GenAIComps/pull/729), [Text-to-SQL](https://github.com/opea-project/GenAIExamples/tree/main/DBQnA) and [Avatar Animation](https://github.com/opea-project/GenAIComps/pull/775) capabilities; introduces [Prediction Guard](https://github.com/opea-project/GenAIComps/pull/677) to enhance AI security by Factuality Check, PII Detection, Prompt Injection Detection and Toxicity Detection; introduces [Generative AI Studio](https://github.com/opea-project/GenAIStudio) that streamlines the creation of enterprise Generative AI applications, offering an alternative to manual processes; expands the hardware support portfolio, like [Intel® Arc™ GPUs](https://github.com/opea-project/GenAIComps/pull/641) and [AMD® GPUs](https://github.com/opea-project/GenAIExamples/issues/1153); introduces enhanced monitoring system, providing real-time insights into runtime status and system resource utilization for [CPU and Intel® Gaudi® AI Accelerator](https://github.com/opea-project/GenAIEval/issues/195), as well as [Horizontal Pod Autoscaler(HPA)](https://github.com/opea-project/GenAIInfra/pull/531); introduces Helm Charts suppport for 7 examples and microservices; integrates benchmark tools for long-context language models (LCLMs), like [LongBench](https://github.com/opea-project/GenAIEval/pull/179) and [HELMET](https://github.com/opea-project/GenAIEval/pull/182). ### Highlights #### New GenAI Examples @@ -41,8 +41,11 @@ This release introduces more scenarios with general availability, including the #### Enhanced Observability Observability offers real-time insights into component performance and system resource utilization. We enhanced this capability by monitoring [key system metrics](https://github.com/opea-project/GenAIInfra/pull/541), including [CPU](https://github.com/opea-project/GenAIInfra/pull/541), host memory, storage, network, and accelerators (such as [Intel Gaudi](https://github.com/opea-project/GenAIInfra/pull/541)), as well as tracking [OPEA application scaling](https://github.com/opea-project/GenAIInfra/pull/541). +#### Helm Charts Support +OPEA examples and microservices support Helm Charts as the packaging format on K8S. The newly supported examples include [AgentQnA](https://github.com/opea-project/GenAIInfra/commit/66de41c), [AudioQnA](https://github.com/opea-project/GenAIInfra/commit/9efacee), [FaqGen](https://github.com/opea-project/GenAIInfra/commit/f847e05), [VisualQnA](https://github.com/opea-project/GenAIInfra/commit/b077d44) etc. The newly supported microservices include [chathistory, mongodb, prompt](https://github.com/opea-project/GenAIInfra/commit/881e2b5), and [Milvus for data-prep and retriever](https://github.com/opea-project/GenAIInfra/commit/d289b4e). + #### Long-context Benchmark Support -We added the following two benchmark kit to response to the community's requirements of long-context language models. +We added the following two benchmark kits to response to the community's requirements of long-context language models. - HELMET: a comprehensive benchmark for long-context language models covering seven diverse categories of tasks. The datasets are application-centric and are designed to evaluate models at different lengths and levels of complexity. - [LongBench](https://github.com/THUDM/LongBench): a benchmark tool for bilingual, multitask, and comprehensive assessment of long context understanding capabilities of large language models. @@ -304,9 +307,9 @@ We added the following two benchmark kit to response to the community's requirem - Add hyperlinks and paths validation. [d8cd3a1](https://github.com/opea-project/GenAIInfra/commit/d8cd3a1)
-### Full Changelog +### Full Changelogs - GenAIExamples: [v1.0...v1.1](https://github.com/opea-project/GenAIExamples/compare/v1.0...v1.1) - GenAIComps: [v1.0...v1.1](https://github.com/opea-project/GenAIComps/compare/v1.0...v1.1) - GenAIInfra: [v1.0...v1.1](https://github.com/opea-project/GenAIInfra/compare/v1.0...v1.1) - GenAIEval: [v1.0...v1.1](https://github.com/opea-project/GenAIEval/compare/v1.0...v1.1) -- GenAIStudio: [...v1.1](https://github.com/opea-project/GenAIStudio/compare/v1.1...v1.1) \ No newline at end of file +- GenAIStudio: [...v1.1](https://github.com/opea-project/GenAIStudio/compare/3df407115e079622428c17978719ce782a25927a...v1.1) \ No newline at end of file From e83bc174fc42653cd6b8667c763fb76292e1a4e7 Mon Sep 17 00:00:00 2001 From: Yi Yao Date: Thu, 21 Nov 2024 13:23:06 +0800 Subject: [PATCH 04/10] Add TGI-Gaudi 2.0.6 commit to the notable changes of v1.1 release notes. --- release_notes/v1.1.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/release_notes/v1.1.md b/release_notes/v1.1.md index b5bb5061..e5b7677f 100644 --- a/release_notes/v1.1.md +++ b/release_notes/v1.1.md @@ -104,7 +104,7 @@ We added the following two benchmark kits to response to the community's require - Dependency Versioning - [gradio] Bump gradio from 4.44.0 to 5.0.0 in /MultimodalQnA/ui/gradio [f2f6c09](https://github.com/opea-project/GenAIExamples/commit/f2f6c09) - [TGI-CPU] Update TGI CPU image to latest official release 2.4.0-intel-cpu [0306c62](https://github.com/opea-project/GenAIExamples/commit/0306c62) - - [TGI-Gaudi] Remove redundant code and update tgi version to 2.0.5 [33b9d4e](https://github.com/opea-project/GenAIExamples/commit/33b9d4e) + - [TGI-Gaudi] Upgrade TGI Gaudi version to v2.0.6 [1ff85f6a](https://github.com/opea-project/GenAIExamples/commit/1ff85f6a852785a3abad7a5b286e020592dfd2b7) - [TEI-Gaudi] Use fixed version(1.5.0) of TEI Gaudi for stability [9ff7df9](https://github.com/opea-project/GenAIExamples/commit/9ff7df9) - [vLLM-Gaudi] align vllm hpu version to latest vllm-fork [e9b1645](https://github.com/opea-project/GenAIExamples/commit/e9b1645) From be31930fa8ef7daf63ab06369df3e40042032bb1 Mon Sep 17 00:00:00 2001 From: Yi Yao Date: Thu, 21 Nov 2024 15:04:19 +0800 Subject: [PATCH 05/10] Add contributors section in v1.1 release notes --- release_notes/v1.1.md | 23 ++++++++++++++++++----- 1 file changed, 18 insertions(+), 5 deletions(-) diff --git a/release_notes/v1.1.md b/release_notes/v1.1.md index e5b7677f..99ae9f92 100644 --- a/release_notes/v1.1.md +++ b/release_notes/v1.1.md @@ -4,7 +4,7 @@ We are pleased to announce the release of OPEA version 1.1, which includes signi More information about how to get started with OPEA v1.1 can be found at [Getting Started](https://opea-project.github.io/latest/index.html) page. All project source code is maintained in the [repository](https://github.com/opea-project). To pull Docker images, please access the [hub](https://hub.docker.com/u/opea). ## What's New in OPEA v1.1 -This release introduces more scenarios with general availability, including the [Image-to-Video](https://github.com/opea-project/GenAIComps/pull/465), [Text-to-Image](https://github.com/opea-project/GenAIComps/pull/729), [Text-to-SQL](https://github.com/opea-project/GenAIExamples/tree/main/DBQnA) and [Avatar Animation](https://github.com/opea-project/GenAIComps/pull/775) capabilities; introduces [Prediction Guard](https://github.com/opea-project/GenAIComps/pull/677) to enhance AI security by Factuality Check, PII Detection, Prompt Injection Detection and Toxicity Detection; introduces [Generative AI Studio](https://github.com/opea-project/GenAIStudio) that streamlines the creation of enterprise Generative AI applications, offering an alternative to manual processes; expands the hardware support portfolio, like [Intel® Arc™ GPUs](https://github.com/opea-project/GenAIComps/pull/641) and [AMD® GPUs](https://github.com/opea-project/GenAIExamples/issues/1153); introduces enhanced monitoring system, providing real-time insights into runtime status and system resource utilization for [CPU and Intel® Gaudi® AI Accelerator](https://github.com/opea-project/GenAIEval/issues/195), as well as [Horizontal Pod Autoscaler(HPA)](https://github.com/opea-project/GenAIInfra/pull/531); introduces Helm Charts suppport for 7 examples and microservices; integrates benchmark tools for long-context language models (LCLMs), like [LongBench](https://github.com/opea-project/GenAIEval/pull/179) and [HELMET](https://github.com/opea-project/GenAIEval/pull/182). +This release introduces more scenarios with general availability, including the [Image-to-Video](https://github.com/opea-project/GenAIComps/pull/465), [Text-to-Image](https://github.com/opea-project/GenAIComps/pull/729), [Text-to-SQL](https://github.com/opea-project/GenAIExamples/tree/main/DBQnA) and [Avatar Animation](https://github.com/opea-project/GenAIComps/pull/775) capabilities; introduces [Generative AI Studio](https://github.com/opea-project/GenAIStudio) that streamlines the creation of enterprise Generative AI applications, offering an alternative to manual processes; expands the hardware support portfolio, like [Intel® Arc™ GPUs](https://github.com/opea-project/GenAIComps/pull/641) and [AMD® GPUs](https://github.com/opea-project/GenAIExamples/issues/1153); introduces enhanced monitoring system, providing real-time insights into runtime status and system resource utilization for [CPU and Intel® Gaudi® AI Accelerator](https://github.com/opea-project/GenAIEval/issues/195), as well as [Horizontal Pod Autoscaler(HPA)](https://github.com/opea-project/GenAIInfra/pull/531); introduces Helm Charts suppport for 7 examples and microservices; integrates benchmark tools for long-context language models (LCLMs), like [LongBench](https://github.com/opea-project/GenAIEval/pull/179) and [HELMET](https://github.com/opea-project/GenAIEval/pull/182). ### Highlights #### New GenAI Examples @@ -35,9 +35,6 @@ This release introduces more scenarios with general availability, including the #### GenAIStudio [GenAI Studio](https://github.com/opea-project/GenAIStudio), a new project of OPEA, streamlines the creation of enterprise Generative AI applications by providing alternate to manual processes with a seamless, end-to-end solution. From GenAI app development and evaluation to performance benchmarking and deployment, GenAI Studio empowers developers to effortlessly build, test, optimize their LLM solutions and create the deployment package. Its intuitive no-code/low-code interface accelerates innovation, enabling rapid development and deployment of cutting-edge AI applications with unparalleled efficiency and precision. -#### Prediction Guard -[Prediction Guard](https://docs.predictionguard.com) allows you to utilize hosted open access LLMs, LVMs, and embedding functionality with seamlessly integrated safeguards. In addition to providing a scalable access to open models, Prediction Guard allows you to configure factual consistency checks, toxicity filters, PII filters, and prompt injection blocking. - #### Enhanced Observability Observability offers real-time insights into component performance and system resource utilization. We enhanced this capability by monitoring [key system metrics](https://github.com/opea-project/GenAIInfra/pull/541), including [CPU](https://github.com/opea-project/GenAIInfra/pull/541), host memory, storage, network, and accelerators (such as [Intel Gaudi](https://github.com/opea-project/GenAIInfra/pull/541)), as well as tracking [OPEA application scaling](https://github.com/opea-project/GenAIInfra/pull/541). @@ -312,4 +309,20 @@ We added the following two benchmark kits to response to the community's require - GenAIComps: [v1.0...v1.1](https://github.com/opea-project/GenAIComps/compare/v1.0...v1.1) - GenAIInfra: [v1.0...v1.1](https://github.com/opea-project/GenAIInfra/compare/v1.0...v1.1) - GenAIEval: [v1.0...v1.1](https://github.com/opea-project/GenAIEval/compare/v1.0...v1.1) -- GenAIStudio: [...v1.1](https://github.com/opea-project/GenAIStudio/compare/3df407115e079622428c17978719ce782a25927a...v1.1) \ No newline at end of file +- GenAIStudio: [...v1.1](https://github.com/opea-project/GenAIStudio/compare/3df407115e079622428c17978719ce782a25927a...v1.1) + +## Contributors +This release would not have been possible without the contributions of the following organizations and individuals. + +### Orgnization Contributors +- `AMD`: AMD CPU/GPU support. +- `Capital One`: Contributions to CI/CD process. +- `China Unicom`: Contributions to the deployment of GenAI examples. +- `Huawei`: Contributions to OPEA services deployment. +- `Intel`: Development and improvements to GenAI examples, components, infrastructure, evaluation and studio. +- `Nascenia Ltd.`: Contributions to documentation. +- `National Chiao Tung University`: Contributions to documentation. +- `Princeton University`: Integration of HELMET. + +### Individual Contributors +For a comprehensive list of individual contributors, please refer to the "Full Changelogs" section. \ No newline at end of file From f4f454316a5daf4b119882d03f10d76237599e00 Mon Sep 17 00:00:00 2001 From: Yi Yao Date: Fri, 22 Nov 2024 09:16:01 +0800 Subject: [PATCH 06/10] Apply Malini's suggestions from code review Co-authored-by: Malini Bhandaru Co-authored-by: Eero Tamminen --- release_notes/v1.1.md | 20 ++++++++++---------- 1 file changed, 10 insertions(+), 10 deletions(-) diff --git a/release_notes/v1.1.md b/release_notes/v1.1.md index 99ae9f92..b6f3701c 100644 --- a/release_notes/v1.1.md +++ b/release_notes/v1.1.md @@ -4,15 +4,15 @@ We are pleased to announce the release of OPEA version 1.1, which includes signi More information about how to get started with OPEA v1.1 can be found at [Getting Started](https://opea-project.github.io/latest/index.html) page. All project source code is maintained in the [repository](https://github.com/opea-project). To pull Docker images, please access the [hub](https://hub.docker.com/u/opea). ## What's New in OPEA v1.1 -This release introduces more scenarios with general availability, including the [Image-to-Video](https://github.com/opea-project/GenAIComps/pull/465), [Text-to-Image](https://github.com/opea-project/GenAIComps/pull/729), [Text-to-SQL](https://github.com/opea-project/GenAIExamples/tree/main/DBQnA) and [Avatar Animation](https://github.com/opea-project/GenAIComps/pull/775) capabilities; introduces [Generative AI Studio](https://github.com/opea-project/GenAIStudio) that streamlines the creation of enterprise Generative AI applications, offering an alternative to manual processes; expands the hardware support portfolio, like [Intel® Arc™ GPUs](https://github.com/opea-project/GenAIComps/pull/641) and [AMD® GPUs](https://github.com/opea-project/GenAIExamples/issues/1153); introduces enhanced monitoring system, providing real-time insights into runtime status and system resource utilization for [CPU and Intel® Gaudi® AI Accelerator](https://github.com/opea-project/GenAIEval/issues/195), as well as [Horizontal Pod Autoscaler(HPA)](https://github.com/opea-project/GenAIInfra/pull/531); introduces Helm Charts suppport for 7 examples and microservices; integrates benchmark tools for long-context language models (LCLMs), like [LongBench](https://github.com/opea-project/GenAIEval/pull/179) and [HELMET](https://github.com/opea-project/GenAIEval/pull/182). +This release introduces more scenarios with general availability, including the [Image-to-Video](https://github.com/opea-project/GenAIComps/pull/465), [Text-to-Image](https://github.com/opea-project/GenAIComps/pull/729), [Text-to-SQL](https://github.com/opea-project/GenAIExamples/tree/main/DBQnA) and [Avatar Animation](https://github.com/opea-project/GenAIComps/pull/775) capabilities; introduces [Generative AI Studio](https://github.com/opea-project/GenAIStudio) that offers a no-code alternative to create enterprise Generative AI applications; expands the portfolio of supported hardware to include [Intel® Arc™ GPUs](https://github.com/opea-project/GenAIComps/pull/641) and [AMD® GPUs](https://github.com/opea-project/GenAIExamples/issues/1153); introduces monitoring enhancements that provide real-time insights into runtime status and system resource utilization for [CPU and Intel® Gaudi® AI Accelerator](https://github.com/opea-project/GenAIEval/issues/195), as well as [Horizontal Pod Autoscaling (HPA)](https://github.com/opea-project/GenAIInfra/pull/531); introduces Helm Chart support for 7 GenAIExamples and their microservices; integrates benchmark tools for long-context language models (LCLMs) such as [LongBench](https://github.com/opea-project/GenAIEval/pull/179) and [HELMET](https://github.com/opea-project/GenAIEval/pull/182). ### Highlights #### New GenAI Examples -- [AvatarChatbot](https://github.com/opea-project/GenAIExamples/tree/main/AvatarChatbot): a chatbot that combines a virtual "avatar" on either Intel Gaudi 2 AI Acclerator or Intel Xeon Scalable Processors. -- [DBQnA](https://github.com/opea-project/GenAIExamples/tree/main/DBQnA): it seamlessly translates natural language queries into SQL and delivers real-time database results. +- [AvatarChatbot](https://github.com/opea-project/GenAIExamples/tree/main/AvatarChatbot): a chatbot that combines a virtual "avatar" that can run on either Intel Gaudi 2 AI Accelerator or Intel Xeon Scalable Processors. +- [DBQnA](https://github.com/opea-project/GenAIExamples/tree/main/DBQnA): for seamless translation of natural language queries into SQL and deliver real-time database results. - [EdgeCraftRAG](https://github.com/opea-project/GenAIExamples/tree/main/EdgeCraftRAG): a customizable and tunable RAG example for edge solutions on Intel® Arc™ GPUs. -- [GraphRAG](https://github.com/opea-project/GenAIExamples/tree/main/GraphRAG): a use case presents a Graph RAG approach to query-focused summarization. -- [Text2Image](https://github.com/opea-project/GenAIExamples/tree/main/Text2Image): an application that generates image conditioning on the provided texts. +- [GraphRAG](https://github.com/opea-project/GenAIExamples/tree/main/GraphRAG): a Graph RAG-based approach to summarization. +- [Text2Image](https://github.com/opea-project/GenAIExamples/tree/main/Text2Image): an application that generates images based on text prompts. - [WorkflowExecAgent](https://github.com/opea-project/GenAIExamples/tree/main/WorkflowExecAgent): a workflow executor example to handle data/AI workflow operations via LangChain agents to execute custom-defined workflow-based tools. #### Enhanced GenAI Examples @@ -33,7 +33,7 @@ This release introduces more scenarios with general availability, including the - Multimedia support ([29ef6426](https://github.com/opea-project/GenAIComps/commit/29ef6426), [baafa402](https://github.com/opea-project/GenAIComps/commit/baafa402)) #### GenAIStudio -[GenAI Studio](https://github.com/opea-project/GenAIStudio), a new project of OPEA, streamlines the creation of enterprise Generative AI applications by providing alternate to manual processes with a seamless, end-to-end solution. From GenAI app development and evaluation to performance benchmarking and deployment, GenAI Studio empowers developers to effortlessly build, test, optimize their LLM solutions and create the deployment package. Its intuitive no-code/low-code interface accelerates innovation, enabling rapid development and deployment of cutting-edge AI applications with unparalleled efficiency and precision. +[GenAI Studio](https://github.com/opea-project/GenAIStudio), a new project of OPEA, streamlines the creation of enterprise Generative AI applications by providing an alternative UI-based processes to create end-to-end solutions. It supports GenAI application definition, evaluation, performance benchmarking, and deployment. The GenAI Studio empowers developers to effortlessly build, test, optimize their LLM solutions, and create a deployment package. Its intuitive no-code/low-code interface accelerates innovation, enabling rapid development and deployment of cutting-edge AI applications with unparalleled efficiency and precision. #### Enhanced Observability Observability offers real-time insights into component performance and system resource utilization. We enhanced this capability by monitoring [key system metrics](https://github.com/opea-project/GenAIInfra/pull/541), including [CPU](https://github.com/opea-project/GenAIInfra/pull/541), host memory, storage, network, and accelerators (such as [Intel Gaudi](https://github.com/opea-project/GenAIInfra/pull/541)), as well as tracking [OPEA application scaling](https://github.com/opea-project/GenAIInfra/pull/541). @@ -54,7 +54,7 @@ We added the following two benchmark kits to response to the community's require #### Newly Supported Hardware - Intel® Arc™ GPU: [vLLM powered by OpenVINO](https://github.com/opea-project/GenAIComps/pull/729) can perform optimal model serving on Intel® Arc™ GPU. -- AMD® GPU: deploy GenAI examples on AMD® GPU using AMD® ROCm™: [CodeTrans](https://github.com/opea-project/GenAIExamples/pull/1138), [CodeGen](https://github.com/opea-project/GenAIExamples/pull/1130), [FaqGen](https://github.com/opea-project/GenAIExamples/pull/1126), [DocSum](https://github.com/opea-project/GenAIExamples/pull/1125), [ChatQnA](https://github.com/opea-project/GenAIExamples/pull/1122). +- AMD® GPU: deploy GenAI examples on AMD® GPUs using AMD® ROCm™: [CodeTrans](https://github.com/opea-project/GenAIExamples/pull/1138), [CodeGen](https://github.com/opea-project/GenAIExamples/pull/1130), [FaqGen](https://github.com/opea-project/GenAIExamples/pull/1126), [DocSum](https://github.com/opea-project/GenAIExamples/pull/1125), [ChatQnA](https://github.com/opea-project/GenAIExamples/pull/1122). ### Notable Changes @@ -259,7 +259,7 @@ We added the following two benchmark kits to response to the community's require - [ChatQnA] Implement the nowrapper version chatqna [71c81d0](https://github.com/opea-project/GenAIInfra/commit/71c81d0) - [FaqGen] Add FaqGen helm chart [f847e05](https://github.com/opea-project/GenAIInfra/commit/f847e05) - [FaqGen] helm: Add llm-faqgen-tgi support [325126e](https://github.com/opea-project/GenAIInfra/commit/325126e) - - [HPA] helm/manifest: Sync HPA related K8S probe settings [c399578](https://github.com/opea-project/GenAIInfra/commit/c399578) + - [HPA] helm/manifest: Sync HPA related k8s probe settings [c399578](https://github.com/opea-project/GenAIInfra/commit/c399578) - [VisualQnA] Add helm chart for VisualQnA example [b077d44](https://github.com/opea-project/GenAIInfra/commit/b077d44) - [UI] support variants for multiple examples [96af2ad](https://github.com/opea-project/GenAIInfra/commit/96af2ad) - [Nginx] helm-chart: Make nginx service type configurable [a5c96ab](https://github.com/opea-project/GenAIInfra/commit/a5c96ab) @@ -300,7 +300,7 @@ We added the following two benchmark kits to response to the community's require - Fix link check failure (#481) (5 weeks ago) [fc87ef3](https://github.com/opea-project/GenAIInfra/commit/fc87ef3) - Fix CI failures (#477) (5 weeks ago) [7e7b8ab](https://github.com/opea-project/GenAIInfra/commit/7e7b8ab) - Optimize path and link validity check. [91bd163](https://github.com/opea-project/GenAIInfra/commit/91bd163) - - enable image build process for memory-bandwidth-exporter [ddeac46](https://github.com/opea-project/GenAIInfra/commit/ddeac46) + -Enable image build process for memory-bandwidth-exporter [ddeac46](https://github.com/opea-project/GenAIInfra/commit/ddeac46) - Add hyperlinks and paths validation. [d8cd3a1](https://github.com/opea-project/GenAIInfra/commit/d8cd3a1) @@ -314,7 +314,7 @@ We added the following two benchmark kits to response to the community's require ## Contributors This release would not have been possible without the contributions of the following organizations and individuals. -### Orgnization Contributors +### Contributing Organizations - `AMD`: AMD CPU/GPU support. - `Capital One`: Contributions to CI/CD process. - `China Unicom`: Contributions to the deployment of GenAI examples. From 6916fa1f93875d20bdd5422876d60a7d737c3a95 Mon Sep 17 00:00:00 2001 From: Yi Yao Date: Fri, 22 Nov 2024 09:17:12 +0800 Subject: [PATCH 07/10] Update release_notes/v1.1.md according to Eero's suggestions Co-authored-by: Eero Tamminen --- release_notes/v1.1.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/release_notes/v1.1.md b/release_notes/v1.1.md index b6f3701c..560632d0 100644 --- a/release_notes/v1.1.md +++ b/release_notes/v1.1.md @@ -109,7 +109,7 @@ We added the following two benchmark kits to response to the community's require - [ChatQnA] Add instructions of modifying reranking docker image for NVGPU [2587179](https://github.com/opea-project/GenAIExamples/commit/2587179) - [ChatQnA] setup ollama service in aipc docker compose [def39cf](https://github.com/opea-project/GenAIExamples/commit/def39cf) - [ChatQnA] Make rerank run on gaudi for hpu docker compose [3c164f3](https://github.com/opea-project/GenAIExamples/commit/3c164f3) - - [ChatQnA] Added the K8s yaml for vLLM support [e2f9037](https://github.com/opea-project/GenAIExamples/commit/e2f9037) + - [ChatQnA] Added the k8s yaml for vLLM support [e2f9037](https://github.com/opea-project/GenAIExamples/commit/e2f9037) - [ChatQnA] manage your own ChatQnA pipelines. [d16c80e](https://github.com/opea-project/GenAIExamples/commit/d16c80e) - [ChatQnA] docker install instruction for csp [75df2c9](https://github.com/opea-project/GenAIExamples/commit/75df2c9) - [ChatQnA] ChatQnA with Remote Inference Endpoints (Kubernetes) [56f770c](https://github.com/opea-project/GenAIExamples/commit/56f770c) From a4494a9e0931961a86aefecef3d40ca11f0f2e69 Mon Sep 17 00:00:00 2001 From: Yi Yao Date: Fri, 22 Nov 2024 10:01:07 +0800 Subject: [PATCH 08/10] Update v1.1 release notes according to review suggestions. --- release_notes/v1.1.md | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/release_notes/v1.1.md b/release_notes/v1.1.md index 560632d0..9f59e8d6 100644 --- a/release_notes/v1.1.md +++ b/release_notes/v1.1.md @@ -1,7 +1,7 @@ # OPEA Release Notes v1.1 We are pleased to announce the release of OPEA version 1.1, which includes significant contributions from the open-source community. This release addresses over 460 pull requests. -More information about how to get started with OPEA v1.1 can be found at [Getting Started](https://opea-project.github.io/latest/index.html) page. All project source code is maintained in the [repository](https://github.com/opea-project). To pull Docker images, please access the [hub](https://hub.docker.com/u/opea). +More information about how to get started with OPEA v1.1 can be found at [Getting Started](https://opea-project.github.io/latest/index.html) page. All project source code is maintained in the [repository](https://github.com/opea-project). To pull Docker images, please access the [Docker Hub](https://hub.docker.com/u/opea). For instructions on deploying Helm Charts, please refer to the [guide](https://github.com/opea-project/GenAIInfra/tree/v1.1/helm-charts#readme). ## What's New in OPEA v1.1 This release introduces more scenarios with general availability, including the [Image-to-Video](https://github.com/opea-project/GenAIComps/pull/465), [Text-to-Image](https://github.com/opea-project/GenAIComps/pull/729), [Text-to-SQL](https://github.com/opea-project/GenAIExamples/tree/main/DBQnA) and [Avatar Animation](https://github.com/opea-project/GenAIComps/pull/775) capabilities; introduces [Generative AI Studio](https://github.com/opea-project/GenAIStudio) that offers a no-code alternative to create enterprise Generative AI applications; expands the portfolio of supported hardware to include [Intel® Arc™ GPUs](https://github.com/opea-project/GenAIComps/pull/641) and [AMD® GPUs](https://github.com/opea-project/GenAIExamples/issues/1153); introduces monitoring enhancements that provide real-time insights into runtime status and system resource utilization for [CPU and Intel® Gaudi® AI Accelerator](https://github.com/opea-project/GenAIEval/issues/195), as well as [Horizontal Pod Autoscaling (HPA)](https://github.com/opea-project/GenAIInfra/pull/531); introduces Helm Chart support for 7 GenAIExamples and their microservices; integrates benchmark tools for long-context language models (LCLMs) such as [LongBench](https://github.com/opea-project/GenAIEval/pull/179) and [HELMET](https://github.com/opea-project/GenAIEval/pull/182). @@ -39,12 +39,12 @@ This release introduces more scenarios with general availability, including the Observability offers real-time insights into component performance and system resource utilization. We enhanced this capability by monitoring [key system metrics](https://github.com/opea-project/GenAIInfra/pull/541), including [CPU](https://github.com/opea-project/GenAIInfra/pull/541), host memory, storage, network, and accelerators (such as [Intel Gaudi](https://github.com/opea-project/GenAIInfra/pull/541)), as well as tracking [OPEA application scaling](https://github.com/opea-project/GenAIInfra/pull/541). #### Helm Charts Support -OPEA examples and microservices support Helm Charts as the packaging format on K8S. The newly supported examples include [AgentQnA](https://github.com/opea-project/GenAIInfra/commit/66de41c), [AudioQnA](https://github.com/opea-project/GenAIInfra/commit/9efacee), [FaqGen](https://github.com/opea-project/GenAIInfra/commit/f847e05), [VisualQnA](https://github.com/opea-project/GenAIInfra/commit/b077d44) etc. The newly supported microservices include [chathistory, mongodb, prompt](https://github.com/opea-project/GenAIInfra/commit/881e2b5), and [Milvus for data-prep and retriever](https://github.com/opea-project/GenAIInfra/commit/d289b4e). +OPEA examples and microservices support Helm Charts as the packaging format on K8S. The newly supported examples include [AgentQnA](https://github.com/opea-project/GenAIInfra/blob/v1.1/helm-charts/agentqna/README.md), [AudioQnA](https://github.com/opea-project/GenAIInfra/tree/v1.1/helm-charts/audioqna/README.md), [FaqGen](https://github.com/opea-project/GenAIInfra/blob/v1.1/helm-charts/faqgen/README.md), [VisualQnA](https://github.com/opea-project/GenAIInfra/blob/v1.1/helm-charts/visualqna/README.md). The newly supported microservices include [chathistory, mongodb, prompt](https://github.com/opea-project/GenAIInfra/commit/881e2b5), and [Milvus for data-prep and retriever](https://github.com/opea-project/GenAIInfra/commit/d289b4e). Helm Charts have now option to get Prometheus metrics from the applications. #### Long-context Benchmark Support We added the following two benchmark kits to response to the community's requirements of long-context language models. -- HELMET: a comprehensive benchmark for long-context language models covering seven diverse categories of tasks. The datasets are application-centric and are designed to evaluate models at different lengths and levels of complexity. +- [HELMET](https://github.com/princeton-nlp/HELMET): a comprehensive benchmark for long-context language models covering seven diverse categories of tasks. The datasets are application-centric and are designed to evaluate models at different lengths and levels of complexity. - [LongBench](https://github.com/THUDM/LongBench): a benchmark tool for bilingual, multitask, and comprehensive assessment of long context understanding capabilities of large language models. #### Newly Supported Models @@ -181,7 +181,7 @@ We added the following two benchmark kits to response to the community's require - [Agent] support custom prompt [3473bfb3](https://github.com/opea-project/GenAIComps/commit/3473bfb3) - agent short & long term memory with langgraph. [e39b08f3](https://github.com/opea-project/GenAIComps/commit/e39b08f3) - support faqgen upload file in UI [453ff726](https://github.com/opea-project/GenAIComps/commit/453ff726) - - Add E2E Promeheus metrics to applications [a6998a1d](https://github.com/opea-project/GenAIComps/commit/a6998a1d) + - Add E2E Prometheus metrics to applications [a6998a1d](https://github.com/opea-project/GenAIComps/commit/a6998a1d) - Multiple models support for LLM TGI [e879366c](https://github.com/opea-project/GenAIComps/commit/e879366c) - Add RAG agent and ReAct agent implemention for llama3.1 served by TGI-gaudi [e7fdf537](https://github.com/opea-project/GenAIComps/commit/e7fdf537) - Support Llama3.2 vision and vision guard model [534c227a](https://github.com/opea-project/GenAIComps/commit/534c227a) From 984b728e158f0dd577b9b6228d8144e13cfd1877 Mon Sep 17 00:00:00 2001 From: Yi Yao Date: Fri, 22 Nov 2024 10:21:10 +0800 Subject: [PATCH 09/10] Re-format "What's New" section of v1.1 release --- release_notes/v1.1.md | 10 ++++++++-- 1 file changed, 8 insertions(+), 2 deletions(-) diff --git a/release_notes/v1.1.md b/release_notes/v1.1.md index 9f59e8d6..aff9ccf3 100644 --- a/release_notes/v1.1.md +++ b/release_notes/v1.1.md @@ -1,10 +1,16 @@ # OPEA Release Notes v1.1 -We are pleased to announce the release of OPEA version 1.1, which includes significant contributions from the open-source community. This release addresses over 460 pull requests. +We are pleased to announce the release of OPEA version 1.1, which includes significant contributions from the open-source community. This release addresses over 470 pull requests. More information about how to get started with OPEA v1.1 can be found at [Getting Started](https://opea-project.github.io/latest/index.html) page. All project source code is maintained in the [repository](https://github.com/opea-project). To pull Docker images, please access the [Docker Hub](https://hub.docker.com/u/opea). For instructions on deploying Helm Charts, please refer to the [guide](https://github.com/opea-project/GenAIInfra/tree/v1.1/helm-charts#readme). ## What's New in OPEA v1.1 -This release introduces more scenarios with general availability, including the [Image-to-Video](https://github.com/opea-project/GenAIComps/pull/465), [Text-to-Image](https://github.com/opea-project/GenAIComps/pull/729), [Text-to-SQL](https://github.com/opea-project/GenAIExamples/tree/main/DBQnA) and [Avatar Animation](https://github.com/opea-project/GenAIComps/pull/775) capabilities; introduces [Generative AI Studio](https://github.com/opea-project/GenAIStudio) that offers a no-code alternative to create enterprise Generative AI applications; expands the portfolio of supported hardware to include [Intel® Arc™ GPUs](https://github.com/opea-project/GenAIComps/pull/641) and [AMD® GPUs](https://github.com/opea-project/GenAIExamples/issues/1153); introduces monitoring enhancements that provide real-time insights into runtime status and system resource utilization for [CPU and Intel® Gaudi® AI Accelerator](https://github.com/opea-project/GenAIEval/issues/195), as well as [Horizontal Pod Autoscaling (HPA)](https://github.com/opea-project/GenAIInfra/pull/531); introduces Helm Chart support for 7 GenAIExamples and their microservices; integrates benchmark tools for long-context language models (LCLMs) such as [LongBench](https://github.com/opea-project/GenAIEval/pull/179) and [HELMET](https://github.com/opea-project/GenAIEval/pull/182). +This release introduces more scenarios with general availability, including: +- Newly supported Generative AI capabilities: [Image-to-Video](https://github.com/opea-project/GenAIComps/pull/465), [Text-to-Image](https://github.com/opea-project/GenAIComps/pull/729), [Text-to-SQL](https://github.com/opea-project/GenAIExamples/tree/main/DBQnA) and [Avatar Animation](https://github.com/opea-project/GenAIComps/pull/775). +- [Generative AI Studio](https://github.com/opea-project/GenAIStudio) that offers a no-code alternative to create enterprise Generative AI applications. +- Expands the portfolio of supported hardware to include [Intel® Arc™ GPUs](https://github.com/opea-project/GenAIComps/pull/641) and [AMD® GPUs](https://github.com/opea-project/GenAIExamples/issues/1153). +- Enhanced monitoring system, providing real-time insights into runtime status and system resource utilization for [CPU and Intel® Gaudi® AI Accelerator](https://github.com/opea-project/GenAIEval/issues/195), as well as [Horizontal Pod Autoscaling (HPA)](https://github.com/opea-project/GenAIInfra/pull/531). +- Helm Chart support for 7 GenAIExamples and their microservices. +- Benchmark tools for long-context language models (LCLMs) such as [LongBench](https://github.com/opea-project/GenAIEval/pull/179) and [HELMET](https://github.com/opea-project/GenAIEval/pull/182). ### Highlights #### New GenAI Examples From c16ba076ecbd0dbfe39e644faaba854401b8132b Mon Sep 17 00:00:00 2001 From: Yi Yao Date: Sat, 23 Nov 2024 08:43:58 +0800 Subject: [PATCH 10/10] Apply suggestions from Eero's review Co-authored-by: Eero Tamminen --- release_notes/v1.1.md | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/release_notes/v1.1.md b/release_notes/v1.1.md index aff9ccf3..3353976b 100644 --- a/release_notes/v1.1.md +++ b/release_notes/v1.1.md @@ -8,8 +8,8 @@ This release introduces more scenarios with general availability, including: - Newly supported Generative AI capabilities: [Image-to-Video](https://github.com/opea-project/GenAIComps/pull/465), [Text-to-Image](https://github.com/opea-project/GenAIComps/pull/729), [Text-to-SQL](https://github.com/opea-project/GenAIExamples/tree/main/DBQnA) and [Avatar Animation](https://github.com/opea-project/GenAIComps/pull/775). - [Generative AI Studio](https://github.com/opea-project/GenAIStudio) that offers a no-code alternative to create enterprise Generative AI applications. - Expands the portfolio of supported hardware to include [Intel® Arc™ GPUs](https://github.com/opea-project/GenAIComps/pull/641) and [AMD® GPUs](https://github.com/opea-project/GenAIExamples/issues/1153). -- Enhanced monitoring system, providing real-time insights into runtime status and system resource utilization for [CPU and Intel® Gaudi® AI Accelerator](https://github.com/opea-project/GenAIEval/issues/195), as well as [Horizontal Pod Autoscaling (HPA)](https://github.com/opea-project/GenAIInfra/pull/531). -- Helm Chart support for 7 GenAIExamples and their microservices. +- Enhanced monitoring support, providing real-time insights into runtime status and system resource utilization for [CPU and Intel® Gaudi® AI Accelerator](https://github.com/opea-project/GenAIEval/issues/195), as well as [Horizontal Pod Autoscaling (HPA)](https://github.com/opea-project/GenAIInfra/pull/531). +- Helm Chart support for 7 new GenAIExamples and their microservices. - Benchmark tools for long-context language models (LCLMs) such as [LongBench](https://github.com/opea-project/GenAIEval/pull/179) and [HELMET](https://github.com/opea-project/GenAIEval/pull/182). ### Highlights @@ -45,7 +45,7 @@ This release introduces more scenarios with general availability, including: Observability offers real-time insights into component performance and system resource utilization. We enhanced this capability by monitoring [key system metrics](https://github.com/opea-project/GenAIInfra/pull/541), including [CPU](https://github.com/opea-project/GenAIInfra/pull/541), host memory, storage, network, and accelerators (such as [Intel Gaudi](https://github.com/opea-project/GenAIInfra/pull/541)), as well as tracking [OPEA application scaling](https://github.com/opea-project/GenAIInfra/pull/541). #### Helm Charts Support -OPEA examples and microservices support Helm Charts as the packaging format on K8S. The newly supported examples include [AgentQnA](https://github.com/opea-project/GenAIInfra/blob/v1.1/helm-charts/agentqna/README.md), [AudioQnA](https://github.com/opea-project/GenAIInfra/tree/v1.1/helm-charts/audioqna/README.md), [FaqGen](https://github.com/opea-project/GenAIInfra/blob/v1.1/helm-charts/faqgen/README.md), [VisualQnA](https://github.com/opea-project/GenAIInfra/blob/v1.1/helm-charts/visualqna/README.md). The newly supported microservices include [chathistory, mongodb, prompt](https://github.com/opea-project/GenAIInfra/commit/881e2b5), and [Milvus for data-prep and retriever](https://github.com/opea-project/GenAIInfra/commit/d289b4e). Helm Charts have now option to get Prometheus metrics from the applications. +OPEA examples and microservices support Helm Charts as the packaging format on Kubernetes (k8s). The newly supported examples include [AgentQnA](https://github.com/opea-project/GenAIInfra/blob/v1.1/helm-charts/agentqna/README.md), [AudioQnA](https://github.com/opea-project/GenAIInfra/tree/v1.1/helm-charts/audioqna/README.md), [FaqGen](https://github.com/opea-project/GenAIInfra/blob/v1.1/helm-charts/faqgen/README.md), [VisualQnA](https://github.com/opea-project/GenAIInfra/blob/v1.1/helm-charts/visualqna/README.md). The newly supported microservices include [chathistory, mongodb, prompt](https://github.com/opea-project/GenAIInfra/commit/881e2b5), and [Milvus for data-prep and retriever](https://github.com/opea-project/GenAIInfra/commit/d289b4e). Helm Charts have now option to get Prometheus metrics from the applications. #### Long-context Benchmark Support We added the following two benchmark kits to response to the community's requirements of long-context language models.