# RAG

Ducky
Ducky is a fully managed AI retrieval service designed for developers who need fast and accurate results. It supports semantic search, including Retrieval-Augmented Generation (RAG), and provides a straightforward Python SDK for quickly building excellent search functionalities.
Search
40.0K

Contextual AI Reranker
Contextual AI Reranker is a revolutionary AI model designed to solve information conflicts and inaccurate ranking issues in enterprise-level Retrieval Augmented Generation (RAG) systems. It can accurately rank retrieval results based on user-provided natural language instructions, ensuring that the most relevant information is displayed first. This product is based on advanced AI technology and has demonstrated excellent performance through industry-standard BEIR benchmark tests and internal dataset validation. Its main advantages include high accuracy, strong instruction-following capabilities, and flexible customization options, suitable for various fields such as finance, technology, and professional services. The product currently offers a free trial and is accessed via API, facilitating quick deployment and use by enterprises.
AI Search
69.0K

Wdoc
wdoc is a RAG system developed by Olicorne (a medical student) to address document querying and summarization using retrieval-augmented generation technology. It supports multiple file types (such as PDFs, web pages, YouTube videos, etc.) and combines various language models to provide high-recall and high-precision query results. wdoc's main advantages include robust support for multiple file types, efficient retrieval capabilities, and flexible extensibility. It is suitable for researchers, students, and professionals, helping them process large amounts of information quickly. wdoc is currently under development, and the developer welcomes user feedback and feature requests to continuously improve the product.
Knowledge Management
59.1K

Site RAG
Site RAG is a Chrome extension designed to assist users in quickly obtaining answers to questions while browsing the web, using natural language processing technology. It supports querying with the current page content as context and allows the indexing of entire website content into a vector database for subsequent retrieval-enhanced generation (RAG). The product runs entirely in the local browser, ensuring user data security, and supports connection to a locally running Ollama instance for inference. It primarily targets users who need to rapidly extract information from web content, such as developers, researchers, and students. This product is currently offered for free, making it suitable for those seeking immediate assistance while browsing the web.
AI search
67.9K

Rag Chat Component
This product is a React component specifically designed for RAG (Retrieval-Augmented Generation) AI assistants. It combines Upstash Vector for similarity searching, Together AI as the LLM (Large Language Model), and Vercel AI SDK for streaming responses. This modular design enables developers to rapidly incorporate RAG capabilities into Next.js applications, greatly simplifying the development process while providing high customizability. Key advantages include responsive design, support for streaming responses, persistent chat history, and dark/light mode support. The component primarily targets developers looking to integrate intelligent chat functionality into Web applications, particularly those working with the Next.js framework, streamlining the integration process and reducing development costs while delivering powerful capabilities.
Development & Tools
56.3K

RAG Logger
RAG-logger is an open-source logging tool designed for Retrieval-Augmented Generation (RAG) applications. It is a lightweight open-source alternative tailored to the specific logging needs of RAG, focusing on providing comprehensive logging features such as query tracking, retrieval result logging, LLM interaction logging, and step-by-step performance monitoring. It utilizes a JSON-based log format, supports daily log organization, automatic file management, and metadata enrichment. With its open-source, lightweight, and RAG-focused features, RAG-logger offers developers an effective tool for monitoring and analyzing the performance of RAG applications.
Development & Tools
57.1K

Command R7B
Command R7B is a high-performance, scalable large language model (LLM) introduced by Cohere, specifically designed for enterprise applications. It delivers top-tier speed, efficiency, and quality while maintaining a compact model size, significantly lowering the production deployment costs of AI applications on standard GPUs, edge devices, or even CPUs. Command R7B excels in multilingual support, retrieval-augmented generation (RAG), reasoning, tool usage, and agent behavior, making it ideal for enterprises focusing on optimizing speed, cost efficiency, and computational resources.
AI Model
53.8K

E2M
E2M is a Python library capable of parsing and converting multiple file types into Markdown format. It employs a parser-converter architecture, supporting the conversion of a variety of file formats including doc, docx, epub, html, htm, url, pdf, ppt, pptx, mp3, and m4a. The ultimate aim of the E2M project is to provide high-quality data for Retrieval-Augmented Generation (RAG) and model training or fine-tuning.
Development & Tools
62.7K

Vision Is All You Need
vision-is-all-you-need is a demonstration project showcasing the Vision RAG (V-RAG) architecture. The V-RAG architecture directly embeds PDF file pages (or other documents) into vectors using Vision Language Models (VLM), eliminating the need for cumbersome chunk processing. This technology enhances the efficiency and accuracy of document retrieval, especially when dealing with large datasets. Background information indicates that this is an innovative tool leveraging the latest AI technologies to improve document processing capabilities. The project is currently open-source and free to use.
Knowledge Management
50.8K

Minima
Minima is an open-source, fully localized Retrieval-Augmented Generation (RAG) model, capable of integrating with ChatGPT and the Model Context Protocol (MCP). It supports three modes: full local installation, querying local documents through ChatGPT, and querying local files using Anthropic Claude. The primary advantages of Minima include localized data processing, privacy protection, and the ability to leverage powerful language models to enhance retrieval and generation tasks. Background information indicates that Minima supports multiple file formats and allows users to customize configurations to fit different usage scenarios. Minima is free and open-source, targeting developers and enterprises seeking localized AI solutions.
Development & Tools
93.6K

Qwen Agent
Qwen-Agent is an Agent framework built on Qwen>=2.0, equipped with capabilities for instruction following, tool usage, planning, and memory. The framework showcases example applications such as browser assistants, code interpreters, and custom assistants. The primary advantages of Qwen-Agent include its high scalability and modular design, allowing developers to integrate various tools and functionalities as needed. Background information indicates that Qwen-Agent aims to provide developers with a robust toolkit for building and deploying applications based on large language models. Qwen-Agent is open-source on GitHub, enabling community contributions and collaboration.
Development and Tools
66.0K

Inquir
Inquir is a powerful tool designed to create personalized search engines tailored to your data. It unlocks features such as custom search solutions, data aggregation, AI-driven Retrieval-Augmented Generation (RAG) systems, and context-aware search capabilities. Take the first step towards enhancing user experience by launching your engine or scheduling a demonstration.
AI Search
50.0K

Chonkie
Chonkie is a text chunking library designed for Retrieval-Augmented Generation (RAG) applications. It is lightweight, fast, and user-friendly. The library provides various text chunking methods, supports multiple tokenizers, and boasts high performance. Key advantages of Chonkie include rich functionality, ease of use, rapid processing speeds, extensive support, and a lightweight design. It is suitable for developers and researchers who require efficient text data processing, especially in natural language processing and machine learning. Chonkie is open-source and complies with the MIT license, making it freely available.
Development & Tools
67.1K
Fresh Picks

Trieve
Trieve is an AI-first infrastructure API that combines language models and tools for fine-tuning ranking and relevance, providing a one-stop solution for search, recommendations, RAG, and analytics. It can continuously improve automatically based on dozens of feedback signals, ensuring relevance. Trieve supports semantic vector search, BM25, and SPLADE full-text search, as well as hybrid search combining full-text and semantic vector searches. Additionally, it offers product promotion and relevance adjustment features to help users fine-tune search results via API or no-code dashboards to achieve KPIs. Built on the best foundations, it utilizes open-source embedding models and LLMs, running on its own servers to ensure data security.
Search
58.0K

Dabarqus
Dabarqus is a Retrieval-Augmented Generation (RAG) framework that allows users to provide private data to large language models (LLMs) in real time. This tool facilitates the storage of various data sources (such as PDFs, emails, and raw data) into semantic indices, referred to as 'memory repositories,' through REST APIs, SDKs, and CLI tools. Dabarqus supports LLM-style prompts, enabling users to interact with memory repositories in a straightforward manner without needing to construct special queries or learn a new query language. Additionally, it allows for the creation and utilization of multiple semantic indices (memory repositories), organizing data by topics, categories, or other grouping methods. The product background of Dabarqus emphasizes its aim to simplify the integration of private data with AI language models, enhancing the efficiency and accuracy of data retrieval.
Development & Tools
50.5K

Vectorize
Vectorize is a platform focused on transforming unstructured data into optimized vector search indices, specifically designed for retrieval-augmented generation (RAG). By connecting various data sources such as content management systems, file systems, CRM, and collaboration tools, it helps users create productivity-enhancing assistant systems and innovative customer experiences. Key advantages of Vectorize include ease of use, rapid deployment, and high-accuracy search results, making it suitable for enterprises that need to handle large amounts of data and wish to quickly realize AI applications.
AI Search
50.2K
English Picks

Epsilla
Epsilla is a no-code Retrieval-Augmented Generation as a Service (RAG-as-a-Service) platform that allows users to build production-ready Large Language Model (LLM) applications based on private or public data. The platform offers a one-stop solution, including data management, RAG tools, CI/CD-style evaluations, and enterprise-level security measures, designed to lower the total cost of ownership (TCO), enhance query speed and throughput, while ensuring timeliness and security of information.
Development Platform
61.0K

Kotaemon
Kotaemon is an open-source tool based on the Retrieval-Augmented Generation (RAG) model designed to interact with user documents through a chat interface. It supports various language model API providers and local language models, offering a clean and customizable user interface suitable for end users conducting document Q&A and developers building their own RAG Q&A workflows.
AI Conversational Agents
78.9K
English Picks

Ragie
Ragie is a Retrieval-Augmented Generation (RAG) as a service product targeted at developers, providing easy-to-use APIs and SDKs to help them quickly launch and implement generative AI applications. Ragie features advanced capabilities such as LLM re-ranking, summary indexing, and entity extraction to ensure accurate and reliable information. It supports direct connections to popular data sources like Google Drive and Notion, along with automatic syncing to keep data current. Led by Craft Ventures, Ragie offers a straightforward pricing strategy with no setup fees or hidden costs.
Development and Tools
56.6K

RAG Techniques
RAG_Techniques is a collection focused on Retrieval-Augmented Generation (RAG) systems, aimed at enhancing the accuracy, efficiency, and contextual richness of these systems. It serves as a cutting-edge technology hub that drives the development and innovation of RAG technology through community contributions and a collaborative environment.
AI Model
56.6K

Easy RAG
Easy-RAG is a Retrieval-Augmented Generation (RAG) system that is ideal for learners to understand and master RAG technology, while also being convenient for developers to use and expand independently. This system enhances retrieval efficiency and generation quality by integrating knowledge graph extraction tools, reranking mechanisms, and the FAISS vector database.
AI Model
97.4K

Ragfoundry
RAGFoundry is a library designed to enhance the ability of large language models (LLMs) to utilize external information by fine-tuning models on specially created RAG-augmented datasets. The library facilitates efficient model training using Parameter-Efficient Fine-Tuning (PEFT), allowing users to easily measure performance improvements with RAG-specific metrics. It features a modular design, enabling workflow customization through configuration files.
AI Development Assistant
54.1K

Korvus
Korvus is a search SDK built on Postgres, unifying the entire RAG (Retrieval Augmented Generation) process into a single database query. It offers high-performance, customizable search capabilities while minimizing infrastructure concerns. Korvus leverages PostgresML's pgml and pgvector extensions, compressing the RAG workflow within Postgres itself. It supports multi-language SDKs, including Python, JavaScript, Rust, and C, allowing developers to seamlessly integrate into existing technology stacks.
AI search engine
56.9K

Learn RAG With Langchain
Retrieval-Augmented Generation (RAG) is a cutting-edge technique that enhances the capabilities of generative models by integrating external knowledge sources, leading to higher quality and reliability in generated content. LangChain is a powerful framework designed specifically for building and deploying robust language model applications. This tutorial series offers a comprehensive, step-by-step guide to help you implement RAG using LangChain. It starts with an introduction to the fundamental RAG process and gradually delves into areas like query transformation, document embedding, routing mechanisms, query construction, indexing strategies, retrieval techniques, and the generation stage. Ultimately, it integrates all these concepts into a practical scenario, showcasing the power and flexibility of RAG.
Development and Tools
56.0K
Fresh Picks

Ragelo
RAGElo is a toolkit that leverages the Elo rating system to help select the best-performing Large Language Model (LLM) agents enhanced with Retrieval Augmented Generation (RAG). While prototyping and integrating generative LLMs in production has become easier, evaluation remains the most challenging aspect of these solutions. RAGElo addresses this by comparing the answers of different RAG pipelines and prompts to multiple questions, calculating rankings for various setups. This provides a clear overview of which configurations are effective and which are not.
AI Model
59.3K
Fresh Picks

DB GPT
DB-GPT is an open-source AI-native data application development framework that, utilizing AWEL (Agentic Workflow Expression Language) and agent technologies, simplifies the integration of large models with data. Through its capabilities in multi-model management, Text2SQL optimization, RAG framework optimization, and multi-agent framework collaboration, DB-GPT empowers enterprises and developers to build customized applications with less code. In the era of Data 3.0, DB-GPT, based on models and databases, provides foundational data intelligence technologies for building enterprise-level reporting, analysis, and business insights.
AI development assistant
66.5K
Fresh Picks

Gomate
GoMate is a model based on the Retrieval-Augmented Generation (RAG) framework, focused on delivering reliable input and trustworthy output. By combining retrieval and generation technologies, it enhances the accuracy and reliability of information retrieval and text generation. GoMate is suitable for fields requiring efficient and accurate information processing, such as natural language processing and question answering.
AI Model
57.7K

Omakase RAG Orchestrator
Omakase RAG Orchestrator is a project aimed at addressing the challenges encountered when building RAG applications. It provides a comprehensive web application and API to encapsulate large language models (LLMs) and their wrappers. The project integrates Django, Llamaindex, and Google Drive to enhance the application's usability, scalability, and data and user access management.
AI Development Assistant
54.6K

Verba
Verba is an open-source application designed to provide an end-to-end, seamless, and user-friendly retrieval-augmented generation (RAG) interface. It combines cutting-edge RAG techniques with Weaviate's context-aware database, supporting both local and cloud deployments, making it easy to explore datasets and extract insights.
AI search engine
63.8K
Fresh Picks

Command R+
Command R+ is an advanced RAG optimization model designed for enterprise-level workloads. It is the first to be launched on Microsoft Azure. This model boasts a 128k token context window, delivering best-in-class retrieval-augmented generation (RAG) performance. It supports 10 key languages for multilingual coverage, and features tool usage capabilities to automate complex business processes. Pricing for Command R+: $3.00/M input tokens, $15.00/M output tokens. It is suitable for a wide range of enterprise scenarios, such as finance, human resources, sales, marketing, and customer support.
AI Model
119.0K
- 1
- 2
Featured AI Tools
English Picks

Jules AI
Jules は、自動で煩雑なコーディングタスクを処理し、あなたに核心的なコーディングに時間をかけることを可能にする異步コーディングエージェントです。その主な強みは GitHub との統合で、Pull Request(PR) を自動化し、テストを実行し、クラウド仮想マシン上でコードを検証することで、開発効率を大幅に向上させています。Jules はさまざまな開発者に適しており、特に忙しいチームには効果的にプロジェクトとコードの品質を管理する支援を行います。
開発プログラミング
48.9K

Nocode
NoCode はプログラミング経験を必要としないプラットフォームで、ユーザーが自然言語でアイデアを表現し、迅速にアプリケーションを生成することが可能です。これにより、開発の障壁を下げ、より多くの人が自身のアイデアを実現できるようになります。このプラットフォームはリアルタイムプレビュー機能とワンクリックデプロイ機能を提供しており、技術的な知識がないユーザーにも非常に使いやすい設計となっています。
開発プラットフォーム
44.7K

Listenhub
ListenHub は軽量級の AI ポッドキャストジェネレーターであり、中国語と英語に対応しています。最先端の AI 技術を使用し、ユーザーが興味を持つポッドキャストコンテンツを迅速に生成できます。その主な利点には、自然な会話と超高品質な音声効果が含まれており、いつでもどこでも高品質な聴覚体験を楽しむことができます。ListenHub はコンテンツ生成速度を改善するだけでなく、モバイルデバイスにも対応しており、さまざまな場面で使いやすいです。情報取得の高効率なツールとして位置づけられており、幅広いリスナーのニーズに応えています。
AI
42.8K
Chinese Picks

腾讯混元画像 2.0
腾讯混元画像 2.0 は腾讯が最新に発表したAI画像生成モデルで、生成スピードと画質が大幅に向上しました。超高圧縮倍率のエンコード?デコーダーと新しい拡散アーキテクチャを採用しており、画像生成速度はミリ秒級まで到達し、従来の時間のかかる生成を回避することが可能です。また、強化学習アルゴリズムと人間の美的知識の統合により、画像のリアリズムと詳細表現力を向上させ、デザイナー、クリエーターなどの専門ユーザーに適しています。
画像生成
43.3K

Openmemory MCP
OpenMemoryはオープンソースの個人向けメモリレイヤーで、大規模言語モデル(LLM)に私密でポータブルなメモリ管理を提供します。ユーザーはデータに対する完全な制御権を持ち、AIアプリケーションを作成する際も安全性を保つことができます。このプロジェクトはDocker、Python、Node.jsをサポートしており、開発者が個別化されたAI体験を行うのに適しています。また、個人情報を漏らすことなくAIを利用したいユーザーにお勧めします。
オープンソース
45.0K

Fastvlm
FastVLM は、視覚言語モデル向けに設計された効果的な視覚符号化モデルです。イノベーティブな FastViTHD ミックスドビジュアル符号化エンジンを使用することで、高解像度画像の符号化時間と出力されるトークンの数を削減し、モデルのスループットと精度を向上させました。FastVLM の主な位置付けは、開発者が強力な視覚言語処理機能を得られるように支援し、特に迅速なレスポンスが必要なモバイルデバイス上で優れたパフォーマンスを発揮します。
画像処理
42.8K
English Picks

ピカ
ピカは、ユーザーが自身の創造的なアイデアをアップロードすると、AIがそれに基づいた動画を自動生成する動画制作プラットフォームです。主な機能は、多様なアイデアからの動画生成、プロフェッショナルな動画効果、シンプルで使いやすい操作性です。無料トライアル方式を採用しており、クリエイターや動画愛好家をターゲットとしています。
映像制作
17.6M
Chinese Picks

Liblibai
LiblibAIは、中国をリードするAI創作プラットフォームです。強力なAI創作能力を提供し、クリエイターの創造性を支援します。プラットフォームは膨大な数の無料AI創作モデルを提供しており、ユーザーは検索してモデルを使用し、画像、テキスト、音声などの創作を行うことができます。また、ユーザーによる独自のAIモデルのトレーニングもサポートしています。幅広いクリエイターユーザーを対象としたプラットフォームとして、創作の機会を平等に提供し、クリエイティブ産業に貢献することで、誰もが創作の喜びを享受できるようにすることを目指しています。
AIモデル
6.9M