I am working on supporting compatibility with other embeddding models, and we will have that soon, for now I had to reduce the compatibility only to Sentence Transformers.
For what concerns page numbers, I am also working toward having better and more extensive metadata: everything is a big work-in-progress and will come in future releases!
Clelia Astra Bertelli
as-cle-bert
AI & ML interests
Biology + Artificial Intelligence = โค๏ธ | AI for sustainable development, sustainable development for AI | Researching on Machine Learning Enhancement | I love automation for everyday things | Blogger | Open Source
Recent Activity
replied to
their
post
about 1 hour ago
Ever dreamt of ingesting into a vector DB that pile of CSVs, Word documents and presentations laying in some remote folders on your PC?๐๏ธ
What if I told you that you can do it within three to six lines of code?๐คฏ
Well, with my latest open-source project, ๐ข๐ง๐ ๐๐ฌ๐ญ-๐๐ง๐ฒ๐ญ๐ก๐ข๐ง๐ (https://github.com/AstraBert/ingest-anything), you can take all your non-PDF files, convert them to PDF, extract their text, chunk, embed and load them into a vector database, all in one go!๐
How? It's pretty simple!
๐ The input files are converted into PDF by PdfItDown (https://github.com/AstraBert/PdfItDown)
๐ The PDF text is extracted using LlamaIndex readers
๐ฆ The text is chunked exploiting Chonkie
๐งฎ The chunks are embedded thanks to Sentence Transformers models
๐๏ธ The embeddings are loaded into a Qdrant vector database
And you're done!โ
Curious of trying it? Install it by running:
๐ฑ๐ช๐ฑ ๐ช๐ฏ๐ด๐ต๐ข๐ญ๐ญ ๐ช๐ฏ๐จ๐ฆ๐ด๐ต-๐ข๐ฏ๐บ๐ต๐ฉ๐ช๐ฏ๐จ
And you can start using it in your python scripts!๐
Don't forget to star it on GitHub and let me know if you have any feedback! โก๏ธ https://github.com/AstraBert/ingest-anything
replied to
their
post
2 days ago
Ever dreamt of ingesting into a vector DB that pile of CSVs, Word documents and presentations laying in some remote folders on your PC?๐๏ธ
What if I told you that you can do it within three to six lines of code?๐คฏ
Well, with my latest open-source project, ๐ข๐ง๐ ๐๐ฌ๐ญ-๐๐ง๐ฒ๐ญ๐ก๐ข๐ง๐ (https://github.com/AstraBert/ingest-anything), you can take all your non-PDF files, convert them to PDF, extract their text, chunk, embed and load them into a vector database, all in one go!๐
How? It's pretty simple!
๐ The input files are converted into PDF by PdfItDown (https://github.com/AstraBert/PdfItDown)
๐ The PDF text is extracted using LlamaIndex readers
๐ฆ The text is chunked exploiting Chonkie
๐งฎ The chunks are embedded thanks to Sentence Transformers models
๐๏ธ The embeddings are loaded into a Qdrant vector database
And you're done!โ
Curious of trying it? Install it by running:
๐ฑ๐ช๐ฑ ๐ช๐ฏ๐ด๐ต๐ข๐ญ๐ญ ๐ช๐ฏ๐จ๐ฆ๐ด๐ต-๐ข๐ฏ๐บ๐ต๐ฉ๐ช๐ฏ๐จ
And you can start using it in your python scripts!๐
Don't forget to star it on GitHub and let me know if you have any feedback! โก๏ธ https://github.com/AstraBert/ingest-anything
posted
an
update
3 days ago
Ever dreamt of ingesting into a vector DB that pile of CSVs, Word documents and presentations laying in some remote folders on your PC?๐๏ธ
What if I told you that you can do it within three to six lines of code?๐คฏ
Well, with my latest open-source project, ๐ข๐ง๐ ๐๐ฌ๐ญ-๐๐ง๐ฒ๐ญ๐ก๐ข๐ง๐ (https://github.com/AstraBert/ingest-anything), you can take all your non-PDF files, convert them to PDF, extract their text, chunk, embed and load them into a vector database, all in one go!๐
How? It's pretty simple!
๐ The input files are converted into PDF by PdfItDown (https://github.com/AstraBert/PdfItDown)
๐ The PDF text is extracted using LlamaIndex readers
๐ฆ The text is chunked exploiting Chonkie
๐งฎ The chunks are embedded thanks to Sentence Transformers models
๐๏ธ The embeddings are loaded into a Qdrant vector database
And you're done!โ
Curious of trying it? Install it by running:
๐ฑ๐ช๐ฑ ๐ช๐ฏ๐ด๐ต๐ข๐ญ๐ญ ๐ช๐ฏ๐จ๐ฆ๐ด๐ต-๐ข๐ฏ๐บ๐ต๐ฉ๐ช๐ฏ๐จ
And you can start using it in your python scripts!๐
Don't forget to star it on GitHub and let me know if you have any feedback! โก๏ธ https://github.com/AstraBert/ingest-anything
Organizations
as-cle-bert's activity

replied to
their
post
about 1 hour ago

replied to
their
post
2 days ago
So, there are two possibilities:
- If you mean customizing the embedder among the ones available within Sentence Transformers, it is very possible, you just have to change the
embedding_model
parameter when calling theingest
method - If you mean that you have your own embedding model (like saved on your PC), that is a tad more difficult. I think Sentence Transformer might allow loading the model from your PC as long as it is compatible with the package. I think that this guide might be useful in that regard
For now the package only supports Sentence Transformers models, in the future it will probably extend its support to other embedding models as well :)

posted
an
update
3 days ago
Post
2701
Ever dreamt of ingesting into a vector DB that pile of CSVs, Word documents and presentations laying in some remote folders on your PC?๐๏ธ
What if I told you that you can do it within three to six lines of code?๐คฏ
Well, with my latest open-source project, ๐ข๐ง๐ ๐๐ฌ๐ญ-๐๐ง๐ฒ๐ญ๐ก๐ข๐ง๐ (https://github.com/AstraBert/ingest-anything), you can take all your non-PDF files, convert them to PDF, extract their text, chunk, embed and load them into a vector database, all in one go!๐
How? It's pretty simple!
๐ The input files are converted into PDF by PdfItDown (https://github.com/AstraBert/PdfItDown)
๐ The PDF text is extracted using LlamaIndex readers
๐ฆ The text is chunked exploiting Chonkie
๐งฎ The chunks are embedded thanks to Sentence Transformers models
๐๏ธ The embeddings are loaded into a Qdrant vector database
And you're done!โ
Curious of trying it? Install it by running:
๐ฑ๐ช๐ฑ ๐ช๐ฏ๐ด๐ต๐ข๐ญ๐ญ ๐ช๐ฏ๐จ๐ฆ๐ด๐ต-๐ข๐ฏ๐บ๐ต๐ฉ๐ช๐ฏ๐จ
And you can start using it in your python scripts!๐
Don't forget to star it on GitHub and let me know if you have any feedback! โก๏ธ https://github.com/AstraBert/ingest-anything
What if I told you that you can do it within three to six lines of code?๐คฏ
Well, with my latest open-source project, ๐ข๐ง๐ ๐๐ฌ๐ญ-๐๐ง๐ฒ๐ญ๐ก๐ข๐ง๐ (https://github.com/AstraBert/ingest-anything), you can take all your non-PDF files, convert them to PDF, extract their text, chunk, embed and load them into a vector database, all in one go!๐
How? It's pretty simple!
๐ The input files are converted into PDF by PdfItDown (https://github.com/AstraBert/PdfItDown)
๐ The PDF text is extracted using LlamaIndex readers
๐ฆ The text is chunked exploiting Chonkie
๐งฎ The chunks are embedded thanks to Sentence Transformers models
๐๏ธ The embeddings are loaded into a Qdrant vector database
And you're done!โ
Curious of trying it? Install it by running:
๐ฑ๐ช๐ฑ ๐ช๐ฏ๐ด๐ต๐ข๐ญ๐ญ ๐ช๐ฏ๐จ๐ฆ๐ด๐ต-๐ข๐ฏ๐บ๐ต๐ฉ๐ช๐ฏ๐จ
And you can start using it in your python scripts!๐
Don't forget to star it on GitHub and let me know if you have any feedback! โก๏ธ https://github.com/AstraBert/ingest-anything

replied to
their
post
5 days ago
Hey @T-2000 , you're absolutely right! I'm in the process of making the application online so for now the repo got a bit messy, tomorrow it will be clean and ready to be spinned up also locally: sorry for the incovenient!

posted
an
update
7 days ago
Post
2933
Finding a job that matches with our resume shouldn't be difficult, especially now that we have AI... And still, we're drowning in unclear announcements, jobs whose skill requirements might not really fit us, and tons of material๐ตโ๐ซ
That's why I decided to build ๐๐๐ฌ๐ฎ๐ฆ๐ ๐๐๐ญ๐๐ก๐๐ซ (https://github.com/AstraBert/resume-matcher), a fully open-source application that scans your resume and searches the web for jobs that match with it!๐
The workflow is very simple:
๐ฆ A LlamaExtract agent parses the resume and extracts valuable data that represent your profile
๐๏ธThe structured data are passed on to a Job Matching Agent (built with LlamaIndex๐) that uses them to build a web search query based on your resume
๐ The web search is handled by Linkup, which finds the top matches and returns them to the Agent
๐ The agent evaluates the match between your profile and the jobs, and then returns a final answer to you
So, are you ready to find a job suitable for you?๐ผ You can spin up the application completely locally and with Docker, starting from the GitHub repo โก๏ธ https://github.com/AstraBert/resume-matcher
Feel free to leave your feedback and let me know in the comments if you want an online version of Resume Matcher as well!โจ
That's why I decided to build ๐๐๐ฌ๐ฎ๐ฆ๐ ๐๐๐ญ๐๐ก๐๐ซ (https://github.com/AstraBert/resume-matcher), a fully open-source application that scans your resume and searches the web for jobs that match with it!๐
The workflow is very simple:
๐ฆ A LlamaExtract agent parses the resume and extracts valuable data that represent your profile
๐๏ธThe structured data are passed on to a Job Matching Agent (built with LlamaIndex๐) that uses them to build a web search query based on your resume
๐ The web search is handled by Linkup, which finds the top matches and returns them to the Agent
๐ The agent evaluates the match between your profile and the jobs, and then returns a final answer to you
So, are you ready to find a job suitable for you?๐ผ You can spin up the application completely locally and with Docker, starting from the GitHub repo โก๏ธ https://github.com/AstraBert/resume-matcher
Feel free to leave your feedback and let me know in the comments if you want an online version of Resume Matcher as well!โจ

replied to
their
post
17 days ago
I used good old Canva (pro :)

posted
an
update
21 days ago
Post
2924
Llama-4 is out and I couldn't resist but to cook something with it... So I came up with ๐๐ฅ๐๐ฆ๐๐๐๐ฌ๐๐๐ซ๐๐ก๐๐ซ (https://llamaresearcher.com), your deep-research AI companion!๐
The workflow behind ๐๐น๐ฎ๐บ๐ฎ๐ฅ๐ฒ๐๐ฒ๐ฎ๐ฟ๐ฐ๐ต๐ฒ๐ฟ is simple:
๐ฌ You submit a query
๐ก๏ธ Your query is evaluated by Llama 3 guard model, which deems it safe or unsafe
๐ง If your query is safe, it is routed to the Researcher Agent
โ๏ธ The Researcher Agent expands the query into three sub-queries, with which to search the web
๐ The web is searched for each of the sub-queries
๐ The retrieved information is evaluated for relevancy against your original query
โ๏ธ The Researcher Agent produces an essay based on the information it gathered, paying attention to referencing its sources
The agent itself is also built with easy-to-use and intuitive blocks:
๐ฆ LlamaIndex provides the agentic architecture and the integrations with the language models
โกGroq makes Llama-4 available with its lightning-fast inference
๐ Linkup allows the agent to deep-search the web and provides sourced answers
๐ช FastAPI does the heavy loading with wrapping everything within an elegant API interface
โฑ๏ธ Redis is used for API rate limiting
๐จ Gradio creates a simple but powerful user interface
Special mention also to Lovable, which helped me build the first draft of the landing page for LlamaResearcher!๐
If you're curious and you want to try LlamaResearcher, you can - completely for free and without subscription - for 30 days from now โก๏ธ https://llamaresearcher.com
And if you're like me, and you like getting your hands in code and build stuff on your own machine, I have good news: this is all open-source, fully reproducible locally and Docker-ready๐
Just go to the GitHub repo: https://github.com/AstraBert/llama-4-researcher and don't forget to star it, if you find it useful!โญ
As always, have fun and feel free to leave your feedbackโจ
The workflow behind ๐๐น๐ฎ๐บ๐ฎ๐ฅ๐ฒ๐๐ฒ๐ฎ๐ฟ๐ฐ๐ต๐ฒ๐ฟ is simple:
๐ฌ You submit a query
๐ก๏ธ Your query is evaluated by Llama 3 guard model, which deems it safe or unsafe
๐ง If your query is safe, it is routed to the Researcher Agent
โ๏ธ The Researcher Agent expands the query into three sub-queries, with which to search the web
๐ The web is searched for each of the sub-queries
๐ The retrieved information is evaluated for relevancy against your original query
โ๏ธ The Researcher Agent produces an essay based on the information it gathered, paying attention to referencing its sources
The agent itself is also built with easy-to-use and intuitive blocks:
๐ฆ LlamaIndex provides the agentic architecture and the integrations with the language models
โกGroq makes Llama-4 available with its lightning-fast inference
๐ Linkup allows the agent to deep-search the web and provides sourced answers
๐ช FastAPI does the heavy loading with wrapping everything within an elegant API interface
โฑ๏ธ Redis is used for API rate limiting
๐จ Gradio creates a simple but powerful user interface
Special mention also to Lovable, which helped me build the first draft of the landing page for LlamaResearcher!๐
If you're curious and you want to try LlamaResearcher, you can - completely for free and without subscription - for 30 days from now โก๏ธ https://llamaresearcher.com
And if you're like me, and you like getting your hands in code and build stuff on your own machine, I have good news: this is all open-source, fully reproducible locally and Docker-ready๐
Just go to the GitHub repo: https://github.com/AstraBert/llama-4-researcher and don't forget to star it, if you find it useful!โญ
As always, have fun and feel free to leave your feedbackโจ

posted
an
update
25 days ago
Post
736
I heard someone saying ๐๐ผ๐ถ๐ฐ๐ฒ assistants are the future, and someone else that ๐ ๐๐ฃ will rule the AI world... So I decided to combine both!๐
Meet ๐๐ฒ๐๐๐ (๐ง๐pe๐ฆcript ๐ฉoice ๐ssistant, https://github.com/AstraBert/TySVA), your (speaking) AI companion for everyday TypeScript programming tasks!๐๏ธ
TySVA is a skilled TypeScript expert and, to provide accurate and up-to-date responses, she leverages the following workflow:
๐ฃ๏ธ If you talk to her, she converts the audio into a textual prompt, and use it a starting point to answer your questions (if you send a message, she'll use directly that๐ฌ)
๐ง She can solve your questions by (deep)searching the web and/or by retrieving relevant information from a vector database containing TypeScript documentation. If the answer is simple, she can also reply directly (no tools needed!)
๐ To ease her life, TySVA has all the tools she needs available through Model Context Protocol (MCP)
๐ Once she's done, she returns her answer to you, along with a voice summary of what she did and what solution she found
But how does she do that? What are her components?๐คจ
๐ Qdrant + HuggingFace give her the documentation knowledge, providing the vector database and the embeddings
๐ Linkup provides her with up-to-date, grounded answers, connecting her to the web
๐ฆ LlamaIndex makes up her brain, with the whole agentic architecture
๐ค ElevenLabs gives her ears and mouth, transcribing and producing voice inputs and outoputs
๐ Groq provides her with speech, being the LLM provider behind TySVA
๐จ Gradio+FastAPI make up her face and fibers, providing a seamless backend-to-frontend integration
If you're now curious of trying her, you can easily do that by spinning her up locally (and with Docker!๐) from the GitHub repo โก๏ธ https://github.com/AstraBert/TySVA
And feel free to leave any feedback!โจ
Meet ๐๐ฒ๐๐๐ (๐ง๐pe๐ฆcript ๐ฉoice ๐ssistant, https://github.com/AstraBert/TySVA), your (speaking) AI companion for everyday TypeScript programming tasks!๐๏ธ
TySVA is a skilled TypeScript expert and, to provide accurate and up-to-date responses, she leverages the following workflow:
๐ฃ๏ธ If you talk to her, she converts the audio into a textual prompt, and use it a starting point to answer your questions (if you send a message, she'll use directly that๐ฌ)
๐ง She can solve your questions by (deep)searching the web and/or by retrieving relevant information from a vector database containing TypeScript documentation. If the answer is simple, she can also reply directly (no tools needed!)
๐ To ease her life, TySVA has all the tools she needs available through Model Context Protocol (MCP)
๐ Once she's done, she returns her answer to you, along with a voice summary of what she did and what solution she found
But how does she do that? What are her components?๐คจ
๐ Qdrant + HuggingFace give her the documentation knowledge, providing the vector database and the embeddings
๐ Linkup provides her with up-to-date, grounded answers, connecting her to the web
๐ฆ LlamaIndex makes up her brain, with the whole agentic architecture
๐ค ElevenLabs gives her ears and mouth, transcribing and producing voice inputs and outoputs
๐ Groq provides her with speech, being the LLM provider behind TySVA
๐จ Gradio+FastAPI make up her face and fibers, providing a seamless backend-to-frontend integration
If you're now curious of trying her, you can easily do that by spinning her up locally (and with Docker!๐) from the GitHub repo โก๏ธ https://github.com/AstraBert/TySVA
And feel free to leave any feedback!โจ

posted
an
update
about 1 month ago
Post
629
Drowning in handouts, documents and presentations from your professors and not knowing where to start?๐๐ตโ๐ซ
Well, I might have a tool for you: ๐๐๐๐๐๐จ๐ญ๐๐ฌ (https://github.com/AstraBert/pdf2notes) is an ๐๐-๐ฝ๐ผ๐๐ฒ๐ฟ๐ฒ๐ฑ, ๐ผ๐ฝ๐ฒ๐ป-๐๐ผ๐๐ฟ๐ฐ๐ฒ solution that lets you turn your unstructured and chaotic PDFs into nice and well-ordered notes in a matter of seconds!๐
๐๐ผ๐ ๐ฑ๐ผ๐ฒ๐ ๐ถ๐ ๐๐ผ๐ฟ๐ธ?
๐ You first upload a document
โ๏ธ LlamaParse by LlamaIndex extracts the text from the document, using DeepMind's Gemini 2 Flash to perform multi-modal parsing
๐ง Llama-3.3-70B by Groq turns the extracted text into notes!
The notes are not perfect or you want more in-depth insights? No problem:
๐ฌ Send a direct message to the chatbot
โ๏ธ The chatbot will retrieve the chat history from a Postgres database
๐ง Llama-3.3-70B will produce the answer you need
All of this is nicely wrapped within a seamless backend-to-frontend framework powered by Gradio and FastAPI๐จ
And you can even spin it up easily and locally, using Docker๐
So, what are you waiting for? Go turn your hundreds of pages of chaotic learning material into neat and elegant notes โก๏ธ https://github.com/AstraBert/pdf2notes
And, if you would like an online demo, feel free to drop a comment - we'll see what we can build๐
Well, I might have a tool for you: ๐๐๐๐๐๐จ๐ญ๐๐ฌ (https://github.com/AstraBert/pdf2notes) is an ๐๐-๐ฝ๐ผ๐๐ฒ๐ฟ๐ฒ๐ฑ, ๐ผ๐ฝ๐ฒ๐ป-๐๐ผ๐๐ฟ๐ฐ๐ฒ solution that lets you turn your unstructured and chaotic PDFs into nice and well-ordered notes in a matter of seconds!๐
๐๐ผ๐ ๐ฑ๐ผ๐ฒ๐ ๐ถ๐ ๐๐ผ๐ฟ๐ธ?
๐ You first upload a document
โ๏ธ LlamaParse by LlamaIndex extracts the text from the document, using DeepMind's Gemini 2 Flash to perform multi-modal parsing
๐ง Llama-3.3-70B by Groq turns the extracted text into notes!
The notes are not perfect or you want more in-depth insights? No problem:
๐ฌ Send a direct message to the chatbot
โ๏ธ The chatbot will retrieve the chat history from a Postgres database
๐ง Llama-3.3-70B will produce the answer you need
All of this is nicely wrapped within a seamless backend-to-frontend framework powered by Gradio and FastAPI๐จ
And you can even spin it up easily and locally, using Docker๐
So, what are you waiting for? Go turn your hundreds of pages of chaotic learning material into neat and elegant notes โก๏ธ https://github.com/AstraBert/pdf2notes
And, if you would like an online demo, feel free to drop a comment - we'll see what we can build๐

posted
an
update
about 2 months ago
Post
1683
๐๐๐๐๐จ๐จ๐ง๐ฆ - ๐๐ ๐๐ง๐ญ๐ข๐ ๐๐๐ ๐ญ๐จ ๐ก๐๐ฅ๐ฉ ๐ฒ๐จ๐ฎ ๐๐ฎ๐ข๐ฅ๐ ๐ฒ๐จ๐ฎ๐ซ ๐ฌ๐ญ๐๐ซ๐ญ๐ฎ๐ฉ
GitHub ๐ https://github.com/AstraBert/ragcoon
Are you building a startup and you're stuck in the process, trying to navigate hundreds of resources, suggestions and LinkedIn posts?๐ถโ๐ซ๏ธ
Well, fear no more, because ๐ฅ๐๐๐ฐ๐ผ๐ผ๐ป๐ฆ is here to do some of the job for you:
๐ It's built on free resources written by successful founders
โ๏ธ It performs complex retrieval operations, exploiting "vanilla" hybrid search, query expansion with an ๐ต๐๐ฝ๐ผ๐๐ต๐ฒ๐๐ถ๐ฐ๐ฎ๐น ๐ฑ๐ผ๐ฐ๐๐บ๐ฒ๐ป๐ approach and ๐บ๐๐น๐๐ถ-๐๐๐ฒ๐ฝ ๐พ๐๐ฒ๐ฟ๐ ๐ฑ๐ฒ๐ฐ๐ผ๐บ๐ฝ๐ผ๐๐ถ๐๐ถ๐ผ๐ป
๐ It evaluates the ๐ฟ๐ฒ๐น๐ถ๐ฎ๐ฏ๐ถ๐น๐ถ๐๐ of the retrieved context, and the ๐ฟ๐ฒ๐น๐ฒ๐๐ฎ๐ป๐ฐ๐ and ๐ณ๐ฎ๐ถ๐๐ต๐ณ๐๐น๐ป๐ฒ๐๐ of its own responses, in an auto-correction effort
RAGcoon๐ฆ is ๐ผ๐ฝ๐ฒ๐ป-๐๐ผ๐๐ฟ๐ฐ๐ฒ and relies on easy-to-use components:
๐นLlamaIndex is at the core of the agent architecture, provisions the integrations with language models and vector database services, and performs evaluations
๐น Qdrant is your go-to, versatile and scalable companion for vector database services
๐นGroq provides lightning-fast LLM inference to support the agent, giving it the full power of ๐ค๐๐ค-๐ฏ๐ฎ๐ by Qwen
๐นHugging Face provides the embedding models used for dense and sparse retrieval
๐นFastAPI wraps the whole backend into an API interface
๐น๐ ๐ฒ๐๐ผ๐ฝ by Google is used to serve the application frontend
RAGcoon๐ฆ can be spinned up locally - it's ๐๐ผ๐ฐ๐ธ๐ฒ๐ฟ-๐ฟ๐ฒ๐ฎ๐ฑ๐๐, and you can find the whole code to reproduce it on GitHub ๐ https://github.com/AstraBert/ragcoon
But there might be room for an online version of RAGcoon๐ฆ: let me know if you would use it - we can connect and build it together!๐
GitHub ๐ https://github.com/AstraBert/ragcoon
Are you building a startup and you're stuck in the process, trying to navigate hundreds of resources, suggestions and LinkedIn posts?๐ถโ๐ซ๏ธ
Well, fear no more, because ๐ฅ๐๐๐ฐ๐ผ๐ผ๐ป๐ฆ is here to do some of the job for you:
๐ It's built on free resources written by successful founders
โ๏ธ It performs complex retrieval operations, exploiting "vanilla" hybrid search, query expansion with an ๐ต๐๐ฝ๐ผ๐๐ต๐ฒ๐๐ถ๐ฐ๐ฎ๐น ๐ฑ๐ผ๐ฐ๐๐บ๐ฒ๐ป๐ approach and ๐บ๐๐น๐๐ถ-๐๐๐ฒ๐ฝ ๐พ๐๐ฒ๐ฟ๐ ๐ฑ๐ฒ๐ฐ๐ผ๐บ๐ฝ๐ผ๐๐ถ๐๐ถ๐ผ๐ป
๐ It evaluates the ๐ฟ๐ฒ๐น๐ถ๐ฎ๐ฏ๐ถ๐น๐ถ๐๐ of the retrieved context, and the ๐ฟ๐ฒ๐น๐ฒ๐๐ฎ๐ป๐ฐ๐ and ๐ณ๐ฎ๐ถ๐๐ต๐ณ๐๐น๐ป๐ฒ๐๐ of its own responses, in an auto-correction effort
RAGcoon๐ฆ is ๐ผ๐ฝ๐ฒ๐ป-๐๐ผ๐๐ฟ๐ฐ๐ฒ and relies on easy-to-use components:
๐นLlamaIndex is at the core of the agent architecture, provisions the integrations with language models and vector database services, and performs evaluations
๐น Qdrant is your go-to, versatile and scalable companion for vector database services
๐นGroq provides lightning-fast LLM inference to support the agent, giving it the full power of ๐ค๐๐ค-๐ฏ๐ฎ๐ by Qwen
๐นHugging Face provides the embedding models used for dense and sparse retrieval
๐นFastAPI wraps the whole backend into an API interface
๐น๐ ๐ฒ๐๐ผ๐ฝ by Google is used to serve the application frontend
RAGcoon๐ฆ can be spinned up locally - it's ๐๐ผ๐ฐ๐ธ๐ฒ๐ฟ-๐ฟ๐ฒ๐ฎ๐ฑ๐๐, and you can find the whole code to reproduce it on GitHub ๐ https://github.com/AstraBert/ragcoon
But there might be room for an online version of RAGcoon๐ฆ: let me know if you would use it - we can connect and build it together!๐

posted
an
update
about 2 months ago
Post
2740
I just released a fully automated evaluation framework for your RAG applications!๐
GitHub ๐ https://github.com/AstraBert/diRAGnosis
PyPi ๐ https://pypi.org/project/diragnosis/
It's called ๐๐ข๐๐๐๐ง๐จ๐ฌ๐ข๐ฌ and is a lightweight framework that helps you ๐ฑ๐ถ๐ฎ๐ด๐ป๐ผ๐๐ฒ ๐๐ต๐ฒ ๐ฝ๐ฒ๐ฟ๐ณ๐ผ๐ฟ๐บ๐ฎ๐ป๐ฐ๐ฒ ๐ผ๐ณ ๐๐๐ ๐ ๐ฎ๐ป๐ฑ ๐ฟ๐ฒ๐๐ฟ๐ถ๐ฒ๐๐ฎ๐น ๐บ๐ผ๐ฑ๐ฒ๐น๐ ๐ถ๐ป ๐ฅ๐๐ ๐ฎ๐ฝ๐ฝ๐น๐ถ๐ฐ๐ฎ๐๐ถ๐ผ๐ป๐.
You can launch it as an application locally (it's Docker-ready!๐) or, if you want more flexibility, you can integrate it in your code as a python package๐ฆ
The workflow is simple:
๐ง You choose your favorite LLM provider and model (supported, for now, are Mistral AI, Groq, Anthropic, OpenAI and Cohere)
๐ง You pick the embedding models provider and the embedding model you prefer (supported, for now, are Mistral AI, Hugging Face, Cohere and OpenAI)
๐ You prepare and provide your documents
โ๏ธ Documents are ingested into a Qdrant vector database and transformed into a synthetic question dataset with the help of LlamaIndex
๐ The LLM is evaluated for the faithfulness and relevancy of its retrieval-augmented answer to the questions
๐ The embedding model is evaluated for hit rate and mean reciprocal ranking (MRR) of the retrieved documents
And the cool thing is that all of this is ๐ถ๐ป๐๐๐ถ๐๐ถ๐๐ฒ ๐ฎ๐ป๐ฑ ๐ฐ๐ผ๐บ๐ฝ๐น๐ฒ๐๐ฒ๐น๐ ๐ฎ๐๐๐ผ๐บ๐ฎ๐๐ฒ๐ฑ: you plug it in, and it works!๐โก
Even cooler? This is all built on top of LlamaIndex and its integrations: no need for tons of dependencies or fancy workarounds๐ฆ
And if you're a UI lover, Gradio and FastAPI are there to provide you a seamless backend-to-frontend experience๐ถ๏ธ
So now it's your turn: you can either get diRAGnosis from GitHub ๐ https://github.com/AstraBert/diRAGnosis
or just run a quick and painless:
To get the package installed (lightning-fast) in your environment๐โโ๏ธ
Have fun and feel free to leave feedback and feature/integrations requests on GitHub issuesโจ
GitHub ๐ https://github.com/AstraBert/diRAGnosis
PyPi ๐ https://pypi.org/project/diragnosis/
It's called ๐๐ข๐๐๐๐ง๐จ๐ฌ๐ข๐ฌ and is a lightweight framework that helps you ๐ฑ๐ถ๐ฎ๐ด๐ป๐ผ๐๐ฒ ๐๐ต๐ฒ ๐ฝ๐ฒ๐ฟ๐ณ๐ผ๐ฟ๐บ๐ฎ๐ป๐ฐ๐ฒ ๐ผ๐ณ ๐๐๐ ๐ ๐ฎ๐ป๐ฑ ๐ฟ๐ฒ๐๐ฟ๐ถ๐ฒ๐๐ฎ๐น ๐บ๐ผ๐ฑ๐ฒ๐น๐ ๐ถ๐ป ๐ฅ๐๐ ๐ฎ๐ฝ๐ฝ๐น๐ถ๐ฐ๐ฎ๐๐ถ๐ผ๐ป๐.
You can launch it as an application locally (it's Docker-ready!๐) or, if you want more flexibility, you can integrate it in your code as a python package๐ฆ
The workflow is simple:
๐ง You choose your favorite LLM provider and model (supported, for now, are Mistral AI, Groq, Anthropic, OpenAI and Cohere)
๐ง You pick the embedding models provider and the embedding model you prefer (supported, for now, are Mistral AI, Hugging Face, Cohere and OpenAI)
๐ You prepare and provide your documents
โ๏ธ Documents are ingested into a Qdrant vector database and transformed into a synthetic question dataset with the help of LlamaIndex
๐ The LLM is evaluated for the faithfulness and relevancy of its retrieval-augmented answer to the questions
๐ The embedding model is evaluated for hit rate and mean reciprocal ranking (MRR) of the retrieved documents
And the cool thing is that all of this is ๐ถ๐ป๐๐๐ถ๐๐ถ๐๐ฒ ๐ฎ๐ป๐ฑ ๐ฐ๐ผ๐บ๐ฝ๐น๐ฒ๐๐ฒ๐น๐ ๐ฎ๐๐๐ผ๐บ๐ฎ๐๐ฒ๐ฑ: you plug it in, and it works!๐โก
Even cooler? This is all built on top of LlamaIndex and its integrations: no need for tons of dependencies or fancy workarounds๐ฆ
And if you're a UI lover, Gradio and FastAPI are there to provide you a seamless backend-to-frontend experience๐ถ๏ธ
So now it's your turn: you can either get diRAGnosis from GitHub ๐ https://github.com/AstraBert/diRAGnosis
or just run a quick and painless:
uv pip install diragnosis
To get the package installed (lightning-fast) in your environment๐โโ๏ธ
Have fun and feel free to leave feedback and feature/integrations requests on GitHub issuesโจ

posted
an
update
2 months ago
Post
2403
I built an AI agent app in less than 8 hours๐คฏ
And, believe me, this is ๐ป๐ผ๐ clickbaitโ
GitHub ๐ https://github.com/AstraBert/PapersChat
Demo ๐ as-cle-bert/PapersChat
The app is called ๐๐๐ฉ๐๐ซ๐ฌ๐๐ก๐๐ญ, and it is aimed at ๐บ๐ฎ๐ธ๐ถ๐ป๐ด ๐ฐ๐ต๐ฎ๐๐๐ถ๐ป๐ด ๐๐ถ๐๐ต ๐๐ฐ๐ถ๐ฒ๐ป๐๐ถ๐ณ๐ถ๐ฐ ๐ฝ๐ฎ๐ฝ๐ฒ๐ฟ๐ ๐ฒ๐ฎ๐๐ถ๐ฒ๐ฟ.
๐๐๐ซ๐ ๐ข๐ฌ ๐ฐ๐ก๐๐ญ ๐ญ๐ก๐ ๐๐ฉ๐ฉ ๐๐จ๐๐ฌ:
๐ Parses the papers that you upload thanks to LlamaIndex๐ฆ (either with LlamaParse or with simpler, local methods)
๐ Embeds documents both with a sparse and with a dense encoder to enable hybrid search
๐ Uploads the embeddings to Qdrant
โ๏ธ Activates an Agent based on mistralai/Mistral-Small-24B-Instruct-2501 that will reply to your prompt
๐ง Retrieves information relevant to your question from the documents
๐ง If no relevant information is found, it searches PubMed and arXiv databases
๐ง Returns a grounded answer to your prompt
๐๐จ๐ฐ ๐๐ข๐ ๐ ๐ฆ๐๐ง๐๐ ๐ ๐ญ๐จ ๐ฆ๐๐ค๐ ๐ญ๐ก๐ข๐ฌ ๐๐ฉ๐ฉ๐ฅ๐ข๐๐๐ญ๐ข๐จ๐ง ๐ข๐ง ๐ ๐ก๐จ๐ฎ๐ซ๐ฌ?
Three key points:
- LlamaIndex๐ฆ provides countless integrations with LLM providers, text embedding models and vectorstore services, and takes care of the internal architecture of the Agent. You just plug it in, and it works!๐โก
- Qdrant is a vector database service extremely easy to set up and use: you just need a one-line Docker command๐
- Gradio makes frontend development painless and fast, while still providing modern and responsive interfaces๐๏ธ
And a bonus point:
- Deploying the demo app couldn't be easier if you use Gradio-based Hugging Face Spaces๐ค
So, no more excuses: build your own AI agent today and do it fast, (almost) for free and effortlessly๐
And if you need a starting point, the code for PapersChat is open and fully reproducible on GitHub ๐ https://github.com/AstraBert/PapersChat
And, believe me, this is ๐ป๐ผ๐ clickbaitโ
GitHub ๐ https://github.com/AstraBert/PapersChat
Demo ๐ as-cle-bert/PapersChat
The app is called ๐๐๐ฉ๐๐ซ๐ฌ๐๐ก๐๐ญ, and it is aimed at ๐บ๐ฎ๐ธ๐ถ๐ป๐ด ๐ฐ๐ต๐ฎ๐๐๐ถ๐ป๐ด ๐๐ถ๐๐ต ๐๐ฐ๐ถ๐ฒ๐ป๐๐ถ๐ณ๐ถ๐ฐ ๐ฝ๐ฎ๐ฝ๐ฒ๐ฟ๐ ๐ฒ๐ฎ๐๐ถ๐ฒ๐ฟ.
๐๐๐ซ๐ ๐ข๐ฌ ๐ฐ๐ก๐๐ญ ๐ญ๐ก๐ ๐๐ฉ๐ฉ ๐๐จ๐๐ฌ:
๐ Parses the papers that you upload thanks to LlamaIndex๐ฆ (either with LlamaParse or with simpler, local methods)
๐ Embeds documents both with a sparse and with a dense encoder to enable hybrid search
๐ Uploads the embeddings to Qdrant
โ๏ธ Activates an Agent based on mistralai/Mistral-Small-24B-Instruct-2501 that will reply to your prompt
๐ง Retrieves information relevant to your question from the documents
๐ง If no relevant information is found, it searches PubMed and arXiv databases
๐ง Returns a grounded answer to your prompt
๐๐จ๐ฐ ๐๐ข๐ ๐ ๐ฆ๐๐ง๐๐ ๐ ๐ญ๐จ ๐ฆ๐๐ค๐ ๐ญ๐ก๐ข๐ฌ ๐๐ฉ๐ฉ๐ฅ๐ข๐๐๐ญ๐ข๐จ๐ง ๐ข๐ง ๐ ๐ก๐จ๐ฎ๐ซ๐ฌ?
Three key points:
- LlamaIndex๐ฆ provides countless integrations with LLM providers, text embedding models and vectorstore services, and takes care of the internal architecture of the Agent. You just plug it in, and it works!๐โก
- Qdrant is a vector database service extremely easy to set up and use: you just need a one-line Docker command๐
- Gradio makes frontend development painless and fast, while still providing modern and responsive interfaces๐๏ธ
And a bonus point:
- Deploying the demo app couldn't be easier if you use Gradio-based Hugging Face Spaces๐ค
So, no more excuses: build your own AI agent today and do it fast, (almost) for free and effortlessly๐
And if you need a starting point, the code for PapersChat is open and fully reproducible on GitHub ๐ https://github.com/AstraBert/PapersChat

posted
an
update
2 months ago
Post
1403
๐๐๐ข๐๐๐ฐ๐ฌ๐๐จ๐ญ - ๐๐๐ฉ๐จ๐ซ๐ญ ๐๐๐ข๐ฅ๐ฒ ๐๐๐ข๐๐ง๐๐ ๐ง๐๐ฐ๐ฌ ๐จ๐ง ๐๐ฅ๐ฎ๐๐๐ค๐ฒ
GitHub ๐ https://github.com/AstraBert/SciNewsBot
BlueSky ๐ https://bsky.app/profile/sci-news-bot.bsky.social
Hi there HF Community!๐ค
I just created a very simple AI-powered bot that shares fact-checked news about Science, Environment, Energy and Technology on BlueSky :)
The bot takes news from Google News, filters out the sources that are not represented in the Media Bias Fact Check database, and then evaluates the reliability of the source based on the MBFC metrics. After that, it creates a catchy headline for the article and publishes the post on BlueSky๐ฐ
The cool thing? SciNewsBot is open-source and is cheap to maintain, as it is based on mistralai/Mistral-Small-24B-Instruct-2501 (via Mistral API). You can reproduce it locally, spinning it up on your machine, and even launch it on cloud through a comfy Docker setup๐
Have fun and spread Science!โจ
GitHub ๐ https://github.com/AstraBert/SciNewsBot
BlueSky ๐ https://bsky.app/profile/sci-news-bot.bsky.social
Hi there HF Community!๐ค
I just created a very simple AI-powered bot that shares fact-checked news about Science, Environment, Energy and Technology on BlueSky :)
The bot takes news from Google News, filters out the sources that are not represented in the Media Bias Fact Check database, and then evaluates the reliability of the source based on the MBFC metrics. After that, it creates a catchy headline for the article and publishes the post on BlueSky๐ฐ
The cool thing? SciNewsBot is open-source and is cheap to maintain, as it is based on mistralai/Mistral-Small-24B-Instruct-2501 (via Mistral API). You can reproduce it locally, spinning it up on your machine, and even launch it on cloud through a comfy Docker setup๐
Have fun and spread Science!โจ

posted
an
update
2 months ago
Post
2766
๐๐ก๐ข๐๐ฐ๐๐ง๐๐๐๐ - ๐ ๐ซ๐๐๐ฌ๐จ๐ง๐ข๐ง๐ ๐๐ฌ๐ฌ๐ข๐ฌ๐ญ๐๐ง๐ญ ๐๐จ๐ซ ๐ฒ๐จ๐ฎ๐ซ ๐๐๐๐ ๐๐๐ฎ๐๐๐ญ๐ข๐จ๐ง
Demo ๐ https://pqstem.org
GitHub ๐ https://github.com/AstraBert/PhiQwenSTEM
Hello HF community!๐ค
Ever struggled with some complex Maths problem or with a very hard Physics question? Well, fear no more, because now you can rely on PhiQwenSTEM, an assistant specialized in answering STEM-related question!
The assistant can count on a knowledge base of ๐ญ๐ฑ๐ธ+ ๐๐ฒ๐น๐ฒ๐ฐ๐๐ฒ๐ฑ ๐ฆ๐ง๐๐ ๐พ๐๐ฒ๐๐๐ถ๐ผ๐ป-๐ฎ๐ป๐๐๐ฒ๐ฟ ๐ฝ๐ฎ๐ถ๐ฟ๐ spanning the domains of Chemistry, Physics, Matemathics and Biochemistry (from EricLu/SCP-116K). It also relies on the combined power of microsoft/Phi-3.5-mini-instruct and Qwen/QwQ-32B-Preview to produce reliable and reasoned answers.
For the next 30 days, you will be able to try for free the web demo: https://pqstem.org
In the GitHub repo you can find all the information to reproduce PhiQwenSTEM ๐ผ๐ป ๐๐ผ๐๐ฟ ๐น๐ผ๐ฐ๐ฎ๐น ๐บ๐ฎ๐ฐ๐ต๐ถ๐ป๐ฒ, ๐ฏ๐ผ๐๐ต ๐๐ถ๐ฎ ๐๐ผ๐๐ฟ๐ฐ๐ฒ ๐ฐ๐ผ๐ฑ๐ฒ ๐ฎ๐ป๐ฑ ๐๐ถ๐๐ต ๐ฎ ๐ฐ๐ผ๐บ๐ณ๐ ๐๐ผ๐ฐ๐ธ๐ฒ๐ฟ๐ ๐๐ฒ๐๐๐ฝ: https://github.com/AstraBert/PhiQwenSTEM
Demo ๐ https://pqstem.org
GitHub ๐ https://github.com/AstraBert/PhiQwenSTEM
Hello HF community!๐ค
Ever struggled with some complex Maths problem or with a very hard Physics question? Well, fear no more, because now you can rely on PhiQwenSTEM, an assistant specialized in answering STEM-related question!
The assistant can count on a knowledge base of ๐ญ๐ฑ๐ธ+ ๐๐ฒ๐น๐ฒ๐ฐ๐๐ฒ๐ฑ ๐ฆ๐ง๐๐ ๐พ๐๐ฒ๐๐๐ถ๐ผ๐ป-๐ฎ๐ป๐๐๐ฒ๐ฟ ๐ฝ๐ฎ๐ถ๐ฟ๐ spanning the domains of Chemistry, Physics, Matemathics and Biochemistry (from EricLu/SCP-116K). It also relies on the combined power of microsoft/Phi-3.5-mini-instruct and Qwen/QwQ-32B-Preview to produce reliable and reasoned answers.
For the next 30 days, you will be able to try for free the web demo: https://pqstem.org
In the GitHub repo you can find all the information to reproduce PhiQwenSTEM ๐ผ๐ป ๐๐ผ๐๐ฟ ๐น๐ผ๐ฐ๐ฎ๐น ๐บ๐ฎ๐ฐ๐ต๐ถ๐ป๐ฒ, ๐ฏ๐ผ๐๐ต ๐๐ถ๐ฎ ๐๐ผ๐๐ฟ๐ฐ๐ฒ ๐ฐ๐ผ๐ฑ๐ฒ ๐ฎ๐ป๐ฑ ๐๐ถ๐๐ต ๐ฎ ๐ฐ๐ผ๐บ๐ณ๐ ๐๐ผ๐ฐ๐ธ๐ฒ๐ฟ๐ ๐๐ฒ๐๐๐ฝ: https://github.com/AstraBert/PhiQwenSTEM

posted
an
update
3 months ago
Post
1040
Hi HuggingFace community!๐ค
I just published an article in which I try to articulate some counter-points to Dario Amodei's post "On DeepSeek and Export Control"๐ https://huggingface.co/blog/as-cle-bert/why-we-dont-need-export-control
I try to address several key passages of the third section from Amodei's post (https://darioamodei.com/on-deepseek-and-export-controls), bringing my perspective on the importance of open source, open knowledge and multipolarity in a crucial field for our future such as Artificial Intelligence.
Happy reading!โจ
I just published an article in which I try to articulate some counter-points to Dario Amodei's post "On DeepSeek and Export Control"๐ https://huggingface.co/blog/as-cle-bert/why-we-dont-need-export-control
I try to address several key passages of the third section from Amodei's post (https://darioamodei.com/on-deepseek-and-export-controls), bringing my perspective on the importance of open source, open knowledge and multipolarity in a crucial field for our future such as Artificial Intelligence.
Happy reading!โจ

replied to
their
post
3 months ago
Hi!
I generally use LangChain + PyPDF, I leave here a code snippet:
from langchain_community.document_loaders import PyPDFLoader
def preprocess(pdf: str) -> list:
"""
Uses LangChain's PyPDFLoader to extract text.
"""
loader = PyPDFLoader(pdf)
documents = loader.load()
for doc in documents:
print(doc.page_content)
This should give a more solid result :)
PS: Langchain is distributed under an MIT license, see their GitHub (https://github.com/langchain-ai/langchain)

posted
an
update
3 months ago
Post
1613
๐๐๐๐ฐ ๐๐๐ฆ๐จ ๐๐ฅ๐๐ซ๐ญ๐
Convert (almost) everything to PDF with ๐๐๐๐๐ญ๐๐จ๐ฐ๐ง, now on Spaces! ๐ as-cle-bert/pdfitdown
You can also install it locally:
Don't forget to star it on GitHub, if you find it useful! ๐ https://www.github.com/AstraBert/PdfItDown
Convert (almost) everything to PDF with ๐๐๐๐๐ญ๐๐จ๐ฐ๐ง, now on Spaces! ๐ as-cle-bert/pdfitdown
You can also install it locally:
python3 -m pip install pdfitdown
Don't forget to star it on GitHub, if you find it useful! ๐ https://www.github.com/AstraBert/PdfItDown

posted
an
update
3 months ago
Post
569
Hi HuggingFace Community๐ค, I am thrilled to announce:
๐ช๐๐ฎ๐ซ๐ฅ๐ฅ๐ฆ ๐๐ท-๐๐.๐ท (https://github.com/AstraBert/qdurllm/tree/january-2025)
Qdurllm (๐ค๐ฑrant, ๐จ๐ฅLs, ๐arge ๐anguage ๐ odels) is a local Gradio (Gradio) application that lets you upload you web content to a local Qdrant (Qdrant) database and search through it or chat with it.
The ๐ป๐ฒ๐ ๐ฝ๐ฟ๐ฒ-๐ฟ๐ฒ๐น๐ฒ๐ฎ๐๐ฒ (https://github.com/AstraBert/qdurllm/releases/tag/v1.0.0-rc.0) implements ๐๐ฝ๐ฎ๐ฟ๐๐ฒ ๐๐ฒ๐ฎ๐ฟ๐ฐ๐ต (with prithivida/Splade_PP_en_v1) + ๐ฟ๐ฒ๐ฟ๐ฎ๐ป๐ธ๐ถ๐ป๐ด (with nomic-ai/modernbert-embed-base by Hugging Face + Nomic AI) and ๐๐ฒ๐บ๐ฎ๐ป๐๐ถ๐ฐ ๐ฐ๐ฎ๐ฐ๐ต๐ถ๐ป๐ด (based on Qdrant) and switched ๐ณ๐ฟ๐ผ๐บ google/gemma-2-2b-it ๐๐ผ Qwen/Qwen2.5-1.5B-Instruct to conform to the SOTA landscape and to finally make the application based ๐ผ๐ป๐น๐ ๐ผ๐ป ๐๐ฟ๐๐น๐ ๐ผ๐ฝ๐ฒ๐ป ๐บ๐ผ๐ฑ๐ฒ๐น๐.
The pre-release is ๐ฎ๐๐ฎ๐ถ๐น๐ฎ๐ฏ๐น๐ฒ ๐ณ๐ผ๐ฟ ๐๐ฒ๐๐๐ถ๐ป๐ด and I would be really really happy if you wanted to give it a try and leave your feedback on the discussion thread on GitHub (https://github.com/AstraBert/qdurllm/discussions/8) or here on Hugging Face forum via comments under this postโจ.
Find all the information to install and launch it here ๐ https://astrabert.github.io/qdurllm/#2-installation
๐ช๐๐ฎ๐ซ๐ฅ๐ฅ๐ฆ ๐๐ท-๐๐.๐ท (https://github.com/AstraBert/qdurllm/tree/january-2025)
Qdurllm (๐ค๐ฑrant, ๐จ๐ฅLs, ๐arge ๐anguage ๐ odels) is a local Gradio (Gradio) application that lets you upload you web content to a local Qdrant (Qdrant) database and search through it or chat with it.
The ๐ป๐ฒ๐ ๐ฝ๐ฟ๐ฒ-๐ฟ๐ฒ๐น๐ฒ๐ฎ๐๐ฒ (https://github.com/AstraBert/qdurllm/releases/tag/v1.0.0-rc.0) implements ๐๐ฝ๐ฎ๐ฟ๐๐ฒ ๐๐ฒ๐ฎ๐ฟ๐ฐ๐ต (with prithivida/Splade_PP_en_v1) + ๐ฟ๐ฒ๐ฟ๐ฎ๐ป๐ธ๐ถ๐ป๐ด (with nomic-ai/modernbert-embed-base by Hugging Face + Nomic AI) and ๐๐ฒ๐บ๐ฎ๐ป๐๐ถ๐ฐ ๐ฐ๐ฎ๐ฐ๐ต๐ถ๐ป๐ด (based on Qdrant) and switched ๐ณ๐ฟ๐ผ๐บ google/gemma-2-2b-it ๐๐ผ Qwen/Qwen2.5-1.5B-Instruct to conform to the SOTA landscape and to finally make the application based ๐ผ๐ป๐น๐ ๐ผ๐ป ๐๐ฟ๐๐น๐ ๐ผ๐ฝ๐ฒ๐ป ๐บ๐ผ๐ฑ๐ฒ๐น๐.
The pre-release is ๐ฎ๐๐ฎ๐ถ๐น๐ฎ๐ฏ๐น๐ฒ ๐ณ๐ผ๐ฟ ๐๐ฒ๐๐๐ถ๐ป๐ด and I would be really really happy if you wanted to give it a try and leave your feedback on the discussion thread on GitHub (https://github.com/AstraBert/qdurllm/discussions/8) or here on Hugging Face forum via comments under this postโจ.
Find all the information to install and launch it here ๐ https://astrabert.github.io/qdurllm/#2-installation

replied to
their
post
4 months ago
Thank you so much for letting me know! This is indeed a very interesting role :)

posted
an
update
4 months ago
Post
1384
Hi HuggingFace community!๐ค
I recently released PrAIvateSearch v2.0-beta.0 (https://github.com/AstraBert/PrAIvateSearch), my privacy-first, AI-powered, user-centered and data-safe application aimed at providing a local and open-source alternative to big AI search engines such as SearchGPT or Perplexity AI.
We have several key changes:
- New chat UI built with NextJS
- DuckDuckGo API used for web search instead of Google
- Qwen/Qwen2.5-1.5B-Instruct as a language model served on API (by FastAPI)
- Crawl4AI crawler used for web scraping
- Optimizations in the data workflow inside the application
Read more in my blog post ๐ https://huggingface.co/blog/as-cle-bert/search-the-web-with-ai
Have fun and feel free to leave feedback about how to improve the application!โจ
I recently released PrAIvateSearch v2.0-beta.0 (https://github.com/AstraBert/PrAIvateSearch), my privacy-first, AI-powered, user-centered and data-safe application aimed at providing a local and open-source alternative to big AI search engines such as SearchGPT or Perplexity AI.
We have several key changes:
- New chat UI built with NextJS
- DuckDuckGo API used for web search instead of Google
- Qwen/Qwen2.5-1.5B-Instruct as a language model served on API (by FastAPI)
- Crawl4AI crawler used for web scraping
- Optimizations in the data workflow inside the application
Read more in my blog post ๐ https://huggingface.co/blog/as-cle-bert/search-the-web-with-ai
Have fun and feel free to leave feedback about how to improve the application!โจ