I’ve been exploring MariaDB 11.8’s new vector search capabilities for building AI-driven applications, particularly with local LLMs for retrieval-augmented generation (RAG) of fully private data that never leaves the computer. I’m curious about how others in the community are leveraging these features in their projects.

I’m especially interested in using it with local LLMs (like Llama or Mistral) to keep data on-premise and avoid cloud-based API costs or security concerns.

Does anyone have experiences to share, in particular what LLMs are you using when generating embeddings to store in MariaDB?

2 points

I used all-minilm which worked well for documents. But it doesn’t work for images (or didn’t when I was using it last time). Although I did in combination with N8N and Qdrant.

permalink
report
reply
1 point

MixedBread is nice.

permalink
report
parent
reply
1 point

Have a look at Ollams embeddings. Easy to set up and the models are much smaller than a typical LLM.

permalink
report
reply

Open Source

!opensource@lemmy.ml

Create post

All about open source! Feel free to ask questions, and share news, and interesting stuff!

Useful Links

Rules

  • Posts must be relevant to the open source ideology
  • No NSFW content
  • No hate speech, bigotry, etc

Related Communities

Community icon from opensource.org, but we are not affiliated with them.

Community stats

  • 3.8K

    Monthly active users

  • 1.4K

    Posts

  • 14K

    Comments