Skip to content

Latest commit

 

History

History
26 lines (19 loc) · 1.85 KB

README.md

File metadata and controls

26 lines (19 loc) · 1.85 KB

README

This is a repository accompanying the blogpost

It ingests a set of meeting notes called meetingbank creates summaries for each of the entries via a local Ollama instance. The Rails application then displays the meeting notes, along with the pre-generated summaries, and a chat window where you can queryan LLM about the specific meeting selected. The LLM uses the currently selected meeting as well as the chat history as context for its replies.

This application also highlights the importance of cleaning your sources, as shown below where the LLM tries to confidently assert that there's some tension tension in the meeting when in fact there was a mistranscription in the notes.

docschat screenshot

Setup

  1. Install Ollama https://ollama.com/download/linux
  2. ollama run gemma2:2b to download the model
  3. docker compose up to run postgres with pgvector
  4. ruby seeds.rb to seed the database (optional if not using the pre-generated seeds)
  5. bin/setup to setup dependencies

Resources, Inspirations, and References

  1. https://github.com/patterns-ai-core/langchainrb/blob/main/lib/langchain/llm/ollama.rb
  2. https://github.com/rails/rails/tree/main/guides/bug_report_templates
  3. https://greg.molnar.io/blog/a-single-file-rails-application/
  4. https://github.com/hopsoft/sr_mini
  5. https://thoughtbot.com/blog/talking-to-actioncable-without-rails
  6. https://www.mintbit.com/blog/subscribing-sending-and-receiving-actioncable-messages-with-js
  7. https://github.com/sonyarianto/react-without-buildsteps