GraphRAG with Ollama - Install Local Models for RAG - Easiest Tutorial

แชร์
ฝัง
  • เผยแพร่เมื่อ 6 ก.ค. 2024
  • This video is a step-by-step tutorial to install Microsoft GraphRAG with Ollama models with your own data.
    🔥 Buy Me a Coffee to support the channel: ko-fi.com/fahdmirza
    🔥 Get 50% Discount on any A6000 or A5000 GPU rental, use following link and coupon:
    bit.ly/fahd-mirza
    Coupon code: FahdMirza
    ▶ Become a Patron 🔥 - / fahdmirza
    #graphrag #ollama
    PLEASE FOLLOW ME:
    ▶ LinkedIn: / fahdmirza
    ▶ TH-cam: / @fahdmirza
    ▶ Blog: www.fahdmirza.com
    RELATED VIDEOS:
    ▶ Resource github.com/microsoft/graphrag
    ▶ Code www.fahdmirza.com/2024/07/ins...
    All rights reserved © Fahd Mirza
  • วิทยาศาสตร์และเทคโนโลยี

ความคิดเห็น • 73

  • @fahdmirza
    @fahdmirza  16 วันที่ผ่านมา +2

    Watch More GraphRAG Videos:
    🔥GraphRAG with Ollama - Install Local Models for RAG - Easiest Tutorial th-cam.com/video/6Yu6JpLMWVo/w-d-xo.htmlsi=ONzq5rT1OSd0l4mD
    🔥Install GraphRAG Locally - Build RAG Pipeline with Local and Global Search th-cam.com/video/Sy5K6Ay46xU/w-d-xo.htmlsi=g5eKWBsWg6zPaN7a
    🔥GraphRAG with Groq - Install Locally with Local and Global Search th-cam.com/video/xkDGpR5g9D0/w-d-xo.htmlsi=QVfnD5tUSnxvPhAH
    🔥GraphRAG with Llama.cpp Locally with Groq th-cam.com/video/9Gp2Qo1NASY/w-d-xo.html

    • @jianjieyin
      @jianjieyin 8 วันที่ผ่านมา

      GraphRAG with Ollama, entity_extraction directory is not empty but errors come... Columns must be same length as key . How to solve?

  • @fahdmirza
    @fahdmirza  20 วันที่ผ่านมา

    🔥Install GraphRAG Locally - Build RAG Pipeline with Local and Global Search th-cam.com/video/Sy5K6Ay46xU/w-d-xo.htmlsi=f-o9SyqE62OgNU14

  • @georgeknerr
    @georgeknerr 5 วันที่ผ่านมา

    Excellent work - got a working example going!

  • @Ayush-tl3ny
    @Ayush-tl3ny 20 วันที่ผ่านมา

    Thank you so much for this video! You are Awesome ❤

    • @fahdmirza
      @fahdmirza  19 วันที่ผ่านมา

      thank you

  • @davidtindell950
    @davidtindell950 19 วันที่ผ่านมา +1

    Thank You from a NEW Subscriber !

    • @fahdmirza
      @fahdmirza  18 วันที่ผ่านมา

      Awesome, thank you!

  • @tollington9414
    @tollington9414 20 วันที่ผ่านมา

    Good stuff!

    • @fahdmirza
      @fahdmirza  19 วันที่ผ่านมา

      thanks

  • @sergeziehi4816
    @sergeziehi4816 20 วันที่ผ่านมา +1

    Graph RAG cost a lot indeed on API calls.
    One of your best video I do believe.thanks a lot

    • @fahdmirza
      @fahdmirza  19 วันที่ผ่านมา +1

      thank you

  • @Salionca
    @Salionca 20 วันที่ผ่านมา

    Thank you for the video.

    • @fahdmirza
      @fahdmirza  19 วันที่ผ่านมา

      You're welcome

  • @TheStuzenz
    @TheStuzenz 17 วันที่ผ่านมา

    Nice video Fahd - GraphRAG looks really good! I plan on trying it out tonight. The querying against it looks quite expensive though. I wonder if they have built in any caching approach with the query engine. I guess I better do some reading.

  • @padhuLP
    @padhuLP 17 วันที่ผ่านมา

    Good tutorial. Thank you for sharing the code.

    • @fahdmirza
      @fahdmirza  16 วันที่ผ่านมา

      Thank you

  • @vitaliiturchenko8101
    @vitaliiturchenko8101 20 วันที่ผ่านมา +1

    Thanks Fahd for your hard work! Very interesting!!! 1) is it possible to link GraphRag to the local ChromaDB database ? 2) local search also works in your method or only global search ?

    • @fahdmirza
      @fahdmirza  19 วันที่ผ่านมา +1

      thaks. You would have to hack the source code to change the vector store. Yes local search also worked. Just have to replace global keyword with local.

  • @mikew2883
    @mikew2883 19 วันที่ผ่านมา +1

    Excellent tutorial! I was wondering if you had a chance to work with the "graphrag-accelerator" Github project that Microsoft also put out. It says it can be used as an API that has all the GraphRAG functionality but in an API.

    • @fahdmirza
      @fahdmirza  19 วันที่ผ่านมา

      I think graphrag-accelerator requires Azure. If its API based, I would rather go directly to OpenAI and I have already done a video on it.

  • @shameekm2146
    @shameekm2146 19 วันที่ผ่านมา

    Thanks for the video.
    Can i use mxbai from ollama for embedding purposes... or is there a limitation on that?

    • @fahdmirza
      @fahdmirza  18 วันที่ผ่านมา +1

      sure you can use it.

  • @ibc--mediators
    @ibc--mediators 20 วันที่ผ่านมา +1

    Hi Fahd …, 1. Where does graphrag store the vectors and graphs in? I.e on local machine… 2. how do we transfer the entire graphrag app from the local machine to into the cloud….once we are done with ingestion and testing

    • @fahdmirza
      @fahdmirza  19 วันที่ผ่านมา

      It has its own built-in vector store. For migration, I would suggest installing it from scratch in cloud.

  • @AdityaSingh-in9lr
    @AdityaSingh-in9lr 6 วันที่ผ่านมา

    hey, i got it working, but it is giving out of context answers when I do local search, any idea what could be wrong?

  • @TheMariolino2005
    @TheMariolino2005 17 วันที่ผ่านมา

    Great job! What if I want to add another document to the GraphRAG? Should I repeat the --init procedure or is there any other method? Great video, thank you.

    • @fahdmirza
      @fahdmirza  16 วันที่ผ่านมา +1

      Yes, you would have to run the index procedure. Thanks.

  • @zhengwu-jw6fm
    @zhengwu-jw6fm 19 วันที่ผ่านมา +1

    When run the code 'python3 -m graphing.index --root./rattiest',showers occurred during the pipeline run, See logs for more details.What to solve this problem?

    • @fahdmirza
      @fahdmirza  19 วันที่ผ่านมา

      plz check the logs in output directory and see what the error is. Also make sure that command is correct

  • @chrishau5556
    @chrishau5556 10 วันที่ผ่านมา

    Does this solution still works for anybody ?

  • @revanthphanisaimedukonduru1177
    @revanthphanisaimedukonduru1177 17 วันที่ผ่านมา

    Thanks for latest information, Can you please also add reference for this point , "GraphRAG don't support if its less than 32k context?" 7:22

    • @fahdmirza
      @fahdmirza  17 วันที่ผ่านมา +1

      That's on basis of trial at the moment of creating video.

  • @unknownu2e56
    @unknownu2e56 17 วันที่ผ่านมา

    hi i have a doubt about this graphRag can it be run in aws ec2 instance

    • @fahdmirza
      @fahdmirza  16 วันที่ผ่านมา +1

      yes

    • @unknownu2e56
      @unknownu2e56 16 วันที่ผ่านมา

      @@fahdmirza if possible please make a video ,that will be helpful for me

  • @EngineerMustaphaSahli
    @EngineerMustaphaSahli 3 วันที่ผ่านมา

    Thanks for sharing! ... Anyone else suffering from this error: "openai.APITimeoutError: Request timed out." ??

  • @khriza4991
    @khriza4991 13 วันที่ผ่านมา

    Thank you for the video. I'm facing the same error as another commenter mentioned: '❌ Errors occurred during the pipeline run, see logs for more details.' Where can I find the logs?

    • @fahdmirza
      @fahdmirza  13 วันที่ผ่านมา +1

      Sure, go to this directory ~/ragtest/output/20240711-055438/reports . The date directory would vary as per your run. You would log files there. Thanks.

    • @jiangnanfan8944
      @jiangnanfan8944 8 วันที่ผ่านมา

      @@fahdmirza raise ValueError(\"Columns must be same length as key\")
      ValueError: Columns must be same length as key
      ", "source": "Columns must be same length as key", "details": null , I FACE SAME ERROR , AND I FOUND THE LOG FILES , THEY SAID

    • @Gadgetwars
      @Gadgetwars 7 วันที่ผ่านมา

      @@jiangnanfan8944 I also face the same error "ValueError(\"Columns must be same length as key\", "details": null)

  • @narendrasingh-tg1mb
    @narendrasingh-tg1mb 17 วันที่ผ่านมา

    hi fahd thanks for video, getting this error : File "C:\Users\Narendrasingh\.conda\envs\graphollama\Lib\site-packages\graphrag\config\create_graphrag_config.py", line 229, in
    create_graphrag_config
    raise ApiKeyMissingError
    graphrag.config.errors.ApiKeyMissingError: API Key is required for Completion API. Please set either the OPENAI_API_KEY, GRAPHRAG_API_KEY or
    GRAPHRAG_LLM_API_KEY environment variable.
    ⠋ GraphRAG Indexer

    • @davidbeauchamp5867
      @davidbeauchamp5867 17 วันที่ผ่านมา

      Same issue here...below states to use "ollama" as API key. In which file should this be indicated?

  • @themax2go
    @themax2go 16 วันที่ผ่านมา

    @fahdmirza FYI on your webpage linked w/ the commands and code snippets for this vid, you have "model: nomic_embed_text" yet "ollama pull nomic-embed-text" which leads to: Error embedding chunk {'OpenAIEmbedding': 'Error code: 404 - {\'error\': "model \'nomic_embed_text\' not found, try pulling it first"}'}

  • @YoussefMohamed-fn6wl
    @YoussefMohamed-fn6wl 20 วันที่ผ่านมา +3

    first of all thank you,
    ZeroDivisionError: Weights sum to zero, can't be normalized when using local method

    • @fahdmirza
      @fahdmirza  19 วันที่ผ่านมา

      which model you are using?

    • @aravindchakrahari8966
      @aravindchakrahari8966 18 วันที่ผ่านมา +2

      I got the same error as well while using local method. And also, Error embedding chunk {'OpenAIEmbedding': "'NoneType' object is not iterable"}
      I am using mistral and nomic-embed-text:latest for embeddings.

    • @ayushjadia6527
      @ayushjadia6527 18 วันที่ผ่านมา +2

      I am also getting same error while using local method

    • @Ayush-tl3ny
      @Ayush-tl3ny 18 วันที่ผ่านมา +1

      same error with groq api llama3 8b and nomic embed text, any solution to this?

    • @guynas7582
      @guynas7582 17 วันที่ผ่านมา

      Issue is that `--method local` does not work out of the box with open source embedding models.
      It is because of the way how OpenAI's `text-embedding-3-small` model is working. It is using token IDs as input, while open source models like `nomic-embed-text` are working with text as input.
      So you need to convert token IDs to text before using open source models.
      Solution is to add one line to package's `graphrag/query/llm/oai/embedding.py` "embed" function :
      ```python
      ...
      def embed(self, text: str, **kwargs: Any) -> list[float]:
      """
      Embed text using OpenAI Embedding's sync function.
      For text longer than max_tokens, chunk texts into max_tokens, embed each chunk, then combine using weighted average.
      Please refer to: github.com/openai/openai-cookbook/blob/main/examples/Embedding_long_inputs.ipynb
      """
      token_chunks = chunk_text(
      text=text, token_encoder=self.token_encoder, max_tokens=self.max_tokens
      )
      chunk_embeddings = []
      chunk_lens = []
      for chunk in token_chunks:
      # decode chunk from token ids to text (added line after row 83)
      chunk = self.token_encoder.decode(chunk)
      try:
      embedding, chunk_len = self._embed_with_retry(chunk, **kwargs)
      chunk_embeddings.append(embedding)
      chunk_lens.append(chunk_len)
      # TODO: catch a more specific exception
      except Exception as e: # noqa BLE001
      self._reporter.error(
      message="Error embedding chunk",
      details={self.__class__.__name__: str(e)},
      )
      continue
      chunk_embeddings = np.average(chunk_embeddings, axis=0, weights=chunk_lens)
      chunk_embeddings = chunk_embeddings / np.linalg.norm(chunk_embeddings)
      return chunk_embeddings.tolist()
      ...
      ```

  • @aa-xn5hc
    @aa-xn5hc 19 วันที่ผ่านมา

    API key for Ollama should be "ollama". also, no need to do the embeddings locally because their cost is not high. The main objective should be to to do the LLM part with Ollama and then enquire both globally and locally.

    • @fahdmirza
      @fahdmirza  19 วันที่ผ่านมา +1

      That can be done too in various ways, but the purpose of this video to do it all in Ollama. Thanks for comment.

    • @davidbeauchamp5867
      @davidbeauchamp5867 17 วันที่ผ่านมา

      would you change this in the .env file or directly in the setting.yaml. I have the same issue as above where _config.py requires API key

  • @codelucky
    @codelucky 16 วันที่ผ่านมา

    Can you create a video on how to use GraphRAG with the GROQ API? Looks like nobody has done it yet. Thank you.

    • @fahdmirza
      @fahdmirza  16 วันที่ผ่านมา +1

      yeah just did. Thanks.

    • @codelucky
      @codelucky 16 วันที่ผ่านมา

      @@fahdmirza Thanks, I appreciate your work.

  • @sharankumar31
    @sharankumar31 17 วันที่ผ่านมา

    Kindly could you show, how to use this Graph RAG with CSV data. Will be super helpful

    • @fahdmirza
      @fahdmirza  16 วันที่ผ่านมา

      Its the same process as any data. The cleaner your data is, the better your responses will be.

  • @Thinker-i8d
    @Thinker-i8d 5 วันที่ผ่านมา

    perfect job. but when i try to use graphrag with ollama, error happened. logs.json shows: {"type": "error", "data": "Error Invoking LLM", "stack": "Traceback (most recent call last), and the index-engine.log shows:graphrag.index.reporting.file_workflow_callbacks INFO Error Invoking LLM
    does anyone know how to fix this error??

  • @ibc--mediators
    @ibc--mediators 19 วันที่ผ่านมา

    Langchain+neo4j+chroma = MS graphrag …. Correct?

    • @fahdmirza
      @fahdmirza  18 วันที่ผ่านมา

      Please explore this repo github.com/microsoft/graphrag for underlying tech. Thanks.

  • @shawnkratos1347
    @shawnkratos1347 19 วันที่ผ่านมา +1

    You only did global search what about local. That is only half the rag. I got this far and thought you figured it out

    • @fahdmirza
      @fahdmirza  19 วันที่ผ่านมา

      Its the same process, you just need to replace global with local

    • @shawnkratos1347
      @shawnkratos1347 19 วันที่ผ่านมา

      @@fahdmirza no it fails to build community reports:just tested again with mistral to make sure i have the exact same set up as you. look in the index-engine.log. 5:48:44,679 graphrag.index.graph.extractors.community_reports.community_reports_extractor ERROR error generating community report
      Traceback (most recent call last):
      File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/index/graph/extractors/community_reports/community_reports_extractor.py", line 58, in __call__
      await self._llm(
      File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/llm/openai/json_parsing_llm.py", line 34, in __call__
      result = await self._delegate(input, **kwargs)
      File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/llm/openai/openai_token_replacing_llm.py", line 37, in __call__
      return await self._delegate(input, **kwargs)
      File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/llm/openai/openai_history_tracking_llm.py", line 33, in __call__
      output = await self._delegate(input, **kwargs)
      File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/llm/base/caching_llm.py", line 104, in __call__
      result = await self._delegate(input, **kwargs)
      File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/llm/base/rate_limiting_llm.py", line 177, in __call__
      result, start = await execute_with_retry()
      File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/llm/base/rate_limiting_llm.py", line 159, in execute_with_retry
      async for attempt in retryer:
      File "/home/shawn/.local/lib/python3.10/site-packages/tenacity/asyncio/__init__.py", line 166, in __anext__
      do = await self.iter(retry_state=self._retry_state)
      File "/home/shawn/.local/lib/python3.10/site-packages/tenacity/asyncio/__init__.py", line 153, in iter
      result = await action(retry_state)
      File "/home/shawn/.local/lib/python3.10/site-packages/tenacity/_utils.py", line 99, in inner
      return call(*args, **kwargs)
      File "/home/shawn/.local/lib/python3.10/site-packages/tenacity/__init__.py", line 398, in
      self._add_action_func(lambda rs: rs.outcome.result())
      File "/usr/lib/python3.10/concurrent/futures/_base.py", line 451, in result
      return self.__get_result()
      File "/usr/lib/python3.10/concurrent/futures/_base.py", line 403, in __get_result
      raise self._exception
      File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/llm/base/rate_limiting_llm.py", line 165, in execute_with_retry
      return await do_attempt(), start
      File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/llm/base/rate_limiting_llm.py", line 147, in do_attempt
      return await self._delegate(input, **kwargs)
      File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/llm/base/base_llm.py", line 48, in __call__
      return await self._invoke_json(input, **kwargs)
      File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/llm/openai/openai_chat_llm.py", line 90, in _invoke_json
      raise RuntimeError(FAILED_TO_CREATE_JSON_ERROR).......python -m graphrag.query --root . --method global "what are the top themes in this story?"
      INFO: Reading settings from settings.yaml
      creating llm client with {'api_key': 'REDACTED,len=56', 'type': "openai_chat", 'model': 'mistral', 'max_tokens': 4000, 'request_timeout': 180.0, 'api_base': 'localhost:11434/v1', 'api_version': None, 'organization': None, 'proxy': None, 'cognitive_services_endpoint': None, 'deployment_name': None, 'model_supports_json': True, 'tokens_per_minute': 0, 'requests_per_minute': 0, 'max_retries': 10, 'max_retry_wait': 10.0, 'sleep_on_rate_limit_recommendation': True, 'concurrent_requests': 25}
      SUCCESS: Global Search Response: In the story, the main themes revolve around the transition of young people from formal education to practical work, specifically through apprenticeship under Ebenezer Scrooge. This transition is evident in various scenes and actions [Data: Scenes (1, 2, 3); Actions (4)].
      During their apprenticeship, the young people are engaged in specific tasks or responsibilities that are likely related to Scrooge's business [Data: Actions (1-5)]. It is also suggested that Scrooge may act as a mentor or supervisor to these apprentices during this period [Data: Relationships (1-23)].
      The young people are involved in various activities related to their apprenticeship, which could include tasks such as bookkeeping, accounting, or business management [Data: Actions (1-5)]. However, the exact nature of these activities is not explicitly detailed in the provided data.
      It is important to note that the information provided is based on the analysis of multiple reports and does not necessarily cover all aspects of the story. For a more comprehensive understanding, additional research or analysis may be required.
      shawn@pop-os:~/Documents/GRAPHRAG$ python -m graphrag.query --root . --method local "who is scrooge, and what are his main relationships?"
      INFO: Reading settings from settings.yaml
      creating llm client with {'api_key': 'REDACTED,len=56', 'type': "openai_chat", 'model': 'mistral', 'max_tokens': 4000, 'request_timeout': 180.0, 'api_base': 'localhost:11434/v1', 'api_version': None, 'organization': None, 'proxy': None, 'cognitive_services_endpoint': None, 'deployment_name': None, 'model_supports_json': True, 'tokens_per_minute': 0, 'requests_per_minute': 0, 'max_retries': 10, 'max_retry_wait': 10.0, 'sleep_on_rate_limit_recommendation': True, 'concurrent_requests': 25}
      creating embedding llm client with {'api_key': 'REDACTED,len=56', 'type': "openai_embedding", 'model': 'nomic_embed_text', 'max_tokens': 4000, 'request_timeout': 180.0, 'api_base': 'localhost:11434/api', 'api_version': None, 'organization': None, 'proxy': None, 'cognitive_services_endpoint': None, 'deployment_name': None, 'model_supports_json': None, 'tokens_per_minute': 0, 'requests_per_minute': 0, 'max_retries': 10, 'max_retry_wait': 10.0, 'sleep_on_rate_limit_recommendation': True, 'concurrent_requests': 25}
      Error embedding chunk {'OpenAIEmbedding': 'Error code: 404 - {\'error\': "model \'nomic_embed_text\' not found, try pulling it first"}'}
      Traceback (most recent call last):
      File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main
      return _run_code(code, main_globals, None,
      File "/usr/lib/python3.10/runpy.py", line 86, in _run_code
      exec(code, run_globals)
      File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/query/__main__.py", line 75, in
      run_local_search(
      File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/query/cli.py", line 154, in run_local_search
      result = search_engine.search(query=query)
      File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/query/structured_search/local_search/search.py", line 118, in search
      context_text, context_records = self.context_builder.build_context(
      File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/query/structured_search/local_search/mixed_context.py", line 139, in build_context
      selected_entities = map_query_to_entities(
      File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/query/context_builder/entity_extraction.py", line 55, in map_query_to_entities
      search_results = text_embedding_vectorstore.similarity_search_by_text(
      File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/vector_stores/lancedb.py", line 118, in similarity_search_by_text
      query_embedding = text_embedder(text)
      File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/query/context_builder/entity_extraction.py", line 57, in
      text_embedder=lambda t: text_embedder.embed(t),
      File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/query/llm/oai/embedding.py", line 96, in embed
      chunk_embeddings = np.average(chunk_embeddings, axis=0, weights=chunk_lens)
      File "/home/shawn/.local/lib/python3.10/site-packages/numpy/lib/function_base.py", line 550, in average
      raise ZeroDivisionError(
      ZeroDivisionError: Weights sum to zero, can't be normalized

  • @themax2go
    @themax2go 16 วันที่ผ่านมา

    python -m graphrag.query --root ./ --method local "explain relationships between the people in the story" leads to: ./graphrag/lib/python3.12/site-packages/numpy/lib/function_base.py", line 550, in average
    raise ZeroDivisionError(ZeroDivisionError: Weights sum to zero, can't be normalized - and before that: Error embedding chunk {'OpenAIEmbedding': "'NoneType' object is not iterable"}

  • @LuZhenxian
    @LuZhenxian 16 วันที่ผ่านมา

    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
    File "/Users/zhenxian/Documents/XXG/github/graphrag-main/graphrag/config/create_graphrag_config.py", line 231, in create_graphrag_config
    raise ApiKeyMissingError
    graphrag.config.errors.ApiKeyMissingError: API Key is required for Completion API. Please set either the OPENAI_API_KEY, GRAPHRAG_API_KEY or GRAPHRAG_LLM_API_KEY
    environment variable.

    • @donzhu4996
      @donzhu4996 16 วันที่ผ่านมา

      got the same error