• NVIDIA released a demo version of a chatbot that runs locally on your PC, giving it access to your files and documents.

• The chatbot, called Chat with RTX, can answer queries and create summaries based on personal data fed into it.

• It supports various file formats and can integrate YouTube videos for contextual queries, making it useful for data research and analysis.

  • General_Effort@lemmy.world
    link
    fedilink
    English
    arrow-up
    50
    ·
    9 months ago

    That was an annoying read. It doesn’t say what this actually is.

    It’s not a new LLM. Chat with RTX is specifically software to do inference (=use LLMs) at home, while using the hardware acceleration of RTX cards. There are several projects that do this, though they might not be quite as optimized for NVIDIA’s hardware.


    Go directly to NVIDIA to avoid the clickbait.

    Chat with RTX uses retrieval-augmented generation (RAG), NVIDIA TensorRT-LLM software and NVIDIA RTX acceleration to bring generative AI capabilities to local, GeForce-powered Windows PCs. Users can quickly, easily connect local files on a PC as a dataset to an open-source large language model like Mistral or Llama 2, enabling queries for quick, contextually relevant answers.

    Source: https://blogs.nvidia.com/blog/chat-with-rtx-available-now/

    Download page: https://www.nvidia.com/en-us/ai-on-rtx/chat-with-rtx-generative-ai/

    • GenderNeutralBro@lemmy.sdf.org
      link
      fedilink
      English
      arrow-up
      14
      ·
      9 months ago

      Pretty much every LLM you can download already has CUDA support via PyTorch.

      However, some of the easier to use frontends don’t use GPU acceleration because it’s a bit of a pain to configure across a wide range of hardware models and driver versions. IIRC GPT4All does not use GPU acceleration yet (might need outdated; I haven’t checked in a while).

      If this makes local LLMs more accessible to people who are not familiar with setting up a CUDA development environment or Python venvs, that’s great news.