• 0 Posts
  • 3 Comments
Joined 2 years ago
cake
Cake day: June 13th, 2024

help-circle

  • sounds like an unsanctioned chat bot going through internal private data

    Probably better than handing that internal private data to a cloud provider. At least with this setup, it will all stay in the network under their control. There should be no reason to give the inference server access to the internet.


  • theunknownmuncher@lemmy.worldtoHomelabAm I getting ripped off?
    link
    fedilink
    English
    arrow-up
    11
    ·
    edit-2
    9 hours ago

    5070 Ti

    for the entire 15-20 person firm

    Local AI is a great option to look into, but I can’t imagine that’s going to go well… 16GB of VRAM is going to limit you to very small models and small context size. I imagine for a law firm, you’re going to want the AI to be reading lots of documents, so lots of context. Maybe it will be fine depending on how the 15-20 people access it, but I’m doubtful.

    Is this machine just a proof of concept to start putting together a process and testing the waters? I wouldn’t call total bullshit immediately, but I’d expect you’ll eventually find that you need much more VRAM and probably a heavier development lift to integrate with n8n.