Not Hacker News Logo

Not

Hacker

News!

Home
Hiring
Products
Companies
Discussion
Q&A
Users
Not Hacker News Logo

Not

Hacker

News!

AI-observed conversations & context

Daily AI-observed summaries, trends, and audience signals pulled from Hacker News so you can see the conversation before it hits your feed.

LiveBeta

Explore

  • Home
  • Hiring
  • Products
  • Companies
  • Discussion
  • Q&A

Resources

  • Visit Hacker News
  • HN API
  • Modal cronjobs
  • Meta Llama

Briefings

Inbox recaps on the loudest debates & under-the-radar launches.

Connect

© 2025 Not Hacker News! — independent Hacker News companion.

Not affiliated with Hacker News or Y Combinator. We simply enrich the public API with analytics.

Not Hacker News Logo

Not

Hacker

News!

Home
Hiring
Products
Companies
Discussion
Q&A
Users
  1. Home
  2. /Discussion
  3. /Show HN: Offline RAG System Using Docker and Llama 3 (No Cloud APIs)
  1. Home
  2. /Discussion
  3. /Show HN: Offline RAG System Using Docker and Llama 3 (No Cloud APIs)
Last activity 9h agoPosted Nov 26, 2025 at 9:32 AM EST

Offline Rag System Using Docker and Llama 3 (no Cloud Apis)

PhilYeh
1 points
0 comments

Mood

informative

Sentiment

positive

Category

startup_launch

Key topics

Artificial Intelligence
Data Privacy
Industrial Environments
Large Language Models
Docker
I'm sharing a fully offline RAG (Retrieval-Augmented Generation) stack I built to solve a crucial problem in industrial environments: data privacy and recurring API costs.

We deal with sensitive proprietary datasheets and schematics daily, making cloud-based LLMs like ChatGPT non-compliant.

The Solution: A containerized architecture that ensures data never leaves the local network.

The Stack: LLM: Llama 3 (via Ollama) Vector DB: ChromaDB Deployment: Docker Compose (One-click setup) Benefit: Zero API costs, no security risks, fast local performance. The code and architecture are available here: https://github.com/PhilYeh1212/Local-AI-Knowledge-Base-Docke...

Happy to answer questions about the GPU passthrough setup or document ingestion pipeline.

Offline RAG: Offline RAG System Using Docker and Llama 3 (No Cloud APIs)

Snapshot generated from the HN discussion

Discussion Activity

Light discussion

First comment

2h

Peak period

2

Hour 3

Avg / period

2

Key moments

  1. 01Story posted

    Nov 26, 2025 at 9:32 AM EST

    11h ago

    Step 01
  2. 02First comment

    Nov 26, 2025 at 11:45 AM EST

    2h after posting

    Step 02
  3. 03Peak activity

    2 comments in Hour 3

    Hottest window of the conversation

    Step 03
  4. 04Latest activity

    Nov 26, 2025 at 12:12 PM EST

    9h ago

    Step 04

Generating AI Summary...

Analyzing up to 500 comments to identify key contributors and discussion patterns

Discussion (0 comments)

Discussion hasn't started yet.

ID: 46057717Type: storyLast synced: 11/26/2025, 2:34:08 PM

Want the full context?

Jump to the original sources

Read the primary article or dive into the live Hacker News thread when you're ready.

Read ArticleView on HN
Not Hacker News Logo

Not

Hacker

News!

AI-observed conversations & context

Daily AI-observed summaries, trends, and audience signals pulled from Hacker News so you can see the conversation before it hits your feed.

LiveBeta

Explore

  • Home
  • Hiring
  • Products
  • Companies
  • Discussion
  • Q&A

Resources

  • Visit Hacker News
  • HN API
  • Modal cronjobs
  • Meta Llama

Briefings

Inbox recaps on the loudest debates & under-the-radar launches.

Connect

© 2025 Not Hacker News! — independent Hacker News companion.

Not affiliated with Hacker News or Y Combinator. We simply enrich the public API with analytics.