← Home

The AI Arsenal

Four world-class AI platforms. One unified workflow. Zero compromises. This is how we build the future.

"We don't pick a side in the AI wars. We pick the best weapon for each battle."

Most teams pick one AI provider and stick with it. We believe that's leaving performance on the table. Each AI platform has unique strengths — Claude's code generation, Gemini's research synthesis, ChatGPT's creative flexibility, and Qwen's local privacy. By orchestrating all four through Project Blue Note AI Rhino, we get capabilities no single model can match.

PRIMARY PARTNER

Anthropic Claude

Claude Opus 4 • Sonnet 4 • Haiku • Claude Code • Claude CoWork

Claude is our primary AI development partner — and for good reason. Anthropic's models consistently deliver the most reliable, nuanced, and technically accurate code generation in the industry.

Claude Code is our secret weapon. Running directly in the CLI, it enables AI-assisted pair programming at a level that transforms how software gets built. From architecture design to full implementation, Claude Code generates production-ready code that actually ships.

Claude CoWork takes collaboration further — enabling persistent AI workspaces where complex, multi-step engineering tasks can be planned, iterated, and delivered with unprecedented efficiency.

How We Use Claude

  • Full-stack code generation (PHP, React, MySQL, CSS)
  • System architecture and API design
  • Code review and refactoring at scale
  • Database schema design and query optimization
  • DevOps automation and server configuration
  • Technical documentation and specification writing
claude-code
$ claude ╭──────────────────────────────────╮ │ Claude Code v1.0 │ │ Model: claude-opus-4 │ │ Mode: Interactive │ ╰──────────────────────────────────╯ You: Build a complete AI inference proxy with PHP 8.4 Claude: I'll create a production-ready AI inference proxy with routing, caching, and multi-model support. Creating files... ✓ src/Proxy/Router.php ✓ src/Proxy/ModelAdapter.php ✓ src/Cache/InferenceCache.php ✓ src/API/Endpoints.php ✓ tests/ProxyTest.php All tests passing. Ship it? 🚀
Opus 4 Primary Model
CLI Claude Code
24/7 Active Usage
RESEARCH ENGINE

Google Gemini

Gemini 2.0 • NotebookLM • Multimodal AI • Research Synthesis

Google's Gemini platform serves as our research and knowledge management engine. Where Claude excels at code generation, Gemini shines at synthesizing complex information from multiple sources into actionable intelligence.

NotebookLM is a game-changer for our workflow. We feed it technical documentation, API specs, research papers, and domain knowledge — and it transforms them into interactive, queryable knowledge bases. It's like having a research assistant that has actually read and understood every document in your organization.

Gemini's multimodal capabilities also power our image analysis pipeline — processing screenshots, diagrams, and visual data alongside text for truly comprehensive understanding.

How We Use Gemini

  • NotebookLM for interactive knowledge bases
  • Research synthesis across technical papers
  • Multimodal analysis (text + images + code)
  • Documentation processing and summarization
  • Competitive analysis and market research
  • Audio overview generation for team briefings

NotebookLM Workflow

📄 Upload Docs
🧠 Gemini Processes
💡 Interactive KB
🎙️ Audio Overviews
🚀 Actionable Insights
2.0 Gemini Version
NLM NotebookLM
Multi Modal
CREATIVE ENGINE

OpenAI / ChatGPT

GPT-4o • o1 Reasoning • DALL-E • ChatGPT Plus

OpenAI's ChatGPT ecosystem is our creative powerhouse. When we need brainstorming, ideation, image generation, or rapid creative prototyping, ChatGPT is the go-to platform.

GPT-4o handles the heavy lifting for creative content generation — marketing copy, product descriptions, user interface text, and the kind of human-sounding prose that makes products feel polished and professional.

o1's reasoning capabilities come into play for complex problem-solving that benefits from chain-of-thought reasoning — architecture decisions, algorithm design, and the kind of deep analytical thinking that requires step-by-step logic.

DALL-E and the image generation pipeline power our visual assets — from concept art and UI mockups to marketing materials and social media graphics.

How We Use OpenAI

  • Creative content and copywriting
  • Image generation with DALL-E
  • Complex reasoning with o1
  • Rapid prototyping and ideation
  • User experience text and microcopy
  • Marketing materials and presentations
✍️

Content

Marketing copy, blog posts, and product descriptions that actually convert.

🎨

Imagery

DALL-E generates visual assets from concept art to production graphics.

🧩

Reasoning

o1 handles complex multi-step reasoning and architectural decisions.

💬

Ideation

ChatGPT as a brainstorming partner for rapid creative exploration.

GPT-4o Primary
o1 Reasoning
DALL-E Images
LOCAL INFERENCE

Alibaba Qwen

Qwen 2.5 • 32B Parameters • 72B Parameters • Self-Hosted via Ollama

This is where things get really interesting. While Claude, Gemini, and ChatGPT are cloud services, Qwen runs entirely on our own hardware — zero cloud dependency, zero data leaving our infrastructure, zero latency overhead.

Through Ollama, we deploy Qwen models ranging from the lightning-fast 32B variant to the heavyweight 72B model — all accelerated by our NVIDIA RTX 5090's 32GB of VRAM. This gives us enterprise-grade AI inference that we fully own and control.

Qwen excels at multilingual tasks (crucial for our Norwegian operations), code generation in multiple languages, and the kind of rapid-fire inference that benefits from zero network latency. When milliseconds matter, local inference wins.

This is the beating heart of Project Blue Note AI Rhino — proving that you don't need to rent your AI from the cloud. You can build it, own it, and run it yourself.

How We Use Qwen

  • Self-hosted inference with zero cloud dependency
  • Multilingual processing (English, Norwegian, +)
  • Privacy-sensitive data processing
  • Rapid-fire inference with zero network latency
  • Offline capability for mission-critical tasks
  • Cost-effective high-volume processing
ollama@bluenote-rhino
$ ollama list NAME SIZE MODIFIED qwen2.5:72b 41GB 2 hours ago qwen2.5:32b 18GB 1 hour ago llama3.1:8b 4.7GB 3 hours ago $ nvidia-smi NVIDIA RTX 5090 32768 MiB VRAM | 92% Utilization Temperature: 62°C | Power: 340W $ ollama run qwen2.5:72b >>> Processing at 45 tokens/sec ✓ Local inference: ACTIVE ✓ Cloud dependency: NONE ✓ Data sovereignty: FULL
72B Parameters
45 Tok/sec
0ms Network Lag

How It All Works Together

The real power isn't in any single model — it's in how we orchestrate them.

   ┌─────────────┐     ┌──────────────┐     ┌─────────────┐     ┌─────────────┐
   │   CLAUDE     │     │    GEMINI     │     │   CHATGPT    │     │    QWEN      │
   │  Code + Dev  │     │  Research +   │     │  Creative +  │     │  Local +     │
   │  Architecture│     │  NotebookLM   │     │  Ideation    │     │  Private     │
   └──────┬───────┘     └──────┬────────┘     └──────┬───────┘     └──────┬───────┘
          │                    │                     │                    │
          └────────────────────┼─────────────────────┼────────────────────┘
                               │                     │
                    ┌──────────┴─────────────────────┴──────────┐
                    │       PROJECT BLUE NOTE AI RHINO           │
                    │       ════════════════════════════         │
                    │                                            │
                    │    Unified Orchestration Layer              │
                    │    Route → Process → Deliver                │
                    │                                            │
                    │    RTX 5090 │ 48 Cores │ 120GB RAM         │
                    └───────────────────┬────────────────────────┘
                                        │
                    ┌───────────────────┴────────────────────────┐
                    │           SHIPPED PRODUCTS                  │
                    │     Trivia & Tunes │ Portfolios │ APIs      │
                    └────────────────────────────────────────────┘
                
1

Task Analysis

Every incoming task is analyzed to determine which AI platform(s) are best suited. Code task? Claude. Research? Gemini. Creative? ChatGPT. Private? Qwen.

2

Intelligent Routing

Blue Note AI Rhino routes tasks to the optimal model through our unified API. Some tasks use a single model; complex tasks may chain multiple models together.

3

Processing & QA

Results are processed, validated, and quality-checked. Code gets tested. Content gets reviewed. Data gets verified. Nothing ships without passing QA.

4

Delivery

Finished products deploy through our CI/CD pipeline. From AI-generated code to published content — everything ships production-ready.

👤 Dave Gilligan Creator & Architect
🎵 Blue Note Logic Inc. Infrastructure & Tech
🇳🇴 Gilligan Tech ENK Local Operations, Norway