This AI Model Runs On Your Phone (With No Internet)!
This AI Model Runs On Your Phone (With No Internet)!
66 days agoMatt Wolfe@mreflow
YouTube11 min 52 sec
Watch on YouTube
Note: AI-generated summary based on third-party content. Not financial advice. Read more.
Quick Insights

The rapid rise of On-Device AI creates a strong bull case for Apple (AAPL), as users must upgrade to the iPhone 15 Pro or newer to handle the high computational and thermal demands of local models. Investors should also look toward semiconductor leaders like Qualcomm (QCOM) and ARM, which provide the essential neural engines required for this shift toward "Edge AI." The release of high-performing open-weight models like Qwen 3.5 suggests that "Sovereign AI" is becoming a viable alternative to paid, cloud-based subscriptions. This trend poses a significant disruption risk to the subscription moats of OpenAI and Google, as casual users may pivot to free, private, local alternatives for everyday tasks. To capitalize on this theme, focus on hardware manufacturers and chipmakers that facilitate high-efficiency processing with low power requirements.

Detailed Analysis

Locally AI (App)

The podcast highlights Locally AI, a mobile application that allows users to run powerful Large Language Models (LLMs) directly on their smartphones without an internet connection.

  • On-Device Processing: Unlike ChatGPT or Claude, which send data to the cloud, this app runs models locally using the phone's hardware.
  • Privacy & Security: Because it functions in Airplane Mode, no data is sent to companies like OpenAI, Google, or Anthropic, making it ideal for sensitive information.
  • Hardware Requirements:
    • iPhone 15 Pro or newer: Recommended for larger models (4 billion parameters).
    • iPhone 15: Suitable for mid-range models (2 billion parameters).
    • iPhone 14 or newer: Can run smaller models (800 million parameters).
  • Features: Includes "Thinking Mode" (Chain of Thought), image recognition (vision), custom instructions, and a voice mode.

Takeaways

  • Privacy-First AI: This represents a shift toward "Sovereign AI" where users control their data. Investors should watch for apps that prioritize local privacy as a competitive advantage over big tech.
  • Hardware Upgrade Cycle: The high computational demand of these models (noted by the phone getting warm and "choppy" performance during long chats) suggests a strong bull case for Apple (AAPL) and high-end smartphone manufacturers, as users will need the latest chips to run local AI effectively.

Qwen 3.5 (Model by Alibaba/Open-Weights)

The discussion focuses on the Qwen 3.5 family of models, which are "open-weight" models recently released (March 2nd).

  • Performance: Benchmarks suggest these small models are on par with or outperform GPT-4o Nano in specific tasks.
  • Variations: Available in 800M, 2B, 4B, and 9B parameter versions.
  • Capabilities: While not as logically advanced as the largest cloud models (e.g., Claude Opus or GPT-4), they are highly effective for brainstorming, basic logic, and real-time assistance.

Takeaways

  • Open-Source/Open-Weight Momentum: The rapid improvement of small models (SMMs) poses a threat to the "moats" of expensive subscription-based AI services.
  • Efficiency over Size: There is a growing investment theme in "efficient AI"—models that provide high utility with low power and memory requirements.

Edge Computing & On-Device AI (Sector Theme)

The transcript highlights a broader shift from cloud-based AI to "Edge AI" (AI that lives on the device).

  • Utility in Connectivity Gaps: The ability to use AI on airplanes or in remote areas without Wi-Fi creates new use cases for travel, emergency services, and parenting.
  • Technical Limitations: The speaker noted that very large state-of-the-art models still require "insane GPU power" and cloud infrastructure, but the gap is closing for everyday tasks.

Takeaways

  • Semiconductor Demand: This trend reinforces the long-term demand for mobile processors capable of AI acceleration (NPU/Neural Engines). Key players include Apple, Qualcomm, and ARM.
  • Disruption of SaaS: If users can get "good enough" AI for free locally on their devices, paid AI subscriptions may see increased churn among casual users.

Risk Factors Mentioned

  • Thermal Issues: Running local AI causes devices to heat up significantly, which could impact battery health over time.
  • Performance Degradation: As conversation context grows, the app becomes "choppy" and slow, indicating that mobile RAM and processing still have limits.
  • Logic Flaws: Smaller local models still struggle with complex logic (e.g., the "strawberry" test or situational reasoning) compared to massive cloud-based models.
Ask about this postAnswers are grounded in this post's content.
Video Description
Trying out locally AI to run AI models on my phone without internet. Discover More: 🛠️ Explore AI Tools & News: https://futuretools.io/ 📰 Weekly Newsletter: https://futuretools.io/newsletter 🎙️ The Next Wave Podcast: https://youtube.com/@TheNextWavePod Socials: ❌ Twiter/X: https://x.com/mreflow 🖼️ Instagram: https://instagram.com/mr.eflow 🧵 Threads: https://www.threads.net/@mr.eflow 🟦 LinkedIn: https://www.linkedin.com/in/matt-wolfe-30841712/ 👍 Facebook: https://www.facebook.com/mattrwolfe Resources From Today's Video: https://locallyai.app/ Let’s work together! - Brand, sponsorship & business inquiries: mattwolfe@smoothmedia.co #AINews #AITools #ArtificialIntelligence
About Matt Wolfe
Matt Wolfe

Matt Wolfe

By @mreflow

AI News Breakdowns every Saturday and other cool nerdy tech and AI stuff in between. Let's work together! - For brand ...