Coder Social home page Coder Social logo

curtisgray / wingman Goto Github PK

View Code? Open in Web Editor NEW
38.0 2.0 1.0 192.1 MB

Wingman is the fastest and easiest way to run Llama models on your PC or Mac.

Home Page: https://curtisgray.github.io/wingman/

License: MIT License

Dockerfile 0.57% TypeScript 86.57% JavaScript 6.89% CSS 0.48% HTML 0.86% PowerShell 3.03% Shell 1.60%
ai chatbot chatgpt linux llama llamacpp llm local macos windows

wingman's Introduction

Wingman is a zero-configuration chatbot for running AI models locally. You can download and use open source models like Meta’s Llama 2, Mistral and phi 2 from Hugging Face directly in the app. Wingman is also compatible with OpenAI AI models. It runs on Windows (Nvidia GPUs or CPU-based) and MacOS (Intel and Apple Silicon).

Wingman - Rooster is the first release of Wingman. Future releases will be named after characters from the Top Gun franchise.


Requirements for Hardware and OS

  • Windows 10+ or MacOS 10.15+
  • 2016 or newer CPU
  • 8GB+ RAM
  • 2016 or newer GPU with at least 3GB of RAM
  • GPU not required, but recommended
  • 100GB free disk space (for AI models)
  • 1GB free disk space (for app)

🔔 KNOWN ISSUES - FIXES ON THE WAY 🔔

  • If you run a model that is too large on an Apple Silicon mac, the app may crash and get into an unrecoverable state. Relaunch the app. If it hangs on the "Ready for takeoff!" screen, you must reboot your computer to relaunch the app.
  • If you run an AI model that is just slightly too large on Windows, the AI model will crash. Wingman will attempt to recover anyway. If Wingman hangs on, "Connecting to the Wingman Service," restarting the app will fix the problem.
  • On Apple Silicon Macs, the "Ready for Takeoff" feature that detects if AI models work on your machine is too aggressive. The algorithm is not tuned for newer Macs, so some AI models may erroneously state that they can be run on your machine. It's pretty accurate, but it needs more adjustments.

Click below to watch the video

Watch the video


Features

  • Easy to use UI with no terminals and no code required.
  • Runs on Windows and Mac (Intel or Apple Silicon).
  • It’s a free, open-source app.
  • Run Large Language Models (LLMs) like Meta’s Llama 2, Mistral, Yi, Microsoft’s phi 2, OpenAI, zephyr and more all in the same app with a familiar chatbot interface.
  • Quick access to LLMs from Hugging Face right inside the app. You can even see what’s popular or trending.
  • Swift Switch: Quickly swap between models mid-conversation for the best results.
  • Private by design: the LLMs all run on your machine, so you can keep your chats private.
  • Wingman will evaluate your machine so you can see at a glance what models may or may not run on your hardware. We won’t stop you from trying any of them, though!
  • Save and organize your chats into folders so you can find them again later.
  • Set system prompts and prompt templates so you can talk with characters and get the best results from all models.

Sponsored by Electric Pipelines: Offering enhanced software engineering. For custom Wingman modifications or deployments, contact us today!


Planned Features

Silk Tuning – Personalized AI Feedback Loop: At its core, Wingman employs an innovative feature called Silk Tuning. This mechanism enables users to rate AI responses in real-time, directly influencing the AI's learning process. The system then uses this feedback to generate a Learning Rate Adjustment (LoRA), which dynamically tailors the AI’s responses to align more closely with individual user preferences and interaction styles. This process of personalized adjustment occurs during periods of low activity on your PC, ensuring continuous improvement and customization of the AI experience.

Flight Formation - Collaborative AI Optimization: Wingman's Flight Formation unlocks the power of community-driven AI model enhancement. Users can rate responses, share optimized prompts and prompt templates, and contribute to a growing repository of knowledge. This collective wisdom improves model outputs, tailoring them to deliver more accurate and relevant results. Flight Formation creates a feedback loop that allows AI models to continuously learn and adapt, ensuring its capabilities soar to new heights through the combined efforts of the Wingman community.

Airborne Server - Take to the Cloud: With Airborne Server, users can deploy their own private Wingman instance on cloud infrastructure or an external server. This innovative feature provides enhanced control, scalability, and customization options. Whether running resource-intensive models, handling sensitive data, or integrating with existing systems, Airborne Server empowers users to take flight with Wingman's capabilities in their own secure, tailored environment.

Genius Primer - Unlock Model Potential: Genius Primer is a groundbreaking feature that amplifies the capabilities of smaller language models, transforming them into high-performing AI powerhouses. By leveraging advanced priming techniques, Genius Primer equips modest models with knowledge and contextual understanding far beyond their natural capacity. This cutting-edge technology opens up new realms of possibility, enabling users to harness the full potential of AI without the need for resource-intensive, large-scale models.

Radar Augmentation - Contextualize External Sources: Wingman's Radar Augmentation increases AI awareness by contextualizing external documents and multimedia directly into your conversations. Whether incorporating PDFs, text files, images, or YouTube videos, this powerful feature allows any open-source model to build upon the insights and knowledge contained within auxiliary sources. With Radar Augmentation, you can simply paste links or upload files, and the AI will ingest and synthesize that supplementary information to produce outputs backed by an enriched, multifaceted context. Expand your AI's perspective by providing access to external documents and media.

Super Context - Defy Context Limitations: Shatter the constraints of traditional context windows with Wingman's Super Context capability. This innovative feature exponentially extends any model's maximum context length, empowering even smaller, lower-capacity models to build upon a vastly expanded knowledge base. By dynamically increasing the conversational context far beyond natural bounds, Super Context facilitates generation of coherent, substantive outputs that seamlessly combine information across broad timelines and subject areas. Unlock new frontiers of hyper-aware AI outputs backed by comprehensive, contextualized understanding that defies ordinary context limitations.

Wingman is more than just an application; it's a gateway to a more accessible and tailored AI experience. By eliminating the barriers of complex AI integration and offering unique features for personalized and flexible interactions, Wingman stands as a valuable tool for those wanting to explore and utilize AI technologies with ease and efficiency.

Wingman is under heavy development. Expect frequent updates.

Progress Updates

  • ✅ Add All GPT Models
  • ✉️ Internationalization
  • ✅ Duplicate existing conversations
  • ✅ Add local Llama support (in progress)
  • ✅ Download Llama models locally
  • ⏳ Desktop app <- (will be out in the next week!)
  • ⏳ Deploy to Windows Store and Mac App Store
  • ✉️ Silk Tuning
  • ✅ Swift Switch
  • ⏳ Radar Augmentation
  • ⏳ Airborne Server
  • ✉️ Super Context
  • ✉️ Genius Primer
  • ✉️ Flight Formation

Legend: ✅ Completed, ⏳ In Progress, ✉️ Not Started

Installation

Note, an OpenAI API key is NOT needed, unless you want to run an OpenAI GPT-style model. Otherwise it is not needed. Wingman is designed to run without it.

Wingman will autodetect whether you have an OpenAI API key and make a GPT AI models available automatically.

Native Installers (Windows and MacOS)

Download the latest release from the Releases page.

Build From Source

See Build Locally for instructions on how to build Wingman locally.

Prerequisites

Build Locally

1. Clone Repo
git clone https://github.com/curtisgray/wingman.git --recurse-submodules
2. Build Front and Back Ends
build.ps1
3. Run App

Windows:

  • Select the Start button, and type wingman, then select Wingman from the list of results.

Mac:

  • Open Finder and navigate to the Applications folder. Double-click Wingman.

Use Wingman

Download a model from within the app, or enter your OpenAI API key to use OpenAI models.

Contact

If you have any questions reach out to Electric Curtis on Twitter.

Acknowledgements

wingman's People

Contributors

saltylightning avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar

Forkers

visioninhope

wingman's Issues

Wingman is unable to detect the GPU in ASUS TUF Laptop

Describe the bug
Wingman does not detect and use my GPU on my ASUS TUF FX505DT. Instead, it uses the CPU.

Additional context
My laptop shows two GPUs in the task manager:
image

Specific Configuration
ASUS TUF FX505DT Gaming Laptop- 15.6"
CPU: AMD Ryzen 5 R5-3550H Processor
GPU: GeForce GTX 1650 Graphics
RAM: 8GB DDR4

Clicking links opens them in Wingman

Describe the bug
Clicking a link that a model provides in an answer opens the link inside of Wingman.

Steps to Reproduce

  • Open Wingman
  • Create a new chat
  • Select the phi 2 dpo model
  • Type the following message to the model: "Give me a link to google.com"
  • Click the link

Expected behavior
Open the link in an external browser.

Actual behavior
The link opens in Wingman.

Additional context
If you hold control and press the link on Windows, it will open another Wingman window with the link. Browser back on a mouse button seems to take you back to regular Wingman.

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.