• unwind ai
  • Posts
  • OpenAI o1 now has an OpenSource Replica

OpenAI o1 now has an OpenSource Replica

PLUS: NoCode Multi-Agent AI Researcher, Opensource OpenAI Voice Mode

  • Explore future trends, infrastructure & scalability, and more

  • Learn from the best in Model Development and Performance Optimization

  • Get inspired by real-world case studies and success stories

Don’t miss out - use code SHUBHAM15 for 15% off your ticket! 

See you in Austin, TX, November 7-8, 2024

Today’s top AI Highlights:

  1. This opensource project is a faster & cheaper alternative to OpenAI Strawberry models

  2. Opensource version of OpenAI’s advanced Voice Mode

  3. Video-to-Video feature in Runway lets you morph videos with simple prompts

  4. AI research engine driven by a network of interconnected AI agents

& so much more!

Read time: 3 mins

Latest Developments

OpenAI’s o1 models have brought a lot of excitement due to their advanced reasoning abilities, but developers have flagged two significant pain points: they’re slow to start generating output (time-to-first-token is high), and the reasoning-heavy responses are expensive due to the cost of output tokens. These tokens also take up a large portion of the model’s context window. Trying to tackle this, a developer has created g1, an open-source project powered by Llama-3.1 on GroqInc that achieves the same structured reasoning as o1 but via a system prompt.

Key Highlights:

  1. System Prompt - o1’s chain-of-thought reasoning comes from extensive training and RL strategies. In contrast, g1 uses a dynamic system prompt that instructs the LLM to perform reasoning in steps, using commands like “explore alternative answers” and “use at least 3 methods to derive the solution.”

  2. Speed and context improvements - g1 reduces the delay in the time to first token while still using reasoning chains to solve problems step by step. This offers faster responses without sacrificing quality.

  3. Improved logic handling - In an initial test on common logic problems (like counting letters in a word) where even the leading LLMs stutter, g1 achieved 70% accuracy - significantly outperforming Llama 3.1 alone, which had 0% accuracy, and GPT-4o, which managed only 30%. Please note that these are not formal benchmarks in any way.

  4. What can you do with it - g1 is a great a prototype to experiment with reasoning improvements. It’s an actionable solution for those seeking to enhance the reasoning ability of opensource models like Llama without the expense and delay of using o1, fine-tuning, or retraining.

Didn’t get access to OpenAI’s advanced Voice Mode? Don’t sweat, here’s an opensource version of it that can process as well as generate both text and audio simultaneously in real-time. Mini-Omni language model eliminates the delays caused by separate speech recognition and text-to-speech components typically found in other models. It is based on the Qwen-2-0.5B model and trained on a combination of speech and text datasets, including a new VoiceAssistant-400K dataset designed specifically for this purpose.

Key Highlights:

  1. End-to-End Speech Interaction - Mini-Omni handles both speech recognition and generation internally. No need to integrate separate ASR/TTS models.

  2. Parallel Text and Audio - The model can "think" (process information) and "speak" (generate audio) concurrently for more natural and responsive interactions.

  3. "Any Model Can Talk" - This method allows you to add speech capabilities to language models you are already using with minimal training and modification, using adapters and a small amount of synthesized data.

  4. Opensource - The GitHub repository provides the model code, training scripts, and the dataset for you to experiment, fine-tune, and build upon Mini-Omni. 

Quick Bites

Runway has released Video-to-Video feature with the Gen-3 Alpha model for its paying customers. It lets you change the style of your videos by using a text prompt.

Google AI Studio now has a new “Compare” feature where you can compare the output of different models with different parameters side-by-side. It is on the top-right menu in the UI.

OpenAI is likely to change its structure from a non-profit to a traditional for-profit business next year. This comes as they are in talks to raise $6.5 billion, with investors pushing for a removal of the current profit cap.

Godmother of AI Fei-Fei Li has raised $230 million to launch World Labs, her startup focused on creating AI with spatial intelligence to understand 3D environments. The company wants to develop advanced models for applications like AR/VR and robotics.

Tools of the Trade

  1. Compute by Hyperspace: Builds and runs AI workflows for research by breaking down your prompt into smaller sub-tasks that are handled by a network of customizable AI agents. It achieves cost-effectiveness and customization by:

    • Using small, open-source AI models instead of large, expensive ones.

    • Allowing you to choose from a variety of AI models and providers, including running models locally.

    • Integrating built-in AI primitives like code execution, web search, and access to a vast vector database.

.

  1. Trustworthy Language Model: Adds a trustworthiness score to each LLM output to help determine which responses are reliable and which need review. It’s designed to reduce errors and hallucinations in enterprise apps.

  2. langgraph_streamlit_codeassistant: AI assistant that integrates Python execution capabilities with React component rendering on the fly, offering a comprehensive environment for data analysis, visualization, and interactive web development.

  3. Awesome LLM Apps: Build awesome LLM apps using RAG to interact with data sources like GitHub, Gmail, PDFs, and YouTube videos through simple text. These apps will let you retrieve information, engage in chat, and extract insights directly from content on these platforms.

Hot Takes

  1. Too many startups/investors are exploring the b2b opportunities with LLMs and not enough are exploring the consumer opportunities. This is the moment consumer founders have been waiting for. ~
    Michael Seibel

  2. If we’re being honest:
    Aside from Nvidia, the people who have benefitted the most financially from AI are spammers.
    We’re not far from most of the content we see on social networks, on search engines, and on phone calls being AI-generated. ~
    Nikita Bier

Meme of the Day

That’s all for today! See you tomorrow with more such AI-filled content.

Real-time AI Updates 🚨

⚡️ Follow me on Twitter @Saboo_Shubham for lightning-fast AI updates and never miss what’s trending!

Unwind AI - Twitter | LinkedIn | Instagram | Facebook

PS: We curate this AI newsletter every day for FREE, your support is what keeps us going. If you find value in what you read, share it with at least one (or 20) of your friends!

Reply

or to participate.