- Superpower Daily
- Posts
- OpenAI reveals 'Strawberry' AI to the feds, powering its new 'Orion' project.
OpenAI reveals 'Strawberry' AI to the feds, powering its new 'Orion' project.
The world's fastest AI is here
In today’s email:
🔥 Google Meet’s automatic AI note-taking is here
😭 ‘Make your health insurance company cry’: One woman’s fight to turn the tables on insurers
📚 Why AI can’t spell ‘strawberry’
🧰 12 new AI-powered tools and resources. Make sure to check the online version for the full list of tools.
OpenAI is making significant strides with a new AI model code-named "Strawberry," which is designed to tackle complex problems without the hallucinations that often plague current AI systems. Although Strawberry is reportedly slower and more expensive at inference time, it excels at solving intricate challenges on the first try. The model's main role is to generate synthetic data for "Orion," OpenAI's upcoming flagship large language model (LLM). This focus on creating accurate synthetic data could help Orion reduce errors and improve its reasoning capabilities, making it a significant advancement in AI development.
OpenAI is also pushing to incorporate a distilled version of Strawberry into ChatGPT, potentially as early as this fall. This smaller, simplified version would allow ChatGPT to benefit from Strawberry's improved reasoning abilities, offering more accurate responses across a range of tasks, including math problems and programming challenges. While this integration could slow down response times, the trade-off would be worth it for users seeking more reliable and thoughtful answers.
In addition to enhancing ChatGPT, OpenAI has demonstrated Strawberry's capabilities to U.S. national security officials, underscoring its potential broader applications. The model's ability to generate synthetic data could address limitations in obtaining high-quality training data from real-world sources, a critical issue for the development of future AI systems. Strawberry's origins trace back to research led by Ilya Sutskever, a former chief scientist at OpenAI, and continued by Jakub Pachocki and Szymon Sidor.
While others scramble, savvy entrepreneurs are quietly transforming their businesses with AI.
Here's the kicker: You don't need a Ph.D. to join them.
Our no-BS guide cuts through the hype, giving you:
A jargon-free breakdown of large language models (the engine behind AI's magic)
Concrete steps to leverage AI for immediate business growth
Real-world examples that turn "prompt engineering" from gibberish into your secret weapon
Don't let the tech heads have all the fun. Grab your share of the AI pie – no coding required.
Google Meet has introduced a new AI-powered feature called "take notes for me," which aims to automate note-taking during meetings. This feature, now rolling out to select Google Workspace customers, can summarize key points from discussions rather than just transcribe the entire conversation. Initially, the feature supports only spoken English and is available to customers with the Gemini Enterprise, Gemini Education Premium, or AI Meetings & Messaging add-ons. Once the meeting concludes, the notes are saved in a Google Doc, which is automatically attached to the calendar event and shared with the meeting organizer and anyone who enabled the feature.
In addition to summarizing conversations, the tool is designed to provide a meeting recap for latecomers, allowing them to catch up quickly without disrupting the flow. If users also opt to record and transcribe the meeting, those files will be linked in the same document as the notes. This integration aims to streamline post-meeting organization and accessibility, particularly benefiting those who may find it difficult to process spoken language while taking notes simultaneously.
While this feature holds promise for improving efficiency and accessibility, there are still concerns about its accuracy. Users of Google's transcription tool have reported issues with incorrect transcriptions, often needing to review and edit the generated content. As a result, there is some skepticism about whether the note-taking tool will offer a significant improvement in capturing the essence of conversations accurately. Despite these concerns, Google is hopeful that the rollout, expected to reach all eligible customers by September 10th, 2024, will meet users' needs.
Cerebras has unveiled its new AI inference solution, which claims to be the fastest in the world. The Cerebras inference system delivers an impressive 1,800 tokens per second for Llama3.1 8B and 450 tokens per second for Llama3.1 70B—20 times faster than traditional GPU-based cloud solutions. The offering also stands out with industry-leading pricing, starting at just 10 cents per million tokens for Llama3.1 8B. Developers can now access Cerebras' inference capabilities via an open API, making it easier to integrate into various applications.
Powered by Cerebras' third-generation Wafer Scale Engine (WSE-3), the system breaks through the memory bandwidth limitations that often slow down traditional LLM inference. By storing the entire model on a single chip, Cerebras can achieve speeds that far exceed GPU-based solutions. With an aggregate memory bandwidth of 21 petabytes per second, the WSE-3 offers 7,000 times more memory bandwidth than leading GPUs like the H100, eliminating bottlenecks and enabling real-time inference for even the largest models.
Unlike some other platforms that compromise accuracy by reducing model precision, Cerebras inference uses Meta's original 16-bit weights to maintain the highest accuracy possible. This dedication to precision ensures that the Llama3.1 models perform optimally in complex tasks, such as multi-turn conversations and reasoning challenges. Cerebras is also planning to add larger models, including Llama3 405B and Mistral Large, in the near future, further expanding the platform's capabilities.
The launch of Cerebras inference signifies a major leap forward in AI technology, offering both unmatched speed and cost-effectiveness. The platform’s high-speed processing opens up new possibilities for AI applications, from real-time decision-making to advanced code generation. By solving long-standing challenges in AI inference, Cerebras is positioning itself as a leader in the future of AI development.
Other stuff
‘Make your health insurance company cry’: One woman’s fight to turn the tables on insurers 🔥🔥
Elon Musk backs California bill to regulate AI
The World’s Call Center Capital Is Gripped by AI Fever and Fear 🔥
Three-quarters of founders in the latest Y Combinator cohort are working on AI startups
Police officers are starting to use AI chatbots to write crime reports. Will they hold up in court? 🔥
Why AI can’t spell ‘strawberry’ 🔥
Anthropic publishes the ‘system prompts’ that make Claude tick
OpenAI exits San Francisco HQ after Elon Musk stops paying rent
All your ChatGPT images in one place 🎉
You can now search for images, see their prompts, and download all images in one place.
Clockwise Prism - A first-of-its-kind intelligent calendar
Astra AI - Universal API for adding any app to your LLM
Hey! - AI-powered pair programming friend
Mimrr - Eliminating technical debt for your startup.
Bento - AI-Generated Movies
Ragie - Fully managed RAG-as-a-Service for developers
CommandDash - AI assistant for open-source libraries
Kypso for Code Reviews - AI code reviews without the noise
Astra AI - Universal API for adding any app to your LLM
GPT Subtitler - Translate your subtitles to any language with LLMs
Seven24.ai - Capture feedback and turn it into tasks
Face Animator - Instant reactions & expressions with AI
Unclassified 🌀
WFH Team - Work from anywhere in the world
How did you like today’s newsletter? |
Help share Superpower
⚡️ Be the Highlight of Someone's Day - Think a friend would enjoy this? Go ahead and forward it. They'll thank you for it!
Hope you enjoyed today's newsletter
Did you know you can add Superpower Daily to your RSS feed https://rss.beehiiv.com/feeds/GcFiF2T4I5.xml
⚡️ Join over 200,000 people using the Superpower ChatGPT extension on Chrome and Firefox.
OR