- AI Collections @Beehiiv
- Posts
- Will Siri team up with ChatGPT?
Will Siri team up with ChatGPT?
PLUS: Eyes peeled for today's OpenAI event, Alibaba's compute-for-equity model
Todayโs top AI Highlights:
Apple is closing a deal with OpenAI to use ChatGPT to power AI features in Apple devices
What to expect at OpenAIโs Spring Event today
Alibaba is growing cloud compute infrastructure revenue with a unique compute-for-equity model
Meta, Microsoft, and UC Berkeleyโs new research for enhancing LLMsโ domain-specific performance
Use Claude to create production-ready prompts optimized for LLMs
& so much more!
Read time: 3 mins
Latest Developments ๐
Will Siri be powered by ChatGPT? ๐ฃ๏ธ
Apple has realized they need to bark the โgenerative AIโ word much louder than before. The company has not trained any LLM in-house but has built powerful chips to run AI on Apple devices. With this, it is nearing a deal with OpenAI to bring ChatGPT to iPhones. The deal would see ChatGPT features integrated into the upcoming iOS 18 operating system and powering AI in more apps.
Key Highlights:
Deal with Google: Apple has also engaged in discussions with Google about potentially licensing its Gemini chatbot, although no agreement has been reached yet.
Siri by GPT-4: Rumors are flying that Siri might start using ChatGPTโs brain to answer your questions. Imagine asking Siri something and getting smarter replies.
Why it matters: If Siri gets this upgrade, it means your iPhone will not just respond faster but also smarter. You could ask for cooking tips, solve math problems, or even get relationship advice.
Stay tuned: Nothingโs official yet, but keep your eyes peeled at the WWDC 2024 on June 10.
What to Expect at OpenAIโs Event Today ๐
OpenAIโs Spring event is scheduled for today and the social media platforms are brimming with speculations. Till last week, it was even rumored that OpenAI is planning to launch a search engine to compete with Google, however, Sam Altman has clearly negated this rumor. Itโll be all about new features and updates to ChatGPT. Letโs see what the speculations are and what we can expect today:
Connected Apps: ChatGPT will soon have the Connected Apps feature, integrating with Google Drive and Microsoft One Drive. This will let you attach documents from the cloud directly in ChatGPT, saving the unnecessary hassle of downloading the document first and uploading it again.
New Voice Assistant: ChatGPT already has a Voice Assistant that uses a transcription model (Whisper), an LLM (GPT 4), and a text-to-speech model (TTS-1). But new reports say that OpenAI has developed an advanced Voice Assistant that is much more capable than the existing one, especially at understanding the mood of the speaker. This would make up for a really good AI customer rep.
Phone Calls by ChatGPT: OpenAI might release a new feature allowing ChatGPT to make phone calls using WebRTC technology. This means users could opt-in to have real-time voice calls with ChatGPT. We would probably receive ChatGPTโs calls to get reminders for meeting or completing a project, or call us back once it is done with its research.
New Models: OpenAI might introduce three new models -
gpt4-lite: a lighter version of the current GPT-4 model to replace GPT-3.5. It might be smaller, possibly faster but still maintains some of the capabilities of GPT-4.
gpt4-auto: This could be a new model endpoint that can automatically gather data from the web and other sources. This would be different from the current capability where you have to explicitly tell GPT-4 to search the web to get the latest information.
gpt4-lite-auto: A combination of the above two features.
Alibaba, the Chinese e-commerce giant, is taking a novel approach to increase its revenue while also promoting Chinaโs generative AI scene. Instead of traditional cash investments, Alibaba is leveraging its vast cloud computing infrastructure to offer computing credits to promising AI startups. This strategy allows Alibaba to secure stakes in these companies while also boosting its cloud business.
Key Highlights:
Compute-for-Equity: Alibaba is providing startups with credits to use its powerful cloud computing resources, specifically for training AI models, in exchange for equity in the companies.
Advantage: This approach addresses the scarcity of advanced computing resources in China due to US chip export restrictions. This also mirrors Microsoftโs successful investment in OpenAI by supporting local equivalents of ChatGPT and other AI applications.
Example: Alibaba led a $1 billion funding round in Moonshot AI that is building an LLM that can handle long inputs. Nearly half of Alibabaโs $800 million contribution was in the form of cloud computing credits.
The Dual Incentive: This not only allows Alibaba to secure a foothold in the rapidly growing AI sector but also drives revenue for its cloud computing division, which has experienced slowing growth in recent quarters.
Technical Research ๐ฌ
RAFTing LLM towards More Domain-Specific Accuracy ๐ฉโโ๏ธ
LLMs are trained on a vast and diverse dataset to make them capable of answering questions on a wide range of topics. However, many usecases require LLMs to specialize more in a specific domain. In this case, supervised fine-tuning becomes very laborious and the quality of RAG depends heavily on the quality of the document retrieval process.
Researchers at UC Berkeley, Meta AI, and Microsoft have proposed a method called Retrieval Augmented Fine-Tuning (RAFT) for LLMs to focus only on the important parts of documents in a specific domain without relying on the document retriever.
Key Highlights:
Selective Training: RAFT trains language models to differentiate between โoracleโ documents that contain answers and โdistractorโ documents that do not, enhancing focus on relevant information.
Model and Tools: RAFT was applied on the Llama 2-7B model, trained on Microsoft AI Studio. It was chosen due to its language understanding, math skills, and ability to parse long documents.
Benchmarks and Performance: After using RAFT, Llama 2-7B was evaluated on datasets like PubMed, HotpotQA, and Gorilla, where it showed major improvements in accuracy for domain-specific Q&A.
Deployment Flexibility: The trained models can be deployed on various platforms, including GPUs or CPUs, via Microsoft AI Studio and llama.cpp, making it adaptable for different enterprise needs.
๐ Enjoying so far, share it with your friends!
Tools of the Trade โ๏ธ
Anthropic Console: Generate production-ready prompts in this Console using Claude-3 itself. Just describe the task, and Claude-3 will turn it into a high-quality prompt that works best with LLMs, specifically Claude-3. You can even invite team members to collaborate and use these prompts.
llm-ui: A React library for building user interfaces for LLMs. It provides features like correcting broken markdown syntax, custom component integration, and output throttling for a smoother user experience. llm-ui offers code block rendering using Shiki and a headless design allowing for complete style customization.
StartKit AI: Provides boilerplate code with OpenAI integration, Node.js API, Mongo Database, and Pinecone vector storage to quickly build AI tools and SaaS products. You can use it to create applications such as ChatGPT clones, PDF analysis tools, image generation apps, and more, with pre-built modules for common AI tasks.
Mock-My-Mockup: Upload a screenshot of a page youโre working on, and get brutally honest feedback. Itโll highlight both the positives and negatives of the page, with a little roast.
Hot Takes ๐ฅ
Our advantage currently lies not in compute, cloud, or chips. ๐๐ฎ๐ซ ๐๐๐ฏ๐๐ง๐ญ๐๐ ๐ ๐ข๐ฌ ๐จ๐ฎ๐ซ ๐ฉ๐จ๐ฉ๐ฎ๐ฅ๐๐ญ๐ข๐จ๐ง ๐๐ง๐ ๐ญ๐ก๐๐ข๐ซ ๐๐ฌ๐ฉ๐ข๐ซ๐๐ญ๐ข๐จ๐ง๐ฌ. ๐๐ก๐ข๐ฌ ๐ข๐ฌ ๐ฐ๐ก๐ฒ ๐ฐ๐ ๐ก๐๐ฏ๐ ๐ญ๐จ ๐๐ซ๐ข๐ง๐ ๐๐จ๐ฐ๐ง ๐ญ๐ก๐ ๐๐จ๐ฌ๐ญ ๐จ๐ ๐ข๐ง๐๐๐ซ๐๐ง๐๐ ๐๐ซ๐จ๐ฆ ๐๐ฌ 100 ๐ญ๐จ ๐๐ฌ 1โฆโฆ๐๐ง๐ง๐จ๐ฏ๐๐ญ๐ ๐๐ซ๐ฎ๐ ๐๐ฅ๐ฅ๐ฒ ๐ญ๐จ ๐๐ซ๐๐ฆ๐๐ญ๐ข๐๐๐ฅ๐ฅ๐ฒ ๐ซ๐๐๐ฎ๐๐ ๐ญ๐ก๐ ๐๐จ๐ฌ๐ญ ๐จ๐ ๐๐. ๐๐ ๐๐๐ง'๐ญ ๐๐๐ฅ๐ข๐ฏ๐๐ซ ๐ข๐ญ ๐ญ๐จ ๐ ๐๐ข๐ฅ๐ฅ๐ข๐จ๐ง ๐ฉ๐๐จ๐ฉ๐ฅ๐ ๐ฎ๐ง๐ฅ๐๐ฌ๐ฌ ๐ฐ๐ ๐๐๐ง ๐๐๐ ๐ข๐ง ๐ญ๐จ ๐๐ก๐๐ซ๐ ๐ 1 ๐ซ๐ฎ๐ฉ๐๐ ๐ฉ๐๐ซ ๐ญ๐ซ๐๐ง๐ฌ๐๐๐ญ๐ข๐จ๐ง. ~Nandan Nilekani on Indiaโs future in AI
If OpenAI really is releasing a voice assistant tomorrow, itโs highly likely to be a true end-to-end system. Itโs the natural way forwardโฆ a somewhat-better multi-model system wonโt be enough to really wow anyone. End-to-end with low latency would be a breakthrough. ~Matt Shumer
Meme of the Day ๐คก
Thatโs all for today! See you tomorrow with more such AI-filled content.
Real-time AI Updates ๐จ
โก๏ธ Follow me on Twitter @Saboo_Shubham for lightning-fast AI updates and never miss whatโs trending!
PS: I curate this AI newsletter every day for FREE, your support is what keeps me going. If you find value in what you read, share it with your friends by clicking the share button below!