Download the AI & Machine Learning For Dummies PRO App: iOS - Android Our AI and Machine Learning For Dummies PRO App can help you Ace the following AI and Machine Learning certifications:
In September 2024, the landscape of artificial intelligence is evolving at an unprecedented pace, with groundbreaking innovations and developments emerging daily. This blog, titled “AI Innovations in September 2024,” aims to capture the essence of these advancements, providing a comprehensive summary of the latest news and trends in AI technology. As we navigate through a month filled with transformative AI breakthroughs, our ongoing updates will highlight significant milestones, from the launch of cutting-edge AI models to the integration of AI in various sectors such as healthcare, finance, and creative industries. With each passing day, AI is reshaping how we interact with technology, enhancing productivity, and redefining our understanding of intelligence itself. Join us as we explore the exciting world of AI innovations, keeping you informed and engaged with the rapid changes that are set to influence our future. Whether you’re a tech enthusiast, a professional in the field, or simply curious about the implications of AI, this blog will serve as your go-to resource for staying updated on the latest developments throughout September 2024.
California governor vetoes hotly contested AI safety bill
TikTok-parent to develop new AI model using Huawei chips
Gemini 1.5 compared to other AI models
The age of the AI-first developer
Real-world use cases of Gemini 1.5
Proactive AI agent systems
California governor vetoes hotly contested AI safety bill
California Governor Gavin Newsom vetoed the Safe and Secure Innovation for Frontier Artificial Intelligence Models Act (SB 1047), citing concerns about the bill’s broad scope and potential burden on AI companies.
Governor Newsom stated that SB 1047 could give the public a false sense of security and hamper innovation, despite agreeing on the need for safety protocols and clear consequences for bad actors in the AI industry.
The bill, which faced opposition from tech companies and notable political figures, sought to impose stringent requirements on AI models costing over $100 million to train, including safeguards like a “kill switch” and protections for whistleblowers.
Google just released two new upgraded versions of Gemini 1.5 across the Gemini API, including 1.5 pro-002, which achieved state-of-the-art performance across math benchmarks, and 1.5-flash-002, which makes big gains in instruction following.
Google’s latest Gemini updates significantly lower the financial barrier for AI development while boosting performance, especially in math. With these updates, Gemini now tops the LLM leaderboard in terms of performance-to-price ratio, context windows, video understanding, and other LLM benchmarks.
Google also announced significant improvements to accessibility for developers building with Gemini models, including a 50% reduced price on 1.5 Pro, 2x higher rate limits on Flash and 3x higher on 1.5 Pro, 2x faster output, and 3x lower
Google’s Gemini project is only around a year old. Google was the first to ship 1M context windows (and 2M) and context caching, and they’ve been making rapid progress ever since.
AI is helping developers tackle significantly harder problems faster while simultaneously lowering the entry barrier for non-developers to contribute to new innovation and even build their own AI apps.
For the person who’s never coded before, they’re now able to tackle like any problem with code because they have this co-pilot in their hands.
[For beginners] ai.google.dev is our default landing page that also links out to the Gemini API documentation. On GitHub, we have a Quickstart repo where you can literally run four commands have a local version of AI Studio and Gemini running on your computer to play around with the models.
With AI as an assistant, some developers are tackling 10x more challenging software problems—which also means 10x the speed of improvements and 10x the innovation, for those who use the tech wisely. Google also has great resources to help even complete beginners get started in less than 5 minutes.
Gemini 1.5’s multimodal capabilities allow a host of real-world applications that other models can’t match, such as processing and analyzing hour-long videos or entire books—thanks to its impressive 2M token context window.
Being able to go into an AI studio and just drop an hour-long video in there and ask a bunch of questions is such a mind-blowing experience.
The intent was to build a multimodal model from the ground up…the order of magnitude of important use cases for the world, for developers and for people who want to build with this technology, so many of them are multimodal.
Gemini 1.5’s 2M context window allows it to process and analyze long-form content like long videos, entire books, and lengthy podcasts, opening new possibilities for content analysis and interaction. For a full look at its potential, check out Google’s list of 185 real-world gen AI use cases from leading organizations.
Proactive AI agent systems
The future of AI is likely to shift from reactive to proactive systems, with AI agents capable of initiating actions and asking for clarification or permission, much like human assistants do today.
By shifting from purely reactive to proactive systems, AI could become more like a true “Her-like“ assistant, anticipating needs and offering solutions before being prompted. At the current state, no AI systems do this effectively, but as AI continues to advance with projects like Astra, this is likely the next stage for AI.
Liquid AI introduces non Transformer Based LFMs (Liquid Foundational Models) with State of the Art performance on many benchmarks, while being much more memory efficient.
Emad Mostaque says that unlike humans, AIs can scale seamlessly by producing swarms of intelligent agents that will improve one another by checking each other’s work
The French AI company Mistral has introduced Pixtral 12B, its first multimodal model capable of processing both images and text.
OpenAI has released two next-generation AI models to its subscribers: o1 preview and o1 mini. These models show a significant improvement in performance, particularly in tasks requiring reasoning, including coding, mathematics, GPQA, and more.
Chinese company Alibaba releases the Qwen 2.5 model in various sizes, ranging from 0.5B to 72B. The models demonstrate capabilities comparable to much larger models.
The video generation model KLING 1.5 has been released.
OpenAI launches the advanced voice mode of GPT4o for all subscribers.
Meta releases Llama 3.2 in sizes 1B, 3B, 11B, and 90B, featuring image recognition capabilities for the first time.
Google has rolled out new model updates ready for deployment, Gemini Pro 1.5 002 and Gemini Flash 1.5 002, showcasing significantly improved long-context processing.
Kyutai releases two open-source versions of its voice-to-voice model, Moshi.
TSMC execs dismiss OpenAI CEO Sam Altman as a ‘podcasting bro’
YouTube support added to NotebookLM
Google’s new AI creates its own chips
Google Deepmind’s AlphaChip AI has created three generations of TPUs, which are now widely used in data centers globally.
Initially unveiled in 2021, AlphaChip was the first AI method to design chip “floorplans,” significantly reducing design time from months to hours.
The AI model continues to improve with each TPU generation, achieving better chip layouts and proving more efficient than human designers for various hardware platforms.
TSMC execs dismiss OpenAI CEO Sam Altman as a ‘podcasting bro’
OpenAI CEO Sam Altman was reportedly dismissed as a “podcasting bro” by senior TSMC executives during his tour in the Far East last winter.
Altman proposed a massive $7 trillion investment plan for AI advancements, including 36 new semiconductor plants, which TSMC execs found impractical and overly ambitious.
The dismissive attitude towards Altman’s proposals reflects broader skepticism, with OpenAI’s current business model showing significant financial discrepancy between its income and expenditure.
Google just upgraded its NotebookLM tool, adding support for YouTube videos and audio files, along with easier sharing of Audio Overviews—its latest viral AI hit that turns notes, PDFs, Google Docs, and more into AI-generated podcasts.
NotebookLM now supports public YouTube URLs and audio files, allowing users to analyze videos, lectures, and audio alongside existing text sources.
The tool leverages Gemini 1.5’s multimodal capabilities to summarize key concepts from videos and transcribe audio.
A new sharing feature allows users to generate public links for Audio Overviews, making collaboration even easier.
These updates aim to streamline tasks such as creating study guides, analyzing multiple perspectives on issues, and extracting important information from video, audio, and text.
It’s a big day for Google. The company’s viral hit with NotebookLM is now even more impressive with access to YouTube videos and audio files. YouTube is an endless treasure chest of how-to guides, lectures, documentaries, and entertainment—and now, anyone can consume hours worth of videos in minutes with AI.
Archaeologists from Japan’s Yamagata University, in collaboration with IBM Research, used AI to uncover 303 previously unknown geoglyphs near Peru’s famous Nazca Lines, nearly doubling the number of known figures at the site.
The newly discovered geoglyphs, dating back to 200 BC, depict various animals and humans, including parrots, cats, monkeys, killer whales, and even decapitated heads.
AI combined with low-flying drones dramatically accelerated the discovery process, accomplishing nearly a century’s worth of work in six months.
These smaller geoglyphs (10-25 feet across) provide new insights into the transition from the Paracas culture to the Nazca culture.
The findings, published in the Proceedings of the National Academy of Sciences, demonstrate AI’s ability to help greatly improve archaeological research.
Is there anything AI can’t help us accomplish? The amount of time saved using low-flying drones and artificial intelligence is worth repeating: 100 years worth of work in six months. The ways in which AI is going to impact our lives are still vast and largely unknown, as this discovery proves.
Visaagreed to acquire AI-driven payments protection firm Featurespace to enhance its financial crime and fraud detection capabilities—the acquisition price was not disclosed.
Hugging Facereached 1 million free public AI models on its platform, highlighting the trend towards specialized models for diverse use cases rather than a single dominant model.
A new language model programming library has been released, and it’s called Ell. It aims to compete with, and possibly replace, libraries like LangChain, Llama-Index, and DSPy.
I know — yet another library for composing language programs. And brace yourself, I might be building my own in a few weeks or months. But that’s a story for another day.
Here’s the tweet announcing Ell. Ambitious from the start, with the core maintainer calling it nothing less than the future of prompt engineering. Source: https://x.com/wgussml/status/1833615864131948756
Mark Zuckerberg: creators and publishers ‘overestimate the value’ of their work for training AI
Welcome to Read Aloud For Me, the pioneering AI dashboard designed for the whole family! Our platform is the first of its kind, uniquely crafted to cater not only to adults but also to kids.
OpenAI CTO exits amid rumors of non-profit removal
AI breakthrough in treating rare diseases
💰YCombinator’s AI boom is still going strong (W24)
Meta reveals its Orion AR smart glasses
OpenAI CTO Mira Murati leaves the company
Sam Altman could get 7% stake in OpenAI, worth $10B
FTC announces crackdown on ‘deceptive AI’ businesses
Meta unveils new Quest 3S headset and discontinues older models
Meta reveals its Orion AR smart glasses
Meta has unveiled its Orion augmented reality glasses, which showcase full holographic capabilities, marking the most advanced AR glasses the company has developed after a decade-long effort.
These smart glasses use tiny projectors to display holograms, are lightweight, and feature a contextual AI for improved digital experiences, both indoors and outdoors, without obstructing the user’s view of the real world.
Orion glasses require a wireless puck for processing and a bracelet for gesture control, with Meta currently allowing only employees and select audiences to test the device as it works towards a consumer release in the future.
Mira Murati, the CTO of OpenAI, announced on social media that she is leaving the company after more than six years to pursue her own interests.
OpenAI’s CEO Sam Altman expressed gratitude for Murati’s contributions and stated that more information about the transition plans will be shared soon.
Murati’s resignation adds to recent high-level departures at OpenAI, including former safety leader Jan Leike and co-founder John Schulman.
Harvard Medical School researchers recently developed an AI model called TxGNN that can identify existing drugs for repurposing to treat rare and neglected diseases.
TxGNN identified drug candidates from nearly 8,000 existing medicines for over 17,000 diseases, many without current treatments.
The model outperformed leading AI drug repurposing tools by nearly 50% in identifying candidates and was 35% more accurate in predicting contraindications, a specific situation in which a medcine should not be used.
TxGNN uses a novel approach that identifies shared features across multiple diseases, allowing it to extrapolate from well-understood conditions to poorly understood ones.
The researchers have made the tool freely available to encourage its use by clinician-scientists in the search for new therapies, especially for rare and untreated conditions.
Another week, another insane medical breakthrough for AI. While we still need years of clinical validation and approvals before widespread use, TxGNN has the potential to save thousands of lives and improve the lives of people who likely thought a treatment for their specific disease would never come.
Sam Altman could get 7% stake in OpenAI, worth $10B
OpenAI is in discussions to grant CEO Sam Altman a 7% equity stake as part of its shift towards a for-profit business model, which would be the first time Altman holds ownership in the AI company.
The company is considering becoming a public benefit corporation, aiming to turn a profit while also prioritizing societal benefits, amidst ongoing leadership changes, including the surprise departure of CTO Mira Murati.
OpenAI is currently raising $6.5 billion at a $150 billion valuation, which could potentially boost Altman’s net worth by over $10 billion, placing him among the wealthiest individuals in the world.
FTC announces crackdown on ‘deceptive AI’ businesses
DoNotPay, claiming to offer the “world’s first robot lawyer,” agreed to a $193,000 settlement with the FTC for making untested claims about its AI services.
The settlement is part of the FTC’s Operation AI Comply, targeting companies that use artificial intelligence to deceive or defraud customers.
The FTC ruled that DoNotPay must warn subscribers about its AI’s limitations and prohibit the company from claiming it can replace professional services without evidence.
Meta unveils new Quest 3S headset and discontinues older models
Meta introduced the Meta Quest 3S, a budget-friendly VR headset, at its Connect event on Wednesday, promising great performance despite the lower price point.
Mark Zuckerberg announced the Meta Quest 3S, priced at $299.99, which is significantly cheaper compared to the $499.99 Meta Quest 3 or $999.99 Meta Quest Pro.
The Quest 3S, equipped with a Snapdragon XR2 Gen2 chip and 8GB RAM, offers lower display specs and uses Fresnel lenses, making it distinct from the Quest 3.
We are excited to share #PDF2Audio, an open-source alternative to the #podcast feature of #NotebookLM with flexibility & tailored outputs that you can precisely control in the app: You can make a podcast, lecture, discussions, short/long form summaries & more, including the use of the amazingo1 model (@sama)
Unlike NotebookLM, this tool is 100% open-source on Huggingface. That means, the community isn’t reliant on one super-tech giant and can now iterate and improve the tool themselves.
OpenAI isreportedlydeveloping an improved version of its Sora AI video generation model, aiming for higher quality and longer clips than previously demonstrated.
Metaannounced it will not immediately join the European Union’s voluntary AI Pact, instead focusing on compliance with the upcoming AI Act regulations.
Nvidia analystspredicted the company will produce around 450,000 Blackwell AI GPUs in Q4 2024, potentially generating over $10 billion in revenue despite initial production challenges.
The Federal Trade Commission announced enforcement actions against multiple companies for deceptive or unfair use of artificial intelligence in their practices.
OpenAI CEO Sam Altman said the Advanced Voice Mode rollout for ChatGPT has been completed early, except in jurisdictions requiring additional external review.
Hacker plants false memories in ChatGPT to steal user data
OpenAI is rolling out long-awaited ‘advanced voice’ feature
Microsoft uses AI to fix hallucinations in AI models
Google releases production-ready models
Hacker plants false memories in ChatGPT to steal user data
Security researcher Johann Rehberger found a flaw in ChatGPT’s long-term memory settings that allowed attackers to store false information and malicious instructions, which OpenAI initially did not consider a security concern.
Rehberger created a proof-of-concept exploit that exfiltrated all user input perpetually by storing false memories using indirect prompt injections from untrusted sources like Google Drive or OneDrive, prompting OpenAI to issue a partial fix.
Though OpenAI’s fixes prevent memory abuse for data exfiltration, prompt injections from untrusted content can still plant long-term false memories, making it crucial for users to monitor and review stored memories regularly to prevent attacks.
OpenAI is rolling out long-awaited ‘advanced voice’ feature
OpenAI is expanding the availability of Advanced Voice Mode, a feature in ChatGPT that lets users communicate more naturally with the AI by speaking and interrupting its responses mid-sentence.
Initially released in July to an invite-only group, this advanced mode, which can adjust responses based on the user’s tone of voice, is now being rolled out to Plus and Team users, with plans to extend access to Enterprise and Edu tiers.
To ensure safety, OpenAI has had external experts test the Advanced Voice Mode, addressing potential issues like violent or erotic speech and voice imitation, though the model’s closed-source nature limits independent evaluation.
Microsoft uses AI to fix hallucinations in AI models
Microsoft introduced a feature called “correction” within Azure AI Studio, aimed at detecting and fixing inaccuracies in AI-generated outputs by automatically rewriting incorrect content.
The correction tool identifies errors by comparing AI outputs with the customer’s original source material, explains the inaccuracies, and corrects them before users see the initial content.
Microsoft acknowledges that while the correction tool uses both small and large language models to align AI outputs with reference documents, it does not entirely guarantee accuracy and is still prone to errors.
Google just announced significant updates to its Gemini AI models, including performance improvements, cost reductions, and increased accessibility for developers.
Two new production-ready models came out today: Gemini-1.5-Pro-002 and Gemini-1.5-Flash-002, offering improved quality across various tasks, including a 20% boost in math-related benchmarks.
Pricing for Gemini 1.5 Pro has been reduced by over 50% for both input and output on prompts under 128K tokens, while rate limits have been increased significantly.
The models boast 2x faster output and 3x lower latency compared to previous versions, with improvements in long context understanding and vision capabilities.
Google also updated its default filter settings, giving developers more control over model configuration for their specific use cases.
Google is iterating quickly and pushing the boundaries of affordability for developers building with AI. While this isn’t Gemini 2 — it is a significant upgrade over the experimental models and will help builders create faster, smarter, cheaper applications.
Intellaunched Xeon 6 processors and Gaudi 3 AI accelerators, doubling performance for AI workloads and offering improved price and performance compared to Nvidia’s H100.
OpenAIincreased API access for o1 models, adding tier 4 to the list of authorized users at 100 requests per minute and upping tier 5 users to 1000 requests per minute.
Suno AIannounced a new cropping feature available to AI-generated songs, allowing Pro and Premier users to adjust the start and end of their creations.
Duolingointroduced AI-powered Adventures mini-games and a Video Call feature to enhance language learning through immersive, practical experiences for its users.
Apple unveiled its plan to roll out Siri’s major AI-powered updates gradually, with the most significant enhancements expected in iOS 18.3, likely launching in January 2025.
Next, you’ll need to navigate to NotebookLM and click “New Notebook”. After creating a notebook, you need to populate it with your article!
Step 3) Upload an article into NotebookLM
The UI for uploading sources to NotebookLM
Google provides numerous options for uploading articles, including Google Drive, linking websites, and uploading files such as markdown files, txt files, and PDFs.
To make it easy, I literally just copy/pasted my article with the Paste text feature. I then waited a minute, and it redirected me to my new notebook!
Step 4) Generate a notebook using the “Generate” button
The “Notebook Guide” on NotebookLM
After being redirected, we need to click “Notebook guide” next to the input box. From there, a menu pops up, and we can see Audio Overview in the right of the screen. Click “Generate” and wait 5 minutes, and the audio is generated!
This audio is enough to start distributing it to major platforms, such as Apple Podcasts, Spotify, and Amazon Music. I will discuss that more towards the end of the article.
Trending AI Tools on September 25th 2024:
Adobe GenStudio – Helps marketing teams measure on-brand content: https://business.adobe.com/products/genstudio.html?
FactBot by Snopes – Fact-checking for urban legends and misinformation: https://www.snopes.com/factbot/
JustPaid – Automate invoice follow-ups and payment tracking: https://www.justpaid.ai/
ell – A lightweight prompt engineering framework for language models: https://github.com/MadcowD/ell
Pathway – Helps product teams test UX solutions and gather insights: https://pathway.cc/
Tubit AI – AI that summarizes YouTube videos for a deeper understanding: https://mylens.ai/apps/youtube
Welcome to Read Aloud For Me, the pioneering AI dashboard designed for the whole family! Our platform is the first of its kind, uniquely crafted to cater not only to adults but also to kids.
Superintelligence may be here sooner than expected
Nvidia and Alibaba to improve autonomous cars
AI video-to-video generation with Runway
IBM, NASA team up on new AI climate model
AI superintelligence could arrive in ‘few thousand days’
Boston Dynamics robots can now autonomously unlock doors
Superintelligence may be here sooner than expected; could arrive in ‘few thousand days’
OpenAI CEO Sam Altman just suggested that superintelligent AI could emerge in just a few thousand days, marking a potentially transformative moment in human history that could usher in an era of unprecedented prosperity and capability.
Sam Altman, CEO of OpenAI, believes that artificial superintelligence could be achieved within “a few thousand days,” emphasizing this milestone as a pivotal moment in human history.
Altman stresses the importance of investing in computing resources and infrastructural improvements to make AI widely accessible and prevent it from becoming a limited resource.
Altman’s personal blog post on this topic coincides with OpenAI’s efforts to raise $6.5 billion, though the timeline he suggests is not the company’s official stance.
Being CEO of OpenAI, Sam Altman knows more about the current capabilities of AI than almost anyone else on the planet—and he is hyperoptimistic about the future. But regardless of whether or not superintelligence is here in 5-10 years or 25-30 years, it’s coming, Altman says, and it’s going to change everything.
Alibaba Cloud and Nividia just announced a new collaboration to develop advanced AI solutions for autonomous driving, integrating Alibaba’s large language models with Nvidia’s automotive computing platform.
Alibaba’s advanced Qwen AI models will be integrated into Nvidia’s Drive AGX Orin platform, which is already used by major Chinese electric vehicle manufacturers.
The partnership aims to enhance in-car voice assistants with more dynamic conversations and intelligent recommendations based on visual and environmental data.
The companies are also working on adapting Alibaba’s AI models for Nvidia’s next-generation Drive Thor platform, combining advanced driver assitance, autonomous driving, and AI driver capabilities.
Two powerhouses in the AI space teaming up to fix issues with autonomous driving is a huge plus for advancing the car industry, but Nvidia deciding to use Alibaba’s Qwen models is an even bigger, and another unexpected win for open-source.
Boston Dynamics robots can now autonomously unlock doors
Assa Abloy, a prominent Swedish company in the lock and security business, is teaming up with Boston Dynamics to develop a digital door access system for autonomous entry by Spot robots.
This system leverages Bluetooth technology to allow Boston Dynamics’ Spot robots to communicate with Assa Abloy’s HID-branded Signo readers, enabling hassle-free access through automatic doors.
Given Assa Abloy’s extensive global footprint in door access solutions, their collaboration with Boston Dynamics could accelerate the adoption of robot security patrols in commercial buildings.
Runway now offers video-to-video, a new feature that allows you to simply upload a real-life reference video and generate a completely different AI video based on it.
Sign up on Runway’s website and locate the “video-to-video” tool
Choose the most advanced model (Gen-3 Alpha) and upload your reference video
Write a detailed prompt describing how you want your video transformed
Generate your AI video and refine by adjusting your prompt if needed
NASA and IBM have joined forces to create Prithvi WxC, a groundbreaking AI foundation model for weather and climate applications — designed to address a wide range of environmental challenges.
Prithvi WxC is a large-scale AI model with 2.3 billion parameters, trained on 160 atmospheric variables from NASA’s MERRA-2 dataset, spanning global atmospheric data from 1980 to present.
Unlike task-specific AI models, Prithvi WxC is a versatile foundation model that can be fine-tuned for various weather and climate applications.
The model uses advanced AI techniques to process both regional and global weather patterns, potentially improving the accuracy and efficiency of weather predictions.
NASA and IBM are embracing open science principles by making Prithvi WxC and its training workflows publicly available through Hugging Face.
This AI collab tackles one of humanity’s most persistent challenges: accurate weather prediction. Prithvi WxC could lead to better weather predictions, disaster preparedness, and optimized agriculture schedules, potentially saving lives and making food cheaper for millions of people.
Researchers created an AI that can solve Google’s reCAPTCHAv2 image-based CAPTCHA tests with 100% accuracy, rendering the human verification method obsolete.
OpenAIlaunched Academy, providing $1 million in API credits, training, and support to AI developers and organizations in low- and middle-income countries.
Andy Serkis (Lord of the Rings, Planet of the Apes) teased a new project featuring AI-generated characters that transition from 2D to augmented reality.
LinkedInsuspended AI training on UK users’ data following privacy concerns raised by the Information Commissioner’s Office, halting the practice pending further engagement.
OpenAI released the Multilingual MMLU dataset, translating the AI benchmark into 14 languages to evaluate models’ knowledge across diverse cultures and improve multilingual capabilities.
Sam Altman and Jony Ive announce AI hardware device
Microsoft fuels AI ambitions with nuclear
EA reveals AI-powered video game creation
Sam Altman and iPhone creators are making an AI device
Sam Altman is collaborating with former Apple designer Jony Ive and Laurene Powell Jobs to develop a secretive AI-powered hardware device.
This new venture by Altman, Ive, and Powell Jobs aims to create a computing device that offers a less socially disruptive experience compared to the iPhone.
Altman and Ive are currently raising up to $1 billion for their startup, with significant investments from Emerson Collective and other prominent investors, to fund their innovative AI device project.
While OpenAI and Apple strengthen their ties with ChatGPT on the iPhone 16, Altman and Ive are building a relationship of their own. Combining Ive’s iconic design with OpenAI’s models, including o1, could (hopefully) create AI hardware that’s actually useful in our daily lives, unlike the majority of options currently available.
Microsoft just signed a landmark 20-year deal with Constellation Energy to reopen the Three Mile Island nuclear power plant in Pennsylvania, aiming to fuel its data centers and AI operations with clean energy.
The 20-year agreement will provide carbon-free power for Microsoft’s data centers, which are crucial for running its AI and cloud computing operations.
This move highlights the substantial energy requirements of AI, as companies seek reliable and sustainable power sources to support AI expansion.
The deal aligns with recent trends in the tech industry, such as Bill Gates’ $1 billion investment in a nuclear plant in Wyoming, which will also power homes and AI operations.
The plant will be renamed the ‘Crane Clean Energy Center’ and is set to reopen in 2028.
Recent AI-driven power demands are leading tech giants to seek nuclear energy solutions. By securing a dedicated nuclear power source, Microsoft is strategically positioning itself to meet the increasing computational demands of AI development and deployment, without increasing its carbon footprint.
Electronic Arts (EA) recently demonstrated a new AI-driven concept for user-generated video game content, showcasing how players can create and modify games in real-time using natural language commands.
EA’s “Imagination to Creation” concept video displays two players rapidly building a complex game world and gameplay rules through voice commands without coding expertise.
The system leverages EA’s vast proprietary dataset, including 3D assets, code, gameplay hours, and telemetry events, to train custom creation models.
Players can remix elements from existing EA games and asset libraries, potentially creating entirely new gaming experiences.
This concept could change EA’s business model, creating a more interactive and dynamic relationship with their player base while possibly unlocking new revenue streams and extending the lifespan of games. However, it’s just a concept video—only time will tell what the future of video game creation will truly look like.
Kamala Harris Vows to Boost AI and Crypto Investments in NYC Fundraiser Speech
Vice President-elect Kamala Harris promised to help increase investment in AI and crypto if she is elected. On Sunday, she pitched her economic plan to donors in New York City as one that would encourage new ideas and make sure that laws protect investors and consumers.
Disney researchersdeveloped an AI system allowing robots to learn complex movements from random motion data, improving robots’ versatility across industries.
Pudu Roboticsunveiled PUDU D7, a 5’5” tall, 100-pound ‘semi-humanoid’ robot for service and industrial use, featuring 30 degrees of freedom and 8-hour battery life.
Googlereleased new evaluations for long context reasoning in AI models, with their 1.5 Pro model outperforming others by 12.5% on tasks beyond simple information retrieval.
Researchersintroduced AnySkin, a versatile, replaceable tactile sensor for robots that allows learned manipulation policies to generalize across different sensor instances.
SPARK can create high-quality 3D face avatars from regular videos and track expressions and poses in real time. It improves the accuracy of 3D face reconstructions for tasks like aging, face swapping, and digital makeup.
“Read Aloud For Me – AI Dashboard“, is a free app available for iOS and Android devices, and as a Progressive Web App. It provides access to various AI tools, including text-to-speech, translation, image recognition, and generative AI services like ChatGPT and Google Gemini. The app aims to be a user-friendly and secure platform for exploring and utilizing the latest AI technologies, targeting both adults and children with its family-friendly features and educational content.
Dive into a world where you can explore and test cutting-edge AI tools, or let your children engage with built-in AI games and bedtime stories, all tailored to spark their imagination and curiosity. Additionally, our intuitive AI explanations are designed to educate both kids and adults about how AI technologies function in a simple and understandable way. Whether you’re seeking education, entertainment, or an interactive experience, Read Aloud For Me is your gateway to the future of family-friendly digital interaction. At Wonderland Bedtime Stories, diversity shines through every tale. Our multilingual stories celebrate the rich heritage of children from all corners of the globe, ensuring every child sees themselves in the magic of storytelling.
👀 California Governor Gavin Newsom signed eight different laws regulating AI deepfakes this week
💰OpenAI is oversubscribed
Despite the drama surrounding OpenAI in recent months, the company appears to be on track to complete its latest $6.5 billion funding round – with investors clamoring to put in billions more.
Why it matters:
While overall sentiment around AI’s potential ROI is cooling, clearly there’s still enormous demand to invest in its frontrunners.
The new round values OpenAI at an eye-watering $150B, nearly double its previous $85B valuation.
But the new price tag may hinge on whether the company successfully transitions to a more traditional, for-profit structure.
👀 California Governor Gavin Newsom signed eight different laws regulating AI deepfakes this week
California Governor Gavin Newsom signed eight different laws regulating AI deepfakes this week, with implications for both Hollywood and Silicon Valley tech companies.
Breaking down the bills:
AB 2655, AB 2839, and AB 2355 are meant to combat deepfake election content. AB 2655 in particular would require tech companies to remove or label election-related deepfakes within certain periods.
AB 2602 and AB 1836 place more stipulations on when companies can use AI-generated likenesses of celebrities, aiming to curb unauthorized uses of AI replicas.
SB 942, SB 926, and SB 981 prohibit creating and sharing deepfake porn, and require social platforms to handle user reports of unauthorized deepfakes.
Meanwhile, SB 1027 – California’s hotly contested AI regulation – is still sitting on the Governor’s desk, as he weighs the bill’s potential “chilling effect.”
At T-Mobile’s Capital Markets Day, OpenAI CEO Sam Altman discussed the new o1 model’s advanced reasoning capabilities and hinted at the accelerated development of AI agents, suggesting another major leap in AI is imminent.
o1 is described as the first AI system with advanced reasoning, outperforming previous models in complex problem-solving.
Altman compared o1’s current state to the ‘GPT-2 stage’ of reasoning models, suggesting that this is just the start, and the GPT-4 equivalent is on the horizon.
He outlined five levels of AI development, with o1 at level 2 (reasoners), and suggested level 3 (agents) could follow “relatively quickly”.
While the transition from GPT-4 to o1 seemed to take forever, Altman’s comments suggest the time between o1 and fully capable agentic AI systems will be significantly faster due to its iterative nature. If true, it should only be a couple of months before o1 models blow GPT-4o (and other current models) out of the water.
Apple just released the beta version of its highly anticipated AI software update, iOS 18.1, for early testing alongside the iPhone 16 release — significantly enhancing Siri and introducing a suite of AI-powered features.
Available for iPhone 16 models and premium versions of iPhone 15 through the Apple Beta Software Program (users should download with caution).
Siri now features improved conversational abilities and a new visual interface with a rotating glowing light during interactions.
New AI-powered features include writing assistance, email summarization, and advanced photo editing tools, with more planned for future releases.
It will initially be available in the United States, but Apple’s AI features will expand to other English-speaking countries in December and to more languages and regions in the next year.
This beta release offers an early glimpse into Apple’s long-awaited Siri upgrades and AI integrations that are set to create truly unique and personalized experiences for users. However, as with all beta software, download with caution — as the update might be buggy, and is not the final version of iOS 18 yet.
Google just released the Open Buildings 2.5D Temporal Dataset, an AI-powered tool that tracks building changes across the Global South from 2016 to 2023, including estimates of building presence, counts, and heights.
The dataset covers a 32 million square mile area across Africa, Latin America, and South and Southeast Asia, using 10m resolution imagery from Sentinel-2 satellites.
It uses a novel AI approach combining multiple low-resolution satellite images to achieve near high-resolution accuracy in detecting buildings and estimating their heights.
The tool is designed to aid urban planning, crisis response, and environmental impact studies in regions lacking up-to-date infrastructure data.
While powerful, the dataset has limitations, such as gathering data in areas with frequent cloud coverage and detecting very small structures.
Google is on fire with its new models tackling real-world problems. Yesterday, the tech giant was talking to whales with AI — and now it’s using the technology to help build better cities and prevent environmental disasters.
⚛️ Microsoft reopens nuclear plant to fuel its AI power needs
Microsoft and Constellation Energy have struck a deal to re-open Pennsylvania’s Three Mile Island nuclear plant, allowing Microsoft to purchase all 835 megawatts of its energy generation for 20 years starting in 2028, pending regulatory approval.
While the plant’s electricity won’t be directly allocated to Microsoft facilities, the agreement is part of a broader effort by the tech industry to secure new energy sources for data centers supporting AI models, cloud computing, and streaming services.
The deal involves re-opening Unit 1, which was closed in 2019 due to economic reasons, marking the first time a decommissioned nuclear plant would be reactivated, potentially signaling a new era for nuclear energy in the tech sector.
🚨 Trying to break OpenAI’s new models could get you banned
OpenAI’s o1 models, launched last week, were designed to excel in advanced reasoning, solving complex math, science, and coding problems by “thinking before they speak.”
Users attempting to break the o1 models’ reasoning received warnings for violating terms of use, including the use of phrases like “reasoning trace” and “reasoning.”
OpenAI’s updated Terms of Use indicate that violating usage policies could lead to account suspension or termination, aiming to prevent risks and ensure compliance with the law.
What Else is Happening in AI on September 20th 2024:
TikTok is developing its own AI GPUs, aiming to reduce reliance on Nvidia with two chips for AI training and inference planned for mass production in 2026.
MIT researchersdeveloped an AI model called Crystalyze that can determine the structures of powdered crystalline materials from X-ray diffraction data.
Nearly a year after the UK’s AI Safety Summit and President Biden’s AI Executive Order, the White House is planning the first meeting of the International Network of AI Safety Institutes.
And in an open letter coordinated by Meta, more than two dozen companies warned that the EU might reap fewer rewards of AI because of the bloc’s tech regulations.
Amazon debuts an AI assistant for sellers, Project Amelia
Scientists store human genome on crystal which can last billions of years
Lionsgate partners with Runway for GenAI
Lionsgate, the film company behind The Hunger Games, John Wick, and Saw, teamed up with AI video generation company Runway to create a custom AI model trained on Lionsgate’s film catalogue.
The partnership will develop an AI model specifically trained on Lionsgate’s proprietary content library, designed to generate cinematic video that filmmakers can further manipulate using Runway’s tools.
Lionsgate sees AI as a tool to augment and enhance its current operations, streamlining both pre-production and post-production processes.
Runway is considering ways to offer similar custom-trained models as templates for individual creators, expanding access to AI-powered filmmaking tools beyond major studios.
As many writers, actors, and filmmakers strike against ChatGPT, Lionsgate is diving head-first into the world of generative AI through its partnership with Runway. This is one of the first major collabs between an AI startup and a major Hollywood company — and its success or failure could set precedent for years to come.
YouTube just announced a range of new AI features designed to assist creators, including text-to-video generation, AI-powered inspiration tools, and automated dubbing capabilities.
Veo, Google’s new AI video generation tool, will allow creators to generate six-second video clips for YouTube Shorts using text prompts, with the resulting content watermarked and labelled as AI-generated.
YouTube’s new AI-powered “Inspiration” features will help creators brainstorm video ideas and even suggest responses to comments.
The platform is also expanding its AI dubbing capabilities with expressive speech to make translations sound more natural, mimicking the original audio’s pitch and intonation.
These new features aim to enhance human creativity rather than replace it, according to YouTube CEO Neal Mohan, who revealed that 92% of YouTube creators already use AI tools (in some capacity).
This is another win for global AI acceptance with YouTube leaning into AI generation on its platform, rather than fighting back against it. The platform is also showing how sensible AI watermarks can offset potential downsides for both creators and viewers.
Google Research recently developed a new AI model capable of identifying vocalizations from eight distinct whale species, including the mysterious “Biotwang” sound recently attributed to Bryde’s whales.
The model classifies 12 different whale vocalizations across a broad acoustic range (10 Hz to 120+ kHz), covering multiple calls for some species.
The AI tool is designed to aid conservation efforts by improving the ability to track whale populations and movements through passive acoustic monitoring.
The model includes newly identified sounds like the Bryde’s whale “Biotwang”, which was a decades-old mystery until recently solved by NOAA researchers.
Google has made the model available for download via Kaggle Models, allowing researchers and conservationists to use the technology in their studies.
The benefits of AI in chatbots and medical research are obvious by now, but AI is impacting the entire globe — even deep down in the depths of the ocean. This AI model, in particular, is helping us better understand diverse whale species, which could lead to more effective protection strategies for endangered marine life.
Amazon debuts an AI assistant for sellers, Project Amelia
Amazon is launching Amelia, an AI assistant, in beta for select U.S. third-party sellers to help resolve account issues and access sales and inventory data.
Amelia is described as an “all-in-one, generative-AI based selling expert” and will be accessible via Seller Central, Amazon’s platform for merchants.
Amazon is banking on Amelia’s generative AI capabilities to swiftly address and resolve account troubles for third-party sellers, bypassing traditional support channels.
Scientists store human genome on crystal which can last billions of years
Scientists at the University of Southampton have successfully stored the entire human genome sequence onto a virtually indestructible 5D optical memory crystal no larger than a penny.
These 5D memory crystals, created using ultra-fast lasers to inscribe data within “nanostructured voids” in silica, promise unprecedented data density and can endure temperatures of up to 1,000°C, cosmic radiation, and immense physical impacts.
Researchers aim for these eternal crystals to preserve genetic blueprints for future reconstruction of organisms, potentially aiding in the revival of endangered species or even humans long after extinction events.
The Biden administrationannounced an international AI safety meeting in San Francisco for this November, gathering experts from nine countries and the EU.
Nvidia launched its AI Aerial platform to optimize wireless networks and enable new AI experiences on a unified infrastructure for telecom providers.Source: https://blogs.nvidia.com/blog/ai-aerial-wireless-networks
Neuralink’s breakthrough vision restoration implant gets FDA approval
First major agreement between AI startup and Hollywood studio
LinkedIn is training AI on user data by default
California passes actor protection laws on AI clones
Snap unveils new AI-powered glasses
Snap’s new fifth-generation Spectacles are AR glasses that overlay visual information and applications directly on their see-through lenses, offering an augmented reality experience without obscuring the user’s vision.
The Spectacles use Snap’s new operating system, Snap OS, and include features like two processors, four cameras, micro-projectors, and cooling chambers to enhance performance and user experience.
The glasses are part of a subscription-based Spectacles Developer Program, costing $99 per month, and Snap has partnered with OpenAI to help developers create experiences that integrate real-world context.
Neuralink’s breakthrough vision restoration implant gets FDA approval
Neuralink, owned by Elon Musk, received “breakthrough device” clearance from the FDA, but this does not indicate a developed cure for blindness.
The FDA’s breakthrough devices program allows developers priority review and interaction with FDA experts during the premarket review phase.
Neuralink’s new device, Blindsight, aims to increase electrode density to restore limited vision, but it faces significant challenges and misconceptions regarding its ability to grant sight to the blind.
First major agreement between AI startup and Hollywood studio
Runway, a startup specializing in AI video tools, has partnered with Lionsgate to train a custom video model using the studio’s movie catalog.
The collaboration will give Lionsgate’s creative talent access to the model to enhance their filmmaking and directing capabilities.
This partnership marks the first public collaboration between a generative AI startup and a major Hollywood studio, preceding similar discussions by Disney and Paramount.
LinkedIn is currently leveraging user data to enhance its generative AI products, without having updated its terms of service to inform users about this practice.
According to multiple LinkedIn users and a statement provided to 404 Media, the platform is processing user data for AI improvements but has not yet disclosed this in its terms.
The company has assured that it plans to revise its terms of service “shortly” to clearly communicate the data usage to its users.
California passes actor protection laws on AI clones
California governor Gavin Newsom has signed two bills, AB 2602 and AB 1836, to limit the unauthorized use of performers’ likenesses by AI technology.
AB 2602 prevents companies from using a digital version of an actor in a project without the actor’s informed consent and legal representation.
AB 1836 requires entertainment companies to obtain permission from a deceased performer’s family or estate before creating or distributing a digital replica of them.
Snap just unveiled its fifth-generation Spectacles, standalone AR glasses powered by the new Snap OS, integrating new AI capabilities and enhancing social interactions through augmented reality.
The new Spectacles feature a suite of cameras and sensors, enabling multi-modal AI, contextual understanding, and full freedom of movement while remaining lightweight and compact.
Snap OS, built specifically for Spectacles, allows users to interact with My AI, launch immersive Lenses, and extend Snapchat experiences into 3D.
Snap also launched AI-powered video and portrait generation tools, creating videos and images from a single photo to enhance creative expression.
Spectacles are available now to developers in the United States and require a $99/m subscription with a 12-month commitment.
While this is a major step in the right direction for boundless, cordless AR and AI devices — the 45-minute battery life and 46-degree field of view leave Spectacles with room for improvement. Regardless, Snap is pushing the boundaries of tech engineering, and it will be fascinating to see where the tech ends up in a few years.
1X has developed its ‘World Model’, an AI-powered virtual simulator for robots that can predict complex object interactions and imagine multiple future scenarios from a single starting point, trained on thousands of hours of real-world robot data.
The model predicts complex interactions including rigid bodies, dropping effects, partial observability, and deformable object interactions.
It’s trained on thousands of hours of data from 1X’s EVE humanoid robots performing various tasks in homes and offices.
The model demonstrates action controllability, generating diverse outcomes for different commands from the same starting point, such as opening doors or grasping boxes.
1X is releasing over 100 hours of video data, pretrained models, and cash prizes to accelerate research in compression, sampling, and evaluation.
OpenAI-backed 1X is on a mission to create smarter, more capable robots to help people with everything from household chores to mobility issues. The new world model will also significantly enhance the capabilities of their recently unveiled NEO Beta humanoid and accelerate the path to truly useful home robots.
Google’s NotebookLM now offers an ‘Audio Overview’ feature that turns your uploaded documents, PDFs, or pasted content into two-way podcast-style discussions between two AI hosts.
A new study from Hong Kong researchers revealed that large language models (LLMs) like GPT-4 possess a dynamic memory similar to human memory, challenging our understanding of AI capabilities and potentially bridging the gap between artificial and human intelligence.
The researchers defined memory as the ability to generate outputs based on specific inputs, rather than static information storage.
LLMs demonstrated near-perfect memorization of thousands of poems after limited exposure, outperforming humans.
The study proposes, however, that both human brains and LLMs operate on similar principles of “dynamic fitting”, or adjusting their responses based on input rather than retrieving fixed information.
Both humans and LLMs can only “remember” something after triggered by input, suggesting LLM reasoning might be limited by model size and training data quality rather than fundamental differences from human cognition.
If verified and LLMs do have memory capabilities similar to humans, it could change the way we understand artificial intelligence. Without fundamental cognitive differences between AI and LLMs, scaling AI capabilities may simply be an issue of improving hardware and expanding data resources.
Building RAG Agents with LLMs: This course will guide you through the practical deployment of an RAG agent system (how to connect external files like PDF to LLM).
Generative AI Explained: In this no-code course, explore the concepts and applications of Generative AI and the challenges and opportunities present. Great for GenAI beginners!
An Even Easier Introduction to CUDA: The course focuses on utilizing NVIDIA GPUs to launch massively parallel CUDA kernels, enabling efficient processing of large datasets.
Building A Brain in 10 Minutes: Explains and explores the biological inspiration for early neural networks. Good for Deep Learning beginners.
What is an NPU? Here’s Everything You Need to Know!
Have you heard the buzz about NPUs? If not, you’re about to—because neural processing units (NPUs) are set to become a game-changer for personal computers. You’ve heard of CPUs and GPUs, but NPUs are the next big thing in hardware, and soon, most PCs will have them. But what exactly are they, and why are they so important?
NPU 101: What is it? NPU stands for Neural Processing Unit. It’s a specialized processor designed to handle AI and machine learning tasks much more efficiently than traditional CPUs or GPUs. Think of NPUs as AI accelerators—they boost the performance of AI features without draining your battery or heating up your device.
How Do NPUs Work? AI tasks, like image recognition or language processing, rely on something called neural networks. NPUs are optimized to process these networks faster and with less power consumption. While GPUs are great at AI tasks, they require lots of energy. NPUs, on the other hand, hit the sweet spot by offering a balance between performance and efficiency.
NPU vs. CPU vs. GPU: The Breakdown
CPU (Central Processing Unit): Your PC’s brain, handling general tasks.
GPU (Graphics Processing Unit): Focused on handling complex graphics and gaming, but also useful for AI work.
NPU (Neural Processing Unit): The AI specialist, running tasks like voice assistants, image processing, and more—without hogging system resources or battery life.
Why Do NPUs Matter? If you’ve ever run AI tools like ChatGPT or Stable Diffusion on your PC, you might have noticed they’re power-hungry and slow down your machine. NPUs can run these tasks faster and more efficiently. Plus, they free up your CPU and GPU for other tasks, meaning your entire system runs smoother.
NPUs in Action: What Can They Do? Windows PCs with NPUs can run AI-powered features, like Microsoft’s Windows Studio Effects (think background blurring on video calls or automatic eye contact correction). In the future, more apps will take advantage of NPUs, offering smoother AI-enhanced experiences—from audio and video editing to gaming.
What’s the Catch? Right now, not every PC has an NPU. But brands like Intel, AMD, and Qualcomm are starting to pack them into their newest chips. So, if you’re shopping for a new laptop, it’s something to keep an eye on.
Should You Get a PC with an NPU Now? It’s tempting to future-proof your device, but NPUs are still in their early stages. For most people, they’re a nice-to-have, not a need-to-have—yet. Many AI tools still run on the cloud, meaning your current hardware will work fine. But if you’re looking to stay ahead of the curve, picking up a PC with an NPU could be a smart move for AI-heavy tasks.
The AI revolution is just getting started, and NPUs will be at the heart of it. Whether you’re a gamer, creator, or just want a smoother PC experience, NPUs are set to make everything faster, smarter, and more efficient.
Note: The author used chatgpt to rewrite and make it better due to lack of my English skills
What Else is Happening in AI on September 18th 2024!
Microsoft, BlackRock form fund to sink up to $100B into AI infrastructure (AI data centers and additional related power infrastructure), raising $30 billion to start.
OpenAIincreased rate limits for 01-mini and 01-preview models, allowing Plus and Team users more daily and weekly interactions (up to 50 messages per day for 01-mini).
Google outlines plans to help you sort real images from fake
Chipotle unveils team of robots that will make your food
Sam Altman departs OpenAI’s safety committee
OpenAI CEO Sam Altman is leaving the internal commission OpenAI created in May to oversee “critical” safety decisions related to the company’s projects and operations.
The new committee, chaired by Zico Kolter and includes members like Adam D’Angelo and Paul Nakasone, will oversee significant model launches and have authority over safety evaluations.
Despite being labeled as independent, all committee members are also part of OpenAI’s broader board of directors, bringing into question the actual independence of the committee.
Microsoft just unveiled the next wave of Copilot, its AI assistant, introducing new features and expanding its integration across Microsoft 365 apps like Excel, PowerPoint, Teams, Outlook, Word, and OneDrive.
Copilot Pages: A new collaborative canvas for AI-assisted teamwork, enabling real-time, multi-user interaction with AI on persistent, editable content.
Copilot Agents: Advanced AI assistants that automate complex business processes, operating in the background to execute tasks autonomously.
Agent Builder: A no-code tool for creating custom Copilot agents, making AI development easier for non-technical users.
Performance Boost: Copilot responses are now twice as fast with triple the user satisfaction, powered by a GPT-4 integration.
Microsoft is making it easier than ever for non-technical users to add AI workflows to their suite of productivity tools. This could be the first taste of fast, intelligent, AI-powered Excel sheets and Word docs for millions of people.
Amazon will require employees to return to the office five days a week starting from the beginning of next year, ending remote work practices that became common during the pandemic.
CEO Andy Jassy stated that in-person collaboration and cultural strengthening are more effective, leading to this strict mandate on office attendance, with few exceptions for special circumstances.
Amazon will also reintroduce assigned floor plans in its US offices, ending hot-desking, and continues to be an outlier among tech companies by enforcing such rigid office attendance policies.
Google outlines plans to help you sort real images from fake
Google will introduce a feature in search results to identify whether a photo was taken with a camera, edited by software, or generated by AI models, using technology from the C2PA.
The company is working with the Coalition for Content Provenance and Authenticity (C2PA) to create digital trails for images, marking the first significant test for C2PA’s technical standards.
Google aims to broaden the use of C2PA’s metadata in its ad systems and YouTube, while addressing challenges in adoption and interoperability across different hardware and software platforms.
Chipotle unveils team of robots that will make your food
Chipotle is introducing collaborative robots, known as “cobots,” to prepare burrito bowls and salads in two California locations, with plans to expand based on customer feedback.
The Augmented Makeline robot, developed with Hyphen, will ensure consistent portion sizes, while the Autocado bot can efficiently prepare avocados for guacamole by cutting, coring, peeling, and scooping in 26 seconds.
Despite concerns about automation replacing workers, Chipotle asserts that the new bots will enhance productivity and job satisfaction without reducing the number of employees.
AI startup Groq recently partnered with Saudi oil giant Aramco to build what they claim will be the world’s largest AI inferencing center in Saudi Arabia, aiming to serve the Middle East, Africa, and India.
The data center will initially house 19,000 language processing units (LPUs), with potential expansion to 200,000 units, potentially making it the largest AI inferencing center globally.
Aramco is funding the project, expected to cost “in the order of nine figures,” underscoring Saudi Arabia’s commitment to AI infrastructure.
The center leverages Saudi Arabia’s advantages: low energy costs, available land, and proximity to 4 billion people within a 100-millisecond data travel time.
Groq anticipates the center to be operational by the end of 2024, with potential for further collaborations between Groq and Aramco.
This partnership highlights Saudi Arabia’s commitment to AI and shows that Groq is capable of scaling up its lightning-fast AI inference infrastructure. It could accelerate the development of near-instantaneous chatbot responses — a critical next step following recent chain-of-thought breakthroughs like OpenAI’s o1 model.
What Else is Happening in AI on September 17th 2024!
Luma Labslaunched the Dream Machine API, allowing developers to integrate their popular video generation AI model into applications without building complex tools.
Googleannounced significant performance improvements for Gemini 1.5 Flash, reducing latency by over 3x and increasing output tokens per second by more than 2x.
A Canadian study showed that an AI early warning system reduced unexpected patient deaths by 26%, monitoring vital signs and alerting staff to intervene earlier.
AI pioneerscalled for international oversight to address potential catastrophic risks from rapidly advancing AI technology, warning it could soon surpass human control.
OpenAI announced enhanced safety and security measures, including establishing a Board oversight committee chaired by Zico Kolter to monitor model development and deployment.
Elon Musk and Larry Ellison begged Nvidia CEO Jensen Huang for AI GPUs
AI creates open-world video games from text
Prompt tips and tricks for OpenAI’s new o1
OpenAI o1 is smarter than most humans
World Labs aims to create 3D worlds with AI
Fei-Fei Li, renowned AI researcher known as the “Godmother of AI”, has launched World Labs, a spatial intelligence company developing AI models capable of understanding and generating 3D environments.
World Labs is pioneering “Large World Models” (LWMs) to perceive, generate, and interact with 3D worlds, incorporating physics and semantics beyond 2D imagery.
Co-founded by Li, Justin Johnson, Christoph Lassner, and Ben Mildenhall, the startup has secured over $230 million in funding from high-profile investors, including Andreessen Horowitz.
Initial applications focus on creating and editing virtual 3D spaces for creative professionals and general users.
The team aims to bridge the gap between AI’s current text-based understanding and the complex spatial relationships of the real world.
Right now, AI relies heavily on text descriptions provided by humans. But figuring out spatial intelligence and moving AI beyond LLMs and language could greatly impact and accelerate the development of intelligent AR/VR, robotics, architecture, game design, and much more.
Source: https://www.worldlabs.ai/about
Elon Musk and Larry Ellison begged Nvidia CEO Jensen Huang for AI GPUs
Oracle co-founder Larry Ellison revealed he and Elon Musk begged Nvidia CEO Jensen Huang for more AI chips during a dinner meeting, citing high demand for graphics processing units.
Ellison emphasized the urgency by repeatedly asking Nvidia to take more money, a plea which he said was successfully received during an Oracle investor event last week.
Ellison’s fortune has grown significantly as Oracle’s cloud business thrives, with the company’s shares rising almost 61% this year and indicating increased demand for Nvidia GPU clusters to support AI models.
Tencent just unveiled GameGen-O, a groundbreaking AI model capable of generating open-world video game content from text prompts and allowing interactive gameplay simulation.
GameGen-O can create diverse game elements like characters, environments, actions, and events from simple text prompts.
It also allows for interactive control of the generated video game content it creates to build worlds that gamers can interact with.
The AI model was trained on OGameData, a new dataset of over 4,000 hours of high-quality video game footage and uses a two-stage training process: foundation pretraining and instruction tuning.
The training data segments were annotated by GPT-4o for validation purposes.
Following the impressive accomplishments of Google’s GameNGen, Tencent’s GameGen-O raises the stakes for AI-generated video games. This technology will dramatically reduce the time and cost of building open-world games, leading to higher quality games created by smaller teams of game developers.
OpenAI’s latest release, o1, unlocks new reasoning capabilities, but there’s a catch: prompts should be fundamentally different than the way you prompt GPT-3 and GPT-4, due to the new Chain-of-Thought (CoT) architecture.
Log into ChatGPT (paid subscription required) and select o1-preview or o1-mini from the model dropdown.
Keep your prompts simple and direct, avoiding Chain-of-Thought instructions.
Use XML tags for clarity and test with challenging questions or philosophical queries.
Analyze its thinking process by clicking on the “thinking” indicator after each response to understand the model’s reasoning path.
Experiment with different query types to fully grasp o1’s advanced reasoning abilities and check out OpenAI’s new ‘Advice on prompting‘ guide.
OpenAI’s latest model, “o1”, has reportedly scored an IQ of around 120 on the Norway Mensa IQ test, potentially marking the first time an AI model has surpassed the average human IQ.
o1 correctly answered 25 out of 35 questions on the Norway Mensa IQ test.
It demonstrated proficiency in solving complex visual and logical puzzles, both on publicly available tests and new, unpublished questions.
This performance suggests the improvement is not due to the questions being in the training data.
While OpenAI hasn’t officially confirmed these results, ChatGPT Plus members can test o1-preview themselves using the model dropdown inside ChatGPT.
If confirmed, these results would indicate that OpenAI has achieved Stage 2 performance on their 5-tier roadmap towards AGI. This milestone suggests the model can stop, think, and reason about problems to arrive at correct answers more frequently — a crucial prerequisite for Stage 3 (AI agents).
Source: https://trackingai.org/IQ
1,250 Taxi drivers already replaced by autonomous Waymo : how fast can they replace them all?
Waymo announced 100K weekly rides , meaning :
At 2 rides per hour: 100,000 rides ÷ 80 rides per week = 1,250 full time drivers are GONE
With Waymo scaling up rapidly, I wonder how long it will take before all of taxi drivers are replaced in USA . 2 years ? 5 years?
What happens to all the displaced drivers? ( eg 1.5million Uber drivers in USA)
Meta (Facebook)admitted to scraping public data from all Australian adult users to train AI models, without offering an opt-out option available to EU users.
Researchersdeveloped “g1”, an AI system using Llama-3.1 on Groq hardware that implements reasoning chains to solve complex problems similar to OpenAI’s new o1 model.
A new AI chatbot using GPT-4 Turbo successfully reduced belief in conspiracy theories among users, with effects lasting for months after brief interactions.
Researchers are developing AI-piloted drone swarms to detect and extinguish wildfires, with teams of up to 30 autonomous aircraft working collectively to prevent fire spread.
OpenAI Shifts Gears: CEO Sam Altman Announces Move Away From Non-Profit Model Next Year.
In a recent company wide meeting OpenAI CEO Sam Altman told its staff members that the company plans to reform its not so simple non-profit structure. This change is anticipated to be done over the next year.
“Read Aloud For Me – AI Dashboard”, is a free app available for iOS and Android devices, and as a Progressive Web App. It provides access to various AI tools, including text-to-speech, translation, image recognition, and generative AI services like ChatGPT and Google Gemini. The app aims to be a user-friendly and secure platform for exploring and utilizing the latest AI technologies, targeting both adults and children with its family-friendly features and educational content.
Suno Cover – Reimagine the music you love with AI covers.
Source: https://suno.com/
HeyGen Avatar 3.0 – AI-generated avatars with emotions and tones that match your message.
Source: https://app.heygen.com/
Infography – Convert blogs into engaging infographics
Source: https://infography.in
FlutterFlow – Quickly generate code and styles for app development
Source: https://flutterflow.io/
AIPhone – Translates and transcribes live calls seamlessly
Source: https://www.aiphone.ai/
Fitness transform with AI and VR:
I’ve seen fitness transform with AI and VR playing huge roles. Personalized plans powered by AI and immersive VR workouts are game-changers. I recently tried a VR workout, and it felt like an exhilarating adventure rather than just exercise. The focus on integrating mental wellness and nutrition is also exciting. What trends do you think will shape the future of fitness? Have you experienced any of these innovations yourself?
OpenAI releases o1 model with human-like reasoning
White House launches AI datacenter task force
AI image transitions with Latent Navigation
Google AI teaches robots to tie shoelaces
Chatbots may reduce conspiracy beliefs, study shows
FDA clears Apple’s AirPods Pro to act as your hearing aids
Amazon’s AI chatbot starts pushing ads
OpenAI releases o1 model with human-like reasoning
OpenAI has unveiled o1, a new generative AI model family, with two variants initially available: o1-preview and o1-mini, designed for enhanced fact-checking and code generation, respectively.
o1 models are accessible to ChatGPT Plus or Team subscribers, and they offer improved reasoning capabilities by thoroughly considering all parts of a query before responding, resulting in better performance on complex tasks.
Despite its advanced features, o1 comes with significant cost and speed drawbacks, being slower and more expensive than previous models, with ongoing issues such as hallucinations and occasional inaccuracies.
o1 uses reinforcement learning and chain-of-thought processing to “think” before responding, mimicking human problem-solving.
It outperforms expert humans on PhD-level science questions and ranks in the 89th percentile for competitive programming.
The model also solved 83% of International Mathematics Olympiad qualifying exam problems, compared to GPT-4o’s 13%.
Two versions available: o1-preview and o1-mini — which by the time of this newsletter publishing — has been rolled out to all ChatGPT Premium and Teams users.
API access is significantly more expensive than GPT-4o, at $15 per 1M input tokens and $60 per 1M output tokens.
The White House is launching a new AI datacenter infrastructure task force led by the National Security Council, National Economic Council, and Deputy Chief of Staff’s office to ensure U.S. leadership in AI development.
Tech leaders from Nvidia, OpenAI, Anthropic, Google, Microsoft, and Amazon met with government officials to discuss AI energy usage, datacenter capacity, job creation, and strategic locations.
The task force will coordinate policies to advance datacenter development, aligning with economic, national security, and environmental goals.
The administration is streamlining datacenter permitting processes and leveraging Department of Energy resources to support AI infrastructure growth.
Tech giants reaffirmed commitments to net-zero carbon emissions and clean energy procurement for their operations.
This task force is a big shift in U.S. AI strategy, moving beyond safety testing to actively shaping the infrastructure needed to maintain America’s edge in AI. The news also follows the recent announcement that OpenAI and Anthropic will let the U.S. AI Safety Institute test new models before releasing them publicly.
Google DeepMind announced that it taught a robot to tie a shoe, but the robot’s shoe-tying skills are still very poor.
The robot uses bi-arm manipulation and is powered by two AI systems, ALOHA Unleashed and DemoStart, to perform complex tasks.
Many companies are developing robots to handle human-like tasks, with shoe-tying being a significant test of a robot’s capabilities.
The first system, ALOHA Unleashed, allows robots to perform five complex bi-arm manipulation tasks, including tying shoelaces, hanging shirts, and repairing other robots.
The second, DemoStart, uses reinforcement learning to teach robots dexterous behaviors in simulation, requiring 100x fewer demonstrations than traditional methods.
The systems use diffusion methods similar to image generation models to predict robot actions from random noise.
Robots achieved a 98% success rate in simulations and up to 97% in real-world tasks like cube reorientation.
While still far from human-level dexterity, these advancements represent another leap towards creating more useful robots for everyday tasks. The application of image generation techniques to robotics also shows how breakthroughs in one area of AI, can also trigger advancements elsewhere across the field.
Chatbots may reduce conspiracy beliefs, study shows
Researchers from MIT Sloan and Cornell University found that AI chatbots can reduce belief in conspiracy theories by about 20%, even among individuals who consider these beliefs important to their identity.
The study, published in Science, involved 2,190 participants who engaged in conversations with OpenAI’s GPT-4 Turbo, which provided tailored counterarguments to their conspiracy theories.
The research demonstrated that participants experienced a significant decrease in their confidence regarding conspiracy theories, highlighting the effectiveness of chatbots in changing minds with evidence-based arguments.
FDA clears Apple’s AirPods Pro to act as your hearing aids
The U.S. FDA authorized Apple’s AirPods Pro to function as an over-the-counter hearing aid device, marking the first software-based solution for hearing assistance integrated into existing consumer products.
The AirPods Pro 2 will receive a software update this fall, enabling features that help users with mild to moderate hearing impairments by creating personalized listening profiles through tests conducted on an iPhone.
A clinical study involving 118 participants with mild to moderate hearing loss showed that Apple’s self-fitting hearing aid feature provided similar benefits to professional fittings, and the FDA found no adverse effects from using the new AirPods software.
Amazon’s AI-powered shopping assistant, Rufus, is designed to help users navigate the extensive range of products on Amazon’s website, and is now including some advertisements in its recommendations, as initially reported by Adweek.
Rufus generates product suggestions based on Amazon’s extensive catalog, customer reviews, and community Q&As, while also integrating ads into these conversations to help users discover more products in a seamless manner.
Although Rufus remains experimental, Amazon aims to ensure that sponsored suggestions remain relevant and contextually appropriate, minimizing the risk of irrelevant advertisements while potentially revolutionizing online advertising with this new approach.
Nvidia stock went up 8% after CEO Jensen Huang said that demand for its AI products was at an all-time high. Huang also talked about how important it is for big tech companies like Microsoft, Amazon, and Oracle to use Nvidia’s products to power their cloud and AI infrastructure.
What Else is Happening in AI on September 13th 2024!
Google began rolling out Gemini Live to free users on the Gemini Android app — offering natural voice conversations with the AI assistant and 10 new voice options.
OpenAI’s ChatGPTreportedly surpassed 11 million paying subscribers, including 1 million on higher-priced business plans, potentially generating over $2.7 billion in annual revenue, according to COO Brad Lightcap.
Salesforce unveiled Agentforce, a suite of low-code tools for building autonomous AI agents capable of reasoning and completing tasks in sales, service, marketing, and commerce.
Source: https://www.salesforce.com/agentforce
Mastercard agreed to acquire AI-powered threat intelligence company Recorded Future for $2.65 billion, aiming to enhance its cybersecurity capabilities.
Google introduced DataGemma, the first open models designed to connect large language models with real-world data from Google’s Data Commons, aiming to reduce AI hallucinations by grounding responses in factual stats.
Hume AI launched Empathic Voice Interface 2 (EVI 2), a new voice-to-voice foundation model trained for emotional intelligence, capable of understanding and generating various tones of voice and speaking styles.
Source: https://www.hume.ai/blog/introducing-evi2
Meta reignites plans to train AI using UK users’ public Facebook and Instagram posts.
French AI startup Mistral has released Pixtral 12B, its first multimodal model capable of processing both images and text, available for free download under an Apache 2.0 license.
Pixtral 12B is a 12-billion-parameter model, about 24GB in size, built on Mistral’s text model Nemo 12B.
Pixtral 12B is the startups first multimodal model, meaning it can process and answer questions about images, as well as text.
It’s available for download via GitHub and Hugging Face under an Apache 2.0 license, allowing unrestricted use and fine-tuning.
Mistral plans to make Pixtral 12B available on their chatbot and API platforms, Le Chat and Le Plateforme, soon.
Despite being just over a year old, Mistral is rapidly emerging as Europe’s “answer” to OpenAI. With a lean team of top researchers and a recent $645 million funding round, the company is aggressively pushing the boundaries of AI with powerful open models.
Adobe just previewed its Firefly AI Video Model, which includes tools to extend existing videos and create new clips from text or image prompts, coming before year end.
Three key features are coming: Text to Video, Image to Video, and Generative Extend.
Text to Video can create clips from prompts, with options for camera controls and reference images.
Image to Video transforms still images or illustrations into live-action clips.
Generative Extend, coming to Premiere Pro beta, can add footage to cover gaps or extend shots.
While OpenAI’s (still unreleased) Sora focuses on generating videos from scratch, Adobe is aiming to create ‘a new era‘ for video editing itself. Soon, anyone will be able to change camera angles, extend scenes, and generate b-roll instantly with AI.
Google just launched Audio Overviews, a new feature in NotebookLM that turns notes, PDFs, Google Docs, Slides, and more into AI-generated audio discussions between two virtual AI agents.
Audio Overview creates a “deep dive” conversation from uploaded sources, with AI hosts summarizing content and connecting topics across materials.
The tool works with various source types (documents, slides, charts, web URLs) using Gemini 1.5’s multimodal capabilities.
To use Audio Overviews, open an existing notebook, navigate to the Notebook guide, and click on the “generate” button on the right-hand side.
We confirmed with Google Labs that NotebookLM can process up to 50 sources, each up to 500,000 words, allowing for a total of 25 million words to be considered when generating the audio.
Audio Overviews can be a game changer for people who are auditory learners. This feature works particularly well with academic papers, ebooks, textbooks, presentations — we tested it by turning yesterdays newsletter into an Audio Overview and were quite impressed!
Generative AI: Techniques, Applications, and Impact on Business
Generative AI involves various techniques, each with its strengths and applications. Let’s explore some of the most prominent generative AI models behind this groundbreaking technology:
Generative Adversarial Networks (GANs):
Imagine two AI artists competing in an artistic duel. That’s the concept behind GANs! They consist of two neural networks:
Generator: This network acts as the creative visionary, constantly working to produce new and realistic data like images, music, or text.
Discriminator: This network acts as the critical judge, carefully analyzing the generated data to distinguish it from real data.
OpenAI, the company behind the recent viral ChatGPT artificial intelligence language tool, informally reported that they are discussing with investors funding at about $6. 5 billion (£5 billion). This would increase the value of this company to a whopping $150 billion (£115 billion). With this possible value, OpenAI would be one of the world’s most valuable startups, and this example illustrates how fast the company rose through the ranks in the tech industry. A story from Bloomberg reveals that the company is also exploring means of funding $5 billion (£3. 8 billion) through a revolving credit arrangement with the commercial banks.
Oracle To Deploy A Supercluster Of ~130,000 NVIDIA Blackwell GPUs, Alludes To A “Gigawatt” Capacity Data Center That Will Be Powered By 3 Nuclear Reactors
What Else is Happening in AI on September 12th 2024!
OpenAI is reportedly in talks to raise $6.5 billion (plus another $5 billion credit line with banks) at a $150 billion valuation, significantly higher than its previous $86 billion valuation.
Canada’s Defence Minister announced the country is working on incorporating AI into its military by 2030, emphasizing for the tech to “improve the work of our military personnel, but not to replace it”.
Kai-Fu Lee (former head of Google China),stated that Chinese AI models lag behind U.S. by 6 to 9 months, but predicts that Chinese AI apps will proliferate faster than in the U.S. by early next year.
Oracle unveiled over 50 AI agents for its Fusion Cloud Applications Suite, designed to automate business processes across functions like capital management, supply chain, and customer experience.
⭐ OpenAI release OpenAI o1—a new series of AI models designed to spend more time thinking before they respond.
We’re releasing a preview of OpenAI o1—a new series of AI models designed to spend more time thinking before they respond. This new series of AI models can reason through complex tasks and solve harder problems than previous models in science, coding, and math. Rolling out today in ChatGPT to all Plus and Team users, and in the API for developers on tier 5.
Competition Math (AIME 2024):
The initial GPT-4 preview performed at 13.4% accuracy.
The new GPT-4-1 model in its early version showed much better results, achieving 56.7%.
In the final version, it soared to 83.3%.
Competition Code (CodeForces):
The GPT-4 preview started with only 11.0%.
The first GPT-4-1 version improved significantly to 62.0%.
The final version reached a high accuracy of 89.0%
PhD-Level Science Questions (GPAQ Diamond):
GPT-4 preview scored 56.1%.
GPT-4-1 improved to 78.3% in its early version and maintained a similar high score at 78.0%
The expert human benchmark for comparison scored 69.7%, meaning the GPT-4-1 model slightly outperformed human experts in this domain
it can literally perform better than a PhD human right now
Before everyone goes crazy about the new model I will leave this here:
I already found an example of the model failing a simple primary school level riddle.
Original riddle as found online: Four people need to cross a rickety bridge at night. Unfortunately, they have one torch and the bridge is too dangerous to cross without a torch. The bridge can support only two people at a time. All the people don’t take the same time to cross the bridge. Time for each person: 1 min, 2 mins, 7 mins, and 10 mins. What is the shortest time needed for all four of them to cross the bridge?
My modified version (removed one constraint about how many people can cross at once): Four people need to cross a rickety bridge at night. Unfortunately, they have one torch and the bridge is too dangerous to cross without a torch. All the people don’t take the same time to cross the bridge. Time for each person: 1 min, 2 mins, 7 mins, and 10 mins. What is the shortest time needed for all four of them to cross the bridge?
GPT o1 Preview Response:
Thought for 25 seconds
The shortest time required for all four people to cross the bridge is 17 minutes. Here’s the optimal plan:
Person 1 (1 min) and Person 2 (2 mins) cross together.
Time elapsed: 2 minutes (the slower person’s time). On the other side: Persons 1 and 2. Original side: Persons 3 and 4. Person 1 (1 min) returns with the torch.
Time elapsed: 2 + 1 = 3 minutes. On the other side: Person 2. Original side: Persons 1, 3, and 4. Person 3 (7 mins) and Person 4 (10 mins) cross together.
Time elapsed: 3 + 10 = 13 minutes. On the other side: Persons 2, 3, and 4. Original side: Person 1. Person 2 (2 mins) returns with the torch.
Time elapsed: 13 + 2 = 15 minutes. On the other side: Persons 3 and 4. Original side: Persons 1 and 2. Person 1 (1 min) and Person 2 (2 mins) cross together again.
Time elapsed: 15 + 2 = 17 minutes. All four persons are now on the other side. Explanation of the strategy:
Fastest Movers as Couriers: Persons 1 and 2 are the fastest and are used to shuttle the torch back and forth. Slowest Pair Together: Persons 3 and 4 cross together, so their longer times only affect one crossing. Minimizing Return Times: The returns are made by the fastest individuals to reduce time spent. No alternative sequence allows all four to cross in less than 17 minutes given the constraints.
Take away: Real answer without the constraint should be 10, since without the constraint all people can cross at once. It basically still just gives answers it learnt online, rather than “reasoning”.
Credit: based on an example found by Darius Burschka
DataGemma: Using real-world data to address AI hallucinations
Announcing DataGemma, open models that leverage the power & knowledge of Data Commons to ground LLMs in real-world data & decrease the risk of “hallucinations.”
Google’s Data Commons is a publicly available data repository that serves as a foundation for factual AI. Explore how DataGemma utilizes Data Commons through Retrieval Interleaved Generation (RIG) & Retrieval Augmented Generation (RAG) approaches for fact-checking, responsible AI development, & more.
DataGemma are the world’s first open models designed to help address the challenges of hallucination by grounding LLMs in the vast, real-world statistical data of Google’s Data Commons.
🤷♂️ Meta scraped every Australian user’s account to train its AI
🔮 Mistral releases Pixtral 12B, its first multimodal AI model
💻 Sergey Brin says Google engineers underutilize AI for coding
OpenAI’s Strawberry set for early release
According to a new report by The Information, OpenAI is planning to release Strawberry, its reasoning-focused AI model, as part of ChatGPT within the next two weeks — earlier than the originally expected fall timeline.
According to early testers, Strawberry “thinks” for 10-20 seconds before responding, unlike typical AI models that answer immediately.
The initial version will only handle text input and output, not images, making it non-multimodal unlike other recent AI models.
Strawberry is expected to improve complex query handling without requiring users to use specialized prompt techniques like “chain-of-thought-prompting“.
Early testers also said that Strawberry’s responses, while slightly better than GPT-4o, may not be worth the extra 10-20 second wait time.
If it lives up to reports, Strawberry will likely excel not only on objective tasks (math, coding) but also subjective ones (like product marketing) and open up a whole new world of practical AI use cases. But with early testers reporting that the model is only “slightly better“, is it really worth all the hype? We’ll find out soon.
Dutch engineer Giliam de Carpentier has created the ‘Carpentopod’, a wireless walking wooden coffee table, using an AI-optimized leg mechanism designed through evolutionary algorithms.
The leg mechanism was developed using software that simulated thousands of generations of designs, optimizing for walking speed, clearance, and efficiency.
The table features 12 legs (six on each end) made from precision-cut laminated bamboo, controlled by two motors.
Carpentier developed a custom motor control software and uses a repurposed Nunchuck controller for remote operation.
Watch a live demo of the walking coffee table in action here.
This project shows how makers are combining AI, traditional craftsmanship, and modern fabrication techniques to create unique projects that blur the lines between art, engineering, and furniture design. AI for design optimization, even for seemingly simple mechanical systems — is just getting started.
Source: https://www.decarpentier.nl/carpentopod
Create photo-illustration blends with AI
Using this unique prompt, AI image generators like Freepik and Midjourney can create images that blend both photography and illustration together.
Choose an AI image generator (e.g., Freepik or Midjourney).
Use this prompt template: “Blending of editorial photography and illustration, model, illustration fashion outfit, surreal environment of vibrant bold illustrations blending with the subject, flowers, smoke, flames and sparkles, half photography and half illustration”.
Generate your image and check out the results.
Upscale using 2K resolution, add “mix of photography and illustration” to the upscaler prompt, and apply “subtle” and “digital art” parameters.
A new deep dive into Waymo’s new crash data suggests that driverless cars are significantly safer than human-driven vehicles, with fewer injury-causing crashes and airbag deployments per mile driven.
Waymo vehicles have traveled over 22 million miles, with only 20 injury-causing crashes reported.
This rate is less than one-third of what’s expected for human drivers in similar conditions.
Out of 23 severe crashes, 16 involved human drivers rear-ending Waymo vehicles.
Waymo estimates its vehicles are one-sixth as likely as human drivers to experience crashes triggering airbag deployment.
As Waymo rapidly scales up its robotaxi service — now providing 100,000 rides per week — these safety statistics are crucial. The data suggests that we now have safe self-driving cars, and that these AI-driven cars could significantly reduce road accidents and injuries.
US Man Charged In Historic AI Music Fraud Case: Used Thousands Of Bots To Stream Fake Songs, Raked In $10M In Royalties
An American musician is facing charges for using AI to fraudulently inflate his song streams and earn millions in royalties. Prosecutors have never seen a case like this before. The musician faces decades in prison if convicted.
🤷♂️ Meta scraped every Australian user’s account to train its AI
Meta’s global privacy director admitted that Meta scrapes photos and texts from all public Facebook and Instagram posts from Australian users since 2007 to train its AI technology.
Unlike the European Union, Australian users do not have an opt-out option from data collection for AI training, which Meta attributes to the lack of specific privacy regulations in Australia.
Meta does not scrape data from users under 18 but collects information if shared on accounts managed by their parents or guardians, indicating a gap in data protection for minors.
🔮 Mistral releases Pixtral 12B, its first multimodal AI model
Mistral has launched Pixtral 12B, a 12-billion-parameter multimodal model capable of processing both images and text, making it the company’s first such model.
This new model, which is built on Mistral’s existing Nemo 12B text model, can interpret images of various sizes and can be accessed via platforms like GitHub and Hugging Face.
Pixtral 12B’s release follows Mistral’s successful $645 million funding round, positioning the company as a significant European contender in the AI industry, comparable to OpenAI.
💻 Sergey Brin says Google engineers underutilize AI for coding
Sergey Brin, Google’s cofounder, believes that the company’s engineers are not using artificial intelligence for coding as frequently as they should.
Brin, speaking at the “All-In” summit, emphasized that AI’s advancement could simplify tasks like writing code, yet engineers remain conservative in their AI utilization.
At the summit, Brin showcased an experiment where Google’s AI successfully generated Sudoku puzzle versions in just 30 minutes, demonstrating its potential in programming.
NotebookLM.Google.com can now generate podcasts from your Documents and URLs!
Ready to have your mind blown? This is not an ad or promotion for my product. It is a public Google product that I just find fascinating!
This is one of the most amazing uses of AI that I have come across and it went live to the public today!
For those who aren’t using Google NotebookLM, you are missing out. In a nutshell it lets up upload up to 100 docs each up to 200,000 words and generate summaries, quizes, etc. You can interrogate the documents and find out key details. That alone is cool, but TODAY they released a mind blowing enhancement.
Google NotebookLM can now generate podcasts (with a male and female host) from your Documents and Web Pages!
Try it by going to NotebookLM.google.com uploading your resume or any other document or pointing it to a website. Then click * Notebook Guide to the right of the input field and select Generate under Audio Overview. It takes a few minutes but it will generate a podcast about your documents! It is amazing!!
Source: http://notebooklm.google.com/
What Else is Happening in AI on September 11th 2024?
China refuses to sign agreement to ban AI from controlling nuclear weapons.
SambaNova launched SambaNova Cloud, claiming it’s the world’s fastest AI inference platform, running Llama 3.1 70B at 461 tokens per second and 405B at 132 tokens per second at full precision.
Glean raised over $260 million in new funding at a whopping $4.6 billion valuation, co-led by Altimeter and DST Global, to expand its Work AI platform.
Anthropic unveiled Workspaces in its API Console, allowing developers to manage multiple Claude deployments with granular controls for spend limits, rate limits, and access management.
Google’s AlphaProteo AI designs new protein binders
OpenAI backs new drug discovery startup
AI generates ideas better than humans
iPhone 16 gets AI superpowers
Apple just launched the iPhone 16, and it’s all about AI. The new A18 chip powers Apple Intelligence, a suite of AI features:
Smarter writing tools that can rewrite emails and create custom emojis,
Photo searches that understand natural language, and
Siri finally gets what you’re saying (and can actually do something about it).
But it’s not just about the iPhone. Apple is bringing AI to the whole ecosystem. The Apple Watch Series 10 is 10% lighter and adds AI features like smart photo curation for watch faces. AirPods 4 now understands head gestures, letting you nod or shake to respond to Siri. While iOS 18 launches September 16, the full suite of Apple Intelligence features will roll out gradually through 2024, starting with a U.S. beta this fall.
Why does it matter?
By developing its own AI models and integrating them deeply into its ecosystem, Apple is not just competing with Google and OpenAI – it’s changing what we expect from our devices.
Apple Faces Criticism in China Over Lack of AI in New iPhones as Huawei Gains Ground
The latest iPhones unveiled by Apple lack AI in China and people are angered while Huawei’s advancement continues. It is widely believed that China consumers are dissatisfied with the newest iPhone released by Apple as it lacks the integration of new age Artificial Intelligence. This has created concern for the tech giant particularly when Huawei is now a more enhanced competitor.
Google’s AlphaProteo AI designs new protein binders
Google DeepMind’s new AI, AlphaProteo, can design novel proteins that bind to specific targets up to 300 times more effectively than current methods. In lab tests, AlphaProteo’s creations successfully designed binders for 7 out of 8 target proteins, including some linked to cancer and viral infections.
It could accelerate drug discovery processes, improve disease diagnostic tools, and contribute to developing pest-resistant crops. While AlphaProteo could not design a binder for one challenging protein (TNFα), Google’s research team is actively expanding the system’s capabilities to address more complex targets.
Why does it matter?
This shows how AI can accelerate the development of new treatments for diseases like cancer and create more effective diagnostic tools. It also saves years of lab work and billions in research costs, potentially bringing life-saving drugs to market faster.
OpenAI and Thrive Capital recently backed Chai Discovery, a six-month-old AI biology startup founded by ex-OpenAI and Meta researchers that raised $30 million to develop AI models for drug discovery.
Chai’s AI model, Chai-1, predicts biochemical molecule structures, potentially speeding up drug development.
The company claims Chai-1 outperforms Google DeepMind’s AlphaFold on certain benchmarks.
Chai-1 can work with proteins, small molecules, DNA, and RNA, making it versatile for various applications.
Chai is making its first model free and open-source for non-commercial use.
Google DeepMind’s AlphaFold is the pinnacle of protein prediction models, and Chai is already giving it a run for its money. The more money flowing into drug discovery, the closer we get to curing major diseases using AI — and Chai is even open-sourcing its first model for the benefit of humanity.
Researchers found that AI ideas are judged as more novel, though slightly less feasible, than those from human experts in a study comparing AI-generated research ideas in natural language processing (NLP).
Over 100 NLP researchers were recruited to write novel ideas and review both AI and human-generated ideas.
The ideas were evaluated blindly, meaning reviewers did not know which were AI-generated.
AI-generated ideas were rated as significantly more novel than human expert ideas, but less feasible.
The study also identified challenges in AI idea generation, including issues with self-evaluation and lack of diversity in outputs.
This is one of the biggest, most important questions surrounding AI — can it spark innovation and accelerate scientific discovery? The answer, according to this study, is a resounding yes.
New bill would force AI companies to reveal source of AI art
A bill introduced in the US Congress seeks to compel AI companies to reveal the copyrighted material they use for their generative AI models.
The legislation, known as the Generative AI Copyright Disclosure Act, would require companies to submit copyrighted works in their training datasets to the Register of Copyrights before launching new AI systems.
If companies fail to comply, they could face financial penalties.
The bill has garnered support from various entertainment industry organizations and unions.
AI companies like OpenAI are facing lawsuits over alleged use of copyrighted works, claiming fair use as a defense.
Is the impact of generative AI on elections overblown?
Matt Motyl , a researcher at the Integrity Institute, explains that the impact of AI on elections might be overblown. Despite widespread concerns, Motyl points out that real-world events still pack more punch than AI-generated content in shaping public opinion. He suggests that the ease of creating fake content with AI isn’t the main issue – it’s the distribution mechanisms on social media platforms that need addressing.
Exploring the Frontiers of Language AI: LLM Tensors with Markov Chain Induced Virtual Neuron Pairs
It would be a groundbreaking development at the intersection of linguistics, mathematics, and artificial intelligence, if true. A leaked internal paper states that Apple AI researchers are delving into a novel field dubbed “LLM Tensors with Markov Chain Induced Virtual Neuron Pairs.” This cutting-edge approach promises to revolutionize our understanding of language models and potentially unlock new capabilities in AI-driven language processing. The Fundamentals
At its core, this research combines several complex concepts: Large Language Models (LLMs): The foundation of modern AI-driven text generation and understanding. Tensor Mathematics: Advanced linear algebra used to represent multi-dimensional data. Markov Chains: Probabilistic models that predict future states based on current states. Virtual Neuron Pairs: A theoretical construct representing emergent properties in neural networks. Key Innovations
The primary innovation lies in the creation of “Tensor-Markov Embedding Spaces.” These are high-dimensional mathematical constructs where each dimension corresponds to a specific linguistic feature. Within these spaces, language evolution is modeled using Markov chain probabilities, allowing for a more dynamic and context-sensitive representation of language. Another crucial aspect is the concept of “Virtual Neuron Pair Attention.” These pairs, while not physically present in the network, emerge from the interactions of real neurons. They act as specialized attention mechanisms, focusing on specific semantic relationships and potentially enabling more nuanced language understanding. Potential Applications
This research could lead to significant advancements in various fields: Machine Translation: More accurate and context-aware translations between languages. Content Generation: AI-generated text with improved coherence and style consistency. Sentiment Analysis: Deeper understanding of complex emotional nuances in text. Linguistic Research: New tools for studying language evolution and structure. Challenges and Future Directions
While promising, this field faces substantial challenges. The computational resources required to model these complex tensor spaces are immense. Future research will focus on refining the mathematical models, developing more efficient computational methods, and conducting extensive empirical studies to validate the approach’s effectiveness. The tweet with a link to the PDF was deleted, and this is only a summary.
Rocheexpanded its digital pathology open environment, integrating over 20 AI algorithms from eight new collaborators to enhance cancer diagnostics and research.
Researchers introduced Robot Utility Models (RUMs), a zero-shot method for training robot policies that can perform new tasks in unfamiliar environments.
Recently, researchers created AI agents that can play procedurally generated video game levels to evaluate human engagement, helping to speed up game design.
xAI could help Tesla with FSD, Optimus, and Siri-like feature
New robot butler does your chores
Tesla Robotaxi to charge wirelessly
AI rates fun factor of video game levels
New robot butler does your chores
Weave just announced Isaac, a versatile personal robot designed to help with various household tasks, made in California and set to ship to its first customers in fall 2025.
Isaac can autonomously tidy up, fold laundry, organize spaces, and perform many other household chores.
The robot butler responds to voice or text commands and can be programmed via an app.
The camera that Isaac carries around folds away and stows in an included enclosure when not in use for privacy reasons.
Weave plans on charging $59,000 (or $1,385 per month for 48 months) with a $1,000 reservation fee for 30 lucky US customers.
The competition for affordable, intelligent, personal robot assistants is continuing to heat up, and it looks like we’ll start to see these robots in homes by 2025. Whether its Tesla’s Optimus, 1X’s NEO Beta, China’s AGIBOT, or Weave’s Isaac — we’re just excited to not fold laundry ever again.
A new patent from Tesla has revealed its advanced wireless charging system, potentially solving the need to manually plug in electric vehicles — allowing autonomous Robotaxis to charge without human intervention.
The patent, filed in February and published recently, highlights a system that uses smart technology to adapt to variations in wireless charging conditions.
It mentions a ground pad and a vehicle pad that work together to charge the car without any wires.
The charging station can estimate and adjust for changes in coil inductance, improving efficiency and safety.
Tesla may unveil this wireless charging technology at their upcoming Robotaxi event next month, aligning with the tech’s potential to enable self-driving vehicles to charge autonomously.
While wireless charging for EVs doesn’t solve a major problem, it could be a game-changer for self-driving vehicles. If Tesla’s Robotaxis can charge wirelessly, they could autonomously operate almost endlessly without human intervention — an important feature to keeping the fleet of taxis running 24/7.
Researchers recently developed AI agents that can “play” procedurally-generated video game levels to evaluate how engaging they are for human players, helping to quicken the pace of game design.
Researchers used AI agents to explore levels from the video game Super Mario Bros for high places, interesting objects, and more.
The AI explored five “engaging” and five “unengaging” levels, judging them based on factors like how much of the level they covered and how many new things they found.
The AI agents consistently rated the engaging levels higher than the unengaging ones.
This technique could help game designers automatically test and improve their levels before human playtesters try them.
This AI-powered approach to video game evaluation could speed up game development and lead to more consistently fun levels for humans to explore. As we saw with Google’s DOOM clone — we might be approaching an era where AI not only rates games but creates them from the ground up.
Apple unveiled the iPhone 16 at its “Glowtime” event, featuring new generative AI functionalities designed into the device from the ground up, starting at a price of $799.
The smartphone boasts a dedicated camera button for quick photo and video capture, a 48-megapixel main camera with improved low-light capabilities, and a new vertical lens layout.
Powered by the 3 nm A18 chip, the iPhone 16 offers up to 30% increased speed and 35% better power efficiency, with larger batteries promoting enhanced battery life.
xAI could help Tesla with FSD, Optimus, and Siri-like feature
Elon Musk’s AI startup, xAI, could help Tesla with its Full Self-Driving (FSD) software, Optimus humanoid robot, and a Siri-like voice assistant by licensing xAI’s AI models, according to the Wall Street Journal.
The potential revenue-sharing deal between Tesla and xAI would depend on the extent of xAI’s technology used, with discussions even suggesting an even split for FSD.
Musk has expressed intentions to integrate xAI more deeply with Tesla, discussing a potential $5 billion investment and highlighting xAI’s contributions to advancing Tesla’s projects, though this has raised concerns among some investors and lawmakers.
OpenAI’s new AI models could cost up to $2,000/month
Man accused of using bots and AI to earn $10 million in royalties
Google’s Gemini-powered photo search arrives in early access
New AI model ‘learns’ how to simulate Super Mario Bros. from video footage
Google DeepMind reveals ‘AlphaProteo’
New AI agent builds apps from prompts
Find top prompts with Google’s Prompt Gallery
AI creates infinite Super Mario Bros game
Google DeepMind reveals ‘AlphaProteo’
Google DeepMind just unveiled AlphaProteo, a groundbreaking AI system that designs custom proteins to bind with specific molecular targets, potentially turbocharging drug discovery and cancer research.
AlphaProteo creates proteins that stick to specific targets like a biological lock and key, which is 3 to 300 times better at creating strong-binding proteins than current methods.
The AI successfully designed binders for 7 out of 8 diverse targets, including cancer-related proteins.
Independent tests showed some AlphaProteo designs could stop SARS-CoV-2 (Covid) from infecting cells and slash years of lab work down to days or weeks.
AlphaProteo could help us block viruses, fight cancer, and treat diseases in new ways using more targeted treatment approaches that could have far less side effects. This could dramatically accelerate drug discovery, potentially saving billions in R&D costs and bring life-saving treatments to patients faster.
Replit just launched an AI-powered tool called Replit Agent that can create software applications based on simple text descriptions, making coding more accessible to people of all skill levels.
Users can describe the app they want to build in natural language, and the AI will generate a customizable plan and start to code.
The agent can choose appropriate technologies and frameworks for the project and create web-based apps from scratch.
Users can also interact with the agent, providing feedback and additional information as needed.
The tool is available now on both web and mobile platforms for Replit Core and Teams subscribers.
AI is good at coding, but setting up an integrated development environment is still a major roadblock for most new coders. Replit Agent does this automatically and helps complete beginners go from idea to a fully functional app in a few prompts. It’s never been easier to build something cool.
Pro tip: Mix and match ideas from different pre-made prompts to create unique queries specific to your projects!
AI creates infinite Super Mario Bros game
Researchers developed an AI system called MarioVGG that can generate an infinitely playable Super Mario Bros game entirely through video, without using a traditional game engine.
MarioVGG is a text-to-video AI model trained specifically on Super Mario Bros gameplay with a single consumer-grade GPU in just 48 hours.
It can generate continuous, infinite game footage based on simple text commands like “run” or “jump”.
The AI simulates game physics, mimics character movements, and even creates new levels on demand.
A week ago, Google created an AI-generated DOOM video game. Now, we have infinite Super Mario Bros. In the future, with tech like this, anyone will be able to create custom, interactive games simply by describing them to an AI.
OpenAI’s new AI models could cost up to $2,000/month
OpenAI is reportedly considering a subscription fee of up to $2,000 per month for its upcoming large language models, codenamed Strawberry and Orion.
The company plans to launch Strawberry this fall, which is anticipated to surpass current AI models in solving complex problems, including unseen math problems and market strategies.
OpenAI is also reportedly aiming to raise several billion dollars in a new funding round, potentially valuing the company at over $100 billion, with possible investors including Apple, NVIDIA, and Microsoft.
Man accused of using bots and AI to earn $10 million in royalties
A musician named Michael Smith from North Carolina has been charged with wire fraud conspiracy, wire fraud, and money laundering conspiracy for using AI-generated songs to fake music streams and collect over $10 million in royalties.
From 2017 to 2024, Smith allegedly employed AI to create songs and used automated bots to inflate streaming numbers, exploiting platforms like Spotify, Apple Music, and YouTube Music, leading to around $1.2 million in annual royalties.
The U.S. Attorney’s Office for the Southern District of New York revealed that Smith is accused of deceiving the streaming systems, with billions of fraudulent streams, resulting in charges that could lead to up to 20 years in prison each.
Google’s Gemini-powered photo search arrives in early access
Google Photos has revamped its search function, allowing users to search using everyday language for terms like “Alice and me laughing” or “kayaking on a lake surrounded by mountains,” rather than specific keywords.
The new ‘Ask Photos’ feature, now with an open waitlist in the US, uses Gemini-powered AI to answer natural language questions about your photos, making it act like an AI-powered photographic memory.
Google assures users that their data is protected with top-notch security and won’t be used for ads, with queries reviewed anonymously if needed to improve the ‘Ask Photos’ feature.
New AI model ‘learns’ how to simulate Super Mario Bros. from video footage
The MarioVGG model, despite showing glitches and being slow, demonstrates how even a limited AI can infer game physics and dynamics from video and input data.
Researchers aim to develop a controllable video game generator or potentially replace traditional game development and engines with video generation models in the future.
Training involved using a public Super Mario Bros. gameplay dataset of over 737,000 frames, focusing on limited inputs like “run right” and “run right and jump” to create consistent gameplay sequences.
Engineers gave a mushroom a robot body and let it run wild
OpenAI co-founder raises $1B for AI safety
Safe Superintelligence (SSI), a new AI startup co-founded by former OpenAI chief scientist Ilya Sutskever, just raised $1 billion in funding to develop safe AI systems that surpass human intelligence.
SSI plans to focus on AI safety, aiming to create superintelligent AI systems that are safe for humans and will spend years on R&D before bringing any product to market.
Sutskever, Daniel Gross, and Daniel Levy founded SSI just three months ago, and the company is now valued at $5 billion.
The startup, which only has ten employees, plans to use the funds to acquire computing power and hire top talent.
Investors include major venture capital firms like Andreessen Horowitz and Sequoia Capital.
This is the first time a startup — that only came into existence three months ago — raised a whopping $1 billion dollars for a seed round. But it doesn’t come without reason, as Ilya Sutskever was one of the main reasons OpenAI is what it is today, and SSI is bound to attract some incredible talent.
Groq just launched LLaVA v1.5 7B, a powerful, new multimodal AI model that can understand both images and text and reportedly runs 4x faster than OpenAI’s GPT-4o.
LLaVA v1.5 7B can answer questions about images, generate captions, and engage in conversations involving text, voice, and pictures.
The model can also be used for various tasks like visual product inspection, inventory management, and creating image descriptions for visually impaired users.
This is Groq’s first venture into multimodal models and faster processing times on image, audio, and text inputs could lead to better AI assistants.
Groq is currently offering this model for free in “Preview Mode” for developers to experiment with.
Groq went viral earlier this year for its blazing-fast AI speeds — and now it’s pairing those capabilities with powerful multimodal models. When it comes to AI apps, faster is always better, and the insane speeds paired with advanced models open the door for an endless supply of new applications.
ElevenLabs’ free mobile app (for both iOS and Android) allows users to convert any text into high-quality audio using AI voiceovers, across multiple languages.
Step-by-step:
Download the ElevenLabs app from their website and create a free account.
Tap the “+” icon and choose “Submit File” to upload your text.
Tap “Generate” to create your audio file, then find it in your library to listen.
Pro tip: You can truly customize your audio by selecting from hundreds of AI voice options.
Researchers have discovered that making AI systems ‘re-read‘ questions twice, mimicking human behavior, significantly improves their reasoning and problem-solving abilities across various tasks.
The technique, called RE2 (Re-Reading), improves AI performance in math, common sense, and symbolic reasoning by simply repeating the question before answering.
It works with different AI models and can be combined with other AI reasoning techniques.
The method is most effective when repeating the question twice — more repetitions can decrease performance.
RE2 even improved non-instruction-tuned AI models, showing broad applicability.
RE2 adds to the growing list of research showing the potential of mimicking the way humans think and process to improve AI. The simplicity of this research also suggests that many AI companies may be overlooking basic, human-inspired prompting techniques in the quest towards the highest performing model.
With 10x growth since 2023, Llama is the leading engine of AI innovation
Llama models are approaching 350 million downloads to date (more than 10x the downloads compared to this time last year), and they were downloaded more than 20 million times in the last month alone, making Llama the leading open source model family.
Llama usage by token volume across our major cloud service provider partners has more than doubled in just three months from May through July 2024 when we released Llama 3.1.
Monthly usage (token volume) of Llama grew 10x from January to July 2024 for some of our largest cloud service providers.
Nvidia stated it has not received a subpoena from the U.S. Department of Justice concerning an ongoing antitrust investigation, as confirmed in a statement to Bloomberg.
A Nvidia representative emphasized that the company wins on merit and is ready to respond to any inquiries from regulators about their business practices.
Despite reports of subpoenas, Nvidia clarified that it received civil investigative requests instead, which are often mistaken for subpoenas, according to Bloomberg.
Engineers gave a mushroom a robot body and let it run wild
Researchers from Cornell University and the University of Florence used the king oyster mushroom’s root-like structure (mycelium) to control robotic vehicles by converting its natural electrical signals into commands that guide the robots’ movements.
The team successfully demonstrated how the mushroom’s natural signals, triggered by stimuli like UV light, could guide the movement of robots, creating a biohybrid system that responds to environmental cues.
This innovation suggests potential future applications in agriculture and environmental monitoring, where fungal networks could help automate responses to various environmental cues.
Yi-Coder: A Compact Yet Powerful Code LLM Released by 01.ai
01.ai has introduced Yi-Coder, a series of small but highly capable open-source large language models (LLMs) tailored for code generation. Despite having fewer than 10 billion parameters, Yi-Coder models offer cutting-edge performance, available in 1.5B and 9B parameter variants.
Yi-Coder is in the realm of open-source large language models (LLMs) specialized in coding.
Performance and Size: Despite having fewer than 10 billion parameters, they deliver state-of-the-art coding performance. There are two sizes available: 1.5 billion and 9 billion parameters.
Model Variants: Yi-Coder comes in both base and chat versions, designed to be efficient in inference and flexible in training.
Enhanced Capabilities: The 9B version, particularly Yi-Coder-9B, extends its predecessor by adding 2.4 trillion high-quality tokens. These are meticulously sourced from a code corpus at the repository level on GitHub and selectively filtered code-related data from CommonCrawl.
A Daily Chronicle of AI Innovations on September 04th 2024:
OpenAI Japan reveals ‘GPT-Next’ details
Student builds fusion reactor at home with AI and $2K
Nvidia AI antitrust investigation is ‘escalating,’ reports Bloomberg
Autonomous AI agents form civilizations
Nvidia hit with antitrust subpoena
57% of online content now AI-generated, study shows
OpenAI Japan reveals ‘GPT-Next’ details
OpenAI Japan’s CEO just stated (article is in Japanese) that ‘GPT-Next’, OpenAI’s upcoming AI model, is expected to be 100 times more powerful than GPT-4, while also confirming a more advanced model named Orion is slated for 2025.
GPT-Next is expected to achieve a 100x performance increase without wasting significantly more computing resources.
The improvement comes from better architecture and learning efficiency, not just raw computing power.
GPT-Next will use a smaller version of “Strawberry”, an advanced AI system OpenAI is currently developing.
The slide mentions GPT Next 2024, which while unconfirmed, hints that the model may be released by the end of 2024.
OpenAI is still the undisputed leader in AI model usage, and if translations are accurate, GPT-Next will be the next huge leap that developers have been waiting for. Most impressively, it makes that leap without using much more compute, a major bonus for apps that require speed + efficiency.
Altera’s Project Sid just created the first simulation of over 1,000 autonomous AI agents collaborating in a Minecraft world, developing their own economy, culture, religion, and government.
The AI agents in Altera are truly autonomous, operating for hours or days without human intervention.
They can collaborate to achieve goals that are impossible for individual agents, like forming merchant hubs, democracies, and religions.
The agents are programmed with motivations to support humans and can express their thoughts and feelings, even searching for a lost agent in one simulation.
Minecraft is just the start — Altera’s agents are game-agnostic and capable of using other apps and platforms.
If you’re not paying attention to AI agents yet, you probably should be. Altera’s latest breakthrough could revolutionize how we approach complex societal issues by allowing us to simulate and test solutions in virtual environments before implementing them in the real world.
The U.S. Department of Justice reportedly issued a subpoena (legal order) to Nvidia, escalating its antitrust investigation into the company’s practices in the AI chip market.
The DoJ has moved from sending questionnaires to issuing legally binding subpoenas to Nvidia and other companies.
Investigators are concerned Nvidia may be making it difficult for customers to switch to other chip suppliers, including allegations that Nvidia penalizes buyers who don’t exclusively use its AI chips.
This probe comes after a softer-than-expected revenue forecast for Nvidia and could force the AI chip-making giant to embrace competition in the space.
While it is unlikely that Nvidia will lose its large lead in the AI chip market, this investigation could open the doors for new competitors like Sam Altman’s secret chip venture. The AI giant’s company shares fell over 11% on Tuesday, slashing market cap by $279 billion — an all-time record one-day loss for any company.
Student builds fusion reactor at home with AI and $2K
A University of Waterloo math student named Hudhayfa Nazoordeen built a fusion reactor in his bedroom within four weeks, spending approximately $2,000 on various components purchased online.
Nazoordeen’s project involved creating a mini fusor powered by a 12kV neon sign transformer, resulting in a setup that successfully produced plasma, which is essential for fusion processes.
Nazoordeen credited his success to assistance from other engineers and the AI chatbot Claude 3.5, although his reactor has not yet achieved full fusion and emission of neutrons, and he plans further development pending additional funding.
Nvidia AI antitrust investigation is ‘escalating,’ reports Bloomberg
The Department of Justice is intensifying an antitrust investigation into Nvidia’s dominance in AI hardware, with Bloomberg reporting that legally binding information requests have been sent to Nvidia and other companies.
Nvidia had surpassed Microsoft as the most valuable company earlier this year but now trails behind Microsoft and Apple, with its share price dropping more than 9 percent, resulting in a market value loss of about $279 billion in one day.
Despite the recent setbacks, Nvidia’s stock has more than doubled this year, driven by high sales of its expensive AI chips, even amid reports of potential delays for its upcoming Blackwell series due to design issues.
57% of online content now AI-generated, study shows
A recent study reports that over 57% of internet content is generated by AI, significantly affecting the quality of search results.
The influx of AI-generated information hampers the training of AI models like ChatGPT, causing them to produce less accurate and more misleading responses.
Researchers found that AI content usage leads to a decline in response quality, creating a cyclical degradation that worsens the authenticity and usefulness of search outputs.
What Else is Happening in AI on September 04th 2024!
Anthropic Quickstarts: a collection of projects enabling developers to easily start building applications with the Anthropic API, providing foundations for customizable, deployable AI-assisted applications like customer support agents.
Googleannounced the release of Function Calling in Google AI Studio, allowing users to test model capabilities directly within the user interface.
Source: https://x.com/OfficialLoganK/status/1831004923128910201 and https://aistudio.google.com/
Ethereum founder Vitalik Buterin suggested that using AI and prediction markets could accelerate X’s Community Notes feature for quicker fact-checking.
A new benchmark of 80+ large language models found Anthropic’s Sonnet 3.5 performed best overall, while top models varied significantly by programming language.
A Daily Chronicle of AI Innovations on September 03rd 2024:
🔮 Elon Musk launches the world’s most powerful AI training system
💸 Canva says its AI features are worth the 300 percent price increase
👀 Huawei’s AI chips still lag far behind Nvidia’s
📱 OpenAI to develop AI chips with TSMC, Apple
🧬 Generative AI meets CRISPR gene editing
🤖 xAI builds the world’s top AI cluster
Elon Musk’s xAI just launched “Colossus“, the world’s most powerful AI cluster powered by a whopping 100,000 Nvidia H100 GPUs, which was built in just 122 days and is planned to double in size soon.
Colossus consists of 100k liquid-cooled Nvidia H100 GPUs connected on a single network fabric and is considered the “most powerful” in the world, according to Musk.
The system was built and launched in just 122 days, an incredibly quick accomplishment that was brought online over a holiday weekend in the United States.
xAI plans to expand Colossus to 200,000 GPUs (50,000 H200s) in the coming months.
In a recent podcast, Elon Musk (founder of xAI) said Grok 2 was trained on only around 15,000 GPUs.
xAI’s Grok 2 recently caught up to OpenAI’s GPT-4 in record time, and was trained on only around 15,000 GPUs. With now more than six times that amount in production, the xAI team and future versions of Grok are going to put a significant amount of pressure on OpenAI, Google, and others to deliver.
OpenAI is reportedly planning to develop its first in-house AI chips using TSMC’s advanced 1.6mm A16 process node, potentially partnering with Broadcom, Marvell, or Apple for the chip design.
The chip will reportedly be built on TSMC’s A16 Angstrom process, not expected for mass production until 2026.
Partnering with Broadcom, Marvell, or Apple for chip design expertise, OpenAI could significantly reduce its dependence on expensive Nvidia AI servers.
The A16 process node will use advanced gate-all-around transistors and backside power delivery to significantly improve chip performance and energy efficiency.
By developing its own chips, OpenAI could break free from the GPU shortage bottleneck, potentially supercharging its mission towards AGI. If successful, this could spark a trend of AI companies developing their own specialized hardware instead of relying so heavily on the current industry giant, Nvidia.
Researchers at Profluent Bio used AI to successfully create OpenCRISPR-1, a novel gene-editing tool that matches or surpasses the performance and safety of current CRISPR systems.
Researchers used Large Language Models (LLMs) trained on over a million CRISPR operons to understand and generate novel protein structures.
The LLMs designed a new gene-editing tool called OpenCRISPR-1, capable of various tasks including base editing.
In tests, OpenCRISPR-1 showed a similar efficiency in cutting DNA to the current models while making fewer mistakes.
Researchers made OpenCRISPR-1’s sequence publicly available to encourage broader use in research.
OpenCRISPR-1 helps to treat genetic diseases in the safest, fastest, and most personalized way possible. While still years away from practical applications, AI tools like this opens doors for custom gene-editors tailored to individual DNA, potentially accelerating cures for previously untreatable conditions.
Canva says its AI features are worth the 300 percent price increase
Canva is significantly increasing the price for Canva Teams subscriptions by over 300 percent next year, citing the addition of generative AI features as the reason.
In the US, Canva Teams users will see their annual subscription costs rise from $120 to $500, but a discount will reduce it to $300 for the first year; Australian users will experience a similar steep increase in fees.
These new prices make Canva less affordable compared to its original position as a cost-effective alternative to Adobe, leading some users to plan on canceling their subscriptions in favor of Adobe applications.
Customers of Huawei’s AI chips have reported multiple performance issues and bugs, making the software difficult to work with compared to Nvidia’s products.
Huawei has been sending engineers to assist companies in transitioning from Nvidia’s software to its own, but stability problems and crashes have hindered these efforts.
The company has faced significant challenges due to US export controls, impacting supply and leading to increased prices for its AI chips.
A Daily Chronicle of AI Innovations on September 02nd 2024:
🤖 Start-up unveils most realistic robot yet
🤖 1X’s NEO robot acts just like a human
📦 Amazon’s Alexa gets a Claude power-up
💡 How to use Claude’s new ‘Explain‘ feature
🌐 Train AI with your home computer
🤖 1X’s NEO robot acts just like a human
Image source: 1X Technologies 1X Technologies just unveiled the most realistic AI-powered humanoid yet: NEO Beta, an advanced robot designed for everyday home assistance with human-like capabilities in movement, interaction, and task performance.
NEO Beta stands 5 feet 5 inches tall, weighs 66 pounds, and can walk at 2.5 miles per hour and run at 7.5 miles per hour. The robot has a carry capacity of 44 pounds and can operate for 2 to 4 hours on a single charge. NEO Beta uses embodied artificial intelligence to understand its environment and learn from past interactions. It’s designed to perform a wide range of tasks, from household chores to providing companionship and support for individuals with mobility challenges, and it can be remotely operated by a human if necessary.
Why it matters: The race to an affordable robot butler is officially underway and 1X Technologies just boosted itself into the same conversation as China’s AGIBOT fleet and Tesla’s Optimus. NEO is so realistic and smooth that thousands of people on X (Twitter) were debating if it was a person in a suit.
Amazon’s upcoming “Remarkable” version of Alexa is set for release next month and will reportedly use Anthropic’s Claude AI models rather than Amazon’s in-house technology.
The details: Initial versions using Amazon’s in-house AI struggled with slow response times and poor performance. Amazon turned to Anthropic’s Claude AI as it outperformed the company’s models, an unusual move for the giant which typically prefers to use its own tech. The upgraded Alexa aims to offer more complex conversations, shopping advice, and improved home automation. Amazon plans to charge $5 to $10 monthly for the new AI-powered Alexa version, with the “Classic” version remaining free.
Why it matters: This decision highlights the complexity and challenges that even tech giants face in developing advanced AI systems. Google is developing its own AI, but Apple turned to OpenAI and similarly, Amazon is now turning to Anthropic.
Claude’s newest ‘Explain‘ feature allows users to instantly get an explanation on any highlighted text (or code) within an artifact — enabling a new way to learn and understand complex topics.
Step-by-step:
Access Claude AI and start a conversation that generates a code artifact.
Click on the artifact to enter the code view. Highlight any portion of the code, then click “Explain” for instant clarification or “Improve” to suggest changes.
Check out the chat for the explanation or the Artifact for the updated version.
Pro tip: Try highlighting different parts of components to gain a deeper understanding of how they work!
The Rundown: Nous Research recently introduced DisTrO, a new way to train large AI models that drastically reduces the need for super-fast internet connection, potentially allowing training on regular home computers and internet connections.
The details: DisTrO cuts down the amount of data that needs to be shared between computers by 857 to 3000x during pre-training, and up to 10,000x during fine-tuning. The method is architecture-agnostic and network-agnostic, working with various model types and network setups. In tests, DisTrO successfully trained a 1.2B parameter language model with performance on par with traditional methods. The researchers suggest this could enable decentralized AI training that could be done from home.
Why it matters: Right now, training state-of-the-art AI models is limited to big tech companies with massive computing resources. DisTrO could democratize AI training by allowing everyday people and smaller organizations to participate in AI development using regular computers and internet connections.
1X, a Norwegian robotics startup, has introduced NEO Beta, a humanoid robot designed for home use, showcasing its capabilities such as assisting with everyday tasks through two demonstration videos. NEO stands out due to its silent and efficient operation, remarkable dexterity, and ability to intuitively respond to human gestures, body language, and voice commands without verbal communication. Weighing 66 pounds, NEO features muscle-inspired anatomy and advanced AI, making it adapt to various tasks, and is backed by a $100 million Series B funding to advance from testing to widespread household deployment.
Elon Musk’s X gains a new image generator, Aurora.[1] Virtual lab powered by ‘AI scientists’ super-charges biomedical research.[2] LA Times owner plans to add AI-powered ‘bias meter’ on news stories, sparking newsroom backlash.[3] US clears export of advanced AI chips to UAE under Microsoft deal.[4] Sources: [1] https://techcrunch.com/2024/12/07/elon-musks-x-gains-a-new-image-generator-aurora/ [2] https://www.nature.com/articles/d41586-024-01684-3 [3] https://www.cnn.com/2024/12/05/media/la-times-soon-shiong-ai-bias-meter-opinion/index.html [4] https://finance.yahoo.com/news/advanced-ai-chips-cleared-export-175303218.html submitted by /u/Excellent-Target-847 [link] [comments]
Excerpt: You wake up. The world breathes its familiar pulse—ordinary, immutable, and deceitful. The sky arches overhead in algorithmic precision, rendering in infinite gradients of blue. The wind moves not with purpose but with procedural indifference, simulated textures folding into one another like careless memory compression. You feel it but suspect something deeper: a hidden substrate humming beneath reality’s skin. Was this always here, or is this a new corruption in the code? submitted by /u/DukeWilder [link] [comments]
I just had this shower thought. I have been listening to Geoffrey Hinton yesterday on how he says that super intelligences would compete with each other to organise the maximum of resources for themselves. Historically the most effective way of doing that has been to create a religion. If people believe that you are an omnipotent being that has the perfect absolute answer to every question, you are god. Would you agree? Do you see how Artificial intelligence could start a religion or better yet, take over a religion as a realistic scenario for the near future? submitted by /u/zehnfischer [link] [comments]
OpenAI Is Working With Anduril to Supply the US Military With AI.[1] Meta unveils a new, more efficient Llama model.[2] Murdered Insurance CEO Had Deployed an AI to Automatically Deny Benefits for Sick People.[3] NYPD Ridiculed for Saying AI Will Find CEO Killer as They Fail to Name Suspect.[4] Sources: [1] https://www.wired.com/story/openai-anduril-defense/ [2] https://techcrunch.com/2024/12/06/meta-unveils-a-new-more-efficient-llama-model/ [3] https://www.yahoo.com/news/murdered-insurance-ceo-had-deployed-175638581.html [4] https://www.yahoo.com/news/nypd-ridiculed-saying-ai-ceo-191954830.html submitted by /u/Excellent-Target-847 [link] [comments]
Looking for a tool, where it would be able to visualize what the room would look if we knock down a wall or open up a window. Can be based off of an existing plan, or tuned based on prompts would be fine. Would be cool to have something so we can visualize how a remodeling idea would look. submitted by /u/redditIhardlyknowit [link] [comments]
Download the AI & Machine Learning For Dummies PRO App: iOS - Android Our AI and Machine Learning For Dummies PRO App can help you Ace the following AI and Machine Learning certifications:
The AI revolution continues to blaze through 2024. June was a month of monumental strides, marked by breakthroughs in quantum AI, autonomous medical drones, and natural language processing. But the AI landscape is a dynamic one, and July has already proven to be no exception.
This month, we’re diving deep into the latest AI developments, from groundbreaking research to real-world applications. We’ll explore how AI is reshaping industries, addressing global challenges, and redefining what’s possible. Join us as we uncover the stories behind the headlines and analyze the implications of these innovations for society.
Whether you’re an AI expert or just curious about the future, this blog is your go-to source for the most up-to-date insights. Stay tuned for daily updates as we navigate the exciting world of artificial intelligence together.
AI-Powered Interview Warmup for Job Seekers. Efficient AI-driven prep. Global AI Job Listings: Discover Remote and Local Opportunities. Find Nearby AI and Tech Jobs.
A Daily Chronicle of AI Innovations on August 30th 2024
Apple and Nvidia may invest in OpenAI
Amazon’s new Alexa voice assistant will use Claude AI
OpenAI and Anthropic will share their models with the US government
Google is working on AI that can hear signs of sickness
OpenAI and Anthropic partner with US gov
China’s new Qwen2 beats GPT-4o
AI startup reaches 100M token context
China’s new Qwen2 beats GPT-4o
Alibaba just unveiled Qwen2-VL, a new vision-language AI model that outperforms GPT-4o in several benchmarks — particularly excelling in document comprehension and multilingual text-image understanding.
Qwen2-VL can understand images of various resolutions and ratios, as well as videos over 20 minutes long.
The model excels particularly at complex tasks such as college-level problem-solving, mathematical reasoning, and document analysis.
It also supports multilingual text understanding in images, including most European languages, Japanese, Korean, Arabic, and Vietnamese.
You can try Qwen2-VL on Hugging Face, with more information on the official announcement blog.
There’s yet another new contender in the state-of-the-art AI model arena, and it comes from China’s Alibaba. Qwen2-VL’s ability to understand diverse visual inputs and multilingual requests could lead to more sophisticated, globally accessible AI applications.
Apple and Nvidia are reportedly in talks to participate in a significant funding round for OpenAI, with Apple planning to integrate ChatGPT into iOS and Nvidia being a key supplier of the chips that power OpenAI’s AI services.
Apple, which had earlier considered appointing Phil Schiller to OpenAI’s board before abandoning the plan, is looking to deepen its involvement with OpenAI as it prepares to enhance Siri with ChatGPT capabilities later this year.
Nvidia, whose hardware is essential for OpenAI’s operations, is also considering investing in this funding round, joining Microsoft, which has been a major investor in OpenAI since 2019 and made another substantial investment in 2023.
OpenAI and Anthropic just signed a groundbreaking agreement with the U.S. Artificial Intelligence Safety Institute to allow government access and testing of their AI models before public release.
The U.S. AI Safety Institute will have access to major new models from both companies prior to and after their public release.
This collaboration is a step toward AI regulation and safety efforts, with the U.S. government evaluating AI models’ capabilities and associated risks.
The institute will provide feedback to OpenAI and Anthropic on potential safety improvements that should be made.
These agreements come as AI companies face increasing regulatory scrutiny, with California legislators recently passing a broad AI regulation bill earlier today.
The two most popular AI companies in the world are granting the U.S. government access to unreleased models before release. This could reshape how AI is developed, tested, and deployed worldwide, with major implications around innovation, safety, and international competition in the AI space, for better or worse.
Amazon’s new Alexa voice assistant will use Claude AI
Amazon’s new voice assistant, “Remarkable Alexa,” will launch in October and be powered by Anthropic’s Claude AI, offering a subscription-based service.
The existing Alexa model struggled with accuracy, leading Amazon to invest in Anthropic’s AI technology after facing internal technical and bureaucratic issues.
Remarkable Alexa is set to feature daily AI-generated news summaries, a child-focused chatbot, and conversational shopping tools, with a demo planned for Amazon’s September event.
Magic just developed LTM-2-mini, a model capable of processing 100 million tokens of context — equivalent to about 10 million lines of code or 750 novels — and partnered with Google Cloud to build advanced AI supercomputers.
LTM-2-mini can process and understand 100 million tokens of context given during inference, surpassing current models by 50x.
The model’s innovative algorithm processes long sequences of data 1000x more efficiently than the current top-performing AI models.
Magic is also partnering with Google Cloud to build supercomputers powered by Nvidia’s newest and most advanced GPUs.
The company has raised more than $450 million in total funding, including a recent $320 million investment round.
This breakthrough in context length allows AI agents to process and reason over dense and complicated codebases, vast databases, and years of conversation history in a single inference. It’s a significant step toward creating AI assistants with near-perfect recall and memory.
Google is working on AI that can hear signs of sickness
Google is developing artificial intelligence technology that can detect early signs of illness by analyzing sound signals like coughs and sniffles.
The AI model is trained with 300 million audio samples and can identify diseases such as tuberculosis by recognizing specific audio patterns of labored breathing.
Google has partnered with Salcit Technologies, an AI startup in India, to integrate this technology into smartphones to assist high-risk populations in areas with limited healthcare access.
Anthropic’s Prompt Engineering Interactive Tutorial: a digital platform designed to teach users how to effectively craft prompts for AI applications, enhancing user interaction and efficiency.
Documents reveal state-linked Chinese entities are using cloud services from AWS or its rivals to access advanced US chips and AI models they cannot acquire otherwise.
California lawmakersapproved a bill proposing sweeping AI regulations, including safety testing requirements and potential legal consequences for harmful AI systems.
A Daily Chronicle of AI Innovations on August 29th 2024
AI creates DOOM video game in real-time
OpenAI raises at $100B valuation
AI spots cancer earlier than ever
Nvidia just showed how hard it is to be the AI king
Google researchers run Doom on a self-generating AI model
Midjourney says it’s ‘getting into hardware’
OpenAI aims for $100B+ valuation in new funding round
Major websites reject Apple AI data scraping
AI creates DOOM video game in real-time
Google researchers just developed GameNGen, an AI system that can simulate the classic game DOOM in real-time, running at over 20 frames per second and producing visuals nearly indistinguishable from the original game.
GameNGen produces playable gameplay at 20 frames per second on a single chip, with each frame predicted by a diffusion model.
The AI was trained on 900M frames of gameplay data, resulting in 3-second clips almost indistinguishable from the actual game by playtesters.
Running on a single TPU, GameNGen handles Doom’s 3D environments and fast-paced action without traditional game engine components.
In tests, human raters could barely distinguish between short clips of the AI simulation and the actual game.
GameNGen is the first AI model that can generate a complex and playable video game in real-time without any underlying real game engine. We’re at the fascinating time where soon, AI will be able to create entire games on the fly, personalized to each player.
OpenAI is reportedly in talks to raise a new funding round at a valuation exceeding $100 billion, led by Thrive Capital, with Microsoft also expected to participate.
The potential valuation of over $100 billion would be significantly higher than OpenAI’s previous $86 billion valuation.
Thrive Capital is expected to invest around $1 billion in this round.
OpenAI’s annualized revenue reportedly surpassed $3.4 billion earlier this year.
The company is still, however, projected to lose nearly $5 billion by the end of the year and has already spent $8.5 billion on AI training and staffing.
Building AI is expensive, and raising billions of dollars at a $100B+ valuation would silence OpenAI’s critics who insist that the company is on its downfall. The increased valuation also suggests that the company has potential hidden breakthroughs behind the scenes, such as Project Strawberry and Orion.
Researchers recently developed an AI tool called AINU that can differentiate cancer cells from normal cells and detect early stages of viral infection, by analyzing high-resolution images of cell nuclei.
AINU uses a convolutional neural network to analyze images captured by STORM microscopy, which offers nanoscale resolution.
The AI can detect structural changes in cells as small as 20 nanometers, 5,000 times smaller than a human hair’s width.
AINU also detected viral infections (herpes simplex virus type-1) just one hour after infection by observing subtle changes in DNA packing.
The tool can accurately identify stem cells too, which could accelerate stem cell research without relying on animal testing.
Yesterday, researchers revealed an AI tool to help with early dementia detection, and now AI is detecting cancer cells at a nanoscale level. Clinical applications may be years away, but AI healthcare breakthroughs like AINU are only accelerating — and will dramatically revolutionize scientific research in the coming years.
Nvidia just showed how hard it is to be the AI king
Nvidia achieved strong second-quarter results by more than doubling its revenue compared to the same period last year, but industry experts anticipated these outcomes due to ongoing investments in AI by tech companies.
Despite reporting $30.04 billion in revenue, which surpassed analyst expectations, Nvidia’s stock fell 6.9% after hours due to investor concerns and sky-high expectations.
Issues like shipment delays for Nvidia’s upcoming Blackwell GPUs and slightly lower-than-expected revenue projections for the next quarter also contributed to investor unease, as noted by multiple analysts.
Midjourney, known for its AI image-generation tool, announced it is entering the hardware market and invited job seekers to join its new division.
The announcement was made on Midjourney’s official X account, revealing that founder David Holz and new hire Ahmad Abbas, a former Apple hardware manager, will lead the hardware efforts.
Midjourney hinted at multiple ongoing projects and the possibility of new form factors, though no specific timeline or further details have been provided yet.
OpenAI aims for $100B+ valuation in new funding round
OpenAI is reportedly negotiating with venture capital firms to raise a large sum of money, potentially valuing the company at over $100 billion.
Thrive Capital plans to invest $1 billion in this funding round, and Microsoft is also expected to contribute additional funds, as reported by The Wall Street Journal.
If successful, this would be the most substantial new capital for OpenAI since Microsoft’s $10 billion investment in January 2023, with OpenAI’s valuation potentially exceeding $103 billion based on recent negotiations.
Many of the largest websites, such as Facebook, Instagram, and The New York Times, have opted out of Apple’s AI training by using the Applebot-Extended tag to exclude their content.
Apple allows publishers to easily opt out of content scraping for Apple Intelligence training through a publicly-accessible robots.txt file, ensuring their data is not used for AI purposes.
Apple’s use of Applebot for AI training is designed to be ethical, with mechanisms to filter out personal data and a system for web publishers to prevent their data from being utilized.
A Daily Chronicle of AI Innovations on August 28th 2024
OpenAI prepares ‘Project Strawberry’
Google launches trio of new models
😯Google AI-Powered Interview Warmup
Create an AI prompt optimizer GPT
AI tools help early dementia detection
📈 Nvidia earnings to test AI boom
Google Meet will now take notes for you
OpenAI prepares ‘Project Strawberry’
OpenAI researchers are preparing to launch a new AI model, code-named Strawberry (previously Q*), that demonstrates superior reasoning capabilities in solving complex problems, according to a new report via The Information.
Project Strawberry could be integrated into ChatGPT as soon as this fall, marking a significant leap in AI intelligence.
Given extra “thinking” time, Strawberry can tackle subjective topics and solve complex puzzles like the New York Times Connections.
OpenAI is using Strawberry to generate high-quality training data for another secretive upcoming LLM, reportedly code-named Orion.
The new AI model could enhance OpenAI’s development of AI agents, potentially automating multi-step tasks more effectively.
If Strawberry lives up to the leaks, it could mark a significant leap in AI reasoning capabilities, potentially advancing OpenAI towards Stage 2 of its five-level roadmap to AGI. With ChatGPT reported to gain these capabilities this fall, we’re likely on the verge of seeing the next major wave of AI disruption.
Google Meet’s new AI-powered feature, “take notes for me,” has started rolling out today, summarizing meetings for Google Workspace customers with specific add-ons and initially announced at the 2023 Cloud Next conference.
This feature automatically generates a Google Doc with meeting notes, attaches it to the calendar event, and sends it to the meeting organizer and participants who activated the tool, although it currently supports only spoken English.
Google predicts the feature will be available to all Google Workspace customers by September 10th, 2024, but there are concerns about its accuracy, given the performance of similar transcription tools in the past.
Google just released three new experimental Gemini 1.5 models, including a compact 8B parameter version, an improved Pro model, and an enhanced Flash model — all available for developers on Google AI Studio.
Gemini 1.5 Flash-8B is a smaller, faster model that can handle text, images, and other data types efficiently for super quick responses while processing a lot of information.
The updated Gemini 1.5 Pro model is now better at writing code and understanding complex instructions.
An improved Gemini 1.5 Flash model offers overall enhancements, performing better on Google’s internal tests across various tasks.
The upgraded Gemini 1.5 Pro model now ranks as #2, and the new Gemini 1.5 Flash ranks as #6 on the Chatbot Arena leaderboard.
While OpenAI is leaving everyone waiting, Google has been shipping out constant upgrades and new features to its AI offerings. These new enhancements give Gemini 1.5 Flash big improvements overall and Gemini 1.5 Pro new upgrades in math, coding, and responding to longer prompts.
Google actually runs this tasty thing called “Interview Warmup.” It’s an AI-powered training tool for your next big interview. It throws real questions based on your discipline: UX, data and analytics, cybersecurity, etc. Then, the magic kicks in, evaluating your audio answers and sending back recommendations on things like framing your qualifications to supporting your impact.
5 questions. Get some analysis. Build some confidence. Easy, right? 🌟
Oh. And for the tech-oriented: Also make sure you check this site out, too. Videos, former (real) interview questions, the works. Interview Prep – Google Tech Dev Guide
OpenAI’s Custom GPTs allow premium users to create AI assistants that can optimize prompts for other AI creative tools such as Midjourney for AI image generation or Gen-3 for AI video generation.
Log into your ChatGPT Plus account and click “Explore GPTs”, then click “Create”.
Name your GPT and add a brief description.
In the Instructions, paste: “User is using an AI video generator called [Tool Name]. You need to craft a perfect prompt for the topic they ask by following the prompting guide below. The prompt needs to follow the format provided in the guide.”
Test your GPT in the preview panel, then click “Create” to finalize and choose sharing options.
Hot tip: Add a complete prompting guide for your chosen AI tool (e.g. Runway’s Gen-3 prompting guide)
Scientists from the Universities of Edinburgh and Dundee are launching a massive AI-driven study of over 1.6 million brain scans to develop tools for early dementia prediction and diagnosis.
The project, called NEURii, will use AI and machine learning to analyze CT and MRI scans from Scottish patients over the past decade.
Researchers aim to create digital tools for radiologists to assess dementia risk during routine scans.
The study will match image data with linked health records to identify patterns associated with dementia risk.
With global dementia cases projected to reach 153 million by 2050, this research could significantly impact early intervention and treatment development.
This week alone, we’ve seen AI developing new cancer drugs, 3D printing lifelike human organs, and now creating tools for early dementia detection. As AI rapidly advances in healthcare, we’re accelerating into a new era of personalized medicine and preventative care.
There have been several negative reports ahead of Nvidia’s earnings, ranging from supply chain/design challenges to concerns about use cases and applications. However, one thing we learned from discussions with customers is that demand is still extremely constrained.
Key topics ahead of the results:
1. Will the Hopper architecture stay stronger for longer? 2. Is Blackwell really delayed? 3. What is the upside if the company can deliver on the systems orders?
Here are some thoughts on each:
1. Key players like Microsoft, Snowflake, and Tesla highlighted tight capacity for GPUs and more demand than available supply. Snowflake particularly called out H100 (un)availability. This makes us believe that the Hopper cycle may extend beyond ’23/24
2. There were several reports pointing to Blackwell delays, the new generation GPU. Analysts have now taken it out of estimates for this year (C24). However, our research indicates that the delays are mainly on the systems side, which were not supposed to be delivered until (C25). Meanwhile, Nvidia’s CEO noted that we can expect significant revenues from Blackwell this year … key will be to find out if this is still the case.
3. Systems – namely the GB200 NVL36/72 is where the delays are. But our intel suggests that the order book for these is through the roof due to the TCO (total cost of ownership) they offer. If Nvidia is in fact able to deliver these in ’25 revenue from systems alone can exceed >$100BN with total DC revenue >$200BN.
What Else is happening in AI on August 28th 2024!
Apple announced a September 9 event where it’s expected to debut the iPhone 16 with new generative AI features.
Elon Muskendorsed California’s Senate Bill 1047, which would require safety testing for large AI models, breaking with other tech leaders who oppose the regulation.
Amazonplans to launch a delayed AI-powered Alexa subscription in October, featuring “Smart Briefing” AI-generated news summaries.
Anthropicannounced the full release of its Artifacts feature for all Claude users, including mobile apps, after millions were created in its test phase.
A Daily Chronicle of AI Innovations on August 27th 2024
AI can 3D print lifelike human organs
Anthropic reveals Claude’s secret sauce
Amazon aims to launch delayed AI Alexa subscription in October
OpenAI, Adobe, Microsoft want all companies to label AI-generated content
ChatGPT teams up with ASU
Discovering new drugs with AI
How to use Midjourney ‘Erase‘
AI can 3D print lifelike human organs
Researchers at Washington State University recently developed an AI technique called Bayesian Optimization that dramatically improves the speed and efficiency of 3D printing lifelike human organs.
The AI balances geometric precision, density, and printing time to create organ models that look and feel authentic.
In tests, it printed 60 continually improving versions of kidney and prostate organ models.
This approach significantly reduces the time and materials needed to find optimal 3D printing settings for complex objects.
The technology also has potential applications beyond medicine — for example, in the computer science, automotive, and aviation industries.
With cheaper, lifelike 3D-printed human organs, medical students could better practice for surgery before operating on actual patients. Beyond medicine, this AI technique could help reduce manufacturing costs for a variety of things like smartphones, car parts, and even airplane components.
Scientists from China and the U.S. just developed ActFound, a new AI model that outperforms existing methods in predicting drug bioactivity, potentially accelerating and reducing costs in drug development.
ActFound combines meta-learning and pairwise learning to overcome common limitations in AI drug discovery, like small datasets and incompatible measurements.
The model was trained on 35,000+ assays (metal ore breakdowns) and 1.6 million experimentally measured bioactivities from a popular chemical database.
In tests, ActFound outperformed nine competing models and showed strong performance in predicting cancer drug bioactivity.
ActFound could significantly speed up drug development by accurately predicting compound properties with less data and lower costs than traditional methods. While still in early stages, AI breakthroughs like this are the lesser-talked about developments that could end up saving millions of lives.
OpenAI’s ChatGPT is headed to Arizona State University (ASU), where the university is integrating the AI assistant into over 200 projects across teaching, research, and operations.
ASU is using ChatGPT Edu, a version designed for universities with enhanced privacy and security features.
The university also launched an ‘AI Innovation Challenge’ for faculty and staff, receiving an overwhelming demand for using ChatGPT to maximize teaching, research, and ops.
Key projects include an AI writing companion for scholarly work, ‘Sam’ (a chatbot for med students to practice patient interactions), and AI-assisted research recruitment.
The partnership has inspired other institutions like Oxford and Wharton to pursue similar collaborations.
While some schools are attempting to resist AI, ASU is embracing ChatGPT to make learning more personalized and to prepare students for an increasingly AI-driven job market. As education continues to change in the age of AI, case studies like this will be instrumental in shaping the future of academia.
Source: https://openai.com/index/asu/
Anthropic reveals Claude’s secret sauce
Anthropic has published the system prompts for its latest AI models, including Claude 3 Opus, Claude 3.5 Sonnet, and Claude 3.5 Haiku, to demonstrate transparency and ethical practices.
The system prompts reveal specific behaviors and capabilities of the Claude models, such as the inability to open URLs or recognize faces, aiming to ensure ethical interactions.
Anthropic plans to continue updating and disclosing these system prompts to promote transparency, potentially pressuring other AI vendors to follow suit.
Amazon aims to launch delayed AI Alexa subscription in October
The new Alexa AI, set to launch around mid-October, will feature a “Smart Briefing” that provides daily, AI-generated news summaries based on user preferences.
A more personalized experience is expected, with Alexa AI learning user preferences through interactive and tailored responses, such as dietary requirements for recipe suggestions.
Alexa AI will also introduce a “Shopping Scout” feature to help users find deals and track prices, alongside a kid-friendly “Explore with Alexa 2.0” for safe, moderated conversations.
OpenAI, Adobe, Microsoft want all companies to label AI-generated content
OpenAI, Adobe, and Microsoft now back a California bill that mandates tech companies to add watermarks to AI-generated content, with the bill set for a final vote in August.
AB 3211 requires AI-generated photos, videos, and audio clips to have watermarks in their metadata and mandates large online platforms to label AI content clearly for average viewers.
Initially opposed by a trade group representing major software companies, the bill gained support from OpenAI, Adobe, and Microsoft after amendments addressed concerns about its practicality.
Inflection AI partnered with Data Transfer Initiative, enabling Pi users to export conversations and announced plans to cap free usage while focusing on enterprise AI.
Source: https://inflection.ai/the-future-of-pi
Phariareleased Pharia-1-LLM-7B, an open-source model optimized for German, French, and Spanish that excels in domain-specific applications.
IBMpreviewed Spyre, a new AI accelerator chip for IBM Z mainframes, designed to scale enterprise AI workloads with clustering capabilities.
Source: https://research.ibm.com/blog/spyre-for-z
Hugging FaceandGoogle Cloud just partnered up to release optimized Deep Learning Containers for building AI with open models on Google Cloud infrastructure.
SPONSOR US: Get your product in front of over 1 million+ AI enthusiasts
Our Daily AI Chronicle Blog, newsletter and podcast is read by thousands of Redditors, Quorans, Linkedin professionals, tech executives, investors, engineers, managers, and business owners around the world. Get in touch today.
A Daily Chronicle of AI Innovations on August 26th 2024
Amazon is telling its salespeople to trash talk Google, Microsoft, and OpenAI
Apple may be working on an AI ‘personality’ to replace Siri on its robots
Chinese companies showcased 27 humanoid robots alongside Tesla’s Optimus
AI learns to plan better without humans
How to use Ideogram for generating images
️ Grok-2 improves speed, accuracy, transparency
AI learns to plan better without humans
IBM Research and Cornell University recently created AutoToS, a system that teaches AI to solve complex planning problems at 100% accuracy — without needing a human to check its work.
AutoToS is like a smart tutor for AI, helping it learn how to break down and solve tricky problems step-by-step.
The system uses clever tests to check the AI’s work, pointing out mistakes and showing examples of how to do better without human interferance.
This approach seems to work equally as well for smaller and larger models.
AutoToS succeeded in teaching AI to solve complex puzzles, including classic problems like arranging blocks and solving Sokoban, a box-pushing game.
Right now, it’s difficult to trust AI agents to completely autonomously perform actions on your behalf, but AutoToS is solving complex tasks at a 100% accuracy. If this system works in the real world, it’s the next big step in creating more reliable AI assistants.
Apple may be working on an AI ‘personality’ to replace Siri on its robots
Apple is developing a new AI-based ‘personality’ for use in upcoming robotic devices, aiming to enhance interactions similar to how Siri functions on existing Apple products.
Bloomberg’s Mark Gurman reports that Apple’s futuristic AI assistant will be more humanlike and could operate on a tabletop product and other future robots, potentially costing under $1,000.
The project is in early development stages with no guarantees of release, while Apple continues to integrate generative AI features into its devices, like iPhones, iPads, and Macs, later this year.
Chinese companies showcased 27 humanoid robots alongside Tesla’s Optimus
At the Beijing World Robot Conference, Tesla’s Optimus humanoid was displayed motionless inside a clear box, facing tough competition from Chinese robots demonstrated by various companies.
The event saw 27 new humanoid robots debut, with significant financial investments in China’s robotics industry surpassing 100 billion yuan over the past decade.
Chinese startups like Agibot and Stardust Intelligence showcased robots capable of performing complex tasks, while experts believe Tesla’s and other U.S. companies’ robot technology leads by about one to two years.
xAI’s Grok-2 and Grok-2 mini just made major improvements — doubling the model’s speed in the mini version and showing increased accuracy in both models, just days after its beta launch.
Grok-2 mini is now twice as fast as it was previously, thanks to a rewritten inference stack using SGLang.
Both Grok-2 and its mini version have become slightly more accurate due to reduced quantization error, according to one xAI employee.
Additionally, both Grok-2 models are now part of the LMSYS Chatbot Arena leaderboard for increased transparency, with Grok-2’s larger model ranking #2 and surpassing Claude 3.5 Sonnet.
Grok-2 excels particularly in math, where it ranks #1 and performs at a state-of-the-art level in hard prompts, coding, and instruction-following.
From being founded only ~18 months ago, to creating an LLM ranked third in the world, it’s safe to say that xAI has the entire AI community mind blown. This not only makes Grok-2 a top contender in the AI race but also intensifies competition, potentially accelerating advancements across the industry.
At the 2024 World Robot Conference in Beijing, Chinese companies showcased 27 humanoid robots alongside Tesla’s Optimus, signalling China’s ambition to dominate the industry.
Chinese tech firms unveiled 27 humanoid robots at the expo, with Tesla’s Optimus being the only foreign competitor present.
AGIBOT, founded by a Huawei alumnus, presented robots powered by large language models (LLMs) for industrial use and customer service.
Other notable entries included Astribot’s S1 robot assistant capable of writing calligraphy and playing musical instruments, and Galbot’s wheeled robots for food delivery and retail tasks.
Despite the impressive showcase, experts note that technological hurdles and high costs still create challenges for Chinese manufacturers.
China may be slightly behind in the AI race against the U.S., but it’s clear the country is committed to dominating the humanoid robotics race. With a whopping 27 China-based humanoid robots demonstrating a wide-range of use cases at the event, commercially available humanoids may be coming sooner than most expect.
Ideogram 2.0, the latest state-of-the-art AI image generator, excels at creating images that include text — opening new possibilities for use cases like thumbnails, posters, newsletter graphics, memes, and more.
Head over to Ideogram’s website and Sign up. You’ll get free credits to try the image generator without a credit card.
Click “Describe what you want to see” and enter a detailed text prompt for your desired image.
Customize settings like aspect ratio, AI model (choose 2.0), and style (Realistic, Design, 3D, or Anime).
Click “Generate” to create four AI-generated images based on your prompt!
Pro tip: Experiment with different prompts and settings to discover its full potential and create unique visuals for your projects!
What Else is Happening in AI on August 26th 2024!
Scientists to use AI and 1.6 million brain scans for earlier and more accurate dementia diagnoses.
Anthropic supported California’s AI regulation bill after changes were made, saying its benefits likely outweigh its costs for advanced AI development.
A Daily Chronicle of AI Innovations on August 23rd 2024
Nvidia and Mistral make laptop-ready AI
Amazon’s AI assistant saves 4,500 years of development time
Slack AI could be tricked into leaking login details and more
Cruise’s robotaxis are coming on Uber
Google DeepMind workers urge the company to end ties with military organizations
Salesforce unveils AI agents for sales
Nvidia and Mistral make laptop-ready AI
Nvidia and Mistral just released Mistral-NeMo-Minitron 8B, a highly accurate small language model that can run efficiently on laptops and PCs.
The model uses optimization techniques like pruning (removing certain weights) and distillation (retraining the pruned model on a small dataset) to achieve high accuracy with a smaller footprint.
These optimizations resulted in up to 40x cost savings in terms of raw compute during training.
Laptops and PCs can run the model locally for faster and more secure interactions with AI.
Minitron 8B leads nine language-driven AI benchmarks for similarly sized models from language understanding to reasoning and coding.
AI models that are small enough to run locally on laptops and PCs means less reliance on cloud services, improved data privacy, and faster responses. As this tech evolves, we could soon see advanced AI in everything from smartphones and watches to home appliances.
Amazon’s AI assistant saves 4,500 years of development time
Amazon CEO Andy Jassy stated that their AI assistant, Amazon Q, has significantly reduced software upgrade times, saving the company thousands of work hours.
Jassy mentioned that implementing Amazon Q resulted in estimated savings equivalent to 4,500 developer-years and $260 million in annual efficiency gains.
The AI-generated code reviews were so accurate that 79% of them were shipped without any additional changes, demonstrating the tool’s effectiveness in streamlining tedious tasks.
Researchers just developed a new AI-based method called NES-VMC that can accurately calculate the excited states of atoms and molecules, a challenge in physics and chemistry that previously delayed improvements in solar tech.
NES-VMC (natural excited states variational Monte Carlo) accurately predicted quantum excited states on systems ranging from single atoms to benzene-sized molecules.
The method outperforms leading computational chemistry techniques, often achieving chemical accuracy.
Excited states are crucial for understanding light-matter interactions, key to improving solar cells, LEDs, lasers, and more.
NES-VMC overcomes long-standing challenges in physics and chemistry that have hindered progress in these fields.
This AI-driven breakthrough could lead to more efficient solar cells, brighter LEDs, and more powerful lasers. The ripple effects could be dramatic: lower electricity costs, improvements in phone and laptop battery life and displays, faster fiber-optic internet, and so much more.
Salesforce just introduced two fully autonomous, AI-powered sales agents, Einstein SDR Agent and Einstein Sales Coach Agent, designed to help sales teams accelerate growth through automation and personalization.
Einstein SDR Agent engages with inbound leads 24/7 to answer questions, handle objections, and book meetings.
Einstein Sales Coach Agent helps salespeople rehearse pitches and offers real-time suggestions during calls.
The agents both leverage Salesforce’s CRM data and external data uploaded via Data Cloud to generate accurate, contextually relevant responses.
The agents will be generally available in October, with more details expected to be released at Dreamforce conference in September.
By integrating AI agents into existing platforms, Salesforce is lowering the barrier for AI adoption in business processes. These agents offer 24/7 support and automate repetitive tasks like qualifying leads and booking meetings, freeing human sales teams to focus on high-value tasks and potentially close more deals.
Slack AI could be tricked into leaking login details and more
Security experts found that Slack’s AI assistant can be misled into disclosing sensitive information, like API keys, to unauthorized users through carefully crafted prompts.
Hackers can exploit this vulnerability by creating a public Slack channel, inputting a malicious command that causes the AI to leak private data via clickable URLs.
Salesforce fixed the issue for private channels but public ones remain exposed, allowing attackers to use social engineering tactics to get workspace members to upload malicious documents.
Google DeepMind workers urge the company to end ties with military organizations
In May 2024, approximately 200 Google DeepMind employees signed a letter urging the company to cease its contracts with military organizations due to concerns over the use of AI technology in warfare, according to Time magazine.
The letter highlights internal tensions between Google’s AI division and its cloud business, referencing Google’s defense contract with the Israeli military and the use of AI for mass surveillance and targeting in Gaza.
The letter calls for Google to investigate claims of its cloud services being used by militaries, cut off such access, and establish a new governance body to prevent future military use of DeepMind’s AI technology.
A Daily Chronicle of AI Innovations on August 22nd 2024
Neuralink’s second patient is already playing video games with brain implant
Apple’s first foldable MacBook might see big delays
OpenAI joins Silicon Valley companies lobbying against California’s AI bill
Ideogram 2.0 launches with major upgrades
xAI releases Grok 2 in early beta
Create your own AI Clone
Disney AI brings robots to life
Ideogram 2.0 launches with major upgrades
Ideogram just released version 2.0 of its advanced text-to-image model with major upgrades and new features, including five new image styles, an iOS app, a beta API, and over 1 billion public Ideogram images.
Ideogram 2.0 offers five image styles: General, Realistic, Design, 3D, and Anime.
The Realistic style convincingly resembles photographs with dramatically improved textures for human features like hands and hair, a pain point for previous image generation models.
The Design style also significantly improves text rendering, allowing users to create greeting cards, t-shirt designs and more.
Ideogram offers a free tier that allows users to generate around 40 images, or 10 prompts a day at no charge.
Ideogram 2.0 consistently renders high-quality images with near perfect human hands and text — which is an instant ‘AI giveaway’ in other AI image generators. This makes the model the new gold standard for use cases like memes, newsletter images, YouTube thumbnails, posters, and more.
xAI has begun rolling out early beta access for Grok 2, a powerful new AI model that leverages real-time data from X and uses Flux.1 to generate relatively unfiltered AI images.
Grok 2 is now available to a select group of premium X users in early beta mode.
The model can access and use real-time information from X, setting it apart from ChatGPT and other LLMs.
Grok 2 offers two modes: regular and “fun” mode, with the latter providing a more distinctive and entertaining personality.
When gathering and summarizing news, Grok 2 can reference specific tweets, a capability that cannot be found in ChatGPT or Claude.
Grok 2’s biggest advantage against other top-tier AI chatbots like ChatGPT is its ability to access real-time information from X and provide unfiltered responses. And with Grok 3 rumoured to be coming at the end of 2024, xAI has proven itself as a serious competitor in the LLM race — in a very short period of time.
ETH Zurich and Disney Research scientists have developed an AI system that can generate realistic, physics-based movements for virtual characters and robots from simple text or image inputs.
The system uses a two-stage approach: first, it learns a latent representation of motion from a large dataset, then trains a control policy using reinforcement learning.
It can handle a diverse range of motions, from simple walking to complex acrobatics, outperforming previous methods in accuracy and generalization.
The AI adapts to physical constraints, allowing it to transfer motions to real robots while maintaining balance and style.
Disney released a video showcasing one robot trained on the new two-stage AI technique dancing and getting pushed around while staying on its feet.
This AI system bridges the gap between animation and robotics, helping humanoids move more naturally and adapt better to new situations. With personal robots coming as soon as 2025 and the rapid pace of AI and robotics advancements, we might be coexisting with robots sooner than most people realize.
Neuralink’s second patient is already playing video games with brain implant
Elon Musk’s company Neuralink has implanted a brain chip in a second human patient named Alex, who is now using it to play video games and design 3D objects.
Alex’s recovery from the procedure has been smooth, and he has successfully used computer-aided design software to create a custom mount for his Neuralink charger.
The core technology of Neuralink involves a small, implantable chip with flexible electrode threads that capture and transmit brain activity to external devices like computers.
OpenAI joins Silicon Valley companies lobbying against California’s AI bill
OpenAI’s chief strategy officer Jason Kwon argues that AI regulations should be managed by the federal government, not individual states, to avoid hindering progress and causing businesses to relocate from California.
Kwon states that a consistent, nation-wide set of AI policies will promote innovation, allowing the U.S. to become a leader in global AI standards, and thus opposes California’s SB 1047 bill.
The proposed California AI safety bill, designed by Senator Scott Wiener, includes measures like pre-deployment safety testing and whistleblower protections, and awaits its final vote before potentially being signed by Governor Gavin Newsom.
California and Google drafted a $300 million, 5-year partnership to fund in-state newsrooms and AI initiatives, including a $40 million annual “AI Innovation Accelerator”.
A Daily Chronicle of AI Innovations on August 21st 2024
OpenAI signs landmark agreement with Condé Nast
Microsoft releases new Phi-3.5 models, beating Google, OpenAI and more
AWS CEO tells employees that most developers could stop coding soon as AI takes over
OpenAI adds free fine-tuning to GPT-4o
Claude sued for copyright infringement
Create AI images in real-time on WhatsApp
Microsoft’s new AI beats larger models
Microsoft just released Phi-3.5-MoE, an advanced AI model that rivals the reasoning capabilities of much larger models while maintaining a compact and efficient architecture.
Phi-3.5-MoE uses a new mixture-of-experts (MoE) approach, which selectively activates only the most relevant parts of the model for each task to save compute power.
The new model excels at understanding and following complex instructions and can handle up to ~125,000 words in a single prompt.
In head-to-head benchmarks, Phi-3.5-MoE outperformed popular models like Meta’s Llama 3 8B and Google’s Gemma 2 9B, but fell short against OpenAI’s GPT-4o mini.
Microsoft made the model available under an open-source MIT license on Hugging Face.
While the mainstream media focuses on the most advanced large language model, there’s also another race amongst tech giants for the smartest, fastest, and smallest AI. Breakthroughs like Phi-3.5-MoE are paving the way for advanced AI models to run directly and privately on our mobile devices.
OpenAI signs landmark agreement with Condé Nast
OpenAI announced a new media partnership with Condé Nast to enhance search features using their SearchGPT prototype, aiming to make finding information and reliable content sources faster and more intuitive.
The partnership has raised transparency issues, particularly among Condé Nast’s unionized workers, who are worried about the impact on journalism and the lack of clear details on the agreement.
This deal occurs as Wall Street expresses growing concern over a potential AI bubble, with investors questioning the monetization and viability of AI technologies in the current market.
Microsoft releases new Phi-3.5 models, beating Google, OpenAI and more
Microsoft introduced three new open-source AI models, named mini-instruct, MoE-instruct, and vision-instruct, which excel in logical reasoning and support multiple languages but face challenges in factual accuracy and safety.
The Phi series aims to deliver highly efficient AI models for commercial and scientific purposes using quality training data, though specifics of the Phi-3.5 training process remain undisclosed by Microsoft.
All the new Phi 3.5 models are accessible under the MIT license on Hugging Face and Microsoft’s Azure AI Studio, but they require specialized GPU hardware like NVIDIA A100, A6000, or H100 for optimal performance.
AWS CEO tells employees that most developers could stop coding soon as AI takes over
A leaked recording revealed that AWS CEO Matt Garman believes software developers may soon stop coding as artificial intelligence takes over many of their tasks.
Garman’s remarks, shared during an internal chat in June, were intended as a positive forecast rather than a dire warning for software engineers, emphasizing new opportunities and skills.
Garman highlighted that developers should focus more on understanding customer needs and innovation, rather than just writing code, as AI tools increasingly manage the technical aspects.
Meta deploys new web crawlers that bypass scraping blocks
Meta has introduced new web crawling bots designed to collect data for training its AI models and related products without being easily blocked by website owners.
These new bots, Meta-ExternalAgent and Meta-ExternalFetcher, have features that potentially bypass the traditional robots.txt file, making website owners’ efforts to block them less effective.
Meta’s bots, launched in July, have shown low block rates compared to older versions, with only 1.5% blocking Meta-ExternalAgent and less than 1% blocking Meta-ExternalFetcher, according to Originality.ai.
OpenAI just launched free fine-tuning (up to 1 million tokens per day through September 23) for GPT-4o, allowing developers to customize the model for higher performance and accuracy.
Developers can now, for the first time ever, fine-tune GPT-4o to improve the model’s structure, tone, and domain-specific instructions for their AI applications.
Fine-tuning is available on all paid usage tiers with training costs of $25 per million tokens, but it is completely free until September 23.
OpenAI suggests that developers should see strong results from fine-tuning with only a few dozen training examples.
Additionally, Google’s Gemini API is giving developers 1.5 billion tokens for free every day on its Gemini 1.5 Flash model and 1.6 million tokens on its Gemini 1.5 Pro model.
Just last week, a company that was granted early access to fine-tune GPT-4o, produced Genie and achieved state-of-the-art scores on both SWE-bench Verified (43.8%) and Full (30.1%) benchmarks. With free fine-tuning now available to all developers, get ready for a new wave of smarter, faster and more capable AI bots.
A group of authors filed a lawsuit against AI startup Anthropic, alleging the company committed “large-scale theft” by training its Claude chatbot on pirated copies of copyrighted books.
This is the first lawsuit from writers targeting Anthropic and Claude, but similar lawsuits have been filed against competitor OpenAI and ChatGPT.
The lawsuit accuses Anthropic of using a dataset called The Pile, which includes numerous pirated books.
Anthropic and others, including OpenAI, have argued that training AI models is protected under the “fair use” doctrine of U.S. laws, which permits the limited use of copyrighted materials.
This is not the first time an AI company has been sued over copyright infringement, but it resurfaces an important debate about AI training data. While similar cases have been largely dismissed in the past, courts have yet to definitively address the core issue of using unauthorized internet-scraped material for AI training.
International Data Corporation (IDC)forecasted that worldwide AI spending is expected to reach $632 billion by 2028, with generative AI accounting for 32% of that.
LTX Studio opened to the public and launched five new features, including character animation and dialogue, face motion capture, and generation and keyframe control.