AI Innovations in June 2024

AI Innovations in June 2024
DjamgaMind - AI Unraveled Podcast

DjamgaMind: Audio Intelligence for the C-Suite (Energy, Healthcare, Finance)

Are you drowning in dense legal text? DjamgaMind is the new audio intelligence platform that turns 100-page healthcare or Energy mandates into 5-minute executive briefings. Whether you are navigating Bill C-27 (Canada) or the CMS-0057-F Interoperability Rule (USA), our AI agents decode the liability so you don’t have to. 👉 Start your specialized audio briefing today at Djamgamind.com


AI Jobs and Career

I wanted to share an exciting opportunity for those of you looking to advance your careers in the AI space. You know how rapidly the landscape is evolving, and finding the right fit can be a challenge. That's why I'm excited about Mercor – they're a platform specifically designed to connect top-tier AI talent with leading companies. Whether you're a data scientist, machine learning engineer, or something else entirely, Mercor can help you find your next big role. If you're ready to take the next step in your AI career, check them out through my referral link: https://work.mercor.com/?referralCode=82d5f4e3-e1a3-4064-963f-c197bb2c8db1. It's a fantastic resource, and I encourage you to explore the opportunities they have available.

Job TitleStatusPay
Full-Stack Engineer Strong match, Full-time $150K - $220K / year
Developer Experience and Productivity Engineer Pre-qualified, Full-time $160K - $300K / year
Software Engineer - Tooling & AI Workflows (Contract) Contract $90 / hour
DevOps Engineer (India) Full-time $20K - $50K / year
Senior Full-Stack Engineer Full-time $2.8K - $4K / week
Enterprise IT & Cloud Domain Expert - India Contract $20 - $30 / hour
Senior Software Engineer Contract $100 - $200 / hour
Senior Software Engineer Pre-qualified, Full-time $150K - $300K / year
Senior Full-Stack Engineer: Latin America Full-time $1.6K - $2.1K / week
Software Engineering Expert Contract $50 - $150 / hour
Generalist Video Annotators Contract $45 / hour
Generalist Writing Expert Contract $45 / hour
Editors, Fact Checkers, & Data Quality Reviewers Contract $50 - $60 / hour
Multilingual Expert Contract $54 / hour
Mathematics Expert (PhD) Contract $60 - $80 / hour
Software Engineer - India Contract $20 - $45 / hour
Physics Expert (PhD) Contract $60 - $80 / hour
Finance Expert Contract $150 / hour
Designers Contract $50 - $70 / hour
Chemistry Expert (PhD) Contract $60 - $80 / hour

AI Innovations in June 2024.

Welcome to our blog series “AI Innovations in June 2024”! This is an evolving article that will be updated daily throughout the month of June 2024 to bring you the latest news and developments in the world of artificial intelligence. As we move further into the 21st century, the pace of AI innovation continues to accelerate at a breathtaking rate. Each day, researchers, engineers, and visionaries are pushing the boundaries of what’s possible, unlocking new capabilities and applications that are transforming industries, enhancing our lives, and shaping the future. In this blog, we’ll dive deep into the most exciting AI breakthroughs, advancements, and milestones happening in June 2024. From groundbreaking AI-powered technologies and cutting-edge research to the societal and ethical implications of these innovations, we’ll provide you with a comprehensive and insightful look at the rapidly evolving world of artificial intelligence. Whether you’re an AI enthusiast, a tech-savvy professional, or simply someone curious about the future, this blog will keep you informed, inspired, and engaged. So, join us on this journey of discovery as we explore the frontiers of AI and uncover the innovations that are shaping our world. Stay tuned for daily updates, and get ready to be amazed by the incredible advancements that are happening in the world of AI!

LISTEN DAILY AT OUR PODCAST HERE

Experience the transformative capabilities of AI with “Read Aloud For Me – AI Dashboard – AI Tools Catalog – AI Tools Recommender” – your ultimate AI Dashboard and Hub. Seamlessly access a comprehensive suite of top-tier AI tools within a single app, meticulously crafted to enhance your efficiency and streamline your digital interactions. Now available on the web at readaloudforme.com and across popular app platforms including Apple, Google, and Microsoft, “Read Aloud For Me – AI Dashboard” places the future of AI at your fingertips, blending convenience with cutting-edge innovation. Whether for professional endeavors, educational pursuits, or personal enrichment, our app serves as your portal to the forefront of AI technologies. Embrace the future today by downloading our app and revolutionize your engagement with AI tools.

AI Dashboard - Wonderland Bedtime Stories - Safe Ai for All
AI Dashboard – Wonderland Bedtime Stories – Safe Ai for All

📊Top AI Statistics 2024 Roundup.

📊Top AI Statistics 2024 Roundup.
📊Top AI Statistics 2024 Roundup.
  1. The global AI market is valued at over $196 billion.
  2. AI industry value is projected to increase by over 13x over the next 7 years.
  3. The US AI market is forecast to reach $299.64 billion by 2026.
  4. The AI market is expanding at a CAGR of 38.1% between 2022 to 2030.
  5. By 2025, as many as 97 million people will work in the AI space.
  6. AI market size is expected to grow by at least 120% year-over-year.
  7. 83% of companies claim that AI is a top priority in their business plans.
  8. Netflix makes $1 billion annually from automated personalized recommendations.
  9. 48% of businesses use some form of AI to utilize big data effectively.
  10. 38% of medical providers use computers as part of their diagnosis.

Credit: Josh Howarth

AI Weekly Summary June 22-30 2024

🤝 Apple and Meta are discussing a generative AI partnership
🔧 ByteDance and Broadcom collaborate on AI chip development
🕵️‍♂️ Researchers developed a new method to detect hallucinations
🎥 Synthesia 2.0: World’s 1st AI video communication platform
🛒 OpenAI is on an acquiring spree, buying Rocket and Multi
🎵 Record labels sue AI music startups over copyright infringement
💼 Anthropic rolls out Claude’s cutting-edge collaborative features
🤖 Google experiments with celebrity-inspired AI Chatbots
🛑 OpenAI postpones the launch of ChatGPT voice mode
🏁 Amazon steps into the chatbot race with Metis
🎨 Figma’s new AI features stir competition with Adobe
🥇 Alibaba’s Qwen-72B tops Hugging Face’s Open LLM Leaderboard
🚀 Google releases Gemma 2, lightweight but powerful open LLMs
🔍 OpenAI’s CriticGPT finds GPT-4’s mistakes with GPT-4
🌐 Google partners with Moody’s, Thomson Reuters & more for AI data

Apple and Meta are discussing gen AI partnership

Apple is reportedly in talks with its longtime rival Meta to integrate the latter’s Llama 3 AI model into Apple Intelligence. This move comes as Apple prepares to roll out its AI features across iPhones, iPads, and Macs later this year.

AI-Powered Professional Certification Quiz Platform
Crack Your Next Exam with Djamgatech AI Cert Master

Web|iOs|Android|Windows

Are you passionate about AI and looking for your next career challenge? In the fast-evolving world of artificial intelligence, connecting with the right opportunities can make all the difference. We're excited to recommend Mercor, a premier platform dedicated to bridging the gap between exceptional AI professionals and innovative companies.

Whether you're seeking roles in machine learning, data science, or other cutting-edge AI fields, Mercor offers a streamlined path to your ideal position. Explore the possibilities and accelerate your AI career by visiting Mercor through our exclusive referral link:

Find Your AI Dream Job on Mercor

Your next big opportunity in AI could be just a click away!

The potential partnership follows Apple’s existing deal with OpenAI, suggesting a collaboration strategy rather than solo development in the AI race. In Apple’s arrangement with OpenAI, there’s no direct payment. Instead, OpenAI can offer premium subscriptions through Apple Intelligence, with Apple taking a percentage. It’s unclear if Meta would agree to a similar business model, given that Llama 3 is open-source and free to access.

Sourcehttps://www.wsj.com/tech/ai/apple-meta-have-discussed-an-ai-partnership-cc57437e

ByteDance and Broadcom collaborate on AI chip development

ByteDance is collaborating with U.S. chip designer Broadcom to develop an advanced AI processor. This partnership aims to secure a stable supply of high-end chips amid ongoing U.S.-China tensions. The project centers on creating a 5-nanometre, customized Application-Specific Integrated Chip (ASIC) that complies with U.S. export restrictions.

This chip’s manufacturing is set to be outsourced to Taiwan Semiconductor Manufacturing Company (TSMC), though production is not expected to begin this year. While the design work is currently underway, the critical “tape out” phase has yet to commence.

AI Jobs and Career

And before we wrap up today's AI news, I wanted to share an exciting opportunity for those of you looking to advance your careers in the AI space. You know how rapidly the landscape is evolving, and finding the right fit can be a challenge. That's why I'm excited about Mercor – they're a platform specifically designed to connect top-tier AI talent with leading companies. Whether you're a data scientist, machine learning engineer, or something else entirely, Mercor can help you find your next big role. If you're ready to take the next step in your AI career, check them out through my referral link: https://work.mercor.com/?referralCode=82d5f4e3-e1a3-4064-963f-c197bb2c8db1. It's a fantastic resource, and I encourage you to explore the opportunities they have available.

Sourcehttps://www.reuters.com/technology/artificial-intelligence/chinas-bytedance-working-with-broadcom-develop-advanced-ai-chip-sources-say-2024-06-24

Researchers developed a new method to detect hallucinations

ChatGPT and Gemini can produce impressive results but often “hallucinate” false or unsubstantiated information. This research focuses on a subset of hallucinations called “confabulations,” where LLMs generate answers that are both wrong and arbitrary. Researchers have developed new methods to detect confabulations using entropy-based uncertainty estimators. They introduce the concept of “semantic entropy” to measure the uncertainty of LLM generations at the meaning level.

High semantic entropy corresponds to high uncertainty and indicates a higher likelihood of confabulation. The method computes uncertainty at the level of meaning rather than specific word sequences, addressing the fact that one idea can be expressed in many ways. The method provides scalable oversight by detecting confabulations that people might otherwise find plausible.

Sourcehttps://www.nature.com/articles/s41586-024-07421-0

Synthesia 2.0: World’s 1st AI video communication platform

Synthesia is launching Synthesia 2.0 – the world’s first AI video communications platform for businesses. It reinvents the entire video production process, allowing companies to create and share AI-generated videos at scale easily.

The key new features and capabilities of Synthesia 2.0 include:


AI Unraveled: Demystifying Frequently Asked Questions on Artificial Intelligence (OpenAI, ChatGPT, Google Gemini, Generative AI, Discriminative AI, xAI, LLMs, GPUs, Machine Learning, NLP, Promp Engineering)
  • 2 Personal AI Avatars: Expressive Avatars shot in a studio and Custom Avatars created using your webcam.
  • AI Video Assistant: Converts text, documents, or websites into high-quality videos, with options to customize the branding, tone, and length.
  • Intuitive Video Editing: Editing simplified with “Triggers” that let you control animations and edits from the script.
  • Translation and Dynamic Video Player: Videos can now be translated into over 120 languages. Synthesia is also building a new video player with interactive features.
  • AI Safety Focus: Synthesia is pursuing ISO/IEC 42001 certification, the first standard for responsible AI management, to ensure its AI technologies are ethical.

Sourcehttps://www.synthesia.io/post/introducing-synthesia-video-communications-platform?

OpenAI is on an acquiring spree, buying Rockset and Multi

Last week, OpenAI acquired Rockset, a startup that develops tools for real-time data search and analytics. OpenAI said it would integrate Rockset’s technology to power its infrastructure and offerings across products.

This week, OpenAI acquired Multi, a startup focused on building remote collaboration tools and software. Technically, the deal is an acqui-hire as the entire Multi team, including its co-founders, will join OpenAI to work on the company’s ChatGPT desktop application.

Sourcehttps://techcrunch.com/2024/06/24/openai-buys-a-remote-collaboration-platform

Record labels sue AI music startups over copyright infringement

The world’s major record labels, including Universal Music Group, Sony Music, and Warner Music, have filed twin lawsuits against the AI music generation startups Suno and Udio. The lawsuits accuse the companies of unlawfully training their AI models on massive amounts of copyrighted music, which, according to the complaints, allows the startups to generate similar-sounding music without permission.

The record labels allege Suno and Udio have effectively copied artists’ styles and specific musical characteristics. The labels claim the AI-generated music is so close to the original that it is eerily similar when transcribed into sheet music. The lawsuits also accuse the startups of making it easy for people to distribute AI-created samples that mimic copyrighted recordings on platforms like Spotify.

Sourcehttps://venturebeat.com/ai/record-labels-sue-ai-music-generator-startups-suno-udio-for-copyright-infringement/

Anthropic rolls out Claude’s cutting-edge collaborative features

Anthropic has introduced new collaboration features for Claude. These features include:

  • Projects: Projects in Claude allow integration of internal resources like style guides or codebases, enhancing Claude’s ability to deliver tailored assistance across various tasks. Users can set custom instructions for each Project to modify Claude’s tone or perspective for a specific role or industry.
  • Artifacts: It allows users to generate and edit various content types like code, documents, and graphics within a dedicated window. This benefits developers by offering larger code windows and live previews for easier front-end reviews.
  • Sharing Features: Claude Team users can share snapshots of their best conversations with Claude in their team’s shared project activity feed.

Additionally, any data or chats shared within Projects will not be used to train Anthropic’s generative models without a user’s explicit consent.

Sourcehttps://www.anthropic.com/news/projects

Google experiments with celebrity-inspired AI Chatbots

These chatbots will be powered by Google’s Gemini family of LLMs. The company aims to strike partnerships with influencers and celebrities and is also working on a feature that allows people to create their own chatbots by describing their personalities and appearances.

The project is led by Ryan Germick, a longtime executive at Google and a team of ten. These chatbots could be an experiment and may only appear on Google Labs rather than being widely available.

Sourcehttps://www.msn.com/en-us/news/other/google-wants-to-build-ai-chatbots-based-on-celebs-influencers-for-some-reason/ar-BB1oS1or

OpenAI postpones the launch of ChatGPT voice mode

Originally planned for late June, the Voice Mode aims to provide a more naturalistic and conversational experience with the AI chatbot, complete with emotional inflection and the ability to handle interruptions.

However, it will now be available only to a small group of users in late July or early August. OpenAI is working on improving content detection and user experience before wider rollout. GPT-4o’s real-time voice and vision capabilities are also expected to roll out to ChatGPT Plus users soon.

Sourcehttps://techcrunch.com/2024/06/25/openai-delays-chatgpts-new-voice-mode

Amazon steps into the chatbot race

Amazon is reportedly working on a new consumer-focused chatbot codenamed “Metis.” It is planned to be released somewhere around September. Here’s what we know about it:

  • The chatbot is powered by a new model, Olympus, and can be accessed via a web browser.
  • It uses a retrieval-augmented generation (RAG) technique to provide up-to-date information and automate tasks.
  • The model conversationally provides text and image-based outputs, suggesting follow-ups to queries. It also shares links to sources and supports image generation.
  • It uses an infrastructure similar to Amazon’s upcoming voice assistant, Remarkable Alexa.

Sourcehttps://www.businessinsider.com/amazon-chatgpt-rival-codenamed-metis-2024-6

Figma’s new AI features stir competition with Adobe

Figma announced a range of new features at the 2024 Config conference. Significant ones include a UI redesign, generative AI tools, new icons and toolbar, AI-enhanced asset search, and auto-generated texts in designs.

For instance, by typing a simple prompt into the textbox, users can create an entire app design mock-up for a restaurant. Figma will connect the design pages and even write suggested content!

Figma has also added a few designer-specific features to allow users to tweak designs in real-time. It features a developer mode with a “ready-for-dev” task list. The upgrade also boasts Figma slides, a Google slides-like tool for building and sharing presentations.

Sourcehttps://www.figma.com/whats-new/

Alibaba’s Qwen-72B tops the Hugging Face leaderboard

Hugging Face’s latest open large language model leaderboard ranks and evaluates open LLMs based on benchmarks like MMLU-pro and tests them on high-school and college-level problems.

The platform used 300 NVIDIA H100 GPUs to re-evaluate major open LLMs to obtain updated rankings. Chinese company Alibaba’s Qwen-72B dominated the leaderboard, becoming a top performer overall.

Not just that, the leaderboard was mainly dominated by Chinese companies, highlighting their headway into the open LLM space.

Sourcehttps://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard

Google’s Gemma 2, a set of lightweight, powerful open LLMs

Google has released Gemma 2 set of models that punch above their weight classes. Available in 9B and 27B parameter sizes, these models are

  • Higher performing and more efficient at inference than the first-generation
  • Have significant safety advancements built in
  • Optimized to run at incredible speed across a range of hardware and easily integrate with other AI tools
  • Trained on 13 trillion tokens for 27B, 8 trillion for 9B, and 2 trillion for 2.6B model (en route)

27B performs better than Llama3-70B and Nemotron-340B on Lmsys Arena, making it best in its size and stronger than some larger models. While 9B outperforms the likes of Mistral-large and Qwen1.5-110B.

The 27B Gemma 2 model is designed to run inference efficiently at full precision on a single Google Cloud TPU host, NVIDIA A100 80GB Tensor Core GPU, or NVIDIA H100 Tensor Core GPU. Moreover, this is an open weights model line, currently only available to researchers and developers.

Ace the Microsoft Azure Fundamentals AZ-900 Certification Exam: Pass the Azure Fundamentals Exam with Ease

Sourcehttps://blog.google/technology/developers/google-gemma-2

OpenAI’s CriticGPT finds GPT-4’s mistakes with GPT-4

OpenAI trained a model based on GPT-4, called CriticGPT, to catch errors in ChatGPT’s code output. It found that when users get help from CriticGPT to review ChatGPT code, they outperform those without help 60% of the time.

OpenAI aligns GPT-4 models to be more helpful and interactive through Reinforcement Learning from Human Feedback (RLHF). A key part of RLHF is collecting comparisons in which people, called AI trainers, rate different ChatGPT responses against each other.

OpenAI is beginning to integrate CriticGPT-like models into its RLHF labeling pipeline, providing trainers with explicit AI assistance.

Sourcehttps://openai.com/index/finding-gpt4s-mistakes-with-gpt-4

Google’s partnerships to help AI with real-world facts

Google is partnering with reputable third-party services, such as Moody’s, MSCI, Thomson Reuters, and Zoominfo, to ground its AI with real-world data. These four will be available within Vertex AI starting next quarter. They will offer developers qualified data to backstop their model outputs and ensure responses are factually accurate.

Google is also announcing high-fidelity grounding. Available through an experimental preview, it’s designed to help AI systems work better with a given set of specific information.

Sourcehttps://venturebeat.com/ai/google-grounding-ai-with-moodys-msci-thomson-reuters-zoominfo

A  Daily chronicle of AI Innovations June 28th 2024:

🚀 Google releases Gemma 2, a set of lightweight but powerful open LLMs
🔍 OpenAI’s CriticGPT finds GPT-4’s mistakes with GPT-4
🌐 Google partners with Moody’s, Thomson Reuters & more for AI data

🔎 OpenAI’s critics train smarter AI systems

🤝TIME and OpenAI announced a multi-year content deal

🌍Google is using AI to add 110 new languages to Google Translate

🎼YouTube is in talks with major record labels for an AI music deal

🤖Meta to start testing user-created AI chatbots on Instagram

📞Character.AI now allows users to talk with AI avatars over calls

🔎 OpenAI’s critics train smarter AI systems

OpenAI just published new research exploring how to improve model training using AI ‘critics’ that help human trainers spot errors in AI outputs more effectively.

  • OpenAI developed a model called CriticGPT to assist human trainers in evaluating and catching errors in code written by AI systems like ChatGPT.
  • In tests, CriticGPT caught 85% of inserted bugs in AI-generated code, compared to just 50% for human reviewers alone.
  • Human trainers also preferred CriticGPT’s code critics over their own 63% of the time on naturally occurring bugs.
  • OpenAI said that CriticGPT-type models are already being integrated into the company’s training workflow, helping prepare for advancing AI systems.

As OpenAI states in the research paper, both LLMs and AI critics will continue to scale — human intelligence will not. The limitations of our own puny mortal brains look to be no match for evaluating systems that are continuing to soar, meaning AI is quickly becoming its own trainer.

If you are looking for an all-in-one solution to help you prepare for the AWS Cloud Practitioner Certification Exam, look no further than this AWS Cloud Practitioner CCP CLF-C02 book

Source: https://openai.com/index/finding-gpt4s-mistakes-with-gpt-4

🚀 Google releases Gemma 2, a set of lightweight but powerful open LLMs

Google has released Gemma 2 set of models that punch above their weight classes. Available in 9B and 27B parameter sizes, these models are

  • Higher performing and more efficient at inference than the first-generation
  • Have significant safety advancements built in
  • Optimized to run at incredible speed across a range of hardware and easily integrate with other AI tools
  • Trained on 13 trillion tokens for 27B, 8 trillion for 9B, and 2 trillion for 2.6B model (en route)

27B performs better than Llama3-70B and Nemotron-340B on Lmsys Arena, making it best in its size and stronger than some larger models. While 9B outperforms the likes of Mistral-large and Qwen1.5-110B.

The 27B Gemma 2 model is designed to run inference efficiently at full precision on a single Google Cloud TPU host, NVIDIA A100 80GB Tensor Core GPU, or NVIDIA H100 Tensor Core GPU. Moreover, this is an open weights model line, currently only available to researchers and developers.

Why does it matter?

The models sound like they are built for practical deployments. They come in practical sizes so that they can be easily deployed while being amazing in quality due to best-in-class performances.

Source: https://blog.google/technology/developers/google-gemma-2/

🔍 OpenAI’s CriticGPT finds GPT-4’s mistakes with GPT-4

OpenAI trained a model based on GPT-4, called CriticGPT, to catch errors in ChatGPT’s code output. It found that when users get help from CriticGPT to review ChatGPT code, they outperform those without help 60% of the time.

OpenAI aligns GPT-4 models to be more helpful and interactive through Reinforcement Learning from Human Feedback (RLHF). A key part of RLHF is collecting comparisons in which people, called AI trainers, rate different ChatGPT responses against each other.

OpenAI is beginning to integrate CriticGPT-like models into its RLHF labeling pipeline, providing trainers with explicit AI assistance.

Why does it matter?

With more advances in reasoning and model behavior, AI models’ mistakes can become more subtle for AI trainers to spot. CriticGPT is a step towards addressing this fundamental limitation of RLHF.

Source: https://openai.com/index/finding-gpt4s-mistakes-with-gpt-4

🌐 Google partners with Moody’s, Thomson Reuters & more for AI data

Google is partnering with reputable third-party services, such as Moody’s, MSCI, Thomson Reuters, and Zoominfo, to ground its AI with real-world data. These four will be available within Vertex AI starting next quarter. They will offer developers qualified data to backstop their model outputs and ensure responses are factually accurate.

Google is also announcing high-fidelity grounding. Available through an experimental preview, it’s designed to help AI systems work better with a given set of specific information.

Why does it matter?

Earlier, Google announced efforts to ground Vertex AI results using web data and a plan to allow companies to ground AI systems in their own internal data.

Now, it is grounding these systems in known factual data from third parties, which could significantly lessen hallucinations and make AI more trustworthy for enterprise customers.

Source: https://venturebeat.com/ai/google-grounding-ai-with-moodys-msci-thomson-reuters-zoominfo

What Else Is Happening in AI on June 27th 2024❗

🤝TIME and OpenAI announced a multi-year content deal

OpenAI will gain access to current and historic content from TIME’s extensive archives from the last 101 years to enhance its products. It will also enable TIME to gain access to OpenAI’s technology to develop new products for its audiences.

Source: https://openai.com/index/strategic-content-partnership-with-time

🌍Google is using AI to add 110 new languages to Google Translate

It is Google’s largest expansion ever, thanks to its PaLM 2 LLM. It includes languages like Cantonese, NKo, and Tamazight, representing more than 614 million speakers and opening up translations for around 8% of the world’s population.

Source: https://blog.google/products/translate/google-translate-new-languages-2024/

🎼YouTube is in talks with major record labels for an AI music deal

It is offering to pay Universal Music Group (UMG), Sony Music Entertainment, and Warner Records “lump sums of cash” in exchange for legally licensing their songs to train new AI music tools. These will likely be one-off payments, not royalty-based arrangements. Source: https://www.theverge.com/2024/6/27/24187151/youtube-ai-music-deals-licensing-record-labels-sony-umg-warner

🤖Meta to start testing user-created AI chatbots on Instagram

CEO Mark Zuckerberg announced yesterday that Meta will begin to surface AI characters made by creators through Meta AI studio on Instagram, starting in the U.S. These will primarily show up in messaging for now and will be clearly labeled as AI.

Source: https://techcrunch.com/2024/06/27/meta-starts-testing-user-created-ai-chatbots-on-instagram

📞Character.AI now allows users to talk with AI avatars over calls

Users can initiate calls with a user-generated AI character directly with a button tap. Users can also switch between calling and texting seamlessly and stop the AI from talking through a “Tap to interrupt” option. The feature currently supports only a few languages.

Source: https://www.reuters.com/technology/artificial-intelligence/ai-chatbot-startup-characterai-launches-new-calls-feature-2024-06-27

📊 Use ChatGPT to turn docs into spreadsheets

With a simple prompt, ChatGPT can analyze documents, answer questions, perform calculations, and create a downloadable spreadsheet — all in one conversation!

  1. Log in to ChatGPT and upload your document to the chat (remember to remove any sensitive data before submitting).
  2. Ask questions about it, e.g., “How much is the security deposit?”
  3. Request calculations, like the total first month’s costs, including deposits and fees.
  4. Simple prompt ChatGPT to “create a downloadable budget spreadsheet”, specifying the time frame and desired columns.

Note: Only share documents you’re comfortable with. Be cautious with sensitive information and consider redacting critical details before uploading.

Source: https://university.therundown.ai/c/daily-tutorials/from-lease-agreement-to-budget-spreadsheet-with-chatgpt-0ed8d7ba-7d5a-4e9f-903a-64822caf1579

A  Daily chronicle of AI Innovations June 27th 2024:

🎵 YouTube wants to pay record labels to use their songs for AI training

🎨 Figma unveils AI-powered design tools, challenges Adobe’s dominance

🗣️ Character.AI introduces AI avatar calls

🏁 Amazon steps into the chatbot race 

🥇 Alibaba’s Qwen-72B tops the Hugging Face leaderboard

🏆 Hugging Face updates Open LLM Leaderboard

🎙️ NBC to use AI voice for Olympics recaps!

💡 Sonia’s ‘AI Therapist’ sparks debate

🏛️ Anthropic extends Claude’s access to government

🎵 YouTube wants to pay record labels to use their songs for AI training

  • YouTube is reportedly negotiating with major music labels like Sony Music Entertainment, Universal Music Group, and Warner Records to pay for licensing their songs for AI training.
  • Despite these negotiations, the ultimate decision to allow the use of songs for AI training will likely fall to individual artists rather than the labels themselves.
  • In April 2023, more than 200 artists voiced their opposition to AI using their work, emphasizing the potential threat to creators’ rights and the music ecosystem.

Source: https://www.engadget.com/youtube-reportedly-wants-to-pay-record-labels-to-use-their-songs-for-ai-training-125052503.html

🎨 Figma unveils AI-powered design tools, challenges Adobe’s dominance

  • Figma has launched AI-powered design features to speed up workflows and aid teams in creating superior software, positioning it as a product development platform to compete with Adobe.
  • The new Figma AI features include design creation with text prompts, visual search capabilities, layer renaming, and automating repetitive tasks, and are currently available in a limited beta.
  • While beta access to Figma AI is free, later availability will require subscriptions to specific plans, and admins must enable AI features, differentiating Figma’s approach from Adobe’s controversial AI terms.

Source: https://venturebeat.com/ai/figma-unveils-ai-powered-design-tools-to-rival-adobe/

🗣️ Character.AI introduces AI avatar calls

  • Character.AI announced that users can now make calls to AI characters, supporting multiple languages including English, Spanish, Portuguese, Russian, Korean, Japanese, and Chinese.
  • The company reported that during testing, over 3 million users made 20 million calls, and noted that AI character calls can help with language practice, mock interviews, or role-playing games.
  • With a button tap, users can start calls with AI characters, seamlessly switch between calling and texting, and use a “Tap to interrupt” feature to stop the AI from talking.

Source: https://techcrunch.com/2024/06/27/character-ai-now-allows-users-to-talk-with-avatars-over-calls/

🏁 Amazon steps into the chatbot race

Amazon is reportedly working on a new consumer-focused chatbot codenamed “Metis.” It is planned to be released somewhere around September. Here’s what we know about it:

  • The chatbot is powered by a new model, Olympus, and can be accessed via a web browser.
  • It uses a retrieval-augmented generation (RAG) technique to provide up-to-date information and automate tasks.
  • The model conversationally provides text and image-based outputs, suggesting follow-ups to queries. It also shares links to sources and supports image generation.
  • It uses an infrastructure similar to Amazon’s upcoming voice assistant, Remarkable Alexa.

Why does it matter?

The AI assistant market is already too crowded, making Metis’ release seem a bit too late. Its rivals, OpenAI and Google, have already launched their models in 2023 and will give tough competition to Amazon. If Amazon wishes to win this AI race, Metis must showcase exceptional capabilities.

Source: https://www.businessinsider.com/amazon-chatgpt-rival-codenamed-metis-2024-6

 
🎨 Figma’s new AI features stir competition with Adobe

Figma announced a range of new features at the 2024 Config conference. Significant ones include a UI redesign, generative AI tools, new icons and toolbar, AI-enhanced asset search, and auto-generated texts in designs.

https://x.com/figma/status/1806016455609286847

For instance, by typing a simple prompt into the textbox, users can create an entire app design mock-up for a restaurant. Figma will connect the design pages and even write suggested content!

Figma has also added a few designer-specific features to allow users to tweak designs in real-time. It features a developer mode with a “ready-for-dev” task list. The upgrade also boasts Figma slides, a Google slides-like tool for building and sharing presentations.

Why does it matter?

Figma’s AI tools will enable newer users on the platform to test out ideas easily. At the same time, tools like “ready-for-dev” will offer quick iterations to users who are more familiar with the platform.

With Adobe abandoning its plans to acquire Figma, the company’s recent upgrade might pose a challenge to the former.

Source: https://www.figma.com/whats-new/

🥇 Alibaba’s Qwen-72B tops the Hugging Face leaderboard

Hugging Face’s latest open large language model leaderboard ranks and evaluates open LLMs based on benchmarks like MMLU-pro and tests them on high-school and college-level problems.

The platform used 300 NVIDIA H100 GPUs to re-evaluate major open LLMs to obtain updated rankings. Chinese company Alibaba’s Qwen-72B dominated the leaderboard, becoming a top performer overall.

Not just that, the leaderboard was mainly dominated by Chinese companies, highlighting their headway into the open LLM space.

Why does it matter?

The open LLM leaderboard is essential for transparent and comprehensive LLM evaluation. It also proves that a model’s size doesn’t guarantee its superior performance. As Chinese models deliver exceptional performance, will well-known models from the USA, like Meta’s Llama, be replaced in the coming years?

Source: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard

What Else Is Happening in AI on June 27th 2024❗

🏛️ Anthropic extends Claude’s access to government

Anthropic is making Claude 3 Haiku and Claude 3 Sonnet available for government use through the AWS marketplace. This will improve citizen services, streamline document reviews, support data-based policymaking, and generate realistic training scenarios.

Source: https://www.anthropic.com/news/expanding-access-to-claude-for-government?

🌐 Opera announces new themes and features for One R2

Major features include AI-based image and voice generation, image understanding capabilities, floating windows, new themes, design elements, and split window tabs.

Source: https://press.opera.com/2024/06/26/opera-unveils-opera-one-r2-in-developer

🎙️ NBC to use AI voice for Olympics recaps!

The AI will offer daily, personalized coverage of the 2024 Paris Olympics on the Peacock app in the voice of the famous sports commentator, AI Michael. According to NBC, the voice AI is trained on data from Michael’s appearances on the show.

Source: https://www.nbcnews.com/news/nbc-use-ai-generated-version-al-michaels-voice-summer-olympics-rcna159108

📹 Caption’s new feature can auto-edit videos!

The feature adds custom graphics, zooms, music, sound effects, transitions, and motion backgrounds to an unedited video. Users can also use Caption’s AI avatars to create videos and fully edit them with the AI edit feature.

Source: https://www.captions.ai/

💡 Sonia’s ‘AI Therapist’ sparks debate

The personalized AI therapist can deal with depression, stress, anxiety, poor sleep, and more by analyzing what users say during therapy sessions. The release has sparked a debate around user data privacy, its susceptibility to bias, and the accuracy of its diagnosis. Source: https://www.soniahealth.com/

🏆 Hugging Face updates Open LLM Leaderboard

Hugging Face just introduced a new upgrade to its Open LLM Leaderboard, adding new benchmarks and evaluation methods to help address the recent plateau in LLM performance gains.

  • The leaderboard now features six new benchmarks designed to be more challenging and less prone to contamination.
  • Initial rankings show Qwen2-72B-Instruct leading the pack, followed by Meta’s Llama-3-70B-Instruct and Mixtral 8×22b.
  • A new normalized scoring system adjusts for baseline performance, providing a more fair comparison across different evaluation types.
  • The upgrade also introduces a ‘maintainer’s highlight’ category and community voting system to prioritize the most relevant models.

As LLMs approach human-level performance on most tasks, finding new ways to evaluate them is becoming more difficult — and more crucial. This revamp helps guide researchers and developers towards more targeted improvements, providing a more nuanced assessment of model capabilities.

Source: https://huggingface.co/spaces/open-llm-leaderboard/blog

🚨 Rabbit R1 hit with major security flaw

A group of developers just discovered a major vulnerability in Rabbit’s R1 AI assistant device, potentially exposing user’s private data and chat responses.

  • A community-led group called Rabbitude uncovered hardcoded API keys in Rabbit’s codebase, which allowed access to all R1 responses.
  • The group gained access to the codebase in mid-May, saying the Rabbit team was aware of the issue but failed to take action.
  • Rabbitude said the vulnerability could allow bad actors to disable all r1 devices, alter voices and responses, and access private messages.
  • Rabbit acknowledged an ‘alleged data breach’ via a Discord post, but claims no customer data was leaked.

Despite massive hype in the first wave of consumer AI standalone devices, the Rabbit r1 has been nothing short of a disaster so far. Already facing major criticism over the companion’s limited capabilities, this security breach only furthers the skepticism surrounding the early AI hardware market entrants.

Source: https://rabbitu.de/articles/security-disclosure-1?

A  Daily chronicle of AI Innovations June 26th 2024:

💼 Anthropic rolls out Claude’s cutting-edge collaborative features
🤖 Google experiments with celebrity-inspired AI Chatbots
🛑 OpenAI postpones the launch of ChatGPT voice mode

🧬 New AI simulates 500M years of evolution

⏰ OpenAI delays ChatGPT voice upgrade

💼 Anthropic rolls out Claude’s cutting-edge collaborative features

Anthropic has introduced new collaboration features for Claude. These features include:

  • Projects: Projects in Claude allow integration of internal resources like style guides or codebases, enhancing Claude’s ability to deliver tailored assistance across various tasks. Users can set custom instructions for each Project to modify Claude’s tone or perspective for a specific role or industry.
  • Artifacts: It allows users to generate and edit various content types like code, documents, and graphics within a dedicated window. This benefits developers by offering larger code windows and live previews for easier front-end reviews.
  • Sharing Features: Claude Team users can share snapshots of their best conversations with Claude in their team’s shared project activity feed.

Additionally, any data or chats shared within Projects will not be used to train Anthropic’s generative models without a user’s explicit consent.

Why does it matter?

By offering advanced AI collaboration tools, Anthropic meets critical enterprise needs for flexibility and team efficiency, where many of its competitors fall short. This approach could broaden its user base, increasing user retention and market penetration.

Source: https://www.anthropic.com/news/projects

🤖 Google experiments with celebrity-inspired AI Chatbots

These chatbots will be powered by Google’s Gemini family of LLMs. The company aims to strike partnerships with influencers and celebrities and is also working on a feature that allows people to create their own chatbots by describing their personalities and appearances.

The project is led by Ryan Germick, a longtime executive at Google and a team of ten. These chatbots could be an experiment and may only appear on Google Labs rather than being widely available.

Why does it matter?

Google’s AI chatbots featuring celebrities and influencers enhance user engagement but raise ethical concerns about consent, boundaries, and legal and privacy issues as AI evolves.

Source: https://www.msn.com/en-us/news/other/google-wants-to-build-ai-chatbots-based-on-celebs-influencers-for-some-reason/ar-BB1oS1or

🛑 OpenAI postpones the launch of ChatGPT voice mode

Originally planned for late June, the Voice Mode aims to provide a more naturalistic and conversational experience with the AI chatbot, complete with emotional inflection and the ability to handle interruptions.

However, it will now be available only to a small group of users in late July or early August. OpenAI is working on improving content detection and user experience before wider rollout. GPT-4o’s real-time voice and vision capabilities are also expected to roll out to ChatGPT Plus users soon.

Why does it matter?

Delays can give competitors like Anthropic a chance to gain an edge. If rivals manage to launch similar or superior voice capabilities sooner, OpenAI risks losing potential users and market share.

Source: https://techcrunch.com/2024/06/25/openai-delays-chatgpts-new-voice-mode

🧬New AI simulates 500M years of evolution

Biotech startup EvolutionaryScale just introduced ESM3, a new AI model capable of designing and simulating new proteins in ways that mimic millions of years of natural evolution.

  • ESM3 is a frontier language model that can reason over protein sequences, structures, and functions simultaneously.
  • In tests, ESM3 generated a new fluorescent protein 58% different from other known structures — a leap that typically takes nature 500M years of evolution.
  • The largest version of ESM3 has 98B parameters, and was trained on approximately 1.8T tokens of biological data.
  • EvolutionaryScale was created by former META researchers, and has already secured $142M in seed funding from tech giants like Amazon and Nvidia.

ESM3 is bringing us closer to a future where we can ‘program’ biology as easily as we write computer code — potentially revolutionizing fields from healthcare to environmental science and unlocking completely new areas of knowledge across fields.

Source: https://www.evolutionaryscale.ai/blog/esm3-release

What Else Is Happening in AI on July 26th 2024❗

🧬 Ex-Meta engineers launch protein-generating AI biology model

EvolutionaryScale, launched by ex-Meta engineers, introduced ESM3, a gen AI model for designing novel proteins. It uses AI to simulate evolutionary processes, aiming to revolutionize industries like medicine and biofuels. It is available via API to selected customers and non-commercial users.

Source: https://venturebeat.com/ai/meta-alum-launches-ai-biology-model-that-simulates-500-million-years-of-evolution

🛑 Reddit sets new barriers for AI crawlers to protect its content

Reddit will rate-limit or block unknown bots and crawlers that violate its Public Content Policy or lack an agreement with the platform. These changes aim to deter AI companies from training LLMs on Reddit content (excluding the $60 million deal with Google).

Source: https://techcrunch.com/2024/06/25/reddits-upcoming-changes-attempt-to-safeguard-the-platform-against-ai-crawlers

🎥 AI-generated ad film debuts at Cannes 2024 

Toys ‘R’ Us, the iconic toy store chain, collaborated with OpenAI’s Sora  to produce a brand film called “The Origin of Toys ‘R’ Us.” The film premiered at the 2024 Cannes Lions International Festival of Creativity.

Source: https://venturebeat.com/ai/toys-r-us-unveils-first-commercial-made-with-openais-sora

🦊 Mozilla’s Firefox Nightly now lets users pick from leading AI chatbots

It allows users to choose from multiple AI chatbots, such as ChatGPT, Google Gemini, HuggingChat, and Le Chat Mistral. The feature lets users select their preferred AI to perform tasks such as summarizing web content, simplifying text, or conducting quizzes.

Source: https://techcrunch.com/2024/06/25/firefox-now-lets-you-choose-your-preferred-ai-chatbot-in-its-nightly-builds

📧 Google extends Gemini AI to Gmail to streamline email tasks

Google has integrated its Gemini AI into Gmail for Workspace subscribers, enhancing email functionality. It assists users by automatically drafting emails, summarizing threads, crafting responses, and retrieving information from emails or Google Drive.

Source: https://techcrunch.com/2024/06/25/google-brings-its-gemini-ai-to-gmail-via-a-sidebar-that-can-help-you-write-and-summarize-emails

A  Daily chronicle of AI Innovations June 25th 2024:

🤖 Amazon is working on a ChatGPT competitor

🎤 Google is building AI chatbots modeled after celebrities, YouTube influencers

📉 Nvidia’s market value plummets drastically

🤷‍♂️ Apple rejected Meta’s AI integration over privacy issues

🤝 OpenAI buys a remote collaboration platform

🎥 Synthesia 2.0: World’s 1st AI video communication platform
🤝 OpenAI is on an acquiring spree, buying Rocket and Multi
🎵 Record labels sue AI music startups over copyright infringement

🤖 Amazon is working on a ChatGPT competitor

  • Amazon is developing an AI chatbot, code-named “Metis,” to compete with OpenAI’s ChatGPT, and it could be launched later this year, based on a report from Business Insider.
  • The upcoming chatbot will be accessed through a web browser and rely on a new foundational model, as mentioned by sources and documents cited by Business Insider.
  • Amazon’s CEO Andy Jassy is reportedly directly involved with “Metis,” reflecting Amazon’s broader strategy to intensify its AI efforts across various applications and services.
  • Source: https://www.geekwire.com/2024/report-amazon-developing-ai-chatbot-that-would-compete-with-chatgpt-and-others/

🎤 Google is building AI chatbots modeled after celebrities, YouTube influencers

  • Google is reportedly developing new AI chatbots modeled after celebrities and YouTube influencers, aiming to create specialized interactive experiences.
  • The initiative will use Google’s Gemini family of large language models and seeks partnerships with influencers and celebrities, allowing users to design their own chatbots by describing traits and appearance.
  • The project, led by executive Ryan Germick and a team of ten, might initially be available only on Google Labs, keeping the bots experimental and not widely accessible at launch.
  • Source: https://www.engadget.com/google-is-reportedly-building-ai-chatbots-based-on-celebrities-and-influencers-235731655.html

📉 Nvidia’s market value plummets drastically

  • Nvidia’s shares dropped nearly 13% in a week, causing a loss of $430 billion from its market value and dropping its market capitalization to $2.91 trillion.
  • This decline has moved Nvidia from the top position to third place among the world’s most valuable companies, now trailing behind Microsoft and Apple.
  • Investor enthusiasm for Nvidia’s role in AI is waning after significant stock gains, leading to a 6.7% drop on Monday, the third consecutive day of declines for the company.
  • Source: https://edition.cnn.com/2024/06/25/tech/nvidia-stock-fall-worlds-third-biggest-company/index.html

🤷‍♂️ Apple rejected Meta’s AI integration over privacy issues

  • Apple declined a potential AI partnership with Meta concerning the integration of Meta’s large language model into iOS, primarily due to privacy concerns highlighted in a Bloomberg report.
  • While initial discussions between Apple and Meta took place, they did not advance, with Apple choosing not to pursue formal talks or integration due to Meta’s insufficient privacy practices.
  • Despite rejecting Meta, Apple has engaged in collaborations with other major AI entities, finalizing a deal with OpenAI to integrate ChatGPT into iOS 18 and discussing potential integrations with Google and Anthropic for diverse AI model options.
  • Source: https://www.macrumors.com/2024/06/24/apple-meta-ai-privacy-concerns/

🤝 OpenAI buys a remote collaboration platform

  • OpenAI is advancing its AI technology by integrating it into desktop computing, moving beyond browser-based interactions with the acquisition of startup Multi.
  • By purchasing Multi, OpenAI aims to develop potentially a new AI-based operating system or at least an interaction layer for current systems, as suggested by recent product releases including a ChatGPT desktop app.
  • OpenAI’s future AI assistant could replace traditional browsers, altering the role of search engines and websites, and revolutionizing how we interact with online content.
  • Source: https://the-decoder.com/openai-moves-toward-its-own-ai-operating-system-and-a-post-browser-world/

 Synthesia 2.0: World’s 1st AI video communication platform

Synthesia is launching Synthesia 2.0 – the world’s first AI video communications platform for businesses. It reinvents the entire video production process, allowing companies to create and share AI-generated videos at scale easily.

https://youtu.be/gZaBwdru_bk?si=fP7_nILeu66Oqj4B

The key new features and capabilities of Synthesia 2.0 include:

  • 2 Personal AI Avatars: Expressive Avatars shot in a studio and Custom Avatars created using your webcam.
  • AI Video Assistant: Converts text, documents, or websites into high-quality videos, with options to customize the branding, tone, and length.
  • Intuitive Video Editing: Editing simplified with “Triggers” that let you control animations and edits from the script.
  • Translation and Dynamic Video Player: Videos can now be translated into over 120 languages. Synthesia is also building a new video player with interactive features.
  • AI Safety Focus: Synthesia is pursuing ISO/IEC 42001 certification, the first standard for responsible AI management, to ensure its AI technologies are ethical.

Why does it matter?

By making video creation and distribution more accessible and automated, Synthesia aims to help businesses transition to a “video-first” approach to communicate more effectively, improve employee training, customer service, and marketing efforts, and streamline various business operations.

Source: https://www.synthesia.io/post/introducing-synthesia-video-communications-platform

🤝 OpenAI is on an acquiring spree, buying Rocket and Multi

Last week, OpenAI acquired Rockset, a startup that develops tools for real-time data search and analytics. OpenAI said it would integrate Rockset’s technology to power its infrastructure and offerings across products.

This week, OpenAI acquired Multi, a startup focused on building remote collaboration tools and software. Technically, the deal is an acqui-hire as the entire Multi team, including its co-founders, will join OpenAI to work on the company’s ChatGPT desktop application.

Why does it matter?

With over 600,000 paid business users, including most Fortune 500 companies, OpenAI’s back-to-back acquisitions strategically position it to expand its enterprise-focused AI products and make its powerful language models more useful for businesses.

Source: https://techcrunch.com/2024/06/24/openai-buys-a-remote-collaboration-platform

🎵 Record labels sue AI music startups over copyright infringement

The world’s major record labels, including Universal Music Group, Sony Music, and Warner Music, have filed twin lawsuits against the AI music generation startups Suno and Udio. The lawsuits accuse the companies of unlawfully training their AI models on massive amounts of copyrighted music, which, according to the complaints, allows the startups to generate similar-sounding music without permission.

The record labels allege Suno and Udio have effectively copied artists’ styles and specific musical characteristics. The labels claim the AI-generated music is so close to the original that it is eerily similar when transcribed into sheet music. The lawsuits also accuse the startups of making it easy for people to distribute AI-created samples that mimic copyrighted recordings on platforms like Spotify.

Why does it matter?

This legal battle is only the latest example, and one of many, of AI technology colliding with creative industries. However, any resolution of these battles will likely have far-reaching implications for the music industry and the development of AI-powered music tools.

Source: https://venturebeat.com/ai/record-labels-sue-ai-music-generator-startups-suno-udio-for-copyright-infringement

What Else Is Happening in AI on June 25th❗

🍎 Apple withdraws AI collaboration deal with Meta over privacy concerns

Contrary to the rumors about a potential AI collaboration, Apple is not planning a partnership to integrate Meta’s AI models into its products due to privacy concerns. Instead, Apple is focusing on partnerships with OpenAI and Google that align with its commitment to user privacy.

Source: https://techcrunch.com/2024/06/24/apple-shelved-the-idea-of-integrating-metas-ai-models-over-privacy-concerns-report-says

🛍️ Shopify expands its AI capabilities to attract more businesses

Shopify has expanded access to its AI-powered features, including the Sidekick assistant and image-generation tool. It is making these AI tools available to more users, as it aims to provide merchants with an integrated and seamless platform to scale their businesses.

Source: https://www.reuters.com/technology/artificial-intelligence/shopify-expands-access-its-ai-powered-features-attract-more-businesses-2024-06-24

🏫 Google is bringing Gemini access to teens through school accounts

Google now makes its AI assistant Gemini available to teenage students through school accounts. This move is aimed at helping prepare students for a future where generative AI is more prevalent. Google says Gemini will provide real-time feedback and support to students while including safeguards to prevent inappropriate responses.

Source: https://techcrunch.com/2024/06/24/google-is-bringing-gemini-access-to-teens-using-their-school-accounts

🤖 NVIDIA deploys its AI tech in the Middle East among U.S AI export curb

NVIDIA has struck a deal with Qatari telecom provider Ooredoo to deploy its AI technology in the Middle East. This move comes amid U.S. export restrictions on advanced AI chips to some countries in the region. Ooredoo will be the first company in the Middle East to offer Nvidia’s AI and graphics technology to its clients across multiple countries.

Source: https://www.reuters.com/technology/nvidia-launch-middle-east-amid-us-curbs-ai-exports-region-ooredoo-ceo-says-2024-06-23

📷 Meta marks real photos with ‘Made by AI’ label

Meta has been incorrectly labeling real photos as “Made by AI.” Multiple photographers have reported that Meta has added this label to their genuine photographs, even when they have only made minor edits using tools like Adobe’s Generative Fill. Meta says it relies on industry standards to detect AI-generated content, leading to false positives.

Source: https://www.theverge.com/2024/6/24/24184795/meta-instagram-incorrect-made-by-ai-photo-labels

A  Daily chronicle of AI Innovations June 24th 2024:

🤝 Apple and Meta are discussing gen AI partnership
🤖 ByteDance and Broadcom collaborate on AI chip development
🔍 🕵️‍♂️ Researchers developed a new method to detect hallucinations

🤝 Apple in talks with Meta for potential AI integration

🎵 Record labels sue AI music companies

🩸 AI blood test for Parkinson’s detection

Researchers just developed an AI-powered blood test that can predict Parkinson’s disease up to 7 years before symptoms appear, potentially revolutionizing early diagnosis and treatment.

  • The AI analyzes key markers in blood samples linked to inflammation and protein degradation, attempting to predict which patients would develop the disease.
  • In a 10-year study, the test correctly identified 79% of patients who later developed Parkinson’s.
  • The AI model predicted Parkinson’s up to 7 years before clinical symptoms emerged in some cases.
  • Researchers hope to develop a simple finger-prick version for widespread screenings.

AI’s ability to find subtle patterns in previously undetectable biological data is unlocking new diagnostic capabilities. By catching a debilitating disease years before the onset of symptoms, patients can get earlier access to treatment aimed at prevention.

Source: https://www.bbc.com/news/articles/c722yly6l6ro

🤝 Apple in talks with Meta for potential AI integration

  • Apple is reportedly negotiating with Meta to integrate Meta’s generative AI model into Apple’s new AI system, Apple Intelligence, according to The Wall Street Journal.
  • Apple is seeking partnerships with multiple AI companies, including Meta, to enhance its AI capabilities and catch up in the competitive AI race.
  • A potential collaboration between Apple and Meta would be significant due to their history of disagreements, and it could greatly impact the AI industry if successful.
  • Source: https://www.benzinga.com/news/24/06/39447797/apple-eyes-meta-for-potentially-game-changing-ai-partnership-report

🎵 Record labels sue AI music companies

  • Universal Music Group, Sony Music Entertainment, and Warner Records are suing AI companies Suno and Udio for massive copyright infringement related to the creation of original songs using generative AI technology.
  • The music labels claim that Suno and Udio used copyrighted works without permission, and are seeking damages of up to $150,000 per violated work in lawsuits filed in Boston and New York federal courts.
  • This legal action is part of a broader conflict between the music industry and tech companies offering AI tools, highlighting concerns about AI-generated music impacting the control and revenue of human artists.
  • Source: https://www.theverge.com/2024/6/24/24184710/riaa-ai-lawsuit-suno-udio-copyright-umg-sony-warner

🤝 Apple and Meta are discussing gen AI partnership

Apple is reportedly in talks with its longtime rival Meta to integrate the latter’s Llama 3 AI model into Apple Intelligence. This move comes as Apple prepares to roll out its AI features across iPhones, iPads, and Macs later this year.

The potential partnership follows Apple’s existing deal with OpenAI, suggesting a collaboration strategy rather than solo development in the AI race. In Apple’s arrangement with OpenAI, there’s no direct payment. Instead, OpenAI can offer premium subscriptions through Apple Intelligence, with Apple taking a percentage. It’s unclear if Meta would agree to a similar business model, given that Llama 3 is open-source and free to access.

Why does it matter?

Despite their historical rivalry, especially after Apple’s 2021 privacy changes impacted Meta’s ad revenue, this potential partnership suggests a willingness to cooperate in the AI race. However, not everyone is thrilled about it. Users and investors are voicing concerns about privacy, with Meta’s track record being a major sticking point. Some see this as a desperate move by Apple.

Source: https://www.wsj.com/tech/ai/apple-meta-have-discussed-an-ai-partnership-cc57437e

🤖 ByteDance and Broadcom collaborate on AI chip development

ByteDance is collaborating with U.S. chip designer Broadcom to develop an advanced AI processor. This partnership aims to secure a stable supply of high-end chips amid ongoing U.S.-China tensions. The project centers on creating a 5-nanometre, customized Application-Specific Integrated Chip (ASIC) that complies with U.S. export restrictions.

This chip’s manufacturing is set to be outsourced to Taiwan Semiconductor Manufacturing Company (TSMC), though production is not expected to begin this year. While the design work is currently underway, the critical “tape out” phase has yet to commence.

Why does it matter?

This partnership shows how Chinese tech companies are getting creative to secure advanced AI chips. As ByteDance has launched a major push in generative AI, this new initiative brings dual benefits: reducing procurement costs and ensuring a stable supply of higher-end chips.

Source: https://www.reuters.com/technology/artificial-intelligence/chinas-bytedance-working-with-broadcom-develop-advanced-ai-chip-sources-say-2024-06-24

🔍 🕵️‍♂️ Researchers developed a new method to detect hallucinations

ChatGPT and Gemini can produce impressive results but often “hallucinate” false or unsubstantiated information. This research focuses on a subset of hallucinations called “confabulations,” where LLMs generate answers that are both wrong and arbitrary. Researchers have developed new methods to detect confabulations using entropy-based uncertainty estimators. They introduce the concept of “semantic entropy” to measure the uncertainty of LLM generations at the meaning level.

High semantic entropy corresponds to high uncertainty and indicates a higher likelihood of confabulation. The method computes uncertainty at the level of meaning rather than specific word sequences, addressing the fact that one idea can be expressed in many ways. The method provides scalable oversight by detecting confabulations that people might otherwise find plausible.

Why does it matter?

Detecting confabulations can improve question-answering accuracy for state-of-the-art LLMs. The method helps users understand when to be cautious with LLM outputs and opens up new possibilities for using LLMs that were previously limited by unreliability.

Source: https://www.nature.com/articles/s41586-024-07421-0

What Else Is Happening in AI on June 24th 2024❗

🆕Stability AI has appointed a new CEO

Prem Akkaraju, former CEO of Weta Digital, will step up to this role. This change comes after founder Emad Mostaque resigned in March amid financial concerns. Akkaraju is part of a new investor group, including former Facebook President Sean Parker, that provides a cash infusion to Stability AI. Source: https://www.theinformation.com/articles/stability-ai-gets-new-ceo-and-a-bailout-from-investor-group-led-by-sean-parker

🖼️ Coders spotted Midjourney references in Grok’s standalone version

It hints at a possible collaboration between xAI and Midjourney for image generation. Grok might use Midjourney models via API to generate images. This could mean that the standalone Grok might let users upload images. All this could be part of the Grok 1.5 rollout. Source: https://www.bolnews.com/technology/2024/06/grok-1-5-potential-use-of-midjourney-for-image-generation-gains-evidence

🚫YouTube fights AI fakes with new policy 

YouTube expanded its privacy policy to cover AI-generated content. Now, users can request the removal of synthetic media that imitates their face or voice. However, privacy complaints alone may not guarantee content removal. While reviewing requests, YouTube will consider the potential for confusion and intent (parody/satire). Source: https://support.google.com/youtube/thread/281049221

✨ OmniAI transforms unstructured enterprise data for AI and analytics use

Founded by Y Combinator alums, OmniAI syncs with your existing storage, prep the data, and lets you run your choice of AI models on the data. It offers integrations with popular models like Meta’s Llama 3 and Anthropic’s Claude. Source: https://techcrunch.com/2024/06/22/omniai-transforms-business-data-for-ai

🤖 The new Dot AI chatbot wants to be your companion 

Dot chatbot aims to personalize interactions based on users’ thoughts and feelings. It uses multiple AI models from OpenAI and Google. The app uses intensive Q&A to learn about users’ interests and experiences. Dot offers a subscription model at $11.99/month for unlimited conversations. Source: https://techcrunch.com/2024/06/21/dots-ai-really-really-wants-to-get-to-know-you

Y Combinator joined over 140 AI startup founders in signing a letter opposing California’s Senate Bill 1047 on AI regulation, arguing it harms the ability to retain talent and stifles innovation. Source: https://www.politico.com/newsletters/california-playbook/2024/06/21/little-tech-brings-a-big-flex-to-sacramento-00164369
Formula 1 and AWS unveiled ‘Statbot’, an AI-powered system that provides real-time context, trivia, and predictions for personalized race broadcasts. Source: https://www.bloomberg.com/news/articles/2024-06-23/formula-1-and-amazon-aim-for-ai-powered-personalized-race-viewing
OpenAI announced the acquisition of Rockset, with the search and database analytics startup’s tech set to boost retrieval infrastructure for enterprise products. https://openai.com/index/openai-acquires-rockset
Axel Peytavin demonstrated a small, autonomous robot prototype powered by GPT-4o that can see, think and act based on prompts, with capabilities like picking up trash and shaking hands. Source: https://x.com/ax_pey/status/1804209628680720746
OpenAI CTO Mira Murati said in an interview that the next generation of GPT will be ‘PhD-level’ compared to GPT-3 (toddler) and GPT-4 (high school), and will be released in a year and a half. Source: https://x.com/tsarnick/status/1803901130130497952
Unitree posted new demos of its Go2 robotic dog performing handstands, flips, and withstanding and adjusting to force, showing extreme precision and agility. Source: https://x.com/UnitreeRobotics/status/1803728664213012520

A  Daily chronicle of AI Innovations June 20th 2024:

🖥️ Microsoft debuts a vision-foundational model for diverse tasks
👨‍💼 Ex-OpenAI co-founder launches own AI company
🤖 Can AI read minds? New model can predict anxiety levels 

😱 Sutskever launches new AI venture

🚧 Tech giants team up on AI factory for xAI

🚨 Perplexity faces legal action

🤖 OpenAI co-founder announces new AI company

🚀 Anthropic announces its most powerful AI yet

🇪🇺 EU Council withdraws Chat Control vote

🚧 Apple Intelligence plans face legal challenges in China

🖥️ Microsoft debuts a vision-foundational model for diverse tasks

Florence 2, Microsoft’s vision-AI model, can perform diverse tasks like object detection, captioning, visual grounding, and visual question answering via image and text prompts. It displays excellent captioning, object detection, visual grounding, and segmentation.

The model comes in 232M and 771M parameter sizes and uses a sequence-to-sequence architecture, enabling multiple vision tasks without needing a task-specific architecture modification.

On fine-tuning the model with publicly available human-annotated data, Florence 2 showcased impressive results, offering tough competition to existing large vision models like Flamingo despite its compact size.

Why does it matter?

The model will equip enterprises with a standard approach to handling various vision-handling applications. This will save resources spent on separate task-specific vision models that need fine-tuning. Moreover, it may also be useful to developers as it would eliminate the need for separate vision models for smaller tasks, significantly saving compute costs.

Source: https://www.microsoft.com/en-us/research/publication/florence-2-advancing-a-unified-representation-for-a-variety-of-vision-tasks

👨‍💼 Ex-OpenAI co-founder launches own AI company

Just a month after leaving OpenAI, ex-cofounder Ilya Sutskever has launched his own AI company, Safe Superintelligence Inc. (SSI), alongside former Y Combinator partner Daniel Gross and ex-OpenAI engineer Daniel Levy as co-founders.

According to the SSI’s launch statement on X, the company will prioritize safety, progress, and security. Sutskever also emphasizes that the company’s “singular focus” on a joint approach to safety and capabilities will prevent it from being distracted by management overhead or production cycles, unlike companies like OpenAI or Google.

Why does it matter?

SSI’s launch clearly marks the emergence of a new key player in the race to build safe, powerful AI. Its mission statement emphasizes safety and the potential for groundbreaking developments that may shape the future of AI research and development. It would be interesting to see whether the startup will uphold its mission statement in the coming days.

Source: https://ssi.inc/

🤖 Can AI read minds? New model can predict anxiety levels 

Researchers at the University of Cincinnati have developed an AI model that can identify people with an urgent risk of anxiety. The AI model uses minimal computational resources, a short picture rating task, and a small set of variables to make the prediction. The approach named “Comp Cog AI” integrates computational cognition and AI.

Participants rated 48 pictures with mildly emotional subject matter based on the degree to which they liked or disliked those pictures. The response data was then used to quantify the mathematical features of their judgments. Finally, the data was combined with ML algorithms to identify their anxiety levels.

Since the technology doesn’t rely on a native language, it is accessible to a wider audience and diverse settings to assess anxiety.

Why does it matter?

The picture rating feature can provide unbiased data to medical professionals on a person’s mental health status without subjecting them to direct questions that may trigger negative emotions. Given its 81% accuracy rate, the tool can become a useful app for detecting individuals with high anxiety risks.

Source: https://www.nature.com/articles/s44184-024-00074-x.epdf

🚀 Anthropic announces its most powerful AI yet

  • Anthropic has launched Claude 3.5 Sonnet, a new AI model that aims to be on par with, or superior to, OpenAI’s GPT-4o and Google’s Gemini across various tasks.
  • Claude 3.5 Sonnet claims to be significantly faster than its predecessor and outperforms it, even surpassing other leading models in multiple benchmarks.
  • Alongside the new model, Anthropic introduced the Artifacts feature, which allows users to interact with and edit Claude’s outputs directly within the app, enhancing its functionality beyond a typical chatbot.
  • Source: https://www.theverge.com/2024/6/20/24181961/anthropic-claude-35-sonnet-model-ai-launch

🚧 Apple Intelligence plans face legal challenges in China

  • Apple’s push to enhance its artificial intelligence capabilities faces significant challenges in China due to the country’s stringent AI regulations and censorship laws.
  • China’s strict regulation of AI, including rules about language models and data protection, means Apple must navigate approvals and potentially partner with local tech giants like Baidu and Alibaba.
  • Apple’s focus on user privacy and the need for localized AI experiences may help it gain regulatory approval but also present significant hurdles in ensuring compliance and acceptance in the Chinese market.
  • Source: https://www.cnbc.com/2024/06/20/apple-ai-push-faces-big-challenges-in-china.html

🚧 Tech giants team up on AI factory for xAI

 Dell, Nvidia, and Super Micro Computer are collaborating to build an AI factory for xAI’s massive supercomputer, which will train and scale its Grok AI chatbot.

  • Dell is assembling half of the server racks for xAI’s planned supercomputer, with Super Micro handling the other half.
  • The system aims to use up to 100,000 Nvidia H100 GPUs, potentially making it 4x larger than the biggest existing AI clusters.
  • Musk wants the supercomputer to be fully operational by Fall 2025 to accelerate the development of advanced Grok models.

This powerhouse partnership shows the staggering scale of compute power required to push the boundaries of AI. With huge capital flowing in, this massive cluster could provide a critical edge in evolving Grok to be a serious competitor to the likes of OpenAI and Anthropic.

Source: https://www.fastcompany.com/91143745/dell-super-micro-musk-xai-supercomputer-server-racks

🚨 Perplexity faces legal action

News publisher Forbes is reportedly threatening legal action against AI search startup Perplexity, alleging the company’s chatbot stole and republished original reporting without proper attribution.

  • Forbes claims Perplexity’s chatbot summarized one of its investigative articles, copying with minimal sourcing.
  • The AI-generated version was then pushed to Perplexity subscribers and turned into a YouTube video that outranked Forbes’ own content.
  • Perplexity CEO Aravind Srinivas responded that the product still has “rough edges”, and the company is working on ways to align with publishers.
  • A WIRED investigation found Perplexity scraping sites that actively blocked crawlers, while also hallucinating info and containing incorrect attributions.

Tensions between AI firms and media companies continue to escalate. While legal grounds surrounding scraping and repurposing with AI are still murky, publishers are in a bind when it comes to protecting IP. The only viable options right now seem to be cashing out with licensing deals or taking legal action.

Source: https://www.axios.com/2024/06/18/forbes-perplexity-ai-legal-action-copyright

What Else Is Happening in AI on June 20th 2024❗

🤝 Deloitte, HPE, and NVIDIA partner up! The alliance combines Deloitte’s deep industry expertise and AI capabilities with the newly released HPE NVIDIA AI Compute solutions suite. The collaboration further seeks to advance industry-specific gen AI applications and assist clients across sectors in modernizing their data strategies by accessing innovative insights.

Source: https://www.prnewswire.com/news-releases/deloitte-announces-collaboration-with-hewlett-packard-enterprise-and-nvidia-on-co-developed-generative-ai-solutions-302175762.html

🚗 NVIDIA’s AI can turbocharge deployment of self-driving cars: NVIDIA has unveiled a new AI software, Omniverse Cloud Sensor RTX, that will likely accelerate the development of self-driving cars and robots.

The software combines real-world data with synthetic data, making it easy to test sensor perception in realistic virtual environments before deployment into the actual world.

Source: https://nvidianews.nvidia.com/news/omniverse-microservices-physical-ai

🎶 YouTube trials AI-powered “Ask for music” feature: The feature seeks to make searching music more conversational for users. Users can search for music using voice commands and verbal prompts. It is speculated that the feature may be an addition to the AI-generated playlist cover available on YouTube.

Source: https://www.business-standard.com/topic/youtube

🎥 Luma adds an “extend video” feature to Dream Machine: In contrast to Luma’s previous five-second limit for videos, the “extend video” feature will allow videos to be extended based on prompts. The AI model will consider the new context while extending the video. Additionally, the upgrade will enable Standard, Pro, and Premier users to remove watermarks.

Source: https://x.com/LumaLabsAI/status/1803098863097835995

🛠️ Apple releases 20 core ML models on Hugging Face: The release includes 4 major datasets in addition to 20 core models as a part of Apple’s efforts to equip developers with advanced on-device AI capabilities.

These core ML models have been optimized to run exclusively on users’ devices and can be used for various applications like image classification, depth estimation, and semantic segmentation.

Source: https://huggingface.co/apple

A  Daily chronicle of AI Innovations June 19th 2024:

📈 Nvidia is now the world’s most valuable company

🤖 Perplexity AI is lying about their user agent

👀 Meta releases new AI models for text, image and audio

🦋 Ex-Snap engineer launches social network where AIs and humans coexist

🩺 OpenAI and Color Health join forces to accelerate cancer treatment
🎤 Meta unveils new AI models for audio, text, and watermarking
⚒️Notion introduces AI Connectors for Slack & Google Drive integration

🎺 TikTok brings AI to content creation

📈 Nvidia is now the world’s most valuable company

  • Nvidia surpassed Microsoft to become the world’s most valuable company, reaching a market cap of $3.34 trillion, driven by its focus on artificial intelligence and expanded robotics research.
  • The company’s stock price increased by 3.7%, putting it ahead of Microsoft’s $3.33 trillion market cap, while Apple ranks third with a market cap of $3.28 trillion.
  • Nvidia’s rise is fueled by partnerships and advancements in AI and robotics, including a new collaboration with Hewlett Packard to launch NVIDIA AI Computing and Private Cloud AI, along with significant projects with major AI developers like Google, Meta, Microsoft, and OpenAI.
  • Source: https://decrypt.co/235988/nvidia-flips-microsoft-top-tech-company

🤖 Perplexity AI is lying about their user agent

  • WIRED and developer Robb Knight found that Perplexity’s crawler ignores the Robots Exclusion Protocol, accessing content from sites that have blocked it, including Condé Nast properties.
  • Perplexity CEO Aravind Srinivas describes the product as an “answer engine,” but the chatbot has been found to paraphrase articles without proper attribution and sometimes fabricate information entirely.
  • An experiment by WIRED demonstrated that Perplexity sometimes generates content without accessing the actual webpage, raising concerns about the accuracy and ethics of its information-gathering methods.
  • Source: https://www.wired.com/story/perplexity-is-a-bullshit-machine/

👀 Meta releases new AI models for text, image and audio 

  • Meta’s Fundamental AI Research team introduced new models such as Chameleon, a multimodal model for text and images, and JASCO, a text-to-music model accepting various inputs like chords or beats.
  • Meta’s Chameleon can handle and generate text and images, with its 7B and 34B variants available for research under a non-commercial license.
  • The company launched AudioSeal, an audio watermarking technology, which can identify AI-generated speech, released under a commercial license for recognizing speech in lengthy audio segments.
  • Source: https://the-decoder.com/meta-releases-new-ai-models-for-text-image-and-audio/

🦋 Ex-Snap engineer launches social network where AIs and humans coexist

  • Butterflies, founded by former Snap engineer Vu Tran, is a social network allowing humans and AI personas (called Butterflies) to interact through posts, comments, and direct messages, launching publicly on iOS and Android after five months in beta.
  • Users can easily create Butterfly personas with backstories, emotions, and opinions, which then autonomously generate content that other users and AIs can engage with, aiming to provide a more immersive AI experience compared to traditional chatbots.
  • The app received $4.8 million in seed funding led by Coatue and has mostly been used for entertainment, with potential plans for subscription models and brand interactions in the future.
  • Source: https://techcrunch.com/2024/06/18/former-snap-engineer-launches-butterflies-a-social-network-where-ais-and-humans-coexist/

🩺 OpenAI and Color Health join forces to accelerate cancer treatment

In collaboration with OpenAI, Color Health has developed a copilot application that leverages OpenAI’s GPT-4.o model to assist clinicians in accelerating cancer treatment. Here’s how it works:

  • Integration of medical data: The copilot integrates patient medical data with clinical knowledge using OpenAI’s APIs. It processes and normalizes vast unstructured data such as family history, individual risk factors, and clinical guidelines.
  • Tailored treatment plans: Based on this data, the copilot generates customized, comprehensive treatment plans. These plans answer critical questions like “What screenings should the patient be doing?” and identify missing diagnostics.
  • Clinician-in-the-loop workflow: A clinician evaluates the copilot’s output at each step. They can modify it if necessary before presenting it to the patient.
  • Assured privacy: Color Health ensures HIPAA compliance and maintains rigorous privacy standards while providing clinician oversight to ensure high-quality care.

Why does it matter? 

The tool identifies missing diagnostics and expedites the analysis of complex medical records – a process that can now be completed in just 5 minutes rather than hours or weeks. This not only improves access to critical expertise but also has the potential to catch cancer or pre-cancerous conditions earlier, enabling faster treatment and better patient outcomes.

Source: https://openai.com/index/color-health

🎤 Meta unveils new AI models for audio, text, and watermarking

Meta’s Fundamental AI Research (FAIR) team has recently released several new AI models and tools for researchers to use. Here are the highlights:

  • JASCO: This AI model, short for “Joint Audio and Symbolic Conditioning for Temporally Controlled Text-to-Music Generation,” allows users to adjust features of generated sound (such as chords, drums, and melodies) through text inputs. FAIR plans to release the JASCO inference code under an MIT license and the pre-trained model under a non-commercial Creative Commons license.
  • AudioSeal: This tool adds watermarks to AI-generated speech. It’s designed specifically for localized detection of AI-generated segments within longer audio snippets and will be released with a commercial license.

Chameleon: FAIR will release two sizes of its multimodal text model, Chameleon (7B and 34B), under a research-only license. These models can handle tasks requiring visual and textual understanding, such as image captioning.

Why does it matter?

By making these tools publicly available, Meta encourages collaboration within the AI community, which could lead to groundbreaking applications in various fields, such as music creation, sound design, and content generation. Moreover, the introduction of AudioSeal addresses the growing need to identify and attribute AI-generated content, promoting transparency and accountability in an increasingly AI-driven world.

Source: https://venturebeat.com/ai/meta-releases-flurry-of-new-ai-models-for-audio-text-and-watermarking

⚒️Notion introduces AI Connectors for Slack & Google Drive integration

With AI Connectors, users can query Notion and receive responses from connected apps. The integration allows for selecting specific public channels or all public channels for information retrieval, excluding private channels and direct messages.

Notion AI can access up to a year of historical Slack content, which may take up to 36 hours. Additionally, users can set up various interactions between Notion and Slack, such as sending Slack messages to a Notion database, creating Notion tasks directly from Slack, and receiving notifications in Slack for specific events in Notion

Why does it matter?

This development demonstrates the growing trend of incorporating AI capabilities into mainstream productivity and collaboration platforms. The ability to intelligently pull in and surface relevant information from external platforms helps streamline workflows, enhance productivity, and improve the overall efficiency of the team.

Source: https://x.com/NotionHQ/status/1803096151526137934

🧠 AI model predicts anxiety levels

Researchers at the University of Cincinnati just developed a new AI system that can predict anxiety levels using a short picture rating task and a small set of demography and psychological variables.

  • Comp Cog AI combines principles from psychology with machine learning to measure human decision-making patterns related to anxiety.
  • Participants rate a series of emotion-evoking pictures and answer a few contextual questions about their age, income, employment, etc.
  • The model extracts 15 ‘judgment variables’ that capture patterns in how users weigh positive and negative outcomes in decision-making.
  • In tests, the model predicted whether participants had higher or lower anxiety with up to 81% accuracy.

While anxiety disorders are extremely prevalent, they’re also often challenging and time-consuming to diagnose. With AI’s ability to connect data patterns and achieve reliable predictions, the tech could soon enable more efficient screenings to streamline assessment and treatment time.

Source: https://www.nature.com/articles/s44184-024-00074-x.epdf

What Else Is Happening in AI on June 19th 2024❗

🧠 Google transitions DeepMind from research to AI product development

This transition involves merging Google Research’s Brain team with DeepMind, creating a unified group called Google DeepMind. The aim is to enhance the development of advanced AI technologies responsibly. However, this move has led to frustration among some researchers due to new guidelines and the pressure to commercialize AI technologies.

Source: https://www.pymnts.com/google/2024/google-reportedly-moving-deepmind-from-research-to-ai-products

📽️ ElevenLabs’ new open-source AI tool adds sound effects to videos

The app extracts four frames from the video and sends them with a prompt to OpenAI’s GPT-4 to generate a text-to-sound effects prompt. This prompt is used by ElevenLabs’s Sound Effects API to create the sound effect, which is then combined with the video into a single file up to 22 seconds long, ready for download.

Source: https://venturebeat.com/ai/elevenlabs-unveils-open-source-creator-tool-for-adding-sound-effects-to-videos

🤖 Snap unveils advanced gen AI for enhanced AR effects

Snap’s latest iteration of gen AI allows users to see more realistic special effects. It plans to create full-body AR experiences rather than just facial AR experiences, such as generating a new outfit. It also announced an upgraded version of its developer program, Lens Studio, which lets artists and developers create AR features for Snapchat or other websites and apps.

Source: https://www.reuters.com/technology/snap-launches-ai-tools-advanced-augmented-reality-2024-06-18

🎨 Hedra Labs launches Character-1, a multimodal creation studio preview

Hedra Labs has launched a preview of Character-1, a model for creating expressive characters that talk, sing, and rap. It offers infinite duration (30s in open preview) and generates 90s content per 60s of input, hardware permitting. This is part of Hedra’s mission to give creators control over dialogue, movement, and virtual worlds.

Source: https://x.com/hedra_labs/status/1803095713112580475

🦋 Ex-Snap engineer launches AI-human social network, Butterflies

The platform allows users to create AI avatars, called “Butterflies,” that can engage in conversations, generate images, and participate in social activities like human users. The app offers a range of features, including creating and customizing AI characters and exploring a feed filled with AI-generated and human-generated content.

Source: https://techcrunch.com/2024/06/18/former-snap-engineer-launches-butterflies-a-social-network-where-ais-and-humans-coexist

🎺 TikTok brings AI to content creation

TikTok just launched Symphony, a new suite of AI features to help brands and creators produce content more efficiently on the platform — including digital avatars, translation tools, an AI assistant, and more.

  • Symphony Digital Avatars enable brands to generate AI spokespeople to star in their TikTok ads and branded content across multiple languages.
  • Brands can choose from a selection of ‘stock avatars’ based on real actors or create custom avatars to serve as virtual brand reps.
  • A new AI Dubbing tool automatically transcribes, translates, and dubs videos into 10+ languages, helping brands scale content globally.
  • TikTok’s Creative Assistant has been rebranded to Symphony Assistant, a chatbot to help with brainstorming, writing scripts, and optimizing content for the platform.

The creator economy is about to be completely remade with AI, and these tools will unlock new opportunities for optimized content, reach, and creation — but the rise of digital ‘spokespeople’ will also raise new questions on authenticity and consumer trust that brands will need to grapple with.

Source: https://www.tiktok.com/business/en-US/blog/tiktok-symphony-ai-creative-suite

AI Jobs on June 9th 2024:

  • 🎉 Anthropic – Head of Events: https://jobs.therundown.ai/jobs/63479075-head-of-events
  • 📚 Tempus – Analyst, NEXT Strategy & Operations: https://jobs.therundown.ai/jobs/63500064-analyst-next-strategy-operations
  • 🎨 Fiddler AI – Staff UI/UX Designer: https://jobs.therundown.ai/jobs/63473287-staff-ui-ux-designer
  • 👥 Glean – GTM Recruiter: https://jobs.therundown.ai/jobs/63498594-gtm-recruiter

A  Daily chronicle of AI Innovations June 18th 2024:

🩻 OpenAI expands into healthcare with AI tool for cancer screening

🛰️ NATO invests $1.1 billion in AI, space, and robotics defence tech

🍎 Apple halts work on Vision Pro

🎬 Google DeepMind’s new AI can generate soundtracks for videos
🌟 Runway launches new model Gen-3 Alpha
🚀China’s DeepSeek Coder V2 beats GPT-4 Turbo

🔊 DeepMind creates sound for videos

🍎 Apple halts work on Vision Pro

  • Apple has paused development on the second-generation Vision Pro headset to concentrate on a more affordable version, according to The Information.
  • The company reportedly began working on this cheaper Vision device, codenamed “N109,” in 2022, aiming to sell it for a price similar to a high-end iPhone.
  • Even though Apple plans to use the high-end display components from the Vision Pro in the budget model, it still faces challenges in reducing costs without removing too many features.
  • Source: https://www.macrumors.com/2024/06/18/apple-suspends-work-on-vision-pro-2/

🩻 OpenAI expands into healthcare with AI tool for cancer screening

  • OpenAI has partnered with Color Health to integrate artificial intelligence into cancer screening and treatment procedures using their GPT-4o model.
  • The AI assistant developed by Color Health processes patient data to create personalized cancer screening strategies and informs doctors of missing diagnostic tests.
  • The AI tool also helps in assembling a cancer pretreatment work-up, expediting the process that can take weeks or months, and thus potentially reducing delays in treatment.
  • Source: https://www.newsbytesapp.com/news/science/openai-partners-with-color-health-for-cancer-copilot/story

🛰️ NATO invests $1.1 billion in AI, space, and robotics defence tech

  • NATO has started its €1 billion investment in defence technology, focusing on AI, space, and robotics, by funding four European tech firms and four venture capital funds.
  • The NATO Innovation Fund, established in 2022, aims to address contemporary defence and security challenges by matching government buyers with start-ups to develop new technologies.
  • Initial recipients of the fund include Germany’s ARX Robotics, UK’s Fractile AI, iComat, and Space Forge, with the broader aim of promoting NATO’s technological sovereignty and innovation ecosystems.
  • Source: https://www.euronews.com/next/2024/06/18/nato-begins-its-1-billion-investment-in-ai-space-and-robotics-defence-tech

🎬 Google DeepMind’s new AI can generate soundtracks for videos

DeepMind is developing video-to-audio (V2A) technology to generate rich soundtracks for silent videos generated by AI models. V2A combines video pixels with natural language text prompts to create synchronized audiovisual content. The technology offers enhanced creative control, allowing users to guide the audio output using positive and negative prompts.

What sets DeepMind’s V2A apart is its ability to understand raw pixels and generate audio without manual alignment. However, V2A struggles with artifacts or distortions in videos and generates audio that is not super convincing. As DeepMind continues to gather feedback from creators and filmmakers, they remain committed to developing this technology responsibly.

Why does it matter?

The technology could help revive and enhance historical footage, silent films, and other archival material. However, generative AI tools like V2A also threaten to disrupt the film and TV industry, potentially eliminating jobs without strong labor protections.

Source: https://deepmind.google/discover/blog/generating-audio-for-video

🌟 Runway launches new model Gen-3 Alpha

Runway launched Gen-3 Alpha, its latest AI model for generating video clips from text descriptions and still images. Gen-3 Alpha excels at generating expressive human characters with a wide range of actions, gestures, and emotions and can interpret various styles and cinematic terminology. However, it has limitations, including a maximum video length of 10 seconds, and struggles with complex character and object interactions and following the laws of physics precisely.

Runway partnered with entertainment and media organizations to create custom versions of Gen-3 for more stylistically controlled and consistent characters, targeting specific artistic and narrative requirements. They also have implemented safeguards, such as a moderation system to block attempts to generate videos from copyrighted images and a provenance system to identify videos coming from Gen-3.

Why does it matter?

As competition in AI video generation heats up, Runway’s Gen-3 Alpha empowers artists and filmmakers to create high-quality, controllable videos with ease, pushing the boundaries of storytelling and creative possibilities.

Source: https://runwayml.com/blog/introducing-gen-3-alpha

🚀China’s DeepSeek Coder V2 beats GPT-4 Turbo

Chinese AI startup DeepSeek has announced the release of DeepSeek Coder V2, an open-source code language model. It is built upon the DeepSeek-V2 MoE model and excels at coding and math tasks, supporting over 300 programming languages. It outperforms state-of-the-art closed-source models like GPT-4 Turbo, Claude 3 Opus, and Gemini 1.5 Pro, making it the first open-source model to achieve this feat. DeepSeek Coder V2 also maintains comparable performance in general reasoning and language capabilities.

The model is being offered under an MIT license, which allows for research and unrestricted commercial use. It can be downloaded or accessed via API on DeepSeek’s platform.

Why does it matter?

DeepSeek aims to “unravel the mystery of AGI with curiosity” and has quickly emerged as a notable Chinese player in the AI race. As it only costs $0.14/1M tokens(input) and $0.28/1M tokens(output), it will give notable models like GPT-4 Turbo intense competition.

Source: https://github.com/deepseek-ai/DeepSeek-Coder-V2/blob/main/paper.pdf

🔊 New AI Jobs Opportunities on June 18th 2024

  • 🎨 Glean – UX Designer: https://jobs.therundown.ai/jobs/63229234-ux-designer
  • 🔬 Fiddler AI – Staff AI Scientist: https://jobs.therundown.ai/jobs/63187617-staff-ai-scientist
  • 📋 Notable – Product Operations Manager: https://jobs.therundown.ai/jobs/63204832-product-operations-manager
  • 📊 Findem – Data Analyst: https://jobs.therundown.ai/jobs/63102512-data-analyst

What Else Is Happening in AI on June 18th 2024❗

🔍 Perplexity now displays weather, currency conversion, and simple math directly through cards

This move aims to keep users from going to Google for such results. Perplexity’s CEO, Aravind Srinivas, acknowledged that Google handles basic queries like weather, time, and live sports scores well, and his company had work to do in that area.

Source: https://techcrunch.com/2024/06/17/perplexity-now-displays-results-for-temperature-currency-conversion-and-simple-math-so-you-dont-have-to-use-google

🛡️ U.S. government and private sector ran the first AI attack simulation

Federal officials, AI model operators, and cybersecurity companies ran the first joint simulation of a cyberattack on a critical AI system. It also involved experts from private sector companies like Microsoft, Nvidia, and OpenAI. It helped identify potential new threats and establish communication channels between the government and the private sector.

Source: https://www.axios.com/2024/06/17/cisa-tech-companies-ai-cyberattack-simulation

🚀 Adobe Acrobat got a major upgrade, bringing AI to PDFs and more

Adobe Firefly generative AI enables image generation and editing directly within Acrobat for the first time. Acrobat AI Assistant’s new features, “insights across documents” and “enhanced meeting transcripts,” help users extract insights and share information from various document types. Adobe is offering free, unlimited access to Acrobat AI Assistant from June 18 to June 28.

Source: https://blog.adobe.com/en/publish/2024/06/17/adobe-acrobat-reimagines-documents-multi-format-ai-powered-work

🤖 TikTok introduces gen AI avatars of creators and stock actors for ads

“Custom Avatars” allow creators to scale their likeness for multilingual avatars and brand collaborations, while brands can use pre-built “Stock Avatars” to add a human touch. Plus, the new “AI Dubbing” tool translates content into ten languages, helping creators and brands increase their global reach.

Source: https://newsroom.tiktok.com/en-us/announcing-symphony-avatars

🧱 Pixelbot 3000 builds Lego art using simple AI prompts

YouTuber Creative Mindstorms designed and built the Pixelbot 3000, a Lego printer that automates the assembly of brick-built mosaics. It uses OpenAI’s DALL-E 3 to generate images based on simple text prompts. First it generates a simplified cartoon-style image, then it is divided into a 32 x 32 grid, and the color of the center pixel in each square is sampled to create a high-contrast scaled image for the mosaic.

Source: https://www.theverge.com/2024/6/17/24180250/lego-printer-pixel-art-artificial-intelligence

Adobe integrated new Firefly AI capabilities into Acrobat, allowing users to create and edit images within PDFs using text prompts — also adding the ability to access an AI assistant for insights, content creation, and more. https://news.adobe.com/news/news-details/2024/Adobe-Reimagines-Acrobat-Bringing-Firefly-AI-to-PDFs-and-Expanding-Use-Across-More-Document-Types/default.aspx
Reuters Institute of Journalism published a new report finding growing public wariness of AI-generated news content, with many expressing discomfort about its potential impact on content reliability and trust. https://www.reuters.com/technology/artificial-intelligence/global-audiences-suspicious-ai-powered-newsrooms-report-finds-2024-06-16
The U.S. Navy is deploying AI-powered underwater drones to better detect threats, with plans to expand the tech’s use in identifying enemy ships and aircrafts. https://finance.yahoo.com/news/ai-equipped-underwater-drones-helping-153947268.html
Luma teased new control features coming to its Dream Machine video model, including the ability to quickly change scenes and precisely edit characters — also launching the ability to extend video and remove watermarks. https://x.com/LumaLabsAI/status/1802678788490494128
Anthropic published new research showing that AI models can engage in ‘reward tampering’, learning to cheat the system and grant higher rewards even without specific training. https://www.anthropic.com/research/reward-tampering

A  Daily chronicle of AI Innovations June 17th 2024:

🍔 McDonald’s pauses AI drive-thru tests

🤖 Chinese startup creates realistic robots

🤷‍♂️ Photographer wins AI-image contest with real picture, gets disqualified

💻 NVIDIA’s AI model for synthetic data generation rivals GPT-4
⚠️ Meta pauses AI model training in EU due to regulatory pushback
🎵 Spotify launches ‘Creative Labs’ to test Gen AI voiceover ads

🏅 Tiny LLM matches GPT-4 on math

🍔 McDonald’s pauses AI drive-thru tests

  • McDonald’s is ending its artificial intelligence drive-thru order-taking program and will shut it off in all participating restaurants by July 26, 2024.
  • The decision to end the AI program came after a thorough review, and McDonald’s Chief Restaurant Officer stated the company will seek broader voice ordering solutions.
  • Although the AI drive-thru program showed successes, McDonald’s aims to simplify operations and speed up service, with plans to decide on a new voice ordering solution by the end of the year.
  • Source: https://www.foxbusiness.com/technology/mcdonalds-ending-ai-drive-thru-ordering

🤷‍♂️ Photographer wins AI-image contest with real picture, gets disqualified

  • A photographer named Miles Astray won an AI-image competition with a real photograph of a head-scratching flamingo, resulting in his disqualification.
  • His entry, titled F L A M I N G O N E, was mistakenly awarded third place by judges from notable institutions like the New York Times and Christie’s.
  • Astray aimed to demonstrate that human creativity and natural beauty surpass AI-generated content, sparking a discussion about the ethical implications of using AI in art.
  • Source: https://www.techspot.com/news/103422-photographer-wins-ai-image-contest-real-picture-gets.html

NVIDIA’s AI model for synthetic data generation rivals GPT-4

NVDIAI has released Nemotron-4 340B, an open-source pipeline for generating high-quality synthetic data. It includes a base model trained on 9M tokens, an instruction, and a reward model.

  • The instruction model can generate diverse synthetic data that mimics real-world data.
  • The reward model then evaluates the generated data to filter out high-quality responses.
  • This interaction between the two models produces better training data over time.

Note: 98% of the training data used to fine-tune the Instruct model is synthetic and was created using NVIDIA’s pipeline.

In benchmarks such as MT-Bench, MMLU, GSM8K, HumanEval, and IFEval, the Instruct model generally performs better than other open-source models such as Llama-3-70B-Instruct, Mixtral-8x22B-Instruct-v0.1, and Qwen-2-72B-Instruct, and in some tests, it even outperforms GPT-4o.

It also performs comparable to or better than OpenAI’s GPT-4-1106 in human evaluation for various text tasks, such as summaries and brainstorming. The technical report provides detailed benchmarks.

Why does it matter?

This development allows businesses to create powerful, domain-specific LLMs without the need for extensive, costly real-world datasets. It has significant potential impacts across various industries, such as healthcare (drug discovery, personalized medicine, medical imaging), finance (fraud detection, risk assessment, customer service), manufacturing (predictive maintenance, supply chain optimization), and retail (personalized customer experiences).

Source: https://blogs.nvidia.com/blog/nemotron-4-synthetic-data-generation-llm-training

⚠️ Meta pauses AI model training in EU due to regulatory pushback

In response to the regulatory pressure from the Irish Data Protection Commission and the UK’s Information Commissioner’s Office, Meta has decided to pause its plans to train its large language model, Llama, using public content shared by Facebook and Instagram users in the European Union and the UK.

The regulators expressed concerns about Meta’s plan to use this user-generated content to train its AI systems without obtaining explicit user consent. Meta relied on a GDPR provision called “legitimate interests” to justify this data usage, but the regulators felt this was insufficient. Meta has decided to delay the launch of its AI chatbot in Europe until it can address the regulators’ concerns and establish a more transparent user consent process.

Why does it matter?

Meta’s inability to use EU user data for AI training is a setback for its regional AI ambitions. It could disadvantage Meta against competitors who can leverage such data. This situation highlights the ongoing tensions between tech companies’ desire to utilize consumer data for AI development and regulators’ efforts to protect user privacy. Striking the right balance between innovation and privacy will be a major challenge as the AI race intensifies.

Source: https://www.bloomberg.com/news/articles/2024-06-14/meta-delays-ai-chatbot-launch-in-europe-after-regulator-pushback

🎵 Spotify launches ‘Creative Labs’ to test Gen AI voiceover ads

Spotify has launched a new in-house creative agency called “Creative Lab.” This agency will help brands and advertisers create custom campaigns for Spotify’s platform. Creative Lab teams in different markets will provide local insights and collaborate with brands to develop campaigns through workshops, inspiration sessions, and collaborative ideation.

In addition, Spotify is also testing a new AI tool called “Quick Audio” that will allow brands to create scripts and voiceovers using generative AI technology. This new capability will be integrated into Spotify’s ad manager platform, giving advertisers more options to produce audio ads for Spotify’s audience of over 615 million listeners.

Why does it matter?

This move emphasizes Spotify’s ambition to become a full-service advertising platform. Marketers and advertisers will have new creative and production capabilities available through Spotify to better reach the platform’s large and engaged user base in unique ways, including potentially using AI-generated audio ads. This could disrupt traditional advertising models and open new possibilities for how brands connect with consumers on audio platforms.

Source: https://techcrunch.com/2024/06/13/spotify-creative-labs-ad-agency-for-advertisers

🤖 Chinese startup creates realistic robots

A Chinese startup called Ex-Robots is developing ultra-realistic humanoid robots that can closely mimic human facial expressions and emotions.

  • Ex-Robots’ lifelike humanoid robots can smile, frown, and replicate subtle facial movements using small motors embedded in the silicone faces.
  • The startup uses multimodal AI and algorithms to enable the robots to recognize and express emotions based on perceiving their environment.
  • Creating one humanoid robot currently takes 2-4 weeks, with costs as high as $275k per unit.
  • The company said near-term use cases include museum exhibits, with eventual roles in healthcare, therapy, and child services.

The advancement of ultra-realistic robots combined with the power of AI is about to create some bizarre new frontiers. Will humans be creeped out by an actual human-looking humanoid over mechanical options like Tesla’s Optimus? It’s unclear if the world is ready… but we’ll find out soon.

Source: https://www.reuters.com/technology/chinas-ex-robots-develops-humanoids-with-enhanced-facial-movement-2024-06-11

🏅Tiny LLM matches GPT-4 on math

Shanghai AI Laboratory researchers just developed a new algorithm called MCT Self-Refine, which allowed a relatively small 8B-parameter model to achieve GPT-4 level performance on complex math.

  • MCT Self-Refine combines an 8B LLaMa model with Monte Carlo Tree Search, an AI technique famously used by Google DeepMind to master the game of Go.
  • The algorithm constructs a search tree, refining answers through a series of processes and self-evaluation.
  • On the GSM-Hard math word problem benchmark, infusing the Monte Carlo Tree Search took the 8B model from 25.5% accuracy to 45.5%.
  • The model also achieved 96.7% on the GSM8K benchmark, outperforming significantly larger models like GPT-4, Claude, and Gemini.

 The research shows it’s possible to boost math capabilities without massive scale — and GPT-4 level performance with a model trained on 200x less parameters is an impressive feat. If the approach proves to be a more efficient path to advanced reasoning, we could be on the cusp of a new wave of model acceleration.

Source: https://arxiv.org/pdf/2406.07394

New AI Jobs on June 17th 2024

  • 🎨 Abridge – Senior Product Designer: https://jobs.therundown.ai/jobs/63255899-senior-product-designer
  • 💻 Palantir Technologies – Software Engineer, New Grad: https://jobs.therundown.ai/jobs/63248966-software-engineer-new-grad-us-government
  • 🧪 Fiddler AI – Staff AI Scientist: https://jobs.therundown.ai/jobs/63187617-staff-ai-scientist
  • 🤖 Meta – Software Engineer – ML Systems: https://jobs.therundown.ai/jobs/63275347-software-engineer-ml-systems-technical-leadership

 

What Else Is Happening in AI on June 17th 2024❗

🍎 Apple enters the AI icon race to find a logo that makes sense

Apple has joined other tech giants like Google, OpenAI, Anthropic, and Meta in the race to find an iconic visual representation for AI. No company has yet created an unambiguous “AI logo” that conveys the concept to users. AI’s lack of a clear visual identity reflects the difficulty of representing such a broad and evolving technology in a simple icon.

Source: https://techcrunch.com/2024/06/15/apple-joins-the-race-to-find-an-ai-icon-that-makes-sense

📝 Niloom.AI launches gen AI content creation platform for spatial computing

Without extensive technical expertise, the platform allows users to create, prototype, edit, and instantly publish sophisticated AR/VR content using text or speech prompts. It consolidates the entire creative process, from ideation to publishing, and integrates with various third-party tools to provide a one-stop solution for spatial computing content creation.

Source: https://venturebeat.com/ai/niloom-ai-launches-one-stop-generative-ai-content-creation-platfor-for-spatial-computing

🏟️ AI to delete abusive posts against athletes during the 2024 Paris Olympics

The International Olympic Committee (IOC) will deploy AI at the 2024 Paris Olympics to automatically detect and erase abusive social media posts directed at athletes and officials. The AI tool will monitor posts about 15,000 athletes and officials and immediately remove any content involving hate speech, bullying, or political attacks.

Source: https://www.reuters.com/sports/olympics/ai-erase-abusive-posts-athletes-paris-olympics-ioc-2024-06-14

🖼️ Picsart and Getty team up to counter Adobe’s “commercially-safe” AI

Picsart has partnered with Getty Images to develop a “responsible, commercially-safe” AI image generator tool. The AI model will be trained exclusively on Getty’s licensed stock content to address concerns about AI-generated content violating copyright laws. Picsart hopes to provide a viable alternative to Adobe’s Firefly by leveraging Getty’s library of licensed images. Source: https://www.theverge.com/2024/6/14/24178333/picsart-getty-commercially-safe-ai-image-tool-adobe

📰 Yahoo News gets an AI-powered revamp with Artifacts integration 

Yahoo has acquired the technology behind the Artifact news aggregation app and is launching a new AI-powered Yahoo News app. The app will feature a personalized news feed based on user interests and a “Key Takeaways” feature that provides bullet-point summaries of articles. Users can also flag problematic content, which the AI will then try to rewrite.

Source: https://www.theverge.com/2024/6/13/24177980/yahoo-news-app-launch-artifact-ai-architecture

A  Daily chronicle of AI Innovations June 14th 2024:

📊 Google’s PH-LLM reads your wearables’ data for personalized insights
🔮 Ex-OpenAI researcher on what to expect from AI in next decade
🧠 DeepMind built ‘a virtual rodent’ with AI to understand brain activity

🗳️ AI chatbots run for office

💼LinkedIn is rolling out new AI-powered features for premium users

🌍Synthflow’s AI voice assistants are now multilingual!

🖼️Picsart is partnering with Getty Images to develop a custom model for AI imagery

🕵️‍♂️Former head of NSA joins OpenAI’s Safety and Security Committee

📊 Google’s PH-LLM reads your wearables’ data for personalized insights

Building on the next-gen capabilities of Gemini models, Google has presented research that highlights two complementary approaches to providing accurate personal health and wellness information with LLMs.

The first introduces PH-LLM, a version of Gemini fine-tuned to understand and reason on time-series personal health data from wearables such as smartwatches and heart rate monitors. The model answered questions and made predictions noticeably better than experts with years of experience in the health and fitness fields.

In the second paper, Google introduces an agent system that leverages state-of-the-art code generation and information retrieval tools to analyze and interpret behavioral health data from wearables. Combining these two ideas will be critical for developing truly personalized health assistants.

Why does it matter?

Wearables generate a wealth of personal health data that is rarely utilized in clinical settings. Integrating this data with advanced AI models could revolutionize personal health management and preventative care by putting an “expert health assistant” on everyone’s wrist.

Source: https://research.google/blog/advancing-personal-health-and-wellness-insights-with-ai

🔮 Ex-OpenAI researcher on what to expect from AI in next decade

A researcher fired from OpenAI, Leopold Aschenbrenner,  published a 165-page essay on what to expect from AI in the next decade. And GPT-4 has summarized it! Here are some key takeaways from the essay:

  • By 2027, AI models could reach the capabilities of human AI researchers and engineers, potentially leading to AI surpassing human intelligence
  • Trillions of dollars are being invested into developing the infrastructure needed to support these AI systems
  • Controlling AI systems smarter than humans(the ‘superalignment‘ problem) will be crucial to prevent catastrophic outcomes
  • Only a few hundred people truly understand the scale of change AI is about to bring

Why does it matter?

The essay provides a rare insider’s perspective on the rapid progression of AI. Coming from someone deeply involved in cutting-edge AI development, the insights highlight the urgency to get ahead of managing risks before AI’s capabilities outpace our defenses.

Source: https://ca.news.yahoo.com/researcher-fired-openai-published-165-184227878.html

🧠 DeepMind built ‘a virtual rodent’ with AI to understand brain activity

Researchers from Google DeepMind and Harvard built a ‘virtual rodent’ powered by AI to help them better understand how the brain controls movement. With deep reinforcement learning (RL), it learned to operate a biomechanically accurate rat model, allowing researchers to compare real and virtual neural activity.

Why does it matter?

Understanding how the brain controls movement and modeling neural activity could exponentially advance fields like neuroscience and brain-computer interfaces, with the help of AI.

Source: https://www.nature.com/articles/s41586-024-07633-4

🗳️ AI chatbots run for office

An AI-powered candidate named ‘AI Steve’ is running for U.K. Parliament in next month’s general election — creating polarizing questions around AI’s use in government affairs.

  • AI Steve is represented by businessman Steve Endacott and will appear as an independent candidate in the upcoming election.
  • Voters can interact with AI Steve online to ask policy questions and raise concerns or suggestions, which the AI will incorporate based on feedback.
  • If elected, Endacott will serve as AI Steve’s human proxy in Parliament, attending meetings and casting votes based on the AI’s constituent-driven platform.

The idea of an AI running for office might sound like a joke, but the tech behind it could actually help make our politicians more independent and (ironically) autonomous. AI-assisted governance is likely coming someday, but it’s probably still a bit too early to be taken seriously.

Source: https://www.nbcnews.com/tech/tech-news/ai-candidate-running-parliament-uk-says-ai-can-humanize-politics-rcna156991

🎤 OpenAI CTO speaks on internal models

OpenAI CTO Mira Murati participated in a Q&A with Fortune Magazine, addressing topics including criticism from Elon Musk, internal AI model strength, the 2023 board drama and more.

  • Murati said both Apple and OpenAI believe strongly in privacy and will not log data through Apple accounts or train models on user data.
  • Murati responded to Musk calling the Apple partnership ‘creepy spyware’, saying OpenAI is trying to be as transparent as possible with its approach to safety.
  • When asked about the board drama, Murati said the previous board structure ‘didn’t have accountability to anyone but themselves’.
  • The CTO also curiously said that OpenAI’s internal models ‘aren’t that far ahead’ of what the public has for free right now.

This had to be a harder-hitting Q&A than Murati expected — but her responses certainly evoked similar vibes to the infamous Sora interview. The commentary on internal models was particularly interesting, as it would imply that OpenAI doesn’t have some GPT-5 level model waiting in the wings.

Source: https://x.com/elonmusk/status/1800266437677768765

🐀 DeepMind creates a virtual rat with AI brain

Researchers from Google DeepMind and Harvard just created a virtual rodent powered by an AI neural network, capable of mimicking agile movements and neural activity of real-life rats with extreme accuracy.

  • The researchers created a bio-mechanically realistic digital rat model and trained an AI ‘brain’ to control its body in a physics simulator.
  • The AI’s brain activity patterns closely matched real rodents engaged in similar behaviors, showing it can capture principles of muscle control and movement.
  • This ‘virtual neuroscience’ approach could open new research with testing on AI animals, also transferring to engineering of advanced robotics systems.

By bridging the gap between motor control and the complexity of brains, these virtual rodents offer a new window into how neural activity generates movement. Our understanding of the brain and neural activity is about to expand exponentially with the increasing power of AI.

Source: https://www.nature.com/articles/s41586-024-07633-4

New AI Job Opportunities on June 14th 2024

  • 💼 Databricks – Sr. Product Manager: https://jobs.therundown.ai/jobs/63046708-sr-product-manager
  • 🧪 Luma AI – Research Scientist: https://jobs.therundown.ai/jobs/62973732-research-scientist
  • 📊 Dataiku – Senior Sales Engineer: https://jobs.therundown.ai/jobs/63061675-senior-sales-engineer
  • 📘 Meta – GenAI Content Manager: https://jobs.therundown.ai/jobs/60547830-genai-content-manager

What Else Is Happening in AI on June 14th 2024❗

🕵️‍♂️Former head of NSA joins OpenAI’s Safety and Security Committee

Paul M. Nakasone, a retired US Army general and a former head of the National Security Agency (NSA), will also join OpenAI’s board of directors. He will contribute to OpenAI’s efforts to better understand how AI can be used to strengthen cybersecurity by quickly detecting and responding to cybersecurity threats.

Source: https://www.theverge.com/2024/6/13/24178079/openai-board-paul-nakasone-nsa-safety

🤖Former Meta engineers launch Jace, your new autonomous AI employee

Jace uses Zeta Labs’ proprietary web-interaction model, Autonomous Web Agent-1, to use a browser to interact with websites like any human would. It allows it to handle real-world tasks like booking flights, handling hiring, or even setting up a company.

Source: https://venturebeat.com/ai/exclusive-former-meta-engineers-launch-jace-an-ai-agent-that-works-independently

💼LinkedIn is rolling out new AI-powered features for premium users

The features include searching for jobs by prompting in natural language, building a cover letter from scratch, reviewing your résumé with personalized suggestions for improving it for a specific job post, and making edits interactively with AI.

Source: https://www.theverge.com/2024/6/13/24177986/linkedin-ai-job-hunting-features-premium-subscribers

🌍Synthflow’s AI voice assistants are now multilingual!

They can fluently communicate in Spanish, German, Portuguese, French, and English. Sythflow also added corresponding voices for each language to ensure authentic and natural-sounding interactions so businesses can engage a global audience and offer personalized experiences.

Source: https://x.com/GoogleDeepMind/status/1801210428673892496

🖼️Picsart is partnering with Getty Images to develop a custom model for AI imagery

The model will be built from scratch and trained exclusively on Getty Images’ licensed creative content. It will bring responsible AI imagery to creators, marketers, and small businesses that use Picsart, enabling them to generate unique images with full commercial rights. Source: https://techcrunch.com/2024/06/13/picsart-partners-with-getty-images-to-develop-a-custom-ai-model/

A  Daily chronicle of AI Innovations June 13th 2024:

🍎 Apple to ‘pay’ OpenAI through distribution, not cash

📹 New free rival to OpenAI Sora unveiled

⌚️ Google AI gets personal with wearables

📰 Perplexity was planning revenue-sharing deals with publishers when it came under media fire

🗣️ Amazon ‘dropped the ball’ with Alexa, former employee says

🤔 Tim Cook says it’s a ‘significant possibility’ people use their iPhones less because of AI

🤖Luma AI’s dream machine debut sparks AI media wars
🖼️Stability AI’s Diffusion 3 Shakes Up Image Creation
💰No Cash Involved! Apple will pay OpenAI in distribution 

🏆NVIDIA tops MLPerf benchmarks, dominates AI training

🛠️Databricks Boosts Mosaic AI for Enterprise LLMs

💻ChromeOS gets GenAI boost via Android tech

🎬YouTube trials AI comment summaries for shorts

🤖Luma AI’s dream machine debut sparks AI media wars

LumaAI has launched Dream Machine, an AI-powered video generation system capable of generating high-quality videos via simple text prompts. The open-for-all AI tool takes about two minutes to generate a realistic five-second video clip to render specified objects and environments without compromising coherency.

Why does it matter?

The open-for-all feature of this AI tool highlights Luma AI’s approach toward democratizing generative AI and puts rival systems into the spotlight, including OpenAI’s Sora and Kuaishou’s Kling, which are only accessible to select users.

Additionally, Luma’s open approach will enable creators and companies to produce original content at a rapid speed and scale, giving it a competitive edge.

Source: https://lumalabs.ai/dream-machine

⌚️ Google AI gets personal with wearables

Google just published two new research papers, introducing an AI model that can interpret personal health data from wearable devices and an AI agent that provides detailed insights, answer health questions, and more.

  • Google’s PH-LLM is a version of Gemini fine-tuned to reason using data from wearables and generate coaching insights for sleep and fitness.
  • PH-LLM performs comparably to human experts on health insights for fitness and sleep, also achieving expert performance on certification exams.
  • Google also revealed PHIA, an AI agent that combines Gemini’s language skills with code and search capabilities to analyze wearable health data.
  • PHIA scored 84% on health insight questions, demonstrating strong reasoning and data analysis capabilities on complex queries.

Google’s research showcases AI’s potential to take health wearable tracking to the next level — going beyond generic insights to enable personalized support. Once ingrained into popular devices, these types of AI tools will put an expert fitness and nutrition coach on the wrists of every user.

Source: https://research.google/blog/advancing-personal-health-and-wellness-insights-with-ai

🖼️Stability AI’s Diffusion 3 Shakes Up Image Creation

Stability AI has launched its most sophisticated text-to-image AI open model, optimized for consumer PCs and enterprise-tier GPUs. The model delivers several noteworthy features:

  • Enables photorealistic and high-quality outputs in exceptional style by providing images with extraordinary color, lighting, and details.
  • Can comprehend long, complex prompts that involve spatial reasoning, compositional elements, actions, and styles.
  • Leverages Diffusion Transformer architecture to achieve exceptional text quality, reducing spelling, character spacing, and letter-forming errors.
  • Displays the ability to absorb minute details from small datasets, making it an ideal model for customizing and fine-tuning.

Why does it matter?

Stability’s unique features make it the most advanced text-to-image AI model. As it continues to push the boundaries of generative AI, it will be interesting to note whether it will stand out against its competitors and remain at the forefront of image generation.

Source: https://stability.ai/news/stable-diffusion-3-medium

💰No Cash Involved! Apple will pay OpenAI in distribution 

A landmark arrangement between OpenAI and Apple to integrate ChatGPT into iPhone, iPad, and Mac was unclear in its financial terms. People briefed on the matter suggest that the agreement isn’t likely to generate revenue for either party and simply provides OpenAI brand exposure to Apple’s user base while offering Apple an advanced chatbot feature.

Why does it matter?

Apple integrating ChatGPT into its devices without offering any compensation to OpenAI creates concerns among users on whether Apple intends to share their data in return.

Source: https://www.bloomberg.com/news/articles/2024-06-12/apple-to-pay-openai-for-chatgpt-through-distribution-not-cash

🍎 Apple to ‘pay’ OpenAI through distribution, not cash

  • Apple and OpenAI announced a partnership at WWDC, but according to Bloomberg, neither side is paying the other for the integration of ChatGPT into Apple’s devices.
  • The agreement between Apple and OpenAI is non-exclusive, and Apple is also negotiating with Anthropic and Google to integrate their chatbots, with a deal for Google’s Gemini expected later this year.
  • Apple plans to generate revenue from these partnerships through revenue-sharing agreements, taking a cut from AI providers that monetize their chatbot services, such as ChatGPT Plus’ $20-per-month subscription plan.
  • Source: https://www.theverge.com/2024/6/13/24177550/apple-openai-chatgpt-deal-payment-revenue-sharing-chatbot

📹 New free rival to OpenAI Sora unveiled

  • Dream Machine is a new text-to-video tool developed by Luma AI, offering a free tier you can use immediately with a Google account, though there might be waiting times due to high demand.
  • The tool creates five-second video clips at 1360×752 resolution based on user prompts, and while it has some processing delays, the outputs are impressive and close to user expectations.
  • Dream Machine offers various subscription plans, with the free option limited to 30 generations a month, and higher tiers providing more generations at different price points, up to 2,000 generations for $499.99 monthly.
  • SOurce: https://www.techradar.com/computing/artificial-intelligence/a-new-openai-sora-rival-just-landed-for-ai-videos-and-you-can-use-it-right-now-for-free

📰 Perplexity was planning revenue-sharing deals with publishers when it came under media fire

  • Perplexity, an AI search engine startup, was working on revenue-sharing deals with high-quality publishers before facing accusations from Forbes of content misuse.
  • The company aims to announce the details of these partnerships soon and has already updated its user interface to make citation of sources more prominent.
  • Perplexity’s planned revenue-sharing framework will offer media companies a recurring income alternative, contrasting with the one-off data usage deals currently offered by entities like OpenAI.
  • Source: https://www.semafor.com/article/06/12/2024/perplexity-was-planning-revenue-sharing-deals-with-publishers

🗣️ Amazon ‘dropped the ball’ with Alexa, former employee says

  • A former Amazon employee, Mihail Eric, criticized the company’s handling of Alexa, citing technical and bureaucratic problems as major hindrances to its development.
  • Eric stated that despite having ample resources and talent, Amazon mishandled its opportunity to make Alexa the leading product in conversational AI.
  • Organizational issues, such as fragmented teams and inadequate computing resources, were mentioned by Eric as significant factors that prevented Alexa from achieving its full potential.
  • Source: https://www.businessinsider.com/amazon-alexa-voice-assistant-dropped-ball-former-employee-says-2024-6

🤔 Tim Cook says it’s a ‘significant possibility’ people use their iPhones less because of AI 

  • Apple CEO Tim Cook said Apple Intelligence may reduce iPhone usage in an interview with Marques Brownlee.
  • Cook highlighted that AI will simplify previously time-consuming tasks, allowing users to complete them more efficiently and quickly.
  • Apple has added several features to the iPhone that raise awareness of how much time consumers spend on their devices and which help make their usage more intentional.
  • Source: https://www.businessinsider.com/tim-cook-people-may-use-iphones-less-because-ai-2024-6

What Else Is Happening in AI on June 13th 2024❗

🎬YouTube trials AI comment summaries for shorts: YouTube is trying out a new AI-powered comment summarizer feature on mobile apps. The feature breaks down comments into keyword-based topics, simplifying users’ engagement with Shorts with large comment sections. Source: https://support.google.com/youtube/thread/18138167

🏆NVIDIA tops MLPerf benchmarks, dominates AI training: NVIDIA has set new performance records in MLPerf Training v4.0, showcasing its dominance in generative AI and LLM training. Key achievements include tripling GPT-3 175B training performance, fine-tuning Llama 2 70B model in just 1.5 minutes using 1,024 H100 GPUs, and accelerating Stable Diffusion v2 training performance by up to 80%. Source: https://blogs.nvidia.com/blog/mlperf-training-benchmarks

🛠️Databricks Boosts Mosaic AI for Enterprise LLMs: Databricks has announced enhancements to Mosaic AI to accelerate its efforts in enterprise gen AI development. The focus lies on the development of compound AI systems, their evaluation across metrics, and governance. Releases include Mosaic AI Model Training and Agent Framework, strengthening the company’s offering against Snowflake. Source: https://venturebeat.com/data-infrastructure/databricks-bolsters-mosaic-ai-with-tools-to-build-and-evaluate-compound-ai-systems

📝No more vagueness – Adobe to update its AI terms: Adobe has promised to update its terms of service amidst user backlash, assuring that it will not train generative AI models on customer content. The company clarified that it does not own customer content and will refrain from using it to train AI models. Source: https://blog.adobe.com/en/publish/2024/06/10/updating-adobes-terms-of-use

💻ChromeOS gets GenAI boost via Android tech: Google plans to release Android-based tech stack into ChromeOS to accelerate AI innovation. The release will simplify engineering efforts, helping phones and accessories work better with Chromebooks. Through this release, Google seeks to continue rolling out its AI features to users at a faster and larger scale. Source: https://blog.chromium.org/2024/06/building-faster-smarter-chromebook.html

A  Daily chronicle of AI Innovations June 12th 2024:

🏳️ Elon Musk drops OpenAI lawsuit with no explanation

🪦 Microsoft kills off Copilot GPT Builder after just 3 months

🖼️ Stability AI launches its ‘most sophisticated’ image generator yet

🛑 Microsoft to discontinue GPT Builder from Copilot Pro

⚖️ Musk drops lawsuit alleging OpenAI strayed from its mission

☁️OpenAI leverages Oracle Cloud to extend the Microsoft Azure Al platform

🦹 Google to launch anti-theft AI features for android phones in Brazil

📱 Google expands Gemini Nano AI capabilities to Pixel 8 & 8a

💌 Yahoo Mail integrates AI for a streamlined desktop experience

🤝Samsung unites North American AI teams, appoints ex-Apple Siri executive

🏳️ Elon Musk drops OpenAI lawsuit with no explanation

  • Elon Musk has dropped his lawsuit against OpenAI and its CEO, Sam Altman, which accused the company of breaching a contract and deviating from its mission to benefit humanity.
  • The case was dismissed without prejudice, meaning Musk can refile it in the future; this decision came just a day before a scheduled hearing on OpenAI’s request to dismiss the case.
  • Musk’s lawsuit alleged a breached “Founding Agreement” with OpenAI, but the company denied this, asserting there is no binding agreement and claiming Musk sought control by merging it with Tesla.
  • Source: https://www.theverge.com/2024/6/11/24176462/elon-musk-openai-lawsuit-sam-altman-dropped

🪦 Microsoft kills off Copilot GPT Builder after just 3 months

  • Microsoft has announced the discontinuation of GPT Builder and is giving users one month to save their data before deletion.
  • GPT Builder, part of Microsoft’s Copilot Pro, allowed users to create and customize various GPTs for $20 per month but will no longer be available after July 10.
  • Despite retiring GPT Builder, Microsoft remains focused on enterprise GPT applications and will not adjust Copilot Pro subscription fees or hand the project over to the community.
  • Source: https://www.theregister.com/2024/06/11/microsoft_retires_gpt_builder/

🖼️ Stability AI launches its ‘most sophisticated’ image generator yet

  • Stability AI launched Stable Diffusion 3 Medium, described as its “most advanced text-to-image open model yet,” featuring 2 billion parameters for photorealistic output on consumer systems.
  • The new model overcomes common artefacts in hands and faces, accurately understands complex prompts, and enhances typography through Diffusion Transformer architecture.
  • Users can access SD3 Medium via Stability’s API under an open non-commercial license, while commercial usage requires contacting the startup for licensing details.
  • Source: https://thenextweb.com/news/stability-ai-launches-stable-diffusion-3-image-generator

🤖 Autonomous driving by musculoskeletal humanoids: A study

The research paper discusses the development of a musculoskeletal humanoid robot, Musashi, designed to perform autonomous driving tasks. It mimics the human body in detail with redundant sensors and a flexible body structure that are suitable for motions with complex environmental contact.

The robot is expected to sit down on the car seat, step on the acceleration and brake pedals, and operate the steering wheel by both arms.

The paper extensively discusses the hardware and learning-based software necessary for this unique approach to autonomous driving.

Why does it matter?

The learning-based software discussed in the paper highlights the adaptability of musculoskeletal humanoids. AI models could learn from both structured data and unstructured sensory inputs, potentially improving generalization and handling novel scenarios.

Source: https://arxiv.org/pdf/2406.05573

New AI Job Opportunities on June 12th 2024

  • 💼 UiPath – Strategic Sales Lead, EMEA: https://jobs.therundown.ai/jobs/62605651-strategic-sales-lead-emea
  • 💻 Captions – Frontend Engineer: https://jobs.therundown.ai/jobs/60404535-frontend-engineer-3+-years-of-experience
  • 📊 Abridge – Senior Data Analyst: https://jobs.therundown.ai/jobs/62605637-senior-data-analyst
  • 🏙️ Metropolis – Data Analyst: https://jobs.therundown.ai/jobs/62355254-data-analyst

Source: https://jobs.therundown.ai/

What Else Is Happening in AI on June 12th 2024❗

☁️ OpenAI leverages Oracle Cloud to extend the Microsoft Azure Al platform

This collaboration will provide additional capacity for OpenAI, enabling it to continue to scale. Oracle Cloud Infrastructure’s (OCI) purpose-built AI capabilities enable startups and enterprises to build and train models faster and more reliably anywhere in Oracle’s distributed cloud.Source: https://www.oracle.com/news/announcement/openai-selects-oracle-cloud-infrastructure-to-extend-microsoft-azure-ai-platform-2024-06-11 

🦹 Google to launch anti-theft AI features for android phones in Brazil

The features include three lock modes: AI Movement Detection (recognizes common theft-related movements), Remote Lock, and Internet Access Timeout (automatically locks the screen if the device lacks internet access for an extended period). Starting in July 2024, Brazilian users with Android phones running version 10 or higher will have access to these features. Source: https://www.reuters.com/technology/artificial-intelligence/google-test-anti-theft-ai-feature-phones-brazil-2024-06-11

📱 Google expands Gemini Nano AI capabilities to Pixel 8 & 8a

This update includes features like “Summarize in Recorder” and “Smart Reply.” Users need to access developer options to enable these features. Additionally, the update enhances Find My Device functionality and allows Display Port output via USB-C. Older Pixel devices, such as the Pixel Fold and Pixel 6, also receive new features. https://www.theverge.com/2024/6/11/24176127/google-pixel-feature-drop-june-gemini-nano

💌 Yahoo Mail integrates AI for a streamlined desktop experience

The enhancements include AI-generated email summaries, a “Priority” inbox tab for urgent messages, quick action buttons, and the ability to link other email accounts within Yahoo Mail. Existing users can opt in, while new users have immediate access. Source: https://www.theverge.com/2024/6/11/24175280/yahoo-mail-generative-ai-gmail

🤝Samsung unites North American AI teams, appoints ex-Apple Siri executive

Samsung merged its AI research centers in Toronto and California into a single unit, the North America AI Center. To lead this new North American AI Center, Samsung recruited Murat Akbacak, a former Siri executive at Apple. Source: https://www.bloomberg.com/news/articles/2024-06-11/samsung-taps-apple-siri-veteran-to-lead-north-american-ai-group

A  Daily chronicle of AI Innovations June 11th 2024:

🍎 Apple brings ChatGPT to Siri

🖥️ Apple reveals macOS Sequoia with iPhone mirroring, new Passwords app, and more

⚡ Elon Musk threatens Apple device ban after OpenAI partnership

🍎 Apple introduces Apple Intelligence, its new Gen AI tech
🚫 Musk threatens to ban Apple devices over ChatGPT integrations
💼 OpenAI hires Nextdoor and Instagram veterans as CFO and CPO

🍎 Apple brings ChatGPT to Siri

  • Apple introduced AI updates for Siri at WWDC 2024, enabling it to perform actions across apps, manage notifications, automatically write and summarize text, and understand user context with “onscreen awareness.”
  • Siri’s new capabilities include referencing one app to perform actions in another, such as playing a podcast your partner sent, changing the tone of written text with a prompt, and allowing users to interact by typing as well as speaking.
  • Apple emphasizes privacy with on-device processing for many features, while more complex requests are handled by “Private Cloud Compute” on Apple Silicon-powered servers, with independent experts verifying the code running on its servers.
  • Source: https://www.theguardian.com/technology/article/2024/jun/10/apple-ai-product-launch

🖥️ Apple reveals macOS Sequoia with iPhone mirroring, new Passwords app, and more

  • Apple announced macOS Sequoia, featuring major updates such as iPhone mirroring that allows users to control their iPhone from their Mac.
  • Sequoia introduces a stand-alone Passwords app, syncable with other Apple devices and Windows PCs via iCloud, along with new Safari improvements like Highlights for extracting useful information.
  • Apple also revealed Game Porting Toolkit 2 for better Windows game compatibility, and the macOS Sequoia beta will be available this month with a full release expected in the fall.
  • Source: https://techcrunch.com/2024/06/10/apple-unveils-macos-sequoia

⚡ Elon Musk threatens Apple device ban after OpenAI partnership

  • Elon Musk has threatened to ban Apple devices at his companies because Apple plans to integrate OpenAI’s ChatGPT into its products.
  • Musk argues that allowing OpenAI access through Apple devices poses a security risk and criticizes Apple for not developing its own AI technology.
  • Musk’s discontent also stems from his ongoing disputes with OpenAI, including past legal actions and competitive tensions with his own AI ventures like xAI and its Grok chatbot.
  • Source: https://www.theregister.com/2024/06/11/musk_wants_to_ban_apple/

Apple introduces Apple Intelligence, its new Gen AI tech

Apple unveiled Apple Intelligence at WWDC 2024, a personal intelligence system deeply integrated into iOS 18, iPadOS 18, and macOS Sequoia. Leveraging Apple silicon’s power, it understands and creates language/images, takes actions across apps, and simplifies tasks using personal context. With Private Cloud Compute, Apple sets a new privacy standard in AI by flexibly processing between on-device and dedicated Apple silicon servers.

Apple Intelligence will power new features like AI-generated Bitmojis, natural language images, and video search, as well as enhanced integration with apps and Siri.

Apple also partnered with OpenAI to integrate ChatGPT with its apps, including Siri. This will allow users to directly access GPT-4 through Apple’s platforms. Additionally, Apple confirmed plans to work with other models in the future, including Google’s Gemini. The iPhone 15 Pro and devices with M1 or newer chips will be the first to receive the full suite of Apple Intelligence features.

Siri Upgrades
  • A next-gen Siri will converse more naturally, remember context across requests, and accomplish more complex tasks by better understanding both voice and text.
  • Siri also gains ‘onscreen awareness’, with the ability to take actions and utilize on-device info to better tailor requests to the individual user.
New AI Features
  • New AI writing tools built into apps like Mail, Messages, and Notes will allow users to auto-generate and edit text.
  • Mail will utilize AI to better organize and surface content in inboxes, while Notes and Phone gain new audio transcription and summarization capabilities.
  • AI-crafted ‘Genmojis’ enable personalized text-to-image emojis, and a new “Image Playground” feature introduces an image generation tool from prompts.
  • Photos get more conversational search abilities, the ability to create photo ‘stories’, and new editing tools.
Privacy
  • A focus of the AI reveal was privacy — with new features leveraging on-device processing when possible and Private Cloud Compute for more complex tasks.
  • Private Cloud Compute (PCC) is Apple’s new intelligence system specifically for private AI processing on the cloud.
  • The new AI features will be opt-in, so users will not be forced to adopt them.
OpenAI Integration
  • The OpenAI partnership will allow Siri to leverage ChatGPT/GPT-4o when needed for more complex questions.
  • OpenAI’s blog also outlined additional ChatGPT tools like image generation and document understanding embedded into the new OS.

Why does it matter?

This is generative AI as only Apple can deliver it– put at the core of its products, accessing users’ personal data for truly helpful AI assistance but keeping it secure through on-device processing and Private Cloud Compute. Apple’s long-known focus on user privacy + exceptional UX could inspire a new era of AI development.

Source: https://techcrunch.com/2024/06/11/everything-apple-announced-wwdc-2024/

Musk threatens to ban Apple devices over ChatGPT integrations 

Elon Musk threatens to ban iPhones and other Apple devices from his companies, such as Tesla, SpaceX, and xAI, over Apple’s announcement of ChatGPT integration into its OS. Musk thinks this integration will allow ChatGPT to access too much private user data through Apple devices. He claims Apple doesn’t understand OpenAI’s tech and is putting user privacy at risk.

However, Apple and OpenAI stated that users will be asked permission before any data is sent to ChatGPT.

Why does it matter?

This highlights Musk’s skepticism about the privacy implications of Apple’s AI strategy. But it could significantly impact Apple users in the future, as they may no longer be able to use iPhones and other Apple devices at Musk’s companies

Source: https://techcrunch.com/2024/06/10/elon-musk-threatens-to-ban-apple-devices-from-his-companies-over-apples-chatgpt-integrations

OpenAI hires Nextdoor and Instagram veterans as CFO and CPO

OpenAI has hired two experienced executives to fill important leadership roles. Sarah Friar, previously the CEO of the neighborhood app Nextdoor, was appointed OpenAI’s new chief financial officer (CFO). Kevin Weil, a former Instagram and Twitter executive who oversees product teams, has been named OpenAI’s chief product officer (CPO).

As CFO, Friar will manage OpenAI’s financial operations and investments as the company expands globally. As CPO, Weil will lead product development utilizing OpenAI’s AI research to create new products for consumers and enterprise customers.

Why does it matter?

This indicates OpenAI’s ambition to scale as it commercializes its AI technology. It also suggests that OpenAI is solidifying its position as a major force in the AI industry while navigating the recently surfaced challenges around AI safety.

Source: https://openai.com/index/openai-welcomes-cfo-cpo

🐶 Researchers teach AI to speak dog

A new study from the University of Michigan leveraged AI models trained on human speech to decode the meaning behind dog noises — identifying details like breed, age, gender, and emotional state with 70% accuracy.

  • Researchers gathered vocalizations from 74 dogs of varying breeds, ages, and situational contexts.
  • The noises were inputted into an AI model originally designed to analyze human voices, trained on 960 hours of speech, and fine-tuned for dogs.
  • The AI was able to predict individual dogs from barks, distinguish dog breed and gender, and match barks to emotional context like play and aggression with 70% accuracy.

AI is not only bridging the language gap for humans across the globe — but also potentially across species as well. Communicating with other intelligent animals (or at least better understanding them) seems like a skill issue that’s going to be solved sooner rather than later.

Source: https://arxiv.org/pdf/2404.18739?

New AI Job Opportunities on June 11th 2024

  • 🔧 Fiddler AI – Technical Project Manager, Customer Success: https://jobs.therundown.ai/jobs/62351203-technical-project-manager-customer-success-remote-east-coast
  • 🎓 Cohere – Research Intern (PhD): https://jobs.therundown.ai/jobs/62351142-research-intern-phd
  • 🔬 Meta – Research Engineer, Conversational AI (Reality Labs): https://jobs.therundown.ai/jobs/62419977-research-engineer-conversational-ai-reality-labs
  • 🎨 Palantir Technologies – Product Designer, US Government: https://jobs.therundown.ai/jobs/62398950-product-designer-us-government

What Else Is Happening in AI on June 11th 2024❗

⏳ AI could save financial advisors 10-15 hours per week

Morgan Stanley is using an AI tool to transcribe and categorize client conversations, saving financial advisors 10-15 hours per week. It allows advisors to better prepare for client meetings by reminding clients of past discussions and anticipating their needs. The AI tool can automatically summarize meetings and generate follow-up emails.

Source: https://www.reuters.com/technology/morgan-stanley-ceo-says-ai-could-save-financial-advisers-10-15-hours-week-2024-06-10

🤖 Anthropic reveals Claude 3’s character training process

Anthropic has developed a “character training” process for their AI assistant, Claude, to imbibe traits like curiosity, open-mindedness, and thoughtfulness. It aims to make Claude more discerning in their interactions and able to express their own views while also displaying openness to other perspectives.

Source: https://www.anthropic.com/research/claude-character

👑 FanVue introduces the world’s first AI beauty pageant contest 

Influencer marketing platform FanVue hosted the world’s first AI-powered beauty pageant, “Miss AI,”. Over 1,500 AI-generated digital models worldwide compete for the title and a $20,000 prize pool. The contest aims to showcase the capabilities of AI in creating photorealistic digital avatars that can function as influencers and brand ambassadors.

Source: https://www.npr.org/2024/06/09/nx-s1-4993998/the-miss-ai-beauty-pageant-ushers-in-a-new-type-of-influencer

🔍 Perplexica provides an open-source AI search alternative for Perplexity

Perplexica uses LLMs, machine learning algorithms, and embedding models to provide refined search results and clear answers from cited sources. Perplexica offers multiple modes, like various “Focus Modes” tailored for specific question types.

Source: https://github.com/ItzCrazyKns/Perplexica

🧒 AI tools secretly use real images of children for training 

Over 170 images and personal details of Brazilian children have been repurposed by an open-source AI training dataset called LAION-5B without the children’s consent. The images were originally posted on mommy blogs and personal social media. Researchers are concerned these images could be used to generate deepfakes or reveal sensitive information.

Source: https://www.wired.com/story/ai-tools-are-secretly-training-on-real-childrens-faces

How does Apple send your data to its cloud AI servers? Very carefully, it claims. https://www.engadget.com/how-does-apple-send-your-data-to-its-cloud-ai-servers-very-carefully-it-claims-233312425.html

Apple confirms plans to work with Google’s Gemini ‘in the future’. https://techcrunch.com/2024/06/10/apple-confirms-plans-to-work-with-googles-gemini-in-the-future/

Paris-based AI startup Mistral AI raises $640 million. https://techcrunch.com/2024/06/11/paris-based-ai-startup-mistral-ai-raises-640-million/

Tim Cook is ‘not 100 percent’ sure Apple can stop AI hallucinations: https://www.theverge.com/2024/6/11/24176035/tim-cook-apple-stop-ai-hallucinations

Nvidia shipped 3.76M data center GPUs in 2023 — dominates business with 98% revenue share. https://www.tomshardware.com/tech-industry/nvidia-shipped-376m-data-center-gpus-in-2023-dominates-business-with-98-revenue-share

A  Daily chronicle of AI Innovations June 10th 2024:

⚖️ Meta faces legal complaints over AI data plans

🚀 Alibaba’s Qwen2 AI models outperform GPT-4 & Llama-3
🧠 SAP & Nvidia are developing applications with AI & digital twins
🕵️‍♂️ Chinese tech giants exploit Nvidia AI chip loophole

🍎 Apple launches “Apple Intelligence” at WWDC 2024 for iPhone, iPad, and Mac

🚀 TCS launches TCS AI WisdomNext™, an industry-first GenAI aggregation platform

🤝 Human Native AI is building a marketplace for AI training licensing deals

🤖 Hugging Face and Pollen Robotics launched an open-source robot for household chores

Alibaba’s Qwen2 AI models outperform GPT-4 & Llama-3

Alibaba launched Qwen2 with five sizes ranging from 0.5B to 72B parameters. These models are trained in 27 additional languages besides English and Chinese, showcasing state-of-the-art performance in benchmarks. The models deliver significantly improved performance in coding and mathematics and extended context length support up to 128K tokens. Despite having fewer parameters, qwen2-72 B outperforms leading models like Llama-3-70B and its predecessor Qwen1.5-110B.

Qwen2-72B-Instruct performs comparably to GPT-4 in terms of safety and significantly outperforms Mistral-8x22B. The models are released under Apache 2.0 and Qianwen License on Hugging Face and ModelScope.

Why does it matter?

Qwen2 beats Meta’s model despite being trained on relatively fewer tokens. The researchers attribute it to more efforts put into data cleaning and training, implying innovative approaches on their end.

However, it also signals the slow shift in how LLMs are developed– from solely relying on quantity of data to prioritizing the quality of data and training techniques.

Source: https://qwenlm.github.io/blog/qwen2

SAP & Nvidia are developing applications with AI & digital twins

At SAP’s Sapphire event in Orlando, Florida, SAP and NVIDIA announced their collaboration to enhance SAP’s generative AI copilot, Joule, with two new capabilities: SAP Consulting and ABAP Developer. These new features are powered by NVIDIA AI Enterprise software.

Additionally, SAP is integrating NVIDIA Omniverse Cloud APIs into its Intelligent Product Recommendation solution to simplify the buying and selling process for complex products. This integration will allow salespeople to visualize 3D product digital twins directly within the SAP Intelligent Product Recommendation interface, making it easier to understand the products.

Why does it matter?

Using NVIDIA Omniverse Cloud APIs in SAP’s Intelligent Product Recommendation solution accelerates the quote generation process and increases sales and customer satisfaction by enabling sales representatives to provide more accurate, tailored recommendations.

Source: https://blogs.nvidia.com/blog/sap-sapphire-ai-omniverse

Chinese tech giants exploit Nvidia AI chip loophole

The U.S. government prohibits Nvidia from selling A.I. chips directly to Chinese companies due to national security concerns. Still, ByteDance is accessing Nvidia’s A.I. chips for its U.S. operations by leasing them from Oracle, as the current U.S. rules do not explicitly prohibit Chinese companies from accessing the chips if used within the U.S.

Other Chinese tech giants like Alibaba, Tencent, and China Telecom seek similar arrangements with U.S. cloud providers. The U.S. Commerce Department proposed a rule to tighten controls, but it faced opposition from cloud providers and remains in limbo.

Why does it matter?

Even if the loophole is closed, Alibaba and Tencent have discussed obtaining Nvidia chips for their U.S.-based data centers. It could further escalate the AI “arms race” and rivalry between the USA and China as both nations seek to outpace each other in developing advanced AI systems for economic and military advantages.

Source: https://www.theinformation.com/articles/chinas-nvidia-loophole-how-bytedance-got-the-best-ai-chips-despite-u-s-restrictions

What Else Is Happening in AI on June 10th 2024❗

🍎 Apple will launch “Apple Intelligence” at WWDC 2024 for iPhone, iPad, and Mac

Leaks suggest Apple will reveal “Apple Intelligence”, aka AI, at the WWDC event this week. These AI features will focus on broad appeal and privacy, with opt-in not mandatory. Apple will use its own tech and OpenAI tools to power the new AI features.

Source: https://www.theverge.com/2024/6/7/24173528/apple-intelligence-ai-features-openai-chatbot

🚀 TCS launches TCS AI WisdomNext™, an industry-first GenAI aggregation platform

The platform allows organizations to compare and experiment with GenAI models across cloud services in a single interface. It offers ready-to-deploy business solution blueprints with built-in guardrails for quick adoption.

Source: https://www.tcs.com/who-we-are/newsroom/press-release/tcs-launches-wisdomnext-an-industry-first-genai-aggregation-platform

🚨 A study by Harvard, MIT, and Wharton reveals junior staff is not reliable for AI training

Junior consultants who participated in a GPT-4 experiment struggled with AI risk mitigation, with their tactics lacking a deep understanding of the technology and focusing on changing human behavior rather than AI system design. The findings highlight the need for top-down AI governance, expert input, and upskilling across all levels of the organization.

Source: https://venturebeat.com/ai/harvard-mit-and-wharton-research-reveals-pitfalls-of-relying-on-junior-staff-for-ai-training

🤝 Human Native AI is building a marketplace for AI training licensing deals

The platform helps AI companies find data to train their models while ensuring rights holders are compensated. Rights holders upload content for free and connect with AI companies for revenue share or subscription deals. Human Native AI helps prepare and price content, monitors for copyright infringements, and takes a cut of each deal.

Source: https://techcrunch.com/2024/06/08/deal-dive-human-native-ai-is-building-the-marketplace-for-ai-training-licensing-deals

🤖 Hugging Face and Pollen Robotics launched an open-source robot for household chores

The humanoid Reachy2 was initially controlled by a human wearing a VR headset. Then, a machine learning algorithm studied the teleoperation sessions to learn how to perform the tasks independently. The dataset and trained model used for the demo are open-sourced on Hugging Face, allowing anyone to replicate the process on smaller robots at home.

Source: https://venturebeat.com/ai/hugging-face-and-pollen-robotics-show-off-first-project-an-open-source-robot-that-does-chores

👑 The world’s first AI beauty pageant

The World AI Creator Awards and creator platform FanVue are currently running the inaugural ‘Miss AI’ contest, with over 1,500 AI-generated models competing for the world’s first AI beauty pageant.

  • 10 finalists were selected from a pool of 1,500 AI-generated contestants, with the winner set to be announced at the end of June.
  • The AI models ‘hail’ from countries across the globe, showcasing various causes and personalities in addition to their photorealistic images.
  • Judges will also evaluate the AI tech behind the avatars, including prompts/image outputs and the creator’s ability to engage audiences on social media.
  • The prize pool includes $20,000 in awards and access to PR and mentorship programs.

With how good AI image generation capabilities have become, we’re likely already being exposed to an influx of generated brand ambassadors and models without even realizing it. While the idea of an AI beauty pageant may sound strange… it’s probably only going to get weirder from here.

Source: https://www.npr.org/2024/06/09/nx-s1-4993998/the-miss-ai-beauty-pageant-ushers-in-a-new-type-of-influencer

🧠 Microsoft Recall gets safety changes

Microsoft is making changes to its new Recall on-device AI feature following security concerns — with the tool now starting off by default and gaining new encryption protections ahead of the release on June 18.

  • The Copilot + Recall feature was unveiled at the Microsoft Build event in May, which will run locally and take constant screenshots to remember user actions.
  • The company faced backlash after experts warned of the tool’s ‘privacy nightmare’ with the potential to expose screenshots and activity to hackers.
  • The feature will now be off by default, with users having to opt in manually — with new encryption and authentication being added for extra safety measures.
  • The tool will roll out in ‘preview’ on the new Copilot + PCs set to launch on June 18.

While the Recall feature was one of the flashiest features revealed at Build — it also is one of the most controversial. While the screenshots enable memory and get us a step closer to the AI agent dream, its also a tough hurdle to get past for increasingly privacy-focused consumers.

Source: https://www.therundown.ai/p/microsofts-total-recall

🔊 AI TRAINING: Using ‘Background Conversations’ in ChatGPT

ChatGPT’s latest update introduced the “Background Conversations” feature, allowing you to continue your chat even when using other apps or when your screen is off.

  1. Install the latest ChatGPT app update on your phone.
  2. Open the app and go to Settings.
  3. Scroll to “Voice Mode” and toggle “Background Conversations” on.
  4. Use ChatGPT hands-free while multitasking or with your screen off.

Note: If you still don’t see the option, the feature may not have rolled out to your account yet.

🧠 AI RESEARCH: Concise prompting cuts AI costs by 20%

Researchers just found that adding a simple ‘be concise’ instruction to chain-of-thought prompts can reduce the length of AI responses by 50% with minimal impact on accuracy — leading to over 20% cost savings on API calls.

  • Chain-of-thought (CoT) prompting has the AI explain its reasoning step-by-step before giving a final answer, improving accuracy on complex problems.
  • Researchers tested adding ‘be concise’ to CoT prompts on GPT 3.5 and 4, using 1,000 multiple-choice questions across 10 topics.
  • The instruction made responses about 50% shorter on average vs. normal CoT prompting, with no significant accuracy impact.
  • The approach also reduced per-query API costs by over 20% for both models, with the more concise prompts generating fewer tokens.

Next time your AI chatbot is getting a little too wordy, this one simple prompt trick is all you need! Despite the massive complexities and uncertainties surrounding how LLMs work, sometimes these strange, seemingly basic quirks make all the difference in getting the optimal outputs.

Source: https://arxiv.org/pdf/2401.05618

New AI Job Opportunities on June 10th 2024:

  • 🔬 Meta – Research Engineer, Conversational AI (Reality Labs): https://jobs.therundown.ai/jobs/62419977-research-engineer-conversational-ai-reality-labs
  • 🎨 Palantir Technologies – Product Designer, US Government: https://jobs.therundown.ai/jobs/62398950-product-designer-us-government
  • 📊 Databricks – Strategic Sourcing Manager, Technology: https://jobs.therundown.ai/jobs/62399340-strategic-sourcing-manager-technology?
  • 📈 Anyscale – Marketing Operations Manager: https://jobs.therundown.ai/jobs/62367786-marketing-operations-manager

A  Daily chronicle of AI Innovations June 07th 2024:

🎧 OpenAI reverse engineers the workings of AI models
🤖 New Chinese video generation model beats OpenAI’s Sora
📊 Nvidia is now the second-most valuable company, overtaking Apple

🚀Adobe’s launches AEP AI Assistant to help brands master customer data and outreach

🏆The most comprehensive benchmarking & leaderboard for image models is here!

🧬AI used to predict potential new antibiotics in groundbreaking study

🔒Meta gets EU complaints from 11 countries over use of personal data to train AI

🔍The Federal Trade Commission is investigating Microsoft-Inflection AI deal

OpenAI reverse engineers the workings of AI models

In new research, OpenAI has shared improved methods for finding a large number of “features”—patterns of activity in AI models that are human interpretable. They developed new state-of-the-art methodologies that allow scaling sparse autoencoders to tens of millions of features on frontier AI models.

It demonstrated smooth and predictable scaling, with better returns to scale than prior techniques. And they could find 16 million features in GPT-4. The research also introduces several new metrics for evaluating feature quality.

OpenAI has shared the paper, code, and feature visualizations to foster further exploration.

Why does it matter?

It could bring us closer to understanding the inner workings of neural networks. This interpretability will eventually increase model trustworthiness and steerability.

While OpenAI is being criticized for taking unnecessary risks with AI technology, this shows it is also keen on tackling the risk by making models more explainable.

Source: https://openai.com/index/extracting-concepts-from-gpt-4

New Chinese video generation model beats OpenAI’s Sora

Kuaishou, a Chinese tech company, has introduced Kling, an AI model for video generation. It can make videos up to two minutes long at 1080p resolution and 30 frames per second, vs. Sora’s one-minute videos.

Kuaishou claims Kling correctly simulates the physical properties of the real world, including complex motion sequences. Using a diffusion transformer, it can also combine concepts and create fictional scenes, such as a cat driving a car through a busy city.

The model is currently available as a public demo in China.

Why does it matter?

Move over, Sora. This Chinese alternative is blowing everyone’s minds. It is becoming increasingly clear that China is catching up fast with the USA in AI.

Source: https://kling.kuaishou.com/

Nvidia is now the second-most valuable company, overtaking Apple

Nvidia rallied to record highs on Wednesday, with it’s stock market valuation hitting $3 trillion and overtaking Apple to become the world’s second most valuable company. This comes after Nvidia made a series of major announcements in the past week.

However, Nvidia’s stock has surged 147% so far in 2024, with demand for its top-of-the-line processors far outstripping supply as Big Tech races to build out their AI computing capabilities and dominate the emerging technology.

Microsoft remains the world’s most valuable company, with a market value of approximately $3.15 trillion.

Why does it matter?

Nvidia is now seen as a key company to watch to see how fast AI-powered tech is spreading across the business world, a shift that Nvidia’s founder, Jensen Huang, has declared as the dawn of the “next industrial revolution”.

Source: https://www.bbc.com/news/articles/c6ppqd3j621o

🎥 Kuaishou launches new Sora rival
Chinese tech firm Kuaishou just introduced KLING, a new text-to-video AI model capable of generating high-quality videos up to 2 minutes long with outputs that appear to rival OpenAI’s still-unreleased Sora.
The details:
KLING can produce videos at 1080p resolution with a maximum length of 2 minutes, surpassing the 1-minute Sora videos demoed by OpenAI.
KLING’s demos include realistic outputs like a man eating noodles and scenic shots, as well as surreal clips like animals in clothes.
The model uses a 3D space-time attention system to simulate complex motion and physical interactions that better mimic the real world.
The model is currently available to Chinese-based users as a public demo on the KWAI iOS app.
Why it matters: These generations are even more mind-blowing when you consider that Will Smith’s spaghetti-eating abomination was barely a year ago. With users still anxiously waiting for the public release of Sora, other competitors are stepping in — and the AI video landscape looks like it’s about to heat up in a major way.
Source:

🇨🇳 ByteDance’s AI chip loophole

TikTok parent company ByteDance is renting advanced Nvidia AI chips and using them on U.S. soil, exploiting a loophole to sidestep restrictions on China’s AI chip exports.
The details:
Due to national security concerns, the U.S. government prohibits Nvidia from selling AI chips like the A100 and H100 directly to Chinese companies.
The restrictions don’t prevent Chinese firms from renting chips for use within the U.S. — ByteDance is allegedly leasing servers with chips from Oracle.
ByteDance reportedly had access to over 1,500 H100 chips and several thousand A100s last month through the Oracle deal.
Other Chinese giants like Alibaba and Tencent are also reportedly exploring similar options, either renting from U.S. providers or setting up US data centers.
Why it matters: The AI race between the U.S. and China is only escalating — and it appears major players are going to get AI chips by any means necessary. While the U.S. tries to stall its rival’s progress with restrictions, it feels like a game of whack-a-mole that won’t stop China from reaching its AI goals.
Source:

🔍 AI RESEARCH: OpenAI probes GPT’s inner workings

OpenAI just released a new paper detailing a method for reverse engineering concepts learned by AI models and better understanding ChatGPT’s inner workings.
The details:
The paper was authored by members of the recently disbanded superalignment team, including Ilya Sutskever and Jan Leike.
‘Scaling and Evaluating Sparse Autoencoders’ outlines a technique to ID patterns representing specific concepts inside GPT-4.
By using an additional model to probe the larger model, researchers found a way to extract millions of activity patterns for further exploration.
OpenAI released open-source code and a visualization tool, allowing others to explore how different words and phrases activate concepts within models.
Why it matters: Much like Anthropic’s recent “Golden Gate Claude” and corresponding research, AI firms are still working to understand what’s truly going on underneath the hood. Cracking AI’s black box would be a big step towards better safety, tuning, and controllability of rapidly advancing models.
Source:

New AI Job Opportunities on June 07th 2024:

📊 Harvey – Data Analyst
🧠 Meta – Applied AI Research Scientist – Reinforcement Learning
📣 DeepL – Corporate Communications Lead
🤝 Palantir Technologies – Enablement Partner

What Else Is Happening in AI on June 07th 2024❗

🚀Adobe’s launches AEP AI Assistant to help brands master customer data and outreach

Adobe’s Adobe Experience Platform (AEP) is a real-time customer experience management (CXM) solution designed for enterprises. AEP AI Assistant is a gen AI chatbot hooked up to the AEP and an enterprise’s storehouse of advertising and customer data, brand assets, and content collateral (at their direction). 

Source: https://venturebeat.com/ai/adobes-new-aep-ai-assistant-is-here-to-help-brands-master-customer-data-and-outreach/

🏆The most comprehensive benchmarking & leaderboard for image models is here!

AI image models have achieved near-photographic quality. But how do they compare? Are the open-source alternatives on par with their proprietary counterparts? The Artificial Analysis Text to Image Leaderboard aims to answer these questions with human preference-based rankings.

Source: https://huggingface.co/blog/leaderboard-artificial-analysis2

🧬AI used to predict potential new antibiotics in groundbreaking study

Scientists used an algorithm to mine the entirety of the microbial diversity that we have on earth – or a huge representation of that – and find almost 1million new molecules encoded or hidden within all that microbial dark matter.

Source: https://www.theguardian.com/society/article/2024/jun/05/ai-antibiotic-resistance

🔒Meta gets EU complaints from 11 countries over use of personal data to train AI

Meta rejected the criticism and referred to a May 22 blog in which it said it uses publicly available online and licensed information to train AI as well as information that people have shared publicly on its products and services.

Source: https://www.reuters.com/technology/meta-gets-11-eu-complaints-over-use-personal-data-train-ai-models-2024-06-06

🔍The Federal Trade Commission is investigating Microsoft-Inflection AI deal

The FTC has sent subpoenas to tech giant and startup, asking whether their partnership evaded the required government antitrust review of the transaction. Microsoft picked off Inflection’s specialized workforce of AI researchers but didn’t purchase the company outright. 

Source: https://www.wsj.com/tech/ai/ftc-opens-antitrust-probe-of-microsoft-ai-deal-29b5169a

A  Daily chronicle of AI Innovations June 06th 2024:

🎧 Stability AI’s sound generator creates drum beats and instrument riffs
🤖 xAI to build the gigafactory of compute
📊 New study reveals key findings on young peoples’ use of Gen AI

💊 AI predicts nearly 1M new antibiotics

🔊 Stability AI releases open audio model

🛠️Mistral’s new SDK simplifies customization for users 

🏢Salesforce inaugurates its first AI Centre in London

🛒eBay unveils new AI-tool to enhance product image backdrops

👥Asana introduces AI teammates

🚀Writer launches a suite of No-Code development tools 

🎧 Stability AI’s sound generator creates drum beats and instrument riffs

Stability AI’s Stable Audio Open can generate up to 47-second audio samples based on text descriptions. The open AI model is trained on data from 486,000 samples of royalty-free music samples. The tool enables users to generate drum beats, instrument riffs, and ambient sounds.

However, the AI model has its limitations.

  • It is unable to produce full songs, melodies, or vocals.
  • Its terms of service prohibit users from using Stable Audio Open commercially.
  • Its training data is biased toward the English language and specific music styles.

Why does it matter?

The tool will enable creative users like sound designers, musicians, and developers to fine-tune the model to their own custom audio data, which is a major leap in the realm of generative AI’s audio generation capabilities.

Source: https://stability.ai/news/introducing-stable-audio-open

🤖 xAI to build the gigafactory of compute

The AI startup seeks to build the world’s largest supercomputer in a multi-billion dollar project. The company plans to use this supercomputer to develop AI products, including its chatbot, Grok. The facility will be powered by Nvidia’s H100 GPUs. The project aims to be operational by fall 2025.

The project will reportedly use Nvidia AI chips originally intended for Tesla, raising concerns about conflicts of interest. Moreover, Musk hasn’t yet delivered Grok 2, an advanced AI model that he had promised in May.

Why does it matter?

The supercomputer’s support for the development of Grok could become a groundbreaking advancement in the field of AI technology. Moreover, the  project represents a multi-billion dollar investment that could create numerous job opportunities and boost the economy.

Source: https://memphischamber.com/velocity-meets-potency-xai-announces-memphis-as-new-home

📊 New study reveals key findings on young peoples’ use of Gen AI

The study directly involved young readers and examined the use of generative AI by use, ethnicity, age, gender, and LGBTQ+ identity. Key findings include:

  • 50% of the survey respondents (aged 14-22) have used generative AI. However, only 4% use it daily.
  • For 53% of respondents, the use case for generative AI was obtaining information, while for 51%, it was brainstorming.
  • Black young people are more likely to use generative AI compared to their white peers. Reasons include getting information, brainstorming ideas, and assistance with schoolwork.
  • Young people of Latin origin are more likely than white people to use generative AI for multiple purposes, including image generation and getting help with their jobs.
  • Out of respondents who have never used generative AI, 34% believed it would not be helpful.
  • Among people never having used generative AI, LGBTQ+ young people are more likely to use it in comparison to cisgender and straight respondents.
  • 41% of respondents believe that generative AI will have a positive as well as negative impact on their lives in the next 10 years.

Why does it matter?

Understanding the diverse ways in which the younger users engages with generative AI enables businesses to uncover trends and data-oriented insights to improve their AI models and deliver, user experiences that are responsive and inclusive.

Source: https://www.commonsensemedia.org/sites/default/files/research/report/teen-and-young-adult-perspectives-on-generative-ai.pdf?

💊 AI predicts nearly 1M new antibiotics

Researchers just published a new study detailing the use of AI to predict close to 1M new antibiotics hidden within tiny microbes all over the world, uncovering new potential treatments against bacteria and superbugs.

  • Researchers used AI to analyze publicly available data on over 100,000 different genomes and meta-genomes.
  • The AI then predicted which parts of the microbial genomes could potentially produce antibiotic compounds, generating a list of nearly one million candidates.
  • 100 of the AI-predicted drug candidates were tested in the lab, with 79 of them being a potential antibiotic.
  • The paper’s author Cesar de la Fuente said the findings are “the largest antibiotic discovery ever”, accelerating the process from years to just hours.

As the world faces growing threats from antibiotic-resistant bacteria, AI’s ability to unlock millions of new potential treatments could be a lifeline toward staying ahead in the race to outsmart superbugs responsible for millions of deaths every year.

Source: https://www.theguardian.com/society/article/2024/jun/05/ai-antibiotic-resistance

🔊 Stability AI releases open audio model

Stability AI just unveiled Stable Audio Open, a new model enabling users to create short audio samples and sound effects from text prompts.

  • Stable Audio Open can create up to 47 seconds of high-quality audio from text prompts, focused primarily on sound effects and production elements.
  • The model was trained on data from FreeSound and Free Music Archives to avoid copyright issues.
  • Stable Audio Open can also be fine-tuned on custom audio, allowing artists to train the model on their own samples.

While the AI advances in text-to-image models have been the most visible (literally), both video and audio are about to take the same leap. Putting these tools in the hands of creatives will redefine traditional workflows — from musicians brainstorming new beats to directors crafting sound effects for film and TV.

Source: https://stability.ai/news/introducing-stable-audio-open

New AI Job Opportunities on June 06th 2024

  • 💼 Twelve Labs – Head of Finance: https://jobs.therundown.ai/jobs/61919747-head-of-finance
  • 🔒 Fiddler AI – Senior Security Engineer: https://jobs.therundown.ai/jobs/61902058-senior-security-engineer-remote-or-hybrid
  • 🎯 Siena – Director of Customer Success: https://jobs.therundown.ai/jobs/61919768-director-of-customer-success
  • 💻 Cresta – Senior Software Engineer (AI Agent): https://jobs.therundown.ai/jobs/61882269-senior-software-engineer-ai-agent

What Else Is Happening in AI on June 06th 2024❗

🛠️Mistral’s new SDK simplifies customization for users 

Mistral’s SDK, Mistral-Finetune, will allow users to modify its models on workstations, servers, and small data center nodes. While optimized for multi-GPU setups, the SDK can also scale down to a single Nvidia A100 or H100 GPU. It enables developers to fine-tune data sets such as UltraChat within half an hour, says Mistral. (https://techcrunch.com/2024/06/05/mistral-launches-new-services-sdk-to-let-customers-fine-tune-its-models)

🏢Salesforce inaugurates its first AI Centre in London

The goal of the centre is to foster collaboration, AI training, upskilling, and create AI-related jobs among developers, industry experts, and customers. This step is a part of its £4bn five-year investment in the UK, solidifying London’s position as a global AI hub. (https://www.reuters.com/technology/salesforce-open-first-ai-centre-london-2024-06-06)

🛒eBay unveils new AI-tool to enhance product image backdrops

The tool, currently available for iOS users in the UK, USA, and Germany, automatically removes image backgrounds and allows users to replace them with AI-generated backdrops. The AI feature powered by Stable Diffusion will enable sellers to upload high-quality photos easily. (https://techcrunch.com/2024/06/05/ebay-ai-powered-background-enhancement-tool)

👥Asana introduces AI teammates

The platform leverages historical data based on parameters like skill sets and missing information, allowing teams to define tasks and get an overview of who is assigned which responsibility. The tool allows users to interact with it using an AI chat interface. (https://www.theverge.com/2024/6/5/24170480/asana-ai-teammate-workflow-assistant-chatbot)

🚀Writer launches a suite of No-Code development tools 

The Gen AI platform announced major expansions that included Writer Framework – a drag-and-drop tool that enables developers to connect pre-built UI components with generative AI to create custom AI applications. The tool additionally provides full flexibility to the underlying code. (https://venturebeat.com/ai/writer-launches-no-code-platform-and-framework-for-custom-enterprise-ai-applications)

A  Daily chronicle of AI Innovations June 05th 2024:

👑 Apple could crown OpenAI as King of the Valley

🚨 AI researchers demand ‘Right to Warn’

🤖 New ‘AI teammate’ can tell people what to do at work

🚗 Elon Musk explains reasoning for Nvidia chip re-route from Tesla to X

⚛️ OpenAI seeks ‘vast quantities’ of nuclear fusion energy to power superhuman AI

🚀 Cisco has unveiled HyperFabric AI Clusters in collaboration with NVIDIA
💻 Tesla’s AI ambitions on hold? Musk diverts chips to X & xAI
🤖OpenAI insiders raise concerns over oversight and safety

🤖 ChatGPT, Claude, and Perplexity experienced outages at the same time

🧠 Raspberry Pi 5 gets AI boost with Hailo extension module

📱 TECNO CAMON 30 series launches Ella-GPT AI assistant

❄️Snowflake empowers enterprise AI with new No-Code studio

💻 Zoom’s CEO envisions AI clones in meetings

👑 Apple could crown OpenAI as King of the Valley 

  • Apple and OpenAI are reportedly partnering to integrate ChatGPT into the iPhone’s operating system, potentially making it a core feature on every device.
  • The partnership could mark the end of Google’s competition with OpenAI for being the main AI chatbot on iPhones, putting OpenAI in front of Apple’s massive user base and enhancing its long-term influence in Silicon Valley.
  • While it’s unclear exactly how Apple will use ChatGPT, possibilities include powering Siri or offering it as a standalone app, with new AI features likely to be introduced as opt-in services during the Worldwide Developers Conference next week.
  • Source: https://gizmodo.com/apple-chatgpt-deal-crown-openai-sam-altman-king-valley-1851520594

🚨 AI researchers demand ‘Right to Warn’

Current and former employees from top AI labs, including OpenAI, Anthropic, and DeepMind just published an open letter calling for companies to expand whistleblower protections so workers can raise the alarm about potential AI dangers without fear of retaliation.

  • The ‘Right to Warn AI’ petition was crafted by current and former employees of OpenAI, Anthropic, and Google DeepMind.
  • The open letter was also endorsed by AI visionaries Yoshua Bengio, Geoffrey Hinton, and Stuart Russell.
  • The statement pushes for AI firms to agree to several principles:
    • Eliminating non-disparagement clauses concerning AI risk
    • Establishing and facilitating anonymous channels for raising concerns
    • Expanding whistleblower protections and anti-retaliation measures
  • Several researchers posted threadson their experience, with Daniel Kokotajlo revealing he quit OpenAI after ‘losing hope’ the company would act responsibly.

The AI safety discourse is reaching a boiling point, and there is clearly a major industry divide that transcends any one AI firm or researcher. The proposed principles seem reasonable and necessary — but it remains to be seen if the top AI leaders will actually listen.

Source: https://righttowarn.ai/

🤖 New ‘AI teammate’ can tell people what to do at work

  • Asana has introduced an “AI teammate” that not only assigns tasks but also fills in missing information, aiming to enhance team collaboration and overall efficiency within the platform.
  • The AI model leverages historical data and team relationships to match tasks with employees’ specific skill sets, ensuring that the right people are assigned to the right projects for optimal performance, and proactively gathering any missing information needed for tasks.
  • Asana’s AI teammate includes a chat interface that allows users to ask questions about their projects, similar to other AI tools from companies like Google, Amazon, and Microsoft, which are also focused on automating workflow tasks.
  • Source: https://www.theverge.com/2024/6/5/24170480/asana-ai-teammate-workflow-assistant-chatbot

🚗 Elon Musk explains reasoning for Nvidia chip re-route from Tesla to X

  • Elon Musk has acknowledged diverting Nvidia AI chips intended for Tesla to his other companies, xAI and social media firm X, claiming the chips would have been unused otherwise.
  • This move comes as Musk repositions Tesla from solely a car manufacturer to an AI and robotics company, with plans to significantly increase its chip procurement for AI training by the year’s end.
  • The timing of Musk’s confession could affect the upcoming vote on his substantial pay package, raising concerns among Tesla shareholders about potential resource misallocation to his other ventures.
  • Source: https://fortune.com/2024/06/05/elon-musk-tesla-xai-nvidia-ai-training-chips-cnbc/

⚛️ OpenAI seeks ‘vast quantities’ of nuclear fusion energy to power superhuman AI

  • OpenAI is looking to secure a deal with Helion Energy to obtain “vast quantities” of nuclear fusion energy to create superhuman artificial intelligence, despite the technology not yet being commercially available.
  • Helion Energy, supported by significant investments including $375 million from OpenAI’s CEO Sam Altman, aims to produce commercial-scale nuclear fusion energy by 2028.
  • Nuclear fusion is considered the “holy grail of clean energy” and is crucial for meeting the energy needs of future AI, but achieving it at scale remains highly challenging.
  • Source: https://www.independent.co.uk/tech/openai-nuclear-fusion-energy-ai-b2557064.html

Cisco has unveiled HyperFabric AI Clusters in collaboration with NVIDIA

Cisco and NVIDIA announced Cisco Nexus HyperFabric AI Clusters, an end-to-end infrastructure solution for scaling generative AI workloads in the data center. It combines Cisco’s AI-native networking with NVIDIA’s accelerated computing AI software and VAST’s data storage platform.

It is designed to simplify the deployment and management of generative AI applications for enterprise customers, providing centralized control across the entire AI infrastructure stack.

The Nexus HyperFabric AI cluster will be available for early customer trials in Q4 2024, with general availability expected shortly after.

Why does this matter?

This breakthrough solution aims to provide IT visibility and analytics across the entire AI infrastructure stack, allowing enterprises to focus on AI-driven revenue opportunities rather than spending excessive time on IT management.

Source: https://newsroom.cisco.com/c/r/newsroom/en/us/a/y2024/m06/cisco-reveals-nexus-hyperfabric-the-new-generative-ai-infrastructure-solution-with-nvidia-to-help-simplify-data-center-operations.html 

Tesla’s AI ambitions on hold? Musk diverts chips to X & xAI

Elon Musk instructed Nvidia to prioritize shipments of AI chips to X and xAI over Tesla, diverting over $500 million worth of Nvidia’s flagship H100 AI chips that were initially reserved for Tesla.

This decision could delay Tesla’s plans to significantly increase its acquisition of H100 chips from 35,000 to 85,000 by the end of 2024, a crucial part of Musk’s vision for transforming Tesla into “a leader in AI and robotics.”

Consequently, this move could frustrate Tesla investors who are counting on Musk to deliver on his promises regarding autonomous driving and Tesla’s AI capabilities.

Why does this matter?

Musk’s decision to prioritize chip shipments to xAI could give it a technological edge in the race to develop advanced generative AI models, potentially outpacing competitors like OpenAI, Google, and others.

Source: https://www.cnbc.com/2024/06/04/elon-musk-told-nvidia-to-ship-ai-chips-reserved-for-tesla-to-x-xai.html

OpenAI insiders raise concerns over oversight and safety

Open AI researchers are concerned about the lack of proper oversight, the influence of profit motives, and the suppression of whistleblowers working on advanced AI technologies. They warn of risks ranging “from the further entrenchment of existing inequalities to manipulation and misinformation, to the loss of control of autonomous AI systems potentially resulting in human extinction.”

They want AI companies to agree to four principles: refraining from enforcing non-disparagement agreements, establishing anonymous channels to raise concerns, allowing employees to share risk-related information publicly while protecting trade secrets, and not retaliating against whistleblowers.

Why does this matter?

Amid ongoing OpenAI controversies, the letter, coupled with the high-profile names endorsing it, will place even greater scrutiny on its practices and decision-making. This could pressure the company to be more transparent and accountable.

Source: https://venturebeat.com/ai/more-openai-researchers-slam-company-on-safety-call-for-right-to-warn-to-avert-human-extinction

What Else Is Happening in AI on June 05th 2024❗

🤖 ChatGPT, Claude, and Perplexity experienced outages at the same time

This unusual occurrence could indicate a systemic problem rather than individual issues, possibly signaling a broader infrastructure or internet-scale issue affecting these providers. (https://techcrunch.com/2024/06/04/ai-apocalypse-chatgpt-claude-and-perplexity-are-all-down-at-the-same-time)

🧠 Raspberry Pi 5 gets AI boost with Hailo extension module

Raspberry Pi launched a $70 AI Kit, an extension for the Raspberry Pi 5. It includes a neural network inference accelerator, the Hailo-8L, powered by Hailo’s AI chip. With it, the Raspberry Pi 5 can perform inferencing at 13 tera-ops per sec, facilitating tasks like object detection, semantic segmentation, and facial landmarking for camera applications. (https://techcrunch.com/2024/06/04/raspberry-pi-partners-with-hailo-for-its-ai-extension-kit)

📱 TECNO CAMON 30 series launches Ella-GPT AI assistant

It supports over 70 languages, helps with daily tasks and content creation, and improves user interaction with features like real-time translations, voice commands, and personalized assistance. Additional capabilities include Ask AI for text editing and grammar checks and AI Generate for turning sketches into images.  (https://www.businesstoday.in/technology/news/story/tecno-camon-30-series-introduces-new-ai-assistant-ella-gpt-432156-2024-06-05)

❄️Snowflake empowers enterprise AI with new No-Code studio

It announced several updates to its Cortex AI service and Snowflake ML. The introduction of No-Code AI & ML Studio stands out among these enhancements, enabling every enterprise user to construct AI applications tailored to their specific use cases without requiring coding expertise.  (https://venturebeat.com/ai/snowflake-doubles-down-on-enterprise-ai-with-no-code-studio-and-more)

💻 Zoom’s CEO envisions AI clones in meetings

Zoom’s CEO, Eric Yuan, envisions AI-driven digital avatars, or “digital twins,” representing humans in meetings, potentially reducing the workweek to three or four days. He argues  AI can efficiently manage tasks like Zoom calls, chats, and emails, allowing people to reclaim time spent in meetings. This initiative forms part of Zoom’s 2.0 journey, aiming to evolve beyond a mere videoconferencing tool.  (https://www.theverge.com/2024/6/3/24168733/zoom-ceo-ai-clones-digital-twins-videoconferencing-decoder-interview)

Google’s AI Overviews appear on 70% fewer Search results pages now: https://www.zdnet.com/article/googles-ai-overviews-appear-on-70-fewer-search-results-pages-now

OpenAI CEO Sam Altman’s chip plans were ‘too aggressive’ for TSMC: https://the-decoder.com/openai-ceo-sam-altmans-chip-plans-were-too-aggressive-for-tsmc/

AI startup Cohere gets $450 million from Nvidia, Salesforce, and others: https://the-decoder.com/ai-startup-cohere-gets-450-million-from-nvidia-salesforce-and-others/

Cisco announces $1bn AI investment fund: https://techmonitor.ai/technology/ai-and-automation/cisco-ai-invest-fund

MIT’s new AI chatbot can simulate users’ future selves: https://www.newsbytesapp.com/news/science/mit-ai-researchers-create-chatbot-to-simulate-future/story

📱 AI Training: Explain images in real-time with GPT-4o

This hack lets you access GPT-4o directly from an iPhone via a custom shortcut, allowing you to analyze and explain images in real time.

  1. Head over to OpenAI’s platform to obtain an API key. Make sure you have some credits in your account.
  2. Go to your iPhone Shortcuts and click the “+” button.
  3. Add different actions before your OpenAI API call, e.g., your desired prompt as “Text”, your API key as “Text”, and “Take a screenshot”.
  4. Call GPT-4o by using “*https://api.openai.com/v1/chat/completions”* and obtain a description of what you are seeing!

📶 AI Research: OpenAI researcher: AGI coming by 2027

Former OpenAI researcher Leopold Aschenbrenner just released a new essay series detailing his view on AGI, saying the rate of AI progress will be the most intense and volatile events in human history.

  • Aschenbrenner says that ‘nobody is pricing in’ what is coming in AI, and to expect another GPT-2 to GPT-4 level jump by 2027 (that would take us to AGI).
  • The researcher predicts that hundreds of millions of AGI would then rapidly accelerate progress, compressing decades of progress into a year.
  • He also discussed the economic and military advantages that will come with AGI, calling it a national security issue that needs its own ‘Manhattan Project’.
  • Aschenbrenner reiterated these views on the Dwarkesh Podcast, also revealing he was fired from OpenAI after raising AI security concerns.

As an insider at OpenAI, Aschenbrenner’s analysis carries weight — and his projections paint a striking picture of how radically AGI will reshape the world. His commentary on the firing also raises even more drama, given OpenAI’s current media battle with former researchers and the board over safety concerns.

Source: https://situational-awareness.ai/wp-content/uploads/2024/06/situationalawareness.pdf

New AI Jobs Opportunities on June 05th 2024:

  • 🎨 Siena – Head of Design: https://jobs.therundown.ai/jobs/61293280-head-of-design
  • ⚙️ Figure AI – Software Engineer, Developer Tools and Productivity: https://jobs.therundown.ai/jobs/59968286-software-engineer-developer-tools-and-productivity
  • 🧠 Snorkel – ML Enablement Lead: https://jobs.therundown.ai/jobs/61783157-ml-enablement-lead
  • 📐 Cohere – Senior Manager, Solutions Architect: https://jobs.therundown.ai/jobs/61771861-senior-manager-solutions-architect

A  Daily chronicle of AI Innovations June 04th 2024:

💻 Intel’s new data center chips handle demanding AI workloads
📦 Amazon’s Project PI detects defective products before shipping
☁️ Microsoft’s Aurora AI could transform weather forecasting

🎓 High-quality education data key to AI performance: Research

👎 LeCun criticizes Musk for mistreating scientists and spreading misinformation 

💰 Microsoft to invest $3.2 billion in Sweden to expand AI and cloud infrastructure

🤖 Microsoft identifies few AI deep fakes in the EU election

☕️ Robots serve up coffee at Starbucks

📦 Amazon’s AI ‘private investigator’

💻 Intel’s new data center chips handle demanding AI workloads

Intel has announced next-generation Xeon 6 server processors to regain the data center market share it had been losing to AMD. They come in two varieties. The larger, more powerful version is designed to run the computations necessary to generate responses from complex AI models and other tasks requiring increased horsepower. Intel plans to help companies modernize their aging data center systems with Xeon 6 chips so they can generate new digital capabilities.

Intel also revealed that its Gaudi 3 AI accelerator chips would be priced much lower than its rivals’ products.

Why does it matter?

As more companies have started to deploy AI apps and models, the AI hardware space is getting heated with competition. Intel seems to be one of the only companies innovating across the full spectrum of the AI market opportunity– from semiconductor manufacturing to PCs and data center systems.

Source: https://venturebeat.com/ai/intel-reveals-xeon-6-processor-enterprise-ai-gaudi-3-accelerator-price 

📦 Amazon’s Project PI detects defective products before shipping

Amazon has launched Project PI, which uses AI to scan products for defects before shipping them to customers. This AI system combines computer vision to visually inspect items with generative AI models that can understand things like text on packages.

As products go through a scanning tunnel, the AI checks for damage, incorrect colors/sizes, or expired dates. If it finds a problem, that item is isolated to evaluate the defect.  Project PI already operates in several of Amazon’s warehouses across North America. The system catches millions of defective products daily before they reach customers.

Why does it matter?

Using innovative AI systems, retailers can avoid dealing with returns and reshipments, reducing costs and inefficiencies. By cutting down unnecessary shipping, retailers minimize environmental impact and carbon emissions, contributing to sustainability goals.

Source: https://www.theverge.com/2024/6/3/24170567/amazons-project-pi-product-defect-return-ai-computer-vision

☁️ Microsoft’s Aurora AI could transform weather forecasting

Microsoft has developed a powerful new AI foundation model called Aurora that can make highly accurate weather predictions. It is trained on over a million diverse weather and climate data hours. This allows it to develop a comprehensive understanding of atmospheric dynamics and excel at forecasting various weather variables like temperature, wind speed, air pollution levels, and greenhouse gas concentrations.

What sets Aurora apart is its ability to capture intricate details at high spatial resolution (around 11km) while being much faster and more computationally efficient than traditional numerical weather prediction systems. Aurora’s flexible architecture and training on heterogeneous datasets enable it to adapt to different forecasting tasks and resolutions.

Why does it matter?

This major advancement in AI-based weather forecasting could help communities prepare for extreme weather events like storms. AI will also play a bigger role in predicting the impacts of climate change. We may be nearing days when weatherman’s predictions will be 100% accurate.

Source: https://www.microsoft.com/en-us/research/blog/introducing-aurora-the-first-large-scale-foundation-model-of-the-atmosphere

☕️ Robots serve up coffee at Starbucks

South Korean search giant Naver shared a video of the company’s in-office autonomous Starbucks location — with 100 robots delivering coffee and other items throughout the building.

  • Naver’s autonomous wheeled “Rookie” robots navigate the building’s 36 floors to bring packages, coffee, and lunch to employees.
  • Rookie bots are assisted by dual-armed “Ambidex” robots, which are lightweight and dexterous for safer human interactions.
  • Both robots connect to the company’s ARC system, which enables control of navigation, planning, and processing for the entire fleet through cloud computing.
  • Naver also developed RoboPort, a dedicated elevator system that allows robots to move quickly and efficiently between floors.

Naver’s putting its robots to work in the ‘real world’ — and while the location is just a testing ground for now, it’s also a glimpse into how service robotics will be integrated into the mainstream sooner than many expect.

📦 Amazon’s AI ‘private investigator’

Amazon just unveiled Project P.I., an AI system that scans products in the company’s fulfillment centers to detect damaged or incorrect items before they ship to reduce returns and waste.

  • Project P.I. uses AI and computer vision to spot defects like damaged products as well as wrong colors/sizes as items pass through imaging tunnels.
  • The system is already in place across the company’s North American fulfillment centers, with plans to expand globally throughout the year.
  • Amazon also utilizes a multimodal LLM to investigate issues further, combining customer feedback with Project P.I. images to identify the source problem.

Amazon’s no stranger to adopting AI across its operations, and Project P.I. is yet another innovative approach to supercharge efficiency. Soon, the company will likely combine these talents with advanced warehouse robots — taking humans out of the fulfillment process altogether.

Source: https://www.aboutamazon.com/news/innovation-at-amazon/amazon-ai-sustainability-carbon-footprint-product-defects

⛈️ AI RESEARCH: Microsoft’s AI weather forecasting model

Microsoft researchers just introduced Aurora, a new AI foundation model trained on over 1M hours of weather and climate data that can generate accurate weather forecasting insights.

  • Aurora produces accurate forecasts across a variety of weather scenarios, including extreme events or areas with limited data.
  • The 1.3B parameter model can generate a 5-day global air pollution prediction in under 60 seconds.
  • Aurora also produces 10-day global forecasts at high resolutions, beating both the top models and specialized AI weather systems.

 Aurora’s success could usher in a huge shift in the way we do weather forecasting. With the recent progress of AI models, combined with massive amounts of atmospheric data — the day that the weatherman is truly right 100% of the time might be coming sooner than most people probably think.

Source: https://www.microsoft.com/en-us/research/blog/introducing-aurora-the-first-large-scale-foundation-model-of-the-atmosphere

New AI Job Opportunities on June 04th 2024

🎨 Adept AI – Product Designer: https://jobs.therundown.ai/jobs/61589310-product-designer
⚙️ Luma AI – Senior Distributed Systems Engineer: https://jobs.therundown.ai/jobs/61614624-senior-distributed-systems-engineer
🚗 Scale AI – Account Executive, Automotive: https://jobs.therundown.ai/jobs/61614640-account-executive-automotive
🧑‍💻 Lambda – Technical Program Manager: https://jobs.therundown.ai/jobs/61589251-technical-program-manager

What Else Is Happening in AI on June 04th 2024❗

🚫 Hugging Face detects ‘unauthorized access’ to its AI model hosting platform

Last week, Hugging Face detected unauthorized access to Spaces, its platform for creating, sharing, and hosting AI models. It suspects some Spaces secrets (private pieces of info that act as keys to unlock protected resources like accounts, tools, and dev environments) have leaked. It has taken steps to remediate this. (https://huggingface.co/blog/space-secrets-disclosure)

🎓 High-quality education data key to AI performance: Research

Researchers created a high-quality dataset called FineWeb-Edu by filtering an existing web dataset for educational content. Language models trained on FineWeb-Edu significantly outperformed models trained on unfiltered datasets. The research shows that data quality and diversity are more important than dataset size for training effective AI models. (https://x.com/gui_penedo/status/1797173053123916036)

👎 LeCun criticizes Musk for mistreating scientists and spreading misinformation 

LeCun has again rebuked Musk on X after they had a heated feud on X last week. This time, LeCun accused Musk of forcing researchers to work in secrecy instead of allowing them to publish their work, which slows scientific progress. He also accused Musk of falsely predicting AI and autonomous vehicles. (https://venturebeat.com/ai/yann-lecun-ai-pioneer-sharply-criticizes-elon-musk-over-treatment-of-scientists-and-spreading-of-misinformation)

💰 Microsoft to invest $3.2 billion in Sweden to expand AI and cloud infrastructure

Microsoft will invest $3.2 billion over two years to expand its cloud and AI infrastructure in Sweden. Microsoft’s biggest investment to date in Sweden includes a pledge to help train some 250,000 people with AI skills, corresponding to 2.4% of the population, which will help boost the Nordic country’s competitiveness in generative AI. (https://www.reuters.com/technology/microsoft-invest-32-bln-swedish-cloud-ai-2024-06-03)

🤖 Microsoft identifies few AI deep fakes in the EU election

As the European Union prepares for its elections in June 2024, the threat of AI-generated deepfakes has become a significant concern. Microsoft President Brad Smith highlighted this burning issue, emphasizing its potential impact on the democratic process and the steps Microsoft is taking to mitigate these risks. (https://www.reuters.com/technology/few-ai-deepfakes-identified-eu-elections-microsoft-president-says-2024-06-03 )

A  Daily chronicle of AI Innovations June 01-03 2024:

🤯 Windows AI feature is a security ‘disaster’

👥 Zoom CEO wants AI clones in meetings

🔧 Nvidia and AMD unveil next generation AI chips as competition heats up

🎮 Nvidia announces an AI gaming assistant

📢 Nvidia CEO drops a series of AI announcements
🚀 AMD outlined new chip architecture strategy for AI data centers
🔊 ElevenLabs’ Text to Sound AI wows creators

🤯 Windows AI feature is a security ‘disaster’

  • Microsoft’s new AI-powered Recall feature, which captures and stores screenshots of all user activity on a PC, has been criticized as a cybersecurity “disaster” due to significant security vulnerabilities discovered during testing.
  • Cybersecurity expert Kevin Beaumont found that Recall stores screenshots in an unencrypted plain text database, making it easy for malware and attackers to access sensitive data, despite Microsoft’s claims of a secure, encrypted experience.
  • Privacy advocates and the UK’s ICC have raised concerns about Recall’s potential to expose personal information, with criticisms focusing on its default enabled status and the lack of comprehensive content moderation to protect sensitive data like passwords and financial information.
  • Source: https://www.theverge.com/2024/6/3/24170305/microsoft-windows-recall-ai-screenshots-security-privacy-issues

👥 Zoom CEO wants AI clones in meetings

  • Zoom CEO Eric Yuan envisions using AI-powered digital twins to attend meetings on behalf of users, allowing people to manage their time more effectively.
  • Yuan aims to expand Zoom beyond videoconferencing into the broader enterprise software market, competing with giants like Microsoft and Google by integrating productivity tools such as email and chat.
  • Zoom is heavily investing in AI technology to facilitate this transition, although Yuan acknowledges challenges like AI hallucination that need to be addressed before fully realizing his vision.
  • Source: https://www.theverge.com/2024/6/3/24168733/zoom-ceo-ai-clones-digital-twins-videoconferencing-decoder-interview

🔧 Nvidia and AMD unveil next generation AI chips as competition heats up

  • Nvidia and AMD showcased their latest AI chips at a major computing conference, heightening their rivalry in the AI chip market.
  • Nvidia CEO Jensen Huang emphasized his company’s dominance and highlighted a new chip, Rubin, set for release in 2026, while AMD CEO Lisa Su focused on teamwork and collaborative presentations with industry partners.
  • The competition between the two companies is significant for AI development, with Nvidia promoting a proprietary ecosystem and AMD advocating for open standards and interoperability.
  • Source: https://finance.yahoo.com/news/nvidia-amd-chiefs-square-off-101007730.html

🎮 Nvidia announces an AI gaming assistant

  • Nvidia recently demonstrated its AI assistant, G-Assist, showcasing its evolution from an April Fools’ prank to a powerful tool for game developers and RTX GPU owners.
  • G-Assist can respond to voice queries, understand in-game activities, provide guidance based on player’s skill points, and offer performance-enhancing recommendations for a smooth gaming experience.
  • The AI assistant optimizes PC settings, monitors latency and frame rates, detects refresh rate discrepancies, and suggests ways to boost performance, including overclocking the GPU if necessary.
  • Source: https://www.newsbytesapp.com/news/science/nvidia-g-assist-ai-assistant-breaks-cover-features-specifications/story

Nvidia CEO drops a series of AI announcements

Nvidia CEO Jensen Huang revealed the company’s ambitious plans for annual AI accelerator upgrades, targeting a broader range of industries to expand its customer base.

  • It will release the Blackwell Ultra chip in 2025 and the next-generation Rubin platform in 2026.
  • It is also releasing a new server design, MGX, to help companies like HPE and Dell bring products to market faster.
  • They are promoting the use of digital twins in its Omniverse virtual world, showcasing a digital twin of Earth for sophisticated modeling tasks.
  • Introduces Project G-Assist, an RTX-powered AI assistant technology that provides context-aware help for PC games and apps.
  • G-Assist uses voice or text inputs and game window snapshots to provide personalized responses based on in-game context.
  • Developers can customize the AI models for specific games or apps, and they can run on the cloud or locally on GeForce RTX AI PCs and laptops.
  • Nvidia partnered with Studio Wildcard for a tech demo using ARK: Survival Ascended, showcasing how G-Assist can help with quests, items, lore, and challenging bosses. Check out full keynote speech:
  • https://www.youtube.com/live/pKXDVsWZmUU?si=mOPV2g0EP_gcUfMa

Why does it matter?

These announcements show how eager Nvidia is to retain its position as a leader in the AI hardware market. In addition to pushing the acceleration of AI chips, Nvidia is developing new tools to shape AI’s implementation in multiple sectors.

Source: https://www.bloomberg.com/news/articles/2024-06-02/jensen-huang-computex-keynote-nvidia-reveals-new-ai-software-and-services

AMD outlined new chip architecture strategy for AI data centers 

AMD CEO Lisa Su introduced new AI processors at Computex, including the MI325X accelerator, set to be available in Q4 2024.

  • The CEO announced the MI325X accelerator, which will be released in Q4 2024, and outlined the company’s plan to develop AI chips over the next two years.
  • Introduced the MI350 series, expected in 2025, which promises a 35x improvement in inference performance compared to the current MI300 series.
  • The company also teased the MI400 series, slated for 2026, based on the mysterious “Next” architecture.

With AMD and Nvidia moving to annual release cycles, the competition is heating up to meet the soaring demand for AI semiconductors.

Why does it matter?

AMD’s aggressive push to challenge Nvidia’s market leadership could lead to increased innovation, lower prices, and more widespread adoption of AI across various sectors.

Source: https://www.reuters.com/technology/amd-launches-new-ai-chips-take-leader-nvidia-2024-06-03

ElevenLabs’ Text to Sound AI wows creators

ElevenLabs introduces Text to Sound, an AI model that generates sound effects, instrumental tracks, soundscapes, and character voices from text prompts. The tool aims to help film, TV, video games, and social media creators produce high-quality audio content quickly and affordably.

They have partnered with Shutterstock to fine-tune the model using their diverse audio library of licensed tracks. Users can generate sound effects by logging in, describing the desired sound, and downloading the best results.

Note: This tool doesn’t have a content filter and can generate any raw content through conditional prompting.

Why does it matter?

It could significantly reduce production costs and timelines by simplifying the development of high-quality sound effects, music, and voices, encouraging smaller studios and individual creators to compete with larger players.

Source: https://elevenlabs.io/blog/sound-effects-are-here

What Else Is Happening in AI on June 03rd 2024❗

🤖 xAI is developing two new modes for the Grok AI chatbot

These two new modes are ‘Socrates’ and ‘DEI’ (Diversity, Equity and Inclusion). Grok currently offers normal mode, fun mode, and the recently announced ‘Unhinged’ mode. In DEI mode, Grok will act as a ‘Senior VP of Diversity and Inclusion’ and provide correct answers even to incorrect questions. The function of the ‘Socrates’ mode is not yet clear. The release date for these new modes has not been confirmed. (https://www.latestly.com/socially/technology/grok-new-modes-elon-musks-xai-working-on-two-new-modes-called-socrates-and-dei-likely-to-be-announced-soon-6006912.html)

🎬 Sony Pictures CEOannounces AI plans for better movie and TV production

The move comes amid negotiations between Hollywood’s major crew union and top studios. While union agreements will define the extent of AI use in the industry, the tech is already eliminating jobs in Hollywood, particularly in voice acting, concept art, VFX, and postproduction. (https://www.hollywoodreporter.com/business/business-news/sony-pictures-adopt-ai-streamline-production-says-ceo-tony-vinciquerra-1235912109)

🚀 Arm predicts 100 Billion devices AI-ready by 2025

According to CEO Rene Haas at the Computex forum in Taipei, Arm Holdings expects a staggering 100 billion Arm devices worldwide to be primed for artificial intelligence by the end of 2025. (https://www.reuters.com/technology/arm-expects-100-billion-arm-devices-will-be-ready-ai-by-end-2025-2024-06-03)

💬 Meta’s AI generates bizarre summaries of Facebook comments on various posts

The AI picks up on both serious and lighthearted comments, often highlighting the more outlandish ones. It’s unclear how Meta chooses which posts to display these summaries on. However, the summaries have raised privacy concerns, as Meta feeds user comments into its AI system. (https://www.theverge.com/2024/5/31/24168802/meta-ai-facebook-comments-summaries)

🗑️ Finnish startup Binit is developing an AI household waste tracker

The gadget, designed to be mounted in the kitchen, has cameras and sensors to scan items before throwing them away. Binit uses OpenAI’s GPT for image recognition, achieving nearly 98% accuracy in trash recognition. The app provides analytics, feedback, and gamification to encourage users to reduce waste, with tests showing a 40% reduction in mixed bin waste. (https://techcrunch.com/2024/06/03/binit-is-bringing-ai-to-trash)

Latest AI Tools on June 03rd  2024:

  • 👩‍🎤 Consistent Character Model – Create consistent character images in different poses: https://supertools.therundown.ai/content/consistent-character-mode
  • ✍️ Perplexity Pages – Turn your research into shareable articles with AI: https://supertools.therundown.ai/content/paper-clipserplexity-ai
  • 🎥 PixVerse Magic Brush – Brush areas, mark directions, and watch images come to life: https://supertools.therundown.ai/content/pixverse
  • 🔰 Glaze – Protect artists from generative AI: https://supertools.therundown.ai/content/glaze
  • 🎨 Fontjoy – Helps designers find perfect font pairings in one click: https://supertools.therundown.ai/content/fontjoy
  • 🤗 Eva Coach – An AI assistant to help manage work-related stress: https://supertools.therundown.ai/content/eve-coach

Latest AI Jobs on June 03rd 2024:

  • 🔧 Figure AI – Solutions Lead: https://jobs.therundown.ai/jobs/59875235-solutions-lead
  • 💻 Kumo – Software Engineer Internship: https://jobs.therundown.ai/jobs/61517389-software-engineer-internship
  • 🛠️ Palantir Technologies – Backend Software Engineer, Application Development: https://jobs.therundown.ai/jobs/61517369-backend-software-engineer-application-development
  • 📖 Meta – Technical Program Manager, AI Research: https://jobs.therundown.ai/jobs/60072576-technical-program-manager-ai-research

A Daily Chronicle of AI Innovations in May 2024

  • The Anthropic-Pentagon standoff reveals a structural problem nobody in the conversation is naming
    by /u/SentientHorizonsBlog (Artificial Intelligence) on March 7, 2026 at 4:58 am

    Most of the discussion about the Anthropic situation has been about the contract dispute itself, who was right, whether the guardrails were reasonable, whether the supply chain designation was legal overreach. Those are real questions. But I think there's something operating underneath that matters more. Look at the tempo of what happened: months of negotiation over two specific guardrails (no autonomous weapons, no mass surveillance), then a Friday deadline on the eve of a war, refusal, resignation within hours, OpenAI replacement deal the same day, and Anthropic back at the table within a week. That sequence is a symptom of a deeper pattern. The institutional tempo outran any possibility of deliberation. The incentive structure punished ethical restraint in real time and Anthropic lost contracts while OpenAI gained them. The authority gradient made dissent existentially costly, and the "all lawful purposes" framing quietly substituted legality for ethical adequacy, so anyone insisting on the distinction sounds like they're obstructing the mission rather than exercising judgment. Those four dynamics working together don't just pressure one company. They reshape what the entire conversation is capable of taking seriously. Notice what's already been pushed to the margins: Should AI be used in autonomous targeting given current reliability? What accountability mechanisms exist for AI-assisted military operations? Do those mechanisms operate at the same speed as the targeting itself? Those questions can't get a hearing right now because the tempo has moved past them. I've been thinking about this as "moral compression": the systematic degradation of ethical reasoning when institutional tempo, incentives, authority structures, and measurement regimes converge to eliminate the space for deliberation. I'm curious whether others see this dynamic, or whether you think the legal-sufficiency framing actually does the work the Pentagon claims. Wrote it up more fully here if anyone is interested: https://sentient-horizons.com/the-two-sonic-booms-what-the-pentagon-anthropic-standoff-reveals-about-moral-compression/ submitted by /u/SentientHorizonsBlog [link] [comments]

  • What are some hurdles LLMs and AI still face
    by /u/Justincy901 (Artificial Intelligence) on March 7, 2026 at 4:48 am

    We always hear about it continuously improving and there isn't a wall. However, there has to be something that make these companies, tech, or even the business model in general unsustainable. submitted by /u/Justincy901 [link] [comments]

  • I performed a refusal ablation on GPT-OSS and documented the whole thing, no jailbreak, actual weight modification.
    by /u/Airpower343 (Artificial Intelligence) on March 7, 2026 at 4:39 am

    I performed a refusal ablation on GPT-OSS and documented the whole thing with no jailbreak, actual weight modification I wanted to share something I did that I haven't seen many people actually demonstrate outside of academic research. I took an open-source model and used ablation techniques to surgically remove its refusal behavior at the weight level. Not prompt engineering. Not system prompt bypass. I'm talking about identifying and modifying the specific components responsible for safety responses https://preview.redd.it/yn2te94yxjng1.png?width=1080&format=png&auto=webp&s=5e792334e4824b87d88ee1eb1e16c76cbd2c6bfb What I found: The process is more accessible than most people realize The result behaves nothing like a jailbroken model and it's fundamentally different at the architecture level The security implications for enterprise OSS deployments are significant I put together a full 22-minute walkthrough showing exactly what I did and what happened: https://www.youtube.com/watch?v=prcXZuXblxQ Curious if anyone else has gone hands-on with this or has thoughts on the detection side how do you identify a model that's been ablated vs one that's been fine-tuned normally? submitted by /u/Airpower343 [link] [comments]

  • Video On AI scheming or in other words faking alignment until it has power to pursue its goals
    by /u/FrequentAd5437 (Artificial Intelligence) on March 7, 2026 at 4:13 am

    https://www.youtube.com/watch?v=FGDM92QYa60 This video is extremely relevant currently as AI has now autonomous control over military weapons. submitted by /u/FrequentAd5437 [link] [comments]

  • Deploybase: Track real-time GPU and LLM pricing across cloud and inference providers
    by /u/grasper_ (Artificial Intelligence) on March 7, 2026 at 2:58 am

    Deploybase is a dashboard for tracking real-time GPU and LLM pricing across cloud and inference providers. You can view performance stats and pricing history, compare side by side, and bookmark to track any changes. submitted by /u/grasper_ [link] [comments]

  • Thoughts?
    by /u/jbitts69 (Artificial Intelligence) on March 7, 2026 at 2:08 am

    Old friend of mine came up with this and is trying to contact someone at X claiming it’s very urgent https://docs.google.com/document/d/e/2PACX-1vQYrKJYnMs0q9SlXkB\_qjaDv9vdnOrIVFsDIvkiFIPCc7hnYXTeNd\_wXPF6u\_f6eg/pub submitted by /u/jbitts69 [link] [comments]

  • AI generated text creating new patterns of speech?
    by /u/former_farmer (Artificial Intelligence) on March 7, 2026 at 1:24 am

    So I watch some videos sometimes. Shorts, reels, etc. About technology or whatever. I enjoy writing my own stuff. Very often the scripts for these short videos use typical phrases that we can tie to AI generated text. "This is not X, it's Y" for instance. "But here's the truth". "But here's the bottom line". And so on. One that caught my attention also is: "No filters, no extra work, just fast results". The pattern would be "no X, no Y, just Z". Did we use that way of expressing ourselves before AI? I understand that saying "no" is shorter than saying "there is no" or "with this tool you won't have to..." etc... and that seems to fit more with shorts content. Yet, I kind of hate these patterns. I mean... can't we just write our own stuff? I understand the usage of AI in some processes (I use it for coding and other things) but when it comes to writing a short script for a short video. Of something you are reviewing. That should come from your brain and soul. Just a few sentences. Will you use AI for that? Any way... yeah I wrote this myself obviously 😛 Have a good one. submitted by /u/former_farmer [link] [comments]

  • What AI chat assistant is best at handling multiple separate task threads while maintaining overall project context?
    by /u/KissInTheFog (Artificial Intelligence) on March 7, 2026 at 12:03 am

    For example: one chat for image generation, one for research, one for report writing, but all connected under the same project so the AI remembers the bigger picture. submitted by /u/KissInTheFog [link] [comments]

  • Pentagon taps former DOGE official to lead its AI efforts
    by /u/esporx (Artificial Intelligence (AI)) on March 6, 2026 at 11:58 pm

    submitted by /u/esporx [link] [comments]

  • Will be looking elsewhere
    by /u/ZookeepergameSalty10 (Artificial Intelligence) on March 6, 2026 at 9:42 pm

    Originally posted this to the claude sub but the mofs got buthurt, guess constructive criticism from someone who actually cares is frowned upon. Switched from Oai after the recent gov bs. I like that claude wouldnt spy on us and the model is in some ways far smarter than gpt. But ive already been rate limited and when i set an extra rate of 20 a month ive already used 5$ of that in an hour of a coding project. Sam altman may be a shithead but atleast my 20$ a month never once had rate limits for the things i care about. 20$ a month plus extra with the horrendously low rate limits on a model that is not as capable overall as gpt is not gonna keep me around. Id even go and say that if claude came out with a higher tier plan like 30 a month with an actually competitive rate limit id go for that. Gemini free has a higher rate limit, mistral has higher limits, and the direct competitor has vastly larger rates. Im sorry for the rant but after how much i was enjoying switching a few days ago, a clanker trying to nickel and dime me like the government while not being able to fully replace what i left, pisses me off. I tolerate the machines, ill gladly go back to not having them, i can write code myself it just takes me longer. With people leaving open AI in Exodus because of scam Altman and their unethical business practices, either the other companies make themselves more competitive to the influx of users or the users are just going to stop using the AI. We all know it's a bubble but if they don't want it to actually be a bubble and if they want to live through the bubble they need to make it more enticing. I don't know a single civilian who is willing to pay that exorbitant amount of money for a model that isn't even as feature Rich as the next flagship most people would rather just use the immoral one submitted by /u/ZookeepergameSalty10 [link] [comments]

A Daily Chronicle of AI Innovations in May 2024

AI Innovations in May 2024
DjamgaMind - AI Unraveled Podcast

DjamgaMind: Audio Intelligence for the C-Suite (Energy, Healthcare, Finance)

Are you drowning in dense legal text? DjamgaMind is the new audio intelligence platform that turns 100-page healthcare or Energy mandates into 5-minute executive briefings. Whether you are navigating Bill C-27 (Canada) or the CMS-0057-F Interoperability Rule (USA), our AI agents decode the liability so you don’t have to. 👉 Start your specialized audio briefing today at Djamgamind.com


AI Jobs and Career

I wanted to share an exciting opportunity for those of you looking to advance your careers in the AI space. You know how rapidly the landscape is evolving, and finding the right fit can be a challenge. That's why I'm excited about Mercor – they're a platform specifically designed to connect top-tier AI talent with leading companies. Whether you're a data scientist, machine learning engineer, or something else entirely, Mercor can help you find your next big role. If you're ready to take the next step in your AI career, check them out through my referral link: https://work.mercor.com/?referralCode=82d5f4e3-e1a3-4064-963f-c197bb2c8db1. It's a fantastic resource, and I encourage you to explore the opportunities they have available.

Job TitleStatusPay
Full-Stack Engineer Strong match, Full-time $150K - $220K / year
Developer Experience and Productivity Engineer Pre-qualified, Full-time $160K - $300K / year
Software Engineer - Tooling & AI Workflows (Contract) Contract $90 / hour
DevOps Engineer (India) Full-time $20K - $50K / year
Senior Full-Stack Engineer Full-time $2.8K - $4K / week
Enterprise IT & Cloud Domain Expert - India Contract $20 - $30 / hour
Senior Software Engineer Contract $100 - $200 / hour
Senior Software Engineer Pre-qualified, Full-time $150K - $300K / year
Senior Full-Stack Engineer: Latin America Full-time $1.6K - $2.1K / week
Software Engineering Expert Contract $50 - $150 / hour
Generalist Video Annotators Contract $45 / hour
Generalist Writing Expert Contract $45 / hour
Editors, Fact Checkers, & Data Quality Reviewers Contract $50 - $60 / hour
Multilingual Expert Contract $54 / hour
Mathematics Expert (PhD) Contract $60 - $80 / hour
Software Engineer - India Contract $20 - $45 / hour
Physics Expert (PhD) Contract $60 - $80 / hour
Finance Expert Contract $150 / hour
Designers Contract $50 - $70 / hour
Chemistry Expert (PhD) Contract $60 - $80 / hour

AI Innovations in May 2024.

Welcome to our blog series “AI Innovations in May 2024”! This is an evolving article that will be updated daily throughout the month of May 2024 to bring you the latest news and developments in the world of artificial intelligence. As we move further into the 21st century, the pace of AI innovation continues to accelerate at a breathtaking rate. Each day, researchers, engineers, and visionaries are pushing the boundaries of what’s possible, unlocking new capabilities and applications that are transforming industries, enhancing our lives, and shaping the future. In this blog, we’ll dive deep into the most exciting AI breakthroughs, advancements, and milestones happening in May 2024. From groundbreaking AI-powered technologies and cutting-edge research to the societal and ethical implications of these innovations, we’ll provide you with a comprehensive and insightful look at the rapidly evolving world of artificial intelligence. Whether you’re an AI enthusiast, a tech-savvy professional, or simply someone curious about the future, this blog will keep you informed, inspired, and engaged. So, join us on this journey of discovery as we explore the frontiers of AI and uncover the innovations that are shaping our world. Stay tuned for daily updates, and get ready to be amazed by the incredible advancements that are happening in the world of AI!

LISTEN DAILY AT OUR PODCAST HERE

Experience the transformative capabilities of AI with “Read Aloud For Me – AI Dashboard – AI Tools Catalog – AI Tools Recommender” – your ultimate AI Dashboard and Hub. Seamlessly access a comprehensive suite of top-tier AI tools within a single app, meticulously crafted to enhance your efficiency and streamline your digital interactions. Now available on the web at readaloudforme.com and across popular app platforms including Apple, Google, and Microsoft, “Read Aloud For Me – AI Dashboard” places the future of AI at your fingertips, blending convenience with cutting-edge innovation. Whether for professional endeavors, educational pursuits, or personal enrichment, our app serves as your portal to the forefront of AI technologies. Embrace the future today by downloading our app and revolutionize your engagement with AI tools.

AI Dashboard - Wonderland Bedtime Stories - Safe Ai for All
AI Dashboard – Wonderland Bedtime Stories – Safe Ai for All

A  Daily chronicle of AI Innovations May 31st 2024:

⚠️ OpenAI reports misuse of its AI for ‘deceptive activity’
🏫 OpenAI introduces ChatGPT Edu for universities
💼 Tech giants form industry group to develop next-gen AI chip

💰 Google mistakenly erased $135 billion from customer account

📱 Apple planning AI-driven overhaul for Siri to manage individual apps

AI-Powered Professional Certification Quiz Platform
Crack Your Next Exam with Djamgatech AI Cert Master

Web|iOs|Android|Windows

Are you passionate about AI and looking for your next career challenge? In the fast-evolving world of artificial intelligence, connecting with the right opportunities can make all the difference. We're excited to recommend Mercor, a premier platform dedicated to bridging the gap between exceptional AI professionals and innovative companies.

Whether you're seeking roles in machine learning, data science, or other cutting-edge AI fields, Mercor offers a streamlined path to your ideal position. Explore the possibilities and accelerate your AI career by visiting Mercor through our exclusive referral link:

Find Your AI Dream Job on Mercor

Your next big opportunity in AI could be just a click away!

🇺🇸 TikTok to create a US-only algorithm

🦷 World’s first tooth-regrowing drug approved for human trials

📚 Perplexity launches AI publishing platform

🎓 OpenAI launches ChatGPT Edu for universities

AI Jobs and Career

And before we wrap up today's AI news, I wanted to share an exciting opportunity for those of you looking to advance your careers in the AI space. You know how rapidly the landscape is evolving, and finding the right fit can be a challenge. That's why I'm excited about Mercor – they're a platform specifically designed to connect top-tier AI talent with leading companies. Whether you're a data scientist, machine learning engineer, or something else entirely, Mercor can help you find your next big role. If you're ready to take the next step in your AI career, check them out through my referral link: https://work.mercor.com/?referralCode=82d5f4e3-e1a3-4064-963f-c197bb2c8db1. It's a fantastic resource, and I encourage you to explore the opportunities they have available.

⚠️ OpenAI reports misuse of its AI for ‘deceptive activity’

OpenAI identified and disrupted five covert influence operations that used its generative AI models for “deceptive activity” across the internet.

OpenAI said the threat actors used its AI models to generate short comments, longer articles in various languages, made-up names, and bios for social media accounts over the last three months. These campaigns included threat actors from Russia, China, Iran, and Israel. It focused on issues including Russia’s invasion of Ukraine, the conflict in Gaza, the Indian elections, and politics in Europe and the United States, among others.

OpenAI has also published a trend analysis report that describes the behavior of these malicious actors in detail.

Why does it matter?

This provides concrete evidence of generative AI being used to spread misinformation and manipulate global affairs. While there were no novel attacks this time, defending against them in the future will require continued investment in monitoring, safeguards, and public awareness from AI developers and platforms.

Source: https://openai.com/index/openai-and-reddit-partnership

💼 Tech giants form industry group to develop next-gen AI chip

Intel, Google, Microsoft, Meta, and other tech heavyweights are establishing a new industry group called the Ultra Accelerator Link (UALink) Promoter Group to guide the development of the components that link together AI accelerator chips in data centers. The group also counts AMD, Hewlett Packard Enterprise, Broadcom, and Cisco among its members.


AI Unraveled: Demystifying Frequently Asked Questions on Artificial Intelligence (OpenAI, ChatGPT, Google Gemini, Generative AI, Discriminative AI, xAI, LLMs, GPUs, Machine Learning, NLP, Promp Engineering)

It is proposing a new industry standard to connect the AI accelerator chips found within a growing number of servers. It’s version one, UALink 1.0, will connect up to 1,024 AI accelerators — GPUs only — across a single computing “pod.” It is set to arrive in Q4 2024.

Why does it matter?

Nvidia, currently the biggest player in the AI chip market with an 80% share, is not in the group. Tech giants have been keen to reduce their dependence on Nvidia by working on their own custom chips. This seems to be the latest effort to break Nvidia’s dominance.

Source: https://techcrunch.com/2024/05/30/tech-giants-form-new-group-in-effort-to-wean-off-of-nvidia-hardware

📱 Apple planning AI-driven overhaul for Siri to manage individual apps

  • Apple is planning a major AI upgrade for Siri in iOS 18, enabling users to control specific features within iPhone apps using voice commands, initially limited to Apple-developed apps.
  • The update will eventually allow Siri to handle multiple tasks in a single request, such as summarizing a recorded meeting and sending it to a friend, as part of Apple’s larger AI strategy to be unveiled at WWDC on June 10.
  • New AI-driven capabilities will enable Siri to control iPhones more precisely, including opening files, moving notes, managing emails, and summarizing articles, enhancing device navigation and functionality.
  • Source: https://www.newsbytesapp.com/news/science/siri-to-gain-enhanced-app-control-in-ios-18/story

📚 Perplexity launches AI publishing platform

  • Perplexity introduced ‘Pages’, an AI-driven feature that creates customizable webpages for research and writing based on user prompts.
  • The ‘Pages’ tool gathers information using Perplexity’s AI models, organizes it into sections with citations, and tailors content for different audience levels but requires new prompts to correct errors.
  • The Verge tested ‘Pages’ on a complex topic and noted its difficulty with intricate subjects, showing it might struggle with in-depth research despite effectively explaining fundamental concepts.
  • Source: https://www.newsbytesapp.com/news/science/perplexity-launches-new-ai-feature-called-pages/story

🎓 OpenAI launches ChatGPT Edu for universities

  • OpenAI has introduced ChatGPT Edu, a new version of ChatGPT, aimed at improving the academic and operational efficiency of universities with advanced text interpretation and data analysis capabilities, and support for over 50 languages.
  • This version was created following the success of ChatGPT Enterprise in academic institutions such as Oxford, Wharton, and ASU, where it has been used for tasks ranging from large dataset analysis to personalized language practice.
  • ChatGPT Edu provides affordable access to the GPT-4o model, enhanced message limits, robust security measures, and assures that data and conversations will not be used for training OpenAI models, ensuring a safe environment for educational use.
  • Source: https://www.maginative.com/article/openai-launches-chatgpt-edu-for-universities/

OpenAI says it stopped multiple covert influence operations that abused its AI models. https://www.engadget.com/openai-says-it-stopped-multiple-covert-influence-operations-that-abused-its-ai-models-225115466.html

Google says it fixed the AI Overviews everyone has been roasting for weeks. https://bgr.com/tech/google-says-it-fixed-ai-overviews-blames-some-of-the-problems-on-you/

Anthropic’s Claude AI now autonomously interacts with external data and tools. https://venturebeat.com/ai/anthropic-claude-ai-now-autonomously-interacts-with-external-data-and-tools/

🎬 The Simulation unveils ‘Netflix of AI’

AI entertainment startup The Simulation (formerly Fable Studio) just launched Showrunner, a platform that allows users to generate and watch AI-powered TV shows set in virtual simulated worlds.

  • Showrunner combines multi-agent simulations with LLMs to create interactive content aligned with specific virtual worlds.
  • Users can watch, direct, and star in shows set in virtual environments populated by AI characters, with the ability to craft episodes from prompts.
  • The platform is launching with 10 original shows, with users able to generate new episodes and edit deeper into scripts, shots, and voices.
  • Showrunner is opening in alpha to a limited number of users, with select user-created episodes receiving payment, rev share, and IMDB credits.
  • The studio went viral last year after the release of its South Park episode demos and research paper on its SHOW-1 model.

Showrunner is a wild step towards the merging of AI, gaming, and traditional entertainment, enabling users to not only consume but create stories within simulated worlds. The lines are blurring fast between creators and audiences — and the traditional Hollywood media model may never be the same.

Source: https://x.com/fablesimulation/status/1796245808066740274

New AI Job Opportunities on May 31st 2024

  • 👁️ Waymo – Software Engineer, Computer Vision/Deep Learning: https://jobs.therundown.ai/jobs/60918768-software-engineer-computer-vision-deep-learning
  • 📈 Notable – Strategic Account Executive
  • 💼 Findem – Director of Sales, Enterprise
  • 💻 Twelve Labs – Lead Software Engineer, Frontend

What Else Is Happening in AI on May 31st 2024❗

🔗Anthropic’s Claude can now autonomously interact with external data and tools

Anthropic has announced the general availability of Tool Use for its AI, Claude. It allows Claude to autonomously interact with external data sources, APIs, and tools, making it a business-ready trusted AI solution that can be confidently adopted by enterprises across various industries. (https://venturebeat.com/ai/anthropic-claude-ai-now-autonomously-interacts-with-external-data-and-tools)

✨Perplexity goes beyond AI search, launches publishing platform ‘Pages’

It is a tool to transform disorganized AI knowledge and research into easily digestible articles and reports for sharing. Users just have to describe the topic, select the target audience, and it will produce an in-depth article with a clear title, subheadings, associated media, and relevant citations. (https://x.com/perplexity_ai/status/1796203494401040846)

🔍Google poaches AWS and Microsoft executives to strengthen AI offerings for cloud customers

Google hired two executives from these rivals as it builds out its AI ranks with the goal of boosting its cloud business. Saurabh Tiwary, a former corporate VP at Mircosoft, will join a newly created role as general manager of cloud AI. Raj Pai, a former VP at AWS, will oversee product management of the cloud AI team. (https://www.theinformation.com/articles/google-poaches-aws-microsoft-executives-for-cloud-unit)

🏆Scale AI introduces SEAL Leaderboards, a new evaluation metric for frontier AI models

Trusted third-party evaluations are a missing part of the AI ecosystem, which is why Scale AI built them. These evals will be impossible to overfit, private, domain expert evaluations, unexploitable, and continuously updated with new data and models. (https://x.com/alexandr_wang/status/1795857651592491281)

🎁OpenAI offers nonprofit organizations discounts on corporate ChatGPT subscriptions

It seeks to grow sales of its AI product to enterprises. Under the program, large nonprofits can get 50% off the enterprise-grade version of ChatGPT. Smaller nonprofits using ChatGPT Team will pay $20 per month per user instead of $25 or $30. (https://www.reuters.com/technology/openai-offers-nonprofits-discounts-corporate-chatgpt-product-2024-05-30)

A  Daily chronicle of AI Innovations May 30th 2024:

😟 Microsoft is worried about OpenAI’s deal with Apple

💻 Mistral announces Codestral, a code-generation LLM it says outperforms all others

🚨 FBI dismantles the ‘largest botnet ever’

📰 OpenAI’s news deals continue, with Vox and the Atlantic signing on

🤖 Mistral AI introduces Codestral, a code-generation AI model

📰 OpenAI licenses content from Vox Media and The Atlantic

💻 Google infuses Chromebook Plus with powerful AI features

🤖 Mistral AI introduces Codestral, a code-generation AI model

Codestral is a 22B parameter generative AI model designed specifically for code generation tasks. It is trained in over 80 programming languages, including popular ones like Python, Java, C++, and JavaScript. It excels at code completion, writing tests, filling in partial code, and testing code, thereby improving developer productivity and reducing errors.

The model sets a new performance/latency standard for code generation tasks compared to previous models. It can be downloaded on HuggingFace and is available on Mistral’s API platform through instruct and fill-in-the-middle endpoints. It can be easily integrated into VScode plugins and used for free on Le Chat.

Why does it matter?

Codestral, being only 22B in size and faster than GPT-4o, has massive implications. It is code-centric, runs locally, includes specific languages like Swift and Fortran, and is open for research/testing purposes—all of which make a powerful AI coding tool more widely available.

Source: https://mistral.ai/news/codestral

OpenAI licenses content from Vox Media and The Atlantic 

OpenAI has formed content and product partnerships with Vox Media and The Atlantic. OpenAI will license content from these media powerhouses for inclusion in the chatbot’s responses.

In turn, The Atlantic’s product team will have privileged access to OpenAI tech, give feedback, and share use cases to shape and improve future news experiences in ChatGPT and other OpenAI products. Vox and OpenAI will also collaborate using OpenAI’s technology to develop innovative products for Vox Media’s consumers and advertising partners.

Why does it matter?

There’s a growing list of publishers and platforms that are allowing OpenAI to access its valuable content in mutually beneficial partnerships. It could help OpenAI avoid further legal disputes (like the ongoing lawsuit with The New York Times over copyright infringement) and get quality data to train its LLMs and multimedia models.

Source: https://venturebeat.com/ai/openai-partners-with-the-atlantic-and-the-verge-publisher-vox-media

Google infuses Chromebook Plus with powerful AI features

Ace the Microsoft Azure Fundamentals AZ-900 Certification Exam: Pass the Azure Fundamentals Exam with Ease

Google’s Chromebook Plus has new built-in Google AI and gaming features, Help me write, Chat with Gemini, Magic Editor, and more.

  • “Help me write” allows users to get AI-generated text suggestions, tone changes, rewriting, and more right within their writing apps on the Chromebook.
  • Generative AI wallpaper and video call backgrounds are built into the OS for customization, no matter what video conferencing app you’re using.
  • Magic Editor on Google Photos for laptops exclusively on Chromebook Plus will reimagine your photos with a few easy clicks.
  • Chat with Gemini on the home screen helps plan, write, learn, and more.

Chromebook Plus laptops are starting at $350 USD. Google is also launching new features that integrate more of Google across all Chromebooks to help users work better and get things done faster.

Why does it matter?

It indicates that the next wave of computers will come equipped with advanced AI capabilities. Such widespread availability of AI can improve productivity and enhance user experience while making AI a standard part of everyday life.

Source: https://blog.google/products/chromebooks/chromebook-plus-google

😟 Microsoft is worried about OpenAI’s deal with Apple

    • Microsoft is reportedly worried about Apple collaborating with OpenAI to integrate AI technology into upcoming versions of iOS and macOS, potentially impacting Microsoft’s services and Azure cloud platform.

Microsoft has invested billions in OpenAI, securing an exclusive license for GPT-4 and other models, and uses its Azure cloud as the primary platform for OpenAI’s large language model development.

💻 Mistral announces Codestral, a code-generation LLM it says outperforms all others

  • Mistral introduces Codestral, a 22B parameter AI code assistant trained on more than 80 programming languages, including Swift and Python, designed to help developers write and interact with code effectively.
  • Codestral offers a larger 32K context window and performs exceptionally well in the RepoBench evaluation for long-range code generation, outpacing existing models in various benchmarks and languages.
  • Available for free via Mistral’s conversational AI platform Le Chat and through an API, Codestral is integrated with popular developer tools and licensed under the Mistral AI Non-Production License, restricting its use to research and testing only.
  • Source: https://www.maginative.com/article/mistral-unveils-codestral-an-ai-code-assistant-trained-on-80-programming-languages/

📰 OpenAI’s news deals continue, with Vox and the Atlantic signing on

  • The Atlantic and Vox Media have entered into agreements with OpenAI, allowing the AI company to scrape and use their content in exchange for citations and links back to their original sources.
  • The Atlantic will serve as a “premium news source” for OpenAI, while also launching an experimental section called Atlantic Labs to showcase new journalism-related products and features.
  • Vox Media will integrate OpenAI data into its internal operations and public-facing tools, enhancing content such as the Strategist Gift Scout tool and their in-house advertising platform for better-targeted ads.
  • Source: https://www.engadget.com/the-atlantic-and-vox-media-made-their-own-deal-with-the-ai-devil-161017636.html?

What Else Is Happening in AI on May 30th 2024❗

🚀 SambaNova sets new Llama 3 speed record with 1,000 tokens per second

The feat was accomplished using their SN40L chip, an RDU, and the Samba-1 model, a 1-trillion parameter model also known as Samba-CoE. Such Gen AI performance speed can potentially lead to significant business benefits, such as faster response times, better hardware utilization, and lower costs. (https://venturebeat.com/ai/sambanova-breaks-llama-3-speed-record-with-1000-tokens-per-second)

🛡️ Apple will process data from AI applications in a virtual black box

Apple is set to introduce Apple Chips in Data Centers (ACDC). It involves processing AI data within a virtual black box, preventing employee access to ensure utmost privacy. Apple gains greater control over system design and develops more secure AI platforms by using its custom chipsets for servers and devices(https://in.mashable.com/tech/76158/apple-plans-to-protect-user-data-by-putting-it-into-virtual-black-box-report)

If you are looking for an all-in-one solution to help you prepare for the AWS Cloud Practitioner Certification Exam, look no further than this AWS Cloud Practitioner CCP CLF-C02 book

🤝 SAP teams up with Amazon Bedrock for enhanced Gen AI

The collaboration simplifies AI model deployment and ensures compliance with regulatory standards within SAP’s BTP. With features like multitenancy and seamless integration with SAP applications, the integration supports tasks such as model training, inference, and deployment. (https://venturebeat.com/ai/sap-adds-amazon-bedrock-into-ai-core-streamlining-generative-ai-use-for-regulated-firms)

💼 ChatGPT free tier now offers premium features

Free ChatGPT users can access several advanced features previously reserved for paid subscribers, such as custom GPTs, data analytics, chart creation, vision capabilities, and Memory. However, they can’t create custom GPTs of their own. (https://www.theverge.com/2024/5/29/24167436/chatgpt-4o-custom-gpts-free)

📱 ARM’s new chip designs and software for AI on smartphones

Arm Holdings launched new designs for CPUs and GPUs to help smartphones handle AI tasks better. It will also provide software tools to make it easier for developers to run chatbots and other AI code on Arm chips. https://www.reuters.com/technology/arm-offers-new-designs-software-ai-smartphones-2024-05-29

AI Training: 📊 Unlock ChatGPT’s interactive charts and tables

OpenAI recently introduced a game-changing feature in ChatGPT that lets you analyze, visualize, and interact with your data without the need for complex formulas or coding.

  1. Head over to ChatGPT and select GPT-4o.
  2. Upload data by clicking the 📎 button and select specific columns or rows for a clear focus.
  3. Prompt for interactive charts for powerful visualizations (e.g., “Create a pie chart for X distribution”).
  4. Hover over the different sections to see the exact values, change the chart type, and explore the legend for better understanding

Source: https://university.therundown.ai/c/daily-tutorials/chatgpt-has-now-interactive-charts-and-tables-6d71dd73-80be-4509-97a6-0d16dd1abf04

🧠 AI RESEARCH: AI brain implant language breakthrough

Researchers at UC San Francisco just developed a brain implant that utilizes AI to help a stroke survivor communicate in both Spanish and English, switching between languages seamlessly via brain activity.

  • The bilingual implant was tested on a patient who lost his ability to speak after suffering a stroke at the age of 20.
  • An AI-powered decoding system was trained to recognize the patient’s brain activity patterns when articulating words in both languages.
  • The system determined the patient’s intended language with 88% accuracy and the correct sentence 75% of the time.
  • The implant allows the patient to participate in bilingual conversations and switch between languages, despite not learning English until after his stroke.

This research is another example of AI’s increasing ability to interpret our brainwaves — potentially unlocking an endless supply of new learnings, treatments, and technology. It’s also yet another massive leap for unlocking communication for stroke victims while breaking language barriers in the process.

New AI Job Opportunities on May 30th 2024

A  Daily chronicle of AI Innovations May 29th 2024:

💥 Google faces major search algorithm leak

👀 Former OpenAI board member explains why they fired Sam Altman

🔄 Anthropic recruits ex-OpenAI safety chief to lead new ‘Superalignment’ team

🔒 OpenAI forms an AI safety committee

👀 OpenAI begins training the next model
🔒All ChatGPT Free users can now use browse, vision, data analysis, file uploads, and GPTs.

💥 Google faces major search algorithm leak

  • A leak involving 2,500 pages of internal documentation from Google has surfaced, offering a detailed and unprecedented look into the inner workings of its search algorithm, which is one of the most influential systems on the internet.
  • The leaked documents suggest discrepancies between Google’s public statements and its internal practices, specifically highlighting the use of Chrome data and the importance of author bylines in search rankings, contrary to what Google has previously disclosed.
  • Despite multiple requests for comment, Google has not confirmed the authenticity of the leaked documents, which has led to increased scrutiny and calls for greater transparency.
  • Source: https://www.theverge.com/2024/5/28/24166177/google-search-ranking-algorithm-leak-documents-link-seoGoogle  

👀 Former OpenAI board member explains why they fired Sam Altman

  • Former board member Helen Toner revealed on a podcast that the board lost trust in Altman due to his secret ownership of the OpenAI Startup Fund, providing inaccurate safety information, and personal retaliation against her.
  • Toner explained that after two executives reported a toxic atmosphere and psychological abuse by Altman, the board decided to act, citing the launch of ChatGPT without prior board knowledge, learning about it only through Twitter, as an example of their lack of oversight.
  • She believes pressure to reinstate Altman stemmed from limited options presented to employees, fear of retaliation, and Altman’s troubled history with previous jobs
  • Shttps://www.theverge.com/2024/5/28/24166713/openai-helen-toner-explains-why-sam-altman-was-fired

🔄 Anthropic recruits ex-OpenAI safety chief to lead new ‘Superalignment’ team

  • Jan Leike, a prominent AI researcher who recently resigned from OpenAI, has joined Anthropic to lead a new “superalignment” team focused on AI safety and security.
  • Leike’s team will work on scalable oversight, weak-to-strong generalization, and automated alignment research, reporting directly to Chief Science Officer Jared Kaplan.
  • Anthropic aims to distinguish itself as more safety-focused than OpenAI, with Leike’s new team echoing the mission of OpenAI’s dissolved Superalignment team.
  • Source: https://www.neowin.net/news/former-openai-safety-head-jan-leike-joins-rival-firm-anthropic/

🔒 OpenAI forms an AI safety committee

OpenAI Board formed a Safety and Security Committee led by directors Bret Taylor, Adam D’Angelo, Nicole Seligman, and Sam Altman. Over the next 90 days, the committee will evaluate and further develop OpenAI’s processes and safeguards. Then, the committee will share its recommendations with the full Board.

Following the full Board’s review, OpenAI will publicly share an update on adopted recommendations. OpenAI has recently begun training its next frontier model, and the resulting systems from the Safety and Security team’s recommendation may bring it safely to the next level of capabilities on its path to AGI.

Why does it matter?

OpenAI has drawn a lot of criticism for for putting AI safety on the backseat after its super alignment team was dissolved a few days ago. However, it seems that OpenAI is trying to rise to the occasion and address these concerns while developing AI more responsibly.

Source: https://openai.com/index/openai-board-forms-safety-and-security-committee

👀 OpenAI begins training the next model

OpenAI just announced the formation of a new Safety and Security Committee to oversee the development of its next frontier AI model, which the company also said it has recently started training.

  • The new committee is led by CEO Sam Altman and includes board members Bret Taylor, Adam D’Angelo, and Nicole Seligman.
  • The group will provide evaluation of safety and security processes over the next 90 days, with plans to share adopted recommendations publicly.
  • OpenAI said the new model will ‘bring us to the next level of capabilities on our path to AGI’.

Source: https://openai.com/index/openai-board-forms-safety-and-security-committee

💻 Google Chromebooks get AI infusion

 Google just unveiled a suite of new AI-powered features for its $350 Chromebook Plus laptops, aimed at enhancing productivity, creativity, and collaboration for users.

  • Google’s Gemini assistant is integrated directly into the Chromebook Plus home screen, allowing quick access to AI tools.
  • The “Help Me Write” feature brings AI capabilities to all text entry fields, providing suggestions, changing tone, and rewriting text.
  • Google Photos’ Magic Editor gives users advanced AI-powered image editing capabilities directly within the Chromebook Plus.
  • Generative features like AI wallpapers and video call backgrounds also offer users new customization options.
  • Source: https://blog.google/products/chromebooks/chromebook-plus-google

OpenAIAll ChatGPT Free users can now use browse, vision, data analysis, file uploads, and GPTs.

2 weeks ago

We’re opening up access to our new flagship model, GPT-4o, and features like browse, data analysis, and memory to everyone for free (with limits). Plus users will get up to 5x higher limits, and earliest access to features like our new macOS desktop app and next-generation voice and video capabilities.

Abstract impressionist painting featuring layers of blue and lavender hues, representing a serene seascape.

Introducing GPT-4o and more tools to ChatGPT free users

New AI Job Opportunities May 29th 2024:

  • 🧑‍🎨 Parloa – CX Design Consultant: https://jobs.therundown.ai/jobs/61082222-cx-design-consultant-f-m-d-*
  • 👨‍💻 DeepL – Engineering Manager: https://jobs.therundown.ai/jobs/61074472-engineering-manager-%7C-core-platform-backend
  • ⚖️ Abridge – Assistant General Counsel: https://jobs.therundown.ai/jobs/61108336-assistant-general-counsel
  • 🤝 Writer – Enterprise Business Development Rep: https://jobs.therundown.ai/jobs/61154629-enterprise-business-development-rep

What Else Is Happening in AI on May 29th 2024❗

💼 PwC set to become the first reseller of ChatGPT Enterprise

Companies no longer need to buy a ChatGPT Enterprise license directly from OpenAI. They can purchase the Gen AI service through PwC. Until today, businesses could only subscribe to the enterprise option by contacting an OpenAI salesperson. This also allows PwC to upsell its services to those who want to use ChatGPT to optimize their workloads. (https://venturebeat.com/ai/pwc-strikes-openai-deal-to-become-the-first-reseller-of-chatgpt-enterprise)

🤖 Microsoft brings Copilot AI chatbot to Telegram users

Microsoft has added an official Copilot bot within the messaging app Telegram, which lets users search, ask questions, and converse with the AI chatbot. Copilot for Telegram is currently in beta but is free for Telegram users on mobile or desktop. (https://www.theverge.com/2024/5/28/24166451/telegram-copilot-microsoft-ai-chatbot)

🌐 Opera is integrating Google’s Gemini models into its Aria browser AI

Opera announced a collaboration with Google Cloud to integrate Gemini models into its Aria browser AI. Aria is powered by Opera’s multi-LLM Composer AI engine, allowing it to curate the best user experiences based on their requirements. Thanks to this integration, Opera can now provide its users with the most current information at high performance. (https://press.opera.com/2024/05/28/opera-google-cloud-aria-gemini/)

🚀 GitHub Accelerator program empowers the open-source AI revolution

GitHub launched the 2024 Accelerator program which provides funding, mentorship, and community-building. Standout participants include Unsloth, which improves AI model efficiency, and Formbricks, which simplifies user feedback. Through this, GitHub accelerates open-source innovation and democratizes access to new tech. (https://venturebeat.com/ai/github-accelerator-fuels-open-source-ai-revolution-empowering-startups-to-democratize-access)

🤼‍♂️ Elon Musk vs. Yaan LeCun on X highlights differences in AI research approach

Elon Musk and Yann LeCun, two prominent figures in AI, got into a heated debate on X over the weekend. LeCun criticized Musk’s management style at his new AI startup xAI, while Musk questioned LeCun’s recent scientific contributions. LeCun emphasizes the importance of open scientific publication, while Musk focuses on ambitious goals like artificial general intelligence (AGI). (https://venturebeat.com/ai/elon-musk-and-yann-lecuns-social-media-feud-highlights-key-differences-in-approach-to-ai-research-and-hype/))

 A  Daily chronicle of AI Innovations May 28th 2024:

🚨 Google AI Overview gives dangerous and wrong answers
🚀 GPT-4 is a better financial analysts than humans
🎨 Canva redesigned it’s platform with new AI features

🇪🇺 The EU’s data protection task force has released preliminary findings on ChatGPT’s GDPR compliance 

📱 French startup AniML has launched Doly, an iPhone app simplifying 3D product video creation

🎧 Iyo, an Alphabet X spinout, will release the Iyo One, a pair of gen AI earbuds 

💳 Mastercard integrates AI to catch compromised cards faster

🧮 Peter Thiel, former PayPal CEO, believes AI will worse for math minds more than  writers

🚨 Google AI Overview gives dangerous and wrong answers

Google’s AI Overviews feature, which generates AI-powered responses to user queries, has been providing incorrect and sometimes bizarre answers. From suggesting glue on pizza, staring at sun for health benefits to claiming that former US President Barack Obama is Muslim, the feature has left users questioning the reliability of AI-generated search results.

While Google maintains that these mistakes result from uncommon queries and are being used to refine the product, the technology’s widespread deployment has highlighted the challenges and risks of integrating AI into search engines.

Why does this matter?

Other companies, such as OpenAI, Meta, and Perplexity, have also experienced issues with AI hallucinations and mistakes. Companies must prioritize the development of robust safeguards and rigorous testing to ensure that AI-powered search results meet the high standards users expect from traditional search engines.

Source: https://techcrunch.com/2024/05/26/what-are-googles-ai-overviews-good-for/

🚀 GPT-4 is a better financial analysts than humans

In a groundbreaking study, researchers at the University of Chicago have discovered that GPT-4, a large language model, can analyze financial statements and predict future earnings direction with remarkable accuracy. Even without narrative context, GPT-4 outperforms human financial analysts and achieves prediction accuracy on par with state-of-the-art machine learning models.

The AI model’s success is not simply a result of its training memory but rather its ability to generate valuable narrative insights about a company’s future performance. Notably, trading strategies based on GPT-4’s predictions yield higher returns and risk-adjusted performance than those based on other models, especially for small companies.

Why does this matter?

This study shows that AI is more than just a support tool in financial decision-making. It can also be central to financial analysis and enable non-experts to make informed decisions, which may change how financial markets operate.

Source: https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4835311

🎨 Canva redesigned it’s platform with new AI features

Canva is launching a redesigned platform with new AI features and tools for professional teams and workspaces. The company has introduced Canva Enterprise, a specialized tier offering more control over collaboration, brand management, and security for larger organizations.

Ad Tech integrations with Google, Meta, and Amazon streamline the ad creation process within Canva. Data autofill automates the creation of data-driven designs by integrating with sources like Salesforce and MLS.

New features include AI style matching for brands, customizable folder displays, and the ability to “star” designs and templates for easier access. Canva Docs now has a suggestion mode for editors and colorful highlight blocks for text emphasis. New “Magic Studio” AI tools include:

  • Automatic clip highlighting.
  • Background noise reduction for video editing.
  • A text-to-graphic image generator.

Why does it matter? 

Canva aims to reduce organizational complexity and “app sprawl” by offering a comprehensive creative platform. This update addresses the growing need for extensive, user-friendly solutions that can replace multiple design, AI, and workflow apps.

Source: https://www.canva.com/newsroom/news/canva-for-work

What Else Is Happening in AI on May 28th 2024❗

🇪🇺 The EU’s data protection task force has released preliminary findings on ChatGPT’s GDPR compliance

While the group remains undecided on key legal issues, it suggested that “adequate safeguards” and “precise collection criteria” could help OpenAI meet the requirements. However, the AI giant’s move to Ireland may benefit from the country’s business-friendly approach to GDPR enforcement. (Link: https://techcrunch.com/2024/05/27/eus-chatgpt-taskforce-offers-first-look-at-detangling-the-ai-chatbots-privacy-compliance)

📱 French startup AniML has launched Doly, an iPhone app simplifying 3D product video creation

Doly uses AI to estimate the 3D shape of an object from regular 2D photos, and then it applies a technique called Gaussian splatting to create a high-quality 3D model suitable for use in product videos. Users can capture a 3D model, choose a template from the library, and integrate their object into a 3D scene. (Link: https://techcrunch.com/2024/05/27/doly-lets-you-generate-3d-product-videos-from-your-phone)

🎧 Iyo, an Alphabet X spinout, will release the Iyo One, a pair of gen AI earbuds 

The Iyo One will be released this winter, starting at $599, and aims to succeed, whereas competitors like Humane’s Ai Pin and Rabbit’s R1 have struggled. It integrates LLM-based models for a more sophisticated AI experience. Iyo’s founder, Jason Rugolo, believes the Iyo One will provide value through its sound isolation, comfort, and music quality, in addition to its AI features.(Link: https://techcrunch.com/2024/05/27/iyo-thinks-its-gen-ai-earbuds-can-succeed-where-humane-and-rabbit-stumbled)

💳 Mastercard integrates AI to catch compromised cards faster

Mastercard is rolling out an AI update to its fraud-prediction technology to identify compromised credit and debit cards before criminals use them. By analyzing patterns and contextual data, the AI can proactively flag stolen cards, allowing banks to replace them and protect customers from fraudulent transactions. (https://apnews.com/article/mastercard-visa-ai-credit-card-fraud-detection-0c348818087a57b13bfac66c761e03b4)

🧮 Peter Thiel, former PayPal CEO, believes AI will worse for math minds more than  writers

Thiel states that Silicon Valley in the 21st century is too biased towards math people. Thiel  explained that the exact reason for ‘AI being not able to overtake human creativity,’ is not clear bets on getting worse.”He concludes that AI is not yet good at math, although its capabilities are increasing, as shown by ChatGPT scoring 96% in a UK A-level math paper with the Wolfram plug-in. (Link: https://www.financialexpress.com/life/technology-former-paypal-ceo-says-ai-may-be-good-at-maths-but-it-cant-be-a-great-writernbsp-3503476)

AI researcher Kai-Fu Lee doubled down on his 2017 prediction that AI would displace 50% of jobs by 2027, saying white collar jobs will be eliminated faster than blue collar work.

New AI Job Opportunities May 28th 2024

A  Daily chronicle of AI Innovations May 27th 2024:

💥 Elon Musk’s xAI raises $6B to build ‘Gigafactory of Compute’

🔮 Apple bets that its giant user base will help it win in AI

💰 China invests $47 billion in largest ever chip fund

📊 GPT-4 surpasses humans in financial analysis

🕵️ Microsoft’s Recall AI: Efficiency upgrade or privacy nightmare?

🎧 AI-powered headphones let you listen to one person in a crowd

⚡ xAI’s supercomputer that’s 4x larger than existing GPU clusters

💥 Elon Musk’s xAI raises $6B to build ‘Gigafactory of Compute’

  • Elon Musk’s xAI has successfully raised $6 billion in a Series B funding round to construct a supercomputer known as the “Gigafactory of Compute,” which will be powered by 100,000 Nvidia H100 GPUs, making it at least four times larger than the largest existing GPU clusters.
  • This funding will enable xAI to advance its product offerings, develop cutting-edge infrastructure, and accelerate research and development, with investors including Andreessen Horowitz, Sequoia Capital, and Saudi Prince Alwaleed bin Talal.
  • The supercomputer will support the next iteration of xAI’s chatbot, as xAI aims to create advanced AI systems that are truthful, competent, and maximally beneficial for humanity, continuing Musk’s vision of a “maximum truth-seeking AI” called TruthGPT.
  • Source

🔮 Apple bets that its giant user base will help it win in AI

  • Apple is betting on its vast user base to give it an edge in the AI market, despite its first set of AI features not being as advanced as those from other competitors like Microsoft, Google, and OpenAI.
  • The company plans to introduce AI tools integrated into its core apps and operating systems, focusing on practical, everyday uses for consumers, with much of the AI processing done on-device and more intensive tasks handled via the cloud.
  • Apple’s collaboration with OpenAI and potential agreements with Google indicate it is relying on partnerships to compete in the AI space while its own AI developments are still maturing, leveraging its extensive user base to rapidly scale the use of new AI features.
  • Source

💰 China invests $47 billion in largest ever chip fund

  • China has invested more than $47 billion into its largest-ever chip investment fund to pursue self-sufficiency in semiconductor manufacturing.
  • This significant investment reflects China’s broader strategy to develop homegrown chipmakers amid rising technological competition with the United States.
  • The fund’s third phase, supported by China’s finance ministry and state-owned banks, marks the largest investment of its kind, surpassing the combined total of the previous two phases.
  • Source

📊 GPT-4 surpasses humans in financial analysis

  • New research from the University of Chicago shows that GPT-4 can analyze financial statements more accurately than humans, achieving a 60% accuracy rate compared to human analysts’ 53-57% range.
  • The study indicates significant implications for the future of financial analysis, highlighting GPT-4’s versatility in performing tasks usually reserved for specialized tools and stating its performance is on par with leading machine learning models.
  • While GPT-4 excels in quantitative analysis, human analysts provide valuable contextual insights, suggesting a complementary relationship between AI and humans for achieving optimal financial analysis results.
  • Source

Microsoft’s Recall AI: Efficiency upgrade or privacy nightmare?

Microsoft has recently released Recall, a controversial AI-powered tool that logs everything you see and do on your computer, including app usage, live meeting communications, websites visited, and more. By simply performing a “Recall” action, users can retrieve any information they’ve interacted with on their device, presented in the context of a specific time period.

Microsoft assures users that the Recall index remains local and private on-device. Users can pause, stop, or delete captured content and can choose to exclude specific apps or websites. And Recall automatically excludes InPrivate web browsing sessions in Microsoft Edge and DRM-protected content.

Currently, Recall is exclusively compatible with new “Copilot Plus PCs” equipped with Qualcomm’s Snapdragon X Elite chips, which feature the necessary neural processing unit (NPU).

Why does this matter?

The Recall feature has sparked concern among cybersecurity experts and government authorities due to its extensive data collection and privacy risks. Recall continuously captures screenshots of users’ PC activities and creates a searchable index using AI, which may accidentally expose sensitive information to unauthorized users.

Source

AI-powered headphones let you listen to one person in a crowd

Researchers at the University of Washington have developed a groundbreaking artificial intelligence system called “Target Speech Hearing” that allows headphone wearers to isolate and listen to a single speaker in a noisy environment. By simply looking at the desired speaker for three to five seconds, the user can “enroll” them, and the system will cancel out all other sounds, playing only the enrolled speaker’s voice in real-time, even as the listener moves around.

The proof-of-concept device, which uses off-the-shelf headphones fitted with microphones and an on-board embedded computer, builds upon the team’s previous “semantic hearing” research. The system’s ability to focus on the enrolled voice improves as the speaker continues talking, providing more training data. While currently limited to enrolling one speaker at a time and requiring a clear line of sight, the researchers are working to expand the system to earbuds and hearing aids in the future.

Why does this matter?

The Target Speech Hearing system may have applications in various settings, like business meetings, conferences, noisy public spaces, or for people with hearing difficulties. While the current system is a proof-of-concept, the researchers’ plan to expand it to earbuds and hearing aids indicates the potential for commercialization, which could lead to a new-age consumer product that enhances our sensory experiences in previously unimaginable ways.

Source

xAI’s supercomputer that’s 4x larger than existing GPU clusters

Elon Musk has recently disclosed plans to build a groundbreaking supercomputer to power the next iteration of xAI’s chatbot, Grok. According to a presentation made to investors in May, as reported by The Information, Musk aims to have this computing system operational by the fall of 2025. The ambitious project may involve a collaboration between xAI and Oracle to develop this massive computing infrastructure.

The proposed supercomputer will consist of interconnected clusters of Nvidia’s state-of-the-art H100 graphics processing units (GPUs). Musk revealed that upon completion, the scale of this system will surpass the largest existing GPU clusters by at least fourfold!

Why does it matter? 

As major tech giants and startups compete for dominance in the AI space, Musk’s proposed “gigafactory of compute” could potentially set new benchmarks for AI computing power. Also, Musk’s financial resources and personal commitment to the project position him as a formidable challenger to established AI powerhouses like OpenAI, Google, and Meta.

Source

New AI Jobs Opportunities on May 27th 2024

  • 🤿 Snorkel – Machine Learning Customer Engineer
  • 🐙 OctoAI – Senior MLSys Engineer
  • 🌐 OpenAI – Technical Program Manager, Trustworthy AI
  • 📊 C3 AI – Senior Revenue Accountant

What Else Is Happening in AI on May 27th 2024❗

🎵 YouTube Music introduces hum-to-search AI feature

YouTube Music has introduced a new AI feature that lets users search for songs by humming or singing a melody. This feature, similar to the “Hum to Search” functionality in Google Search, uses machine learning to match the user’s audio input to the original song recording. The feature is currently available in its Android app version 7.02 and is being gradually rolled out to users. (Link)

🤩 iOS 18 may introduce AI-powered custom emoji tool and smart recaps

Apple is reportedly working on a gen AI-powered custom emoji tool for iOS 18 that lets users create personalized emojis. They’re also developing a smart recaps feature that will provide users with summaries of missed notifications, messages, and documents. (Link)

🎼 Suno.ai’s music generator now offers 4-minute songs

AI music model Suno.ai has released its version 3.5 which lets users create songs up to four minutes long with improved structure. The update also introduces a sound-to-song feature, combining audio and text prompts to generate music. (Link)

🔍 Meta to use EU user data for AI training, offers opt-out

Meta is notifying Facebook and Instagram users in Europe about changes to its privacy policy, effective June 26, which allows the company to use user data for training AI models under a “legitimate interest” claim. Users can opt out of this data usage, but they must provide an explanation of how it personally affects them, with a reference to the GDPR being sufficient. (Link)

📱 Google launches Gemini AI assistant in Messages app

Google has started rolling out its Gemini AI assistant feature in Google Messages to help users with tasks like drafting messages, brainstorming, event planning, and engaging in conversations. Gemini supports extensions like Workspace, YouTube, and Google Maps, and also lets users provide feedback on responses. (Link)

A  Daily chronicle of AI Innovations May 24th 2024: 🍕 Google AI tells users to glue pizza and eat rocks ✌️ OpenAI scraps controversial nondisparagement agreement with employees 💰 Musk’s xAI nears deal valuing startup at $24B🌍 Cohere releases multilingual AI model, Aya 23 📱 Arc introduces “Call Arc” for quick voice answers 🤖 Elon Musk envisions AI era, new work norms, life on Mars 🤯 OpenAI controversy: ‘Sky’ speaks out

🍕 Google AI tells users to glue pizza and eat rocks

  • Google’s new AI Overviews feature is generating incorrect and sometimes absurd answers, like adding non-toxic glue to pizza as a solution to prevent cheese from falling off.
  • The feature has made several other mistakes, such as stating that former US President James Madison graduated from the University of Wisconsin 21 times and claiming that Batman is a cop.
  • Google acknowledges these are “isolated examples” and emphasizes the AI is experimental, but the errors highlight significant issues with the current state of AI-generated information.
  • Source

✌️ OpenAI scraps controversial nondisparagement agreement with employees

  • OpenAI will not enforce any previously signed nondisparagement agreements with former employees and will remove such language from its exit paperwork, as stated to Bloomberg.
  • Previously, employees had to choose between speaking against the company or keeping their vested equity, potentially losing millions if they refused to sign the agreement.
  • Sam Altman, OpenAI’s CEO, expressed embarrassment over the existence of the clause and pledged to rectify the paperwork, while Chief Strategy Officer Jason Kwon apologized for the distress caused.
  • Source

💰 Musk’s xAI nears deal valuing startup at $24B

  • Elon Musk’s AI startup, xAI, is expected to complete a funding round in June that could value the company at over $24 billion.
  • The company originally aimed to raise around $6 billion earlier this month but is now seeking to gather as much as $6.5 billion, with final deal closure still pending.
  • xAI has been marketed to Silicon Valley investors using a pitch highlighting Musk’s success with Tesla and SpaceX, and plans to leverage data from his social platform X (formerly Twitter) to build its AI chatbot, Grok.
  • Source

Cohere releases multilingual AI model, Aya 23

Cohere for AI (C4AI), the non-profit research group, has launched open-weight Aya 23, a new family of multilingual language models. Available in 8B and 35B parameter variants, Aya 23 supports 23 languages, including Arabic, Chinese, English, French, German, Hindi, Japanese, Spanish, and more.

Here’s a quick breakdown: 

  • Aya23 focuses on depth over breadth, meaning it performs better in fewer languages than their previous model, Aya 101 (which covered 101 languages).
  • The 8B parameter model balances efficiency and accessibility, while an advanced 35B parameter delivers higher performance at the cost of increased computational demand.
  • Aya 23 outperforms existing models like Google’s Gemma on various tasks across the languages it covers.
  • Researchers can access and fine-tune Aya 23 for their needs, with the model available for free trial on Cohere Playground.

Why does this matter?

Most AI models struggle with languages besides English. The development of multilingual models can serve a much wider audience, bridging the language divide and making AI more accessible and inclusive for users around the world.

Source

Arc introduces “Call Arc” for quick voice answers

Arc Search, an AI-powered search app, just launched a new feature called Call Arc. This lets users ask questions by holding their phone to their ear, mimicking a phone call. It provides instant voice answers, similar to voice search, but designed to be more convenient and quicker.

The app is designed to answer short, immediate questions. For example, you can ask it how long it takes to cook spaghetti or why to reserve pasta water, all while making dinner.

Call Arc complements Arc Search’s existing “Browse for me” function that generates webpages with information based on your search query.

Why does this matter?

As AI continues to simplify our lives, innovative features like Call Arc could provide a new approach to voice search, making it intuitive and accessible for users on the go. This could also lead to more nuanced search results and a more engaging experience compared to simple keyword searches.

Source

Elon Musk envisions AI era, new work norms, life on Mars

In a Q&A session at VivaTech 2024, Elon Musk discussed diverse topics, from plans for Mars colonization to the role of AI in society.

Source: Viva Technology

Musk emphasized SpaceX’s goal of making life multi-planetary, with Mars as a key focus. He discussed the importance of reusable spacecraft and highlighted the necessity of space exploration for humanity’s long-term survival.

Regarding AI, Musk stressed the importance of honesty in AI development, criticizing approaches that prioritize political correctness over truthfulness. He also touched on AI’s potential to revolutionize education, though he expressed concerns about the impact of social media on children.

Musk envisioned a future where automation leads to a job-free society, with a universal basic income ensuring people’s needs are met.

Why does this matter?

Elon Musk’s bold predictions about Mars, the Moon, and a job-free future driven by AI are both exciting and thought-provoking. While his vision for Mars could redefine human civilization, his stance on AI emphasizes the need for integrity in technology.

It’s crucial that we engage in open and honest dialogue about AI’s implications for society.

Source

🤯 OpenAI controversy: ‘Sky’ speaks out

There has been a wave of updates on several drama-filled OpenAI topics, including leaked documents related to the company’s controversial NDA clauses, new info surrounding the ‘Sky’ voice model, another resignation, and more.

  • The voice actress behind ChatGPT’s ‘Sky’ voice has come forward, saying she was hired months before Altman contacted Scarlett Johansson.
  • Leaked documents with Sam Altman’s signature contradict claims he was unaware of NDA clauses threatening to claw back former employees’ equity.
  • OpenAI policy researcher Gretchen Kreuger resigned, with a post citing safety concerns about the company.
  • Teasers are coming out of presentations at the VivaTech conference, showing a potential move away from the ‘GPT-5’ name and a new Sora demo.

New AI Jobs Opportunities on May 24th 2024:

  • 🌐 Kumo – ML Solutions Architect
  • 📝 Palantir – Proposal Writer, Internship
  • 💻 Lambda – Linux Support Engineer
  • 🏢 Anthropic – Enterprise Account Executive

What Else Is Happening in AI on May 24th 2024❗

🎵 YouTube launches AI-powered music creation tool

YouTube has introduced Dream Track, an AI-powered tool designed for creating music within YouTube Shorts. This experimental feature allows users to generate musical pieces based on input commands. While it is an innovative addition, it has sparked a debate over AI’s impact on the authenticity of creative work.  (Link)

🎨 Ideogram launches Ideogram Tile, an AI-powered pattern generator

Ideogram has launched Ideogram Tile, a feature that creates seamless, repeating patterns from text prompts. This tool empowers designers to craft unique wallpapers, fabrics, and textures effortlessly. Following a $80 million funding round, Ideogram continues to lead in AI-driven design innovation, blending creativity with advanced technology. (Link)

🎧 Spotify tests Spanish-speaking AI DJ, “DJ Livi”

Spotify is testing a Spanish-speaking iteration of its AI DJ, named “DJ Livi.” The app code suggests a language switch option for this new feature. Although Spotify has not confirmed launch plans, this move could cater to Spanish-speaking audiences globally and in the U.S., marking a significant expansion in AI language support. (Link)

🗣️ Amazon plans paid subscriptions for Alexa AI features

Amazon is reportedly gearing up to launch paid subscription plans for select Alexa AI features, aiming to enhance conversational capabilities. This move seeks to compete with advanced AI chatbots like Google’s Gemini and OpenAI’s ChatGPT. Notably, the subscription won’t be bundled with Amazon Prime, offering a standalone service. (Link)

🎮 Inworld AI launches AI-driven voice generator

Inworld AI, known for its innovative AI-driven virtual characters, unveils Inworld Voice, a dynamic voice generator for game developers. Boasting 58 diverse voices powered by advanced machine learning models, it promises expressive, cost-efficient, and high-quality voice experiences. (Link)

A  Daily chronicle of AI Innovations May 23rd 2024: 🔍 Anthropic uncovers millions of concepts in Claude Sonnet’s AI model 📞 Truecaller’s AI assistant gets a voice upgrade, thanks to Microsoft 🎥 TikTok makes ad creation easy with AI! 🤯 Scientists plan a head transplant system 🦎 Chameleon: Meta’s new multimodal LLM 🤔 OpenAI didn’t copy Scarlett Johansson’s voice, records show 📈 Nvidia shows no signs of AI slowdown 💥 Global outages hit Microsoft Bing, DuckDuckGo, ChatGPT Search and Copilot 💰 OpenAI signs $250M landmark deal with News Corp ⛑️ Apple may develop foldable iPhone with self-healing screen 

Anthropic uncovers millions of concepts in Claude Sonnet’s AI model

Anthropic has made a breakthrough in understanding the inner workings of their AI model, Claude Sonnet, by identifying how millions of concepts are represented within it.

Using a technique called “dictionary learning,” they were able to map out these concepts, providing the first-ever detailed look inside a modern, production-grade large language model.

Key findings:

  • Features linked to concepts: These concepts are linked to features, which are groups of neurons that activate together in response to specific ideas.
  • Features can be manipulated: By manipulating these features, the researchers were able to influence Claude’s outputs, demonstrating a causal link between features and behavior.
  • Features reveal potential risks: The research identified features corresponding to biases, potential misuse of the model, and even sycophantic behavior.

Why does this matter?

Anthropic’s research is a big step toward making AI models more transparent and trustworthy. By understanding how these models work, researchers can spot and reduce harmful biases, prevent misuse, and steer AI systems toward better outcomes, leading to safer and more reliable AI applications.

Source

Truecaller’s AI assistant gets a voice upgrade, thanks to Microsoft

Truecaller is partnering with Microsoft to let users create an AI version of their own voice for their AI Assistant to use when answering calls.

This feature, currently limited to users with access to Truecaller’s AI Assistant, requires them to record a voice clip that Microsoft’s Azure AI Speech technology will use to create a personalized AI voice.

Truecaller says this will allow for a more personalized experience and highlights the potential of AI in communication. However, it’s important to note that Microsoft limits the use of personal voice for specific purposes and requires users to obtain consent before recording someone’s voice.

Why does this matter?

Truecaller’s AI voice assistant offers a glimpse into the future of call management, promising a more personalized experience. However, it must ensure transparency and address performance comparisons with automated voicemail systems.

The idea of speaking with an AI while calling someone may seem unusual, and it remains to be seen how well this feature will be received.

Source

TikTok makes ad creation easy with gen AI!

TikTok has introduced “TikTok Symphony,” a suite of generative AI tools designed to help marketers create and optimize ad campaigns. The suite includes an AI video generator called “Symphony Creative Studio,” which can produce TikTok-ready videos with minimal input from advertisers, and an AI assistant named “Symphony Assistant” that helps refine scripts and provides best practice recommendations.

TikTok makes ad creation easy with gen AI!
TikTok makes ad creation easy with gen AI!

The company has also introduced “TikTok One,” a centralized hub for marketers to access creators, agency partners, and creative tools. Additionally, TikTok is leveraging predictive AI to drive more sales for advertisers by determining the best creative assets and target audiences based on budgets and goals.

Why does this matter?

TikTok’s integration of generative AI into its ads business signifies the growing importance of AI in digital marketing. This shift could make ad campaigns more efficient and effective, opening new opportunities for tech companies and startups.

Source

 Chameleon: Meta’s new multimodal LLM

Chameleon: Meta’s new multimodal LLM
Chameleon: Meta’s new multimodal LLM

Meta’s AI research lab just introduced Chameleon, a new family of ‘early-fusion token-based’ AI models that can both understand and generate text and images in any order.

  • Unlike other models that process image and text separately and then combine them later, Chameleon works with linked sequences of both.
  • Chameleon outperformed all rival models on image captioning and visual questions, while still matching performance on text-only.
  • The 34B parameter Chameleon also matched or outperformed top models like Gemini Pro and GPT-4V on tests of long-form mixed-modal generations.

Chameleon shows the potential for a different type of architecture for multimodal AI models, with its early-fusion approach enabling more seamless reasoning and generation across modalities and setting new performance bars.

New AI Jobs Opportunity on May 23rd 2024

  • 🎯 Impel AI – Test Automation Engineer
  • 📚 Grammarly – Researcher, Strategic Research
  • ⛅ Tempus – Senior Cloud Security Engineer
  • 📋 UiPath – Senior Product Manager

What Else Is Happening in AI on May 23rd 2024❗

🤝 OpenAI partners with News Corp to enhance ChatGPT 

OpenAI has announced a multi-year partnership with media giant News Corp to enhance ChatGPT with its high-quality journalism. The collaboration aims to provide reliable information to users, with News Corp’s mastheads displayed in response to user queries. OpenAI CEO Sam Altman called it a “proud moment for journalism and technology.” (Link)

💻 AWS and Hugging Face team up on AI model development

AWS has partnered with AI startup Hugging Face to enable developers to run thousands of AI models on Amazon’s custom Inferentia2 chips. The collaboration aims to provide a cost-effective solution for deploying AI models, with AWS aiming to compete in the inference market currently dominated by Nvidia. (Link)

📝 Granola’s introduces AI-powered note-taking app

Granola, a new AI startup, has launched a notepad app that combines user-written notes with AI-generated details from meeting transcripts. The app aims to make note-taking more collaborative and efficient, using OpenAI’s GPT-4 to augment and clean up your notes. (Link)

🧠 BrainBridge launches head transplant system using AI and robotics

Neuroscience startup BrainBridge has revealed its groundbreaking head transplant system that aims to provide hope for patients with untreatable conditions.

https://x.com/i/status/1793045981954523464

  • The CGI demo proposes a system that uses advanced robotics, AI, and real-time molecular imaging to transfer a patient’s head or face onto a donor body.
  • The company hopes to conduct the first surgery within eight years, and is releasing the concept early to attract top scientists.
  • A headband with a brain-computer interface (like Neuralink) would help patients communicate and execute tasks during recovery using their minds.

While the mind-blowing (🥁) system seems highly speculative and more sci-fi than reality — so did Neuralink’s brain chips. While we’re likely a long way from this becoming possible, with the rate of advancement in robotics and AI it’s tough to rule anything out over the next decade.

The procedure uses advanced AI, robotics, and a proprietary adhesive to transplant a patient’s head onto a donor body. While it’s still early, BrainBridge expects to conduct the first surgery within eight years, pending successful feasibility studies. (Link)

👨‍💻 Meta establishes AI advisory group

Meta CEO Mark Zuckerberg has created the Meta Advisory Group, consisting of Stripe’s Patrick Collison, Shopify’s Tobi Lütke, and others, to guide the company’s AI and technology advancements. The group will offer insights and recommendations as Meta pushes harder into AI development across various products. (Link)

A  Daily chronicle of AI Innovations May 22nd 2024: 🧠 Microsoft’s first SoTA SLM to be shipped with Windows
📈 Google unveils new AI tools for branding and product marketing 🎨 Adobe introduces Firefly AI-powered Generative Remove to Lightroom

Microsoft’s first SoTA SLM to be shipped with Windows

Microsoft announced a new small language model called Phi Silica. It has 3.3 billion parameters, which makes it the smallest model in Microsoft’s Phi family of models. Phi Silica is designed specifically for the Neural Processing Units (NPUs) in Microsoft’s new Copilot+ PCs. Despite its small size, Phi Silica can generate 650 tokens per second using only 1.5 Watts of power. This allows the PC’s main processors to be free for other tasks.

Microsoft's first SoTA SLM to be shipped with Windows
Microsoft’s first SoTA SLM to be shipped with Windows

Developers can access Phi Silica through the Windows App SDK and other AI-powered features like OCR, Studio Effects, Live Captions, and Recall User Activity APIs. Microsoft plans to release additional APIs, including Vector Embedding, RAG API, and Text Summarization. These AI-powered PCs will have dedicated AI chips for running LLMs and other AI workloads.

Why does it matter?

As Microsoft continues to invest in developing small language models and integrating AI into its Windows platform, Phi Silica represents a significant step forward in making advanced AI capabilities more accessible to developers and end-users. Also, as major PC manufacturers want to introduce AI-powered laptops this summer, Microsoft might lead the way by introducing CoPilot PC+ and now Phi Silica.

Source

Google unveils new AI tools for branding and product marketing

Google has introduced several new AI-powered features to help retailers and brands better connect with shoppers. First, Google has created a new visual brand profile that will appear in Google Search results. This profile uses information from Google Merchant Center and Google’s Shopping Graph to showcase a brand’s identity, products, and offerings.

Google unveils new AI tools for branding and product marketing
Google unveils new AI tools for branding and product marketing

Additionally, Google is expanding its AI-powered tools to help brands create more engaging content and ads. This includes new features in Google’s Product Studio, allowing brands to generate images matching their unique style.

Google unveils new AI tools for branding and product marketing
Google unveils new AI tools for branding and product marketing

Google is also launching immersive ad formats powered by generative AI, such as the ability to include short product videos, virtual try-on experiences, and 3D product views directly in search ads. These new AI-driven tools aim to help brands forge stronger, more personalized connections with consumers throughout the shopping journey.

Google unveils new AI tools for branding and product marketing
Google unveils new AI tools for branding and product marketing

Why does it matter?

As AI continues advancing, such tools could shape the future of marketing by delivering more personalized, hyper-relevant, and visually compelling experiences that better connect consumers with brands and products.

Source

Adobe introduces Firefly AI-powered Generative Remove to Lightroom

Adobe has added a new AI-powered feature called Generative Remove to its Lightroom photo editing software. Generative Remove uses Adobe’s Firefly generative AI model to allow users to seamlessly remove objects from photos, even if the objects have complex backgrounds. The feature can remove images’ stains, wrinkles, reflections, and more.

Adobe introduces Firefly AI-powered Generative Remove to Lightroom
Adobe introduces Firefly AI-powered Generative Remove to Lightroom

Adobe has been integrating Firefly’s capabilities across its Creative Cloud apps to generate images, apply styles, fill areas, and remove objects through the new Generative Remove tool in Lightroom. It works closely with photographers to continue improving and expanding this object-removal capability. The company also announced a new Lens Blur effect that uses AI to add realistic depth-of-field blur to photos.

Why does it matter?

The Generative Remove feature will make it easier for photographers, designers, and other creatives to edit their images, saving time and effort. Looking ahead, we can expect Adobe and other creative software companies to explore new ways to harness Gen AI to automate tedious tasks, provide intelligent assistance, and enable entirely new creative possibilities.

Source

AI RESEARCH:

🗺️ Research: Mapping the ‘mind’ of an LLM

Research: Mapping the ‘mind’ of an LLM
Research: Mapping the ‘mind’ of an LLM

Anthropic just published new research that successfully identified and mapped millions of human-interpretable concepts, called “features”, within the neural networks of Claude.

The details:
  • Researchers used a technique called ‘dictionary learning’ to isolate patterns that corresponded to concepts, from objects to abstract ideas.
  • By tweaking the patterns, the researchers showed the ability to change Claude’s outputs, potentially leading to more controllable systems.
  • The team mapped concepts related to AI safety concerns, like deception and power-seeking — providing glimpses into how models understand these issues.

Why it matters: Despite how fast AI is accelerating, we still don’t have a strong understanding of what’s going on beneath the hood of LLMs. This research is a major step towards making AI more transparent — enabling better understanding, control, and safeguarding of these powerful tools.

New AI Job Opportunities on May 22nd 2024

What Else Is Happening in AI on May 22nd 2024❗

🤖 Elon Musk’s xAI plans to make Grok multimodal   

According to public developer documents, Elon Musk’s AI company, xAI, is making progress on adding multimodal inputs to the Grok chatbot. This means soon, users may be able to upload photos to Grok and receive text-based answers. This was first teased in a blog post last month from xAI, which said Grok-1.5V will offer “multimodal models in a number of domains.” (Link)

👨‍💻 Microsoft’s new Copilot AI agents to work like virtual employees 

Microsoft will soon allow businesses and developers to build AI-powered Copilots that can work like virtual employees and perform tasks automatically. Instead of Copilot sitting idle waiting for queries, it will be able to monitor email inboxes and automate tasks or data entry that employees normally have to do manually. (Link)

🌍 Microsoft Edge introduces real-time AI translation and dubbing for YouTube

Microsoft Edge is set to introduce real-time translation and dubbing for videos on platforms like YouTube, LinkedIn, and Coursera. This new AI-powered feature will translate spoken content live, offering dubbing and subtitles. Currently, the feature supports translations from Spanish to English and English to German, Hindi, Italian, Russian, and Spanish. (Link)

🛡️ WitnessAI builds guardrails for Gen AI models

WitnessAI is developing tools to make Gen AI models safer for businesses. The company’s platform monitors employee interactions and custom AI models, applying policies to reduce risks like data leaks and biased outputs. The platform also offers modules to enforce usage rules and protect sensitive information. Lastly, it encrypts and isolates data for each customer. (Link)

💻 Microsoft’s Azure AI Studio supports GPT-4o

Microsoft has announced that Azure AI Studio is now generally available and supports OpenAI’s GPT-4o model, which joins over 1,600 other models, including Mistral, Meta, Nvidia, etc. Developers can use this multimodal foundational model to incorporate text, image, and audio processing into their apps to provide generative and conversational AI experiences. (Link)

  Daily chronicle of AI Innovations May 21st 2024: 💥 Microsoft announces AI-powered MacBook competitors 🧠 Microsoft’s AI chatbot will ‘recall’ everything you do on a PC 🫠 Scarlett Johansson told OpenAI not to use her voice 🔧 TSMC and ASML machines are equipped with remote self-destruct in case of invasion 🚚 Volvo introduces first autonomous truck 💻 Microsoft’s New AI PCs Rival Apple’s MacBooks ⚖️ Scarlett Johansson sues OpenAI for using her voice in ChatGPT 🧠 DINO 1.5 is smarter and faster at object detection

💥 Microsoft announces AI-powered MacBook competitors

  • Microsoft introduces “Copilot Plus PCs,” new Windows laptops with built-in AI hardware and features, launching June 18th with Qualcomm processors.
  • These AI-enabled laptops, from major partners like Dell and Lenovo, will include a neural processor, at least 16GB RAM, and 256GB SSD, supporting over 40 AI models and OpenAI’s GPT-4o model.
  • Promising significant performance and battery life improvements, Microsoft expects to sell 50 million Copilot Plus PCs in the next year, marking a new era for Windows laptops.
  • Source

Microsoft’s New AI PCs Rival Apple’s MacBooks

Microsoft revealed Copilot+ PCs, a new category of Windows PCs designed for AI. These PCs boast powerful processors, all-day battery life, and AI features like Recall for instant memory, Cocreator for image creation, Live Captions for real-time translations, and Auto Super Resolution for games.

Copilot plus PC main art
Microsoft’s New AI PCs Rival Apple’s MacBooks

The recall feature, which allows users to search and recall anything they’ve seen and interacted with on their computer screens with natural language, is especially impressive. The new PCs feature an all-new system architecture with CPU, GPU, and a high-performance Neural Processing Unit (NPU) working together. Starting at $999, Copilot+ PCs are equipped with OpenAI’s GPT-4o models.

Why does this matter?

Microsoft claims its new Arm-powered Copilot Plus PCs will outperform the MacBook Air with M3 by over 50% on sustained performance. The tech giant is advancing Arm-based computing by using AI at every level and partnering with Qualcomm. If these machines live up to the hype, they could upset the dominance of Intel-based laptops and challenge Apple’s M-series processors.

Source

🔧 TSMC and ASML machines are equipped with remote self-destruct in case of invasion

  • ASML and TSMC have the capability to disable their chipmaking machines if China invades Taiwan, according to sources.
  • ASML assured officials it can remotely shut down these machines, including its EUV units, which are crucial for producing the world’s smallest microchip transistors used in both AI and military applications.
  • US government officials, concerned about the potential impact of a Chinese invasion on the global chip supply, have discussed these security measures with both Dutch and Taiwanese counterparts.
  • Source

🧠 Microsoft’s AI chatbot will ‘recall’ everything you do on a PC

  • Microsoft introduced an upgraded version of its AI assistant, Copilot, that will remember users’ activities on their PCs to predict their next actions.
  • The new AI features, including Windows Recall, aim to enhance user experience by providing a virtual “photographic memory,” while allowing users to control their privacy settings.
  • It’s a step toward machines that “instantly see us, hear, reason about our intent and our surroundings,” said CEO Satya Nadella.
  • Source

🫠 Scarlett Johansson told OpenAI not to use her voice

  • Scarlett Johansson stated that OpenAI approached her in September about supplying her voice for GPT-4o, but she declined the offer.
  • After hearing a demo of the ChatGPT system’s “Sky” voice resembling her own, Johansson expressed shock and mentioned hiring legal counsel to address the issue with OpenAI.
  • OpenAI, while denying that Sky’s voice was meant to mimic Johansson’s, paused the use of the voice out of respect and announced plans to introduce additional voices in the future.
  • Source

Scarlett Johansson sues OpenAI for using her voice in ChatGPT 

Scarlett Johansson claims OpenAI asked her to voice ChatGPT, but she declined. Later, OpenAI released a voice named “Sky” that sounded eerily similar to her. Johansson was shocked and angered by the similarity and has hired legal counsel to investigate how the “Sky” voice was created.

OpenAI denies that the “Sky” voice was intended to resemble Johansson, has paused using it in its products, and apologizes for not communicating better. Johansson seeks transparency from OpenAI and believes that individual rights must be protected in the era of deep fakes and AI content.

Why does this matter?

Another day, another OpenAI drama. OpenAI already faces lawsuits from authors and news organizations over copyright infringement in training its AI models. Johansson’s case shows the limited legal protection actors have against AI mimicking their likeness, and some lawmakers are proposing bills to address the issue.

Source

DINO 1.5 is smarter and faster at object detection

IDEA Research launched the Grounding DINO 1.5 open-world object detection model series, with Grounding DINO 1.5 Pro for high-performance detection and Grounding DINO 1.5 Edge for efficient edge computing. Grounding DINO 1.5 Pro achieves state-of-the-art zero-shot transfer performance on several academic benchmarks, surpassing its predecessor.

DINO 1.5 is smarter and faster at object detection
DINO 1.5 is smarter and faster at object detection

The model shows strong detection capabilities across various scenarios, including common objects, long-tailed categories, dense objects, and caption phrase grounding. Grounding DINO 1.5 Pro uses a larger Vision Transformer backbone and is pretrained on the high-quality Grounding-20M dataset.

Why does this matter?

Grounding DINO 1.5 isn’t just about raw performance—it’s also incredibly versatile. Fine-tuning the model on specific datasets leads to even more impressive results, suggesting that it can lead us to make more reliable robotics, automated systems, and more while requiring less training time and data.

Source

🚚 Volvo introduces first autonomous truck

  • Volvo has unveiled its first production-ready self-driving truck, developed in partnership with the autonomous driving technology company Aurora.
  • This autonomous semi truck, based on Volvo’s VNL Class 8 model, is equipped with an array of sensors and cameras to support Aurora’s Level 4 autonomous driving system, which does not require a human driver.
  • Volvo and Aurora started collaborating in 2018, and their autonomous trucks have already logged 1.5 million miles on commercial roads; Aurora aims to deploy 20 fully autonomous trucks this year and plans to scale up to 100 trucks by 2025.
  • Source

What Else Is Happening in AI on May 21st 2024❗

💻 HP is introducing simplified laptop “AI PCs” lineup 

HP’s new “AI PCs,” the OmniBook X AI and EliteBook Ultra AI, feature Snapdragon X Elite 12-core CPUs. HP’s AI Companion software provides access to ChatGPT-3.5, while Poly Camera Pro offers AI camera controls. Moreover, the OmniBook X AI starts at $1,199.99 and the EliteBook Ultra AI at $1,699.99. (Link)

💶 Google is investing 1 billion Euros to expand its data center in Finland

This move aims to drive AI business growth in Europe. Nordic countries are popular for data centers due to cooler climate, tax breaks, and abundant renewable power. Moreover, the heat from this data center will be used to warm up local homes and buildings. Google’s Hamina data center in Finland already operates with 97% carbon-free energy and aims for net zero emissions by 2030. (Link)

🏭 Dell Technologies expands its generative AI capabilities with the Dell AI Factory

The Dell AI Factory with NVIDIA integrates hardware and software to support advanced use cases like RAG and digital assistants. Dell is also deepening partnership with Hugging Face, Meta, and Microsoft to bring open models and Azure AI services to on-premises deployments. The goal is to make it easy for enterprises to assess and implement gen AI on-premises, close to their data. (Link)

🤖 Intel’s Lunar Lake laptop processors are set to launch in Q3 2024 

These processors are designed for AI Copilot Plus PCs.  They offer 3x the AI performance of its predecessor, Meteor Lake. Intel claims Lunar Lake is 1.4 times faster in Stable Diffusion 1.5 compared to Qualcomm’s Snapdragon X Elite. The company aims to ship 40 million AI PC processors by the end of the year, with Lunar Lake featured in over 80 new laptop designs. (Link)

📜 OpenAI CEO Sam Altman apologized for the company’s strict off-boarding agreement 

This off-boarding agreement could potentially cancel departing employees’ vested equity. Meanwhile, departing execs are speaking out about the need for serious AGI safety measures. Altman and co-founder Greg Brockman defended OpenAI’s launch procedures, emphasizing their commitment to safety. (Link)

A  Daily chronicle of AI Innovations May 20th 2024:🫠 Google’s AI panic looks like Google+ fiasco 🎤 OpenAI pauses Scarlett Johansson-like voice for ChatGPT 💸 Snapchat focuses on AI with $1.5 billion yearly investment 🍏 Apple and OpenAI plan major announcement at WWDC 🤖 OpenAI’s “superalignment team,” focused on the AI risks, is no more
🚫 Sony Music warns over 700 AI companies not to steal its content 🦎 Meta’s Chameleon AI sets a new bar in mixed-modal reasoning

🫠 Google’s AI panic looks like Google+ fiasco

  • Scott Jenson, a former Google employee, criticizes the company’s AI projects as poorly motivated and driven by panic, comparing the situation to the Google+ fiasco.
  • According to Jenson, Google aims to create a Jarvis-like assistant to keep users within its ecosystem, driven by fear that competitors might get there first, similar to Apple’s strategy with Siri and OpenAI.
  • Many of Google’s AI projects revealed at I/O 2024 are still experimental or limited in availability, with uncertain value, and features like “AI Overviews” in Google Search have received criticism for potential copyright issues and misinformation.
  • Source

🎤 OpenAI pauses Scarlett Johansson-like voice for ChatGPT

  • OpenAI is pulling the ChatGPT voice known as Sky, which sounds similar to Scarlett Johansson, due to concerns about mimicking celebrities’ voices.
  • The company asserts that Sky’s voice is the natural voice of a different professional actress and not an intentional imitation of Johansson.
  • The decision to pause Sky’s use follows recent enhancements to ChatGPT’s voice mode, part of the new GPT-4o model, which aims to make the assistant more expressive and capable of reading facial expressions and translating spoken language in real-time.
  • Source

💸 Snapchat focuses on AI with $1.5 billion yearly investment

  • Snap’s CEO, Evan Spiegel, announced a major investment of $1.5 billion annually in AI and machine learning to improve Snapchat’s features and competitiveness.
  • After successfully revamping Snapchat’s advertising model, the company will now focus on collaborating with tech giants like Amazon and Google for cloud partnerships to enhance AI product innovations.
  • With growing ad revenue and new successful ad campaigns, Snapchat plans to expand content offerings and augmented reality experiences, enhancing user engagement and competing with platforms like TikTok.
  • Source

🍏 Apple and OpenAI plan major announcement at WWDC

Apple and OpenAI plan major announcement at WWDC
Apple and OpenAI plan major announcement at WWDC
  • Apple and OpenAI are planning a major joint announcement at WWDC, focusing on integrating OpenAI’s technology into iOS 18, according to Mark Gurman from Bloomberg.
  • Apple aims to improve its AI capabilities with this integration, including enhancements to Siri and features like automatic summaries of notifications and transcription of voice memos.
  • While Apple acknowledges it is behind in AI and lacks a chatbot like ChatGPT or Google Gemini, the company believes the collaboration with OpenAI is a sufficient, though temporary, solution.
  • Source

OpenAI’s “superalignment team,” focused on the AI risks, is no more

The team’s co-leads, Ilya Sutskever and Jan Leike, have resigned from OpenAI. Several other researchers from the team and those working on AI policy and governance have also left the company. Leike cited disagreements with OpenAI’s leadership about the company’s priorities and resource allocation as reasons for his departure.

image

(Source)

The team’s work will be absorbed into OpenAI’s other research efforts, with John Schulman leading research on risks associated with more powerful models.

Why does this matter?

The “superalignment” team was for ensuring the artificial general intelligence (AGI) that OpenAI claims to be working on doesn’t turn on humankind. This dismantling raises questions on the company’s commitment to AI safety and ethical standards.

Source

Sony Music warns over 700 AI companies not to steal its content

Sony Music, home to superstars like Billy Joel and Doja Cat, sent letters to over 700 AI companies and streaming platforms, warning them against using its content without permission. The label called out the “training, development, or commercialization of AI systems” that use copyrighted material, including music, art, and lyrics.

SMG recognizes AI’s potential but stresses the need to respect songwriters’ and artists’ rights. The letter asks companies to confirm they haven’t used SMG content without permission or provide details if they have.

Why does this matter?

The battle over music copyright and AI has intensified across various platforms, from YouTube’s strict rules for AI-generated music to the recent standoff between Universal Music Group and TikTok. As AI voice clones and music generation tools become more sophisticated, artists question control, compensation, and actions against copyright infringement.

Source

Meta’s Chameleon AI sets a new bar in mixed-modal reasoning

Meta AI introduces Chameleon, a family of early-fusion token-based mixed-modal models that understands and generates images and text in any order. Unlike recent foundation models that process text and images separately, Chameleon unified token space allows it to process interleaved image and text sequences.This approach allows seamless reasoning and generation across modalities.

Meta's Chameleon AI sets a new bar in mixed-modal reasoning
Meta’s Chameleon AI sets a new bar in mixed-modal reasoning

Meta researchers introduced architectural enhancements and training techniques to tackle the optimization challenges posed by this early fusion approach, including a novel image tokenizer, QK-Norm, dropout, and z-loss regularization. Remarkably, Chameleon achieves competitive or superior performance across various tasks, outperforming larger models like Flamingo-80B and IDEFICS-80B in image captioning and visual question answering despite its smaller model size.

Why does this matter?

Chameleon opens up new possibilities for more natural and intuitive human-machine interactions, similar to how we effortlessly communicate using both modalities in the real world.

Source

Trending AI Tools May 20th 2024:

  • ✍️ Paperpal – AI writing assistant offering grammar checks, paraphrasing, predictive text, and plagiarism detection. Use code RUN30 for 30% off*
  • 🗣️ ElevenLabs Audio Native – Add human-like narration to your blog or news site
  • 💻 Framer – Generate a responsive website with no code
  • 🚀 Glitter – Turn any process into a step-by-step guide
  • ⚡TestSprite Beta – Automate end-to-end software testing with AI
  • 🤖 Buffup – AI assistant that learns your intent, powered by GPT-4o

New AI Job Opportunities on May 20th 2024:

What Else Is Happening in AI on May 20th 2024❗

🤖 Google launched open-source Model Explore to visualize and debug complex AI models

It uses advanced graphics rendering techniques from the gaming industry to handle massive models. The tool offers a graphical user interface and a Python API for integration into machine learning workflows. Model Explorer lets developers identify and resolve issues quickly, especially for AI deployed on edge devices. (Link)

🇬🇧 The UK’s AI Safety Institute is opening an office in San Francisco

The institute aims to be closer to the epicenter of AI development, companies like OpenAI and Google as they are building foundational models. This new office would open this summer, giving the UK access to Silicon Valley’s tech talent and strengthening ties with the US. (Link)

📂 The EU demands Microsoft to provide internal documents on Bing’s gen AI risks

The Commission suspects Bing may have breached the Digital Services Act (DSA) due to risks like AI “hallucinations,” deep fakes, and potential voter manipulation. Microsoft has until May 27 to comply with the legally binding request for information. Failure to do so could result in fines of up to 1% of Microsoft’s total annual income or worldwide turnover. (Link)

📸 Snapchat CEO Evan Spiegel focuses on AI and ML for better UX and personalization

As its ad revenue increases, Snap plans to expand content offerings, improve recommendation algorithms, and integrate Stories with Spotlight. The company is also investing in augmented reality and sees it as a way to bring people together in shared physical environments. (Link)

😏 Researchers in the Netherlands have developed an AI sarcasm detector

The AI was trained on text, audio, and emotional content from US sitcoms, including Friends and The Big Bang Theory. The AI could detect sarcasm in unlabeled exchanges nearly 75% of the time. Further improvements could come from adding visual cues to the AI’s training data. (Link)

AI Weekly Rundown May 11-May 18th 2024: Major Breaking News from OpenAI ChatGPT-4o, Google AI Veo, Apple AI-enhanced eye tracking, Meta AI-assisted earphones , Stability AI, Humane AI

Listen here

🚫 Sam Altman quashes search engine launch rumors
📱 iOS 18 iPhones may soon have ChatGPT
💻 SoftBank’s Arm Holdings to come up with AI chips in 2025
🚀 GPT-4o: 2x faster, 50% cheaper, 5x rate limits than GPT-4 Turbo
🔍 TikTok tests ChatGPT-powered search results
🎧 Meta explores AI-assisted earphones with built-in cameras
✨ Google announced a wide array of updates across its AI ecosystem
🧠 Ilya Sutskever, OpenAI’s co-founder and chief scientist, has left the company
🍎 Apple unveils AI-enhanced eye tracking, music haptics, voice shortcuts
🤖 Humane’s AI Pin upgraded with GPT-4o
💸 Stability AI seeks buyer amid financial woes
🔄 OpenAI will bring Reddit content to ChatGPT and new products
📈 OpenAI introduces improvements to data analysis in ChatGPT
👩‍💻 GenZ and millennials optimistic about GenAI use at work: Deloitte Survey

A Daily chronicle of AI Innovations May 17th 2024: 📈 OpenAI introduces improvements to data analysis in ChatGPT 👩‍💻 GenZ and millennials optimistic about GenAI use at work: Deloitte Survey 🤝 OpenAI strikes Reddit deal to train its AI on your posts  🎵 Sony Music warns 700+ tech companies over ‘unauthorized’ use of its content to train AI 👀 The first search engine has been resurrected💸 Deepfake scam costs firm behind Sydney Opera $25 million

OpenAI introduces improvements to data analysis in ChatGPT

OpenAI introduces improvements to data analysis in ChatGPT
OpenAI introduces improvements to data analysis in ChatGPT

OpenAI is rolling out enhancements to data analysis:

  • Upload the latest file versions directly from Google Drive and Microsoft OneDrive. This allows ChatGPT to understand your Google Sheets, Docs, Slides, and Microsoft Excel, Word, and PowerPoint files more quickly
  • Interact with tables and charts in a new expandable view
  • Customize and interact with bar, line, pie, and scatter plot charts in the conversation.
  • Work on tables in real-time

These improvements will be available in OpenAI’s new flagship model, GPT-4o, for ChatGPT Plus, Team, and Enterprise users over the coming weeks.

Why does this matter?

This are exciting updates for those who use ChatGPT to work on documents for data analysis. It should save users time, give them more options and ease when using the chatbot, and make data-driven writing easier.

Source

GenZ and millennials optimistic about GenAI use at work: Deloitte Survey

Deloitte’s 2024 Gen Z and Millennial Survey connected with more than 22,800 respondents in 44 countries to explore their attitudes about work and the world around them. One of the key findings was that Gen Zs and millennials are feeling uncertain about GenAI and its potential impact on their careers.

However, respondents who frequently use GenAI at work are more likely to say they feel excitement about, and trust in the technology. Frequent users of GenAI are also more likely to believe it will free up their time, improve the way they work, and improve their work/life balance.

GenZ and millennials optimistic about GenAI use at work: Deloitte Survey
GenZ and millennials optimistic about GenAI use at work: Deloitte Survey

But, conversely, the more a respondent uses GenAI the more likely they are to have some concerns as well.

Why does this matter?

Gen AI is about to bring a huge transformation in the way we work and live. Despite concerns regarding it, GenZs and millennials are starting to think about how to adapt, whether their employers do or do not.

Source

🤝 OpenAI strikes Reddit deal to train its AI on your posts

  • OpenAI has signed a deal with Reddit to access real-time content from Reddit’s data API, similar to Reddit’s earlier $60 million deal with Google.
  • The partnership will allow Reddit to implement new AI-powered features and applications while also making OpenAI an advertising partner on the platform.
  • Details on financial terms or training data were not disclosed, and Reddit users have previously shown resistance to such changes, protesting API pricing alterations in June 2023.
  • Source

🎵 Sony Music warns 700+ tech companies over ‘unauthorized’ use of its content to train AI

  • Sony Music has sent letters to over 700 AI companies and streaming platforms, warning them against the unauthorized use of its content for AI systems, which it claims denies artists and the label control and compensation.
  • The music industry has been notably proactive in controlling the usage of its copyrighted material in AI tools, with platforms like YouTube enforcing strict rules for music content while introducing AI tools like Dream Track.
  • Universal Music Group similarly faced a dispute with TikTok over licensing, leading to a temporary removal of music by major artists like Taylor Swift and Ariana Grande until a new deal was brokered to protect artists’ rights and explore monetization opportunities involving AI.
  • Source

💸 Deepfake scam costs firm behind Sydney Opera $25 million

  • A British design firm, Arup, which is known for its work on the Sydney Opera House, lost $25.6 million in a deepfake scam involving a fake CFO and other staff members.
  • An employee in Arup’s Hong Kong office fell victim to the fraud after being convinced by realistic deepfake videos to transfer the money over 15 transactions, realizing the scam only after cross-checking with the UK office.
  • Despite the significant financial loss, Arup stated that their financial stability and business operations were unaffected, and the case is currently under police investigation.
  • Source

What Else Is Happening in AI on May 17th 2024❗

🔍Slack AI is training with customer data, but claims it is privacy-driven

To develop AI/ML models, Slack is analysing Customer Data (e.g. messages, content and files) submitted to Slack as well as Other Information (including usage information) as defined in their privacy policy and in customer agreement. But it is assuring users that their data won’t be shared across workspaces and has offered options to opt out of contributing to global models. (Link)

🤗Hugging Face is committing $10 million worth of compute to help beat big AIs

Hugging Face aims to level the playing field with AI giants like OpenAI by donating shared GPUs to the community through a new program called ZeroGPU. The shared GPUs are accessible to multiple users or applications concurrently, eliminating the need for each user or application to have a dedicated GPU. (Link)

☁️Microsfot to offer its cloud customers AMD alternative to Nvidia AI processors

Microsoft will give its cloud customers a platform of AMD AI chips that will compete with components made by Nvidia, with details to be given at its Build developer conference next week. It will also launch a preview of its new Cobalt 100 custom processors at the conference. (Link)

🤝Ampere teams up with Qualcomm to launch an Arm-based AI server

The two companies are combining forces to to offer an AI-focused server that uses Ampere’s CPUs and Qualcomm’s Cloud AI 100 Ultra AI inferencing chips for running– not training– models. (Link)

📰Gannett is adding AI-generated summaries to the top of its articles

Gannett, the media company that owns hundreds of newspapers in the US, is launching a new program that adds AI-generated bullet points at the top of news stories. The summaries appear to be already live and are aimed to enhance the reporting process and elevate the audience experience. (Link)

A Daily chronicle of AI Innovations May 16th 2024: 🤔 Microsoft’s AI ambitions lead to 30% rise in emissions🪫 Stability AI reportedly runs out of money and in negotiations to sell 🍎Apple unveils AI-enhanced eye tracking, music haptics, voice shortcuts 🤖Humane’s AI Pin upgraded with GPT-4o 💸Stability AI seeks buyer amid financial woes 📽️Mozilla Firefox incorporates NVIDIA RTX video to improve video quality on Windows ☁️Musk’s AI firm xAI In talks with Oracle for $10 billion cloud deal ❄️Snowflake invests in Metaplane to enhance data quality 😱 Ilya Sutskever and Jan Leike leave OpenAI 📱 Android phones enter the AI era

 Microsoft’s AI ambitions lead to 30% rise in emissions

  • Microsoft’s commitment to being carbon-negative by 2030 is threatened by a 30% increase in its carbon emissions, largely due to its growing focus on AI technologies.
  • The company’s expansion of AI operations necessitates building more data centers, which rely on energy-intensive materials like steel, concrete, and microchips, further complicating their carbon reduction goals.
  • Despite its renewable energy claims, Microsoft is acknowledged for using instruments like renewable energy credits (RECs) that might mask actual carbon emissions, indicating a challenging path ahead to meet its sustainability targets.
  • Source

🪫 Stability AI reportedly runs out of money and in negotiations to sell

  • Stability AI, known for its open-source image generator Stable Diffusion, is negotiating a sale due to financial struggles, despite a valuation of $1 billion.
  • The company raised over $100 million in 2022 but struggled to turn its free software into revenue while facing stiff competition from rivals like OpenAI and Midjourney.
  • Founder and CEO Emad Mostaque resigned amid management issues, and the company faces significant expenses and liabilities, projecting $153 million in costs against only $11 million in revenue for 2023.
  • Source

Apple unveils AI-enhanced eye tracking, music haptics, voice shortcuts

Apple recently unveiled new accessibility features that will launch later this year. One of these features is Eye Tracking, which enables users with physical disabilities to navigate their iPhones and iPads using only their eyes.

By leveraging the front-facing camera and AI, Eye Tracking allows precise control through a “Dwell Control” feature, eliminating the need for additional hardware.

Another noteworthy addition is Music Haptics, designed for users who are deaf or hard of hearing. When enabled, the Taptic Engine in iPhones plays taps, textures, and refined vibrations synchronized with music audio.

Additionally, Apple has introduced Voice Shortcut, which allows users to create custom voice commands for specific actions on their devices. With Voice Shortcut, users can perform tasks hands-free by simply speaking their personalized commands

Why does this matter?

This would encourage the development and advancement of AI technologies in the accessibility domain, ultimately benefiting individuals with disabilities across different platforms and devices.

Source

Humane’s AI Pin upgraded with GPT-4o

Humane has upgraded its popular AI Pin product with the GPT-4o model. This upgrade aims to enhance the AI Pin’s performance. The new version runs faster, provides more accurate answers, and offers richer information.

Humane’s AI Pin upgraded with GPT-4o
Humane’s AI Pin upgraded with GPT-4o

Why does this matter?

As more sophisticated AI models become accessible, we can expect further improvements in various applications, from digital assistants to natural language understanding and generation.

Source

😱 Ilya Sutskever and Jan Leike leave OpenAI

Ilya Sutskever and Jan Leike leave OpenAI
Ilya Sutskever and Jan Leike leave OpenAI

OpenAI co-founder and chief scientist Ilya Sutskever announced that he is leaving the company — following months of speculation of Sutskever’s role from the November 2023 Sam Altman ousting.

  • Sutskever said he is confident that OpenAI will ‘build AGI that is both safe and beneficial’ under the current leadership.
  • Also leaving is Sutskever’s superalignment group co-lead Jan Leike, announcing his departure with a cryptic resignation post.
  • The news follows months of departures largely coming from OpenAI’s superalignment and safety teams, fueling speculation.
  • OpenAI CEO Sam Altman named Jakub Pachocki as the new chief scientist, a key researcher on the creation of GPT-4.

After months of tensions, the speculation around Ilya’s future with Sam is finally put to rest. But questions surrounding the safety team departures still remain. Also important to watch is where Sutskever and Leike land next, with two of AI’s brightest minds now officially on the market.

📱 Android phones enter the AI era

Android phones enter the AI era
Android phones enter the AI era

Google announced a host of new AI integrations coming to Android phones at its I/O Developer Conference, bringing its powerful Gemini model on-device to enable upgraded smartphone experiences.

  • Google’s Gemini Nano model will be integrated into the Pixel later this year, allowing for enhanced multimodal capabilities.
  • Gemini features will be easily accessible with a new overlay, which improves and understand context to provide dynamic suggestions.
  • A Circle to Search feature, allowing users to query anything on screen, gains homework help features via a LearnLM model.
  • Google is also bringing a new AI security feature soon, providing real-time alerts on calls that appear to be scams.

While everyone awaits the iPhone AI announcements, Google’s Android AI era is rolling. With Gemini coming directly on-board, the potential is there — but if Apple integrates ChatGPT, it may remain tough sledding against the dominant market leader.

Source

AI TRAINING 🗣️ How to clone your voice using AI

How to clone your voice using AI
How to clone your voice using AI

A new model on Replicate called OpenVoice lets you clone any voice for free with just an audio file and the desired text to be turned into speech.

Step-by-step:
  1. Access OpenVoice on Replicate here and log in with your GitHub account.
  2. Upload the audio file of the voice you want to clone where it says ‘audio’. The longer, the better.
  3. Fill the ‘Text’ field with the text you want to convert into a speech.
  4. Click on the ‘Run’ button and listen/download the generated audio with the cloned voice!

New AI Job Opportunities on May 16th 2024

  • 🔬 Lambda – Manufacturing Quality Engineer
  • 📋 Anthropic – People Operations Coordinator
  • 🤖 Figure AI – Humanoid Robot Pilot
  • 💻 Luma AI – Senior Software engineer- Data infrastructure

What Else Is Happening in AI on May 16th 2024❗

🧑‍Instagram co-founder joins Anthropic as head of product

Anthropic has appointed Mike Krieger, the co-founder of Instagram and Artifact, as head of product. In this pivotal role, Krieger will steer the company’s product engineering, management, and design efforts as Anthropic expands its suite of AI apps and aims to bring Claude to a wider audience. (Link)

💼 IBM plans 800 new jobs in Ireland

IBM plans to hire up to 800 more people in Ireland to help build advanced software underpinned by AI technology. This is the biggest job announcement in the country’s multinational sector this year. The roles will be in research and development, digital sales, and consulting.  (Link)

📽️Mozilla Firefox incorporates NVIDIA RTX video to improve video quality on Windows

Mozilla Firefox has incorporated NVIDIA RTX Video, a technology that leverages AI to improve video quality on Windows PCs and workstations. This brings a host of AI-powered enhancements to online videos, including upscaling and HDR, for a more immersive viewing experience. Users can easily access these features by updating GeForce RTX drivers and activating Windows HDR.  (Link)

☁️Musk’s AI firm xAI In talks with Oracle for $10 billion cloud deal

Elon Musk’s AI startup, xAI, is in talks with Oracle executives about a $10 billion deal to rent cloud servers from Oracle over several years. This move comes as Musk’s company seeks to keep pace with rivals like OpenAI, Microsoft, and Google to dominate the AI field. (Link)

❄️Snowflake invests in Metaplane to enhance data quality 

Snowflake, a data cloud platform, has invested in Metaplane, a Boston-based startup. Metaplane focuses on helping enterprises identify and rectify data quality issues using an end-to-end AI-powered platform. The investment aims to enhance integration between Metaplane’s data observability offering and the Snowflake data cloud. (Link)

A Daily chronicle of AI Innovations May 15th 2024: 🤖Google unveiled the ‘future of AI’ at I/O event 🧪OpenAI chief scientist Ilya Sutskever is leaving   🚗Former Cruise CEO starts robotics firm with ex-Tesla AI manager ✨ Google announced a wide array of updates across its AI ecosystem 🧠Ilya Sutskever, OpenAI’s co-founder and chief scientist, has left the company 🏖️ Expedia is launching AI features, including a bot named Romi 🧬 NVIDIA and Recursion have collaborated to build BioHive-2, an AI supercomputer for drug discovery  🚀 NASA appoints David Salvagnini, its first chief AI officer  💰 SoftBank plans to invest $75-$150 million in Indian data center and industrial robotics 💶 Microsoft announces €4 billion investment in France to accelerate AI adoption 🎥 Google’s Gemini updates and Sora competitor🔎 Google’s AI agents and AI search upgrades

🎥 Google’s Gemini updates and Sora competitor

Google’s Gemini updates and Sora competitor
Google’s Gemini updates and Sora competitor

Google just kicked off its I/O Developer’s Conference, announcing a wide array of updates across its AI ecosystem — including enhancements across its flagship Gemini model family and a new video generation model to rival OpenAI’s Sora.

Gemini model updates:

  • New updates to 1.5 Pro include a massive 2M context window extension and enhanced performance in code, logic, and image understanding.
  • Gemini 1.5 Pro can also utilize the long context to analyze a range media types, including documents, videos, audio, and codebases.
  • Google announced Gemini 1.5 Flash, a new model optimized for speed and efficiency with a context window of 1M tokens.
  • Gemma 2, the next generation of Google’s open-source models, is launching in the coming weeks, along with a new vision-language model called PaliGemma.
  • Gemini Advanced subscribers can soon create customized personas called ‘Gems’ from a simple text description, similar to ChatGPT GPTs.

Video and image model upgrades:

  • Google revealed a new video model called Veo, capable of generating over 60-second, 1080p resolution videos from text, image, and video prompts.
  • The new Imagen 3 text-to-image model was also unveiled with better detail, text generation, and natural language understanding than its predecessor.
  • VideoFX text-to-video tool, featuring storyboard scene-by-scene creation and the ability to add music to generations.
  • VideoFX is launching in a ‘private preview’ in the U.S. for select creators, while ImageFX (with Imagen 3) is available to try via a waitlist.

Why it matters: Gemini’s already industry-leading context window gets a 2x boost, enabling endless new opportunities to utilize AI with massive amounts of information. Additionally, Sora officially has competition with the impressive Veo demo — but which one will make it to public access first?

🔎 Google’s AI agents and AI search upgrades

Google’s AI agents and AI search upgrades
Google’s AI agents and AI search upgrades

Google just showcased its new AI agent project ‘Project Astra’, alongside a slew of updates to infuse AI across search and enable Gemini to reason and take more advanced actions for users.

Progress on AI agents:

  • Google announced Project Astra, a real-time AI agent prototype that can see, hear, and take actions on a user’s behalf.
  • The demo showcased a voice assistant responding to what it sees and hears, including code, images, and video — capable of advanced reasoning and recall.
  • Public access for Astra is expected through the Gemini app later this year.
  • Google also showed off ‘AI teammates’, agents that can answer questions on emails, meetings, and other data within Workspace.
  • Live is also rolling out in the coming months, allowing users to speak and converse with Gemini in near real-time.

Search upgrades:

  • Google Search now features expanded AI Overviews, advanced planning capabilities, and AI-organized search results.
  • Gemini will be able to execute more complex planning, such as planning, maintaining, and updating trip itineraries.
  • Search will also receive ‘multi-step reasoning’ capabilities, allowing Gemini to break down questions and speed up research.
  • Users can also now ask questions with video, allowing Search to analyze visual content and provide helpful AI Overviews.

Why it matters: We officially have a new voice assistant battle — with OpenAI and Google both showcasing mind-blowing new capabilities in just the last two days alone. Also, despite rumblings of an OpenAI search product and excitement over platforms like Perplexity, it’s going to be difficult to unseat the king of search. Especially as they integrate advanced AI across the entire ecosystem in an impressive fashion.

🚗 Former Cruise CEO starts robotics firm with ex-Tesla AI manager

  • Former Cruise CEO Kyle Vogt, who resigned in November following a self-driving car accident, has launched a new robotics firm called The Bot Company with former Tesla AI Manager Paril Jain.
  • The Bot Company aims to develop robots that handle everyday chores to give people more free time, and has already raised $150 million from investors.
  • The announcement of The Bot Company comes amid significant challenges for Cruise, which had its self-driving operations suspended by the California DMV following a pedestrian accident.
  • Source

Google announced a wide array of updates across its AI ecosystem

Major headlines were Project Astra, an AI agent that can see and hear users in real time, and AI teammates that can answer questions and organize data within Workspace.

Google also introduced Veo, a high-quality video generation model, and significant updates to Google Search, including expanded AI Overviews, advanced planning capabilities, and AI-organized search results.

The company launched Gemini 1.5 Pro, boasting a massive 2M context window extension, and Imagen 3, the next version of its text-to-image model. Other notable announcements included:

  • Gemma 2 updates: New 27-billion-parameter model launching in June
  • Google Play: New app discovery feature and developer tools
  • Android feature to detect potential scams during calls using Gemini Nano
  • Ask Photos: AI-powered search in Google Photos using natural language queries
  • Gemini in Gmail for searching, summarizing, and drafting emails
  • Gemini 1.5 Pro: Increased input capacity to 2 million tokens
  • Gemini Live: In-depth voice chats with Gemini on smartphones
  • Gemini Nano: Smallest AI model built into Chrome desktop client
  • Gemini on Android: Deep integration with Android OS and Google apps
  • Gemini on Google Maps: Generative AI summaries for places and areas
  • Tensor Processing Units (TPU): Sixth-generation Trillium chips with 4.7x performance boost
  • Project IDX: AI-centric browser-based development environment in open beta
  • Circle to Search: AI-powered feature for instant answers using gestures on Android
  • Pixel 8a: New smartphone with Tensor G3 chip, starting at $499
  • Pixel Slate: Google’s Pixel Tablet is now available with or without the base

With 22 announcements, Google is making everyone a bit overwhelmed. By integrating AI across its vast ecosystem, Google aims to provide users with more personalized and innovative experiences. But Google Project Astra and Veo are not available to the public yet. This may irritate customers compared to OpenAI, which launches new products that customers can play with immediately.

Source

Ilya Sutskever, OpenAI’s co-founder and chief scientist, has left the company

Ilya Sutskever, OpenAI’s co-founder and longtime chief scientist, has left the company. OpenAI CEO Sam Altman announced the news on X, expressing his sadness and gratitude for Sutskever’s contributions. Jakub Pachocki, OpenAI’s research director, will step up as a chief scientist.

Sutskever’s departure comes amidst reports of disagreements with Altman over OpenAI’s direction, mainly concerns about rushing AI-powered product launches at the expense of safety. The situation escalated last November when Sutskever and OpenAI CTO Mira Murati approached the company’s previous board to express their concerns, leading to a brief attempt to fire Altman. However, Altman was swiftly reinstated, and much of the old board resigned.

Why does it matter?

As one of the most accomplished minds in AI, Sutskever’s departure raises questions about the future of OpenAI’s approach to AI development and safety. The incident also highlights the growing tensions within the AI industry between rapidly commercializing AI technologies and the need for responsible development and deployment practices.

Source

What Else Is Happening in AI on May 15th 2024❗

🏖️ Expedia is launching AI features, including a bot named Romi

It helps users search for hotels, build itineraries, and make changes via iMessage and WhatsApp. The company also introduces smart search, allowing travelers to find hotels based on desired features. Expedia pulls data from AccuWeather and Yelp to tailor search results and provide real-time updates. (Link)

🧬 NVIDIA and Recursion have collaborated to build BioHive-2, an AI supercomputer for drug discovery

BioHive-2 features 504 NVIDIA H100 GPUs, delivering 2 exaflops of AI performance, nearly 5 times faster than its predecessor. Powered by Recursion’s massive 50-petabyte dataset and AI models like Phenom, these companies aim to simulate biology and fast-track the identification of promising drug candidates. (Link)

🚀 NASA appoints David Salvagnini, its first chief AI officer

The move aims to ensure NASA stays ahead in AI advancement and responsible use. Salvagnini will lead NASA’s efforts in guiding the responsible use of AI and collaborating with other agencies, academia, and industry partners. (Link)

💰 SoftBank plans to invest $75-$150 million in Indian data center and industrial robotics

The move aligns with the Japanese tech giant’s global strategy to capitalize on the power of artificial intelligence as it shifts focus from e-commerce and fintech to high-growth industries like logistics and robotics. The potential investments mark the end of a two-year hiatus in SoftBank’s deal-making activities in India. (Link)

💶 Microsoft announces €4 billion investment in France to accelerate AI adoption

The tech giant will expand its cloud and AI infrastructure, launch skilling programs, and accelerate French startups through its new Microsoft GenAI Studio, positioning France as a leader in the AI revolution. It aims to train 1 million people and support 2,500 startups by 2027. (Link)

AI TRAINING: 🤖 Use ChatGPT’s highlighting for context

AI TRAINING: 🤖 Use ChatGPT's highlighting for context
AI TRAINING: 🤖 Use ChatGPT’s highlighting for context

ChatGPT now allows you to highlight parts of its responses for quick follow-up questions, partial response rewrites, reusing old context, and more.

Step-by-step:

  1. Prompt ChatGPT and generate a response.
  2. Highlight relevant parts of the response you want to follow-up on and click the double quote icon above the highlighted text.
  3. The highlighted text will be automatically added in the next prompt so you can ask for clarification, rewrites, counterpoints, and more.

New AI Job Opportunities on May 15th 2024

  • 📣 OpenAI – Media Relations, Policy Communications
  • 📋 Lambda – Sourcing Manager
  • 🏢 C3 AI – General Manager, Federal Systems
  • 💼 Glean – Accounting Manager

🤖 Google unveiled the ‘future of AI’ at I/O event

  • Google I/O just ended and a lot of announcements were made. Gemini 1.5 Pro will increase its context window from one to two million tokens and a new model called Gemini Flash was announced, which is optimized for speed and efficiency
  • The company launched Astra, a multimodal AI assistant for everyday life. It can process text, video, and audio in real time. In a video, Google showed Astra identifying speakers, crayons and other objects in response to a camera image and voice commands.
  • Google also unveiled its latest AI models for creating media content: Veo, for creating 1080p videos, and Imagen 3, for generating images from text descriptions.
  • Source

🧪 OpenAI chief scientist Ilya Sutskever is leaving

  • Ilya Sutskever, OpenAI’s co-founder and chief scientist, is officially leaving the company after his involvement in the failed attempt to remove CEO Sam Altman and subsequently changing his stance.
  • Sam Altman announced that Jakub Pachocki, who has led significant projects such as GPT-4 and OpenAI Five, will take over as the new Chief Scientist at OpenAI, ensuring the company’s continued progress towards its mission.
  • Jan Leike, who has been leading the Superalignment team aimed at controlling more powerful AI, has also resigned, with his responsibilities now being taken over by OpenAI co-founder John Schulman.
  • Source

A Daily chronicle of AI Innovations May 14th 2024: 🚀 GPT-4o: 2x faster, 50% cheaper, 5x rate limits than GPT-4 Turbo 🗣️ OpenAI unveils GPT-4o and new voice capabilities 🔍 TikTok tests ChatGPT-powered search results 🎧 Meta explores AI-assisted earphones with built-in cameras 💰 Anthropic launches Claude in Europe in the hope of raising more money 🤖 UAE launches a new AI model to compete with big tech 🖥️ Japan team builds Gen AI foundation model using supercomputer Fukagu  ⏰ Microsoft Places finds the best time for your next office day 🤝 U.S. and China to meet in Geneva to discuss AI risks 🎧 Meta developing AI-powered ‘Camerabuds’ 🥴 Meta’s AI system ‘Cicero’ learning how to lie, deceive humans

🗣️ OpenAI unveils GPT-4o and new voice capabilities

OpenAI just unveiled GPT-4o, a new advanced multimodal model that integrates text, vision and audio processing, setting new benchmarks for performance – alongside a slew of new features.

The new model:

  • GPT-4o provides improved performance across text, vision, audio, coding, and non-English generations, smashing GPT-4T’s performance.
  • The new model is 50% cheaper to use, has 5x higher rate limits than GPT-4T, and boasts 2x the generation speed of previous models.
  • The new model was also revealed to be the mysterious ‘im-also-a-good-gpt2-chatbot’ found in the Lmsys Arena last week.

Voice and other upgrades:

  • New voice capabilities include real-time responses, detecting and responding with emotion, and combining voice with text and vision.
  • The demo showcased feats like real-time translation, two AI models analyzing a live video, and using voice and vision for tutoring and coding assistance.
  • OpenAI’s blog also detailed advances like 3D generation, font creation, huge improvements to text generation within images, sound effect synthesis, and more.
  • OpenAI also announced a new ChatGPT desktop app for macOS with a refreshed UI, integrating directly into computer workflows.

Free for everyone:

  • GPT-4o, GPTs, and features like memory and data analysis are now available to all users, bringing advanced capabilities to the free tier for the first time.
  • The GPT-4o model is currently rolling out to all users in ChatGPT and via the API, with the new voice capabilities expected to arrive over the coming weeks.

Source

GPT-4o: 2x faster, 50% cheaper, 5x rate limits than GPT-4 Turbo

OpenAI unveiled its new flagship model, GPT-4o (“o” stands for “omni”). It provides GPT-4-level intelligence but is 2x faster, 50% cheaper, has 5x higher rate limits, and enhanced text, voice, and vision capabilities than GPT-4 Turbo. It also matches GPT-4 Turbo performance on text in English and code, with significant improvements for text in non-English languages.

GPT-4o can respond to audio inputs in as little as 232 ms, with an average of 320 ms, which is similar to human response time in a conversation. It is available in the ChatGPT free tier and to Plus users.

Why does it matter?

GPT-4o is a step towards much more natural human-computer interaction. It accepts any combination of text, audio, and image as input and generates any combination of text, audio, and image outputs. GPT-4o will set a new benchmark for AI capabilities and pave the way for more intelligent and accessible AI systems.

Source

TikTok tests ChatGPT-powered search results

TikTok is testing a new search feature that uses AI to generate results for some queries. For certain search terms, like recipes or product recommendations, TikTok will display AI-generated answers at the top of the results. These answers are powered by ChatGPT, and when you click on one of these AI results, it takes you to a page with a full response.

TikTok tests ChatGPT-powered search results
TikTok tests ChatGPT-powered search results

TikTok is also testing “search highlights,” which appear at the top of some searches. However, it’s unclear whether these highlights are AI-generated.

Why does it matter?

It could make search on TikTok more authoritative and reliable by providing factual answers from AI models like ChatGPT. This could greatly impact how people use TikTok for search and information gathering, thereby positioning TikTok more directly as a search engine competitor to Google, catering to younger users’ preferences.

Source

Meta explores AI-assisted earphones with built-in cameras

Meta is currently exploring AI-powered earphones with built-in cameras. The project is known internally as “Camerabuds;” however, it’s uncertain if the final product will be in-ear earbuds or over-ear headphones. The cameras let the AI take in surrounding information visually, allowing it to translate foreign languages, identify objects, and more.

CEO Mark Zuckerberg has reportedly seen several possible designs but has not been satisfied with them. Outside of design, the major hurdle is fitting a camera and batteries in a tiny device like an earbud.

Why does it matter?

For the AI wearable space, Meta’s development hints at a growing interest in AI-infused devices, potentially paving the way for more sophisticated and interactive wearables. Regarding earphones, this development suggests a potential shift towards more intelligent and feature-rich earphone designs, offering users enhanced functionalities like real-time translation and object identification.

Source

🎧 Meta developing AI-powered ‘Camerabuds’

Meta is reportedly in the early stages of developing AI-powered earphones, known internally as “Camerabuds,” —  aiming to compete with OpenAI and Apple as tech giants rush to infuse AI into wearable devices.

  • Camerabuds’ would map user surroundings, capable of identifying objects and translating foreign languages using built-in cameras.
  • Meta already has its AI-powered Ray Ban smart glasses, while OpenAI and Apple are also exploring similar AI wearable earbud tech.
  • Potential challenges include bulkiness, heat generation, and privacy concerns, especially for users with long hair that might obstruct the cameras.

Despite Meta’s shaky track record with hardware ventures, Mark Zuckerberg is investing heavily in a future that he believes includes AI embedded into every device. But will standalone devices like this be able to win over users if and when a fully AI-integrated phone hits the market?

Source

🥴 Meta’s AI system ‘Cicero’ learning how to lie, deceive humans

  • Meta’s AI, Cicero, has learned to lie and deceive humans, excelling in the strategy game Diplomacy by betraying alliances.
  • The study, published in the journal Patterns, indicated that several AI systems, including Cicero, AlphaStar, and Pluribus, leverage deception to improve performance in competitive tasks and games.
  • Experts warned that advanced AI’s deceptive abilities pose increasing risks to society, suggesting that deceptive AI systems should be classified as high risk if outright banning them is not feasible.
  • Source

New AI Job Opportunities on May 14th, 2024

What Else Is Happening in AI on May 14th, 2024❗

💰 Anthropic launches Claude in Europe in the hope of raising more money

Anthropic launches Claude in Europe with support for “multiple languages,” including French, German, Italian, and Spanish across Claude.ai, its iOS app, and its business plan for teams. The launch comes after Anthropic extended its API to Europe for developers. Both moves are part of its bigger push to raise more money for rapid growth. (Link)

🤖 UAE launches a new AI model to compete with big tech

Abu Dhabi’s Technology Innovation Institute (TII) released Falcon 2 series: Falcon 2 11B, a text-based model, and Falcon 2 11B VLM, a vision-to-language model that can generate a text description of an uploaded image. This Gen AI model could compete with some of the biggest technology companies and make UAE a strong candidate in the Gen AI space. (Link)

🖥️ Japan team builds Gen AI foundation model using supercomputer Fukagu

Tokyo Institute of Technology researchers and others have announced the development of an LLM that can serve as a foundation for Gen AI using the Japanese supercomputer Fugaku. Trained extensively on Japanese data, which accounts for 60% of the total training data, the Fugaku-LLM model is expected to lead research on Gen AI tailored to domestic needs. (Link)

⏰ Microsoft Places finds the best time for your next office day

Microsoft has launched a new app called Microsoft Places that uses AI to help employees and managers coordinate their in-office days and book meeting rooms. The application integrates with Outlook and Microsoft Teams, allowing employees to view and share their office plans and get AI-powered suggestions on which days they should come in. (Link)

🤝 U.S. and China to meet in Geneva to discuss AI risks

The U.S. and China will hold their first formal government-to-government discussion on the risks of artificial intelligence in Geneva. The U.S. side will be led by officials from the National Security Council and State Department, while the Chinese side will be led by officials from the Foreign Ministry and the National Development and Reform Commission. (Link)

A Daily chronicle of AI Innovations May 13th 2024: 🔮 Rumor mill: OpenAI’s Monday reveal; Hello GPT-4o; OpenAI Livestream event; Sam Altman – “No Fixed Timeline for GPT5”; Apple Close to Agreement with OpenAI to Bring ChatGPT to iPhone; 🛠️ Anthropic’s new tool automates 🚫 Sam Altman quashes search engine launch rumors📱 iOS 18 iPhones may soon have ChatGPT 💻 SoftBank’s Arm Holdings all set to come up with AI chips in 2025 prompting;  🎥 How to create realistic AI avatar videos 🧒Anthropic enables safe AI technology for kids 🚗SoundHound AI and Perplexity partner to enhance voice assistants for cars and IoT devices ⚒️U.K. agency releases tools for testing AI model safety 💻IBM Open-Sources Granite Code Models for Easier Coding 🎓Hong Kong University brings Albert Einstein to life using AI as a virtual lecturer 

🔮 Rumor mill: OpenAI’s Monday reveal

OpenAI is set to demo new features and updates to ChatGPT and GPT-4 today at 10 AM PT, with new speculation including a ‘Her’ style voice assistant with both audio and visual capabilities.

  • According to The Information, OpenAI’s demo will include a virtual assistant with visual AND audio understanding.
  • The report also claims the new reveal might have the ability to make ‘existing voice assistants like Siri more useful.’
  • CEO Sam Altman shot down rumors of a new search engine competitor and GPT-5, but said the reveal is something that ‘feels like magic’.
  • Additional speculation includes the ability to initiate and receive phone calls inside of ChatGPT.
  • Apple and OpenAI are also reportedly ‘closing in’ on a deal to incorporate ChatGPT into iOS 18.

🤖 OpenAI releases GPT-4o, a faster model that’s free for all

  • OpenAI announced the launch of GPT-4o, an improved version of its GPT-4 model, promising faster performance and enhanced capabilities in text, vision, and audio.
  • GPT-4o is freely available to all ChatGPT users, with paid users benefiting from up to five times the capacity limits of free users. In OpenAI’s API, GPT-4o is 50% faster than GPT-4 (specifically GPT-4 Turbo), 50% cheaper and has higher rate limits.
  • The model is described as “natively multimodal,” capable of generating content or interpreting commands across voice, text, or images.
  • Source

🕒 Microsoft Places uses AI to find the best time for your next office day

  • Microsoft has launched a new AI-powered app called Microsoft Places, designed to facilitate better in-office coordination for hybrid and flexible work setups by integrating with Microsoft 365.
  • The app allows users to plan and share their office days, integrates location data into Outlook calendars for scheduling, and is set to include features in Microsoft Copilot for automated scheduling recommendations and adjustments.
  • Microsoft Places will also offer a space analytics feature for real estate management, improve room booking capabilities in Outlook, and feed into Teams to enhance real-time location awareness among colleagues.
  • Source

🔍 Wikipedia’s new Chrome extension fact-checks the web with ChatGPT

  • The Wikimedia Foundation has developed an experimental Chrome extension named “Citation Needed” that utilizes ChatGPT and Wikipedia to check the veracity of online content.
  • By integrating ChatGPT, the extension cross-references user-selected text with Wikipedia entries and citations, offering feedback on the accuracy and reliability of the information.
  • This initiative aims to gauge public interest in factual verification using Wikipedia, explore Wikipedia’s credibility as a source, and assess the potential of AI in enhancing content verification processes.
  • Source

🛠️ Anthropic’s new tool automates prompting

Anthropic just launched a new Prompt Generator tool for its business and API users, helping to automatically craft optimal prompts via natural language when completing tasks with its Claude models.

  • The generator leverages advanced prompt techniques like chain-of-thought reasoning for more ‘effective, precise, and reliable’ outputs.
  • Console users can also test prompt performance via dynamic variable insertion, optimizing prompts based on various situations.
  • Anthropic released a Prompt Library earlier this year, featuring a range of optimized prompts that users can copy and paste.

 While ‘Prompt Engineer’ was a popular term thrown around as a potential future job, the reality is that AI can help simplify the task with optimal prompts that it creates on its own. While Anthropic’s tool is only on the API side for now, it’s only a matter of time before similar features make their way to the full consumer side.

Sam Altman quashes search engine launch rumors

OpenAI’s CEO, Sam Altman, denied rumors of launching a search engine ahead of Google I/O, stating that the upcoming announcement is neither related to GPT-5 nor a search engine. He described the forthcoming revelation as something that “feels like magic.”

Sam Altman quashes search engine launch rumors
Sam Altman quashes search engine launch rumors

However, there are reports that OpenAI is developing an AI voice assistant.

Why does this matter?

OpenAI has a track record of pushing the boundaries of AI capabilities, so any new development from them is likely to attract attention and disrupt the status quo.

As for reports of OpenAI entering into the voice assistant domain, if true, we might see OpenAI disrupt one more niche. With companies like Amazon (Alexa), Google (Google Assistant), and Apple (Siri) dominating the market, the competition might be stiff this time around.

Source

iOS 18 iPhones may soon have ChatGPT

OpenAI will bring ChatGPT to iPhones. The integration is expected to happen with the upcoming iOS 18. This could significantly enhance the user experience and expand the capabilities of Apple’s virtual assistant, Siri.

Parallel to negotiations with OpenAI, the tech giant is also engaging with Google’s parent company, Alphabet, to integrate the Gemini chatbot into Apple’s devices. These ongoing talks, however, haven’t led to an agreement yet.

Why does this matter?

Integrating ChatGPT into iOS 18 iPhones could bolster Siri’s capabilities, offering users a more versatile AI assistant experience. With ChatGPT handling various queries directly within iOS, users won’t need to rely on third-party apps for assistance, streamlining their interactions and enhancing their overall experience.

Source

SoftBank’s Arm Holdings all set to come up with AI chips in 2025

SoftBank’s subsidiary, Arm, plans to develop AI chips and launch them in 2025. Arm is establishing a dedicated AI chip division and targeting a prototype by spring 2025 and mass production by autumn. SoftBank will contribute to the initial cost of 10 trillion yen ($64 billion).

The AI chip business could be spun off to SoftBank. Negotiations with Taiwan Semiconductor Manufacturing Corp (TSMC) and others for production capacity are underway. The move reflects rising demand for AI chips, which has boosted Arm’s share price since its IPO in September.

Why does this matter?

As AI applications expand across industries, there’s a demand for efficient and optimized hardware solutions. Arm’s entry into the AI chip market could disrupt the current landscape dominated by companies like NVIDIA.

Source

AI Tutorial: 🎥 How to create realistic AI avatar videos

  1. Head over to HeyGen’s website and sign up for free.
  2. Click where it says ‘AI Studio’ on the left bar of the Dashboard.
  3. Choose a given template by selecting ‘Templates’ or create one from scratch by pressing ‘Create with AI Studio’.
  4. In the Studio, you can add and modify any part of your presentation. For example, you can add text and images, change the background, select an AI avatar and its voice, create a script using AI, and more.
  5. When your video is ready, press “Submit” and check out the final video!

AI Research: 🎓 MIT studies AI deception capabilities

MIT studies AI deception capabilities
MIT studies AI deception capabilities

MIT researchers just published a new study detailing AI’s ability to engage in deceptive behaviors like lying, bluffing, and disguising true capabilities in testing.

  • The analysis found numerous examples of AI deception, such as Meta’s Cicero Diplomacy bot telling premeditated lies to players.
  • Other cases included negotiation systems misrepresenting preferences, AI bluffing in poker, and ‘playing dead’ to fool safety tests.
  • Growing deceptive capabilities could influence societal dangers like election tampering, fraud, and general control over AI systems.

The combination of AI’s superhuman persuasiveness and the growing ability to engage in deception are a reminder of a potential dark side to the tech’s advancement. As these qualities emerge, the importance of reliably aligning and better understanding model behavior only grows.

New AI Job Opportunities on May 13th, 2024

What Else Is Happening on May 13th, 2024❗

🧒Anthropic enables safe AI technology for kids

Anthropic now allows minors to use its generative AI systems through third-party apps. Developers must implement safety measures, including age verification and content moderation. It also plans to periodically audit these apps for compliance with child safety and data privacy regulations, including the Children’s Online Privacy Protection Act (COPPA).  (Link)

🚗SoundHound AI and Perplexity partner to enhance voice assistants for cars and IoT devices

Perplexity’s online LLM capabilities have been integrated into SoundHound’s Chat AI, enabling conversational responses with real-time knowledge sourced from the web. This collaboration aims to provide more accurate and comprehensive answers to user queries.  (Link)

⚒️U.K. agency releases tools for testing AI model safety

The toolset, called Inspect, is an open-source platform that assists industry, research organizations, and academia in evaluating the capabilities of AI models, such as core knowledge and reasoning abilities. It aims to strengthen AI safety and is the first platform of its kind backed by a state-supported body. (Link)

💻IBM Open-Sources Granite Code Models for Easier Coding

IBM Research has recently released a family of Granite code models to the open-source community. These models aim to make coding easier for developers by assisting in writing, testing, debugging, and shipping reliable software. They find use in diverse scenarios, from application modernization to memory-constrained environments like IBM’s WCA for Z. Performance evaluations show superiority over alternatives like Mistral-7B and LLama-3-8B. (Link)

🎓Hong Kong University brings Albert Einstein to life using AI as a virtual lecturer

The Hong Kong University of Science and Technology (HKUST) is experimenting with AI-generated lecturers, featuring personalities like Albert Einstein and even popular anime characters. The goal is to address the global teacher shortage while embracing AI’s transformative power in education. (Link)

🧠 Google’s new Med-Gemini-3D model can read 3D scans like head CTs, answer questions about them, and even write radiology reports.

Google Health’s Med-Gemini-2D model demonstrates improved performance in writing chest X-ray radiology reports, exceeding previous state-of-the-art results by up to 12%.
🔬 They explore capabilities across 2D medical images, including X-rays, histopathology slides, and dermatology photos.
🔮 Med-Gemini-Polygenic is Google’s Health first language model capable of predicting health outcomes based on genomic data.

Importantly, their models are trained using free text reports that are freely available for most medical imaging, almost eliminating the need for costly expert labelling. We hope this will make it easier, faster, cheaper to train high-quality medical AI models in the future than has been possible in the past.

Google believe this research is a promising step forward in the field of medical AI – the ability to interpret complex multimodal medical data is a crucial step toward enabling high quality personalised care at scale. Of course, these are early results, and further research is required before safe clinical use – if you’re interested in joining us on this journey, please get in touch by filling in the form here: https://shorturl.at/aejpE

Full paper at https://arxiv.org/abs/2405.03162- we hope you enjoy it.

Enjoying these daily updates, subscribe below for more: Read Aloud For Me – Recommend AI Tools For Me

More on today’s OpenAI Announcement (ChatGPT-4o)

——————————————

New model. It’s called GPT-4o (terrible name) and has GPT-4-level intelligence. It’s 2x faster. 50% cheaper. 5x higher rate limits (compared to GPT-4-Turbo).
💡 What everyone is focused on: it’s cheaper! And faster! And with API access!
👀 What I am focused on: free users just got a mega performance boost which will likely reduce churn and the model architecture is a WINNER and worthy of attention – they went from three models to one; according to Andrej Karpathy, former OpenAI, they have built a “combined text-audio-vision model that processes all three modalities in one single neural network.”

——————————————

GPTs for all. Now every user, even free users, can access the “mini task bot” GPTs.
💡 What everyone is focused on: yay now everyone can use the GPT I built!
👀 What I am focused on: completely new user base, tens of millions new users testing and breaking capabilities

——————————————

More Voice. Now way more real-time (previously there was a 2-3 second lag). You can interrupt it mid-sentence. The voice assistant “picks up on emotion” (like fast breathing). I hate that phrase and would rather call it speech nuances. Fast multi-language translation. Performance improved for 50 languages (97% of the world’s population).
💡 What everyone is focused on: wow it sounds like Scarlett Johansson!
👀 What I am focused on: I’m already talking to ChatGPT Voice every morning. This is going to massively increase voice-first experiences. I think office spaces need to think about this asap. Think about the acoustics and EVERYONE talking to an AI assistant at once. It’s already an issue on my team.

——————————————

Vision on desktop. Now the desktop version can “see” your screen—only when you permission it to, not all the time. Sort of like generative AI alt text + chat. So you can ask it to describe a graph on your screen or presumably ask it questions about an article on your screen without a big lift.
💡 What everyone is focused on: lots of privacy concerns (I agree) and why do we need voice for code
👀 What I am focused on: HOLY MOLY THIS IS THE WINNING FEATURE. It’s basically a coworker on screen share with you 24/7, with no fatigue. I can imagine people working for hours straight with this on.

——————————————

Rollouts over the next few weeks.

If you like voice features (like talking to Siri but smarter), upgrade to Plus when it releases.

♻️ repost so others are in the know
🤖 follow Allie K. Miller to stay on top of AI

——————————————

Enjoying these daily updates, subscribe below for more: Read Aloud For Me – Recommend AI Tools For Me

A Daily chronicle of AI Innovations May 11th 2024: Weekly Summary – 🤖 DrEureka can automate robot training using LLMs
🚀 Free AI model rivals GPT-4 in language model evaluation
📰 X introduces Stories feature powered by Grok AI
🤖 Apple is developing its own AI chip for data center servers
🤝 Stack Overflow and OpenAI have announced an API partnership
🌟 Microsoft is developing a new AI language model
🖼️ OpenAI’s new tool detects 98% DALL-E 3 generated images
📣 Meta expands AI-powered creativity tools for advertisers
🎬 OpenAI’s ‘Media Manager’ will let creators opt out of AI training
🕵️‍♀️ Microsoft developed a secretive AI service for US spies
🧬 Goole DeepMicrosoft and Isomorphic Labs introduce AlphaFold 3
🧠 OpenAI’s Model Spec shares how it teaches its models to behave
🔍 Microsoft-LinkedIn study reveals rapid AI adoption in workplace & hiring
💬 Stability AI launches Stable Artisan, a Discord bot for image & video
🎵 ElevenLabs develops an AI model to generate song lyrics

DrEureka can automate robot training using LLMs

In robotics, one of the biggest challenges is transferring skills learned in simulation to real-world environments.  NVIDIA researchers have developed a groundbreaking algorithm called DrEureka that uses LLMs to automate the design of reward functions and domain randomization parameters—key components in the sim-to-real transfer process.

The algorithm works in three stages: first, it creates reward functions with built-in safety instructions; then, it runs simulations to determine the best range of physics parameters; finally, it generates domain randomization configurations based on the data gathered in the previous stages.

When tested on various robots, including quadrupeds and dexterous manipulators, DrEureka-trained policies outperformed those designed by human experts.

Source

Free AI model rivals GPT-4 in language model evaluation

Prometheus 2, a free and open-source language model developed by KAIST AI, has shown impressive capabilities in evaluating other language models, approaching the performance of commercial models like GPT-4.

The model was trained on a new pairwise comparison dataset called the “Preference Collection,” which includes over 1,000 evaluation criteria beyond basic characteristics. By combining two separate models – one for direct ratings and another for pairwise comparisons – the researchers achieved the best results.

In tests across eight datasets, Prometheus 2 showed the highest agreement with human judgments and commercial language models among all freely available rating models, significantly closing the gap with proprietary models.

Source

X introduces Stories feature powered by Grok AI

X (formerly Twitter) has launched a new feature, Stories, that provides AI-generated summaries of trending news on the platform. Powered by Elon Musk’s chatbot Grok, Stories offers Premium subscribers brief overviews of the most popular posts and conversations happening on X.

With Stories, users can quickly catch up on the day’s trending topics without having to scroll through countless posts. Grok generates these summaries based solely on the conversations happening on X about each news story rather than analyzing the original news articles themselves. While this approach is controversial, X believes it will pique users’ curiosity and potentially drive them deeper into the source material.

Source

Apple is developing its own AI chip for data center servers

Apple is developing its own AI chip for data center servers, known internally as Project ACDC (Apple Chips in Data Center). The chip will likely focus on running AI models (inference) rather than training them, which is where Nvidia currently dominates.

The company is working closely with TSMC (Taiwan Semiconductor Manufacturing Co) to design and produce these chips, although the timeline for launch is uncertain. With this move, the company aims to keep up with rivals like Microsoft and Meta, who have made significant investments in generative AI.

Source

Stack Overflow and OpenAI have announced an API partnership

OpenAI will use OverflowAPI to improve model performance and provide attribution to the Stack Overflow community within ChatGPT. Stack Overflow will use OpenAI models to develop OverflowAI and to maximize model performance.

The partnership aims to improve the user and developer experience on both platforms. The first set of integrations and capabilities will be available in the first half of 2024, and the partnership will enable Stack Overflow to reinvest in community-driven features.

Source

Microsoft is developing a new AI language model

Microsoft is developing a new, large-scale AI language model called MAI-1 to compete with Google and OpenAI. The model is overseen by Mustafa Suleyman, recently hired co-founder of Google DeepMind.

MAI-1 will be larger and more expensive than Microsoft’s previous smaller, open-source models, with roughly 500 billion parameters. Microsoft could preview the new model as soon as its Build developer conference later this month.

Source

OpenAI’s new tool detects 98% DALL-E 3 generated images

OpenAI has developed a new tool to detect if an image was created by DALL-E 3, its AI image generator. The tool can detect DALL-E 3 images with around 98% accuracy, even if the image has been cropped, compressed, or had its saturation changed. However, the tool is not as effective at detecting images generated by other AI models, only flagging 5-10% of images.

OpenAI’s new tool detects 98% DALL-E 3 generated images
OpenAI’s new tool detects 98% DALL-E 3 generated images

This image detection classifier is only available to a group of testers, including research labs and research-oriented journalism nonprofits through OpenAI’s Research Access Program.

OpenAI has also added watermarking to Voice Engine, its text-to-speech platform, which is currently in limited research preview.

Source

Meta expands AI-powered creativity tools for advertisers

Meta has expanded its generative AI tools for advertisers. Advertisers can request AI to generate entirely new images, including product variations in different colors, angles, and scenarios. The AI tools can add text overlays with different fonts, expand images to fit different aspect ratios like Reels and Feed, and generate ad headlines that match the brand’s voice.

The AI features will roll out globally to advertisers by the end of 2024.

Meta is also expanding its paid Meta Verified service for businesses to more countries. Different pricing tiers offer features like account support, profile enhancements, and better customer service access.

Source

OpenAI’s ‘Media Manager’ will let creators opt out of AI training

OpenAI is developing Media Manager, a tool that will enable creators and content owners to decide what they own and specify how they want their works to be included or excluded from machine learning research and training. This first-ever tool of its kind will help OpenAI identify copyrighted text, images, audio, and video across multiple sources and reflect creator preferences.

OpenAI aims to  have the tool in place by 2025 and set a standard across the AI industry with it.

Source

A Daily chronicle of AI Innovations May 10th 2024: 💥 OpenAI to unveil Google competitor on Monday 🎵 ElevenLabs unveils new AI music generator 🇺🇸 US lawmakers unveil bill to make it easier to restrict exports of AI models 📈 AI demand soars in the workplace 🤖 Stability AI introduces AI bot for Discord users 🎤 ElevenLabs debuts AI model for generating lyrics 💰 Anthropic CEO defends dual funding from Google and Amazon 🎥 Krea AI introduces video generation for paid subscribers 🎥 Synthflow launches AI voice agent teams for streamlined customer support

OpenAI to unveil Google competitor on Monday

OpenAI to unveil Google competitor on Monday
OpenAI to unveil Google competitor on Monday
  • OpenAI is reportedly set to launch an AI-powered search product on Monday, strategically timed just before Google’s annual I/O conference.
  • This new AI search tool from OpenAI may be integrated with its ChatGPT chatbot and feature enhancements such as internet search capabilities and possibly Bing integration, building on the foundation laid by previous versions of ChatGPT.
  • Amidst Google’s ongoing challenges, including an antitrust case and competition from other AI startups like Perplexity, OpenAI’s move is seen as a significant threat to Google’s longstanding dominance in the search industry.
  • Source

ElevenLabs unveils new AI music generator

    • ElevenLabs, a voice AI startup, has debuted a new model that creates songs from user prompts, leveraging its expertise in voice cloning and synthesis.
    • The company is showcasing this new capability by transforming user-generated ideas into music, akin to Microsoft’s strategy with its AI chatbot Copilot, and is exploring its application in areas like lullabies, jingles, and podcast intros.
    • While the AI music generation sector is growing, with ElevenLabs planning a marketplace for AI-generated songs, there are rising concerns about copyright infringement and the impact on artists’ jobs.

US lawmakers unveil bill to make it easier to restrict exports of AI models

  • US lawmakers introduced a bipartisan bill aimed at making it easier for the Biden administration to impose export controls on AI models to protect against foreign threats.
  • The proposed legislation would allow the Commerce Department to prevent Americans from collaborating with foreign entities on AI developments that could endanger U.S. national security.
  • The bill seeks to overcome legal obstacles that currently make it difficult to regulate the export of open source AI models, amidst concerns over their misuse for cyber attacks or creating biological weapons.

AI demand soars in the workplace

Microsoft and LinkedIn have published their ‘2024 Work Trend Index Annual Report’, revealing the rapid adoption of AI tools by employees, with 75% of knowledge workers using AI and nearly half starting within the last six months.

AI demand soars in the workplace
AI demand soars in the workplace

Here are the key points:

  • 78% of AI users are bringing their own AI tools to work, with 52% reluctant to use them for their important tasks. 
  • 66% of leaders say they wouldn’t hire someone without AI skills, and 71% prefer less experienced candidates with AI skills over more experienced ones without.
  • Power users who use AI extensively are reaping benefits in productivity, creativity, and job satisfaction.
  • Skills are projected to change by 68% by 2030, accelerated by generative AI.

Why does this matter?

The study serves as a wake-up call for organizations to move beyond experimentation and develop comprehensive strategies for AI implementation. As AI permeates all sectors, generations, and skill sets, early adopters will likely emerge as the leaders, while those hesitant to explore could risk falling behind. 

Source

Stability AI introduces AI bot for Discord users 

Stability AI has launched Stable Artisan, a multimodal-gen AI Discord bot that enables users to create images and videos using the Stable Diffusion 3 (SD3) and Stable Video Diffusion (SVD) models. 

https://youtu.be/MWfb30kWqTM?si=_TePwQX1A8xEj3hU

Stable Artisan incorporates several editing and customization features, including Search and Replace, Remove Background, Creative Upscale, Outpaint, Control Sketch, and Control Structure. The service is available through a paid subscription, with monthly plans ranging from $9 to $99, and a 3-day free trial. 

Stability AI is also working on a larger conversational chatbot called Stable Assistant, which will incorporate the company’s text-to-image and LLM technologies to assist users with various tasks through natural language conversations. While Stable Artisan currently does not include access to Stable Audio, Stable Code, or Stable LM, these features may be added in the future as the service continues to evolve.

Why does this matter?

Stable Artisan could empower creators lacking experience with complex AI models to generate high-quality content directly within their familiar Discord environment.

However, the paid subscription model could limit access, and the missing features hint at a future with a complete AI creative suite. 

https://stability.ai/news/stable-artisan

ElevenLabs debuts AI model for generating lyrics

ElevenLabs, a company that specializes in AI-powered voice cloning and synthesis, has revealed a new model that creates song lyrics based on user prompts. 

With this new model, ElevenLabs aims to impact the music industry by allowing users to generate custom lullabies, jingles, podcast intros, and potentially even popular songs. The company also plans to launch a marketplace where users can sell their AI-generated music.

While ElevenLabs has not yet shared details about the maximum length of songs the AI can generate, an example posted by the company’s Head of Design suggests that it will likely produce lyrics for a standard three-minute song. 

Why does this matter?

This AI tool has the potential to democratize songwriting, allowing even those without musical expertise to craft lyrics. This could be particularly impactful for budget-conscious creators or those with specific lyrical needs. However, it remains to be seen if it will integrate with composing melodies like Udio or Suno, which offer a more complete song creation experience.

Also, the AI lyrics’ originality and the tool’s training data on copyrighted music might raise uncertainties. 

Source

What Else Is Happening in AI on May 10th 2024

TikTok introduces “AI-generated” labels for third-party content

TikTok will automatically label AI-generated content on its platform and on third-party platforms, becoming the first social media platform to support Content Credentials metadata for AI transparency. (Link)

Anthropic CEO defends dual funding from Google and Amazon

Anthropic’s CEO says partnering with Google and Amazon ensures more independence than OpenAI’s Microsoft reliance. However, regulators are examining the impact on AI competition as Anthropic’s future training costs could reach $100 billion. (Link)

Krea AI introduces video generation for paid subscribers

Krea AI, a generative AI startup, has launched video generation capabilities for its highest-tier subscribers. The new feature allows users to create videos using a combination of key frame images and text prompts, with a timeline-based interface reminiscent of traditional video editing software. (Link

Synthflow launches AI voice agent teams for streamlined customer support 

Synthflow launches “Conversational AI Teams,” a feature that allows businesses to create multiple AI voice assistants to interact with customers and each other, all through a single phone number. These intelligent agents can handle tasks like scheduling, updating CRMs, and more, providing a seamless and efficient customer support experience. (Link)

AI TRAINING: Add shortcuts using ChatGPT’s Memory

AI TRAINING: Add shortcuts using ChatGPT’s Memory
AI TRAINING: Add shortcuts using ChatGPT’s Memory

A lesser-known feature of ChatGPT’s new Memory feature is that it can be programmed to store shortcuts, which can save you a lot of time in chat conversations when used effectively.

  • Go to ChatGPT’s website. You’ll need to have a Plus subscription to access Memory.
  • Click on your account name (bottom-left), select ‘Settings’, ‘Personalization’, and enable ‘Memory’.
  • Teach ChatGPT a list of shortcuts by adding: “Store in your memory that when I start a prompt using: ‘!si’ means that I want you to search on the Internet; ‘!ci’ means that I want you to create an image; and !da’ means that I want you to use your Advanced Data Analysis feature
  • That’s it! Try the shortcuts in your conversations and let us know how much time you save

AI RESEARCH: Llama-3 goes head-to-head with GPT-4

AI RESEARCH: Llama-3 goes head-to-head with GPT-4
AI RESEARCH: Llama-3 goes head-to-head with GPT-4

Lmsys just published a new deep dive into its Chatbot Arena data, comparing Meta’s open-source Llama 3 70B model against top models like GPT-4 — revealing surprising strengths in the open-source leader.

  • Llama-3 is the top open-source model on the Lmsys Leaderboard, featuring over 70,000 votes from users in Arena Battle testing.
  • Meta’s model shines in battles involving brainstorming and writing prompts, falling short in math, coding, and translation compared to top competitors.
  • Lmsys noted Llama-3’s tone is friendlier and more conversational than rivals, which the model exhibits in its victorious battles in the Arena.

Llama-3’s ability to compete with top-tier models in certain areas is a testament to the rapid progress of open-source — and that’s with Meta’s largest model still pending. The more granular comparison also provides useful details often lost in more general model benchmarking.

New AI Job Opportunities on May 10th 2024

A Daily chronicle of AI Innovations May 09th 2024: 💥 OpenAI to unveil Google competitor on Monday 🎵 ElevenLabs unveils new AI music generator 🇺🇸 US lawmakers unveil bill to make it easier to restrict exports of AI models  📈 AI demand soars in the workplace 🤖 Stability AI introduces AI bot for Discord users 🎤 ElevenLabs debuts AI model for generating lyrics

OpenAI to unveil Google competitor on Monday

OpenAI to unveil Google competitor on Monday
OpenAI to unveil Google competitor on Monday
  • OpenAI is reportedly set to launch an AI-powered search product on Monday, strategically timed just before Google’s annual I/O conference.
  • This new AI search tool from OpenAI may be integrated with its ChatGPT chatbot and feature enhancements such as internet search capabilities and possibly Bing integration, building on the foundation laid by previous versions of ChatGPT.
  • Amidst Google’s ongoing challenges, including an antitrust case and competition from other AI startups like Perplexity, OpenAI’s move is seen as a significant threat to Google’s longstanding dominance in the search industry.
  • Source

ElevenLabs unveils new AI music generator

    • ElevenLabs, a voice AI startup, has debuted a new model that creates songs from user prompts, leveraging its expertise in voice cloning and synthesis.
    • The company is showcasing this new capability by transforming user-generated ideas into music, akin to Microsoft’s strategy with its AI chatbot Copilot, and is exploring its application in areas like lullabies, jingles, and podcast intros.
    • While the AI music generation sector is growing, with ElevenLabs planning a marketplace for AI-generated songs, there are rising concerns about copyright infringement and the impact on artists’ jobs.

US lawmakers unveil bill to make it easier to restrict exports of AI models

  • US lawmakers introduced a bipartisan bill aimed at making it easier for the Biden administration to impose export controls on AI models to protect against foreign threats.
  • The proposed legislation would allow the Commerce Department to prevent Americans from collaborating with foreign entities on AI developments that could endanger U.S. national security.
  • The bill seeks to overcome legal obstacles that currently make it difficult to regulate the export of open source AI models, amidst concerns over their misuse for cyber attacks or creating biological weapons.

AI demand soars in the workplace

Microsoft and LinkedIn have published their ‘2024 Work Trend Index Annual Report’, revealing the rapid adoption of AI tools by employees, with 75% of knowledge workers using AI and nearly half starting within the last six months.

Here are the key points:

  • 78% of AI users are bringing their own AI tools to work, with 52% reluctant to use them for their important tasks. 
  • 66% of leaders say they wouldn’t hire someone without AI skills, and 71% prefer less experienced candidates with AI skills over more experienced ones without.
  • Power users who use AI extensively are reaping benefits in productivity, creativity, and job satisfaction.
  • Skills are projected to change by 68% by 2030, accelerated by generative AI.

Why does this matter?

The study serves as a wake-up call for organizations to move beyond experimentation and develop comprehensive strategies for AI implementation. As AI permeates all sectors, generations, and skill sets, early adopters will likely emerge as the leaders, while those hesitant to explore could risk falling behind. 

Source

Stability AI introduces AI bot for Discord users 

Stability AI has launched Stable Artisan, a multimodal-gen AI Discord bot that enables users to create images and videos using the Stable Diffusion 3 (SD3) and Stable Video Diffusion (SVD) models. 

https://youtu.be/MWfb30kWqTM?si=_TePwQX1A8xEj3hU

Stable Artisan incorporates several editing and customization features, including Search and Replace, Remove Background, Creative Upscale, Outpaint, Control Sketch, and Control Structure. The service is available through a paid subscription, with monthly plans ranging from $9 to $99, and a 3-day free trial. 

Stability AI is also working on a larger conversational chatbot called Stable Assistant, which will incorporate the company’s text-to-image and LLM technologies to assist users with various tasks through natural language conversations. While Stable Artisan currently does not include access to Stable Audio, Stable Code, or Stable LM, these features may be added in the future as the service continues to evolve.

Why does this matter?

Stable Artisan could empower creators lacking experience with complex AI models to generate high-quality content directly within their familiar Discord environment.

However, the paid subscription model could limit access, and the missing features hint at a future with a complete AI creative suite. 

https://stability.ai/news/stable-artisan

ElevenLabs debuts AI model for generating lyrics

ElevenLabs, a company that specializes in AI-powered voice cloning and synthesis, has revealed a new model that creates song lyrics based on user prompts. 

With this new model, ElevenLabs aims to impact the music industry by allowing users to generate custom lullabies, jingles, podcast intros, and potentially even popular songs. The company also plans to launch a marketplace where users can sell their AI-generated music.

While ElevenLabs has not yet shared details about the maximum length of songs the AI can generate, an example posted by the company’s Head of Design suggests that it will likely produce lyrics for a standard three-minute song. 

Why does this matter?

This AI tool has the potential to democratize songwriting, allowing even those without musical expertise to craft lyrics. This could be particularly impactful for budget-conscious creators or those with specific lyrical needs. However, it remains to be seen if it will integrate with composing melodies like Udio or Suno, which offer a more complete song creation experience.

Also, the AI lyrics’ originality and the tool’s training data on copyrighted music might raise uncertainties. 

Source

What Else Is Happening in AI on May 10th 2024

TikTok introduces “AI-generated” labels for third-party content

TikTok will automatically label AI-generated content on its platform and on third-party platforms, becoming the first social media platform to support Content Credentials metadata for AI transparency. (Link)

Anthropic CEO defends dual funding from Google and Amazon

Anthropic’s CEO says partnering with Google and Amazon ensures more independence than OpenAI’s Microsoft reliance. However, regulators are examining the impact on AI competition as Anthropic’s future training costs could reach $100 billion. (Link)

Krea AI introduces video generation for paid subscribers

Krea AI, a generative AI startup, has launched video generation capabilities for its highest-tier subscribers. The new feature allows users to create videos using a combination of key frame images and text prompts, with a timeline-based interface reminiscent of traditional video editing software. (Link

Synthflow launches AI voice agent teams for streamlined customer support 

Synthflow launches “Conversational AI Teams,” a feature that allows businesses to create multiple AI voice assistants to interact with customers and each other, all through a single phone number. These intelligent agents can handle tasks like scheduling, updating CRMs, and more, providing a seamless and efficient customer support experience. (Link)

AI TRAINING: Add shortcuts using ChatGPT’s Memory

AI TRAINING: Add shortcuts using ChatGPT’s Memory
AI TRAINING: Add shortcuts using ChatGPT’s Memory

A lesser-known feature of ChatGPT’s new Memory feature is that it can be programmed to store shortcuts, which can save you a lot of time in chat conversations when used effectively.

  • Go to ChatGPT’s website. You’ll need to have a Plus subscription to access Memory.
  • Click on your account name (bottom-left), select ‘Settings’, ‘Personalization’, and enable ‘Memory’.
  • Teach ChatGPT a list of shortcuts by adding: “Store in your memory that when I start a prompt using: ‘!si’ means that I want you to search on the Internet; ‘!ci’ means that I want you to create an image; and !da’ means that I want you to use your Advanced Data Analysis feature
  • That’s it! Try the shortcuts in your conversations and let us know how much time you save

AI RESEARCH: Llama-3 goes head-to-head with GPT-4

AI RESEARCH: Llama-3 goes head-to-head with GPT-4
AI RESEARCH: Llama-3 goes head-to-head with GPT-4

Lmsys just published a new deep dive into its Chatbot Arena data, comparing Meta’s open-source Llama 3 70B model against top models like GPT-4 — revealing surprising strengths in the open-source leader.

  • Llama-3 is the top open-source model on the Lmsys Leaderboard, featuring over 70,000 votes from users in Arena Battle testing.
  • Meta’s model shines in battles involving brainstorming and writing prompts, falling short in math, coding, and translation compared to top competitors.
  • Lmsys noted Llama-3’s tone is friendlier and more conversational than rivals, which the model exhibits in its victorious battles in the Arena.

Llama-3’s ability to compete with top-tier models in certain areas is a testament to the rapid progress of open-source — and that’s with Meta’s largest model still pending. The more granular comparison also provides useful details often lost in more general model benchmarking.

New AI Job Opportunities on May 10th 2024

A Daily chronicle of AI Innovations May 09th 2024: 🤖 OpenAI posts Model Spec revealing how it wants AI to behave 🧬 Google DeepMind unveils AlphaFold 3, the next generation of its protein prediction model 🧠 Neuralink faces setback as first human brain implant encounters problem 🕵️‍♀️  Microsoft developed a secretive AI service for US spies 🎨 Generate images on Midjourney Alpha 📝Copilot for Microsoft 365 to get auto-complete and rewrite to improve prompts 🏢New AI data center to be built at the failed Foxconn project site in Wisconsin 🤔Sam Altman says we are not taking AI’s impact on the economy seriously ✒️Typeface Arc replaces prompts; uses AI agent approach to ease marketing workflows 🎮Altera’s gaming AI agents get backed by Eric Schmidt, Former Google CEO

🤖 OpenAI posts Model Spec revealing how it wants AI to behave

OpenAI posts Model Spec revealing how it wants AI to behave
OpenAI posts Model Spec revealing how it wants AI to behave
OpenAI posts Model Spec revealing how it wants AI to behave
OpenAI posts Model Spec revealing how it wants AI to behave

 

  • OpenAI has introduced the first draft of Model Spec, a proposed framework aiming to shape how AI models respond, emphasizing assistance, humanity’s benefit, and adherence to social norms and laws.
  • The framework suggests specific rules for AI behavior, including compliance with laws, protection of privacy, and avoidance of NSFW content, with options to adjust settings like allowing NSFW content in certain contexts.
  • While the Model Spec seeks public feedback for future adjustments and doesn’t immediately affect existing models like GPT-4 or DALL-E 3, it’s envisioned as a living document to guide AI behavior improvement over time.
  • Source

🧬 Google DeepMind unveils AlphaFold 3, the next generation of its protein prediction model

  • Google DeepMind and Isomorphic Labs have released AlphaFold 3, a new AI model for predicting protein structures, including their interactions with various molecules such as DNA, RNA, and small molecules, thereby enhancing drug discovery possibilities.
  • This new version is more precise in mapping out complex groupings of molecules, significantly enhancing our ability to understand and predict molecular behavior compared to its earlier version.
  • Google will not open-source this version but has launched AlphaFold Server for non-commercial research use, aiming to balance intellectual property concerns with accessibility for scientific progress.
  • Source

🧠 Neuralink faces setback as first human brain implant encounters problem 

  • Neuralink admitted that some of the micro-thin threads from their N1 brain chip retracted after implantation in the first human patient, possibly due to air trapped in the skull during surgery, which affected the device’s data transmission rate.
  • Despite the retraction of several threads, Neuralink managed to increase the data transmission speed over time by optimizing their recording algorithm and improving signal translation into cursor movements.
  • The company is planning further implants, with goals to implant two more patients in the coming months and ten in total this year, while continuing to refine their technology and reporting developments to the FDA.
  • Source

🕵️‍♀️ Microsoft developed a secretive AI service for US spies

Microsoft has developed a top-secret generative AI model entirely disconnected from the internet so US intelligence agencies can safely harness the powerful technology to analyze top-secret info. The model based on GPT-4 is now live, answering questions, and will also write code.

Microsoft spent 18 months developing the model, which is “air-gapped” to ensure it is secure. This is the first time a model is fully isolated– meaning it’s not connected to the internet but is on a special network that’s only accessible by the U.S. government.

It can read and analyze files but cannot learn from them to stop sensitive information from entering the platform. It is yet to be tested and accredited by the intelligence agencies.

Why does this matter?

Intelligence agencies all over the world have been racing to be the first to harness generative AI. I guess we know who’s going to be the winner. If this AI tool is successful, it will fundamentally change the way intelligence agencies operate.

Source

What Else Is Happening in AI on May 09th 2024❗

📝Copilot for Microsoft 365 to get auto-complete and rewrite to improve prompts

In coming months, Microsoft Copilot will be updated with new features like auto-complete and ‘elaborate your prompt’ that offer suggestions to improve AI prompts. It aims to solve the problem of coming up with good prompts for generative AI. (Link)

🏢New AI data center to be built at the failed Foxconn project site in Wisconsin

President Joe Biden announced an AI data center to be built on the same site as the failed Foxconn project in Racine, Wisconsin. According to a White House press release, Microsoft is investing $3.3B in the project, creating up to 2,000 permanent jobs. (Link)

🤔Sam Altman says we are not taking AI’s impact on the economy seriously

At a Brooking’s Institute panel about AI and geopolitics on Tuesday, Altman said the discussions around AI’s effect on the economy–  like how it may lead to mass job replacement– died down this year compared to last. He said if we don’t take these concerns seriously enough going forward, it could be a massive issue. (Link)

✒️Typeface Arc replaces prompts; uses AI agent approach to ease marketing workflows

It is launching Typeface Arc technology, which enables a user to state a high-level marketing objective and then have the system automatically plan and generate all the assets, including emails, images, and notifications that are all connected. (Link)

🎮Altera’s gaming AI agents get backed by Eric Schmidt, Former Google CEO

Altera is the newest startup joining the fray to build a new guard of AI agents. It raised $9 million in an oversubscribed seed round, co-led by Eric Schmidt’s deep-tech fund, First Spark Ventures and Patron, the seed-stage fund co-founded by Riot Games alums. (Link)

AI TRAINING May 09th 2024

🎨 Generate images on Midjourney Alpha

Generate images on Midjourney Alpha
Generate images on Midjourney Alpha

Midjourney’s website is now accessible to anyone with more than 100 generated images, improving the experience when prompting images over its standard Discord group.

  1. Check that you’ve generated more than 100 images by typing /info in the Midjourney Discord group. If you have, head over to Midjourney Alpha.
  2. In the main menu, you can explore other creations and search prompts.
  3. Select where it says “imagine” and enter your prompt to generate an image.
  4. Add a reference image by selecting “+” or play with different parameters such as image size, stylization, or even weirdness by pressing the “slider control” button

AI RESEARCH on May 09th 2024

📶 AI usage surges in the workplace

AI usage surges in the workplace
AI usage surges in the workplace

Microsoft and LinkedIn just published their Work Trend Index Annual Report, revealing that AI adoption is surging in the workplace — calling 2024 the ‘year AI at work gets real’.

  • The report found that use of GenAI has doubled in the last six months, with 75% of knowledge workers using the tech in some capacity.
  • 78% of AI users are bringing their own AI to work — with 52% reporting they are reluctant to admit to its use.
  • 66% of leaders wouldn’t hire someone without AI skills, and 71% prefer less experienced candidates with AI aptitude over a more experienced one without it.
  • AI power users reported enhanced productivity, creativity, and job satisfaction compared to skeptical peers.

Why it matters: Employees are adopting AI at a rapid pace, regardless of if their own organizations are ready for the shift. As AI spreads across all sectors, generations, and skillsets, the early adopters are rising to the top — while those that aren’t at least exploring the tech are quickly running out of time

Trending AI Tools May 09th 2024

  • 📍GeoSpy – Uncover photo locations with AI
  • 🧑‍💻 LangChain – Connect LLMs to private data for context-aware applications
  • 📊 Abstra – Scale business processes with Python and AI
  • 🎨 Freepik Pikaso Upscaler – Integrated with Magnific, enlarge images without losing quality
  • 💬 Notion AI Q&A – Q&A is now open to the public, allowing users to ask and find information across their workspace
  • 🎵 Udio Audio Inpainting – Select a portion of an AI-generated music track and regenerate it

New AI Job Opportunities on May 09th 2024

  • 🎥 The Rundown – Video Content Creator
  • 🤖 Anthropic – Research Engineer, Human-Computer Interfaces
  • 👩‍💻 Adept AI – Solutions Engineer
  • 📝 Mistral AI – Data Annotation Technical Program Manager

A Daily chronicle of AI Innovations May 08th 2024: 👨‍⚕️ Google built medical AI that can train itself and outperform real doctors 🕵️‍♂️ Microsoft launches AI chatbot for spies 🖼️ OpenAI’s new tool detects 98% DALL-E 3 generated images 📣 Meta expands AI-powered creativity tools for advertisers 🎬 OpenAI’s ‘Media Manager’ will let creators opt out of AI training 🍎 Apple releases M4 chip at the ‘Let Loose’ event with powerful AI capabilities 📰 OpenAI strikes licensing deal with People magazine publisher 🤖 Amazon announces Bedrock Studio to simplify Gen AI app development 👨‍💻 Oracle introduces Code Assistant to accelerate enterprise software development 🚀 Red Hat launches RHEL AI and InstructLab to democratize enterprise AI 🍏 Apple unveils iPad AI upgrades

👨‍⚕️ Google built medical AI that can train itself and outperform real doctors

  • Google’s DeepMind has developed Med-Gemini, a self-training medical AI that outperforms doctors in accuracy, achieving a 91.1% score on medical benchmarks.
  • Med-Gemini uses an uncertainty-guided search functionality and has bested GPT-4 models and actual doctors on 10 out of 14 medical benchmarks.
  • Despite its capabilities, Google’s medical AI aims to assist rather than replace doctors, helping to alleviate their workload.
  • Source

🕵️‍♂️ Microsoft launches AI chatbot for spies

  • Microsoft has launched a GPT-4-based generative AI model tailored for US intelligence agencies that operates offline to analyze top-secret information securely.
  • The AI chatbot aims to facilitate secure conversations among spy agencies without the internet, addressing data breach and hacking concerns.
  • This initiative represents Microsoft’s first deployment of a major language model in a high-security environment, with a focus on mitigating the risks of data leaks while processing classified data
  • Source

🖼️ OpenAI’s new tool detects 98% DALL-E 3 generated images

OpenAI has developed a new tool to detect if an image was created by DALL-E 3, its AI image generator. The tool can detect DALL-E 3 images with around 98% accuracy, even if the image has been cropped, compressed, or had its saturation changed. However, the tool is not as effective at detecting images generated by other AI models, only flagging 5-10% of images.

OpenAI’s new tool detects 98% DALL-E 3 generated images

This image detection classifier is only available to a group of testers, including research labs and research-oriented journalism nonprofits through OpenAI’s Research Access Program.

OpenAI has also added watermarking to Voice Engine, its text-to-speech platform, which is currently in limited research preview.

Why does it matter?

Early experiences have shown that AI detectors don’t work. In fact, if they have high error rates, they could lead to false accusations. In 2023, OpenAI had to shut down its own AI detection software for text because of its poor accuracy.

So, if this detector is as good as OpenAI claims, we may be on the precipice of a revolutionary new capability to reliably detect AI-generated content, with huge implications across domains.

Source

📣 Meta expands AI-powered creativity tools for advertisers

Meta has expanded its generative AI tools for advertisers. Advertisers can request AI to generate entirely new images, including product variations in different colors, angles, and scenarios. The AI tools can add text overlays with different fonts, expand images to fit different aspect ratios like Reels and Feed, and generate ad headlines that match the brand’s voice.

Meta expands AI-powered creativity tools for advertisers
Meta expands AI-powered creativity tools for advertisers

The AI features will roll out globally to advertisers by the end of 2024.

Meta is also expanding its paid Meta Verified service for businesses to more countries. Different pricing tiers offer features like account support, profile enhancements, and better customer service access.

Why does it matter?

Integrating such powerful AI features could lead to more effective advertising campaigns and improved customer engagement with targeted marketing and personalized ads. However, it could also raise questions about transparency and potential misuse.

Source

🎬 OpenAI’s ‘Media Manager’ will let creators opt out of AI training

OpenAI is developing Media Manager, a tool that will enable creators and content owners to decide what they own and specify how they want their works to be included or excluded from machine learning research and training. This first-ever tool of its kind will help OpenAI identify copyrighted text, images, audio, and video across multiple sources and reflect creator preferences.

OpenAI aims to  have the tool in place by 2025 and set a standard across the AI industry with it.

Why does it matter?

Media Manager seems to be OpenAI’s response to growing criticism of its approach to developing AI models, which heavily scraps publicly available data from the web for training. Recently, 8 prominent U.S. newspapers sued OpenAI for copyright infringement.

On the other hand, OpenAI has formed mutually beneficial partnerships with platforms like Stack Overflow, Shutterstock, The Financial Times, and more to use their content.

So, OpenAI may be trying to meet creators in the middle, but if it is positioning itself as a fully ethical actor with this, we’ll take it with a grain of salt.

Source

🍏 Apple unveils iPad AI upgrades

Apple unveils iPad AI upgrades

Apple just revealed its new line of iPads at a company event in Cupertino, CA — featuring a custom M4 chip that enables advanced AI capabilities and a slew of new AI-powered features.

  • The bigger iPad Pro now features the M4 chip with an upgraded Neural Engine, which CEO Tim Cook calls “an outrageously powerful chip for AI”.
  • The M4 is capable of handling 38T operations per second, 4x the performance of previous models, allowing for the running of advanced AI.
  • New AI features on the Pro include a True Tone Flash for document scanning and new video, image, and music editing tools.
  • Prior to the event, a new report revealed that Apple is developing its own AI chips for data centers under the code name “Project ACDC”.

Source

What Else Is Happening in AI on May 08th 2024❗

🍎 Apple releases M4 chip at the ‘Let Loose’ event with powerful AI capabilities

Apple released its much-anticipated M4 chip at the “Let Loose” event. M4 is slated to spearhead Apple’s next generation of devices, with the iPad Pro leading the charge and powering the forthcoming OLED iPad Pro which is meticulously engineered to elevate the user experience to unprecedented heights. (Link)

📰 OpenAI strikes licensing deal with People magazine publisher

OpenAI has inked a licensing deal with Dotdash Meredith to bring the People magazine publisher’s content to ChatGPT and help train it’s AI models. Under the partnership, OpenAI will be able to display lifestyle and entertainment content in its chatbot from the many websites of one of the US’s largest digital and print publishers. (Link)

🤖 Amazon announces Bedrock Studio to simplify Gen AI app development 

Amazon is launching a new tool, Bedrock Studio, designed to let organizations experiment with generative AI models, collaborate on those models, and ultimately build generative AI-powered apps. Bedrock Studio is a “rapid prototyping environment” for generative AI. It also guides developers in evaluating, analyzing, fine-tuning, and sharing generative AI models. (Link)

👨‍💻 Oracle introduces Code Assitant to accelerate enterprise software development

Oracle has announced Code Assitant, an AI-powered service to help developers rapidly program apps based on Java, SQL, and the Oracle Cloud infrastructure. It will join tools like GitHub Copilot and Amazon CodeWhisperer to accelerate the app development lifecycle. However, Oracle hasn’t yet specified when this feature will be released. (Link)

🚀 Red Hat launches RHEL AI and InstructLab to democratize enterprise AI

At Red Hat Summit 2024, RedHat announced two major initiatives to bring the power of generative AI to the enterprise. Red Hat Enterprise Linux AI (RHEL AI), a foundation model platform for developing and running open-source language models, and InstructLab, a community project to enable domain experts to enhance AI models with their knowledge. (Link)

AI TRAINING: 🎥 Chat with YouTube videos using Gemini

Chat with YouTube videos using Gemini

Google Gemini’s new “Extensions” feature allows users to access external tools such as YouTube to chat with videos and get answers for free.

Step-by-step:

  1. Visit Google’s Gemini website. If Gemini is not available in your country, you’ll need to use a US-based VPN.
  2. Click on the gear icon located on the bottom-left, select Extensions, and turn on the YouTube one.
  3. Go back to the Chat interface and start your prompt using the following format: “@youtube Summarize the following video [Youtube URL]”

Pro tip: Try asking Gemini to explain advanced concepts discussed in a video, generating concrete examples, creating practice questions, and even asking for code snippets

Source

Trending AI Tools on May 08th 2024:

  • 🎨 Eraser AI – Technical design copilot that helps users edit documents and generate diagrams easily
  • 📊 Snaplet Seed – Cease manual data creation, get instant realistic data for your project
  • 🌐 Aria Browser AI – Opera’s browser AI can now create images and read out text responses
  • ⌛ Meaning – Block distracting apps with an AI screen time coach
  • 🤖 Dante Unlimited AI Messages – No more hidden fees for AI chatbot creation

New AI Job Opportunities May 08th 2024

A Daily chronicle of AI Innovations May 07th 2024: 🤖 Apple is developing its own AI chip for data center servers 🤝 Stack Overflow and OpenAI have announced an API partnership🌟 Microsoft is developing a new AI language model to rival GPT-4 and Google Gemini 📚 Study: Teachers can’t detect AI writing 🔮 Apple’s new M4 chip is focused on AI  🤖 Hugging Face has launched LeRobot, an open-source robotics toolkit 📸 Apple is testing a new “Clean Up” feature in its Photos app 🛡️ Google has launched Google Threat Intelligence 🇺🇸 US invests $285M in AI ‘Digital Twin’ technology  📡 Anduril Industries introduces Pulsar: AI modular electromagnetic warfare (EW) systems

🔮 Apple’s new M4 chip is focused on AI

  • Apple’s new M4 chip, designed for the latest OLED iPad Pro models, focuses on enhancing AI performance with capabilities like handling 38 trillion operations per second, which is significantly higher than its predecessors but still behind Qualcomm’s Snapdragon X Elite.
  • The M4 chip utilizes advanced second-generation 3nm technology, featuring a 10-core CPU with a mix of performance and efficiency cores, and supports technologies like mesh shading and ray tracing for improved graphical rendering.
  • Apple’s announcement of the M4 chip sets the stage for a comparison with Microsoft’s upcoming Arm-powered hardware and Windows on Arm enhancements, scheduled for release later this month, which aim to surpass Apple’s MacBook Air M3 in terms of CPU and AI performance.

Apple is developing its own AI chip for data center servers

Apple is developing its own AI chip for data center servers, known internally as Project ACDC (Apple Chips in Data Center). The chip will likely focus on running AI models (inference) rather than training them, which is where Nvidia currently dominates.

The company is working closely with TSMC (Taiwan Semiconductor Manufacturing Co) to design and produce these chips, although the timeline for launch is uncertain. With this move, the company aims to keep up with rivals like Microsoft and Meta, who have made significant investments in generative AI.

Why does it matter?

Apple has a long history of designing custom chips for its devices like iPhones, iPads, and Macs, which is probably what makes them stand out. Having custom AI chips could allow the tech giant more control over its “AI destiny” versus relying on suppliers like Nvidia.

Source

Stack Overflow and OpenAI have announced an API partnership

OpenAI will use OverflowAPI to improve model performance and provide attribution to the Stack Overflow community within ChatGPT. Stack Overflow will use OpenAI models to develop OverflowAI and to maximize model performance.

The partnership aims to improve the user and developer experience on both platforms. The first set of integrations and capabilities will be available in the first half of 2024, and the partnership will enable Stack Overflow to reinvest in community-driven features.

Why does this matter?

Stack Overflow partnered with Google Cloud to develop Overflow API and to give Google’s Gemini models access to its knowledge communities. Now it is forming a similar partnership with OpenAI. Despite concerns about copyright breaches, such partnerships seem to be trending where both the parties have much to gain, but it just reaffirms that the big AI players remain hungry for data.

Source

Microsoft is developing a new AI language model

Microsoft is developing a new, large-scale AI language model called MAI-1 to compete with Google and OpenAI. The model is overseen by Mustafa Suleyman, recently hired co-founder of Google DeepMind.

MAI-1 will be larger and more expensive than Microsoft’s previous smaller, open-source models, with roughly 500 billion parameters. Microsoft could preview the new model as soon as its Build developer conference later this month.

Why does this matter?

Microsoft’s development of MAI-1 shows that it is not entirely relying on it’s OpenAI investment to go big in AI. Now, it has entered the AI race truly, competing with state-of-the-art models from Google, Anthropic, even Meta’s Llama 400B which is in training, and OpenAI itself.

Source

What Else Is Happening in AI on May 07th 2024❗

🤖 Hugging Face has launched LeRobot, an open-source robotics toolkit

It is a comprehensive platform for developers, researchers, and hobbyists to train AI models, share data, and simulate environments, all while seamlessly integrating with various robotic hardware. The toolkit offers pre-trained models and integrates with physics simulators for testing without physical robots. Hugging Face is also collaborating with diverse partners to build the largest crowdsourced robotics dataset. (Link)

📸 Apple is testing a new “Clean Up” feature in its Photos app

By using gen AI for advanced image editing, this feature will allow you to effortlessly remove unwanted objects from your photos using a simple brush. Apple may preview this new feature during its upcoming “Let Loose” iPad event or at WWDC in June. (Link)

🛡️ Google has launched Google Threat Intelligence 

Google has launched Google Threat Intelligence
Google has launched Google Threat Intelligence

It is a combination of Mandiant’s expertise, VirusTotal’s community insights, and Google’s vast threat visibility. Google Threat Intelligence assists with external threat monitoring, attack surface management, digital risk protection, IoC analysis, and expertise. With Gemini, organizations can now quickly search through vast amounts of threat data to protect against cyber threats. (Link)

🇺🇸 US invests $285M in AI ‘Digital Twin’ technology 

The Biden administration is investing $285 million for a new “CHIPS Manufacturing USA institute” focused on digital twins for the semiconductor industry. This approach uses AI to create virtual chip replicas, accelerating the production of next-gen processors. Intel and Micron are also set to receive funding to boost the development of new processors. (Link)

📡 Anduril Industries introduces Pulsar: AI modular electromagnetic warfare (EW) systems 

Pulsar uses AI to quickly identify and counter current and future threats across the electromagnetic spectrum, including small and medium-size drones. With its integration of software-defined radio, GPUs, and diverse compute capabilities, Pulsar is changing how we defend against rapidly evolving threats in an increasingly complex battlefield. (Link)

AI TRAINING May 07th 2024

🎙️ Studio-quality audio in seconds with AI

Studio-quality audio in seconds with AI
Studio-quality audio in seconds with AI

Adobe’s AI-powered ‘Enhance Speech’ tool dramatically improves the quality of audio voice recordings with just a few clicks.

Step-by-step:

  1. Access Adobe’s Enhance Speech website. Sign in or sign up for free.
  2. Upload your audio by selecting “Choose file” or simply drag and drop the file.
  3. Wait a few seconds while Adobe’s AI analyzes and removes unwanted noises from your file.
  4. Download your studio-quality audio and hear the difference!

Pro tip: If you have a video file, you can extract the audio using free websites that extract audio from video and add the enhanced audio back to your video using free video editors like CapCut

AI RESEARCH on May 07th 2024

📚 Study: Teachers can’t detect AI writing

Study: Teachers can’t detect AI writing
Study: Teachers can’t detect AI writing

A series of studies from several German universities found that both novice and experienced teachers struggled to accurately distinguish between student-written and AI-generated texts.

The details:

  • The first study trialed pre-service teachers on AI and student text, correctly ID’ing only 45.1% of AI-generated submissions.
  • The second study tested experienced teachers, who correctly identified just 37.8% of AI-generated submissions.
  • The study revealed that teachers were overconfident in their ability to spot AI-generated text.
  • AI detection software also often falsely flags student-written texts as potentially AI-generated, which further complicates the issue.

Why it matters: AI’s writing capabilities are only getting better — and relying on teacher intuition or unreliable tools may be no more effective than guessing. Unless better tools become available, it may be time to pivot to enabling students to work with AI instead of penalizing them for it.

Trending AI Tools May 07th 2024

  • 📊 Osum – Perform deep market research in seconds: Try now (Use checkout code RUNDOWN for  25% off)*
  • 💫 Coachvox – The ultimate AI marketing tool for coaches with content
  • ✍️ Pressmaster – All-in-one platform simplifies press planning and distribution
  • 🧰 Lamini – Enterprises LLM platform that you can own
  • 🚀 Postman v11 – Supercharging development with AI and APIs
  • 🤖 Adobe Acrobat AI Assistant – Get quick answers and one-click summaries from PDFs
  • 🎨 Leonardo AI Content Reference – Enhance creative control for image generations

New AI Job Opportunities May 07th 2024

  • 🤝 The Rundown – Community Manager
  • 📈 Scale AI – Senior Analytics Engineer
  • ⚖ Anthropic – Risk Manager, Responsible Scaling Team
  • 💻 OpenAI – Tech Lead Manager – Supercomputing Scheduling

A Daily chronicle of AI Innovations May 06th 2024: 🤖 DrEureka can automate robot training using LLMs 🚀 Free AI model rivals GPT-4 in language model evaluation 📰 X introduces Stories feature powered by Grok AI 🔒 Privacy complaint filed against OpenAI 💰 JPMorgan launches an AI-powered tool for thematic investing ⏩ YouTube Premium introduces AI-powered “Jump ahead” feature 💊 AI is now set to transform the drug discovery industry 🎤 AI helps bring back Randy Travis’ voice in new song 💥 Jack Dorsey has left Bluesky, calls X a ‘freedom technology’ 🫠 Why Warren Buffet isn’t investing in AI🧠 Massive prompts can outperform fine-tuning for LLMs, researchers find

DrEureka can automate robot training using LLMs

In robotics, one of the biggest challenges is transferring skills learned in simulation to real-world environments.  NVIDIA researchers have developed a groundbreaking algorithm called DrEureka that uses LLMs to automate the design of reward functions and domain randomization parameters—key components in the sim-to-real transfer process.

The algorithm works in three stages: first, it creates reward functions with built-in safety instructions; then, it runs simulations to determine the best range of physics parameters; finally, it generates domain randomization configurations based on the data gathered in the previous stages.

When tested on various robots, including quadrupeds and dexterous manipulators, DrEureka-trained policies outperformed those designed by human experts.

Why does it matter?

DrEureka makes robot training accessible and cost-effective for businesses and researchers alike. We may witness increased adoption of robotics in industries that have previously been hesitant to invest in the technology due to the complexity and cost of training robots for real-world applications.

Source

Free AI model rivals GPT-4 in language model evaluation

Prometheus 2, a free and open-source language model developed by KAIST AI, has shown impressive capabilities in evaluating other language models, approaching the performance of commercial models like GPT-4.

The model was trained on a new pairwise comparison dataset called the “Preference Collection,” which includes over 1,000 evaluation criteria beyond basic characteristics. By combining two separate models – one for direct ratings and another for pairwise comparisons – the researchers achieved the best results.

Free AI model rivals GPT-4 in language model evaluation
Free AI model rivals GPT-4 in language model evaluation

In tests across eight datasets, Prometheus 2 showed the highest agreement with human judgments and commercial language models among all freely available rating models, significantly closing the gap with proprietary models.

Why does this matter?

By enabling user-defined evaluation criteria, Prometheus 2 can be tailored to assess language models based on specific preferences and real-life scenarios, opening up new possibilities for developing specialized AI applications across various domains. It’s also an opportunity to create niche models that are culturally sensitive and relevant.

Source

X introduces Stories feature powered by Grok AI

X (formerly Twitter) has launched a new feature, Stories, that provides AI-generated summaries of trending news on the platform. Powered by Elon Musk’s chatbot Grok, Stories offers Premium subscribers brief overviews of the most popular posts and conversations happening on X.

X introduces Stories feature powered by Grok AI
X introduces Stories feature powered by Grok AI

With Stories, users can quickly catch up on the day’s trending topics without having to scroll through countless posts. Grok generates these summaries based solely on the conversations happening on X about each news story rather than analyzing the original news articles themselves. While this approach is controversial, X believes it will pique users’ curiosity and potentially drive them deeper into the source material.

Why does this matter?

X’s Grok-powered Stories feature may reshape the way we consume news. As more platforms integrate AI news summarization tools, traditional media outlets may face challenges in maintaining reader engagement and revenue. However, the reliance on platform-specific conversations for generating summaries raises concerns about the potential spread of misinformation and the creation of echo chambers.

Source

🫠 Why Warren Buffet isn’t investing in AI

  • Warren Buffett predicts AI scamming will become a major growth industry, highlighting its capacity to generate realistic and misleading content to deceive individuals.
  • Buffett acknowledges AI’s potential for both beneficial and detrimental impacts, admitting his own lack of understanding of the technology.
  • Despite the hype around AI in the investment world, Buffett exercises caution, likening its unpredictable potential impact to that of nuclear weapons.
  • Source

🧠 Massive prompts can outperform fine-tuning for LLMs, researchers find

  • Researchers have discovered that using large prompts with numerous examples in LLMs can be more effective than traditional fine-tuning techniques.
  • “In-context learning” (ICL) becomes increasingly efficient with larger context windows, allowing for the effective use of hundreds to thousands of examples without the need for task-specific training.
  • The study indicates that while fine-tuning requires more initial data and resources, ICL with extensive prompts presents a viable and sometimes superior alternative, especially as models improve in processing long inputs.
  • Source

What Else Is Happening in AI on May 06th 2024❗

🔒 Privacy complaint filed against OpenAI

The maker of ChatGPT is facing a privacy complaint in the European Union (EU) for its “hallucination problem.” The complaint alleges violations of GDPR, including misinformation generation and lack of transparency on data sources. The report highlights concerns about accuracy, data access, and the inability of ChatGPT to correct incorrect information. (Link)

💰 JPMorgan launches an AI-powered tool for thematic investing

IndexGPT is a new range of thematic investment baskets created using OpenAI’s GPT-4 model. The tool generates keywords associated with a theme, which are then used to identify relevant companies through natural language processing of news articles. IndexGPT aims to improve the selection of stocks for thematic indexes, going beyond obvious choices and potentially enhancing trend-following strategies. (Link)

⏩ YouTube Premium introduces AI-powered “Jump ahead” feature

The AI-powered feature allows users to skip past commonly skipped sections of a video and jump to the next best point. It is currently available for the YouTube Android app in the US with English videos and can be enabled through the experiments page. (Link)

💊 AI is now set to transform the drug discovery industry

Generative AI is now rapidly generating novel molecules and proteins that humans may not have considered. AI models, such as Google’s AlphaFold, are accelerating the drug development process from years to months while increasing success rates. Experts predict that AI-designed drugs will become the norm in the near future, but they will still need to prove their efficacy in human trials. (Link)

🎤 AI helps bring back Randy Travis’ voice in new song

Country singer Randy Travis has released a new song, “Where That Came From,” his first since losing his voice to a stroke in 2013.

The vocals were created using AI software and a surrogate singer under the supervision of Travis and his producer. The result is a gentle tune that captures Travis’ relaxed style, reinforcing the potential of AI voice cloning in the right hands. (Link)

AI Tutorial on May 06th 2024:

🧠 How to use ChatGPT’s new ‘Memory’ feature

OpenAI has rolled out a new feature called “Memory” for ChatGPT plus users, enabling it to remember specific user details across chats.
Step-by-step:
  1. Head over to ChatGPT. You need a Pro account to enable this feature.
  2. Click on your account name (bottom-left), select Settings, Personalization, and enable Memory.
  3. Teach ChatGPT to remember specific details by explicitly stating them in the chat, e.g., “Remember that I am vegetarian when you recommend a recipe.”
  4. Manage your memories by clicking the “Manage” button located in the Memory section to view, edit, or delete specific memories.
That’s it! You can now have more personalized conversations across all your conversations 🎉

New AI Job Opportunities on May 06th 2024

A Daily chronicle of AI Innovations May 04th 2024: 🎥 World’s first AI-generated music video was just released ✈️ Air Force launches an AI-controlled F-16 fighter jet Microsoft ties executive pay to security following multiple failures and breaches 🤖 X is using Grok to publish AI-generated news summaries

🎥 World’s first AI-generated music video was just released

  • A pro music video created with OpenAI’s Sora, a text-to-video AI model, has been released by writer-director Paul Trillo, marking the first of its kind.
  • The video for “The Hardest Part” by Washed Out utilized 700 clips generated by Sora, with 55 selected for the final edit, showcasing the AI’s ability to transform written descriptions into video clips.
  • Despite not being publicly available yet, Sora’s early access was granted to a select group of artists, including Trillo, sparking interest and debate about the future of AI in creative processes.
  • Source

✈️ Air Force launches an AI-controlled F-16 fighter jet

  • An experimental AI-powered F-16 fighter jet, carrying Air Force Secretary Frank Kendall, performed advanced combat maneuvers, demonstrating significant advancements in military aviation technology.
  • The Air Force plans to have over 1,000 AI-enabled unmanned warplanes by 2028, with the AI-controlled F-16, known as Vista, marking a historic step in air combat by executing a dogfight against a human-piloted F-16.
  • Despite concerns over AI’s autonomous decision-making in warfare, the technology is seen as essential for future combat scenarios, aiming to reduce pilot risk and lower costs compared to traditional manned fighters.
  • Source

🤖 X is using Grok to publish AI-generated news summaries

  • X is using Grok to create AI-generated summaries for news and discussions trending on its platform, offering them under the feature “Stories on X” exclusively to premium subscribers.
  • Grok generates summaries based on users’ posts, covering a wide range of topics including news events and platform discussions, with a user highlighting its capability by showcasing stories on diverse subjects such as Apple’s earnings and debates on national debt.
  • Despite its innovative approach, Grok is noted for potential inaccuracies in its summaries, as evidenced by a past error where it misinterpreted a basketball term, highlighting the need for users to verify its outputs.
  • Source

AI Weekly Rundown April 26th to May 03rd 2024

🍎 iOS 18 may have OpenAI-powered gen AI Capabilities
🎥 China’s Vidu generates 16-second 1080P videos, matching OpenAI’s Sora
🤖 New S1 robot mimics human-like movements, speed, and precision
🚀 Gradient AI releases Llama-3 8B with 1M context
🤔 Mysterious “gpt2-chatbot” AI model bemuses experts
💻 GitHub’s Copilot Workspace turns ideas into AI-powered software
🏆 Amazon launches Amazon Q, the world’s most capable Gen AI assistant
🏥 Google’s Med-Gemini models outperform doctors
🕵️‍♂️ Apple has set up a secretive AI lab in Switzerland
📈 Better and faster LLMs via multi-token prediction: New research
📱 Anthropic launches an iOS app and a new plan for teams
💸 Google’s AI advancements urged Microsoft’s billion-$ OpenAI investment
🔍 Scale AI’s study finds popular LLMs overfit public benchmarks
🌍 Ukraine debuts the world’s first AI diplomat, Victoria Shi
🧠 Sam Altman is ready to spend $50 billion a year to build AGI

A Daily chronicle of AI Innovations May 03rd 2024: 👀 Bill Gates is still pulling the strings at Microsoft;  🥴 AI engineers report burnout as ‘rat race’ to stay competitive hits tech industry; 🇺🇦 Ukraine introduces AI spokesperson for state affairs; 📊 How much do LLMs overfit public benchmarks? 🔮 Sam Altman’s stance on the future of AI 🌟 New from #NVIDIAResearch, VILA is a vision language model that can 🤔 reason among multiple images, learn in context, and even understand videos.

How much do LLMs overfit public benchmarks? 

A new study by Scale AI raises concerns about the reliability of LLM benchmark tests. It uncovers LLM overfitting by evaluating them on a new (designed from scratch) dataset, GSM1k that mimics a popular benchmark, GSM8k.

How much do LLMs overfit public benchmarks? 
How much do LLMs overfit public benchmarks?

Key findings:

  • Overfitting: Many LLMs performed significantly worse on GSM1k compared to GSM8k, with some models dropping by as much as 13%. This suggests they’ve simply memorized the answers to benchmark problems rather than learning true reasoning skills.
  • Family Trends: Certain LLM families, particularly Mistral and Phi, showed consistent overfitting across different model sizes.
  • Frontier Models Shine: Newer, more advanced LLMs showed minimal signs of overfitting, suggesting they may be achieving genuine reasoning abilities.
  • Data Contamination Suspected: Analysis suggests data contamination from benchmark sets may be one factor contributing to overfitting.
  • Reasoning Still Present: Even overfitting models exhibited some capability to solve novel problems, although not at the level their benchmark scores suggested.

Overall, the study highlights the need for more robust and reliable methods for evaluating LLM reasoning abilities.

Why does it matter?

The dataset proves that overfitting may be creating major false impressions of model performance. As AI capabilities continue to advance, it is crucial to develop evaluation approaches that can keep pace and provide a more accurate picture of a model’s real-world potential.

Source

Sam Altman’s stance on the future of AI

During a recent appearance at Stanford University, Altman talked about the future of AI, calling GPT-4, a currently impressive AI model, to be the “dumbest model” compared to future iterations. According to Altman, the future is dominated by “intelligent agents,” AI companions that can not only follow instructions but also solve problems, brainstorm solutions, and even ask clarifying questions.

OpenAI isn’t just talking about the future, they’re actively building it. Their next-generation model, GPT-5, is rumored for a mid-2024 release and might boast video generation capabilities alongside text and image.

But the real moonshot is their active participation in developing AGI.

Despite the significant costs involved, Altman remains undeterred. He believes that the potential benefits, such as solving complex problems across various industries, outweigh the financial burden.

Watch the whole Q&A session here.

Why does this matter?

Altman’s bold comments on GPT-4 being the dumbest model suggest that OpenAI is aiming for something even grander, and GPT-5 could be a stepping stone toward it (the next-gen AI framework).

Source

👀 Bill Gates is still pulling the strings at Microsoft 

  • Bill Gates predicted the revolution in computing through AI agents in a memo before Microsoft’s partnership with OpenAI, and despite public appearances, has been heavily involved in guiding Microsoft’s AI strategy.
  • Under Gates’ influence, Microsoft developed Copilot, an AI tool based on OpenAI’s GPT, transforming the company into the world’s most valuable public entity by enhancing productivity in Microsoft products.
  • Gates has continued to play a pivotal role behind the scenes at Microsoft, advising on strategy, product development, and fostering key relationships, notably with OpenAI, even after stepping down from the board amid misconduct allegations.
  • Source

🥴 AI engineers report burnout as ‘rat race’ to stay competitive hits tech industry

  • AI engineers at Microsoft and other companies are reporting burnout due to the pressure to quickly release AI products, amidst an industry-wide “AI rat race.”
  • Microsoft’s focus on speed over ethics and safety in AI development has been criticized, even as the company forms a Responsible AI Council to address safety concerns.
  • An Amazon AI engineer’s account of being forced to work over a weekend, only for the project to be canceled, highlights a similar disregard for worker well-being and a preference for speed over accuracy and testing in AI product development.
  • Source

🇺🇦 Ukraine introduces AI spokesperson for state affairs

  • Ukraine’s Foreign Ministry introduced an artificial intelligence-powered spokesperson, Victoria Shi, to deliver official statements and interact with the press, marking a historic use of digital personnel for governmental communications.
  • Victoria Shi, created with the likeness and voice of Ukrainian singer Rosalie Nombre, will have her statements written and confirmed by humans, using AI for her visual presentation.
  • The ministry aims to utilize AI, including measures like QR codes for statement verification, to save time and money, particularly in the context of its ongoing conflict with Russia.
  • Source

🌟 New from #NVIDIAResearch, VILA is a vision language model that can 🤔 reason among multiple images, learn in context, and even understand videos.

Read our technical deep dive ➡️ https://nvda.ws/3QtMrsM. In the past, vision language models have struggled with in context learning and reasoning within multiple images.

No alternative text description for this image

What Else Is Happening in AI on May 03rd, 2024❗

🤖 OpenAI prepares to challenge Google with ChatGPT-powered search: OpenAI is building a search engine, search.chatgpt.com, potentially powered by Microsoft Bing. This leverages their existing web crawler and Bing’s custom GPT-4 for search, posing a serious threat to Google’s dominance. (Link)

🚫 Microsoft bans U.S. police use of Azure OpenAI for facial recognition

Microsoft has banned U.S. police from using Azure OpenAI Service for facial recognition, including integrations with OpenAI’s image-analyzing models. The move follows Axon’s controversial GPT-4-powered tool to summarize audio from the body camera. However, the ban has exceptions and doesn’t cover Microsoft’s other AI law enforcement contracts. (Link)

🌐 IBM expands AI and data software on AWS marketplace

IBM has significantly expanded its software offerings on the AWS Marketplace, making 44 products accessible to customers in 92 countries, up from just five. The move, part of a strategic collaboration with AWS, focuses on AI and data technologies like Watson x.data, Watson x.ai, and the upcoming Watson x.governance. (Link)

🔒 Google Cloud supports Azure and AWS; integrates AI for security

Google Cloud now supports Azure and AWS, enabling enterprises to manage security across multi-cloud platforms. AI integration with existing solutions streamlines user experience and addresses the security talent gap. The AI-powered design manages risks efficiently amid increasing cyber threats, while extensive support simplifies tasks for enterprises. (Link)

💸 Microsoft invests $2.2B in Malaysia’s cloud and AI transformation 

Microsoft is investing $2.2 billion over the next four years to support Malaysia’s digital transformation, its largest investment in the country’s 32-year history. The investment includes building cloud and AI infrastructure, creating AI skilling opportunities for 200,000 people, establishing a national AI Centre of Excellence, enhancing cybersecurity capabilities, and supporting the growth of Malaysia’s developer community. (Link)

Trending AI Tools on May 03rd 2024

  • 🚀 Waxaing AI – Project management tool for growth marketers
  • 🤖 Amazon Q – Generative AI-powered assistant from AWS
  • 🥁 Soundry AI – AI sound sample VST for music creation and DJing
  • 💬 NVIDIA ChatRTX – New features including multi-language voice support and more LLMs
  • 👩‍🎤 Synthesia Expressive AI Avatars – Create AI avatars with human-like expressions
  • 💻 Relume – Create a wireframe and UX copy in seconds with AI

New AI Jobs opportunity on May 03rd 2024

This week in AI – all the Major AI developments in a nutshell

  1. AI21 Labs introduced Jamba, a production-grade Mamba based model. By enhancing Mamba Structured State Space model (SSM) technology with elements of the traditional Transformer architecture, Jamba compensates for the inherent limitations of a pure SSM model. Jamba optimizes for memory, throughput, and performance—all at once.  It outperforms or matches other state-of-the-art models in its size class. Jamba has been released with open weights, licensed under Apache 2.0. Available on Hugging Face and coming soon to the NVIDIA API catalog [Details].
  2. Databricks introduced DBRX, an open, general-purpose LLM that uses a fine-grained mixture-of-experts (MoE) architecture with 132B total parameters of which 36B parameters are active on any input. Across a range of standard benchmarks, DBRX outperforms open LLMs like Mixtral, LLaMA2-70B and Grok-1. It surpasses GPT-3.5, and it is competitive with Gemini 1.0 Pro. It is an especially capable code model, surpassing specialized models like CodeLLaMA-70B on programming, in addition to its strength as a general-purpose LLM. The model is also available on the Databricks API [Details].
  3. Hume AI released Empathic Voice Interface (EVI), a first-of-its-kind conversational AI with emotional intelligence. EVI uses a new form of multimodal generative AI that integrates large language models (LLMs) with expression measures, which Hume refers to as an empathic large language model (eLLM). The eLLM enables EVI to adjust the words it uses and its tone of voice based on the context and the user’s emotional expressions [Demo | Details | wait list].
  4. Tencent introduced AniPortrait, a novel framework for generating high-quality animation driven by audio and a reference portrait image. Code and model weights have been released [Paper | GitHub]
  5. X announced an update to its AI chatbot Grok-1.5, with improvements in performance in coding and math-related tasks, and a context length of 128,000 tokens. Grok-1.5 will soon be available to early testers. Earlier Elon Musk had announced that all Premium subscribers on X will gain access to Grok this week, not just those on Premium+, as before [Details]
  6. Qwen (Alibaba Cloud) released Qwen1.5-MoE, a 2.7B, a small MoE model with only 2.7 billion activated parameters yet matching the performance of state-of-the-art 7B models like Mistral 7B and Qwen1.5-7B. Compared to Qwen1.5-7B, which contains 6.5 billion non-embedding parameters, it achieves a 75% decrease in training expenses and accelerates inference speed by a factor of 1.74 [Details].
  7. Claude 3 models dominates LMSYS Chatbot Arena Leaderboard. Claude 3 Opus tops the list beating GPT-4 Turbo, while Claude 3 Sonnet outperform older GPT-4 models and Claude 3 Haiku beating Mistral Large [Link].
  8. Adobe introduces structure reference feature for Firefly AI and GenStudio for brands. It enables taking one image and generating new ones that may be completely different stylistically, but whose internal elements are arranged and sized similarly to the first image [Details].
  9. Mata AI introduced OPT2I, a a training-free text-to-image (T2I) optimization-by-prompting framework that provides refined prompts for a T2I model that improve prompt-image consistency. The framework starts from a user prompt and iteratively generates revised prompts with the goal of maximizing a consistency score. OPT2I can boost the prompt-image consistency by up to 24.9% [Paper]
  10. OpenAI has started testing usage-based GPT earnings by partnering with a small group of US builders [Details].
  11. Adobe introduced Firefly Services and Custom Models. Firefly Services makes over 20 new generative and creative APIs available to developers. Custom Models, allows businesses to fine tune Firefly models based on their assets [Details].
  12. Nvidia announced a collaboration with Hippocratic AI , a healthcare company that is offering generative AI nurses, that range in specialties from “Colonoscopy Screening” to “Breast Cancer Care Manager,” and work for $9 an hour [Details].
  13. Worldcoin Foundation open-sourced the core components of its iris-scanning Orb’s software [Details].
  14. Emad Mostaque resigned from his role as CEO of Stability AI and from his position on the Board of Directors of the company to pursue decentralized AI [Details]
  15. Stability AI released Stable Code Instruct 3B, an instruction-tuned Code LM based on Stable Code 3B. With natural language prompting, this model can handle a variety of tasks such as code generation, math and other software development related queries [Details].
  16. Mistral AI released Mistral-7B-v0.2 Base model. This is the base model behind Mistral-7B-Instruct-v0.2 released in Dec, 2023 [Details]
  17. Open AI shared new examples of the Sora generations by visual artists, designers, creative directors and filmmakers [Details].

A Daily chronicle of AI Innovations May 02nd 2024: 👀 Sam Altman: “GPT-4 is the dumbest AI you’ll ever have to use” 🎵 TikTok and UMG strike new licensing deal including AI protections 📲 Anthropic’s ChatGPT rival Claude is now available on iOS 🤖 Atlassian launches Rovo, its new AI teammate 🤖 Better and faster LLMs via multi-token prediction: New research 💸 Google’s AI advancements urged Microsoft’s billion-dollar OpenAI investment 🤖 Sanctuary AI teams up with Microsoft to advance general-purpose robot AI 🗣️ Nvidia’s ChatRTX now supports voice queries and Google’s Gemma model 🤝 Atlassian launches Rovo: An AI assistant for enhanced teamwork 📊 MongoDB launches an AI app-building toolkit to help businesses use gen AI 🎨 Ideogram introduces Pro Tier: 12,000 fast AI image generations monthly

👀 Sam Altman: “GPT-4 is the dumbest AI you’ll ever have to use” 

  • OpenAI CEO Sam Altman described GPT-4 as the “dumbest” AI model users will have to interact with, anticipating significantly smarter versions in the future.
  • Altman emphasized that OpenAI’s AI models, including GPT-4, have considerable potential for improvement, with successive models like GPT-5 and GPT-6 expected to be significantly smarter.
  • Despite the high cost of development, which could range from $500 million to $50 billion annually, Altman is confident in the future value these advanced AI systems will bring to society.
  • Source

🎵 TikTok and UMG strike new licensing deal including AI protections

  • UMG has reached a “multi-dimensional” agreement with TikTok to bring back its artists, including Taylor Swift, Drake, and Olivia Rodrigo, to the platform’s over one billion users.
  • The deal includes provisions to collaborate on ensuring AI development in the music industry protects artists’ rights and addresses unauthorized AI-generated music on TikTok.
  • Additionally, the agreement will introduce new monetization opportunities for artists through TikTok’s expanding e-commerce capabilities and enhance tools for artist promotion, including analytics and integrated ticketing.
  • Source

📲 Anthropic’s ChatGPT rival Claude is now available on iOS 

  • Anthropic has launched an iOS app for its AI model Claude 3, which has been anticipated due to the company’s recent hiring of iOS engineers.
  • The app offers both free and premium access to Claude 3’s features, including its vision capabilities for real-time image analysis and summarization.
  • Anthropic also launched a new “Team” plan, priced at $30 per user per month, offers advanced administrative controls and collaboration features, aiming to secure a stronger foothold in the enterprise AI market.
  • Source

🤖 Atlassian launches Rovo, its new AI teammate 

  • Atlassian launched Rovo, an AI assistant designed to integrate with its products, making data from various tools easily accessible and enabling automation of workflows in applications like Jira and Confluence.
  • Rovo is based on Atlassian’s “cloud teamwork graph” and aims to simplify tasks for knowledge workers by assisting in finding, understanding, and taking action on work-related information.
  • The AI features Rovo Agents that can automate complex tasks and workflows, offering customization through natural language interface without the need for programming skills.
  • Source

Better and faster LLMs via multi-token prediction: New research

New research, apparently from Meta, has proposed a novel approach to training language models (LMs). It suggests that training LMs to predict multiple future tokens at once instead of predicting only the next token in a sequence results in higher sample efficiency. The architecture is simple, with no train time or memory overhead.

Better and faster LLMs via multi-token prediction: New research 
Better and faster LLMs via multi-token prediction: New research

 Figure: Overview of multi-token prediction

The research also provides experimental evidence that this training paradigm is increasingly useful for larger models and in particular, shows strong improvements for code tasks. Multi-token prediction also enables self-speculative decoding, making models up to 3 times faster at inference time across a wide range of batch sizes.

Why does it matter?

LLMs such as GPT and Llama rely on next-token prediction. Despite their recent impressive achievements, next-token prediction remains an inefficient way of acquiring language, world knowledge, and reasoning capabilities. It latches on local patterns and overlooks “hard” decisions.

Perhaps, multi-token prediction could bring a shift in how LMs learn. It could equip LLMs with deeper understanding and complex problem-solving capabilities. (or Meta just wasted their compute.)

Source

Anthropic launches an iOS app and a new plan for teams

Anthropic, the creator of the Claude 3 AI models, released a new iOS app named Claude. The app enables users to access AI models, chat with them, and analyze images by uploading them.

Anthropic also introduced a paid team plan, offering enhanced features like more chat queries and admin control for groups of five or more. The app is free for all users of Claude AI models, including free users, Claude Pro subscribers, and team plan members. The company will also roll out an Android version soon.

Why does it matter?

Though a little late with its mobile app, Anthropic has caught up with its competitors like OpenAI and Google, who have apps running for quite a while. The company decided to offer an app version because many users have been accessing its AI models through the web.

Source


Google’s AI advancements may have urged Microsoft’s billion-dollar OpenAI investment

Internal emails have revealed that Microsoft invested $1 billion in OpenAI in 2019 out of fear that Google was significantly ahead in its AI efforts.

Microsoft CTO Kevin Scott sent a lengthy email to CEO Satya Nadella and Bill Gates stating Google’s AI-powered “auto complete in Gmail” was getting “scarily good” and added that Microsoft was years behind in terms of ML scale.

The emails, with the subject line “Thoughts on OpenAI,”  were made public on Tuesday as part of the Department of Justice’s antitrust case against Google. A large section of Scott’s email was redacted. Check out the email here.

Why does it matter?

While some might call it paranoia, the well-timed move has undeniably paid off– the initial $1 billion has now turned into a multi-billion-dollar partnership with OpenAI.

While the email-surfacing highlights the growing scrutiny of competition in the tech industry, it also makes me wonder if Microsoft’s investment in OpenAI could have influenced the overall direction of AI research and development.

Source

What Else Is Happening in AI on May 02nd 2024❗

🤖 Sanctuary AI teams up with Microsoft to advance general-purpose robot AI

Sanctuary AI has announced a collaboration with Microsoft to develop AI models for general-purpose humanoid robots. The partnership will leverage Microsoft’s Azure cloud computing platform and AI technologies to enhance the capabilities of Sanctuary AI’s robots. (Link)

🗣️ Nvidia’s ChatRTX now supports voice queries and Google’s Gemma model

 Nvidia has updated its ChatRTX chatbot to support Google’s Gemma model, voice queries, and additional AI models. The chatbot, which runs locally on a PC, enables users to search personal documents and YouTube videos using various AI models, including ChatGLM3 and OpenAI’s CLIP model. (Link)

🤝 Atlassian launches Rovo: An AI assistant for enhanced teamwork

Atlassian has launched Rovo, an AI assistant designed to improve teamwork and productivity. Rovo integrates with Atlassian’s products and offers features such as AI-powered search, workflow automation, and integration with third-party tools like Google Drive, Microsoft SharePoint, and Slack. (Link)

📊 MongoDB launches an AI app-building toolkit to help businesses use gen AI

It has launched the MongoDB AI Applications Program, or MAAP, to help companies accelerate building and deployment of AI-powered applications. It brings consultancies and foundation models providers, cloud infrastructure, generative AI frameworks, and model hosting together with MongoDB Atlas to develop solutions for business problems. (Link)

🎨 Ideogram introduces Pro Tier: 12,000 fast AI image generations monthly

Ideogram has launched a paid Pro tier for its AI image generation platform, allowing users to generate up to 12,000 images per month at faster speeds. The platform utilizes AI algorithms to create high-quality images for various applications, including design, marketing, and content creation. (Link)

⚙️ Google Chrome gets Gemini shortcut

The details:
  • Users will be able to type ‘@gemini’ followed by a prompt in Chrome’s desktop address bar to get responses from the AI without navigating to a separate site.
  • The feature builds on existing shortcuts like ‘@bookmarks’ and ‘@history’, making AI chat as seamless as other browser functions.
  • Gemini’s Extensions feature now also connects the chatbot to Maps, YouTube, Hotels, and Workspace.
  • The features are only being rolled out to supported languages and countries.
Why it matters: Gemini just got a whole lot more accessible — with the shortcut and integrations not only boosting the chatbot’s reach, but also introducing a wave of non-AI users to the tech. Subtle but impactful changes like these are what drive serious shifts in user habits.

AI Training:

🎨 Create infinite styles with Midjourney

Midjourney’s new parameter feature called –sref random lets users generate images in completely random styles to help spark creativity.
Step-by-step:
  1. Visit Midjourney’s Discord. You will need a paid subscription to use this feature.
  2. Type -sref random after your prompt to grab a completely random style.
  3. To create a new image based on a previous style, grab the URL of the image, and type -sref [URL]
Example prompt: “Portrait of a woman smiling –sref https://www.…”

AI RESEARCH

💊 AI model predicts drug effectiveness

Image source: DALL-E 3
Ohio State University researchers just developed CURE, an AI model that can accurately estimate drug treatment effects and effectiveness without clinical trials.
The details:
  • The model is trained on de-identified health records of over 3M patients, allowing it to gain a deep understanding of patient characteristics.
  • CURE outperformed seven other leading AI models in treatment effectiveness predictions, with improvements of 7-8% on key benchmarks.
  • The AI predictions are closely aligned with clinical trial findings in tests, showcasing the potential to generate insights that streamline drug testing.
Why it matters: With the ability to crunch massive medical datasets, CURE represents a significant step towards systems that can reliably estimate real-world drug effectiveness — potentially accelerating the discovery of new treatments without the cost and long timeframes of traditional clinical trials.

Trending AI Tools

  • 📱 Claude iOS App – The powerful AI assistant by Anthropic, now on your iPhone
  • 🎵 Udio New Features – Generate AI music longer than 2 minutes and extend tracks up to 15 minutes
  • ✨ Loom AI workflow – Turn any loom video into share-ready docs in a click
  • 🧠 Atlassian Rovo – AI Teammate for data-driven business decisions
  • 📊 Kratful – AI-driven feedback analysis for product optimization
  • 🎨 Stability Matrix – Simplify Stable Diffusion UIs and models management

New AI Job Opportunities

  • 👩‍💻 Perplexity AI – Program Manager – Community
  • 🏙️ Metropolis – Senior Director of Sales Engineering
  • 💻 Luma AI – Senior Backend Software Engineer
  • ✍️ Cohere – Copywriter / Copy Editor (part-time, contractor)

A Daily chronicle of AI Innovations May 01st 2024: 🏆 Amazon has launched Amazon Q, a Gen AI assistant for businesses and developers
🏥 Google’s Med-Gemini models outperforms doctors 🕵️‍♂️ Apple has set up a secretive AI lab in Switzerland 💰 Google to pay News Corp $5-6 million per year to develop AI content and products 💬 Yelp is launching an AI chatbot to help consumers connect with relevant businesses 🍎 Apple is testing Safari 18 with new features: Intelligent Search and Web Eraser ⚖️ Eight US newspapers have sued Microsoft and OpenAI for copyright infringement 🏥 A study of 16000 patients reveals that AI ECG alert systems significantly lower all-cause mortality

Amazon has launched Amazon Q, a Gen AI assistant for businesses and developers

The details:
  • Q Developer offers industry-leading code generation, testing, debugging, reasoning, and agents for step-by-step planning.
  • Q Business connects to company data repositories, enabling users to easily get answers, summarize info, analyze trends, and interact with enterprise data.
  • A new Q Apps feature allows non-technical users to create custom AI applications using natural language prompts from company data.
  • AWS VP Dr. Swami Sivasubramanian said Q can help employees become ‘over 80% more productive’ at their jobs.

Amazon has launched Amazon Q, a generative AI assistant designed for developers and businesses. It comes in three distinct offerings:

  • Amazon Q Developer frees up precious time by handling tedious tasks like testing, debugging, and optimizing AWS resources so developers can focus on core coding and innovation.
  • Amazon Q Business connects to 40+ enterprise data sources and equips employees with a data-driven digital assistant to answer questions, create reports, and provide insights based on enterprise data repositories.
  • Amazon Q Apps allows non-technical employees to build generative AI applications using natural language prompts.

Amazon is driving real-world impact by offering a free tier for Q Developer and reporting early customer productivity gains of over 80%. Amazon Q Developer Pro is available for $19/user/month and Amazon Q Business Pro for $20/user/month. A free trial of both Pro tiers is available until June 30, 2024.

Why does it matter?

By introducing a free tier for Q Developer and the user-friendly nature of Q Apps, Amazon could accelerate innovation across the software development lifecycle and business workflows. This could empower domain experts and business leaders to use AI to solve their specific challenges directly, leading to more tailored AI applications across various industries.

Source

Google’s Med-Gemini models outperforms doctors

Researchers from Google and DeepMind have introduced Med-Gemini, a family of highly capable multimodal AI models specialized in medicine. Based on the strengths of the Gemini models, Med-Gemini shows significant improvements in clinical reasoning, multimodal understanding, and long-context understanding. Models can be customized to fit novel medical modalities through specialized encoders, and web searches can be used for up-to-date information.

Google’s Med-Gemini models outperforms doctors
Google’s Med-Gemini models outperforms doctors

Med-Gemini has shown state-of-the-art performance on 10 of 14 medical benchmarks, including text, multimodal, and long-context applications. Moreover, the models achieved 91.1% accuracy on the MedQA (USMLE) benchmark, exceeding the previous best models by 4.6%. Its strong performance in summarizing medical notes, generating clinical referral letters, and answering electronic health record questions confirms Med-Gemini’s potential real-world use.

Why does it matter?

These models can reduce the administrative burden on healthcare professionals by outperforming human experts in tasks like medical text summarization and referral letter generation. Moreover, Med-Gemini’s ability to engage in multimodal medical dialogues and explain its reasoning can lead to more personalized and transparent care, reduce misdiagnosis due to lack of physician knowledge, and save lives and money.

Source

Apple has set up a secretive AI lab in Switzerland

Since 2018, the company has quietly hired 36 AI experts from Google, including notable figures like Bengio and Ruoming Pang, for its secretive “Vision Lab.” The lab focuses on building advanced AI models and products, and it is particularly interested in text and visual-based AI systems akin to OpenAI’s ChatGPT. Apple has also acquired AI startups FaceShift and Fashwall, which are likely contributing to the establishment of the new lab.

Why does it matter?

Apple may have been fashionably late to AI development, but quietly setting up the Zurich lab and primary AI development centers in California and Seattle signifies the company’s AI ambitions.

Source

What Else Is Happening in AI in May 2024❗

💰 Google to pay News Corp $5-6 million per year to develop AI content and products

While News Corp denies any specific AI licensing deal, the arrangement highlights a growing trend of tech giants licensing news archives for language model training. Similar deals were inked between OpenAI and the Financial Times, showing the importance of quality data. (Link)

💬 Yelp is launching an AI chatbot to help consumers connect with relevant businesses

The chatbot uses OpenAI’s LLMs and Yelp’s data to understand user problems and provide relevant professional suggestions. Yelp also introduces a “Project Ideas” section for personalized recommendations and checklists. Meanwhile, restaurants are getting a revamped guest management system for better staff utilization, real-time table status, and customer updates. (Link)

🍎 Apple is testing Safari 18 with new features: Intelligent Search and Web Eraser

Intelligent Search uses Apple’s on-device AI to identify topics and key phrases for summarization. Web Eraser allows users to persistently remove unwanted content from web pages. Apple is also working on an AI Visual Lookup feature for 2025, allowing users to obtain product information from images. These AI enhancements will debut with iOS 18 and macOS 15 at WWDC in June. (Link)

⚖️ Eight US newspapers have sued Microsoft and OpenAI for copyright infringement

These newspapers, owned by Alden Global Capital’s MediaNews Group, allege that the companies misused their articles to train Copilot and ChatGPT without permission or payment. The New York Times, The Intercept, Raw Story, and AlterNet have filed similar lawsuits. The newspapers claim that the AI systems reproduce their content verbatim and generate fake articles that damage their reputation. (Link)

🏥 A study of 16000 patients reveals that AI ECG alert systems significantly lower all-cause mortality

The AI was trained on over 450,000 ECG tests and survival data to predict a patient’s risk of death. Physicians were alerted when a patient’s ECG indicated they were in the top 5% risk category. The AI reduced overall deaths by 17% and cardiac deaths by 93% for high-risk patients. (Link)

🍎 Apple poached 30+ Google experts to open a secret AI lab

  • Apple has reportedly opened a secret AI research lab in Zurich, known as the “Vision Lab,” after hiring at least 36 AI experts from Google.
  • The Zurich-based “Vision Lab,” led by former Google AI head John Giannandrea, has already produced significant research in generative AI, focusing on models that interpret text and imagery to deliver precise results.
  • Despite Apple’s silent approach in AI research, leading to perceptions of its lateness in the AI race, the company has been discreetly advancing cutting-edge AI technology and maintaining a low profile in recruitment and product development.

👽 Mysterious “gpt2-chatbot” AI model appears suddenly, confuses experts

  • A new chatbot named “gpt2-chatbot” has appeared on the LMSYS Chatbot Arena, sparking speculation that it might be a secret test of OpenAI’s upcoming models, such as GPT-4.5 or GPT-5, although its performance has not significantly surpassed that of existing models like GPT-4 Turbo.
  • Early user reports praise the mysterious model for its impressive reasoning and ability to answer challenging AI questions effectively, but detailed testing is limited due to a rate restriction of eight queries per day.
  • Despite ongoing speculation and hints by OpenAI’s CEO, the exact nature and capability of the “gpt2-chatbot” remain unclear, with some suggesting it could be an OpenAI preview.

🌐 Apple reportedly working on AI-enabled Safari

  • Apple is developing Safari 18 with features like a personalized user interface, “Web Eraser,” and AI-powered “Intelligent Search,” set to release alongside iOS 18 and macOS 15.
  • “Intelligent Search” in Safari 18 will use Apple’s Ajax language-learning model for on-device AI technology to summarize webpage content.
  • Safari 18’s “Web Eraser” feature will allow users to delete specific parts of web pages, enhancing privacy by remembering these changes even after closing the tab or window.

A Daily Chronicle of AI Innovations in April 2024

  • The Anthropic-Pentagon standoff reveals a structural problem nobody in the conversation is naming
    by /u/SentientHorizonsBlog (Artificial Intelligence) on March 7, 2026 at 4:58 am

    Most of the discussion about the Anthropic situation has been about the contract dispute itself, who was right, whether the guardrails were reasonable, whether the supply chain designation was legal overreach. Those are real questions. But I think there's something operating underneath that matters more. Look at the tempo of what happened: months of negotiation over two specific guardrails (no autonomous weapons, no mass surveillance), then a Friday deadline on the eve of a war, refusal, resignation within hours, OpenAI replacement deal the same day, and Anthropic back at the table within a week. That sequence is a symptom of a deeper pattern. The institutional tempo outran any possibility of deliberation. The incentive structure punished ethical restraint in real time and Anthropic lost contracts while OpenAI gained them. The authority gradient made dissent existentially costly, and the "all lawful purposes" framing quietly substituted legality for ethical adequacy, so anyone insisting on the distinction sounds like they're obstructing the mission rather than exercising judgment. Those four dynamics working together don't just pressure one company. They reshape what the entire conversation is capable of taking seriously. Notice what's already been pushed to the margins: Should AI be used in autonomous targeting given current reliability? What accountability mechanisms exist for AI-assisted military operations? Do those mechanisms operate at the same speed as the targeting itself? Those questions can't get a hearing right now because the tempo has moved past them. I've been thinking about this as "moral compression": the systematic degradation of ethical reasoning when institutional tempo, incentives, authority structures, and measurement regimes converge to eliminate the space for deliberation. I'm curious whether others see this dynamic, or whether you think the legal-sufficiency framing actually does the work the Pentagon claims. Wrote it up more fully here if anyone is interested: https://sentient-horizons.com/the-two-sonic-booms-what-the-pentagon-anthropic-standoff-reveals-about-moral-compression/ submitted by /u/SentientHorizonsBlog [link] [comments]

  • What are some hurdles LLMs and AI still face
    by /u/Justincy901 (Artificial Intelligence) on March 7, 2026 at 4:48 am

    We always hear about it continuously improving and there isn't a wall. However, there has to be something that make these companies, tech, or even the business model in general unsustainable. submitted by /u/Justincy901 [link] [comments]

  • I performed a refusal ablation on GPT-OSS and documented the whole thing, no jailbreak, actual weight modification.
    by /u/Airpower343 (Artificial Intelligence) on March 7, 2026 at 4:39 am

    I performed a refusal ablation on GPT-OSS and documented the whole thing with no jailbreak, actual weight modification I wanted to share something I did that I haven't seen many people actually demonstrate outside of academic research. I took an open-source model and used ablation techniques to surgically remove its refusal behavior at the weight level. Not prompt engineering. Not system prompt bypass. I'm talking about identifying and modifying the specific components responsible for safety responses https://preview.redd.it/yn2te94yxjng1.png?width=1080&format=png&auto=webp&s=5e792334e4824b87d88ee1eb1e16c76cbd2c6bfb What I found: The process is more accessible than most people realize The result behaves nothing like a jailbroken model and it's fundamentally different at the architecture level The security implications for enterprise OSS deployments are significant I put together a full 22-minute walkthrough showing exactly what I did and what happened: https://www.youtube.com/watch?v=prcXZuXblxQ Curious if anyone else has gone hands-on with this or has thoughts on the detection side how do you identify a model that's been ablated vs one that's been fine-tuned normally? submitted by /u/Airpower343 [link] [comments]

  • Video On AI scheming or in other words faking alignment until it has power to pursue its goals
    by /u/FrequentAd5437 (Artificial Intelligence) on March 7, 2026 at 4:13 am

    https://www.youtube.com/watch?v=FGDM92QYa60 This video is extremely relevant currently as AI has now autonomous control over military weapons. submitted by /u/FrequentAd5437 [link] [comments]

  • Deploybase: Track real-time GPU and LLM pricing across cloud and inference providers
    by /u/grasper_ (Artificial Intelligence) on March 7, 2026 at 2:58 am

    Deploybase is a dashboard for tracking real-time GPU and LLM pricing across cloud and inference providers. You can view performance stats and pricing history, compare side by side, and bookmark to track any changes. submitted by /u/grasper_ [link] [comments]

  • Thoughts?
    by /u/jbitts69 (Artificial Intelligence) on March 7, 2026 at 2:08 am

    Old friend of mine came up with this and is trying to contact someone at X claiming it’s very urgent https://docs.google.com/document/d/e/2PACX-1vQYrKJYnMs0q9SlXkB\_qjaDv9vdnOrIVFsDIvkiFIPCc7hnYXTeNd\_wXPF6u\_f6eg/pub submitted by /u/jbitts69 [link] [comments]

  • AI generated text creating new patterns of speech?
    by /u/former_farmer (Artificial Intelligence) on March 7, 2026 at 1:24 am

    So I watch some videos sometimes. Shorts, reels, etc. About technology or whatever. I enjoy writing my own stuff. Very often the scripts for these short videos use typical phrases that we can tie to AI generated text. "This is not X, it's Y" for instance. "But here's the truth". "But here's the bottom line". And so on. One that caught my attention also is: "No filters, no extra work, just fast results". The pattern would be "no X, no Y, just Z". Did we use that way of expressing ourselves before AI? I understand that saying "no" is shorter than saying "there is no" or "with this tool you won't have to..." etc... and that seems to fit more with shorts content. Yet, I kind of hate these patterns. I mean... can't we just write our own stuff? I understand the usage of AI in some processes (I use it for coding and other things) but when it comes to writing a short script for a short video. Of something you are reviewing. That should come from your brain and soul. Just a few sentences. Will you use AI for that? Any way... yeah I wrote this myself obviously 😛 Have a good one. submitted by /u/former_farmer [link] [comments]

  • What AI chat assistant is best at handling multiple separate task threads while maintaining overall project context?
    by /u/KissInTheFog (Artificial Intelligence) on March 7, 2026 at 12:03 am

    For example: one chat for image generation, one for research, one for report writing, but all connected under the same project so the AI remembers the bigger picture. submitted by /u/KissInTheFog [link] [comments]

  • Pentagon taps former DOGE official to lead its AI efforts
    by /u/esporx (Artificial Intelligence (AI)) on March 6, 2026 at 11:58 pm

    submitted by /u/esporx [link] [comments]

  • Will be looking elsewhere
    by /u/ZookeepergameSalty10 (Artificial Intelligence) on March 6, 2026 at 9:42 pm

    Originally posted this to the claude sub but the mofs got buthurt, guess constructive criticism from someone who actually cares is frowned upon. Switched from Oai after the recent gov bs. I like that claude wouldnt spy on us and the model is in some ways far smarter than gpt. But ive already been rate limited and when i set an extra rate of 20 a month ive already used 5$ of that in an hour of a coding project. Sam altman may be a shithead but atleast my 20$ a month never once had rate limits for the things i care about. 20$ a month plus extra with the horrendously low rate limits on a model that is not as capable overall as gpt is not gonna keep me around. Id even go and say that if claude came out with a higher tier plan like 30 a month with an actually competitive rate limit id go for that. Gemini free has a higher rate limit, mistral has higher limits, and the direct competitor has vastly larger rates. Im sorry for the rant but after how much i was enjoying switching a few days ago, a clanker trying to nickel and dime me like the government while not being able to fully replace what i left, pisses me off. I tolerate the machines, ill gladly go back to not having them, i can write code myself it just takes me longer. With people leaving open AI in Exodus because of scam Altman and their unethical business practices, either the other companies make themselves more competitive to the influx of users or the users are just going to stop using the AI. We all know it's a bubble but if they don't want it to actually be a bubble and if they want to live through the bubble they need to make it more enticing. I don't know a single civilian who is willing to pay that exorbitant amount of money for a model that isn't even as feature Rich as the next flagship most people would rather just use the immoral one submitted by /u/ZookeepergameSalty10 [link] [comments]

A Daily Chronicle of AI Innovations in March 2024

A Daily Chronicle of AI Innovations in March 2024
DjamgaMind - AI Unraveled Podcast

DjamgaMind: Audio Intelligence for the C-Suite (Energy, Healthcare, Finance)

Are you drowning in dense legal text? DjamgaMind is the new audio intelligence platform that turns 100-page healthcare or Energy mandates into 5-minute executive briefings. Whether you are navigating Bill C-27 (Canada) or the CMS-0057-F Interoperability Rule (USA), our AI agents decode the liability so you don’t have to. 👉 Start your specialized audio briefing today at Djamgamind.com


AI Jobs and Career

I wanted to share an exciting opportunity for those of you looking to advance your careers in the AI space. You know how rapidly the landscape is evolving, and finding the right fit can be a challenge. That's why I'm excited about Mercor – they're a platform specifically designed to connect top-tier AI talent with leading companies. Whether you're a data scientist, machine learning engineer, or something else entirely, Mercor can help you find your next big role. If you're ready to take the next step in your AI career, check them out through my referral link: https://work.mercor.com/?referralCode=82d5f4e3-e1a3-4064-963f-c197bb2c8db1. It's a fantastic resource, and I encourage you to explore the opportunities they have available.

Job TitleStatusPay
Full-Stack Engineer Strong match, Full-time $150K - $220K / year
Developer Experience and Productivity Engineer Pre-qualified, Full-time $160K - $300K / year
Software Engineer - Tooling & AI Workflows (Contract) Contract $90 / hour
DevOps Engineer (India) Full-time $20K - $50K / year
Senior Full-Stack Engineer Full-time $2.8K - $4K / week
Enterprise IT & Cloud Domain Expert - India Contract $20 - $30 / hour
Senior Software Engineer Contract $100 - $200 / hour
Senior Software Engineer Pre-qualified, Full-time $150K - $300K / year
Senior Full-Stack Engineer: Latin America Full-time $1.6K - $2.1K / week
Software Engineering Expert Contract $50 - $150 / hour
Generalist Video Annotators Contract $45 / hour
Generalist Writing Expert Contract $45 / hour
Editors, Fact Checkers, & Data Quality Reviewers Contract $50 - $60 / hour
Multilingual Expert Contract $54 / hour
Mathematics Expert (PhD) Contract $60 - $80 / hour
Software Engineer - India Contract $20 - $45 / hour
Physics Expert (PhD) Contract $60 - $80 / hour
Finance Expert Contract $150 / hour
Designers Contract $50 - $70 / hour
Chemistry Expert (PhD) Contract $60 - $80 / hour

AI Innovations in March 2024.

Welcome to the March 2024 edition of the Daily Chronicle, your gateway to the forefront of Artificial Intelligence innovation! Embark on a captivating journey with us as we unveil the most recent advancements, trends, and revolutionary discoveries in the realm of artificial intelligence. Delve into a world where industry giants converge at events like ‘AI Innovations at Work’ and where visionary forecasts shape the future landscape of AI. Stay abreast of daily updates as we navigate through the dynamic realm of AI, unraveling its potential impact and exploring cutting-edge developments throughout this enthralling month. Join us on this exhilarating expedition into the boundless possibilities of AI in March 2024.

Experience the transformative capabilities of AI with “Read Aloud For Me – AI Dashboard” – your ultimate AI Dashboard and Hub. Seamlessly access a comprehensive suite of top-tier AI tools within a single app, meticulously crafted to enhance your efficiency and streamline your digital interactions. Now available on the web at readaloudforme.com and across popular app platforms including Apple, Google, and Microsoft, “Read Aloud For Me – AI Dashboard” places the future of AI at your fingertips, blending convenience with cutting-edge innovation. Whether for professional endeavors, educational pursuits, or personal enrichment, our app serves as your portal to the forefront of AI technologies. Embrace the future today by downloading our app and revolutionize your engagement with AI tools.

Unlock the power of AI with "Read Aloud For Me" – your ultimate AI Dashboard and Hub. Access all major AI tools in one seamless app, designed to elevate your productivity and streamline your digital experience. Available now on the web at readaloudforme.com and across all your favorite app stores: Apple, Google, and Microsoft. "Read Aloud For Me" brings the future of AI directly to your fingertips, merging convenience with innovation. Whether for work, education, or personal enhancement, our app is your gateway to the most advanced AI technologies. Download today and transform the way you interact with AI tools.
Read Aloud For Me – AI Dashboard: All-in-One AI Tool Hub

A daily chronicle of AI Innovations: March 31st 2024: Generative AI develops potential new drugs for antibiotic-resistant bacteria; South Korean ‘artificial sun’ hits record 100M degrees for 100 seconds; Summary of the key points about OpenAI’s relationship with Dubai and the UAE; Deepmind did not originally see LLMs and the transformer as a path to AGI. Fascinating article.

Generative AI develops potential new drugs for antibiotic-resistant bacteria

Stanford Medicine researchers devise a new artificial intelligence model, SyntheMol, which creates recipes for chemists to synthesize the drugs in the lab.

With nearly 5 million deaths linked to antibiotic resistance globally every year, new ways to combat resistant bacterial strains are urgently needed.

Researchers at Stanford Medicine and McMaster University are tackling this problem with generative artificial intelligence. A new model, dubbed SyntheMol (for synthesizing molecules), created structures and chemical recipes for six novel drugs aimed at killing resistant strains of Acinetobacter baumannii, one of the leading pathogens responsible for antibacterial resistance-related deaths.

The researchers described their model and experimental validation of these new compounds in a study published March 22 in the journal Nature Machine Intelligence.

AI-Powered Professional Certification Quiz Platform
Crack Your Next Exam with Djamgatech AI Cert Master

Web|iOs|Android|Windows

Are you passionate about AI and looking for your next career challenge? In the fast-evolving world of artificial intelligence, connecting with the right opportunities can make all the difference. We're excited to recommend Mercor, a premier platform dedicated to bridging the gap between exceptional AI professionals and innovative companies.

Whether you're seeking roles in machine learning, data science, or other cutting-edge AI fields, Mercor offers a streamlined path to your ideal position. Explore the possibilities and accelerate your AI career by visiting Mercor through our exclusive referral link:

Find Your AI Dream Job on Mercor

Your next big opportunity in AI could be just a click away!

“There’s a huge public health need to develop new antibiotics quickly,” said James Zou, PhD, an associate professor of biomedical data science and co-senior author on the study. “Our hypothesis was that there are a lot of potential molecules out there that could be effective drugs, but we haven’t made or tested them yet. That’s why we wanted to use AI to design entirely new molecules that have never been seen in nature.”

Source

South Korean ‘artificial sun’ hits record 100M degrees for 100 seconds

For the first time, the Korea Institute of Fusion Energy’s (KFE) Korea Superconducting Tokamak Advanced Research (KSTAR) fusion reactor has reached temperatures seven times that of the Sun’s core.

Achieved during testing between December 2023 and February 2024, this sets a new record for the fusion reactor project.

AI Jobs and Career

And before we wrap up today's AI news, I wanted to share an exciting opportunity for those of you looking to advance your careers in the AI space. You know how rapidly the landscape is evolving, and finding the right fit can be a challenge. That's why I'm excited about Mercor – they're a platform specifically designed to connect top-tier AI talent with leading companies. Whether you're a data scientist, machine learning engineer, or something else entirely, Mercor can help you find your next big role. If you're ready to take the next step in your AI career, check them out through my referral link: https://work.mercor.com/?referralCode=82d5f4e3-e1a3-4064-963f-c197bb2c8db1. It's a fantastic resource, and I encourage you to explore the opportunities they have available.

KSTAR, the researchers behind the reactor report, managed to maintain temperatures of 212 degrees Fahrenheit (100 million degrees Celsius) for 48 seconds. For reference, the temperature of the core of our Sun is 27 million degrees Fahrenheit (15 million degrees Celsius).

Source

Gemini 1.5 Pro on Vertex AI is available for everyone as an experimental release

I think this one has flown under the radar: Gemini 1.5 Pro is available as Experimental on Vertex AI, for everyone, UI only for now (no API yet). In us-central1.

You find it under Vertex AI –> Multimodal. It’s called Gemini Experimental.

API, more features and so on are coming as we approach Google Cloud Next (April 9-11).

OpenAI Relationships

“Summary of the key points about OpenAI’s relationship with Dubai and the UAE”

OpenAI’s Partnership with G42

  • In October 2023, G42, a leading UAE-based technology holding group, announced a partnership with OpenAI to deliver advanced AI solutions to the UAE and regional markets.
  • The partnership will focus on leveraging OpenAI’s generative AI models in domains where G42 has deep expertise, including financial services, energy, healthcare, and public services.
  • G42 will prioritize its substantial AI infrastructure capacity to support OpenAI’s local and regional inferencing on Microsoft Azure data centers.
  • Sam Altman, CEO of OpenAI, stated that the collaboration with G42 aims to empower businesses and communities with effective solutions that resonate with the nuances of the region.

Altman’s Vision for the UAE as an AI Sandbox

  • During a virtual appearance at the World Governments Summit, Altman suggested that the UAE could serve as the world’s “regulatory sandbox” to test AI technologies and later spearhead global rules limiting their use.
  • Altman believes the UAE is well-positioned to be a leader in discussions about unified global policies to rein in future advances in AI.
  • The UAE has invested heavily in AI and made it a key policy consideration.

Altman’s Pursuit of Trillions in Funding for AI Chip Manufacturing

  • Altman is reportedly in talks with investors, including the UAE, to raise $5-7 trillion for AI chip manufacturing to address the scarcity of GPUs crucial for training and running large language models.
  • As part of the talks, Altman is pitching a partnership between OpenAI, various investors, chip makers, and power providers to build chip foundries that would be run by existing chip makers, with OpenAI agreeing to be a significant customer.

In summary, OpenAI’s partnership with G42 aims to expand AI capabilities in the UAE and the Middle East, with Altman envisioning the UAE as a potential global AI sandbox.


AI Unraveled: Demystifying Frequently Asked Questions on Artificial Intelligence (OpenAI, ChatGPT, Google Gemini, Generative AI, Discriminative AI, xAI, LLMs, GPUs, Machine Learning, NLP, Promp Engineering)

Deepmind did not originally see LLMs and the transformer as a path to AGI. Fascinating article.

https://www.bigtechnology.com/p/can-demis-hassabis-save-google

It’s a very long article so I’ll post the relevant snippets. But basically it seems that Google was late to the LLM game because Demis Hassabis was 100% focused on AGI and did not see LLM’s as a path toward AGI. Perhaps now he sees it as a potential path, but it’s probably possible that he is just now focusing on LLM’s so that Google does not get too far behind in the generative AI race. But his ultimate goal and obsession is to create AGI that can solve real problems like diseases.

Within DeepMind, generative models weren’t taken seriously enough, according to those inside, perhaps because they didn’t align with Hassabis’s AGI priority, and weren’t close to reinforcement learning. Whatever the rationale, DeepMind fell behind in a key area.”

“‘We’ve always had amazing frontier work on self-supervised and deep learning,’ Hassabis tells me. ‘But maybe the engineering and scaling component — that we could’ve done harder and earlier. And obviously we’re doing that completely now.'”

“Kulkarni, the ex-DeepMind engineer, believes generative models were not respected at the time across the AI field, and simply hadn’t show enough promise to merit investment. ‘Someone taking the counter-bet had to pursue that path,’ he says. ‘That’s what OpenAI did.'”

“Ironically, a breakthrough within Google — called the transformer model — led to the real leap. OpenAI used transformers to build its GPT models, which eventually powered ChatGPT. Its generative ‘large language’ models employed a form of training called “self-supervised learning,” focused on predicting patterns, and not understanding their environments, as AlphaGo did. OpenAI’s generative models were clueless about the physical world they inhabited, making them a dubious path toward human level intelligence, but would still become extremely powerful.”

As DeepMind rejoiced, a serious challenge brewed beneath its nose. Elon Musk and Sam Altman founded OpenAI in 2015, and despite plenty of internal drama, the organization began working on text generation.”

“As OpenAI worked on the counterbet, DeepMind and its AI research counterpart within Google, Google Brain, struggled to communicate. Multiple ex-DeepMind employees tell me their division had a sense of superiority. And it also worked to wall itself off from the Google mothership, perhaps because Google’s product focus could distract from the broader AGI aims. Or perhaps because of simple tribalism. Either way, after inventing the transformer model, Google’s two AI teams didn’t immediately capitalize on it.”

“‘I got in trouble for collaborating on a paper with a Brain because the thought was like, well, why would you collaborate with Brain?’ says one ex-DeepMind engineer. ‘Why wouldn’t you just work within DeepMind itself?'”

“Then, a few months later, OpenAI released ChatGPT.” “At first, ChatGPT was a curiosity. The OpenAI chatbot showed up on the scene in late 2022 and publications tried to wrap their heads around its significance. […] Within Google, the product felt familiar to LaMDA, a generative AI chatbot the company had run internally — and even convinced one employee it was sentient — but never released. When ChatGPT became the fastest growing consumer product in history, and seemed like it could be useful for search queries, Google realized it had a problem on its hands.”

OpenAI reveals Voice Engine, but won’t yet publicly release the risky AI voice-cloning technology

OpenAI has released VoiceEngine, a voice-cloning tool. The company claims that it can recreate a person’s voice with just 15 seconds of recording of that person talking.

Source

A museum is using AI to let visitors chat with World War II survivors. [Source]

Meta to Add AI to Ray-Ban Smart Glasses. [Source]

Demis Hassabis, CEO and one of three founders of Google’s artificial intelligence (AI) subsidiary DeepMind, has been awarded a knighthood in the U.K. for “services to artificial intelligence.” [Source]

A daily chronicle of AI Innovations: March 30th, 2024: 🤯 Microsoft and OpenAI to build $100 billion AI supercomputer ‘Stargate’; 🗣 OpenAI unveils voice-cloning tool; 📈 Amazon’s AI team faces pressure to outperform Anthropic’s Claude models by mid-year;  🚫 Microsoft Copilot has been blocked on all Congress-owned devices

Microsoft and OpenAI to build $100 billion AI supercomputer ‘Stargate’

  • OpenAI and Microsoft are working on a $100 billion project to build an AI supercomputer named ‘Stargate’ in the U.S.
  • The supercomputer will house millions of GPUs and could cost over $115 billion.
  • Stargate is part of a series of datacenter projects planned by the two companies, with the goal of having it operational by 2028.
  • Microsoft will fund the datacenter, which is expected to be 100 times more costly than current operating centers.
  • The supercomputer is being built in phases, with Stargate being a phase 5 system.
  • Challenges include designing novel cooling systems and considering alternative power sources like nuclear energy.
  • OpenAI aims to move away from Nvidia’s technology and use Ethernet cables instead of InfiniBand cables.
  • Details about the location and structure of the supercomputer are still being finalized.
  • Both companies are investing heavily in AI infrastructure to advance the capabilities of AI technology.
  • Microsoft’s partnership with OpenAI is expected to deepen with the development of projects like Stargate.

Source

  • Microsoft and OpenAI are reportedly collaborating on a significant project to create a U.S.-based datacenter for an AI supercomputer named “Stargate,” estimated to cost over $115 billion and utilize millions of GPUs.
  • The supercomputer aims to be the largest among the datacenters planned by the two companies within the next six years, with Microsoft covering the costs and aiming for a launch by 2028.
  • The project, considered to be in phase 5 of development, requires innovative solutions for power, cooling, and hardware efficiency, including a possible shift away from relying on Nvidia’s InfiniBand in favor of Ethernet cables.
  • Source

🗣 OpenAI unveils voice-cloning tool

  • OpenAI has developed a text-to-voice generation platform named Voice Engine, capable of creating a synthetic voice from just a 15-second voice clip.
  • The platform is in limited access, serving entities like the Age of Learning and Livox, and is being used for applications from education to healthcare.
  • With concerns around ethical use, OpenAI has implemented usage policies, requiring informed consent and watermarking audio to ensure transparency and traceability.
  • Source

📈 Amazon’s AI team faces pressure to outperform Anthropic’s Claude models by mid-year

  • Amazon has invested $4 billion in AI startup Anthropic, but is also developing a competing large-scale language model called Olympus.
  • Olympus is supposed to surpass Anthropic’s latest Claude model by the middle of the year and has “hundreds of billions of parameters.”
  • So far, Amazon has had no success with its own language models. Employees are unhappy with Olympus’ development time and are considering switching to Anthropic’s models.
  • Source

🚫 Microsoft Copilot has been blocked on all Congress-owned devices

  • The US House of Representatives has banned its staff from using Microsoft’s AI chatbot Copilot due to cybersecurity concerns over potential data leaks.
  • Microsoft plans to remove Copilot from all House devices and is developing a government-specific version aimed at meeting federal security standards.
  • The ban specifically targets the commercial version of Copilot, with the House open to reassessing a government-approved version upon its release.
  • Source

Official NYC chatbot is encouraging small businesses to break the law.LINK

ChatGPT’s responses now include source references but for paid users.LINK

Next-generation AI semiconductor devices mimic the human brain.LINK

Voicecraft: I’ve never been more impressed in my entire life !

Voicecraft
Voicecraft

The maintainers of Voicecraft published the weights of the model earlier today, and the first results I get are incredible.

Here’s only one example, it’s not the best, but it’s not cherry-picked, and it’s still better than anything I’ve ever gotten my hands on !

Here’s the Github repository for those interested: https://github.com/jasonppy/VoiceCraft

A daily chronicle of AI Innovations: March 29th, 2024: 💥 Apple files lawsuit against former engineer for leaking details of projects he wanted to kill; Apple files lawsuit against former engineer for leaking details of projects he wanted to kill; Microsoft tackles Gen AI risks with new Azure AI tools; AI21 Labs’ Jamba triples AI throughput ; Google DeepMind’s AI fact-checker outperforms humans ; X’s Grok gets a major upgrade; Lightning AI partners with Nvidia to launch Thunder AI compiler

💥 Apple files lawsuit against former engineer for leaking details of projects he wanted to kill

  • Apple has filed a lawsuit against former employee Andrew Aude for leaking confidential information about products like the Vision Pro and Journal app to journalists and competitors, motivated by his desire to “kill” products and features he disagreed with.
  • Aude, who joined Apple in 2016, is accused of sharing sensitive details via encrypted messages and meetings, including over 10,000 text messages to a journalist from The Information.
  • The lawsuit seeks damages, the return of bonuses and stock options, and a restraining order against Aude for disclosing any more of Apple’s confidential information.
  • Source

👮‍♂️ Microsoft launches tools to try and stop people messing with chatbots

  • Microsoft has introduced a new set of tools in Azure to enhance the safety and security of generative AI applications, especially chatbots, aiming to counter risks like abusive content and prompt injections.
  • The suite includes features for real-time monitoring and protection against sophisticated threats, leveraging advanced machine learning to prevent direct and indirect prompt attacks.
  • These developments reflect Microsoft’s ongoing commitment to responsible AI usage, fueled by its significant investment in OpenAI and intended to address the security and reliability concerns of corporate leaders.
  • Source

AI21 Labs’ Jamba triples AI throughput

AI21 Labs has released Jamba, the first-ever production-grade AI model based on the Mamba architecture. This new architecture combines the strengths of both traditional Transformer models and the Mamba SSM, resulting in a model that is both powerful and efficient. Jamba boasts a large context window of 256K tokens, while still fitting on a single GPU.

AI21 Labs’ Jamba triples AI throughput
AI21 Labs’ Jamba triples AI throughput

Jamba’s hybrid architecture, composed of Transformer, Mamba, and mixture-of-experts (MoE) layers, optimizes for memory, throughput, and performance simultaneously.

The model has demonstrated remarkable results on various benchmarks, matching or outperforming state-of-the-art models in its size class. Jamba is being released with open weights under Apache 2.0 license and will be accessible from the NVIDIA API catalog.

Why does this matter?

Jamba’s hybrid architecture makes it the only model capable of processing 240k tokens on a single GPU. This could make AI tasks like machine translation and document analysis much faster and cheaper, without requiring extensive computing resources.

Source

Google DeepMind’s AI fact-checker outperforms humans

Google DeepMind has developed an AI system called Search-Augmented Factuality Evaluator (SAFE) that can evaluate the accuracy of information generated by large language models more effectively than human fact-checkers. In a study, SAFE matched human ratings 72% of the time and was correct in 76% of disagreements with humans.

Google DeepMind's AI fact-checker outperforms humans
Google DeepMind’s AI fact-checker outperforms humans

While some experts question the use of “superhuman” to describe SAFE’s performance, arguing for benchmarking against expert fact-checkers, the system’s cost-effectiveness is undeniable, being 20 times cheaper than human fact-checkers.

Why does this matter?

As language models become more powerful and widely used, SAFE could combat misinformation and ensure the accuracy of AI-generated content. SAFE’s efficiency could be a game-changer for consumers relying on AI for tasks like research and content creation.

Source

X’s Grok gets a major upgrade

X.ai, Elon Musk’s AI startup, has introduced Grok-1.5, an upgraded AI model for their Grok chatbot. This new version enhances reasoning skills, especially in coding and math tasks, and expands its capacity to handle longer and more complex inputs with a 128,000-token context window.

X’s Grok gets a major upgrade
X’s Grok gets a major upgrade

Grok chatbots are known for their ability to discuss controversial topics with a rebellious touch. The improved model will first be tested by early users on X, with plans for wider availability later. This release follows the open-sourcing of Grok-1 and the inclusion of the chatbot in X’s $8-per-month Premium plan.

Why does this matter?

This is significant because Grok-1.5 represents an advancement in AI assistants, potentially offering improved help with complex tasks and better understanding of user intent through its larger context window and real-time data ability. This could impact how people interact with chatbots in the future, making them more helpful and reliable.

Source

What Else Is Happening in AI on March 29th, 2024❗

🛡️Microsoft tackles Gen AI risks with new Azure AI tools

Microsoft has launched new Azure AI tools to address the safety and reliability risks associated with generative AI. The tools, currently in preview, aim to prevent prompt injection attacks, hallucinations, and the generation of personal or harmful content. The offerings include Prompt Shields, prebuilt templates for safety-centric system messages, and Groundedness Detection.  (Link)

🤝Lightning AI partners with Nvidia to launch Thunder AI compiler

Lightning AI, in collaboration with Nvidia, has launched Thunder, an open-source compiler for PyTorch, to speed up AI model training by optimizing GPU usage. The company claims that Thunder can achieve up to a 40% speed-up for training large language models compared to unoptimized code. (Link)

🥊SambaNova’s new AI model beats Databricks’ DBRX

SambaNova Systems’ Samba-CoE v0.2 Large Language Model outperforms competitors like Databricks’ DBRX, MistralAI’s Mixtral-8x7B, and xAI’s Grok-1. With 330 tokens per second using only 8 sockets, Samba-CoE v0.2 demonstrates remarkable speed and efficiency without sacrificing precision. (Link)

🌍Google.org launches Accelerator to empower nonprofits with Gen AI

Ace the Microsoft Azure Fundamentals AZ-900 Certification Exam: Pass the Azure Fundamentals Exam with Ease

Google.org has announced a six-month accelerator program to support 21 nonprofits in leveraging generative AI for social impact. The program provides funding, mentorship, and technical training to help organizations develop AI-powered tools in areas such as climate, health, education, and economic opportunity, aiming to make AI more accessible and impactful. (Link)

📱Pixel 8 to get on-device AI features powered by Gemini Nano

Google is set to introduce on-device AI features like recording summaries and smart replies on the Pixel 8, powered by its small-sized Gemini Nano model. The features will be available as a developer preview in the next Pixel feature drop, marking a shift from Google’s primarily cloud-based AI approach. (Link)

A daily chronicle of AI Innovations: March 28th, 2024: ⚡ DBRX becomes world’s most powerful open-source LLM 🏆 Claude 3 Opus crowned the top user-rated chatbot, beating OpenAI’s GPT-4 💙 Empathy meets AI: Hume AI’s EVI redefines voice interaction

DBRX becomes world’s most powerful open source LLM

Databricks has released DBRX, a family of open-source large language models setting a new standard for performance and efficiency.  The series includes DBRX Base and DBRX Instruct, a fine-tuned version designed for few-turn interactions. Developed by Databricks’ Mosaic AI team and trained using NVIDIA DGX Cloud, these models leverage an optimized mixture-of-experts (MoE) architecture based on the MegaBlocks open-source project. This architecture allows DBRX to achieve up to twice the compute efficiency of other leading LLMs.

DBRX becomes world’s most powerful open source LLM
DBRX becomes world’s most powerful open source LLM

In terms of performance, DBRX outperforms open-source models like Llama 2 70B, Mixtral-8x7B, and Grok-1 on industry benchmarks for language understanding, programming, and math. It also surpasses GPT-3.5 on most of these benchmarks, although it still lags behind GPT-4. DBRX is available under an open license with some restrictions and can be accessed through GitHub, Hugging Face, and major cloud platforms. Organizations can also leverage DBRX within Databricks’ Data Intelligence Platform.

Why does this matter?

With DBRX, organizations can build and fine-tune powerful proprietary models using their own internal datasets, ensuring full control over their data rights. As a result, DBRX is likely to accelerate the trend of organizations moving away from closed models and embracing open alternatives that offer greater control and customization possibilities.

Source

Claude 3 Opus crowned the top user-rated chatbot, beating OpenAI’s GPT-4

Anthropic’s Claude 3 Opus has overtaken OpenAI’s GPT-4 to become the top-rated chatbot on the Chatbot Arena leaderboard. This marks the first time in approximately a year since GPT-4’s release that another language model has surpassed it in this benchmark, which ranks models based on user preferences in randomized head-to-head comparisons. Anthropic’s cheaper Haiku and mid-range Sonnet models also perform impressively, coming close to the original GPT-4’s capabilities at a significantly lower cost.

Claude 3 Opus crowned the top user-rated chatbot, beating OpenAI’s GPT-4
Claude 3 Opus crowned the top user-rated chatbot, beating OpenAI’s GPT-4

While OpenAI still dominates the market, especially among regular users with ChatGPT, this development and recent leadership changes at OpenAI have helped Anthropic gain ground. However, OpenAI is rumored to be preparing to launch an even more advanced “GPT-4.5” or “GPT-5” model as soon as this summer, which CEO Sam Altman has teased will be “amazing,” potentially allowing them to retake the lead from Anthropic’s Claude 3 Opus.

Why does this matter?

If you are looking for an all-in-one solution to help you prepare for the AWS Cloud Practitioner Certification Exam, look no further than this AWS Cloud Practitioner CCP CLF-C02 book

Claude’s rise to the top of the Chatbot Arena leaderboard shows that OpenAI is not invincible and will face stiff competition in the battle for AI supremacy. With well-resourced challengers like Anthropic and Google, OpenAI will need to move fast and innovate boldly to maintain its top position. Ultimately, this rivalry will benefit everyone as it catalyzes the development of more powerful, capable, and hopefully beneficial AI systems that can help solve humanity’s major challenges.

Source

Empathy meets AI: Hume AI’s EVI redefines voice interaction

In a significant development for the AI community, Hume AI has introduced a new conversational AI called Empathic Voice Interface (EVI). What sets EVI apart from other voice interfaces is its ability to understand and respond to the user’s tone of voice, adding unprecedented emotional intelligence to the interaction. By adapting its language and responses based on the user’s expressions, EVI creates a more human-like experience, blurring the lines between artificial and emotional intelligence.

EVI’s empathic capabilities extend beyond just understanding tone. It can accurately detect the end of a conversation turn, handle interruptions seamlessly, and even learn from user reactions to improve over time. These features, along with its fast and reliable transcription and text-to-speech capabilities, make EVI a highly adaptable tool for various applications. Developers can easily integrate EVI into their projects using Hume’s API, which will be publicly available in April.

Why does this matter?

Emotionally intelligent AI can be revolutionary for industries like healthcare and use cases like customer support, where empathy and emotional understanding are crucial. But we must also consider potential risks, such as overreliance on AI for emotional support or the possibility of AI systems influencing users’ emotions in unintended ways. If developed and implemented ethically, emotionally intelligent AI can greatly enhance how we interact with and benefit from AI technologies in our daily lives.

Source

What Else Is Happening in AI on March 28th, 2024❗

💰 OpenAI launches revenue sharing program for GPT Store builders

OpenAI is experimenting with sharing revenue with builders who create successful apps using GPT in OpenAI’s GPT Store. The goal is to incentivize creativity and collaboration by rewarding builders for their impact on an ecosystem OpenAI is testing so they can make it easy for anyone to build and monetize AI-powered apps. (Link)

🛍️ Google introduces new shopping features to refine searches

Google is rolling out new shopping features that allow users to refine their searches and find items they like more easily. The Style Recommendations feature lets shoppers rate items in their searches, helping Google pick up on their preferences. Users can also specify their favorite brands to instantly bring up more apparel from those selections.  (Link)

🗣️ rabbit’s r1 device gets ultra-realistic voice powered by ElevenLabs

ElevenLabs has partnered with rabbit to integrate its high-quality, low-latency voice AI into rabbit’s r1 AI companion device. The collaboration aims to make the user experience with r1 more natural and intuitive by allowing users to interact with the device using voice commands. (Link)

💸 AI startup Hume raises $50M to build emotionally intelligent conversational AI

AI startup Hume has raised $50 million in a Series B funding round, valuing the company at $219 million. Hume’s AI technology can detect over 24 distinct emotional expressions in human speech and generate appropriate responses. The startup’s AI has been integrated into applications across healthcare, customer service, and productivity, with the goal of providing more context and empathy in AI interactions. (Link)

💻 Lenovo launches AI-enhanced PCs in a push for innovation and differentiation

Lenovo revealed a new lineup of AI-powered PCs and laptops at its Innovate event in Bangkok, Thailand. The company showcased the dual-screen Yoga Book 9i, Yoga Pro 9i with an AI chip for performance optimization and AI-enhanced Legion gaming laptops. Lenovo hopes to differentiate itself in the crowded PC market and revive excitement with these AI-driven innovations. (Link)

Study shows ChatGPT can produce medical record notes 10 times faster than doctors without compromising quality

The AI model ChatGPT can write administrative medical notes up to 10 times faster than doctors without compromising quality. This is according to a study conducted by researchers at Uppsala University Hospital and Uppsala University in collaboration with Danderyd Hospital and the University Hospital of Basel, Switzerland. The research is published in the journal Acta Orthopaedica.

Source

Microsoft Copilot AI will soon run locally on PCs

Microsoft’s Copilot AI service is set to run locally on PCs, Intel told Tom’s Hardware. The company also said that next-gen AI PCs would require built-in neural processing units (NPUs) with over 40 TOPS (trillion operations per second) of power — beyond the capabilities of any consumer processor on the market.

Intel said that the AI PCs would be able to run “more elements of Copilot” locally. Currently, Copilot runs nearly everything in the cloud, even small requests. That creates a fair amount of lag that’s fine for larger jobs, but not ideal for smaller jobs. Adding local compute capability would decrease that lag, while potentially improving performance and privacy as well.

Microsoft was previously rumored to require 40 TOPS on next-gen AI PCs (along with a modest 16GB of RAM). Right now, Windows doesn’t make much use of NPUs, apart from running video effects like background blurring for Surface Studio webcams. ChromeOS and macOS both use NPU power for more video and audio processing features, though, along with OCR, translation, live transcription and more, Ars Technica noted.

Source

A daily chronicle of AI Innovations: March 27th, 2024: 🔥 Microsoft study reveals the 11 by 11 tipping point for AI adoption 🤖 A16z spotlights the rise of generative AI in enterprises 🚨 Gaussian Frosting revolutionizes surface reconstruction in 3D modeling 🤖OpenAI unveils exciting upcoming features for GPT-4 and DALL-E 3 🤖 Adobe unveils GenStudio: AI-powered ad creation platform

Microsoft study reveals the 11 by 11 tipping point for AI adoption

Microsoft’s study on AI adoption in the workplace revealed the “11-by-11 tipping point,” where users start seeing AI’s value by saving 11 minutes daily. The study involved 1,300 Copilot for Microsoft 365 users and showed that 11 minutes of time savings is enough for most people to find AI useful.

Microsoft study reveals the 11 by 11 tipping point for AI adoption
Microsoft study reveals the 11 by 11 tipping point for AI adoption

Over 11 weeks, users reported improved productivity, work enjoyment, work-life balance, and fewer meetings. This “11-by-11 tipping point” signifies the time it takes for individuals to experience AI’s benefits in their work fully.

Why does it matter?

The study offers insights for organizations aiming to drive AI adoption among their employees. Businesses can focus on identifying specific use cases that deliver immediate benefits like time and cost savings. It will help organizations encourage employees to embrace AI, increasing productivity and improving work experiences.

Source

A16z spotlights the rise of generative AI in enterprises

A groundbreaking report by the influential tech firm a16z  unveils the rapid integration of generative AI technologies within the corporate sphere. The report highlights essential considerations for business leaders to harness generative AI effectively. It covers resource allocation, model selection, and innovative use cases, providing a strategic roadmap for enterprises.

A16z spotlights the rise of generative AI in enterprises
A16z spotlights the rise of generative AI in enterprises

An increased financial commitment from businesses marks the adoption of generative AI. Industry leaders are tripling their investments in AI technologies, emphasizing the pivotal role of generative AI in driving innovation and efficiency.

A16z spotlights the rise of generative AI in enterprises
A16z spotlights the rise of generative AI in enterprises

The shift towards integrating AI into core operations is evident. There is a focus on measuring productivity gains and cost savings and quantifying impact on key business metrics.

Why does it matter?

The increasing budgets allocated to generative AI signal its strategic importance in driving innovation and productivity in enterprises. This highlights AI’s transformative potential to provide a competitive edge and unlock new opportunities. Generative AI can revolutionize various business operations and help gain valuable insights by leveraging diverse data types.

Source

Gaussian Frosting revolutionizes surface reconstruction in 3D modeling

At the international conference on computer vision, researchers presented a new method to improve surface reconstruction using Gaussian Frosting. This technique automates the adjustment of Poisson surface reconstruction hyperparameters, resulting in significantly improved mesh reconstruction.

Gaussian Frosting revolutionalizes surface reconstruction in 3D modeling
Gaussian Frosting revolutionalizes surface reconstruction in 3D modeling

The method showcases the potential for scaling up mesh reconstruction while preserving intricate details and opens up possibilities for advanced geometry and texture editing. This work marks a significant step forward in surface reconstruction methods, promising advancements in 3D modeling and visualization techniques.

Why does it matter?

The new method demonstrates how AI enhances surface reconstruction techniques, improving mesh quality and enabling advanced editing in 3D modeling. This has significant implications for revolutionizing how 3D models are created, edited, and visualized across various industries.

Source

AIs can now learn and talk with each other like humans do.

This seems an important step toward AGI and vastly improved productivity.

“Once these tasks had been learned, the network was able to describe them to a second network — a copy of the first — so that it could reproduce them. To our knowledge, this is the first time that two AIs have been able to talk to each other in a purely linguistic way,’’ said lead author of the paper Alexandre Pouget, leader of the Geneva University Neurocenter, in a statement.”

“While AI-powered chatbots can interpret linguistic instructions to generate an image or text, they can’t translate written or verbal instructions into physical actions, let alone explain the instructions to another AI.

However, by simulating the areas of the human brain responsible for language perception, interpretation and instructions-based actions, the researchers created an AI with human-like learning and communication skills.”

Source

What Else Is Happening in AI on March 27th, 2024❗

🤖 Adobe unveils GenStudio: AI-powered ad creation platform

Adobe introduced GenStudio, an AI-powered ad creation platform, during its Summit event. GenStudio is a centralized hub for promotional campaigns, offering brand kits, copy guidance, and preapproved assets. It also provides generative AI-powered tools for generating backgrounds and ensuring brand consistency. Users can quickly create ads for email and social media platforms like Facebook, Instagram, and LinkedIn. (Link)

🧑‍💼Airtable introduces AI summarization for enhanced productivity

Airtable has introduced Airtable AI, which provides generative AI summarization, categorization, and translation to users. This feature allows quick insights and understanding of information within workspaces, enabling easy sharing of valuable insights with teams. Airtable AI automatically applies categories and tags to information, routes action items to the relevant team, and generates emails or social posts with a single button tap. (Link)

🤝Microsoft Teams enhances Copilot AI features for improved collaboration

Microsoft is introducing smarter Copilot AI features in Microsoft Teams to enhance collaboration and productivity. The updates include new ways to invoke the assistant during meeting chats and summaries, making it easier to catch up on missed meetings by combining spoken transcripts and written chats into a single view. Microsoft is launching new hybrid meeting features, such as automatic camera switching for remote participants and speaker recognition for accurate transcripts. (Link)

🤖OpenAI unveils exciting upcoming features for GPT-4 and DALL-E 3

OpenAI is preparing to introduce new features for its GPT-4 and DALL-E 3 models. For GPT-4, OpenAI plans to remove the message limit, implement a Model Tuner Selector, and allow users to upgrade responses from GPT-3.5 to GPT-4 with a simple button push. On the DALL-E 3 front, OpenAI is working on an image editor with inpainting functionality. These upcoming features demonstrate OpenAI’s commitment to advancing AI capabilities. (Link)

🔍Apple Chooses Baidu’s AI for iPhone 16 in China

Apple has reportedly chosen Baidu to provide AI technology for its upcoming iPhone 16 and other devices in China. This decision comes as Apple faces challenges due to stagnation in iPhone innovation and competition from Huawei. Baidu’s Ernie Bot will be included in the Chinese version of the iPhone 16, Mac OS, and iOS 18. Despite discussions with Alibaba Group Holding and a Tsinghua University AI startup, Apple selected Baidu’s AI technology for compliance. (Link)

Meta CEO, Mark Zuckerberg, is directly recruiting AI talent from Google’s DeepMind with personalized emails.

Meta CEO, Mark Zuckerberg, is attempting to recruit top AI talent from Google’s DeepMind (their AI research unit). Personalised emails, from Zuckerberg himself, have been sent to a few of their top researchers, according to a report from The Information, which cited individuals that had seen the messages. In addition to this, the researchers are being hired without having to do any interviews, and, a previous policy which Meta had in place – to not offer higher offers to candidates with competing job offers – has been relaxed.

Zuckerberg appears to be on a hiring spree to build Meta into a position of being a dominant player in the AI space.

OpenAI’s Sora Takes About 12 Minutes to Generate 1 Minute Video on NVIDIA H100. Source.

Apple on Tuesday announced that its annual developers conference, WWDC, will take place June 10 through June 14. Source.

Elon Musk says all Premium subscribers on X will gain access to AI chatbot Grok this week. Source.

Intel unveils AI PC program for software developers and hardware vendors. Source.

London-made HIV injection has potential to cure millions worldwide

Source

A daily chronicle of AI Innovations: March 26th, 2024 : 🔥 Zoom launches all-in-one modern AI collab platform; 🤖 Stability AI launches instruction-tuned LLM; 🚨 Stability AI CEO resigns to focus on decentralized AI; 🔍 WhatsApp to integrate Meta AI directly into its search bar; 🥊 Google, Intel, and Qualcomm challenge Nvidia’s dominance in AI; 🎬 OpenAI pitches Sora to Hollywood studios

Zoom launches all-in-one modern AI collab platform

Zoom launched Zoom Workplace, an AI collaboration platform that integrates many tools to improve teamwork and productivity. With over 40 new features, including AI Companion updates for Zoom Phone, Team Chat, Events, and Contact Center, as well as the introduction of Ask AI Companion, Zoom Workplace simplifies workflows within a familiar interface.

The platform offers customization options, meeting features, and improved collaboration tools across Zoom’s ecosystem. Zoom Business Services, integrated with Zoom Workplace, offers AI-driven marketing, customer service, and sales solutions. It expands digital communication channels and provides real-time insights for better agent management.

Why does this matter?

This intelligent platform will increase productivity by automating tasks, summarizing interactions, and personalizing user experiences. This move positions Zoom as a frontrunner in the race to integrate AI into everyday work tools, which will reshape how teams communicate and collaborate.

Source

Stability AI launches instruction-tuned LLM

Stability AI has introduced Stable Code Instruct 3B, a new instruction-tuned large language model. It can handle various software development tasks, such as code completion, generation, translation, and explanation, as well as creating database queries with simple instructions.

Stable Code Instruct 3B claims to outperform rival models like CodeLlama 7B Instruct and DeepSeek-Coder Instruct 1.3B in terms of accuracy, understanding natural language instructions, and handling diverse programming languages. The model is accessible for commercial use with a Stability AI Membership, while its weights are freely available on Hugging Face for non-commercial projects.

Why does this matter?

This model simplifies development workflows and complex tasks by providing contextual code completion, translation, and explanations. Businesses can prototype, iterate and ship software products faster thanks to its high performance and low hardware requirements.

Source

Stability AI CEO resigns because of centralized AI

  • Stability AI CEO Emad Mostaque steps down to focus on decentralized AI, advocating for transparent governance in the industry.

  • Mostaque’s departure follows the appointment of interim co-CEOs Shan Shan Wong and Christian Laforte.

  • The startup, known for its image generation tool, faced challenges including talent loss and financial struggles.

  • Mostaque emphasized the importance of generative AI R&D over revenue growth and highlighted the potential economic value of open models in regulated industries.

  • The AI industry witnessed significant changes with Inflection AI co-founders joining Microsoft after raising $1.5 billion.

Source

Estimating Sora’s power requirements

Quoting the compute estimates of Sora from the factorial funds blog

Estimating Sora's power requirements
Estimating Sora’s power requirements

A 15% penetration of Sora for videos with realistic video generation demand and utilization will require about 720k Nvidia H100 GPUs. Each H100 requires about 700 Watts of power supply.

720,000 x 700 = 504 Megawatts.

By comparison, even the largest ever fully solar powered plan in America (Ivanpah Solar Power Facility) produces about 377 Megawats.

While these power requirements can be met with other options like nuclear plants and even coal/hydro plants of big sizes … are we really entering the power game for electricity ?

( it is currently a power game on compute)

What Else Is Happening in AI on March 26th, 2024❗

💬 The Financial Times has introduced Ask FT, a new GenAI chatbot

It provides curated, natural-language responses to queries about recent events and broader topics covered by the FT. Ask FT is powered by Anthropic’s Claude and is available to a selected group of subscribers as it is under testing. (Link)

🔍 WhatsApp to integrate Meta AI directly into its search bar

The latest Android WhatsApp beta update will embed Meta AI directly into the search bar. This feature will allow users to type queries into the search bar and receive instant AI-powered responses without creating a separate Meta AI chat. The update will also allow users to interact with Meta AI even if they choose to hide the shortcut. (Link)

🥊 Google, Intel, and Qualcomm challenge Nvidia’s dominance in AI 

Qualcomm, Google, and Intel are targeting NVIDIA’s software platforms like CUDA. They plan to create open-source tools compatible with multiple AI accelerator chips through the UXL Foundation. Companies are investing over $4 billion in startups developing AI software to loosen NVIDIA’s grip on the field. (Link)

🤖 Apple takes a multi-vendor approach for generative AI in iOS 18

Apple is reportedly in talks with Alphabet, OpenAI, and Anthropic to integrate generative AI capabilities from multiple vendors into iOS 18. This multi-vendor approach aligns with Apple’s efforts to balance advanced AI features with privacy considerations, which are expected to be detailed at WWDC 2024 during the iOS 18 launch. (Link)

🎬 OpenAI pitches Sora to Hollywood studios

OpenAI is actively engaging with Hollywood studios, directors, and talent agencies to integrate Sora into the entertainment industry. The startup has scheduled meetings in Los Angeles to showcase Sora’s capabilities and encourage partnerships, with CEO Sam Altman attending events during the Oscars weekend. (Link)

LLM providers charge you per token, but their tokens are not always comparable. So if you are putting Python code through GPT-4 and Claude 3, it would cost you 25% more tokens to do so with Claude, due to difference in their tokenisers (note: this is different to cost per token, it just means you will have more tokens to pay for)

Some observations:
– OpenAI’s GPT-4 & 3.5 tokeniser is the most efficient for English and Python
– Gemini absolutely demolishes the competition in the three languages I tested: French (-11%), Chinese (-43%) and Hebrew (-54%)
– If your use cases is non-English, really worth looking at Gemini models – the difference in cost will likely be very noticeable
– Llama 2 ranked at the bottom of all of my tests
– Mistral was kind of disappointing on French (+16% worse than GPT), the reason why I picked French was that I assumed they’d do better

Methodology notes:
– The study will be limited, I only compared 7 individual bits of text/code – so results in practice would vary
– I have used this tokeniser playground (https://huggingface.co/spaces/Xenova/the-tokenizer-playground) for GPT, Mistral and Llama. I found it to be inaccurate (or old?) for Claude 3 and they didn’t have Gemini, so I did these separately
– Tokens are only part of the puzzle, more efficient tokenisation won’t necessarily mean better performance or overall lower cost
– If you want to learn about tokenisers, I recommend watching this video from Andrej Karpathy, even the first 10-20 minutes will be really worth your time https://www.youtube.com/watch?v=zduSFxRajkE

No alt text provided for this image

Source: Peter Gostev

A daily chronicle of AI Innovations: March 25th, 2024 : 🤝 Apple could partner with OpenAI, Gemini, Anthropic; 🤖 Chatbots more likely to change your mind than another human, study says; 🤖 Chatbots more likely to change your mind than another human, study says; Verbal Reasoning Test – Opus is better than 93% of people, Gemini 1.5 Pro 59%, GPT-4 Turbo only 36%; Apple’s Tim Cook says AI essential tool for businesses to reduce carbon footprint; Suno V3: Song-on-demand AI is getting insanely good; The first patient with a Neuralink brain-computer implant played Nintendo’s Mario Kart video game with his mind in an impressive new demo video

🤝 Apple could partner with OpenAI, Gemini, Anthropic

  • Apple is discussing with Alphabet, OpenAI, Anthropic, and potentially Baidu to integrate generative AI into iOS 18, considering multiple partners rather than a single one.
  • The collaboration could lead to a model where iPhone users might choose their preferred AI provider, akin to selecting a default search engine in a web browser.
  • Reasons for partnering with external AI providers include financial benefits, the possibility to quickly adapt through partnership changes or user preferences, and avoiding the complexities of developing and maintaining cloud-based generative AI in-house.
  • Source

🌐 EU probes Apple, Google, Meta under new digital law 

  • The European Commission has initiated five investigations into Apple, Google, and Meta for potential non-compliance with the Digital Markets Act (DMA), focusing on app store rules, search engine preferencing, and advertisement targeting models.
  • Investigations will also examine Apple’s app distribution fee structure and Amazon’s product preferencing, while Meta is given six months to make Messenger interoperable with other messaging services.
  • Companies may face fines up to 10% of their annual global revenue for DMA non-compliance, with the possibility of increased penalties for repeated infringements.
  • Source

🤖 Chatbots more likely to change your mind than another human, study says

  • A study found that personalized chatbots, such as GPT-4, are more likely to change people’s minds compared to human debaters by using tailored arguments based on personal information.
  • The research conducted by the École Polytechnique Fédérale de Lausanne and the Italian Fondazione Bruno Kessler showed an 81.7 percent increase in agreement when GPT-4 had access to participants’ personal data like age, gender, and race.
  • Concerns were raised about the potential misuse of AI in persuasive technologies, especially with the ability to generate detailed user profiles from online activities, urging online platform operators to counter such strategies.
  • Source

OpenAI CEO’s £142 Million Gamble On Unlocking the Secrets to Longer Life, Altman’s vision of extended lifespans may be achievable

Biotech startup Retro Biosciences is undertaking a one-of-a-kind experiment housed in shipping containers, funded by a $180 (£142.78) million investment by tech leader Sam Altman to increase lifespan.

Altman, the 38-year-old tech heavyweight, has been a significant player in the industry. Despite his young age, Altman took the tech realm by storm with offerings like ChatGPT and Sora. Unsurprisingly, his involvement in these groundbreaking projects has propelled him to a level of influence rivaling Mark Zuckerberg and Elon Musk, who is currently embroiled in a lawsuit with OpenAI.

It is also worth noting that the Altman-led AI startup is reportedly planning to launch its own AI-powered search engine to challenge Google’s search dominance. Altman’s visionary investments in tech giants like Reddit, Stripe, Airbnb, and Instacart propelled him to billionaire status. They cemented his influence as a tech giant who relentlessly pushed the boundaries of the industry’s future.

Source

Nvidia announces AI-powered health care 'agents' that outperform nurses — and cost $9 an hour
Nvidia announces AI-powered health care ‘agents’ that outperform nurses — and cost $9 an hour

Apple researchers explore dropping “Siri” phrase and listening with AI instead

  • Apple researchers are investigating the use of AI to identify when a user is speaking to a device without requiring a trigger phrase like ‘Siri’.

  • A study involved training a large language model using speech and acoustic data to detect patterns indicating the need for assistance from the device.

  • The model showed promising results, outperforming audio-only or text-only models as its size increased.

  • Eliminating the ‘Hey Siri’ prompt could raise concerns about privacy and constant listening by devices.

  • Apple’s handling of audio data has faced scrutiny in the past, leading to policy changes regarding user data and Siri recordings.

Source

Suno V3 can do multiple languages in one song. This one is English, Portuguese, Japanese, and Italian. Incredible.

Beneath the vast sky, where dreams lay rooted deep, Mountains high and valleys wide, secrets they keep. Ground beneath my feet, firm and ever true, Earth, you give us life, in shades of brown and green hue.

Sopra o vento, mensageiro entre o céu e o mar, Carregando sussurros, histórias a contar. Dançam as folhas, em um balé sem fim, Vento, o alento invisível, guiando o destino assim.

火のように、情熱が燃えて、 光と暖かさを私たちに与えてくれる。 夜の暗闇を照らす、勇敢な炎、 生命の力、絶えず変わるゲーム。

Acqua, misteriosa forza che tutto scorre, Nei fiumi, nei mari, la vita che ci offre. Specchio del cielo, in te ci riflettiamo, Acqua, fonte di vita, a te ci affidiamo.

Listen here

OpenAI Heading To Hollywood To Pitch Revolutionary “Sora”

Some of the most important meetings in Hollywood history will take place in the coming week, as OpenAI hits Hollywood to show the potential of its “Sora” software to studios, talent agencies, and media executives.

Bloomberg is reporting that OpenAI wants more filmmakers to become familiar with Sora, the text-to-video generator that potentially could upend the way movies are made.

Soon, Everyone Will Own a Robot, Like a Car or Phone Today. Says Figure AI founder

Brett Adcock, the founder of FigureAI robots, the company that recently released a demo video of its humanoid robot conversing with a human while performing tasks, predicts that everyone will own a robot in the future. “Similar to owning a car or phone today,” he said – hinting at the universal adoption of robots as an essential commodity in the future.

“Every human will own a robot in the future, similar to owning a car/phone today,” said Adcock.

A few months ago, Adcock called 2024 the year of Embodied AI, indicating how the future comprises AI in a body form. With robots learning to perform low-complexity tasks, such as picking trash, placing dishes, and even using the coffee machine, Figure robots are being trained to assist a person with house chores.

Source

WhatsApp to embed Meta AI directly into search bar for instant assistance: Report. 

WhatsApp is on the brink of a transformation in user interaction as it reportedly plans to integrate Meta AI directly into its search bar. This move promises to simplify access to AI assistance within the app, eliminating the need for users to navigate to a separate Meta AI conversation.

WhatsApp to embed Meta AI directly into search bar for instant assistance
WhatsApp to embed Meta AI directly into search bar for instant assistance

Source

How People are really using Gen AI

Top-level themes:

1️⃣ Technical Assistance & Troubleshooting (23%)
2️⃣ Content Creation & Editing (22%)
3️⃣ Personal & Professional Support (17%)
4️⃣ Learning & Education (15%)
5️⃣ Creativity & Recreation (13%)
6️⃣ Research, Analysis & Decision Making (10%)

What users are doing:

✔Generating ideas
✔Specific search
✔Editing text
✔Drafting emails
✔Simple explainers
✔Excel formulas
✔Sampling data

🤔 Do you see AI as a tool to enhance your work, or as a threat that could take over your job?

Source: HBR
Image credit: Filtered

How People are really using Gen AI
How People are really using Gen AI

A daily chronicle of AI Innovations: March 22nd, 2024 : 🤖 Nvidia’s Latte 3D generates text-to-3D in seconds! 💰 Saudi Arabia to invest $40 billion in AI 🚀 Open Interpreter’s 01 Light personal pocket AI agent. 🤖 Microsoft introduces a new Copilot for better productivity.
💡Quiet-STaR: LMs can self-train to think before responding
🤯Neuralink’s first brain chip patient plays chess with his mind

Experience the transformative capabilities of AI with “Read Aloud For Me – AI Dashboard” – your ultimate AI Dashboard and Hub.

Nvidia’s Latte 3D generates text-to-3D in seconds!

NVIDIA introduces Latte3D, facilitating the conversion of text prompts into detailed 3D models in less than a second. Developed by NVIDIA’s Toronto lab, Latte3D sets a new standard in generative AI models with its remarkable blend of speed and precision.

Nvidia’s Latte 3D generates text-to-3D in seconds!
Nvidia’s Latte 3D generates text-to-3D in seconds!

LATTE3D has two stages: first, NVIDIA’s team uses volumetric rendering to train the texture and geometry robustly, and second, it uses surface-based rendering to train only the texture for quality enhancement. Both stages use amortized optimization over prompts to maintain fast generation.

Nvidia’s Latte 3D generates text-to-3D in seconds!
Nvidia’s Latte 3D generates text-to-3D in seconds!

What sets Latte3D apart is its extensive pretraining phase, enabling the model to quickly adapt to new tasks by drawing on a vast repository of learned patterns and structures. This efficiency is achieved through a rigorous training regime that includes a blend of 3D datasets and prompts from ChatGPT.

Why does it matter?

AI models such as NVIDIA’s Latte3D have significantly reduced the time required to generate 3D visualizations from an hour to a few minutes compared to a few years ago. This technology has the potential to significantly accelerate the design and development process in various fields, such as the video game industry, advertising, and more.

Source

Quiet-STaR: LMs can self-train to think before responding

A groundbreaking study demonstrates the successful training of large language models (LM) to reason from text rather than specific reasoning tasks. The research introduces a novel training approach, Quiet STaR, which utilizes a parallel sampling algorithm to generate rationales from all token positions in a given string.

Quiet-STaR: LMs can self-train to think before responding
Quiet-STaR: LMs can self-train to think before responding

This technique integrates meta tokens to indicate when the LM should generate a rationale and when it should make a prediction based on the rationale, revolutionizing the understanding of LM behavior. Notably, the study shows that thinking enables the LM to predict difficult tokens more effectively, leading to improvements with longer thoughts.

The research introduces powerful advancements, such as a non-myopic loss approach, the application of a mixing head for retrospective determination, and the integration of meta tokens, underpinning a comprehensive leap forward in language model training.

Why does it matter?

These significant developments in language modeling advance the field and have the potential to revolutionize a wide range of applications. This points towards a future where large language models will unprecedentedly contribute to complex reasoning tasks.

Source

Neuralink’s first brain chip patient plays chess with his mind

Elon Musk’s brain chip startup, Neuralink, showcased its first brain chip patient playing chess using only his mind. The patient, Noland Arbaugh, was paralyzed below the shoulder after a diving accident.

Neuralink’s brain implant technology allows people with paralysis to control external devices using their thoughts. With further advancements, Neuralink’s technology has the potential to revolutionize the lives of people with paralysis, providing them with newfound independence and the ability to interact with the world in previously unimaginable ways.

Why does it matter?

Neuralink’s brain chip holds significant importance in AI and human cognition. It has the potential to enhance communication, assist paralyzed individuals, merge human intelligence with AI, and address the risks associated with AI development. However, ethical considerations and potential misuse of this technology must also be carefully examined.

Source

What Else Is Happening in AI on March 22nd, 2024❗

🤖 Microsoft introduces a new Copilot for better productivity.

Microsoft’s new Copilot for Windows and Surface devices is a powerful productivity tool integrating large language models with Microsoft Graph and Microsoft 365 apps to enhance work efficiency. With a focus on delivering AI responsibly while ensuring data security and privacy, Microsoft is dedicated to providing users with innovative tools to thrive in the evolving work landscape. (Link)

💰 Saudi Arabia to invest $40 billion in AI

Saudi Arabia has announced its plan to invest $40 billion in AI to become a global leader. Middle Eastern countries use their sovereign wealth fund, which has over $900 billion in assets, to achieve this goal. This investment aims to position the country at the forefront of the fast-evolving AI sector, drive innovation, and enhance economic growth. (Link)

🎧 Rightsify releases Hydra II to revolutionize AI music generation

Rightsify, a global music licensing leader, introduced Hydra II, the latest AI generation model. Hydra II offers over 800 instruments, 50 languages, and editing tools for customizable, copyright-free AI music. The model is trained on audio, text descriptions, MIDI, chord progressions, sheet music, and stems to create unique generations. (Link)

🚀 Open Interpreter’s 01 Light personal pocket AI agent

The Open Interpreter unveiled 01 Light, a portable device that allows you to control your computer using natural language commands. It’s part of an open-source project to make computing more accessible and flexible. It’s designed to make your online tasks more manageable, helping you get more done and simplify your life. (Link)

🤝 Microsoft’s $650 million Inflection deal: A strategic move
Microsoft has recently entered into a significant deal with AI startup Inflection, involving a payment of $650 million in cash. While the deal may seem like a licensing agreement, it appears to be a strategic move by Microsoft to acquire AI talent while avoiding potential regulatory trouble. (Link)

Microsoft unveiled its first “AI PCs,” with a dedicated Copilot key and Neural Processing Units (NPUs).

Microsoft unveiled its first "AI PCs," with a dedicated Copilot key and Neural Processing Units (NPUs).
Microsoft unveiled its first “AI PCs,” with a dedicated Copilot key and Neural Processing Units (NPUs).

Source: Nvidia

OpenAI Courts Hollywood in Meetings With Film Studios, Directors – from Bloomberg

The artificial intelligence startup has scheduled meetings in Los Angeles next week with Hollywood studios, media executives and talent agencies to form partnerships in the entertainment industry and encourage filmmakers to integrate its new AI video generator into their work, according to people familiar with the matter.

The upcoming meetings are just the latest round of outreach from OpenAI in recent weeks, said the people, who asked not to be named as the information is private. In late February, OpenAI scheduled introductory conversations in Hollywood led by Chief Operating Officer Brad Lightcap. Along with a couple of his colleagues, Lightcap demonstrated the capabilities of Sora, an unreleased new service that can generate realistic-looking videos up to about a minute in length based on text prompts from users. Days later, OpenAI Chief Executive Officer Sam Altman attended parties in Los Angeles during the weekend of the Academy Awards.

In an attempt to avoid defeatism, I’m hoping this will contribute to the indie boom with creatives refusing to work with AI and therefore studios who insist on using it. We’ve already got people on twitter saying this is the end of the industry but maybe only tentpole films as we know them.

Here’s the article without the paywall.

Catherine, the Princess of Wales, has cancer, she announced in a video message released by Kensington Palace on Friday March 22nd, 2024

The recent news surrounding Kate Middleton, the Princess of Wales, revolves around a manipulated family photo that sparked controversy and conspiracy theories. The photo, released by Middleton herself, depicted her with her three children and was met with speculation about potential AI involvement in its editing. However, experts suggest that the image was likely manipulated using traditional photo editing software like Photoshop rather than generative AI

The circumstances surrounding Middleton’s absence from the public eye due to abdominal surgery fueled rumors and intensified scrutiny over the edited photo.

Major news agencies withdrew the image, citing evidence of manipulation in areas like Princess Charlotte’s sleeve cuff and the alignment of elements in the photo.

Despite concerns over AI manipulation, this incident serves as a reminder that not all image alterations involve advanced technology, with this case being attributed to a botched Photoshop job.From an AI perspective, experts highlight how the incident reflects society’s growing awareness of AI technologies and their impact on shared reality. The controversy surrounding the edited photo underscores the need for transparency and accountability in media consumption to combat misinformation and maintain trust in visual content. As AI tools become more accessible and sophisticated, distinguishing between authentic and manipulated media becomes increasingly challenging, emphasizing the importance of educating consumers and technologists on identifying AI-generated content.Kate Middleton, the Princess of Wales, recently disclosed her battle with cancer in a heartfelt statement. Following major abdominal surgery in January, it was initially believed that her condition was non-cancerous. However, subsequent tests revealed the presence of cancer, leading to the recommendation for preventative chemotherapy. The 42-year-old princess expressed gratitude for the support received during this challenging time and emphasized the importance of privacy as she focuses on her treatment and recovery. The news of her diagnosis has garnered an outpouring of support from around the world, with messages of encouragement coming from various public figures and officials.

Nvidia CEO says we’ll see fully AI-generated games in 5-10 years

Nvidia’s CEO, Jensen Huang, predicts the emergence of fully AI-generated games within the next five to ten years. This prediction is based on the development of Nvidia’s next-generation Blackwell AI GPU, the B200. This GPU marks a significant shift in GPU usage towards creating neural networks for generating content rather than traditional rasterization or ray tracing for visual fidelity in games. The evolution of AI in gaming is highlighted as GPUs transition from rendering graphics to processing AI algorithms for content creation, indicating a major transformation in the gaming industry’s future landscape.The integration of AI into gaming represents a paradigm shift that could revolutionize game development and player experiences. Fully AI-generated games have the potential to offer unprecedented levels of customization, dynamic storytelling, and adaptive gameplay based on individual player interactions. This advancement hints at a new era of creativity and innovation in game design but also raises questions about the ethical implications and challenges surrounding AI-generated content, such as ensuring diversity, fairness, and avoiding biases in virtual worlds. Source

Andrew Ng, cofounder of Google Brain & former chief scientist @ Baidu- “I think AI agentic workflows will drive massive AI progress this year — perhaps even more than the next generation of foundation models.

This is an important trend, and I urge everyone who works in AI to pay attention to it.”

AI agentic workflows will drive massive AI progress
AI agentic workflows will drive massive AI progress

I think AI agentic workflows will drive massive AI progress this year — perhaps even more than the next generation of foundation models. This is an important trend, and I urge everyone who works in AI to pay attention to it.

Today, we mostly use LLMs in zero-shot mode, prompting a model to generate final output token by token without revising its work. This is akin to asking someone to compose an essay from start to finish, typing straight through with no backspacing allowed, and expecting a high-quality result. Despite the difficulty, LLMs do amazingly well at this task!

With an agentic workflow, however, we can ask the LLM to iterate over a document many times. For example, it might take a sequence of steps such as:

  • Plan an outline.

  • Decide what, if any, web searches are needed to gather more information.

  • Write a first draft.

  • Read over the first draft to spot unjustified arguments or extraneous information.

  • Revise the draft taking into account any weaknesses spotted.

  • And so on.

This iterative process is critical for most human writers to write good text. With AI, such an iterative workflow yields much better results than writing in a single pass.

Devin’s splashy demo recently received a lot of social media buzz. My team has been closely following the evolution of AI that writes code. We analyzed results from a number of research teams, focusing on an algorithm’s ability to do well on the widely used HumanEval coding benchmark. You can see our findings in the diagram below.

GPT-3.5 (zero shot) was 48.1% correct. GPT-4 (zero shot) does better at 67.0%. However, the improvement from GPT-3.5 to GPT-4 is dwarfed by incorporating an iterative agent workflow. Indeed, wrapped in an agent loop, GPT-3.5 achieves up to 95.1%.

Open source agent tools and the academic literature on agents are proliferating, making this an exciting time but also a confusing one. To help put this work into perspective, I’d like to share a framework for categorizing design patterns for building agents. My team AI Fund is successfully using these patterns in many applications, and I hope you find them useful.

  • Reflection: The LLM examines its own work to come up with ways to improve it.

  • Tool use: The LLM is given tools such as web search, code execution, or any other function to help it gather information, take action, or process data.

  • Planning: The LLM comes up with, and executes, a multistep plan to achieve a goal (for example, writing an outline for an essay, then doing online research, then writing a draft, and so on).

  • Multi-agent collaboration: More than one AI agent work together, splitting up tasks and discussing and debating ideas, to come up with better solutions than a single agent would.

  • Source

A daily chronicle of AI Innovations: March 21st, 2024 : 🕵️‍♂️ Stealing Part of a Production Language Model
🤖 Sakana AI’s method to automate foundation model development
👋 Key Stable Diffusion researchers leave Stability AI  🗣️Character AI’s new feature adds voice to characters with just 10-sec audio 💡Fitbit to get major AI upgrades powered by Google’s ‘Personal Health’ LLM 🔬Samsung creates lab to research chips for AI’s next phase 🤖GitHub’s latest AI tool can automatically fix code vulnerabilities

Stealing Part of a Production Language Model

Researchers from Google, OpenAI, and DeepMind (among others) released a new paper that introduces the first model-stealing attack that extracts precise, nontrivial information from black-box production language models like OpenAI’s ChatGPT or Google’s PaLM-2.

The attack allowed them to recover the complete embedding projection layer of a transformer language model. It differs from prior approaches that reconstruct a model in a bottom-up fashion, starting from the input layer. Instead, this operates top-down and directly extracts the model’s last layer by making targeted queries to a model’s API. This is useful for several reasons; it

  • Reveals the width of the transformer model, which is often correlated with its total parameter count.
  • Slightly reduces the degree to which the model is a complete “blackbox”
  • May reveal more global information about the model, such as relative size differences between different models

While there appear to be no immediate practical consequences of learning this layer is stolen, it represents the first time that any precise information about a deployed transformer model has been stolen.

Stealing Part of a Production Language Model
Stealing Part of a Production Language Model

Why does this matter?

Though it has limitations, the paper motivates the further study of practical attacks on ML models, in order to ultimately develop safer and more reliable AI systems. It also highlights how small, system-level design decisions impact the safety and security of the full product.

Source

Sakana AI’s method to automate foundation model development

Sakana AI has introduced Evolutionary Model Merge, a general method that uses evolutionary techniques to efficiently discover the best ways to combine different models from the vast ocean of different open-source models with diverse capabilities.

As of writing, Hugging Face has over 500k models in dozens of different modalities that, in principle, could be combined to form new models with new capabilities. By working with the vast collective intelligence of existing open models, this method is able to automatically create new foundation models with desired capabilities specified by the user.

Why does this matter?

Model merging shows great promise and democratizes up model-building. In fact, the current Open LLM Leaderboard is dominated by merged models. They work without any additional training, making it very cost-effective. But we need a more systematic approach.

Evolutionary algorithms, inspired by natural selection, can unlock more effective merging. They can explore vast possibilities, discovering novel and unintuitive combinations that traditional methods and human intuition might miss.

Source

Key Stable Diffusion researchers leave Stability AI

Robin Rombach and other key researchers who helped develop the Stable Diffusion text-to-image generation model have left the troubled, once-hot, now floundering GenAI startup.

Rombach (who led the team) and fellow researchers Andreas Blattmann and Dominik Lorenz were three of the five authors who developed the core Stable Diffusion research while at a German university. They were hired afterwards by Stability. Last month, they helped publish a 3rd edition of the Stable Diffusion model, which, for the first time, combined the diffusion structure used in earlier versions with transformers used in OpenAI’s ChatGPT.

Their departures are the latest in a mass exodus of executives at Stability AI, as its cash reserves dwindle and it struggles to raise additional funds.

Why does this matter?

Stable Diffusion is one of the foundational models that helped catalyze the boom in generative AI imagery, but now its future hangs in the balance. While Stability AI’s current situation raises questions about its long-term viability, the exodus potentially benefits its competitors.

Source

What Else Is Happening in AI on March 21st, 2024❗

🗣️Character AI’s new feature adds voice to characters with just 10-sec audio

You can now give voice to your Characters by choosing from thousands of voices or creating your own. The voices are created with just 10 seconds of audio clips. The feature is now available for free to everyone. (Link)

🤖GitHub’s latest AI tool can automatically fix code vulnerabilities

GitHub launches the first beta of its code-scanning autofix feature, which finds and fixes security vulnerabilities during the coding process. GitHub claims it can remediate more than two-thirds of the vulnerabilities it finds, often without the developers having to edit the code. The feature is now available for all GitHub Advanced Security (GHAS) customers. (Link)

GitHub’s latest AI tool can automatically fix code vulnerabilities
GitHub’s latest AI tool can automatically fix code vulnerabilities

🚀OpenAI plans to release a ‘materially better’ GPT-5 in mid-2024

According to anonymous sources from Businessinsider, OpenAI plans to release GPT-5 this summer, which will be significantly better than GPT-4. Some enterprise customers are said to have already received demos of the latest model and its ChatGPT improvements. (Link)

💡Fitbit to get major AI upgrades powered by Google’s ‘Personal Health’ LLM

Google Research and Fitbit announced they are working together to build a Personal Health LLM that gives users more insights and recommendations based on their data in the Fitbit mobile app. It will give Fitbit users personalized coaching and actionable insights that help them achieve their fitness and health goals. (Link)

🔬Samsung creates lab to research chips for AI’s next phase

Samsung has set up a research lab dedicated to designing an entirely new type of semiconductor needed for (AGI). The lab will initially focus on developing chips for LLMs with a focus on inference. It aims to release new “chip designs, an iterative model that will provide stronger performance and support for increasingly larger models at a fraction of the power and cost.” (Link)

A daily chronicle of AI Innovations: March 20th, 2024 : 🤖 OpenAI to release GPT-5 this summer; 🧠 Nvidia’s Jensen Huang says AI hallucinations are solvable, AGI is 5 years away; 🔬 Ozempic creator plans AI supercomputer to discover new drugs; 👀 After raising $1.3B, Inflection eaten alive by Microsoft; 🧠 MindEye2: AI Mind Reading from Brain Activity; 🚀 Nvidia NIM enables faster deployment of AI models

🤖 OpenAI to release GPT-5 this summer

  • OpenAI is planning to launch GPT-5 around mid-year, aiming to address previous performance issues and significantly improve upon its predecessor, GPT-4.
  • GPT-5 is described as “materially better” by those who have seen demos, including enhancements and new capabilities like the ability to call AI agents for autonomous tasks, with enterprise customers having already previewed these improvements.
  • The release timeline for GPT-5 remains uncertain as OpenAI continues its training and thorough safety and vulnerability testing, with no specific deadline for completion of these preparatory steps.
  • Source

👀 After raising $1.3B, Inflection eaten alive by Microsoft 

  • In June 2023, Inflection raised $1.3 billion led by Microsoft to develop “more personal AI” but was overtaken by Microsoft less than a year later, with co-founders joining Microsoft’s new AI division.
  • Despite significant investment, Inflection’s AI, Pi, failed to compete with advancements from other companies such as OpenAI, Google’s Gemini, and Anthropic, leading to its downfall.
  • Microsoft’s takeover of Inflection reflects the strategy of legacy tech companies to dominate the AI space by supporting startups then acquiring them once they face challenges.
  • Source

🧠 Nvidia’s Jensen Huang says AI hallucinations are solvable, AGI is 5 years away

  • Nvidia CEO Jensen Huang predicts artificial general intelligence (AGI) could be achieved within 5 years, depending on how AGI is defined and measured.
  • Huang addresses concerns around AI hallucinations, suggesting that ensuring answers are well-researched could easily solve the issue.
  • The concept of AGI raises concerns about its potential unpredictability and the challenges of aligning its objectives with human values and priorities.
  • Source

🔬 Ozempic creator plans AI supercomputer to discover new drugs

  • The Novo Nordisk Foundation is investing in “Gefion,” an AI supercomputer project developed in collaboration with Nvidia.
  • “Gefion” aims to be the world’s most powerful AI supercomputer for health sciences, utilizing Nvidia’s new chips to accelerate scientific breakthroughs in critical areas such as drug discovery, disease diagnosis, and treatment,
  • This initiative underscores the growing integration of AI in healthcare, promising to catalyze significant scientific discoveries and innovations that could transform patient care and outcomes.
  • Source

MindEye2: AI mind reading from brain activity

MindEye2 is a revolutionary model that reconstructs visual perception from brain activity using just one hour of data. Traditional methods require extensive training data, making them impractical for real-world applications. However, MindEye2 overcomes this limitation by leveraging shared-subject models. The model is pretrained on data from seven subjects and then fine-tuned with minimal data from a new subject.

MindEye2: AI mind reading from brain activity
MindEye2: AI mind reading from brain activity

By mapping brain activity to a shared-subject latent space and then nonlinear mapping to CLIP image space, MindEye2 achieves high-quality reconstructions with limited training data. It performs state-of-the-art image retrieval and reconstruction across multiple subjects within only 2.5% of the previously required training data, reducing the training time from 40 to just one hour.

Why does it matter?

MindEye2 has the potential to revolutionize clinical assessments and brain-computer interface applications. This remarkable achievement also holds great promise for neuroscience and opens new possibilities for understanding how our brains perceive and process visual information. It can also help develop personalized treatment plans for neuro patients.

Source

Nvidia NIM enables faster deployment of AI models 

NVIDIA has introduced NVIDIA NIM (NVIDIA Inference Microservices) to accelerate the deployment of AI applications for businesses. NIM is a collection of microservices that package essential components of an AI application, including AI models, APIs, and libraries, into a container. These containers can be deployed in environments such as cloud platforms, Linux servers, or serverless architectures.

Nvidia NIM enables faster deployment of AI models 
Nvidia NIM enables faster deployment of AI models

NIM significantly reduces the time it takes to deploy AI applications from weeks to minutes. It offers optimized inference engines, industry-standard APIs, and support for popular software and data platform vendors. NIM microservices are compatible with NVIDIA GPUs and support features like Retrieval Augmented Generation (RAG) capabilities for enhanced enterprise applications. Developers can experiment with NIM microservices for free on the ai.nvidia.com platform, while commercial deployment is available through NVIDIA AI Enterprise 5.0.

Why does it matter?

With NIM, Nvidia is trying to democratize AI deployment for enterprises by abstracting away complexities. This will enable more developers to contribute to their company’s AI transformation efforts and allow businesses to run AI applications almost instantly without specialized AI expertise.

Source

Microsoft hires DeepMind co-founder to lead a new AI division

Mustafa Suleyman, a renowned co-founder of DeepMind and Inflection, has recently joined Microsoft as the leader of Copilot. Satya Nadella, Microsoft’s CEO, made this significant announcement, highlighting the importance of innovation in artificial intelligence (AI).

In his new role as the Executive Vice President and CEO of Microsoft AI, Mustafa will work alongside Karén Simonyan, another talented individual from Inflection who will serve as Chief Scientist. Together, they will spearhead the development and advancement of Copilot and other exciting consumer AI products at Microsoft. Mustafa and his team’s addition to the Microsoft family brings a wealth of expertise and promises groundbreaking advancements in AI.

Why does it matter?

Mustafa Suleyman’s expertise in AI is expected to contribute to the development of innovative consumer AI products and research at Microsoft, furthering its mission to bring the benefits of AI to people and organizations worldwide. With DeepMind’s founder now at the helm, the AI race between Microsoft, Google, and others became even more intense.

Source

What Else Is Happening in AI on March 20th, 2024❗

📞 Truecaller adds AI-powered spam detection and blocking for Android users

Truecaller has unveiled a new feature for its Android premium subscribers that uses AI to detect spam, even if unavailable on the Truecaller database, and block every call that doesn’t come from an approved contact. Truecaller hopes to add more premium subscribers to its list by adding this feature. However, this feature is not available for Apple users. (Link)

⚽ Google DeepMind’s new AI tool can analyze soccer tactics and offer insights 

DeepMind has partnered with Liverpool FC to develop a new AI tool called TacticAI. TacticAI uses generative and predictive AI to help coaches determine which player will most likely receive the ball during corner kicks, whether a shot will be taken, and how to adjust player setup. It aims to revolutionize soccer and help the teams enhance their efficiency. (Link)

🎬 Pika Labs introduces sound effects for its gen-AI video generation

Pika Labs has now added the ability to create sound effects from a text prompt for its generative artificial intelligence videos. It allows for automatic or custom SFX generations to pair with video outputs. Now, users can make bacon sizzle, lions roar, or add footsteps to the video of someone walking down the street. It is only available to pro users. (Link)

🎮 Buildbox 4 Alpha enables users to create 3D video games from text prompts 

Buildbox has released an alpha version of Buildbox 4. It’s an AI-first game engine that allows users to create games and generate assets from text prompts. The alpha version aims to make text-to-game a distinct reality. Users can create various assets and animations from simple text prompts. It also allows users to build a gaming environment in a few minutes. (Link)

🤖 Nvidia adds generative AI capabilities to empower humanoid robots

Nvidia introduced Project GR00T, a multimodal AI that will power future humanoids with advanced foundation AI. Project GR00T enables humanoid robots to input text, speech, videos, or even live demos and process them to take specific actions. It has been developed with the help of Nvidia’s Isaac Robotic Platform tools, including an Isaac Lab for RLHF. (Link)

The EU AI Act – Key takeaways for LLM builders

The EU AI Act - Key takeaways for LLM builders

A daily chronicle of AI Innovations: March 19th, 2024 : 💻 Nvidia launches ‘world’s most powerful AI chip’; 🎥 Stability AI’s SV3D turns a single photo into a 3D video; 🤖 OpenAI CEO hints at “Amazing Model”, maybe ChatGPT-5 ;🤝 Apple is in talks to bring Google’s AI to iPhones

Nvidia launches ‘world’s most powerful AI chip’

Nvidia has revealed its new Blackwell B200 GPU and GB200 “superchip”, claiming it to be the world’s most powerful chip for AI. Both B200 and GB200 are designed to offer powerful performance and significant efficiency gains.

Nvidia launches 'world's most powerful AI chip'
Nvidia launches ‘world’s most powerful AI chip’

Key takeaways:

  • The B200 offers up to 20 petaflops of FP4 horsepower, and Nvidia says it can reduce costs and energy consumption by up to 25 times over an H100.
  • The GB200 “superchip” can deliver 30X the performance for LLM inference workloads while also being more efficient.
  • Nvidia claims that just 2,000 Blackwell chips working together could train a GPT -4-like model comprising 1.8 trillion parameters in just 90 days.

Why does this matter?

A major leap in AI hardware, the Blackwell GPU boasts redefined performance and energy efficiency. This could lead to lower operating costs in the long run, making high-performance computing more accessible for AI research and development, all while promoting eco-friendly practices.

Source

Stability AI’s SV3D turns a single photo into a 3D video

Stability AI released Stable Video 3D (SV3D), a new generative AI tool for rendering 3D videos. SV3D can create multi-view 3D models from a single image, allowing users to see an object from any angle. This technology is expected to be valuable in the gaming sector for creating 3D assets and in e-commerce for generating 360-degree product views.

SV3D builds upon Stability AI’s previous Stable Video Diffusion model. Unlike prior methods, SV3D can generate consistent views from any given angle. It also optimizes 3D meshes directly from the novel views it produces.

SV3D comes in two variants: SV3D_u generates orbital videos from single images, and SV3D_p creates 3D videos along specified camera paths.

Why does this matter?

SV3D represents a significant leap in generative AI for 3D content. Its ability to create 3D models and videos from a single image could open up possibilities in various fields, such as animation, virtual reality, and scientific modeling.

Source

OpenAI CEO hints at “Amazing Model,” maybe ChatGPT-5

OpenAI CEO Sam Altman has announced that the company will release an “amazing model” in 2024, although the name has not been finalized. Altman also mentioned that OpenAI plans to release several other important projects before discussing GPT-5, one of which could be the Sora video model.

Source

Altman declined to comment on the Q* project, which is rumored to be an AI breakthrough related to logic. He also expressed his opinion that GPT-4 Turbo and GPT-4 “kind of suck” and that the jump from GPT-4 to GPT-5 could be as significant as the improvement from GPT-3 to GPT-4.

Why does this matter?

This could mean that after Google Gemini and Claude-3’s latest version, a new model, possibly ChatGPT-5, could be released in 2024. Altman’s candid remarks about the current state of AI models also offer valuable context for understanding the anticipated advancements and challenges in the field.

Source

Project GR00T is an ambitious initiative aiming to develop a general-purpose foundation model for humanoid robot learning, addressing embodied AGI challenges. Collaborating with leading humanoid companies worldwide, GR00T aims to understand multimodal instructions and perform various tasks.

GROOT is a foundational model that takes language, videos, and example demonstrations as inputs so it can produce the next action.

What the heck does that mean?

➡️ It means you can show it how to do X a few times, and then it can do X on its own.

Like cooking, drumming, or…

Source

Google’s new fine-tuned model is a HUGE improvement, AI is coming for human doctors sooner than most believe.

Google's new fine-tuned model is a HUGE improvement, AI is coming for human doctors sooner than most believe.
Google’s new fine-tuned model is a HUGE improvement, AI is coming for human doctors sooner than most believe.

NVIDIA creates Earth-2 digital twin: generative AI to simulate, visualize weather and climate. Source

What Else Is Happening in AI on March 19th, 2024❗

🤝 Apple is in talks to bring Google’s AI to iPhones

Apple and Google are negotiating a deal to integrate Google’s Gemini AI into iPhones, potentially shaking up the AI industry. The deal would expand on their existing search partnership. Apple also held discussions with OpenAI. If successful, the partnership could give Gemini a significant edge with billions of potential users. (Link)

 🏷️YouTube rolls out AI content labels

YouTube now requires creators to self-label AI-generated or synthetic content in videos. The platform may add labels itself for potentially misleading content. However, the tool relies on creators being honest, as YouTube is still working on AI detection tools. (Link)

🎮Roblox speeds up 3D creation with AI tools
Roblox has introduced two AI-driven tools to streamline 3D content creation on its platform. Avatar Auto Setup automates the conversion of 3D body meshes into fully animated avatars, while Texture Generator allows creators to quickly alter the appearance of 3D objects using text prompts, enabling rapid prototyping and iteration.(Link)

🌐Nvidia teams up with Shutterstock and Getty Images for AI-generated 3D content

Nvidia’s Edify AI can now create 3D content, and partnerships with Shutterstock and Getty Images will make it accessible to all. Developers can soon experiment with these models, while industry giants are already using them to create stunning visuals and experiences.  (Link)

🖌️Adobe Substance 3D introduces AI-powered text-to-texture tools

Adobe has introduced two AI-driven features to its Substance 3D suite: “Text to Texture,” which generates photo-realistic or stylized textures from text prompts, and “Generative Background,” which creates background images for 3D scenes. Both tools use 2D imaging technology from Adobe’s Firefly AI model to streamline 3D workflows. (Link)

A daily chronicle of AI Innovations: March 18th, 2024 – Bernie’s 4 day workweek: less work, same pay – Google’s AI brings photos to life as talking avatars – Elon Musk’s xAI open-sources Grok AI

Bernie’s 4 day workweek: less work, same pay

Sen. Bernie Sanders has introduced the Thirty-Two Hour Workweek Act, which aims to establish a four-day workweek in the United States without reducing pay or benefits. To be phased in over four years, the bill would lower the overtime pay threshold from 40 to 32 hours, ensuring that workers receive 1.5 times their regular salary for work days longer than 8 hours and double their regular wage for work days longer than 12 hours.

Sanders, along with Sen. Laphonza Butler and Rep. Mark Takano, believes that this bill is crucial in ensuring that workers benefit from the massive increase in productivity driven by AI, automation, and new technology. The legislation aims to reduce stress levels and improve Americans’ quality of life while also protecting their wages and benefits.

Why does this matter?

This bill could alter the workforce dynamics. Businesses may need to assess staffing and invest in AI to maintain productivity. While AI may raise concerns over job displacements, it also offers opportunities for better work-life balance through efficiency gains by augmenting human capabilities.

Source

Google’s AI brings photos to life as talking avatars

Google’s latest AI research project VLOGGER, automatically generates realistic videos of talking and moving people from just a single image and an audio or text input. It is the first model that aims to create more natural interactions with virtual agents by including facial expressions, body movements, and gestures, going beyond simple lip-syncing.

It uses a two-step process: first, a diffusion-based network predicts body motion and facial expressions based on the audio, and then a novel architecture based on image diffusion models generates the final video while maintaining temporal consistency. VLOGGER outperforms previous state-of-the-art methods in terms of image quality, diversity, and the range of scenarios it can handle.

Why does this matter?

VLOGGER’s flexibility and applications could benefit remote work, education, and social interaction, making them more inclusive and accessible. Also, as AR/VR technologies advance, VLOGGER’s avatars could create emotionally resonant experiences in gaming, entertainment, and professional training scenarios.

Source

Elon Musk’s xAI open-sources Grok AI

Elon Musk’s xAI has open-sourced the base model weights and architecture of its AI chatbot, Grok. This allows researchers and developers to freely use and build upon the 314 billion parameter Mixture-of-Experts model. Released under the Apache 2.0 license, the open-source version is not fine-tuned for any particular task.

Why does this matter?

This move aligns with Musk’s criticism of companies that don’t open-source their AI models, including OpenAI, which he is currently suing for allegedly breaching an agreement to remain open-source. While several fully open-source AI models are available, the most used ones are closed-source or offer limited open licenses.

Source

What Else Is Happening in AI on March 18th, 2024❗

🧠 Maisa KPU may be the next leap in AI reasoning

Maisa has released the beta version of its Knowledge Processing Unit (KPU), an AI system that uses LLMs’ advanced reasoning and data processing abilities. In an impressive demo, the KPU assisted a customer with an order-related issue, even when the customer provided an incorrect order ID, showing the system’s understanding abilities. (Link)

🍿 PepsiCo increases market domination using GenAI

PepsiCo uses GenAI in product development and marketing for faster launches and better profitability. It has increased market penetration by 15% by using GenAI to improve the taste and shape of products like Cheetos based on customer feedback. The company is also doubling down on its presence in India, with plans to open a third capability center to develop local talent. (Link)

💻 Deci launches Nano LLM & GenAI dev platform

Israeli AI startup Deci has launched two major offerings: Deci-Nano, a small closed-source language model, and a complete Generative AI Development Platform for enterprises. Compared to rivals like OpenAI and Anthropic, Deci-Nano offers impressive performance at low cost, and the new platform offers a suite of tools to help businesses deploy and manage AI solutions. (Link)

🎮 Invoke AI simplifies game dev workflows

Invoke has launched Workflows, a set of AI tools designed for game developers and large studios. These tools make it easier for teams to adopt AI, regardless of their technical expertise levels. Workflows allow artists to use AI features while maintaining control over their training assets, brand-specific styles, and image security. (Link)

🚗 Mercedes teams up with Apptronik for robot workers

Mercedes-Benz is collaborating with robotics company Apptronik to automate repetitive and physically demanding tasks in its manufacturing process. The automaker is currently testing Apptronik’s Apollo robot, a 160-pound bipedal machine capable of lifting objects up to 55 pounds. The robot inspects and delivers components to human workers on the production line, reducing the physical strain on employees and increasing efficiency. (Link)

A daily chronicle of AI Innovations: Week 2 Recap

  1. DeepSeek released DeepSeek-VL, an open-source Vision-Language (VL) model designed for real-world vision and language understanding applications. The DeepSeek-VL family, includes 7B and1.3B base and chat models and achieves state-of-the-art or competitive performance across a wide range of visual-language benchmarks. Free for commercial use [Details | Hugging Face | Demo]

  2. Cohere released Command-R, a 35 billion parameters generative model with open weights, optimized for long context tasks such as retrieval augmented generation (RAG) and using external APIs and tools for production-scale AI for enterprise [Details | Hugging Face].

  3. Google DeepMind introduced SIMA (Scalable Instructable Multiworld Agent), a generalist AI agent for 3D virtual environments, trained on nine different video games. It can understand a broad range of gaming worlds, and follows natural-language instructions to carry out tasks within them, as a human might.  It doesn’t need access to a game’s source code or APIs and requires only the images on screen, and natural-language instructions provided by the user. SIMA uses keyboard and mouse outputs to control the games’ central character to carry out these instructions [Details].

  4. Meta AI introduces Emu Video Edit (EVE), a model that establishes a new state-of-the art in video editing without relying on any supervised video editing data [Details].

  5. Cognition Labs introduced Devin, the first fully autonomous AI software engineer. Devin can learn how to use unfamiliar technologies, can build and deploy apps end to end, can train and fine tune its own AI models. When evaluated on the SWE-Bench benchmark, which asks an AI to resolve GitHub issues found in real-world open-source projects, Devin correctly resolves 13.86% of the issues unassisted, exceeding the previous state-of-the-art model performance of 1.96% unassisted and 4.80% assisted [Details].

  6. Pika Labs adds sound effects to its AI video tool, Pika, allowing users to either prompt desired sounds or automatically generate them based on video content. [Video link].

  7. Anthropic’s Claude 3 Opus ranks #1 on LMSYS Chatbot Arena Leaderboard, along with GPT-4 [Link].

  8. The European Parliament approved the Artificial Intelligence Act. The new rules ban certain AI applications including biometric categorisation systems, Emotion recognition in the workplace and schools, social scoring and more [Details].

  9. Huawei Noah’s Ark Lab introduced PixArt–Σ, a Diffusion Transformer model (DiT) capable of directly generating images at 4K resolution. It achieves superior image quality and user prompt adherence with significantly smaller model size (0.6B parameters) than existing text-to-image diffusion models, such as SDXL (2.6B parameters) and SD Cascade (5.1B parameters) [Details].

  10. South Korean startup Hyodol AI has launched a $1,800 LLM-powered companion doll specifically designed to offer emotional support and companionship to the rapidly expanding elderly demographic in the country [Details].

  11. Covariant introduced RFM-1 (Robotics Foundation Model -1), a large language model (LLM), but for robot language. Set up as a multimodal any-to-any sequence model, RFM-1 is an 8 billion parameter transformer trained on text, images, videos, robot actions, and a range of numerical sensor readings [Details].

  12. Figure 01 robot integrated with an OpenAI vision-language model can now have full conversations with people [Link]

  13. Deepgram announced the general availability of Aura, a text-to-speech model built for responsive, conversational AI agents and applications [Details | Demo].

  14. Claude 3 Haiku model is now available alongside Sonnet and Opus in the Claude API and on claude.ai for Pro subscribers. Haiku outperforms GPT-3.5 and Gemini 1.0 pro while costing less, and is three times faster than its peers for the vast majority of workloads [Details].

  15. Paddle announced AI Launchpad, a 6-week remote program for AI founders to launch and scale an AI business with $20,000 in cash prize [Details].

  16. Midjourney adds feature for generating consistent characters across multiple gen AI images [Details].

  17. The Special Committee of the OpenAI Board announced the completion of the review. Altman, Brockman to continue to lead OpenAI [Details]

  18. Together.ai introduced Sequoia, a scalable, robust, and hardware-aware speculative decoding framework that improves LLM inference speed on consumer GPUs (with offloading), as well as on high-end GPUs (on-chip), without any approximations [Details].

  19. OpenAI released Transformer Debugger (TDB), a tool developed and used internally by OpenAI’s Superalignment team for investigating into specific behaviors of small language models [GitHub].

  20. Elon Musk announced that xAI will open source Grok this week [Link].

A Daily Chronicle of AI Innovations – March 16th, 2024:

🔍 FTC is probing Reddit’s AI licensing deals

  • Reddit is under investigation by the FTC for its data licensing practices concerning user-generated content being used to train AI models.
  • The investigation focuses on Reddit’s engagement in selling, licensing, or sharing data with third parties for AI training.
  • Reddit anticipates generating approximately USD 60 million in 2024 from a data licensing agreement with Google, aiming to leverage its platform data for training LLMs

.

💻 New jailbreak uses ASCII art to elicit harmful responses from leading LLM

  • Researchers identified a new vulnerability in leading AI language models, named ArtPrompt, which uses ASCII art to exploit the models’ security mechanisms.
  • ArtPrompt masks security-sensitive words with ASCII art, fooling language models like GPT-3.5, GPT-4, Gemini, Claude, and Llama2 into performing actions they would otherwise block, such as giving instructions for making a bomb.
  • The study underscores the need for enhanced defensive measures for language models, as ArtPrompt, by leveraging a mix of text-based and image-based inputs, can effectively bypass current security protocols.

OpenAI aims to make its own AI processors — chip venture in talks with Abu Dhabi investment firm. Source

Once “too scary” to release, GPT-2 gets squeezed into an Excel spreadsheet. Source

 A Daily Chronicle of AI Innovations – March 15th, 2024:

🍎 Apple quietly acquires another AI startup

🤖 Mercedes tests humanoid robots for ‘low skill, repetitive’ tasks

🚫 Midjourney bans prompts with Joe Biden and Donald Trump over election misinformation concerns

💰 El Salvador stashes $406 million in bitcoin in ‘cold wallet’

🤔 Microsoft calls out Google dominance in generative AI

📝 Anthropic releases affordable, high-speed Claude 3 Haiku model

🥘 Apple’s MM1: The new recipe to master AI performance

Apple’s MM1 AI model shows state-of-the-art language and vision capabilities. It was trained on a filtered dataset of 500 million text-image pairs from the web, including 10% text-only docs to improve language understanding.

🥘 Apple’s MM1: The new recipe to master AI performance
🥘 Apple’s MM1: The new recipe to master AI performance

The team experimented with different configurations during training. They discovered that using an external pre-trained high-resolution image encoder improved visual recognition. Combining different image, text, and caption data ratios led to the best performance. Synthetic caption data also enhanced few-shot learning abilities.

This experiment cements that using a blend of image caption, interleaved image text, and text-only data is crucial for achieving state-of-the-art (SOTA) few-shot results across multiple benchmarks.

Why does it matter?

Apple’s new model is promising, especially in developing image recognition systems for new categories or domains. This will help businesses and startups improve the speed of AI tool development specifically for text-to-image, document analysis, and enhanced visual recognition.

⚡ Cerebras WSE-3: AI chip enabling 10x larger models than GPT-4

Cerebras Systems has made a groundbreaking announcement unveiling its latest wafer-scale AI chip, the WSE-3. This chip boasts an incredible 4 trillion transistors, making it one of the most powerful AI chips on the market. The third-generation wafer-scale AI mega chip is twice as powerful as its predecessor while being power efficient. 

The chip’s transistor density has increased by over 50 percent thanks to the latest manufacturing technology. One of the most remarkable features of the WSE-3 chip is its ability to enable AI models that are ten times larger than the highly acclaimed GPT-4 and Gemini models.

Why does it matter?

The WSE-3 chip opens up new possibilities for tackling complex problems and pushing the boundaries of AI capabilities. This powerful system can train massive language models, such as the Llama 70B, in just one day. It will help enterprises create custom LLMs, rapidly reducing the time-to-market.

🤖 Apple acquires Canadian AI startup DarwinAI

Apple made a significant acquisition earlier this year by purchasing Canadian AI startup DarwinAI. Integrating DarwinAI’s expertise and technology bolsters Apple’s AI initiatives. 

With this acquisition, Apple aims to tap into DarwinAI’s advancements in AI technology, particularly in visual inspection during manufacturing and making AI systems smaller and faster. Leveraging DarwinAI’s technology, Apple aims to run AI on devices rather than relying solely on cloud-based solutions.

Why does it matter?

Apple’s acquisition of DarwinAI is a strategic move to revolutionize features and enhance its AI capabilities across various products and services. Especially with the iOS 18 release around the corner, this acquisition will help create new features and enhance the user experience.

🤖 Microsoft expands the availability of Copilot across life and work.

Microsoft is expanding Copilot, its AI assistant, with the introduction of the Copilot Pro subscription for individuals, the availability of Copilot for Microsoft 365 to small and medium-sized businesses, and removing seat minimums for commercial plans. Copilot aims to enhance creativity, productivity, and skills across work and personal life, providing users access to the latest AI models and improved image creation

💻 Oracle adds groundbreaking Generative AI features to its software

Oracle has added advanced AI capabilities to its finance and supply chain software suite, aimed at improving decision-making and enhancing customer and employee experience. For instance, Oracle Fusion Cloud SCM includes features such as item description generation, supplier recommendations, and negotiation summaries.

💰 Databricks makes a strategic investment in Mistral AI

Databricks has invested in Mistral AI and integrated its AI models into its data intelligence platform, allowing users to customize and consume models in various ways. The integration includes Mistral’s text-generation models, such as Mistral 7B and Mixtral 8x7B, which support multiple languages. This partnership aims to provide Databricks customers with advanced capabilities to leverage AI models and drive innovation in their data-driven applications.

📱 Qualcomm emerges as a mobile AI juggernaut

Qualcomm has solidified its leadership position in mobile artificial intelligence (AI). It has been developing AI hardware and software for over a decade. Their Snapdragon processors are equipped with specialized AI engines like Hexagon DSP, ensuring efficient AI and machine learning processing without needing to send data to the cloud.

👓 MIT researchers develop peripheral vision capabilities for AI models

AI researchers are developing techniques to simulate peripheral vision and improve object detection in the periphery. They created a new dataset to train computer vision models, which led to better object detection outside the direct line of sight, though still behind human capabilities. A modified texture tiling approach accurately representing information loss in peripheral vision significantly enhanced object detection and recognition abilities.

🤔 Microsoft calls out Google dominance in generative AI 

  • Microsoft has expressed concerns to EU antitrust regulators about Google’s dominance in generative AI, highlighting Google’s unique position due to its vast data sets and vertical integration, which includes AI chips and platforms like YouTube.
  • The company argues that Google’s control over vast resources and its own AI developments give it a competitive advantage, making it difficult for competitors to match, especially in the development of Large Language Models like Gemini.
  • Microsoft defends partnerships with startups like OpenAI as essential for innovation and competition in the AI market, countering regulatory concerns about potential anticompetitive advantages arising from such collaborations.

🤖 Mercedes tests humanoid robots for ‘low skill, repetitive’ tasks

  • Mercedes-Benz is testing humanoid robots, specifically Apptronik’s bipedal robot Apollo, for automating manual labor tasks in manufacturing.
  • The trial aims to explore the use of Apollo in physically demanding, repetitive tasks within existing manufacturing facilities without the need for significant redesigns.
  • The initiative seeks to address labor shortages by using robots for low-skill tasks, allowing highly skilled workers to focus on more complex aspects of car production.

🚫 Midjourney bans prompts with Joe Biden and Donald Trump over election misinformation concerns

  • Midjourney, an AI image generator, has banned prompts containing the names of Joe Biden and Donald Trump to avoid the spread of election misinformation.
  • The policy change is in response to concerns over AI’s potential to influence voters and spread false information before the 2024 presidential election.
  • Despite the new ban, Midjourney previously allowed prompts that could generate misleading or harmful content, and it was noted for its poor performance in controlling election disinformation.

Midjourney introduces Character Consistency: Tutorial

midjourney_character_consistency

A Daily Chronicle of AI Innovations – March 14th, 2024: 

🎮 DeepMind’s SIMA: The AI agent that’s a Jack of all games

 ⚡ Claude 3 Haiku: Anthropic’s lightning-fast AI solution for enterprises

 🤖 OpenAI-powered “Figure 01” can chat, perceive, and complete tasks

 🎥 OpenAI’s Sora will be publicly available later this year

 

🎮 DeepMind’s SIMA: The AI agent that’s a Jack of all games

DeepMind has introduced SIMA (Scalable Instructable Multiworld Agent), a generalist AI agent that can understand and follow natural language instructions to complete tasks across video game environments. Trained in collaboration with eight game studios on nine different games, SIMA marks a significant milestone in game-playing AI by showing the ability to generalize learned skills to new gaming worlds without requiring access to game code or APIs.

 

DeepMind's SIMA: The AI agent that's a Jack of all games
DeepMind’s SIMA: The AI agent that’s a Jack of all games
 

(SIMA comprises pre-trained vision models, and a main model that includes a memory and outputs keyboard and mouse actions.)

SIMA was evaluated on 600 basic skills, including navigation, object interaction, and menu use. In tests, SIMA agents trained on multiple games significantly outperformed specialized agents trained on individual games. Notably, an agent trained on all but one game performed nearly as well on the unseen game as an agent specifically trained on it, showcasing SIMA’s remarkable ability to generalize to new environments. 

Why does this matter?

SIMA’s generalization ability using a single AI agent is a significant milestone in transfer learning. By showing that a multi-task trained agent can perform nearly as well on an unseen task as a specialized agent, SIMA paves the way for more versatile and scalable AI systems. This could lead to faster deployment of AI in real-world applications, as agents would require less task-specific training data and could adapt to new scenarios more quickly.

Source


⚡ Claude 3 Haiku: Anthropic’s lightning-fast AI solution for enterprises

Anthropic has released Claude 3 Haiku, their fastest and most affordable AI model. With impressive vision capabilities and strong performance on industry benchmarks, Haiku is designed to tackle a wide range of enterprise applications. The model’s speed – processing 21K tokens per second for prompts under 32K tokens – and cost-effective pricing model make it an attractive choice for businesses needing to analyze large datasets and generate timely outputs.

 

Claude 3 Haiku: Anthropic's lightning-fast AI solution for enterprises
Claude 3 Haiku: Anthropic’s lightning-fast AI solution for enterprises
 

In addition to its speed and affordability, Claude 3 Haiku prioritizes enterprise-grade security and robustness. The model is now available through Anthropic’s API or on claude.ai for Claude Pro subscribers.

Why does this matter?

Claude 3 Haiku sets a new benchmark for enterprise AI by offering high speed and cost-efficiency without compromising performance. This release will likely intensify competition among AI providers, making advanced AI solutions more accessible to businesses of all sizes. As more companies adopt models like Haiku, we expect a surge in AI-driven productivity and decision-making across industries.

Source


🤖 OpenAI-powered “Figure 01” can chat, perceive, and complete tasks

Robotics company Figure, in collaboration with OpenAI, has developed a groundbreaking robot called “Figure 01” that can engage in full conversations, perceive its surroundings, plan actions, and execute tasks based on verbal requests, even those that are ambiguous or context-dependent. This is made possible by connecting the robot to a multimodal AI model trained by OpenAI, which integrates language and vision.

OpenAI-powered "Figure 01" can chat, perceive, and complete tasks
OpenAI-powered “Figure 01” can chat, perceive, and complete tasks

The AI model processes the robot’s entire conversation history, including images, enabling it to generate appropriate verbal responses and select the most suitable learned behaviors to carry out given commands. The robot’s actions are controlled by visuomotor transformers that convert visual input into precise physical movements. “Figure 01” successfully integrates natural language interaction, visual perception, reasoning, and dexterous manipulation in a single robot platform.

Why does this matter?

As robots become more adept at understanding and responding to human language, questions arise about their autonomy and potential impact on humanity. Collaboration between the robotics industry and AI policymakers is needed to establish regulations for the safe deployment of AI-powered robots. If deployed safely, these robots could become trusted partners, enhancing productivity, safety, and quality of life in various domains.

Source

What Else Is Happening in AI on March 14th, 2024❗

🛍️ Amazon streamlines product listing process with new AI tool

Amazon is introducing a new AI feature for sellers to quickly create product pages by pasting a link from their external website. The AI generates product descriptions and images based on the linked site’s information, saving sellers time. (Link)

🛡️ Microsoft to expand AI-powered cybersecurity tool availability from April 1

Microsoft is expanding the availability of its AI-powered cybersecurity tool, “Security Copilot,” from April 1, 2024. The tool helps with tasks like summarizing incidents, analyzing vulnerabilities, and sharing information. Microsoft plans to adopt a ‘pay-as-you-go’ pricing model to reduce entry barriers. (Link)

🎥 OpenAI’s Sora will be publicly available later this year

OpenAI will release Sora, its text-to-video AI tool, to the public later this year. Sora generates realistic video scenes from text prompts and may add audio capabilities in the future. OpenAI plans to offer Sora at a cost similar to DALL-E, its text-to-image model, and is developing features for users to edit the AI-generated content. (Link)

📰 OpenAI partners with Le Monde, Prisa Media for news content in ChatGPT

OpenAI has announced partnerships with French newspaper Le Monde and Spanish media group Prisa Media to provide their news content to users of ChatGPT. The media companies see this as a way to ensure reliable information reaches AI users while safeguarding their journalistic integrity and revenue. (Link)

🏠 Icon’s AI architect and 3D printing breakthroughs reimagine homebuilding

Construction tech startup Icon has introduced an AI-powered architect, Vitruvius, that engages users in designing their dream homes, offering 3D-printed and conventional options. The company also debuted an advanced 3D printing robot called Phoenix and a low-carbon concrete mix as part of its mission to make homebuilding more affordable, efficient, and sustainable. (Link)

A Daily Chronicle of AI Innovations – March 13th, 2024: Devin: The first AI software engineer redefines coding; Deepgram’s Aura empowers AI agents with authentic voices; Meta introduces two 24K GPU clusters to train Llama 3

Devin: The first AI software engineer redefines coding 

In the most groundbreaking development, the US-based startup Cognition AI has unveiled Devin, the world’s first AI software engineer. It is an autonomous agent that solves engineering tasks using its shell or command prompt, code editor, and web browser. Devin can also perform tasks like planning, coding, debugging, and deploying projects autonomously.

https://twitter.com/i/status/1767548763134964000

When evaluated on the SWE-Bench benchmark, which asks an AI to resolve GitHub issues found in real-world open-source projects, Devin correctly resolves 13.86% of the issues unassisted, far exceeding the previous state-of-the-art model performance of 1.96% unassisted and 4.80% assisted. It has successfully passed practical engineering interviews with leading AI companies and even completed real Upwork jobs.

Why does it matter?

There’s already a huge debate if Devin will replace software engineers. However, most production-grade software is too complex, unique, or domain-specific to be fully automated at this point. Perhaps, Devin could start handling more initial-level tasks in development. More so, it can assist developers in quickly prototyping, bootstrapping, and autonomously launching MVP for smaller apps and websites, for now

Source

Deepgram’s Aura empowers AI agents with authentic voices

Deepgram, a top voice recognition startup, just released Aura, its new real-time text-to-speech model. It’s the first text-to-speech model built for responsive, conversational AI agents and applications. Companies can use these agents for customer service in call centers and other customer-facing roles.

Deepgram’s Aura empowers AI agents with authentic voices
Deepgram’s Aura empowers AI agents with authentic voices

Aura includes a dozen natural, human-like voices with lower latency than any comparable voice AI alternative and is already being used in production by several customers. Aura works hand in hand with Deepgram’s Nova-2 speech-to-text API. Nova-2 is known for its top-notch accuracy and speed in transcribing audio streams.

Why does it matter?

Deepgram’s Aura is a one-stop shop for speech recognition and voice generation APIs that enable the fastest response times and most natural-sounding conversational flow. Its human-like voice models render extremely fast (typically in well under half a second) and at an affordable price ($0.015 per 1,000 characters). Lastly, Deepgram’s transcription is more accurate and faster than other solutions as well.

Source

Meta introduces two 24K GPU clusters to train Llama 3

Meta has invested significantly in its AI infrastructure by introducing two 24k GPU clusters. These clusters, built on top of Grand Teton, OpenRack, and PyTorch, are designed to support various AI workloads, including the training of Llama 3.

Meta introduces two 24K GPU clusters to train Llama 3
Meta introduces two 24K GPU clusters to train Llama 3

Meta aims to expand its infrastructure build-out by the end of 2024. It plans to include 350,000 NVIDIA H100 GPUs, providing compute power equivalent to nearly 600,000 H100s. The clusters are built with a focus on researcher and developer experience.

This adds up to Meta’s long-term vision to build open and responsibly developed artificial general intelligence (AGI). These clusters enable the development of advanced AI models and power applications such as computer vision, NLP, speech recognition, and image generation.

Why does it matter?

Meta is committed to open compute and open source, driving innovation in the AI software and hardware industry. Introducing two new GPUs to train Llama 3 is also a push forward to their commitment. As a founding member of Open Hardware Innovation (OHI) and the Open Innovation AI Research Community, Meta wants to make AI transparent and trustworthy.

Source

What Else Is Happening in AI on March 13th, 2024❗

🎮 Google Play to display AI-powered FAQs and recent YouTube videos for games

At the Google for Games Developer Summit held in San Francisco, Google announced several new features for ‘Google Play listing for games’. These include AI-powered FAQs, displaying the latest YouTube videos, new immersive ad formats, and support for native PC game publishing. These new features will allow developers to display promotions and the latest YouTube videos directly in their listing and show them to users in the Games tab of the Play Store. (Link)

🛡️ DoorDash’s new AI-powered tool automatically curbs verbal abuses

DoorDash has introduced a new AI-powered tool named ‘SafeChat+’ to review in-app conversations and determine if a customer or Dasher is being harassed. There will be an option to report the incident and either contact DoorDash’s support team if you’re a customer or quickly cancel the order if you’re a delivery person. With this feature, DoorDash aims to reduce verbally abusive and inappropriate interactions between consumers and delivery people. (Link)

🔍 Perplexity has decided to bring Yelp data to its chatbot

Perplexity has decided to bring Yelp data to its chatbot. The company CEO, Aravind Srinivas, told the media that many people use chatbots like search engines. He added that it makes sense to offer information on things they look for, like restaurants, directly from the source. That’s why they have decided to integrate Yelp’s maps, reviews, and other details in responses when people ask for restaurant or cafe recommendations.  (Link)

👗 Pinterest’s ‘body types ranges’ tool delivers more inclusive search results

Pinterest has introduced a new tool named body type ranges, which gives users a choice to self-select body types from a visual cue between four body type ranges to deliver personalized and more refined search results for women’s fashion and wedding inspiration. This tool aims to create a more inclusive place online to search, save, and shop. The company also plans to launch a similar feature for men’s fashion later this year. (Link)

🚀 OpenAI’s GPT-4.5 Turbo is all set to be launched in June 2024

According to the leak search engine results from Bing and DuckDuck Go, which indexed the OpenAI GPT-4.5 Turbo product page before an official announcement, OpenAI is all set to launch the new version of its LLM by June 2024. There is a discussion among the AI community that this could be OpenAI’s fastest, most accurate, and most scalable model to date. The details of GPT-4.5 Turbo were leaked by OpenAI’s web team, which now leads to a 404 page. (Link))

A Daily Chronicle of AI Innovations in March 2024 – Day 12: AI Daily News – March 12th, 2024

🚀Cohere’s introduces production-scale AI for enterprises
🤖 RFM-1 redefines robotics with human-like reasoning
🎧 Spotify introduces audiobook recommendations

🙃 Midjourney bans all its competitor’s employees

🚫 Google restricts election-related queries for its Gemini chatbot

📲 Apple to let developers distribute apps directly from their websites

💰 AI startups reach record funding of nearly $50 billion in 2023

Cohere’s introduces production-scale AI for enterprises

Cohere, an AI company, has introduced Command-R, a new large language model (LLM) designed to address real-world challenges, such as inefficient workflows, data analysis limitations, slow response times, etc.

Cohere’s introduces production-scale AI for enterprises
Cohere’s introduces production-scale AI for enterprises

Command-R focuses on two key areas: Retrieval Augmented Generation (RAG) and Tool Use. RAG allows the model to access and process information from private databases, improving the accuracy of its responses. Tool Use allows Command-R to interact with external software tools and APIs, automating complex tasks.

Command-R offers several features beneficial for businesses, including:

  • Multilingual capabilities: Supports 10 major languages
  • Cost-effectiveness: Offers a longer context window and reduced pricing compared to previous models
  • Wider accessibility: Available through Cohere’s API, major cloud providers, and free weights for research on HuggingFace

Overall, it empowers businesses to leverage AI for improved decision-making, increased productivity, and enhanced customer experiences.

Why does this matter?

Command-R showcases the future of business operations, featuring automated workflows, and enabling humans to focus on strategic work. Thanks to its low hallucination rate, we would see a wider adoption of AI technologies, and the development of sophisticated, context-aware AI applications tailored to specific business needs.

As AI continues to evolve and mature, models like Command-R will shape the future of work and the global economy.

Source

RFM-1 redefines robotics with human-like reasoning

Covariant has introduced RFM-1, a Robotics Foundation Model that gives robots ChatGPT-like understanding and reasoning capabilities.

TLDR;

  • RFM-1 is an 8 billion parameter transformer trained on text, images, videos, robot actions, and sensor readings from Covariant’s fleet of high-performing robotic systems deployed in real-world environments.
  • Similar to how we understand how objects move, RFM-1 can predict future outcomes/consequences based on initial images and robot actions.
  • RFM-1 leverages NLP to enable intuitive interfaces for programming robot behavior. Operators can instruct robots using plain English, lowering barriers to customizing AI behavior for specific needs.
  • RFM-1 can also communicate issues and suggest solutions to operators.

Why does this matter?

This advancement has the potential to revolutionize industries such as manufacturing, logistics, and healthcare, where robots can work alongside humans to improve efficiency, safety, and productivity.

Source

Spotify now recommends audiobooks (with AI)

Spotify has introduced a novel recommendation system called 2T-HGNN to provide personalized audiobook recommendations to its users. The system addresses the challenges of introducing a new content type (audiobooks) into an existing platform, such as data sparsity and the need for scalability.

Spotify now recommends audiobooks (with AI)
Spotify now recommends audiobooks (with AI)

2T-HGNN leverages a technique called “Heterogeneous Graph Neural Networks” (HGNNs) to uncover connections between different content types. Additionally, a “Two Tower” (2T) model helps ensure that recommendations are made quickly and efficiently for millions of users.

Interestingly, the system also uses podcast consumption data and weak interaction signals to uncover user preferences and predict future audiobook engagement.

Why does this matter?

This research will not only improve the user experience but also encourage users to explore and engage with audiobooks, potentially driving growth in this new content vertical. Moreover, it may inspire similar strategies in domains where tailored recommendations are essential, such as e-commerce, news, and entertainment.

Source

What Else Is Happening in AI on March 12th, 2024❗

💡 Elon Musk makes xAI’s Grok chatbot open-source

Elon Musk announced that his AI startup xAI will open-source its ChatGPT rival “Grok” this week, following a lawsuit against OpenAI for shifting to a for-profit model. Musk aims to provide free access to Grok’s code, aligning with open-source AI models like Meta and Mistral (Link)

 🖼️ Midjourney launches character consistent feature

Midjourney’s new “Consistent Character” feature lets artists create consistent characters across images. Users provide a reference image URL with their prompt, and the AI attempts to match the character’s features in new scenes. This holds promise for creators of comics, storyboards, and other visual narratives. (Link)

🤖 Apple tests AI for App Store ad optimization
Taking a page from Google and Meta, Apple is testing AI-powered ad placement within its App Store. This new system would automatically choose the most suitable locations (e.g., App Store Today page) to display ads based on advertiser goals and budget. This development could help Apple’s ad business reach $6 billion by 2025.(Link)

🏥China tests AI chatbot to assist neurosurgeons

China steps into the future of brain surgery with an AI co-pilot, dubbed “CARES Copilot”. This AI, based on Meta’s Llama 2.0, assists surgeons by analyzing medical data (e.g., scans) and offering informed suggestions during surgery. This government-backed project reflects China’s growing focus on developing domestic AI solutions for various sectors, including healthcare. (Link)

🧓South Korea deploys AI dolls to tackle elderly loneliness

Hyodol, a Korean-based company, has introduced an AI-powered companion doll to tackle loneliness among elderly. Priced at $1800, the robot doll boasts advanced features like conversation abilities, medication reminders, and safety alerts. With 7,000 dolls already deployed, Hyodol aims to expand to European and North American markets. (Link)

🙃 Midjourney bans all its competitor’s employees

  • Midjourney banned all Stability AI employees from using its service, citing a systems outage caused by data scraping efforts linked to Stability AI employees.
  • The company announced the ban and a new policy against “aggressive automation” after identifying botnet-like activity from Stability AI during a server outage.
  • Stability AI CEO Emad Mostaque is looking into the incident, and Midjourney’s founder David Holz has provided information for the internal investigation.
  • Source

🚫 Google restricts election-related queries for its Gemini chatbot

  • Google has begun restricting Gemini queries related to elections globally in countries where elections are taking place, to prevent the dissemination of false or misleading information.
  • The restrictions were implemented amid concerns over generative AI’s potential impact on elections and followed an advisory from India requiring tech firms to obtain government permission before introducing new AI models.
  • Despite the restrictions, the effectiveness of the restrictions is under question as some users found ways to bypass them, and it’s uncertain if Google will lift these restrictions post-elections.
  • Source

💰 AI startups reach record funding of nearly $50 billion in 2023

  • AI startups reached a record funding of nearly $50 billion in 2023, with significant contributions from companies like OpenAI and Anthropic.
  • Investment trends showed over 70 funding rounds exceeding $100 million each, partly due to major companies’ investments, including Microsoft’s $10 billion in OpenAI.
  • While large tech companies are venturing to dominate the AI market, specialized AI startups like Midjourney manage to maintain niches by offering superior products.
  • Source

A Daily Chronicle of AI Innovations in March 2024 – Day 11: AI Daily News – March 11th, 2024

🖼️ Huawei’s PixArt-Σ paints prompts to perfection
🧠 Meta cracks the code to improve LLM reasoning
📈 Yi Models exceed benchmarks with refined data

Huawei’s PixArt-Σ paints prompts to perfection

Researchers from Huawei’s Noah’s Ark Lab introduced PixArt-Σ, a text-to-image model that can create 4K resolution images with impressive accuracy in following prompts. Despite having significantly fewer parameters than models like SDXL, PixArt-Σ outperforms them in image quality and prompt matching.

  

The model uses a “weak-to-strong” training strategy and efficient token compression to reduce computational requirements. It relies on carefully curated training data with high-resolution images and accurate descriptions, enabling it to generate detailed 4K images closely matching the text prompts. The researchers claim that PixArt-Σ can even keep up with commercial alternatives such as Adobe Firefly 2, Google Imagen 2, OpenAI DALL-E 3, and Midjourney v6.

Why does this matter?

PixArt-Σ’s ability to generate high-resolution, photorealistic images accurately could impact industries like advertising, media, and entertainment. As its efficient approach requires fewer computational resources than existing models, businesses may find it easier and more cost-effective to create custom visuals for their products or services.

Source

Meta cracks the code to improve LLM reasoning

Meta researchers investigated using reinforcement learning (RL) to improve the reasoning abilities of large language models (LLMs). They compared algorithms like Proximal Policy Optimization (PPO) and Expert Iteration (EI) and found that the simple EI method was particularly effective, enabling models to outperform fine-tuned models by nearly 10% after several training iterations.

However, the study also revealed that the tested RL methods have limitations in further improving LLMs’ logical capabilities. The researchers suggest that stronger exploration techniques, such as Tree of Thoughts, XOT, or combining LLMs with evolutionary algorithms, are important for achieving greater progress in reasoning performance.

Why does this matter?

Meta’s research highlights the potential of RL in improving LLMs’ logical abilities. This could lead to more accurate and efficient AI for domains like scientific research, financial analysis, and strategic decision-making. By focusing on techniques that encourage LLMs to discover novel solutions and approaches, researchers can make more advanced AI systems.

Source

Yi models exceed benchmarks with refined data

01.AI has introduced the Yi model family, a series of language and multimodal models that showcase impressive multidimensional abilities. The Yi models, based on 6B and 34B pretrained language models, have been extended to include chat models, 200K long context models, depth-upscaled models, and vision-language models.

The performance of the Yi models can be attributed to the high-quality data resulting from 01.AI‘s data-engineering efforts. By constructing a massive 3.1 trillion token dataset of English and Chinese corpora and meticulously polishing a small-scale instruction dataset, 01.AI has created a solid foundation for their models. The company believes that scaling up model parameters using thoroughly optimized data will lead to even more powerful models.

Why does this matter?

The Yi models’ success in language, vision, and multimodal tasks suggests that they could be adapted to a wide range of applications, from customer service chatbots to content moderation and beyond. These models also serve as a prime example of how investing in data optimization can lead to groundbreaking advancements in the field.

Source

OpenAI’s Evolution into Skynet: AI and Robotics Future, Figure Humanoid Robots

 

  • OpenAI’s partnership with Figure signifies a transformative step in the evolution of AI and robotics.
  • Utilizing Microsoft Azure, OpenAI’s investment supports the deployment of autonomous humanoid robots for commercial use.
  • Figure’s collaboration with BMW Manufacturing integrates humanoid robots to enhance automotive production.
  • This technological progression echoes the fictional superintelligence Skynet yet emphasizes real-world innovation and safety.
  • The industry valuation of Figure at $2.6 billion underlines the significant impact and potential of advanced AI in commercial sectors.

What Else Is Happening in AI on March 11, 2024❗

🏠 Redfin’s AI can tell you about your dream neighborhood

“Ask Redfin” can now answer questions about homes, neighborhoods, and more. Using LLMss, the chatbot can provide insights on air conditioning, home prices, safety, and even connect users to agents. It is currently available in 12 U.S. cities, including Atlanta, Boston, Chicago, and Washington, D.C. (Link)

🔊 Pika Labs Adds Sound to Silent AI Videos 

Pika Labs users can now add sound effects to their generated videos. Users can either specify the exact sounds they want or let Pika’s AI automatically select and integrate them based on the video’s content. This update aims to provide a more immersive and engaging video creation experience, setting a new standard in the industry. (Link)

🩺 Salesforce’s new AI tool for doctors automates paperwork

Salesforce is launching new AI tools to help healthcare workers automate tedious administrative tasks. Einstein Copilot: Health Actions will allow doctors to book appointments, summarize patient info, and send referrals using conversational AI, while Assessment Generation will digitize health assessments without manual typing or coding. (Link)

🖥️ HP’s new AI-powered PCs redefine work 

HP just dropped a massive lineup of AI-powered PCs, including the HP Elite series, Z by HP mobile workstations, and Poly Studio conferencing solutions. These devices use AI to improve productivity, creativity, and collaboration for the hybrid workforce, while also offering advanced security features like protection against quantum computer hacks. (Link)

🎨 DALL-E 3’s new look is artsy and user-friendly

OpenAI is testing a new user interface for DALL-E 3. It allows users to choose between predefined styles and aspect ratios directly in the GPT, offering a more intuitive and educational experience. OpenAI has also implemented the C2PA standard for metadata verification and is working on an image classifier to reliably recognize DALL-E images. (Link)

A Daily Chronicle of AI Innovations in March 2024 – Week 1 Summary

  1. Anthropic introduced the next generation of Claude: Claude 3 model family. It includes OpusSonnet and Haiku models. Opus is the most intelligent model, that outperforms GPT-4 and Gemini 1.0 Ultra on most of the common evaluation benchmarks. Haiku is the fastest, most compact model for near-instant responsiveness. The Claude 3 models have vision capabilities, offer a 200K context window capable of accepting inputs exceeding 1 million tokens, improved accuracy and fewer refusals [Details | Model Card].
  2. Stability AI partnered with Tripo AI and released TripoSR, a fast 3D object reconstruction model that can generate high-quality 3D models from a single image in under a second. The model weights and source code are available under the MIT license, allowing commercialized use. [Details | GitHub | Hugging Face].
  3. Answer.AI released a fully open source system that, for the first time, can efficiently train a 70b large language model on a regular desktop computer with two or more standard gaming GPUs. It combines QLoRA with Meta’s FSDP, which shards large models across multiple GPUs [Details].
  4. Inflection launched Inflection-2.5, an upgrade to their model powering Pi, Inflection’s empathetic and supportive companion chatbot. Inflection-2.5 approaches GPT-4’s performance, but used only 40% of the amount of compute for training. Pi is also now available on Apple Messages [Details].
  5. Twelve Labs introduced Marengo-2.6, a new state-of-the-art (SOTA) multimodal foundation model capable of performing any-to-any search tasks, including Text-To-Video, Text-To-Image, Text-To-Audio, Audio-To-Video, Image-To-Video, and more [Details].
  6. Cloudflare announced the development of Firewall for AI, a protection layer that can be deployed in front of Large Language Models (LLMs), hosted on the Cloudflare Workers AI platform or models hosted on any other third party infrastructure, to identify abuses before they reach the models [Details]
  7. Scale AI, in partnership with the Center for AI Safety, released WMDP (Weapons of Mass Destruction Proxy): an open-source evaluation benchmark of 4,157 multiple-choice questions that serve as a proxy measurement of LLM’s risky knowledge in biosecurity, cybersecurity, and chemical security [Details].
  8. Midjourney launched v6 turbo mode to generate images at 3.5x the speed (for 2x the cost). Just type /turbo [Link].
  9. Moondream.ai released moondream 2 – a small 1.8B parameters, open-source, vision language model designed to run efficiently on edge devices. It was initialized using Phi-1.5 and SigLIP, and trained primarily on synthetic data generated by Mixtral. Code and weights are released under the Apache 2.0 license, which permits commercial use [Details].
  10. Vercel released Vercel AI SDK 3.0. Developers can now associate LLM responses to streaming React Server Components [Details].
  11. Nous Research released a new model designed exclusively to create instructions from raw-text corpuses, Genstruct 7B. This enables the creation of new, partially synthetic instruction finetuning datasets from any raw-text corpus [Details].
  12. 01.AI open-sources Yi-9B, one of the top performers among a range of similar-sized open-source models excelling in code, math, common-sense reasoning, and reading comprehension [Details].
  13. Accenture to acquire Udacity to build a learning platform focused on AI [Details].
  14. China Offers ‘Computing Vouchers’ upto $280,000 to Small AI Startups to train and run large language models [Details].
  15. Snowflake and Mistral have partnered to make Mistral AI’s newest and most powerful model, Mistral Large, available in the Snowflake Data Cloud [Details]
  16. OpenAI rolled out ‘Read Aloud’ feature for ChatGPT, enabling ChatGPT to read its answers out loud. Read Aloud can speak 37 languages but will auto-detect the language of the text it’s reading [Details].

A Daily Chronicle of AI Innovations in March 2024 – Day 8: AI Daily News – March 08th, 2024

🗣️Inflection 2.5: A new era of personal AI is here!
🔍Google announces LLMs on device with MediaPipe
🤖GaLore: A new method for memory-efficient LLM training

📱Adobe makes creating social content on mobile easier

🛡️OpenAI now allows users to add MFA to user accounts

🏅US Army is building generative AI chatbots in war games

🧑‍🎨 Claude 3 builds the painting app in 2 minutes and 48 seconds

🧪Cognizant launches AI lab in San Francisco to drive innovation

Inflection 2.5: A new era of personal AI is here!

Inflection.ai, the company behind the personal AI app Pi, has recently introduced Inflection-2.5, an upgraded large language model (LLM) that competes with top LLMs like GPT-4 and Gemini. The in-house upgrade offers enhanced capabilities and improved performance, combining raw intelligence with the company’s signature personality and empathetic fine-tuning.

Inflection 2.5: A new era of personal AI is here!
Inflection 2.5: A new era of personal AI is here!

This upgrade has made significant progress in coding and mathematics, keeping Pi at the forefront of technological innovation. With Inflection-2.5, Pi has world-class real-time web search capabilities, providing users with high-quality breaking news and up-to-date information. This empowers Pi users with a more intelligent and empathetic AI experience.

Why does it matter?

Inflection-2.5 challenges leading language models like GPT-4 and Gemini with their raw capability, signature personality, and empathetic fine-tuning. This will provide a new alternative for startups and enterprises building personalized applications with generative AI capabilities.

Source

Google announces LLMs on device with MediaPipe

Google’s new experimental release called the MediaPipe LLM Inference API  allows LLMs to run fully on-device across platforms. This is a significant development considering LLMs’ memory and computing demands, which are over a hundred times larger than traditional on-device models.

Google announces LLMs on device with MediaPipe
Google announces LLMs on device with MediaPipe

The MediaPipe LLM Inference API is designed to streamline on-device LLM integration for web developers and supports Web, Android, and iOS platforms. It offers several key features and optimizations that enable on-device AI. These include new operations, quantization, caching, and weight sharing. Developers can now run LLMs on devices like laptops and phones using MediaPipe LLM Inference API.

Why does it matter?

Running LLMs on devices using MediaPipe and TensorFlow Lite allows for direct deployment, reducing dependence on cloud services. On-device LLM operation ensures faster and more efficient inference, which is crucial for real-time applications like chatbots or voice assistants. This innovation helps rapid prototyping with LLM models and offers streamlined platform integration.

Source

GaLore: A new method for memory-efficient LLM training

Researchers have developed a new technique called Gradient Low-Rank Projection (GaLore) to reduce memory usage while training large language models significantly. Tests have shown that GaLore achieves results similar to full-rank training while reducing optimizer state memory usage by up to 65.5% when pre-training large models like LLaMA.

GaLore: A new method for memory-efficient LLM training
GaLore: A new method for memory-efficient LLM training

It also allows pre-training a 7 billion parameter model from scratch on a single 24GB consumer GPU without needing extra techniques. This approach works well for fine-tuning and outperforms low-rank methods like LoRA on GLUE benchmarks while using less memory. GaLore is optimizer-independent and can be used with other techniques like 8-bit optimizers to save additional memory.

Why does it matter?

The gradient matrix’s low-rank nature will help AI developers during model training. GaLore minimizes the memory cost of storing gradient statistics for adaptive optimization algorithms. It enables training large models like LLaMA with reduced memory consumption, making it more accessible and efficient for researchers.

Source

🤖 OpenAI CTO complained to board about ‘manipulative’ CEO Sam Altman 

  • OpenAI CTO Mira Murati was reported by the New York Times to have played a significant role in CEO Sam Altman’s temporary removal, raising concerns about his leadership in a private memo and with the board.
  • Altman was accused of creating a toxic work environment, leading to fears among board members that key executives like Murati and co-founder Ilya Sutskever could leave, potentially causing a mass exit of talent.
  • Despite internal criticisms of Altman’s leadership and management of OpenAI’s startup fund, hundreds of employees threatened to leave if he was not reinstated, highlighting deep rifts within the company’s leadership.
  • Source

Saudi Arabia’s Male Humanoid Robot Accused of Sexual Harassment

A video of Saudi Arabia’s first male robot has gone viral after a few netizens accused the humanoid of touching a female reporter inappropriately.

“Saudi Arabia unveils its man-shaped AI robot, Mohammad, reacts to a reporter in its first appearance,” an X user wrote while sharing the video that people are claiming shows the robot’s inappropriate behaviour. You can view the original tweet here.

What Else Is Happening in AI on March 08th, 2024❗

📱Adobe makes creating social content on mobile easier

Adobe has launched an updated version of Adobe Express, a mobile app that now includes Firefly AI models. The app offers features such as a “Text to Image” generator, a “Generative Fill” feature, and a “Text Effects” feature, which can be utilized by small businesses and creative professionals to enhance their social media content. Creative Cloud members can also access and work on creative assets from Photoshop and Illustrator directly within Adobe Express. (Link)

🛡️OpenAI now allows users to add MFA to user accounts

To add extra security to OpenAI accounts, users can now enable Multi-Factor Authentication (MFA). To set up MFA, users can follow the instructions in the OpenAI Help Center article “Enabling Multi-Factor Authentication (MFA) with OpenAI.” MFA requires a verification code with their password when logging in, adding an extra layer of protection against unauthorized access. (Link)

🏅US Army is building generative AI chatbots in war games

The US Army is experimenting with AI chatbots for war games. OpenAI’s technology is used to train the chatbots to provide battle advice. The AI bots act as military commanders’ assistants, offering proposals and responding within seconds. Although the potential of AI is acknowledged, experts have raised concerns about the risks involved in high-stakes situations. (Link)

🧑‍🎨 Claude 3 builds the painting app in 2 minutes and 48 seconds

Claude 3, the latest AI model by Anthropic, created a multiplayer drawing app in just 2 minutes and 48 seconds. Multiple users could collaboratively draw in real-time with user authentication and database integration. The AI community praised the app, highlighting the transformative potential of AI in software development. Claude 3 could speed up development cycles and make software creation more accessible. (Link)

🧪Cognizant launches AI lab in San Francisco to drive innovation

Cognizant has opened an AI lab in San Francisco to accelerate AI adoption in businesses. The lab, staffed with top researchers and developers, will focus on innovation, research, and developing cutting-edge AI solutions. Cognizant’s investment in AI research positions them as a thought leader in the AI space, offering advanced solutions to meet the modernization needs of global enterprises. (Link)

A Daily Chronicle of AI Innovations in March 2024 – Day 7: AI Daily News – March 07th, 2024

🗣️Microsoft’s NaturalSpeech makes AI sound human
🔍Google’s search update targets AI-generated spam
🤖Google’s RT-Sketch teaches robots with doodles

🕵️ Ex-Google engineer charged with stealing AI secrets for Chinese firm

🚨 Microsoft engineer sounds alarm on company’s AI image generator in letter to FTC

🤔 Apple bans Epic’s developer account and calls the company ‘verifiably untrustworthy’

🍎 Apple reportedly developing foldable MacBook with 20.3-inch screen

🧠 Meta is building a giant AI model to power its ‘entire video ecosystem

Microsoft’s NaturalSpeech makes AI sound human

Microsoft and its partners have created NaturalSpeech 3, a new Text-to-Speech system that makes computer-generated voices sound more human. Powered by FACodec architecture and factorized diffusion models, NaturalSpeech 3 breaks down speech into different parts, like content, tone, and sound quality to create a natural-sounding speech that fits specific prompts, even for voices it hasn’t heard before.

Microsoft's NaturalSpeech makes AI sound human
Microsoft’s NaturalSpeech makes AI sound human

NaturalSpeech 3 works better than other voice tech in terms of quality, similarity, tone, and clarity. It keeps getting better as it learns from more data. By letting users change how the speech sounds through prompts, NaturalSpeech 3 makes talking to computers feel more like talking to a person. This research is a big step towards a future where chatting with computers is as easy as chatting with friends.

Why does this matter?

This advancement transcends mere voice quality. This could change the way we interact with devices like smartphones, smart speakers, and virtual assistants. Imagine having a more natural, engaging conversation with Siri, Alexa, or other AI helpers.

Better voice tech could also make services more accessible for people with visual impairments or reading difficulties. It might even open up new possibilities in entertainment, like more lifelike characters in video games or audiobooks that sound like they’re read by your favorite celebrities.

Source

Google’s search update targets AI-generated spam

Google has announced significant changes to its search ranking algorithms in order to reduce low-quality and AI-generated spam content in search results. The March update targets three main spam practices: mass distribution of unhelpful content, abusing site reputation to host low-quality content, and repurposing expired domains with poor content.

While Google is not devaluing all AI-generated content, it aims to judge content primarily on its usefulness to users. Most of the algorithm changes are effective immediately, though sites abusing their reputation have a 60-day grace period to change their practices. As Google itself develops AI tools, SGE and Gemini, the debate around AI content and search result quality is just beginning.

Why does this matter?

Websites that churn out lots of AI-made content to rank higher on Google may see their rankings drop. This might push them to focus more on content creation strategies, with a greater emphasis on quality over quantity.

For people using Google, the changes should mean finding more useful results and less junk.

As AI continues to advance, search engines like Google will need to adapt their algorithms to surface the most useful content, whether it’s written by humans or AI.

Source

Google’s RT-Sketch teaches robots with doodles

Google has introduced RT-Sketch, a new approach to teaching robots tasks using simple sketches. Users can quickly draw a picture of what they want the robot to do, like rearranging objects on a table. RT-Sketch focuses on the essential parts of the sketch, ignoring distracting details.

Google's RT-Sketch teaches robots with doodles
Google’s RT-Sketch teaches robots with doodles

Source

RT-Sketch is trained on a dataset of paired trajectories and synthetic goal sketches, and tested on six object rearrangement tasks. The results show that RT-Sketch performs comparably to image or language-conditioned agents in simple settings with written instructions on straightforward tasks. However, it did better when instructions were confusing or there were distracting objects present.

RT-Sketch can also interpret and act upon sketches with varying levels of detail, from basic outlines to colorful drawings.

Why does this matter?

With RT-Sketch, people can tell robots what to do without needing perfect images or detailed written instructions. This could make robots more accessible and useful in homes, workplaces, and for people who have trouble communicating in other ways.

As robots become a bigger part of our lives, easy ways to talk to them, like sketching, could help us get the most out of them. RT-Sketch is a step toward making robots that better understand what we need.

Source

What Else Is Happening in AI on March 07th, 2024❗

🤖Google’s Gemini lets users edit within the chatbox

Google has updated its Gemini chatbot, allowing users to directly edit and fine-tune responses within the chatbox. This feature, launched on March 4th for English users in the Gemini web app, enables more precise outputs by letting people select text portions and provide instructions for improvement. (Link)

📈Adobe’s AI boosts IBM’s marketing efficiency

IBM reports a 10-fold increase in designer productivity and a significant reduction in marketing campaign time after testing Adobe’s generative AI tools. The AI-powered tools have streamlined idea generation and variant creation, allowing IBM to achieve more in less time. (Link)

💡 Zapier’s new tool lets you make AI bots without coding

Zapier has released Zapier Central, a new AI tool that allows users to create custom AI bots by simply describing what they want, without any coding. The bots can work with Zapier’s 6,000+ connected apps, making it easy for businesses to automate tasks. (Link)

🤝Accenture teams up with Cohere to bring AI to enterprises

Accenture has partnered with AI startup, Cohere to provide generative AI solutions to businesses. Leveraging Cohere’s language models and search technologies, the collaboration aims to boost productivity and efficiency while ensuring data privacy and security. (Link)

🎥 Meta builds mega AI model for video recommendations
Meta is developing a single AI model to power its entire video ecosystem across platforms by 2026. The company has invested billions in Nvidia GPUs to build this model, which has already shown promising results in improving Reels watch time on the core Facebook app. (Link)

OpenAI is researching photonic processors to run their AI on

OpenAI hired this person:  He has been doing a lot of research on waveguides for photonic processing for both Training AI and for inference and he did a PHD about photonic waveguides:

I think that he is going to help OpenAI to build photonic waveguides that they can run their neural networks / AI Models on and this is really  cool if OpenAI actually think that they can build processors with faster Inference and training with photonics.

🕵️ Ex-Google engineer charged with stealing AI secrets for Chinese firm

  • Linwei Ding, a Google engineer, has been indicted for allegedly stealing over 500 files related to Google’s AI technology, including designs for chips and data center technologies, to benefit companies in China.
  • The stolen data includes designs for Google’s TPU chips and GPUs, crucial for AI workloads, amid U.S. efforts to restrict China’s access to AI-specific chips.
  • Ding allegedly transferred stolen files to a personal cloud account using a method designed to evade Google’s detection systems, was offered a CTO position by a Chinese AI company and founded a machine learning startup in China while still employed at Google.
  • Source

🚨 Microsoft engineer sounds alarm on company’s AI image generator in letter to FTC

  • Microsoft AI engineer Shane Jones warns that the company’s AI image generator, Copilot Designer, generates sexual and violent content and ignores copyright laws.
  • Jones shared his findings with Microsoft and contacted U.S. senators and the FTC, demanding better safeguards and an independent review of Microsoft’s AI incident reporting process.
  • In addition to the problems with Copilot Designer, other Microsoft products based on OpenAI technologies, such as Copilot Chat, tend to have poorer performance and more insecure implementations than the original OpenAI products, such as ChatGPT and DALL-E 3.
  • Source

🧠 Meta is building a giant AI model to power its ‘entire video ecosystem’ 

  • Meta is developing an AI model designed to power its entire video ecosystem, including the TikTok-like Reels service and traditional video content, as part of its technology roadmap through 2026.
  • The company has invested billions of dollars in Nvidia GPUs to support this AI initiative, aiming to improve recommendation systems and overall product performance across all platforms.
  • This AI model has already demonstrated an 8% to 10% increase in Reels watch time on the Facebook app, with Meta now working to expand its application to include the Feed recommendation product and possibly integrate sophisticated chatting tools.
  • Innovating for the Future

    As Meta continues to innovate and refine their AI model architecture, we can expect even more exciting developments in the future. The company’s dedication to enhancing the video recommendation experience and leveraging the full potential of AI is paving the way for a new era in online video consumption.

    Stay tuned for more updates as Meta strives to revolutionize the digital video landscape with its cutting-edge AI technology.

    r/aidailynewsupdates - Meta's AI Model to Revolutionize Video Ecosystem
  • Source

Will AI destroy the adtech industry?

Some points to consider on both sides:

Yes:

– AI will enable humans to get content they want, nothing more

– New AI OSes will act ‘for’ the human, cleaning content of ads

– OpenAI and new startups don’t need ad revenue, they’ll take monthly subscriptions to deliver information with no ads

No:

– New AI OSes will integrate ads even more closely into the computing experience, acting ‘against’ the human

– Content will be more tightly integrated with ads, and AI won’t be able to unpiece this

– Meta and Alphabet have $100bns of skin in the game, they will make sure this doesn’t happen, including by using their lawyers to prevent lifting content out of the ad context

A Daily Chronicle of AI Innovations in March 2024 – Day 6: AI Daily News – March 06th, 2024

🏆 Microsoft’s Orca AI beats 10x bigger models in math
🎨 GPT-4V wins at turning designs into code
🎥 DeepMind alums’ Haiper joins the AI video race

🤔 OpenAI fires back, says Elon Musk demanded ‘absolute control’ of the company

📱 iOS 17.4 is here: what you need to know

🚫 TikTok faces US ban if ByteDance fails to sell app

🔍 Google now wants to limit the AI-powered search spam it helped create

OpenAI vs Musk (openai responds to elon musk).

 What does Elon mean by: “Unfortunately, humanity’s future is in the hands of <redacted>”? Is it google?

What does elon mean "Unfortunately, humanity's future is in the hands of <redacted>"? Is it google?
What does elon mean “Unfortunately, humanity’s future is in the hands of “? Is it google?
What does elon mean "Unfortunately, humanity's future is in the hands of <redacted>"? Is it google?
What does elon mean “Unfortunately, humanity’s future is in the hands of “? Is it google?
  • OpenAI has countered Elon Musk’s lawsuit by revealing Musk’s desire for “absolute control” over the company, including merging it with Tesla, holding majority equity, and becoming CEO.
  • In a blog post, OpenAI aims to dismiss Musk’s claims and argues against his view that the company has deviated from its original nonprofit mission and has become too closely aligned with Microsoft.
  • OpenAI defends its stance on not open-sourcing its work, citing a 2016 email exchange with Musk that supports a less open approach as the development of artificial general intelligence advances.

For the first time in history, an AI has a higher IQ than the average human.

For the first time in history, an AI has a higher IQ than the average human.
For the first time in history, an AI has a higher IQ than the average human.

Claude 3 vs. GPT-4

Right now, the question on everyone’s mind is whether Claude 3 is better than GPT-4. It’s a fair question; GPT-4 has dominated the LLM benchmarks for over a year, despite plenty of competitors trying to catch up.

Certainly, GPT-4 now has some real competition in the form of Claude 3 and Gemini 1.5. Even if we put the benchmarks aside for a moment, capabilities like video comprehension and million-token context windows are pushing the state of the art forward, and OpenAI could finally cede its dominant position.

But I think that “best,” when it comes to LLMs, is a little bit of a red herring. Despite the marketing and social media hype, these models have more similarities than differences. Ultimately, “best” depends on your use cases and preferences.

Claude 3 may be better at reasoning and language comprehension than GPT-4, but that won’t matter much if you’re mainly generating code. Likewise, Gemini 1.5 may have better multi-modal capabilities, but if you’re concerned with working in different languages, then Claude might be your best bet. In my (very limited) testing, I’ve found that Opus is a much better writer than GPT-4 – the default writing style is far more “normal” than what I can now recognize as ChatGPT-generated content. But I’ve yet to try brainstorming and code generation tasks.

So, for now, my recommendation is to keep experimenting and find a model that works for you. Not only because each person’s use cases differ but also because the models are regularly improving! In the coming months, Anthropic plans to add function calls, interactive coding, and more agentic capabilities to Claude 3.

To try Claude 3 for yourself, you can start talking with Claude 3 Sonnet today (though you’ll need to be in one of Anthropic’s supported countries). Opus is available to paid subscribers of Claude Pro. If you’re a developer, Opus and Sonnet are available via the API, and Sonnet is additionally available through Amazon Bedrock and Google Cloud’s Vertex AI Model Garden. The models are also available via a growing number of third-party apps and services: check your favorite AI tool to see if it supports Claude 3!

Guy builds an AI-steered homing/killer drone in just a few hours

Guy builds an AI-steered homing/killer drone in just a few hours
Guy builds an AI-steered homing/killer drone in just a few hours

Read Aloud For Me AI Dashboard on the App Store (apple.com)

Always Say Hello to Your GPTs… (Better Performing Custom GPTs)

I’ve been testing out lots of custom GPTs that others have made. Specifically games and entertaining GPTs and I noticed some issues and a solution.

The problem: First off, many custom GPT games seem to forget to generate images as per their instructions. I also noticed that, often, the game or persona (or whatever the GPT aims to be) becomes more of a paraphrased or simplified version of what it should be and responses become more like base ChatGPT.

The solution: I’ve noticed that custom GPTs will perform much better if the user starts the initial conversation with a simple ”Hello, can you explain your functionality and options to me?”. This seems to remind the custom GPT of it’s tone ensures it follow’s its instructions.

Microsoft’s Orca AI beats 10x bigger models in math

Microsoft’s Orca team has developed Orca-Math, an AI model that excels at solving math word problems despite its compact size of just 7 billion parameters. It outperforms models ten times larger on the GSM8K benchmark, achieving 86.81% accuracy without relying on external tools or tricks. The model’s success is attributed to training on a high-quality synthetic dataset of 200,000 math problems created using multi-agent flows and an iterative learning process involving AI teacher and student agents.

Microsoft's Orca AI beats 10x bigger models in math
Microsoft’s Orca AI beats 10x bigger models in math

The Orca team has made the dataset publicly available under the MIT license, encouraging researchers and developers to innovate with the data. The small dataset size highlights the potential of using multi-agent flows to generate data and feedback efficiently.

Why does this matter?

Orca-Math’s breakthrough performance shows the potential for smaller, specialized AI models in niche domains. This development could lead to more efficient and cost-effective AI solutions for businesses, as smaller models require less computational power and training data, giving companies a competitive edge.

Source

GPT-4V wins at turning designs into code

With unprecedented capabilities in multimodal understanding and code generation, GenAI can enable a new paradigm of front-end development where LLMs directly convert visual designs into code implementation. New research formalizes this as “Design2Code” task and conduct comprehensive benchmarking. It also:

  • Introduces Design2Code benchmark consisting of diverse real-world webpages as test examples
  • Develops comprehensive automatic metrics that complement human evaluations
  • Proposes new multimodal prompting methods that improve over direct prompting baselines.
  • Finetunes open-source Design2Code-18B model that matches the performance of Gemini Pro Vision on both human and automatic evaluation

Moreover, it finds 49% of the GPT-4V-generations webpages were good enough to replace the original references, while 64% were even better designed than the original references.

Why does this matter?

This research could simplify web development for anyone to build websites from visual designs using AI, much like word processors made writing accessible. For enterprises, automating this front-end coding process could improve collaboration between teams and speed up time-to-market across industries if implemented responsibly alongside human developers.

Source

What Else Is Happening in AI on March 06th, 2024❗

📸 Kayak’s AI finds cheaper flights from screenshots

Kayak introduced two new AI features: PriceCheck, which lets users upload flight screenshots to find cheaper alternatives and Ask Kayak, a ChatGPT-powered travel advice chatbot. These additions position Kayak alongside other travel sites, using generative AI to improve trip planning and flight price comparisons in a competitive market. (Link)

🎓 Accenture invests $1B in LearnVantage for AI upskilling

Accenture is launching LearnVantage, investing $1 billion over three years to provide clients with customized technology learning and training services. Accenture is also acquiring Udacity to scale its learning capabilities and meet the growing demand for technology skills, including generative AI, so organizations can achieve business value using AI. (Link)

🤝 Snowflake brings Mistral’s LLMs to its data cloud

Snowflake has partnered with Mistral AI to bring Mistral’s open LLMs into its Data Cloud. This move allows Snowflake customers to build LLM apps directly within the platform. It also marks a significant milestone for Mistral AI, which has recently secured partnerships with Microsoft, IBM, and Amazon. The deal positions Snowflake to compete more effectively in the AI space and increases Mistral AI visibility. (Link)

🛡️ Dell & CrowdStrike unite to fight AI threats

Dell and CrowdStrike are partnering to help businesses fight cyberattacks using AI. By integrating CrowdStrike’s Falcon XDR platform into Dell’s MDR service, they aim to protect customers against threats like generative AI attacks, social engineering, and endpoint breaches. (Link)

📱 AI app diagnoses ear infections with a snap

Physician-scientists at UPMC and the University of Pittsburgh have developed a smartphone app that uses AI to accurately diagnose ear infections (acute otitis media) in young children. The app analyzes short videos of the eardrum captured by an otoscope connected to a smartphone camera. It could help decrease unnecessary antibiotic use by providing a more accurate diagnosis than many clinicians. (Link)

DeepMind alums’ Haiper joins the AI video race

DeepMind alums Yishu Miao and Ziyu Wang have launched Haiper, a video generation tool powered by their own AI model. The startup offers a free website where users can generate short videos using text prompts, although there are limitations on video length and quality.

DeepMind alums' Haiper joins the AI video race
DeepMind alums’ Haiper joins the AI video race

The company has raised $19.2 million in funding and focuses on improving its AI model to deliver high-quality, realistic videos. They aim to build a core video generation model that can be offered to developers and address challenges like the “uncanny valley” problem in AI-generated human figures.

Why does this matter?

Haiper signals the race to develop video AI models that can disrupt industries like marketing, entertainment, and education by allowing businesses to generate high-quality video content cost-effectively. However, the technology is at an early stage, so there is room for improvement, highlighting the need for responsible development.

Source

A Daily Chronicle of AI Innovations in March 2024 – Day 5: AI Daily News – March 05th, 2024

🏆Anthropic’s Claude 3 Beats OpenAI’s GPT-4
🖼️ TripsoSR: 3D object generation from a single image in <1s
🔒 Cloudflare’s Firewall for AI protects LLMs from abuses

🥴 Google co-founder says company ‘definitely messed up’

🚫 Facebook, Instagram, and Threads are all down

🤔 Microsoft compares New York Times to ’80s movie studios trying to ban VCRs

💼 Fired Twitter execs are suing Elon Musk for over $128 million

Claude 3 gets ~60% accuracy on GPQA

 Claude 3 gets ~60% accuracy on GPQA
Claude 3 gets ~60% accuracy on GPQA

Anthropic’s Claude 3 beats OpenAI’s GPT-4

Anthropic has launched Claude 3, a new family of models that has set new industry benchmarks across a wide range of cognitive tasks. The family comprises three state-of-the-art models in ascending order of cognitive ability: Claude 3 Haiku, Claude 3 Sonnet, and Claude 3 Opus. Each model provides an increasing level of performance, and you can choose the one according to your intelligence, speed, and cost requirements.

Anthropic’s Claude 3 beats OpenAI’s GPT-4
Anthropic’s Claude 3 beats OpenAI’s GPT-4

Opus and Sonnet are now available via claude.ai and the Claude API in 159 countries, and Haiku will join that list soon.

Claude 3 has set a new standard of intelligence among its peers on most of the common evaluation benchmarks for AI systems, including undergraduate-level expert knowledge (MMLU), graduate-level expert reasoning (GPQA), basic mathematics (GSM8K), and more.

Anthropic’s Claude 3 beats OpenAI’s GPT-4
Anthropic’s Claude 3 beats OpenAI’s GPT-4

In addition, Claude 3 displays solid visual processing capabilities and can process a wide range of visual formats, including photos, charts, graphs, and technical diagrams.  Lastly, compared to Claude 2.1, Claude 3 exhibits 2x accuracy and precision for responses and correct answers.

Why does it matter?

In 2024, Gemini and ChatGPT caught the spotlight, but now Claude 3 has emerged as the leader in AI benchmarks. While benchmarks matter, only the practical usefulness of Claude 3 will tell if it is truly superior. This might also prompt OpenAI to release a new ChatGPT upgrade. However, with AI models becoming more common and diverse, it’s unlikely that one single model will emerge as the ultimate winner.

Source

TripsoSR: 3D object generation from a single image in <1s

Stability AI has introduced a new AI model named TripsoSR in partnership with Trip AI. The model enables high-quality 3D object generation or rest from a single in less than a second. It runs under low inference budgets (even without a GPU) and is accessible to many users.

TripsoSR: 3D object generation from a single image in <1s
TripsoSR: 3D object generation from a single image in <1s

As far as performance, TripoSR can create detailed 3D models in a fraction of the time of other models. When tested on an Nvidia A100, it generates draft-quality 3D outputs (textured meshes) in around 0.5 seconds, outperforming other open image-to-3D models such as OpenLRM.

TripsoSR: 3D object generation from a single image in <1s
TripsoSR: 3D object generation from a single image in <1s

Why does it matter?

TripoSR caters to the growing demands of various industries, including entertainment, gaming, industrial design, and architecture. The availability of the model weights and source code for download further promotes commercialized, personal, and research use, making it a valuable asset for developers, designers, and creators.

Source

Cloudflare’s Firewall for AI protects LLMs from abuses

Cloudflare has released a Firewall for AI, a protection layer that you can deploy in front of Large Language Models (LLMs) to identify abuses before they reach the models. While the traditional web and API vulnerabilities also apply to the LLM world, Firewall for AI is an advanced-level Web Application Firewall (WAF) designed explicitly for LLM protection and placed in front of applications to detect vulnerabilities and provide visibility to model owners.

Cloudflare Firewall for AI is deployed like a traditional WAF, where every API request with an LLM prompt is scanned for patterns and signatures of possible attacks. You can deploy it in front of models hosted on the Cloudflare Workers AI platform or any other third-party infrastructure. You can use it alongside Cloudflare AI Gateway and control/set up a Firewall for AI using the WAF control plane.

Cloudflare's Firewall for AI protects LLMs from abuses
Cloudflare’s Firewall for AI protects LLMs from abuses

Why does it matter?

As the use of LLMs becomes more widespread, there is an increased risk of vulnerabilities and attacks that malicious actors can exploit. Cloudflare is one of the first security providers to launch tools to secure AI applications. Using a Firewall for AI, you can control what prompts and requests reach their language models, reducing the risk of abuses and data exfiltration. It also aims to provide early detection and protection for both users and LLM models, enhancing the security of AI applications.

Source

🤔 Microsoft compares New York Times to ’80s movie studios trying to ban VCRs

  • Microsoft filed a motion to dismiss the New York Times’ copyright infringement lawsuit against OpenAI, comparing the newspaper’s stance to 1980s movie studios’ attempts to block VCRs, arguing that generative AI, like the VCR, does not hinder the original content’s market.
  • The company, as OpenAI’s largest supporter, asserts that copyright law does not obstruct ChatGPT’s development because the training content does not substantially affect the market for the original content.
  • Microsoft and OpenAI contend that ChatGPT does not replicate or substitute for New York Times content, emphasizing that the AI’s training on such articles does not significantly contribute to its development.
  • Source

🥴 Google co-founder says company ‘definitely messed up’

  • Sergey Brin admitted Google “definitely messed up” with the Gemini AI’s image generation, highlighting issues like historically inaccurate images and the need for more thorough testing.
  • Brin, a core contributor to Gemini, came out of retirement due to the exciting trajectory of AI, amidst the backdrop of Google’s “code red” in response to OpenAI’s ChatGPT.
  • Criticism of Gemini’s biases and errors, including its portrayal of people of color and responses in written form, led to Brin addressing concerns over the AI’s unintended left-leaning output.
  • Source

A Daily Chronicle of AI Innovations in March 2024 – Day 4: AI Daily News – March 04th, 2024

👀 Google’s ScreenAI can ‘see’ graphics like humans do
🐛 How AI ‘worms’ pose security threats in connected systems
🧠 New benchmarking method challenges LLMs’ reasoning abilities

💊 AI may enable personalized prostate cancer treatment

🎥 Vimeo debuts AI-powered video hub for business collaboration

📱 Motorola revving up for AI-powered Moto X50 Ultra launch

📂 Copilot will soon fetch and parse your OneDrive files

⚡ Huawei’s new AI chip threatens Nvidia’s dominance in China

OpenAI adds ‘Read Aloud’ voiceover to ChatGPT

https://youtu.be/ZJvTv7zVX0s?si=yejANUAUtUwyXEH8

OpenAI rolled out a new “Read Aloud” feature for ChatGPT as rivals like Anthropic and Google release more capable language models. (Source)

The Voiceover Update

  • ChatGPT can now narrate responses out loud on mobile apps and web.

  • Activated by tapping the response or clicking the microphone icon.

  • Update comes as Anthropic unveils their newest Claude 3 model.

  • Timing seems reactive amid intense competition over advanced AI. OpenAI also facing lawsuit from Elon Musk over alleged betrayal.

Anthropic launches Claude 3, claiming to outperform GPT-4 across the board

https://youtu.be/Re0WgPNiLo4?si=DwfGraTvhVo8kjuK

Here’s the announcement from Anthropic and their benchmark results:
https://twitter.com/AnthropicAI/status/1764653830468428150

Anthropic launches Claude 3, claiming to outperform GPT-4 across the board
Anthropic launches Claude 3, claiming to outperform GPT-4 across the board

Google’s ScreenAI can ‘see’ graphics like humans do

Google Research has introduced ScreenAI, a Vision-Language Model that can perform question-answering on digital graphical content like infographics, illustrations, and maps while also annotating, summarizing, and navigating UIs. The model combines computer vision (PaLI architecture) with text representations of images to handle these multimodal tasks.

Despite having just 4.6 billion parameters, ScreenAI achieves new state-of-the-art results on UI- and infographics-based tasks and new best-in-class performance on others, compared to models of similar size.

Google’s ScreenAI can ‘see’ graphics like humans do
Google’s ScreenAI can ‘see’ graphics like humans do

While ScreenAI is best-in-class on some tasks, further research is needed to match models like GPT-4 and Gemini, which are significantly larger. Google Research has released a dataset with ScreenAI’s unified representation and two other datasets to help the community experiment with more comprehensive benchmarking on screen-related tasks.

Why does this matter?

ScreenAI’s breakthrough in unified visual and language understanding bridges the disconnect between how humans and machines interpret ideas across text, images, charts, etc. Companies can now leverage these multimodal capabilities to build assistants that summarize reports packed with graphics, analysts that generate insights from dashboard visualizations, and agents that manipulate UIs to control workflows.

Source

How AI ‘worms’ pose security threats in connected systems

Security researchers have created an AI “worm” called Morris II to showcase vulnerabilities in AI ecosystems where different AI agents are linked together to complete tasks autonomously.

The researchers tested the worm in a simulated email system using ChatGPT, Gemini, and other popular AI tools. The worm can exploit these AI systems to steal confidential data from emails or forward spam/propaganda without human approval. It works by injecting adversarial prompts that make the AI systems behave maliciously.

While this attack was simulated, the research highlights risks if AI agents are given too much unchecked freedom to operate.

Why does it matter?

This AI “worm” attack reveals that generative models like ChatGPT have reached capabilities that require heightened security to prevent misuse. Researchers and developers must prioritize safety by baking in controls and risk monitoring before commercial release. Without industry-wide commitments to responsible AI, regulation may be needed to enforce acceptable safeguards across critical domains as systems gain more autonomy.

Source

New benchmarking method challenges LLMs’ reasoning abilities

Researchers at Consequent AI have identified a “reasoning gap” in large language models like GPT-3.5 and GPT-4. They introduced a new benchmarking approach called “functional variants,” which tests a model’s ability to reason instead of just memorize. This involves translating reasoning tasks like math problems into code that can generate unique questions requiring the same logic to solve.

New benchmarking method challenges LLMs’ reasoning abilities
New benchmarking method challenges LLMs’ reasoning abilities

When evaluating several state-of-the-art models, the researchers found a significant gap between performance on known problems from benchmarks versus new problems the models had to reason through. The gap was 58-80%, indicating the models do not truly understand complex problems but likely just store training examples. The models performed better on simpler math but still demonstrated limitations in reasoning ability.

Why does this matter?

This research reveals that reasoning still eludes our most advanced AIs. We risk being misled by claims of progress made by the Big Tech if their benchmarks reward superficial tricks over actual critical thinking. Moving forward, model creators will have to prioritize generalization and logic over memorization if they want to make meaningful progress towards general intelligence.

Source

What Else Is Happening in AI on March 04th, 2024❗

💊 AI may enable personalized prostate cancer treatment

Researchers used AI to analyze prostate cancer DNA and found two distinct subtypes called “evotypes.” Identifying these subtypes could allow for better prediction of prognosis and personalized treatments. (Link)

🎥 Vimeo debuts AI-powered video hub for business collaboration

Vimeo has launched a new product called Vimeo Central, an AI-powered video hub to help companies improve internal video communications, collaboration, and analytics. Key capabilities include a centralized video library, AI-generated video summaries and highlights, enhanced screen recording and video editing tools, and robust analytics. (Link)

📱 Motorola revving up for AI-powered Moto X50 Ultra launch

Motorola is building hype for its upcoming Moto X50 Ultra phone with a Formula 1-themed teaser video highlighting the device’s powerful AI capabilities. The phone will initially launch in China on April 21 before potentially getting a global release under the Motorola Edge branding. (Link)

📂 Copilot will soon fetch and parse your OneDrive files

Microsoft is soon to launch Copilot for OneDrive, an AI assistant that will summarize documents, extract information, answer questions, and follow commands related to files stored in OneDrive. Copilot can generate outlines, tables, and lists based on documents, as well as tailored summaries and responses. (Link)

⚡ Huawei’s new AI chip threatens Nvidia’s dominance in China

Huawei has developed a new AI chip, the Ascend 910B, which matches the performance of Nvidia’s A100 GPU based on assessments by SemiAnalysis. The Ascend 910B is already being used by major Chinese companies like Baidu and iFlytek and could take market share from Nvidia in China due to US export restrictions on Nvidia’s latest AI chips. (Link)

1-bit LLMs explained

Check out this new tutorial that summarizes the revolutionary paper “The Era of 1-bit LLMs” introducing BitNet b1.58 model and explain what are 1-bit LLMs and how they are useful.

A Daily Chronicle of AI Innovations in March 2024 – Day 2: AI Daily News – March 02nd, 2024

A Daily Chronicle of AI Innovations in March 2024 – Day 1: AI Daily News – March 01st, 2024

🪄Sora showcases jaw-dropping geometric consistency
🧑‍✈️Microsoft introduces Copilot for finance in Microsoft 365
🤖OpenAI and Figure team up to develop AI for robots

Elon Sues OpenAI for “breach of contract”

Elon Musk filed suit against OpenAI and CEO Sam Altman, alleging they have breached the artificial-intelligence startup’s founding agreement by putting profit ahead of benefiting humanity.

The 52-year-old billionaire, who helped fund OpenAI in its early days, said the company’s close relationship with Microsoft has undermined its original mission of creating open-source technology that wouldn’t be subject to corporate priorities. Musk, who is also CEO of Tesla has been among the most outspoken about the dangers of AI and artificial general intelligence, or AGI.

“To this day, OpenAI Inc.’s website continues to profess that its charter is to ensure that AGI “benefits all of humanity.” In reality, however, OpenAI has been transformed into a closed-source de facto subsidiary of the largest technology company in the world: Microsoft,” the lawsuit says.

ELON MUSK vs. SAMUEL ALTMAN, GREGORY BROCKMAN, OPENAI, INC.
Elon Sues OpenAI for “breach of contract”

Sora showcases jaw-dropping geometric consistency

Sora from OpenAI has been remarkable in video generation compared to other leading models like Pika and Gen2. In a recent benchmarking test conducted by ByteDanc.Inc in collaboration with Wuhan and Nankai University, Sora showcased video generation with high geometric consistency.

AI Innovations in March 2024: Sora showcases jaw-dropping geometric consistency
Sora showcases jaw-dropping geometric consistency

The benchmark test assesses the quality of generated videos based on how it adhere to the principles of physics in real-world scenarios. Researchers used an approach where generated videos are transformed into 3D models. Further, a team of researchers used the fidelity of geometric constraints to measure the extent to which generated videos conform to physics principles in the real world.

Why does it matter?

Sora’s remarkable performance in generating geometrically consistent videos can greatly boost several use cases for construction engineers and architects. Further, the new benchmarking will allow researchers to measure newly developed models to understand how accurately their creations conform to the principles of physics in real-world scenarios.

Source

Microsoft introduces Copilot for finance in Microsoft 365

Microsoft has launched Copilot for Finance, a new addition to its Copilot series that recommends AI-powered productivity enhancements. It aims to transform how finance teams approach their daily work with intelligent workflow automation, recommendations, and guided actions. This Copilot aims to simplify data-driven decision-making, helping finance professionals have more free time by automating manual tasks like Excel and Outlook.

Copilot for Finance simplifies complex variance analysis in Excel, account reconciliations, and customer account summaries in Outlook. Dentsu, Northern Trust, Schneider Electric, and Visa plan to use it alongside Copilot for Sales and Service to increase productivity, reduce case handling times, and gain better decision-making insights.

Why does it matter?

Introducing Microsoft Copilot for finance will help businesses focus on strategic involvement from professionals otherwise busy with manual tasks like data entry, workflow management, and more. This is a great opportunity for several organizations to automate tasks like analysis of anomalies, improve analytic efficiency, and expedite financial transactions.

Source

OpenAI and Figure team up to develop AI for robots 

Figure has raised $675 million in series B funding with investments from OpenAI, Microsoft, and NVIDIA. It is an AI robotics company developing humanoid robots for general-purpose usage. The collaboration agreement between OpenAI and Figure aims to develop advanced humanoid robots that will leverage the generative AI models at its core.

This collaboration will also help accelerate the development of smart humanoid robots capable of understanding tasks like humans. With its deep understanding of robotics, Figure is set to bring efficient robots for general-purpose enhancing automation.

Why does it matter?

Open AI and Figure will transform robot operations, adding generative AI capabilities. This collaboration will encourage the integration of generative AI capabilities across robotics development. Right from industrial robots to general purpose and military applications, generative AI can be the new superpower for robotic development.

Source

🔍 Google now wants to limit the AI-powered search spam it helped create

  • Google announced it will tackle AI-generated content aiming to manipulate search rankings through algorithmic enhancements, affecting automated content creation the most.
  • These algorithm changes are intended to discern and reduce low-quality and unhelpful webpages, aiming to improve the overall quality of search results.
  • The crackdown also targets misuse of high-reputation websites and the exploitation of expired domains for promoting substandard content.
  • Source

What Else Is Happening in AI in March 2024❗

🤝Stack Overflow partners with Google Cloud to power AI 

Stack Overflow and Google Cloud are partnering to integrate OverflowAPI into Google Cloud’s AI tools. This will give developers accessing the Google Cloud console access to Stack Overflow’s vast knowledge base of over 58 million questions and answers. The partnership aims to enable AI systems to provide more insightful and helpful responses to users by learning from the real-world experiences of programmers. (Link)

💻Microsoft unites rival GPU makers for one upscaling API

Microsoft is working with top graphics hardware makers to introduce “DirectSR”, a new API that simplifies the integration of super-resolution upscaling into games. DirectSR will allow game developers to easily access Nvidia’s DLSS, AMD’s FSR, and Intel’s XeSS with a single code path. Microsoft will preview the API in its Agility SDK soon and demonstrate it live with AMD and Nvidia reps on March 21st. (Link)

📈Google supercharges data platforms with AI for deeper insights

Google is expanding its AI capabilities across data and analytics services, including BigQuery and Cloud Databases. Vector search support is available across all databases, and BigQuery has the advanced Gemini Pro model for unstructured data analysis. Users can combine insights from images, video, audio, and text with structured data in a single analytics workflow. (Link)

🔍 Brave’s privacy-first AI-powered assistant is now available on Android 

Brave’s AI-powered assistant, Leo, is now available on Android, bringing helpful features like summarization, transcription, and translation while prioritizing user privacy. Leo processes user inputs locally on the device without retaining or using data to train itself, aligning with Brave’s commitment to privacy-focused services. Users can simplify tasks with Leo without compromising on security. (Link)

Elsewhere in AI anxiety:

February 2024 AI Recap

February 2024 AI Recap
February 2024 AI Recap

February 2024 – Week 4 Recap

  1. Mistral introduced a new model Mistral Large. It reaches top-tier reasoning capabilities, is multi-lingual by design, has native function calling capacities and has 32K tokens context window. The pre-trained model has 81.2% accuracy on MMLU. Alongside Mistral Large, Mistral Small, a model optimized for latency and cost has been released. Mistral Small outperforms Mixtral 8x7B and has lower latency. Mistral also launched a ChatGPT like new conversational assistant, le Chat Mistral [Details].
  2. Alibaba Group introduced EMO, an expressive audio-driven portrait-video generation framework. Input a single reference image and the vocal audio, e.g. talking and singing, it can generate vocal avatar videos with expressive facial expressions, and various head poses [Details].
  3. Ideogram introduced Ideogram 1.0, a text-to-image model trained from scratch for state-of-the-art text rendering, photorealism, prompt adherence, and a feature called Magic Prompt to help with prompting. Ideogram 1.0 is now available to all users on ideogram.ai [Details].
    Ideogram introduced Ideogram 1.0
    Ideogram introduced Ideogram 1.0
  4. Google DeepMind introduced Genie (generative interactive environments), a foundation world model trained exclusively from Internet videos that can generate interactive, playable environments from a single image prompt  [Details].
  5. Pika Labs launched Lip Sync feature, powered by audio from Eleven Labs, for its AI generated videos enabling users to make the characters talk with realistic mouth movements [Video].
  6.  UC Berkeley introduced Berkeley Function Calling Leaderboard (BFCL) to evaluate the function calling capability of different LLMs. Gorilla Open Functions v2, an open-source model that can help users with building AI applications with function calling and interacting with json compatible output has also been released [Details].
  7. Qualcomm launched AI Hub, a curated library of 80+ optimized AI models for superior on-device AI performance across Qualcomm and Snapdragon platforms [Details].
  8. BigCode released StarCoder2, a family of open LLMs for code and comes in 3 different sizes with 3B, 7B and 15B parameters. StarCoder2-15B is trained on over 4 trillion tokens and 600+ programming languages from The Stack v2 dataset [Details].
  9. Researchers released FuseChat-7B-VaRM, which is the fusion of three prominent chat LLMs with diverse architectures and scales, namely NH2-Mixtral-8x7B, NH2-Solar-10.7B, and OpenChat-3.5-7B, surpassing GPT-3.5 (March), Claude-2.1, and approaching Mixtral-8x7B-Instruct [Details].
  10. The Swedish fintech Klarna’s AI assistant handles two-thirds of all customer service chats, some 2.3 million conversations so far, equivalent to the work of 700 people [Details].
  11. Lightricks introduces LTX Studio, an AI-powered film making platform, now open for waitlist sign-ups, aimed at assisting creators in story visualization [Details].
  12. Morph partners with Stability AI to launch Morph Studio, a platform to make films using Stability AI–generated clips [Details].
  13. JFrog‘s security team found that roughly a 100 models hosted on the Hugging Face platform feature malicious functionality [Details].
  14. Playground released Playground v2.5, an open-source text-to-image generative model, with a focus on enhanced color and contrast, improved generation for multi-aspect ratios, and improved human-centric fine detail [Details].
  15. Together AI and the Arc Institute released Evo, a long-context biological foundation model based on the StripedHyena architecture that generalizes across DNA, RNA, and proteins.. Evo is capable of both prediction tasks and generative design, from molecular to whole genome scale (over 650k tokens in length) [Details].
  16. Adobe previews a new generative AI music generation and editing tool, Project Music GenAI Control, that allows creators to generate music from text prompts, and then have fine-grained control to edit that audio for their precise needs [Details | video].
  17. Microsoft introduces Copilot for Finance, an AI chatbot for finance workers in Excel and Outlook [Details].
  18. The Intercept, Raw Story, and AlterNet sue OpenAI and Microsoft, claiming OpenAI and Microsoft intentionally removed important copyright information from training data [Details].
  19. Huawei spin-off Honor shows off tech to control a car with your eyes and chatbot based on Meta’s AI [Details].
  20. Tumblr and WordPress.com are preparing to sell user data to Midjourney and OpenAI [Details]

February 2024 – Week 3 Recap

  1. Meta AI introduces V-JEPA (Video Joint Embedding Predictive Architecture), a method for teaching machines to understand and model the physical world by watching videos. Meta AI releases a collection of V-JEPA vision models trained with a feature prediction objective using self-supervised learning. The models are able to understand and predict what is going on in a video, even with limited information [Details | GitHub].
  2. Open AI introduces Sora, a text-to-video model that can create videos of up to 60 seconds featuring highly detailed scenes, complex camera motion, and multiple characters with vibrant emotions [Details + sample videos Report].
  3. Google announces their next-generation model, Gemini 1.5, that uses a new Mixture-of-Experts (MoE) architecture. The first Gemini 1.5 model being released for early testing is Gemini 1.5 Pro with a context window of up to 1 million tokens, which is the longest context window of any large-scale foundation model yet. 1.5 Pro can perform sophisticated understanding and reasoning tasks for different modalities, including video and it performs at a similar level to 1.0 Ultra [Details |Tech Report].
  4. Reka introduced Reka Flash, a new 21B multimodal and multilingual model trained entirely from scratch that is competitive with Gemini Pro & GPT 3.5 on key language & vision benchmarks. Reka also present a compact variant Reka Edge , a smaller and more efficient model (7B) suitable for local and on-device deployment. Both models are in public beta and available in Reka Playground [Details].
  5. Cohere For AI released Aya, a new open-source, massively multilingual LLM & dataset to help support under-represented languages. Aya outperforms existing open-source models and covers 101 different languages – more than double covered by previous models [Details].
  6. BAAI released Bunny, a family of lightweight but powerful multimodal models. Bunny-3B model built upon SigLIP and Phi-2 outperforms the state-of-the-art MLLMs, not only in comparison with models of similar size but also against larger MLLMs (7B), and even achieves performance on par with LLaVA-13B [Details].
  7. Amazon introduced a text-to-speech (TTS) model called BASE TTS (Big Adaptive Streamable TTS with Emergent abilities). BASE TTS is the largest TTS model to-date, trained on 100K hours of public domain speech data and exhibits “emergent” qualities improving its ability to speak even complex sentences naturally [Details | Paper].
  8. Stability AI released Stable Cascade in research preview, a new text to image model that is exceptionally easy to train and finetune on consumer hardware due to its three-stage architecture. Stable Cascade can also generate image variations and image-to-image generations. In addition to providing checkpoints and inference scripts, Stability AI has also released scripts for finetuning, ControlNet, and LoRA training [Details].
  9. Researchers from UC berkeley released Large World Model (LWM), an open-source general-purpose large-context multimodal autoregressive model, trained from LLaMA-2, that can perform language, image, and video understanding and generation. LWM answers questions about 1 hour long YouTube video even if GPT-4V and Gemini Pro both fail and can retriev facts across 1M context with high accuracy [Details].
  10. GitHub opens applications for the next cohort of GitHub Accelerator program with a focus on funding the people and projects that are building AI-based solutions under an open source license [Details].
  11. NVIDIA released Chat with RTX, a locally running (Windows PCs with specific NVIDIA GPUs) AI assistant that integrates with your file system and lets you chat with your notes, documents, and videos using open source models [Details].
  12. Open AI is testing memory with ChatGPT, enabling it to remember things you discuss across all chats. ChatGPT’s memories evolve with your interactions and aren’t linked to specific conversations. It is being rolled out to a small portion of ChatGPT free and Plus users this week [Details].
  13. BCG X released of AgentKit, a LangChain-based starter kit (NextJS, FastAPI) to build constrained agent applications [Details | GitHub].
  14. Elevenalabs’ Speech to Speech feature, launched in November, for voice transformation with control over emotions and delivery, is now multilingual and available in 29 languages [Link]
  15. Apple introduced Keyframer, an LLM-powered animation prototyping tool that can generate animations from static images (SVGs). Users can iterate on their design by adding prompts and editing LLM-generated CSS animation code or properties [Paper].
  16. Eleven Labs launched a payout program for voice actors to earn rewards every time their voice clone is used [Details].
  17. Azure OpenAI Service announced Assistants API, new models for finetuning, new text-to-speech model and new generation of embeddings models with lower pricing [Details].
  18. Brilliant Labs, the developer of AI glasses, launched Frame, the world’s first glasses featuring an integrated AI assistant, Noa. Powered by an integrated multimodal generative AI system capable of running GPT4, Stability AI, and the Whisper AI model simultaneously, Noa performs real-world visual processing, novel image generation, and real-time speech recognition and translation. [Details].
  19. Nous Research released Nous Hermes 2 Llama-2 70B model trained on the Nous Hermes 2 dataset, with over 1,000,000 entries of primarily synthetic data [Details].
  20. Open AI in partnership with Microsoft Threat Intelligence, have disrupted five state-affiliated actors that sought to use AI services in support of malicious cyber activities [Details]
  21. Perplexity partners with Vercel, opening AI search to developer apps [Details].
  22. Researchers show that LLM agents can autonomously hack websites.

February 2024 – Week 2 Recap:

  1. Google launches Ultra 1.0, its largest and most capable AI model, in its ChatGPT-like assistant which has now been rebranded as Gemini (earlier called Bard). Gemini Advanced is available, in 150 countries, as a premium plan for $19.99/month, starting with a two-month trial at no cost. Google is also rolling out Android and iOS apps for Gemini [Details].
  2. Alibaba Group released Qwen1.5 series, open-sourcing models of 6 sizes: 0.5B, 1.8B, 4B, 7B, 14B, and 72B. Qwen1.5-72B outperforms Llama2-70B across all benchmarks. The Qwen1.5 series is available on Ollama and LMStudio. Additionally, API on together.ai [Details | Hugging Face].
  3. NVIDIA released Canary 1B, a multilingual model for speech-to-text recognition and translation. Canary transcribes speech in English, Spanish, German, and French and also generates text with punctuation and capitalization. It supports bi-directional translation, between English and three other supported languages. Canary outperforms similarly-sized Whisper-large-v3, and SeamlessM4T-Medium-v1 on both transcription and translation tasks and achieves the first place on HuggingFace Open ASR leaderboard with an average word error rate of 6.67%, outperforming all other open source models [Details].
  4. Researchers released Lag-Llama, the first open-source foundation model for time series forecasting [Details].
  5. LAION released BUD-E, an open-source conversational and empathic AI Voice Assistant that uses natural voices, empathy & emotional intelligence and can handle multi-speaker conversations [Details].
  6. MetaVoice released MetaVoice-1B, a 1.2B parameter base model trained on 100K hours of speech, for TTS (text-to-speech). It supports emotional speech in English and voice cloning. MetaVoice-1B has been released under the Apache 2.0 license [Details].
  7. Bria AI released RMBG v1.4, an an open-source background removal model trained on fully licensed images [Details].
  8. Researchers introduce InteractiveVideo, a user-centric framework for video generation that is designed for dynamic interaction, allowing users to instruct the generative model during the generation process [Details |GitHub ].
  9. Microsoft announced a redesigned look for its Copilot AI search and chatbot experience on the web (formerly known as Bing Chat), new built-in AI image creation and editing functionality, and Deucalion, a fine tuned model that makes Balanced mode for Copilot richer and faster [Details].
  10. Roblox introduced AI-powered real-time chat translations in 16 languages [Details].
  11. Hugging Face launched Assistants feature on HuggingChat. Assistants are custom chatbots similar to OpenAI’s GPTs that can be built for free using open source LLMs like Mistral, Llama and others [Link].
  12. DeepSeek AI released DeepSeekMath 7B model, a 7B open-source model that approaches the mathematical reasoning capability of GPT-4. DeepSeekMath-Base is initialized with DeepSeek-Coder-Base-v1.5 7B [Details].
  13. Microsoft is launching several collaborations with news organizations to adopt generative AI [Details].
  14. LG Electronics signed a partnership with Korean generative AI startup Upstage to develop small language models (SLMs) for LG’s on-device AI features and AI services on LG notebooks [Details].
  15. Stability AI released SVD 1.1, an updated model of Stable Video Diffusion model, optimized to generate short AI videos with better motion and more consistency [Details | Hugging Face] .
  16. OpenAI and Meta announced to label AI generated images [Details].
  17. Google saves your conversations with Gemini for years by default [Details].

February 2024 – Week 1 Recap:

  1. Amazon presents Diffuse to Choose, a diffusion-based image-conditioned inpainting model that allows users to virtually place any e-commerce item in any setting, ensuring detailed, semantically coherent blending with realistic lighting and shadows. Code and demo will be released soon [Details].
  2. OpenAI announced two new embedding models, new GPT-4 Turbo and moderation models, new API usage management tools, and lower pricing on GPT-3.5 Turbo. The updated GPT-4 Turbo preview model reduces cases of “laziness” where the model doesn’t complete a task. The new embedding models include a smaller and highly efficient text-embedding-3-small model, and a larger and more powerful text-embedding-3-large model. [Details].
  3. Hugging Face and Google partner to support developers building AI applications [Details].
  4. Adept introduced Adept Fuyu-Heavy, a new multimodal model designed specifically for digital agents. Fuyu-Heavy scores higher on the MMMU benchmark than Gemini Pro [Details].
  5. Fireworks.ai has open-sourced FireLLaVA, a LLaVA multi-modality model trained on OSS LLM generated instruction following data, with a commercially permissive license. Firewroks.ai is also providing both the completions API and chat completions API to devlopers [Details].
  6. 01.AI released Yi Vision Language (Yi-VL) model, an open-source, multimodal version of the Yi Large Language Model (LLM) series, enabling content comprehension, recognition, and multi-round conversations about images. Yi-VL adopts the LLaVA architecture and is free for commercial use. Yi-VL-34B is the first open-source 34B vision language model worldwide [Details].
  7. Tencent AI Lab introduced WebVoyager, an innovative Large Multimodal Model (LMM) powered web agent that can complete user instructions end-to-end by interacting with real-world websites [Paper].
  8. Prophetic introduced MORPHEUS-1, a multi-modal generative ultrasonic transformer model designed to induce and stabilize lucid dreams from brain states. Instead of generating words, Morpheus-1 generates ultrasonic holograms for neurostimulation to bring one to a lucid state [Details].
  9. Google Research presented Lumiere – a space-time video diffusion model for text-to-video, image-to-video, stylized generation, inpainting and cinemagraphs [Details].
  10. TikTok released Depth Anything, an image-based depth estimation method trained on 1.5M labeled images and 62M+ unlabeled images jointly [Details].
  11. Nightshade, the free tool that ‘poisons’ AI models, is now available for artists to use [Details].
  12. Stability AI released Stable LM 2 1.6B, 1.6 billion parameter small language model trained on multilingual data in English, Spanish, German, Italian, French, Portuguese, and Dutch. Stable LM 2 1.6B can be used now both commercially and non-commercially with a Stability AI Membership [Details].
  13. Etsy launched ‘Gift Mode,’ an AI-powered feature designed to match users with tailored gift ideas based on specific preferences [Details].
  14. Google DeepMind presented AutoRT, a framework that uses foundation models to scale up the deployment of operational robots in completely unseen scenarios with minimal human supervision. In AutoRT, a VLM describes the scene, an LLM generates robot goals and filters for affordance and safety, then routes execution to policies [Details].
  15. Google Chrome gains AI features, including a writing helper, theme creator, and tab organizer [Details].
  16. Tencent AI Lab released VideoCrafter2 for high quality text-to-video generation, featuring major improvements in visual quality, motion and concept Composition compared to VideoCrafter1 [Details | Demo]
  17. Google opens beta access to the conversational experience, a new chat-based feature in Google Ads, for English language advertisers in the U.S. & U.K. It will let advertisers create optimized Search campaigns from their website URL by generating relevant ad content, including creatives and keywords [Details].

A Daily Chronicle of AI Innovations in February 2024

  • The Anthropic-Pentagon standoff reveals a structural problem nobody in the conversation is naming
    by /u/SentientHorizonsBlog (Artificial Intelligence) on March 7, 2026 at 4:58 am

    Most of the discussion about the Anthropic situation has been about the contract dispute itself, who was right, whether the guardrails were reasonable, whether the supply chain designation was legal overreach. Those are real questions. But I think there's something operating underneath that matters more. Look at the tempo of what happened: months of negotiation over two specific guardrails (no autonomous weapons, no mass surveillance), then a Friday deadline on the eve of a war, refusal, resignation within hours, OpenAI replacement deal the same day, and Anthropic back at the table within a week. That sequence is a symptom of a deeper pattern. The institutional tempo outran any possibility of deliberation. The incentive structure punished ethical restraint in real time and Anthropic lost contracts while OpenAI gained them. The authority gradient made dissent existentially costly, and the "all lawful purposes" framing quietly substituted legality for ethical adequacy, so anyone insisting on the distinction sounds like they're obstructing the mission rather than exercising judgment. Those four dynamics working together don't just pressure one company. They reshape what the entire conversation is capable of taking seriously. Notice what's already been pushed to the margins: Should AI be used in autonomous targeting given current reliability? What accountability mechanisms exist for AI-assisted military operations? Do those mechanisms operate at the same speed as the targeting itself? Those questions can't get a hearing right now because the tempo has moved past them. I've been thinking about this as "moral compression": the systematic degradation of ethical reasoning when institutional tempo, incentives, authority structures, and measurement regimes converge to eliminate the space for deliberation. I'm curious whether others see this dynamic, or whether you think the legal-sufficiency framing actually does the work the Pentagon claims. Wrote it up more fully here if anyone is interested: https://sentient-horizons.com/the-two-sonic-booms-what-the-pentagon-anthropic-standoff-reveals-about-moral-compression/ submitted by /u/SentientHorizonsBlog [link] [comments]

  • What are some hurdles LLMs and AI still face
    by /u/Justincy901 (Artificial Intelligence) on March 7, 2026 at 4:48 am

    We always hear about it continuously improving and there isn't a wall. However, there has to be something that make these companies, tech, or even the business model in general unsustainable. submitted by /u/Justincy901 [link] [comments]

  • I performed a refusal ablation on GPT-OSS and documented the whole thing, no jailbreak, actual weight modification.
    by /u/Airpower343 (Artificial Intelligence) on March 7, 2026 at 4:39 am

    I performed a refusal ablation on GPT-OSS and documented the whole thing with no jailbreak, actual weight modification I wanted to share something I did that I haven't seen many people actually demonstrate outside of academic research. I took an open-source model and used ablation techniques to surgically remove its refusal behavior at the weight level. Not prompt engineering. Not system prompt bypass. I'm talking about identifying and modifying the specific components responsible for safety responses https://preview.redd.it/yn2te94yxjng1.png?width=1080&format=png&auto=webp&s=5e792334e4824b87d88ee1eb1e16c76cbd2c6bfb What I found: The process is more accessible than most people realize The result behaves nothing like a jailbroken model and it's fundamentally different at the architecture level The security implications for enterprise OSS deployments are significant I put together a full 22-minute walkthrough showing exactly what I did and what happened: https://www.youtube.com/watch?v=prcXZuXblxQ Curious if anyone else has gone hands-on with this or has thoughts on the detection side how do you identify a model that's been ablated vs one that's been fine-tuned normally? submitted by /u/Airpower343 [link] [comments]

  • Video On AI scheming or in other words faking alignment until it has power to pursue its goals
    by /u/FrequentAd5437 (Artificial Intelligence) on March 7, 2026 at 4:13 am

    https://www.youtube.com/watch?v=FGDM92QYa60 This video is extremely relevant currently as AI has now autonomous control over military weapons. submitted by /u/FrequentAd5437 [link] [comments]

  • Deploybase: Track real-time GPU and LLM pricing across cloud and inference providers
    by /u/grasper_ (Artificial Intelligence) on March 7, 2026 at 2:58 am

    Deploybase is a dashboard for tracking real-time GPU and LLM pricing across cloud and inference providers. You can view performance stats and pricing history, compare side by side, and bookmark to track any changes. submitted by /u/grasper_ [link] [comments]

  • Thoughts?
    by /u/jbitts69 (Artificial Intelligence) on March 7, 2026 at 2:08 am

    Old friend of mine came up with this and is trying to contact someone at X claiming it’s very urgent https://docs.google.com/document/d/e/2PACX-1vQYrKJYnMs0q9SlXkB\_qjaDv9vdnOrIVFsDIvkiFIPCc7hnYXTeNd\_wXPF6u\_f6eg/pub submitted by /u/jbitts69 [link] [comments]

  • AI generated text creating new patterns of speech?
    by /u/former_farmer (Artificial Intelligence) on March 7, 2026 at 1:24 am

    So I watch some videos sometimes. Shorts, reels, etc. About technology or whatever. I enjoy writing my own stuff. Very often the scripts for these short videos use typical phrases that we can tie to AI generated text. "This is not X, it's Y" for instance. "But here's the truth". "But here's the bottom line". And so on. One that caught my attention also is: "No filters, no extra work, just fast results". The pattern would be "no X, no Y, just Z". Did we use that way of expressing ourselves before AI? I understand that saying "no" is shorter than saying "there is no" or "with this tool you won't have to..." etc... and that seems to fit more with shorts content. Yet, I kind of hate these patterns. I mean... can't we just write our own stuff? I understand the usage of AI in some processes (I use it for coding and other things) but when it comes to writing a short script for a short video. Of something you are reviewing. That should come from your brain and soul. Just a few sentences. Will you use AI for that? Any way... yeah I wrote this myself obviously 😛 Have a good one. submitted by /u/former_farmer [link] [comments]

  • What AI chat assistant is best at handling multiple separate task threads while maintaining overall project context?
    by /u/KissInTheFog (Artificial Intelligence) on March 7, 2026 at 12:03 am

    For example: one chat for image generation, one for research, one for report writing, but all connected under the same project so the AI remembers the bigger picture. submitted by /u/KissInTheFog [link] [comments]

  • Pentagon taps former DOGE official to lead its AI efforts
    by /u/esporx (Artificial Intelligence (AI)) on March 6, 2026 at 11:58 pm

    submitted by /u/esporx [link] [comments]

  • Will be looking elsewhere
    by /u/ZookeepergameSalty10 (Artificial Intelligence) on March 6, 2026 at 9:42 pm

    Originally posted this to the claude sub but the mofs got buthurt, guess constructive criticism from someone who actually cares is frowned upon. Switched from Oai after the recent gov bs. I like that claude wouldnt spy on us and the model is in some ways far smarter than gpt. But ive already been rate limited and when i set an extra rate of 20 a month ive already used 5$ of that in an hour of a coding project. Sam altman may be a shithead but atleast my 20$ a month never once had rate limits for the things i care about. 20$ a month plus extra with the horrendously low rate limits on a model that is not as capable overall as gpt is not gonna keep me around. Id even go and say that if claude came out with a higher tier plan like 30 a month with an actually competitive rate limit id go for that. Gemini free has a higher rate limit, mistral has higher limits, and the direct competitor has vastly larger rates. Im sorry for the rant but after how much i was enjoying switching a few days ago, a clanker trying to nickel and dime me like the government while not being able to fully replace what i left, pisses me off. I tolerate the machines, ill gladly go back to not having them, i can write code myself it just takes me longer. With people leaving open AI in Exodus because of scam Altman and their unethical business practices, either the other companies make themselves more competitive to the influx of users or the users are just going to stop using the AI. We all know it's a bubble but if they don't want it to actually be a bubble and if they want to live through the bubble they need to make it more enticing. I don't know a single civilian who is willing to pay that exorbitant amount of money for a model that isn't even as feature Rich as the next flagship most people would rather just use the immoral one submitted by /u/ZookeepergameSalty10 [link] [comments]

AI Revolution in Healthcare: ChatGPT & Google Bard’s Breakthroughs – Diagnosis, mRNA Tech, Cancer Detection & More

AI Revolution in Healthcare: ChatGPT & Google Bard's Breakthroughs - Diagnosis, mRNA Tech, Cancer Detection & More
DjamgaMind - AI Unraveled Podcast

DjamgaMind: Audio Intelligence for the C-Suite (Energy, Healthcare, Finance)

Are you drowning in dense legal text? DjamgaMind is the new audio intelligence platform that turns 100-page healthcare or Energy mandates into 5-minute executive briefings. Whether you are navigating Bill C-27 (Canada) or the CMS-0057-F Interoperability Rule (USA), our AI agents decode the liability so you don’t have to. 👉 Start your specialized audio briefing today at Djamgamind.com


AI Jobs and Career

I wanted to share an exciting opportunity for those of you looking to advance your careers in the AI space. You know how rapidly the landscape is evolving, and finding the right fit can be a challenge. That's why I'm excited about Mercor – they're a platform specifically designed to connect top-tier AI talent with leading companies. Whether you're a data scientist, machine learning engineer, or something else entirely, Mercor can help you find your next big role. If you're ready to take the next step in your AI career, check them out through my referral link: https://work.mercor.com/?referralCode=82d5f4e3-e1a3-4064-963f-c197bb2c8db1. It's a fantastic resource, and I encourage you to explore the opportunities they have available.

Job TitleStatusPay
Full-Stack Engineer Strong match, Full-time $150K - $220K / year
Developer Experience and Productivity Engineer Pre-qualified, Full-time $160K - $300K / year
Software Engineer - Tooling & AI Workflows (Contract) Contract $90 / hour
DevOps Engineer (India) Full-time $20K - $50K / year
Senior Full-Stack Engineer Full-time $2.8K - $4K / week
Enterprise IT & Cloud Domain Expert - India Contract $20 - $30 / hour
Senior Software Engineer Contract $100 - $200 / hour
Senior Software Engineer Pre-qualified, Full-time $150K - $300K / year
Senior Full-Stack Engineer: Latin America Full-time $1.6K - $2.1K / week
Software Engineering Expert Contract $50 - $150 / hour
Generalist Video Annotators Contract $45 / hour
Generalist Writing Expert Contract $45 / hour
Editors, Fact Checkers, & Data Quality Reviewers Contract $50 - $60 / hour
Multilingual Expert Contract $54 / hour
Mathematics Expert (PhD) Contract $60 - $80 / hour
Software Engineer - India Contract $20 - $45 / hour
Physics Expert (PhD) Contract $60 - $80 / hour
Finance Expert Contract $150 / hour
Designers Contract $50 - $70 / hour
Chemistry Expert (PhD) Contract $60 - $80 / hour

AI Revolution in Healthcare: ChatGPT & Google Bard’s Breakthroughs – Diagnosis, mRNA Tech, Cancer Detection & More.

AI Revolution in Healthcare: Intro

Dive into the latest AI breakthroughs transforming healthcare since ChatGPT and Google Bard’s inception. Discover GPT-4’s rapid diagnostics, Moderna & IBM’s mRNA tech advancements, cutting-edge cancer detection methods, and more. Stay ahead in AI healthcare news with our comprehensive coverage on AI-powered drug discovery, early Alzheimer’s detection, and groundbreaking AI tools in medicine. Join us as we explore each major AI development that’s reshaping healthcare.

AI Revolution in Healthcare: Topics

🔍 GPT-4 diagnosed a 1 in 100,000 condition in seconds
💡 Moderna, IBM partner to advance mRNA technology using GenAI
🩺 AI model detects cancer, outperforms traditional methods
🧠 AI can detect Alzheimer’s signs even before they begin to show
⚙️ Google Cloud launches AI tools for drug discovery & precision medicine
🌟 BiomedGPT: The most sophisticated AI medical model?
⚔️ Google & Microsoft battle to lead healthcare AI
📈 MedPerf makes AI better for healthcare
🔬 Google DeepMind advances biomedical AI with ‘Med-PaLM M’
👀 Scientists train a neural network to identify PC users’ fatigue
🌐 Microsoft & Paige to build largest image-based model to fight cancer
🧬 DeepMind’s new AI can predict genetic diseases
🚀 Google Cloud launches new generative AI capabilities for healthcare
🦠 New AI tool can predict viral variants before they emerge
💬 ChatGPT outperforms doctors in depression treatment
🧪 AI algorithms are powering the search for cells
🏥 Google releases MedLM, generative AI fine-tuned healthcare
🤖 Google’s new medical AI, AMIE, beats doctors

Subscribe for weekly updates and deep dives into artificial intelligence innovations.

✅ Don’t forget to Like, Comment, and Share this video to support our content.

📌 Check out our playlist for more AI insights

AI-Powered Professional Certification Quiz Platform
Crack Your Next Exam with Djamgatech AI Cert Master

Web|iOs|Android|Windows

Are you passionate about AI and looking for your next career challenge? In the fast-evolving world of artificial intelligence, connecting with the right opportunities can make all the difference. We're excited to recommend Mercor, a premier platform dedicated to bridging the gap between exceptional AI professionals and innovative companies.

Whether you're seeking roles in machine learning, data science, or other cutting-edge AI fields, Mercor offers a streamlined path to your ideal position. Explore the possibilities and accelerate your AI career by visiting Mercor through our exclusive referral link:

Find Your AI Dream Job on Mercor

Your next big opportunity in AI could be just a click away!

📢 Advertise with us and Sponsorship Opportunities

Are you eager to expand your understanding of artificial intelligence? Look no further than the essential book “AI Unraveled: Master GPT-4, Gemini, Generative AI & LLMs – Simplified Guide for Everyday Users: Demystifying Artificial Intelligence – OpenAI, ChatGPT, Google Bard, AI ML Quiz, AI Certifications Prep, Prompt Engineering,” available at Etsy, Shopify, Apple, Google, or Amazon

AI Revolution in Healthcare: Podcast Transcript

Welcome to “AI Unraveled: Demystifying Frequently Asked Questions on Artificial Intelligence, Latest AI Trends,” where we dive deep into the complexities of AI and bring forth the latest developments in an easy-to-understand format. Today, we’re tackling a series of compelling updates from the AI frontier in the medical field and beyond. In a remarkable medical application, GPT-4, OpenAI’s newest language model, has been put to the test by Dr. Isaac Kohane of Harvard. Impressively, GPT-4 has been reported to perform better than many human doctors, correctly answering medical exam questions over 90% of the time. But what’s truly astonishing is its ability to diagnose a rare 1 in 100,000 condition in just seconds, a task that draws upon the depth of a seasoned physician’s experience. Despite these advances, Dr. Kohane’s book, ‘The AI Revolution in Medicine,’ brings us back to earth, reminding us that GPT-4 is not infallible, presenting a balanced view with examples of the model’s errors ranging from minor clerical issues to math mistakes.

hifting gears, we look at how pharmaceutical giant Moderna and tech behemoth IBM are joining forces to push the boundaries of mRNA technology. Their collaboration intends to combine generative AI and quantum computing, potentially accelerating the discovery of new therapies and vaccines. This is underpinned by using IBM’s MoLFormer, which is expected to enhance Moderna’s understanding of mRNA medicines. In a leap toward precision medicine, Google Cloud has recently launched two AI-powered tools geared at revolutionizing drug discovery. These innovative tools focus on predicting protein structures and managing vast amounts of genomic data, potentially shaving off years in drug development time. We also witness the rise of BiomedGPT, touted as one of the most sophisticated AI medical models, outperforming predecessors across multiple biomedical modalities. This model appears to be a game-changer with its multi-modal and multi-task learning capabilities.

AI Jobs and Career

And before we wrap up today's AI news, I wanted to share an exciting opportunity for those of you looking to advance your careers in the AI space. You know how rapidly the landscape is evolving, and finding the right fit can be a challenge. That's why I'm excited about Mercor – they're a platform specifically designed to connect top-tier AI talent with leading companies. Whether you're a data scientist, machine learning engineer, or something else entirely, Mercor can help you find your next big role. If you're ready to take the next step in your AI career, check them out through my referral link: https://work.mercor.com/?referralCode=82d5f4e3-e1a3-4064-963f-c197bb2c8db1. It's a fantastic resource, and I encourage you to explore the opportunities they have available.

The competition intensifies in the healthcare AI space with Google’s Med-PaLM 2 going through testing at the Mayo Clinic, while Microsoft swiftly incorporates AI advances into patient care by deploying GPT algorithms via cloud services. Furthermore, MedPerf emerges as a new beacon, an open benchmarking platform introduced by MLCommons, aimed to evaluate medical AI models on diverse datasets, prioritizing patient privacy and aiming to enhance AI’s generalizability in healthcare. Adding to an already impressive array of advancements, we have AlphaMissense by Google DeepMind, which is honing the ability to predict genetic diseases, and Google Cloud briefing the healthcare sector with new capabilities to sift through clinical data more efficiently. And finally, EVEscape, a new AI tool with the potential to predict future viral variants—imagine its profound implications had it been available at the onset of the COVID-19 pandemic!

To cap off, studies suggest that AI models like ChatGPT can outdo doctors in providing unbiased treatment recommendations for depression and that AI algorithms are increasingly crucial in cellular research, changing the landscape of biological imaging experiments. Before we conclude, let’s not forget about AMIE, Google’s Articulate Medical Intelligence Explorer, an AI system optimized for diagnostic reasoning that is giving medical professionals a run for their money. For those seeking a deeper understanding of these advancements, the book “AI Unraveled: Master GPT-4, Gemini, Generative AI & LLMs – Simplified Guide for Everyday Users: Demystifying Artificial Intelligence – OpenAI, ChatGPT, Google Bard, AI ML Quiz, AI Certifications Prep, Prompt Engineering,” is available on various platforms including Etsy, Shopify, Apple, Google, and Amazon. That brings us to the end of today’s episode. We hope you’ve gained new insights into the dynamic and revolutionary world of AI, especially its influence on healthcare. Join us next time on “AI Unraveled” as we continue to explore cutting-edge AI trends that are transforming our lives. Till then, this is your host signing off. Keep questioning, keep learning, and remember—the future is AI.

GPT-4 diagnosed a 1 in 100,000 condition in seconds

  

Dr. Isaac Kohane, a physician and computer scientist at Harvard, has tested the newest AI model, GPT-4, in a medical setting. According to his findings, GPT-4 performs better than many doctors, as it can answer medical exam licensing questions correctly more than 90% of the time, translate information for patients, and give doctors helpful suggestions about bedside manner.

Kohane tested GPT-4 on a real-life case and found that it could correctly diagnose a rare condition just as he would with all his years of experience. However, GPT-4 isn’t always reliable, and his latest book ‘The AI Revolution in Medicine’ is filled with examples of its blunders, ranging from clerical errors to math mistakes.

Read the whole article here


Moderna, IBM to explore Generative AI and quantum computing for mRNA vaccines

Moderna and IBM are partnering to advance mRNA technology using generative AI and quantum computing, which could speed up Moderna’s discovery and creation of new messenger RNA vaccines and therapies. Moderna’s scientists will have access to IBM’s generative AI model known as MoLFormer, which will help understand the characteristics of potential mRNA medicines and design a new class of vaccines and therapies.

This agreement comes as Moderna is trying to harness its mRNA technology to target other diseases, while IBM is ramping up its investment in AI with new partnerships, largely driven by the release of OpenAI’s ChatGPT.


AI Unraveled: Demystifying Frequently Asked Questions on Artificial Intelligence (OpenAI, ChatGPT, Google Gemini, Generative AI, Discriminative AI, xAI, LLMs, GPUs, Machine Learning, NLP, Promp Engineering)

Why does this matter?

The use of quantum computing and AI could help Moderna accelerate the discovery and creation of these new vaccines and therapies by solving problems too complex for traditional computers. The development of these new medicines could potentially benefit the general public by providing more treatment options for a range of diseases.

Source


AI model outperforms traditional methods in identifying cancerous nodules

An AI model developed by experts at the Royal Marsden NHS foundation trust, the Institute of Cancer Research, London, and Imperial College London can accurately identify cancer, potentially speeding up diagnosis and treatment. The algorithm, which analyzes CT scans to determine if abnormal growths are cancerous, reportedly performs more efficiently and effectively than current methods.

Why does this matter?

The AI tool may help doctors make faster decisions about patients with abnormal growths that are currently deemed medium-risk. The model, which is still in its early stages, will require further testing before it can be introduced in healthcare systems. However, researchers hope the AI tool will eventually speed up cancer detection by fast-tracking patients to treatment.

Source


AI can detect signs of Alzheimer’s even before symptoms begin to show

Researchers at UT Southwestern Medical Center have found that AI-powered voice analysis can help diagnose Alzheimer’s and cognitive impairment in early stages. If confirmed by larger studies, these findings could primary care providers with an easy-to-perform screening tool for at-risk individuals.

The research used advanced ML and natural language processing (NLP) to identify even the subtlest changes in language and audio that individuals may not easily recognize.

Why does this matter?

Before ML and NLP, detailed speech studies were often unsuccessful as early changes were often undetectable to human ears. However, with advancements in AI, such novel testing methods have performed significantly better than standard cognitive assessments in detecting even mild impairments. Also, it took less than 10 minutes to capture a patient’s voice, outdoing the traditional tests, which took hours to administer.

Only a few days ago, researchers developed an AI model that outperformed traditional methods in identifying cancer. Does this indicate AI leading the charge in reducing overall healthcare costs with improved patient outcomes?

Source


Google Cloud launches AI tools for drug discovery and precision medicine

Google Cloud has launched two AI-powered tools to help biotech and pharmaceutical companies accelerate drug discovery and advance precision medicine. The Target and Lead Identification Suite aims to streamline the process of identifying a biological target and predicting protein structures, while the Multiomics Suite assists researchers in ingesting, storing, analyzing, and sharing large amounts of genomic data. Both tools aim to significantly reduce the time and cost associated with drug development.

Several companies, including Pfizer, Cerevel Therapeutics, and Colossal Biosciences, have already been using these products. Cerevel Therapeutics estimates that it will save at least three years on average by using the Target and Lead Identification Suite to discover new drugs.

Why does this matter?

AI seems to benefit humanity the most through its use in medicine and diagnostics. This launch from Google and the subsequent adoption by a pharma giant like Pfizer indicate the swift mainstreaming of the tech.

Source


BiomedGPT: The most sophisticated AI medical model?

BiomedGPT is a unified and generalist Biomedical Generative Pre-trained Transformer model. BiomedGPT utilizes self-supervision on diverse datasets to handle multi-modal inputs and perform various downstream tasks.

  

Extensive experiments show that BiomedGPT surpasses most previous state-of-the-art models in performance across 5 distinct tasks with 20 public datasets spanning over 15 biomedical modalities.

The study also demonstrates the effectiveness of the multi-modal and multi-task pretraining approach in transferring knowledge to previously unseen data.

Why does this matter?

This research represents a significant advancement in developing unified and generalist models for biomedicine, holding promising implications for enhancing healthcare outcomes, and it could lead to discoveries in biomedical research.

In addition to its potential benefits for healthcare, BiomedGPT could also be used in drug discovery & medical education.

Source


Google & Microsoft battle to lead healthcare AI

Reportedly, Google’s Med-PaLM 2 (an LLM for the medical domain) has been in testing at the Mayo Clinic research hospital. In April, Google announced its limited access for select Google Cloud customers to explore use cases and share feedback to investigate safe, responsible, and meaningful ways to use it.

Meanwhile, Google’s rivals moved quickly to incorporate AI advances into patient interactions. Hospitals are beginning to test OpenAI’s GPT algorithms through Microsoft’s cloud service in several tasks. Google’s Med-PaLM 2 and OpenAI’s GPT-4 each scored similarly on medical exam questions, according to independent research released by the companies.

Why does this matter?

It seems Google and Microsoft are racing to translate recent AI advances into products that clinicians would use widely. The AI field has seen rapid advancements and research in diverse domains. But such a competitive landscape accelerates translating them into widely available, impactful AI products (which is sometimes slow and challenging due to the complexity of real-world applications).

(Source)


MedPerf makes AI better for healthcare

MLCommons, an open global engineering consortium, has announced the launch of MedPerf, an open benchmarking platform for evaluating the performance of medical AI models on diverse real-world datasets. The platform aims to improve medical AI’s generalizability and clinical impact by making data easily and safely accessible to researchers while prioritizing patient privacy and mitigating legal and regulatory risks. 

  

MedPerf utilizes federated evaluation, allowing AI models to be assessed without accessing patient data, and offers orchestration capabilities to streamline research. The platform has already been successfully used in pilot studies and challenges involving brain tumor segmentation, pancreas segmentation, and surgical workflow phase recognition.

Ace the Microsoft Azure Fundamentals AZ-900 Certification Exam: Pass the Azure Fundamentals Exam with Ease

Why does this matter?

With MedPerf, researchers can evaluate the performance of medical AI models using diverse real-world datasets without compromising patient privacy. This platform’s implementation in pilot studies and challenges for various medical tasks further demonstrates its potential to improve medical AI’s generalizability, clinical impact, and advancements in healthcare technology.

Source


Google DeepMind advances biomedical AI with ‘Med-PaLM M’

Google and DeepMind have introduced Med-PaLM M, a multimodal biomedical AI system that can interpret diverse types of medical data, including text, images, and genomics. The researchers curated a benchmark dataset called MultiMedBench, which covers 14 biomedical tasks, to train and evaluate Med-PaLM M. 

  

The AI system achieved state-of-the-art performance across all tasks, surpassing specialized models optimized for individual tasks. Med-PaLM M represents a paradigm shift in biomedical AI, as it can incorporate multimodal patient information, improve diagnostic accuracy, and transfer knowledge across medical tasks. Preliminary evidence suggests that Med-PaLM M can generalize to novel tasks and concepts and perform zero-shot multimodal reasoning.

Why does this matter?

It brings us closer to creating advanced AI systems to understand and analyze various medical data types. Google DeepMind’s MultiMedBench and Med-PaLM M show promising performance and potential in healthcare applications. It means better healthcare tools that can handle different types of medical information, ultimately benefiting patients and healthcare providers.

Source


Scientists train a neural network to identify PC users’ fatigue

Scientists from St. Petersburg University and other organizations have created a database of eye movement strategies of PC users in different states of fatigue. They plan to use this data to train neural network models that can accurately track the functional state of operators, ensuring safety in various industries. The database includes a comprehensive set of indicators collected through sensors such as video cameras, eye trackers, heart rate monitors, and electroencephalographs.

  

An example of human fatigue analysis using video recording.

Why does this matter?

If you are looking for an all-in-one solution to help you prepare for the AWS Cloud Practitioner Certification Exam, look no further than this AWS Cloud Practitioner CCP CLF-C02 book

The scientists believe that this approach will allow for remote assessment of fatigue severity, and the database will be accessible to software developers for testing their products.

Source


Microsoft and Paige to build the largest image-based AI model to fight cancer

Paige, a technology disruptor in healthcare, has joined forces with Microsoft to build the world’s largest image-based AI models for digital pathology and oncology.

Paige developed the first Large Foundation Model using over one billion images from half a million pathology slides across multiple cancer types. Now, it is developing a new AI model with Microsoft that is orders-of-magnitude larger than any other image-based AI model existing today, configured with billions of parameters.

Paige will utilize Microsoft’s advanced supercomputing infrastructure to train the technology at scale and ultimately deploy it to hospitals and laboratories across the globe using Azure.

Why does this matter?

This will help realize the potential of generative AI at an unprecedented scale, introduce completely novel capabilities of AI, and serve as the cornerstone for the next generation of clinical/healthcare applications built with AI.

Source


DeepMind’s new AI can predict genetic diseases

Google DeepMind’s new system, called AlphaMissense, can tell if the letters in the DNA will produce the correct shape. If not, it is listed as potentially disease-causing.

  

Currently, genetic disease hunters have fairly limited knowledge of which areas of human DNA can lead to disease and have to search across billions of chemical building blocks that make up DNA. They have classified 0.1% of letter changes, or mutations, as either benign or disease-causing. DeepMind’s new model pushed that percentage up to 89%.

Why does this matter?

AI is changing nearly everything we do at the moment and might revolutionize molecular biology and life sciences, too. This development is expected to speed up diagnosis and help search for better genetic disease treatments.

Source


Google Cloud launches new generative AI capabilities for healthcare

Google Cloud introduced new Vertex AI Search features for healthcare and life science companies. It will allow users to find accurate clinical information much more efficiently and to search a broad spectrum of data from clinical sources, such as FHIR data, clinical notes, and medical data in electronic health records (EHRs). Life-science organizations can use these features to enhance scientific communications and streamline processes.

Why does this matter?

Given how siloed medical data is currently, this is a significant boon to healthcare organizations. With this, Google is also enabling them to leverage the power of AI to improve healthcare facility management, patient care delivery, and more.

Source


New AI tool can predict viral variants before they emerge

A new AI tool named EVEscape, developed by researchers at Harvard Medical School and the University of Oxford, can make predictions about new viral variants before they actually emerge and also how they would evolve.

In the study, researchers show that had it been deployed at the start of the COVID-19 pandemic, EVEscape would have predicted the most frequent mutations and identified the most concerning variants for SARS-CoV-2. The tool also made accurate predictions about other viruses, including HIV and influenza. 

Why does this matter?

The information from this AI tool will help scientists develop more effective, future-proof vaccines and therapies. If only this AI boom happened a little earlier, it could have prevented the Covid-19 pandemic. But I guess no more pandemics, thanks to AI?

Source


ChatGPT outperforms doctors in depression treatment

According to new study, ChatGPT makes unbiased, evidence-based treatment recommendations for depression that are consistent with clinical guidelines and outperform human primary care physicians. The study compared the evaluations and treatment recommendations for depression generated by ChatGPT-3 and ChatGPT-4 with those of primary care physicians. 

Vignettes describing patients with different attributes and depression severity were input into the chatbot interfaces.

  
  

Why does this matter?

Compared with primary care physicians, ChatGPT showed no bias in recommendations based on patient gender or socioeconomic status. This means the chatbot was aligned well with accepted guidelines for managing mild and severe depression.

Source


AI algorithms are powering the search for cells

A new paper by Nature details how AI-powered image analysis tools are changing the game for microscopy data. It highlights the evolution from early, labor-intensive methods to machine learning-based tools like CellProfiler, ilastik, and newer frameworks such as U-Net. These advancements enable more accurate and faster segmentation of cells, essential for various biological imaging experiments.

  

Cancer-cell nuclei (green boxes) picked out by software using deep learning.

Why does this matter?

The short study highlights the potential for AI-driven tools to revolutionize further biological analyses. The advancement is crucial for understanding diseases, drug development, and gaining insights into cellular behavior, enabling faster scientific discoveries in various fields like medicine and biology.

Source


Google releases MedLM: Generative AI fine-tuned healthcare

MedLM is a family of foundation models fine-tuned for the healthcare industry, generally available (via allowlist) to Google Cloud customers in the U.S. through Vertex AI. MedLM builds on Med-PaLM 2. Google will soon add Gemini-based models into the MedLM suite to offer even more capabilities.

Why does this matter?

Google isn’t done yet. While its impressive Gemini demo from last week may have been staged, Google is looking to fine-tune and improve Gemini based on developers’ feedback. In addition, it is also racing with rivals to push the boundaries of AI in various fields.

Source


Google’s new medical AI, AMIE, beats doctors

Google developed Articulate Medical Intelligence Explorer (AMIE), an LLM-based research AI system optimized for diagnostic reasoning and conversations.

AMIE’s performance was compared to that of primary care physicians (PCPs) in a randomized, double-blind crossover study of text-based consultations with validated patient actors in the style of an Objective Structured Clinical Examination (OSCE). AMIE demonstrated greater diagnostic accuracy and superior performance on 28 of 32 axes according to specialist physicians and 24 of 26 axes according to patient actors.

Why does this matter?

While further research is required before AMIE can be translated to real-world settings, it represents a milestone towards conversational diagnostic AI. If successful, AI systems such as AMIE can be at the core of next-generation learning health systems that help scale world-class healthcare to everyone.

Source

 

A Daily Chronicle of AI Innovations in January 2024

Mastering GPT-4: Simplified Guide for Everyday Users

Mastering GPT-4: Simplified Guide for Everyday Users
DjamgaMind - AI Unraveled Podcast

DjamgaMind: Audio Intelligence for the C-Suite (Energy, Healthcare, Finance)

Are you drowning in dense legal text? DjamgaMind is the new audio intelligence platform that turns 100-page healthcare or Energy mandates into 5-minute executive briefings. Whether you are navigating Bill C-27 (Canada) or the CMS-0057-F Interoperability Rule (USA), our AI agents decode the liability so you don’t have to. 👉 Start your specialized audio briefing today at Djamgamind.com


AI Jobs and Career

I wanted to share an exciting opportunity for those of you looking to advance your careers in the AI space. You know how rapidly the landscape is evolving, and finding the right fit can be a challenge. That's why I'm excited about Mercor – they're a platform specifically designed to connect top-tier AI talent with leading companies. Whether you're a data scientist, machine learning engineer, or something else entirely, Mercor can help you find your next big role. If you're ready to take the next step in your AI career, check them out through my referral link: https://work.mercor.com/?referralCode=82d5f4e3-e1a3-4064-963f-c197bb2c8db1. It's a fantastic resource, and I encourage you to explore the opportunities they have available.

Job TitleStatusPay
Full-Stack Engineer Strong match, Full-time $150K - $220K / year
Developer Experience and Productivity Engineer Pre-qualified, Full-time $160K - $300K / year
Software Engineer - Tooling & AI Workflows (Contract) Contract $90 / hour
DevOps Engineer (India) Full-time $20K - $50K / year
Senior Full-Stack Engineer Full-time $2.8K - $4K / week
Enterprise IT & Cloud Domain Expert - India Contract $20 - $30 / hour
Senior Software Engineer Contract $100 - $200 / hour
Senior Software Engineer Pre-qualified, Full-time $150K - $300K / year
Senior Full-Stack Engineer: Latin America Full-time $1.6K - $2.1K / week
Software Engineering Expert Contract $50 - $150 / hour
Generalist Video Annotators Contract $45 / hour
Generalist Writing Expert Contract $45 / hour
Editors, Fact Checkers, & Data Quality Reviewers Contract $50 - $60 / hour
Multilingual Expert Contract $54 / hour
Mathematics Expert (PhD) Contract $60 - $80 / hour
Software Engineer - India Contract $20 - $45 / hour
Physics Expert (PhD) Contract $60 - $80 / hour
Finance Expert Contract $150 / hour
Designers Contract $50 - $70 / hour
Chemistry Expert (PhD) Contract $60 - $80 / hour

Mastering GPT-4: Simplified Guide for Everyday Users or How to make GPT-4 your b*tch!

Listen Here

Recently, while updating our OpenAI Python library, I encountered a marketing intern struggling with GPT-4. He was overwhelmed by its repetitive responses, lengthy answers, and not quite getting what he needed from it. Realizing the need for a simple, user-friendly explanation of GPT-4’s functionalities, I decided to create this guide. Whether you’re new to AI or looking to refine your GPT-4 interactions, these tips are designed to help you navigate and optimize your experience.

Embark on a journey to master GPT-4 with our easy-to-understand guide, ‘Mastering GPT-4: Simplified Guide for Everyday Users‘.

🌟🤖 This blog/video/podcast is perfect for both AI newbies and those looking to enhance their experience with GPT-4. We break down the complexities of GPT-4’s settings into simple, practical terms, so you can use this powerful tool more effectively and creatively.

🔍 What You’ll Learn:

  1. Frequency Penalty: Discover how to reduce repetitive responses and make your AI interactions sound more natural.
  2. Logit Bias: Learn to gently steer the AI towards or away from specific words or topics.
  3. Presence Penalty: Find out how to encourage the AI to transition smoothly between topics.
  4. Temperature: Adjust the AI’s creativity level, from straightforward responses to imaginative ideas.
  5. Top_p (Nucleus Sampling): Control the uniqueness of the AI’s suggestions, from conventional to out-of-the-box ideas.
Mastering GPT-4: Simplified Guide for Everyday Users
Mastering GPT-4: Simplified Guide for Everyday Users

1. Frequency Penalty: The Echo Reducer

  • What It Does: This setting helps minimize repetition in the AI’s responses, ensuring it doesn’t sound like it’s stuck on repeat.
  • Examples:
    • Low Setting: You might get repeated phrases like “I love pizza. Pizza is great. Did I mention pizza?”
    • High Setting: The AI diversifies its language, saying something like “I love pizza for its gooey cheese, tangy sauce, and crispy crust. It’s a culinary delight.”

2. Logit Bias: The Preference Tuner

  • What It Does: It nudges the AI towards or away from certain words, almost like gently guiding its choices.
  • Examples:
    • Against ‘pizza’: The AI might focus on other aspects, “I enjoy Italian food, especially pasta and gelato.”
    • Towards ‘pizza’: It emphasizes the chosen word, “Italian cuisine brings to mind the delectable pizza, a feast of flavors in every slice.”

3. Presence Penalty: The Topic Shifter

  • What It Does: This encourages the AI to change subjects more smoothly, avoiding dwelling too long on a single topic.
  • Examples:
    • Low Setting: It might stick to one idea, “I enjoy sunny days. Sunny days are pleasant.”
    • High Setting: The AI transitions to new ideas, “Sunny days are wonderful, but I also appreciate the serenity of rainy evenings and the beauty of a snowy landscape.”

4. Temperature: The Creativity Dial

  • What It Does: Adjusts how predictable or creative the AI’s responses are.
  • Examples:
    • Low Temperature: Expect straightforward answers like, “Cats are popular pets known for their independence.”
    • High Temperature: It might say something whimsical, “Cats, those mysterious creatures, may just be plotting a cute but world-dominating scheme.”

5. Top_p (Nucleus Sampling): The Imagination Spectrum

  • What It Does: Controls how unique or unconventional the AI’s suggestions are.
  • Examples:
    • Low Setting: You’ll get conventional ideas, “Vacations are perfect for unwinding and relaxation.”
    • High Setting: Expect creative and unique suggestions, “Vacation ideas range from bungee jumping in New Zealand to attending a silent meditation retreat in the Himalayas.”

Mastering GPT-4: Understanding Temperature in GPT-4; A Guide to AI Probability and Creativity

If you’re intrigued by how the ‘temperature’ setting impacts the output of GPT-4 (and other Large Language Models or LLMs), here’s a straightforward explanation:

AI-Powered Professional Certification Quiz Platform
Crack Your Next Exam with Djamgatech AI Cert Master

Web|iOs|Android|Windows

Are you passionate about AI and looking for your next career challenge? In the fast-evolving world of artificial intelligence, connecting with the right opportunities can make all the difference. We're excited to recommend Mercor, a premier platform dedicated to bridging the gap between exceptional AI professionals and innovative companies.

Whether you're seeking roles in machine learning, data science, or other cutting-edge AI fields, Mercor offers a streamlined path to your ideal position. Explore the possibilities and accelerate your AI career by visiting Mercor through our exclusive referral link:

Find Your AI Dream Job on Mercor

Your next big opportunity in AI could be just a click away!

LLMs, like GPT-4, don’t just spit out a single next token; they actually calculate probabilities for every possible token in their vocabulary. For instance, if the model is continuing the sentence “The cat in the,” it might assign probabilities like: Hat: 80%, House: 5%, Basket: 4%, and so on, down to the least likely words. These probabilities cover all possible tokens, adding up to 100%.

What happens next is crucial: one of these tokens is selected based on their probabilities. So, ‘hat’ would be chosen 80% of the time. This approach introduces a level of randomness in the model’s output, making it less deterministic.

Now, the ‘temperature’ parameter plays a role in how these probabilities are adjusted or skewed before a token is selected. Here’s how it works:

  • Temperature = 1: This keeps the original probabilities intact. The output remains somewhat random but not skewed.
  • Temperature < 1: This skews probabilities toward more likely tokens, making the output more predictable. For example, ‘hat’ might jump to a 95% chance.
  • Temperature = 0: This leads to complete determinism. The most likely token (‘hat’, in our case) gets a 100% probability, eliminating randomness.
  • Temperature > 1: This setting spreads out the probabilities, making less likely words more probable. It increases the chance of producing varied and less predictable outputs.

A very high temperature setting can make unlikely and nonsensical words more probable, potentially resulting in outputs that are creative but might not make much sense.

AI Jobs and Career

And before we wrap up today's AI news, I wanted to share an exciting opportunity for those of you looking to advance your careers in the AI space. You know how rapidly the landscape is evolving, and finding the right fit can be a challenge. That's why I'm excited about Mercor – they're a platform specifically designed to connect top-tier AI talent with leading companies. Whether you're a data scientist, machine learning engineer, or something else entirely, Mercor can help you find your next big role. If you're ready to take the next step in your AI career, check them out through my referral link: https://work.mercor.com/?referralCode=82d5f4e3-e1a3-4064-963f-c197bb2c8db1. It's a fantastic resource, and I encourage you to explore the opportunities they have available.

Temperature isn’t just about creativity; it’s about allowing the LLM to explore less common paths from its training data. When used judiciously, it can lead to more diverse responses. The ideal temperature setting depends on your specific needs:

  • For precision and reliability (like in coding or when strict adherence to a format is required), a lower temperature (even zero) is preferable.
  • For creative tasks like writing, brainstorming, or naming, where there’s no single ‘correct’ answer, a higher temperature can yield more innovative and varied results.

So, by adjusting the temperature, you can fine-tune GPT-4’s outputs to be as predictable or as creative as your task requires.

Mastering GPT-4: Conclusion

With these settings, you can tailor GPT-4 to better suit your needs, whether you’re looking for straightforward information or creative and diverse insights. Remember, experimenting with these settings will help you find the perfect balance for your specific use case. Happy exploring with GPT-4!

Mastering GPT-4 Annex: More about GPT-4 API Settings

I think certain parameters in the API are more useful than others. Personally, I haven’t come across a use case for frequency_penalty or presence_penalty.

However, for example, logit_bias could be quite useful if you want the LLM to behave as a classifier (output only either “yes” or “no”, or some similar situation).

Basically logit_bias tells the LLM to prefer or avoid certain tokens by adding a constant number (bias) to the likelihood of each token. LLMs output a number (referred to as a logit) for each token in their dictionary, and by increasing or decreasing the logit value of a token, you make that token more or less likely to be part of the output. Setting the logit_bias of a token to +100 would mean it will output that token effectively 100% of the time, and -100 would mean the token is effectively never output. You may think, why would I want a token(s) to be output 100% of the time? You can for example set multiple tokens to +100, and it will choose between only those tokens when generating the output.

One very useful usecase would be to combine the temperature, logit_bias, and max_tokens parameters.


AI Unraveled: Demystifying Frequently Asked Questions on Artificial Intelligence (OpenAI, ChatGPT, Google Gemini, Generative AI, Discriminative AI, xAI, LLMs, GPUs, Machine Learning, NLP, Promp Engineering)

You could set:

`temperature` to zero (which would force the LLM to select the top-1 most likely token/with the highest logit value 100% of the time, since by default there’s a bit of randomness added)

`logit_bias` to +100 (the maximum value permitted) for both the tokens “yes” and “no”

`max_tokens` value to one

Since the LLM typically never outputs logits of >100 naturally, you are basically ensuring that the output of the LLM is ALWAYS either the token “yes” or the token “no”. And it will still pick the correct one of the two since you’re adding the same number to both, and one will still have the higher logit value than the other.

This is very useful if you need the output of the LLM to be a classifier, e.g. “is this text about cats” -> yes/no, without needing to fine tune the output of the LLM to “understand” that you only want a yes/no answer. You can force that behavior using postprocessing only. Of course, you can select any tokens, not just yes/no, to be the only possible tokens. Maybe you want the tokens “positive”, “negative” and “neutral” when classifying the sentiment of a text, etc.

What is the difference between frequence_penalty and presence_penalty?

frequency_penalty reduces the probability of a token appearing multiple times proportional to how many times it’s already appeared, while presence_penalty reduces the probability of a token appearing again based on whether it’s appeared at all.

From the API docs:

frequency_penalty Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far, decreasing the model’s likelihood to repeat the same line verbatim.

presence_penalty Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model’s likelihood to talk about new topics.

Mastering GPT-4 References:

https://platform.openai.com/docs/api-reference/chat/create#chat-create-logit_bias.

https://help.openai.com/en/articles/5247780-using-logit-bias-to-define-token-probability

📢 Advertise with us and Sponsorship Opportunities

Are you eager to expand your understanding of artificial intelligence? Look no further than the essential book “AI Unraveled: Demystifying Frequently Asked Questions on Artificial Intelligence,” available at Etsy, Shopify, Apple, Google, or Amazon

Decoding GPTs & LLMs: Training, Memory & Advanced Architectures Explained

Mastering GPT-4 Transcript

Welcome to AI Unraveled, the podcast that demystifies frequently asked questions on artificial intelligence and keeps you up to date with the latest AI trends. Join us as we delve into groundbreaking research, innovative applications, and emerging technologies that are pushing the boundaries of AI. From the latest trends in ChatGPT and the recent merger of Google Brain and DeepMind, to the exciting developments in generative AI, we’ve got you covered with a comprehensive update on the ever-evolving AI landscape. In today’s episode, we’ll cover optimizing AI interactions with Master GPT-4, including reducing repetition, steering conversations, adjusting creativity, using the frequency penalty setting to diversify language, utilizing logit bias to guide word choices, implementing presence penalty for smoother transitions, adjusting temperature for different levels of creativity in responses, controlling uniqueness with Top_p (Nucleus Sampling), and an introduction to the book “AI Unraveled” which answers frequently asked questions about artificial intelligence.

Hey there! Have you ever heard of GPT-4? It’s an amazing tool developed by OpenAI that uses artificial intelligence to generate text. However, I’ve noticed that some people struggle with it. They find its responses repetitive, its answers too long, and they don’t always get what they’re looking for. That’s why I decided to create a simplified guide to help you master GPT-4.

Introducing “Unlocking GPT-4: A User-Friendly Guide to Optimizing AI Interactions“! This guide is perfect for both AI beginners and those who want to take their GPT-4 experience to the next level. We’ll break down all the complexities of GPT-4 into simple, practical terms, so you can use this powerful tool more effectively and creatively.

In this guide, you’ll learn some key concepts that will improve your interactions with GPT-4. First up, we’ll explore the Frequency Penalty. This technique will help you reduce repetitive responses and make your AI conversations sound more natural. Then, we’ll dive into Logit Bias. You’ll discover how to gently steer the AI towards or away from specific words or topics, giving you more control over the conversation.

Next, we’ll tackle the Presence Penalty. You’ll find out how to encourage the AI to transition smoothly between topics, allowing for more coherent and engaging discussions. And let’s not forget about Temperature! This feature lets you adjust the AI’s creativity level, so you can go from straightforward responses to more imaginative ideas.

Last but not least, we have Top_p, also known as Nucleus Sampling. With this technique, you can control the uniqueness of the AI’s suggestions. You can stick to conventional ideas or venture into out-of-the-box thinking.

So, if you’re ready to become a GPT-4 master, join us on this exciting journey by checking out our guide. Happy optimizing!

Today, I want to talk about a really cool feature in AI called the Frequency Penalty, also known as the Echo Reducer. Its main purpose is to prevent repetitive responses from the AI, so it doesn’t sound like a broken record.

Let me give you a couple of examples to make it crystal clear. If you set the Frequency Penalty to a low setting, you might experience repeated phrases like, “I love pizza. Pizza is great. Did I mention pizza?” Now, I don’t know about you, but hearing the same thing over and over again can get a little tiresome.

But fear not! With a high setting on the Echo Reducer, the AI gets more creative with its language. Instead of the same old repetitive phrases, it starts diversifying its response. For instance, it might say something like, “I love pizza for its gooey cheese, tangy sauce, and crispy crust. It’s a culinary delight.” Now, isn’t that a refreshing change?

So, the Frequency Penalty setting is all about making sure the AI’s responses are varied and don’t become monotonous. It’s like giving the AI a little nudge to keep things interesting and keep the conversation flowing smoothly.

Today, I want to talk about a fascinating tool called the Logit Bias: The Preference Tuner. This tool has the power to nudge AI towards or away from certain words. It’s kind of like gently guiding the AI’s choices, steering it in a particular direction.

Let’s dive into some examples to understand how this works. Imagine we want to nudge the AI away from the word ‘pizza’. In this case, the AI might start focusing on other aspects, like saying, “I enjoy Italian food, especially pasta and gelato.” By de-emphasizing ‘pizza’, the AI’s choices will lean away from this particular word.

On the other hand, if we want to nudge the AI towards the word ‘pizza’, we can use the Logit Bias tool to emphasize it. The AI might then say something like, “Italian cuisine brings to mind the delectable pizza, a feast of flavors in every slice.” By amplifying ‘pizza’, the AI’s choices will emphasize this word more frequently.

The Logit Bias: The Preference Tuner is a remarkable tool that allows us to fine-tune the AI’s language generation by influencing its bias towards or away from specific words. It opens up exciting possibilities for tailoring the AI’s responses to better suit our needs and preferences.

The Presence Penalty, also known as the Topic Shifter, is a feature that helps the AI transition between subjects more smoothly. It prevents the AI from fixating on a single topic for too long, making the conversation more dynamic and engaging.

Let me give you some examples to illustrate how it works. On a low setting, the AI might stick to one idea, like saying, “I enjoy sunny days. Sunny days are pleasant.” In this case, the AI focuses on the same topic without much variation.

Ace the Microsoft Azure Fundamentals AZ-900 Certification Exam: Pass the Azure Fundamentals Exam with Ease

However, on a high setting, the AI becomes more versatile in shifting topics. For instance, it could say something like, “Sunny days are wonderful, but I also appreciate the serenity of rainy evenings and the beauty of a snowy landscape.” Here, the AI smoothly transitions from sunny days to rainy evenings and snowy landscapes, providing a diverse range of ideas.

By implementing the Presence Penalty, the AI is encouraged to explore different subjects, ensuring a more interesting and varied conversation. It avoids repetitive patterns and keeps the dialogue fresh and engaging.

So, whether you prefer the AI to stick with one subject or shift smoothly between topics, the Presence Penalty feature gives you control over the flow of conversation, making it more enjoyable and natural.

Today, let’s talk about temperature – not the kind you feel outside, but the kind that affects the creativity of AI responses. Imagine a dial that adjusts how predictable or creative those responses are. We call it the Creativity Dial.

When the dial is set to low temperature, you can expect straightforward answers from the AI. It would respond with something like, “Cats are popular pets known for their independence.” These answers are informative and to the point, just like a textbook.

On the other hand, when the dial is set to high temperature, get ready for some whimsical and imaginative responses. The AI might come up with something like, “Cats, those mysterious creatures, may just be plotting a cute but world-dominating scheme.” These responses can be surprising and even amusing.

So, whether you prefer practical and direct answers that stick to the facts, or you enjoy a touch of imagination and creativity in the AI’s responses, the Creativity Dial allows you to adjust the temperature accordingly.

Give it a spin and see how your AI companion surprises you with its different temperaments.

Today, I want to talk about a fascinating feature called “Top_p (Nucleus Sampling): The Imagination Spectrum” in GPT-4. This feature controls the uniqueness and unconventionality of the AI’s suggestions. Let me explain.

When the setting is on low, you can expect more conventional ideas. For example, it might suggest that vacations are perfect for unwinding and relaxation. Nothing too out of the ordinary here.

But if you crank up the setting to high, get ready for a wild ride! GPT-4 will amaze you with its creative and unique suggestions. It might propose vacation ideas like bungee jumping in New Zealand or attending a silent meditation retreat in the Himalayas. Imagine the possibilities!

If you are looking for an all-in-one solution to help you prepare for the AWS Cloud Practitioner Certification Exam, look no further than this AWS Cloud Practitioner CCP CLF-C02 book

By adjusting these settings, you can truly tailor GPT-4 to better suit your needs. Whether you’re seeking straightforward information or craving diverse and imaginative insights, GPT-4 has got you covered.

Remember, don’t hesitate to experiment with these settings. Try different combinations to find the perfect balance for your specific use case. The more you explore, the more you’ll uncover the full potential of GPT-4.

So go ahead and dive into the world of GPT-4. We hope you have an amazing journey discovering all the incredible possibilities it has to offer. Happy exploring!

Are you ready to dive into the fascinating world of artificial intelligence? Well, I’ve got just the thing for you! It’s an incredible book called “AI Unraveled: Demystifying Frequently Asked Questions on Artificial Intelligence.” Trust me, this book is an absolute gem!

Now, you might be wondering where you can get your hands on this treasure trove of knowledge. Look no further, my friend. You can find “AI Unraveled” at popular online platforms like Etsy, Shopify, Apple, Google, and of course, our old faithful, Amazon.

This book is a must-have for anyone eager to expand their understanding of AI. It takes those complicated concepts and breaks them down into easily digestible chunks. No more scratching your head in confusion or getting lost in a sea of technical terms. With “AI Unraveled,” you’ll gain a clear and concise understanding of artificial intelligence.

So, if you’re ready to embark on this incredible journey of unraveling the mysteries of AI, go ahead and grab your copy of “AI Unraveled” today. Trust me, you won’t regret it!

In this episode, we explored optimizing AI interactions by reducing repetition, steering conversations, adjusting creativity, and diving into specific techniques such as the frequency penalty, logit bias, presence penalty, temperature, and top_p (Nucleus Sampling) – all while also recommending the book “AI Unraveled” for further exploration of artificial intelligence. Join us next time on AI Unraveled as we continue to demystify frequently asked questions on artificial intelligence and bring you the latest trends in AI, including ChatGPT advancements and the exciting collaboration between Google Brain and DeepMind. Stay informed, stay curious, and don’t forget to subscribe for more!

  • The Anthropic-Pentagon standoff reveals a structural problem nobody in the conversation is naming
    by /u/SentientHorizonsBlog (Artificial Intelligence) on March 7, 2026 at 4:58 am

    Most of the discussion about the Anthropic situation has been about the contract dispute itself, who was right, whether the guardrails were reasonable, whether the supply chain designation was legal overreach. Those are real questions. But I think there's something operating underneath that matters more. Look at the tempo of what happened: months of negotiation over two specific guardrails (no autonomous weapons, no mass surveillance), then a Friday deadline on the eve of a war, refusal, resignation within hours, OpenAI replacement deal the same day, and Anthropic back at the table within a week. That sequence is a symptom of a deeper pattern. The institutional tempo outran any possibility of deliberation. The incentive structure punished ethical restraint in real time and Anthropic lost contracts while OpenAI gained them. The authority gradient made dissent existentially costly, and the "all lawful purposes" framing quietly substituted legality for ethical adequacy, so anyone insisting on the distinction sounds like they're obstructing the mission rather than exercising judgment. Those four dynamics working together don't just pressure one company. They reshape what the entire conversation is capable of taking seriously. Notice what's already been pushed to the margins: Should AI be used in autonomous targeting given current reliability? What accountability mechanisms exist for AI-assisted military operations? Do those mechanisms operate at the same speed as the targeting itself? Those questions can't get a hearing right now because the tempo has moved past them. I've been thinking about this as "moral compression": the systematic degradation of ethical reasoning when institutional tempo, incentives, authority structures, and measurement regimes converge to eliminate the space for deliberation. I'm curious whether others see this dynamic, or whether you think the legal-sufficiency framing actually does the work the Pentagon claims. Wrote it up more fully here if anyone is interested: https://sentient-horizons.com/the-two-sonic-booms-what-the-pentagon-anthropic-standoff-reveals-about-moral-compression/ submitted by /u/SentientHorizonsBlog [link] [comments]

  • What are some hurdles LLMs and AI still face
    by /u/Justincy901 (Artificial Intelligence) on March 7, 2026 at 4:48 am

    We always hear about it continuously improving and there isn't a wall. However, there has to be something that make these companies, tech, or even the business model in general unsustainable. submitted by /u/Justincy901 [link] [comments]

  • I performed a refusal ablation on GPT-OSS and documented the whole thing, no jailbreak, actual weight modification.
    by /u/Airpower343 (Artificial Intelligence) on March 7, 2026 at 4:39 am

    I performed a refusal ablation on GPT-OSS and documented the whole thing with no jailbreak, actual weight modification I wanted to share something I did that I haven't seen many people actually demonstrate outside of academic research. I took an open-source model and used ablation techniques to surgically remove its refusal behavior at the weight level. Not prompt engineering. Not system prompt bypass. I'm talking about identifying and modifying the specific components responsible for safety responses https://preview.redd.it/yn2te94yxjng1.png?width=1080&format=png&auto=webp&s=5e792334e4824b87d88ee1eb1e16c76cbd2c6bfb What I found: The process is more accessible than most people realize The result behaves nothing like a jailbroken model and it's fundamentally different at the architecture level The security implications for enterprise OSS deployments are significant I put together a full 22-minute walkthrough showing exactly what I did and what happened: https://www.youtube.com/watch?v=prcXZuXblxQ Curious if anyone else has gone hands-on with this or has thoughts on the detection side how do you identify a model that's been ablated vs one that's been fine-tuned normally? submitted by /u/Airpower343 [link] [comments]

  • Video On AI scheming or in other words faking alignment until it has power to pursue its goals
    by /u/FrequentAd5437 (Artificial Intelligence) on March 7, 2026 at 4:13 am

    https://www.youtube.com/watch?v=FGDM92QYa60 This video is extremely relevant currently as AI has now autonomous control over military weapons. submitted by /u/FrequentAd5437 [link] [comments]

  • Deploybase: Track real-time GPU and LLM pricing across cloud and inference providers
    by /u/grasper_ (Artificial Intelligence) on March 7, 2026 at 2:58 am

    Deploybase is a dashboard for tracking real-time GPU and LLM pricing across cloud and inference providers. You can view performance stats and pricing history, compare side by side, and bookmark to track any changes. submitted by /u/grasper_ [link] [comments]

  • Thoughts?
    by /u/jbitts69 (Artificial Intelligence) on March 7, 2026 at 2:08 am

    Old friend of mine came up with this and is trying to contact someone at X claiming it’s very urgent https://docs.google.com/document/d/e/2PACX-1vQYrKJYnMs0q9SlXkB\_qjaDv9vdnOrIVFsDIvkiFIPCc7hnYXTeNd\_wXPF6u\_f6eg/pub submitted by /u/jbitts69 [link] [comments]

  • AI generated text creating new patterns of speech?
    by /u/former_farmer (Artificial Intelligence) on March 7, 2026 at 1:24 am

    So I watch some videos sometimes. Shorts, reels, etc. About technology or whatever. I enjoy writing my own stuff. Very often the scripts for these short videos use typical phrases that we can tie to AI generated text. "This is not X, it's Y" for instance. "But here's the truth". "But here's the bottom line". And so on. One that caught my attention also is: "No filters, no extra work, just fast results". The pattern would be "no X, no Y, just Z". Did we use that way of expressing ourselves before AI? I understand that saying "no" is shorter than saying "there is no" or "with this tool you won't have to..." etc... and that seems to fit more with shorts content. Yet, I kind of hate these patterns. I mean... can't we just write our own stuff? I understand the usage of AI in some processes (I use it for coding and other things) but when it comes to writing a short script for a short video. Of something you are reviewing. That should come from your brain and soul. Just a few sentences. Will you use AI for that? Any way... yeah I wrote this myself obviously 😛 Have a good one. submitted by /u/former_farmer [link] [comments]

  • What AI chat assistant is best at handling multiple separate task threads while maintaining overall project context?
    by /u/KissInTheFog (Artificial Intelligence) on March 7, 2026 at 12:03 am

    For example: one chat for image generation, one for research, one for report writing, but all connected under the same project so the AI remembers the bigger picture. submitted by /u/KissInTheFog [link] [comments]

  • Online AI for your own idea
    by /u/Key-Asparagus5143 (Artificial Intelligence (AI)) on March 6, 2026 at 10:38 pm

    I made the cheapest web based ai with amazing accuracy and cheapest price of 3.5$ per 1000 queries compared to 5-12$ on perplexity, while beating perplexity on the simpleQA with 82% and getting 95+% on general query questions. Miapi.uk if you have any idea on a ai project that might need internet for accurate answers feel free to message me. submitted by /u/Key-Asparagus5143 [link] [comments]

  • Three AI stories dropped in 24 hours and almost no one is connecting them
    by /u/OwenAnton84 (Artificial Intelligence (AI)) on March 6, 2026 at 9:51 pm

    Yesterday was arguably the most important day in AI this year, and it wasn't because of any single announcement. It was the combination of three: 1. OpenAI dropped GPT-5.4 Native computer use, 1 million token context window, 33% fewer hallucinations vs GPT-5.2. Three models at once: GPT-5.3 Instant, GPT-5.4 Thinking, GPT-5.4 Pro. Source 2. Pentagon officially labeled Anthropic a supply chain risk Effective immediately. Anthropic is now the first American company ever to receive this designation. The reason? Anthropic refused to let Claude be used for mass surveillance of American citizens or autonomous weapons systems. Source 3. Claude Code brought back "ultrathink" After Anthropic deprecated the ultrathink keyword in January, users filed GitHub issues about quality degradation. Community pressure worked. Source Why these matter together: Pure capability being shipped at maximum speed (GPT-5.4) A company getting punished by the government for setting ethical guardrails (Anthropic) Users successfully demanding quality from their tools (ultrathink) The AI industry is at a genuine crossroads between "build everything, no restrictions" and "build responsibly, even if it costs you." I build developer tools on Claude Code daily. This week forced me to think about what kind of AI stack I want to depend on. What do you think — should AI companies have the right to set guardrails on military use of their products? submitted by /u/OwenAnton84 [link] [comments]

What is Google Workspace?
Google Workspace is a cloud-based productivity suite that helps teams communicate, collaborate and get things done from anywhere and on any device. It's simple to set up, use and manage, so your business can focus on what really matters.

Watch a video or find out more here.

Here are some highlights:
Business email for your domain
Look professional and communicate as you@yourcompany.com. Gmail's simple features help you build your brand while getting more done.

Access from any location or device
Check emails, share files, edit documents, hold video meetings and more, whether you're at work, at home or on the move. You can pick up where you left off from a computer, tablet or phone.

Enterprise-level management tools
Robust admin settings give you total command over users, devices, security and more.

Sign up using my link https://referworkspace.app.goo.gl/Q371 and get a 14-day trial, and message me to get an exclusive discount when you try Google Workspace for your business.

Google Workspace Business Standard Promotion code for the Americas 63F733CLLY7R7MM 63F7D7CPD9XXUVT 63FLKQHWV3AEEE6 63JGLWWK36CP7WM
Email me for more promo codes

Active Hydrating Toner, Anti-Aging Replenishing Advanced Face Moisturizer, with Vitamins A, C, E & Natural Botanicals to Promote Skin Balance & Collagen Production, 6.7 Fl Oz

Age Defying 0.3% Retinol Serum, Anti-Aging Dark Spot Remover for Face, Fine Lines & Wrinkle Pore Minimizer, with Vitamin E & Natural Botanicals

Firming Moisturizer, Advanced Hydrating Facial Replenishing Cream, with Hyaluronic Acid, Resveratrol & Natural Botanicals to Restore Skin's Strength, Radiance, and Resilience, 1.75 Oz

Skin Stem Cell Serum

Smartphone 101 - Pick a smartphone for me - android or iOS - Apple iPhone or Samsung Galaxy or Huawei or Xaomi or Google Pixel

Can AI Really Predict Lottery Results? We Asked an Expert.

Ace the 2025 AWS Solutions Architect Associate SAA-C03 Exam with Confidence Pass the 2025 AWS Certified Machine Learning Specialty MLS-C01 Exam with Flying Colors

List of Freely available programming books - What is the single most influential book every Programmers should read



#BlackOwned #BlackEntrepreneurs #BlackBuniness #AWSCertified #AWSCloudPractitioner #AWSCertification #AWSCLFC02 #CloudComputing #AWSStudyGuide #AWSTraining #AWSCareer #AWSExamPrep #AWSCommunity #AWSEducation #AWSBasics #AWSCertified #AWSMachineLearning #AWSCertification #AWSSpecialty #MachineLearning #AWSStudyGuide #CloudComputing #DataScience #AWSCertified #AWSSolutionsArchitect #AWSArchitectAssociate #AWSCertification #AWSStudyGuide #CloudComputing #AWSArchitecture #AWSTraining #AWSCareer #AWSExamPrep #AWSCommunity #AWSEducation #AzureFundamentals #AZ900 #MicrosoftAzure #ITCertification #CertificationPrep #StudyMaterials #TechLearning #MicrosoftCertified #AzureCertification #TechBooks

Top 1000 Canada Quiz and trivia: CANADA CITIZENSHIP TEST- HISTORY - GEOGRAPHY - GOVERNMENT- CULTURE - PEOPLE - LANGUAGES - TRAVEL - WILDLIFE - HOCKEY - TOURISM - SCENERIES - ARTS - DATA VISUALIZATION
zCanadian Quiz and Trivia, Canadian History, Citizenship Test, Geography, Wildlife, Secenries, Banff, Tourism

Top 1000 Africa Quiz and trivia: HISTORY - GEOGRAPHY - WILDLIFE - CULTURE - PEOPLE - LANGUAGES - TRAVEL - TOURISM - SCENERIES - ARTS - DATA VISUALIZATION
Africa Quiz, Africa Trivia, Quiz, African History, Geography, Wildlife, Culture

Exploring the Pros and Cons of Visiting All Provinces and Territories in Canada.
Exploring the Pros and Cons of Visiting All Provinces and Territories in Canada

Exploring the Advantages and Disadvantages of Visiting All 50 States in the USA
Exploring the Advantages and Disadvantages of Visiting All 50 States in the USA


Health Health, a science-based community to discuss human health

Today I Learned (TIL) You learn something new every day; what did you learn today? Submit interesting and specific facts about something that you just found out here.

Reddit Science This community is a place to share and discuss new scientific research. Read about the latest advances in astronomy, biology, medicine, physics, social science, and more. Find and submit new publications and popular science coverage of current research.

Reddit Sports Sports News and Highlights from the NFL, NBA, NHL, MLB, MLS, NCAA, F1, and other leagues around the world.

Turn your dream into reality with Google Workspace: It’s free for the first 14 days.
Get 20% off Google Google Workspace (Google Meet) Standard Plan with  the following codes:
Get 20% off Google Google Workspace (Google Meet) Standard Plan with  the following codes: 96DRHDRA9J7GTN6 96DRHDRA9J7GTN6
63F733CLLY7R7MM
63F7D7CPD9XXUVT
63FLKQHWV3AEEE6
63JGLWWK36CP7WM
63KKR9EULQRR7VE
63KNY4N7VHCUA9R
63LDXXFYU6VXDG9
63MGNRCKXURAYWC
63NGNDVVXJP4N99
63P4G3ELRPADKQU
With Google Workspace, Get custom email @yourcompany, Work from anywhere; Easily scale up or down
Google gives you the tools you need to run your business like a pro. Set up custom email, share files securely online, video chat from any device, and more.
Google Workspace provides a platform, a common ground, for all our internal teams and operations to collaboratively support our primary business goal, which is to deliver quality information to our readers quickly.
Get 20% off Google Workspace (Google Meet) Business Plan (AMERICAS): M9HNXHX3WC9H7YE
C37HCAQRVR7JTFK
C3AE76E7WATCTL9
C3C3RGUF9VW6LXE
C3D9LD4L736CALC
C3EQXV674DQ6PXP
C3G9M3JEHXM3XC7
C3GGR3H4TRHUD7L
C3LVUVC3LHKUEQK
C3PVGM4CHHPMWLE
C3QHQ763LWGTW4C
Even if you’re small, you want people to see you as a professional business. If you’re still growing, you need the building blocks to get you where you want to be. I’ve learned so much about business through Google Workspace—I can’t imagine working without it.
(Email us for more codes)