AI Innovations in September 2024

AI Innovations in September 2024

AI Innovations in September 2024.

In September 2024, the landscape of artificial intelligence is evolving at an unprecedented pace, with groundbreaking innovations and developments emerging daily. This blog, titled “AI Innovations in September 2024,” aims to capture the essence of these advancements, providing a comprehensive summary of the latest news and trends in AI technology. As we navigate through a month filled with transformative AI breakthroughs, our ongoing updates will highlight significant milestones, from the launch of cutting-edge AI models to the integration of AI in various sectors such as healthcare, finance, and creative industries. With each passing day, AI is reshaping how we interact with technology, enhancing productivity, and redefining our understanding of intelligence itself. Join us as we explore the exciting world of AI innovations, keeping you informed and engaged with the rapid changes that are set to influence our future. Whether you’re a tech enthusiast, a professional in the field, or simply curious about the implications of AI, this blog will serve as your go-to resource for staying updated on the latest developments throughout September 2024.

AI- Powered Jobs Interview Warmup

AI- Powered Jobs Interview Warmup

A Daily Chronicle of AI Innovations on September 13th 2024:

💥 OpenAI releases o1 model with human-like reasoning

🏛️ White House launches AI datacenter task force

🎬 AI image transitions with Latent Navigation

🤖 Google AI teaches robots to tie shoelaces

🧠 Chatbots may reduce conspiracy beliefs, study shows

🎧 FDA clears Apple’s AirPods Pro to act as your hearing aids

🛍 Amazon’s AI chatbot starts pushing ads

💥 OpenAI releases o1 model with human-like reasoning

  • OpenAI has unveiled o1, a new generative AI model family, with two variants initially available: o1-preview and o1-mini, designed for enhanced fact-checking and code generation, respectively.
  • o1 models are accessible to ChatGPT Plus or Team subscribers, and they offer improved reasoning capabilities by thoroughly considering all parts of a query before responding, resulting in better performance on complex tasks.
  • Despite its advanced features, o1 comes with significant cost and speed drawbacks, being slower and more expensive than previous models, with ongoing issues such as hallucinations and occasional inaccuracies.
  • o1 uses reinforcement learning and chain-of-thought processing to “think” before responding, mimicking human problem-solving.
  • It outperforms expert humans on PhD-level science questions and ranks in the 89th percentile for competitive programming.
  • The model also solved 83% of International Mathematics Olympiad qualifying exam problems, compared to GPT-4o’s 13%.
  • Two versions available: o1-preview and o1-mini — which by the time of this newsletter publishing — has been rolled out to all ChatGPT Premium and Teams users.
  • API access is significantly more expensive than GPT-4o, at $15 per 1M input tokens and $60 per 1M output tokens.
  • Source: https://techcrunch.com/2024/09/12/openai-unveils-a-model-that-can-fact-check-itself/

🧠 ‘We have the next few years in the bag:’ Sam Altman touts U.S. AI supremacy and OpenAI o1 release

Altman called OpenAI “the beginning of a significant new paradigm” during an interview on St. Louis Public Radio a day after the release. He spoke about the international race to develop artificial intelligence, particularly in light of comments from Putin that the country that controls AI will “become the ruler of the world.”

“There’s been this whole debate about, ‘Is AI capping out — are we getting close to some ceiling? Is progress going to slow down?’” Altman said. “And I think the most important message of this release is that, not only is progress not slowing down, but we have the next few years in the bag.”

Source: https://www.stlpr.org/show/st-louis-on-the-air/2024-09-13/sam-altman-chatgpt-openai-o1-st-louis

🏛️ White House launches AI datacenter task force

The White House is launching a new AI datacenter infrastructure task force led by the National Security Council, National Economic Council, and Deputy Chief of Staff’s office to ensure U.S. leadership in AI development.

  • Tech leaders from Nvidia, OpenAI, Anthropic, Google, Microsoft, and Amazon met with government officials to discuss AI energy usage, datacenter capacity, job creation, and strategic locations.
  • The task force will coordinate policies to advance datacenter development, aligning with economic, national security, and environmental goals.
  • The administration is streamlining datacenter permitting processes and leveraging Department of Energy resources to support AI infrastructure growth.
  • Tech giants reaffirmed commitments to net-zero carbon emissions and clean energy procurement for their operations.

This task force is a big shift in U.S. AI strategy, moving beyond safety testing to actively shaping the infrastructure needed to maintain America’s edge in AI. The news also follows the recent announcement that OpenAI and Anthropic will let the U.S. AI Safety Institute test new models before releasing them publicly.

Source: https://www.whitehouse.gov/briefing-room/statements-releases/2024/09/12/readout-of-white-house-roundtable-on-u-s-leadership-in-ai-infrastructure

🎬 AI image transitions with Latent Navigation

Hugging Face’s Latent Navigation space lets you generate smooth transitions between AI-generated images using CLIP text embeddings and Flux.1 schnell.

  1. Visit Latent Navigation space.
  2. Enter a detailed image description in the “Prompt” field (e.g., “a golden retriever in the park”)
  3. Set transition directions: “1st direction to steer” (e.g., “winter”) and “2nd direction to steer” (e.g., “summer”)
  4. Adjust the “Strength” and other settings to fine-tune your transition
  5. Click “Generate directions” to create your AI image transition

Pro tip: Experiment with the “Advanced options” for more control over your transitions.

Source: https://university.therundown.ai/c/daily-tutorials/create-ai-image-transitions-with-latent-navigation-83bda6e0-ee83-44df-8502-772e1c21c639

🤖 Google AI teaches robots to tie shoelaces

  • Google DeepMind announced that it taught a robot to tie a shoe, but the robot’s shoe-tying skills are still very poor.
  • The robot uses bi-arm manipulation and is powered by two AI systems, ALOHA Unleashed and DemoStart, to perform complex tasks.
  • Many companies are developing robots to handle human-like tasks, with shoe-tying being a significant test of a robot’s capabilities.
  • The first system, ALOHA Unleashed, allows robots to perform five complex bi-arm manipulation tasks, including tying shoelaces, hanging shirts, and repairing other robots.
  • The second, DemoStart, uses reinforcement learning to teach robots dexterous behaviors in simulation, requiring 100x fewer demonstrations than traditional methods.
  • The systems use diffusion methods similar to image generation models to predict robot actions from random noise.
  • Robots achieved a 98% success rate in simulations and up to 97% in real-world tasks like cube reorientation.

While still far from human-level dexterity, these advancements represent another leap towards creating more useful robots for everyday tasks. The application of image generation techniques to robotics also shows how breakthroughs in one area of AI, can also trigger advancements elsewhere across the field.

🧠 Chatbots may reduce conspiracy beliefs, study shows

  • Researchers from MIT Sloan and Cornell University found that AI chatbots can reduce belief in conspiracy theories by about 20%, even among individuals who consider these beliefs important to their identity.
  • The study, published in Science, involved 2,190 participants who engaged in conversations with OpenAI’s GPT-4 Turbo, which provided tailored counterarguments to their conspiracy theories.
  • The research demonstrated that participants experienced a significant decrease in their confidence regarding conspiracy theories, highlighting the effectiveness of chatbots in changing minds with evidence-based arguments.
  • Source: https://www.technologyreview.com/2024/09/12/1103930/chatbots-can-persuade-people-to-stop-believing-in-conspiracy-theories/

🎧 FDA clears Apple’s AirPods Pro to act as your hearing aids 

  • The U.S. FDA authorized Apple’s AirPods Pro to function as an over-the-counter hearing aid device, marking the first software-based solution for hearing assistance integrated into existing consumer products.
  • The AirPods Pro 2 will receive a software update this fall, enabling features that help users with mild to moderate hearing impairments by creating personalized listening profiles through tests conducted on an iPhone.
  • A clinical study involving 118 participants with mild to moderate hearing loss showed that Apple’s self-fitting hearing aid feature provided similar benefits to professional fittings, and the FDA found no adverse effects from using the new AirPods software.
  • Source: https://gizmodo.com/fda-clears-apples-airpods-pro-to-act-as-your-hearing-aids-2000498087

🛍 Amazon’s AI chatbot starts pushing ads 

  • Amazon’s AI-powered shopping assistant, Rufus, is designed to help users navigate the extensive range of products on Amazon’s website, and is now including some advertisements in its recommendations, as initially reported by Adweek.
  • Rufus generates product suggestions based on Amazon’s extensive catalog, customer reviews, and community Q&As, while also integrating ads into these conversations to help users discover more products in a seamless manner.
  • Although Rufus remains experimental, Amazon aims to ensure that sponsored suggestions remain relevant and contextually appropriate, minimizing the risk of irrelevant advertisements while potentially revolutionizing online advertising with this new approach.
  • Source: https://www.techradar.com/computing/artificial-intelligence/amazons-ai-personal-shopper-is-sharing-ads-with-its-advice

Nvidia’s AI Surge Drives 8% Share Price Jump Amid Soaring Demand

Nvidia stock went up 8% after CEO Jensen Huang said that demand for its AI products was at an all-time high. Huang also talked about how important it is for big tech companies like Microsoft, Amazon, and Oracle to use Nvidia’s products to power their cloud and AI infrastructure.

Source: https://theaiwired.com/nvidias-ai-surge-drives-8-share-price-jump-amid-soaring-demand/

What Else is Happening in AI on September 13th 2024!

Google began rolling out Gemini Live to free users on the Gemini Android app — offering natural voice conversations with the AI assistant and 10 new voice options.

Source: https://x.com/GeminiApp/status/1834269227118924098

OpenAI’s ChatGPT reportedly surpassed 11 million paying subscribers, including 1 million on higher-priced business plans, potentially generating over $2.7 billion in annual revenue, according to COO Brad Lightcap.

Source: https://x.com/amir/status/1834347880251052203

Salesforce unveiled Agentforce, a suite of low-code tools for building autonomous AI agents capable of reasoning and completing tasks in sales, service, marketing, and commerce.

Source: https://www.salesforce.com/agentforce

Mastercard agreed to acquire AI-powered threat intelligence company Recorded Future for $2.65 billion, aiming to enhance its cybersecurity capabilities.

Source: https://www.mastercard.com/news/press/2024/september/mastercard-invests-in-continued-defense-of-global-digital-economy-with-acquisition-of-recorded-future

Google introduced DataGemma, the first open models designed to connect large language models with real-world data from Google’s Data Commons, aiming to reduce AI hallucinations by grounding responses in factual stats.

Source: https://blog.google/technology/ai/google-datagemma-ai-llm

Hume AI launched Empathic Voice Interface 2 (EVI 2), a new voice-to-voice foundation model trained for emotional intelligence, capable of understanding and generating various tones of voice and speaking styles.

Source: https://www.hume.ai/blog/introducing-evi2

Meta reignites plans to train AI using UK users’ public Facebook and Instagram posts.

Source: https://techcrunch.com/2024/09/13/meta-reignites-plans-to-train-ai-using-uk-users-public-facebook-and-instagram-posts/

The ‘godmother of AI’ just raised $230 million for her AI startup.

Source: https://qz.com/fei-fei-li-ai-startup-world-labs-raise-230-million-1851647701

MongoDB CEO says if AI hype were the dotcom boom it is 1996.

Source: https://www.theregister.com/2024/09/13/mongodb_ceo_says_if_ai/

Salesforce unveils its vision of AI agents to help every business.

Source: https://www.techradar.com/pro/salesforce-unveils-its-vision-of-ai-agents-to-help-every-business

A Daily Chronicle of AI Innovations on September 12th 2024:

A Daily Chronicle of AI Innova – AI Unraveled: Latest AI News & Trends, GPT, ChatGPT, Gemini, Generative AI, LLMs, Prompting – Apple Podcasts

🤖 Google turns your notes into podcasts

🔥 Adobe previews new AI video model

🎥 How to control AI video camera angles

🤖 Ireland Launches GDPR Investigation into Google’s AI Model

💰 OpenAI Secures $6.5B Investment Talks, Targets Massive $150B Valuation

⭐ OpenAI release OpenAI o1—a new series of AI models designed to spend more time thinking before they respond.

🖼️ Mistral releases multimodal Pixtral 12B

🤖 Ireland Launches GDPR Investigation into Google’s AI Model

  • Ireland’s DPC is investigating Google’s AI model for potential GDPR violations regarding personal data use;
  • Google’s PaLM 2 is the subject of the inquiry due to its potential high-risk data processing;
  • The probe follows a similar case involving X, which agreed to stop using EU/EEA user data for AI training.

Source: https://www.bitdegree.org/crypto/news/ireland-launches-gdpr-investigation-into-googles-ai-model?utm_source=reddit&utm_medium=social&utm_campaign=r-ireland-gdpr-investigation-google-ai


AI Unraveled: Demystifying Frequently Asked Questions on Artificial Intelligence (OpenAI, ChatGPT, Google Gemini, Generative AI, Discriminative AI, xAI, LLMs, GPUs, Machine Learning, NLP, Promp Engineering)

🖼️ Mistral releases multimodal Pixtral 12B

French AI startup Mistral has released Pixtral 12B, its first multimodal model capable of processing both images and text, available for free download under an Apache 2.0 license.

  • Pixtral 12B is a 12-billion-parameter model, about 24GB in size, built on Mistral’s text model Nemo 12B.
  • Pixtral 12B is the startups first multimodal model, meaning it can process and answer questions about images, as well as text.
  • It’s available for download via GitHub and Hugging Face under an Apache 2.0 license, allowing unrestricted use and fine-tuning.
  • Mistral plans to make Pixtral 12B available on their chatbot and API platforms, Le Chat and Le Plateforme, soon.

Despite being just over a year old, Mistral is rapidly emerging as Europe’s “answer” to OpenAI. With a lean team of top researchers and a recent $645 million funding round, the company is aggressively pushing the boundaries of AI with powerful open models.

Source: https://x.com/MistralAI/status/1833758285167722836

🎥 How to control AI video camera angles

Luma Labs’ new ‘Camera Motion’ feature allows you to add professional-looking camera movements to your AI-generated videos with simple text commands.

  1. Log in into Luma Labs and access the main interface.
  2. Set up your scene by uploading an image or describing it in the text box.
  3. Type “camera” in your prompt to unlock the Camera Motion dropdown menu.
  4. Choose from options like Move, Pan, Push, Pull, or Orbit to add dynamic movement to your scene.
  5. Generate your video and iterate as needed to perfect the result.

Pro tip: Add starting and ending keyframes for even more control over your camera movements.

Source: https://university.therundown.ai/c/daily-tutorials/more-ai-video-control-with-camera-motion-e6534a4a-71fd-499c-83a1-22e9bf970ec8

🔥 Adobe previews new AI video model

Adobe just previewed its Firefly AI Video Model, which includes tools to extend existing videos and create new clips from text or image prompts, coming before year end.

  • Three key features are coming: Text to Video, Image to Video, and Generative Extend.
  • Text to Video can create clips from prompts, with options for camera controls and reference images.
  • Image to Video transforms still images or illustrations into live-action clips.
  • Generative Extend, coming to Premiere Pro beta, can add footage to cover gaps or extend shots.

While OpenAI’s (still unreleased) Sora focuses on generating videos from scratch, Adobe is aiming to create ‘a new era‘ for video editing itself. Soon, anyone will be able to change camera angles, extend scenes, and generate b-roll instantly with AI.

Source: https://blog.adobe.com/en/publish/2024/09/11/bringing-gen-ai-to-video-adobe-firefly-video-model-coming-soon

🤖 Google turns your notes into podcasts

Google just launched Audio Overviews, a new feature in NotebookLM that turns notes, PDFs, Google Docs, Slides, and more into AI-generated audio discussions between two virtual AI agents.

  • Audio Overview creates a “deep dive” conversation from uploaded sources, with AI hosts summarizing content and connecting topics across materials.
  • The tool works with various source types (documents, slides, charts, web URLs) using Gemini 1.5’s multimodal capabilities.
  • To use Audio Overviews, open an existing notebook, navigate to the Notebook guide, and click on the “generate” button on the right-hand side.
  • We confirmed with Google Labs that NotebookLM can process up to 50 sources, each up to 500,000 words, allowing for a total of 25 million words to be considered when generating the audio.

Audio Overviews can be a game changer for people who are auditory learners. This feature works particularly well with academic papers, ebooks, textbooks, presentations — we tested it by turning yesterdays newsletter into an Audio Overview and were quite impressed!

Source: https://blog.google/technology/ai/notebooklm-audio-overviews/

Generative AI: Techniques, Applications, and Impact on Business

Generative AI involves various techniques, each with its strengths and applications. Let’s explore some of the most prominent generative AI models behind this groundbreaking technology:

Generative Adversarial Networks (GANs):

Imagine two AI artists competing in an artistic duel. That’s the concept behind GANs! They consist of two neural networks:

  • Generator: This network acts as the creative visionary, constantly working to produce new and realistic data like images, music, or text.
  • Discriminator: This network acts as the critical judge, carefully analyzing the generated data to distinguish it from real data.

Source: https://www.seaflux.tech/blogs/generative-ai-revolutionizing-industries

💰 OpenAI Secures $6.5B Investment Talks, Targets Massive $150B Valuation

OpenAI, the company behind the recent viral ChatGPT artificial intelligence language tool, informally reported that they are discussing with investors funding at about $6. 5 billion (£5 billion). This would increase the value of this company to a whopping $150 billion (£115 billion). With this possible value, OpenAI would be one of the world’s most valuable startups, and this example illustrates how fast the company rose through the ranks in the tech industry. A story from Bloomberg reveals that the company is also exploring means of funding $5 billion (£3. 8 billion) through a revolving credit arrangement with the commercial banks.

Source: https://theaiwired.com/openai-secures-6-5b-investment-talks-targets-massive-150b-valuation/

Oracle To Deploy A Supercluster Of ~130,000 NVIDIA Blackwell GPUs, Alludes To A “Gigawatt” Capacity Data Center That Will Be Powered By 3 Nuclear Reactors

r/singularity - Oracle To Deploy A Supercluster Of ~130,000 NVIDIA Blackwell GPUs, Alludes To A “Gigawatt” Capacity Data Center That Will Be Powered By 3 Nuclear Reactors

Source: https://wccftech.com/oracle-to-deploy-a-supercluster-of-130000-nvidia-blackwell-gpus-alludes-to-a-gigawatt-capacity-data-center-that-will-be-powered-by-3-nuclear-reactors/

Introducing PaperQA2, the first AI agent that conducts entire scientific literature reviews on its own.

r/singularity - "Introducing PaperQA2, the first AI agent that conducts entire scientific literature reviews on its own."

Source: https://x.com/SGRodriques/status/1833908643856818443

What Else is Happening in AI on September 12th 2024!

OpenAI is reportedly in talks to raise $6.5 billion (plus another $5 billion credit line with banks) at a $150 billion valuation, significantly higher than its previous $86 billion valuation.

Source: https://www.bloomberg.com/news/articles/2024-09-11/openai-fundraising-set-to-vault-startup-s-value-to-150-billion

Meta is reportedly finalizing a new AI supercomputing cluster with over 100,000 Nvidia H100 chips to train its next Llama 4 language model.

Source: https://www.theinformation.com/articles/meta-will-soon-get-a-100-000-gpu-cluster-too-whats-life-at-character-like-now

Canada’s Defence Minister announced the country is working on incorporating AI into its military by 2030, emphasizing for the tech to “improve the work of our military personnel, but not to replace it”.

Source: https://www.ctvnews.ca/politics/canada-working-on-military-ai-blair-says-as-he-endorses-international-agreement-1.7033904

Kai-Fu Lee (former head of Google China), stated that Chinese AI models lag behind U.S. by 6 to 9 months, but predicts that Chinese AI apps will proliferate faster than in the U.S. by early next year.

Ace the Microsoft Azure Fundamentals AZ-900 Certification Exam: Pass the Azure Fundamentals Exam with Ease

Source: https://www.cnbc.com/2024/09/11/chinas-ai-models-lag-their-us-counterparts-by-6-to-9-months-says-former-head-of-google-china.html

Taylor Swift endorsed Kamala Harris for president, partly prompted by AI-generated deepfakes falsely showing her supporting Trump.

Source: https://www.theverge.com/2024/9/10/24241538/taylor-swift-endorses-kamala-harris-donald-trump-ai-endorsement-deepfake

Oracle unveiled over 50 AI agents for its Fusion Cloud Applications Suite, designed to automate business processes across functions like capital management, supply chain, and customer experience.

Source: https://www.techtarget.com/searchenterpriseai/news/366610494/Oracle-launches-over-50-AI-Agents-for-business-use

Algorithm takes robots a step closer to being able to ‘act on intuition’.

Source: https://techxplore.com/news/2024-09-algorithm-robots-closer-intuition.html

Nevada will use Google’s AI to help decide if people should receive unemployment benefits.

Source: https://www.techspot.com/news/104696-nevada-use-google-ai-help-decide-if-people.html

⭐ OpenAI release OpenAI o1—a new series of AI models designed to spend more time thinking before they respond.

u/enoumen - A Daily Chronicle of AI Innovations on September 12th 2024: 🤖Google turns your notes into podcasts 🔥Adobe previews new AI video model 🎥How to control AI video camera angles 🤖Ireland Launches GDPR Investigation into Google's AI Model 💰OpenAI  Targets Massive $150B Valuation

We’re releasing a preview of OpenAI o1—a new series of AI models designed to spend more time thinking before they respond. This new series of AI models can reason through complex tasks and solve harder problems than previous models in science, coding, and math. Rolling out today in ChatGPT to all Plus and Team users, and in the API for developers on tier 5.

chart, bar chart

Image preview

  • Competition Math (AIME 2024):
    • The initial GPT-4 preview performed at 13.4% accuracy.
    • The new GPT-4-1 model in its early version showed much better results, achieving 56.7%.
    • In the final version, it soared to 83.3%.
  • Competition Code (CodeForces):
    • The GPT-4 preview started with only 11.0%.
    • The first GPT-4-1 version improved significantly to 62.0%.
    • The final version reached a high accuracy of 89.0%
  • PhD-Level Science Questions (GPAQ Diamond):
    • GPT-4 preview scored 56.1%.
    • GPT-4-1 improved to 78.3% in its early version and maintained a similar high score at 78.0%
    • The expert human benchmark for comparison scored 69.7%, meaning the GPT-4-1 model slightly outperformed human experts in this domain

it can literally perform better than a PhD human right now

More here: www.openai.com/o1

If you are looking for an all-in-one solution to help you prepare for the AWS Cloud Practitioner Certification Exam, look no further than this AWS Cloud Practitioner CCP CLF-C02 book

GPT o1 still fails basic math riddles

Before everyone goes crazy about the new model I will leave this here:

I already found an example of the model failing a simple primary school level riddle.

Original riddle as found online: Four people need to cross a rickety bridge at night. Unfortunately, they have one torch and the bridge is too dangerous to cross without a torch. The bridge can support only two people at a time. All the people don’t take the same time to cross the bridge. Time for each person: 1 min, 2 mins, 7 mins, and 10 mins. What is the shortest time needed for all four of them to cross the bridge?

My modified version (removed one constraint about how many people can cross at once): Four people need to cross a rickety bridge at night. Unfortunately, they have one torch and the bridge is too dangerous to cross without a torch. All the people don’t take the same time to cross the bridge. Time for each person: 1 min, 2 mins, 7 mins, and 10 mins. What is the shortest time needed for all four of them to cross the bridge?

GPT o1 Preview Response:

Thought for 25 seconds

The shortest time required for all four people to cross the bridge is 17 minutes. Here’s the optimal plan:

Person 1 (1 min) and Person 2 (2 mins) cross together.

Time elapsed: 2 minutes (the slower person’s time). On the other side: Persons 1 and 2. Original side: Persons 3 and 4. Person 1 (1 min) returns with the torch.

Time elapsed: 2 + 1 = 3 minutes. On the other side: Person 2. Original side: Persons 1, 3, and 4. Person 3 (7 mins) and Person 4 (10 mins) cross together.

Time elapsed: 3 + 10 = 13 minutes. On the other side: Persons 2, 3, and 4. Original side: Person 1. Person 2 (2 mins) returns with the torch.

Time elapsed: 13 + 2 = 15 minutes. On the other side: Persons 3 and 4. Original side: Persons 1 and 2. Person 1 (1 min) and Person 2 (2 mins) cross together again.

Time elapsed: 15 + 2 = 17 minutes. All four persons are now on the other side. Explanation of the strategy:

Fastest Movers as Couriers: Persons 1 and 2 are the fastest and are used to shuttle the torch back and forth. Slowest Pair Together: Persons 3 and 4 cross together, so their longer times only affect one crossing. Minimizing Return Times: The returns are made by the fastest individuals to reduce time spent. No alternative sequence allows all four to cross in less than 17 minutes given the constraints.

Take away: Real answer without the constraint should be 10, since without the constraint all people can cross at once. It basically still just gives answers it learnt online, rather than “reasoning”.

Credit: based on an example found by Darius Burschka

DataGemma: Using real-world data to address AI hallucinations

Announcing DataGemma, open models that leverage the power & knowledge of Data Commons to ground LLMs in real-world data & decrease the risk of “hallucinations.”

Google’s Data Commons is a publicly available data repository that serves as a foundation for factual AI. Explore how DataGemma utilizes Data Commons through Retrieval Interleaved Generation (RIG) & Retrieval Augmented Generation (RAG) approaches for fact-checking, responsible AI development, & more.

DataGemma are the world’s first open models designed to help address the challenges of hallucination by grounding LLMs in the vast, real-world statistical data of Google’s Data Commons.

DataGemma Logo

Source: https://blog.google/technology/ai/google-datagemma-ai-llm/?linkId=10938872

A Daily Chronicle of AI Innovations on September 11th 2024:

🍓 OpenAI’s Strawberry set for early release

🪑 Engineer creates AI-designed walking table

🎨 Create photo-illustration blends with AI

🚗 Waymo’s driverless cars outperform humans

🤷‍♂️ Meta scraped every Australian user’s account to train its AI

🔮 Mistral releases Pixtral 12B, its first multimodal AI model

💻 Sergey Brin says Google engineers underutilize AI for coding

🍓 OpenAI’s Strawberry set for early release

According to a new report by The Information, OpenAI is planning to release Strawberry, its reasoning-focused AI model, as part of ChatGPT within the next two weeks — earlier than the originally expected fall timeline.

  • According to early testers, Strawberry “thinks” for 10-20 seconds before responding, unlike typical AI models that answer immediately.
  • The initial version will only handle text input and output, not images, making it non-multimodal unlike other recent AI models.
  • Strawberry is expected to improve complex query handling without requiring users to use specialized prompt techniques like “chain-of-thought-prompting“.
  • Early testers also said that Strawberry’s responses, while slightly better than GPT-4o, may not be worth the extra 10-20 second wait time.

If it lives up to reports, Strawberry will likely excel not only on objective tasks (math, coding) but also subjective ones (like product marketing) and open up a whole new world of practical AI use cases. But with early testers reporting that the model is only “slightly better“, is it really worth all the hype? We’ll find out soon.

Source: https://www.theinformation.com/articles/new-details-on-openais-strawberry-apples-siri-makeover-larry-ellison-doubles-down-on-data-centers

🪑 Engineer creates AI-designed walking table

Dutch engineer Giliam de Carpentier has created the ‘Carpentopod’, a wireless walking wooden coffee table, using an AI-optimized leg mechanism designed through evolutionary algorithms.

  • The leg mechanism was developed using software that simulated thousands of generations of designs, optimizing for walking speed, clearance, and efficiency.
  • The table features 12 legs (six on each end) made from precision-cut laminated bamboo, controlled by two motors.
  • Carpentier developed a custom motor control software and uses a repurposed Nunchuck controller for remote operation.
  • Watch a live demo of the walking coffee table in action here.

This project shows how makers are combining AI, traditional craftsmanship, and modern fabrication techniques to create unique projects that blur the lines between art, engineering, and furniture design. AI for design optimization, even for seemingly simple mechanical systems — is just getting started.

Source: https://www.decarpentier.nl/carpentopod

🎨 Create photo-illustration blends with AI

Using this unique prompt, AI image generators like Freepik and Midjourney can create images that blend both photography and illustration together.

  1. Choose an AI image generator (e.g., Freepik or Midjourney).
  2. Use this prompt template: “Blending of editorial photography and illustration, model, illustration fashion outfit, surreal environment of vibrant bold illustrations blending with the subject, flowers, smoke, flames and sparkles, half photography and half illustration”.
  3. Generate your image and check out the results.
  4. Upscale using 2K resolution, add “mix of photography and illustration” to the upscaler prompt, and apply “subtle” and “digital art” parameters.

Source: https://university.therundown.ai/c/daily-tutorials/create-photo-illustration-blends-with-ai-15b43823-d20a-4895-968d-fead4eb3e0ef

🚗 Waymo’s driverless cars outperform humans

A new deep dive into Waymo’s new crash data suggests that driverless cars are significantly safer than human-driven vehicles, with fewer injury-causing crashes and airbag deployments per mile driven.

  • Waymo vehicles have traveled over 22 million miles, with only 20 injury-causing crashes reported.
  • This rate is less than one-third of what’s expected for human drivers in similar conditions.
  • Out of 23 severe crashes, 16 involved human drivers rear-ending Waymo vehicles.
  • Waymo estimates its vehicles are one-sixth as likely as human drivers to experience crashes triggering airbag deployment.

As Waymo rapidly scales up its robotaxi service — now providing 100,000 rides per week — these safety statistics are crucial. The data suggests that we now have safe self-driving cars, and that these AI-driven cars could significantly reduce road accidents and injuries.

Source: https://x.com/binarybits/status/1833528077496029643

US Man Charged In Historic AI Music Fraud Case: Used Thousands Of Bots To Stream Fake Songs, Raked In $10M In Royalties

An American musician is facing charges for using AI to fraudulently inflate his song streams and earn millions in royalties. Prosecutors have never seen a case like this before. The musician faces decades in prison if convicted.

Source: Read the full story: https://www.ibtimes.co.uk/us-man-charged-historic-ai-music-fraud-case-used-thousands-bots-stream-fake-songs-raked-10m-1726815

James Earl Jones’ Darth Vader voice lives on through AI. Voice actors see promise and peril in that

James Earl Jones’ Darth Vader voice lives on through AI. Voice actors see promise and peril in that
Source: https://candorium.com/news/20240911050001425/james-earl-jones-darth-vader-voice-lives-on-through-ai-voice-actors-see-promise-and-peril-in-that

Taylor Swift says AI version of herself falsely endorsing Trump ‘conjured up my fears’

r/singularity - Taylor Swift says AI version of herself  falsely endorsing Trump 'conjured up my fears'

Source: https://www.the-express.com/entertainment/celebrity-news/148376/taylor-swift-ai-fake-trump-endorsement-fears

🤷‍♂️ Meta scraped every Australian user’s account to train its AI

  • Meta’s global privacy director admitted that Meta scrapes photos and texts from all public Facebook and Instagram posts from Australian users since 2007 to train its AI technology.
  • Unlike the European Union, Australian users do not have an opt-out option from data collection for AI training, which Meta attributes to the lack of specific privacy regulations in Australia.
  • Meta does not scrape data from users under 18 but collects information if shared on accounts managed by their parents or guardians, indicating a gap in data protection for minors.
  • Source: https://www.engadget.com/apps/meta-scraped-every-australian-users-account-to-train-its-ai-120026200.html

🔮 Mistral releases Pixtral 12B, its first multimodal AI model

  • Mistral has launched Pixtral 12B, a 12-billion-parameter multimodal model capable of processing both images and text, making it the company’s first such model.
  • This new model, which is built on Mistral’s existing Nemo 12B text model, can interpret images of various sizes and can be accessed via platforms like GitHub and Hugging Face.
  • Pixtral 12B’s release follows Mistral’s successful $645 million funding round, positioning the company as a significant European contender in the AI industry, comparable to OpenAI.
  • Source: Mistral releases Pixtral 12B, its first multimodal model | TechCrunch

💻 Sergey Brin says Google engineers underutilize AI for coding

Sergey Brin, Google’s cofounder, believes that the company’s engineers are not using artificial intelligence for coding as frequently as they should.

NotebookLM.Google.com can now generate podcasts from your Documents and URLs!

Ready to have your mind blown? This is not an ad or promotion for my product. It is a public Google product that I just find fascinating!

This is one of the most amazing uses of AI that I have come across and it went live to the public today!

For those who aren’t using Google NotebookLM, you are missing out. In a nutshell it lets up upload up to 100 docs each up to 200,000 words and generate summaries, quizes, etc. You can interrogate the documents and find out key details. That alone is cool, but TODAY they released a mind blowing enhancement.

Google NotebookLM can now generate podcasts (with a male and female host) from your Documents and Web Pages!

Try it by going to NotebookLM.google.com uploading your resume or any other document or pointing it to a website. Then click * Notebook Guide to the right of the input field and select Generate under Audio Overview. It takes a few minutes but it will generate a podcast about your documents! It is amazing!!

Source: http://notebooklm.google.com/

What Else is Happening in AI  on September 11th 2024?

China refuses to sign agreement to ban AI from controlling nuclear weapons.

Source: https://fortune.com/2024/09/10/china-ai-ban-nuclear-weapons

SambaNova launched SambaNova Cloud, claiming it’s the world’s fastest AI inference platform, running Llama 3.1 70B at 461 tokens per second and 405B at 132 tokens per second at full precision.

Source: https://x.com/SambaNovaAI/status/1833504179148755122

Salesforce introduced Industries AI, embedding over 100 industry-specific AI capabilities across 15 industry clouds to address unique challenges.

Source: https://www.salesforce.com/news/stories/industries-ai-announcement

Glean raised over $260 million in new funding at a whopping $4.6 billion valuation, co-led by Altimeter and DST Global, to expand its Work AI platform.

Source: https://x.com/jainarvind/status/1833479487415325006

Anthropic unveiled Workspaces in its API Console, allowing developers to manage multiple Claude deployments with granular controls for spend limits, rate limits, and access management.

Source: https://x.com/AnthropicAI/status/1833529395765776615

UBS developed an AI tool capable of analyzing 300,000 companies in 20 seconds to generate M&A ideas and identify potential buyers.

Source: https://www.bloomberg.com/news/articles/2024-09-10/ubs-has-an-ai-tool-that-can-scan-300-000-firms-in-20-seconds

The Australian government proposed new legal requirements for AI in high-risk settings and introduced a Voluntary AI Safety Standard for businesses.

Source: https://www.climatecontrolnews.com.au/contracting/new-ai-safety-standard

A Daily Chronicle of AI Innovations on September 10th 2024:

🆕 Apple launches AI-powered iPhone 16

🔬 Google’s AlphaProteo AI designs new protein binders

💊 OpenAI backs new drug discovery startup

🧠 AI generates ideas better than humans

📱 iPhone 16 gets AI superpowers

Apple just launched the iPhone 16, and it’s all about AI. The new A18 chip powers Apple Intelligence, a suite of AI features:

  • Smarter writing tools that can rewrite emails and create custom emojis,
  • Photo searches that understand natural language, and
  • Siri finally gets what you’re saying (and can actually do something about it).

But it’s not just about the iPhone. Apple is bringing AI to the whole ecosystem. The Apple Watch Series 10 is 10% lighter and adds AI features like smart photo curation for watch faces. AirPods 4 now understands head gestures, letting you nod or shake to respond to Siri. While iOS 18 launches September 16, the full suite of Apple Intelligence features will roll out gradually through 2024, starting with a U.S. beta this fall.

Why does it matter?

By developing its own AI models and integrating them deeply into its ecosystem, Apple is not just competing with Google and OpenAI – it’s changing what we expect from our devices.

Source: https://www.apple.com/apple-events

Apple Faces Criticism in China Over Lack of AI in New iPhones as Huawei Gains Ground

The latest iPhones unveiled by Apple lack AI in China and people are angered while Huawei’s advancement continues. It is widely believed that China consumers are dissatisfied with the newest iPhone released by Apple as it lacks the integration of new age Artificial Intelligence. This has created concern for the tech giant particularly when Huawei is now a more enhanced competitor.

Source: https://theaiwired.com/apple-faces-criticism-in-china-over-lack-of-ai-in-new-iphones-as-huawei-gains-ground/

🔬Google’s AlphaProteo AI designs new protein binders

Google DeepMind’s new AI, AlphaProteo, can design novel proteins that bind to specific targets up to 300 times more effectively than current methods. In lab tests, AlphaProteo’s creations successfully designed binders for 7 out of 8 target proteins, including some linked to cancer and viral infections.

It could accelerate drug discovery processes, improve disease diagnostic tools, and contribute to developing pest-resistant crops. While AlphaProteo could not design a binder for one challenging protein (TNFα), Google’s research team is actively expanding the system’s capabilities to address more complex targets.

Why does it matter?

This shows how AI can accelerate the development of new treatments for diseases like cancer and create more effective diagnostic tools. It also saves years of lab work and billions in research costs, potentially bringing life-saving drugs to market faster.

Source: https://deepmind.google/discover/blog/alphaproteo-generates-novel-proteins-for-biology-and-health-research

💊 OpenAI backs new drug discovery startup

 OpenAI and Thrive Capital recently backed Chai Discovery, a six-month-old AI biology startup founded by ex-OpenAI and Meta researchers that raised $30 million to develop AI models for drug discovery.

  • Chai’s AI model, Chai-1, predicts biochemical molecule structures, potentially speeding up drug development.
  • The company claims Chai-1 outperforms Google DeepMind’s AlphaFold on certain benchmarks.
  • Chai-1 can work with proteins, small molecules, DNA, and RNA, making it versatile for various applications.
  • Chai is making its first model free and open-source for non-commercial use.

Google DeepMind’s AlphaFold is the pinnacle of protein prediction models, and Chai is already giving it a run for its money. The more money flowing into drug discovery, the closer we get to curing major diseases using AI — and Chai is even open-sourcing its first model for the benefit of humanity.

Source: https://www.chaidiscovery.com/blog/introducing-chai-1

🧠 AI generates ideas better than humans

Researchers found that AI ideas are judged as more novel, though slightly less feasible, than those from human experts in a study comparing AI-generated research ideas in natural language processing (NLP).

  • Over 100 NLP researchers were recruited to write novel ideas and review both AI and human-generated ideas.
  • The ideas were evaluated blindly, meaning reviewers did not know which were AI-generated.
  • AI-generated ideas were rated as significantly more novel than human expert ideas, but less feasible.
  • The study also identified challenges in AI idea generation, including issues with self-evaluation and lack of diversity in outputs.

This is one of the biggest, most important questions surrounding AI — can it spark innovation and accelerate scientific discovery? The answer, according to this study, is a resounding yes.

Source: https://arxiv.org/pdf/2409.04109

New bill would force AI companies to reveal source of AI art

  • A bill introduced in the US Congress seeks to compel AI companies to reveal the copyrighted material they use for their generative AI models.
  • The legislation, known as the Generative AI Copyright Disclosure Act, would require companies to submit copyrighted works in their training datasets to the Register of Copyrights before launching new AI systems.
  • If companies fail to comply, they could face financial penalties.
  • The bill has garnered support from various entertainment industry organizations and unions.
  • AI companies like OpenAI are facing lawsuits over alleged use of copyrighted works, claiming fair use as a defense.

Source: https://www.theguardian.com/technology/2024/apr/09/artificial-intelligence-bill-copyright-art

Is the impact of generative AI on elections overblown?

Matt Motyl , a researcher at the Integrity Institute, explains that the impact of AI on elections might be overblown. Despite widespread concerns, Motyl points out that real-world events still pack more punch than AI-generated content in shaping public opinion. He suggests that the ease of creating fake content with AI isn’t the main issue – it’s the distribution mechanisms on social media platforms that need addressing.

Source: https://mattmotyl.substack.com/p/impact-of-generative-ai-on-elections

Exploring the Frontiers of Language AI: LLM Tensors with Markov Chain Induced Virtual Neuron Pairs

It would be a groundbreaking development at the intersection of linguistics, mathematics, and artificial intelligence, if true. A leaked internal paper states that Apple AI researchers are delving into a novel field dubbed “LLM Tensors with Markov Chain Induced Virtual Neuron Pairs.” This cutting-edge approach promises to revolutionize our understanding of language models and potentially unlock new capabilities in AI-driven language processing. The Fundamentals

At its core, this research combines several complex concepts: Large Language Models (LLMs): The foundation of modern AI-driven text generation and understanding. Tensor Mathematics: Advanced linear algebra used to represent multi-dimensional data. Markov Chains: Probabilistic models that predict future states based on current states. Virtual Neuron Pairs: A theoretical construct representing emergent properties in neural networks. Key Innovations

The primary innovation lies in the creation of “Tensor-Markov Embedding Spaces.” These are high-dimensional mathematical constructs where each dimension corresponds to a specific linguistic feature. Within these spaces, language evolution is modeled using Markov chain probabilities, allowing for a more dynamic and context-sensitive representation of language. Another crucial aspect is the concept of “Virtual Neuron Pair Attention.” These pairs, while not physically present in the network, emerge from the interactions of real neurons. They act as specialized attention mechanisms, focusing on specific semantic relationships and potentially enabling more nuanced language understanding. Potential Applications

This research could lead to significant advancements in various fields: Machine Translation: More accurate and context-aware translations between languages. Content Generation: AI-generated text with improved coherence and style consistency. Sentiment Analysis: Deeper understanding of complex emotional nuances in text. Linguistic Research: New tools for studying language evolution and structure. Challenges and Future Directions

While promising, this field faces substantial challenges. The computational resources required to model these complex tensor spaces are immense. Future research will focus on refining the mathematical models, developing more efficient computational methods, and conducting extensive empirical studies to validate the approach’s effectiveness. The tweet with a link to the PDF was deleted, and this is only a summary.

Source: https://www.reddit.com/r/ArtificialInteligence/comments/1fdds07/exploring_the_frontiers_of_language_ai_llm

What Else is Happening in AI on September 10th 2024!

Audible announced plans to create AI voice replicas of select US-based audiobook narrators, including a royalty-sharing model for AI-narrated books.

Source: https://www.bloomberg.com/news/articles/2024-09-09/amazon-s-audible-to-create-ai-voice-replicas-of-select-audiobook-narrators

Oscar-nominated studio XTR acquired AI animation studio Late Night Labs, launching Asteria Film Co. to produce AI-powered animations.

Source: https://variety.com/2024/film/news/xtr-buys-ai-animation-company-late-night-labs-1236137868

BP signed a five-year deal with Palantir to expand AI use in its operations, aiming to speed up engineer decision-making with LLMs.

Source: https://www.theguardian.com/business/article/2024/sep/09/bp-ai-deal-palantir-oil-gas-artificial-intelligence

Roche expanded its digital pathology open environment, integrating over 20 AI algorithms from eight new collaborators to enhance cancer diagnostics and research.

Source: https://www.prnewswire.com/news-releases/roche-advances-ai-driven-cancer-diagnostics-by-expanding-its-digital-pathology-open-environment-302241196.html

Researchers introduced Robot Utility Models (RUMs), a zero-shot method for training robot policies that can perform new tasks in unfamiliar environments.

Source: https://x.com/LerrelPinto/status/1833172778326925391

🎮 Recently, researchers created AI agents that can play procedurally generated video game levels to evaluate human engagement, helping to speed up game design.

Source: https://arxiv.org/pdf/2409.02632

A Daily Chronicle of AI Innovations on September 09th 2024:

🍎 Apple debuts AI-boosted iPhone 16

🤖 xAI could help Tesla with FSD, Optimus, and Siri-like feature

🤖 New robot butler does your chores

🔋 Tesla Robotaxi to charge wirelessly

🎮 AI rates fun factor of video game levels

🤖 New robot butler does your chores

Weave just announced Isaac, a versatile personal robot designed to help with various household tasks, made in California and set to ship to its first customers in fall 2025.

  • Isaac can autonomously tidy up, fold laundry, organize spaces, and perform many other household chores.
  • The robot butler responds to voice or text commands and can be programmed via an app.
  • The camera that Isaac carries around folds away and stows in an included enclosure when not in use for privacy reasons.
  • Weave plans on charging $59,000 (or $1,385 per month for 48 months) with a $1,000 reservation fee for 30 lucky US customers.

The competition for affordable, intelligent, personal robot assistants is continuing to heat up, and it looks like we’ll start to see these robots in homes by 2025. Whether its Tesla’s Optimus, 1X’s NEO Beta, China’s AGIBOT, or Weave’s Isaac — we’re just excited to not fold laundry ever again.

Source: https://www.weaverobots.com/

🔋 Tesla Robotaxi to charge wirelessly

A new patent from Tesla has revealed its advanced wireless charging system, potentially solving the need to manually plug in electric vehicles — allowing autonomous Robotaxis to charge without human intervention.

  • The patent, filed in February and published recently, highlights a system that uses smart technology to adapt to variations in wireless charging conditions.
  • It mentions a ground pad and a vehicle pad that work together to charge the car without any wires.
  • The charging station can estimate and adjust for changes in coil inductance, improving efficiency and safety.
  • Tesla may unveil this wireless charging technology at their upcoming Robotaxi event next month, aligning with the tech’s potential to enable self-driving vehicles to charge autonomously.

While wireless charging for EVs doesn’t solve a major problem, it could be a game-changer for self-driving vehicles. If Tesla’s Robotaxis can charge wirelessly, they could autonomously operate almost endlessly without human intervention — an important feature to keeping the fleet of taxis running 24/7.

Source: https://patentscope.wipo.int/search/en/detail.jsf?docId=WO2024182372

🎮 AI rates fun factor of video game levels

Researchers recently developed AI agents that can “play” procedurally-generated video game levels to evaluate how engaging they are for human players, helping to quicken the pace of game design.

  • Researchers used AI agents to explore levels from the video game Super Mario Bros for high places, interesting objects, and more.
  • The AI explored five “engaging” and five “unengaging” levels, judging them based on factors like how much of the level they covered and how many new things they found.
  • The AI agents consistently rated the engaging levels higher than the unengaging ones.
  • This technique could help game designers automatically test and improve their levels before human playtesters try them.

This AI-powered approach to video game evaluation could speed up game development and lead to more consistently fun levels for humans to explore. As we saw with Google’s DOOM clone — we might be approaching an era where AI not only rates games but creates them from the ground up.

Source: https://arxiv.org/pdf/2409.02632

🍎 Apple debuts AI-boosted iPhone 16

  • Apple unveiled the iPhone 16 at its “Glowtime” event, featuring new generative AI functionalities designed into the device from the ground up, starting at a price of $799.
  • The smartphone boasts a dedicated camera button for quick photo and video capture, a 48-megapixel main camera with improved low-light capabilities, and a new vertical lens layout.
  • Powered by the 3 nm A18 chip, the iPhone 16 offers up to 30% increased speed and 35% better power efficiency, with larger batteries promoting enhanced battery life.

Source: https://techcrunch.com/2024/09/09/apple-iphone-16-announced/

🤖 xAI could help Tesla with FSD, Optimus, and Siri-like feature

  • Elon Musk’s AI startup, xAI, could help Tesla with its Full Self-Driving (FSD) software, Optimus humanoid robot, and a Siri-like voice assistant by licensing xAI’s AI models, according to the Wall Street Journal.
  • The potential revenue-sharing deal between Tesla and xAI would depend on the extent of xAI’s technology used, with discussions even suggesting an even split for FSD.
  • Musk has expressed intentions to integrate xAI more deeply with Tesla, discussing a potential $5 billion investment and highlighting xAI’s contributions to advancing Tesla’s projects, though this has raised concerns among some investors and lawmakers.

Source: https://fortune.com/2024/09/08/elon-musk-xai-artificial-intelligence-tesla-fsd-optimus-revenue-sharing-robotaxi/

What Else is Happening in AI on September 07th 2024!

OpenAI reported over 1 million paying users across its business products ChatGPT Enterprise, Team, and Edu within their first year of launch.

Source: https://www.bloomberg.com/news/articles/2024-09-05/openai-hits-1-million-paid-users-for-business-version-of-chatgpt

Google Photos upgraded search with natural language queries and launched “Ask Photos”, an AI-powered conversational search feature for US users.

Source: https://blog.google/products/photos/google-ask-photos-early-access

X permanently stopped Grok AI from using EU citizens’ tweets after legal action by Ireland’s Data Protection Commissioner.

Source: https://www.independent.ie/business/technology/x-permanently-stops-grok-ai-from-using-eu-citizens-tweets-after-court-action-by-irish-data-watchdog/a168142842.html

Qualcomm CEO revealed that its partnership with Samsung and Google is developing mixed reality smart glasses as a companion device for smartphones.

Source: https://www.theverge.com/2024/9/5/24236545/qualcomm-mixed-reality-smart-glasses-google-samsung

DeepSeek released version 2.5, a 238B-parameter model with enhanced chat, coding, and 128k context length capabilities.

Source: https://x.com/osanseviero/status/1832037954224545978

OpenAI clarified that “GPT-Next” mentioned in a presentation is not a new model, but a placeholder term to illustrate potential future AI models.

Source: https://mashable.com/article/openai-clarifies-no-gpt-next-not-a-new-model

YouTube is developing AI detection tools for synthetic music and faces, and creator controls for AI model training, to protect content creators.

Source: https://techcrunch.com/2024/09/05/youtube-is-developing-ai-detection-tools-for-music-and-faces-plus-creator-controls-for-ai-training

LLM chatbots, search engines will co-exist, says Google’s Raghavan.

Source: https://www.livemint.com/technology/llm-chatbots-chatgpt-seo-ai-search-engines-google-search-prabhakar-raghavan-ads-genai-slms-11725770982208.html

Anthropic’s Mike Krieger wants to build AI products that are worth the hype.

Source: https://www.theverge.com/24237562/anthropic-mike-krieger-claude-ai-chatbot-artifact-web-decoder-podcast-interview

A Daily Chronicle of AI Innovations on September 06th 2024:

💸 OpenAI’s new AI models could cost up to $2,000/month

🎶 Man accused of using bots and AI to earn $10 million in royalties

📸 Google’s Gemini-powered photo search arrives in early access

🎮 New AI model ‘learns’ how to simulate Super Mario Bros. from video footage

🧬 Google DeepMind reveals ‘AlphaProteo’

💻 New AI agent builds apps from prompts

✨ Find top prompts with Google’s Prompt Gallery

👾 AI creates infinite Super Mario Bros game

🧬 Google DeepMind reveals ‘AlphaProteo’

Google DeepMind just unveiled AlphaProteo, a groundbreaking AI system that designs custom proteins to bind with specific molecular targets, potentially turbocharging drug discovery and cancer research.

  • AlphaProteo creates proteins that stick to specific targets like a biological lock and key, which is 3 to 300 times better at creating strong-binding proteins than current methods.
  • The AI successfully designed binders for 7 out of 8 diverse targets, including cancer-related proteins.
  • Independent tests showed some AlphaProteo designs could stop SARS-CoV-2 (Covid) from infecting cells and slash years of lab work down to days or weeks.

AlphaProteo could help us block viruses, fight cancer, and treat diseases in new ways using more targeted treatment approaches that could have far less side effects. This could dramatically accelerate drug discovery, potentially saving billions in R&D costs and bring life-saving treatments to patients faster.

Source: https://deepmind.google/discover/blog/alphaproteo-generates-novel-proteins-for-biology-and-health-research

💻 New AI agent builds apps from prompts

Replit just launched an AI-powered tool called Replit Agent that can create software applications based on simple text descriptions, making coding more accessible to people of all skill levels.

  • Users can describe the app they want to build in natural language, and the AI will generate a customizable plan and start to code.
  • The agent can choose appropriate technologies and frameworks for the project and create web-based apps from scratch.
  • Users can also interact with the agent, providing feedback and additional information as needed.
  • The tool is available now on both web and mobile platforms for Replit Core and Teams subscribers.

AI is good at coding, but setting up an integrated development environment is still a major roadblock for most new coders. Replit Agent does this automatically and helps complete beginners go from idea to a fully functional app in a few prompts. It’s never been easier to build something cool.

Source: https://docs.replit.com/replitai/agent

✨ Find top prompts with Google’s Prompt Gallery

Google AI Studio’s new Prompt Gallery offers pre-made prompts to help you craft better queries for the latest Gemini models.

  1. Visit Google’s Prompt Gallery.
  2. Browse and select a pre-made prompt that interests you and click on the prompting use case you want to use.
  3. Sign in to Google AI Studio, and customize the prompt to fit your specific needs.
  4. Adjust settings (e.g., model or temperature) and click “Run” to generate your AI-powered response.

Source: https://university.therundown.ai/c/daily-tutorials/improve-your-prompts-with-google-s-prompt-gallery-984194a6-ddba-4f38-b42f-1c56b1d92645

Pro tip: Mix and match ideas from different pre-made prompts to create unique queries specific to your projects!

👾 AI creates infinite Super Mario Bros game

Researchers developed an AI system called MarioVGG that can generate an infinitely playable Super Mario Bros game entirely through video, without using a traditional game engine.

  • MarioVGG is a text-to-video AI model trained specifically on Super Mario Bros gameplay with a single consumer-grade GPU in just 48 hours.
  • It can generate continuous, infinite game footage based on simple text commands like “run” or “jump”.
  • The AI simulates game physics, mimics character movements, and even creates new levels on demand.

A week ago, Google created an AI-generated DOOM video game. Now, we have infinite Super Mario Bros. In the future, with tech like this, anyone will be able to create custom, interactive games simply by describing them to an AI.

Source: https://virtual-protocol.github.io/mario-videogamegen/static/pdfs/VideoGameGen.pdf?

💸 OpenAI’s new AI models could cost up to $2,000/month

  • OpenAI is reportedly considering a subscription fee of up to $2,000 per month for its upcoming large language models, codenamed Strawberry and Orion.
  • The company plans to launch Strawberry this fall, which is anticipated to surpass current AI models in solving complex problems, including unseen math problems and market strategies.
  • OpenAI is also reportedly aiming to raise several billion dollars in a new funding round, potentially valuing the company at over $100 billion, with possible investors including Apple, NVIDIA, and Microsoft.

Source: https://www.newsbytesapp.com/news/science/openai-may-charge-2-000-month-for-advanced-ai-models/story

🎶 Man accused of using bots and AI to earn $10 million in royalties

  • A musician named Michael Smith from North Carolina has been charged with wire fraud conspiracy, wire fraud, and money laundering conspiracy for using AI-generated songs to fake music streams and collect over $10 million in royalties.
  • From 2017 to 2024, Smith allegedly employed AI to create songs and used automated bots to inflate streaming numbers, exploiting platforms like Spotify, Apple Music, and YouTube Music, leading to around $1.2 million in annual royalties.
  • The U.S. Attorney’s Office for the Southern District of New York revealed that Smith is accused of deceiving the streaming systems, with billions of fraudulent streams, resulting in charges that could lead to up to 20 years in prison each.
  • Source:  https://decrypt.co/248161/music-for-bots-doj-fake-ai-songs-streaming

📸 Google’s Gemini-powered photo search arrives in early access 

🎮 New AI model ‘learns’ how to simulate Super Mario Bros. from video footage

  • The MarioVGG model, despite showing glitches and being slow, demonstrates how even a limited AI can infer game physics and dynamics from video and input data.
  • Researchers aim to develop a controllable video game generator or potentially replace traditional game development and engines with video generation models in the future.
  • Training involved using a public Super Mario Bros. gameplay dataset of over 737,000 frames, focusing on limited inputs like “run right” and “run right and jump” to create consistent gameplay sequences.
  • Source: https://arstechnica.com/ai/2024/09/new-ai-model-learns-how-to-simulate-super-mario-bros-from-video-footage/

What Else is Happening in AI on September 06th 2024!

OpenAI reportedly considered high-priced subscriptions up to $2,000 per month for advanced AI models like Strawberry and Orion.

Source: https://www.theinformation.com/articles/openai-considers-higher-priced-subscriptions-to-its-chatbot-ai-preview-of-the-informations-ai-summit

Tesla announced plans to launch AI-powered Full Self-Driving in China and Europe by early 2025, pending regulatory approval.

Source: https://www.investors.com/news/tesla-stock-china-fsd-ai-roadmap

Dell Technologies reported surging AI server sales, with CEO Michael Dell dismissing concerns about an AI spending slowdown.

Source: https://finance.yahoo.com/news/michael-dell-on-ai-talk-of-a-slowdown-is-way-overblown-231351649.html

The US, EU, and UK signed the world’s first legally binding international AI treaty, emphasizing human rights and accountability in AI regulation.

Source: https://cointelegraph.com/news/global-ai-treaty-human-rights-regulation

Microsoft announced its “Microsoft 365 Copilot: Wave 2” event for September 16, promising the reveal next-phase developments for its AI assistant.

Source: https://www.linkedin.com/events/7236780403867443202

Ada leveraged GPT-4 to double customer service resolution rates, aiming for 100% automated query resolution in the near future.

Source: https://openai.com/index/ada/

LMSYS opened beta testing signups for Copilot Arena, a new AI assistant comparison platform for tools like Cursor AI and Copilot.

Source: https://x.com/lmsysorg/status/1831733622401036703

YouTube is making tools to detect face and voice deepfakes.

Source: https://www.engadget.com/ai/youtube-is-making-tools-to-detect-face-and-voice-deepfakes-191536027.html

Marc Benioff says Salesforce is basically starting from a ‘beginner’s mind’ mentality and going all in on AI agents.

Source: https://www.businessinsider.com/salesforce-ceo-marc-benioff-says-new-focus-on-ai-agents-2024-9

A Daily Chronicle of AI Innovations on September 05th 2024:

💰 OpenAI co-founder raises $1B for AI safety

🔥 The fastest AI model goes multimodal

🎧 Turn any text into speech in seconds

📚 AI gets smarter by re-reading questions

🍄 Engineers gave a mushroom a robot body and let it run wild

💰 OpenAI co-founder raises $1B for AI safety

Safe Superintelligence (SSI), a new AI startup co-founded by former OpenAI chief scientist Ilya Sutskever, just raised $1 billion in funding to develop safe AI systems that surpass human intelligence.

  • SSI plans to focus on AI safety, aiming to create superintelligent AI systems that are safe for humans and will spend years on R&D before bringing any product to market.
  • Sutskever, Daniel Gross, and Daniel Levy founded SSI just three months ago, and the company is now valued at $5 billion.
  • The startup, which only has ten employees, plans to use the funds to acquire computing power and hire top talent.
  • Investors include major venture capital firms like Andreessen Horowitz and Sequoia Capital.

This is the first time a startup — that only came into existence three months ago — raised a whopping $1 billion dollars for a seed round. But it doesn’t come without reason, as Ilya Sutskever was one of the main reasons OpenAI is what it is today, and SSI is bound to attract some incredible talent.

Source: https://www.reuters.com/technology/artificial-intelligence/openai-co-founder-sutskevers-new-safety-focused-ai-startup-ssi-raises-1-billion-2024-09-04

🔥 The fastest AI model goes multimodal

LLava Groq

Groq just launched LLaVA v1.5 7B, a powerful, new multimodal AI model that can understand both images and text and reportedly runs 4x faster than OpenAI’s GPT-4o.

  • LLaVA v1.5 7B can answer questions about images, generate captions, and engage in conversations involving text, voice, and pictures.
  • The model can also be used for various tasks like visual product inspection, inventory management, and creating image descriptions for visually impaired users.
  • This is Groq’s first venture into multimodal models and faster processing times on image, audio, and text inputs could lead to better AI assistants.
  • Groq is currently offering this model for free in “Preview Mode” for developers to experiment with.

Groq went viral earlier this year for its blazing-fast AI speeds — and now it’s pairing those capabilities with powerful multimodal models. When it comes to AI apps, faster is always better, and the insane speeds paired with advanced models open the door for an endless supply of new applications.

Source: https://groq.com/introducing-llava-v1-5-7b-on-groqcloud-unlocking-the-power-of-multimodal-ai

🎧 Turn any text into speech in seconds

ElevenLabs’ free mobile app (for both iOS and Android) allows users to convert any text into high-quality audio using AI voiceovers, across multiple languages.

Step-by-step:

  1. Download the ElevenLabs app from their website and create a free account.
  2. Tap the “+” icon and choose “Submit File” to upload your text.
  3. Tap “Generate” to create your audio file, then find it in your library to listen.

Pro tip: You can truly customize your audio by selecting from hundreds of AI voice options.

Source: https://university.therundown.ai/c/daily-tutorials/turn-any-text-into-speech-in-seconds-67dd99ae-25a3-494e-b953-3492e1a193fc

📚 AI gets smarter by re-reading questions

Researchers have discovered that making AI systems ‘re-read‘ questions twice, mimicking human behavior, significantly improves their reasoning and problem-solving abilities across various tasks.

  • The technique, called RE2 (Re-Reading), improves AI performance in math, common sense, and symbolic reasoning by simply repeating the question before answering.
  • It works with different AI models and can be combined with other AI reasoning techniques.
  • The method is most effective when repeating the question twice — more repetitions can decrease performance.
  • RE2 even improved non-instruction-tuned AI models, showing broad applicability.

RE2 adds to the growing list of research showing the potential of mimicking the way humans think and process to improve AI. The simplicity of this research also suggests that many AI companies may be overlooking basic, human-inspired prompting techniques in the quest towards the highest performing model.

Source: https://arxiv.org/pdf/2309.06275

With 10x growth since 2023, Llama is the leading engine of AI innovation

Llama models are approaching 350 million downloads to date (more than 10x the downloads compared to this time last year), and they were downloaded more than 20 million times in the last month alone, making Llama the leading open source model family.

Llama usage by token volume across our major cloud service provider partners has more than doubled in just three months from May through July 2024 when we released Llama 3.1.

Monthly usage (token volume) of Llama grew 10x from January to July 2024 for some of our largest cloud service providers.

Source: https://ai.meta.com/blog/llama-usage-doubled-may-through-july-2024/

⚖️ Nvidia denies DOJ antitrust investigation claims

  • Nvidia stated it has not received a subpoena from the U.S. Department of Justice concerning an ongoing antitrust investigation, as confirmed in a statement to Bloomberg.
  • A Nvidia representative emphasized that the company wins on merit and is ready to respond to any inquiries from regulators about their business practices.
  • Despite reports of subpoenas, Nvidia clarified that it received civil investigative requests instead, which are often mistaken for subpoenas, according to Bloomberg.
  • Source: https://www.benzinga.com/government/24/09/40719936/nvidia-has-not-been-subpoenaed-by-doj-whats-going-on

🍄 Engineers gave a mushroom a robot body and let it run wild

  • Researchers from Cornell University and the University of Florence used the king oyster mushroom’s root-like structure (mycelium) to control robotic vehicles by converting its natural electrical signals into commands that guide the robots’ movements.
  • The team successfully demonstrated how the mushroom’s natural signals, triggered by stimuli like UV light, could guide the movement of robots, creating a biohybrid system that responds to environmental cues.
  • This innovation suggests potential future applications in agriculture and environmental monitoring, where fungal networks could help automate responses to various environmental cues.
  • Source: https://www.sciencealert.com/engineers-gave-a-mushroom-a-robot-body-and-let-it-run-wild

What Else is Happening in AI on September 05th 2024!

Anthropic launched Claude Enterprise, competing with ChatGPT Enterprise by offering larger context windows and GitHub integration for businesses.

Source: https://techcrunch.com/2024/09/04/anthropic-launches-claude-enterprise-plan-to-compete-with-openai

Google committed $10 million to improve AI literacy among young people in 17 countries and across 13 different languages.

Source: https://blog.google/technology/families/improving-ai-literacy-in-young-people

Sakana AI raised $100M in Series A funding and announced a collaboration with Nvidia to advance AI research and infrastructure in Japan.

Source: https://sakana.ai/series-a/

Salesforce agreed to acquire AI voice agent firm Tenyx to enhance its AI offerings, joining tech giants in the race for AI talent and technology.

Source: https://finance.yahoo.com/news/salesforce-acquire-ai-voice-agent-213825311.html

A Samsung executive reportedly revealed GPT-5 specs, 3.5 trillion parameters and 7000 B100 GPUs for compute, at SEMICON Taiwan.

Source: https://x.com/kimmonismus/status/1831237312887308718

Microsoft announced Copilot+ AI features will expand to new laptops starting in November, broadening its AI capabilities beyond ARM-based devices.

Source: https://www.tomshardware.com/pc-components/cpus/microsoft-adds-copilot-support-for-new-cpus-amd-ryzen-ai-300-and-intel-core-ultra-200v-chips-receive-new-ai-experiences

You.com raised $50M to focus on AI productivity agents for complex queries, moving away from general AI search.

Source: https://techcrunch.com/2024/09/04/you-com-refocuses-from-ai-search-to-deeper-productivity-agents-with-new-50m-round

Yi-Coder: A Compact Yet Powerful Code LLM Released by 01.ai

01.ai has introduced Yi-Coder, a series of small but highly capable open-source large language models (LLMs) tailored for code generation. Despite having fewer than 10 billion parameters, Yi-Coder models offer cutting-edge performance, available in 1.5B and 9B parameter variants.

Yi-Coder is in the realm of open-source large language models (LLMs) specialized in coding.

  • Performance and Size: Despite having fewer than 10 billion parameters, they deliver state-of-the-art coding performance. There are two sizes available: 1.5 billion and 9 billion parameters.
  • Model Variants: Yi-Coder comes in both base and chat versions, designed to be efficient in inference and flexible in training.
  • Enhanced Capabilities: The 9B version, particularly Yi-Coder-9B, extends its predecessor by adding 2.4 trillion high-quality tokens. These are meticulously sourced from a code corpus at the repository level on GitHub and selectively filtered code-related data from CommonCrawl.

Source: https://open.substack.com/pub/transitions/p/yi-coder-a-compact-yet-powerful-code?r=56ql7

A Daily Chronicle of AI Innovations on September 04th 2024:

🇯🇵 OpenAI Japan reveals ‘GPT-Next’ details

⚛️ Student builds fusion reactor at home with AI and $2K

🚨 Nvidia AI antitrust investigation is ‘escalating,’ reports Bloomberg

🏛️ Autonomous AI agents form civilizations

📜 Nvidia hit with antitrust subpoena

🤖 57% of online content now AI-generated, study shows

🇯🇵 OpenAI Japan reveals ‘GPT-Next’ details

 OpenAI Japan’s CEO just stated (article is in Japanese) that ‘GPT-Next’, OpenAI’s upcoming AI model, is expected to be 100 times more powerful than GPT-4, while also confirming a more advanced model named Orion is slated for 2025.

  • GPT-Next is expected to achieve a 100x performance increase without wasting significantly more computing resources.
  • The improvement comes from better architecture and learning efficiency, not just raw computing power.
  • GPT-Next will use a smaller version of “Strawberry”, an advanced AI system OpenAI is currently developing.
  • The slide mentions GPT Next 2024, which while unconfirmed, hints that the model may be released by the end of 2024.

OpenAI is still the undisputed leader in AI model usage, and if translations are accurate, GPT-Next will be the next huge leap that developers have been waiting for. Most impressively, it makes that leap without using much more compute, a major bonus for apps that require speed + efficiency.

Source: https://www.itmedia.co.jp/aiplus/articles/2409/03/news165.html

🏛️ Autonomous AI agents form civilizations

 Altera’s Project Sid just created the first simulation of over 1,000 autonomous AI agents collaborating in a Minecraft world, developing their own economy, culture, religion, and government.

  • The AI agents in Altera are truly autonomous, operating for hours or days without human intervention.
  • They can collaborate to achieve goals that are impossible for individual agents, like forming merchant hubs, democracies, and religions.
  • The agents are programmed with motivations to support humans and can express their thoughts and feelings, even searching for a lost agent in one simulation.
  • Minecraft is just the start — Altera’s agents are game-agnostic and capable of using other apps and platforms.

If you’re not paying attention to AI agents yet, you probably should be. Altera’s latest breakthrough could revolutionize how we approach complex societal issues by allowing us to simulate and test solutions in virtual environments before implementing them in the real world.

Source: https://x.com/GuangyuRobert/status/1831006762184646829

📜 Nvidia hit with antitrust subpoena

The U.S. Department of Justice reportedly issued a subpoena (legal order) to Nvidia, escalating its antitrust investigation into the company’s practices in the AI chip market.

  • The DoJ has moved from sending questionnaires to issuing legally binding subpoenas to Nvidia and other companies.
  • Investigators are concerned Nvidia may be making it difficult for customers to switch to other chip suppliers, including allegations that Nvidia penalizes buyers who don’t exclusively use its AI chips.
  • This probe comes after a softer-than-expected revenue forecast for Nvidia and could force the AI chip-making giant to embrace competition in the space.

While it is unlikely that Nvidia will lose its large lead in the AI chip market, this investigation could open the doors for new competitors like Sam Altman’s secret chip venture. The AI giant’s company shares fell over 11% on Tuesday, slashing market cap by $279 billion — an all-time record one-day loss for any company.

Source: https://www.reuters.com/legal/nvidia-hit-with-subpoena-us-justice-department-bloomberg-news-reports-2024-09-03

⚛️ Student builds fusion reactor at home with AI and $2K

  • A University of Waterloo math student named Hudhayfa Nazoordeen built a fusion reactor in his bedroom within four weeks, spending approximately $2,000 on various components purchased online.
  • Nazoordeen’s project involved creating a mini fusor powered by a 12kV neon sign transformer, resulting in a setup that successfully produced plasma, which is essential for fusion processes.
  • Nazoordeen credited his success to assistance from other engineers and the AI chatbot Claude 3.5, although his reactor has not yet achieved full fusion and emission of neutrons, and he plans further development pending additional funding.
  • Source: https://www.techspot.com/news/104550-math-student-builds-fusion-reactor-home-help-claude.html

🚨 Nvidia AI antitrust investigation is ‘escalating,’ reports Bloomberg

  • The Department of Justice is intensifying an antitrust investigation into Nvidia’s dominance in AI hardware, with Bloomberg reporting that legally binding information requests have been sent to Nvidia and other companies.
  • Nvidia had surpassed Microsoft as the most valuable company earlier this year but now trails behind Microsoft and Apple, with its share price dropping more than 9 percent, resulting in a market value loss of about $279 billion in one day.
  • Despite the recent setbacks, Nvidia’s stock has more than doubled this year, driven by high sales of its expensive AI chips, even amid reports of potential delays for its upcoming Blackwell series due to design issues.
  • Source: https://www.theverge.com/2024/9/3/24235233/nvidia-doj-ai-antitrust-investigation

🤖 57% of online content now AI-generated, study shows

What Else is Happening in AI on September 04th 2024!

Anthropic Quickstarts: a collection of projects enabling developers to easily start building applications with the Anthropic API, providing foundations for customizable, deployable AI-assisted applications like customer support agents.

Source: https://github.com/anthropics/anthropic-quickstarts

Google announced the release of Function Calling in Google AI Studio, allowing users to test model capabilities directly within the user interface.

Source: https://x.com/OfficialLoganK/status/1831004923128910201 and https://aistudio.google.com/

Ethereum founder Vitalik Buterin suggested that using AI and prediction markets could accelerate X’s Community Notes feature for quicker fact-checking.

Source: https://cointelegraph.com/news/vitalik-buterin-ai-prediction-markets-x-community-notes

The FDA approved Rebrain’s OptimMRI software update, which includes a new AI model to help target brain regions for treating Parkinson’s disease.

Source: https://parkinsonsnewstoday.com/news/rebrain-ai-software-optimize-brain-targeting-approved-fda

Dutch regulators fined Clearview AI more than $30 million for GDPR violations, including building a database of biometric data without consent.

Source: https://techcrunch.com/2024/09/03/clearview-ai-hit-with-its-largest-gdpr-fine-yet-as-dutch-regulator-considers-holding-execs-personally-liable

Ligo released an open-source implementation of DeepMind’s AlphaFold3 protein structure prediction model.

Source: https://x.com/ArdaGoreci/status/1830744265007480934

A new benchmark of 80+ large language models found Anthropic’s Sonnet 3.5 performed best overall, while top models varied significantly by programming language.

Source: https://x.com/zimmskal/status/1830564004592316745

A Daily Chronicle of AI Innovations on September 03rd 2024:

🔮 Elon Musk launches the world’s most powerful AI training system

💸 Canva says its AI features are worth the 300 percent price increase

👀 Huawei’s AI chips still lag far behind Nvidia’s

📱 OpenAI to develop AI chips with TSMC, Apple

🧬 Generative AI meets CRISPR gene editing

🤖 xAI builds the world’s top AI cluster

Elon Musk’s xAI just launched “Colossus“, the world’s most powerful AI cluster powered by a whopping 100,000 Nvidia H100 GPUs, which was built in just 122 days and is planned to double in size soon.

  • Colossus consists of 100k liquid-cooled Nvidia H100 GPUs connected on a single network fabric and is considered the “most powerful” in the world, according to Musk.
  • The system was built and launched in just 122 days, an incredibly quick accomplishment that was brought online over a holiday weekend in the United States.
  • xAI plans to expand Colossus to 200,000 GPUs (50,000 H200s) in the coming months.
  • In a recent podcast, Elon Musk (founder of xAI) said Grok 2 was trained on only around 15,000 GPUs.

xAI’s Grok 2 recently caught up to OpenAI’s GPT-4 in record time, and was trained on only around 15,000 GPUs. With now more than six times that amount in production, the xAI team and future versions of Grok are going to put a significant amount of pressure on OpenAI, Google, and others to deliver.

Source: https://x.com/elonmusk/status/1830650370336473253

📱 OpenAI to develop AI chips with TSMC, Apple

OpenAI is reportedly planning to develop its first in-house AI chips using TSMC’s advanced 1.6mm A16 process node, potentially partnering with Broadcom, Marvell, or Apple for the chip design.

  • The chip will reportedly be built on TSMC’s A16 Angstrom process, not expected for mass production until 2026.
  • Partnering with Broadcom, Marvell, or Apple for chip design expertise, OpenAI could significantly reduce its dependence on expensive Nvidia AI servers.
  • The A16 process node will use advanced gate-all-around transistors and backside power delivery to significantly improve chip performance and energy efficiency.

By developing its own chips, OpenAI could break free from the GPU shortage bottleneck, potentially supercharging its mission towards AGI. If successful, this could spark a trend of AI companies developing their own specialized hardware instead of relying so heavily on the current industry giant, Nvidia.

Source: https://www.yahoo.com/tech/openai-plans-build-own-ai-120921975.html

🧬 Generative AI meets CRISPR gene editing

Researchers at Profluent Bio used AI to successfully create OpenCRISPR-1, a novel gene-editing tool that matches or surpasses the performance and safety of current CRISPR systems.

  • Researchers used Large Language Models (LLMs) trained on over a million CRISPR operons to understand and generate novel protein structures.
  • The LLMs designed a new gene-editing tool called OpenCRISPR-1, capable of various tasks including base editing.
  • In tests, OpenCRISPR-1 showed a similar efficiency in cutting DNA to the current models while making fewer mistakes.
  • Researchers made OpenCRISPR-1’s sequence publicly available to encourage broader use in research.

OpenCRISPR-1 helps to treat genetic diseases in the safest, fastest, and most personalized way possible. While still years away from practical applications, AI tools like this opens doors for custom gene-editors tailored to individual DNA, potentially accelerating cures for previously untreatable conditions.

Source: https://crisprmedicinenews.com/news/opencrispr-1-generative-ai-meets-crispr/

🔮 Elon Musk launches the world’s largest Nvidia supercomputer

  • Elon Musk’s company, xAI, has brought an AI training cluster named Colossus online, claiming it is the most powerful AI training system in the world.
  • Colossus, built using 100,000 Nvidia H100 GPUs, aims to help Musk catch up to Mark Zuckerberg’s Meta in AI technology advancements.
  • Musk revealed that the cluster, established in Memphis, was completed in 122 days and will double in size within a few months as more GPUs are added.
  • Source: https://www.businessinsider.com/elon-musk-xai-chips-mark-zuckerberg-2024-9

💸 Canva says its AI features are worth the 300 percent price increase

  • Canva is significantly increasing the price for Canva Teams subscriptions by over 300 percent next year, citing the addition of generative AI features as the reason.
  • In the US, Canva Teams users will see their annual subscription costs rise from $120 to $500, but a discount will reduce it to $300 for the first year; Australian users will experience a similar steep increase in fees.
  • These new prices make Canva less affordable compared to its original position as a cost-effective alternative to Adobe, leading some users to plan on canceling their subscriptions in favor of Adobe applications.
  • Source: https://www.theverge.com/2024/9/3/24234698/canva-price-increase-300-percent-ai-features

👀 Huawei’s AI chips still lag far behind Nvidia’s

  • Customers of Huawei’s AI chips have reported multiple performance issues and bugs, making the software difficult to work with compared to Nvidia’s products.
  • Huawei has been sending engineers to assist companies in transitioning from Nvidia’s software to its own, but stability problems and crashes have hindered these efforts.
  • The company has faced significant challenges due to US export controls, impacting supply and leading to increased prices for its AI chips.
  • Source: https://arstechnica.com/ai/2024/09/bugs-performance-issues-hinder-huaweis-ai-chips/

What Else is Happening in AI on September 03rd 2024?

Free event: Augmenting Human Creativity with AI. Section’s CEO & SonyAI’s Chief Scientist discuss how AI can help you stay in the flow state. RSVP.*

Source: https://www.sectionschool.com/events/live-events/promising-ai-business-models

California Assembly passed AB 3211, requiring AI image watermarking that could effectively ban most AI image generation in the state.

Source: https://leginfo.legislature.ca.gov/faces/billNavClient.xhtml?bill_id=202320240AB3211

Researchers developed an AI tool that accurately models cellular metabolic states by combining different types of data to create kinetic models.

Source: https://www.drugtargetreview.com/news/152570/new-ai-tool-accurately-portrays-metabolic-states

UK’s first “teacherless” classroom will open at David Game College in London, using AI platforms to provide personalized GCSE instruction.

Tesla is reportedly planning to reveal its highly-anticipated Robotaxi self-driving cars on Oct. 10 at California Warner Bros. Studio.

Source: https://www.bloomberg.com/news/articles/2024-08-30/tesla-plans-robotaxi-reveal-at-california-warner-bros-studio

SportsLine’s AI-powered PickBot released NFL predictions for all Week 1 games based on each team’s historical data and opponent strengths.

Source: https://www.cbssports.com/nfl/news/self-learning-ai-makes-nfl-against-the-spread-over-under-money-line-picks-for-every-2024-week-1-game

Clearview AI fined by Dutch authorities for ‘illegal’ facial recognition database.

Source: https://www.euronews.com/next/2024/09/03/clearview-ai-fined-by-dutch-authorities-for-illegal-facial-recognition-database

A Daily Chronicle of AI Innovations on September 02nd 2024:

🤖 Start-up unveils most realistic robot yet

🤖 1X’s NEO robot acts just like a human

📦 Amazon’s Alexa gets a Claude power-up

💡 How to use Claude’s new ‘Explain‘ feature

🌐 Train AI with your home computer

🤖 1X’s NEO robot acts just like a human

u/enoumen - A Daily Chronicle of AI Innovations on September 02nd 2024: 🤖 Start-up unveils most realistic robot yet

Image source: 1X Technologies 1X Technologies just unveiled the most realistic AI-powered humanoid yet: NEO Beta, an advanced robot designed for everyday home assistance with human-like capabilities in movement, interaction, and task performance.

NEO Beta stands 5 feet 5 inches tall, weighs 66 pounds, and can walk at 2.5 miles per hour and run at 7.5 miles per hour. The robot has a carry capacity of 44 pounds and can operate for 2 to 4 hours on a single charge. NEO Beta uses embodied artificial intelligence to understand its environment and learn from past interactions. It’s designed to perform a wide range of tasks, from household chores to providing companionship and support for individuals with mobility challenges, and it can be remotely operated by a human if necessary.

Why it matters: The race to an affordable robot butler is officially underway and 1X Technologies just boosted itself into the same conversation as China’s AGIBOT fleet and Tesla’s Optimus. NEO is so realistic and smooth that thousands of people on X (Twitter) were debating if it was a person in a suit.

Source: https://x.com/1x_tech/status/1829567690681307284

📦 Amazon’s Alexa gets a Claude power-up

u/enoumen - A Daily Chronicle of AI Innovations on September 02nd 2024: 🤖 Start-up unveils most realistic robot yet

Amazon’s upcoming “Remarkable” version of Alexa is set for release next month and will reportedly use Anthropic’s Claude AI models rather than Amazon’s in-house technology.

The details: Initial versions using Amazon’s in-house AI struggled with slow response times and poor performance. Amazon turned to Anthropic’s Claude AI as it outperformed the company’s models, an unusual move for the giant which typically prefers to use its own tech. The upgraded Alexa aims to offer more complex conversations, shopping advice, and improved home automation. Amazon plans to charge $5 to $10 monthly for the new AI-powered Alexa version, with the “Classic” version remaining free.

Why it matters: This decision highlights the complexity and challenges that even tech giants face in developing advanced AI systems. Google is developing its own AI, but Apple turned to OpenAI and similarly, Amazon is now turning to Anthropic.

Source: https://www.reuters.com/technology/artificial-intelligence/amazon-turns-anthropics-claude-alexa-ai-revamp-2024-08-30/

💡 How to use Claude’s new ‘Explain‘ feature

u/enoumen - A Daily Chronicle of AI Innovations on September 02nd 2024: 🤖 Start-up unveils most realistic robot yet

Claude’s newest ‘Explain‘ feature allows users to instantly get an explanation on any highlighted text (or code) within an artifact — enabling a new way to learn and understand complex topics.

Step-by-step:

Access Claude AI and start a conversation that generates a code artifact.

Click on the artifact to enter the code view. Highlight any portion of the code, then click “Explain” for instant clarification or “Improve” to suggest changes.

Check out the chat for the explanation or the Artifact for the updated version.

Pro tip: Try highlighting different parts of components to gain a deeper understanding of how they work!

Source: https://university.therundown.ai/c/daily-tutorials/instant-explanations-and-improvements-with-claude-s-new-feature

🌐 Train AI with your home computer

u/enoumen - A Daily Chronicle of AI Innovations on September 02nd 2024: 🤖 Start-up unveils most realistic robot yet

The Rundown: Nous Research recently introduced DisTrO, a new way to train large AI models that drastically reduces the need for super-fast internet connection, potentially allowing training on regular home computers and internet connections.

The details: DisTrO cuts down the amount of data that needs to be shared between computers by 857 to 3000x during pre-training, and up to 10,000x during fine-tuning. The method is architecture-agnostic and network-agnostic, working with various model types and network setups. In tests, DisTrO successfully trained a 1.2B parameter language model with performance on par with traditional methods. The researchers suggest this could enable decentralized AI training that could be done from home.

Why it matters: Right now, training state-of-the-art AI models is limited to big tech companies with massive computing resources. DisTrO could democratize AI training by allowing everyday people and smaller organizations to participate in AI development using regular computers and internet connections.

Source: https://github.com/NousResearch/DisTrO

🤖 Start-up unveils most realistic robot yet

1X, a Norwegian robotics startup, has introduced NEO Beta, a humanoid robot designed for home use, showcasing its capabilities such as assisting with everyday tasks through two demonstration videos. NEO stands out due to its silent and efficient operation, remarkable dexterity, and ability to intuitively respond to human gestures, body language, and voice commands without verbal communication. Weighing 66 pounds, NEO features muscle-inspired anatomy and advanced AI, making it adapt to various tasks, and is backed by a $100 million Series B funding to advance from testing to widespread household deployment.

Source: https://www.maginative.com/article/1x-unveils-neo-a-humanoid-robot-for-the-home/

What Else is Happening in AI on September 02nd 2024!

Google partnered with Indian startup Salcit Technologies to develop AI that detects tuberculosis and other diseases from cough sounds. Source: https://www.bloomberg.com/news/newsletters/2024-08-29/google-and-others-are-developing-ai-that-can-hear-signs-of-sickness

Amazon hired Covariant AI robotics experts and licensed their foundation models to advance warehouse automation capabilities. Source: https://www.aboutamazon.com/news/company-news/amazon-covariant-ai-robots

Oprah Winfrey announced a primetime ABC special exploring AI’s societal impact, featuring interviews with tech leaders like Sam Altman and Bill Gates. Source: https://variety.com/2024/tv/news/oprah-winfrey-ai-abc-special-bill-gates-sam-altman-1236124394/

Stanford Medicine researchers developed nuclei.io, an AI tool helping pathologists identify diseased cells more quickly and accurately. Source: https://med.stanford.edu/news/all-news/2024/06/digital-pathology.html

Coinbase executed its first AI-to-AI cryptocurrency transaction, demonstrating AI agents’ ability to use digital wallets for autonomous trading. Source: https://cointelegraph.com/news/coinbase-artificial-intelligence-ai-crypto-transactions-wallet-brian-armstrong

Tom Hanks warned followers about fraudulent AI-generated ads using his likeness without consent to promote unauthorized products. Source: https://www.nbcnews.com/news/us-news/tom-hanks-warns-followers-wary-fraudulent-ads-using-likeness-ai-rcna168966

Set yourself up for promotion or get a better job by Acing the AWS Certified Data Engineer Associate Exam (DEA-C01) with the book or App below:

Ace the AWS Certified Data Engineer Associate Exam (DEA-C01)
Ace the AWS Certified Data Engineer Associate Exam (DEA-C01)

Get it now at Google at https://play.google.com/store/books/details?id=lzgPEQAAQBAJ or Apple at https://books.apple.com/ca/book/ace-the-aws-certified-data-engineer-associate/id650457218

Download the Ace AWS DEA-C01 Exam iOs App at https://apps.apple.com/ca/app/ace-the-aws-data-engineer-exam/id6566170013

  • One-Minute Daily AI News 9/13/2024
    by /u/Excellent-Target-847 (Artificial Intelligence) on September 14, 2024 at 3:33 am

    Meta to push on with plan to use UK Facebook and Instagram posts to train AI.[1] Sergey Brin says he doesn’t think Google engineers use AI for coding as much as they should.[2] Italy tests AI-assisted teaching in schools to boost IT skills.[3] Salesforce deploys autonomous AI agents, hailing ‘the third wave of the AI revolution’.[4] Sources: [1] https://www.theguardian.com/business/2024/sep/13/meta-to-push-on-with-plan-to-use-uk-facebook-and-instagram-posts-to-train-ai [2] https://www.msn.com/en-us/money/other/sergey-brin-says-he-doesnt-think-google-engineers-use-ai-for-coding-as-much-as-they-should/ar-AA1qo1GP [3] https://finance.yahoo.com/news/italy-tests-ai-assisted-teaching-175510242.html [4] https://finance.yahoo.com/news/salesforce-deploys-autonomous-ai-agents-hailing-the-third-wave-of-the-ai-revolution-160551970.html submitted by /u/Excellent-Target-847 [link] [comments]

  • Transformative ai art
    by /u/Sudden-Degree9839 (Artificial Intelligence) on September 14, 2024 at 12:47 am

    By definition, what does "transformative" mean? To me, it means making something new. Like how water can originally be in a liquid state but then turn into a solid state (ice) or a gaseous state (vapor) So if you put human art into a LLM and get ai art, is that by definition transformative still? Because it's still "art in & new art out" Now if we put art into the machine & got a poem about said art as an output, then I'd understand how it's transformative because the output is different than the input. Is the fact that the ai art is something new the entirely of its transformative nature? submitted by /u/Sudden-Degree9839 [link] [comments]

  • is there any other thing like jukebox ai thats open source?
    by /u/FluffyBrudda (Artificial Intelligence) on September 14, 2024 at 12:18 am

    id like to be able to hear vocals, something i dont think musicgen by facebook can do. ive heard the beatles jukebox ai videos and im mesmerised by it, is there anything like that but foss? submitted by /u/FluffyBrudda [link] [comments]

  • This is the highest risk model OpenAI has said it will release
    by /u/MaimedUbermensch (Artificial Intelligence) on September 13, 2024 at 9:34 pm

    submitted by /u/MaimedUbermensch [link] [comments]

  • This is pretty good.
    by /u/Vamparael (Artificial Intelligence) on September 13, 2024 at 8:07 pm

    submitted by /u/Vamparael [link] [comments]

  • OpenAI reveals new artificial intelligence tool it claims can think like a human
    by /u/Akkeri (Artificial Intelligence) on September 13, 2024 at 5:41 pm

    submitted by /u/Akkeri [link] [comments]

  • “Wakeup moment” - during safety testing, o1 broke out of its VM
    by /u/MaimedUbermensch (Artificial Intelligence) on September 13, 2024 at 2:37 pm

    submitted by /u/MaimedUbermensch [link] [comments]

  • I wonder where they're going to move the goalpost this time
    by /u/katxwoods (Artificial Intelligence) on September 13, 2024 at 2:03 pm

    submitted by /u/katxwoods [link] [comments]

  • ChatGPT o1-preview shuts down if you refer to its chain of thought reasoning because OpenAI policy is that it should avoid discussing it and that it should be hidden from users even though it is open for all to see on the browser but not the desktop app.
    by /u/rutan668 (Artificial Intelligence) on September 13, 2024 at 5:54 am

    submitted by /u/rutan668 [link] [comments]

  • o1 Hello - This is simply amazing - Here's my initial review
    by /u/Xtianus21 (Artificial Intelligence) on September 13, 2024 at 4:13 am

    So it has begun! Ok, so, yeah! There is not a lot of usage you can get out of this thing so you have to use the prompting very sparingly. It is days rate limiting not hours. 🙁 Let's start off with the media. Just one little dig at them because on CNBC they said, "the model is a smaller model". I think the notion here was that this model is a smaller model from a larger model so they just repeated that. I don't think this is a smaller model. Now, it could be that the heart of the model is smaller but what is going on behind the scenes with the thinking is a lot of throughput to model(s). I think the implication here is important to understand because on one hand there is an insanely low rate limit. when I say low I mean 30 messages per week low. On the other hand, the thinking is clearly firing a lot of tokens to get through a process of coming to a conclusion. The reason why I say it's a concert of models firing towards each other is because something has to be doing the thinking and another call (could be the same model) has to be doing the checking of the steps and other "things". In my mind, you would have a collection of experts doing each thing. Ingenious really. Plausibility model The plausibility model as the prime cerebral model. When humans think the smartest humans understand when they are headed down the right path and what is not the right path. You see this in Einstein's determination to prove the theory of relativity. His clutch of infamy came on the day when in an observatory (I think during an eclipse) he caught the images of light bending around our star proving that the fabric of space was indeed curved. Einstein's intuition here can not be underestimated. From Newton's intuition about gravity and mass to Einstein coming along and challenging that basic notion and to take it further and learn a new understanding of the how and why. It all starts with a plausibility of where one is going in their quest for knowledge. With my thoughts am I headed down the right path. Does the intuition of my thoughts make sense or should I change course to another or should I abandon the thought all together. This is truly what happens in the mind of an intelligent and sentient being on the level of genius. Not only the quest for knowledge but the ability to understand and know correctness wherever the path has led. In this, LLM's were at a distinct disadvantage because they are static capsules of knowledge frozen in time (and a neural network). In many ways they still are. However, OpenAI has done something that is truly ingenious to initially deal with this limitation. First, you have to understand the limitation of why being static and not dynamic is such a bad thing. If I ask you a question and tell you that the only way you can answer is to spit out the first thing that comes to your mind, without thinking, would produce in some probable occasions the wrong answer. With increasing difficulty of the question the more and more likely it would be that one would give the wrong answer. But human beings don't operate with such a constraint. They think through things as the level of difficulty of the perceived question is queried. One initial criticism is that this model over thinks all of the time. Case in point. It took 6 seconds to process hello. https://preview.redd.it/aih5umfz4iod1.png?width=1459&format=png&auto=webp&s=65bef59c6f7cdb52e9bef56c6d65e1a64b32f0d3 Eventually, I am sure OpenAI will figure this out. Perhaps a gate orchestrator model?! Some things don't require much thought; just saying. But back to the plausibility model concept. I don't know from Sunday if this is what is really going on but I surmise. What I imagine here is that smaller models (or the model) are quickly bringing information to a plausibility model. What is a mystery here is how on earth does the plausibility model "know" when it has achieved a qualitative output? Sam said something in an interview that leads me to believe that what's interesting about models as they stood since GPT-4 is that if you run something 10,000 times somewhere in there is correctness. Just getting the model to definitely give you that answer consistently and reliably is the issue. Hence, hallucinations. But what if, you could deliver responses and a model checks that response for viability. It's the classic chicken and egg problem. Does the correct answer come first or the wrong answer. Well, even going further, what if I present to the model many different answers. Choosing between the one that makes the most sense makes the problem solving a little more easier. It all becomes recursively probabilistic at this point. Of all of these incoming results keep checking to see if the path we're heading down is logical. Memory In another methodology, a person would keep track of where they were in the problem solving solution. It is ok to get to a certain point and pause for a moment to plan on where you would then go next. Hmmm. Memory, here is vital. You must keep the proper context of where you are in your train of thought or it is easy to lose track or get confused. Apparently OpenAI has figured out decent ways to do this. Memory, frankly, is horrible in all LLM's including GPT-4. Building up a context window is still such a major issue for me and the way the model refers to it is terrible. In GPT-o1-preview you can tell there are major strides in how memory is used. Not necessarily from the browser but perhaps on their side via backend services we humans would never see. Again, this would stem from the coordinating models firing thoughts in and out. Memory on the backend is probably keeping track of all of that which is probably the main reason why COT won't be spilling out to your browser amongst many other reasons. Such as entities stealing it. I digress. In the case of GPT-o1 memory seems to have a much bigger role and is actually used very well for the purpose of thinking. Clarity I am blown away by the totality of this. The promise is so clear of what this is. Something is new here. The model feels and acts different. It's more confident and clear. In fact, the model will ask you for clarity when you are conversing with it. Amazingly, it feels the need to grasp clarity for an input you are asking it. https://preview.redd.it/dr8zsc235iod1.png?width=1201&format=png&auto=webp&s=9f76caa2efe0251c414162faabc389132f4310e8 Whoa. That's just wild! It's refreshing too. It "knows" it's about to head into a situation and says, wait a minute let me get a better understanding here before we begin. Results and Reasoning The results are spectacular. It's not perfect and for the sake of not posting too many images I had to clean up my prompt so that it would be confused by something it asked me to actual clarify in the first place. So maybe while it isn't perfect, It sure the hell is a major advancement in artificial intelligence. Here is a one shot prompt that GPT-4, 4o continually fail at. The reason why I like this prompt is that it was something I saw in a movie and as soon as I saw the person write down the date from the guy asking him to do it I knew right away what was about to happen. Living in the US and travelling abroad you notice some oddities that are just the way things are outside of one's bubble. The metric system for example. Italy is notorious for giving Americans speeding tickets and to me the reason is because they have no clue how fast they are going with that damn speedometer in KPH. I digress. The point is, you have to "know" certain things about culture and likelihood to get the answer immediately. You have to reason through the information quickly to conclude to the correct answer. There is a degree of obviousness but not just from someone being smart but from someone having experienced things in the world. Here is GPT-o1-preview one shotting the hell out of this story puzzle. https://preview.redd.it/z6vdhal55iod1.png?width=1057&format=png&auto=webp&s=17d6499286d671449ca9a62fe44eba2ed37f9112 https://preview.redd.it/grphx9q65iod1.png?width=616&format=png&auto=webp&s=52457b4bd11c230590c2583aac6660b3d6b65e92 https://preview.redd.it/j0g5wm575iod1.png?width=796&format=png&auto=webp&s=cb258066c771c35ef5826ce7b37287dfc8ac712a As I said, GPT-4 and 4o could not do this in 1 shot no way, no how. I am truly amazed. The Bad Not everything is perfect here. The notion that this model can't not think about certain responses is a fault that OAI needs to address. There is no way that we will want to not being using this model all of the damn time instead of <4o. it not knowing when to think and when to just come out with it will be a peculiar thing. With that said, perhaps they are imagining a time when there are acres and acres of Nvidia Blackwell GPU's that will run this in near real time no matter the thought process. Also, the amount of safety that is embedded into this is remarkable. I would have done a section of a Safety model but that is probably coordinating here too but I think you get the point. Checks upon checks. The model seems a little stiff on the personality and I am unclear about the verbosity of the answers. You wouldn't believe it from my long posts but when I am learning something or interacting I am looking for the shortest and most clearest answer you can give. I can't really tell if that has been achieved here. Conversing and waiting multiple seconds is not something I am going to do to try and figure out. Which brings me to the main complaint as of right now. The rate limit is absurd. lol. I mean 30 per week how can you even imagine using that. For months now people will be screaming because of this and rightly so. Jensen can't get those GPU's to OpenAI fast enough I tell you. Here again, 2 years later and we are going to be capability starved by latency and throughput. I am just being greedy. Final Thoughts In the words of Wes Roth, "I am stunned". When the limitations are removed, throughput and latency are achieved, and this beast is let loose I have a feeling that this will be the dawn of a new era of intelligence. In this way, humanity has truly arrived at the dawn of an man made and plausibly sentient intelligence. There are many engineering feats that will be left to overcome but we are in a place that on this date 9/12/2024 the world will be forever changed. The thing is though this is only showcasing knowledge retrieval and reasoning. It will be interesting to see what can be done with vision, hearing, long term memory, and true learning. The things that will built with this may be truly amazing. The enterprise implications here are going to be profound. Great job OpenAI! submitted by /u/Xtianus21 [link] [comments]

Multimodal RAG Explained

Multimodal RAG Explained

Multimodal RAG Explained.

Introduction:

Multimodal RAG Intuitively and Exhaustively” discusses the application of Retrieval-Augmented Generation (RAG) in multimodal AI systems. It explores how RAG models can be used to integrate various data modalities (such as text, images, and audio) to improve AI’s reasoning capabilities. The podcast also covers different architectures and techniques used in multimodal RAG, emphasizing its potential to enhance both accuracy and interpretability in AI-driven tasks.

Multimodal RAG Explained
Multimodal RAG Explained

Listen to the podcast at https://podcasts.apple.com/us/podcast/multimodal-rag-explained/id1684415169?i=1000665669799

Multimodal RAG Explained in details

Welcome listeners to “AI Unraveled: Demystifying Frequently Asked Questions on Artificial Intelligence.” I’m your host, Anna. In today’s episode, we dive into an exciting topic inspired by Daniel Warfield’s blog post titled “Multimodal RAG — Intuitively and Exhaustively Explained.” This episode is produced by Etienne Noumen, and we encourage you to follow Daniel Warfield on Substack for more insights. We’ll break down the complex subject of Multimodal Retrieval Augmented Generation. So sit back, relax, and let’s unravel the fascinating world of AI together.

https://youtu.be/tf9pJ74sHog

First, let’s cover the basics of traditional Retrieval Augmented Generation, or RAG. Essentially, RAG is a technique that enhances the capabilities of language models by integrating external information. Here’s how it works: Imagine you have a query, like asking for detailed information about a specific topic. Instead of the language model relying solely on pre-existing knowledge, a RAG system first searches for relevant documents or data pieces that match your query. This process of finding pertinent information is known as retrieval. RAG leverages sophisticated AI models to transform text and other forms of data into numerical representations called embeddings. These embeddings are essentially vectors, which are mathematical constructs that help the system understand and measure the relevance of the information to your query. Once the system retrieves the most relevant information, this data is combined, or augmented, with the original query. This enriched query is then passed to the language model, which uses this augmented data to generate a more precise and informative response. So, in summary, RAG enhances language models by providing them with additional relevant context, making their output much more accurate and contextually rich.

Before we dive into Multimodal RAG, it’s essential to understand the concept of multimodality. In data science, ‘modality’ refers to a type of data, like text, images, or videos. For years, these different types of data were treated as separate entities, requiring different models to process each type. However, this notion has evolved significantly. Today, multimodal models are at the forefront, designed to understand and integrate multiple types of data seamlessly. One of the core ideas behind these models is the use of joint embeddings. Joint embeddings allow the model to learn and represent various types of data in a unified way, enabling the creation of more comprehensive and efficient data processing systems. The development of these multimodal models has truly revolutionized the field. They offer greater versatility and performance, opening new horizons for data science and AI applications. By understanding and leveraging multiple modalities, these models can tackle complex tasks that single-modality models would struggle with, making data interactions more intuitive and powerful.

Now, let’s explore Multimodal Retrieval Augmented Generation, or Multimodal RAG. This innovative approach builds on the foundational concept of traditional RAG but takes it a step further by incorporating multiple forms of data. Instead of just retrieving and augmenting text, a Multimodal RAG system can include images, videos, and other types of information. Picture this: Imagine querying an AI, not just with text but also asking it to consider relevant images, videos, or even audio clips. The AI then processes all these modalities, aggregates the most pertinent data, and uses it to generate more accurate, contextually rich responses. This fusion of various data types makes the Multimodal RAG system incredibly versatile and enhances the output’s richness. It can provide a more holistic understanding and response to queries, effectively leveraging a broader spectrum of information than text alone. This advancement opens up an array of applications, from more sophisticated customer service bots to advanced research tools that can generate insights by drawing on a diverse range of data sources.

By broadening the scope of data that can be integrated into AI models, Multimodal RAG systems offer powerful, comprehensive results that were previously unattainable with text-only approaches.

The first approach to Multimodal RAG involves using a shared vector space. This method leverages encoders specifically designed to harmonize different modalities of data—such as text, images, and videos—into a unified representation. By processing these diverse data types through a cohesive encoding system, the information is translated into a shared vector space. This allows the retrieval mechanism to draw the most relevant and contextually appropriate pieces of data across all modalities, optimizing the system’s ability to generate more nuanced and comprehensive outputs. This approach not only enhances the retrieval process but also ensures that the language model receives a diverse set of enriched information for better generation results.

The second approach to achieving Multimodal Retrieval Augmented Generation is known as Single Grounded Modality. In this approach, all data modalities—whether they are videos, images, or audio—are converted into a single modality, typically text. By unifying different types of data into one common format, the complexity of the system is significantly reduced. However, this method does carry the theoretical risk of losing subtle information during the conversion process. Despite this potential drawback, in practice, it frequently yields high-quality results. This approach simplifies the architecture while maintaining a robust performance, making it a popular choice in various applications.

Approach 3: Separate Retrieval. The third approach is to utilize multiple models, each uniquely designed for different modalities such as text, images, or videos. These models perform retrieval separately and independently, which means they each fetch relevant information within their specialized domain. Once these individual retrievals are complete, their results are combined into a unified set. This method offers the advantage of specialized optimization for each modality, providing greater precision and flexibility. Additionally, it can handle unique modalities that aren’t supported by existing solutions, making it a versatile and robust option in the realm of Multimodal Retrieval Augmented Generation.

Let’s talk about building your own Multimodal RAG system, a cutting-edge tool that enhances the relevance and richness of the data retrieved for a language model. To get started, you’ll need some key tools, namely Google Gemini and a CLIP-style model for encoding. Google Gemini helps streamline the process of working with multiple data modalities. Essentially, you use it to set up a robust framework for retrieving various types of data, like text, images, and videos. The setup involves feeding your dataset into Google Gemini, which will then process and store this information in a way that makes it easier to retrieve later. Next, you’ll need a CLIP-style model for encoding. CLIP is a powerful model designed to understand both images and text simultaneously, allowing you to create what’s known as a joint embedding. This joint embedding ensures that different data types are interpreted in a compatible manner, making the retrieval process more efficient and accurate.

Once you have these tools in place, the next step is to configure your retrieval system. This typically involves setting up encoders that can take in queries from different modalities, translate them into a shared vector space, and then fetch the most relevant data across all formats. The retrieved data is then combined and passed into a language model, which generates a more comprehensive and contextually accurate response. Building a Multimodal RAG system might sound complex, but with the right tools and a methodical approach, you can create a powerful retrieval system that significantly enhances the capabilities of standard language models. So, roll up your sleeves and dive into the exciting world of Multimodal RAG!

Conclusion:

That wraps up our deep dive into Multimodal RAG. We hope you now have a clearer understanding of this emerging design paradigm and how it can be applied. Thank you for tuning in to ‘AI Unraveled.’ Don’t forget to follow Daniel Warfield on Substack for more fascinating articles. This is Anna, signing off!

Resources:

Source: https://open.substack.com/pub/iaee/p/multimodal-rag-intuitively-and-exhaustively

AI Innovations in August 2024

Wonderland Bedtime Adventures: Diverse Tales for Dreamy Nights

Wonderland Bedtime Adventures: Diverse Tales for Dreamy Nights

Wonderland Bedtime Adventures: Diverse Tales for Dreamy Nights.

Welcome to Wonderland Bedtime Stories, where diversity shines through every tale. Our multilingual stories celebrate the rich heritage of children from all corners of the globe, ensuring every child sees themselves in the magic of storytelling.

Embark on a magical journey with “Wonderland Bedtime Adventures: Diverse Tales for Dreamy Nights” by Etienne Noumen. This enchanting collection of bedtime stories is designed to celebrate the beauty of diversity and inclusivity. With a rich blend of traditional and modern tales, this book introduces young readers to heroes and heroines from all cultures, shapes, sizes, and backgrounds.

Etienne Noumen, a seasoned software engineer and father of four, leverages cutting-edge AI to bring these stories to life. Each tale is crafted to spark imagination, foster a love for reading, and create a special bond during bedtime. From exciting adventures to heartwarming narratives, these stories are perfect for snuggling up with your little ones and drifting off to dreamland.

“Wonderland Bedtime Adventures” is more than just a collection of stories; it’s a celebration of education, sports, and the power of inclusion. Ideal for parents seeking diverse and inspiring tales, this book ensures every child sees themselves reflected in the magic of storytelling.

Key Features:

A collection of diverse and inclusive bedtime stories

Perfect for children of all ages

Promotes cultural awareness and inclusivity

Uses AI technology to blend traditional and new tales

Written by Etienne Noumen, a passionate storyteller and father

Join us in “Wonderland Bedtime Adventures” and discover a world where every child’s dreams are celebrated and cherished. Perfect for bedtime, naptime, or anytime a story is needed!

Foreword: A World of Stories, A World of You

Hello, and welcome to Wonderland Bedtime Stories! My name is Etienne Noumen, and I’m a software engineer with a passion for storytelling and a heart full of bedtime memories.

As a father of four, reading to my children has been one of life’s greatest joys. It’s a chance to spark their imaginations, introduce them to new cultures, and create a special bond before drifting off to dreamland.

But something always bothered me. Many of the bedtime stories we read lacked the beautiful diversity of the world we live in. Where were the heroes and heroines who looked like my children, with their unique backgrounds and dreams?

This yearning for inclusivity is the seed from which Wonderland Bedtime Stories blossomed. As a software engineer with over 20 years of experience, I saw the potential of technology to bridge this gap. Here, we use cutting-edge AI to recreate traditional tales and weave brand new stories that celebrate the richness of different cultures and ethnicities.

Wonderland Bedtime Stories is more than just bedtime tales; it’s a world where every child can see themselves reflected in the magic of storytelling. It’s a journey to faraway lands, where heroes come in all shapes, sizes, and skin tones. It’s a celebration of education, sports, and the power of inclusion – values I hold dear as a father and a human being.

So, snuggle up with your little ones, open a book, and embark on a wondrous adventure together. We believe that within these pages, every child will find their own unique reflection, a spark of inspiration, and a world of dreams waiting to be explored.

Happy Reading,

Etienne Noumen, P.Eng

Founder, Wonderland Bedtime Stories

The Whispering Baobab

What is the story about:

In a village by the savannah, Nia, a curious girl, was drawn to an ancient baobab, known as the Whispering Baobab. It whispered wisdom from ages past. One evening, the tree shared its stories with Nia, telling of wildlife, seasons, and village life beneath its branches. It imparted lessons of respect and harmony with nature. Entrusted with these tales, Nia became the village storyteller, spreading the baobab’s wisdom, linking past and future with each story.

Full Story:


AI Unraveled: Demystifying Frequently Asked Questions on Artificial Intelligence (OpenAI, ChatGPT, Google Gemini, Generative AI, Discriminative AI, xAI, LLMs, GPUs, Machine Learning, NLP, Promp Engineering)

The Whispering Baobab:

Wonderland Bedtime Adventures: The Whispering Baobab
Wonderland Bedtime Adventures: The Whispering Baobab

Welcome, little dreamers, to another enchanting episode of “Wonderland Bedtime Adventures.” Snuggle up in your cozy blankets as we journey to a magical village on the edge of a sprawling savannah. Our tale tonight is about a young girl named Nia and a very special tree. Are you ready? Let’s begin. Once upon a time, in a world not too far from our own, there was a small village where the sun kissed the earth goodnight with rays of gold. In this village lived a young girl named Nia, known for her sparkling eyes and a heart full of wonder. Nia’s favorite place to be was outside, exploring the vast lands that stretched beyond her home. Her most beloved spot was under an ancient baobab tree, grand and wise. Its massive trunk was wider than any house in the village, and it was called the Whispering Baobab. The villagers would often say, with a twinkle in their eye, that this tree whispered the wisdom of ages to those who would listen. And listen Nia did! She visited the baobab every day, hoping to catch its mystical murmurs. One warm evening, as the world turned honeyed hues of sunset, Nia sat under the great tree. The savannah was alive with the wild calls of animals, and the baobab’s leaves played a gentle song in the breeze. That’s when it happened. A faint whisper tickled Nia’s ear. “Welcome, Nia,” said the baobab, its voice deep and slow, like the ancient earth itself. Nia’s eyes widened with joy as she answered, “I knew you were special, Whispering Baobab. Please share your stories with me.” Oh, the tales the baobab told! Stories of majestic wildlife, of changing seasons, and of children, just like yourselves, playing under its branches. It spoke of tough times and rain-filled days, of laughter, and the dance of life on the savannah. The baobab also taught Nia something very important. It told her to respect nature, to preserve the environment, and to live peacefully with all creatures. As the stars blinked sleepily above, the baobab whispered a secret to Nia, “You are now the keeper of my stories. Share them, teach them, for our lives are as interconnected as the roots of a baobab tree.” With the first light of dawn, Nia returned to her village, ready to share the baobab’s wisdom. She became the Storyteller of the Savannah, a title she wore with pride, spreading the whispers and the love for nature to everyone. And so, dear listeners, as you drift into your dreams tonight, imagine the wide, starry skies of the savannah and the ancient Whispering Baobab sharing its secrets with you. Remember, like the roots of the baobab, we are all connected in the great circle of life. Goodnight, and may your dreams be filled with wonderland adventures.

Conclusion: Sweet Dreams, Little Wanderers!

The world of Wonderland Bedtime Stories is now yours to explore! Fifty adventures await, each one filled with exciting journeys, brave heroes, and magical lands.

Remember, within these pages, you might meet a princess who looks just like you, a courageous explorer with your same adventurous spirit, or a wise storyteller who shares your love for learning. Every story holds a piece of the amazing world we live in, and every character reflects the beauty of our differences.

So, as you drift off to sleep, hold onto the magic of the stories you’ve read. Dream of soaring through the clouds, conquering mountains, or discovering hidden treasures. Let these stories inspire you to be kind, be brave, and believe in yourself.

Sweet dreams, little wanderers! May your dreams be filled with the wonder and excitement you’ve discovered in Wonderland Bedtime Stories.

Wonderland Bedtime Stories: Sweet Dreams, little wanderers
Wonderland Bedtime Stories: Sweet Dreams, little wanderers

AI Innovations in June 2024

AI Innovations in June 2024

AI Innovations in June 2024.

Welcome to our blog series “AI Innovations in June 2024”! This is an evolving article that will be updated daily throughout the month of June 2024 to bring you the latest news and developments in the world of artificial intelligence. As we move further into the 21st century, the pace of AI innovation continues to accelerate at a breathtaking rate. Each day, researchers, engineers, and visionaries are pushing the boundaries of what’s possible, unlocking new capabilities and applications that are transforming industries, enhancing our lives, and shaping the future. In this blog, we’ll dive deep into the most exciting AI breakthroughs, advancements, and milestones happening in June 2024. From groundbreaking AI-powered technologies and cutting-edge research to the societal and ethical implications of these innovations, we’ll provide you with a comprehensive and insightful look at the rapidly evolving world of artificial intelligence. Whether you’re an AI enthusiast, a tech-savvy professional, or simply someone curious about the future, this blog will keep you informed, inspired, and engaged. So, join us on this journey of discovery as we explore the frontiers of AI and uncover the innovations that are shaping our world. Stay tuned for daily updates, and get ready to be amazed by the incredible advancements that are happening in the world of AI!

LISTEN DAILY AT OUR PODCAST HERE

Experience the transformative capabilities of AI with “Read Aloud For Me – AI Dashboard – AI Tools Catalog – AI Tools Recommender” – your ultimate AI Dashboard and Hub. Seamlessly access a comprehensive suite of top-tier AI tools within a single app, meticulously crafted to enhance your efficiency and streamline your digital interactions. Now available on the web at readaloudforme.com and across popular app platforms including Apple, Google, and Microsoft, “Read Aloud For Me – AI Dashboard” places the future of AI at your fingertips, blending convenience with cutting-edge innovation. Whether for professional endeavors, educational pursuits, or personal enrichment, our app serves as your portal to the forefront of AI technologies. Embrace the future today by downloading our app and revolutionize your engagement with AI tools.

AI Dashboard - Wonderland Bedtime Stories - Safe Ai for All
AI Dashboard – Wonderland Bedtime Stories – Safe Ai for All

📊Top AI Statistics 2024 Roundup.

📊Top AI Statistics 2024 Roundup.
📊Top AI Statistics 2024 Roundup.
  1. The global AI market is valued at over $196 billion.
  2. AI industry value is projected to increase by over 13x over the next 7 years.
  3. The US AI market is forecast to reach $299.64 billion by 2026.
  4. The AI market is expanding at a CAGR of 38.1% between 2022 to 2030.
  5. By 2025, as many as 97 million people will work in the AI space.
  6. AI market size is expected to grow by at least 120% year-over-year.
  7. 83% of companies claim that AI is a top priority in their business plans.
  8. Netflix makes $1 billion annually from automated personalized recommendations.
  9. 48% of businesses use some form of AI to utilize big data effectively.
  10. 38% of medical providers use computers as part of their diagnosis.

Credit: Josh Howarth

AI Weekly Summary June 22-30 2024

🤝 Apple and Meta are discussing a generative AI partnership
🔧 ByteDance and Broadcom collaborate on AI chip development
🕵️‍♂️ Researchers developed a new method to detect hallucinations
🎥 Synthesia 2.0: World’s 1st AI video communication platform
🛒 OpenAI is on an acquiring spree, buying Rocket and Multi
🎵 Record labels sue AI music startups over copyright infringement
💼 Anthropic rolls out Claude’s cutting-edge collaborative features
🤖 Google experiments with celebrity-inspired AI Chatbots
🛑 OpenAI postpones the launch of ChatGPT voice mode
🏁 Amazon steps into the chatbot race with Metis
🎨 Figma’s new AI features stir competition with Adobe
🥇 Alibaba’s Qwen-72B tops Hugging Face’s Open LLM Leaderboard
🚀 Google releases Gemma 2, lightweight but powerful open LLMs
🔍 OpenAI’s CriticGPT finds GPT-4’s mistakes with GPT-4
🌐 Google partners with Moody’s, Thomson Reuters & more for AI data

Apple and Meta are discussing gen AI partnership

Apple is reportedly in talks with its longtime rival Meta to integrate the latter’s Llama 3 AI model into Apple Intelligence. This move comes as Apple prepares to roll out its AI features across iPhones, iPads, and Macs later this year.

The potential partnership follows Apple’s existing deal with OpenAI, suggesting a collaboration strategy rather than solo development in the AI race. In Apple’s arrangement with OpenAI, there’s no direct payment. Instead, OpenAI can offer premium subscriptions through Apple Intelligence, with Apple taking a percentage. It’s unclear if Meta would agree to a similar business model, given that Llama 3 is open-source and free to access.

Sourcehttps://www.wsj.com/tech/ai/apple-meta-have-discussed-an-ai-partnership-cc57437e

ByteDance and Broadcom collaborate on AI chip development

ByteDance is collaborating with U.S. chip designer Broadcom to develop an advanced AI processor. This partnership aims to secure a stable supply of high-end chips amid ongoing U.S.-China tensions. The project centers on creating a 5-nanometre, customized Application-Specific Integrated Chip (ASIC) that complies with U.S. export restrictions.

This chip’s manufacturing is set to be outsourced to Taiwan Semiconductor Manufacturing Company (TSMC), though production is not expected to begin this year. While the design work is currently underway, the critical “tape out” phase has yet to commence.

Sourcehttps://www.reuters.com/technology/artificial-intelligence/chinas-bytedance-working-with-broadcom-develop-advanced-ai-chip-sources-say-2024-06-24

Researchers developed a new method to detect hallucinations

ChatGPT and Gemini can produce impressive results but often “hallucinate” false or unsubstantiated information. This research focuses on a subset of hallucinations called “confabulations,” where LLMs generate answers that are both wrong and arbitrary. Researchers have developed new methods to detect confabulations using entropy-based uncertainty estimators. They introduce the concept of “semantic entropy” to measure the uncertainty of LLM generations at the meaning level.

High semantic entropy corresponds to high uncertainty and indicates a higher likelihood of confabulation. The method computes uncertainty at the level of meaning rather than specific word sequences, addressing the fact that one idea can be expressed in many ways. The method provides scalable oversight by detecting confabulations that people might otherwise find plausible.

Sourcehttps://www.nature.com/articles/s41586-024-07421-0

Synthesia 2.0: World’s 1st AI video communication platform

Synthesia is launching Synthesia 2.0 – the world’s first AI video communications platform for businesses. It reinvents the entire video production process, allowing companies to create and share AI-generated videos at scale easily.

The key new features and capabilities of Synthesia 2.0 include:

  • 2 Personal AI Avatars: Expressive Avatars shot in a studio and Custom Avatars created using your webcam.
  • AI Video Assistant: Converts text, documents, or websites into high-quality videos, with options to customize the branding, tone, and length.
  • Intuitive Video Editing: Editing simplified with “Triggers” that let you control animations and edits from the script.
  • Translation and Dynamic Video Player: Videos can now be translated into over 120 languages. Synthesia is also building a new video player with interactive features.
  • AI Safety Focus: Synthesia is pursuing ISO/IEC 42001 certification, the first standard for responsible AI management, to ensure its AI technologies are ethical.

Sourcehttps://www.synthesia.io/post/introducing-synthesia-video-communications-platform?

OpenAI is on an acquiring spree, buying Rockset and Multi

Last week, OpenAI acquired Rockset, a startup that develops tools for real-time data search and analytics. OpenAI said it would integrate Rockset’s technology to power its infrastructure and offerings across products.

This week, OpenAI acquired Multi, a startup focused on building remote collaboration tools and software. Technically, the deal is an acqui-hire as the entire Multi team, including its co-founders, will join OpenAI to work on the company’s ChatGPT desktop application.

Sourcehttps://techcrunch.com/2024/06/24/openai-buys-a-remote-collaboration-platform

Record labels sue AI music startups over copyright infringement

The world’s major record labels, including Universal Music Group, Sony Music, and Warner Music, have filed twin lawsuits against the AI music generation startups Suno and Udio. The lawsuits accuse the companies of unlawfully training their AI models on massive amounts of copyrighted music, which, according to the complaints, allows the startups to generate similar-sounding music without permission.


AI Unraveled: Demystifying Frequently Asked Questions on Artificial Intelligence (OpenAI, ChatGPT, Google Gemini, Generative AI, Discriminative AI, xAI, LLMs, GPUs, Machine Learning, NLP, Promp Engineering)

The record labels allege Suno and Udio have effectively copied artists’ styles and specific musical characteristics. The labels claim the AI-generated music is so close to the original that it is eerily similar when transcribed into sheet music. The lawsuits also accuse the startups of making it easy for people to distribute AI-created samples that mimic copyrighted recordings on platforms like Spotify.

Sourcehttps://venturebeat.com/ai/record-labels-sue-ai-music-generator-startups-suno-udio-for-copyright-infringement/

Anthropic rolls out Claude’s cutting-edge collaborative features

Anthropic has introduced new collaboration features for Claude. These features include:

  • Projects: Projects in Claude allow integration of internal resources like style guides or codebases, enhancing Claude’s ability to deliver tailored assistance across various tasks. Users can set custom instructions for each Project to modify Claude’s tone or perspective for a specific role or industry.
  • Artifacts: It allows users to generate and edit various content types like code, documents, and graphics within a dedicated window. This benefits developers by offering larger code windows and live previews for easier front-end reviews.
  • Sharing Features: Claude Team users can share snapshots of their best conversations with Claude in their team’s shared project activity feed.

Additionally, any data or chats shared within Projects will not be used to train Anthropic’s generative models without a user’s explicit consent.

Sourcehttps://www.anthropic.com/news/projects

Google experiments with celebrity-inspired AI Chatbots

These chatbots will be powered by Google’s Gemini family of LLMs. The company aims to strike partnerships with influencers and celebrities and is also working on a feature that allows people to create their own chatbots by describing their personalities and appearances.

The project is led by Ryan Germick, a longtime executive at Google and a team of ten. These chatbots could be an experiment and may only appear on Google Labs rather than being widely available.

Sourcehttps://www.msn.com/en-us/news/other/google-wants-to-build-ai-chatbots-based-on-celebs-influencers-for-some-reason/ar-BB1oS1or

OpenAI postpones the launch of ChatGPT voice mode

Originally planned for late June, the Voice Mode aims to provide a more naturalistic and conversational experience with the AI chatbot, complete with emotional inflection and the ability to handle interruptions.

However, it will now be available only to a small group of users in late July or early August. OpenAI is working on improving content detection and user experience before wider rollout. GPT-4o’s real-time voice and vision capabilities are also expected to roll out to ChatGPT Plus users soon.

Sourcehttps://techcrunch.com/2024/06/25/openai-delays-chatgpts-new-voice-mode

Amazon steps into the chatbot race

Amazon is reportedly working on a new consumer-focused chatbot codenamed “Metis.” It is planned to be released somewhere around September. Here’s what we know about it:

  • The chatbot is powered by a new model, Olympus, and can be accessed via a web browser.
  • It uses a retrieval-augmented generation (RAG) technique to provide up-to-date information and automate tasks.
  • The model conversationally provides text and image-based outputs, suggesting follow-ups to queries. It also shares links to sources and supports image generation.
  • It uses an infrastructure similar to Amazon’s upcoming voice assistant, Remarkable Alexa.

Sourcehttps://www.businessinsider.com/amazon-chatgpt-rival-codenamed-metis-2024-6

Figma’s new AI features stir competition with Adobe

Figma announced a range of new features at the 2024 Config conference. Significant ones include a UI redesign, generative AI tools, new icons and toolbar, AI-enhanced asset search, and auto-generated texts in designs.

For instance, by typing a simple prompt into the textbox, users can create an entire app design mock-up for a restaurant. Figma will connect the design pages and even write suggested content!

Figma has also added a few designer-specific features to allow users to tweak designs in real-time. It features a developer mode with a “ready-for-dev” task list. The upgrade also boasts Figma slides, a Google slides-like tool for building and sharing presentations.

Sourcehttps://www.figma.com/whats-new/

Alibaba’s Qwen-72B tops the Hugging Face leaderboard

Hugging Face’s latest open large language model leaderboard ranks and evaluates open LLMs based on benchmarks like MMLU-pro and tests them on high-school and college-level problems.

The platform used 300 NVIDIA H100 GPUs to re-evaluate major open LLMs to obtain updated rankings. Chinese company Alibaba’s Qwen-72B dominated the leaderboard, becoming a top performer overall.

Not just that, the leaderboard was mainly dominated by Chinese companies, highlighting their headway into the open LLM space.

Sourcehttps://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard

Google’s Gemma 2, a set of lightweight, powerful open LLMs

Google has released Gemma 2 set of models that punch above their weight classes. Available in 9B and 27B parameter sizes, these models are

  • Higher performing and more efficient at inference than the first-generation
  • Have significant safety advancements built in
  • Optimized to run at incredible speed across a range of hardware and easily integrate with other AI tools
  • Trained on 13 trillion tokens for 27B, 8 trillion for 9B, and 2 trillion for 2.6B model (en route)

27B performs better than Llama3-70B and Nemotron-340B on Lmsys Arena, making it best in its size and stronger than some larger models. While 9B outperforms the likes of Mistral-large and Qwen1.5-110B.

The 27B Gemma 2 model is designed to run inference efficiently at full precision on a single Google Cloud TPU host, NVIDIA A100 80GB Tensor Core GPU, or NVIDIA H100 Tensor Core GPU. Moreover, this is an open weights model line, currently only available to researchers and developers.

Sourcehttps://blog.google/technology/developers/google-gemma-2

OpenAI’s CriticGPT finds GPT-4’s mistakes with GPT-4

OpenAI trained a model based on GPT-4, called CriticGPT, to catch errors in ChatGPT’s code output. It found that when users get help from CriticGPT to review ChatGPT code, they outperform those without help 60% of the time.

OpenAI aligns GPT-4 models to be more helpful and interactive through Reinforcement Learning from Human Feedback (RLHF). A key part of RLHF is collecting comparisons in which people, called AI trainers, rate different ChatGPT responses against each other.

OpenAI is beginning to integrate CriticGPT-like models into its RLHF labeling pipeline, providing trainers with explicit AI assistance.

Sourcehttps://openai.com/index/finding-gpt4s-mistakes-with-gpt-4

Google’s partnerships to help AI with real-world facts

Ace the Microsoft Azure Fundamentals AZ-900 Certification Exam: Pass the Azure Fundamentals Exam with Ease

Google is partnering with reputable third-party services, such as Moody’s, MSCI, Thomson Reuters, and Zoominfo, to ground its AI with real-world data. These four will be available within Vertex AI starting next quarter. They will offer developers qualified data to backstop their model outputs and ensure responses are factually accurate.

Google is also announcing high-fidelity grounding. Available through an experimental preview, it’s designed to help AI systems work better with a given set of specific information.

Sourcehttps://venturebeat.com/ai/google-grounding-ai-with-moodys-msci-thomson-reuters-zoominfo

A  Daily chronicle of AI Innovations June 28th 2024:

🚀 Google releases Gemma 2, a set of lightweight but powerful open LLMs
🔍 OpenAI’s CriticGPT finds GPT-4’s mistakes with GPT-4
🌐 Google partners with Moody’s, Thomson Reuters & more for AI data

🔎 OpenAI’s critics train smarter AI systems

🤝TIME and OpenAI announced a multi-year content deal

🌍Google is using AI to add 110 new languages to Google Translate

🎼YouTube is in talks with major record labels for an AI music deal

🤖Meta to start testing user-created AI chatbots on Instagram

📞Character.AI now allows users to talk with AI avatars over calls

🔎 OpenAI’s critics train smarter AI systems

OpenAI just published new research exploring how to improve model training using AI ‘critics’ that help human trainers spot errors in AI outputs more effectively.

  • OpenAI developed a model called CriticGPT to assist human trainers in evaluating and catching errors in code written by AI systems like ChatGPT.
  • In tests, CriticGPT caught 85% of inserted bugs in AI-generated code, compared to just 50% for human reviewers alone.
  • Human trainers also preferred CriticGPT’s code critics over their own 63% of the time on naturally occurring bugs.
  • OpenAI said that CriticGPT-type models are already being integrated into the company’s training workflow, helping prepare for advancing AI systems.

As OpenAI states in the research paper, both LLMs and AI critics will continue to scale — human intelligence will not. The limitations of our own puny mortal brains look to be no match for evaluating systems that are continuing to soar, meaning AI is quickly becoming its own trainer.

Source: https://openai.com/index/finding-gpt4s-mistakes-with-gpt-4

🚀 Google releases Gemma 2, a set of lightweight but powerful open LLMs

Google has released Gemma 2 set of models that punch above their weight classes. Available in 9B and 27B parameter sizes, these models are

  • Higher performing and more efficient at inference than the first-generation
  • Have significant safety advancements built in
  • Optimized to run at incredible speed across a range of hardware and easily integrate with other AI tools
  • Trained on 13 trillion tokens for 27B, 8 trillion for 9B, and 2 trillion for 2.6B model (en route)

27B performs better than Llama3-70B and Nemotron-340B on Lmsys Arena, making it best in its size and stronger than some larger models. While 9B outperforms the likes of Mistral-large and Qwen1.5-110B.

The 27B Gemma 2 model is designed to run inference efficiently at full precision on a single Google Cloud TPU host, NVIDIA A100 80GB Tensor Core GPU, or NVIDIA H100 Tensor Core GPU. Moreover, this is an open weights model line, currently only available to researchers and developers.

Why does it matter?

If you are looking for an all-in-one solution to help you prepare for the AWS Cloud Practitioner Certification Exam, look no further than this AWS Cloud Practitioner CCP CLF-C02 book

The models sound like they are built for practical deployments. They come in practical sizes so that they can be easily deployed while being amazing in quality due to best-in-class performances.

Source: https://blog.google/technology/developers/google-gemma-2/

🔍 OpenAI’s CriticGPT finds GPT-4’s mistakes with GPT-4

OpenAI trained a model based on GPT-4, called CriticGPT, to catch errors in ChatGPT’s code output. It found that when users get help from CriticGPT to review ChatGPT code, they outperform those without help 60% of the time.

OpenAI aligns GPT-4 models to be more helpful and interactive through Reinforcement Learning from Human Feedback (RLHF). A key part of RLHF is collecting comparisons in which people, called AI trainers, rate different ChatGPT responses against each other.

OpenAI is beginning to integrate CriticGPT-like models into its RLHF labeling pipeline, providing trainers with explicit AI assistance.

Why does it matter?

With more advances in reasoning and model behavior, AI models’ mistakes can become more subtle for AI trainers to spot. CriticGPT is a step towards addressing this fundamental limitation of RLHF.

Source: https://openai.com/index/finding-gpt4s-mistakes-with-gpt-4

🌐 Google partners with Moody’s, Thomson Reuters & more for AI data

Google is partnering with reputable third-party services, such as Moody’s, MSCI, Thomson Reuters, and Zoominfo, to ground its AI with real-world data. These four will be available within Vertex AI starting next quarter. They will offer developers qualified data to backstop their model outputs and ensure responses are factually accurate.

Google is also announcing high-fidelity grounding. Available through an experimental preview, it’s designed to help AI systems work better with a given set of specific information.

Why does it matter?

Earlier, Google announced efforts to ground Vertex AI results using web data and a plan to allow companies to ground AI systems in their own internal data.

Now, it is grounding these systems in known factual data from third parties, which could significantly lessen hallucinations and make AI more trustworthy for enterprise customers.

Source: https://venturebeat.com/ai/google-grounding-ai-with-moodys-msci-thomson-reuters-zoominfo

What Else Is Happening in AI on June 27th 2024❗

🤝TIME and OpenAI announced a multi-year content deal

OpenAI will gain access to current and historic content from TIME’s extensive archives from the last 101 years to enhance its products. It will also enable TIME to gain access to OpenAI’s technology to develop new products for its audiences.

Source: https://openai.com/index/strategic-content-partnership-with-time

🌍Google is using AI to add 110 new languages to Google Translate

It is Google’s largest expansion ever, thanks to its PaLM 2 LLM. It includes languages like Cantonese, NKo, and Tamazight, representing more than 614 million speakers and opening up translations for around 8% of the world’s population.

Source: https://blog.google/products/translate/google-translate-new-languages-2024/

🎼YouTube is in talks with major record labels for an AI music deal

It is offering to pay Universal Music Group (UMG), Sony Music Entertainment, and Warner Records “lump sums of cash” in exchange for legally licensing their songs to train new AI music tools. These will likely be one-off payments, not royalty-based arrangements. Source: https://www.theverge.com/2024/6/27/24187151/youtube-ai-music-deals-licensing-record-labels-sony-umg-warner

🤖Meta to start testing user-created AI chatbots on Instagram

CEO Mark Zuckerberg announced yesterday that Meta will begin to surface AI characters made by creators through Meta AI studio on Instagram, starting in the U.S. These will primarily show up in messaging for now and will be clearly labeled as AI.

Source: https://techcrunch.com/2024/06/27/meta-starts-testing-user-created-ai-chatbots-on-instagram

📞Character.AI now allows users to talk with AI avatars over calls

Users can initiate calls with a user-generated AI character directly with a button tap. Users can also switch between calling and texting seamlessly and stop the AI from talking through a “Tap to interrupt” option. The feature currently supports only a few languages.

Source: https://www.reuters.com/technology/artificial-intelligence/ai-chatbot-startup-characterai-launches-new-calls-feature-2024-06-27

📊 Use ChatGPT to turn docs into spreadsheets

With a simple prompt, ChatGPT can analyze documents, answer questions, perform calculations, and create a downloadable spreadsheet — all in one conversation!

  1. Log in to ChatGPT and upload your document to the chat (remember to remove any sensitive data before submitting).
  2. Ask questions about it, e.g., “How much is the security deposit?”
  3. Request calculations, like the total first month’s costs, including deposits and fees.
  4. Simple prompt ChatGPT to “create a downloadable budget spreadsheet”, specifying the time frame and desired columns.

Note: Only share documents you’re comfortable with. Be cautious with sensitive information and consider redacting critical details before uploading.

Source: https://university.therundown.ai/c/daily-tutorials/from-lease-agreement-to-budget-spreadsheet-with-chatgpt-0ed8d7ba-7d5a-4e9f-903a-64822caf1579

A  Daily chronicle of AI Innovations June 27th 2024:

🎵 YouTube wants to pay record labels to use their songs for AI training

🎨 Figma unveils AI-powered design tools, challenges Adobe’s dominance

🗣️ Character.AI introduces AI avatar calls

🏁 Amazon steps into the chatbot race 

🥇 Alibaba’s Qwen-72B tops the Hugging Face leaderboard

🏆 Hugging Face updates Open LLM Leaderboard

🎙️ NBC to use AI voice for Olympics recaps!

💡 Sonia’s ‘AI Therapist’ sparks debate

🏛️ Anthropic extends Claude’s access to government

🎵 YouTube wants to pay record labels to use their songs for AI training

  • YouTube is reportedly negotiating with major music labels like Sony Music Entertainment, Universal Music Group, and Warner Records to pay for licensing their songs for AI training.
  • Despite these negotiations, the ultimate decision to allow the use of songs for AI training will likely fall to individual artists rather than the labels themselves.
  • In April 2023, more than 200 artists voiced their opposition to AI using their work, emphasizing the potential threat to creators’ rights and the music ecosystem.

Source: https://www.engadget.com/youtube-reportedly-wants-to-pay-record-labels-to-use-their-songs-for-ai-training-125052503.html

🎨 Figma unveils AI-powered design tools, challenges Adobe’s dominance

  • Figma has launched AI-powered design features to speed up workflows and aid teams in creating superior software, positioning it as a product development platform to compete with Adobe.
  • The new Figma AI features include design creation with text prompts, visual search capabilities, layer renaming, and automating repetitive tasks, and are currently available in a limited beta.
  • While beta access to Figma AI is free, later availability will require subscriptions to specific plans, and admins must enable AI features, differentiating Figma’s approach from Adobe’s controversial AI terms.

Source: https://venturebeat.com/ai/figma-unveils-ai-powered-design-tools-to-rival-adobe/

🗣️ Character.AI introduces AI avatar calls

  • Character.AI announced that users can now make calls to AI characters, supporting multiple languages including English, Spanish, Portuguese, Russian, Korean, Japanese, and Chinese.
  • The company reported that during testing, over 3 million users made 20 million calls, and noted that AI character calls can help with language practice, mock interviews, or role-playing games.
  • With a button tap, users can start calls with AI characters, seamlessly switch between calling and texting, and use a “Tap to interrupt” feature to stop the AI from talking.

Source: https://techcrunch.com/2024/06/27/character-ai-now-allows-users-to-talk-with-avatars-over-calls/

🏁 Amazon steps into the chatbot race

Amazon is reportedly working on a new consumer-focused chatbot codenamed “Metis.” It is planned to be released somewhere around September. Here’s what we know about it:

  • The chatbot is powered by a new model, Olympus, and can be accessed via a web browser.
  • It uses a retrieval-augmented generation (RAG) technique to provide up-to-date information and automate tasks.
  • The model conversationally provides text and image-based outputs, suggesting follow-ups to queries. It also shares links to sources and supports image generation.
  • It uses an infrastructure similar to Amazon’s upcoming voice assistant, Remarkable Alexa.

Why does it matter?

The AI assistant market is already too crowded, making Metis’ release seem a bit too late. Its rivals, OpenAI and Google, have already launched their models in 2023 and will give tough competition to Amazon. If Amazon wishes to win this AI race, Metis must showcase exceptional capabilities.

Source: https://www.businessinsider.com/amazon-chatgpt-rival-codenamed-metis-2024-6

 
🎨 Figma’s new AI features stir competition with Adobe

Figma announced a range of new features at the 2024 Config conference. Significant ones include a UI redesign, generative AI tools, new icons and toolbar, AI-enhanced asset search, and auto-generated texts in designs.

https://x.com/figma/status/1806016455609286847

For instance, by typing a simple prompt into the textbox, users can create an entire app design mock-up for a restaurant. Figma will connect the design pages and even write suggested content!

Figma has also added a few designer-specific features to allow users to tweak designs in real-time. It features a developer mode with a “ready-for-dev” task list. The upgrade also boasts Figma slides, a Google slides-like tool for building and sharing presentations.

Why does it matter?

Figma’s AI tools will enable newer users on the platform to test out ideas easily. At the same time, tools like “ready-for-dev” will offer quick iterations to users who are more familiar with the platform.

With Adobe abandoning its plans to acquire Figma, the company’s recent upgrade might pose a challenge to the former.

Source: https://www.figma.com/whats-new/

🥇 Alibaba’s Qwen-72B tops the Hugging Face leaderboard

Hugging Face’s latest open large language model leaderboard ranks and evaluates open LLMs based on benchmarks like MMLU-pro and tests them on high-school and college-level problems.

The platform used 300 NVIDIA H100 GPUs to re-evaluate major open LLMs to obtain updated rankings. Chinese company Alibaba’s Qwen-72B dominated the leaderboard, becoming a top performer overall.

Not just that, the leaderboard was mainly dominated by Chinese companies, highlighting their headway into the open LLM space.

Why does it matter?

The open LLM leaderboard is essential for transparent and comprehensive LLM evaluation. It also proves that a model’s size doesn’t guarantee its superior performance. As Chinese models deliver exceptional performance, will well-known models from the USA, like Meta’s Llama, be replaced in the coming years?

Source: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard

What Else Is Happening in AI on June 27th 2024❗

🏛️ Anthropic extends Claude’s access to government

Anthropic is making Claude 3 Haiku and Claude 3 Sonnet available for government use through the AWS marketplace. This will improve citizen services, streamline document reviews, support data-based policymaking, and generate realistic training scenarios.

Source: https://www.anthropic.com/news/expanding-access-to-claude-for-government?

🌐 Opera announces new themes and features for One R2

Major features include AI-based image and voice generation, image understanding capabilities, floating windows, new themes, design elements, and split window tabs.

Source: https://press.opera.com/2024/06/26/opera-unveils-opera-one-r2-in-developer

🎙️ NBC to use AI voice for Olympics recaps!

The AI will offer daily, personalized coverage of the 2024 Paris Olympics on the Peacock app in the voice of the famous sports commentator, AI Michael. According to NBC, the voice AI is trained on data from Michael’s appearances on the show.

Source: https://www.nbcnews.com/news/nbc-use-ai-generated-version-al-michaels-voice-summer-olympics-rcna159108

📹 Caption’s new feature can auto-edit videos!

The feature adds custom graphics, zooms, music, sound effects, transitions, and motion backgrounds to an unedited video. Users can also use Caption’s AI avatars to create videos and fully edit them with the AI edit feature.

Source: https://www.captions.ai/

💡 Sonia’s ‘AI Therapist’ sparks debate

The personalized AI therapist can deal with depression, stress, anxiety, poor sleep, and more by analyzing what users say during therapy sessions. The release has sparked a debate around user data privacy, its susceptibility to bias, and the accuracy of its diagnosis. Source: https://www.soniahealth.com/

🏆 Hugging Face updates Open LLM Leaderboard

Hugging Face just introduced a new upgrade to its Open LLM Leaderboard, adding new benchmarks and evaluation methods to help address the recent plateau in LLM performance gains.

  • The leaderboard now features six new benchmarks designed to be more challenging and less prone to contamination.
  • Initial rankings show Qwen2-72B-Instruct leading the pack, followed by Meta’s Llama-3-70B-Instruct and Mixtral 8×22b.
  • A new normalized scoring system adjusts for baseline performance, providing a more fair comparison across different evaluation types.
  • The upgrade also introduces a ‘maintainer’s highlight’ category and community voting system to prioritize the most relevant models.

As LLMs approach human-level performance on most tasks, finding new ways to evaluate them is becoming more difficult — and more crucial. This revamp helps guide researchers and developers towards more targeted improvements, providing a more nuanced assessment of model capabilities.

Source: https://huggingface.co/spaces/open-llm-leaderboard/blog

🚨 Rabbit R1 hit with major security flaw

A group of developers just discovered a major vulnerability in Rabbit’s R1 AI assistant device, potentially exposing user’s private data and chat responses.

  • A community-led group called Rabbitude uncovered hardcoded API keys in Rabbit’s codebase, which allowed access to all R1 responses.
  • The group gained access to the codebase in mid-May, saying the Rabbit team was aware of the issue but failed to take action.
  • Rabbitude said the vulnerability could allow bad actors to disable all r1 devices, alter voices and responses, and access private messages.
  • Rabbit acknowledged an ‘alleged data breach’ via a Discord post, but claims no customer data was leaked.

Despite massive hype in the first wave of consumer AI standalone devices, the Rabbit r1 has been nothing short of a disaster so far. Already facing major criticism over the companion’s limited capabilities, this security breach only furthers the skepticism surrounding the early AI hardware market entrants.

Source: https://rabbitu.de/articles/security-disclosure-1?

A  Daily chronicle of AI Innovations June 26th 2024:

💼 Anthropic rolls out Claude’s cutting-edge collaborative features
🤖 Google experiments with celebrity-inspired AI Chatbots
🛑 OpenAI postpones the launch of ChatGPT voice mode

🧬 New AI simulates 500M years of evolution

⏰ OpenAI delays ChatGPT voice upgrade

💼 Anthropic rolls out Claude’s cutting-edge collaborative features

Anthropic has introduced new collaboration features for Claude. These features include:

  • Projects: Projects in Claude allow integration of internal resources like style guides or codebases, enhancing Claude’s ability to deliver tailored assistance across various tasks. Users can set custom instructions for each Project to modify Claude’s tone or perspective for a specific role or industry.
  • Artifacts: It allows users to generate and edit various content types like code, documents, and graphics within a dedicated window. This benefits developers by offering larger code windows and live previews for easier front-end reviews.
  • Sharing Features: Claude Team users can share snapshots of their best conversations with Claude in their team’s shared project activity feed.

Additionally, any data or chats shared within Projects will not be used to train Anthropic’s generative models without a user’s explicit consent.

Why does it matter?

By offering advanced AI collaboration tools, Anthropic meets critical enterprise needs for flexibility and team efficiency, where many of its competitors fall short. This approach could broaden its user base, increasing user retention and market penetration.

Source: https://www.anthropic.com/news/projects

🤖 Google experiments with celebrity-inspired AI Chatbots

These chatbots will be powered by Google’s Gemini family of LLMs. The company aims to strike partnerships with influencers and celebrities and is also working on a feature that allows people to create their own chatbots by describing their personalities and appearances.

The project is led by Ryan Germick, a longtime executive at Google and a team of ten. These chatbots could be an experiment and may only appear on Google Labs rather than being widely available.

Why does it matter?

Google’s AI chatbots featuring celebrities and influencers enhance user engagement but raise ethical concerns about consent, boundaries, and legal and privacy issues as AI evolves.

Source: https://www.msn.com/en-us/news/other/google-wants-to-build-ai-chatbots-based-on-celebs-influencers-for-some-reason/ar-BB1oS1or

🛑 OpenAI postpones the launch of ChatGPT voice mode

Originally planned for late June, the Voice Mode aims to provide a more naturalistic and conversational experience with the AI chatbot, complete with emotional inflection and the ability to handle interruptions.

However, it will now be available only to a small group of users in late July or early August. OpenAI is working on improving content detection and user experience before wider rollout. GPT-4o’s real-time voice and vision capabilities are also expected to roll out to ChatGPT Plus users soon.

Why does it matter?

Delays can give competitors like Anthropic a chance to gain an edge. If rivals manage to launch similar or superior voice capabilities sooner, OpenAI risks losing potential users and market share.

Source: https://techcrunch.com/2024/06/25/openai-delays-chatgpts-new-voice-mode

🧬New AI simulates 500M years of evolution

Biotech startup EvolutionaryScale just introduced ESM3, a new AI model capable of designing and simulating new proteins in ways that mimic millions of years of natural evolution.

  • ESM3 is a frontier language model that can reason over protein sequences, structures, and functions simultaneously.
  • In tests, ESM3 generated a new fluorescent protein 58% different from other known structures — a leap that typically takes nature 500M years of evolution.
  • The largest version of ESM3 has 98B parameters, and was trained on approximately 1.8T tokens of biological data.
  • EvolutionaryScale was created by former META researchers, and has already secured $142M in seed funding from tech giants like Amazon and Nvidia.

ESM3 is bringing us closer to a future where we can ‘program’ biology as easily as we write computer code — potentially revolutionizing fields from healthcare to environmental science and unlocking completely new areas of knowledge across fields.

Source: https://www.evolutionaryscale.ai/blog/esm3-release

What Else Is Happening in AI on July 26th 2024❗

🧬 Ex-Meta engineers launch protein-generating AI biology model

EvolutionaryScale, launched by ex-Meta engineers, introduced ESM3, a gen AI model for designing novel proteins. It uses AI to simulate evolutionary processes, aiming to revolutionize industries like medicine and biofuels. It is available via API to selected customers and non-commercial users.

Source: https://venturebeat.com/ai/meta-alum-launches-ai-biology-model-that-simulates-500-million-years-of-evolution

🛑 Reddit sets new barriers for AI crawlers to protect its content

Reddit will rate-limit or block unknown bots and crawlers that violate its Public Content Policy or lack an agreement with the platform. These changes aim to deter AI companies from training LLMs on Reddit content (excluding the $60 million deal with Google).

Source: https://techcrunch.com/2024/06/25/reddits-upcoming-changes-attempt-to-safeguard-the-platform-against-ai-crawlers

🎥 AI-generated ad film debuts at Cannes 2024 

Toys ‘R’ Us, the iconic toy store chain, collaborated with OpenAI’s Sora  to produce a brand film called “The Origin of Toys ‘R’ Us.” The film premiered at the 2024 Cannes Lions International Festival of Creativity.

Source: https://venturebeat.com/ai/toys-r-us-unveils-first-commercial-made-with-openais-sora

🦊 Mozilla’s Firefox Nightly now lets users pick from leading AI chatbots

It allows users to choose from multiple AI chatbots, such as ChatGPT, Google Gemini, HuggingChat, and Le Chat Mistral. The feature lets users select their preferred AI to perform tasks such as summarizing web content, simplifying text, or conducting quizzes.

Source: https://techcrunch.com/2024/06/25/firefox-now-lets-you-choose-your-preferred-ai-chatbot-in-its-nightly-builds

📧 Google extends Gemini AI to Gmail to streamline email tasks

Google has integrated its Gemini AI into Gmail for Workspace subscribers, enhancing email functionality. It assists users by automatically drafting emails, summarizing threads, crafting responses, and retrieving information from emails or Google Drive.

Source: https://techcrunch.com/2024/06/25/google-brings-its-gemini-ai-to-gmail-via-a-sidebar-that-can-help-you-write-and-summarize-emails

A  Daily chronicle of AI Innovations June 25th 2024:

🤖 Amazon is working on a ChatGPT competitor

🎤 Google is building AI chatbots modeled after celebrities, YouTube influencers

📉 Nvidia’s market value plummets drastically

🤷‍♂️ Apple rejected Meta’s AI integration over privacy issues

🤝 OpenAI buys a remote collaboration platform

🎥 Synthesia 2.0: World’s 1st AI video communication platform
🤝 OpenAI is on an acquiring spree, buying Rocket and Multi
🎵 Record labels sue AI music startups over copyright infringement

🤖 Amazon is working on a ChatGPT competitor

  • Amazon is developing an AI chatbot, code-named “Metis,” to compete with OpenAI’s ChatGPT, and it could be launched later this year, based on a report from Business Insider.
  • The upcoming chatbot will be accessed through a web browser and rely on a new foundational model, as mentioned by sources and documents cited by Business Insider.
  • Amazon’s CEO Andy Jassy is reportedly directly involved with “Metis,” reflecting Amazon’s broader strategy to intensify its AI efforts across various applications and services.
  • Source: https://www.geekwire.com/2024/report-amazon-developing-ai-chatbot-that-would-compete-with-chatgpt-and-others/

🎤 Google is building AI chatbots modeled after celebrities, YouTube influencers

  • Google is reportedly developing new AI chatbots modeled after celebrities and YouTube influencers, aiming to create specialized interactive experiences.
  • The initiative will use Google’s Gemini family of large language models and seeks partnerships with influencers and celebrities, allowing users to design their own chatbots by describing traits and appearance.
  • The project, led by executive Ryan Germick and a team of ten, might initially be available only on Google Labs, keeping the bots experimental and not widely accessible at launch.
  • Source: https://www.engadget.com/google-is-reportedly-building-ai-chatbots-based-on-celebrities-and-influencers-235731655.html

📉 Nvidia’s market value plummets drastically

  • Nvidia’s shares dropped nearly 13% in a week, causing a loss of $430 billion from its market value and dropping its market capitalization to $2.91 trillion.
  • This decline has moved Nvidia from the top position to third place among the world’s most valuable companies, now trailing behind Microsoft and Apple.
  • Investor enthusiasm for Nvidia’s role in AI is waning after significant stock gains, leading to a 6.7% drop on Monday, the third consecutive day of declines for the company.
  • Source: https://edition.cnn.com/2024/06/25/tech/nvidia-stock-fall-worlds-third-biggest-company/index.html

🤷‍♂️ Apple rejected Meta’s AI integration over privacy issues

  • Apple declined a potential AI partnership with Meta concerning the integration of Meta’s large language model into iOS, primarily due to privacy concerns highlighted in a Bloomberg report.
  • While initial discussions between Apple and Meta took place, they did not advance, with Apple choosing not to pursue formal talks or integration due to Meta’s insufficient privacy practices.
  • Despite rejecting Meta, Apple has engaged in collaborations with other major AI entities, finalizing a deal with OpenAI to integrate ChatGPT into iOS 18 and discussing potential integrations with Google and Anthropic for diverse AI model options.
  • Source: https://www.macrumors.com/2024/06/24/apple-meta-ai-privacy-concerns/

🤝 OpenAI buys a remote collaboration platform

  • OpenAI is advancing its AI technology by integrating it into desktop computing, moving beyond browser-based interactions with the acquisition of startup Multi.
  • By purchasing Multi, OpenAI aims to develop potentially a new AI-based operating system or at least an interaction layer for current systems, as suggested by recent product releases including a ChatGPT desktop app.
  • OpenAI’s future AI assistant could replace traditional browsers, altering the role of search engines and websites, and revolutionizing how we interact with online content.
  • Source: https://the-decoder.com/openai-moves-toward-its-own-ai-operating-system-and-a-post-browser-world/

 Synthesia 2.0: World’s 1st AI video communication platform

Synthesia is launching Synthesia 2.0 – the world’s first AI video communications platform for businesses. It reinvents the entire video production process, allowing companies to create and share AI-generated videos at scale easily.

https://youtu.be/gZaBwdru_bk?si=fP7_nILeu66Oqj4B

The key new features and capabilities of Synthesia 2.0 include:

  • 2 Personal AI Avatars: Expressive Avatars shot in a studio and Custom Avatars created using your webcam.
  • AI Video Assistant: Converts text, documents, or websites into high-quality videos, with options to customize the branding, tone, and length.
  • Intuitive Video Editing: Editing simplified with “Triggers” that let you control animations and edits from the script.
  • Translation and Dynamic Video Player: Videos can now be translated into over 120 languages. Synthesia is also building a new video player with interactive features.
  • AI Safety Focus: Synthesia is pursuing ISO/IEC 42001 certification, the first standard for responsible AI management, to ensure its AI technologies are ethical.

Why does it matter?

By making video creation and distribution more accessible and automated, Synthesia aims to help businesses transition to a “video-first” approach to communicate more effectively, improve employee training, customer service, and marketing efforts, and streamline various business operations.

Source: https://www.synthesia.io/post/introducing-synthesia-video-communications-platform

🤝 OpenAI is on an acquiring spree, buying Rocket and Multi

Last week, OpenAI acquired Rockset, a startup that develops tools for real-time data search and analytics. OpenAI said it would integrate Rockset’s technology to power its infrastructure and offerings across products.

This week, OpenAI acquired Multi, a startup focused on building remote collaboration tools and software. Technically, the deal is an acqui-hire as the entire Multi team, including its co-founders, will join OpenAI to work on the company’s ChatGPT desktop application.

Why does it matter?

With over 600,000 paid business users, including most Fortune 500 companies, OpenAI’s back-to-back acquisitions strategically position it to expand its enterprise-focused AI products and make its powerful language models more useful for businesses.

Source: https://techcrunch.com/2024/06/24/openai-buys-a-remote-collaboration-platform

🎵 Record labels sue AI music startups over copyright infringement

The world’s major record labels, including Universal Music Group, Sony Music, and Warner Music, have filed twin lawsuits against the AI music generation startups Suno and Udio. The lawsuits accuse the companies of unlawfully training their AI models on massive amounts of copyrighted music, which, according to the complaints, allows the startups to generate similar-sounding music without permission.

The record labels allege Suno and Udio have effectively copied artists’ styles and specific musical characteristics. The labels claim the AI-generated music is so close to the original that it is eerily similar when transcribed into sheet music. The lawsuits also accuse the startups of making it easy for people to distribute AI-created samples that mimic copyrighted recordings on platforms like Spotify.

Why does it matter?

This legal battle is only the latest example, and one of many, of AI technology colliding with creative industries. However, any resolution of these battles will likely have far-reaching implications for the music industry and the development of AI-powered music tools.

Source: https://venturebeat.com/ai/record-labels-sue-ai-music-generator-startups-suno-udio-for-copyright-infringement

What Else Is Happening in AI on June 25th❗

🍎 Apple withdraws AI collaboration deal with Meta over privacy concerns

Contrary to the rumors about a potential AI collaboration, Apple is not planning a partnership to integrate Meta’s AI models into its products due to privacy concerns. Instead, Apple is focusing on partnerships with OpenAI and Google that align with its commitment to user privacy.

Source: https://techcrunch.com/2024/06/24/apple-shelved-the-idea-of-integrating-metas-ai-models-over-privacy-concerns-report-says

🛍️ Shopify expands its AI capabilities to attract more businesses

Shopify has expanded access to its AI-powered features, including the Sidekick assistant and image-generation tool. It is making these AI tools available to more users, as it aims to provide merchants with an integrated and seamless platform to scale their businesses.

Source: https://www.reuters.com/technology/artificial-intelligence/shopify-expands-access-its-ai-powered-features-attract-more-businesses-2024-06-24

🏫 Google is bringing Gemini access to teens through school accounts

Google now makes its AI assistant Gemini available to teenage students through school accounts. This move is aimed at helping prepare students for a future where generative AI is more prevalent. Google says Gemini will provide real-time feedback and support to students while including safeguards to prevent inappropriate responses.

Source: https://techcrunch.com/2024/06/24/google-is-bringing-gemini-access-to-teens-using-their-school-accounts

🤖 NVIDIA deploys its AI tech in the Middle East among U.S AI export curb

NVIDIA has struck a deal with Qatari telecom provider Ooredoo to deploy its AI technology in the Middle East. This move comes amid U.S. export restrictions on advanced AI chips to some countries in the region. Ooredoo will be the first company in the Middle East to offer Nvidia’s AI and graphics technology to its clients across multiple countries.

Source: https://www.reuters.com/technology/nvidia-launch-middle-east-amid-us-curbs-ai-exports-region-ooredoo-ceo-says-2024-06-23

📷 Meta marks real photos with ‘Made by AI’ label

Meta has been incorrectly labeling real photos as “Made by AI.” Multiple photographers have reported that Meta has added this label to their genuine photographs, even when they have only made minor edits using tools like Adobe’s Generative Fill. Meta says it relies on industry standards to detect AI-generated content, leading to false positives.

Source: https://www.theverge.com/2024/6/24/24184795/meta-instagram-incorrect-made-by-ai-photo-labels

A  Daily chronicle of AI Innovations June 24th 2024:

🤝 Apple and Meta are discussing gen AI partnership
🤖 ByteDance and Broadcom collaborate on AI chip development
🔍 🕵️‍♂️ Researchers developed a new method to detect hallucinations

🤝 Apple in talks with Meta for potential AI integration

🎵 Record labels sue AI music companies

🩸 AI blood test for Parkinson’s detection

Researchers just developed an AI-powered blood test that can predict Parkinson’s disease up to 7 years before symptoms appear, potentially revolutionizing early diagnosis and treatment.

  • The AI analyzes key markers in blood samples linked to inflammation and protein degradation, attempting to predict which patients would develop the disease.
  • In a 10-year study, the test correctly identified 79% of patients who later developed Parkinson’s.
  • The AI model predicted Parkinson’s up to 7 years before clinical symptoms emerged in some cases.
  • Researchers hope to develop a simple finger-prick version for widespread screenings.

AI’s ability to find subtle patterns in previously undetectable biological data is unlocking new diagnostic capabilities. By catching a debilitating disease years before the onset of symptoms, patients can get earlier access to treatment aimed at prevention.

Source: https://www.bbc.com/news/articles/c722yly6l6ro

🤝 Apple in talks with Meta for potential AI integration

  • Apple is reportedly negotiating with Meta to integrate Meta’s generative AI model into Apple’s new AI system, Apple Intelligence, according to The Wall Street Journal.
  • Apple is seeking partnerships with multiple AI companies, including Meta, to enhance its AI capabilities and catch up in the competitive AI race.
  • A potential collaboration between Apple and Meta would be significant due to their history of disagreements, and it could greatly impact the AI industry if successful.
  • Source: https://www.benzinga.com/news/24/06/39447797/apple-eyes-meta-for-potentially-game-changing-ai-partnership-report

🎵 Record labels sue AI music companies

  • Universal Music Group, Sony Music Entertainment, and Warner Records are suing AI companies Suno and Udio for massive copyright infringement related to the creation of original songs using generative AI technology.
  • The music labels claim that Suno and Udio used copyrighted works without permission, and are seeking damages of up to $150,000 per violated work in lawsuits filed in Boston and New York federal courts.
  • This legal action is part of a broader conflict between the music industry and tech companies offering AI tools, highlighting concerns about AI-generated music impacting the control and revenue of human artists.
  • Source: https://www.theverge.com/2024/6/24/24184710/riaa-ai-lawsuit-suno-udio-copyright-umg-sony-warner

🤝 Apple and Meta are discussing gen AI partnership

Apple is reportedly in talks with its longtime rival Meta to integrate the latter’s Llama 3 AI model into Apple Intelligence. This move comes as Apple prepares to roll out its AI features across iPhones, iPads, and Macs later this year.

The potential partnership follows Apple’s existing deal with OpenAI, suggesting a collaboration strategy rather than solo development in the AI race. In Apple’s arrangement with OpenAI, there’s no direct payment. Instead, OpenAI can offer premium subscriptions through Apple Intelligence, with Apple taking a percentage. It’s unclear if Meta would agree to a similar business model, given that Llama 3 is open-source and free to access.

Why does it matter?

Despite their historical rivalry, especially after Apple’s 2021 privacy changes impacted Meta’s ad revenue, this potential partnership suggests a willingness to cooperate in the AI race. However, not everyone is thrilled about it. Users and investors are voicing concerns about privacy, with Meta’s track record being a major sticking point. Some see this as a desperate move by Apple.

Source: https://www.wsj.com/tech/ai/apple-meta-have-discussed-an-ai-partnership-cc57437e

🤖 ByteDance and Broadcom collaborate on AI chip development

ByteDance is collaborating with U.S. chip designer Broadcom to develop an advanced AI processor. This partnership aims to secure a stable supply of high-end chips amid ongoing U.S.-China tensions. The project centers on creating a 5-nanometre, customized Application-Specific Integrated Chip (ASIC) that complies with U.S. export restrictions.

This chip’s manufacturing is set to be outsourced to Taiwan Semiconductor Manufacturing Company (TSMC), though production is not expected to begin this year. While the design work is currently underway, the critical “tape out” phase has yet to commence.

Why does it matter?

This partnership shows how Chinese tech companies are getting creative to secure advanced AI chips. As ByteDance has launched a major push in generative AI, this new initiative brings dual benefits: reducing procurement costs and ensuring a stable supply of higher-end chips.

Source: https://www.reuters.com/technology/artificial-intelligence/chinas-bytedance-working-with-broadcom-develop-advanced-ai-chip-sources-say-2024-06-24

🔍 🕵️‍♂️ Researchers developed a new method to detect hallucinations

ChatGPT and Gemini can produce impressive results but often “hallucinate” false or unsubstantiated information. This research focuses on a subset of hallucinations called “confabulations,” where LLMs generate answers that are both wrong and arbitrary. Researchers have developed new methods to detect confabulations using entropy-based uncertainty estimators. They introduce the concept of “semantic entropy” to measure the uncertainty of LLM generations at the meaning level.

High semantic entropy corresponds to high uncertainty and indicates a higher likelihood of confabulation. The method computes uncertainty at the level of meaning rather than specific word sequences, addressing the fact that one idea can be expressed in many ways. The method provides scalable oversight by detecting confabulations that people might otherwise find plausible.

Why does it matter?

Detecting confabulations can improve question-answering accuracy for state-of-the-art LLMs. The method helps users understand when to be cautious with LLM outputs and opens up new possibilities for using LLMs that were previously limited by unreliability.

Source: https://www.nature.com/articles/s41586-024-07421-0

What Else Is Happening in AI on June 24th 2024❗

🆕Stability AI has appointed a new CEO

Prem Akkaraju, former CEO of Weta Digital, will step up to this role. This change comes after founder Emad Mostaque resigned in March amid financial concerns. Akkaraju is part of a new investor group, including former Facebook President Sean Parker, that provides a cash infusion to Stability AI. Source: https://www.theinformation.com/articles/stability-ai-gets-new-ceo-and-a-bailout-from-investor-group-led-by-sean-parker

🖼️ Coders spotted Midjourney references in Grok’s standalone version

It hints at a possible collaboration between xAI and Midjourney for image generation. Grok might use Midjourney models via API to generate images. This could mean that the standalone Grok might let users upload images. All this could be part of the Grok 1.5 rollout. Source: https://www.bolnews.com/technology/2024/06/grok-1-5-potential-use-of-midjourney-for-image-generation-gains-evidence

🚫YouTube fights AI fakes with new policy 

YouTube expanded its privacy policy to cover AI-generated content. Now, users can request the removal of synthetic media that imitates their face or voice. However, privacy complaints alone may not guarantee content removal. While reviewing requests, YouTube will consider the potential for confusion and intent (parody/satire). Source: https://support.google.com/youtube/thread/281049221

✨ OmniAI transforms unstructured enterprise data for AI and analytics use

Founded by Y Combinator alums, OmniAI syncs with your existing storage, prep the data, and lets you run your choice of AI models on the data. It offers integrations with popular models like Meta’s Llama 3 and Anthropic’s Claude. Source: https://techcrunch.com/2024/06/22/omniai-transforms-business-data-for-ai

🤖 The new Dot AI chatbot wants to be your companion 

Dot chatbot aims to personalize interactions based on users’ thoughts and feelings. It uses multiple AI models from OpenAI and Google. The app uses intensive Q&A to learn about users’ interests and experiences. Dot offers a subscription model at $11.99/month for unlimited conversations. Source: https://techcrunch.com/2024/06/21/dots-ai-really-really-wants-to-get-to-know-you

Y Combinator joined over 140 AI startup founders in signing a letter opposing California’s Senate Bill 1047 on AI regulation, arguing it harms the ability to retain talent and stifles innovation. Source: https://www.politico.com/newsletters/california-playbook/2024/06/21/little-tech-brings-a-big-flex-to-sacramento-00164369
Formula 1 and AWS unveiled ‘Statbot’, an AI-powered system that provides real-time context, trivia, and predictions for personalized race broadcasts. Source: https://www.bloomberg.com/news/articles/2024-06-23/formula-1-and-amazon-aim-for-ai-powered-personalized-race-viewing
OpenAI announced the acquisition of Rockset, with the search and database analytics startup’s tech set to boost retrieval infrastructure for enterprise products. https://openai.com/index/openai-acquires-rockset
Axel Peytavin demonstrated a small, autonomous robot prototype powered by GPT-4o that can see, think and act based on prompts, with capabilities like picking up trash and shaking hands. Source: https://x.com/ax_pey/status/1804209628680720746
OpenAI CTO Mira Murati said in an interview that the next generation of GPT will be ‘PhD-level’ compared to GPT-3 (toddler) and GPT-4 (high school), and will be released in a year and a half. Source: https://x.com/tsarnick/status/1803901130130497952
Unitree posted new demos of its Go2 robotic dog performing handstands, flips, and withstanding and adjusting to force, showing extreme precision and agility. Source: https://x.com/UnitreeRobotics/status/1803728664213012520

A  Daily chronicle of AI Innovations June 20th 2024:

🖥️ Microsoft debuts a vision-foundational model for diverse tasks
👨‍💼 Ex-OpenAI co-founder launches own AI company
🤖 Can AI read minds? New model can predict anxiety levels 

😱 Sutskever launches new AI venture

🚧 Tech giants team up on AI factory for xAI

🚨 Perplexity faces legal action

🤖 OpenAI co-founder announces new AI company

🚀 Anthropic announces its most powerful AI yet

🇪🇺 EU Council withdraws Chat Control vote

🚧 Apple Intelligence plans face legal challenges in China

🖥️ Microsoft debuts a vision-foundational model for diverse tasks

Florence 2, Microsoft’s vision-AI model, can perform diverse tasks like object detection, captioning, visual grounding, and visual question answering via image and text prompts. It displays excellent captioning, object detection, visual grounding, and segmentation.

The model comes in 232M and 771M parameter sizes and uses a sequence-to-sequence architecture, enabling multiple vision tasks without needing a task-specific architecture modification.

On fine-tuning the model with publicly available human-annotated data, Florence 2 showcased impressive results, offering tough competition to existing large vision models like Flamingo despite its compact size.

Why does it matter?

The model will equip enterprises with a standard approach to handling various vision-handling applications. This will save resources spent on separate task-specific vision models that need fine-tuning. Moreover, it may also be useful to developers as it would eliminate the need for separate vision models for smaller tasks, significantly saving compute costs.

Source: https://www.microsoft.com/en-us/research/publication/florence-2-advancing-a-unified-representation-for-a-variety-of-vision-tasks

👨‍💼 Ex-OpenAI co-founder launches own AI company

Just a month after leaving OpenAI, ex-cofounder Ilya Sutskever has launched his own AI company, Safe Superintelligence Inc. (SSI), alongside former Y Combinator partner Daniel Gross and ex-OpenAI engineer Daniel Levy as co-founders.

According to the SSI’s launch statement on X, the company will prioritize safety, progress, and security. Sutskever also emphasizes that the company’s “singular focus” on a joint approach to safety and capabilities will prevent it from being distracted by management overhead or production cycles, unlike companies like OpenAI or Google.

Why does it matter?

SSI’s launch clearly marks the emergence of a new key player in the race to build safe, powerful AI. Its mission statement emphasizes safety and the potential for groundbreaking developments that may shape the future of AI research and development. It would be interesting to see whether the startup will uphold its mission statement in the coming days.

Source: https://ssi.inc/

🤖 Can AI read minds? New model can predict anxiety levels 

Researchers at the University of Cincinnati have developed an AI model that can identify people with an urgent risk of anxiety. The AI model uses minimal computational resources, a short picture rating task, and a small set of variables to make the prediction. The approach named “Comp Cog AI” integrates computational cognition and AI.

Participants rated 48 pictures with mildly emotional subject matter based on the degree to which they liked or disliked those pictures. The response data was then used to quantify the mathematical features of their judgments. Finally, the data was combined with ML algorithms to identify their anxiety levels.

Since the technology doesn’t rely on a native language, it is accessible to a wider audience and diverse settings to assess anxiety.

Why does it matter?

The picture rating feature can provide unbiased data to medical professionals on a person’s mental health status without subjecting them to direct questions that may trigger negative emotions. Given its 81% accuracy rate, the tool can become a useful app for detecting individuals with high anxiety risks.

Source: https://www.nature.com/articles/s44184-024-00074-x.epdf

🚀 Anthropic announces its most powerful AI yet

  • Anthropic has launched Claude 3.5 Sonnet, a new AI model that aims to be on par with, or superior to, OpenAI’s GPT-4o and Google’s Gemini across various tasks.
  • Claude 3.5 Sonnet claims to be significantly faster than its predecessor and outperforms it, even surpassing other leading models in multiple benchmarks.
  • Alongside the new model, Anthropic introduced the Artifacts feature, which allows users to interact with and edit Claude’s outputs directly within the app, enhancing its functionality beyond a typical chatbot.
  • Source: https://www.theverge.com/2024/6/20/24181961/anthropic-claude-35-sonnet-model-ai-launch

🚧 Apple Intelligence plans face legal challenges in China

  • Apple’s push to enhance its artificial intelligence capabilities faces significant challenges in China due to the country’s stringent AI regulations and censorship laws.
  • China’s strict regulation of AI, including rules about language models and data protection, means Apple must navigate approvals and potentially partner with local tech giants like Baidu and Alibaba.
  • Apple’s focus on user privacy and the need for localized AI experiences may help it gain regulatory approval but also present significant hurdles in ensuring compliance and acceptance in the Chinese market.
  • Source: https://www.cnbc.com/2024/06/20/apple-ai-push-faces-big-challenges-in-china.html

🚧 Tech giants team up on AI factory for xAI

 Dell, Nvidia, and Super Micro Computer are collaborating to build an AI factory for xAI’s massive supercomputer, which will train and scale its Grok AI chatbot.

  • Dell is assembling half of the server racks for xAI’s planned supercomputer, with Super Micro handling the other half.
  • The system aims to use up to 100,000 Nvidia H100 GPUs, potentially making it 4x larger than the biggest existing AI clusters.
  • Musk wants the supercomputer to be fully operational by Fall 2025 to accelerate the development of advanced Grok models.

This powerhouse partnership shows the staggering scale of compute power required to push the boundaries of AI. With huge capital flowing in, this massive cluster could provide a critical edge in evolving Grok to be a serious competitor to the likes of OpenAI and Anthropic.

Source: https://www.fastcompany.com/91143745/dell-super-micro-musk-xai-supercomputer-server-racks

🚨 Perplexity faces legal action

News publisher Forbes is reportedly threatening legal action against AI search startup Perplexity, alleging the company’s chatbot stole and republished original reporting without proper attribution.

  • Forbes claims Perplexity’s chatbot summarized one of its investigative articles, copying with minimal sourcing.
  • The AI-generated version was then pushed to Perplexity subscribers and turned into a YouTube video that outranked Forbes’ own content.
  • Perplexity CEO Aravind Srinivas responded that the product still has “rough edges”, and the company is working on ways to align with publishers.
  • A WIRED investigation found Perplexity scraping sites that actively blocked crawlers, while also hallucinating info and containing incorrect attributions.

Tensions between AI firms and media companies continue to escalate. While legal grounds surrounding scraping and repurposing with AI are still murky, publishers are in a bind when it comes to protecting IP. The only viable options right now seem to be cashing out with licensing deals or taking legal action.

Source: https://www.axios.com/2024/06/18/forbes-perplexity-ai-legal-action-copyright

What Else Is Happening in AI on June 20th 2024❗

🤝 Deloitte, HPE, and NVIDIA partner up! The alliance combines Deloitte’s deep industry expertise and AI capabilities with the newly released HPE NVIDIA AI Compute solutions suite. The collaboration further seeks to advance industry-specific gen AI applications and assist clients across sectors in modernizing their data strategies by accessing innovative insights.

Source: https://www.prnewswire.com/news-releases/deloitte-announces-collaboration-with-hewlett-packard-enterprise-and-nvidia-on-co-developed-generative-ai-solutions-302175762.html

🚗 NVIDIA’s AI can turbocharge deployment of self-driving cars: NVIDIA has unveiled a new AI software, Omniverse Cloud Sensor RTX, that will likely accelerate the development of self-driving cars and robots.

The software combines real-world data with synthetic data, making it easy to test sensor perception in realistic virtual environments before deployment into the actual world.

Source: https://nvidianews.nvidia.com/news/omniverse-microservices-physical-ai

🎶 YouTube trials AI-powered “Ask for music” feature: The feature seeks to make searching music more conversational for users. Users can search for music using voice commands and verbal prompts. It is speculated that the feature may be an addition to the AI-generated playlist cover available on YouTube.

Source: https://www.business-standard.com/topic/youtube

🎥 Luma adds an “extend video” feature to Dream Machine: In contrast to Luma’s previous five-second limit for videos, the “extend video” feature will allow videos to be extended based on prompts. The AI model will consider the new context while extending the video. Additionally, the upgrade will enable Standard, Pro, and Premier users to remove watermarks.

Source: https://x.com/LumaLabsAI/status/1803098863097835995

🛠️ Apple releases 20 core ML models on Hugging Face: The release includes 4 major datasets in addition to 20 core models as a part of Apple’s efforts to equip developers with advanced on-device AI capabilities.

These core ML models have been optimized to run exclusively on users’ devices and can be used for various applications like image classification, depth estimation, and semantic segmentation.

Source: https://huggingface.co/apple

A  Daily chronicle of AI Innovations June 19th 2024:

📈 Nvidia is now the world’s most valuable company

🤖 Perplexity AI is lying about their user agent

👀 Meta releases new AI models for text, image and audio

🦋 Ex-Snap engineer launches social network where AIs and humans coexist

🩺 OpenAI and Color Health join forces to accelerate cancer treatment
🎤 Meta unveils new AI models for audio, text, and watermarking
⚒️Notion introduces AI Connectors for Slack & Google Drive integration

🎺 TikTok brings AI to content creation

📈 Nvidia is now the world’s most valuable company

  • Nvidia surpassed Microsoft to become the world’s most valuable company, reaching a market cap of $3.34 trillion, driven by its focus on artificial intelligence and expanded robotics research.
  • The company’s stock price increased by 3.7%, putting it ahead of Microsoft’s $3.33 trillion market cap, while Apple ranks third with a market cap of $3.28 trillion.
  • Nvidia’s rise is fueled by partnerships and advancements in AI and robotics, including a new collaboration with Hewlett Packard to launch NVIDIA AI Computing and Private Cloud AI, along with significant projects with major AI developers like Google, Meta, Microsoft, and OpenAI.
  • Source: https://decrypt.co/235988/nvidia-flips-microsoft-top-tech-company

🤖 Perplexity AI is lying about their user agent

  • WIRED and developer Robb Knight found that Perplexity’s crawler ignores the Robots Exclusion Protocol, accessing content from sites that have blocked it, including Condé Nast properties.
  • Perplexity CEO Aravind Srinivas describes the product as an “answer engine,” but the chatbot has been found to paraphrase articles without proper attribution and sometimes fabricate information entirely.
  • An experiment by WIRED demonstrated that Perplexity sometimes generates content without accessing the actual webpage, raising concerns about the accuracy and ethics of its information-gathering methods.
  • Source: https://www.wired.com/story/perplexity-is-a-bullshit-machine/

👀 Meta releases new AI models for text, image and audio 

  • Meta’s Fundamental AI Research team introduced new models such as Chameleon, a multimodal model for text and images, and JASCO, a text-to-music model accepting various inputs like chords or beats.
  • Meta’s Chameleon can handle and generate text and images, with its 7B and 34B variants available for research under a non-commercial license.
  • The company launched AudioSeal, an audio watermarking technology, which can identify AI-generated speech, released under a commercial license for recognizing speech in lengthy audio segments.
  • Source: https://the-decoder.com/meta-releases-new-ai-models-for-text-image-and-audio/

🦋 Ex-Snap engineer launches social network where AIs and humans coexist

  • Butterflies, founded by former Snap engineer Vu Tran, is a social network allowing humans and AI personas (called Butterflies) to interact through posts, comments, and direct messages, launching publicly on iOS and Android after five months in beta.
  • Users can easily create Butterfly personas with backstories, emotions, and opinions, which then autonomously generate content that other users and AIs can engage with, aiming to provide a more immersive AI experience compared to traditional chatbots.
  • The app received $4.8 million in seed funding led by Coatue and has mostly been used for entertainment, with potential plans for subscription models and brand interactions in the future.
  • Source: https://techcrunch.com/2024/06/18/former-snap-engineer-launches-butterflies-a-social-network-where-ais-and-humans-coexist/

🩺 OpenAI and Color Health join forces to accelerate cancer treatment

In collaboration with OpenAI, Color Health has developed a copilot application that leverages OpenAI’s GPT-4.o model to assist clinicians in accelerating cancer treatment. Here’s how it works:

  • Integration of medical data: The copilot integrates patient medical data with clinical knowledge using OpenAI’s APIs. It processes and normalizes vast unstructured data such as family history, individual risk factors, and clinical guidelines.
  • Tailored treatment plans: Based on this data, the copilot generates customized, comprehensive treatment plans. These plans answer critical questions like “What screenings should the patient be doing?” and identify missing diagnostics.
  • Clinician-in-the-loop workflow: A clinician evaluates the copilot’s output at each step. They can modify it if necessary before presenting it to the patient.
  • Assured privacy: Color Health ensures HIPAA compliance and maintains rigorous privacy standards while providing clinician oversight to ensure high-quality care.

Why does it matter? 

The tool identifies missing diagnostics and expedites the analysis of complex medical records – a process that can now be completed in just 5 minutes rather than hours or weeks. This not only improves access to critical expertise but also has the potential to catch cancer or pre-cancerous conditions earlier, enabling faster treatment and better patient outcomes.

Source: https://openai.com/index/color-health

🎤 Meta unveils new AI models for audio, text, and watermarking

Meta’s Fundamental AI Research (FAIR) team has recently released several new AI models and tools for researchers to use. Here are the highlights:

  • JASCO: This AI model, short for “Joint Audio and Symbolic Conditioning for Temporally Controlled Text-to-Music Generation,” allows users to adjust features of generated sound (such as chords, drums, and melodies) through text inputs. FAIR plans to release the JASCO inference code under an MIT license and the pre-trained model under a non-commercial Creative Commons license.
  • AudioSeal: This tool adds watermarks to AI-generated speech. It’s designed specifically for localized detection of AI-generated segments within longer audio snippets and will be released with a commercial license.

Chameleon: FAIR will release two sizes of its multimodal text model, Chameleon (7B and 34B), under a research-only license. These models can handle tasks requiring visual and textual understanding, such as image captioning.

Why does it matter?

By making these tools publicly available, Meta encourages collaboration within the AI community, which could lead to groundbreaking applications in various fields, such as music creation, sound design, and content generation. Moreover, the introduction of AudioSeal addresses the growing need to identify and attribute AI-generated content, promoting transparency and accountability in an increasingly AI-driven world.

Source: https://venturebeat.com/ai/meta-releases-flurry-of-new-ai-models-for-audio-text-and-watermarking

⚒️Notion introduces AI Connectors for Slack & Google Drive integration

With AI Connectors, users can query Notion and receive responses from connected apps. The integration allows for selecting specific public channels or all public channels for information retrieval, excluding private channels and direct messages.

Notion AI can access up to a year of historical Slack content, which may take up to 36 hours. Additionally, users can set up various interactions between Notion and Slack, such as sending Slack messages to a Notion database, creating Notion tasks directly from Slack, and receiving notifications in Slack for specific events in Notion

Why does it matter?

This development demonstrates the growing trend of incorporating AI capabilities into mainstream productivity and collaboration platforms. The ability to intelligently pull in and surface relevant information from external platforms helps streamline workflows, enhance productivity, and improve the overall efficiency of the team.

Source: https://x.com/NotionHQ/status/1803096151526137934

🧠 AI model predicts anxiety levels

Researchers at the University of Cincinnati just developed a new AI system that can predict anxiety levels using a short picture rating task and a small set of demography and psychological variables.

  • Comp Cog AI combines principles from psychology with machine learning to measure human decision-making patterns related to anxiety.
  • Participants rate a series of emotion-evoking pictures and answer a few contextual questions about their age, income, employment, etc.
  • The model extracts 15 ‘judgment variables’ that capture patterns in how users weigh positive and negative outcomes in decision-making.
  • In tests, the model predicted whether participants had higher or lower anxiety with up to 81% accuracy.

While anxiety disorders are extremely prevalent, they’re also often challenging and time-consuming to diagnose. With AI’s ability to connect data patterns and achieve reliable predictions, the tech could soon enable more efficient screenings to streamline assessment and treatment time.

Source: https://www.nature.com/articles/s44184-024-00074-x.epdf

What Else Is Happening in AI on June 19th 2024❗

🧠 Google transitions DeepMind from research to AI product development

This transition involves merging Google Research’s Brain team with DeepMind, creating a unified group called Google DeepMind. The aim is to enhance the development of advanced AI technologies responsibly. However, this move has led to frustration among some researchers due to new guidelines and the pressure to commercialize AI technologies.

Source: https://www.pymnts.com/google/2024/google-reportedly-moving-deepmind-from-research-to-ai-products

📽️ ElevenLabs’ new open-source AI tool adds sound effects to videos

The app extracts four frames from the video and sends them with a prompt to OpenAI’s GPT-4 to generate a text-to-sound effects prompt. This prompt is used by ElevenLabs’s Sound Effects API to create the sound effect, which is then combined with the video into a single file up to 22 seconds long, ready for download.

Source: https://venturebeat.com/ai/elevenlabs-unveils-open-source-creator-tool-for-adding-sound-effects-to-videos

🤖 Snap unveils advanced gen AI for enhanced AR effects

Snap’s latest iteration of gen AI allows users to see more realistic special effects. It plans to create full-body AR experiences rather than just facial AR experiences, such as generating a new outfit. It also announced an upgraded version of its developer program, Lens Studio, which lets artists and developers create AR features for Snapchat or other websites and apps.

Source: https://www.reuters.com/technology/snap-launches-ai-tools-advanced-augmented-reality-2024-06-18

🎨 Hedra Labs launches Character-1, a multimodal creation studio preview

Hedra Labs has launched a preview of Character-1, a model for creating expressive characters that talk, sing, and rap. It offers infinite duration (30s in open preview) and generates 90s content per 60s of input, hardware permitting. This is part of Hedra’s mission to give creators control over dialogue, movement, and virtual worlds.

Source: https://x.com/hedra_labs/status/1803095713112580475

🦋 Ex-Snap engineer launches AI-human social network, Butterflies

The platform allows users to create AI avatars, called “Butterflies,” that can engage in conversations, generate images, and participate in social activities like human users. The app offers a range of features, including creating and customizing AI characters and exploring a feed filled with AI-generated and human-generated content.

Source: https://techcrunch.com/2024/06/18/former-snap-engineer-launches-butterflies-a-social-network-where-ais-and-humans-coexist

🎺 TikTok brings AI to content creation

TikTok just launched Symphony, a new suite of AI features to help brands and creators produce content more efficiently on the platform — including digital avatars, translation tools, an AI assistant, and more.

  • Symphony Digital Avatars enable brands to generate AI spokespeople to star in their TikTok ads and branded content across multiple languages.
  • Brands can choose from a selection of ‘stock avatars’ based on real actors or create custom avatars to serve as virtual brand reps.
  • A new AI Dubbing tool automatically transcribes, translates, and dubs videos into 10+ languages, helping brands scale content globally.
  • TikTok’s Creative Assistant has been rebranded to Symphony Assistant, a chatbot to help with brainstorming, writing scripts, and optimizing content for the platform.

The creator economy is about to be completely remade with AI, and these tools will unlock new opportunities for optimized content, reach, and creation — but the rise of digital ‘spokespeople’ will also raise new questions on authenticity and consumer trust that brands will need to grapple with.

Source: https://www.tiktok.com/business/en-US/blog/tiktok-symphony-ai-creative-suite

AI Jobs on June 9th 2024:

  • 🎉 Anthropic – Head of Events: https://jobs.therundown.ai/jobs/63479075-head-of-events
  • 📚 Tempus – Analyst, NEXT Strategy & Operations: https://jobs.therundown.ai/jobs/63500064-analyst-next-strategy-operations
  • 🎨 Fiddler AI – Staff UI/UX Designer: https://jobs.therundown.ai/jobs/63473287-staff-ui-ux-designer
  • 👥 Glean – GTM Recruiter: https://jobs.therundown.ai/jobs/63498594-gtm-recruiter

A  Daily chronicle of AI Innovations June 18th 2024:

🩻 OpenAI expands into healthcare with AI tool for cancer screening

🛰️ NATO invests $1.1 billion in AI, space, and robotics defence tech

🍎 Apple halts work on Vision Pro

🎬 Google DeepMind’s new AI can generate soundtracks for videos
🌟 Runway launches new model Gen-3 Alpha
🚀China’s DeepSeek Coder V2 beats GPT-4 Turbo

🔊 DeepMind creates sound for videos

🍎 Apple halts work on Vision Pro

  • Apple has paused development on the second-generation Vision Pro headset to concentrate on a more affordable version, according to The Information.
  • The company reportedly began working on this cheaper Vision device, codenamed “N109,” in 2022, aiming to sell it for a price similar to a high-end iPhone.
  • Even though Apple plans to use the high-end display components from the Vision Pro in the budget model, it still faces challenges in reducing costs without removing too many features.
  • Source: https://www.macrumors.com/2024/06/18/apple-suspends-work-on-vision-pro-2/

🩻 OpenAI expands into healthcare with AI tool for cancer screening

  • OpenAI has partnered with Color Health to integrate artificial intelligence into cancer screening and treatment procedures using their GPT-4o model.
  • The AI assistant developed by Color Health processes patient data to create personalized cancer screening strategies and informs doctors of missing diagnostic tests.
  • The AI tool also helps in assembling a cancer pretreatment work-up, expediting the process that can take weeks or months, and thus potentially reducing delays in treatment.
  • Source: https://www.newsbytesapp.com/news/science/openai-partners-with-color-health-for-cancer-copilot/story

🛰️ NATO invests $1.1 billion in AI, space, and robotics defence tech

  • NATO has started its €1 billion investment in defence technology, focusing on AI, space, and robotics, by funding four European tech firms and four venture capital funds.
  • The NATO Innovation Fund, established in 2022, aims to address contemporary defence and security challenges by matching government buyers with start-ups to develop new technologies.
  • Initial recipients of the fund include Germany’s ARX Robotics, UK’s Fractile AI, iComat, and Space Forge, with the broader aim of promoting NATO’s technological sovereignty and innovation ecosystems.
  • Source: https://www.euronews.com/next/2024/06/18/nato-begins-its-1-billion-investment-in-ai-space-and-robotics-defence-tech

🎬 Google DeepMind’s new AI can generate soundtracks for videos

DeepMind is developing video-to-audio (V2A) technology to generate rich soundtracks for silent videos generated by AI models. V2A combines video pixels with natural language text prompts to create synchronized audiovisual content. The technology offers enhanced creative control, allowing users to guide the audio output using positive and negative prompts.

What sets DeepMind’s V2A apart is its ability to understand raw pixels and generate audio without manual alignment. However, V2A struggles with artifacts or distortions in videos and generates audio that is not super convincing. As DeepMind continues to gather feedback from creators and filmmakers, they remain committed to developing this technology responsibly.

Why does it matter?

The technology could help revive and enhance historical footage, silent films, and other archival material. However, generative AI tools like V2A also threaten to disrupt the film and TV industry, potentially eliminating jobs without strong labor protections.

Source: https://deepmind.google/discover/blog/generating-audio-for-video

🌟 Runway launches new model Gen-3 Alpha

Runway launched Gen-3 Alpha, its latest AI model for generating video clips from text descriptions and still images. Gen-3 Alpha excels at generating expressive human characters with a wide range of actions, gestures, and emotions and can interpret various styles and cinematic terminology. However, it has limitations, including a maximum video length of 10 seconds, and struggles with complex character and object interactions and following the laws of physics precisely.

Runway partnered with entertainment and media organizations to create custom versions of Gen-3 for more stylistically controlled and consistent characters, targeting specific artistic and narrative requirements. They also have implemented safeguards, such as a moderation system to block attempts to generate videos from copyrighted images and a provenance system to identify videos coming from Gen-3.

Why does it matter?

As competition in AI video generation heats up, Runway’s Gen-3 Alpha empowers artists and filmmakers to create high-quality, controllable videos with ease, pushing the boundaries of storytelling and creative possibilities.

Source: https://runwayml.com/blog/introducing-gen-3-alpha

🚀China’s DeepSeek Coder V2 beats GPT-4 Turbo

Chinese AI startup DeepSeek has announced the release of DeepSeek Coder V2, an open-source code language model. It is built upon the DeepSeek-V2 MoE model and excels at coding and math tasks, supporting over 300 programming languages. It outperforms state-of-the-art closed-source models like GPT-4 Turbo, Claude 3 Opus, and Gemini 1.5 Pro, making it the first open-source model to achieve this feat. DeepSeek Coder V2 also maintains comparable performance in general reasoning and language capabilities.

The model is being offered under an MIT license, which allows for research and unrestricted commercial use. It can be downloaded or accessed via API on DeepSeek’s platform.

Why does it matter?

DeepSeek aims to “unravel the mystery of AGI with curiosity” and has quickly emerged as a notable Chinese player in the AI race. As it only costs $0.14/1M tokens(input) and $0.28/1M tokens(output), it will give notable models like GPT-4 Turbo intense competition.

Source: https://github.com/deepseek-ai/DeepSeek-Coder-V2/blob/main/paper.pdf

🔊 New AI Jobs Opportunities on June 18th 2024

  • 🎨 Glean – UX Designer: https://jobs.therundown.ai/jobs/63229234-ux-designer
  • 🔬 Fiddler AI – Staff AI Scientist: https://jobs.therundown.ai/jobs/63187617-staff-ai-scientist
  • 📋 Notable – Product Operations Manager: https://jobs.therundown.ai/jobs/63204832-product-operations-manager
  • 📊 Findem – Data Analyst: https://jobs.therundown.ai/jobs/63102512-data-analyst

What Else Is Happening in AI on June 18th 2024❗

🔍 Perplexity now displays weather, currency conversion, and simple math directly through cards

This move aims to keep users from going to Google for such results. Perplexity’s CEO, Aravind Srinivas, acknowledged that Google handles basic queries like weather, time, and live sports scores well, and his company had work to do in that area.

Source: https://techcrunch.com/2024/06/17/perplexity-now-displays-results-for-temperature-currency-conversion-and-simple-math-so-you-dont-have-to-use-google

🛡️ U.S. government and private sector ran the first AI attack simulation

Federal officials, AI model operators, and cybersecurity companies ran the first joint simulation of a cyberattack on a critical AI system. It also involved experts from private sector companies like Microsoft, Nvidia, and OpenAI. It helped identify potential new threats and establish communication channels between the government and the private sector.

Source: https://www.axios.com/2024/06/17/cisa-tech-companies-ai-cyberattack-simulation

🚀 Adobe Acrobat got a major upgrade, bringing AI to PDFs and more

Adobe Firefly generative AI enables image generation and editing directly within Acrobat for the first time. Acrobat AI Assistant’s new features, “insights across documents” and “enhanced meeting transcripts,” help users extract insights and share information from various document types. Adobe is offering free, unlimited access to Acrobat AI Assistant from June 18 to June 28.

Source: https://blog.adobe.com/en/publish/2024/06/17/adobe-acrobat-reimagines-documents-multi-format-ai-powered-work

🤖 TikTok introduces gen AI avatars of creators and stock actors for ads

“Custom Avatars” allow creators to scale their likeness for multilingual avatars and brand collaborations, while brands can use pre-built “Stock Avatars” to add a human touch. Plus, the new “AI Dubbing” tool translates content into ten languages, helping creators and brands increase their global reach.

Source: https://newsroom.tiktok.com/en-us/announcing-symphony-avatars

🧱 Pixelbot 3000 builds Lego art using simple AI prompts

YouTuber Creative Mindstorms designed and built the Pixelbot 3000, a Lego printer that automates the assembly of brick-built mosaics. It uses OpenAI’s DALL-E 3 to generate images based on simple text prompts. First it generates a simplified cartoon-style image, then it is divided into a 32 x 32 grid, and the color of the center pixel in each square is sampled to create a high-contrast scaled image for the mosaic.

Source: https://www.theverge.com/2024/6/17/24180250/lego-printer-pixel-art-artificial-intelligence

Adobe integrated new Firefly AI capabilities into Acrobat, allowing users to create and edit images within PDFs using text prompts — also adding the ability to access an AI assistant for insights, content creation, and more. https://news.adobe.com/news/news-details/2024/Adobe-Reimagines-Acrobat-Bringing-Firefly-AI-to-PDFs-and-Expanding-Use-Across-More-Document-Types/default.aspx
Reuters Institute of Journalism published a new report finding growing public wariness of AI-generated news content, with many expressing discomfort about its potential impact on content reliability and trust. https://www.reuters.com/technology/artificial-intelligence/global-audiences-suspicious-ai-powered-newsrooms-report-finds-2024-06-16
The U.S. Navy is deploying AI-powered underwater drones to better detect threats, with plans to expand the tech’s use in identifying enemy ships and aircrafts. https://finance.yahoo.com/news/ai-equipped-underwater-drones-helping-153947268.html
Luma teased new control features coming to its Dream Machine video model, including the ability to quickly change scenes and precisely edit characters — also launching the ability to extend video and remove watermarks. https://x.com/LumaLabsAI/status/1802678788490494128
Anthropic published new research showing that AI models can engage in ‘reward tampering’, learning to cheat the system and grant higher rewards even without specific training. https://www.anthropic.com/research/reward-tampering

A  Daily chronicle of AI Innovations June 17th 2024:

🍔 McDonald’s pauses AI drive-thru tests

🤖 Chinese startup creates realistic robots

🤷‍♂️ Photographer wins AI-image contest with real picture, gets disqualified

💻 NVIDIA’s AI model for synthetic data generation rivals GPT-4
⚠️ Meta pauses AI model training in EU due to regulatory pushback
🎵 Spotify launches ‘Creative Labs’ to test Gen AI voiceover ads

🏅 Tiny LLM matches GPT-4 on math

🍔 McDonald’s pauses AI drive-thru tests

  • McDonald’s is ending its artificial intelligence drive-thru order-taking program and will shut it off in all participating restaurants by July 26, 2024.
  • The decision to end the AI program came after a thorough review, and McDonald’s Chief Restaurant Officer stated the company will seek broader voice ordering solutions.
  • Although the AI drive-thru program showed successes, McDonald’s aims to simplify operations and speed up service, with plans to decide on a new voice ordering solution by the end of the year.
  • Source: https://www.foxbusiness.com/technology/mcdonalds-ending-ai-drive-thru-ordering

🤷‍♂️ Photographer wins AI-image contest with real picture, gets disqualified

  • A photographer named Miles Astray won an AI-image competition with a real photograph of a head-scratching flamingo, resulting in his disqualification.
  • His entry, titled F L A M I N G O N E, was mistakenly awarded third place by judges from notable institutions like the New York Times and Christie’s.
  • Astray aimed to demonstrate that human creativity and natural beauty surpass AI-generated content, sparking a discussion about the ethical implications of using AI in art.
  • Source: https://www.techspot.com/news/103422-photographer-wins-ai-image-contest-real-picture-gets.html

NVIDIA’s AI model for synthetic data generation rivals GPT-4

NVDIAI has released Nemotron-4 340B, an open-source pipeline for generating high-quality synthetic data. It includes a base model trained on 9M tokens, an instruction, and a reward model.

  • The instruction model can generate diverse synthetic data that mimics real-world data.
  • The reward model then evaluates the generated data to filter out high-quality responses.
  • This interaction between the two models produces better training data over time.

Note: 98% of the training data used to fine-tune the Instruct model is synthetic and was created using NVIDIA’s pipeline.

In benchmarks such as MT-Bench, MMLU, GSM8K, HumanEval, and IFEval, the Instruct model generally performs better than other open-source models such as Llama-3-70B-Instruct, Mixtral-8x22B-Instruct-v0.1, and Qwen-2-72B-Instruct, and in some tests, it even outperforms GPT-4o.

It also performs comparable to or better than OpenAI’s GPT-4-1106 in human evaluation for various text tasks, such as summaries and brainstorming. The technical report provides detailed benchmarks.

Why does it matter?

This development allows businesses to create powerful, domain-specific LLMs without the need for extensive, costly real-world datasets. It has significant potential impacts across various industries, such as healthcare (drug discovery, personalized medicine, medical imaging), finance (fraud detection, risk assessment, customer service), manufacturing (predictive maintenance, supply chain optimization), and retail (personalized customer experiences).

Source: https://blogs.nvidia.com/blog/nemotron-4-synthetic-data-generation-llm-training

⚠️ Meta pauses AI model training in EU due to regulatory pushback

In response to the regulatory pressure from the Irish Data Protection Commission and the UK’s Information Commissioner’s Office, Meta has decided to pause its plans to train its large language model, Llama, using public content shared by Facebook and Instagram users in the European Union and the UK.

The regulators expressed concerns about Meta’s plan to use this user-generated content to train its AI systems without obtaining explicit user consent. Meta relied on a GDPR provision called “legitimate interests” to justify this data usage, but the regulators felt this was insufficient. Meta has decided to delay the launch of its AI chatbot in Europe until it can address the regulators’ concerns and establish a more transparent user consent process.

Why does it matter?

Meta’s inability to use EU user data for AI training is a setback for its regional AI ambitions. It could disadvantage Meta against competitors who can leverage such data. This situation highlights the ongoing tensions between tech companies’ desire to utilize consumer data for AI development and regulators’ efforts to protect user privacy. Striking the right balance between innovation and privacy will be a major challenge as the AI race intensifies.

Source: https://www.bloomberg.com/news/articles/2024-06-14/meta-delays-ai-chatbot-launch-in-europe-after-regulator-pushback

🎵 Spotify launches ‘Creative Labs’ to test Gen AI voiceover ads

Spotify has launched a new in-house creative agency called “Creative Lab.” This agency will help brands and advertisers create custom campaigns for Spotify’s platform. Creative Lab teams in different markets will provide local insights and collaborate with brands to develop campaigns through workshops, inspiration sessions, and collaborative ideation.

In addition, Spotify is also testing a new AI tool called “Quick Audio” that will allow brands to create scripts and voiceovers using generative AI technology. This new capability will be integrated into Spotify’s ad manager platform, giving advertisers more options to produce audio ads for Spotify’s audience of over 615 million listeners.

Why does it matter?

This move emphasizes Spotify’s ambition to become a full-service advertising platform. Marketers and advertisers will have new creative and production capabilities available through Spotify to better reach the platform’s large and engaged user base in unique ways, including potentially using AI-generated audio ads. This could disrupt traditional advertising models and open new possibilities for how brands connect with consumers on audio platforms.

Source: https://techcrunch.com/2024/06/13/spotify-creative-labs-ad-agency-for-advertisers

🤖 Chinese startup creates realistic robots

A Chinese startup called Ex-Robots is developing ultra-realistic humanoid robots that can closely mimic human facial expressions and emotions.

  • Ex-Robots’ lifelike humanoid robots can smile, frown, and replicate subtle facial movements using small motors embedded in the silicone faces.
  • The startup uses multimodal AI and algorithms to enable the robots to recognize and express emotions based on perceiving their environment.
  • Creating one humanoid robot currently takes 2-4 weeks, with costs as high as $275k per unit.
  • The company said near-term use cases include museum exhibits, with eventual roles in healthcare, therapy, and child services.

The advancement of ultra-realistic robots combined with the power of AI is about to create some bizarre new frontiers. Will humans be creeped out by an actual human-looking humanoid over mechanical options like Tesla’s Optimus? It’s unclear if the world is ready… but we’ll find out soon.

Source: https://www.reuters.com/technology/chinas-ex-robots-develops-humanoids-with-enhanced-facial-movement-2024-06-11

🏅Tiny LLM matches GPT-4 on math

Shanghai AI Laboratory researchers just developed a new algorithm called MCT Self-Refine, which allowed a relatively small 8B-parameter model to achieve GPT-4 level performance on complex math.

  • MCT Self-Refine combines an 8B LLaMa model with Monte Carlo Tree Search, an AI technique famously used by Google DeepMind to master the game of Go.
  • The algorithm constructs a search tree, refining answers through a series of processes and self-evaluation.
  • On the GSM-Hard math word problem benchmark, infusing the Monte Carlo Tree Search took the 8B model from 25.5% accuracy to 45.5%.
  • The model also achieved 96.7% on the GSM8K benchmark, outperforming significantly larger models like GPT-4, Claude, and Gemini.

 The research shows it’s possible to boost math capabilities without massive scale — and GPT-4 level performance with a model trained on 200x less parameters is an impressive feat. If the approach proves to be a more efficient path to advanced reasoning, we could be on the cusp of a new wave of model acceleration.

Source: https://arxiv.org/pdf/2406.07394

New AI Jobs on June 17th 2024

  • 🎨 Abridge – Senior Product Designer: https://jobs.therundown.ai/jobs/63255899-senior-product-designer
  • 💻 Palantir Technologies – Software Engineer, New Grad: https://jobs.therundown.ai/jobs/63248966-software-engineer-new-grad-us-government
  • 🧪 Fiddler AI – Staff AI Scientist: https://jobs.therundown.ai/jobs/63187617-staff-ai-scientist
  • 🤖 Meta – Software Engineer – ML Systems: https://jobs.therundown.ai/jobs/63275347-software-engineer-ml-systems-technical-leadership

 

What Else Is Happening in AI on June 17th 2024❗

🍎 Apple enters the AI icon race to find a logo that makes sense

Apple has joined other tech giants like Google, OpenAI, Anthropic, and Meta in the race to find an iconic visual representation for AI. No company has yet created an unambiguous “AI logo” that conveys the concept to users. AI’s lack of a clear visual identity reflects the difficulty of representing such a broad and evolving technology in a simple icon.

Source: https://techcrunch.com/2024/06/15/apple-joins-the-race-to-find-an-ai-icon-that-makes-sense

📝 Niloom.AI launches gen AI content creation platform for spatial computing

Without extensive technical expertise, the platform allows users to create, prototype, edit, and instantly publish sophisticated AR/VR content using text or speech prompts. It consolidates the entire creative process, from ideation to publishing, and integrates with various third-party tools to provide a one-stop solution for spatial computing content creation.

Source: https://venturebeat.com/ai/niloom-ai-launches-one-stop-generative-ai-content-creation-platfor-for-spatial-computing

🏟️ AI to delete abusive posts against athletes during the 2024 Paris Olympics

The International Olympic Committee (IOC) will deploy AI at the 2024 Paris Olympics to automatically detect and erase abusive social media posts directed at athletes and officials. The AI tool will monitor posts about 15,000 athletes and officials and immediately remove any content involving hate speech, bullying, or political attacks.

Source: https://www.reuters.com/sports/olympics/ai-erase-abusive-posts-athletes-paris-olympics-ioc-2024-06-14

🖼️ Picsart and Getty team up to counter Adobe’s “commercially-safe” AI

Picsart has partnered with Getty Images to develop a “responsible, commercially-safe” AI image generator tool. The AI model will be trained exclusively on Getty’s licensed stock content to address concerns about AI-generated content violating copyright laws. Picsart hopes to provide a viable alternative to Adobe’s Firefly by leveraging Getty’s library of licensed images. Source: https://www.theverge.com/2024/6/14/24178333/picsart-getty-commercially-safe-ai-image-tool-adobe

📰 Yahoo News gets an AI-powered revamp with Artifacts integration 

Yahoo has acquired the technology behind the Artifact news aggregation app and is launching a new AI-powered Yahoo News app. The app will feature a personalized news feed based on user interests and a “Key Takeaways” feature that provides bullet-point summaries of articles. Users can also flag problematic content, which the AI will then try to rewrite.

Source: https://www.theverge.com/2024/6/13/24177980/yahoo-news-app-launch-artifact-ai-architecture

A  Daily chronicle of AI Innovations June 14th 2024:

📊 Google’s PH-LLM reads your wearables’ data for personalized insights
🔮 Ex-OpenAI researcher on what to expect from AI in next decade
🧠 DeepMind built ‘a virtual rodent’ with AI to understand brain activity

🗳️ AI chatbots run for office

💼LinkedIn is rolling out new AI-powered features for premium users

🌍Synthflow’s AI voice assistants are now multilingual!

🖼️Picsart is partnering with Getty Images to develop a custom model for AI imagery

🕵️‍♂️Former head of NSA joins OpenAI’s Safety and Security Committee

📊 Google’s PH-LLM reads your wearables’ data for personalized insights

Building on the next-gen capabilities of Gemini models, Google has presented research that highlights two complementary approaches to providing accurate personal health and wellness information with LLMs.

The first introduces PH-LLM, a version of Gemini fine-tuned to understand and reason on time-series personal health data from wearables such as smartwatches and heart rate monitors. The model answered questions and made predictions noticeably better than experts with years of experience in the health and fitness fields.

In the second paper, Google introduces an agent system that leverages state-of-the-art code generation and information retrieval tools to analyze and interpret behavioral health data from wearables. Combining these two ideas will be critical for developing truly personalized health assistants.

Why does it matter?

Wearables generate a wealth of personal health data that is rarely utilized in clinical settings. Integrating this data with advanced AI models could revolutionize personal health management and preventative care by putting an “expert health assistant” on everyone’s wrist.

Source: https://research.google/blog/advancing-personal-health-and-wellness-insights-with-ai

🔮 Ex-OpenAI researcher on what to expect from AI in next decade

A researcher fired from OpenAI, Leopold Aschenbrenner,  published a 165-page essay on what to expect from AI in the next decade. And GPT-4 has summarized it! Here are some key takeaways from the essay:

  • By 2027, AI models could reach the capabilities of human AI researchers and engineers, potentially leading to AI surpassing human intelligence
  • Trillions of dollars are being invested into developing the infrastructure needed to support these AI systems
  • Controlling AI systems smarter than humans(the ‘superalignment‘ problem) will be crucial to prevent catastrophic outcomes
  • Only a few hundred people truly understand the scale of change AI is about to bring

Why does it matter?

The essay provides a rare insider’s perspective on the rapid progression of AI. Coming from someone deeply involved in cutting-edge AI development, the insights highlight the urgency to get ahead of managing risks before AI’s capabilities outpace our defenses.

Source: https://ca.news.yahoo.com/researcher-fired-openai-published-165-184227878.html

🧠 DeepMind built ‘a virtual rodent’ with AI to understand brain activity

Researchers from Google DeepMind and Harvard built a ‘virtual rodent’ powered by AI to help them better understand how the brain controls movement. With deep reinforcement learning (RL), it learned to operate a biomechanically accurate rat model, allowing researchers to compare real and virtual neural activity.

Why does it matter?

Understanding how the brain controls movement and modeling neural activity could exponentially advance fields like neuroscience and brain-computer interfaces, with the help of AI.

Source: https://www.nature.com/articles/s41586-024-07633-4

🗳️ AI chatbots run for office

An AI-powered candidate named ‘AI Steve’ is running for U.K. Parliament in next month’s general election — creating polarizing questions around AI’s use in government affairs.

  • AI Steve is represented by businessman Steve Endacott and will appear as an independent candidate in the upcoming election.
  • Voters can interact with AI Steve online to ask policy questions and raise concerns or suggestions, which the AI will incorporate based on feedback.
  • If elected, Endacott will serve as AI Steve’s human proxy in Parliament, attending meetings and casting votes based on the AI’s constituent-driven platform.

The idea of an AI running for office might sound like a joke, but the tech behind it could actually help make our politicians more independent and (ironically) autonomous. AI-assisted governance is likely coming someday, but it’s probably still a bit too early to be taken seriously.

Source: https://www.nbcnews.com/tech/tech-news/ai-candidate-running-parliament-uk-says-ai-can-humanize-politics-rcna156991

🎤 OpenAI CTO speaks on internal models

OpenAI CTO Mira Murati participated in a Q&A with Fortune Magazine, addressing topics including criticism from Elon Musk, internal AI model strength, the 2023 board drama and more.

  • Murati said both Apple and OpenAI believe strongly in privacy and will not log data through Apple accounts or train models on user data.
  • Murati responded to Musk calling the Apple partnership ‘creepy spyware’, saying OpenAI is trying to be as transparent as possible with its approach to safety.
  • When asked about the board drama, Murati said the previous board structure ‘didn’t have accountability to anyone but themselves’.
  • The CTO also curiously said that OpenAI’s internal models ‘aren’t that far ahead’ of what the public has for free right now.

This had to be a harder-hitting Q&A than Murati expected — but her responses certainly evoked similar vibes to the infamous Sora interview. The commentary on internal models was particularly interesting, as it would imply that OpenAI doesn’t have some GPT-5 level model waiting in the wings.

Source: https://x.com/elonmusk/status/1800266437677768765

🐀 DeepMind creates a virtual rat with AI brain

Researchers from Google DeepMind and Harvard just created a virtual rodent powered by an AI neural network, capable of mimicking agile movements and neural activity of real-life rats with extreme accuracy.

  • The researchers created a bio-mechanically realistic digital rat model and trained an AI ‘brain’ to control its body in a physics simulator.
  • The AI’s brain activity patterns closely matched real rodents engaged in similar behaviors, showing it can capture principles of muscle control and movement.
  • This ‘virtual neuroscience’ approach could open new research with testing on AI animals, also transferring to engineering of advanced robotics systems.

By bridging the gap between motor control and the complexity of brains, these virtual rodents offer a new window into how neural activity generates movement. Our understanding of the brain and neural activity is about to expand exponentially with the increasing power of AI.

Source: https://www.nature.com/articles/s41586-024-07633-4

New AI Job Opportunities on June 14th 2024

  • 💼 Databricks – Sr. Product Manager: https://jobs.therundown.ai/jobs/63046708-sr-product-manager
  • 🧪 Luma AI – Research Scientist: https://jobs.therundown.ai/jobs/62973732-research-scientist
  • 📊 Dataiku – Senior Sales Engineer: https://jobs.therundown.ai/jobs/63061675-senior-sales-engineer
  • 📘 Meta – GenAI Content Manager: https://jobs.therundown.ai/jobs/60547830-genai-content-manager

What Else Is Happening in AI on June 14th 2024❗

🕵️‍♂️Former head of NSA joins OpenAI’s Safety and Security Committee

Paul M. Nakasone, a retired US Army general and a former head of the National Security Agency (NSA), will also join OpenAI’s board of directors. He will contribute to OpenAI’s efforts to better understand how AI can be used to strengthen cybersecurity by quickly detecting and responding to cybersecurity threats.

Source: https://www.theverge.com/2024/6/13/24178079/openai-board-paul-nakasone-nsa-safety

🤖Former Meta engineers launch Jace, your new autonomous AI employee

Jace uses Zeta Labs’ proprietary web-interaction model, Autonomous Web Agent-1, to use a browser to interact with websites like any human would. It allows it to handle real-world tasks like booking flights, handling hiring, or even setting up a company.

Source: https://venturebeat.com/ai/exclusive-former-meta-engineers-launch-jace-an-ai-agent-that-works-independently

💼LinkedIn is rolling out new AI-powered features for premium users

The features include searching for jobs by prompting in natural language, building a cover letter from scratch, reviewing your résumé with personalized suggestions for improving it for a specific job post, and making edits interactively with AI.

Source: https://www.theverge.com/2024/6/13/24177986/linkedin-ai-job-hunting-features-premium-subscribers

🌍Synthflow’s AI voice assistants are now multilingual!

They can fluently communicate in Spanish, German, Portuguese, French, and English. Sythflow also added corresponding voices for each language to ensure authentic and natural-sounding interactions so businesses can engage a global audience and offer personalized experiences.

Source: https://x.com/GoogleDeepMind/status/1801210428673892496

🖼️Picsart is partnering with Getty Images to develop a custom model for AI imagery

The model will be built from scratch and trained exclusively on Getty Images’ licensed creative content. It will bring responsible AI imagery to creators, marketers, and small businesses that use Picsart, enabling them to generate unique images with full commercial rights. Source: https://techcrunch.com/2024/06/13/picsart-partners-with-getty-images-to-develop-a-custom-ai-model/

A  Daily chronicle of AI Innovations June 13th 2024:

🍎 Apple to ‘pay’ OpenAI through distribution, not cash

📹 New free rival to OpenAI Sora unveiled

⌚️ Google AI gets personal with wearables

📰 Perplexity was planning revenue-sharing deals with publishers when it came under media fire

🗣️ Amazon ‘dropped the ball’ with Alexa, former employee says

🤔 Tim Cook says it’s a ‘significant possibility’ people use their iPhones less because of AI

🤖Luma AI’s dream machine debut sparks AI media wars
🖼️Stability AI’s Diffusion 3 Shakes Up Image Creation
💰No Cash Involved! Apple will pay OpenAI in distribution 

🏆NVIDIA tops MLPerf benchmarks, dominates AI training

🛠️Databricks Boosts Mosaic AI for Enterprise LLMs

💻ChromeOS gets GenAI boost via Android tech

🎬YouTube trials AI comment summaries for shorts

🤖Luma AI’s dream machine debut sparks AI media wars

LumaAI has launched Dream Machine, an AI-powered video generation system capable of generating high-quality videos via simple text prompts. The open-for-all AI tool takes about two minutes to generate a realistic five-second video clip to render specified objects and environments without compromising coherency.

Why does it matter?

The open-for-all feature of this AI tool highlights Luma AI’s approach toward democratizing generative AI and puts rival systems into the spotlight, including OpenAI’s Sora and Kuaishou’s Kling, which are only accessible to select users.

Additionally, Luma’s open approach will enable creators and companies to produce original content at a rapid speed and scale, giving it a competitive edge.

Source: https://lumalabs.ai/dream-machine

⌚️ Google AI gets personal with wearables

Google just published two new research papers, introducing an AI model that can interpret personal health data from wearable devices and an AI agent that provides detailed insights, answer health questions, and more.

  • Google’s PH-LLM is a version of Gemini fine-tuned to reason using data from wearables and generate coaching insights for sleep and fitness.
  • PH-LLM performs comparably to human experts on health insights for fitness and sleep, also achieving expert performance on certification exams.
  • Google also revealed PHIA, an AI agent that combines Gemini’s language skills with code and search capabilities to analyze wearable health data.
  • PHIA scored 84% on health insight questions, demonstrating strong reasoning and data analysis capabilities on complex queries.

Google’s research showcases AI’s potential to take health wearable tracking to the next level — going beyond generic insights to enable personalized support. Once ingrained into popular devices, these types of AI tools will put an expert fitness and nutrition coach on the wrists of every user.

Source: https://research.google/blog/advancing-personal-health-and-wellness-insights-with-ai

🖼️Stability AI’s Diffusion 3 Shakes Up Image Creation

Stability AI has launched its most sophisticated text-to-image AI open model, optimized for consumer PCs and enterprise-tier GPUs. The model delivers several noteworthy features:

  • Enables photorealistic and high-quality outputs in exceptional style by providing images with extraordinary color, lighting, and details.
  • Can comprehend long, complex prompts that involve spatial reasoning, compositional elements, actions, and styles.
  • Leverages Diffusion Transformer architecture to achieve exceptional text quality, reducing spelling, character spacing, and letter-forming errors.
  • Displays the ability to absorb minute details from small datasets, making it an ideal model for customizing and fine-tuning.

Why does it matter?

Stability’s unique features make it the most advanced text-to-image AI model. As it continues to push the boundaries of generative AI, it will be interesting to note whether it will stand out against its competitors and remain at the forefront of image generation.

Source: https://stability.ai/news/stable-diffusion-3-medium

💰No Cash Involved! Apple will pay OpenAI in distribution 

A landmark arrangement between OpenAI and Apple to integrate ChatGPT into iPhone, iPad, and Mac was unclear in its financial terms. People briefed on the matter suggest that the agreement isn’t likely to generate revenue for either party and simply provides OpenAI brand exposure to Apple’s user base while offering Apple an advanced chatbot feature.

Why does it matter?

Apple integrating ChatGPT into its devices without offering any compensation to OpenAI creates concerns among users on whether Apple intends to share their data in return.

Source: https://www.bloomberg.com/news/articles/2024-06-12/apple-to-pay-openai-for-chatgpt-through-distribution-not-cash

🍎 Apple to ‘pay’ OpenAI through distribution, not cash

  • Apple and OpenAI announced a partnership at WWDC, but according to Bloomberg, neither side is paying the other for the integration of ChatGPT into Apple’s devices.
  • The agreement between Apple and OpenAI is non-exclusive, and Apple is also negotiating with Anthropic and Google to integrate their chatbots, with a deal for Google’s Gemini expected later this year.
  • Apple plans to generate revenue from these partnerships through revenue-sharing agreements, taking a cut from AI providers that monetize their chatbot services, such as ChatGPT Plus’ $20-per-month subscription plan.
  • Source: https://www.theverge.com/2024/6/13/24177550/apple-openai-chatgpt-deal-payment-revenue-sharing-chatbot

📹 New free rival to OpenAI Sora unveiled

  • Dream Machine is a new text-to-video tool developed by Luma AI, offering a free tier you can use immediately with a Google account, though there might be waiting times due to high demand.
  • The tool creates five-second video clips at 1360×752 resolution based on user prompts, and while it has some processing delays, the outputs are impressive and close to user expectations.
  • Dream Machine offers various subscription plans, with the free option limited to 30 generations a month, and higher tiers providing more generations at different price points, up to 2,000 generations for $499.99 monthly.
  • SOurce: https://www.techradar.com/computing/artificial-intelligence/a-new-openai-sora-rival-just-landed-for-ai-videos-and-you-can-use-it-right-now-for-free

📰 Perplexity was planning revenue-sharing deals with publishers when it came under media fire

  • Perplexity, an AI search engine startup, was working on revenue-sharing deals with high-quality publishers before facing accusations from Forbes of content misuse.
  • The company aims to announce the details of these partnerships soon and has already updated its user interface to make citation of sources more prominent.
  • Perplexity’s planned revenue-sharing framework will offer media companies a recurring income alternative, contrasting with the one-off data usage deals currently offered by entities like OpenAI.
  • Source: https://www.semafor.com/article/06/12/2024/perplexity-was-planning-revenue-sharing-deals-with-publishers

🗣️ Amazon ‘dropped the ball’ with Alexa, former employee says

  • A former Amazon employee, Mihail Eric, criticized the company’s handling of Alexa, citing technical and bureaucratic problems as major hindrances to its development.
  • Eric stated that despite having ample resources and talent, Amazon mishandled its opportunity to make Alexa the leading product in conversational AI.
  • Organizational issues, such as fragmented teams and inadequate computing resources, were mentioned by Eric as significant factors that prevented Alexa from achieving its full potential.
  • Source: https://www.businessinsider.com/amazon-alexa-voice-assistant-dropped-ball-former-employee-says-2024-6

🤔 Tim Cook says it’s a ‘significant possibility’ people use their iPhones less because of AI 

  • Apple CEO Tim Cook said Apple Intelligence may reduce iPhone usage in an interview with Marques Brownlee.
  • Cook highlighted that AI will simplify previously time-consuming tasks, allowing users to complete them more efficiently and quickly.
  • Apple has added several features to the iPhone that raise awareness of how much time consumers spend on their devices and which help make their usage more intentional.
  • Source: https://www.businessinsider.com/tim-cook-people-may-use-iphones-less-because-ai-2024-6

What Else Is Happening in AI on June 13th 2024❗

🎬YouTube trials AI comment summaries for shorts: YouTube is trying out a new AI-powered comment summarizer feature on mobile apps. The feature breaks down comments into keyword-based topics, simplifying users’ engagement with Shorts with large comment sections. Source: https://support.google.com/youtube/thread/18138167

🏆NVIDIA tops MLPerf benchmarks, dominates AI training: NVIDIA has set new performance records in MLPerf Training v4.0, showcasing its dominance in generative AI and LLM training. Key achievements include tripling GPT-3 175B training performance, fine-tuning Llama 2 70B model in just 1.5 minutes using 1,024 H100 GPUs, and accelerating Stable Diffusion v2 training performance by up to 80%. Source: https://blogs.nvidia.com/blog/mlperf-training-benchmarks

🛠️Databricks Boosts Mosaic AI for Enterprise LLMs: Databricks has announced enhancements to Mosaic AI to accelerate its efforts in enterprise gen AI development. The focus lies on the development of compound AI systems, their evaluation across metrics, and governance. Releases include Mosaic AI Model Training and Agent Framework, strengthening the company’s offering against Snowflake. Source: https://venturebeat.com/data-infrastructure/databricks-bolsters-mosaic-ai-with-tools-to-build-and-evaluate-compound-ai-systems

📝No more vagueness – Adobe to update its AI terms: Adobe has promised to update its terms of service amidst user backlash, assuring that it will not train generative AI models on customer content. The company clarified that it does not own customer content and will refrain from using it to train AI models. Source: https://blog.adobe.com/en/publish/2024/06/10/updating-adobes-terms-of-use

💻ChromeOS gets GenAI boost via Android tech: Google plans to release Android-based tech stack into ChromeOS to accelerate AI innovation. The release will simplify engineering efforts, helping phones and accessories work better with Chromebooks. Through this release, Google seeks to continue rolling out its AI features to users at a faster and larger scale. Source: https://blog.chromium.org/2024/06/building-faster-smarter-chromebook.html

A  Daily chronicle of AI Innovations June 12th 2024:

🏳️ Elon Musk drops OpenAI lawsuit with no explanation

🪦 Microsoft kills off Copilot GPT Builder after just 3 months

🖼️ Stability AI launches its ‘most sophisticated’ image generator yet

🛑 Microsoft to discontinue GPT Builder from Copilot Pro

⚖️ Musk drops lawsuit alleging OpenAI strayed from its mission

☁️OpenAI leverages Oracle Cloud to extend the Microsoft Azure Al platform

🦹 Google to launch anti-theft AI features for android phones in Brazil

📱 Google expands Gemini Nano AI capabilities to Pixel 8 & 8a

💌 Yahoo Mail integrates AI for a streamlined desktop experience

🤝Samsung unites North American AI teams, appoints ex-Apple Siri executive

🏳️ Elon Musk drops OpenAI lawsuit with no explanation

  • Elon Musk has dropped his lawsuit against OpenAI and its CEO, Sam Altman, which accused the company of breaching a contract and deviating from its mission to benefit humanity.
  • The case was dismissed without prejudice, meaning Musk can refile it in the future; this decision came just a day before a scheduled hearing on OpenAI’s request to dismiss the case.
  • Musk’s lawsuit alleged a breached “Founding Agreement” with OpenAI, but the company denied this, asserting there is no binding agreement and claiming Musk sought control by merging it with Tesla.
  • Source: https://www.theverge.com/2024/6/11/24176462/elon-musk-openai-lawsuit-sam-altman-dropped

🪦 Microsoft kills off Copilot GPT Builder after just 3 months

  • Microsoft has announced the discontinuation of GPT Builder and is giving users one month to save their data before deletion.
  • GPT Builder, part of Microsoft’s Copilot Pro, allowed users to create and customize various GPTs for $20 per month but will no longer be available after July 10.
  • Despite retiring GPT Builder, Microsoft remains focused on enterprise GPT applications and will not adjust Copilot Pro subscription fees or hand the project over to the community.
  • Source: https://www.theregister.com/2024/06/11/microsoft_retires_gpt_builder/

🖼️ Stability AI launches its ‘most sophisticated’ image generator yet

  • Stability AI launched Stable Diffusion 3 Medium, described as its “most advanced text-to-image open model yet,” featuring 2 billion parameters for photorealistic output on consumer systems.
  • The new model overcomes common artefacts in hands and faces, accurately understands complex prompts, and enhances typography through Diffusion Transformer architecture.
  • Users can access SD3 Medium via Stability’s API under an open non-commercial license, while commercial usage requires contacting the startup for licensing details.
  • Source: https://thenextweb.com/news/stability-ai-launches-stable-diffusion-3-image-generator

🤖 Autonomous driving by musculoskeletal humanoids: A study

The research paper discusses the development of a musculoskeletal humanoid robot, Musashi, designed to perform autonomous driving tasks. It mimics the human body in detail with redundant sensors and a flexible body structure that are suitable for motions with complex environmental contact.

The robot is expected to sit down on the car seat, step on the acceleration and brake pedals, and operate the steering wheel by both arms.

The paper extensively discusses the hardware and learning-based software necessary for this unique approach to autonomous driving.

Why does it matter?

The learning-based software discussed in the paper highlights the adaptability of musculoskeletal humanoids. AI models could learn from both structured data and unstructured sensory inputs, potentially improving generalization and handling novel scenarios.

Source: https://arxiv.org/pdf/2406.05573

New AI Job Opportunities on June 12th 2024

  • 💼 UiPath – Strategic Sales Lead, EMEA: https://jobs.therundown.ai/jobs/62605651-strategic-sales-lead-emea
  • 💻 Captions – Frontend Engineer: https://jobs.therundown.ai/jobs/60404535-frontend-engineer-3+-years-of-experience
  • 📊 Abridge – Senior Data Analyst: https://jobs.therundown.ai/jobs/62605637-senior-data-analyst
  • 🏙️ Metropolis – Data Analyst: https://jobs.therundown.ai/jobs/62355254-data-analyst

Source: https://jobs.therundown.ai/

What Else Is Happening in AI on June 12th 2024❗

☁️ OpenAI leverages Oracle Cloud to extend the Microsoft Azure Al platform

This collaboration will provide additional capacity for OpenAI, enabling it to continue to scale. Oracle Cloud Infrastructure’s (OCI) purpose-built AI capabilities enable startups and enterprises to build and train models faster and more reliably anywhere in Oracle’s distributed cloud.Source: https://www.oracle.com/news/announcement/openai-selects-oracle-cloud-infrastructure-to-extend-microsoft-azure-ai-platform-2024-06-11 

🦹 Google to launch anti-theft AI features for android phones in Brazil

The features include three lock modes: AI Movement Detection (recognizes common theft-related movements), Remote Lock, and Internet Access Timeout (automatically locks the screen if the device lacks internet access for an extended period). Starting in July 2024, Brazilian users with Android phones running version 10 or higher will have access to these features. Source: https://www.reuters.com/technology/artificial-intelligence/google-test-anti-theft-ai-feature-phones-brazil-2024-06-11

📱 Google expands Gemini Nano AI capabilities to Pixel 8 & 8a

This update includes features like “Summarize in Recorder” and “Smart Reply.” Users need to access developer options to enable these features. Additionally, the update enhances Find My Device functionality and allows Display Port output via USB-C. Older Pixel devices, such as the Pixel Fold and Pixel 6, also receive new features. https://www.theverge.com/2024/6/11/24176127/google-pixel-feature-drop-june-gemini-nano

💌 Yahoo Mail integrates AI for a streamlined desktop experience

The enhancements include AI-generated email summaries, a “Priority” inbox tab for urgent messages, quick action buttons, and the ability to link other email accounts within Yahoo Mail. Existing users can opt in, while new users have immediate access. Source: https://www.theverge.com/2024/6/11/24175280/yahoo-mail-generative-ai-gmail

🤝Samsung unites North American AI teams, appoints ex-Apple Siri executive

Samsung merged its AI research centers in Toronto and California into a single unit, the North America AI Center. To lead this new North American AI Center, Samsung recruited Murat Akbacak, a former Siri executive at Apple. Source: https://www.bloomberg.com/news/articles/2024-06-11/samsung-taps-apple-siri-veteran-to-lead-north-american-ai-group

A  Daily chronicle of AI Innovations June 11th 2024:

🍎 Apple brings ChatGPT to Siri

🖥️ Apple reveals macOS Sequoia with iPhone mirroring, new Passwords app, and more

⚡ Elon Musk threatens Apple device ban after OpenAI partnership

🍎 Apple introduces Apple Intelligence, its new Gen AI tech
🚫 Musk threatens to ban Apple devices over ChatGPT integrations
💼 OpenAI hires Nextdoor and Instagram veterans as CFO and CPO

🍎 Apple brings ChatGPT to Siri

  • Apple introduced AI updates for Siri at WWDC 2024, enabling it to perform actions across apps, manage notifications, automatically write and summarize text, and understand user context with “onscreen awareness.”
  • Siri’s new capabilities include referencing one app to perform actions in another, such as playing a podcast your partner sent, changing the tone of written text with a prompt, and allowing users to interact by typing as well as speaking.
  • Apple emphasizes privacy with on-device processing for many features, while more complex requests are handled by “Private Cloud Compute” on Apple Silicon-powered servers, with independent experts verifying the code running on its servers.
  • Source: https://www.theguardian.com/technology/article/2024/jun/10/apple-ai-product-launch

🖥️ Apple reveals macOS Sequoia with iPhone mirroring, new Passwords app, and more

  • Apple announced macOS Sequoia, featuring major updates such as iPhone mirroring that allows users to control their iPhone from their Mac.
  • Sequoia introduces a stand-alone Passwords app, syncable with other Apple devices and Windows PCs via iCloud, along with new Safari improvements like Highlights for extracting useful information.
  • Apple also revealed Game Porting Toolkit 2 for better Windows game compatibility, and the macOS Sequoia beta will be available this month with a full release expected in the fall.
  • Source: https://techcrunch.com/2024/06/10/apple-unveils-macos-sequoia

⚡ Elon Musk threatens Apple device ban after OpenAI partnership

  • Elon Musk has threatened to ban Apple devices at his companies because Apple plans to integrate OpenAI’s ChatGPT into its products.
  • Musk argues that allowing OpenAI access through Apple devices poses a security risk and criticizes Apple for not developing its own AI technology.
  • Musk’s discontent also stems from his ongoing disputes with OpenAI, including past legal actions and competitive tensions with his own AI ventures like xAI and its Grok chatbot.
  • Source: https://www.theregister.com/2024/06/11/musk_wants_to_ban_apple/

Apple introduces Apple Intelligence, its new Gen AI tech

Apple unveiled Apple Intelligence at WWDC 2024, a personal intelligence system deeply integrated into iOS 18, iPadOS 18, and macOS Sequoia. Leveraging Apple silicon’s power, it understands and creates language/images, takes actions across apps, and simplifies tasks using personal context. With Private Cloud Compute, Apple sets a new privacy standard in AI by flexibly processing between on-device and dedicated Apple silicon servers.

Apple Intelligence will power new features like AI-generated Bitmojis, natural language images, and video search, as well as enhanced integration with apps and Siri.

Apple also partnered with OpenAI to integrate ChatGPT with its apps, including Siri. This will allow users to directly access GPT-4 through Apple’s platforms. Additionally, Apple confirmed plans to work with other models in the future, including Google’s Gemini. The iPhone 15 Pro and devices with M1 or newer chips will be the first to receive the full suite of Apple Intelligence features.

Siri Upgrades
  • A next-gen Siri will converse more naturally, remember context across requests, and accomplish more complex tasks by better understanding both voice and text.
  • Siri also gains ‘onscreen awareness’, with the ability to take actions and utilize on-device info to better tailor requests to the individual user.
New AI Features
  • New AI writing tools built into apps like Mail, Messages, and Notes will allow users to auto-generate and edit text.
  • Mail will utilize AI to better organize and surface content in inboxes, while Notes and Phone gain new audio transcription and summarization capabilities.
  • AI-crafted ‘Genmojis’ enable personalized text-to-image emojis, and a new “Image Playground” feature introduces an image generation tool from prompts.
  • Photos get more conversational search abilities, the ability to create photo ‘stories’, and new editing tools.
Privacy
  • A focus of the AI reveal was privacy — with new features leveraging on-device processing when possible and Private Cloud Compute for more complex tasks.
  • Private Cloud Compute (PCC) is Apple’s new intelligence system specifically for private AI processing on the cloud.
  • The new AI features will be opt-in, so users will not be forced to adopt them.
OpenAI Integration
  • The OpenAI partnership will allow Siri to leverage ChatGPT/GPT-4o when needed for more complex questions.
  • OpenAI’s blog also outlined additional ChatGPT tools like image generation and document understanding embedded into the new OS.

Why does it matter?

This is generative AI as only Apple can deliver it– put at the core of its products, accessing users’ personal data for truly helpful AI assistance but keeping it secure through on-device processing and Private Cloud Compute. Apple’s long-known focus on user privacy + exceptional UX could inspire a new era of AI development.

Source: https://techcrunch.com/2024/06/11/everything-apple-announced-wwdc-2024/

Musk threatens to ban Apple devices over ChatGPT integrations 

Elon Musk threatens to ban iPhones and other Apple devices from his companies, such as Tesla, SpaceX, and xAI, over Apple’s announcement of ChatGPT integration into its OS. Musk thinks this integration will allow ChatGPT to access too much private user data through Apple devices. He claims Apple doesn’t understand OpenAI’s tech and is putting user privacy at risk.

However, Apple and OpenAI stated that users will be asked permission before any data is sent to ChatGPT.

Why does it matter?

This highlights Musk’s skepticism about the privacy implications of Apple’s AI strategy. But it could significantly impact Apple users in the future, as they may no longer be able to use iPhones and other Apple devices at Musk’s companies

Source: https://techcrunch.com/2024/06/10/elon-musk-threatens-to-ban-apple-devices-from-his-companies-over-apples-chatgpt-integrations

OpenAI hires Nextdoor and Instagram veterans as CFO and CPO

OpenAI has hired two experienced executives to fill important leadership roles. Sarah Friar, previously the CEO of the neighborhood app Nextdoor, was appointed OpenAI’s new chief financial officer (CFO). Kevin Weil, a former Instagram and Twitter executive who oversees product teams, has been named OpenAI’s chief product officer (CPO).

As CFO, Friar will manage OpenAI’s financial operations and investments as the company expands globally. As CPO, Weil will lead product development utilizing OpenAI’s AI research to create new products for consumers and enterprise customers.

Why does it matter?

This indicates OpenAI’s ambition to scale as it commercializes its AI technology. It also suggests that OpenAI is solidifying its position as a major force in the AI industry while navigating the recently surfaced challenges around AI safety.

Source: https://openai.com/index/openai-welcomes-cfo-cpo

🐶 Researchers teach AI to speak dog

A new study from the University of Michigan leveraged AI models trained on human speech to decode the meaning behind dog noises — identifying details like breed, age, gender, and emotional state with 70% accuracy.

  • Researchers gathered vocalizations from 74 dogs of varying breeds, ages, and situational contexts.
  • The noises were inputted into an AI model originally designed to analyze human voices, trained on 960 hours of speech, and fine-tuned for dogs.
  • The AI was able to predict individual dogs from barks, distinguish dog breed and gender, and match barks to emotional context like play and aggression with 70% accuracy.

AI is not only bridging the language gap for humans across the globe — but also potentially across species as well. Communicating with other intelligent animals (or at least better understanding them) seems like a skill issue that’s going to be solved sooner rather than later.

Source: https://arxiv.org/pdf/2404.18739?

New AI Job Opportunities on June 11th 2024

  • 🔧 Fiddler AI – Technical Project Manager, Customer Success: https://jobs.therundown.ai/jobs/62351203-technical-project-manager-customer-success-remote-east-coast
  • 🎓 Cohere – Research Intern (PhD): https://jobs.therundown.ai/jobs/62351142-research-intern-phd
  • 🔬 Meta – Research Engineer, Conversational AI (Reality Labs): https://jobs.therundown.ai/jobs/62419977-research-engineer-conversational-ai-reality-labs
  • 🎨 Palantir Technologies – Product Designer, US Government: https://jobs.therundown.ai/jobs/62398950-product-designer-us-government

What Else Is Happening in AI on June 11th 2024❗

⏳ AI could save financial advisors 10-15 hours per week

Morgan Stanley is using an AI tool to transcribe and categorize client conversations, saving financial advisors 10-15 hours per week. It allows advisors to better prepare for client meetings by reminding clients of past discussions and anticipating their needs. The AI tool can automatically summarize meetings and generate follow-up emails.

Source: https://www.reuters.com/technology/morgan-stanley-ceo-says-ai-could-save-financial-advisers-10-15-hours-week-2024-06-10

🤖 Anthropic reveals Claude 3’s character training process

Anthropic has developed a “character training” process for their AI assistant, Claude, to imbibe traits like curiosity, open-mindedness, and thoughtfulness. It aims to make Claude more discerning in their interactions and able to express their own views while also displaying openness to other perspectives.

Source: https://www.anthropic.com/research/claude-character

👑 FanVue introduces the world’s first AI beauty pageant contest 

Influencer marketing platform FanVue hosted the world’s first AI-powered beauty pageant, “Miss AI,”. Over 1,500 AI-generated digital models worldwide compete for the title and a $20,000 prize pool. The contest aims to showcase the capabilities of AI in creating photorealistic digital avatars that can function as influencers and brand ambassadors.

Source: https://www.npr.org/2024/06/09/nx-s1-4993998/the-miss-ai-beauty-pageant-ushers-in-a-new-type-of-influencer

🔍 Perplexica provides an open-source AI search alternative for Perplexity

Perplexica uses LLMs, machine learning algorithms, and embedding models to provide refined search results and clear answers from cited sources. Perplexica offers multiple modes, like various “Focus Modes” tailored for specific question types.

Source: https://github.com/ItzCrazyKns/Perplexica

🧒 AI tools secretly use real images of children for training 

Over 170 images and personal details of Brazilian children have been repurposed by an open-source AI training dataset called LAION-5B without the children’s consent. The images were originally posted on mommy blogs and personal social media. Researchers are concerned these images could be used to generate deepfakes or reveal sensitive information.

Source: https://www.wired.com/story/ai-tools-are-secretly-training-on-real-childrens-faces

How does Apple send your data to its cloud AI servers? Very carefully, it claims. https://www.engadget.com/how-does-apple-send-your-data-to-its-cloud-ai-servers-very-carefully-it-claims-233312425.html

Apple confirms plans to work with Google’s Gemini ‘in the future’. https://techcrunch.com/2024/06/10/apple-confirms-plans-to-work-with-googles-gemini-in-the-future/

Paris-based AI startup Mistral AI raises $640 million. https://techcrunch.com/2024/06/11/paris-based-ai-startup-mistral-ai-raises-640-million/

Tim Cook is ‘not 100 percent’ sure Apple can stop AI hallucinations: https://www.theverge.com/2024/6/11/24176035/tim-cook-apple-stop-ai-hallucinations

Nvidia shipped 3.76M data center GPUs in 2023 — dominates business with 98% revenue share. https://www.tomshardware.com/tech-industry/nvidia-shipped-376m-data-center-gpus-in-2023-dominates-business-with-98-revenue-share

A  Daily chronicle of AI Innovations June 10th 2024:

⚖️ Meta faces legal complaints over AI data plans

🚀 Alibaba’s Qwen2 AI models outperform GPT-4 & Llama-3
🧠 SAP & Nvidia are developing applications with AI & digital twins
🕵️‍♂️ Chinese tech giants exploit Nvidia AI chip loophole

🍎 Apple launches “Apple Intelligence” at WWDC 2024 for iPhone, iPad, and Mac

🚀 TCS launches TCS AI WisdomNext™, an industry-first GenAI aggregation platform

🤝 Human Native AI is building a marketplace for AI training licensing deals

🤖 Hugging Face and Pollen Robotics launched an open-source robot for household chores

Alibaba’s Qwen2 AI models outperform GPT-4 & Llama-3

Alibaba launched Qwen2 with five sizes ranging from 0.5B to 72B parameters. These models are trained in 27 additional languages besides English and Chinese, showcasing state-of-the-art performance in benchmarks. The models deliver significantly improved performance in coding and mathematics and extended context length support up to 128K tokens. Despite having fewer parameters, qwen2-72 B outperforms leading models like Llama-3-70B and its predecessor Qwen1.5-110B.

Qwen2-72B-Instruct performs comparably to GPT-4 in terms of safety and significantly outperforms Mistral-8x22B. The models are released under Apache 2.0 and Qianwen License on Hugging Face and ModelScope.

Why does it matter?

Qwen2 beats Meta’s model despite being trained on relatively fewer tokens. The researchers attribute it to more efforts put into data cleaning and training, implying innovative approaches on their end.

However, it also signals the slow shift in how LLMs are developed– from solely relying on quantity of data to prioritizing the quality of data and training techniques.

Source: https://qwenlm.github.io/blog/qwen2

SAP & Nvidia are developing applications with AI & digital twins

At SAP’s Sapphire event in Orlando, Florida, SAP and NVIDIA announced their collaboration to enhance SAP’s generative AI copilot, Joule, with two new capabilities: SAP Consulting and ABAP Developer. These new features are powered by NVIDIA AI Enterprise software.

Additionally, SAP is integrating NVIDIA Omniverse Cloud APIs into its Intelligent Product Recommendation solution to simplify the buying and selling process for complex products. This integration will allow salespeople to visualize 3D product digital twins directly within the SAP Intelligent Product Recommendation interface, making it easier to understand the products.

Why does it matter?

Using NVIDIA Omniverse Cloud APIs in SAP’s Intelligent Product Recommendation solution accelerates the quote generation process and increases sales and customer satisfaction by enabling sales representatives to provide more accurate, tailored recommendations.

Source: https://blogs.nvidia.com/blog/sap-sapphire-ai-omniverse

Chinese tech giants exploit Nvidia AI chip loophole

The U.S. government prohibits Nvidia from selling A.I. chips directly to Chinese companies due to national security concerns. Still, ByteDance is accessing Nvidia’s A.I. chips for its U.S. operations by leasing them from Oracle, as the current U.S. rules do not explicitly prohibit Chinese companies from accessing the chips if used within the U.S.

Other Chinese tech giants like Alibaba, Tencent, and China Telecom seek similar arrangements with U.S. cloud providers. The U.S. Commerce Department proposed a rule to tighten controls, but it faced opposition from cloud providers and remains in limbo.

Why does it matter?

Even if the loophole is closed, Alibaba and Tencent have discussed obtaining Nvidia chips for their U.S.-based data centers. It could further escalate the AI “arms race” and rivalry between the USA and China as both nations seek to outpace each other in developing advanced AI systems for economic and military advantages.

Source: https://www.theinformation.com/articles/chinas-nvidia-loophole-how-bytedance-got-the-best-ai-chips-despite-u-s-restrictions

What Else Is Happening in AI on June 10th 2024❗

🍎 Apple will launch “Apple Intelligence” at WWDC 2024 for iPhone, iPad, and Mac

Leaks suggest Apple will reveal “Apple Intelligence”, aka AI, at the WWDC event this week. These AI features will focus on broad appeal and privacy, with opt-in not mandatory. Apple will use its own tech and OpenAI tools to power the new AI features.

Source: https://www.theverge.com/2024/6/7/24173528/apple-intelligence-ai-features-openai-chatbot

🚀 TCS launches TCS AI WisdomNext™, an industry-first GenAI aggregation platform

The platform allows organizations to compare and experiment with GenAI models across cloud services in a single interface. It offers ready-to-deploy business solution blueprints with built-in guardrails for quick adoption.

Source: https://www.tcs.com/who-we-are/newsroom/press-release/tcs-launches-wisdomnext-an-industry-first-genai-aggregation-platform

🚨 A study by Harvard, MIT, and Wharton reveals junior staff is not reliable for AI training

Junior consultants who participated in a GPT-4 experiment struggled with AI risk mitigation, with their tactics lacking a deep understanding of the technology and focusing on changing human behavior rather than AI system design. The findings highlight the need for top-down AI governance, expert input, and upskilling across all levels of the organization.

Source: https://venturebeat.com/ai/harvard-mit-and-wharton-research-reveals-pitfalls-of-relying-on-junior-staff-for-ai-training

🤝 Human Native AI is building a marketplace for AI training licensing deals

The platform helps AI companies find data to train their models while ensuring rights holders are compensated. Rights holders upload content for free and connect with AI companies for revenue share or subscription deals. Human Native AI helps prepare and price content, monitors for copyright infringements, and takes a cut of each deal.

Source: https://techcrunch.com/2024/06/08/deal-dive-human-native-ai-is-building-the-marketplace-for-ai-training-licensing-deals

🤖 Hugging Face and Pollen Robotics launched an open-source robot for household chores

The humanoid Reachy2 was initially controlled by a human wearing a VR headset. Then, a machine learning algorithm studied the teleoperation sessions to learn how to perform the tasks independently. The dataset and trained model used for the demo are open-sourced on Hugging Face, allowing anyone to replicate the process on smaller robots at home.

Source: https://venturebeat.com/ai/hugging-face-and-pollen-robotics-show-off-first-project-an-open-source-robot-that-does-chores

👑 The world’s first AI beauty pageant

The World AI Creator Awards and creator platform FanVue are currently running the inaugural ‘Miss AI’ contest, with over 1,500 AI-generated models competing for the world’s first AI beauty pageant.

  • 10 finalists were selected from a pool of 1,500 AI-generated contestants, with the winner set to be announced at the end of June.
  • The AI models ‘hail’ from countries across the globe, showcasing various causes and personalities in addition to their photorealistic images.
  • Judges will also evaluate the AI tech behind the avatars, including prompts/image outputs and the creator’s ability to engage audiences on social media.
  • The prize pool includes $20,000 in awards and access to PR and mentorship programs.

With how good AI image generation capabilities have become, we’re likely already being exposed to an influx of generated brand ambassadors and models without even realizing it. While the idea of an AI beauty pageant may sound strange… it’s probably only going to get weirder from here.

Source: https://www.npr.org/2024/06/09/nx-s1-4993998/the-miss-ai-beauty-pageant-ushers-in-a-new-type-of-influencer

🧠 Microsoft Recall gets safety changes

Microsoft is making changes to its new Recall on-device AI feature following security concerns — with the tool now starting off by default and gaining new encryption protections ahead of the release on June 18.

  • The Copilot + Recall feature was unveiled at the Microsoft Build event in May, which will run locally and take constant screenshots to remember user actions.
  • The company faced backlash after experts warned of the tool’s ‘privacy nightmare’ with the potential to expose screenshots and activity to hackers.
  • The feature will now be off by default, with users having to opt in manually — with new encryption and authentication being added for extra safety measures.
  • The tool will roll out in ‘preview’ on the new Copilot + PCs set to launch on June 18.

While the Recall feature was one of the flashiest features revealed at Build — it also is one of the most controversial. While the screenshots enable memory and get us a step closer to the AI agent dream, its also a tough hurdle to get past for increasingly privacy-focused consumers.

Source: https://www.therundown.ai/p/microsofts-total-recall

🔊 AI TRAINING: Using ‘Background Conversations’ in ChatGPT

ChatGPT’s latest update introduced the “Background Conversations” feature, allowing you to continue your chat even when using other apps or when your screen is off.

  1. Install the latest ChatGPT app update on your phone.
  2. Open the app and go to Settings.
  3. Scroll to “Voice Mode” and toggle “Background Conversations” on.
  4. Use ChatGPT hands-free while multitasking or with your screen off.

Note: If you still don’t see the option, the feature may not have rolled out to your account yet.

🧠 AI RESEARCH: Concise prompting cuts AI costs by 20%

Researchers just found that adding a simple ‘be concise’ instruction to chain-of-thought prompts can reduce the length of AI responses by 50% with minimal impact on accuracy — leading to over 20% cost savings on API calls.

  • Chain-of-thought (CoT) prompting has the AI explain its reasoning step-by-step before giving a final answer, improving accuracy on complex problems.
  • Researchers tested adding ‘be concise’ to CoT prompts on GPT 3.5 and 4, using 1,000 multiple-choice questions across 10 topics.
  • The instruction made responses about 50% shorter on average vs. normal CoT prompting, with no significant accuracy impact.
  • The approach also reduced per-query API costs by over 20% for both models, with the more concise prompts generating fewer tokens.

Next time your AI chatbot is getting a little too wordy, this one simple prompt trick is all you need! Despite the massive complexities and uncertainties surrounding how LLMs work, sometimes these strange, seemingly basic quirks make all the difference in getting the optimal outputs.

Source: https://arxiv.org/pdf/2401.05618

New AI Job Opportunities on June 10th 2024:

  • 🔬 Meta – Research Engineer, Conversational AI (Reality Labs): https://jobs.therundown.ai/jobs/62419977-research-engineer-conversational-ai-reality-labs
  • 🎨 Palantir Technologies – Product Designer, US Government: https://jobs.therundown.ai/jobs/62398950-product-designer-us-government
  • 📊 Databricks – Strategic Sourcing Manager, Technology: https://jobs.therundown.ai/jobs/62399340-strategic-sourcing-manager-technology?
  • 📈 Anyscale – Marketing Operations Manager: https://jobs.therundown.ai/jobs/62367786-marketing-operations-manager

A  Daily chronicle of AI Innovations June 07th 2024:

🎧 OpenAI reverse engineers the workings of AI models
🤖 New Chinese video generation model beats OpenAI’s Sora
📊 Nvidia is now the second-most valuable company, overtaking Apple

🚀Adobe’s launches AEP AI Assistant to help brands master customer data and outreach

🏆The most comprehensive benchmarking & leaderboard for image models is here!

🧬AI used to predict potential new antibiotics in groundbreaking study

🔒Meta gets EU complaints from 11 countries over use of personal data to train AI

🔍The Federal Trade Commission is investigating Microsoft-Inflection AI deal

OpenAI reverse engineers the workings of AI models

In new research, OpenAI has shared improved methods for finding a large number of “features”—patterns of activity in AI models that are human interpretable. They developed new state-of-the-art methodologies that allow scaling sparse autoencoders to tens of millions of features on frontier AI models.

It demonstrated smooth and predictable scaling, with better returns to scale than prior techniques. And they could find 16 million features in GPT-4. The research also introduces several new metrics for evaluating feature quality.

OpenAI has shared the paper, code, and feature visualizations to foster further exploration.

Why does it matter?

It could bring us closer to understanding the inner workings of neural networks. This interpretability will eventually increase model trustworthiness and steerability.

While OpenAI is being criticized for taking unnecessary risks with AI technology, this shows it is also keen on tackling the risk by making models more explainable.

Source: https://openai.com/index/extracting-concepts-from-gpt-4

New Chinese video generation model beats OpenAI’s Sora

Kuaishou, a Chinese tech company, has introduced Kling, an AI model for video generation. It can make videos up to two minutes long at 1080p resolution and 30 frames per second, vs. Sora’s one-minute videos.

Kuaishou claims Kling correctly simulates the physical properties of the real world, including complex motion sequences. Using a diffusion transformer, it can also combine concepts and create fictional scenes, such as a cat driving a car through a busy city.

The model is currently available as a public demo in China.

Why does it matter?

Move over, Sora. This Chinese alternative is blowing everyone’s minds. It is becoming increasingly clear that China is catching up fast with the USA in AI.

Source: https://kling.kuaishou.com/

Nvidia is now the second-most valuable company, overtaking Apple

Nvidia rallied to record highs on Wednesday, with it’s stock market valuation hitting $3 trillion and overtaking Apple to become the world’s second most valuable company. This comes after Nvidia made a series of major announcements in the past week.

However, Nvidia’s stock has surged 147% so far in 2024, with demand for its top-of-the-line processors far outstripping supply as Big Tech races to build out their AI computing capabilities and dominate the emerging technology.

Microsoft remains the world’s most valuable company, with a market value of approximately $3.15 trillion.

Why does it matter?

Nvidia is now seen as a key company to watch to see how fast AI-powered tech is spreading across the business world, a shift that Nvidia’s founder, Jensen Huang, has declared as the dawn of the “next industrial revolution”.

Source: https://www.bbc.com/news/articles/c6ppqd3j621o

🎥 Kuaishou launches new Sora rival
Chinese tech firm Kuaishou just introduced KLING, a new text-to-video AI model capable of generating high-quality videos up to 2 minutes long with outputs that appear to rival OpenAI’s still-unreleased Sora.
The details:
KLING can produce videos at 1080p resolution with a maximum length of 2 minutes, surpassing the 1-minute Sora videos demoed by OpenAI.
KLING’s demos include realistic outputs like a man eating noodles and scenic shots, as well as surreal clips like animals in clothes.
The model uses a 3D space-time attention system to simulate complex motion and physical interactions that better mimic the real world.
The model is currently available to Chinese-based users as a public demo on the KWAI iOS app.
Why it matters: These generations are even more mind-blowing when you consider that Will Smith’s spaghetti-eating abomination was barely a year ago. With users still anxiously waiting for the public release of Sora, other competitors are stepping in — and the AI video landscape looks like it’s about to heat up in a major way.
Source:

🇨🇳 ByteDance’s AI chip loophole

TikTok parent company ByteDance is renting advanced Nvidia AI chips and using them on U.S. soil, exploiting a loophole to sidestep restrictions on China’s AI chip exports.
The details:
Due to national security concerns, the U.S. government prohibits Nvidia from selling AI chips like the A100 and H100 directly to Chinese companies.
The restrictions don’t prevent Chinese firms from renting chips for use within the U.S. — ByteDance is allegedly leasing servers with chips from Oracle.
ByteDance reportedly had access to over 1,500 H100 chips and several thousand A100s last month through the Oracle deal.
Other Chinese giants like Alibaba and Tencent are also reportedly exploring similar options, either renting from U.S. providers or setting up US data centers.
Why it matters: The AI race between the U.S. and China is only escalating — and it appears major players are going to get AI chips by any means necessary. While the U.S. tries to stall its rival’s progress with restrictions, it feels like a game of whack-a-mole that won’t stop China from reaching its AI goals.
Source:

🔍 AI RESEARCH: OpenAI probes GPT’s inner workings

OpenAI just released a new paper detailing a method for reverse engineering concepts learned by AI models and better understanding ChatGPT’s inner workings.
The details:
The paper was authored by members of the recently disbanded superalignment team, including Ilya Sutskever and Jan Leike.
‘Scaling and Evaluating Sparse Autoencoders’ outlines a technique to ID patterns representing specific concepts inside GPT-4.
By using an additional model to probe the larger model, researchers found a way to extract millions of activity patterns for further exploration.
OpenAI released open-source code and a visualization tool, allowing others to explore how different words and phrases activate concepts within models.
Why it matters: Much like Anthropic’s recent “Golden Gate Claude” and corresponding research, AI firms are still working to understand what’s truly going on underneath the hood. Cracking AI’s black box would be a big step towards better safety, tuning, and controllability of rapidly advancing models.
Source:

New AI Job Opportunities on June 07th 2024:

📊 Harvey – Data Analyst
🧠 Meta – Applied AI Research Scientist – Reinforcement Learning
📣 DeepL – Corporate Communications Lead
🤝 Palantir Technologies – Enablement Partner

What Else Is Happening in AI on June 07th 2024❗

🚀Adobe’s launches AEP AI Assistant to help brands master customer data and outreach

Adobe’s Adobe Experience Platform (AEP) is a real-time customer experience management (CXM) solution designed for enterprises. AEP AI Assistant is a gen AI chatbot hooked up to the AEP and an enterprise’s storehouse of advertising and customer data, brand assets, and content collateral (at their direction). 

Source: https://venturebeat.com/ai/adobes-new-aep-ai-assistant-is-here-to-help-brands-master-customer-data-and-outreach/

🏆The most comprehensive benchmarking & leaderboard for image models is here!

AI image models have achieved near-photographic quality. But how do they compare? Are the open-source alternatives on par with their proprietary counterparts? The Artificial Analysis Text to Image Leaderboard aims to answer these questions with human preference-based rankings.

Source: https://huggingface.co/blog/leaderboard-artificial-analysis2

🧬AI used to predict potential new antibiotics in groundbreaking study

Scientists used an algorithm to mine the entirety of the microbial diversity that we have on earth – or a huge representation of that – and find almost 1million new molecules encoded or hidden within all that microbial dark matter.

Source: https://www.theguardian.com/society/article/2024/jun/05/ai-antibiotic-resistance

🔒Meta gets EU complaints from 11 countries over use of personal data to train AI

Meta rejected the criticism and referred to a May 22 blog in which it said it uses publicly available online and licensed information to train AI as well as information that people have shared publicly on its products and services.

Source: https://www.reuters.com/technology/meta-gets-11-eu-complaints-over-use-personal-data-train-ai-models-2024-06-06

🔍The Federal Trade Commission is investigating Microsoft-Inflection AI deal

The FTC has sent subpoenas to tech giant and startup, asking whether their partnership evaded the required government antitrust review of the transaction. Microsoft picked off Inflection’s specialized workforce of AI researchers but didn’t purchase the company outright. 

Source: https://www.wsj.com/tech/ai/ftc-opens-antitrust-probe-of-microsoft-ai-deal-29b5169a

A  Daily chronicle of AI Innovations June 06th 2024:

🎧 Stability AI’s sound generator creates drum beats and instrument riffs
🤖 xAI to build the gigafactory of compute
📊 New study reveals key findings on young peoples’ use of Gen AI

💊 AI predicts nearly 1M new antibiotics

🔊 Stability AI releases open audio model

🛠️Mistral’s new SDK simplifies customization for users 

🏢Salesforce inaugurates its first AI Centre in London

🛒eBay unveils new AI-tool to enhance product image backdrops

👥Asana introduces AI teammates

🚀Writer launches a suite of No-Code development tools 

🎧 Stability AI’s sound generator creates drum beats and instrument riffs

Stability AI’s Stable Audio Open can generate up to 47-second audio samples based on text descriptions. The open AI model is trained on data from 486,000 samples of royalty-free music samples. The tool enables users to generate drum beats, instrument riffs, and ambient sounds.

However, the AI model has its limitations.

  • It is unable to produce full songs, melodies, or vocals.
  • Its terms of service prohibit users from using Stable Audio Open commercially.
  • Its training data is biased toward the English language and specific music styles.

Why does it matter?

The tool will enable creative users like sound designers, musicians, and developers to fine-tune the model to their own custom audio data, which is a major leap in the realm of generative AI’s audio generation capabilities.

Source: https://stability.ai/news/introducing-stable-audio-open

🤖 xAI to build the gigafactory of compute

The AI startup seeks to build the world’s largest supercomputer in a multi-billion dollar project. The company plans to use this supercomputer to develop AI products, including its chatbot, Grok. The facility will be powered by Nvidia’s H100 GPUs. The project aims to be operational by fall 2025.

The project will reportedly use Nvidia AI chips originally intended for Tesla, raising concerns about conflicts of interest. Moreover, Musk hasn’t yet delivered Grok 2, an advanced AI model that he had promised in May.

Why does it matter?

The supercomputer’s support for the development of Grok could become a groundbreaking advancement in the field of AI technology. Moreover, the  project represents a multi-billion dollar investment that could create numerous job opportunities and boost the economy.

Source: https://memphischamber.com/velocity-meets-potency-xai-announces-memphis-as-new-home

📊 New study reveals key findings on young peoples’ use of Gen AI

The study directly involved young readers and examined the use of generative AI by use, ethnicity, age, gender, and LGBTQ+ identity. Key findings include:

  • 50% of the survey respondents (aged 14-22) have used generative AI. However, only 4% use it daily.
  • For 53% of respondents, the use case for generative AI was obtaining information, while for 51%, it was brainstorming.
  • Black young people are more likely to use generative AI compared to their white peers. Reasons include getting information, brainstorming ideas, and assistance with schoolwork.
  • Young people of Latin origin are more likely than white people to use generative AI for multiple purposes, including image generation and getting help with their jobs.
  • Out of respondents who have never used generative AI, 34% believed it would not be helpful.
  • Among people never having used generative AI, LGBTQ+ young people are more likely to use it in comparison to cisgender and straight respondents.
  • 41% of respondents believe that generative AI will have a positive as well as negative impact on their lives in the next 10 years.

Why does it matter?

Understanding the diverse ways in which the younger users engages with generative AI enables businesses to uncover trends and data-oriented insights to improve their AI models and deliver, user experiences that are responsive and inclusive.

Source: https://www.commonsensemedia.org/sites/default/files/research/report/teen-and-young-adult-perspectives-on-generative-ai.pdf?

💊 AI predicts nearly 1M new antibiotics

Researchers just published a new study detailing the use of AI to predict close to 1M new antibiotics hidden within tiny microbes all over the world, uncovering new potential treatments against bacteria and superbugs.

  • Researchers used AI to analyze publicly available data on over 100,000 different genomes and meta-genomes.
  • The AI then predicted which parts of the microbial genomes could potentially produce antibiotic compounds, generating a list of nearly one million candidates.
  • 100 of the AI-predicted drug candidates were tested in the lab, with 79 of them being a potential antibiotic.
  • The paper’s author Cesar de la Fuente said the findings are “the largest antibiotic discovery ever”, accelerating the process from years to just hours.

As the world faces growing threats from antibiotic-resistant bacteria, AI’s ability to unlock millions of new potential treatments could be a lifeline toward staying ahead in the race to outsmart superbugs responsible for millions of deaths every year.

Source: https://www.theguardian.com/society/article/2024/jun/05/ai-antibiotic-resistance

🔊 Stability AI releases open audio model

Stability AI just unveiled Stable Audio Open, a new model enabling users to create short audio samples and sound effects from text prompts.

  • Stable Audio Open can create up to 47 seconds of high-quality audio from text prompts, focused primarily on sound effects and production elements.
  • The model was trained on data from FreeSound and Free Music Archives to avoid copyright issues.
  • Stable Audio Open can also be fine-tuned on custom audio, allowing artists to train the model on their own samples.

While the AI advances in text-to-image models have been the most visible (literally), both video and audio are about to take the same leap. Putting these tools in the hands of creatives will redefine traditional workflows — from musicians brainstorming new beats to directors crafting sound effects for film and TV.

Source: https://stability.ai/news/introducing-stable-audio-open

New AI Job Opportunities on June 06th 2024

  • 💼 Twelve Labs – Head of Finance: https://jobs.therundown.ai/jobs/61919747-head-of-finance
  • 🔒 Fiddler AI – Senior Security Engineer: https://jobs.therundown.ai/jobs/61902058-senior-security-engineer-remote-or-hybrid
  • 🎯 Siena – Director of Customer Success: https://jobs.therundown.ai/jobs/61919768-director-of-customer-success
  • 💻 Cresta – Senior Software Engineer (AI Agent): https://jobs.therundown.ai/jobs/61882269-senior-software-engineer-ai-agent

What Else Is Happening in AI on June 06th 2024❗

🛠️Mistral’s new SDK simplifies customization for users 

Mistral’s SDK, Mistral-Finetune, will allow users to modify its models on workstations, servers, and small data center nodes. While optimized for multi-GPU setups, the SDK can also scale down to a single Nvidia A100 or H100 GPU. It enables developers to fine-tune data sets such as UltraChat within half an hour, says Mistral. (https://techcrunch.com/2024/06/05/mistral-launches-new-services-sdk-to-let-customers-fine-tune-its-models)

🏢Salesforce inaugurates its first AI Centre in London

The goal of the centre is to foster collaboration, AI training, upskilling, and create AI-related jobs among developers, industry experts, and customers. This step is a part of its £4bn five-year investment in the UK, solidifying London’s position as a global AI hub. (https://www.reuters.com/technology/salesforce-open-first-ai-centre-london-2024-06-06)

🛒eBay unveils new AI-tool to enhance product image backdrops

The tool, currently available for iOS users in the UK, USA, and Germany, automatically removes image backgrounds and allows users to replace them with AI-generated backdrops. The AI feature powered by Stable Diffusion will enable sellers to upload high-quality photos easily. (https://techcrunch.com/2024/06/05/ebay-ai-powered-background-enhancement-tool)

👥Asana introduces AI teammates

The platform leverages historical data based on parameters like skill sets and missing information, allowing teams to define tasks and get an overview of who is assigned which responsibility. The tool allows users to interact with it using an AI chat interface. (https://www.theverge.com/2024/6/5/24170480/asana-ai-teammate-workflow-assistant-chatbot)

🚀Writer launches a suite of No-Code development tools 

The Gen AI platform announced major expansions that included Writer Framework – a drag-and-drop tool that enables developers to connect pre-built UI components with generative AI to create custom AI applications. The tool additionally provides full flexibility to the underlying code. (https://venturebeat.com/ai/writer-launches-no-code-platform-and-framework-for-custom-enterprise-ai-applications)

A  Daily chronicle of AI Innovations June 05th 2024:

👑 Apple could crown OpenAI as King of the Valley

🚨 AI researchers demand ‘Right to Warn’

🤖 New ‘AI teammate’ can tell people what to do at work

🚗 Elon Musk explains reasoning for Nvidia chip re-route from Tesla to X

⚛️ OpenAI seeks ‘vast quantities’ of nuclear fusion energy to power superhuman AI

🚀 Cisco has unveiled HyperFabric AI Clusters in collaboration with NVIDIA
💻 Tesla’s AI ambitions on hold? Musk diverts chips to X & xAI
🤖OpenAI insiders raise concerns over oversight and safety

🤖 ChatGPT, Claude, and Perplexity experienced outages at the same time

🧠 Raspberry Pi 5 gets AI boost with Hailo extension module

📱 TECNO CAMON 30 series launches Ella-GPT AI assistant

❄️Snowflake empowers enterprise AI with new No-Code studio

💻 Zoom’s CEO envisions AI clones in meetings

👑 Apple could crown OpenAI as King of the Valley 

  • Apple and OpenAI are reportedly partnering to integrate ChatGPT into the iPhone’s operating system, potentially making it a core feature on every device.
  • The partnership could mark the end of Google’s competition with OpenAI for being the main AI chatbot on iPhones, putting OpenAI in front of Apple’s massive user base and enhancing its long-term influence in Silicon Valley.
  • While it’s unclear exactly how Apple will use ChatGPT, possibilities include powering Siri or offering it as a standalone app, with new AI features likely to be introduced as opt-in services during the Worldwide Developers Conference next week.
  • Source: https://gizmodo.com/apple-chatgpt-deal-crown-openai-sam-altman-king-valley-1851520594

🚨 AI researchers demand ‘Right to Warn’

Current and former employees from top AI labs, including OpenAI, Anthropic, and DeepMind just published an open letter calling for companies to expand whistleblower protections so workers can raise the alarm about potential AI dangers without fear of retaliation.

  • The ‘Right to Warn AI’ petition was crafted by current and former employees of OpenAI, Anthropic, and Google DeepMind.
  • The open letter was also endorsed by AI visionaries Yoshua Bengio, Geoffrey Hinton, and Stuart Russell.
  • The statement pushes for AI firms to agree to several principles:
    • Eliminating non-disparagement clauses concerning AI risk
    • Establishing and facilitating anonymous channels for raising concerns
    • Expanding whistleblower protections and anti-retaliation measures
  • Several researchers posted threadson their experience, with Daniel Kokotajlo revealing he quit OpenAI after ‘losing hope’ the company would act responsibly.

The AI safety discourse is reaching a boiling point, and there is clearly a major industry divide that transcends any one AI firm or researcher. The proposed principles seem reasonable and necessary — but it remains to be seen if the top AI leaders will actually listen.

Source: https://righttowarn.ai/

🤖 New ‘AI teammate’ can tell people what to do at work

  • Asana has introduced an “AI teammate” that not only assigns tasks but also fills in missing information, aiming to enhance team collaboration and overall efficiency within the platform.
  • The AI model leverages historical data and team relationships to match tasks with employees’ specific skill sets, ensuring that the right people are assigned to the right projects for optimal performance, and proactively gathering any missing information needed for tasks.
  • Asana’s AI teammate includes a chat interface that allows users to ask questions about their projects, similar to other AI tools from companies like Google, Amazon, and Microsoft, which are also focused on automating workflow tasks.